Job Title: Senior ETL Data Engineer (Ab Initio / Informatica)
Experience: 4+ Years
Location: Lahore / Karachi / Islamabad (Hybrid)
Job Type: Full-Time
About the Role:
We are looking for a highly skilled Senior ETL Data Engineer (Ab Initio / Informatica) with strong experience in building robust data pipelines, working with large-scale datasets, and leveraging modern Big Data and cloud technologies. The ideal candidate should have hands-on expertise in ETL frameworks, distributed data processing, and data modeling within cloud environments such as AWS. If you have a passion for working with data and enjoy designing scalable systems, we’d like to meet you.
Key Responsibilities:
- Design and develop complex ETL pipelines and data solutions using Big Data and cloud-native technologies.
- Leverage tools such as Ab Initio, Informatica, DBT, and Apache Spark to build scalable data workflows.
- Implement distributed data processing using Hadoop, Hive, Kafka, and Spark.
- Build and optimize data pipelines in AWS using services like EMR, Glue, Lambda, Athena, and S3.
- Work with various structured and unstructured data sources to perform efficient data ingestion and transformation.
- Write optimized SQL queries and manage stored procedures for complex data processing tasks.
- Orchestrate workflows using Airflow, AWS Step Functions, or similar schedulers.
- Collaborate with cross-functional teams to understand data needs and deliver high-quality datasets for analytics and reporting.
- Deploy data models into production environments and ensure robust monitoring and resource management.
- Mentor junior engineers and contribute to the team’s knowledge sharing and continuous improvement efforts.
- Identify and recommend process and technology improvements to enhance data pipeline performance and reliability.
Required Skills & Qualifications:
- Bachelor’s degree in Computer Science, Engineering, or a related field.
- 4+ years of hands-on experience in ETL development, data engineering, and data pipeline orchestration.
- Strong working knowledge of Ab Initio, Informatica, or similar ETL tools.
- Expertise in Python, PySpark, or Scala for data processing.
- Proven experience in Big Data technologies (Hadoop, Hive, Spark, Kafka).
- Proficient with AWS services related to data engineering (EMR, Glue, Lambda, Athena, S3).
- In-depth understanding of data modeling, ETL cycle, data warehousing, and data management principles.
- Hands-on experience with relational (PostgreSQL, MySQL) and columnar databases (Redshift, HBase, Snowflake).
- Familiarity with containerization (Docker), CI/CD pipelines (Jenkins), and Agile tools (Jira).
- Ability to troubleshoot complex data issues and propose scalable solutions.
- Excellent communication and collaboration skills.
Nice to Have:
- Experience with open table formats such as Apache Iceberg.
- Working knowledge of Snowflake and its data warehousing capabilities.
- Familiarity with GDE, Collect > IT, or other components of Ab Initio.
Why Join Us?
- Hybrid work model with flexibility to work from home and office.
- Exposure to cutting-edge technologies and high-impact projects.
- Collaborative team environment with opportunities for growth and innovation.
- Culture that values ownership, continuous learning, and mutual respect.
Read Full Description