Senior ETL Data Engineer (Ab Initio / Informatica)

NorthBay Solutions

Education
Benefits
Qualifications
Special Commitments
Skills

Job Title: Senior ETL Data Engineer (Ab Initio / Informatica)

Experience: 4+ Years

Location: Lahore / Karachi / Islamabad (Hybrid)

Job Type: Full-Time

About the Role:

We are looking for a highly skilled Senior ETL Data Engineer (Ab Initio / Informatica) with strong experience in building robust data pipelines, working with large-scale datasets, and leveraging modern Big Data and cloud technologies. The ideal candidate should have hands-on expertise in ETL frameworks, distributed data processing, and data modeling within cloud environments such as AWS. If you have a passion for working with data and enjoy designing scalable systems, we’d like to meet you.

Key Responsibilities:

  • Design and develop complex ETL pipelines and data solutions using Big Data and cloud-native technologies.
  • Leverage tools such as Ab Initio, Informatica, DBT, and Apache Spark to build scalable data workflows.
  • Implement distributed data processing using Hadoop, Hive, Kafka, and Spark.
  • Build and optimize data pipelines in AWS using services like EMR, Glue, Lambda, Athena, and S3.
  • Work with various structured and unstructured data sources to perform efficient data ingestion and transformation.
  • Write optimized SQL queries and manage stored procedures for complex data processing tasks.
  • Orchestrate workflows using Airflow, AWS Step Functions, or similar schedulers.
  • Collaborate with cross-functional teams to understand data needs and deliver high-quality datasets for analytics and reporting.
  • Deploy data models into production environments and ensure robust monitoring and resource management.
  • Mentor junior engineers and contribute to the team’s knowledge sharing and continuous improvement efforts.
  • Identify and recommend process and technology improvements to enhance data pipeline performance and reliability.

Required Skills & Qualifications:

  • Bachelor’s degree in Computer Science, Engineering, or a related field.
  • 4+ years of hands-on experience in ETL development, data engineering, and data pipeline orchestration.
  • Strong working knowledge of Ab Initio, Informatica, or similar ETL tools.
  • Expertise in Python, PySpark, or Scala for data processing.
  • Proven experience in Big Data technologies (Hadoop, Hive, Spark, Kafka).
  • Proficient with AWS services related to data engineering (EMR, Glue, Lambda, Athena, S3).
  • In-depth understanding of data modeling, ETL cycle, data warehousing, and data management principles.
  • Hands-on experience with relational (PostgreSQL, MySQL) and columnar databases (Redshift, HBase, Snowflake).
  • Familiarity with containerization (Docker), CI/CD pipelines (Jenkins), and Agile tools (Jira).
  • Ability to troubleshoot complex data issues and propose scalable solutions.
  • Excellent communication and collaboration skills.

Nice to Have:

  • Experience with open table formats such as Apache Iceberg.
  • Working knowledge of Snowflake and its data warehousing capabilities.
  • Familiarity with GDE, Collect > IT, or other components of Ab Initio.

Why Join Us?

  • Hybrid work model with flexibility to work from home and office.
  • Exposure to cutting-edge technologies and high-impact projects.
  • Collaborative team environment with opportunities for growth and innovation.
  • Culture that values ownership, continuous learning, and mutual respect.
Read Full Description
Confirmed 11 hours ago. Posted 3 days ago.

Discover Similar Jobs

Suggested Articles