Sr. Azure Data Engineer

Fractal

Education
Benefits
Skills

It's fun to work in a company where people truly BELIEVE in what they are doing!

We're committed to bringing passion and customer focus to the business.

JOB DESCRIPTION:

If you are an extraordinary developer and who loves to push the boundaries to solve complex business problems using creative solutions, then we wish to talk with you. As a Data Engineer - Azure, you will work in the Technology team that helps deliver our Data Engineering offerings at large scale to our Fortune clients worldwide. The role is responsible for innovating, building, and maintaining technology services.

RESPONSIBILITIES:

  • Be an integral part of large-scale client business development and delivery engagements.
  • Develop the software and systems needed for end-to-end execution on large projects.
  • Work across all phases of SDLC and use Software Engineering principles to build scaled solutions.
  • Build the knowledge base required to deliver increasingly complex technology projects.
  • Developing Data Pipelines (Batch/Streaming)
  • Developing Complex data transformations
  • ETL Orchestration
  • Data Migration
  • Develop and Maintain Datawarehouse / Data Lakes

EXPERIENCE:

  • 6 to 9 years of experience in setting up data pipelines to injest streaming and batch data from many sources
  • Experience on optimized data collection procedures
  • Good experience in performing extract, transform and load (ETL) processes to make the data more usable with cloud infrastructures
  • Expertise in data storage structures, data mining, and data cleansing
  • Hands-on experience using Big Data components (Hadoop, Spark) & programming languages.
  • Experience in using, or migrating to cloud databases
  • Proficient in relational database design and development
  • Proficient in Data warehousing concepts and up to date Data Integration patterns/technologies
  • Experience in Azure data lake, Azure SQL Database, Data Bricks, Azure SQL Data warehouse and Azure Data factory.
  • Experience in Developing Spark applications using Spark - SQL in Databricks for data extraction, transformation, and aggregation from multiple file formats for analyzing & transforming the data to uncover insights.
  • Systematic problem-solving approach with strong communication skills and a sense of ownership and drive

TECHNICAL SKILLS:

  • Strong Experience in Azure Cloud
  • Experience with distributed data/computing tools: Map/Reduce, Hadoop, Hive, Spark
  • Strong knowledge in Azure Analytics Products like Data Factory, HD Insights, Azure Data Lake Storage, Azure SQL Datawarehouse & Data Bricks
  • Strong in Python Programming Skills
  • Strong experience in networking and security measures
  • Good knowledge in using NoSQL databases like MongoDB, PostgreSQL and Cassandra running huge volumes of data
  • Proficient in with SQl Databases like MSSql, Postgres.
  • Expertise in Data Warehousing – Hadoop, MapReduce, HIVE, PIG, Apache Spark, Kafka
  • Good understanding of Spark Architecture including Spark Core, Spark SQL, Data Frames, Spark Streaming, Driver Node, Worker Node, Stages, Executors and Tasks.
  • Good understanding of Big Data Hadoop and Yarn architecture along with various Hadoop Demons such as Job Tracker, Task Tracker, Name Node, Data Node, Resource/Cluster Manager, and Kafka (distributed stream-processing).

JOB RESPONSIBILITES:

  • Building and maintaining the data architecture for data science projects.
  • Extract Transform and Load data from sources Systems to Azure Data Storage services using a combination of Azure Data Factory, T-SQL, Spark SQL and U-SQL Azure Data Lake Analytics.
  • Data Ingestion to one or more Azure Services - (Azure Data Lake, Azure Storage, Azure SQL, Azure DW) and processing the data in In Azure Data bricks.
  • Define table schema and quickly adapt with the pipeline.
  • Improve data foundational procedures, integrating new data management technologies and softwares into the existing system, building data collection pipelines
  • Identify valuable data sources and automate collection processes.
  • Implement concepts of programming such as data structures, error handing, data manipulation and I/O processing. Implements database concepts and practices including definition and query language.
  • Implement adequate networking and security measures for the data pipeline.
  • Implement monitoring solution for the data pipeline.
  • Develop Spark applications using Pyspark and Spark-SQL for data extraction, transformation and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns.
  • Support the design, and implement data engineering solutions
  • Maintain excellent documentation for understanding and accessing data storage
  • Work independently as well as in teams to deliver transformative solutions to clients
  • Be proactive and constantly pay attention to the scalability, performance and availability of our systems
  • Establishes privacy/security hierarchy and regulates access
  • Undertake preprocessing of structured and unstructured data
  • Support in collecting, organizing, and interpreting data along with fellow colleagues.
  • Collaborate with engineering and product development teams

EDUCATION:

  • B.E/B.Tech/M.Tech in Computer Science or related technical degree OR Equivalent

If you like wild growth and working with happy, enthusiastic over-achievers, you'll enjoy your career with us!

Not the right fit? Let us know you're interested in a future opportunity by clicking Introduce Yourself in the top-right corner of the page or create an account to set up email alerts as new job postings become available that meet your interest!

Read Full Description
Confirmed 11 hours ago. Posted 30+ days ago.

Discover Similar Jobs

Suggested Articles