Infogain - AWS Data Engineer (Senior)

apartmentInfogain placeGurgaon calendar_month 

Job Description

Roles & Responsibilities

Job Requirements:

We are seeking a highly skilled and motivated Data Engineer to join our dynamic team. The ideal candidate will have extensive experience in ETL, Data Modelling, and Data Architecture. Proficiency in ETL optimization, designing, coding, and tuning big data processes using Python is essential.

Experience in using Scala is an add-on.

Additionally, the candidate should have extensive experience in building data platforms using a variety of technologies, including Python, PostgreSQL, Spark, Parquet/ORC, Data Modelling (Relational Dimensional E-R Modelling), ETL, Splunk, DataDog, Airflow, Git, CI/CD Jenkins, JIRA, Confluence, IntelliJ Idea, Agile - Scrum/Kanban, Code Review, RCP Framework, Query book, Build, Deployment CI/CD & Release Process, Backstage, PagerDuty, and Spinnaker.

Key Responsibilities

Hands-on experience on developing Data platform and its components Data Lake, cloud Datawarehouse, APIs, Batch and streaming data pipeline Experience with building data pipelines and applications to stream and process large datasets at low latency.

  • Develop and maintain batch and stream processing data solutions using Apache Spark and Apache Hive leveraging the RCP Framework to create robust, modular applications.
  • Work on orchestration using Airflow to automate and manage data workflows.
  • Utilize project management tools like JIRA and Confluence to track progress and collaborate with the team.
  • Develop data processing workflows utilizing Spark, SQL/PLSQL, and Python to transform and cleanse raw data into a usable format implementing data storage solutions leveraging Parquet/ORC formats.
  • Develop and manage scalable data pipelines and applications using containerization with Docker and orchestration with Kubernetes.
  • Optimize data storage and retrieval performance through efficient data modelling techniques, including Relational, Dimensional, and E-R modelling.
  • Maintain data integrity and quality by implementing robust validation and error handling mechanisms within ETL processes.
  • Automate deployment processes using CI/CD tools like Jenkins and Spinnaker to ensure reliable and consistent releases.
  • Monitor and troubleshoot data pipelines using monitoring tools like DataDog and Splunk to identify performance bottlenecks and ensure system reliability.
  • Participate in Agile development methodologies such as Scrum/Kanban, including sprint planning, daily stand-ups, and retrospective meetings.
  • Conduct code reviews to ensure adherence to coding standards, best practices, and scalability considerations.
  • Manage and maintain documentation using tools like Confluence to ensure clear and up-to-date documentation of data pipelines, schemas, and processes.
  • Provide on-call support for production data pipelines, responding to incidents and resolving issues in a timely manner.
  • Collaborate with cross-functional teams including developers, data scientists, and operations teams to address complex data engineering challenges.
  • Stay updated on emerging technologies and industry trends to continuously improve data engineering processes and tools.
  • Contribute to the development of reusable components and frameworks to streamline data engineering tasks across projects.
  • Utilize version control systems like Git to manage codebase and collaborate effectively with team members.
  • Leverage IDEs like IntelliJ IDEA for efficient development and debugging of data engineering code.
  • Adhere to security best practices in handling sensitive data and implementing access controls within the data lake environment.
Skills Required
  • Programming Languages: Python, Bash/Unix/Linux
  • Big Data Technologies: Apache Spark, Apache Hive
  • Cloud Services: EC2, ECS, S3, SNS, CloudWatch
  • Databases: Postgres
  • Application development: RCP Framework
  • Containerization and Orchestration: Docker, Kubernetes
  • CI/CD Tools: Github, Spinnaker, Jenkins
  • Additional Skills: Scala, Maven
Experience
  • 6-8 Years
Skills
  • Primary Skill: Data Engineering
  • Sub Skill(s): Data Engineering
  • Additional Skill(s): Python, postgreSQL Development, ETL, Apache Hive, postgreSQL DBA, AWS-Apps, SQL, Datadog, Pyspark

About The Company

Infogain is a human-centered digital platform and software engineering company based out of Silicon Valley. We engineer business outcomes for Fortune 500 companies and digital natives in the technology, healthcare, insurance, travel, telecom, and retail & CPG industries using technologies such as cloud, microservices, automation, IoT, and artificial intelligence.

We accelerate experience-led transformation in the delivery of digital platforms. Infogain is also a Microsoft (NASDAQ: MSFT) Gold Partner and Azure Expert Managed Services Provider (MSP).

Infogain, an Apax Funds portfolio company, has offices in California, Washington, Texas, the UK, the UAE, and Singapore, with delivery centers in Seattle, Houston, Austin, Krakw, Noida, Gurgaon, Mumbai, Pune, and Bengaluru.

local_fire_departmentUrgent

Data Engineer(Knime)

apartmentSteerLean ConsultingplaceGurgaon
Job Description We're Hiring a Data Engineer(Knime)! Are you an expert in Knime with a knack for seamless workflow conversion We're on the lookout for a talented Data Engineer to join our team and make an impact! Experience Required :5 years...
apartmentRandstadplaceGurgaon
One of our esteem client is hiring for Data Eng. Location Gurgaon Position - Data Engineer QualificationsRequired Data Engineer Job Responsibilities:Develops and maintains scalable data pipelines and builds out new API integrations to support...
check_circleNew offer

Junior Data Engineer

apartmentXpertHuntplaceNoida, 33 km from Gurgaon
Job Description Skills: SQL, SSIS, SQL Server Integration Services (SSIS), Microsoft SQL Server, Data Engineering, SQL query, Good Day Dear Aspirants, We are seeking for a talented and motivated Junior Data Engineer for our client. please find...