Manager -Data Engineer (AWS, Python, Spark, Databricks for ETL - Agentic AI

5 days ago


Gurgaon, Haryana, India Genpact Full time
Job Description

Inviting applications for the role of Manager -Data Engineer (AWS, Python, Spark, Databricks for ETL - Agentic AI

In this role, you%27ll be part of Genpact%27s transformation under GenpactNext, as we lead the shift to Agentic AI Solutions-domain-specific, autonomous systems that redefine how we deliver value to clients. You%27ll help drive the adoption of innovations like the Genpact AP Suite in finance and accounting, with more Agentic AI products set to expand across service lines.

Responsibilities

- Design, develop, and manage scalable ETL pipelines using AWS Glue, Databricks, Apache Spark, and Python to process structured and unstructured data from diverse sources.

- Manage releases, Oversee testing, PoCs , including cost evaluations, of various AWS services and other tools, AWS, Python, Spark, Databricks for ETL-Agentic AI, Strong experience in Python and SQL , Proven expertise in AWS services such as S3, Lambda, Glue, EMR, and Redshift. Strong understanding of security principles and best practices for cloud-based environments.

- Build and orchestrate data workflows integrating with services such as AWS Lambda, Step Functions, S3, and Redshift, ensuring high availability and performance.

- Optimize Spark jobs for performance and cost-efficiency across Databricks and AWS Glue environments using partitioning, job bookmarks, and dynamic frame operations.

- Maintain secure data solutions in AWS, leveraging IAM roles, KMS encryption, and VPC-based security to meet compliance and governance standards.

- Migrate legacy ETL jobs and data from on-prem systems to cloud-native architectures on AWS Glue, Redshift, and DynamoDB.

- Implement/monitor data pipeline performance, performing debugging and tuning of Spark jobs to ensure reliable execution and minimal downtime.

- Contribute in the design and review of technical solutions, translating business requirements and user stories into scalable data engineering architectures.

- Conduct unit testing and data validation to ensure functional correctness of pipelines before deployment.

- Contribute to production deployment and collaborate with release management to ensure seamless delivery of data solutions.

- Recommend cost-effective, secure, and high-performing cloud-based data solutions, reducing manual overhead and operational burden.

Qualifications we seek in you

Minimum Qualifications

- Experience in designing, implementing data pipelines, build data applications, data migration on AWS

- Strong experience of implementing data lake using AWS services like Glue, Lambda, Step, Redshift

- Experience of Databricks will be added advantage

- Strong experience in Python and SQL

- Proven expertise in AWS services such as S3, Lambda, Glue, EMR, and Redshift.

- Advanced programming skills in Python for data processing and automation.

- Hands-on experience with Apache Spark for large-scale data processing.

- Proficiency in SQL for data querying and transformation.

- Strong understanding of security principles and best practices for cloud-based environments.

- Experience with monitoring tools and implementing proactive measures to ensure system availability and performance.

- Excellent problem-solving skills and ability to troubleshoot complex issues in a distributed, cloud-based environment.

- Strong communication and collaboration skills to work effectively with cross-functional teams.

Preferred Qualifications/ Skills

- Bachelor's degree in business information systems (IS), computer science or related field, or equivalent-related IT experience.

- AWS Data Engineering & Cloud certifications, Databricks certifications

- Familiar with multiple data integration technologies and cloud platforms

- Knowledge of Change & Incident Management process

-

-

-

-

-

-

  • Gurgaon, Haryana, India Genpact Full time

    Job DescriptionInviting applications for the role of Consultant -Data Engineer ( AWS, Python, Spark, Databricks for ETL -Agentic AIIn this role, you%27ll be part of Genpact%27s transformation under GenpactNext, as we lead the shift to Agentic AI Solutions-domain-specific, autonomous systems that redefine how we deliver value to clients. You%27ll help drive...


  • Gurgaon, Haryana, India beBeeDataEngineer Full time ₹ 15,00,000 - ₹ 25,00,000

    Job OverviewA senior data engineer is needed to lead the development of our ETL processes using AWS, Python, Spark, and Databricks. The ideal candidate will have expertise in designing and implementing scalable data pipelines and collaborating with cross-functional teams.


  • Gurgaon, Haryana, India Genpact Full time

    Job DescriptionReady to build the future with AIAt Genpact, we don&rsquot just keep up with technology&mdashwe set the pace. AI and digital innovation are redefining industries, and we&rsquore leading the charge. Genpact&rsquos AI Gigafactory, our industry-first accelerator, is an example of how we&rsquore scaling advanced technology solutions to help global...

  • Databricks Engineer

    5 days ago


    Gurgaon, Haryana, India Capgemini Full time

    Job DescriptionChoosing Capgemini means choosing a company where you will be empowered to shape your career in the way you'd like, where you'll be supported and inspired by a collaborative community of colleagues around the world, and where you'll be able to reimagine what's possible. Join us and help the world's leading organizations unlock the value of...


  • Gurgaon, Haryana, India beBeeDataEngineer Full time ₹ 1,50,00,000 - ₹ 2,50,00,000

    **Job Summary**We are seeking an experienced Data Engineer to design, develop, and manage scalable data pipelines using AWS services.Key Responsibilities:Design, develop, and maintain ETL pipelines using AWS Glue, Databricks, Apache Spark, and Python to process structured and unstructured data from diverse sources.Build and orchestrate data workflows...

  • Cloud Data Engineer

    4 days ago


    Gurgaon, Haryana, India beBeeDataEngineer Full time ₹ 1,50,00,000 - ₹ 2,50,00,000

    Cloud Data EngineerWe are seeking a highly skilled Cloud Data Engineer to design and develop scalable data pipelines using cloud-based technologies.The ideal candidate will have strong experience in AWS Glue, Databricks, Apache Spark, and Python, with a proven track record of delivering high-quality data solutions.Design and Develop Scalable ETL Pipelines:...


  • Gurgaon, Haryana, India Darwix AI Full time ₹ 15,00,000 - ₹ 20,00,000 per year

    We're Hiring:Senior Python Developer – Backend Engineering Gurgaon (On-site) | Full-Time | 2–6 Years ExperienceAbout Darwix AIDarwix AI is one of India's fastest-growing enterprise AI startups. Our GenAI-powered conversational intelligence and real-time agent assist platform helps leading enterprises across India, MENA, and Southeast Asia supercharge...


  • Gurgaon, Haryana, India beBeeCloudDataEngineer Full time ₹ 2,00,00,000 - ₹ 2,50,00,000

    About the Position:We are seeking a Cloud Data Engineer to lead our data engineering initiatives.Job Summary:The ideal candidate will have experience in designing and developing big data processing systems using Databricks and Spark.This role requires expertise in writing optimized and efficient code using PySpark, Spark SQL, and Python. The successful...

  • Data Engineer

    2 days ago


    Gurgaon, Haryana, India Bain Full time ₹ 15,00,000 - ₹ 20,00,000 per year

    Position Summary:We are seeking a highly skilled and experienced Data Engineer to join our team. The ideal candidate will have strong expertise in Python, SQL, and PySpark, with proven experience working on Databricks and cloud platforms such as Azure, AWS, or GCP.A solid understanding of ETL tools like Python as well as basic knowledge of DevOps practices...


  • Gurgaon, Haryana, India S&P Global Market Intelligence Full time

    Job DescriptionResponsibilities- Design, develop, and maintain scalable ETL/ELT pipelines.- Optimize and automate data ingestion, transformation, and storage processes.- Work with structured and unstructured data sources, ensuring data quality and consistency.- Develop and maintain data models, warehouses, and databases.- Collaborate with cross-functional...