Consultant -Data Engineer ( AWS, Python, Spark, Databricks for ETL -Agentic AI

3 weeks ago


Gurgaon, Haryana, India Genpact Full time
Job Description

Inviting applications for the role of Consultant -Data Engineer ( AWS, Python, Spark, Databricks for ETL -Agentic AI

In this role, you%27ll be part of Genpact%27s transformation under GenpactNext, as we lead the shift to Agentic AI Solutions-domain-specific, autonomous systems that redefine how we deliver value to clients. You%27ll help drive the adoption of innovations like the Genpact AP Suite in finance and accounting, with more Agentic AI products set to expand across service lines.

Responsibilities

- Design, develop, and manage scalable ETL pipelines using AWS Glue, Databricks, Apache Spark, and Python to process structured and unstructured data from diverse sources.

- AWS, Python, Spark, Databricks for ETL-Agentic AI, Strong experience in Python and SQL , Proven expertise in AWS services such as S3, Lambda, Glue, EMR, and Redshift. Strong understanding of security principles and best practices for cloud-based environments. Write clean testable code, engage in code reviews and agile ceremonies

- Build and orchestrate data workflows integrating with services such as AWS Lambda, Step Functions, S3, and Redshift, ensuring high availability and performance.

- Optimize Spark jobs for performance and cost-efficiency across Databricks and AWS Glue environments using partitioning, job bookmarks, and dynamic frame operations.

- Develop and maintain secure data solutions in AWS, leveraging IAM roles, KMS encryption, and VPC-based security to meet compliance and governance standards.

- Migrate legacy ETL jobs and data from on-prem systems to cloud-native architectures on AWS Glue, Redshift, and DynamoDB.

- Implement and monitor data pipeline performance, performing debugging and tuning of Spark jobs to ensure reliable execution and minimal downtime.

- Collaborate in the design and review of technical solutions, translating business requirements and user stories into scalable data engineering architectures.

- Perform unit testing and data validation to ensure functional correctness of pipelines before deployment.

- Lead production deployment and coordinate with release management to ensure seamless delivery of data solutions.

- Recommend cost-effective, secure, and high-performing cloud-based data solutions, reducing manual overhead and operational burden.

- Contribute to backup, disaster recovery, and business continuity strategies for critical data assets.

- Participate in code reviews, technical design discussions, and DevOps integration for CI/CD of data pipelines using tools like Git, CodePipeline, or Databricks Repos.

Qualifications we seek in you

Minimum Qualifications

- Experience in designing, implementing data pipelines, build data applications, data migration on AWS

- Strong experience of implementing data lake using AWS services like Glue, Lambda, Step, Redshift

- Experience of Databricks will be added advantage

- Strong experience in Python and SQL

- Proven expertise in AWS services such as S3, Lambda, Glue, EMR, and Redshift.

- Advanced programming skills in Python for data processing and automation.

- Hands-on experience with Apache Spark for large-scale data processing.

- Proficiency in SQL for data querying and transformation.

- Strong understanding of security principles and best practices for cloud-based environments.

- Experience with monitoring tools and implementing proactive measures to ensure system availability and performance.

- Excellent problem-solving skills and ability to troubleshoot complex issues in a distributed, cloud-based environment.

- Strong communication and collaboration skills to work effectively with cross-functional teams.

Preferred Qualifications/ Skills

- Master's Degree-Computer Science, Electronics, Electrical or equivalent

- AWS Data Engineering & Cloud certifications, Databricks certifications

- Experience with multiple data integration technologies and cloud platforms

- Knowledge of Change & Incident Management process

-

-

-

-

-

-

  • Gurgaon, Haryana, India Genpact Full time

    Job DescriptionInviting applications for the role of Manager -Data Engineer (AWS, Python, Spark, Databricks for ETL - Agentic AIIn this role, you%27ll be part of Genpact%27s transformation under GenpactNext, as we lead the shift to Agentic AI Solutions-domain-specific, autonomous systems that redefine how we deliver value to clients. You%27ll help drive the...


  • Gurgaon, Haryana, India S&P Global Market Intelligence Full time ₹ 12,00,000 - ₹ 36,00,000 per year

    Summary:As a Data Engineer & AI Developer, you will play a key role in transforming complex data into clear, actionable insights. You will be responsible for building robust data pipelines, developing interactive dashboards, and integrating cutting-edge AI techniques to enhance our data processes and insights. This is a unique opportunity to blend expertise...


  • Gurgaon, Haryana, India Genpact Full time ₹ 10,00,000 - ₹ 25,00,000 per year

    Principal Consultant- Databricks Developer with experience in Unity Catalog Python , Spark , Kafka for ETL-COR Ready to shape the future of work? At Genpact, we don't just adapt to change—we drive it. AI and digital innovation are redefining industries, and we're leading the charge. Genpact's AI Gigafactory, our industry-first accelerator, is an...

  • AWS Data Engineer

    3 days ago


    Gurgaon, Haryana, India Deqode Full time ₹ 20,00,000 - ₹ 25,00,000 per year

    Profile :AWS Data EngineerMandate Skills :AWS + Databricks + Pyspark + SQL roleLocation: Bangalore /Pune /Hyderabad /Chennai /GurgaonNotice Period :ImmediateKey ResponsibilitiesDesign, build, and maintain scalable data pipelines to collect, process, and store data from multiple datasetsOptimize data storage solutions for performance, scalability, and...


  • Gurgaon, Haryana, India Genpact Full time

    Job DescriptionReady to build the future with AIAt Genpact, we don&rsquot just keep up with technology&mdashwe set the pace. AI and digital innovation are redefining industries, and we&rsquore leading the charge. Genpact&rsquos AI Gigafactory, our industry-first accelerator, is an example of how we&rsquore scaling advanced technology solutions to help global...

  • Databricks Engineer

    3 weeks ago


    Gurgaon, Haryana, India Capgemini Full time

    Job DescriptionChoosing Capgemini means choosing a company where you will be empowered to shape your career in the way you'd like, where you'll be supported and inspired by a collaborative community of colleagues around the world, and where you'll be able to reimagine what's possible. Join us and help the world's leading organizations unlock the value of...


  • Gurgaon, Haryana, India Weekday AI Full time ₹ 12,00,000 - ₹ 36,00,000 per year

    This role is for one of the Weekday's clientsMin Experience: 5 yearsLocation: GurgaonJobType: full-timeWe are seeking an experienced Senior Data Engineer with strong expertise in building and managing large-scale data pipelines within the AWS ecosystem. The ideal candidate will have a solid background in SQL, cloud-native data platforms, and orchestration...


  • Gurgaon, Haryana, India Weekday AI Full time ₹ 20,00,000 - ₹ 25,00,000 per year

    This role is for one of the Weekday's clientsMin Experience: 5 yearsLocation: GurgaonJobType: full-timeWe are seeking an experienced Senior Data Engineer with strong expertise in building and managing large-scale data pipelines within the AWS ecosystem. The ideal candidate will have a solid background in SQL, cloud-native data platforms, and orchestration...


  • Gurgaon, Haryana, India Sirius AI Full time ₹ 12,00,000 - ₹ 36,00,000 per year

    Key Responsibilities:Design, develop, and manage scalable, secure, and efficient data pipelines on Azure to process and transform large datasets from multiple sources.Implement and optimize Azure data services such as Azure Synapse Analytics, Azure Data Lake Storage (ADLS), Azure SQL Database, Azure Data Factory (ADF), Azure Databricks, and Azure...


  • Gurgaon, Haryana, India S&P Global Market Intelligence Full time

    Job DescriptionResponsibilities- Design, develop, and maintain scalable ETL/ELT pipelines.- Optimize and automate data ingestion, transformation, and storage processes.- Work with structured and unstructured data sources, ensuring data quality and consistency.- Develop and maintain data models, warehouses, and databases.- Collaborate with cross-functional...