
Consultant -Data Engineer ( AWS, Python, Spark, Databricks for ETL -Agentic AI
3 weeks ago
Inviting applications for the role of Consultant -Data Engineer ( AWS, Python, Spark, Databricks for ETL -Agentic AI
In this role, you%27ll be part of Genpact%27s transformation under GenpactNext, as we lead the shift to Agentic AI Solutions-domain-specific, autonomous systems that redefine how we deliver value to clients. You%27ll help drive the adoption of innovations like the Genpact AP Suite in finance and accounting, with more Agentic AI products set to expand across service lines.
Responsibilities
- Design, develop, and manage scalable ETL pipelines using AWS Glue, Databricks, Apache Spark, and Python to process structured and unstructured data from diverse sources.
- AWS, Python, Spark, Databricks for ETL-Agentic AI, Strong experience in Python and SQL , Proven expertise in AWS services such as S3, Lambda, Glue, EMR, and Redshift. Strong understanding of security principles and best practices for cloud-based environments. Write clean testable code, engage in code reviews and agile ceremonies
- Build and orchestrate data workflows integrating with services such as AWS Lambda, Step Functions, S3, and Redshift, ensuring high availability and performance.
- Optimize Spark jobs for performance and cost-efficiency across Databricks and AWS Glue environments using partitioning, job bookmarks, and dynamic frame operations.
- Develop and maintain secure data solutions in AWS, leveraging IAM roles, KMS encryption, and VPC-based security to meet compliance and governance standards.
- Migrate legacy ETL jobs and data from on-prem systems to cloud-native architectures on AWS Glue, Redshift, and DynamoDB.
- Implement and monitor data pipeline performance, performing debugging and tuning of Spark jobs to ensure reliable execution and minimal downtime.
- Collaborate in the design and review of technical solutions, translating business requirements and user stories into scalable data engineering architectures.
- Perform unit testing and data validation to ensure functional correctness of pipelines before deployment.
- Lead production deployment and coordinate with release management to ensure seamless delivery of data solutions.
- Recommend cost-effective, secure, and high-performing cloud-based data solutions, reducing manual overhead and operational burden.
- Contribute to backup, disaster recovery, and business continuity strategies for critical data assets.
- Participate in code reviews, technical design discussions, and DevOps integration for CI/CD of data pipelines using tools like Git, CodePipeline, or Databricks Repos.
Qualifications we seek in you
Minimum Qualifications
- Experience in designing, implementing data pipelines, build data applications, data migration on AWS
- Strong experience of implementing data lake using AWS services like Glue, Lambda, Step, Redshift
- Experience of Databricks will be added advantage
- Strong experience in Python and SQL
- Proven expertise in AWS services such as S3, Lambda, Glue, EMR, and Redshift.
- Advanced programming skills in Python for data processing and automation.
- Hands-on experience with Apache Spark for large-scale data processing.
- Proficiency in SQL for data querying and transformation.
- Strong understanding of security principles and best practices for cloud-based environments.
- Experience with monitoring tools and implementing proactive measures to ensure system availability and performance.
- Excellent problem-solving skills and ability to troubleshoot complex issues in a distributed, cloud-based environment.
- Strong communication and collaboration skills to work effectively with cross-functional teams.
Preferred Qualifications/ Skills
- Master's Degree-Computer Science, Electronics, Electrical or equivalent
- AWS Data Engineering & Cloud certifications, Databricks certifications
- Experience with multiple data integration technologies and cloud platforms
- Knowledge of Change & Incident Management process
-
-
-
-
-
-
-
Gurgaon, Haryana, India Genpact Full timeJob DescriptionInviting applications for the role of Manager -Data Engineer (AWS, Python, Spark, Databricks for ETL - Agentic AIIn this role, you%27ll be part of Genpact%27s transformation under GenpactNext, as we lead the shift to Agentic AI Solutions-domain-specific, autonomous systems that redefine how we deliver value to clients. You%27ll help drive the...
-
Lead AI Data Engineer
5 days ago
Gurgaon, Haryana, India S&P Global Market Intelligence Full time ₹ 12,00,000 - ₹ 36,00,000 per yearSummary:As a Data Engineer & AI Developer, you will play a key role in transforming complex data into clear, actionable insights. You will be responsible for building robust data pipelines, developing interactive dashboards, and integrating cutting-edge AI techniques to enhance our data processes and insights. This is a unique opportunity to blend expertise...
-
Gurgaon, Haryana, India Genpact Full time ₹ 10,00,000 - ₹ 25,00,000 per yearPrincipal Consultant- Databricks Developer with experience in Unity Catalog Python , Spark , Kafka for ETL-COR Ready to shape the future of work? At Genpact, we don't just adapt to change—we drive it. AI and digital innovation are redefining industries, and we're leading the charge. Genpact's AI Gigafactory, our industry-first accelerator, is an...
-
AWS Data Engineer
3 days ago
Gurgaon, Haryana, India Deqode Full time ₹ 20,00,000 - ₹ 25,00,000 per yearProfile :AWS Data EngineerMandate Skills :AWS + Databricks + Pyspark + SQL roleLocation: Bangalore /Pune /Hyderabad /Chennai /GurgaonNotice Period :ImmediateKey ResponsibilitiesDesign, build, and maintain scalable data pipelines to collect, process, and store data from multiple datasetsOptimize data storage solutions for performance, scalability, and...
-
Assistant Manager-Data Engineer
3 weeks ago
Gurgaon, Haryana, India Genpact Full timeJob DescriptionReady to build the future with AIAt Genpact, we don&rsquot just keep up with technology&mdashwe set the pace. AI and digital innovation are redefining industries, and we&rsquore leading the charge. Genpact&rsquos AI Gigafactory, our industry-first accelerator, is an example of how we&rsquore scaling advanced technology solutions to help global...
-
Databricks Engineer
3 weeks ago
Gurgaon, Haryana, India Capgemini Full timeJob DescriptionChoosing Capgemini means choosing a company where you will be empowered to shape your career in the way you'd like, where you'll be supported and inspired by a collaborative community of colleagues around the world, and where you'll be able to reimagine what's possible. Join us and help the world's leading organizations unlock the value of...
-
Senior Data Engineer
3 days ago
Gurgaon, Haryana, India Weekday AI Full time ₹ 12,00,000 - ₹ 36,00,000 per yearThis role is for one of the Weekday's clientsMin Experience: 5 yearsLocation: GurgaonJobType: full-timeWe are seeking an experienced Senior Data Engineer with strong expertise in building and managing large-scale data pipelines within the AWS ecosystem. The ideal candidate will have a solid background in SQL, cloud-native data platforms, and orchestration...
-
Senior Data Engineer
3 days ago
Gurgaon, Haryana, India Weekday AI Full time ₹ 20,00,000 - ₹ 25,00,000 per yearThis role is for one of the Weekday's clientsMin Experience: 5 yearsLocation: GurgaonJobType: full-timeWe are seeking an experienced Senior Data Engineer with strong expertise in building and managing large-scale data pipelines within the AWS ecosystem. The ideal candidate will have a solid background in SQL, cloud-native data platforms, and orchestration...
-
Gurgaon, Haryana, India Sirius AI Full time ₹ 12,00,000 - ₹ 36,00,000 per yearKey Responsibilities:Design, develop, and manage scalable, secure, and efficient data pipelines on Azure to process and transform large datasets from multiple sources.Implement and optimize Azure data services such as Azure Synapse Analytics, Azure Data Lake Storage (ADLS), Azure SQL Database, Azure Data Factory (ADF), Azure Databricks, and Azure...
-
Senior Data Engineer(Python, AWS)
3 weeks ago
Gurgaon, Haryana, India S&P Global Market Intelligence Full timeJob DescriptionResponsibilities- Design, develop, and maintain scalable ETL/ELT pipelines.- Optimize and automate data ingestion, transformation, and storage processes.- Work with structured and unstructured data sources, ensuring data quality and consistency.- Develop and maintain data models, warehouses, and databases.- Collaborate with cross-functional...