
Manager -Data Engineer (AWS, Python, Spark, Databricks for ETL - Agentic AI
1 day ago
Job Description
Inviting applications for the role of Manager -Data Engineer (AWS, Python, Spark, Databricks for ETL - Agentic AI
In this role, you%27ll be part of Genpact%27s transformation under GenpactNext, as we lead the shift to Agentic AI Solutions-domain-specific, autonomous systems that redefine how we deliver value to clients. You%27ll help drive the adoption of innovations like the Genpact AP Suite in finance and accounting, with more Agentic AI products set to expand across service lines.
Responsibilities
- Design, develop, and manage scalable ETL pipelines using AWS Glue, Databricks, Apache Spark, and Python to process structured and unstructured data from diverse sources.
- Manage releases, Oversee testing, PoCs , including cost evaluations, of various AWS services and other tools, AWS, Python, Spark, Databricks for ETL-Agentic AI, Strong experience in Python and SQL , Proven expertise in AWS services such as S3, Lambda, Glue, EMR, and Redshift. Strong understanding of security principles and best practices for cloud-based environments.
- Build and orchestrate data workflows integrating with services such as AWS Lambda, Step Functions, S3, and Redshift, ensuring high availability and performance.
- Optimize Spark jobs for performance and cost-efficiency across Databricks and AWS Glue environments using partitioning, job bookmarks, and dynamic frame operations.
- Maintain secure data solutions in AWS, leveraging IAM roles, KMS encryption, and VPC-based security to meet compliance and governance standards.
- Migrate legacy ETL jobs and data from on-prem systems to cloud-native architectures on AWS Glue, Redshift, and DynamoDB.
- Implement/monitor data pipeline performance, performing debugging and tuning of Spark jobs to ensure reliable execution and minimal downtime.
- Contribute in the design and review of technical solutions, translating business requirements and user stories into scalable data engineering architectures.
- Conduct unit testing and data validation to ensure functional correctness of pipelines before deployment.
- Contribute to production deployment and collaborate with release management to ensure seamless delivery of data solutions.
- Recommend cost-effective, secure, and high-performing cloud-based data solutions, reducing manual overhead and operational burden.
Qualifications we seek in you
Minimum Qualifications
- Experience in designing, implementing data pipelines, build data applications, data migration on AWS
- Strong experience of implementing data lake using AWS services like Glue, Lambda, Step, Redshift
- Experience of Databricks will be added advantage
- Strong experience in Python and SQL
- Proven expertise in AWS services such as S3, Lambda, Glue, EMR, and Redshift.
- Advanced programming skills in Python for data processing and automation.
- Hands-on experience with Apache Spark for large-scale data processing.
- Proficiency in SQL for data querying and transformation.
- Strong understanding of security principles and best practices for cloud-based environments.
- Experience with monitoring tools and implementing proactive measures to ensure system availability and performance.
- Excellent problem-solving skills and ability to troubleshoot complex issues in a distributed, cloud-based environment.
- Strong communication and collaboration skills to work effectively with cross-functional teams.
Preferred Qualifications/ Skills
- Bachelor's degree in business information systems (IS), computer science or related field, or equivalent-related IT experience.
- AWS Data Engineering & Cloud certifications, Databricks certifications
- Familiar with multiple data integration technologies and cloud platforms
- Knowledge of Change & Incident Management process
-
-
-
-
-
-
-
Data Engineer
2 days ago
Gurugram, India SK HR Consultants Full timeJob Summary :- Understanding of how to design technological solutions to complex data problems, developing & testing modular, reusable, efficient and scalable code to implement those solutions- Understand, implement, and automate ETL pipelines with better industry standards- Identify, design, and implement internal process improvements : automating manual...
-
AWS Databricks Engineer
2 weeks ago
Bengaluru, Gurugram, Pune, India Infogain Full time US$ 1,25,000 - US$ 1,75,000 per yearRequired Skills & Experience6+ years of experience in Data Engineering (cloud-based preferred).Strong hands-on expertise in Databricks (PySpark, Delta Lake, SQL).Proficiency in AWS data services: S3, Glue, Lambda, Kinesis, Athena, Redshift.Experience in ETL/ELT design patterns and best practices.Strong Python and SQL skills for data manipulation and...
-
Data Engineer
2 days ago
Gurugram, India Sureminds Solutions Full timeJob Description : Data Engineer Python + AWS + AIExp Level : 5 to 8 yearsLocation : Gurgaon & RemoteInterview Process : Assessment OR 1st Technical interview , 2 - Client interview , 1 HR Round.Role Overview : Lead design and delivery of LLM-powered, agentic AI solutions with robust RAG pipelines and prompt-engineering best practices.Key Responsibilities :-...
-
Consultant, Data Engineer
1 day ago
Gurugram, Gurugram, India Genpact Full timeJob Description Inviting applications for the role of Consultant, Data Engineer (AWS, Python, UI and Web Engineer! - Agentic AI! Responsibilities - Design and develop scalable backend services and RESTful APIs using Node.js, integrating with AWS Lambda, DynamoDB, and PostgreSQL (PGDB) to process structured and semi-structured data. - JavaScript,...
-
Azure Data Engineer
2 days ago
Gurugram, India upGrad Full timeAbout the Job : We're hiring an Azure Data Engineer to design, develop, and maintain scalable data solutions on Azure. The role requires a hands-on professional with solid experience in Python, Azure data services, ETL processes, and cloud data infrastructure. You'll work closely with data scientists, analysts, and engineering teams to enable...
-
Python, Aws
2 weeks ago
Gurugram, Haryana, India Cognizant Full time**Exp: 9.5 to 13 years** **Skill: Python, AWS (Glue/Lambda)** **Location : Bangalore/Hyderabad/Kolkota/Pune/Chennai** **Technical Skills**: Apache Spark, Amazon S3,AWS Glue Catalog, AWS Glue ETL,AWS Glue Studio **Responsibilities**: - Design and develop scalable ETL processes using AWS Glue Studio to meet business requirements. - Oversee the...
-
Data Engineer
2 days ago
Gurugram, India NPG Consultants Full timeSeeking a skilled Data Engineer to drive integration and analytics for digital marketing and web platforms. This role involves working with clickstream data, building customer intelligence, and optimizing big data ecosystems for enhanced decision-making. Responsibilities : - Data Processing & Integration : Extract clickstream data from Adobe stack and merge...
-
Lead Cloud Data Engineer
2 days ago
Gurugram, India upGrad Full timeAbout the Job :We're hiring an cloud data engineering (preferably Azure) data pipelines and Spark. - Work with Databricks platform using Spark for big data processing and analytics. - Write optimized and efficient code using PySpark, Spark SQL and Python. - Develop and maintain ETL processes using Databricks notebooks and workflows. - Implement and...
-
Sr Databrick Engineer
13 hours ago
Delhi, Gurugram, NCR, Pune, India Notus Services Full time ₹ 15,00,000 - ₹ 25,00,000 per yearRole & responsibilitiesMinimum seven plus years of experience in data analytics fieldExperience with Azure/AWS DatabricksExperience in building and optimizing data pipelines, architectures and data setsExcellent experience in Scala or Python, PySpark and SQLAbility to troubleshoot and optimize complex queries on the Spark platformKnowledgeable on structured...
-
Databricks Expert
1 week ago
Noida, Gurgaon / Gurugram, India beBeeData Full time US$ 1,00,000 - US$ 1,50,000We're seeking an experienced Databricks Expert to join our team. This role requires a strong understanding of Databricks administration, coupled with expertise in AWS infrastructure and data warehousing concepts.Key ResponsibilitiesDatabricks Infrastructure Management: Manage and optimize Databricks clusters at scale, ensuring high availability and...