Pyspark Data Engineer
2 weeks ago
About Client :- Our client is a French multinational information technology (IT) services and consulting company, headquartered in Paris, France. Founded in 1967, It has been a leader in business transformation for over 50 years, leveraging technology to address a wide range of business needs, from strategy and design to managing operations. The company is committed to unleashing human energy through technology for an inclusive and sustainable future, helping organisations accelerate their transition to a digital and sustainable world. They provide a variety of services, including consulting, technology, professional, and outsourcing services. Job Details:- location: Pune , Chennai Mode Of Work : Hybrid Notice Period : Immediate Joiners Experience : 6-8 yrs Type Of Hire : Contract to Hire JOB DESCRIPTION: Must have skills : Pyspark and Python Good to have: Familiarity with orchestration tools like Apache Airflow Grade(s) : C1/C2 Location(s) : Pune/Chennai Detailed JD : "Design, develop, and maintain scalable and efficient data processing pipelines using PySpark and Python. Build and implement ETL (Extract, Transform, Load) processes to ingest data from various sources and load it into target destinations. Optimize PySpark applications for performance and troubleshoot existing code. Ensure data integrity and quality throughout the data lifecycle. Collaborate with cross-functional teams, including data engineers and data scientists, to understand and fulfill data needs. Provide technical leadership, conduct code reviews, and mentor junior team members. Translate business requirements into technical solutions and contribute to architectural discussions. Stay current with the latest industry trends in big data and distributed computing. Mandatory skills and experience PySpark and Python: Advanced proficiency and extensive experience in building data processing applications. Distributed Computing: In-depth understanding of principles, including performance tuning. Big Data Ecosystem: Experience with technologies such as Hadoop, Hive, Sqoop, and Spark. Cloud Services: Hands-on experience with cloud platforms like AWS (e.g., Glue, Lambda, Kinesis) is often required. Database and SQL: Strong knowledge of SQL and experience with relational databases and data warehousing. Software Development: Experience with software development best practices, including version control (Git), unit testing, and code reviews. Desired or "nice-to-have" skills Familiarity with orchestration tools like Apache Airflow. Experience with other data processing tools like Kafka or Pandas/Numpy. Knowledge of API development and creating RESTful services. Experience with data file formats like Parquet, ORC, and Avro. Experience with Agile development methodologies.
-
Data Engineer
3 weeks ago
Pune, India RSquareSoft Technologies Full timeRole: Data Engineer Location: Pune (On-site)Job Type: Full-TimeExperience: 4–7 YearsKey ResponsibilitiesDesign, develop, and maintain scalable data pipelines and ETL workflows.Work with large datasets using PySpark , Python , and SQL to ensure efficient data transformation and integration.Implement data solutions on AWS, leveraging services like S3, Glue,...
-
Data Engineer
3 weeks ago
Pune, India RSquareSoft Technologies Full timeRole: Data Engineer Location: Pune (On-site) Job Type: Full-Time Experience: 4–7 Years Key Responsibilities Design, develop, and maintain scalable data pipelines and ETL workflows. Work with large datasets using PySpark , Python , and SQL to ensure efficient data transformation and integration. Implement data solutions on AWS, leveraging services like S3,...
-
Data Engineer
3 weeks ago
Pune, India RSquareSoft Technologies Full timeRole: Data Engineer Location: Pune (On-site) Job Type: Full-Time Experience: 4–7 Years Key Responsibilities - Design, develop, and maintain scalable data pipelines and ETL workflows. - Work with large datasets using PySpark, Python, and SQL to ensure efficient data transformation and integration. - Implement data solutions on AWS, leveraging services like...
-
Data Engineer
3 weeks ago
Pune, India RSquareSoft Technologies Full timeRole: Data Engineer Location: Pune (On-site) Job Type: Full-Time Experience: 4–7 Years Key Responsibilities Design, develop, and maintain scalable data pipelines and ETL workflows. Work with large datasets using PySpark , Python , and SQL to ensure efficient data transformation and integration. Implement data solutions on AWS, leveraging services like S3,...
-
Data Engineer
3 weeks ago
Pune, India RSquareSoft Technologies Full timeRole: Data EngineerLocation: Pune (On-site)Job Type: Full-TimeExperience: 4–7 YearsKey ResponsibilitiesDesign, develop, and maintain scalable data pipelines and ETL workflows.Work with large datasets using PySpark, Python, and SQL to ensure efficient data transformation and integration.Implement data solutions on AWS, leveraging services like S3, Glue,...
-
Data Engineer
3 weeks ago
Pune, India RSquareSoft Technologies Full timeRole: Data EngineerLocation: Pune (On-site)Job Type: Full-TimeExperience: 4–7 YearsKey ResponsibilitiesDesign, develop, and maintain scalable data pipelines and ETL workflows.Work with large datasets using PySpark, Python, and SQL to ensure efficient data transformation and integration.Implement data solutions on AWS, leveraging services like S3, Glue,...
-
Data Engineer
3 weeks ago
Pune, India RSquareSoft Technologies Full timeRole: Data EngineerLocation: Pune (On-site)Job Type: Full-TimeExperience: 4–7 YearsKey ResponsibilitiesDesign, develop, and maintain scalable data pipelines and ETL workflows.Work with large datasets using PySpark, Python, and SQL to ensure efficient data transformation and integration.Implement data solutions on AWS, leveraging services like S3, Glue,...
-
Data Engineer
3 weeks ago
Pune, India RSquareSoft Technologies Full timeRole: Data EngineerLocation: Pune (On-site)Job Type: Full-TimeExperience: 4–7 YearsKey ResponsibilitiesDesign, develop, and maintain scalable data pipelines and ETL workflows.Work with large datasets using PySpark , Python , and SQL to ensure efficient data transformation and integration.Implement data solutions on AWS, leveraging services like S3, Glue,...
-
Data Engineer
3 weeks ago
Pune, India RSquareSoft Technologies Full timeRole: Data EngineerLocation: Pune (On-site)Job Type: Full-TimeExperience: 4–7 YearsKey Responsibilities- Design, develop, and maintain scalable data pipelines and ETL workflows.- Work with large datasets using PySpark, Python, and SQL to ensure efficient data transformation and integration.- Implement data solutions on AWS, leveraging services like S3,...
-
Pyspark Data Engineer
2 weeks ago
Pune, India People Prime Worldwide Full timeAbout Client :- Our client is a French multinational information technology (IT) services and consulting company, headquartered in Paris, France. Founded in 1967, It has been a leader in business transformation for over 50 years, leveraging technology to address a wide range of business needs, from strategy and design to managing operations. The company is...