
Pyspark developer
4 weeks ago
Job Description
ROLE SUMMARY
We are seeking a highly skilledPySpark Developerwith hands-on experience inDatabricksto join Sompo's IT Systems Development unit in an offshore capacity. This role focuses on designing, building, and optimizing large-scale data pipelines and processing solutions on theDatabricks Unified Analytics Platform. The ideal candidate will have expertise inbig data frameworks,distributed computing, and cloud platforms, with a deep understanding ofDatabricks architecture. This is an excellent opportunity to work with cutting-edge technologies in a dynamic, fast-paced environment.
ROLE RESPONSIBILITIES
Data Engineering and Processing:
Develop and manage data pipelines usingPySparkonDatabricks.
Implement ETL/ELT processes to process structured and unstructured data at scale.
Optimize data pipelines for performance, scalability, and cost-efficiency in Databricks.
Databricks Platform Expertise:
Experience in Perform Design, Development & Deployment using Azure Services (Data Factory, Databricks, PySpark, SQL)
Develop and maintain scalable data pipelines and build new Data Source integrations to support increasing data volume and complexity.
Leverage theDatabricks Lakehouse architecturefor advanced analytics and machine learning workflows.
ManageDelta Lakefor ACID transactions and data versioning.
Developnotebooksand workflows for end-to-end data solutions.
Cloud Platforms and Deployment:
Deploy and manage Databricks on Azure(e.g., Azure Databricks).
Use DatabricksJobs,Clusters, andWorkflowsto orchestrate data pipelines.
Optimize resource utilization and troubleshoot performance issues on the Databricks platform.
CI/CD and Testing:
Build and maintainCI/CD pipelinesfor Databricks workflows using tools likeAzure DevOps,GitHub Actions, orJenkins.
Write unit and integration tests for PySpark code using frameworks likePytestorunittest.
Collaboration and Documentation:
Work closely with data scientists, data analysts, and IT teams to deliver robust data solutions.
Document Databricks workflows, configurations, and best practices for internal use.
TECHNICAL QUALIFICATIONS
Experience:
4+ years of experience in data engineering or distributed systems development.
Strong programming skills inPythonandPySpark.
Hands-on experience withDatabricksand its ecosystem, includingDelta LakeandDatabricks SQL.
Knowledge of big data frameworks likeHadoop,Spark, andKafka.
Databricks Expertise:
Proficiency in setting up and managingDatabricks Workspaces,Clusters, andJobs.
Familiarity with DatabricksMLflowfor machine learning workflows is a plus.
Cloud Platforms:
Expertise in deploying Databricks solutions Azure(e.g., Data Lake, Synapse).
Knowledge ofKubernetesfor managing containerized workloads is advantageous.
Database Knowledge:
Experience with both SQL (e.g.,PostgreSQL,SQL Server) and NoSQL databases (e.g.,MongoDB,Cosmos DB).
GENERAL QUALIFICATIONS
Strong analytical and problem-solving skills.
Ability to manage multiple tasks in a high-intensity, deadline-driven environment.
Excellent communication and organizational skills.
Experience in regulated industries like insurance is a plus.
EDUCATION REQUIREMENTS
ABachelor's Degreein Computer Science, Data Engineering, or a related field is preferred.
Relevant certifications inDatabricks,PySpark, or cloud platforms are highly desirable.
-
Pyspark Developer
2 weeks ago
Bengaluru, Chennai, Pune, India VHR Solutions Full time ₹ 5,00,000 - ₹ 12,00,000 per yearJob Description:We are looking for a skilled PySpark Developer to join our data engineering team. The ideal candidate will have strong experience in building scalable data pipelines using Apache Spark (PySpark), integrating with big data platforms, and working with large datasets in distributed environments. You will work closely with data engineers,...
-
Pyspark Developer
5 days ago
Chennai, Hyderabad, Pune, India Tata Consultancy Services Full time ₹ 15,00,000 - ₹ 25,00,000 per yearTCS is Hiring Location :Bangalore, Chennai, Kolkata, Hyderabad, PuneExp : 4-8 YrsFunctional Skills: Experience in Credit Risk/Regulatory risk domainTechnical Skills: Spark ,PySpark, Python, Hive, Scala, MapReduce, Unix shell scriptingGood to Have Skills: Exposure to Machine Learning TechniquesJob Description:4+ Years of experience with Developing/Fine tuning...
-
Pyspark Developer
4 weeks ago
Chennai, Tamil Nadu, India MP DOMINIC AND CO Full timeJob Summary - Design develop and implement scalable data pipelines and streaming use cases using PySpark and Spark on a distributed computing platform - Possess strong programming skills in Spark streaming - Have familiarity with cloud platforms like GCP - Gain experience in big data technologies such as Hadoop Hive and HDFS - Perform ETL operations...
-
PySpark Developer
4 weeks ago
Hyderabad, Bengaluru, Chennai, India Coders Brain Technology Private Limited Full timeJob Description ROLE RESPONSIBILITIES Data Engineering and Processing: Develop and manage data pipelines using PySpark on Databricks. Implement ETL/ELT processes to process structured and unstructured data at scale. Optimize data pipelines for performance, scalability, and cost-efficiency in Databricks. Databricks Platform Expertise: Experience in...
-
Pyspark Developer
6 days ago
Bengaluru, Chennai, Kolkata, India Tata Consultancy Services Full time ₹ 15,00,000 - ₹ 25,00,000 per yearJob Description:5+ Years of experience with Developing/Fine tuning and implementing programs/applicationsUsing Python/PySpark/Scala on Big Data/Hadoop Platform.Roles and Responsibilities:a) Work with a Leading Banks Risk Management team on specific projects/requirements pertaining to risk Models inconsumer and wholesale bankingb) Enhance Machine Learning...
-
Pyspark Developer
6 days ago
Bengaluru, India Tata Consultancy Services Full timeGreetings from TCS!! TCS is hiring for Pyspark Developer Required Skill Set: Pyspark, Python, SQL and relational databases, SparkSQL, Spark Scripting, UNIX Shell Scripting, ETL, Data Warehousing, CI/CD Desired Experience Range: 4 to 10 Years Job Location: Hyderabad, Bangalore, Chennai, Kolkata, Pune Must Have: Data Engineer, Python developer with specialty...
-
Pyspark Developer
6 days ago
Bengaluru, India Tata Consultancy Services Full timeGreetings from TCS!! TCS is hiring for Pyspark Developer Required Skill Set: Pyspark, Python, SQL and relational databases, SparkSQL, Spark Scripting, UNIX Shell Scripting, ETL, Data Warehousing, CI/CD Desired Experience Range: 4 to 10 Years Job Location: Hyderabad, Bangalore, Chennai, Kolkata, Pune Must Have: Data Engineer, Python developer with specialty...
-
Pyspark Developer
2 weeks ago
Bengaluru, India Tata Consultancy Services Full timeGreetings from TCS!!TCS is hiring for Pyspark DeveloperRequired Skill Set: Pyspark, Python, SQL and relational databases, SparkSQL, Spark Scripting, UNIX Shell Scripting, ETL, Data Warehousing, CI/CDDesired Experience Range: 4 to 10 YearsJob Location: Hyderabad, Bangalore, Chennai, Kolkata, PuneMust Have:Data Engineer, Python developer with specialty in...
-
Pyspark Developer
1 week ago
Bengaluru, India Tata Consultancy Services Full timeGreetings from TCS!!TCS is hiring for Pyspark DeveloperRequired Skill Set: Pyspark, Python, SQL and relational databases, SparkSQL, Spark Scripting, UNIX Shell Scripting, ETL, Data Warehousing, CI/CDDesired Experience Range: 4 to 10 YearsJob Location: Hyderabad, Bangalore, Chennai, Kolkata, PuneMust Have:Data Engineer, Python developer with specialty in...
-
pyspark developer
20 hours ago
Pune, Maharashtra, India Capgemini Engineering Full time ₹ 9,00,000 - ₹ 12,00,000 per yearAt Capgemini Engineering, the world leader in engineering services, we bring together a global team of engineers, scientists, and architects to help the world's most innovative companies unleash their potential. From autonomous cars to life-saving robots, our digital and software technology experts think outside the box as they provide unique R&D and...