Engineer (A2 DES, Databricks, PySpark, Python)

5 hours ago


Bengaluru, Karnataka, India BSR & Co Full time ₹ 10,00,000 - ₹ 18,00,000 per year
Description

Roles & responsibilities

Role Overview: The Associate 2 - "Data Engineer with Databricks/Python skills" will be part of the GDC Technology Solutions (GTS) team, working in a technical role in the Audit Data & Analytics domain that requires developing expertise in KPMG proprietary D&A (Data and analytics)) tools and audit methodology. He/she will be a part of the team responsible for extracting and processing datasets from client ERP systems (SAP/Oracle/Microsoft Dynamics) or other sources to provide insights through data warehousing, ETL and dashboarding solutions to Audit/internal teams and be involved in developing solutions using a variety of tools & technologies

The Associate 2 - "Data Engineer" will be predominantly responsible for:

Data Engineering


·Understand requirements, validate assumptions, and develop solutions using Azure Databricks, Azure Data Factory or Python. Able to handle any data mapping changes and customizations within Databricks using PySpark

·Build Azure Databricks notebooks to perform data transformations, create tables, and ensure data quality and consistency. Leverage Unity Catalog for data governance and maintaining a unified data view across the organization

·Analyze enormous volumes of data using Azure Databricks and Apache Spark. Create pipelines and workflows to support data analytics, machine learning, and other data-driven applications

·Able to integrate Azure Databricks with ERP systems or third part systems using APIs and build Python or PySpark notebooks to apply business transformation logic as per the common data model

·Debug, optimize and performance tune and resolve issues, if any, with limited guidance, when processing large data sets and propose possible solutions

·Must have experience in concepts like Partitioning, optimization, and performance tuning for improving the performance of the process

·Implement best practices of Azure Databricks design, development, Testing and documentation

·Work with Audit engagement teams to interpret the results and provide meaningful audit insights from the reports

·Participate in team meetings, brainstorming sessions, and project planning activities

·Stay up-to-date with the latest advancements in Azure Databricks, Cloud and AI development, to drive innovation and maintain a competitive edge

·Enthusiastic to learn and use Azure AI services in business processes.

·Work experience on using Microsoft Fabric is an added advantage

·Write production ready code

·Design, develop, and maintain scalable and efficient data pipelines to process large datasets from various sources using Azure Data Factory (ADF).

·Integrate data from multiple data sources and ensure data consistency, quality, and accuracy, leveraging Azure Data Lake Storage (ADLS).

·Design and implement ETL (Extract, Transform, Load) processes to ensure seamless data flow across systems using Azure

·Work experience on Microsoft Fabric is an added advantage

·Enthusiastic to learn, adapt and integrate Gen AI into the business process and should have experience working with Azure AI services

·Optimize data storage and retrieval processes to enhance system performance and reduce latency.

Technical Skills

Primary Skills:


Ø2-4 years of experience in data engineering, with a strong focus on Databricks, PySpark, Python and Spark SQL.

ØProven experience in implementing ETL processes and data pipelines

ØHands-on experience with Azure Databricks, Azure Data Factory (ADF), Azure Data Lake Storage (ADLS)

ØAbility to write reusable, testable, and efficient code

ØDevelop low-latency, high-availability, and high-performance applications

ØUnderstanding of fundamental design principles behind a scalable application

ØGood knowledge of Azure cloud services

ØFamiliarity with Generative AI and its applications in data engineering

ØKnowledge of Microsoft Fabric and Azure AI services is an added advantage

 Enabling Skills


·Excellent analytical, and problem-solving skills

·Quick learning ability and adaptability

·Effective communication skills

·Attention to detail and good team player

·Willingness and ability to deliver within tight timelines

·Flexible to work timings and willingness to work on different projects/technologies

Responsibilities

Roles & responsibilities

Role Overview: The Associate 2 - "Data Engineer with Databricks/Python skills" will be part of the GDC Technology Solutions (GTS) team, working in a technical role in the Audit Data & Analytics domain that requires developing expertise in KPMG proprietary D&A (Data and analytics)) tools and audit methodology. He/she will be a part of the team responsible for extracting and processing datasets from client ERP systems (SAP/Oracle/Microsoft Dynamics) or other sources to provide insights through data warehousing, ETL and dashboarding solutions to Audit/internal teams and be involved in developing solutions using a variety of tools & technologies

The Associate 2 - "Data Engineer" will be predominantly responsible for:

Data Engineering


·Understand requirements, validate assumptions, and develop solutions using Azure Databricks, Azure Data Factory or Python. Able to handle any data mapping changes and customizations within Databricks using PySpark

·Build Azure Databricks notebooks to perform data transformations, create tables, and ensure data quality and consistency. Leverage Unity Catalog for data governance and maintaining a unified data view across the organization

·Analyze enormous volumes of data using Azure Databricks and Apache Spark. Create pipelines and workflows to support data analytics, machine learning, and other data-driven applications

·Able to integrate Azure Databricks with ERP systems or third part systems using APIs and build Python or PySpark notebooks to apply business transformation logic as per the common data model

·Debug, optimize and performance tune and resolve issues, if any, with limited guidance, when processing large data sets and propose possible solutions

·Must have experience in concepts like Partitioning, optimization, and performance tuning for improving the performance of the process

·Implement best practices of Azure Databricks design, development, Testing and documentation

·Work with Audit engagement teams to interpret the results and provide meaningful audit insights from the reports

·Participate in team meetings, brainstorming sessions, and project planning activities

·Stay up-to-date with the latest advancements in Azure Databricks, Cloud and AI development, to drive innovation and maintain a competitive edge

·Enthusiastic to learn and use Azure AI services in business processes.

·Work experience on using Microsoft Fabric is an added advantage

·Write production ready code

·Design, develop, and maintain scalable and efficient data pipelines to process large datasets from various sources using Azure Data Factory (ADF).

·Integrate data from multiple data sources and ensure data consistency, quality, and accuracy, leveraging Azure Data Lake Storage (ADLS).

·Design and implement ETL (Extract, Transform, Load) processes to ensure seamless data flow across systems using Azure

·Work experience on Microsoft Fabric is an added advantage

·Enthusiastic to learn, adapt and integrate Gen AI into the business process and should have experience working with Azure AI services

·Optimize data storage and retrieval processes to enhance system performance and reduce latency.

Technical Skills

Primary Skills:


Ø2-4 years of experience in data engineering, with a strong focus on Databricks, PySpark, Python and Spark SQL.

ØProven experience in implementing ETL processes and data pipelines

ØHands-on experience with Azure Databricks, Azure Data Factory (ADF), Azure Data Lake Storage (ADLS)

ØAbility to write reusable, testable, and efficient code

ØDevelop low-latency, high-availability, and high-performance applications

ØUnderstanding of fundamental design principles behind a scalable application

ØGood knowledge of Azure cloud services

ØFamiliarity with Generative AI and its applications in data engineering

ØKnowledge of Microsoft Fabric and Azure AI services is an added advantage

 Enabling Skills


·Excellent analytical, and problem-solving skills

·Quick learning ability and adaptability

·Effective communication skills

·Attention to detail and good team player

·Willingness and ability to deliver within tight timelines

·Flexible to work timings and willingness to work on different projects/technologies

Qualifications

Education Requirements


·B. Tech/B.E/MCA (Computer Science / Information Technology)

Primary Skills:


Ø2-4 years of experience in data engineering, with a strong focus on Databricks, PySpark, Python and Spark SQL.

ØProven experience in implementing ETL processes and data pipelines

ØHands-on experience with Azure Databricks, Azure Data Factory (ADF), Azure Data Lake Storage (ADLS)

ØAbility to write reusable, testable, and efficient code

ØDevelop low-latency, high-availability, and high-performance applications

ØUnderstanding of fundamental design principles behind a scalable application

ØGood knowledge of Azure cloud services

ØFamiliarity with Generative AI and its applications in data engineering

ØKnowledge of Microsoft Fabric and Azure AI services is an added advantage


  • Bengaluru, Karnataka, India PradeepIT Consulting Services Full time ₹ 12,00,000 - ₹ 36,00,000 per year

    Job description:Role: Databricks + Pyspark + SQLYear of Experience:4-8 YearsResponsibilities:Collaborate with cross-functional teams to understand data requirements and design efficient data processing solutions.Develop and maintain ETL processes using Databricks and PySpark for large-scale data processing.Optimize and tune existing data pipelines to ensure...

  • Azure Data Engineer

    16 hours ago


    Bengaluru, Karnataka, India AKIRA Insights Full time ₹ 6,00,000 - ₹ 18,00,000 per year

    Job Description :Data EngineerLocation :BengaluruShift :IST Shift (9 :30 AM to 6 :30 PM)Experience :3-4 YearsAbout The RoleWe are seeking a highly skilled Data Engineer with ~3 years of experience to join our dynamic team. The ideal candidate will have a strong background in data engineering technologies, especially SQL, Databricks, PySpark, Azure Data...


  • Bengaluru, Karnataka, India AKIRA Insights Full time ₹ 12,00,000 - ₹ 36,00,000 per year

    DescriptionJob Description :Data EngineerLocation :BengaluruShift :IST Shift (9 :30 AM to 6 :30 PM)Experience :5-6 YearsAbout The RoleWe are seeking a highly skilled Data Engineer with ~5 years of experience to join our dynamic team.The ideal candidate will have a strong background in data engineering technologies, especially SQL, Databricks, PySpark, Azure...

  • Databricks Engineer

    2 weeks ago


    Bengaluru, Karnataka, India Tata Consultancy Services Full time ₹ 15,00,000 - ₹ 25,00,000 per year

    Greetings from TCSWe are hiringDatabricks EngineerforPAN IndiaLocation.Skills: Azure Databricks, Pyspark, ADLS, ADFLocation: PAN INDIAExp: 4+ yrs.JOB DESCRIPTION:MUST HAVE:Azure Data Bricks (Python)Pyspark.ADLSComplex problem solving skill.GOOD TO HAVE:Azure Devops.Azure Data Factory.Azure Synpasis AnalyticsResponsibility of / Expectations from the...

  • Databricks Engineer

    2 weeks ago


    Bengaluru, Karnataka, India 6d9439f4-30e6-4c3b-a55f-c97083047f58 Full time ₹ 12,00,000 - ₹ 36,00,000 per year

    Qualifications & ExperienceBachelors degree in Computer Science, Information Systems, Statistics, or related field.510 years of experience in data quality, data engineering, analytics operations.Strong command of SQL, Python, and PySpark for data validation and troubleshooting.Hands-on exposure to Databricks for building and running data pipelines.Working...


  • Bengaluru, Karnataka, India Changeleaders Consulting Full time ₹ 9,00,000 - ₹ 12,00,000 per year

    Looking for candidate having experience in AWS, Databricks ,Python, Pyspark.

  • Python Pyspark

    7 hours ago


    Bengaluru, Karnataka, India Tata Consultancy Services (TCS) Full time ₹ 5,00,000 - ₹ 15,00,000 per year

    Greetings from Tata Consultancy Services (TCS)Location : Chennai, Bangalore, Pune, DelhiExperience : 5-10 YearsRole: Python Pyspark DeveloperRequired Skill set Python, Pyspark, Scala, SparkMust HaveStrong analytical skills in conducting sophisticated statistical analysis using bureau/vendor data, customer performances.Working experience in languages PySpark...


  • Bengaluru, Karnataka, India 6d9439f4-30e6-4c3b-a55f-c97083047f58 Full time ₹ 10,000 - ₹ 60,000 per year

    Data Engineer Data QualityBuild automated data quality checks using Databricks (Delta Lake) and PySpark.Perform profiling, anomaly detection, and reconciliation across systems.Investigate data issues using SQL, Python, drive root cause fixes.Collaborate on governance, metadata, and data catalog integration.Automate validations via Airflow, ADF, or Databricks...

  • DataBricks Developer

    9 hours ago


    Bengaluru, Karnataka, India Growel Softech Pvt Ltd Full time ₹ 20,00,000 - ₹ 25,00,000 per year

    DataBricks Developer-Key skills-Databricks, Pyspark, Python, SQLLocation : BangaloreExp : 5 to 8 yearsNP : Immediate


  • Bengaluru, Karnataka, India Growel Softech Pvt. Ltd. Full time ₹ 15,00,000 - ₹ 25,00,000 per year

    1.Total 3 to 6 years of experience as a Pyspark and Databricks Developer with a strong portfolio demonstrating your expertise 2. Good hands on knowledge with database technologies such as SQL and NoSQL 3. Familiarity with Snowflake is added advantage 4. Good in Troubleshooting debugging maintaining and improving existing software 5. Excellent problem-solving...