PySpark Data Engineer

5 hours ago


india, IN EXTRAGIG Full time

Contract Assistant – Data Engineer Support (Remote, EST Hours)

Start Date: Sept 10, 2025

Duration: 6 months (extendable)

Pay: $1,000/month

Work Hours: 8:00 AM – 5:30 PM EST

We’re looking for a Contract Assistant to support a PySpark Data Engineer with daily activities. This is a remote contract role (not formal employment).


What You’ll Do:

  • Execute creative software and data solutions, including design, development, and technical troubleshooting, by thinking beyond routine approaches to build solutions or break down technical problems.
  • Develop secure, high-quality production code and data pipelines, reviewing and debugging processes implemented by others.
  • Identify opportunities to eliminate or automate remediation of recurring issues to improve operational stability of software applications and systems.
  • Lead evaluation sessions with external vendors, startups, and internal teams to drive outcomes-oriented probing of architectural designs and technical credentials.
  • Work with business stakeholders to understand requirements and design appropriate solutions, producing architecture and design artifacts for complex applications.
  • Implement robust monitoring and alerting systems to proactively identify and address data ingestion issues, optimizing performance and throughput.
  • Implement data quality checks and validation processes to ensure accuracy and reliability of data.
  • Design and implement scalable data frameworks to manage end-to-end data pipelines for workforce data analytics.
  • Share and develop best practices with Platform and Architecture teams to improve data pipeline framework and modernize the workforce data analytics platform.
  • Gather, analyze, and synthesize large, diverse data sets to continuously improve capabilities and user experiences, leveraging data-driven insights.
  • Contribute to software engineering communities of practice and events that explore new and emerging technologies, fostering a culture of diversity, opportunity, inclusion, and respect.


What You’ll Need:

  • Ability to work alongside me during US EST business hours.
  • 5+ years of applied experience in data engineering, including design, application development, testing, and operational stability.
  • Advanced proficiency in data processing frameworks and tools, including Parquet, Iceberg, PySpark, Glue, Lambda, Databricks, and AWS data services like EMR, Athena, and Redshift.
  • Proficiency in programming languages like Python, Java, or Scala for data processing and application development.
  • Proficiency in automation and continuous delivery methods, utilizing CI/CD pipelines with tools like Git/Bitbucket, Jenkins, or Spinnaker for automated deployment and version control.
  • Hands-on practical experience delivering system design, application development, testing, and operational stability, with advanced understanding of agile methodologies, application resiliency, and security.
  • Demonstrated proficiency in software applications and technical processes within technical disciplines like cloud, artificial intelligence, machine learning, and mobile.
  • In-depth knowledge of the financial services industry and their IT systems.
  • Proficiency in database management and data modeling, working with relational databases like Oracle or SQL Server. Skilled in writing SQL queries for efficient data management and retrieval, utilizing SQL tools such as DML for data handling, DDL for schema management, and PL/SQL for procedural extensions in Oracle databases.
  • Experience with scheduling tools like Autosys to automate and manage job scheduling for efficient workflow execution.

This role offers hands-on exposure to real-world BA activities, stakeholder engagement, and technology evaluation — perfect for someone looking to strengthen their BA career while working remotely.

Interested? Send me a DM or comment below

#NowHiring #RemoteJobs #TechTalent #DataEngineer #HiringNow #WorkFromHome #TechOpportunities #JobAlert #PySpark #AWSData #Databricks #Glue #Lambda #EMR #Athena #Redshift #SQL #PLSQL #Python #Scala #Java #Automation #CICD #Jenkins #Bitbucket #DataPipelines #DataQuality #DataArchitecture


  • Data Engineer

    5 hours ago


    india, IN IntraEdge Full time

    We are seeking a highly skilled Data Engineer with strong experience in Python, PySpark, Snowflake, and AWS Glue to join our growing data team. You will be responsible for building scalable and reliable data pipelines that drive business insights and operational efficiency. This role requires a deep understanding of data modeling, ETL frameworks, and...

  • Data Engineer

    5 hours ago


    india, IN NP Group Full time

    Data Engineer - Palantir Foundry, Workshop, Pyspark & TypescriptFully Remote -Long Term (initially 6 months) full time contractc$12.00 per hourWe have an immediate requirement for an experienced Data Engineer to join the global engineering team for a International Enterprise Organisation.You will bring Data Engineering expertise onto greenfield multi year...

  • Data Engineer

    5 hours ago


    india, IN Vriba Solutions Full time

    Role: Data EngineeringRemoteLooking for 5-10 years of ExpSkill Set:AWS, Snowflake, Kafka, Airflow, GitHub, PySpark, PythonKey Responsibilities:Design, develop, and maintain scalable ETL/ELT pipelinesIngest data from various sources (APIs, databases, files, etc.)Implement both real-time and batch processing solutions based on use case requirementsEnsure data...

  • Data Engineer

    5 hours ago


    Gurugram, Haryana, India, IN GMG Full time

    What we do:GMG is a global well-being company retailing, distributing and manufacturing a portfolio of leading international and home-grown brands across sport, food and health sectors. Its vision is to inspire people to win in ways that make the world better. Today, GMG’s investments span four key verticals: GMG Sports, GMG Food, GMG Health, and GMG...


  • Gurugram, Haryana, India, IN Impetus Full time

    Job Title: Lead Data Engineer – GCP (BigQuery • Composer • Python • PySpark)Location: GurgaonExperience: 8+ years (data engineering / analytics engineering), with previous lead responsibilitiesAbout the Role:You will lead the design, build and operation of large-scale data platforms on the Google Cloud Platform. You will manage a team of data...

  • Senior Data Engineer

    5 hours ago


    india, IN SAIVA AI Full time

    We are building the future of healthcare analytics. Join us to design, build, and scale robust data pipelines that power nationwide analytics and support our machine learning systems. Our goal: pipelines that are reliable, observable, and continuously improving in production.This is a fully remote role, open to candidates based in Europe or India, with...

  • Data Engineer

    5 hours ago


    india, IN TerraGiG Full time

    Responsibilities Lead the design, development, and implementation of data solutions using AWS and Snowflake. Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions. Develop and maintain data pipelines, ensuring data quality, integrity, and security. Optimize data storage and retrieval processes...

  • Data Engineer

    5 hours ago


    Gurugram, Haryana, India, IN EXL Full time

    Job Description We are looking for a Python & PySpark developer and data engineer who can design and build solutions for one of our Fortune 500 Client programs in the realm of Financial Master & Reference Data Management. This is high visibility, fast-paced key initiative will integrate data across internal and external sources, provide analytical insights,...

  • Data Engineer

    5 hours ago


    Gurugram, Haryana, India, IN IGT Solutions Full time

    Job Title: Senior Data EngineerExperience: 6 to 9 YearsLocation: Hybrid (Gurgaon)Employment Type: Full-timeJob Summary:We are seeking a highly skilled Senior Data Engineer with hands-on experience in Databricks, PySpark, ETL development, and SQL. The ideal candidate will be responsible for designing, developing, and optimizing scalable data pipelines and...


  • india, IN Blue Cloud Softech Solutions Limited Full time

    Job Summary:BCSS is seeking a Databricks Data Engineer to support its enterprise-wide Sustainability initiative. The engineer will be responsible for building data pipelines and models to support product-level carbon footprint analysis. This role involves integrating structured engineering, manufacturing, and supplier data into a unified model using...