Data Engineer
6 days ago
Req number: R6413
Employment type: Full time
Worksite flexibility: Remote Who we are
CAI is a global technology services firm with over 8,500 associates worldwide and a yearly revenue of $1 billion+. We have over 40 years of excellence in uniting talent and technology to power the possible for our clients, colleagues, and communities. As a privately held company, we have the freedom and focus to do what is right—whatever it takes. Our tailor-made solutions create lasting results across the public and commercial sectors, and we are trailblazers in bringing neurodiversity to the enterprise.
Job Summary We are seeking a motivated Data Engineer who has strong experience in building cloud-based data lake and analytics architectures using AWS and Databricks, and is proficient in Python programming for data processing and automation. This is a Full-time and Remote position.
Job Description
What You'll Do
Design, develop, and maintain data lakes and data pipelines on AWS using ETL frameworks and Databricks.
Integrate and transform large-scale data from multiple heterogeneous sources into a centralized data lake environment.
Implement and manage Delta Lake architecture using Databricks Delta or Apache Hudi.
Develop end-to-end data workflows using PySpark, Databricks Notebooks, and Python scripts for ingestion, transformation, and enrichment.
Design and develop data warehouses and data marts for analytical workloads using Snowflake, Redshift, or similar systems.
Design and evaluate data models (Star, Snowflake, Flattened) for analytical and transactional systems.
Optimize data storage, query performance, and cost across the AWS and Databricks ecosystem.
Build and maintain CI/CD pipelines for Databricks notebooks, jobs, and Python-based data processing scripts.
Collaborate with data scientists, analysts, and stakeholders to deliver high-performance, reusable data assets.
Maintain and manage code repositories (Git) and promote best practices in version control, testing, and deployment.
Participate in making major technical and architectural decisions for data engineering initiatives.
Monitor and troubleshoot Databricks clusters, Spark jobs, and ETL processes for performance and reliability.
Coordinate with business and technical teams through all phases of the software development life cycle.
What You'll Need
5+ years of experience building and managing Data Lake Architecture on AWS Cloud
3+ years of experience with AWS Data services such as S3, Glue, Lake Formation, EMR, Kinesis, RDS, DMS, and Redshift.
3+ years of experience building Data Warehouses on Snowflake, Redshift, HANA, Teradata, or Exasol.
3+ years of hands-on experience working with Apache Spark or PySpark, on Databricks.
3+ years of experience implementing Delta Lakes using Databricks Delta or Apache Hudi.
3+ years of experience in ETL development using Databricks, AWS Glue, or other modern frameworks.
Proficiency in Python for data engineering, automation, and API integrations.
Experience in Databricks Jobs, Workflows, and Cluster Management.
Experience with CI/CD pipelines and Infrastructure as Code (IaC) tools like Terraform or CloudFormation is a plus.
Bachelor's degree in computer science, Information Technology, Data Science, or related field
Physical Demands
This role involves mostly sedentary work, with occasional movement around the office to attend meetings, etc.
Ability to perform repetitive tasks on a computer, using a mouse, keyboard, and monitor.
Reasonable accommodation statement
If you require a reasonable accommodation in completing this application, interviewing, completing any pre-employment testing, or otherwise participating in the employment selection process, please direct your inquiries to or – 8111.
-
Operation Analyst
1 week ago
Remote, India Data Engineer Academy LLP Full time ₹ 6,00,000 - ₹ 12,00,000 per yearJob Title: Operations AnalystLocation: Remote | Experience: 0-1 YearsType: Full-TimeNotice-Immediate JoinerAbout Us: Data Engineering Academy is an edtech startup led by FAANG data engineers. We equip aspiring data professionals with the tools, training, and support to succeed in the tech industry.About the Role:We're looking for a proactive Operations...
-
Senior Data Engineer
1 week ago
Remote, India AABM Cloud Data Solutions Full time ₹ 20,00,000 per yearSenior Data Engineer (Remote, India)10+ Years Data Engineer Experience RequiredSQL, Python, Snowflake - Mandatory hands-on experienceExperience with any ETL Tools (SSIS Preferred), Cloud Environments, and DB2/DBT is a plus.Job Type: Full-timePay: From ₹2,000,000.00 per yearApplication Question(s):How many years of Python Experience do you have?How many...
-
▷ High Salary! Data
1 week ago
Remote, India NTT Data Full timeJob Description NTT DATA strives to hire exceptional, innovative and passionate individuals who want to grow with us. If you want to be part of an inclusive, adaptable, and forward-thinking organization, apply now. We are currently seeking a Data & AI Engineer Lead to join our team in Remote, Karntaka (IN-KA), India (IN). Job Duties: Role Overview The Data...
-
Data engineer
8 hours ago
Remote, India INNONEXIS TECHNOLOGY PRIVATE LIMITED Full time ₹ 4,80,000 - ₹ 6,00,000 per yearWe're Hiring: Freelance Data EngineerWe are looking for a talented Freelance Data Engineer to help us design, build, and optimize scalable data pipelines and platforms. Key Responsibilities:Design, develop, and maintain ETL pipelines and workflowsWork with structured and unstructured data across multiple sourcesImplement and optimize data solutions on AWS...
-
Data Engineer
4 days ago
Remote, India Mig Staffing Full time ₹ 12,00,000 - ₹ 30,00,000 per yearWe are seeking a dynamic and highly skilled Data Engineer who has extensive experience building enterprise scale data platforms and lead these foundational efforts. This role demands someone who not only possesses a profound understanding of the data engineering landscape but is also at the forefront of their game. The ideal candidate will contribute...
-
Data Engineer
5 hours ago
Remote, India NS Global Corporation Full time ₹ 12,00,000 - ₹ 36,00,000 per yearJob Description : We are looking for an experienced GCP Data Engineer with a minimum of 5 years of professional experience in data engineering, cloud-based data solutions, and large-scale distributed systems. This role is fully remote and requires a hands-on professional who can design, build, and optimize data pipelines and solutions on Google Cloud...
-
Data Engineer
6 hours ago
Remote, India Codash Solutions Full time ₹ 10,80,000 - ₹ 12,00,000 per yearWe are seeking a highly skilled and motivated Data Engineer to join our growing team. The ideal candidate will have a strong background in building and maintaining scalable data pipelines, with hands-on experience in ETL processes, Python programming, Apache Airflow, Apache Spark, and Neo4j graph database technology. You will play a crucial role in...
-
Data Engineer
6 days ago
Remote, India Deltacubes Technologies Full time ₹ 13,20,000 - ₹ 14,40,000 per yearJD(1) for Data Engineer with strong Data Vault 2.0 - Chennai basedImmediate Joiners only· Experience: 6 -8yearsWFO/WFH/ Hybrid· Strong expertise in data transformation and pipeline development (SQL, Python, DBT)· Strong expertise and experience in Data Vault 2.0 modeling framework· IDE-savvy, working with dbt CLI locally and github copilot· Build dbt...
-
Data Engineer
6 days ago
Remote, India Panapps Full time ₹ 8,00,000 - ₹ 24,00,000 per yearJob TitleData EngineerJob LocationRemote/Work From AnywhereJob TypeFull TimeNo. of Positions2Date PostedSeptember 8, 2025Preferred Experience6+Job DescriptionWe are seeking a skilled and experienced Data Engineer to join our team. The ideal candidate will have a strong background in data engineering, ETL processes, and data integration, with hands-on...
-
Data Engineer
2 days ago
Remote, India AdmeeSync Full time ₹ 2,40,000 - ₹ 4,80,000 per yearKey ResponsibilitiesDesign, develop, and maintain scalable data pipelines and architectures.Build and optimize ETL/ELT processes for structured and unstructured data.Work with stakeholders to understand business requirements and translate them into technical solutions.Ensure data quality, governance, and compliance across platforms.Optimize data workflows...