Senior Data Engineer
1 week ago
Senior Data Engineer Location: Coimbatore Experience: 3–5 Years Work Mode: On-site About the Role We are seeking a Senior Data Engineer who will take end-to-end ownership of designing, building, and scaling data pipelines, integrations, and analytics layers for our Enterprise AI Operational Platform . The role involves integrating data from construction-focused enterprise tools such as Sage 300, Paperless, WorkMax, Procore, Kojo, CRM platforms , and other financial, procurement, and field applications to create a unified operational data ecosystem . You will architect scalable systems, build automated ETL/ELT workflows, standardize data governance, and enable real-time insights, predictive analytics, and automated operational intelligence across the organization. This position suits someone who brings strong technical depth, system architecture capability, and hands-on experience building large-scale data systems using Microsoft Azure and/or Snowflake . Key Responsibilities Data Architecture & System Design Architect modular, scalable, and secure data platforms. Design centralized data models to support analytics, forecasting, AI, and automation workloads . Define data governance standards including lineage, quality, and retention. Produce HLD/LLD architecture diagrams, integration maps, and technical documentation. Build frameworks that support complex cross-system workflows. Integrations Across Enterprise Tools (Sage 300, Paperless, WorkMax, Procore, Kojo, CRM, etc.) Lead integrations across construction and operations software including: Sage 300 (Financials, Job Costing, AP/AR, GL) Paperless (Document workflows, invoice automation) WorkMax (Timecards, units, production, field data) Procore (Project management, RFIs, submittals, schedules) Kojo (Procurement, materials management) CRM systems (Sales pipeline, customer records) Develop connectors using REST APIs, GraphQL, Webhooks, Secure File Transfer, and Database links . Standardize schemas across tools to create a single source of truth . Ensure secure and reliable data ingestion from on-prem and cloud systems. Implement robust incremental sync, batch loads, and real-time event-based ingestion. ETL/ELT Development & Data Pipelines Build automated ETL/ELT workflows using Airflow, NiFi, Azure Data Factory, or custom frameworks . Ensure pipelines are fault-tolerant, versioned, and optimized for performance . Handle complex transformations, reconciliations, and multi-system data validation. Develop end-to-end data ingestion frameworks supporting large structured and unstructured datasets. Deploy pipelines into production with CI/CD practices. Data Warehousing, Modeling & Real-Time Reporting Build enterprise-wide data warehouse layers using Snowflake (preferred) , Azure SQL, or cloud data stores. Maintain real-time operational data models for job cost, purchasing, productivity, CRM, safety, compliance, and field activity. Create unified analytical datasets for dashboards and business intelligence. Support executive and departmental reporting by eliminating manual spreadsheets and fragmented data silos. Ensure high data accuracy, reliability, and performance at scale. Predictive Analytics & AI Enablement Prepare AI-ready datasets for forecasting, anomaly detection, productivity modeling, and operational risk scoring. Collaborate with data scientists/AI engineers to build and deploy ML pipelines. Develop streaming/near-real-time data flows to support AI-driven decision automation. Integrate predictive insights back into operation systems (alerts, workflows, dashboards). Automation & Operational Workflows Automate KPI tracking, threshold-based alerts, and data validation rules. Build automated workflows for: RFIs CORs Contracts Vendor documents Field reporting Implement monitoring frameworks for pipeline health, error handling, retry logic, and SLA compliance. Documentation & Knowledge Sharing Create detailed documentation for data models, transformations, integrations, and pipelines. Conduct internal training and handover sessions to engineering, analytics, and operations teams. Maintain continuous improvement of data standards and reusable components. Skills & Experience Required Core Experience 3–5 years of hands-on Data Engineering / Data Platform development . Strong expertise in API-based integrations , especially with enterprise applications. Practical experience with: Python , FastAPI, Node.js ETL frameworks: Airflow / NiFi / Azure Data Factory Snowflake (strongly preferred) or Azure-based data warehousing SQL + cloud databases Azure Functions, Storage, Data Lake environments Data Architecture & Modeling Strong data modeling principles (Kimball, Data Vault, or custom patterns). Building scalable data lakes, warehouse schemas, and analytical data marts. Tools & Platforms Dashboarding tools: Power BI / Tableau Experience working with large structured, semi-structured, and unstructured datasets . Background in financial or field operations workflows (preferred). Preferred Knowledge Working with industry tools such as: Sage 300 Procore WorkMax Kojo Paperless Experience in building ML/AI data pipelines. Exposure to Azure DevOps, CI/CD , and cloud deployment best practices. Key Competencies Strong system design and architecture thinking. High ownership, accountability, and delivery focus. Excellent analytical and problem-solving abilities. Clear communication and documentation skills. Ability to work with leadership, operations, finance, field teams, and engineering. Commitment to data accuracy, reliability, and operational performance. What We Offer Opportunity to build a next-generation Enterprise AI Data Platform . Career growth into Data Platform Lead / Engineering Manager roles . Medical insurance and employee benefits. Supportive leadership, high visibility, and impactful responsibilities. Competitive compensation aligned with expertise and contribution.
-
Senior Data Bricks Data Engineer
2 weeks ago
Coimbatore, Tamil Nadu, India Squash Apps Full time ₹ 6,00,000 - ₹ 7,50,000 per yearSenior Databricks Data EngineerWe're looking for a Senior Databricks Data Engineer to lead large-scale data pipeline development on the Databricks Lakehouse Platform. If you're strong in Spark, cloud platforms, and modern data engineering practices—this role is for you. Responsibilities● Build & optimize ETL/ELT pipelines using Databricks (PySpark, SQL,...
-
Senior Data Engineer
1 week ago
Coimbatore, India AppXcess Technologies Full timeSenior Data EngineerLocation: CoimbatoreExperience: 3–5 YearsWork Mode: On-siteAbout the RoleWe are seeking a Senior Data Engineer who will take end-to-end ownership of designing, building, and scaling data pipelines, integrations, and analytics layers for our Enterprise AI Operational Platform.The role involves integrating data from construction-focused...
-
Senior Data Engineer
1 week ago
Coimbatore, India AppXcess Technologies Full timeSenior Data EngineerLocation: CoimbatoreExperience: 3–5 YearsWork Mode: On-siteAbout the RoleWe are seeking a Senior Data Engineer who will take end-to-end ownership of designing, building, and scaling data pipelines, integrations, and analytics layers for our Enterprise AI Operational Platform .The role involves integrating data from ...
-
Senior Data Engineer
1 week ago
Coimbatore, India AppXcess Technologies Full timeSenior Data Engineer Location: Coimbatore Experience: 3–5 Years Work Mode: On-siteAbout the RoleWe are seeking a Senior Data Engineer who will take end-to-end ownership of designing, building, and scaling data pipelines, integrations, and analytics layers for our Enterprise AI Operational Platform.The role involves integrating data from...
-
Senior Data Engineer
1 week ago
Coimbatore, India AppXcess Technologies Full timeSenior Data Engineer Location: Coimbatore Experience: 3–5 Years Work Mode: On-siteAbout the RoleWe are seeking a Senior Data Engineer who will take end-to-end ownership of designing, building, and scaling data pipelines, integrations, and analytics layers for our Enterprise AI Operational Platform.The role involves integrating data from...
-
Senior Data Engineer
1 week ago
Coimbatore, India AppXcess Technologies Full timeSenior Data Engineer Location: Coimbatore Experience: 3–5 Years Work Mode: On-siteAbout the RoleWe are seeking a Senior Data Engineer who will take end-to-end ownership of designing, building, and scaling data pipelines, integrations, and analytics layers for our Enterprise AI Operational Platform.The role involves integrating data from...
-
Senior Data Engineer
1 week ago
Coimbatore, India AppXcess Technologies Full timeSenior Data Engineer Location: Coimbatore Experience: 3–5 Years Work Mode: On-site About the Role We are seeking a Senior Data Engineer who will take end-to-end ownership of designing, building, and scaling data pipelines, integrations, and analytics layers for our Enterprise AI Operational Platform. The role involves integrating data from...
-
Senior Data Engineer
1 week ago
Coimbatore, India AppXcess Technologies Full timeSenior Data Engineer Location: Coimbatore Experience: 3–5 Years Work Mode: On-site About the Role We are seeking a Senior Data Engineer who will take end-to-end ownership of designing, building, and scaling data pipelines, integrations, and analytics layers for our Enterprise AI Operational Platform . The role involves integrating data from...
-
Senior Data Engineer
1 week ago
coimbatore, India AppXcess Technologies Full timeSenior Data EngineerLocation: CoimbatoreExperience: 3–5 YearsWork Mode: On-siteAbout the RoleWe are seeking a Senior Data Engineer who will take end-to-end ownership of designing, building, and scaling data pipelines, integrations, and analytics layers for our Enterprise AI Operational Platform.The role involves integrating data from construction-focused...
-
Senior Data Engineer
1 week ago
Coimbatore, Tamil Nadu, India Numentica Consulting Group Full time ₹ 8,00,000 - ₹ 12,00,000 per yearPreference to candidates residing @ ChennaiThis is a Fulltime and Onsite role @ Chennai.NuStartz is seeking an experienced Senior Data Engineer to drive the modernization of our data infrastructure. The ideal candidate will play a key role in designing, building, and optimizing data pipelines as we migrate to Google Cloud Platform (GCP)—leveraging...