
Data Pipeline Specialist
5 days ago
An experienced Business Analyst is required to design and develop scalable data pipelines using PySpark and Delta Lake for efficient processing of large, structured, and semi-structured datasets.
The ideal candidate will apply analytical skills to identify trends, patterns, and opportunities for process improvement by analyzing complex data sets using SQL, Python (Pandas, NumPy, scikit-learn, Seaborn).
They will optimize the performance of data processing systems and ensure high levels of accuracy. Additionally, they will create interactive dashboards and reports using tools like Metabase, Tableau, or Power BI.
The Business Analyst will collaborate with management to prioritize business and information needs.
Required Skills and Qualifications- Bachelor's or Master's degree in Statistics, Mathematics, Engineering, Computer Science, or a related field.
- At least 2 years of experience as a Business Analyst with strong expertise in PySpark, Delta Lake, and SQL.
- Solid understanding of data models, database design, and development.
- Experience building incremental data pipelines and implementing watermark-based processing.
- Proficiency with Python and libraries like Pandas, NumPy, scikit-learn, and Seaborn for data analysis and visualization.
- Experience developing dashboards and interactive reports using Metabase, Tableau, or similar tools.
- Strong analytical skills with a keen eye for detail, accuracy, and quality.
-
Data Pipeline Specialist
5 days ago
Gurgaon, Haryana, India beBeeDataEngineer Full time ₹ 15,00,000 - ₹ 20,10,000About the Role:We are seeking a skilled professional to design, build, and optimize data pipelines while ensuring security, scalability, and high performance.This role will involve working closely with global teams on large-scale data projects in a collaborative and innovation-driven environment.Location: Gurgaon (Hybrid)Experience: 7+ yearsEmployment Type:...
-
Senior Data Pipeline Specialist
5 days ago
Gurgaon, Haryana, India beBeeDataEngineer Full time ₹ 20,00,000 - ₹ 25,00,000Job OverviewA Data Engineer is responsible for designing, building, and maintaining scalable and secure data pipelines and infrastructure on GCP. This includes working with various GCP services like BigQuery, Dataflow, Cloud Storage, and Pub/Sub to create efficient data solutions.Key responsibilities include developing ETL processes, implementing data...
-
Senior Data Pipeline Engineer
24 hours ago
Gurgaon, Haryana, India beBeeData Full time ₹ 1,50,000 - ₹ 28,00,000Job Description:A Data Integration Specialist designs, develops and maintains complex data pipelines to extract, transform and load data into target systems. Collaborating with cross-functional teams, this role identifies business requirements and creates solutions that meet those needs.Develops ETL processes using Informatica PowerCenter to optimize...
-
Data Pipeline Architect
3 days ago
Gurgaon, Haryana, India beBeeDataEngineer Full time ₹ 15,00,000 - ₹ 28,00,000Job Overview:We are seeking a skilled professional to fill the role of Senior Data Engineer, responsible for designing and implementing data pipelines and extracting information from various sources.Key Responsibilities:Data Extraction and Parsing: Develop automated systems to extract relevant data from websites, ensuring accuracy and uptime.Pipeline...
-
AI/ML Engineer
5 days ago
Gurgaon, Haryana, India beBeeDataPipeline Full time ₹ 40,00,000 - ₹ 50,00,000Job Title: AI/ML Engineer - Data Pipeline Specialist We are seeking a skilled AI/ML engineer to join our team. The successful candidate will be responsible for building, optimizing and scaling AI/ML pipelines for regulatory compliance solutions. Key Responsibilities:Build and optimize data pipelines to transform and clean data for RAG (Retrieval Augmented...
-
Data Pipeline Engineer
1 week ago
Gurgaon, Haryana, India beBeeDataPipeline Full time ₹ 9,00,000 - ₹ 12,00,000Job SummaryWe are seeking a highly skilled Data Pipeline Engineer to join our team. In this role, you will be responsible for designing, building, and maintaining complex data pipelines that enable us to extract insights from large datasets.About the RoleThis is a fantastic opportunity for someone who has experience with ETL tools and technologies, as well...
-
Data Pipeline Developer
4 days ago
Gurgaon, Haryana, India ScoreMe Solutions Full time ₹ 1,04,000 - ₹ 1,30,878 per yearData Engineer with expertise in MuleSoft APIs, Azure Data Factory, and GoAnywhere Managed File Transfer services.Resource will be responsible for developing, implementing, and maintaining various applications and systems to support our state agencies. This role also includes operations and support activities, such as fielding and resolving customer tickets...
-
Senior Data Pipeline Engineer
6 days ago
Gurgaon, Haryana, India beBeeKafkaIntegration Full time US$ 1,04,000 - US$ 1,30,878We are seeking a highly skilled Kafka Integration Specialist with extensive experience in designing, developing and integrating Apache Kafka solutions.This role requires a strong understanding of distributed systems and data streaming concepts to deliver robust real-time integration solutions. We are looking for an expert in creating and maintaining...
-
Data Pipeline Architect
7 days ago
Gurgaon, Haryana, India beBeeEtldata Full time ₹ 20,00,000 - ₹ 25,00,000Job Opportunity:We are seeking an experienced professional to fill a pivotal role in our organization. As an ETL Developer, you will be responsible for designing and implementing data pipelines that drive business growth.About the Role:The ideal candidate will have hands-on experience with data warehouse design, data modeling, and ETL tools such as...
-
Senior Data Pipeline Specialist
4 days ago
Gurgaon, Haryana, India beBeeDataEngineer Full time ₹ 80,00,000 - ₹ 1,50,00,000Key Data Engineer RoleAs a key member of our team, we are seeking an experienced Data Engineer to join us in driving innovation and delivering high-quality solutions. The ideal candidate will have a strong background in data engineering, with expertise in designing, developing, and maintaining robust data pipelines using Python, PySpark, and SQL.Essential...