
Big Data
6 days ago
Hiring for Big Data PySpark Engineer - Chennai Bangalore Job Overview We are looking for a Big Data PySpark Engineer to join our dynamic data engineering team The ideal candidate will have hands-on experience with big data frameworks particularly Apache Spark with a focus on PySpark and be capable of building optimizing and managing large-scale data processing systems As a Big Data Engineer you will work closely with Data Scientists Analysts and Architects to design and implement efficient data pipelines and ensure that our data infrastructure is scalable and robust Responsibilities Data Pipeline Development Design develop and maintain scalable and efficient data pipelines using PySpark and related big data technologies Data Processing Transformation Write complex data transformations using PySpark to process large datasets and extract meaningful insights Optimization Performance Tuning Optimize Spark jobs for performance manage memory and data partitioning and troubleshoot performance bottlenecks ETL Workflows Build and manage ETL Extract Transform Load workflows that process data from various sources e g databases APIs file systems Data Integration Work with different data sources like HDFS AWS S3 Google Cloud Storage and relational or NoSQL databases Data Quality Monitoring Implement data validation checks and monitor data pipeline health logging and error handling Collaboration Collaborate with data scientists analysts and business teams to ensure data infrastructure meets the company s needs Cloud Computing Work with cloud platforms such as AWS Azure or GCP to deploy and manage big data workflows Documentation Best Practices Maintain clear documentation for code processes and system designs Promote best practices in coding version control and testing
-
Senior Data Pipeline Specialist
6 days ago
Bijapur, Karnataka, India beBeeDataEngineer Full time ₹ 20,00,000 - ₹ 25,00,000Job SummaryWe are seeking an experienced Data Engineer to join our team and contribute to the design and implementation of efficient data pipelines.Main ResponsibilitiesDesign, develop, and maintain scalable and efficient data pipelines using big data technologies such as PySpark.Write complex data transformations using PySpark to process large datasets and...
-
Senior Consultant
3 weeks ago
Bijapur, Karnataka, India HCLTech Full time8+ years of strong proficiency in designing and implementing the solution involvingDesigning data models, Planning databases, Ensuring data security, Improving databases, Creating a data catalog, Staying current5+ years of experience with testing tools related to big data technologies (e.g., Spark, Databricks)3+ years of expertise with big data technologies...