
Python Data Architect
1 week ago
As a skilled Python data engineering professional, you will design, develop, and optimize ETL pipelines using open-source or cloud ETL tools. You will also work on exciting projects that require developing scalable data pipelines, real-time data streaming, and optimizing data processing tasks using Python.
Key Responsibilities:
- Design, develop, and optimize ETL pipelines using Apache Nifi, Talend, Pentaho, Airflow, and AWS Glue.
- Write Python scripts to automate data extraction, transformation, and loading (ETL) processes, ensuring that the code is optimized for performance and scalability.
- Work with Apache Spark and PySpark to process large datasets in a distributed computing environment, optimizing Spark jobs for performance and resource efficiency.
- Use Apache Airflow or other orchestration tools to schedule, monitor, and automate data pipeline workflows.
- Design and implement real-time data streaming solutions using technologies like Apache Kafka or AWS Kinesis for high-throughput, low-latency data processing.
- Work with open-source table formats like Apache Parquet, Apache Avro, or Delta Lake, and other structured/unstructured data formats for efficient data storage and access.
- Work with relational databases like PostgreSQL, MySQL, or SQL Server for data storage, management, and optimization, understanding database concepts such as normalization, indexing, and query optimization.
- Write and optimize complex SQL queries for data extraction, transformations, and aggregation across large datasets, ensuring queries are efficient and scalable.
Required Skills & Qualifications:
- Experience working with open-source ETL tools like Apache Nifi, Talend, or Pentaho, with cloud-based tools like AWS Glue or Azure Data Factory being beneficial.
- Proficiency in Python for automating data processing tasks, writing data pipelines, and working with libraries like Pandas, Dask, PySpark, etc.
- Experience with Apache Spark and PySpark for distributed data processing, along with optimization techniques.
- Experience using Apache Airflow or similar tools for scheduling and automating data pipelines.
- Experience with Apache Kafka or AWS Kinesis for building and managing real-time data pipelines.
- Knowledge of Apache Parquet, Apache Avro, Delta Lake, or similar open-source table formats for efficient data storage and retrieval.
- Strong experience with at least one relational database like PostgreSQL, MySQL, or SQL Server, with a solid understanding of database concepts like indexing, normalization, and query optimization.
- Strong skills in writing and optimizing complex SQL queries for data extraction, transformations, and aggregation.
Nice to Have:
- Familiarity with BI tools like Power BI, Tableau, Looker, or similar reporting and data visualization platforms.
- Knowledge of data warehousing principles, schema design (e.g., star/snowflake), and optimization techniques for large datasets.
- Experience with cloud data platforms like Databricks, Snowflake, or Azure Synapse is beneficial, though the role is focused on on-prem environments.
- Familiarity with containerization tools like Docker or Kubernetes for deploying data engineering workloads.
Educational Qualifications:
- Bachelor's or Master's degree in Computer Science, Engineering, Information Systems, or a related field (or equivalent work experience).
Additional Qualities:
- Excellent problem-solving and troubleshooting skills.
- Ability to work both independently and in a collaborative environment.
- Strong communication skills, both written and verbal.
- Detail-oriented with a focus on data quality and performance optimization.
- Proactive attitude and the ability to take ownership of projects.
-
Data Architect
7 days ago
Nashik, Maharashtra, India beBeeData Full time ₹ 25,00,000 - ₹ 30,00,000Cloud Data ArchitectJob Overview:We are seeking a highly skilled Cloud Data Architect to design, implement, and optimize data warehouses on the Snowflake platform.Key Responsibilities:Design and develop cloud-based data architecture for scalable and high-performance data storage.Ensure effective utilization of Snowflake features for efficient data processing...
-
Senior Data Architect
7 days ago
Nashik, Maharashtra, India beBeeDataScientist Full time ₹ 15,00,000 - ₹ 25,00,000Senior Data ArchitectWe are seeking a highly skilled Senior Data Architect to join our team.Design and implement data pipelines and ETL processes using AWS services.Develop efficient, reusable, and scalable code in Python.Work with SQL databases for data modeling, queries, and performance optimization.Implement and optimize solutions using AWS services such...
-
Data Architect
1 week ago
Nashik, Maharashtra, India beBeeData Full time ₹ 1,00,00,000 - ₹ 2,50,00,000Job Title: Data ArchitectOverview:We are seeking an experienced Data Architect to lead the development of our big data ecosystem. As a key member of our team, you will be responsible for designing and implementing scalable data solutions that drive business growth.Key Responsibilities:Design and develop distributed computing systems using Apache Spark and...
-
Data Architect
6 days ago
Nashik, Maharashtra, India beBeeData Full time ₹ 1,50,00,000 - ₹ 2,50,00,000Job Title: Data ArchitectKey Responsibilities:Develop predictive models and algorithms that improve business outcomes.Collaborate with cross-functional teams to integrate AI solutions into business workflows.Design and deploy machine learning models that solve business problems (customer insights, forecasting, optimization).Stay updated with industry trends...
-
Enterprise Data Architect
1 week ago
Nashik, Maharashtra, India beBeeData Full time ₹ 90,00,000 - ₹ 1,20,00,000Job Title: Enterprise Data ArchitectDescriptionWe are seeking an experienced and skilled enterprise data architect to join our team. The ideal candidate will work on designing, developing, and maintaining scalable back-end services, APIs, and databases, particularly focused on handling complex graph data.The role involves designing and implementing back-end...
-
Cloud Data Architect
7 days ago
Nashik, Maharashtra, India beBeeData Full time ₹ 2,00,00,000 - ₹ 2,50,00,000Job Title: Cloud Data ArchitectWe are seeking a highly skilled cloud data architect to join our team.The ideal candidate will have experience in designing and implementing data architecture for all AWS data services. They will work with stakeholders to identify business needs and requirements for data-related projects, design and implement ETL processes to...
-
Cloud Data Architect
7 days ago
Nashik, Maharashtra, India beBeeData Full time ₹ 1,50,00,000 - ₹ 2,50,00,000Job Title: Cloud Data Architect Job Summary We are seeking an experienced and skilled cloud data architect to design and develop robust data pipelines and scalable cloud-native data solutions. The ideal candidate will be proficient in GCP services like Pub/Sub, Dataflow, Cloud Storage, and BigQuery, with a foundational understanding of AI/ML workflows using...
-
Chief Data Architect
7 days ago
Nashik, Maharashtra, India beBeeDataArchitect Full timeJob OverviewWe seek a seasoned Azure Data Architect to drive data-driven innovation. This role entails designing and implementing scalable data architectures, ensuring optimal performance, cost efficiency, and adherence to governance standards.Key ResponsibilitiesArchitect robust data solutions using Azure Databricks and ADLS aligned with the medallion...
-
Master Data Architect
1 week ago
Nashik, Maharashtra, India beBeeDataEngineer Full time ₹ 20,00,000 - ₹ 25,00,000Job DescriptionWe are seeking an experienced data engineer to lead the design, development, and deployment of scalable data pipelines using Scala and Apache Spark (Databricks). The successful candidate will develop and maintain data processing workflows using PySpark and Python.Key ResponsibilitiesDesign and implement scalable data pipelines using Scala and...
-
Cloud-Based Data Architect
7 days ago
Nashik, Maharashtra, India beBeeData Full time ₹ 1,50,00,000 - ₹ 2,50,00,000Job Title: Data ArchitectWe are looking for an experienced data engineer to join our team. The ideal candidate will have a strong background in AWS IoT and Python, with at least 5 years of experience in designing and implementing scalable data pipelines.Key Responsibilities:Design, develop and deploy scalable data pipelines using AWS servicesWrite efficient,...