Big Data/ Hadoop Admin
16 hours ago
**Detailed Job Description: Big Data Admin/DevOps**
**Key Objective**:
As a Big Data Admin/DevOps professional, your primary role will be to manage the end-to-end Hadoop infrastructure, ensuring its efficiency, security, and stability for large-scale data processing and analysis.
**Primary Responsibilities**:
Maintain End-to-End Hadoop Infrastructure:
l You will oversee the entire Hadoop ecosystem, ensuring that all components like HDFS (Hadoop Distributed File System), YARN (Yet Another Resource Negotiator), HBase, Hive, Spark, and other related tools are functioning seamlessly.
l You'll be responsible for managing the cluster, ensuring its performance, reliability, and scalability as data processing needs grow.
**Roles and Responsibilities**:
**Minimum 6 Years of Experience in Hadoop Environments**:
l You must have a solid background working with Hadoop, particularly in administrative or development capacities, over a significant period to manage large, complex systems.
l This experience ensures a deep understanding of the Hadoop ecosystem, including best practices for deployment, optimization, and troubleshooting.
**Advanced Knowledge of the Hadoop Ecosystem and Its Components**:
l The role demands comprehensive knowledge of Hadoop components like HDFS, YARN, MapReduce, Hive, Pig, HBase, Spark, and others. You should understand how these components integrate and function within the system.
l Your expertise should include tuning, scaling, and optimizing these components to meet business needs.
**Installing, Configuring, and Supporting Hadoop**:
l You will be tasked with the complete lifecycle of Hadoop components, from installation to configuration and ongoing support.
l This includes setting up cluster nodes, ensuring proper networking, configuring resource management, and managing storage.
**Real-Time Troubleshooting of Hadoop Infrastructure**:
l You will need practical, hands-on experience in resolving infrastructure issues as they arise in real
- time.
l This could involve fixing connectivity issues, optimizing slow queries, or troubleshooting node failures in a multi-node environment.
**Understanding and Implementing Hadoop Security Mechanisms**:
l Hadoop security can be complex, with elements like Kerberos authentication, encryption of data at rest and in transit, and role-based access control (RBAC).
l You should have the ability to configure and implement these security protocols to protect the integrity and confidentiality of data.
**Ensuring the Chosen Hadoop Solution is Deployed Without Hindrance**:
l You will ensure that the deployment of Hadoop-based solutions, such as a data processing pipeline or a data analytics platform, proceeds smoothly without technical or operational roadblocks.
l This requires strong project management skills, in addition to technical knowledge.
**Hands-On Experience with GIT and CI/CD**:
l Experience with GIT version control is essential for managing codebases used in Hadoop and Spark jobs.
l CI/CD (Continuous Integration/Continuous Deployment) pipelines will be required to automate deployment and testing processes. Knowledge of tools like Jenkins, Docker, or Kubernetes will also be helpful.
**System Test for Big Data Admin/DevOps**:
This system test is designed to assess your hands-on technical abilities in a real-world scenario:
**Install and Configure a Hadoop Cluster and Spark Cluster**:
l You will be required to install and configure a Hadoop cluster (for distributed storage and processing) and a Spark cluster (for fast, in-memory processing).
l This involves setting up nodes, configuring resource allocation, and ensuring communication between nodes.
**Install and Configure Hadoop Components**:
l **HDFS (Hadoop Distributed File System)**:Manage distributed storage for large datasets.
l **YARN (Resource Manager)**:Allocate system resources to different tasks.
l **Hive Metastore with RDBMS**:Hive is used for querying and managing large datasets. You’ll configure it with an RDBMS (like MySQL or PostgreSQL) as its metastore.
l **Sqoop**:Integrate Sqoop to transfer data between Hadoop and traditional databases (e.g., importing/exporting data).
l **Kafka Cluster**:Set up Kafka, which is essential for real-time data ingestion into the Hadoop ecosystem.
**Import Data to HDFS**:
l After configuring the clusters and components, you’ll import data into HDFS. This is a critical task as you’ll manage and ensure the integrity and flow of large datasets into the Hadoop cluster for processing.
**Key Skills and Tools Needed**:
- Hadoop ecosystem knowledge (HDFS, YARN, Hive, HBase, Pig, etc.)
- Strong hands-on experience with Hadoop/Spark cluster configuration and administration.
- GIT version control for source code management.
- CI/CD pipeline management for smooth and automated deployments.
- Security implementation expertise (Kerberos, encryption, etc.).
- Experience with data ingestion tools like Kafka and Sqoop.
This role combines deep tec
-
Big Data Developer
1 week ago
Hyderabad, India Ara Resources Pvt Ltd Full timeAbout The Company :ARAs client is a leading IT solutions provider, offering Applications, Business Process Outsourcing (BPO) and Infrastructure services globally through a combination of technology knowhow, domain, and process expertise. They have also been ranked 29 of 100 largest financial technology vendors by American Banker and BAI in the FinTech...
-
Avp Hadoop Admin
4 days ago
Hyderabad, India Han Digital Full time8 to 14 Years Hyderabad Having minimum of 4+ years of relevant experience in Hadoop Administration, Hortonworks Data Platform, Apache Hadoop and Cloudera Enterprise distribution of Hadoop Eco-system supporting end users Having minimum experience working on 50 nodes cluster, preferably with Anaconda and Python scripting **Key Job...
-
Hadoop Admin
2 weeks ago
Hyderabad, India Tata Consultancy Services Full timeLocation: Chennai/Hyderabad - Hands-on experience in Hadoop Admin. - Experience in Monitoring of the environment, User provisioning, Security Policies and providing the best practices.
-
Sunware Technologies
2 weeks ago
Hyderabad, India Sunware Technologies Full timeDescription : Location : HyderabadExperience : 610 yearsKey Responsibilities : - Design, develop, and maintain scalable big data systems and pipelines.- Implement data processing frameworks and optimize large datasets using tools such as Hadoop, Spark, and Hive.- Develop and maintain ETL processes to ensure data availability, accuracy, and quality for...
-
Hyderabad, Telangana, India HSBC Global Services Limited Full timeSome careers shine brighter than others If you re looking for a career that will help you stand out join HSBC and fulfil your potential Whether you want a career that could take you to the top or simply take you in an exciting new direction HSBC offers opportunities support and rewards that will take you further HSBC is one of the largest banking and...
-
Hadoop Admin
6 days ago
Kondapur, Hyderabad, Telangana, India TechSophy Full timeAt Techsophy, we believe that technology has the power to elevate lives. We’re not just building solutions; we’re building a future where everyone has the tools to thrive in four crucial dimensions of well-being: **Physical Health**:Offering accessible, high-quality healthcare that reaches everyone, everywhere, ensuring no one is left...
-
Big Data Engineer
3 weeks ago
Hyderabad, India Impacteers Full timePosition Overview :We are looking for a skilled and detail-oriented Big Data Engineer to design, develop, and maintain scalable data pipelines and architectures. The role involves working with large datasets, integrating diverse data sources, and ensuring data availability for analytics, machine learning, and business intelligence. The ideal candidate will...
-
Big Data
3 days ago
Hyderabad, Telangana, India RiskInsight Consulting Pvt Ltd Full time ₹ 8,00,000 - ₹ 24,00,000 per yearResponsibilitiesProvide technical support and troubleshooting for Big Data applications and systems built on the Hadoop ecosystem. Monitor system performance, analyze logs, and identify potential issues before they impact services. Collaborate with engineering teams to deploy and configure Hadoop clusters and related components. Assist in maintenance and...
-
Big Data Engineer
2 days ago
Hyderabad, Telangana, India HARP Technologies And Services Pvt Ltd Full time ₹ 12,00,000 - ₹ 36,00,000 per yearDescription : - Skill : Bigdata + Pyspark - Location : Hyderabad only - Experience : 5-9 Years - Notice Period : Immediate to 30 Days - Interview Mode : L1 Virtual / L2 F2F (Mandatory) - Work Mode : Hybrid - Skills required : Bigdata (Hadoop Hive, Impala, Spark) , pyspark, Python, Oracle, Exadata (RDBMS),...
-
Big Data Engineer
3 weeks ago
Hyderabad, Bengaluru, India Fusion Plus Solutions Full timeJob Description We are looking for a skilled Big Data Engineer with 5 to 7 years of hands-on experience in Big Data technologies. The ideal candidate should have strong expertise in Hadoop, HDFS, Hive, SQL, and Unix, and be able to work on large-scale data processing systems. Roles and Responsibilities: - Work on design, development, and optimization of Big...