Big Data/ Hadoop Admin

5 days ago


Begumpet Hyderabad Telangana, India Dhanush Infotech pvt ltd Full time

**Detailed Job Description: Big Data Admin/DevOps**

**Key Objective**:
As a Big Data Admin/DevOps professional, your primary role will be to manage the end-to-end Hadoop infrastructure, ensuring its efficiency, security, and stability for large-scale data processing and analysis.

**Primary Responsibilities**:
Maintain End-to-End Hadoop Infrastructure:
l You will oversee the entire Hadoop ecosystem, ensuring that all components like HDFS (Hadoop Distributed File System), YARN (Yet Another Resource Negotiator), HBase, Hive, Spark, and other related tools are functioning seamlessly.

l You'll be responsible for managing the cluster, ensuring its performance, reliability, and scalability as data processing needs grow.

**Roles and Responsibilities**:
**Minimum 6 Years of Experience in Hadoop Environments**:
l You must have a solid background working with Hadoop, particularly in administrative or development capacities, over a significant period to manage large, complex systems.

l This experience ensures a deep understanding of the Hadoop ecosystem, including best practices for deployment, optimization, and troubleshooting.

**Advanced Knowledge of the Hadoop Ecosystem and Its Components**:
l The role demands comprehensive knowledge of Hadoop components like HDFS, YARN, MapReduce, Hive, Pig, HBase, Spark, and others. You should understand how these components integrate and function within the system.

l Your expertise should include tuning, scaling, and optimizing these components to meet business needs.

**Installing, Configuring, and Supporting Hadoop**:
l You will be tasked with the complete lifecycle of Hadoop components, from installation to configuration and ongoing support.

l This includes setting up cluster nodes, ensuring proper networking, configuring resource management, and managing storage.

**Real-Time Troubleshooting of Hadoop Infrastructure**:
l You will need practical, hands-on experience in resolving infrastructure issues as they arise in real
- time.

l This could involve fixing connectivity issues, optimizing slow queries, or troubleshooting node failures in a multi-node environment.

**Understanding and Implementing Hadoop Security Mechanisms**:
l Hadoop security can be complex, with elements like Kerberos authentication, encryption of data at rest and in transit, and role-based access control (RBAC).

l You should have the ability to configure and implement these security protocols to protect the integrity and confidentiality of data.

**Ensuring the Chosen Hadoop Solution is Deployed Without Hindrance**:
l You will ensure that the deployment of Hadoop-based solutions, such as a data processing pipeline or a data analytics platform, proceeds smoothly without technical or operational roadblocks.

l This requires strong project management skills, in addition to technical knowledge.

**Hands-On Experience with GIT and CI/CD**:
l Experience with GIT version control is essential for managing codebases used in Hadoop and Spark jobs.

l CI/CD (Continuous Integration/Continuous Deployment) pipelines will be required to automate deployment and testing processes. Knowledge of tools like Jenkins, Docker, or Kubernetes will also be helpful.

**System Test for Big Data Admin/DevOps**:
This system test is designed to assess your hands-on technical abilities in a real-world scenario:
**Install and Configure a Hadoop Cluster and Spark Cluster**:
l You will be required to install and configure a Hadoop cluster (for distributed storage and processing) and a Spark cluster (for fast, in-memory processing).

l This involves setting up nodes, configuring resource allocation, and ensuring communication between nodes.

**Install and Configure Hadoop Components**:
l **HDFS (Hadoop Distributed File System)**:Manage distributed storage for large datasets.

l **YARN (Resource Manager)**:Allocate system resources to different tasks.

l **Hive Metastore with RDBMS**:Hive is used for querying and managing large datasets. You’ll configure it with an RDBMS (like MySQL or PostgreSQL) as its metastore.

l **Sqoop**:Integrate Sqoop to transfer data between Hadoop and traditional databases (e.g., importing/exporting data).

l **Kafka Cluster**:Set up Kafka, which is essential for real-time data ingestion into the Hadoop ecosystem.

**Import Data to HDFS**:
l After configuring the clusters and components, you’ll import data into HDFS. This is a critical task as you’ll manage and ensure the integrity and flow of large datasets into the Hadoop cluster for processing.

**Key Skills and Tools Needed**:

- Hadoop ecosystem knowledge (HDFS, YARN, Hive, HBase, Pig, etc.)
- Strong hands-on experience with Hadoop/Spark cluster configuration and administration.
- GIT version control for source code management.
- CI/CD pipeline management for smooth and automated deployments.
- Security implementation expertise (Kerberos, encryption, etc.).
- Experience with data ingestion tools like Kafka and Sqoop.

This role combines deep tec


  • Avp Hadoop Admin

    1 week ago


    Hyderabad, India Han Digital Full time

    8 to 14 Years Hyderabad Having minimum of 4+ years of relevant experience in Hadoop Administration, Hortonworks Data Platform, Apache Hadoop and Cloudera Enterprise distribution of Hadoop Eco-system supporting end users Having minimum experience working on 50 nodes cluster, preferably with Anaconda and Python scripting **Key Job...


  • Hyderabad, Telangana, India Tata Consultancy Services Full time US$ 90,000 - US$ 1,20,000 per year

    Role : Big data Hadoop Administration/Cloudera CDH AdministrationLocation: HyderabadExperience: 4 to 10 YearsMust Have skills:Installing Hadoop in Linux environment.Deployment in a Hadoop cluster and its maintenance.Health check of a Hadoop cluster monitoring whether it is up and running all the time.Analyze the storage data volume and allocating the space...

  • Hadoop Developer

    2 days ago


    Hyderabad, India S T M TECHNOLOGIES Full time

    Job Title : Hadoop Developer (Python / PySpark / Scala / Advanced SQL)Experience : 5 to 8 YearsNotice Period : Immediate to 15 DaysLocation : Any Brillio Location Bangalore, Pune, Hyderabad, Chennai, GurugramRole Overview :We are looking for an experienced Hadoop Developer with strong programming skills and a deep understanding of Big Data ecosystems. The...

  • Hadoop Admin

    2 weeks ago


    Kondapur, Hyderabad, Telangana, India TechSophy Full time

    At Techsophy, we believe that technology has the power to elevate lives. We’re not just building solutions; we’re building a future where everyone has the tools to thrive in four crucial dimensions of well-being: **Physical Health**:Offering accessible, high-quality healthcare that reaches everyone, everywhere, ensuring no one is left...

  • Big Data

    1 week ago


    Hyderabad, Telangana, India Galaxy i technologies Inc Full time ₹ 15,00,000 - ₹ 20,00,000 per year

    Hi, Everyone******CTH Contract Only*******CTH Contract Only*******Job Title: Big Data & Hadoop EcosystemLoc: Hyderabad (WFO)Job Description:Skill Required: Digital: Bigdata and Hadoop EcosystemsExperience Range: 6-8 Years.Job Description:Design, develop, and maintain scalable big data systems and pipelines.Implement data processing frameworks and optimize...


  • Hyderabad / Secunderabad, Telangana, India beBeeData Full time ₹ 1,04,000 - ₹ 1,30,878

    Big Data Analytics ProfessionalWe are seeking an experienced Big Data Analytics professional to lead our team in designing and implementing a cutting-edge analytics architecture. The ideal candidate will have a strong background in Data Warehousing, Big Data, and Hadoop implementation in Azure environments.Key Responsibilities:Design and implement...

  • Big Data Engineer

    2 weeks ago


    Hyderabad / Secunderabad, Telangana, India beBeeDataEngineering Full time ₹ 15,00,000 - ₹ 20,00,000

    Job Title: Big Data EngineerLead the development and implementation of scalable data pipelines using Hadoop ecosystem tools.Design, build, and maintain ETL/ELT pipelines for efficient data processing.Write complex Hive queries to transform and analyze large datasets.Collaborate with cross-functional teams to deliver solutions that meet business...


  • Hyderabad / Secunderabad, Telangana, Pune, India beBeeHadoop Full time ₹ 15,000 - ₹ 28,00,000

    Big Data DeveloperWe are seeking an experienced Big Data Developer to join our team. In this role, you will be responsible for designing, developing and maintaining large-scale data processing systems using Hadoop and Spark.The ideal candidate will have a strong background in computer science, with experience in working with big data technologies such as...


  • Hyderabad / Secunderabad, Telangana, India beBeeData Full time ₹ 1,04,000 - ₹ 1,30,878

    Job Overview:We are seeking a highly skilled Big Data Cluster Specialist to join our team. In this role, you will be responsible for designing and administering high-performance Hadoop clusters to meet the data processing needs of our organization.Key Responsibilities:Design and implement scalable Hadoop cluster architectures to support large-scale data...

  • Big Data Professional

    2 weeks ago


    Hyderabad / Secunderabad, Telangana, India beBeeDataEngineer Full time ₹ 9,00,000 - ₹ 12,00,000

    Big Data Engineer OpportunityWe are seeking a skilled Big Data Engineer to join our team.">Develop and maintain big data pipelines and ETL workflows using PySpark, Hadoop (HDFS, MapReduce, Hive, HBase).Design and implement data ingestion, transformation, and integration processes on Google Cloud Platform services such as BigQuery, Dataflow, Dataproc, and...