Sr Hadoop Admin Apply
Key Responsibilities:
1. Install, configure, and manage Hadoop clusters (HDFS, MapReduce, YARN, etc.)
2. Monitor cluster performance, identify bottlenecks, and optimize resource utilization.
3. Implement and manage Hadoop security (Kerberos, ACLs, encryption).
4. Develop and maintain scripts for automation (Python, Shell).
5. Collaborate with developers to optimize data processing workflows.
6. Troubleshoot Hadoop-related issues and provide Level 3 support.
7. Plan and execute cluster upgrades, patches, and backups.
8. Ensure data quality, integrity, and compliance with organizational policies.
9. Document cluster architecture, configuration, and procedures.
10. Work closely with data engineers, analysts, and scientists to meet business requirements.
2. Monitor cluster performance, identify bottlenecks, and optimize resource utilization.
3. Implement and manage Hadoop security (Kerberos, ACLs, encryption).
4. Develop and maintain scripts for automation (Python, Shell).
5. Collaborate with developers to optimize data processing workflows.
6. Troubleshoot Hadoop-related issues and provide Level 3 support.
7. Plan and execute cluster upgrades, patches, and backups.
8. Ensure data quality, integrity, and compliance with organizational policies.
9. Document cluster architecture, configuration, and procedures.
10. Work closely with data engineers, analysts, and scientists to meet business requirements.
Requirements:
Technical Skills:
1. 3+ years of experience in Hadoop administration.
2. Strong understanding of Hadoop ecosystem (HDFS, MapReduce, YARN, Hive, Pig, Sqoop, Flume).
3. Experience with cluster management tools (Ambari, Cloudera Manager).
4. Proficient in Linux/Unix and scripting languages (Python, Shell).
5. Knowledge of Hadoop security and authentication protocols.
6. Familiarity with data processing frameworks (Spark, Flink).
7. Experience with cloud-based Hadoop deployments (AWS, Azure, GCP).
2. Strong understanding of Hadoop ecosystem (HDFS, MapReduce, YARN, Hive, Pig, Sqoop, Flume).
3. Experience with cluster management tools (Ambari, Cloudera Manager).
4. Proficient in Linux/Unix and scripting languages (Python, Shell).
5. Knowledge of Hadoop security and authentication protocols.
6. Familiarity with data processing frameworks (Spark, Flink).
7. Experience with cloud-based Hadoop deployments (AWS, Azure, GCP).