Cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise, Dell Open Manage and other tools.
Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
Responsible for implementation and ongoing administration of Hadoop infrastructure, Data science infrastructure and Lakehouses
Automate manual tasks.
Data modelling, design & implementation based on recognized standards.
The Hadoop Administrator administers database systems to protect the confidentiality, integrity and availability of data. The Hadoop Administrator is responsible for installations, upgrades, backups, and configuration. They also maintain query language for established database systems including Starburst Trino , design and execute backup and recovery schemes and implement disaster recovery BDR for Hadoop procedures. The Administrator works with end-users and project team members to understand and advise on database Impala, Hive and Hbase and query requirements and data science query capabilities. They are assigned to complex database systems including those that work on Data science Linux servers and are relied upon to optimize database performance, R configuration, access and security and to serve on project teams contributing database subject matter expertise.
Database backup and recovery.
Point of Contact for Vendor escalation
Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
Working with data delivery teams to setup new Hadoop users. This job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users.
Manage and review Hadoop log files.
File system management and monitoring.
Performance monitoring and tuning.
Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments to Lakehouse technologies
5-8 Years of Hadoop Admin Experience
Related Bachelor's degree in an IT related field or relevant work experience
Monitor Hadoop cluster connectivity and security
Screen Hadoop cluster job performances and capacity planning
Database connectivity and security.
3-5 Years of DBA experience
Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
Disk space management.
Strong communication skills
Benefits