
Hadoop Administrator
- Gurgaon, Haryana
- Permanent
- Full-time
- Installing, upgrade Cloudera Hadoop ecosystem in Linux environment.
- Deployment in a Hadoop cluster and its maintenance.
- Health check of a Hadoop cluster monitoring & alerting whether it is up and running all the time.
- Analyse the storage data volume and allocating the space in HDFS.
- Resource management in a cluster environment. This involves new node creation and removal of unused ones.
- Configuring Name Node to ensure its high availability
- Implementing and administering Hadoop infrastructure on an ongoing basis.
- Required hardware and software deployment in Hadoop environment. Furthermore, to expanding of existing environments.
- Production support experience with SQL Server, MariaDB/MySQL, PostgreSQL, or MongoDB is a plus
- Knowledge in Big query is plus.
- User creation in Linux for Hadoop and its components in the ecosystem. Moreover, setting up Kerberos principals is a part of Hadoop administration.
- Performance tuning and running jobs in Hadoop clusters.
- Capacity planning
- Monitoring connectivity and security of Hadoop cluster
- Managing and reviewing log files in Hadoop.
- Management of HDFS file system and monitoring them.
- Maintaining HDFS and providing necessary supports.
- Backup and recovery tasks.
- Communicating with other development, administrating and business teams. They include infrastructure, application, network, database, and business intelligence teams. Effective communication plays a key role in high quality and availability of data.
- Coordinating with application teams. Installing the operating system and Hadoop related updates as and when required.
- Working as a key person for Vendor escalation
- Troubleshooting
- General operational excellence. This includes good troubleshooting skills, understanding of system’s capacity and bottlenecks. Furthermore, you also need a basic understanding of memory management areas.
- Proper knowledge and hands-on experience in Hadoop ecosystem components.
- Hadoop administration demands one’s deployment skill in Hadoop cluster. Moreover, the job needs adding and removing nodes. It may need tracking of jobs and monitor the important parts of the cluster.
- Configuring high availability of name-node.
- Scheduling and taking backups
- Linux corn is the running platform of Hadoop. Hence, a good hands-on experience of Linux, its commands and scripting are a must.
- Configuration management and deployment exposure in Open-source environment.
- Knowledge of Core Java is an added advantage in performing the job
- 3+ years of administrator experience on Big Data Hadoop Ecosystem and related technologies (Impala, Spark, Sqoop, Hive, Yarn…)
- Prior experience with performance tuning, capacity planning, and workload mapping a plus and experiences around the Cloudera auto scaling.
- Understanding of architecture and support around Big Data and BI solutions is required
- 3+ years of strong Linux and Hadoop administration background in a production support setting is required
- Working knowledge of clustering, replication, and log shipping techniques is a plus
- Expert in SQL and/or Shell scripting, stored procedures, and construction of highly complex queries is required
- Hands on experience working with unstructured data & hardware and configuration is required
- Strong understanding of high availability, fault tolerance, load balancing, and clustering is preferred