
Big data-Pyspark Developer-Hadoop
- Bangalore, Karnataka
- Permanent
- Full-time
2. Implementing, managing and administering the overall hadoop infrastructure.
3. Knowledge all the components in the Hadoop ecosystem like Apache Spark, Apache Hive, HBase, Kafka, Sqoop, Yarn, Zookeeper etc.
4. Takes care of the day-to-day running of Hadoop clusters.
5. Candidate will have to work closely with the database team, network team, BI team and application teams to make sure that all the big data applications are highly available and performing as expected.
6. Candidate is responsible for capacity planning and estimating the requirements for lowering or increasing the capacity of the hadoop cluster.
7. Ensure that the hadoop cluster is up and running all the time.
8. Monitoring the cluster connectivity and performance.
9. Manage and review Hadoop log files.
10. Backup and recovery tasks
11. Resource and security management
12. Troubleshooting application errors and ensuring that they do not occur again."Technical and Professional Requirements :"Skillset required:
1. Excellent knowledge of UNIX/LINUX OS.
2. Knowing of core java is a plus but not mandatory.
3. Good understanding of OS concepts, process management and resource scheduling.
4. Basics of networking, CPU, memory and storage.
5. Good hold of shell scripting"Preferred Skills :Technology->Big Data - Hadoop->Hadoop Administration->HadoopEducational Requirements :Bachelor of EngineeringService Line :India Business Unit