Primary Skills: Hadoop Map/Reduce, SPRAK, Druid, Hive, YARN, Oozie, Scala/Java, Web Services, MySQL, Ansible scripts
• Experience designing and building Big Data applications in on-premise, hosted and Public cloud (like, AWS) operating in under stringent SLAs
• Experience in designing (cluster topology, technology stack) and defining (sizing, tuning, non-functional goals) of large-scale Hadoop (preferably Hortonworks Data Platform) as a Service cluster (multi-tenant)
• Deep architecture knowledge and hands-on experience in productionizing Hadoop based products using technologies, like, Kafka, Spark, HDFS, Hive, YARN, Druid, Kerberos, Oozie, SQL, Webservices (Must have), Superset, Kubernetes, dockers (Good to have)
• Should have very strong performance engineering focus in Building High performing bigdata applications with scalable design. Work towards establishing performance engineering best practices both at Infra level and at application level and get into application level tuning when required.
• Implement system management strategies for monitoring, optimization, rapid feedback and high availability for Bigdata platform. Conduct capacity planning on periodic basis and forecast capacity demands.
• Collaborating with engineering, DevOps & Hadoop admin team to ensure adherence to the optimal design practices, and troubleshoot and resolve issues in dev, test and production environments from infrastructure is highly available and performing as expected.
• Collaborate with Hadoop administrators to deploy/upgrade Hadoop cluster, manage (nodes, services, users, etc), cluster and job tuning, implement non-functional features, like, High availability, security, backup and disaster recovery, monitoring, etc.
• Should have expert level knowledge in one of JVM based languages, like, Java, Scala. Well versed with JVM design and tuning.
• Good knowledge of Linux based environments.