Big Data DevOps Developer/Admin

  • Competitive
  • Hyderabad, Andhra Pradesh, India
  • Permanent, Full time
  • Experian Singapore
  • 21 Apr 19

Big Data DevOps Developer/Admin

Role Purpose:

As a key aide to both the IT Infrastructure and Development teams, you will help support existing systems 24x7 and responsible for administering current Big Data environments. The candidate will be responsible for managing BigData Cluster environments and w ill work with teammates to maintain, optimize, develop, and integrate working solutions for our big data tech stack . To support the product development process in line with the product roadmap for product maintenance and enhancement such that the quality of software deliverables maintains excellent customer relationships and increases the customer base.

If you have the skills and "can do" attitude, we would love to talk to you!


Major Job Functions:

  1. Responsibilities:
  • Responsible for implementation and ongoing administration of Hadoop infrastructure
  • Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments
  • Expert knowledge with delivering Big Data Cloudera Solutions in the cloud with AWS or Azure.
  • Hands-on day-to-day expert experience in administering a Cloudera cluster with Cloudera Manager, Cloudera Director, Cloudera Navigator
  • Working with data delivery teams to setup new Hadoop users. This job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, HBase and Yarn access for the new users
  • Cluster maintenance as well as creation and removal of nodes using tools like Cloudera Manager Enterprise, etc.
  • Performance tuning of Hadoop clusters and Hadoop MapReduce routines
  • Screen Hadoop cluster job performances and capacity planning
  • Monitor Hadoop cluster connectivity and security
  • Manage and review Hadoop log files, File system management and monitoring
  • HDFS support and maintenance
  • Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability
  • Collaborating with application teams to perform Hadoop updates, patches, version upgrades when required
  • General operational expertise such as good troubleshooting skills, understanding of system's capacity, bottlenecks, basics of memory, CPU, OS, storage, and networks
  • The most essential requirements are: They should be able to deploy Hadoop cluster, add and remove nodes, keep track of jobs, monitor critical parts of the cluster, configure name-node high availability, schedule and configure it and take backups
  • Solid Understanding on premise and Cloud network architectures
  • Additional Hadoop skills like Sentry, Spark, Kafka, Oozie, etc
  • Advanced experience with AD/LDAP security integration with Cloudera, including Sentry and ACL configurations
  • Ability to configure and support API and OpenSource integrations
  • Experience working with DevOps environment, developing solutions utilizing Ansible, etc.
  • Will collaborate and communication with all levels of technical and senior business management
  • Will require on-call 24X7 support of production systems on a rotation basis with other team members
  • Pro-actively evaluate evolving technologies and recommend solutions to business problems.

  1. Teamwork
  • To work as part of a team to achieve the departmental standards.
  • To work together with other departments and divisions within the organisation to identify any process improvements and improve standards, efficiency and profitability.


  1. Required Knowledge / Skills / Experience:
  • Typically requires a bachelor's degree (in Computer Science or related field) or equivalent.
  • 5+ years of Linux (Redhat) system administration
  • 3+ years of Hadoop infrastructure administration
  • Cloud Platforms IaaS/PaaS - Cloud solutions: AWS, Azure, VMWare
  • Kerberos administration skills
  • Experience with Cloudera distribution
  • Familiarity with open source configuration management and deployment tools such as Puppet or Chef and Shell scripting
  • Must have strong experience in DevOps tools like Ansible and HAProxy . Automating deployments and monitoring/alerting tasks using Ansible.
  • Working Knowledge of YARN, HBase, Hive, Spark, Flume, Kafka etc.
  • Strong Problem Solving and creative thinking skills
  • Effective oral and written communications
  • Experience working with geographically distributed teams
  • Bachelors or Master's degree in Computer Science or equivalent experience
  • Knowledge and understanding of the business strategy and use of back office applications.
  • Ability to adapt to multi-lingual and multicultural environment, additional language skills are a bonus.
  • Ability to handle conflicting priorities.
  • Ability to learn.
  • Adaptability.
  • Receptive to change.
  • Ability to communicate with business users at all levels
  • Analytical skills
  • Self-motivated and pro-active