We are seeking a Big Data Administrator to manage and support enterprise big data platforms. The role ensures platform
availability, performance, security, and stability
across production and non-production environments, working closely with data engineering, IAM, network, security, and infrastructure teams.
Key Responsibilities
-
Deploy, configure, and administer Cloudera CDP/Hadoop clusters using Cloudera Manager
-
Operate and support Hadoop ecosystem services: HDFS, YARN, Spark, Hive/Impala, HBase, Zookeeper
-
Monitor platform health and performance; perform capacity planning and performance tuning
-
Implement and manage security: Kerberos, Ranger policies, LDAP/AD integration, and TLS/SSL hardening
-
Troubleshoot production incidents (authentication issues, service failures, disk/log issues, stuck Spark/YARN jobs) and perform root cause analysis
-
Manage upgrades, patching, and configuration changes with minimal downtime
-
Build automation using Bash/Python for routine tasks and operational efficiency
-
Administer Kubernetes platforms (Red Hat OpenShift): basic cluster operations, namespace/project management, pod/service troubleshooting, log analysis, and support for platform applications
-
Maintain operational documentation, runbooks, and support procedures; participate in on-call support as required
Requirements
3+ years
of hands-on experience as a
Cloudera/Hadoop Administrator
in production.
-
Strong Linux administration and command-line skills
-
Proven experience with Cloudera Manager and cluster operations
-
Working knowledge of Kubernetes/OpenShift administration (oc/kubectl, pods, deployments, services, logs, troubleshooting)
-
Solid understanding of Kerberos, Ranger, LDAP/AD, and TLS/SSL
-
Experience with monitoring tools such as Grafana/Prometheus (or similar)
Strong troubleshooting, communication, and collaboration skills