DigiFocal IT Solutions Pvt Ltd
Website:
digifocal.in
Job details:
Key Responsibilities Install, configure, and manage Cloudera clusters using Cloudera Manager Administer Hadoop ecosystem components such as HDFS, YARN, Hive, Impala, and Spark Manage and support modern storage layers including Apache Ozone as an alternative to HDFS Implement and manage table formats like Apache Iceberg for large-scale analytical workloads Monitor cluster health, performance, and capacity planning Troubleshoot and resolve issues related to cluster operations, job failures, and performance bottlenecks Implement and manage data security using Kerberos, Apache Ranger, and Apache Atlas Perform upgrades, patches, and version migrations across Cloudera distributions (CDH/CDP) Automate administrative tasks using scripting (Shell/Python) Manage backup, replication, and disaster recovery strategies across storage layers (HDFS/Ozone) Collaborate with data engineering teams to optimize workloads on Iceberg tables Ensure compliance with enterprise data governance and security policies Required Skills & Qualifications Strong experience with Cloudera Distribution (CDH/CDP) Deep understanding of Hadoop ecosystem components Hands-on experience with Apache Ozone and/or object storage concepts Working knowledge of Apache Iceberg including schema evolution, partitioning, and time travel Strong Linux/Unix system administration skills Experience with cluster sizing, tuning, and performance optimization Knowledge of distributed systems and parallel processing Familiarity with cloud platforms like AWS, Azure, or GCP Experience with scheduling tools like Apache Oozie or Apache Airflow Strong scripting skills (Bash, Python) Preferred Qualifications Cloudera Certified Administrator (CCA/CCP) Experience transitioning from HDFS to Ozone-based architectures Experience implementing Iceberg on Spark, Hive, or Impala Exposure to Docker and Kubernetes Experience with Prometheus and Grafana Understanding of DevOps practices and CI/CD pipelines Key Competencies Strong troubleshooting and root cause analysis skills Ability to manage high-availability distributed systems Collaboration with cross-functional data teams Proactive monitoring and platform optimization mindset Nice to Have (Value Add) Exposure to streaming technologies like Apache Kafka Experience with data lakehouse architectures Knowledge of metadata/catalog systems and governance frameworks
Interested candidates kindly share resumes on shikha@digifocal.in
Click on Apply to know more.