Procom-posted 8 months ago
Remote • San Antonio, TX
Professional, Scientific, and Technical Services

The Hadoop Administrator-Cloudera is responsible for managing and maintaining the Hadoop ecosystem to support enterprise data platform initiatives. This role involves ensuring system reliability, security, and performance while collaborating with various IT teams to support long-term data strategies.

  • Maintain and support Hadoop clusters and related ecosystem tools (HDFS, YARN, Hive, LLAP, Druid, Impala, Spark, Kafka, HBase, Cloudera Workbench, etc.)
  • Perform installations, patching, upgrades, and configuration of Hadoop software and components
  • Monitor system performance, conduct tuning, and manage storage and capacity planning
  • Manage high availability configurations, backup scheduling, and data recovery solutions
  • Ingest and manage data flows into Hadoop using tools like Sqoop and Flume
  • Administer and troubleshoot Hadoop security configurations, including Kerberos integration
  • Design and implement data lakes and data warehousing solutions
  • Support job scheduling and workflow management within the ecosystem
  • Collaborate across IT teams to support ERP and data-driven initiatives
  • Participate in on-call rotation for 24/7 support, including occasional work outside business hours
  • Ensure compliance with industry standards, best practices, and regulatory requirements
  • Bachelor's degree in Computer Science, Information Systems, Engineering, or related field
  • Intermediate-level experience in Hadoop administration in a production environment
  • Proven ability to manage Cloudera Hadoop platforms in Linux and virtual environments
  • Strong understanding of Hadoop architecture and ecosystem components
  • Hands-on experience with at least 4 of the following areas: Cloudera distributions and component management, Data lake and warehouse design, Scripting in Linux, Data movement strategies (Sqoop, Flume), System performance tuning and storage optimization, Backup/recovery configuration, High availability (NameNode configuration)
  • Familiarity with DevOps tools and configuration management in open-source environments
  • Strong organizational and task management skills
  • Excellent communication and problem-solving abilities
  • Experience in project management concepts, tools (MS Project), and techniques
  • A record of working effectively with application and infrastructure teams
  • Good hands-on scripting experience in a Linux environment
  • Experience reviewing existing DB and Hadoop infrastructure and determining areas of improvement
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service