Intermediate experience in a Hadoop production environment.
Must have intermediate experience and expert knowledge with at least 4 of the following:
Hands on experience with Hadoop administration in Linux and virtual environments.
Well versed in installing & managing distributions of Hadoop (Cloudera).
Expert knowledge and hands-on experience in Hadoop ecosystem components; including HDFS, Yarn, Hive, LLAP, Druid, Impala, Spark, Kafka, HBase, Cloudera Work Bench, etc.
Thorough knowledge of Hadoop overall architecture.
Experience using and troubleshooting Open Source technologies including configuration management and deployment.
Data Lake and Data Warehousing design and development.
Experience reviewing existing DB and Hadoop infrastructure and determine areas of improvement.
Implementing software lifecycle methodology to ensure supported release and roadmap adherence.
Configuring high availability of name-nodes.
Scheduling and taking backups for Hadoop ecosystem.
Data movement in and out of Hadoop clusters.
Good hands-on scripting experience in a Linux environment.
Experience in project management concepts, tools (MS Project) and techniques.
A record of working effectively with application and infrastructure teams.
Strong ability to organize information, manage tasks and use available tools to effectively contribute to a team and the organization.
Valid Class C Texas Driver’s License.
Makes independent recommendations.