Responsibilities Operational management and architecture of Hadoop ecosystem, managing 100s 1000s of nodes globally Build out clusters in data centers around the world? Tuning multi-tenant Hadoop ecosystem for operational efficiency, balancing various workloads and optimizing Yarn and Impala accordingly Implement security, encryption, authentication, and authorization controls to adhere to corporate security policies Support Data Governance and data lineage on the cluster Enable High Availability and resiliency in the cluster, achieving 99.9999 uptime Understand network optimization and DR strategies Support and help to drive our hybrid cloud strategy, develop strategies for compute burst Work with data architects on the logical data models and physical database designs optimized for performance, availability and reliability Helping to tuning and optimization of backend and frontend data operations Serve as a query tuning and optimization technical expert, providing feedback to team Scripting and automation to support development, QA and production database environments, deployments to production and management of services and infrastructure Mentors development team members Proactively helps to resolve difficult technical issues Provide technical knowledge to teams during project discovery and architecture phases Keep management informed of work activities and schedules Assess new initiatives to determine the work effort and estimate the necessary time-to- completion Document new development, procedures or test plans as needed Participate in data builds and deployment efforts. Help mature our Continuous Integration and Continuous Deployment methodologies Participate in projects through various phases Performs other related duties as assigned Partner with the business units to develop effective solutions that solve business challenges Requirements Cloudera CDP CM API Auto TLS - trusted certs Encryption at Rest Horton-Works HDP Experience with Atlas Ranger policies setup Installation and configuration of the Ranger-KMS Migration Upgrades from HDP to CDP Hadoop, Hive, Impala, HBase and related technologies SparkSpark2 MPP, shared nothing database systems, NoSQL systems Object Oriented and Functional Programming Experience Excellent knowledge of Linux, AIX, or other Unix flavors Experience with scripting (Bash, Linux scripting) Data Warehousing design and concepts Exposure to Infrastructure as Code (Ansible, Terraform) About Matlen Silver Matlen Silver is the hardest working staffing team in the U.S. We do what we know is right for consultants and companies, creating a unique and powerful recruiting and talent experience. When the rubber meets the road, Matlen Silver is the powerhouse that cuts through the nonsense and gets you the job you want and deserve. If you are a rock star go-getter with a proven track record of success, put us to the test.