Description et exigences
Position Summary:
A skilled Big Data (Hadoop) Administrator responsible for the installation, configuration, and maintenance of Cloudera Data Platform (CDP) and Cloudera Flow Management (CFM) streaming clusters on RedHat Linux. Proficiency in DevOps practices, scripting, and infrastructure-as-code for automating routine tasks and improving operational efficiency is desirable. Experience working with cross-functional teams, including application development, infrastructure, and operations, is preferred.
Job Responsibilities:
- Manages the design, distribution, performance, replication, security, availability, and access requirements for large and complex Big Data clusters.
- Designs and develops the architecture and configurations to support various application needs; implements backup, recovery, archiving, conversion strategies, and performance tuning; manages job scheduling, application release, cluster changes, and compliance.
- Identifies and resolves issues utilizing structured tools and techniques.
- Provides technical assistance and mentoring to staff in all aspects of Hadoop cluster management; consults and advises application development teams on security, query optimization, and performance.
- Writes scripts to automate routine cluster management tasks and documents maintenance processing flows per standards.
- Implement industry best practices while performing Hadoop cluster administration tasks.
- Works in an Agile model with a strong understanding of Agile concepts.
- Collaborates with development teams to provide and implement new features.
- Debugs production issues by analyzing logs directly and using tools like Splunk and Elastic.
- Address organizational obstacles to enhance processes and workflows.
- Adopts and learns new technologies based on demand and supports team members by coaching and assisting.
Education:
Bachelor’s degree in computer science, Information Systems, or another related field with 10+ years of IT and Infrastructure engineering work experience.
Experience:
10+ Years Total IT experience & 7+ Years relevant experience in Big Data database
Technical Skills:
- Big Data Platform Management: Big Data Platform Management: Expertise in managing and optimizing the Cloudera Data Platform, including components such as Apache Hadoop (YARN and HDFS), Apache HBase, Apache Solr, Apache Hive, Apache Kafka, Apache NiFi, Apache Ranger, Apache Spark, as well as JanusGraph and IBM BigSQL.
- Data Infrastructure & Security: Proficient in designing and implementing robust data infrastructure solutions with a strong focus on data security, utilizing tools like Apache Ranger and Kerberos.
- Performance Tuning & Optimization: Skilled in performance tuning and optimization of big data environments, leveraging advanced techniques to enhance system efficiency and reduce latency.
- Backup & Recovery: Experienced in developing and executing comprehensive backup and recovery strategies to safeguard critical data and ensure business continuity.
- Linux & Troubleshooting: Strong knowledge of Linux operating systems, with proven ability to troubleshoot and resolve complex technical issues, collaborating effectively with cross-functional teams.
- DevOps & Scripting: Proficient in scripting and automation using tools like Ansible, enabling seamless integration and automation of cluster operations. Experienced in infrastructure-as-code practices and observability tools such as Elastic.
- Agile & Collaboration: Strong understanding of Agile SAFe for Teams, with the ability to work effectively in Agile environments and collaborate with cross-functional teams.
- ITSM Process & Tools: Knowledgeable in ITSM processes and tools such as ServiceNow.
Other critical Requirement:
- Automation and Scripting: Proficiency in automation tools and programming languages such as Ansible and Python to streamline operations and improve efficiency.
- Analytical and Problem-Solving Skills: Strong analytical and problem-solving abilities to address complex technical challenges in a dynamic enterprise environment.
- Communication Skills: Exceptional written and oral communication skills, with the ability to clearly articulate technical and functional issues, conclusions, and recommendations to stakeholders at all levels.
- 24x7 Support: Ability to work in a 24x7 rotational shift to support Hadoop platforms and ensure high availability.
- Stakeholder Management: Prior experience in effectively managing both onshore and offshore stakeholders, ensuring alignment and collaboration across teams.
- Business Presentations: Skilled in creating and delivering impactful business presentations to communicate key insights and recommendations.
- Collaboration and Independence: Demonstrated ability to work independently as well as collaboratively within a team environment, ensuring successful project delivery in a complex enterprise setting.