We are seeking a highly skilled Kafka DevOps Engineer to take ownership of our Kafka service within a complex IT ecosystem. The ideal candidate will be responsible for implementing, maintaining, and optimizing Kafka infrastructure to ensure high availability, scalability, and performance. This role requires expertise in Apache Kafka, strong automation skills, and experience managing distributed data streaming platforms in an enterprise environment.
Key Responsibilities :
Kafka Administration & Operations: Deploy, configure, monitor, and maintain Kafka clusters in a high-availability production environment.
Performance Optimization: Tune Kafka configurations, partitions, replication, and producers/consumers to ensure efficient message streaming.
Infrastructure as Code (IaC): Automate Kafka infrastructure deployment and management using Terraform, Ansible, or similar tools.
Monitoring & Incident Management: Implement robust monitoring solutions (e.g., Dynatrace) and troubleshoot performance bottlenecks, latency issues, and failures.
Security & Compliance: Ensure secure data transmission, access control, and compliance with security best practices (SSL/TLS, RBAC, Kerberos).
CI/CD & Automation: Integrate Kafka with CI/CD pipelines and automate deployment processes to improve efficiency and reliability.
Capacity Planning & Scalability: Analyze workloads and plan for horizontal scaling, resource optimization, and failover strategies.
Collaboration: Work closely with development teams to support Kafka-based applications and ensure seamless data flow.
Training & Technical Support: Provide training and technical assistance to end users and stakeholders.
Documentation: Write and update procedures, contribute to the knowledge base, and ensure best practices are documented.
Teamwork: Work collaboratively with the team to ensure project success.
Required Skills & Experience :
5+ years of experience in DevOps, Site Reliability Engineering (SRE), or Kafka administration.
Strong hands-on experience with Apache Kafka (setup, tuning, and troubleshooting).
Proficiency in scripting languages like Python, Bash and automation tools like Terraform, Ansible.
Experience with cloud environments (AWS, Azure, or GCP) and Kubernetes-based Kafka deployments.
Familiarity with Kafka Connect, KSQL, Schema Registry, Zookeeper.
Knowledge of logging and monitoring tools (Dynatrace, ELK, Splunk).
Strong understanding of networking, security, and access control for Kafka clusters.
Experience with CI/CD tools (Jenkins, GitLab, ArgoCD).
Ability to analyze logs, debug issues, and propose proactive improvements.
ITIL certification is a plus.
Strong communication skills (oral and written) for interaction with internal and external stakeholders.
Ability to work in an international and multicultural environment.
Ability to work both independently and collaboratively.
Nice-to-Have :
Experience with Confluent Kafka or other managed Kafka solutions.
Knowledge of event-driven architectures and stream processing tools (Flink, Spark, Kafka Streams).
Experience with service mesh technologies (Istio, Linkerd) for Kafka networking.
Certifications in Kafka, Kubernetes, or cloud platforms.
Obtenez un examen gratuit et confidentiel de votre CV.
Sélectionnez le fichier ou faites-le glisser pour le déposer