We are seeking a highly skilled AI Engineer with expertise in LLMs, data-driven pipeline implementation, and real-time AI inference to develop and optimize AI models tailored for industrial applications.
Qualifications
5+ years of experience in AI/ML development, specializing in LLMs and NLP-based models.
Proficiency in Python, PyTorch, TensorFlow, and Hugging Face Transformers.
Experience designing and optimizing data pipelines using Apache Spark, Airflow, Kafka, or similar frameworks.
Strong understanding of vector search, RAG, prompt engineering, custom fine-tuning, and knowledge graph-based AI implementations.
Familiarity with multi-modal data integration (text, image, and sensor data).
Experience with containerized MLOps frameworks (Kubeflow, MLflow, TFX) and CI/CD for AI deployments.
Expertise in cloud AI services (AWS SageMaker, Azure ML, Google Vertex AI) and distributed training.
Experience deploying AI models at the edge using NVIDIA Jetson, TensorRT, OpenVINO, or Coral TPUs.
Provable experience in writing code for embedded GPUs and NPU/TPU accelerators, optimizing AI inference workloads for edge computing.
Knowledge of time-series forecasting, anomaly detection, and predictive maintenance models is a plus.
Preferred Qualifications
Experience in mining, industrial automation, or large-scale infrastructure projects.
Knowledge of real-time AI applications in mission-critical environments.
Familiarity with multi-agent AI systems and reinforcement learning.
Knowledge of Computer Vision techniques and image processing.
Soft Skills
Strong problem-solving mindset and ability to optimize AI solutions for industrial challenges.
Ability to work cross-functionally with engineers, data scientists, and business stakeholders.
Excellent communication skills in English and Portuguese (Spanish is a plus).
Responsibilities
Design, implement, and optimize LLM-based AI solutions for industrial and mining use cases.
Develop data-driven pipelines for processing, transforming, and analyzing large-scale operational data from IoT sensors, edge devices, and cloud platforms.
Fine-tune and deploy transformer-based architectures (GPT, BERT, Llama, T5, etc.) for domain-specific AI applications.
Implement real-time AI inference models at the edge and in the cloud to support mission-critical decision-making.
Optimize model performance, latency, and cost efficiency through techniques such as quantization, pruning, and distillation.
Collaborate with data engineers and DevOps teams to integrate AI models into production-grade environments using MLOps best practices.