Haldorix, a startup studio focused on turning industrial challenges into scalable AI ventures, is looking for a highly skilled MLOps Engineer – Edge AI Specialist to join our team in Casablanca. The position is a full‑time CDI with a hybrid working model, offering a competitive salary range of 18 000 MAD to 25 000 MAD per month.
Key Responsibilities:
- Architecture & Infrastructure: Design and deploy a hybrid edge/cloud architecture optimized for real‑time video analytics. Define hardware specifications (Jetson Orin, RTX A2000, Intel NUC) and ensure reliable communication between edge servers and the cloud.
- Model Optimization: Convert and optimize deep learning models for embedded GPUs using ONNX Runtime and TensorRT. Apply quantization (INT8, FP16) and pruning techniques to reduce latency and memory footprint.
- MLOps Pipeline: Build and maintain a CI/CD pipeline tailored for edge deployment – containerized models, version control, automated OTA updates, and proactive performance monitoring.
- Orchestration & Deployment: Deploy and manage fleets of edge servers using K3s/MicroK8s. Implement declarative deployments (ArgoCD/Flux) and centralized management via KubeEdge or AWS IoT Greengrass.
- Security & Compliance: Enforce full data locality, end‑to‑end encryption (TLS/mTLS), and anonymization pipelines to ensure GDPR compliance.
- Monitoring & Reliability: Set up comprehensive dashboards (Prometheus, Grafana, Loki) to track inference performance, GPU utilization, and uptime (>99%).
- LLM Integration: Support deployment of a centralized LLM server (Claude, GPT‑4, or open‑source) powering RAG‑based analytics and real‑time conversational interfaces for clients.
- Field Operations: Conduct on‑site installations, validations, and troubleshooting sessions with client teams. Train local technicians and maintain up‑to‑date documentation for reproducibility and scalability.
Required Qualifications:
- 3–5 years of experience deploying AI models in production.
- Strong expertise in MLOps, edge computing, and embedded GPU environments.
- Proven track record with TensorRT, ONNX Runtime, quantization (INT8/FP16), and model pruning.
- Proficiency in Python (PyTorch, TensorFlow, FastAPI) and DevOps tools (Docker, CI/CD, Ansible).
- Solid understanding of Kubernetes/K3s, networking, and Linux administration.
- Experience with Prometheus, Grafana, and GPU performance profiling.
- Excellent documentation and troubleshooting skills.
Nice to Have:
- Familiarity with NVIDIA Jetson and other embedded AI hardware.
- Experience with Fleet Management Systems (AWS IoT Greengrass, KubeEdge, Balena).
- Knowledge of Stable Diffusion and LLM pipelines (RAG, Pinecone, Weaviate, ChromaDB).
- Background in industrial computer vision, IoT, or real‑time systems.
- Understanding of GDPR compliance and data anonymization for on‑prem AI systems.
Benefits:
- Join a startup studio scaling high‑impact AI ventures from prototype to production.
- Work on cutting‑edge Edge AI systems deployed across industrial sites.
- Collaborate with an agile, expert team blending AI, hardware, and DevOps engineering.
- Gain hands‑on experience with inference optimization, GPU benchmarking, and large‑scale orchestration.
- Be part of a project delivering tangible cost and performance breakthroughs in manufacturing AI.
Recruitment Process:
- Jobzyn AI interview (25–45 min).
- Technical interview (1 h) with the Lead Developer or Technical Architect.
- Practical test (2–3 h) simulating a real‑world MLOps deployment case.
- Final interview with the NITRA team and Haldorix partners.
We look forward to receiving your application and exploring how you can contribute to our mission of transforming industrial challenges into scalable AI solutions.