As an MLOps Engineer, you will work at the intersection of AI enablement and production-grade operations-helping teams run inference services that are dependable, scalable, and secure. Define and operate validation gates covering functional checks, performance benchmarks, and safety/ security requirements before production rollout. * Scalability, Performance & Efficiency: Tune inference services to optimize latency, throughput, batching, GPU utilization, and multi-tenant fairness. We are looking for a colleague who can confidently operate modern infrastructure and understands what it takes to run AI inference services in production. * Experience with web service deployment, such as reverse proxies, TLS certificates, load balancing, API endpoints - In addition, our company medical service and an experienced social counseling team are available to assist you on site
mehr