Help to revolutionise a fast-moving industry with cutting-edge AI:
Our client is a globally recognised brand with deep-rooted expertise. They are heavily invested in leveraging AI to combine their domain expertise with SOTA techniques, solidifying their position as a leader in the field. You'll join a global team with a distributed set of skills including Research, Applied AI and Engineering.
They are seeking MLOps Engineers to help architect the future of communication through AI. This isn't just another engineering role – it's an opportunity to pioneer systems that transform how companies connect with their customers
What You’ll Be Doing
You'll be designing and optimising production-grade MLOps pipelines that bring cutting-edge Generative AI and LLMs from experimentation to real-world impact. Your expertise will directly influence how some of the world's leading brands enhance their strategies.
What You'll Build
Production-Ready GenAI Infrastructure : Design and deploy scalable MLOps pipelines specifically optimized for GenAI applications and large language models
State-of-the-Art Model Deployment : Implement and fine-tune advanced models like GPT and similar architectures in production environments
Hybrid AI Systems : Create solutions that integrate traditional ML techniques with cutting-edge LLMs to deliver powerful insights
Automated MLOps Workflows : Build robust CI/CD pipelines for ML, enabling seamless testing, validation, and deployment
Cost-Efficient Cloud Infrastructure : Optimize cloud resources to maximize performance while maintaining cost efficiency
Governance and Versioning Systems : Establish best practices for model versioning, reproducibility, and responsible AI deployment
Integrated Data Pipelines : Utilize Databricks to construct and manage sophisticated data and ML pipelines
Monitoring Ecosystems : Implement comprehensive monitoring systems to ensure reliability and performance
What You’ll Need
5+ years of hands-on experience in MLOps, DevOps, or ML Engineering roles
Proven expertise deploying and scaling Generative AI models (GPT, Stable Diffusion, BERT)
Proficiency with Python and ML frameworks (TensorFlow, PyTorch, Hugging Face)
Strong cloud platform experience (AWS, GCP, Azure) and managed AI/ML services
Practical experience with Docker, Kubernetes, and container orchestration
Databricks expertise, including ML workflows and data pipeline integration
Familiarity with MLflow, DVC, Prometheus, and Grafana for versioning and monitoring
Bachelor's or Master's degree in Computer Science, Engineering, or related field (or equivalent experience)
Fluency in written and spoken English
The Person We're Looking For
You're a builder at heart – someone who loves creating scalable, production-ready systems
You balance technical excellence with pragmatic delivery
You're excited about pushing boundaries in GenAI and LLM technologies
You can communicate complex concepts effectively to diverse stakeholders
You enjoy mentoring junior team members and elevating the entire technical organization
What Makes This Opportunity Special
You'll be working with a modern data stack designed to process large-scale information, automate analysis pipelines, and integrate seamlessly with AI-driven workflows. This is your chance to make a significant impact on projects that push the boundaries of AI-powered insights and automation in industry.