About the Role
You will work closely with ML engineers, data scientists, and DevOps teams to support model training, deployment, and high-performance AI operations.
Responsibilities
- Design, implement, and manage cloud and on-prem infrastructure for AI/ML training and inference.
- Build and maintain GPU/TPU clusters, high-performance computing environments, and distributed systems.
- Develop and optimise data pipelines to support large-scale model training.
- Work closely with ML engineers to deploy and operationalise AI models in production.
- Improve system performance, reliability, and efficiency for AI workloads.
- Automate infrastructure using IaC and CI/CD tools.
Requirements
- Strong experience with cloud platforms (AWS, GCP, Azure) and GPU-based compute.
- Proficiency with Linux, scripting (Python/Bash), and infrastructure automation.
- Hands-...
Ready to Apply?
Submit your application today and take the next step in your career journey with Odiin.AI.
Apply Now