The DevOps Engineer will play a mission-critical role owning the deployment, scalability, security, and reliability of AI systems and digital platforms. This role has a strong focus on
LLM deployments, AI workloads, and cloud-native infrastructure, ensuring that all AI and software systems operate with enterprise-grade availability, performance, and compliance.
Key Responsibilities
CI/CD & Automation Engineering
- Design, build, and maintain CI/CD pipelines for AI models, LLM services, and software applications
- Automate build, test, deployment, and environment configuration workflows to enable rapid and reliable releases
AI & LLM Deployment Operations
- Deploy, operate, and scale AI systems, LLM APIs, inference workloads, and cloud-based AI services
- Ensure high availability, horizontal scalability, and low-latency inference across all production environments
Infrastructure, Reliability & Cost Optimization
- Monitor infrastructure performance, system health, and AI workloads using observability and monitoring tools
- Optimize infrastructure for reliability, performance, and cloud cost efficiency
Security, Compliance & Governance
- Implement and enforce security best practices, access controls, secrets management, and environment isolation
- Ensure infrastructure and deployment processes align with national data governance, compliance, and cybersecurity standards
Cross-Functional Enablement
- Collaborate closely with AI Engineers, Full-Stack Engineers, and Product teams to enable seamless, scalable deployments
- Act as the primary technical owner for production reliability during mission-critical deployments
Documentation & Architecture Standards
- Maintain comprehensive documentation for DevOps workflows, system architecture, environments, and deployment standards
- Ensure operational readiness, auditability, and knowledge transfer across teams
Required Qualifications
- Minimum 5 years of hands-on DevOps engineering experience in production environments
- Mandatory: Proven experience deploying and operating AI systems and LLM-based workloads in production
- Strong hands-on expertise with Docker, Kubernetes, CI/CD platforms, and cloud services
- Experience with monitoring, observability, logging, and infrastructure-as-code (e.g., Terraform, similar tools)
- Strong understanding of networking, security, and cloud-native architecture principles
- Excellent troubleshooting and incident response capabilities in high-availability systems
Preferred Qualifications
- Experience with MLOps platforms such as MLflow, SageMaker, Vertex AI, or similar
- Proven experience scaling AI and LLM applications in high-traffic production environments
- Exposure to AI model lifecycle management, retraining pipelines, and operational governance
- Experience in government, regulated, or national-scale enterprise environments
KPIs & Deliverables
- Uptime, reliability, and stability of AI platforms and production systems
- Deployment speed, automation maturity, and release reliability
- Infrastructure performance, scalability, and cost optimization efficiency
- Security posture and compliance readiness across all environments
- Quality, completeness, and audit readiness of DevOps documentation and workflows