Oracle Cloud Infrastructure (OCI) is building the core cloud services that power modern enterprise applications, and the Generative AI Service team is focused on building and scaling systems that enable Large Language Models (LLMs) and agent-based AI applications to run reliably and efficiently in production environments.
Requirements
- Proficiency in one or more programming languages: Python, Go, Java, or C++.
- Familiarity with REST APIs, microservices, and cloud-native architectures.
- Experience with CI/CD workflows and container-based deployments (e.g., Docker, Kubernetes).
- Exposure to AI/ML infrastructure or LLM serving technologies.
- Interest in distributed systems, observability, and scalability challenges.
Responsibilities
- Contribute to backend services that support LLM inference and agent execution.
- Write clean, maintainable, and testable code in languages such as Python, Go, or Java.
- Participate in code reviews, design discussions, and cross-team collaboration.
- Assist in deploying and monitoring services using modern cloud-native tools.
- Learn and apply best practices for building reliable and performant distributed systems.
- developing and maintaining components of our LLM-serving and agentic AI infrastructure.
- implement scalable services, automate deployment processes, and support operational excellence in production environments.
Other
- BS in Computer Science or a related technical field.
- 3+ years of software development experience in backend or cloud environments.
- MS in Computer Science or related field.
- Willingness to learn from senior engineers and contribute to high-impact projects.
- Certain US customer or client-facing roles may be required to comply with applicable requirements, such as immunization and occupational health mandates.