We are looking for a Principal Solution Engineer to join our Ecosystem Engineering team, focusing on the optimization of AI/ML models inference and serving. In this role, you will identify, build, and optimize emerging use-case patterns within the AI and vertical industries. These patterns will leverage our products, partner offerings, and open source projects deployed on our portfolio, with a specific emphasis on model serving, inference, and MLOps workflows. You will collaborate closely with our engineering and product management teams, using insights gained from interactions with partners and customers to influence product adoption and development. As part of a geographically distributed team, you will engage with multiple engineering teams and open source communities globally. Success in this role requires strong motivation, curiosity, a passion for problem-solving, and hands-on experience with Linux technologies and open source.
What you will do:
Identify emerging patterns for applying our offerings to business problems
Discover and describe what differentiates our solutions from competitive alternatives by working directly with key partners and customers who deploy and operate our solutions
Create reference architectures for optimized AI workloads running on/with our portfolio
Create and conduct on-demand demo labs, providing an initial practical experience for key partners and customers
Provide technical vision and leadership on critical and high impact projects
Communicate and promote the results of reference architectures globally through publication of blogs and speaking at webinars and conferences
Contribute to a culture of continuous improvement by sharing recommendations and technical knowledge.
Requirements: Passionate about technology and continuous learning; an innovator with the ability to quickly master new tools and technologies and work independently.
Technical leadership acumen
Advanced experience with Kubernetes, OpenShift, or other cloud-native technologies
Experience with AI and Machine Learning platforms, tools, and frameworks, such as LangChain/LangGraph, PyTorch, vLLM, MCP, and Kubeflow
Ability to work on your own in a fast-paced, ever-changing environment
Excellent written and verbal communication skills, with proven experience in technical leadership, publishing content, or presenting at industry events
Passion for open source and community-based software development models
Bachelor's degree in a technical field or equivalent experience
The following will be considered a plus:
Bachelor's degree in statistics, mathematics, computer science, operations research, or a related quantitative field, or equivalent expertise; Masters or PhD is a big plus
Experience in a customer-facing role, such as: solution architecture or consulting, focused on deploying complex AI/ML solutions on Kubernetes.
Advanced development experience in Python or Go.
Familiarity with AI/ML services across major public clouds (AWS, Azure, GCP) and/or hardware accelerators (CUDA, ROCm).
Experience working with automation tools/frameworks (e.g., Ansible, GitOps) and with MLOps/LLMOps platforms
Knowledge and interest in developing tools and solutions using Agentic workflows.
This position is open to all candidates.