דרושים » הנדסה » DevOps Engineer-IDP Team

משרות על המפה
 
בדיקת קורות חיים
VIP
הפוך ללקוח VIP
רגע, משהו חסר!
נשאר לך להשלים רק עוד פרט אחד:
 
שירות זה פתוח ללקוחות VIP בלבד
AllJObs VIP
כל החברות >
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a DevOps Engineer-IDP Team
What Youll Do:
Streamline Delivery Build and run self-service pipelines with our internal Platform Orchestrator, which uses ArgoCD & Argo Workflows for build, test ,deploy, validate, and scan.
Elevate DevEx Grow Backstage into the single entry point for engineers through catalogs, golden-path templates, and clear docs.
Software Rollout Strategies - Define safe rollout strategies (blue/green, canary, Argo Rollouts) and automate promotions across environments.
Incident-Ready Platform Participate in platform on-call, drive post-mortems, and turn findings into hardening and auto-healing features.
Automate Everything Craft Python/Go tools and CLIs that eliminate toil and bake best practices into everyday workflows.
Requirements:
5+ years as a DevOps / Platform / Infrastructure engineer in a tech company.
Proven success designing and scaling internal CI/CD orchestration systems leveraging CICD tools.
hands-on expertise with Argo CD, Argo Workflows, Backstage, and Kubernetes.
Strong coding skills in Python (bonus: Go or TypeScript).
Solid understanding of containerization, IaC , k8s operators and cloud-native observability (Prometheus, Grafana, OpenTelemetry).
A true platform-as-a-product mindset: you measure success by the productivity and happiness of fellow engineers.
Excellent communicationclear docs, engaging workshops, and persuasive storytelling.
Nice to Have:
Service mesh (Istio, Linkerd) and progressive delivery (Argo Rollouts, Flagger).
Experience migrating traditional pipelines to Git-Ops workflows.
Backstage plug-in development or broader CNCF toolchain familiarity (Crossplane, Helm, Kustomize).
This position is open to all candidates.
 
Hide
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8326566
סגור
שירות זה פתוח ללקוחות VIP בלבד
משרות דומות שיכולות לעניין אותך
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a Senior DevOps Engineer to join our R&D team in developing the next rising product in the health tech landscape. If you are looking for a challenging, influential position and are passionate about making an impact, this might be the role for you.
As a Senior DevOps Engineer , youll play a key role in the design, development, testing, deployment, and monitoring of our infrastructure and products. In this position, you'll make significant contributions to our observability stack, helping build and maintain robust systems for logs, metrics, traces, and alerting.
Our ideal candidate is passionate about DevOps and observability, has strong communication skills, and thrives on constant improvement for both technology and processes. If you enjoy working on multiple projects in parallel and are a proactive team player, youll fit right in.
This is a unique opportunity to join the core team of a fast-growing startup, where your contributions will have a direct impact on our product and success.
Responsibilities:
Support and collaborate with cross-functional engineering teams using cutting-edge technologies.
Contribute to the design, implementation, and maintenance of monitoring, logging, and alerting systems (e.g., Prometheus, Grafana, Loki)
Secure, scale, and manage our cloud environments (AWS and GCP)
Design and implement automation solutions for both development and production
Manage and improve our CI/CD pipelines for fast and safe delivery
Lead best practices in infrastructure, observability, configuration management, and system hardening
Continuously assess and improve existing infrastructure in line with industry standards
Requirements:
BSc in Computer Science, Engineering, or equivalent experience
5+ years of experience as a DevOps Engineer or similar software engineering role
Proven experience with Docker and Kubernetes (EKS preferred)
Hands-on experience with monitoring and observability tools, including Prometheus, Grafana, Datadog, or similar.
Expertise in Terraform for AWS infrastructure-as-code deployments
Strong collaboration and interpersonal communication skills
Excellent analytical thinking and problem-solving mindset
Proficiency with relational databases
Solid knowledge of Python and Bash scripting
Experience with test automation an advantage
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8320472
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
05/08/2025
Location: Tel Aviv-Yafo
Job Type: Full Time
As a Principal DevOps Engineer in our Platform Engineering team, you will lead the design and implementation of cutting-edge CI/CD pipelines and cloud architecture that powers our development environment. You'll drive initiatives to enhance developer productivity through automation, tooling, and infrastructure improvements, working with a modern tech stack including Kubernetes, Python, cloud-native and high-scale technologies.
Your Impact
Architect and implement scalable, resilient CI/CD pipelines and cloud infrastructure that supports our engineering organization's evolving needs
Design and develop internal developer tools and platforms that significantly improve developer experience and productivity
Drive the evolution of our Kubernetes-based deployment infrastructure in Google Cloud Platform, ensuring security, reliability and performance
Optimize and scale our CI/CD infrastructure including Jenkins, GitLab, TeamCity, and artifact management systems
Mentor and guide other engineers on DevOps best practices, infrastructure design, and implementation strategies
Drive adoption of infrastructure-as-code, automated testing, and deployment methodologies
Collaborate with development teams to understand their needs and implement solutions that accelerate their workflow
Establish standards and best practices for infrastructure reliability, observability, and performance.
Requirements:
7+ years of experience in DevOps, Site Reliability Engineering, or Platform Engineering roles
Extensive experience with CI/CD pipeline design and implementation in complex environments
Advanced knowledge of Kubernetes administration, deployment patterns, and ecosystem tools
Strong programming skills in Python with solid understanding of OOP principles and design patterns
Deep understanding of cloud architecture, specifically with Google Cloud Platform services
Proven track record designing and implementing developer tooling and automation
Experience managing containerized applications and services in production environments
Strong system design skills with focus on scalability, reliability, and security
Knowledge of GitOps workflows and infrastructure-as-code using tools like Terraform, Pulumi, or equivalent
Familiarity with GitLab CI administration and pipeline development
participate in an on call rotation for working and non-working hours
Nice-to-Have
Knowledge of observability platforms and practices (Prometheus, Grafana, distributed tracing)
Familiarity with TeamCity administration and pipeline development
Experience implementing security best practices in CI/CD pipelines
Understanding of compliance requirements in software delivery pipelines
Experience with Infrastructure as Code testing frameworks
Knowledge of software architecture patterns and microservices design.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8290390
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
1 ימים
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a DevOps DevOps Engineer to take ownership of our Cloud Infrastructure and Platform Engineering strategy, enabling high-scale, cutting-edge GenAI products running across 40+ Kubernetes clusters on GCP and AWS.
This role is a hands-on engineering , requiring deep expertise in cloud-native technologies, Kubernetes at scale, and modern DevOps principles. You will work closely with engineering teams to design and implement scalable infrastructure solutions, optimize developer workflows, and ensure reliability and efficiency across our platform.
Role and Responsibilities:
Cloud & Kubernetes Expertise: Design and implement highly scalable multi-cluster Kubernetes environments across GCP & AWS.
Developer Experience & Enablement: Lead the development of self-service tools and automation that improve efficiency for R&D teams.
Incident & Reliability Engineering: Work with engineering teams to optimize cost, performance, and reliability of production infrastructure through monitoring, capacity planning, and scaling strategies.
Security & Governance: Contribute to best practices for RBAC, IAM, cloud security, and compliance while ensuring infrastructure reliability.
Automation & Infrastructure as Code: Drive adoption of GitOps workflows and Infrastructure as Code (Terraform, Helm, Crossplane) to enhance automation and consistency.
Mentorship & Team Growth: Provide technical mentorship within the platform engineering team and contribute to knowledge-sharing across R&D.
Cross-Team Collaboration: Work closely with engineering teams to align cloud infrastructure goals with business needs and reliability requirements.
Requirements:
5+ years of DevOps, or SRE experience
3+ years working with public cloud platforms (AWS, GCP) at scale
Deep Kubernetes expertise, including managing large-scale, multi-cluster enterprise-grade Kubernetes environments
Experience designing and managing Custom Resource Definitions (CRDs) and custom controllers
Strong background in Infrastructure as Code (Terraform, Helm) and GitOps principles (ArgoCD, Crossplane, FluxCD, etc.)
Hands-on experience in observability & monitoring (Prometheus, Grafana, Datadog, OpenTelemetry, etc.)
Proficiency in scripting & automation (Python, Go, Bash) for infrastructure automation
Expertise in cloud networking (VPC, load balancers, service meshes) and security best practices (RBAC, IAM, security groups, network policies, etc.)
Experience with CI/CD pipelines, optimizing for performance, security, and developer velocity
Nice-to-Have:
Experience with self-hosted on-prem deployments and managed private VPC deployments (Bring Your Own Cloud models)
Advanced expertise in Helm and Crossplane for Kubernetes resource management.
Other cloud provider experience
Experience in GenAI or large-scale SaaS platforms
Familiarity with SQL/NoSQL databases and distributed systems
DevSecOps experience, with a strong understanding of security automation and compliance frameworks
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8326421
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
20/08/2025
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Were looking for a Senior DevOps Engineer to join our newly formed Foundations Teama small, high-impact group responsible for the infrastructure, tools, and shared services that power our entire R&D organization.
In this role, youll design, build, and evolve internal platform infrastructure, CI/CD systems, and developer enablement tooling. Your mission is to empower developers across the company to work autonomously, by creating self-service tools, automation, and clear standards that reduce friction and increase reliability.
Youll collaborate closely with engineers across disciplines and partner with the Foundations Team Lead to shape DevOps practices that scale. This is a hands-on role for someone who thrives in high-velocity, mission-critical environments and is passionate about building tools that make developers faster, more productive, and confident in running their own services.
What Youll Do
Design and maintain scalable, developer-friendly CI/CD pipelines and deployment workflows.
Build self-service tooling and automation that enables teams to manage deployments, environments, secrets, and observability independently
Be responsible for cloud infrastructure and operations foundations
Implement and promote best practices for monitoring, logging, and alerting across services.
Operate and optimize Kubernetes-based production environments, ensuring performance, security, and stability.
Manage infrastructure using Infrastructure as Code (IaC) and ensure repeatability and traceability through tools like Terraform.
Collaborate with R&D teams to support onboarding to internal tooling and promote a culture of enablement over dependency.
Monitor cloud cost, ensuring our cloud operates efficiently.
Requirements:
4+ years of hands-on experience in DevOps or infrastructure engineering, ideally in high-velocity, mission-critical production environments.
Deep expertise in Kubernetes and containerized infrastructure, with experience deploying and managing workloads at scale.
Strong understanding of cloud infrastructure and operations, including networking, storage, compute, and securityGCP experience preferred.
Proficiency with Infrastructure as Code tools, especially Terraform, with a focus on automation and operational excellence.
Experience developing and managing CI/CD processes and tools, with a passion for improving developer workflows and release quality.
Strong debugging and problem-solving skills, with the ability to troubleshoot complex systems across the stack.
Highly self-motivated and organized, able to work independently in a fast-paced, collaborative environment.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8311657
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are seeking an experienced cloud solution architect (5+ years) to design, implement, and manage cloud infrastructure and CI/CD pipelines for big data solutions using Kubernetes and AWS. If you are passionate about DevOps and big data, wed love to hear from you! Availability to work from the Jerusalem office at least one day a week possibly more during the initial onboarding period.
What will your job look like?
Design and manage cloud infrastructure on AWS using IaC tools like Terraform or CloudFormation.
Optimize Kubernetes clusters for big data workloads.
Develop and maintain CI/CD pipelines using tools like ArgoCD and GitLab CI/CD.
Implement monitoring and logging solutions for big data systems.
Ensure compliance with security best practices.
Collaborate with development and operations teams.
Troubleshoot and resolve complex technical issues.
Requirements:
5+ years of DevOps experience, focusing on big data.
Expertise in Kubernetes, AWS, and big data technologies (e.g., Spark, Athena, Argo Workflows).
Proficiency in scripting (Bash, Python, SQL).
Experience with CI/CD tools, containerization (Docker), and configuration management (e.g., Ansible, Puppet).
Strong problem-solving and communication skills.
Preferred Qualifications:
Certifications in AWS or Kubernetes.
Familiarity with cloud-native technologies (Helm, Operators).
Knowledge of security best practices for cloud and big data.
Experience with monitoring tools (Prometheus, Grafana).
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8316102
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
This is a great opportunity to be part of one of the fastest-growing infrastructure companies in history, an organization that is in the center of the hurricane being created by the revolution in artificial intelligence.
"our company's data management vision is the future of the market."- Forbes
we are the data platform company for the AI era. We are building the enterprise software infrastructure to capture, catalog, refine, enrich, and protect massive datasets and make them available for real-time data analysis and AI training and inference. Designed from the ground up to make AI simple to deploy and manage, our company takes the cost and complexity out of deploying enterprise and AI infrastructure across data center, edge, and cloud.
Our success has been built through intense innovation, a customer-first mentality and a team of fearless company ronauts who leverage their skills & experiences to make real market impact. This is an opportunity to be a key contributor at a pivotal time in our companys growth and at a pivotal point in computing history.
The DevOps Engineer position is an operational engineering role and is an integrated part of our development team. You will be responsible for improving the efficiency of our processes, software, and infrastructure, and will be assisting RnD Team with product development. If you are DevOps Engineer that is passionate about automating and scaling everything, this job is for you.
Responsibilities
Monitor and optimize cloud infrastructure for performance, scalability, and cost-efficiency.
Manage and Maintain CI Infrastructure (GitLab CI and Jenkins).
Manage, Maintain and Improve our Release and Development Environments.
Support critical production infrastructure deployed in Multiple Clouds (AWS, Azure, and GCP).
Develop and Support RnD toolchain and implement best practices for code deployment, testing, and maintenance.
Automate On-Premises Labs Infrastructure by adopting IaC practices.
Lead and Develop Monitoring, Telemetry, Alerting, and Logging Production services.
Requirements:
Desired Qualifications:
Proven hands-on experience with Docker and Kubernetes in production. Hands-on experience deploying and managing complex Kubernetes environments, including services, ingresses, load balancers, and Helm charts
Solid understanding of Linux/Unix Internals and experience with handling complex performance and configuration problems in Linux/Unix environment.
Multi-Cloud Expertise: Deep familiarity with both GCP and AWS for provisioning, networking, and cost-optimization strategies
Experience in DSL Configuration tools like Ansible, Chef, or Puppet.
Experienced with programming languages (Python is preferred).
Shell scripting experience.
Proficient in SRE\Monitoring methodologies (Monitoring stacks with emphasis on Prometheus)
Nice To Have Skills
Experienced with CI/CD tools and frameworks.
Experience with managing binary repositories (RPMs, Pypi, NPM and etc)
Experience with developing Ansible collections, roles, and modules.
Experience with managing GitLab and GitLab CI.
Experience with Hashicorp Products: Terraform, Packer, Consul, Vault, and Vagrant.
Experience with automating configuration and deployment of On-Premises Lab Hardware.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8325791
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time
were looking for a DevOps engineer to take the lead on one of the most critical technical challenges in our business: how we deploy software at scale.
Today, deploying our core system into each store is a complex, multi-stage process. Tomorrow, it needs to be seamless, automated, and capable of onboarding dozens of stores per weekwithout sacrificing versatility and quality
This role isnt just about CI/CD or scripting. Its about refining the automation infrastructure that enables repeatable, self-service deployments across hundreds of live, mission-critical environments. Youll sit at the heart of our Ops Technology team, working at the intersection of system engineering and in-store execution, and serving as the technical backbone for deployment scale.
You'll help shape and evolve our DevOps toolsetworking hands-on with cutting-edge technologies to streamline deployments, boost reliability, and scale our platform with speed and confidence. For the right person, this role is a stepping stone toward technical leadership in one of most strategic teams.
A day in the life:
Design and build the software deployment framework that powers in-store systems at scale
Lead remote deployments using Octopus Deploy, driving repeatability and automation
Set up and manage Kubernetes clusters for scalable microservices in diverse environments
Deploy and monitor services with a focus on resilience, observability, and recovery
Troubleshoot complex issues across CI/CD, environments, services, and infrastructure
Collaborate with System Engineering and SRE peers to ensure smooth end-to-end deployment flows
Continuously evolve our CI/CD pipelines, deployment logic, and infrastructure-as-code practices
Build tooling, templates, and documentation to enable fast, low-touch deployments by others
Serve as a technical leader and force multiplier within the Ops Tech team
Requirements:
3+ years in DevOps, Infrastructure, or SRE roles with deep end-to-end ownership
Solid background in Kubernetes, container orchestration, and microservices
Experience deploying and supporting systems in live production environments
Strong CI/CD skills with tools like GitHub Actions, GitLab CI, or Jenkins
Scripting proficiency in Bash, PowerShell, or Python
Familiarity with monitoring, alerting, and diagnostics (Prometheus, Grafana, etc.)
Experience with infrastructure-as-code tools like Terraform, Helm, etc.
Excellent troubleshooting skills and a bias toward automation and scale
Strong communication and the ability to work cross-functionally and independently
Nice to have:
Experience with Azure and GCP
Understanding of K8s networking, service meshes, ingress controllers
Passion for enabling othersthrough tools, documentation, or mentorship
Proven ability to drive change in fast-moving, operationally complex environments
Expertise with Octopus Deploy and experience deploying software remotely
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8325803
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time
We are seeking an experienced Solutions Data Engineer who possess both technical depth and strong interpersonal skills to partner with internal and external teams to develop scalable, flexible, and cutting-edge solutions. Solutions Engineers collaborate with operations and business development to help craft solutions to meet customer business problems.
A Solutions Engineer works to balance various aspects of the project, from safety to design. Additionally, a Solutions Engineer researches advanced technology regarding best practices in the field and seek to find cost-effective solutions.
Job Description:
Were looking for a Solutions Engineer with deep experience in Big Data technologies, real-time data pipelines, and scalable infrastructuresomeone whos been delivering critical systems under pressure, and knows what it takes to bring complex data architectures to life. This isnt just about checking boxes on tech stacksits about solving real-world data problems, collaborating with smart people, and building robust, future-proof solutions.
In this role, youll partner closely with engineering, product, and customers to design and deliver high-impact systems that move, transform, and serve data at scale. Youll help customers architect pipelines that are not only performant and cost-efficient but also easy to operate and evolve.
We want someone whos comfortable switching hats between low-level debugging, high-level architecture, and communicating clearly with stakeholders of all technical levels.
Key Responsibilities:
Build distributed data pipelines using technologies like Kafka, Spark (batch & streaming), Python, Trino, Airflow, and S3-compatible data lakesdesigned for scale, modularity, and seamless integration across real-time and batch workloads.
Design, deploy, and troubleshoot hybrid cloud/on-prem environments using Terraform, Docker, Kubernetes, and CI/CD automation tools.
Implement event-driven and serverless workflows with precise control over latency, throughput, and fault tolerance trade-offs.
Create technical guides, architecture docs, and demo pipelines to support onboarding, evangelize best practices, and accelerate adoption across engineering, product, and customer-facing teams.
Integrate data validation, observability tools, and governance directly into the pipeline lifecycle.
Own end-to-end platform lifecycle: ingestion → transformation → storage (Parquet/ORC on S3) → compute layer (Trino/Spark).
Benchmark and tune storage backends (S3/NFS/SMB) and compute layers for throughput, latency, and scalability using production datasets.
Work cross-functionally with R&D to push performance limits across interactive, streaming, and ML-ready analytics workloads.
Requirements:
24 years in software / solution or infrastructure engineering, with 24 years focused on building / maintaining large-scale data pipelines / storage & database solutions.
Proficiency in Trino, Spark (Structured Streaming & batch) and solid working knowledge of Apache Kafka.
Coding background in Python (must-have); familiarity with Bash and scripting tools is a plus.
Deep understanding of data storage architectures including SQL, NoSQL, and HDFS.
Solid grasp of DevOps practices, including containerization (Docker), orchestration (Kubernetes), and infrastructure provisioning (Terraform).
Experience with distributed systems, stream processing, and event-driven architecture.
Hands-on familiarity with benchmarking and performance profiling for storage systems, databases, and analytics engines.
Excellent communication skillsyoull be expected to explain your thinking clearly, guide customer conversations, and collaborate across engineering and product teams.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8325726
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
1 ימים
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for an exceptional MLOps Team Lead to own, build, and scale the infrastructure and automation that powers state-of-the-art Large Language Models (LLMs) and AI systems.
This is a technical leadership role that blends hands-on engineering with strategic vision. You will define MLOps best practices, build high-performance ML infrastructure, and lead a world-class team working at the intersection of AI research and production-grade ML systems.
You will work closely with LLM Algorithm Researchers, ML Engineers, and Data Scientists to enable fast, scalable, and reliable ML workflows covering everything from distributed training to real-time inference optimization.
If you have deep technical expertise, thrive in high-scale AI environments, and want to lead the next generation of MLOps, we want to hear from you.
Requirements:
3+ years of experience in MLOps, ML infrastructure, or AI platform engineering.
2+ years of hands-on experience in ML pipeline automation, large-scale model deployment, and infrastructure scaling.
Expertise in deep learning frameworks (like PyTorch, TensorFlow, JAX) and MLOps platforms (like Kubeflow, MLflow, TFX).
Proven track record of building production-grade ML systems that scale to billions of predictions daily.
Deep knowledge of Kubernetes, cloud-native architectures (AWS/GCP), and infrastructure as code (Terraform, Helm, ArgoCD).
Strong software engineering skills in Python, Bash, and Go, with a focus on writing clean, maintainable, and scalable code.
Experience with observability & monitoring stacks (Prometheus, Grafana, Datadog, OpenTelemetry).
Strong background in security, compliance, and model governance for AI/ML systems.
Leadership & Execution:
Proven ability to lead high-impact engineering teams in a fast-paced AI environment.
Ability to drive technical strategy while remaining hands-on in critical areas.
Strong cross-functional collaboration skills, working closely with research and engineering teams.
Passion for automation, efficiency, and designing scalable self-service MLOps solutions.
Experience in mentoring and coaching engineers, fostering a culture of innovation and continuous learning.
It Would Be Great If You Have:
Experience working with LLMs and large-scale generative AI models in production.
Expertise in optimizing model inference latency and cost at scale.
Contributions to open-source MLOps tools or AI infrastructure projects.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8326471
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
10/08/2025
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a Staff Devops Engineer.
As a Devops Staff Engineer, you will not be assigned a specific R&D group, but will serve as a focal point for the DevOps engineers, to help and support with any issue.
Youll be leading projects that cross DevOps, push forward technical discussions and interact with each DevOps engineer as needed to solve diverse complex problems of high scale.
Youll support multi-region environments, build and maintain tools for automation, deployment, monitoring, and operations.
Youll troubleshoot and resolve issues in our various environments.
Youll play a key role in designing and enforcing infrastructure patterns that support zero-downtime deployments, high resilience, and compliance standards.
Youll collaborate with teams across the company to define and drive forward scalable, production-grade architecture.
Youll conduct periodic on-call duties and emergency response.
Requirements:
10+ years of experience in the industry, including 6+ years of hands-on experience in high-scale SaaS companies or zero-downtime/disaster recovery enterprise environments (e.g., banking, cybersecurity, healthcare, or large-scale cloud platform providers).
5+ years of experience in DevOps roles across a minimum of 2 different companies, with strong hands-on experience in Kubernetes and AWS. Experience with hybrid or multi-cloud architectures is a strong plus.
Experience with on-call duties to manage critical infrastructure and application issues outside business hours, ensuring high availability and reliability.
3+ years of experience with CI/CD tools such as GitLab, GitHub Actions, CircleCI, or similar.
2+ years of experience with programming languages such as Python or TypeScript. Strong Linux administration skills, including debugging and Bash scripting.
2+ years of experience with Terraform (experience with Terragrunt is a plus), as well as GitOps systems such as ArgoCD.
2+ years of experience with configuration management tools such as Ansible, Chef, or Puppet, and monitoring and alerting systems such as Datadog, Splunk, New Relic, or Grafana.
Strong understanding of networking concepts, including VPC, service meshes, routing, DNS, TLS, and firewalls.
Production-oriented mindset with a strong sense of ownership over reliability, scalability, and incident response.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8296098
סגור
שירות זה פתוח ללקוחות VIP בלבד