דרושים » הנדסה » Senior AI Ops / MLOps Engineer

משרות על המפה
 
בדיקת קורות חיים
VIP
הפוך ללקוח VIP
רגע, משהו חסר!
נשאר לך להשלים רק עוד פרט אחד:
 
שירות זה פתוח ללקוחות VIP בלבד
AllJObs VIP
כל החברות >
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 14 שעות
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
we aren't building a single, generic chatbot. We are building a Composable AI Microservice Architecture, a swarm of hundreds of hyper-specialized AI services, each meticulously "programmed" to solve small, focused tasks with high precision. This fleet powers Ava, our AI support engine, and a suite of cutting-edge generative tools for travel and expense management.
As a Senior AI Ops / MLOps Engineer, you are the architect of the platform that makes this scale possible. You will move beyond traditional MLOps to manage a "factory" of Language Models. Your challenge is one of orchestration and standardization, ensuring that every service in the swarm meets a rigorous bar for quality, reliability, and cost-efficiency.
What You'll Do:
Orchestrate the AI Fleet: Build and own the runtime environment for 100+ specialized AI services. Manage model routing, context versioning, and standardized memory/history stores.
High-Density Inference Optimization: Design and implement SageMaker Multi-Model Endpoints (MME) and Inference Components to serve multiple tuned SLMs per GPU, maximizing hardware utilization while minimizing latency.
Deterministic Service Excellence: Treat reliability as a layered engineering problem. Build deterministic "shells" around probabilistic LM outputs, prioritizing data-layer validation and strict serialization.
Automated Evaluation & Observability: Implement "LLM-as-a-judge" patterns and automated benchmarking to detect semantic drift and hallucinations across the fleet before they impact the user.
Standardize the Workflow: Obsess over building reusable patterns and Terraform-based infrastructure that eliminate "snowflake" configurations, allowing us to deploy new specialized AI tasks in minutes.
Agency Strategy: Partner with AI Researchers to find the "Goldilocks zone" for agentic autonomy-balancing the flexibility of LLM tool-use with the precision required for production stability.
Requirements:
Experience: 5+ years in SRE, Platform Engineering, or MLOps, with at least 2 years focused on deploying LLMs/SLMs in production environments.
SageMaker Mastery: Deep hands-on expertise with AWS SageMaker, specifically configuring Multi-Model Endpoints (MME), Inference Components, and GPU-backed instances (G5/P4).
SLM Expertise: Proven experience with Small Language Models (e.g., Mistral, Llama 3, Phi) and parameter-efficient fine-tuning (PEFT) deployment strategies like LoRA/QLoRA.
Technical Stack: * Languages: Strong proficiency in Python and Terraform.
Orchestration: Experience with Docker, Kubernetes (EKS), or AWS ECS/Fargate.
Data: Familiarity with Snowflake and Vector Databases.
The "AI Ops" Mindset: You understand that AI at scale is a statistical challenge. You are comfortable debugging issues at the data/serialization layer rather than defaulting to prompt tweaks.
CI/CD & Automation: Experience building robust pipelines (Jenkins, GitHub Actions) for non-deterministic software, including automated "eval" stages.
Education: BS or MS in Computer Science, Engineering, Mathematics, or a related technical field.
This position is open to all candidates.
 
Hide
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8567347
סגור
שירות זה פתוח ללקוחות VIP בלבד
משרות דומות שיכולות לעניין אותך
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
11/02/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Join our companys AI research group, a cross-functional team of ML engineers, researchers and security experts building the next generation of AI-powered security capabilities. Our mission is to leverage large language models to understand code, configuration, and human language at scale, and to turn this understanding into security AI capabilities that will drive our companys future security solutions.
We foster a hands-on, research-driven culture where youll work with large-scale data, modern ML infrastructure, and a global product footprint that impacts over 100,000 organizations worldwide.
Key Responsibilities
Your Impact & Responsibilities
As a Data Engineer - AI Technologies, you will be responsible for building and operating the data foundation that enables our LLM and ML research: from ingestion and augmentation, through labeling and quality control, to efficient data delivery for training and evaluation.
You will:
Own data pipelines for LLM training and evaluation
Design, build and maintain scalable pipelines to ingest, transform and serve large-scale text, log, code and semi-structured data from multiple products and internal systems.
Drive data augmentation and synthetic data generation
Implement and operate pipelines for data augmentation (e.g., prompt-based generation, paraphrasing, negative sampling, multi-positive pairs) in close collaboration with ML Research Engineers.
Build tagging, labeling and annotation workflows
Support human-in-the-loop labeling, active learning loops and semi-automated tagging. Work with domain experts to implement tools, schemas and processes for consistent, high-quality annotations.
Ensure data quality, observability and governance
Define and monitor data quality checks (coverage, drift, anomalies, duplicates, PII), manage dataset versions, and maintain clear documentation and lineage for training and evaluation datasets.
Optimize training data flows for efficiency and cost
Design storage layouts and access patterns that reduce training time and cost (e.g., sharding, caching, streaming). Work with ML engineers to make sure the right data arrives at the right place, in the right format.
Build and maintain data infrastructure for LLM workloads
Work with cloud and platform teams to develop robust, production-grade infrastructure: data lakes / warehouses, feature stores, vector stores, and high-throughput data services used by training jobs and offline evaluation.
Collaborate closely with ML Research Engineers and security experts
Translate modeling and security requirements into concrete data tasks: dataset design, splits, sampling strategies, and evaluation data construction for specific security use.
דרישות:
What You Bring
3+ years of hands-on experience as a Data Engineer or ML/Data Engineer, ideally in a product or platform team.
Strong programming skills in Python and experience with at least one additional language commonly used for data / backend (e.g., SQL, Scala, or Java).
Solid experience building ETL / ELT pipelines and batch/stream processing using tools such as Spark, Beam, Flink, Kafka, Airflow, Argo, or similar.
Experience working with cloud data platforms (e.g., AWS, GCP, Azure) and modern data storage technologies (object stores, data warehouses, data lakes).
Good understanding of data modeling, schema design, partitioning strategies and performance optimization for large datasets.
Familiarity with ML / LLM workflows: train/validation/test splits, dataset versioning, and the basics of model training and evaluation (you dont need to be the primary model researcher, but you understand what the models need from the data).
Strong software engineering practices: version control, code review, testing, CI/CD, and documentation.
Ability to work independently and in collaboration with ML engineers, researchers and security experts, and to translate high-level requirements into concrete data engineering tasks.
Nice to Have המשרה מיועדת לנשים ולגברים כאחד.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8541065
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
Required Software Engineer - Al Platform
What this role is really about:
You're building our AI platform. The internal system that powers AI capabilities across Product, Customer Success, Sales, Operations, Data, and IT.
This is a full-stack role where you'll own features end-to-end: design React interfaces for AI workflows, build Lambda functions that orchestrate multi-agent processes, integrate with enterprise systems (Salesforce, Workato, Snowflake), and optimize costs and performance at scale. You'll work with cutting-edge AI while building production-grade systems that handle real business operations.
If you want to build something that directly enables business growth, work across the full stack with modern tech, and have ownership over a platform that the entire company depends on, this is your opportunity.
Job responsibilities:
Build core AI platform services - Design and implement agent orchestration, prompt management, RAG, Connectors, and evaluation pipelines that power AI experiences across the company.
Develop complex agentic process - Develop a multi-step workflow that coordinates tools and services with proper observability, guardrails, and cost controls (using OpenAI Agent SDK, LangGraph, or a similar framework).
Build LLM evaluation and optimization process -Develop evaluation harnesses, offline/online experiments, prompt-testing frameworks, and dashboards to balance quality, latency, and spend across all AI services.
Requirements:
5+ years of hands‑on software engineering experience building production systems at scale.
Strong proficiency in Python, with Practical knowledge of databases.
Strong grounding of LLM/AI application patterns (RAG, tool use, function calling, guardrails) and vendor APIs (OpenAI or similar).
Experience with vector store (pgvector, Pinecone, OpenSearch), feature/semantic layers, or retrieval pipelines
Familiarity with: eval frameworks, prompt/version management, offline/online A/B testing, and cost/latency optimization.
Clear written and verbal communication; able to drive alignment with concise design docs and reviews.
Nice to have:
Experience building developer platforms or internal tooling
Hands-on experience with model optimization, fine-tuning, or distillation techniques.
Deep experience with cloud infrastructure (AWS), containers (Docker, Kubernetes), and distributed systems.
Frontend development frameworks such as React.
Background in SaaS/enterprise environments with compliance requirements (SOC2, GDPR).
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8557290
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
11/02/2026
Location: Tel Aviv-Yafo
Job Type: Full Time
Join our companys AI research group, a cross-functional team of ML engineers, researchers and security experts building the next generation of AI-powered security capabilities. Our mission is to leverage large language models to understand code, configuration, and human language at scale, and to turn this understanding into security AI capabilities which will drive our company AI future security solutions.
We foster a hands-on, research-driven culture where youll work with large-scale data, modern ML infrastructure, and a global product footprint that impacts over 100,000 organizations worldwide.
Key Responsibilities
Your Impact & Responsibilities
As a Senior ML Research Engineer, you will be responsible for the end-to-end lifecycle of large language models: from data definition and curation, through training and evaluation, to providing robust models that can be consumed by product and platform teams.
Own training and fine-tuning of LLMs / seq2seq models: Design and execute training pipelines for transformer-based models (encoder-decoder, decoder-only, retrievalaugmented, etc.), and fine-tune open-source LLMs on our company-specific data (security content, logs, incidents, customer interactions).
Apply advanced LLM training techniques such as instruction tuning, preference / contrastive learning, LoRA / PEFT, continual pre-training, and domain adaptation where appropriate.
Work deeply with data: define data strategies with product, research and domain experts; build and maintain data pipelines for collecting, cleaning, de-duplicating and labeling large-scale text, code and semi-structured data; and design synthetic data generation and augmentation pipelines.
Build robust evaluation and experimentation frameworks: define offline metrics for LLM quality (task-specific accuracy, calibration, hallucination rate, safety, latency and cost); implement automated evaluation suites (benchmarks, regression tests, redteaming scenarios); and track model performance over time.
Scale training and inference: use distributed training frameworks (e.g. DeepSpeed, FSDP, tensor/pipeline parallelism) to efficiently train models on multi-GPU / multi-node clusters, and optimize inference performance and cost with techniques such as quantization, distillation and caching.
Collaborate closely with security researchers and data engineers to turn domain knowledge and threat intelligence into high-value training and evaluation data, and to expose your models through well-defined interfaces to downstream product and platform teams.
Requirements:
What You Bring
5+ years of hands-on work in machine learning / deep learning, including 3+ years focused on NLP / language models.
Proven track record of training and fine-tuning transformer-based models (BERT-style, encoder-decoder, or LLMs), not just consuming hosted APIs.
Strong programming skills in Python and at least one major deep learning framework (PyTorch preferred; TensorFlow).
Solid understanding of transformer architectures, attention mechanisms, tokenization, positional encodings, and modern training techniques.
Experience building data pipelines and tools for large-scale text / log / code processing (e.g. Spark, Beam, Dask, or equivalent frameworks).
Practical experience with ML infrastructure, such as experiment tracking (Weights & Biases, MLflow or similar), job orchestration (Airflow, Argo, Kubeflow, SageMaker, etc.), and distributed training on multi-GPU systems.
Strong software engineering practices: version control, code review, testing, CI/CD, and documentation.
Ability to own research and engineering projects end-to-end: from idea, through prototype and controlled experiments, to models ready for integration by product and platform teams.
Good communication skills and the ability to work closely with non-ML stakeholders (security experts, product managers, engineers).
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8541239
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
15/02/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are seeking an experienced Generative AI Engineer to join our AI squad at our company. This is a unique opportunity to wear multiple hats - serving as both a developer of cutting-edge GenAI solutions and an advisory expert helping organizations transform their AI capabilities. You'll build end-to-end GenAI projects from conception to production while staying at the forefront of this rapidly evolving field.
Key Responsibilities
GenAI Development & Implementation
End-to-End Development: Build GenAI solutions from POC through production deployment, handling all backend development responsibilities
Client Engagement: Participate in technical discussions with clients, gather requirements, and help translate business visions into feasible technical solutions through presentations and consultations
Backend Development: Design and implement production-grade microservices architectures for GenAI applications using Python
Cloud Implementation: Deploy and manage GenAI solutions across GCP, Azure, and AWS platforms, leveraging cloud-native AI services
Cross-functional Collaboration: Work closely with project managers, full-stack developers, and Power Automate teams to deliver complete solutions
System Evaluation: Assess and optimize production-grade GenAI systems for performance, scalability, and reliability
Continuous Learning & Innovation
Technology Scouting: Continuously explore and evaluate new GenAI models, frameworks, and techniques as they emerge
Best Practices Development: Establish and refine methodologies for GenAI solution development and deployment.
Requirements:
Technical Expertise:
Programming: Advanced proficiency in Python for backend development and AI applications
GenAI Mastery: Deep understanding of large language models (LLMs) and experience with major model APIs (OpenAI, Anthropic, Google, etc.)
Multi-Agent Systems: Expertise in designing and implementing GenAI multi-agent architectures
Prompt Engineering: Advanced skills in prompt design, optimization, and engineering techniques
Cloud Platforms:
Required: Hands-on experience with AI services in at least one major cloud platform (GCP, Azure, or AWS)
Advantage: Experience across multiple cloud platforms (AI Search, Vertex AI, SageMaker, etc.)
Development Frameworks: Experience with GenAI frameworks like LangChain and cloud-based retrieval services
Software Engineering: Strong background in microservices architecture, API development, and production system design
AI/ML Fundamentals: Solid understanding of deep learning principles and GenAI techniques
Containerization (Advantage): Experience with Docker and Kubernetes for deployment and orchestration
OCR Technologies (Advantage): Experience with Optical Character Recognition systems and document processing
Data Pipelines (Advantage): Experience building and maintaining data processing pipelines
Professional Experience
Mid+ Level Experience: 2+ years in AI/ML development with significant GenAI project experience
Production Systems: Proven track record of deploying and maintaining AI solutions in production environments
Client-Facing Experience: Comfortable with technical presentations and requirement gathering sessions
Education & Background
Preferred: Bachelor's or Master's degree in Computer Science, AI, Machine Learning, or related technical field
Alternative: Demonstrated industrial experience in developing deep learning and GenAI solutions (degree not required with strong portfolio)
Soft Skills
Problem-Solving: Excellent analytical and creative problem-solving abilities
Communication: Strong technical communication skills for both technical and non-technical audiences
Collaboration: Proven ability to work effectively in cross-functional teams
Adaptability: Thrives in fast-paced environments and eager to learn emerging technologies
Consulting Mindset: Ability to understand client needs and provide strategic technical guidance.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8546141
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Machine Learning Engineer II - GenAI Applications
26947
About the team:
This opening is for the GenAI Applications Team within the Data & AI Marketplace department.
The GenAI Applications team is responsible for designing and delivering agentic, ML-powered solutions for some of our most impactful products, including booking search experiences, trip planning, and trip helpfulness. The team builds AI-driven applications and conversational agents, such as chatbots and intelligent assistants, that significantly enhance the end-to-end customer experience.
Role Description:
As a Machine Learning Engineer, you will work closely with experienced engineers and ML scientists to build scalable, production-grade GenAI applications. Your work will focus on designing, training, and deploying ML systems leveraging LLMs,, recommendation systems, and agent-based architectures, using state-of-the-art technologies. These solutions will directly power customer-facing experiences and play a key role in shaping the future of AI-driven travel products.
Key Job Responsibilities and Duties:
Deploying machine learning models: Design, develop and deploy in collaboration with scientists, scalable machine learning models and algorithms that provide content related insights and generative AI applications, ensuring scalability, efficiency, and accuracy.
Evaluating possible architecture solutions by taking into account cost, business requirements, emerging technologies, and technology requirements, like latency, throughput, and scale.
Generative AI Development: Contribute to the development of generative models such as GPT (Generative Pre-trained Transformer) variants or similar architectures for creative content generation, Q&A, chatbots, translation or other innovative applications.
Deployment and integration: Work closely with software engineers to integrate machine learning models into production systems. Ensure seamless deployment and efficient model inference in real-time environments. Collaborate with DevOps to implement effective monitoring and maintenance strategies.
Owning a service end to end by actively monitoring application health and performance, setting and monitoring relevant metrics and acting accordingly when violated.
Maintain clean, scalable code, ensuring reproducibility and easy integration of models into production environments, including CI/CD.
Collaborate with multidisciplinary teams: Collaborate with product managers, data scientists, and analysts to understand business requirements and translate them into machine learning solutions.
Requirements:
We are looking for driven MLEs who enjoy solving problems, who initiate solutions and discussions and who believe that any challenge can be scaled with the right mindset and tools.
We have found that people who match the following requirements are the ones who fit us best:
Bachelors or masters degree in computer science, Engineering, Statistics, or a related field.
Minimum of 4 years of experience as a Machine Learning Engineer or a similar role, with a consistent record of successfully delivering ML solutions.
Strong programming skills in languages such as Python and Java.
Experience with cloud frameworks like AWS sagemaker for training, evaluation and serving models using TensorFlow, PyTorch, or scikit-learn.
Experience with big data processing frameworks such, Pyspark, Apache Flink, Snowflake or similar frameworks.
Experience with data at scale using MySQL, Pyspark, Snowflake and similar frameworks.
Demonstrable experience with MySQL, Cassandra, DynamoDB or similar relational/NoSQL database systems.
Deep understanding of machine learning algorithms, statistical models, and data structures.
Experience in deploying large-scale language models like GPT, BERT, or similar architectures - an advantage.
Proficiency in data manipulation, analysis, and visualization using tools like NumPy, pandas, and matplotlib - an advantage.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8560104
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
29/01/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for an experienced ML/AI engineer to anchor our ML Platform team. You will lead the effort to productize our Large Language Models, transforming experimental code into robust, efficient services. Your goal is to build the infrastructure that allows our Data Scientists to move fast without breaking production.

What you'll be doing:
Model Development & Optimization

Fine-tuning & Engineering: Develop and fine-tune LLMs/classic models for specific medical tasks (condition detection, dialogue systems) using internal datasets.
Research to Production: Partner with Data Scientists to develop experimental code into scalable, production-ready modules.
Platform Engineering & Infrastructure

Pipeline Orchestration: Build and maintain complex ML workflows using Kubeflow Pipelines (KFP) or similar orchestration tools.
Internal Tooling: Develop and manage the internal Python ecosystem-libraries, SDKs, and utilities-that the Data Science team uses for daily development.
CI/CD & Automation: Write and maintain CI/CD scripts to automate the testing, versioning, and deployment of machine learning artifacts.
Production Standards & Integration

System Integration: Work with backend developers to integrate trained models/agents into the core application architecture.
Code Quality: Enforce high engineering standards through code reviews, ensuring that research code meets production reliability and maintainability requirements.
Requirements Engineering: Translate evolving data science requirements into concrete infrastructure and platform features.
Requirements:
Experience: 10+ years in software engineering with 5+ years in backend/platform roles.
Languages: Expert-level Python; proficiency in another language, such as C++, Rust, Java, or Go, is an advantage.
Cloud & Infra: 4+ years with GCP (preferred) or AWS, including Docker, Kubernetes, and pipelines (KFP/Vertex).
ML Core: Production experience with PyTorch, Transformers, and low-level libraries (CUDA).
LLM Stack: Experience with inference optimization (e.g: vLLM/NGC) and fine-tuning (Axolotl/Huggingface).
Key Traits: Strong focus on code optimization, system reliability, and collaborative problem-solving.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8523767
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
5 ימים
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
Were looking for a Generative AI Developer to join our forward-thinking engineering team. This role is perfect for someone with a passion for cutting-edge AI, a strong software engineering background, and the creative spark to identify and implement novel use cases within our product.

You will play a critical role in adding AI capabilities to our FinOps SaaS platform. Whether it's enhancing user workflows, automating insights, or inventing entirely new product experiences, youll have both the freedom and support to experiment and execute.

We provide a uniquely rich dataset covering the full scope of a companys cloud spend. This expansive data playground offers a powerful foundation for experimentation and insight generation, enabling the development of intelligent, value-driven features.

Responsibilities:
Lead the charge in transforming our product and preparing it for the agentic age.
Design, build, and deploy generative AI-powered features across our product.

Identify opportunities for AI integration by proactively exploring FinOps use cases and user needs.

Prototype and validate new AI use cases quickly and iterate based on internal and external feedback.

Collaborate cross-functionally with product, design, and backend teams to drive innovation from concept to production.

Stay current with the fast-moving generative AI landscape and evaluate new models, APIs, and tools (e.g., OpenAI, Anthropic, Hugging Face, AWS Bedrock, open-source LLMs).

Live in the future and track new innovations and paradigms in this fast evolving field and identify opportunities to integrate them into the product.

Implement safeguards, prompt engineering techniques, and usage monitoring to ensure high-quality AI outputs.

Optimize model performance, inference time, and cost efficiency within AWS infrastructure.
Requirements:
Requirements:

3+ years of hands-on experience in software engineering, with at least 1-2 years working on generative AI projects (LLMs, diffusion models, multimodal models, etc.).

Proven ability to go from idea to production-ideally with examples of real-world AI features youve shipped.

Fluency in Python, Node.js, or similar languages used in ML and full-stack development.

Experience with prompt engineering, fine-tuning, or embedding models using frameworks like LangChain, LlamaIndex, or similar.

Familiarity with AWS services and best practices, including Lambda, S3, SageMaker, ECS/EKS, Bedrock etc.

Experience with MLOps and model deployment practices (e.g., containerization, GPU inference, vector databases).

Creativity and initiative-able to pitch and prototype ideas with minimal oversight.

Strong communication skills and the ability to explain technical concepts to non-technical stakeholders.

Nice-to-Haves:

Prior experience integrating generative AI in FinOps or cloud cost optimization tools.

Background in NLP, computer vision, or other relevant ML fields.

Contributions to open-source AI tools or research.

Knowledge of responsible AI principles and handling model risks.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8563659
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
22/02/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Were seeking an experienced and skilled Data and AI Infra Engineer to join our Data Infrastructure team and drive the companys data capabilities at scale.
As the company is fast growing, the mission of the data and AI infrastructure team is to ensure the company can manage data at scale efficiently and seamlessly through robust and reliable data infrastructure.
A day in the life and how youll make an impact:
As a Senior Engineer, you are required to independently lead the design, development, and optimization of our data infrastructure, collaborating closely with software engineers, data scientists, data engineers, and other key stakeholders. You are expected to own critical initiatives, influence architectural decisions, and mentor engineers to foster a high-performing team
You will:
Lead the design and development of scalable, reliable, and secure data storage, processing, and access systems.
Define and drive best practices for CI/CD processes, ensuring seamless deployment and automation of data services.
Oversee and optimize our machine learning platform for training, releasing, serving, and monitoring models in production.
Own and develop the company-wide LLM infrastructure, enabling teams to efficiently build and deploy projects leveraging LLM capabilities.
Own the company's feature store, ensuring high-quality, reusable, and consistent features for ML and analytics use cases.
Architect and implement real-time event processing and data enrichment solutions, empowering teams with high-quality, real-time insights.
Partner with cross-functional teams to integrate data and machine learning models into products and services.
Ensure that our data systems are compliant with the data governance requirements of our customers and industry best practices.
Mentor and guide engineers, fostering a culture of innovation, knowledge sharing, and continuous improvement.
Requirements:
7+ years of experience in data infra or backend engineering.
Strong knowledge of data services architecture, and ML Ops.
Experience with cloud-based data infrastructure in the cloud, such as AWS, GCP, or Azure.
Deep experience with SQL and NoSQL databases.
Experience with Data Warehouse technologies such as Snowflake and Databricks.
Proficiency in backend programming languages like Python, NodeJS, or an equivalent.
Proven leadership experience, including mentoring engineers and driving technical initiatives.
Strong communication, collaboration, and stakeholder management skills.
Bonus Points:
Experience leading teams working with serverless technologies like AWS Lambda.
Hands-on experience with TypeScript in backend environments.
Familiarity with Large Language Models (LLMs) and AI infrastructure.
Experience building infrastructure for Data Science and Machine Learning.
Experience collaborating with BI developers and analysts to drive business value.
Expertise in administering and managing Databricks clusters.
Experience with streaming technologies such as Amazon Kinesis and Apache Kafka.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8555763
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Senior Machine Learning Engineer I - GenAI Applications
20031
Leadership/Team Quote:
This opening is for the GenAI Infra team in the Marketplace AI department.
The GenAI Infra team builds the Agents platform which is used for all agnetic and non-agentic flows. This team is responsible for both the GenAI agents and the orchestration around them, helping support applications such as the AI Trip Planner, Free text search, etc.
Role Description:
As Senior Machine Learning Engineer, youll work with top notch engineers and data scientists from the team on bringing it to the next level and enabling optimal user experience. The work will focus on building, deploying and serving GenAI capabilities (Agents, Tools and the orchestration between them) using the most advanced technologies and models.
Key Job Responsibilities and Duties:
Deploying machine learning models: Design, develop and deploy in collaboration with scientists, scalable machine learning models and algorithms that provide content related insights and generative AI applications, ensuring scalability, efficiency, and accuracy.
Evaluating possible architecture solutions by taking into account cost, business requirements, emerging technologies, and technology requirements, like latency, throughput, and scale.
Generative AI Development: Contribute to the development of generative models such as GPT (Generative Pre-trained Transformer) variants or similar architectures for creative content generation, Q&A, translation or other innovative applications.
Deployment and integration: Work closely with software engineers to integrate machine learning models into production systems. Ensure seamless deployment and efficient model inference in real-time environments. Collaborate with DevOps to implement effective monitoring and maintenance strategies.
Owning a service end to end by actively monitoring application health and performance, setting and monitoring relevant metrics and acting accordingly when violated.
Maintain clean, scalable code, ensuring reproducibility and easy integration of models into production environments, including CI/CD.
Collaborate with multidisciplinary teams: Collaborate with product managers, data scientists, and analysts to understand business requirements and translate them into machine learning solutions.
Requirements:
Bachelors or masters degree in computer science, Engineering, Statistics, or a related field.
Minimum of 6 years of experience as a Machine Learning Engineer or a similar role, with a consistent record of successfully delivering ML solutions.
Strong programming skills in languages such as Python and Java.
Experience with cloud frameworks like AWS sagemaker for training, evaluation and serving models using TensorFlow, PyTorch, or scikit-learn.
Experience with LLMs, Agents and MCP in production environments.
Experience with big data processing frameworks such, Pyspark, Apache Flink, Snowflake or similar frameworks.
Experience with data at scale using MySQL, Pyspark, Snowflake and similar frameworks.
Demonstrable experience with MySQL, Cassandra, DynamoDB or similar relational/NoSQL database systems.
Deep understanding of machine learning algorithms, statistical models, and data structures.
Experience in deploying large-scale language models like GPT, BERT, or similar architectures - an advantage.
Proficiency in data manipulation, analysis, and visualization using tools like NumPy, pandas, and matplotlib - an advantage.
Experience with experimental design, A/B testing, and evaluation metrics for ML models - an advantage.
Experience of working on products that impact a large customer base - an advantage.
Excellent communication in English; written and spoken.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8560111
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
17/02/2026
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
Required Machine learning operations engineer
Your Mission:
As an MLOps Engineer, your mission is to design, build, and operate the platforms that power our machine learning and generative AI products spanning real-time use cases such as large-scale fraud scoring, MCP & agentic workflows support. Youll create reliable CI/CD for models and Agents, robust data/feature pipelines, secure model serving, and comprehensive observability. You will also support our agentic AI ecosystem and Model Context Protocol (MCP) services so that models can safely use tools, data, and actions across.
You will partner closely with Data Scientists, Data/Platform Engineers, Product, and SRE to ensure every model from classic ML to LLM/RAG agents moves from prototype to production with strong reliability, governance, cost efficiency, and measurable business impact.
Responsibilities:
Operate & Develop ML/LLM platforms on Kubernetes + cloud (Azure; AWS/GCP ok) with Docker, Terraform, and other relevant tools
Manage object storage, GPUs, and autoscaling for training & low-latency model serving
Manage cloud environment, networking, service mesh, secrets, and policies to meet PCI-DSS and data-residency requirements
Build end-to-end CI/CD for models/agents/MCP tooling (versioning, tests, approvals)
Deliver real-time fraud/risk scoring & agent signals under strict latency SLOs.
Maintain MCP servers/clients: tool/resource definitions, versioning, quotas, isolation, access controls
Integrate agents with microservices, event streams, and rule engines; provide SLAs, tracing, and on-call runbooks
Measure operational metrics of ML/LLM (latency, throughput, cost, tokens, tool success, safety events)
Enforce governance: RBAC/ABAC, row-level security, encryption, PII/secrets management, audit trails.
Partner with DS on packaging (wheels/conda/containers), feature contracts, and reproducible experiments.
lead incident response and post-mortems.
Drive FinOps: right-sizing, GPU utilization, batching/caching, budget alerts.
Requirements:
4+ years in DevOps/MLOps/Platform roles building and operating production ML systems (batch and real-time)
Strong hands-on with Kubernetes, Docker, Terraform/IaC, and CI/CD
Practical experience with Spark/Databricks and scalable data processing
Proficiency in Python & Bash
Ability to operate DS code and optimize runtime performance.
Experience with model registries (MLflow or similar), experiment tracking, and artifact management.
Production model serving using FastAPI/Ray Serve/Triton/TorchServe, including autoscaling and rollout strategies
Monitoring and tracing with Prometheus/Grafana/OpenTelemetry; alerting tied to SLOs/SLAs
Solid understanding of PCI-DSS/GDPR considerations for data and ML systems
Experience with the Azure cloud environment is a big plus
Operating LLM/agent workloads in production (prompt/config versioning, tool execution reliability, fallback/retry policies)
Building/maintaining RAG stacks (indexing pipelines, vector DBs, retrieval evaluation, hybrid search)
Implementing guardrails (policy checks, content filters, allow/deny lists) and human-in-the-loop workflows
Experience with feature stores - Qwak Feature Store, Feast
A/B testing for models and agents, offline/online evaluation frameworks
Payments/fraud/risk domain experience; integrating ML outputs with rule engines and operational systems - Advantage
Familiarity with Databricks Unity Catalog, dbt, or similar tooling.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8550121
סגור
שירות זה פתוח ללקוחות VIP בלבד