דרושים » AI » Senior AI Engineer - Applied AI Engineering Group

משרות על המפה
 
בדיקת קורות חיים
VIP
הפוך ללקוח VIP
רגע, משהו חסר!
נשאר לך להשלים רק עוד פרט אחד:
 
שירות זה פתוח ללקוחות VIP בלבד
AllJObs VIP
כל החברות >
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 3 שעות
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Senior AI Engineer - Applied AI Engineering Group
The Dream Job
It starts with you - an engineer driven to build the agentic AI platform that turns LLMs into reliable, production-grade capabilities. You care about clean APIs, well-defined service boundaries, and systems that teams can build on with confidence. Dream is AI-first across the board - every team builds and operates agents. You'll architect and ship the platform that makes this possible: agent orchestration frameworks, LLM gateways, evaluation pipelines, tool-calling infrastructure, and retrieval systems. Without this platform, agents don't ship - you own the layer that turns AI research into Sovereign AI products, deployed across cloud and on-prem environments.
If you want to make a meaningful impact, join our mission and build the agentic AI platform that drives Sovereign AI products - this role is for you.
The Dream-Maker Responsibilities
Design and build agentic systems - single and multi-agent workflows with planning, memory, context engineering, and tool use - for both internal automation and product-facing autonomous capabilities operating over long time horizons.
Build and operate the AI platform layer - LLM gateways, prompt management, structured output handling, tool-calling infrastructure, and cost/latency optimization - deployed on Kubernetes, consumed by every team for their agentic work.
Own the agent framework layer - orchestration primitives, execution environments, state management, and sandboxed tool execution - giving every team the building blocks to create and operate their own agents.
Build evaluation infrastructure that gives teams confidence in agent behavior - automated LLM and agent evals for quality, correctness, safety, latency, cost, and regressions, including human-in-the-loop oversight for mission-critical workflows.
Productionize and harden backend services (APIs, gRPC, async workers) that integrate LLMs - with proper error handling, retries, circuit breakers, and high-availability patterns.
Own RAG pipelines and retrieval systems - indexing, chunking, embedding, vector database management, filtering, and relevance tuning for production retrieval.
Optimize performance and cost across the AI stack - model routing, caching, batching, and inference cost management.
Ship shared tooling - libraries, SDKs, agent templates, and documentation - while working closely with ML Platform, Data Platform, DevOps, and other teams across the Applied AI Engineering group. Own architecture, documentation, and operations end-to-end.
דרישות:
5+ years in backend or distributed systems engineering, with 2+ years focused on production systems that integrate AI/ML models or LLMs.
Engineering craft - Strong Python, Go, or Java, system architecture, API design, testing, and secure coding practices.
Agentic systems - Experience designing and building agent orchestration, tool-use systems, and autonomous workflows; familiarity with frameworks like LangGraph or similar, or having built equivalent from scratch
Backend engineering - Experience building production APIs and services (FastAPI or similar); async programming, service architecture, high-availability, and reliability patterns (retries, circuit breakers, backpressure)
LLM integration - Hands-on experience integrating LLMs via SDKs and APIs; context engineering, structured outputs, tool calling, and model routing
RAG & retrieval - Experience with embedding pipelines, vector databases (e.g., Milvus, Qdrant, Pinecone), chunking strategies, and relevance tuning
Evaluation & observability - Experience designing LLM and agent evals, monitoring AI system quality, and building observability for non-deterministic systems
Nice to Have:
Platform & infra - Kubernetes, AWS, Terraform or similar IaC, CI/CD, container orchestration, deploying and operating production services
Experience with MCP or similar tool-use protocols for agent-to-service communication
Hands-on ML experience - המשרה מיועדת לנשים ולגברים כאחד.
 
Hide
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8603620
סגור
שירות זה פתוח ללקוחות VIP בלבד
משרות דומות שיכולות לעניין אותך
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 3 שעות
Location: Tel Aviv-Yafo
Job Type: Full Time
Required AI Engineering Team Lead - Applied AI Engineering Group
Tel Aviv Full-time
The Dream Job
It starts with you - a technical leader driven to build both the agentic AI platform and the engineering team behind it. You care about backend quality, platform reliability, and growing engineers through real ownership. We are AI-first across the board - every team builds and operates agents. You'll set the technical direction for the platform that makes this possible: agent orchestration frameworks, LLM gateways, evaluation infrastructure, tool-calling systems, and retrieval pipelines. Without this platform, agents don't ship - you own the layer that turns AI research into Sovereign AI products, deployed across cloud and on-prem environments. You stay close enough to the codebase to debug production incidents, unblock your engineers, and make sound architecture calls.
If you want to make a meaningful impact, join our mission and lead the team that builds the agentic AI platform driving Sovereign AI products - this role is for you.
The Dream-Maker Responsibilities
Architect and evolve the AI platform - agent orchestration, LLM gateways, context engineering pipelines, evaluation infrastructure, tool-calling systems, and retrieval pipelines - through RFCs, prototypes, and design reviews.
Lead and grow a small team of AI Engineers building the agent framework, production backend services, and AI platform infrastructure - hire, mentor, pair on hard problems, and raise the bar through hands-on code and design reviews.
Contribute to critical systems, debug production incidents, and maintain enough codebase context to make sound technical calls.
Own reliability across AI and agent services - set and enforce SLAs, build observability for non-deterministic systems, and harden tool execution environments for cost and security.
Set the standard for AI engineering practices - agent testing strategies, evaluation frameworks with human-in-the-loop oversight, retrieval quality benchmarks, and CI/CD for AI systems.
Work closely with ML Platform, Data Platform, DevOps, Data Science, and Product teams across the Applied AI Engineering group - ensure the AI platform evolves to serve teams building agentic workflows across the organization.
Measure and improve developer experience - deploy friction, onboarding time, CI turnaround - as seriously as system performance.
Requirements:
6+ years in backend software engineering, with 4+ years focused on production systems that integrate AI/ML models or LLMs.
2+ years leading an engineering team - hiring, mentoring, conducting design reviews, and shipping alongside your team.
Engineering craft - Strong Python, Go, or Java, system architecture, API design, testing, and secure coding practices.
Agentic systems & LLM integration - Deep understanding of agent orchestration, tool-use architectures, LLM integration patterns, context engineering, and frameworks like LangGraph or similar, or custom-built equivalents
Backend & platform engineering - Experience building and operating production APIs, services, and platform infrastructure at scale; comfortable working with relational databases, message queues, and event-driven architectures
RAG & retrieval - Experience with production RAG pipelines, vector databases, embedding systems, and retrieval quality
Evaluation & observability - Experience building LLM and agent eval infrastructure, monitoring AI quality, and observability for non-deterministic systems
Nice to Have:
Platform & infra - Kubernetes, AWS, Terraform or similar IaC, CI/CD, service architecture, incident management
Experience with MCP or similar tool-use protocols for agent-to-service communication
Hands-on ML experience - model training, fine-tuning, or working directly with ML pipelines.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8603446
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 3 שעות
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Senior ML Engineer - Applied AI Engineering Group
The Dream Job
It starts with you - an engineer driven to build the ML platform that turns research into reliable, production-grade intelligence. You care about reproducibility, low-friction experimentation, and infrastructure that earns the trust of the scientists and researchers who depend on it daily. You'll architect and ship our ML platform - training pipelines, model serving, feature stores, experiment tracking, and compute orchestration - turning models into production capabilities across cloud and on-prem, including air-gapped deployments. A significant part of the platform supports large language models, with unique challenges across training, evaluation, and inference in mission-critical environments.
If you want to make a meaningful impact, join our mission and build the ML platform that drives Sovereign AI products - this role is for you.
The Dream-Maker Responsibilities
Build and operate ML training infrastructure - distributed training pipelines, compute scheduling, and reproducible experiment workflows that data scientists rely on daily.
Own model serving and inference systems - packaging, deployment, autoscaling, A/B testing, canary rollouts, and latency/cost optimization for production models.
Run feature stores, model registries, and dataset versioning - enabling self-serve feature engineering, model lineage, and reproducible experiments across teams.
Build experiment tracking and evaluation infrastructure - automated evals, comparison dashboards, drift detection, and monitoring that give teams visibility into model behavior and performance.
Build and maintain production pipelines for training, fine-tuning workflows, and serving domain models - owning reliability, reproducibility, and scale.
Build and maintain the monitoring and observability layer - model performance tracking, data and prediction drift detection, data quality validation, and alerting.
Improve performance and cost across the ML stack - training throughput, inference latency, batch vs. real-time tradeoffs, and compute cost management.
Ship shared tooling - libraries, templates, CI/CD for models, IaC, and runbooks - while collaborating across Data Platform, AI, Data Science, Engineering, and DevOps. Own architecture, documentation, and operations end-to-end.
Requirements:
5+ years in software engineering, with 2+ years focused on ML infrastructure, MLOps, or data-intensive systems
Engineering craft - Strong Python, distributed systems design, testing, secure coding, API design, CI/CD discipline, and production ownership.
ML platform & serving - Model serving frameworks (e.g., Triton, TorchServe, vLLM, Ray Serve); model packaging, deployment pipelines, and inference optimization
Training infrastructure - Distributed training pipelines (e.g., frameworks like PyTorch, JAX) experiment orchestration and reproducibility
ML lifecycle tooling - Feature stores, model registries, experiment tracking (e.g., MLflow, Weights & Biases); dataset versioning and lineage
Data pipelines - Building training and inference data pipelines; familiarity with tools like Spark, Airflow/Dagster, and streaming ingestion
Comfortable with AI coding tools like Cursor, Claude Code, or Copilot
Nice to Have:
Experience operating in constrained environments - on-premise, private cloud, or air-gapped deployments
Hands-on experience with simulation environments, synthetic data generation, or reinforcement learning workflows
Platform & infra - Kubernetes, AWS, Terraform or similar IaC, CI/CD, observability, incident response
Hands-on data science or applied ML experience.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8603632
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a Senior AI Engineer to join our Cybersecurity team in Tel Aviv. You will design, build, and productionize LLM-powered applications, multi-agent systems, and MLOps infrastructure that power our company's next-generation cybersecurity capabilities. This is a high-impact, hands-on role at the intersection of applied AI, agentic systems, and network securit
What You'll Do
Design and develop LLM-powered security features and internal AI tools, including RAG pipelines, multi-agent workflows, and prompt-engineered systems tailored for cybersecurity use cases
Architect and operate multi-agent systems in production - including agent orchestration, inter-agent communication, task delegation, and failure handling at scale
Build robust agent monitoring and observability pipelines: tracing agent execution, detecting drift or failure, alerting on anomalous behavior, and maintaining agent reliability SLAs
Build and maintain scalable MLOps infrastructure: model serving, evaluation frameworks, experiment tracking, and CI/CD for ML models
Work with internal datasets (network telemetry, security logs, threat intelligence) to fine-tune and adapt foundation models for domain-specific detection and response tasks
Partner with the Cybersecurity, R&D, and infrastructure teams to define AI-driven security features and deliver them end-to-end
Establish best practices for model observability, safety, and responsible AI deployment within the organization
Stay current with the fast-moving LLM/GenAI and agentic AI ecosystem and evaluate emerging frameworks, models, and tools for adoption.
Requirements:
Must-Have
5-8 years of software engineering experience, with at least 2-3 years focused on AI/ML engineering
Hands-on experience building production-grade LLM applications - RAG, agents, tool use, or fine-tuning
Proven experience designing and running multi-agent systems in production: orchestration patterns, agent state management, retries, and graceful degradation
Experience monitoring and observing AI agents in production - execution tracing, latency tracking, failure detection, and alerting (e.g., LangSmith, Arize, custom observability stacks)
Proficiency with agentic frameworks: LangChain, LangGraph, and/or AWS Bedrock AgentCore
Strong Python skills and comfort working across the full AI application stack
Experience designing and operating MLOps pipelines (model versioning, deployment, monitoring)
Solid understanding of transformer-based models, embeddings, and vector databases (e.g., Pinecone, Weaviate, pgvector)
Comfortable working in cloud environments (AWS, GCP, or Azure) and containerized deployments (Docker, Kubernetes)
Strong problem-solving skills and ability to work autonomously in a fast-paced environment
Nice-to-Have
Background in cybersecurity - threat detection, SIEM, SOC automation, or security data analysis - a significant plus for this role
Familiarity with networking concepts (SDN, cloud-native networking, BGP, telemetry)
Experience with model evaluation and benchmarking (LLM-as-judge, RAGAS, or custom eval harnesses)
Exposure to MCP (Model Context Protocol) for tool-augmented agentic workflows
Prior experience in enterprise SaaS, networking, or telecom domains
Publications, open-source contributions, or projects in the LLM/GenAI or agentic AI space
Our Stack
Python PyTorch OpenAI / Anthropic APIs LangChain LangGraph AWS Bedrock AgentCore LangSmith Kubernetes Kafka Elasticsearch AWS PostgreSQL GitHub Jira Confluence.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8595648
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
30/03/2026
חברה חסויה
Location: Tel Aviv-Yafo and Netanya
Job Type: Full Time
We are seeking an experienced, hands-on Senior AI Engineer to join the Generative AI applications Platform group and lead the backend implementation and architecture of AI/LLM solutions - from agent graphs and tooling to RAG, streaming, and production deployment.
As a Senior ML Engineer you will
Design and own agent architectures - Build and evolve graph-based agent workflows (multi-node LLM flows, tool execution, routing, human-in-the-loop review gates) using LangGraph, with clear state schemas, checkpointing, and streaming to production.
Turn product and user needs into backend AI - Work with Engineers, Product, and Analysts to translate business problems into technical requirements and implementations, including agent types, tools, RAG pipelines, and configuration-driven behavior.
Design, develop, and deploy GenAI capabilities end-to-end - LangChain tools and integrations, RAG (retrievers, vector stores, agentic flows), structured outputs, and APIs for chat, Copilot-style integrations, and MCP.
Raise the bar on quality and reliability - Establish patterns for observability (e.g., LangSmith), error handling, content safety, bounded autonomy (tool schemas, review workflows), and evaluation systems so that AI behavior is predictable and auditable.
Mentor and align the team - Provide technical guidance on LLM backend architecture and LangGraph/LangChain best practices so the team can iterate quickly and safely.
Requirements:
Backend-LLM & agent architecture - 5+ years in production ML/AI and backend systems; recent hands-on experience with backend LLM systems, including agent workflows (e.g., LangGraph or similar), LangChain tooling and chains, state management, and streaming (e.g., SSE). You think in terms of nodes, state schemas, routing, and human-in-the-loop.
Technical stack - Proficient in Python; comfortable with LangGraph, LangChain, FastAPI, PostgreSQL, and optionally Azure AI Search or similar. Experience with LLM providers (OpenAI/Azure, Google Vertex AI, etc.) and RAG (retrievers, chunking, reranking) expected.
Generative AI in production - Proven track record building production GenAI applications, including multi-step agents, RAG, tool-augmented LLMs, and ideally human-in-the-loop or review flows. You care about observability, validation, and safe rollout.
Bachelor's degree or higher in Computer Science or a related field, and strong communication and collaboration skills.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8579725
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
4 ימים
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Were looking for a highly motivated AI Developer to help design, build, and deploy intelligent agentic systems across our product ecosystem. In this role, you'll work at the intersection of machine learning, backend systems, and modern frontend technologies to deliver AI-first features that feel magical to users.
This is a hands-on, cross-functional role ideal for engineers who love building full-fledged features-from data pipelines and LLM orchestration to intuitive UI experiences-with a strong product mindset.
Responsibilities:
AI Agent Design & Integration
Design and implement autonomous or semi-autonomous agents using LLMs (e.g., OpenAI, Anthropic, open-source models).
Work with prompt engineering, RAG pipelines, and tool/plugin integrations to enable agents to interact with internal and external systems.
Build scalable agent runtimes and orchestration layers (e.g., LangChain, Semantic Kernel, ReAct-based agents).
Fullstack Product Development
Own full-stack features end-to-end: from backend APIs and data models to React-based frontend interfaces.
Integrate AI/agent capabilities into customer-facing products with clean UX and measurable performance.
Collaborate closely with design, product, and data teams to bring ideas from concept to production.
Systems & Infrastructure
Build and maintain backend services and pipelines that support AI agents, including vector search, embeddings, function calling, and observability.
Optimize inference flows for performance and cost, potentially using streaming, caching, or local model inference.
Ensure systems are secure, reliable, and compliant with InfoSec standards.
Experimentation & Continuous Improvement
Rapidly prototype and iterate on new AI capabilities and user experiences.
Analyze performance and usage metrics to drive product and model improvements.
Stay up to date with the evolving AI toolchain and emerging agent architectures.
Requirements:
8+ years of fullstack development experience with strong skills in TypeScript/JavaScript, React, and Python (or Node/Go for backend).
Solid understanding of LLM APIs, agent frameworks (e.g., LangChain, AutoGPT, CrewAI), or custom AI pipelines- Advantage
Experience with modern cloud infrastructure (e.g., AWS, GCP, Docker, CI/CD).
Familiarity with vector databases (e.g., Pinecone, Weaviate, FAISS) and retrieval-augmented generation (RAG)- Advantage
Product-oriented mindset: you care deeply about building things that work well for users.
Bonus: experience with observability, feedback loops for AI agents, or embedded AI evaluation techniques.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8600287
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
05/03/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are seeking a Backend Team Lead to spearhead the development of ludeo.ai, our GenAI-powered product that enables users to generate interactive (gaming experiences) directly from prompts or video content. This is a high-impact leadership role at the intersection of backend architecture, multimodal AI, and real-time systems. You will architect and lead the AI engine that transforms unstructured inputs (text/video) into structured, interactive gaming playable moments.

What Youll Do

Lead & Mentor: Build and manage a high-performing backend/AI engineering team, drive architectural decisions, and foster rapid innovation while maintaining production-grade reliability.
Design AI-Native Systems: Architect scalable microservices powering complex AI workflows. Design and implement Retrieval-Augmented Generation (RAG) pipelines, embedding strategies, and vector database infrastructure (e.g., Pinecone, Weaviate, Milvus, PGVector). Optimize retrieval, prompt orchestration, latency, and cost.
Agentic Workflows: Design multi-agent systems using planner/executor/tool-calling patterns. Implement stateful, multi-step AI workflows with frameworks such as LangChain, CrewAI, AutoGen, or similar. Build evaluation, observability, and safety mechanisms for LLM systems.
Multimodal AI: Integrate multimodal models (vision + text) to understand video and translate it into structured form.
Scale & Infrastructure: Ensure robustness, security, and high availability on AWS/Kubernetes. Design distributed systems that handle real-time data and AI workloads efficiently.
Collaborate: Work closely with Product and Design to translate GenAI capabilities into stable, scalable production features.
Requirements:
Expreince leading engineering teams in fast-paced environments with strong ownership and architectural responsibility.
Backend Expertise: 6+ years of backend development experience with deep expertise in Node.js and microservices. Strong distributed systems and API design experience.
GenAI Systems Experience: Hands-on experience building production LLM systems. Proven experience with RAG architectures, vector databases, embedding pipelines, and prompt orchestration. Experience designing multi-step or agentic AI workflows.
Infrastructure: Strong experience with AWS and Kubernetes in production environments. Deep knowledge of SQL & NoSQL systems.
Communication: Ability to translate complex AI systems into clear product and business decisions.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8569784
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Senior Software Engineer, AI Platform
About the role
As a Senior Software Engineer - AI, youll design, build, and own production grade AI agents that operate at the core of our cloud security platform. Youll work on distributed, cloud native services that embed agentic AI workflows into our existing microservices architecture.
This role goes beyond building AI logic: youll be responsible for operating AI systems in production, ensuring they are observable, reliable, and continuously improving through systematic evaluation and data driven iteration.
On a typical day youll:
Design and implement cloud-native, distributed services that power our AI-driven security features
Build and maintain agentic AI systems that reason over large-scale cloud security data and interact with multiple internal services
Own AI agents in production, including deployment, monitoring, troubleshooting, and performance optimization
Implement observability for AI systems, including metrics, logging, tracing, and alerting for agent behavior, quality, latency, and cost
Develop continuous evaluation pipelines for agentic solutions, including offline testing, regression detection, and production feedback loops
Design and optimize RAG pipelines that operate reliably over high-volume, high-variance security data
Apply strong software engineering practices: clear APIs, clean abstractions, robust error handling, and scalable data flows
Lead services end to end - from design and implementation to deployment and long-term operation
Collaborate closely with Data Platform, Product, and Security Research teams to ensure AI behavior is correct, explainable, and trustworthy.
Requirements:
5+ years of professional software engineering experience building and operating production systems
Strong proficiency in Python & Typescript and experience designing backend services
Solid experience building cloud-native, distributed systems in a microservices architecture
Hands-on experience building, deploying, and maintaining AI systems in production
Proven hands-on experience building AI systems using LLM and agentic frameworks in production
Practical experience with agentic AI workflows, including tool use, multi-step reasoning, and orchestration
Experience implementing observability and monitoring for complex systems (metrics, logs, traces)
Experience designing or working with evaluation frameworks for AI systems (quality, drift, latency, cost)
Ability to reason about tradeoffs and continuously improve systems based on real-world data
Big advantage
Experience evaluating AI systems in high-stakes domains (security, reliability, correctness)
Background in cloud security, cybersecurity, or large-scale SaaS platforms
Familiarity with RAG evaluation techniques, prompt versioning, and regression testing
Experience operating AI-enabled services at scale in AWS or similar cloud environments.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8588579
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
4 ימים
Location: Tel Aviv-Yafo
Job Type: Full Time
We're hiring for a new AI Engineering team in Tel Aviv, and you would be the first infrastructure hire. You will own the platform layer for AI agents the team builds: deployment architecture, observability, and production reliability.
The team's first two projects: an agent that automates internal governance processes (vendor reviews, security questionnaires, tool provisioning), and an agent that helps engineering teams prepare for architecture reviews. Both integrate with external APIs (LLM providers, OneTrust, ServiceNow), handle structured decision logic, and manage sensitive data flows with audit requirements.
Highlights
- Greenfield, but with real constraints. You're building on Azure/AWS with enterprise security requirements. The challenge is designing deployment and observability for LLM-backed services. You need to track output quality, cost per invocation, and model drift.
- Enterprise complexity, startup autonomy. Ownership and greenfield environment of a startup, with the integration challenges of a Fortune 200: connecting AI services to real enterprise systems.
- More than infrastructure. Your core is SRE, but you'll also write agent code in TypeScript and Python, work with data pipelines, and ship features alongside the team.
What the Work Looks Like
AI Service Infrastructure - Design and maintain deployment and release infrastructure for AI agents. The stack is cloud-native (Azure/AWS), with services that call LLM APIs, connect to enterprise systems, and handle structured data.
Observability & Reliability - Build monitoring and observability for AI services. Ensure model response quality doesn't degrade silently by tracking errors, logging cost spikes, and monitoring upstream API changes.
Security & Compliance - These agents handle sensitive workflows with elevated security requirements. You will work with our company's security team on standards, but you own how they're implemented in the infrastructure.
Developer Experience - Create tooling that makes it easy for the team to build, test, and deploy. The patterns you set become the team's defaults.
Requirements:
Required:
- 5+ years in SRE, platform engineering, DevOps, or infrastructure roles, with experience owning infrastructure end-to-end
- Strong experience with cloud platforms (Azure or AWS), containerization (Docker, Kubernetes), and CI/CD pipelines
- Infrastructure-as-code experience (Terraform, CDK, or CloudFormation)
- Monitoring and observability (Datadog, Splunk, CloudWatch, or similar)
- Infrastructure fundamentals: Linux, networking, security
- Incident management experience: on-call, production incidents, post-mortems
- Comfortable working independently with broad ownership and high accountability
- Strong written and verbal English for async collaboration with distributed teams
Preferred:
- Experience with AI/ML infrastructure: model serving, LLM API integration, vector databases, or evaluation pipelines
- Comfortable writing production code in TypeScript or Python, not just scripts
- Experience building self-service developer tooling or internal platforms
- Cost optimization for cloud and API-based workloads
- Security engineering experience, especially in enterprise or compliance-heavy environments.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8600507
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
30/03/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a highly motivated AI Full stack Engineer with GenAI background in production to join our team and help us shape the future of the Agentic engineering platform (AEP).
What youll do:
At our company, were a platform by developers, for developers. Your role will encompass end-to-end design, implementation, and daily feature delivery across both backend and frontend systems.
You will:
Implement high scale AI-powered features deeply integrated into our platform
Design and build production-grade backend systems serving a wide and growing user base
Build agent-based workflows using frameworks such as AI SDK
Integrate LLMs into real production systems with attention to reliability, latency, observability, and cost
Work across frontend (React + TypeScript) and backend (NodeJS, Python, Go) to deliver complete AI-driven user experiences
Own features end-to-end: design, implementation, testing, deployment, and monitoring
Help define standards and best practices around AI reliability and evaluation
Contribute to technical planning, mentor teammates, and help recruit top talent
Develop retrieval-augmented generation (RAG) pipelines over structured and unstructured data
Our stack includes React + TypeScript on the frontend, and NodeJS + TypeScript, Python, and Golang on the backend, and Vercels AI-SDK + AWS Bedrock + Azure OpenAI for GenAI. We use Kafka + Kafka Connect, Redis, PostgreSQL, MongoDB and other modern infrastructure components.
Requirements:
5+ years of professional software engineering experience
Experience in NodeJS + TypeScript
Strong experience designing and developing complex systems from design to production
Experience dealing with scale and performance-related challenges
Experience building or integrating AI/LLM-powered applications in production or meaningful production systems
Experience building agent workflows and tool integrations
Ability to think critically about model limitations, hallucinations, latency, and cost tradeoffs
A collaborative team player with a can-do approach
Strong written and verbal communication skills in English and Hebrew
Advantages:
Experience with AWS or other cloud platforms
Experience with vercels AI SDK
Experience with embeddings, vector databases, or semantic search
Expierence with AWS Bedrock / Azure Open-AI
Experience building tool-using agents or workflow engines
Experience with AI evaluation, observability, and monitoring
Experience in DevOps-related tools
Experience with PostgreSQL, Kafka, DocumentDB, OpenSearch, Redis.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8597066
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
05/03/2026
Location: Tel Aviv-Yafo
Job Type: Full Time
We are seeking a Backend Engineer to join the development of ludeo.ai, our GenAI-powered product that enables users to generate interactive (gaming experiences) directly from prompts or video content. This is a high-impact role at the intersection of backend architecture, multimodal AI, and real-time systems. You will contribute to the AI engine that transforms unstructured inputs (text/video) into structured, interactive gaming playable moments.

What Youll Do

Design AI-Native Systems: Design and implement scalable microservices powering complex AI workflows. Design and implement Retrieval-Augmented Generation (RAG) pipelines, embedding strategies, and vector database infrastructure (e.g., Pinecone, Weaviate, Milvus, PGVector). Optimize retrieval, prompt orchestration, latency, and cost.
Agentic Workflows: Design and implement multi-agent systems using planner/executor/tool-calling patterns. Implement stateful, multi-step AI workflows with frameworks such as LangChain, CrewAI, AutoGen, or similar. Build evaluation, observability, and safety mechanisms for LLM systems.
Multimodal AI: Integrate multimodal models (vision + text) to understand video and translate it into structured form.
Scale & Infrastructure: Ensure robustness, security, and high availability on AWS/Kubernetes. Contribute to distributed systems that handle real-time data and AI workloads efficiently.
Collaborate: Work closely with Product and Design to translate GenAI capabilities into stable, scalable production features.
Requirements:
Strong Python proficiency, particularly in AI/ML production environments
Hands-on experience with multimodal LLMs (vision-language models) and processing pipelines for image/video + text
Experience designing autonomous or semi-autonomous AI systems (planner/executor architectures, tool-calling, long-running agents)
Experience evaluating and benchmarking LLM systems (quality, hallucination mitigation, latency, cost optimization)
Strong DevOps capabilities including Docker, CI/CD pipelines, and deploying AI services/models to production
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8569780
סגור
שירות זה פתוח ללקוחות VIP בלבד