דרושים » הנדסה » AI & LLM Data Analyst (Customer-Facing)

משרות על המפה
 
בדיקת קורות חיים
VIP
הפוך ללקוח VIP
רגע, משהו חסר!
נשאר לך להשלים רק עוד פרט אחד:
 
שירות זה פתוח ללקוחות VIP בלבד
AllJObs VIP
כל החברות >
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
24/03/2026
Location: Tel Aviv-Yafo
Job Type: Full Time
Were looking for a Data Analyst to join the Data for AI team. This is a hands-on, customer-facing role focused on working with leading AI companies to turn real-world data into inputs that support model development and evaluation.
Youll collaborate closely with external AI teams and internal engineering and product partners to deliver data-driven solutions for specific AI use cases. The work is fast-paced, technical, and often open-ended, requiring comfort with large datasets, ambiguous requirements, and end-to-end ownership.
What does the day-to-day looks like:
Own end-to-end delivery of data solutions for AI use cases, from understanding model and product requirements to analysis, implementation, quality, and automation
Work hands-on with large, raw datasets to create high-quality data inputs that support model training, evaluation, and iteration
Apply strong quantitative analysis and data exploration skills to assess coverage, quality, and behavior of data used in AI systems
Build scripts, analyses, and reusable components in Python and SQL to support scalable and repeatable workflows
Collaborate closely with Engineering to ensure solutions are reliable, scalable, and production-ready
Partner directly with external AI teams and internal stakeholders to translate open-ended questions into concrete data outputs.
Requirements:
4+ years of hands-on experience working with large-scale data using SQL and Spark or BigQuery
Strong Python skills for data analysis, scripting, and building reusable workflows
Experience working with raw, imperfect data and turning it into reliable, high-quality outputs
Strong analytical and problem-solving skills, with the ability to break down open-ended or ambiguous requirements
Ability to take end-to-end ownership of data projects, from exploration to delivery
Some hands-on experience with LLM-based systems, such as running inference via APIs, experimenting with prompts, or participating in basic evaluation or testing workflows
Clear communication skills in English and experience working directly with external stakeholders
Nice to have:
Deeper hands-on experience with LLMs in production or experimentation, for example prompt engineering, batch inference, or structured evaluation using APIs such as OpenAI, Anthropic, or similar providers
Familiarity with agent frameworks or orchestration layers (for example LangChain, LlamaIndex)
Experience with LLM evaluation or monitoring workflows, including offline evals, prompt regression testing, or tools such as LangSmith, Weights & Biases, TruLens, or Ragas
Experience experimenting with open-source or local models (for example via Ollama, vLLM, or Hugging Face tooling)
Familiarity with cloud-based data infrastructure, including AWS.
This position is open to all candidates.
 
Hide
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8590074
סגור
שירות זה פתוח ללקוחות VIP בלבד
משרות דומות שיכולות לעניין אותך
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
30/03/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
Required ML Data Engineer
Israel: Tel Aviv/ Hybrid (Israel)
R&D | Full Time | Job Id: 24792
Key Responsibilities
Your Impact & Responsibilities:
As a Data Engineer - AI Technologies, you will be responsible for building and operating the data foundation that enables our LLM and ML research: from ingestion and augmentation, through labeling and quality control, to efficient data delivery for training and evaluation.
You will:
Own data pipelines for LLM training and evaluation
Design, build and maintain scalable pipelines to ingest, transform and serve large-scale text, log, code and semi-structured data from multiple products and internal systems.
Drive data augmentation and synthetic data generation
Implement and operate pipelines for data augmentation (e.g., prompt-based generation, paraphrasing, negative sampling, multi-positive pairs) in close collaboration with ML Research Engineers.
Build tagging, labeling and annotation workflows
Support human-in-the-loop labeling, active learning loops and semi-automated tagging. Work with domain experts to implement tools, schemas and processes for consistent, high-quality annotations.
Ensure data quality, observability and governance
Define and monitor data quality checks (coverage, drift, anomalies, duplicates, PII), manage dataset versions, and maintain clear documentation and lineage for training and evaluation datasets.
Optimize training data flows for efficiency and cost
Design storage layouts and access patterns that reduce training time and cost (e.g., sharding, caching, streaming). Work with ML engineers to make sure the right data arrives at the right place, in the right format.
Build and maintain data infrastructure for LLM workloads
Work with cloud and platform teams to develop robust, production-grade infrastructure: data lakes / warehouses, feature stores, vector stores, and high-throughput data services used by training jobs and offline evaluation.
Collaborate closely with ML Research Engineers and security experts
Translate modeling and security requirements into concrete data tasks: dataset design, splits, sampling strategies, and evaluation data construction for specific security use.
Requirements:
3+ years of hands-on experience as a Data Engineer or ML/Data Engineer, ideally in a product or platform team.
Strong programming skills in Python and experience with at least one additional language commonly used for data / backend (e.g., SQL, Scala, or Java).
Solid experience building ETL / ELT pipelines and batch/stream processing using tools such as Spark, Beam, Flink, Kafka, Airflow, Argo, or similar.
Experience working with cloud data platforms (e.g., AWS, GCP, Azure) and modern data storage technologies (object stores, data warehouses, data lakes).
Good understanding of data modeling, schema design, partitioning strategies and performance optimization for large datasets.
Familiarity with ML / LLM workflows: train/validation/test splits, dataset versioning, and the basics of model training and evaluation (you dont need to be the primary model researcher, but you understand what the models need from the data).
Strong software engineering practices: version control, code review, testing, CI/CD, and documentation.

Ability to work independently and in collaboration with ML engineers, researchers and security experts, and to translate high-level requirements into concrete data engineering tasks. 
Nice to Have 
Experience supporting LLM or NLP workloads, including dataset construction for pre-training / fine-tuning, or retrieval-augmented generation (RAG) pipelines. 
Familiarity with ML tooling such as experiment tracking (e.g., Weights & Biases, MLflow) and ML-focused data tooling (feature stores, vector databases). 
Background in security / cyber domains (logs, alerts, incidents, SOC workflows) or other high-volume, high-variance data environments. 
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8597480
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
05/04/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Were looking for a highly motivated AI Developer to help design, build, and deploy intelligent agentic systems across our product ecosystem. In this role, you'll work at the intersection of machine learning, backend systems, and modern frontend technologies to deliver AI-first features that feel magical to users.
This is a hands-on, cross-functional role ideal for engineers who love building full-fledged features-from data pipelines and LLM orchestration to intuitive UI experiences-with a strong product mindset.
Responsibilities:
AI Agent Design & Integration
Design and implement autonomous or semi-autonomous agents using LLMs (e.g., OpenAI, Anthropic, open-source models).
Work with prompt engineering, RAG pipelines, and tool/plugin integrations to enable agents to interact with internal and external systems.
Build scalable agent runtimes and orchestration layers (e.g., LangChain, Semantic Kernel, ReAct-based agents).
Fullstack Product Development
Own full-stack features end-to-end: from backend APIs and data models to React-based frontend interfaces.
Integrate AI/agent capabilities into customer-facing products with clean UX and measurable performance.
Collaborate closely with design, product, and data teams to bring ideas from concept to production.
Systems & Infrastructure
Build and maintain backend services and pipelines that support AI agents, including vector search, embeddings, function calling, and observability.
Optimize inference flows for performance and cost, potentially using streaming, caching, or local model inference.
Ensure systems are secure, reliable, and compliant with InfoSec standards.
Experimentation & Continuous Improvement
Rapidly prototype and iterate on new AI capabilities and user experiences.
Analyze performance and usage metrics to drive product and model improvements.
Stay up to date with the evolving AI toolchain and emerging agent architectures.
Requirements:
8+ years of fullstack development experience with strong skills in TypeScript/JavaScript, React, and Python (or Node/Go for backend).
Solid understanding of LLM APIs, agent frameworks (e.g., LangChain, AutoGPT, CrewAI), or custom AI pipelines- Advantage
Experience with modern cloud infrastructure (e.g., AWS, GCP, Docker, CI/CD).
Familiarity with vector databases (e.g., Pinecone, Weaviate, FAISS) and retrieval-augmented generation (RAG)- Advantage
Product-oriented mindset: you care deeply about building things that work well for users.
Bonus: experience with observability, feedback loops for AI agents, or embedded AI evaluation techniques.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8600287
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a Senior AI Engineer to join our Cybersecurity team in Tel Aviv. You will design, build, and productionize LLM-powered applications, multi-agent systems, and MLOps infrastructure that power our company's next-generation cybersecurity capabilities. This is a high-impact, hands-on role at the intersection of applied AI, agentic systems, and network securit
What You'll Do
Design and develop LLM-powered security features and internal AI tools, including RAG pipelines, multi-agent workflows, and prompt-engineered systems tailored for cybersecurity use cases
Architect and operate multi-agent systems in production - including agent orchestration, inter-agent communication, task delegation, and failure handling at scale
Build robust agent monitoring and observability pipelines: tracing agent execution, detecting drift or failure, alerting on anomalous behavior, and maintaining agent reliability SLAs
Build and maintain scalable MLOps infrastructure: model serving, evaluation frameworks, experiment tracking, and CI/CD for ML models
Work with internal datasets (network telemetry, security logs, threat intelligence) to fine-tune and adapt foundation models for domain-specific detection and response tasks
Partner with the Cybersecurity, R&D, and infrastructure teams to define AI-driven security features and deliver them end-to-end
Establish best practices for model observability, safety, and responsible AI deployment within the organization
Stay current with the fast-moving LLM/GenAI and agentic AI ecosystem and evaluate emerging frameworks, models, and tools for adoption.
Requirements:
Must-Have
5-8 years of software engineering experience, with at least 2-3 years focused on AI/ML engineering
Hands-on experience building production-grade LLM applications - RAG, agents, tool use, or fine-tuning
Proven experience designing and running multi-agent systems in production: orchestration patterns, agent state management, retries, and graceful degradation
Experience monitoring and observing AI agents in production - execution tracing, latency tracking, failure detection, and alerting (e.g., LangSmith, Arize, custom observability stacks)
Proficiency with agentic frameworks: LangChain, LangGraph, and/or AWS Bedrock AgentCore
Strong Python skills and comfort working across the full AI application stack
Experience designing and operating MLOps pipelines (model versioning, deployment, monitoring)
Solid understanding of transformer-based models, embeddings, and vector databases (e.g., Pinecone, Weaviate, pgvector)
Comfortable working in cloud environments (AWS, GCP, or Azure) and containerized deployments (Docker, Kubernetes)
Strong problem-solving skills and ability to work autonomously in a fast-paced environment
Nice-to-Have
Background in cybersecurity - threat detection, SIEM, SOC automation, or security data analysis - a significant plus for this role
Familiarity with networking concepts (SDN, cloud-native networking, BGP, telemetry)
Experience with model evaluation and benchmarking (LLM-as-judge, RAGAS, or custom eval harnesses)
Exposure to MCP (Model Context Protocol) for tool-augmented agentic workflows
Prior experience in enterprise SaaS, networking, or telecom domains
Publications, open-source contributions, or projects in the LLM/GenAI or agentic AI space
Our Stack
Python PyTorch OpenAI / Anthropic APIs LangChain LangGraph AWS Bedrock AgentCore LangSmith Kubernetes Kafka Elasticsearch AWS PostgreSQL GitHub Jira Confluence.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8595648
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
10/04/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a Data Analyst with strong Python and engineering capabilities to join our Data team.

This is not a traditional analytics role - it combines data analysis with hands-on development. You will work extensively with Python as part of your day-to-day work, building scalable data processes, evaluation frameworks, and classification logic that directly impacts our product.

Our team works on large-scale data systems, leveraging ML capabilities such as NER and LLMs, alongside robust engineering and analytics to deliver insights across millions of files and tables daily.


Responsibilities
Design and implement data classification logic to expand coverage of sensitive data types, owning the process end-to-end - from research and analysis to production deployment
Develop Python-based workflows and tools for data processing, evaluation, and automation
Build and maintain data pipelines and monitoring systems to track product performance and data quality
Create automated frameworks to evaluate and benchmark model performance
Collaborate closely with R&D and take part in discussions around AI/ML solutions and product direction
Requirements:
2+ years of experience in a data-focused role within a B2B SaaS company
Strong hands-on experience with Python - writing code as part of daily work
Experience with SQL and working with large-scale data
Experience building data pipelines / ETL processes (Airflow, Dagster, or similar)
Familiarity with Spark or distributed data processing - an advantage
Understanding of software development best practices (Git, CI/CD, testing)
Degree in a quantitative field (Computer Science, Engineering, Mathematics, Statistics)
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8604543
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
13/04/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Were looking for an Applied Data Scientist to join one of our product squads. Youll design, build, and deploy data-driven solutions that combine machine learning, statistical methods, and SQL/rules-based decision logic to power autonomous supply chain intelligence platform. Youll work closely with data science, engineering, product, and supply chain experts and own solutions end-to-end-from problem definition to production monitoring and iteration.
Responsibilities:
Deliver data science solutions end-to-end within a product squad: problem framing → data prep/labeling → modeling → deployment support → monitoring → iteration
Build, train, and improve ML models for supply chain use cases (e.g., inventory risk prediction, demand anomalies, root-cause analysis)
Define success metrics and evaluation plans with support from senior DS/PM; run error analysis and document learnings
Work with stakeholders to create and maintain ground truth (label definitions, labeling workflows, QA checks, feedback loops)
Implement hybrid decision logic by combining ML outputs with statistical methods and SQL/rules-based logic for robustness and explainability
Analyze large, multi-source operational datasets to identify trends, anomalies, and drivers of performance
Collaborate with software engineers to productionize solutions (batch and/or real-time), including testing, logging, and basic monitoring
Monitor deployed models/rules, investigate performance issues (data quality, drift, edge cases), and iterate based on outcomes
Contribute to team practices: reproducible notebooks/code, documentation, and experiment tracking
Requirements:
MSc in Computer Science, Data Science, Mathematics, Statistics, Engineering, (or equivalent practical experience)
3+ years of experience in applied data science / ML in a product environment (or equivalent practical experience)
Strong Python skills and experience with common DS libraries (pandas, NumPy, scikit-learn); familiarity with PyTorch/TensorFlow is a plus
Solid SQL skills (joins, aggregations, window functions) and comfort working with production data in a warehouse/lake
Experience building predictive or anomaly detection models and performing rigorous evaluation (baselines, cross-validation where relevant, error analysis)
Ability to translate business questions into measurable metrics and a clear analytical plan (with guidance when needed)
Experience working with messy real-world data: data validation, debugging pipelines, and collaborating on labeling/ground truth
Familiarity with taking models to production: packaging/hand-off to engineers, versioning, and understanding monitoring/drift concepts
Strong communication and collaboration skills with engineering, product, and domain experts; comfortable receiving feedback and iterating fast
Nice to Have (Advantages)
Experience designing or deploying agentic workflows, AI agents, or multi-step decision systems
Cloud + Docker + production engineering practices (CI/CD, testing, monitoring)
Experience publishing academic or applied research (peer-reviewed papers, conference publications, technical whitepapers, or open research work)
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8608560
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
09/04/2026
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Senior AI Engineer - Applied AI Engineering Group
The Dream Job
It starts with you - an engineer driven to build the agentic AI platform that turns LLMs into reliable, production-grade capabilities. You care about clean APIs, well-defined service boundaries, and systems that teams can build on with confidence. Dream is AI-first across the board - every team builds and operates agents. You'll architect and ship the platform that makes this possible: agent orchestration frameworks, LLM gateways, evaluation pipelines, tool-calling infrastructure, and retrieval systems. Without this platform, agents don't ship - you own the layer that turns AI research into Sovereign AI products, deployed across cloud and on-prem environments.
If you want to make a meaningful impact, join our mission and build the agentic AI platform that drives Sovereign AI products - this role is for you.
The Dream-Maker Responsibilities
Design and build agentic systems - single and multi-agent workflows with planning, memory, context engineering, and tool use - for both internal automation and product-facing autonomous capabilities operating over long time horizons.
Build and operate the AI platform layer - LLM gateways, prompt management, structured output handling, tool-calling infrastructure, and cost/latency optimization - deployed on Kubernetes, consumed by every team for their agentic work.
Own the agent framework layer - orchestration primitives, execution environments, state management, and sandboxed tool execution - giving every team the building blocks to create and operate their own agents.
Build evaluation infrastructure that gives teams confidence in agent behavior - automated LLM and agent evals for quality, correctness, safety, latency, cost, and regressions, including human-in-the-loop oversight for mission-critical workflows.
Productionize and harden backend services (APIs, gRPC, async workers) that integrate LLMs - with proper error handling, retries, circuit breakers, and high-availability patterns.
Own RAG pipelines and retrieval systems - indexing, chunking, embedding, vector database management, filtering, and relevance tuning for production retrieval.
Optimize performance and cost across the AI stack - model routing, caching, batching, and inference cost management.
Ship shared tooling - libraries, SDKs, agent templates, and documentation - while working closely with ML Platform, Data Platform, DevOps, and other teams across the Applied AI Engineering group. Own architecture, documentation, and operations end-to-end.
דרישות:
5+ years in backend or distributed systems engineering, with 2+ years focused on production systems that integrate AI/ML models or LLMs.
Engineering craft - Strong Python, Go, or Java, system architecture, API design, testing, and secure coding practices.
Agentic systems - Experience designing and building agent orchestration, tool-use systems, and autonomous workflows; familiarity with frameworks like LangGraph or similar, or having built equivalent from scratch
Backend engineering - Experience building production APIs and services (FastAPI or similar); async programming, service architecture, high-availability, and reliability patterns (retries, circuit breakers, backpressure)
LLM integration - Hands-on experience integrating LLMs via SDKs and APIs; context engineering, structured outputs, tool calling, and model routing
RAG & retrieval - Experience with embedding pipelines, vector databases (e.g., Milvus, Qdrant, Pinecone), chunking strategies, and relevance tuning
Evaluation & observability - Experience designing LLM and agent evals, monitoring AI system quality, and building observability for non-deterministic systems
Nice to Have:
Platform & infra - Kubernetes, AWS, Terraform or similar IaC, CI/CD, container orchestration, deploying and operating production services
Experience with MCP or similar tool-use protocols for agent-to-service communication
Hands-on ML experience - המשרה מיועדת לנשים ולגברים כאחד.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8603620
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
we are looking for a Data Engineer.
As a Data Engineer, youll collaborate with top-notch engineers and data scientists to elevate our platform to the next level and deliver exceptional user experiences. Your primary focus will be on the data engineering aspects-ensuring the seamless flow of high-quality, relevant data to train and optimize content models, including GenAI foundation models, supervised fine-tuning, and ore.
Youll work closely with teams across the company to ensure the availability of high-quality data from ML platforms, powering decisions across all departments. With access to petabytes of data through MySQL, Snowflake, Cassandra, S3, and other platforms, your challenge will be to ensure that this data is applied even more effectively to support business decisions, train and monitor ML models and improve our products.
Key Job Responsibilities and Duties:
Rapidly developing next-generation scalable, flexible, and high-performance data pipelines.
Dealing with massive textual sources to train GenAI foundation models.
Solving issues with data and data pipelines, prioritizing based on customer impact.
End-to-end ownership of data quality in our core datasets and data pipelines.
Experimenting with new tools and technologies to meet business requirements regarding performance, scaling, and data quality.
Providing tools that improve Data Quality company-wide, specifically for ML scientists.
Providing self-organizing tools that help the analytics community discover data, assess quality, explore usage, and find peers with relevant expertise.
Acting as an intermediary for problems, with both technical and non-technical audiences.
Promote and drive impactful and innovative engineering solutions
Technical, behavioral and interpersonal competence advancement via on-the-job opportunities, experimental projects, hackathons, conferences, and active community participation
Collaborate with multidisciplinary teams: Collaborate with product managers, data scientists, and analysts to understand business requirements and translate them into machine learning solutions. Provide technical guidance and mentorship to junior team members.
20718
Requirements:
Bachelors or masters degree in computer science, Engineering, Statistics, or a related field.
Minimum of 3 years of experience as a Data Engineer or a similar role, with a consistent record of successfully delivering ML/Data solutions
You have built production data pipelines in the cloud, setting up data-lake and server-less solutions; ‌ you have hands-on experience with schema design and data modeling and working with ML scientists and ML engineers to provide production level ML solutions.
You have experience designing systems E2E and knowledge of basic concepts (lb, db, caching, NoSQL, etc)
Strong programming skills in languages such as Python and Java.
Experience with big data processing frameworks such, Pyspark, Apache Flink, Snowflake or similar frameworks.
Demonstrable experience with MySQL, Cassandra, DynamoDB or similar relational/NoSQL database systems.
Experience with Data Warehousing and ETL/ELT pipelines
Experience in data processing for large-scale language models like GPT, BERT, or similar architectures - an advantage.
Proficiency in data manipulation, analysis, and visualization using tools like NumPy, pandas, and matplotlib - an advantage.
Experience with experimental design, A/B testing, and evaluation metrics for ML models - an advantage.
Experience of working on products that impact a large customer base - an advantage.
Excellent communication in English; written and spoken.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8627494
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
30/03/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
Required Senior ML Research Engineer
Israel: Tel Aviv/ Hybrid
R&D | Full Time | Job Id: 24793
Your Impact & Responsibilities:
As a Senior ML Research Engineer, you will be responsible for the end-to-end lifecycle of large language models: from data definition and curation, through training and evaluation, to providing robust models that can be consumed by product and platform teams.
Own training and fine-tuning of LLMs / seq2seq models: Design and execute training pipelines for transformer-based models (encoder-decoder, decoder-only, retrievalaugmented, etc.), and fine-tune open-source LLMs -specific data (security content, logs, incidents, customer interactions).
Apply advanced LLM training techniques such as instruction tuning, preference / contrastive learning, LoRA / PEFT, continual pre-training, and domain adaptation where appropriate.
Work deeply with data: define data strategies with product, research and domain experts; build and maintain data pipelines for collecting, cleaning, de-duplicating and labeling large-scale text, code and semi-structured data; and design synthetic data generation and augmentation pipelines.
Build robust evaluation and experimentation frameworks: define offline metrics for LLM quality (task-specific accuracy, calibration, hallucination rate, safety, latency and cost); implement automated evaluation suites (benchmarks, regression tests, redteaming scenarios); and track model performance over time.
Scale training and inference: use distributed training frameworks (e.g. DeepSpeed, FSDP, tensor/pipeline parallelism) to efficiently train models on multi-GPU / multi-node clusters, and optimize inference performance and cost with techniques such as quantization, distillation and caching.
Collaborate closely with security researchers and data engineers to turn domain knowledge and threat intelligence into high-value training and evaluation data, and to expose your models through well-defined interfaces to downstream product and platform teams.
Requirements:
5+ years of hands-on work in machine learning / deep learning, including 3+ years focused on NLP / language models.
Proven track record of training and fine-tuning transformer-based models (BERT-style, encoder-decoder, or LLMs), not just consuming hosted APIs.
Strong programming skills in Python and at least one major deep learning framework (PyTorch preferred; TensorFlow).
Solid understanding of transformer architectures, attention mechanisms, tokenization, positional encodings, and modern training techniques.
Experience building data pipelines and tools for large-scale text / log / code processing (e.g. Spark, Beam, Dask, or equivalent frameworks).
Practical experience with ML infrastructure, such as experiment tracking (Weights & Biases, MLflow or similar), job orchestration (Airflow, Argo, Kubeflow, SageMaker, etc.), and distributed training on multi-GPU systems.
Strong software engineering practices: version control, code review, testing, CI/CD, and documentation.
Ability to own research and engineering projects end-to-end: from idea, through prototype and controlled experiments, to models ready for integration by product and platform teams.
Good communication skills and the ability to work closely with non-ML stakeholders (security experts, product managers, engineers).
Nice to have:
Experience with RLHF / preference optimization, safety alignment, or other humanfeedback-in-the-loop approaches to training LLMs.
Experience with retrieval-augmented generation (RAG), dense retrieval, vector databases, and embedding training.
Background in security / cyber domains such as threat detection, malware analysis, logs, or SOC tools.
Experience with multilingual models (e.g., Hebrew + English) and cross-lingual training.
Experience in a product environment where models must meet reliability, scale, and cost constraints.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8597461
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time and English Speakers
we are looking for a Senior Data Engineer I.
As a Senior Data Engineer, youll collaborate with top-notch engineers and data scientists to elevate our platform to the next level and deliver exceptional user experiences. Your primary focus will be on the data engineering aspects-ensuring the seamless flow of high-quality, relevant data to train and optimize content models, including GenAI foundation models, supervised fine-tuning, and more.
Youll work closely with teams across the company to ensure the availability of high-quality data from ML platforms, powering decisions across all departments. With access to petabytes of data through MySQL, Snowflake, Cassandra, S3, and other platforms, your challenge will be to ensure that this data is applied even more effectively to support business decisions, train and monitor ML models and improve our products.
Key Job Responsibilities and Duties:
Rapidly developing next-generation scalable, flexible, and high-performance data pipelines.
Dealing with massive textual sources to train GenAI foundation models.
Solving issues with data and data pipelines, prioritizing based on customer impact.
End-to-end ownership of data quality in our core datasets and data pipelines.
Experimenting with new tools and technologies to meet business requirements regarding performance, scaling, and data quality.
Providing tools that improve Data Quality company-wide, specifically for ML scientists.
Providing self-organizing tools that help the analytics community discover data, assess quality, explore usage, and find peers with relevant expertise.
Acting as an intermediary for problems, with both technical and non-technical audiences.
Promote and drive impactful and innovative engineering solutions
Technical, behavioral and interpersonal competence advancement via on-the-job opportunities, experimental projects, hackathons, conferences, and active community participation
Collaborate with multidisciplinary teams: Collaborate with product managers, data scientists, and analysts to understand business requirements and translate them into machine learning solutions. Provide technical guidance and mentorship to junior team members.
21679
Requirements:
Bachelors or masters degree in computer science, Engineering, Statistics, or a related field.
Minimum of 6 years of experience as a Data Engineer or a similar role, with a consistent record of successfully delivering ML/Data solutions.
You have built production data pipelines in the cloud, setting up data-lake and server-less solutions; ‌ you have hands-on experience with schema design and data modeling and working with ML scientists and ML engineers to provide production level ML solutions.
You have experience designing systems E2E and knowledge of basic concepts (lb, db, caching, NoSQL, etc)
Strong programming skills in languages such as Python and Java.
Experience with big data processing frameworks such, Pyspark, Apache Flink, Snowflake or similar frameworks.
Demonstrable experience with MySQL, Cassandra, DynamoDB or similar relational/NoSQL database systems
Experience with Data Warehousing and ETL/ELT pipelines
Experience in data processing for large-scale language models like GPT, BERT, or similar architectures - an advantage.
Proficiency in data manipulation, analysis, and visualization using tools lke NumPy, pandas, and matplotlib - an advantage.
Experience with experimental design, A/B testing, and evaluation metrics for ML models - an advantage.
Experience of working on products that impact a large customer base - an advantage.
Excellent communication in English; written and spoken.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8627496
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
20/04/2026
Location: Tel Aviv-Yafo
Job Type: Full Time and Multilingual
We are looking for a native French speaker with a high level of proficiency in English to join our Annotation and Analysis team at the as a Data Annotator and Analyst.
In this role, you will work with large volumes of data in both French and English, annotating and analyzing datasets used to evaluate and train machine learning models, including large language models (LLMs). Your work will involve reviewing model outputs and preparing high-quality training data that supports the development and continuous improvement of AI-driven products. In addition, as part of the job, you will conduct complex analysis on the annotated data and provide insights to the other stakeholders.
This position is ideal for someone passionate about linguistics and data, and who enjoys working on detail-oriented tasks that directly impact model quality. As part of a collaborative, cross-functional team, you will work closely with Data Scientists and Product Managers to ensure that datasets are accurately labeled and annotation standards support the development of reliable machine learning systems.
Responsibilities :
Evaluate the quality of machine-translated text, with a focus on French language accuracy and fluency.
Annotate and assess AI model responses to support systematic evaluation and improvement of model performance.
Create high-quality annotated datasets used to train and evaluate LLMs and other NLP models.
Write and refine LLM prompts that support scalable human annotation workflows.
Collaborate with cross-functional teams, including Data Scientists and Product Managers, to communicate linguistic nuances and implications of French translations on model behavior and user experience.
Analyze annotated data and evaluation results to identify patterns, assess model quality, and surface insights for model improvement.
Requirements:
Native French speaker and high proficiency in English.
Strong attention to detail and ability to perform precise, consistent work across large volumes of data.
Interest in linguistics, language quality, and the interaction between language and AI systems.
Ability to follow detailed annotation guidelines and maintain high accuracy and consistency.
Ability to analyze annotated data and evaluation results to identify patterns and assess model performance.
Ability to work independently with strong communication skills.
Experience with using excel for data analysis
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8618803
סגור
שירות זה פתוח ללקוחות VIP בלבד