דרושים » תוכנה » Senior ML Data Engineer

משרות על המפה
 
בדיקת קורות חיים
VIP
הפוך ללקוח VIP
רגע, משהו חסר!
נשאר לך להשלים רק עוד פרט אחד:
 
שירות זה פתוח ללקוחות VIP בלבד
AllJObs VIP
כל החברות >
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 4 שעות
חברה חסויה
Location: Jerusalem
Job Type: Full Time
Required Senior ML Data Engineer
Jerusalem
Full time
The AI Engineering group builds modern infrastructure and solutions that improve how algorithms are developed.
We are a small, independent team of experienced engineers with a mix of skills in algorithms, software, and infrastructure. We work in a DevOps style and build cross-team solutions that support research and development of advanced perception algorithms.
Our flagship project is a unified AV dataset used to train and evaluate next-generation models. We take large volumes of multi-camera video, object labels, HD maps, and sensor data from across the organization, and turn it into a curated, high-quality training set - at scale.
We are looking for someone who brings ML and computer-vision depth to the team - someone who can help shape the intelligence layer that decides what data is worth training on.
What will your job look like:
Work collaboratively with shared ownership. Your focus area will be the curation and ML side of our data pipeline, but you will contribute across the full stack alongside the rest of the team.
Build and improve the curation pipeline - from vision-model embeddings and scene detection, through VLM-based scene analysis, to scoring, deduplication, and sampling that produces a balanced and diverse dataset.
Run and optimize GPU inference at scale (embedding extraction, VLM inference) across thousands of driving sessions using workflow orchestration.
Develop scoring and sampling strategies that ensure rare but important scenarios (night driving, adverse weather, hazardous situations) are well-represented in the final dataset.
Work with algorithm teams to understand what data gaps hurt model performance and translate those into curation criteria.
Build validation and diagnostics that measure dataset quality - not just pipeline health, but whether the data is actually good for training.
Contribute to the core dataset SDK, converter, and 3D-geometry tooling (camera projection, calibration, coordinate transforms).
Requirements:
4+ years in ML engineering, applied CV, or a similar role combining model work with production data systems.
Hands-on experience with vision models - embeddings, VLMs, or object detection/segmentation.
Strong Python and comfort with the PyData stack (NumPy, PyArrow, Pandas, DuckDB).
Experience building data or ML pipelines that run at scale (not just notebooks).
Solid understanding of 3D geometry and camera models - or the mathematical background to ramp up quickly.
Good understanding of LLM agents and agentic workflows, with genuine interest in applying them to data and engineering problems.
Ability to work across team boundaries with algorithm and infrastructure people.
Strong advantage:
Experience with autonomous-driving datasets or perception pipelines.
Familiarity with dataset curation techniques (active learning, hard-example mining, distribution balancing).
Experience with GPU inference serving (vLLM, Triton, TensorRT).
Familiarity with vector databases or columnar analytics (LanceDB, DuckDB).
Experience with workflow orchestration (Argo, Airflow, Kubeflow).
This position is open to all candidates.
 
Hide
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8636125
סגור
שירות זה פתוח ללקוחות VIP בלבד
משרות דומות שיכולות לעניין אותך
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 5 שעות
Location: Jerusalem
Job Type: Full Time
We are looking for someone who brings ML and computer-vision depth to the team - someone who can help shape the intelligence layer that decides what data is worth training on.
What will your job look like:
Work collaboratively with shared ownership. Your focus area will be the curation and ML side of our data pipeline, but you will contribute across the full stack alongside the rest of the team.
Build and improve the curation pipeline - from vision-model embeddings and scene detection, through VLM-based scene analysis, to scoring, deduplication, and sampling that produces a balanced and diverse dataset.
Run and optimize GPU inference at scale (embedding extraction, VLM inference) across thousands of driving sessions using workflow orchestration.
Develop scoring and sampling strategies that ensure rare but important scenarios (night driving, adverse weather, hazardous situations) are well-represented in the final dataset.
Work with algorithm teams to understand what data gaps hurt model performance and translate those into curation criteria.
Build validation and diagnostics that measure dataset quality - not just pipeline health, but whether the data is actually good for training.
Contribute to the core dataset SDK, converter, and 3D-geometry tooling (camera projection, calibration, coordinate transforms).
Requirements:
4+ years in ML engineering, applied CV, or a similar role combining model work with production data systems.
Hands-on experience with vision models - embeddings, VLMs, or object detection/segmentation.
Strong Python and comfort with the PyData stack (NumPy, PyArrow, Pandas, DuckDB).
Experience building data or ML pipelines that run at scale (not just notebooks).
Solid understanding of 3D geometry and camera models - or the mathematical background to ramp up quickly.
Good understanding of LLM agents and agentic workflows, with genuine interest in applying them to data and engineering problems.
Ability to work across team boundaries with algorithm and infrastructure people.
Strong advantage:
Experience with autonomous-driving datasets or perception pipelines.
Familiarity with dataset curation techniques (active learning, hard-example mining, distribution balancing).
Experience with GPU inference serving (vLLM, Triton, TensorRT).
Familiarity with vector databases or columnar analytics (LanceDB, DuckDB).
Experience with workflow orchestration (Argo, Airflow, Kubeflow).
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8635950
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
12/04/2026
חברה חסויה
Location: Jerusalem
Job Type: Full Time
We're looking for a Data Engineer with 4+ year experience to join our Data Engineering team and help us build and scale our production-grade data platform. You'll work on high-performance systems built on self-hosted ClickHouse, optimize complex data pipelines, and collaborate closely with Product, Analytics, and Infrastructure teams to deliver reliable, fast, and scalable data solutions.

This is a hands-on technical role where you'll have a significant impact on how we ingest, model, store, and serve data that powers our analytics and AI-driven products.
Youll play a key role in shaping the direction of our data platform and have meaningful ownership over critical components of our architecture.

What You'll Do:
Data Modeling & Architecture
Design and evolve data models that reflect business logic and support analytical use cases
Collaborate with the BI and Analytics teams to understand data requirements and translate them into efficient schemas
Performance Optimization
Optimize ClickHouse schemas, partitioning strategies, indexing, and compression
Profile and tune slow queries to improve performance and reduce costs
Implement systems that ensure data quality, consistency, and operational efficiency (e.g., deduplication, validation, anomaly detection)
Monitor pipeline health, data freshness, and query performance with appropriate alerting mechanisms
SQL Compiler Development
Develop and maintain the SQL Compiler layer that translates high-level queries into optimized ClickHouse execution plansImplement query optimization and rewriting strategies to improve performanceDebug and resolve compiler issues to ensure accurate and efficient query translation

Data Pipeline Development & Collaboration
Review and advise the Integration team on pipeline architecture, performance, and best practices.
Provide guidance on data modeling, schema design, and optimization for new data sources.
Troubleshoot and maintain existing pipelines when issues arise or optimization is needed
Ensure data freshness, reliability, and quality across all ingestion pipelines.
Collaboration & Support
Work closely with the Integration team to ensure smooth data ingestion from new sources.
Partner with Infrastructure to support high availability and disaster recovery
Support other teams across the company in accessing and using data effectively.
Requirements:
Excellent communication and collaboration skills
English at a high level, written and spoken required
Ability to work from our Jerusalem office (located in the Central Bus Station next to the train) 2 times a week (Monday & Wednesday) is required
Strong attention to detail, ownership mentality, and ability to work independently
Quick learner who can dive into new codebases, technologies, and systems independently
Hands-on mentality - not afraid to roll up your sleeves, dig into unfamiliar code, and work across the stack (including backend when needed)
4+ years of experience as a Data Engineer
Strong problem-solving skills for complex data challenges at scale - ability to debug performance issues, data inconsistencies, and system bottlenecks in high-volume environments
Experience with data modeling and schema design for analytical workloads
Strong proficiency in SQL and experience with complex analytical queries
Hands-on experience building and maintaining data pipelines (ETL/ELT)
Ability to troubleshoot and optimize systems handling large data volumes (millions+ rows, complex queries, high throughput)
Knowledge of query optimization techniques and execution planning
Familiarity with columnar databases (ClickHouse, BigQuery, Redshift, Snowflake, or similar). Columnar DB experience is a big plus.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8605397
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
25/03/2026
Location: Jerusalem
Job Type: Full Time
As an ML Software Engineer with a focus on low-level and CUDA-based optimizations, you will play a key role in shaping the design, performance, and scalability of machine learning inference systems. Youll work on deeply technical challenges at the intersection of GPU acceleration, systems architecture, and ML deployment.
Your expertise in CUDA, C/C++, and performance tuning will be crucial in enhancing runtime efficiency across heterogeneous computing environments. Youll collaborate with designers, researchers, and backend engineers to build production-grade ML pipelines that are optimized for latency, throughput, and memory use, contributing directly to the infrastructure powering next-generation AI products.This role is ideal for an engineer with strong systems-level thinking, deep familiarity with GPU internals, and a passion for pushing the boundaries of performance and efficiency in machine learning infrastructure.

Responsibilities
Design and implement highly optimized GPU-accelerated ML inference systems using CUDA and low-level parallelism techniques
Optimize memory, compute, and data flow to meet real-time or high-throughput constraints
Improve the performance, reliability, and observability of our inference backend across diverse compute targets (CPU/GPU)
Collaborate with cross-functional teams (including researchers, developers, and designers) to deliver efficient and scalable inference solutions
Contribute to ComfyUI and internal infrastructure to improve usability and performance of model execution flows
Investigate performance bottlenecks at all levels of the stack-from Python to kernel-level execution
Navigate and enhance a large, complex, production-grade codebase
Drive innovation in low-level system design to support future ML workloads
Requirements:
5+ years of experience in high-performance software engineering
Advanced proficiency in CUDA, C/C++, and Python, especially in production environments
Deep understanding of GPU architecture, memory hierarchies, and optimization techniques
Proven track record of optimizing compute-intensive systems
Strong system architecture fundamentals, especially around performance, concurrency, and parallelism
Ability to independently lead deep technical investigations and deliver clean, maintainable solutions
Collaborative and team-oriented mindset, with experience working across functional teams
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8591920
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
25/03/2026
Location: Jerusalem
Job Type: Full Time
The Model Evaluation team is the central nervous system of the LTX Foundation Model group. We don't just measure performance; we define what "good" looks like across a vast array of use cases. While we power the next generation of creative tools, LTX is also a foundational engine for simulation pipelines, game engines, synthetic data generation, architectural rendering, and digital avatars. We act as the critical bridge between raw research and industrial-grade reliability, building the benchmarks that ensure our models are world-class for both artists and engineers.

The Role
As a Research Scientist in Model Evaluation, you are the ultimate authority on model quality and utility. You will design the automated judges, reward models, evaluation datasets, and benchmarking ecosystems that determine the future of LTX. Your mission is to provide the "ground truth" for our pre-training and post-training teams. You will blend the rigor of a researcher with the intuition of a product-thinker, developing metrics that capture both the aesthetic soul of a video and the functional precision required for high-stakes professional use.

Key Responsibilities
Steer Training & Research: Systematically evaluate model checkpoints to provide actionable insights that guide training experiments and architectural decisions.
Design Benchmark Ecosystems: Develop and run rigorous benchmarks for release candidates against competitive models, ensuring LTX-2 remains world-class.
Build Next-Gen Metrics: Develop robust automatic metrics and Reward Models (e.g., for RL, ITS, auto-research agents) that quantify complex attributes like temporal coherence, physical correctness, spatial accuracy, and foley synchronization.
Diagnose & Analyze: Perform deep root-cause analysis on model failures, providing the diagnostic clarity needed for researchers to implement targeted fixes.
Scale Evaluation: Collaborate with platform engineers to deploy evaluation frameworks across large-scale GPU clusters.
Requirements:
Technical Depth: Masters or PhD in Computer Vision, ML, or a related field, with strong software engineering skills and comfort in complex ML training environments.
The "Metric" Mindset: Deep expertise in evaluation methodology and statistical rigor. You know why standard metrics often fail and how to build better ones.
Perceptual Intuition: A sharp "eye and ear" for quality. You can articulate subtle nuances in motion or sound that automated systems might miss and use that intuition to improve our reward models.
Data-Driven Detective: You love diving into datasets to find the "why" behind the numbers, taking pride in curating and specializing data for specific evaluation tasks.
Product-Minded Scientist: You can think like an end-user. You care that our models don't just "beat the benchmark" but actually work reliably in professional pipelines.
Statistical Rigor: You understand experimental design, significance testing, and the nuances of perceptual quality assessment.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8591911
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 6 שעות
Location: Jerusalem
Job Type: Full Time
Our CTO Group is looking for an outstanding Physical-AI Applied Researcher to join our team.
The CTO Group is a small, elite research unit shaping the next generation of algorithmic foundations behind our autonomous driving systems. The group operates at the core of the decision-making and planning stack, addressing some of the most challenging problems in real-world autonomy.
We are seeking a researcher who thrives at the intersection of machine learning, decision-making, and algorithmic rigor - someone who is excited about advancing learning-based approaches for safety-critical, large-scale physical systems.
In this role, you will develop novel approaches for planning and decision-making in interactive, multi-agent driving environments. You will combine deep & reinforcement learning with classical algorithmic structure and formal reasoning. The problems are open-ended, scientifically challenging, and deployed at unprecedented scale.
This is a rare opportunity to conduct high-impact applied research, taking ideas from theory and papers into real-world autonomous systems at scale. If youre excited about pushing the boundaries of learning-based decision-making, wed love you to join us and help shape the future of Physical AI.
What will your job look like:
Design and develop novel learning-based algorithms for decision-making and planning in complex physical environments.
Advance model architectures for long-horizon reasoning, multi-agent interaction, and uncertainty-aware prediction.
Integrate deep learning components into structured planning pipelines with clear formal objectives and safety constraints.
Formulate problems mathematically and derive principled learning objectives grounded in real-world system requirements.
Lead research directions from conception to full-scale production.
Develop using Python (PyTorch or similar frameworks) as well as C++/GPU/Cuda.
Requirements:
M.Sc/Ph.D. in Computer Science, Electrical Engineering, Robotics, Machine Learning, Applied Mathematics, or a related field.
Proven experience in machine learning and deep learning.
Demonstrated ability to conduct independent research (publications in top-tier venues such as NeurIPS, ICML, ICLR, CVPR, CoRL, etc. - advantage).
Strong programming skills in Python; solid C++ experience - advantage.
Experience in training large-scale models and working with real-world data.
Intellectual curiosity, scientific ownership, and comfort operating in open-ended research environments.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8635841
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
25/03/2026
Location: Jerusalem
Job Type: Full Time
This role is designed for individuals who are not only technically proficient but also deeply passionate about pushing the boundaries of AI and machine learning through innovative engineering and collaborative research.

Key Responsibilities
Profile and optimize the training process to ensure efficiency and effectiveness, including optimizing multimodal data pipelines and data storage methods.
Develop high-performance TPU/GPU/CPU kernels and integrate advanced techniques into our training framework to maximize hardware efficiency.
Utilize knowledge of hardware features to make aggressive optimizations and advise on hardware/software co-designs.
Collaboratively develop model architectures with researchers that facilitate efficient training and inference.
Design, maintain, and evolve a high-quality, shared codebase that emphasizes correctness, readability, extensibility, testing, and long-term maintainability, while balancing performance requirements.
Requirements:
Industry experience with small to large-scale ML experiments and multi-modal ML pipelines.
Strong software engineering skills, proficient in Python, and experienced with modern C++.
Deep understanding of GPU, CPU, TPU, or other AI accelerator architectures.
Enjoy diving deep into system implementations to improve performance without compromising code quality and maintainability.
Passion for driving ML large-scale training workloads efficiently and optimizing compute kernels.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8591910
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 4 שעות
Location: Jerusalem
Job Type: Full Time
We are seeking a Backend & Data Engineer to join our Innovation Team within our Mapping Division. This role is best suited for engineers with strong system-level thinking, a can-do approach, and a hands-on mindset, with the ability to design, build, and optimize complex systems operating at scale.
What your job will look like:
Develop and maintain backend and data-processing components in large-scale systems
Design, implement, and optimize data pipelines and distributed processing flows
Work with large-scale storage systems (e.g., S3) and high-volume data access patterns
Optimize systems and code across multiple layers - from architecture to implementation
Identify performance bottlenecks, debug complex issues, and drive root-cause solutions
Work across teams and domains, reading, improving, and refactoring existing code
Take part in technical design and decision-making, balancing performance, scalability, and maintainability.
Requirements:
5+ years of experience in software development, with a strong backend and/or data focus
Experience building backend services (APIs) and working with databases and storage systems
Experience using AI as a core part of the development workflow
Hands-on experience with large-scale data processing and distributed systems
Experience with Spark / PySpark - a strong advantage
Experience with Python or Node.js
Strong understanding of performance optimization and system behavior (CPU, memory, concurrency)
Proven debugging skills and ability to move from symptoms to root cause
A strong can-do approach - proactive, hands-on, and not afraid to dive into complex systems.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8636130
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
25/03/2026
Location: Jerusalem
Job Type: Full Time
This role focuses on pioneering model architecture and pre-training algorithms, shaping the next generation of our foundational generative AI models.

What you will be doing
Pre-train and fine-tune video, audio, and image generative models to pursue state-of-the-art results.
Publish papers and open source models to benefit the research community and advance the field.
Design and implement machine learning models for text-to-audio and text-to-video generation.
Collaborate with data engineers to curate and preprocess text and video data.
Optimize models for high performance, ensuring efficient training and inference.
Build new controls and capabilities into generative text-to-audio and text-to-video models.
Stay updated with the latest developments in Generative AI, particularly in the fields of image, video, and audio.
Work closely with product teams to integrate AI models into applications and services.
Conduct experiments and prototype new concepts to advance the capabilities of our AI tools.
Requirements:
Track record of coming up with new ideas or improving upon existing ideas in generative AI, demonstrated by accomplishments such as first-author publications or projects.
Excellence in engineering as well as research with strong programming skills in Python, and deep familiarity with machine learning frameworks.
Experience in training large diffusion transformer models from scratch.
Proven track record of handling large-scale datasets to train neural networks effectively.
PhD or equivalent experience in the field of generative AI - a plus.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8591908
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
25/03/2026
חברה חסויה
Location: Jerusalem
Job Type: Full Time
As the ML Platform Tech Lead, youll be responsible for the development of AI infrastructure, taking an active role in mentoring backend engineers directly.

This is a hands-on tech- lead position that blends system design and technical strategy. You will be expected to work closely with our ML research teams, product, and engineering stakeholders to take cutting-edge AI models into production, powering features across our creative apps used by millions worldwide.

You will own the full development lifecycle, from system architecture through deployment and operation, while also being accountable for team delivery, collaboration, and growth.
Requirements:
8+ years of experience in backend systems as a tech lead or senior engineer with leadership responsibilities)
Demonstrated ownership of complex projects from design to production in high-scale environments
Strong backend development skills with focus on scalable APIs, distributed systems, and observability
Deep understanding of cloud infrastructure and deployment strategies
Excellent cross-functional communication and stakeholder alignment abilities
Experience mentoring team members and fostering a culture of technical excellence
Exposure to (or strong interest in) ML model serving, GPU-based systems, or ML platform tooling - a plus
Comfortable working in distributed teams and across time zones
B.Sc. in Computer Science or equivalent technical background
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8591919
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
25/03/2026
Location: Jerusalem
Job Type: Full Time
As a Large Scale Video Understanding Research Scientist, you will play a key role in improving video generation quality and efficiency by improving video and audio understanding pipelines used for both training data construction and model evaluation.. This role demands hands-on work with large-scale Video Language Models (VLLMs), including fine-tuning, post-training, and control, alongside implementing classic computer vision and signal processing algorithms and applying strong research skills. Your expertise in post-training and controlling large scale foundational models, understanding statistics, implementing complex systems and eliminating bugs will be crucial, as our video training sets consist of petabytes of data processed across hundreds to thousands of virtual machines.

What you will be doing
Fine-tune and control VLLMs for video and audio understanding.
Design algorithms for balancing, filtering, and curating training and evaluation datasets, informed by model behavior and failure modes.
Implement classic and modern algorithms for processing, clustering, evaluation and filtering of large scale datasets.
Work within high-performance, scalable distributed systems capable of handling petabytes of data, with attention to throughput, correctness, and reproducibility..
Collaborate with other researchers and product stakeholders to iteratively improve training sets and evaluation protocols through tight feedback loops driven by model performance.
Requirements:
Experience training, fine-tuning, or post-training large-scale VLLMs or multimodal foundation models.
Strong software engineering skills, proficient in Jax or PyTorch.
Ability to develop and implement computer vision models for data filtering and evaluation.
Understanding of relevant topics in statistics, clustering.
Enjoys delving into system implementations to enhance performance and maintainability.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8591906
סגור
שירות זה פתוח ללקוחות VIP בלבד