דרושים » דאטה » Senior Backend Developer, Data Platform IL

משרות על המפה
 
בדיקת קורות חיים
VIP
הפוך ללקוח VIP
רגע, משהו חסר!
נשאר לך להשלים רק עוד פרט אחד:
 
שירות זה פתוח ללקוחות VIP בלבד
AllJObs VIP
כל החברות >
משרה זו סומנה ע"י המעסיק כלא אקטואלית יותר
מיקום המשרה: תל אביב יפו
סוג משרה: משרה מלאה
משרות דומות שיכולות לעניין אותך
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for an Experienced Data Engineer to join our marketing team and take end-to-end ownership of our data platform and production data pipelines. In this role, you will be responsible for building robust, scalable, and observable data systems that power analytics, reporting, and downstream business use cases. You will work deeply hands-on with data infrastructure, modeling, and orchestration, and act as a key technical partner to Marketing, Sales Product and Business and Finance teams.
This role suits someone who enjoys working close to the metal, designing systems that scale, and solving ambiguous data problems in a dynamic startup environment. You will play a critical role in shaping how data flows through the company, setting engineering standards, and ensuring data is trustworthy, performant, and ready for growth.
What You'll Do:
Design, build, and maintain scalable, reliable data pipelines and data warehouse architectures to support analytics and business intelligence needs.
Own the end-to-end ETL/ELT processes - ingesting data from internal and external sources, transforming it, and making it analytics-ready.
Model and optimize data structures (fact tables, dimensions, semantic layers) to support performant querying and reporting.
Ensure high standards of data quality, integrity, observability, and reliability across all data assets.
Partner closely with Analytics, Product, Marketing, and Finance teams to understand data requirements and deliver robust data solutions.
Implement monitoring, alerting, and testing frameworks to proactively identify data issues.
Optimize warehouse performance and cost efficiency (query optimization, partitioning, clustering, etc.).
Identify gaps in data collection and work with engineering teams to improve instrumentation and data availability.
Support experimentation and analytics use cases by enabling clean, trustworthy datasets for A/B testing and analysis.
Document data models, pipelines, and best practices to support scale and knowledge sharing.
Requirements:
Bachelors or Masters degree in Computer Science, Data Engineering, Software Engineering, or a related technical field.
3-5 years of hands-on experience as a Data Engineer, preferably in a SaaS or technology-driven environment.
Strong experience designing and maintaining data warehouses (e.g., Snowflake, BigQuery, Redshift).
Proven expertise with ETL/ELT tools and frameworks (e.g., Airflow, dbt, Talend, SSIS, Informatica, or similar).
Advanced SQL skills and solid proficiency in Python (or similar languages) for data processing and orchestration.
Strong understanding of data modeling, warehousing best practices, and analytics engineering concepts.
Experience integrating data from business systems such as Salesforce, HubSpot, or other SaaS platforms.
Familiarity with SaaS metrics and business concepts (ARR, churn, LTV, CAC) - from a data modeling perspective.
Experience supporting BI tools and analytics consumers (Tableau, Looker, Power BI, etc.).
Strong problem-solving skills, attention to detail, and a passion for building reliable data foundations.
Excellent communication skills and the ability to collaborate across technical and non-technical teams.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8563348
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
03/03/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are seeking a skilled Senior Backend Engineer to join our team and help us build Cognition, a multi-agent AI platform. The ideal candidate will have strong backend engineering expertise and enthusiasm for working with cutting-edge AI technologies.You'll be responsible for designing and implementing backend services that integrate AI capabilities including large language models (LLMs), retrieval-augmented generation (RAG), and conversational interfaces into our consumer-facing products. This is an opportunity to expand your expertise into AI while leveraging your strong engineering foundation.
What Youll Do:
Design and develop scalable, high-performance backend services and APIs
Implement integrations with AI platforms and services
Work with AI specialists to translate AI capabilities into production-ready features
Ensure high reliability, performance, and security of AI-powered systems
Collaborate with product and engineering teams to gather requirements and deliver solutions
Requirements:
Proven Backend Expertise: 5+ years of professional experience, with astrong foundation in building and maintaining robust backend systems for large-scale consumer or enterprise products.
Technical Mastery: Advanced proficiency in at least one modern backend language (TypeScript and Node.js), with hands-on experience developing production-grade APIs and microservices. Experience with Lambda functions is a big plus.
Scalable Systems Builder: Demonstrated experience architecting and delivering high-availability services and distributed systems, leveraging patterns for reliability, performance, and security.
Cloud-Native Engineering: Deep familiarity with cloud platforms (AWS, GCP, or Azure), including hands-on experience with serverless technologies (e.g., AWS Lambda), infrastructure-as-code, and automated deployment pipelines.
Data Fluency: Solid experience with both SQL and NoSQL databases; knowledge of vector databases and embedding/search technologies is a big plus.
AI Platform Integration: Experience integrating with modern AI services (OpenAI, Hugging Face, etc.), and enthusiasm for collaborating with ML/AI specialists to bring LLMs, RAG, and conversational AI capabilities into production.
Results-Oriented Delivery: A track record of shipping consumer-facing features at scale, collaborating closely with both product and engineering teams.
Forward-Thinking Problem Solver: Eager to explore new technologies, solving complex problems using both solid engineering fundamentals and innovative approaches.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8567344
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time
we are looking for a Senior Backend & LLM Applications Developer.
As a Senior Backend & LLM Applications Developer, you will design and build scalable backend systems while integrating advanced AI capabilities into production. You will work with Large Language Models (LLMs), agentic AI, and AI agents, making key architectural and technical decisions that shape how AI powers our products for the Care Network.
As part of our fast-growing R&D team, you will drive core infrastructure development, contribute to platform scalability, and help advance cutting-edge AI technology that improves the lives of older adults.
What Youll Do:
Design and develop scalable backend services and cloud-based infrastructure (AWS) using microservices architecture
Take end-to-end ownership of complex backend systems and development projects, from design to production
Drive performance, reliability, security, and scalability of core platform services
Integrate AI and LLM capabilities into backend systems, including prompt optimization and production readiness
Collaborate with Product and Data teams to deliver AI-enabled features
Research, prototype, and lead the adoption of new backend technologies, with selective adoption of AI models and tools into production
Requirements:
6+ years of experience in backend / software engineering, with proven experience designing and developing high-performance, distributed systems
Strong proficiency in Python and Node.js
Proven experience working in cloud environments (AWS preferred; GCP/Azure acceptable)
Hands-on experience with microservices, containerized environments (Kubernetes), and CI/CD pipelines (GitHub Actions, TeamCity)
Experience with message queuing and streaming systems such as Kafka and/or SQS
Strong understanding of SQL and NoSQL databases, large-scale data flows, and data-driven systems
Experience in developing and deploying LLM agents to production (via Langgraph, Langchain, etc.)
Strong collaboration and communication skills, both Hebrew & English
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8555224
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
16/02/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a Senior Data Platform Engineer to design, build, and scale our next-generation data platform, the backbone powering our AI-driven insights.
This role sits at the intersection of data engineering, infrastructure, and MLOps, owning the architecture and reliability of our data ecosystem end-to-end.
Youll work closely with data scientists,r&d teams, analysts to create a robust platform that supports varying use cases, complex ingestion, and AI-powered analytics.
Responsibilities:
Architect and evolve a scalable, cloud-native data platform that supports batch, streaming, analytics, and AI/LLM workloads across R&D.
Help define and implement standards for how data is modeled, stored, governed, and accessed
Design and build data lakes and data warehouses
Develop and maintain complex, reliable, and observable data pipelines
Implement data quality, validation, and monitoring frameworks
Collaborate with ML and data science teams to connect AI/LLM workloads to production data pipelines, enabling RAG, embeddings, and feature engineering flows.
Manage and optimize relational and non-relational datastores (Postgres, Elasticsearch, vector DBs, graph DBs).
Build internal tools and self-service capabilities that enable teams to easily ingest, transform, and consume data.
Contribute to data observability, governance, documentation, and platform visibility
Drive strong engineering practices
Evaluate and integrate emerging technologies that enhance scalability, reliability, and AI integration in the platform.
Requirements:
7+ years experience building/operating data platforms
Strong Python programming skills
Proven experience with cloud data lakes and warehouses (Databricks, Snowflake, or equivalent).
Data orchestration experience (Airflow)
Solid understanding of AWS services
Proficiency with relational databases and search/analytics stores
Experience designing complex data pipelines, managing data quality, lineage, and observability in production.
Familiarity with CI/CD, GitOps, and IaC
Excellent understanding of distributed systems, data partitioning, and schema evolution.
Strong communication skills, ability to document and present technical designs clearly.
Advantages:
Experience with vector databases and graph databases
Experience integrating AI/LLM workloads into data pipelines (feature stores, retrieval pipelines, embeddings).
Familiarity with event streaming and CDC patterns.
Experience with data catalog, lineage, or governance tools
Knowledge of monitoring and alerting stacks
Hands-on experience with multi-source data product architectures.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8548254
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
10/02/2026
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a talented Senior Backend Developer to join our development team. As part of the team, you will contribute to building and enhancing high-scale, multi-tenant SaaS products that support customers worldwide. You will work in a hybrid environment, collaborating closely with highly skilled engineers in an agile, sprint-based methodology. This role is critical to the continued success of our next-generation security and monitoring platforms, ensuring performance, scalability, and reliability across multiple regions and tenants.
The Mobile Solution Group builds our companys mobile security platform, combining on-device protection with scalable backend services. Our backend systems power policy management, integrations, data processing, and enforcement across mobile clients and multiple our company products, operating at scale and under strict reliability and security requirements.
Key Responsibilities
Architect, design, and develop robust, high-quality backend services and APIs for our large-scale SaaS platform.
Own the end-to-end delivery of features in an agile, CI/CD environment, from conception to deployment.
Ensure the scalability, resilience, and high performance of our applications to support a rapidly growing, multi-tenant customer base across multiple regions.
Integrate with and optimize a diverse set of distributed data systems, including PostgreSQL, Redis, RabbitMQ, and Elasticsearch.
Champion and implement observability best practices (metrics, logging, tracing) using tools like Prometheus, Grafana, and OpenTelemetry to ensure system health and proactive monitoring.
Collaborate closely with product managers, frontend developers, and DevOps engineers to translate customer needs into impactful technical solutions.
Continuously explore, evaluate, and adopt new technologies to enhance our architecture and development efficiency.
Mentor and guide other developers, fostering a culture of technical excellence and knowledge sharing.
Requirements:
Must Have:
BA/BSc in Computer Science or a related field, or equivalent practical experience.
6-7 years of professional experience in backend software development.
Proven expertise in Python (including experience with the Django framework) and Java for backend systems.
Experience with relational databases (e.g., PostgreSQL, MySQL, SQL Server).
Strong experience with cloud-native technologies, containers (Docker), orchestration (Kubernetes), and CI/CD pipelines.
Solid understanding of multi-tenant, distributed system architecture.
Strong collaboration, communication, and problem-solving skills.
Nice to Have:
Experience with frontend frameworks such as React or Angular.
Experience with NoSQL databases (e.g., MongoDB, Elasticsearch, Couchbase).
Background in large-scale SaaS performance optimization.
Hands-on experience with observability tools and practices (e.g., Prometheus, Grafana, OpenTelemetry).
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8540395
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: More than one
We are seeking an exceptional Senior Backend Engineer to join our Platform Group, where you will architect and develop shared infrastructure components that power Semperis's enterprise security and identity products. This is a highly collaborative role requiring both technical excellence and strong cross-functional partnership skills.
What You'll Do:
As a Senior Backend Engineer on our Core Platform team, you will:
Design and build shared platform components used across multiple product teams, ensuring scalability, reliability, and maintainability
Architect distributed systems using microservices and event-driven patterns that support enterprise-scale workloads
Own critical backend services from conception through deployment, including code quality, performance optimization, and operational excellence
Collaborate extensively with product engineers, DevOps, and architects to define platform capabilities and technical standards
Mentor junior engineers through code reviews, pair programming, and technical guidance
Drive technical initiatives that improve developer productivity, system observability, and platform resilience
Contribute to architectural decisions and establish best practics for backend development across the organization
Analyze system performance and data patterns to identify optimization opportunities and inform future platform investments
Requirements:
7+ years of professional software engineering experience, with strong expertise in backend development
Deep proficiency in C# and .NET Core, including modern framework features and performance optimization
Production experience with Kubernetes and container orchestration in cloud environments (Azure/AWS/GCP preferred)
Proven track record designing and implementing microservices architectures and event-driven systems at scale
Experience developing shared libraries, frameworks, or platform components consumed by multiple teams
Demonstrated ability building enterprise SaaS applications serving high-volume, multi-tenant environments
Strong collaborative mindset with excellent communication skills and experience working across teams to drive consensus
Bachelor's degree in Computer Science or equivalent practical experience
Technical leadership experience, including mentoring engineers and leading technical initiatives
Fluency in English (written and verbal)
Preferred Qualifications:
Experience in the Enterprise Identity and Access Management (IAM) domain
Hands-on expertise with Kafka, RabbitMQ, or similar message brokers
Knowledge of ElasticSearch or other distributed search/analytics platforms
Open-source contributions or experience maintaining shared component libraries
Experience with observability tools (Prometheus, Grafana, distributed tracing)
Background in API design and governance for platform services
Familiarity with CI/CD pipelines and infrastructure-as-code practices
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8553547
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Senior Algo Data Engineer
Realize your potential by joining the leading performance-driven advertising company!
As a Senior Algo Data Engineer on the Infra group, youll play a vital role in develop, enhance and maintain highly scalable Machine-Learning infrastructures and tools.
About Algo platform:
The objective of the algo platform group is to own the existing algo platform (including health, stability, productivity and enablement), to facilitate and be involved in new platform experimentation within the algo craft and lead the platformization of the parts which should graduate into production scale. This includes support of ongoing ML projects while ensuring smooth operations and infrastructure reliability, owning a full set of capabilities, design and planning, implementation and production care.
The group has deep ties with both the algo craft as well as the infra group. The group reports to the infra department and has a dotted line reporting to the algo craft leadership.
The group serves as the professional authority when it comes to ML engineering and ML ops, serves as a focal point in a multidisciplinary team of algorithm researchers, product managers, and engineers and works with the most senior talent within the algo craft in order to achieve ML excellence.
How youll make an impact:
As a Senior Algo Data Engineer, youll bring value by:
Develop, enhance and maintain highly scalable Machine-Learning infrastructures and tools, including CI/CD, monitoring and alerting and more
Have end to end ownership: Design, develop, deploy, measure and maintain our machine learning platform, ensuring high availability, high scalability and efficient resource utilization
Identify and evaluate new technologies to improve performance, maintainability, and reliability of our machine learning systems
Work in tandem with the engineering-focused and algorithm-focused teams in order to improve our platform and optimize performance
Optimize machine learning systems to scale and utilize modern compute environments (e.g. distributed clusters, CPU and GPU) and continuously seek potential optimization opportunities.
Build and maintain tools for automation, deployment, monitoring, and operations.
Troubleshoot issues in our development, production and test environments
Influence directly on the way billions of people discover the internet
Our tech stack:
Java, Python, TensorFlow, Spark, Kafka, Cassandra, HDFS, vespa.ai, ElasticSearch, AirFlow, BigQuery, Google Cloud Platform, Kubernetes, Docker, git and Jenkins.
Requirements:
Experience developing large scale systems. Experience with filesystems, server architectures, distributed systems, SQL and No-SQL. Experience with Spark and Airflow / other orchestration platforms is a big plus.
Highly skilled in software engineering methods. 5+ years experience.
Passion for ML engineering and for creating and improving platforms
Experience with designing and supporting ML pipelines and models in production environment
Excellent coding skills - in Java & Python
Experience with TensorFlow - a big plus
Possess strong problem solving and critical thinking skills
BSc in Computer Science or related field.
Proven ability to work effectively and independently across multiple teams and beyond organizational boundaries
Deep understanding of strong Computer Science fundamentals: object-oriented design, data structures systems, applications programming and multi threading programming
Strong communication skills to be able to present insights and ideas, and excellent English, required to communicate with our global teams.
Bonus points if you have:
Experience in leading Algorithms projects or teams.
Experience in developing models using deep learning techniques and tools
Experience in developing software within a distributed computation framework.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8559383
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Staff Algo Data Engineer
Realize your potential by joining the leading performance-driven advertising company!
As a Staff Algo Data Engineer on the Infra group, youll play a vital role in develop, enhance and maintain highly scalable Machine-Learning infrastructures and tools.
About Algo platform:
The objective of the algo platform group is to own the existing algo platform (including health, stability, productivity and enablement), to facilitate and be involved in new platform experimentation within the algo craft and lead the platformization of the parts which should graduate into production scale. This includes support of ongoing ML projects while ensuring smooth operations and infrastructure reliability, owning a full set of capabilities, design and planning, implementation and production care.
The group has deep ties with both the algo craft as well as the infra group. The group reports to the infra department and has a dotted line reporting to the algo craft leadership.
The group serves as the professional authority when it comes to ML engineering and ML ops, serves as a focal point in a multidisciplinary team of algorithm researchers, product managers, and engineers and works with the most senior talent within the algo craft in order to achieve ML excellence.
How youll make an impact:
As a Staff Algo Data Engineer Engineer, youll bring value by:
Develop, enhance and maintain highly scalable Machine-Learning infrastructures and tools, including CI/CD, monitoring and alerting and more
Have end to end ownership: Design, develop, deploy, measure and maintain our machine learning platform, ensuring high availability, high scalability and efficient resource utilization
Identify and evaluate new technologies to improve performance, maintainability, and reliability of our machine learning systems
Work in tandem with the engineering-focused and algorithm-focused teams in order to improve our platform and optimize performance
Optimize machine learning systems to scale and utilize modern compute environments (e.g. distributed clusters, CPU and GPU) and continuously seek potential optimization opportunities.
Build and maintain tools for automation, deployment, monitoring, and operations.
Troubleshoot issues in our development, production and test environments
Influence directly on the way billions of people discover the internet
Our tech stack:
Java, Python, TensorFlow, Spark, Kafka, Cassandra, HDFS, vespa.ai, ElasticSearch, AirFlow, BigQuery, Google Cloud Platform, Kubernetes, Docker, git and Jenkins.
Requirements:
Experience developing large scale systems. Experience with filesystems, server architectures, distributed systems, SQL and No-SQL. Experience with Spark and Airflow / other orchestration platforms is a big plus.
Highly skilled in software engineering methods. 5+ years experience.
Passion for ML engineering and for creating and improving platforms
Experience with designing and supporting ML pipelines and models in production environment
Excellent coding skills - in Java & Python
Experience with TensorFlow - a big plus
Possess strong problem solving and critical thinking skills
BSc in Computer Science or related field.
Proven ability to work effectively and independently across multiple teams and beyond organizational boundaries
Deep understanding of strong Computer Science fundamentals: object-oriented design, data structures systems, applications programming and multi threading programming
Strong communication skills to be able to present insights and ideas, and excellent English, required to communicate with our global teams.
Bonus points if you have:
Experience in leading Algorithms projects or teams.
Experience in developing models using deep learning techniques and tools
Experience in developing software within a distributed computation framework
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8559783
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a motivated and experienced Senior Developer to join our Cloud and K8s Group. The successful candidate will possess a strong technical background in Kubernetes and different cloud environments, and will excel in developing advanced k8s-related features. Proficiency in NodeJS (Typescript), MongoDB, Kubernetes, Docker, and data queues is essential for this role, as you will be instrumental in refining our cloud CTEM offering.
Key Responsibilities:
Design and implement advanced k8s-related features to enhance our existing offerings.
Streamline deployment processes and ensure system reliability by demonstrating expertise in Kubernetes and Docker.
Develop modular and scalable solutions while leveraging microservices and jobs architecture.
Utilize advanced data processing techniques with tools such as Spark , Flink, written in Scala to improve system performance.
Contribute to the enhancement of system reliability at scale through hands-on experience with large-scale applications.
Develop attack techniques for K8S, cloud and AI technologies.
Requirements:
Bachelor's degree in Computer Science, a related field, or equivalent practical experience.
Minimum of 5 years of experience in software development, with a strong focus on cloud technologies.
Familiarity with at least one cloud platform (GCP, AWS, Azure).
Familiarity with Kubernetes is a big advantage.
Familiarity with networking concepts.
Proficient in Scala (Or alternatively Java, Kotlin), NodeJS (Typescript), Kubernetes, Docker, and data queue technologies (Kafka/RabbitMQ/SQS).
Understanding of microservices and jobs architecture.
Experience with data processing frameworks such as Flink, Spark is a big advantage.
Experience working on large-scale applications is an advantage.
Knowledge of cybersecurity principles and practices is a plus.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8559954
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time
This is a great opportunity to be part of one of the fastest-growing infrastructure companies in history, an organization that is in the center of the hurricane being created by the revolution in artificial intelligence.
We are seeking an experienced Solutions Data Engineer who possess both technical depth and strong interpersonal skills to partner with internal and external teams to develop scalable, flexible, and cutting-edge solutions. Solutions Engineers collaborate with operations and business development to help craft solutions to meet customer business problems.
A Solutions Engineer works to balance various aspects of the project, from safety to design. Additionally, a Solutions Engineer researches advanced technology regarding best practices in the field and seek to find cost-effective solutions.
Job Description:
Were looking for a Solutions Engineer with deep experience in Big Data technologies, real-time data pipelines, and scalable infrastructure-someone whos been delivering critical systems under pressure, and knows what it takes to bring complex data architectures to life. This isnt just about checking boxes on tech stacks-its about solving real-world data problems, collaborating with smart people, and building robust, future-proof solutions.
In this role, youll partner closely with engineering, product, and customers to design and deliver high-impact systems that move, transform, and serve data at scale. Youll help customers architect pipelines that are not only performant and cost-efficient but also easy to operate and evolve.
We want someone whos comfortable switching hats between low-level debugging, high-level architecture, and communicating clearly with stakeholders of all technical levels.
Key Responsibilities:
Build distributed data pipelines using technologies like Kafka, Spark (batch & streaming), Python, Trino, Airflow, and S3-compatible data lakes-designed for scale, modularity, and seamless integration across real-time and batch workloads.
Design, deploy, and troubleshoot hybrid cloud/on-prem environments using Terraform, Docker, Kubernetes, and CI/CD automation tools.
Implement event-driven and serverless workflows with precise control over latency, throughput, and fault tolerance trade-offs.
Create technical guides, architecture docs, and demo pipelines to support onboarding, evangelize best practices, and accelerate adoption across engineering, product, and customer-facing teams.
Integrate data validation, observability tools, and governance directly into the pipeline lifecycle.
Own end-to-end platform lifecycle: ingestion → transformation → storage (Parquet/ORC on S3) → compute layer (Trino/Spark).
Benchmark and tune storage backends (S3/NFS/SMB) and compute layers for throughput, latency, and scalability using production datasets.
Work cross-functionally with R&D to push performance limits across interactive, streaming, and ML-ready analytics workloads.
Requirements:
2-4 years in software / solution or infrastructure engineering, with 2-4 years focused on building / maintaining large-scale data pipelines / storage & database solutions.
Proficiency in Trino, Spark (Structured Streaming & batch) and solid working knowledge of Apache Kafka.
Coding background in Python (must-have); familiarity with Bash and scripting tools is a plus.
Deep understanding of data storage architectures including SQL, NoSQL, and HDFS.
Solid grasp of DevOps practices, including containerization (Docker), orchestration (Kubernetes), and infrastructure provisioning (Terraform).
Experience with distributed systems, stream processing, and event-driven architecture.
Hands-on familiarity with benchmarking and performance profiling for storage systems, databases, and analytics engines.
Excellent communication skills-youll be expected to explain your thinking clearly, guide customer conversations, and collaborate across engineering and product teams.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8572794
סגור
שירות זה פתוח ללקוחות VIP בלבד