דרושים » הנדסה » Senior SW Engineer - Data Team

משרות על המפה
 
בדיקת קורות חיים
VIP
הפוך ללקוח VIP
רגע, משהו חסר!
נשאר לך להשלים רק עוד פרט אחד:
 
שירות זה פתוח ללקוחות VIP בלבד
AllJObs VIP
כל החברות >
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 9 שעות
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
Responsibilities:

Own, enhance, and improve our core platform.

Focus on system efficiency and resilient infrastructure.

Integrate well-known SAAS platforms to our big data repository e.g., the major cloud providers, DataDog, and Snowflake.

Optimize Spark and airflow processes.

Contribute to data design and platform architecture while working closely with other business units and engineering teams.

Face the challenges of testing and monitoring your large-scale data pipelines.
Requirements:
7+ years developing and operating large-scale systems with high availability.

7+ years of experience with Python or equivalent.

Experience working with cloud environments (AWS preferred) and big data technologies (Spark, Airflow, S3, Snowflake, EMR).

Autodidact and self-motivated team player with solid communication skills who is interested in tackling the challenge of working with data at a scale.
This position is open to all candidates.
 
Hide
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8563734
סגור
שירות זה פתוח ללקוחות VIP בלבד
משרות דומות שיכולות לעניין אותך
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 9 שעות
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
If youre looking to build something big, solve real problems, and grow fast-wed love to meet you.

Responsibilities:

Own, enhance, and improve core platform

Focus on system efficiency and resilient infrastructure

Integrate well-known SAAS platforms to big data repository e.g., the major cloud providers, DataDog, and Snowflake

Optimize Spark and airflow processes

Contribute to data design and platform architecture while working closely with other business units and engineering teams

Face the challenges of testing and monitoring your large-scale data pipelines
Requirements:
7+ years developing and operating large-scale systems with high availability

7+ years of experience with Python or equivalent

Experience working with cloud environments (AWS preferred) and big data technologies (Spark, Airflow, S3, Snowflake, EMR)

Autodidact and self-motivated team player with solid communication skills who is interested in tackling the challenge of working with data at a scale
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8563639
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 9 שעות
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
Responsibilities:

Own, enhance, and improve our core platform.

Build and maintain systems for collecting and processing metrics from client environments.

Focus on system efficiency and resilient infrastructure.

Integrate well-known SaaS platforms into our big data repository (e.g., major cloud providers, Datadog, Snowflake).

Optimize Spark and Airflow processes.

Contribute to data design and platform architecture while working closely with other business units and engineering teams.

Face the challenges of testing and monitoring large-scale data pipelines.
Requirements:
7+ years of experience developing and operating large-scale, high-availability systems.

7+ years of experience with Python (Experience with Go is a plus).

Experience working with cloud environments (AWS preferred) and big data technologies (Spark, Airflow, S3, Snowflake, EMR).

Familiarity with metrics systems (e.g., Prometheus, cloud monitoring APIs) or time-series data - a strong plus.

Autodidact, self-motivated team player with strong communication skills and a passion for solving challenges at scale.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8563716
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 9 שעות
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
If youre looking to build something big, solve real problems, and grow fast-wed love to meet you.

Responsibilities:

Own, enhance, and improve core platform

Build and maintain systems for collecting and processing metrics from client environments

Focus on system efficiency and resilient infrastructure

Integrate well-known SaaS platforms into big data repository (e.g., major cloud providers, Datadog, Snowflake)

Optimize Spark and Airflow processes

Contribute to data design and platform architecture while working closely with other business units and engineering teams

Face the challenges of testing and monitoring large-scale data pipelines
Requirements:
7+ years of experience developing and operating large-scale, high-availability systems

7+ years of experience with Python (Experience with Go is a plus)

Experience working with cloud environments (AWS preferred) and big data technologies (Spark, Airflow, S3, Snowflake, EMR)

Familiarity with metrics systems (e.g., Prometheus, cloud monitoring APIs) or time-series data - a strong plus

Autodidact, self-motivated team player with strong communication skills and a passion for solving challenges at scale
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8563646
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
03/02/2026
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a highly skilled and motivated Senior II Software Engineer to join the Operational Experience engineering team. The team is part of the Customer Experience group, which is responsible for the platform, tools, and customer-facing experiences that power how our customers interact with our ecosystem. This is a high-impact, hands-on role, in which youll be working closely with product managers, designers, customer-facing teams, and engineering partners across the company.

You will operate at the intersection of backend engineering, data-intensive systems, platform development, and customer experience. The ideal candidate brings strong expertise in Node.js and TypeScript, along with deep experience working with large-scale data stores, event-driven pipelines, data models, and high-throughput infrastructure. You will work closely with cross-functional partners to design and implement robust backend services, data-access patterns, and operational workflows that power the portal and internal tools. As we invest heavily in Agentic AI, you will also play a central role in shaping and implementing AI-driven capabilities across the platform. While the role is primarily backend, you will occasionally contribute across the full stack when it supports end-to-end delivery.

If you enjoy owning complex problems end to end, improving systems at scale, and building experiences that bring real value to customers, we would love to meet you.

What you'll be doing:
Drive technical direction and architecture within the OX team and across the broader CX organization. You will proactively identify opportunities to improve performance, resilience, cost, scalability, and developer experience, primarily in backend systems but with influence across the stack.
Lead the development of AI-driven and Agentic AI capabilities. Define how LLMs integrate into our platform, build AI-powered workflows, and establish strong engineering patterns for safe and reliable adoption.
Own and evolve the data foundations behind the portal. Optimize pipelines, improve data quality and freshness, and design resilient data-access patterns across Snowflake, Elasticsearch, Kafka, Redis, MySQL, and related systems.
Work closely with product, design, customer-facing teams, and partner engineering groups. Turn ambiguous problems into clear execution plans and ensure alignment with customer and business goals.
Shape shared standards and platform best practices. Guide other teams on backend services, data integration patterns, portal development approaches, and AI-enabled workflows.
Mentor and elevate engineers across the CX group. Promote engineering excellence, share knowledge openly, and help teams adopt effective modern development practices.
Own delivery of high-impact initiatives. Contribute hands-on when needed, remove blockers, maintain execution momentum, and drive projects from concept to production.
Requirements:
What you'll need:
6+ years of experience as a software engineer with strong expertise in backend development using Node.js and TypeScript, with the ability to work across the stack when needed.
Experience building customer-facing products and working closely with product managers, designers, and customer-facing stakeholders.
Strong familiarity with cloud-native environments. AWS experience is a significant advantage.
Hands-on experience with distributed systems, event-driven architectures, and datastores such as Redis, Kafka, SQS, Elasticsearch, MySQL, and Snowflake.
Demonstrated impact in senior engineering roles. You have led complex technical initiatives, influenced product decisions, and helped drive architecture across teams.
Deep systems thinking with the ability to design and scale robust, performant, and maintainable services.
Excellent communication and collaboration skills. You can discuss architecture with engineers, roadmap with product managers, and explain tradeoffs to non-technical stakeholders.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8530019
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Senior Algo Data Engineer
Realize your potential by joining the leading performance-driven advertising company!
As a Senior Algo Data Engineer on the Infra group, youll play a vital role in develop, enhance and maintain highly scalable Machine-Learning infrastructures and tools.
About Algo platform:
The objective of the algo platform group is to own the existing algo platform (including health, stability, productivity and enablement), to facilitate and be involved in new platform experimentation within the algo craft and lead the platformization of the parts which should graduate into production scale. This includes support of ongoing ML projects while ensuring smooth operations and infrastructure reliability, owning a full set of capabilities, design and planning, implementation and production care.
The group has deep ties with both the algo craft as well as the infra group. The group reports to the infra department and has a dotted line reporting to the algo craft leadership.
The group serves as the professional authority when it comes to ML engineering and ML ops, serves as a focal point in a multidisciplinary team of algorithm researchers, product managers, and engineers and works with the most senior talent within the algo craft in order to achieve ML excellence.
How youll make an impact:
As a Senior Algo Data Engineer, youll bring value by:
Develop, enhance and maintain highly scalable Machine-Learning infrastructures and tools, including CI/CD, monitoring and alerting and more
Have end to end ownership: Design, develop, deploy, measure and maintain our machine learning platform, ensuring high availability, high scalability and efficient resource utilization
Identify and evaluate new technologies to improve performance, maintainability, and reliability of our machine learning systems
Work in tandem with the engineering-focused and algorithm-focused teams in order to improve our platform and optimize performance
Optimize machine learning systems to scale and utilize modern compute environments (e.g. distributed clusters, CPU and GPU) and continuously seek potential optimization opportunities.
Build and maintain tools for automation, deployment, monitoring, and operations.
Troubleshoot issues in our development, production and test environments
Influence directly on the way billions of people discover the internet
Our tech stack:
Java, Python, TensorFlow, Spark, Kafka, Cassandra, HDFS, vespa.ai, ElasticSearch, AirFlow, BigQuery, Google Cloud Platform, Kubernetes, Docker, git and Jenkins.
Requirements:
Experience developing large scale systems. Experience with filesystems, server architectures, distributed systems, SQL and No-SQL. Experience with Spark and Airflow / other orchestration platforms is a big plus.
Highly skilled in software engineering methods. 5+ years experience.
Passion for ML engineering and for creating and improving platforms
Experience with designing and supporting ML pipelines and models in production environment
Excellent coding skills - in Java & Python
Experience with TensorFlow - a big plus
Possess strong problem solving and critical thinking skills
BSc in Computer Science or related field.
Proven ability to work effectively and independently across multiple teams and beyond organizational boundaries
Deep understanding of strong Computer Science fundamentals: object-oriented design, data structures systems, applications programming and multi threading programming
Strong communication skills to be able to present insights and ideas, and excellent English, required to communicate with our global teams.
Bonus points if you have:
Experience in leading Algorithms projects or teams.
Experience in developing models using deep learning techniques and tools
Experience in developing software within a distributed computation framework.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8559383
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Senior Data Engineer I - GenAI Foundation Models
21679
Leadership/Team Quote:
This opening is for the Content Intelligence team within the Marketplace AI department.
The Content Intelligence team is at the forefront of Generative AI innovation, driving solutions for travel-related chatbots, text generation and summarization applications, Q&A systems, and free-text search. Beyond this, the team is building a cutting-edge platform that processes millions of images and textual inputs daily, enriching them with ML capabilities. These enriched datasets power downstream applications, helping personalize the customer experience-for example, selecting and displaying the most relevant images and reviews as customers plan and book their next vacation.
Role Description:
As a Senior Data Engineer, youll collaborate with top-notch engineers and data scientists to elevate our platform to the next level and deliver exceptional user experiences. Your primary focus will be on the data engineering aspects-ensuring the seamless flow of high-quality, relevant data to train and optimize content models, including GenAI foundation models, supervised fine-tuning, and more.
Youll work closely with teams across the company to ensure the availability of high-quality data from ML platforms, powering decisions across all departments. With access to petabytes of data through MySQL, Snowflake, Cassandra, S3, and other platforms, your challenge will be to ensure that this data is applied even more effectively to support business decisions, train and monitor ML models and improve our products.
Key Job Responsibilities and Duties:
Rapidly developing next-generation scalable, flexible, and high-performance data pipelines.
Dealing with massive textual sources to train GenAI foundation models.
Solving issues with data and data pipelines, prioritizing based on customer impact.
End-to-end ownership of data quality in our core datasets and data pipelines.
Experimenting with new tools and technologies to meet business requirements regarding performance, scaling, and data quality.
Providing tools that improve Data Quality company-wide, specifically for ML scientists.
Providing self-organizing tools that help the analytics community discover data, assess quality, explore usage, and find peers with relevant expertise.
Acting as an intermediary for problems, with both technical and non-technical audiences.
Promote and drive impactful and innovative engineering solutions
Technical, behavioral and interpersonal competence advancement via on-the-job opportunities, experimental projects, hackathons, conferences, and active community participation
Collaborate with multidisciplinary teams: Collaborate with product managers, data scientists, and analysts to understand business requirements and translate them into machine learning solutions. Provide technical guidance and mentorship to junior team members.
Requirements:
Bachelors or masters degree in computer science, Engineering, Statistics, or a related field.
Minimum of 6 years of experience as a Data Engineer or a similar role, with a consistent record of successfully delivering ML/Data solutions.
You have built production data pipelines in the cloud, setting up data-lake and server-less solutions; ‌ you have hands-on experience with schema design and data modeling and working with ML scientists and ML engineers to provide production level ML solutions.
You have experience designing systems E2E and knowledge of basic concepts (lb, db, caching, NoSQL, etc)
Strong programming skills in languages such as Python and Java.
Experience with big data processing frameworks such, Pyspark, Apache Flink, Snowflake or similar frameworks.
Demonstrable experience with MySQL, Cassandra, DynamoDB or similar relational/NoSQL database systems.
Experience with Data Warehousing and ETL/ELT pipelines.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8560108
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
04/02/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Were looking for a Senior Data Engineer in the Data Collection Ingest team to contribute to the design and coding of data ingestion services & pipelines. This role involves working as part of a team that handles millions of requests per minute across multiple servers, and is responsible for a wide-range of data pipelines, processing billions of events each day.
Why is this role so important?
As the most trusted platform for measuring online behavior, millions of people rely on insights daily as the ground truth for their knowledge of the digital world. Producing these insights requires large scale raw data to be ingested reliably in high scale to provide stable signals for analysis. As a Data Engineer you will have the opportunity to perform hands-on work and own raw data ingestion pipeline end-to-end. Your work will have a direct impact on the quality and reliability of our data and the insights that our products are delivering to our customers.
Requirements:
Has 5+ years of experience in developing code for big data infrastructure. Proficiency in technologies such as: Databricks, Spark, Airflow, Firehose, SQS, or other similar tools.
Proven experience working with high scale on AWS or any other cloud provider. Experience in architecture and design of large-scale and high performance production systems.
Comfortable taking challenges and learning new technologies.
Excellent communication skills with the ability to provide constant dialog between teams.
Ability to take business requirements and translate them to technical alternatives by performing risk management and evaluating tradeoffs.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8531695
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Were looking for a Senior SW Engineer to lead development of algorithmic data processing pipelines at high scale and build the next generation of our data processor.
Responsibilities
Data Centric Software Development: Lead development of algorithmic data processing pipelines at high scale
Monitoring and Performance Tuning: Optimize resource consumption and monitor performance of complex data pipelines handling high throughput
Collaboration with Data Scientists: Work closely with data scientists and to integrate ML models and other algorithmic artifacts into a production grade data platform and application ecosystem. Facilitate seamless collaboration between teams to ensure ML and data engineering efforts are aligned with business objectives.
Peer collaboartion : Bring knowledge of architecture, design patterns and best practices to help improve quality within the team
Security and Compliance: Ensure the data platform adheres to security best practices and compliance requirements, protecting sensitive data and privacy.
Technology Evaluation and Innovation: Stay abreast of the latest developments in Data processing, ML technologies, and cloud services to drive innovation and adopt new technologies where beneficial. 
Requirements:
Experience: 5+ of software engineering experience, of which 2+ years are in data centric projects
Technology experience: Strong proficiency in Flink, k8s.
Cloud Platforms: Solid experience with cloud platforms (AWS, GCP)
Monitoring Tools: Knowledge of monitoring and logging tools as well logging patterns fit for data pipelines.
Teamwork and Communication: Excellent teamwork and communication skills, with the ability to work effectively in a cross-functional team environment.
Self-Learning: High capacity for self-learning to stay updated with the latest in technology practices.
Education: Bachelor's degree in Computer Science, Engineering, or a related field.
Language: Strong spoken and written English skills.
Preferred Skills:
Experience with ML engineering.
Knowledge of data analytics stack: Python, Pandas, Spark
Knowledge of data governance, security, and privacy best practices in data processing
Experience working with coding agents
Global Experience: Experience working in or with global teams is a plus.
Experience with CI/CD tools - GH actions, Argo CD etc.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8525628
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time
We are seeking an experienced Solutions Data Engineer who possess both technical depth and strong interpersonal skills to partner with internal and external teams to develop scalable, flexible, and cutting-edge solutions. Solutions Engineers collaborate with operations and business development to help craft solutions to meet customer business problems.
A Solutions Engineer works to balance various aspects of the project, from safety to design. Additionally, a Solutions Engineer researches advanced technology regarding best practices in the field and seek to find cost-effective solutions.
Job Description:
Were looking for a Solutions Engineer with deep experience in Big Data technologies, real-time data pipelines, and scalable infrastructure-someone whos been delivering critical systems under pressure, and knows what it takes to bring complex data architectures to life. This isnt just about checking boxes on tech stacks-its about solving real-world data problems, collaborating with smart people, and building robust, future-proof solutions.
In this role, youll partner closely with engineering, product, and customers to design and deliver high-impact systems that move, transform, and serve data at scale. Youll help customers architect pipelines that are not only performant and cost-efficient but also easy to operate and evolve.
We want someone whos comfortable switching hats between low-level debugging, high-level architecture, and communicating clearly with stakeholders of all technical levels.
Key Responsibilities:
Build distributed data pipelines using technologies like Kafka, Spark (batch & streaming), Python, Trino, Airflow, and S3-compatible data lakes-designed for scale, modularity, and seamless integration across real-time and batch workloads.
Design, deploy, and troubleshoot hybrid cloud/on-prem environments using Terraform, Docker, Kubernetes, and CI/CD automation tools.
Implement event-driven and serverless workflows with precise control over latency, throughput, and fault tolerance trade-offs.
Create technical guides, architecture docs, and demo pipelines to support onboarding, evangelize best practices, and accelerate adoption across engineering, product, and customer-facing teams.
Integrate data validation, observability tools, and governance directly into the pipeline lifecycle.
Own end-to-end platform lifecycle: ingestion → transformation → storage (Parquet/ORC on S3) → compute layer (Trino/Spark).
Benchmark and tune storage backends (S3/NFS/SMB) and compute layers for throughput, latency, and scalability using production datasets.
Work cross-functionally with R&D to push performance limits across interactive, streaming, and ML-ready analytics workloads.
Operate and debug object store-backed data lake infrastructure, enabling schema-on-read access, high-throughput ingestion, advanced searching strategies, and performance tuning for large-scale workloads.
Requirements:
2-4 years in software / solution or infrastructure engineering, with 2-4 years focused on building / maintaining large-scale data pipelines / storage & database solutions.
Proficiency in Trino, Spark (Structured Streaming & batch) and solid working knowledge of Apache Kafka.
Coding background in Python (must-have); familiarity with Bash and scripting tools is a plus.
Deep understanding of data storage architectures including SQL, NoSQL, and HDFS.
Solid grasp of DevOps practices, including containerization (Docker), orchestration (Kubernetes), and infrastructure provisioning (Terraform).
Experience with distributed systems, stream processing, and event-driven architecture.
Hands-on familiarity with benchmarking and performance profiling for storage systems, databases, and analytics engines.
Excellent communication skills-youll be expected to explain your thinking clearly, guide customer conversations, and collaborate across engineering and product teams.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8512434
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
we are looking for a talented Backend Developer.
Join the VRM team to build high-performance vulnerability management capabilities that translate large-scale security data and research findings into prioritized, actionable remediation paths. Youll work closely with researchers, product and operations teams to drive measurable reductions in customer risk.
This role provides the opportunity to lead end-to-end design & implementation of advanced server-side features using a wide range of technologies: Apache Spark, Apache Airflow, Scala, k8s, Ts, Python, Redis, Kafka, and Dockers.
If you are up for the challenge come and join us!
Key Responsibilities:
Streamline deployment processes and ensure system reliability by demonstrating expertise in Kubernetes and Docker.
Develop modular and scalable solutions while leveraging microservices and jobs architecture.
Utilize advanced data processing with tools such as Spark , Flink, written in Scala to improve system performance.
Contribute to the enhancement of system reliability at scale through hands-on experience with large-scale applications.
Requirements:
Bachelor's degree in Computer Science, a related field, or equivalent practical experience.
Minimum of 3 years of experience in software development, with a strong focus on cloud technologies.
Proficient in Python, Typescript or other modern languages, and data queue technologies (Kafka/RabbitMQ/SQS).
Familiarity with at least one cloud platform (GCP, AWS, Azure).
Familiarity with Kubernetes/Docker is a big advantage.
Understanding of microservices and jobs architecture.
Experience with data processing frameworks such as Flink, Spark is a big advantage.
Experience working on large-scale applications is an advantage.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8559971
סגור
שירות זה פתוח ללקוחות VIP בלבד