דרושים » דאטה » Data Engineer

משרות על המפה
 
בדיקת קורות חיים
VIP
הפוך ללקוח VIP
רגע, משהו חסר!
נשאר לך להשלים רק עוד פרט אחד:
 
שירות זה פתוח ללקוחות VIP בלבד
AllJObs VIP
כל החברות >
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
2 ימים
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a brilliant, quick-learner Data Engineer for our data engineering team - an independent, logical thinker who understands the importance of data structuring for macro-business decisions.
The position combines high technical skills with a business orientation.
It involves working closely with the analysts, product, and the R&D team and directly affecting the company's cross-department decisions.
Our Data Engineer should be able to design and build a data flow from the API or source requirements with the most suitable tools to fit the data product requirements.
They need to speak in technical and practical terms and, more importantly, lead from one to the other while dealing with challenges, independent learning, and creating them to make our team even better.
Roles and Responsibilities:
Designing and building full data pipelines: From defining source structures to delivering clean, organized data ready for analysis, your work ensures analysts have everything they need to make smart, data-driven decisions
Translating business needs into scalable data solutions involves staying close to the roadmap, understanding technical nuances, and delivering purpose-built pipelines and tools
Writing high-quality, maintainable code: Following best practices while leveraging modern data engineering tooling and CI/CD principles
Solving complex data challenges creatively: Whether it's device identity (Device Graph), online-to-offline matching, privacy compliance, or server-to-server integrations
Managing multi-source data environments: We bring in data from over 50 sources (Marketing, Product, CS, CRM, Ops, and more), and you will assist in tying it all together into a reliable, insight-ready system
Keeping quality and reliability top of mind: Monitor, validate, and improve data quality while ensuring robust processes across the stack
Requirements:
B.A / B.Sc. degree in a highly quantitative field
4+ years of hands-on experience in data engineering, building data pipelines, writing complex SQL, and structuring data at scale
Fast learner with high attention to detail, and proven ability and passion to multitask on several projects at a time
Strong communication skills and a proven ability to collaborate effectively with different stakeholders on various projects and business/technical goals
Google Cloud Data tools (BigQuery, Cloud Composer/Airflow, Pub/Sub, Cloud Functions) or parallel tools on AWS
Fluent in Python, including experience working with APIs, building infrastructure tools (like custom Airflow operators), and managing data streams
High business intuition and analytical mindset, with a strong sense of how to turn raw data into insights and impact
Experience in designing and building scalable data systems for various data applications - an advantage
Experience in designing scalable data systems and analyzing data for insights - an advantage
Background in data-driven companies in large-scale environments - an advantage
This position is open to all candidates.
 
Hide
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8336266
סגור
שירות זה פתוח ללקוחות VIP בלבד
משרות דומות שיכולות לעניין אותך
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a Senior Data Engineer.
As a Senior Data Engineer, youll be more than just a coder - youll be the architect of our data ecosystem. Were looking for someone who can design scalable, future-proof data pipelines and connect the dots between DevOps, backend engineers, data scientists, and analysts.
Youll lead the design, build, and optimization of our data infrastructure, from real-time ingestion to supporting machine learning operations. Every choice you make will be data-driven and cost-conscious, ensuring efficiency and impact across the company.
Beyond engineering, youll be a strategic partner and problem-solver, sometimes diving into advanced analysis or data science tasks. Your work will directly shape how we deliver innovative solutions and support our growth at scale.
Responsibilities:
Design and Build Data Pipelines: Architect, build, and maintain our end-to-end data pipeline infrastructure to ensure it is scalable, reliable, and efficient.
Optimize Data Infrastructure: Manage and improve the performance and cost-effectiveness of our data systems, with a specific focus on optimizing pipelines and usage within our Snowflake data warehouse. This includes implementing FinOps best practices to monitor, analyze, and control our data-related cloud costs.
Enable Machine Learning Operations (MLOps): Develop the foundational infrastructure to streamline the deployment, management, and monitoring of our machine learning models.
Support Data Quality: Optimize ETL processes to handle large volumes of data while ensuring data quality and integrity across all our data sources.
Collaborate and Support: Work closely with data analysts and data scientists to support complex analysis, build robust data models, and contribute to the development of data governance policies.
Requirements:
Bachelor's degree in Computer Science, Engineering, or a related field.
Experience: 5+ years of hands-on experience as a Data Engineer or in a similar role.
Data Expertise: Strong understanding of data warehousing concepts, including a deep familiarity with Snowflake.
Technical Skills:
Proficiency in Python and SQL.
Hands-on experience with workflow orchestration tools like Airflow.
Experience with real-time data streaming technologies like Kafka.
Familiarity with container orchestration using Kubernetes (K8s) and dependency management with Poetry.
Cloud Infrastructure: Proven experience with AWS cloud services (e.g., EC2, S3, RDS).
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8320416
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
01/09/2025
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking a highly skilled and motivated Senior Data Engineer to join our dynamic team.
The ideal candidate will be a great team player that can lead and also be responsible for designing, developing, and maintaining robust data pipelines and analytical solutions to support our business objectives.
This role requires a blend of engineering and analytical skills to ensure data integrity, optimize data workflows, and provide actionable insights.
This role requires a deep understanding of financial data, system integration, and analytics to support strategic decision-making and regulatory compliance.
Your Day-to-Day:
Design, develop, and maintain scalable data pipelines and ETL processes.
Collaborate with product, analysts, and other stakeholders to understand data requirements and translate business needs into technical requirements.
Ensure data quality and integrity across various data sources.
Develop, maintain and own data models, schemas, and documentation.
Optimize database performance and troubleshoot issues.
Stay updated with the latest industry trends and best practices in data engineering and analytics.
Requirements:
Proven experience as a Data Engineer at least 3-5 years.
Expert Proficiency in SQL.
Advanced programming skills in Python.
Develop data monitoring process.
Hands-on experience with cloud data platforms (Snowflake- Advantage, OCI ect).
Understanding of Kafka and event-driven architectures for real-time financial data processing.
Familiarity with financial data models, accounting principles, and regulatory reporting.
Proven experience with cloud architecture principles.
Experience with data visualization and BI tools (Power BI, Looker, BI modeling).
Strong communication and collaboration skills.
Advanced troubleshooting skills- Excellent problem-solving skills, attention to detail and ability to analyze complex data structures.
Advantages:
Experience in the banking or fintech industry.
Experience with API integrations and financial transaction data processing.
Exposure to machine learning and predictive analytics in financial risk modeling.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8327823
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
As a customer-centric tech company, we created an insurance experience that is smart, instant, and delightful.
youll be working with a group of like-minded makers, who get a kick out of moving fast and delivering great products. We surround ourselves with some of the smartest, most motivated, creative people who are filled with positive energy and good karma.

Unlike most publicly traded companies, were nimble and efficient. We take pride in the fact that we still think and operate like a startup. We dont care much about titles and hierarchy and instead focus on innovation, bold moves, and challenging the status quo.

Were built as a lean, data-driven organization that relies on a common understating of objectives and goals to provide teams with autonomy and ownership. We dont like spending our days in meetings and we skip committees altogether. theres no such thing as going over someones head. We have zero tolerance for bureaucracy, office politics, and lean-back personalities.

As a Public Benefit Corporation and a certified B-Corp, we deliver environmental and social impact using our products and tech. Through our Giveback program, we partner with organizations such as the ACLU, New Story, The Humane Society, Malala Fund, American Red Cross, 360.org, charity: water, and dozens of others, and have donated millions towards reforestation, education, animal rights, LGBTQ+ causes, access to water, and more.
Requirements:
Were looking for an experienced and passionate Staff Data Engineer to join our Data Platform group in TLV as a Tech Lead. As the Groups Tech Lead, youll shape and implement the technical vision and architecture while staying hands-on across three specialized teams: Data Engineering Infra, Machine Learning Platform, and Data Warehouse Engineering, forming the backbone of data ecosystem.
The groups mission is to build a state-of-the-art Data Platform that drives toward becoming the most precise and efficient insurance company on the planet. By embracing Data Mesh principles, we create tools that empower teams to own their data while leveraging a robust, self-serve data infrastructure. This approach enables Data Scientists, Analysts, Backend Engineers, and other stakeholders to seamlessly access, analyze, and innovate with reliable, well-modeled, and queryable data, at scale.
In this role youll
Technically lead the group by shaping the architecture, guiding design decisions, and ensuring the technical excellence of the Data Platforms three teams
Design and implement data solutions that address both applicative needs and data analysis requirements, creating scalable and efficient access to actionable insights
Drive initiatives in Data Engineering Infra, including building robust ingestion layers, managing streaming ETLs, and guaranteeing data quality, compliance, and platform performance

Develop and maintain the Data Warehouse, integrating data from various sources for optimized querying, analysis, and persistence, supporting informed decision-makingLeverage data modeling and transformations to structure, cleanse, and integrate data, enabling efficient retrieval and strategic insights
Build and enhance the Machine Learning Platform, delivering infrastructure and tools that streamline the work of Data Scientists, enabling them to focus on developing models while benefiting from automation for production deployment, maintenance, and improvements. Support cutting-edge use cases like feature stores, real-time models, point-in-time (PIT) data retrieval, and telematics-based solutions
Collaborate closely with other Staff Engineers across to align on cross-organizational initiatives and technical strategies
Work seamlessly with Data Engineers, Data Scientists, Analysts, Backend Engineers, and Product Managers to deliver impactful solutions
Share knowledge, mentor team members, and champion engineering standards and technical excellence acros
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8297111
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
01/09/2025
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for a hands-on Data Specialist to join our growing data group, working on the practical backbone of high-scale, financial-grade systems. Youll work closely with engineers, BI, product, and business stakeholders, expert in design, build, and optimize data pipelines and integrations in a cloud-native environment.
If you thrive on solving complex data challenges, enjoy getting deep into code, and want to make an impact on fintech infrastructure, wed love to meet you.
Your Day-to-Day:
Develop, maintain, and optimize robust data pipelines and integrations across multiple systems
Build and refine data models to support analytics and operational needs
Work hands-on with data orchestration, transformation, and cloud infrastructure (AWS/Azure)
Collaborate with engineering, BI, and business teams to translate requirements into scalable data solutions
Contribute to data governance, data quality, and monitoring initiatives
Support implementation of best practices in data management and observability
Requirements:
8+ years in data engineering, data architecture, or similar roles
Deep hands-on experience with PostgreSQL, Snowflake, Oracle etc
Strong experience with ETL/ELT, data integration (Kafka, Airflow)
Proven SQL and Python skills (must)
Experience with AWS or Azure cloud environments
Familiarity with BI tools (Looker, Power BI)
Knowledge of Kubernetes and distributed data systems
Experience in financial systems or fintech (advantage)
Strong ownership, problem-solving ability, and communication skills
Comfort working in a fast-paced, multi-system environment
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8327844
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Realize your potential by joining the leading performance-driven advertising company!
As a Staff Algo Data Engineer on the Infra group, youll play a vital role in develop, enhance and maintain highly scalable Machine-Learning infrastructures and tools.
About Algo platform:
The objective of the algo platform group is to own the existing algo platform (including health, stability, productivity and enablement), to facilitate and be involved in new platform experimentation within the algo craft and lead the platformization of the parts which should graduate into production scale. This includes support of ongoing ML projects while ensuring smooth operations and infrastructure reliability, owning a full set of capabilities, design and planning, implementation and production care.
The group has deep ties with both the algo craft as well as the infra group. The group reports to the infra department and has a dotted line reporting to the algo craft leadership.
The group serves as the professional authority when it comes to ML engineering and ML ops, serves as a focal point in a multidisciplinary team of algorithm researchers, product managers, and engineers and works with the most senior talent within the algo craft in order to achieve ML excellence.
How youll make an impact:
As a Staff Algo Data Engineer Engineer, youll bring value by:
Develop, enhance and maintain highly scalable Machine-Learning infrastructures and tools, including CI/CD, monitoring and alerting and more
Have end to end ownership: Design, develop, deploy, measure and maintain our machine learning platform, ensuring high availability, high scalability and efficient resource utilization
Identify and evaluate new technologies to improve performance, maintainability, and reliability of our machine learning systems
Work in tandem with the engineering-focused and algorithm-focused teams in order to improve our platform and optimize performance
Optimize machine learning systems to scale and utilize modern compute environments (e.g. distributed clusters, CPU and GPU) and continuously seek potential optimization opportunities.
Build and maintain tools for automation, deployment, monitoring, and operations.
Troubleshoot issues in our development, production and test environments
Influence directly on the way billions of people discover the internet.
Requirements:
Experience developing large scale systems. Experience with filesystems, server architectures, distributed systems, SQL and No-SQL. Experience with Spark and Airflow / other orchestration platforms is a big plus.
Highly skilled in software engineering methods. 5+ years experience.
Passion for ML engineering and for creating and improving platforms
Experience with designing and supporting ML pipelines and models in production environment
Excellent coding skills in Java & Python
Experience with TensorFlow a big plus
Possess strong problem solving and critical thinking skills
BSc in Computer Science or related field.
Proven ability to work effectively and independently across multiple teams and beyond organizational boundaries
Deep understanding of strong Computer Science fundamentals: object-oriented design, data structures systems, applications programming and multi threading programming
Strong communication skills to be able to present insights and ideas, and excellent English, required to communicate with our global teams.
Bonus points if you have:
Experience in leading Algorithms projects or teams.
Experience in developing models using deep learning techniques and tools
Experience in developing software within a distributed computation framework.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8335929
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
31/08/2025
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
We are looking for an experienced Senior Data Engineer to join our Data team.

In this role, you will lead and strengthen our Data Team, drive innovation, and ensure the robustness of our data and analytics platforms.

A day in the life and how youll make an impact:
Design and develop high-performance data pipelines and ETL processes to support diverse business needs.
Work closely with business intelligence, sales, and other teams to integrate data solutions, ensuring seamless alignment and collaboration across functions.
Continuously improve our data analytics platforms, optimizing system performance while ensuring a robust and reliable data infrastructure.
Oversee the entire data lifecycle, from infrastructure setup and data acquisition to detailed analysis and automated reporting, driving business growth through data-driven insights.
Implement robust data quality checks, monitoring mechanisms, and data governance policies to maintain data integrity and security, troubleshooting and resolving any data-related issues efficiently.
Requirements:
B.Sc. in computer science/information systems engineering.
5+ years of experience in data engineering (Preferably from a startup company).
Familiarity with data engineering tech stack, including ETL tools (Airflow, Spark, Flink, Kafka, Pubsub).
Strong SQL expertise, working with various databases (relational and NoSQL) such as MySQL, FireStore, Redis, and ElasticSearch.
Experience with cloud-based data warehouse solutions like BigQuery, Snowflake, and Oracle, and proficiency in working with public clouds (AWS/GCP).
Coding experience with Python.
Experience with dashboard tools.
Ability to communicate ideas and analyze results effectively, both verbally and in writing.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8325317
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
22/08/2025
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
make the pursuit of creative hobbies simple and fun. Our top-grossing apps Piano, Guitar, Sing, and Draw put us on the fast track to building the worlds largest consumer subscription service, revolutionizing how people spend their time at home. With over a million monthly downloads and hundreds of thousands of daily learners worldwide, were reinventing how people engage with creative learning at home.

What Youll Be Doing:
Help improve our analytics practices by gathering better data, making it more trustworthy and accessible, and nurturing a top-grade data culture across the company.
Design, develop, scale, and optimize the data pipelines, ETLs and various data models.
Use SQL, DBT, Python and any other relevant tools and technologies.
Identify blind spots in how we use data and lead initiatives to illuminate them.
Collaborate with various stakeholders across the company to translate requirements into high-quality data pipelines and models.
We are currently rebuilding and redesigning much of our tech-stack and processes. Play a major role in designing the future data-work within the company.
Requirements:
Have 5+ years of experience as a BI Developer or Data Engineer including ETL design, implementation, optimization and maintenance.
Very strong understanding of business processes and the ability to translate business needs into relevant data products.
An independent, self-motivated learner with a strong passion for data.
Experienced in writing complex SQL queries, Data Modeling and working with DBT.
Experience analyzing user behavior and modeling data.
Proven track record of merging multiple data sources to develop a single source of truth for reporting/analysis/research needs.
Experience preparing data for business intelligence tools such as Tableau, Looker, or similar platforms.
Bachelors or Masters Degree in Computer Science, Information Systems or a related field.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8314725
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: Full Time
We are seeking an experienced Solutions Data Engineer who possess both technical depth and strong interpersonal skills to partner with internal and external teams to develop scalable, flexible, and cutting-edge solutions. Solutions Engineers collaborate with operations and business development to help craft solutions to meet customer business problems.
A Solutions Engineer works to balance various aspects of the project, from safety to design. Additionally, a Solutions Engineer researches advanced technology regarding best practices in the field and seek to find cost-effective solutions.
Job Description:
Were looking for a Solutions Engineer with deep experience in Big Data technologies, real-time data pipelines, and scalable infrastructuresomeone whos been delivering critical systems under pressure, and knows what it takes to bring complex data architectures to life. This isnt just about checking boxes on tech stacksits about solving real-world data problems, collaborating with smart people, and building robust, future-proof solutions.
In this role, youll partner closely with engineering, product, and customers to design and deliver high-impact systems that move, transform, and serve data at scale. Youll help customers architect pipelines that are not only performant and cost-efficient but also easy to operate and evolve.
We want someone whos comfortable switching hats between low-level debugging, high-level architecture, and communicating clearly with stakeholders of all technical levels.
Key Responsibilities:
Build distributed data pipelines using technologies like Kafka, Spark (batch & streaming), Python, Trino, Airflow, and S3-compatible data lakesdesigned for scale, modularity, and seamless integration across real-time and batch workloads.
Design, deploy, and troubleshoot hybrid cloud/on-prem environments using Terraform, Docker, Kubernetes, and CI/CD automation tools.
Implement event-driven and serverless workflows with precise control over latency, throughput, and fault tolerance trade-offs.
Create technical guides, architecture docs, and demo pipelines to support onboarding, evangelize best practices, and accelerate adoption across engineering, product, and customer-facing teams.
Integrate data validation, observability tools, and governance directly into the pipeline lifecycle.
Own end-to-end platform lifecycle: ingestion → transformation → storage (Parquet/ORC on S3) → compute layer (Trino/Spark).
Benchmark and tune storage backends (S3/NFS/SMB) and compute layers for throughput, latency, and scalability using production datasets.
Work cross-functionally with R&D to push performance limits across interactive, streaming, and ML-ready analytics workloads.
Requirements:
24 years in software / solution or infrastructure engineering, with 24 years focused on building / maintaining large-scale data pipelines / storage & database solutions.
Proficiency in Trino, Spark (Structured Streaming & batch) and solid working knowledge of Apache Kafka.
Coding background in Python (must-have); familiarity with Bash and scripting tools is a plus.
Deep understanding of data storage architectures including SQL, NoSQL, and HDFS.
Solid grasp of DevOps practices, including containerization (Docker), orchestration (Kubernetes), and infrastructure provisioning (Terraform).
Experience with distributed systems, stream processing, and event-driven architecture.
Hands-on familiarity with benchmarking and performance profiling for storage systems, databases, and analytics engines.
Excellent communication skillsyoull be expected to explain your thinking clearly, guide customer conversations, and collaborate across engineering and product teams.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8325726
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
Required Data Engineer
About the Data Team:
Our Data & Analytics team believes theres a better way to make data useful than just creating endless dashboards. We focus on in-depth analysis and building scalable, trustworthy data solutions that help every team make faster, smarter decisions. From analytics and business intelligence to data pipelines and predictive models, we turn raw information into real impact. If youre passionate about finding radical new ways to leverage data, youll fit right in.
On your day to day:
In our day-to-day work, we design and maintain reliable data pipelines using Python and modern orchestration frameworks to move and transform raw data from a variety of sources. This includes working with APIs, event streams, and databases to ingest and process data at scale. We implement and maintain models, manage complex data transformations, and collaborate closely with other teams to ensure that the right data is available at the right time. Whether we're tuning a performance, debugging a transformation step, or shipping an AI-driven feature, every day is a chance to build something impactful.
Requirements:
Strong proficiency building data pipelines and managing data warehouses (2+ years)
Working knowledge of Python for data extraction, processing and automation tasks.
Experience with event tracking and NoSQL databases.
Experience with dbt or similar tools.
Solid grasp of software engineering best practices, including query optimization, version control (e.g. Git), code reviews, and documentation.
Analytical mindset with strong problem-solving skills.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8322355
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time
we are looking for a Data Engineer to join our growing team!
This is a great opportunity to be part of one of the fastest-growing infrastructure companies in history, an organization that is in the center of the hurricane being created by the revolution in artificial intelligence.
"our company's data management vision is the future of the market."- Forbes
we are the data platform company for the AI era. We are building the enterprise software infrastructure to capture, catalog, refine, enrich, and protect massive datasets and make them available for real-time data analysis and AI training and inference. Designed from the ground up to make AI simple to deploy and manage, our company takes the cost and complexity out of deploying enterprise and AI infrastructure across data center, edge, and cloud.
Our success has been built through intense innovation, a customer-first mentality and a team of fearless company ronauts who leverage their skills & experiences to make real market impact. This is an opportunity to be a key contributor at a pivotal time in our companys growth and at a pivotal point in computing history.
In this role, you will be responsible for:
Designing, building, and maintaining scalable data pipeline architectures
Developing ETL processes to integrate data from multiple sources
Creating and optimizing data models for efficient storage and retrieval
Implementing data quality controls and monitoring systems
Collaborating with data scientists and analysts to deliver data solutions
Building and maintaining data warehouses and data lakes
Performing in-depth data analysis and providing insights to stakeholders
Taking full ownership of data quality, documentation, and governance processes
Building and maintaining comprehensive reports and dashboards
Ensuring data security and regulatory compliance.
Requirements:
Bachelor's degree in Computer Science, Engineering, or related field
3+ years experience in data engineering
Strong proficiency in SQL and Python
Experience with ETL tools and data warehousing solutions
Knowledge of big data technologies (Hadoop, Spark, etc.)
Experience with cloud platforms (AWS, Azure, or GCP)
Understanding of data modeling and database design principles
Familiarity with data visualization tools - Tableau, Sisense
Strong problem-solving and analytical skills
Excellent communication and collaboration abilities
Experience with version control systems (Git).
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8326022
סגור
שירות זה פתוח ללקוחות VIP בלבד