דרושים » מערכות מידע » Student Position - BI Systems Analyst (Internal Audit)

משרות על המפה
 
בדיקת קורות חיים
VIP
הפוך ללקוח VIP
רגע, משהו חסר!
נשאר לך להשלים רק עוד פרט אחד:
 
שירות זה פתוח ללקוחות VIP בלבד
AllJObs VIP
כל החברות >
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
דרושים באתוסיה חברת השמה לתחום ההיי טק וביוטק
Location: Tel Aviv-Yafo
:The Project Manager of the Global Strategy Planning (S P) Team plays a pivotal role in supporting the Head of Global Internal Audit S P in managing audit systems and enhancing business intelligence ( BI ) capabilities. This role focuses on optimizing audit tools, data governance, and leveraging BI and AI solutions to enable data -driven insights, improve operational performance, and ensure system functionality.

Required Education
3rd year student (end of 2nd year / beginning of 3rd year) in:
o Industrial Management Engineering
o Information Systems / IT Engineering
From a recognized Israeli or international academic institution
Requirements:
Required Skills Experience

Strong organizational, planning, and time management skills
Self driven, proactive, with the ability to manage multiple tasks and projects
Excellent analytical thinking and attention to detail
Strong computer skills, including advanced use of MS Office (Excel, PowerPoint, Word)
Ability to gather, analyze, and synthesize data accurately and efficiently
Exposure to AI tools, including:
o Writing effective prompts for AI / copilots
o Working with AI?based agents or automation workflows
Fluent written and spoken English, including ability to draft professional documents
Advantages (Nice to Have)
Experience with data analytics, automation, or reporting tools
Familiarity with advanced information systems ( ERP, GRC, BI tools, etc.)
Experience working in a large, global organization, preferably in pharma, life sciences, or manufacturing
This position is open to all candidates.
 
Hide
הגשת מועמדות
עדכון קורות החיים לפני שליחה
8597211
סגור
שירות זה פתוח ללקוחות VIP בלבד
משרות דומות שיכולות לעניין אותך
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
דרושים בNishapro
Job Type: Full Time and Hybrid work
This is a full-time, hybrid role for a business intelligence Developer based in Ramat Gan, Israel.
The Developer will be responsible for for end-to-end BI solutions, including managing ETL processes, architecting data models, and designing impactful dashboards.
In this position, you will work closely with senior management to translate data into actionable business insights.
Requirements:
4+ years of experience with QlikSense and Nprinting
Strong proficiency in SQL and a deep understanding of ETL processes
Academic degree in a relevant field ( Computer Science, Industrial Engineering, etc.)
High level of self-learning ability, independence, and a proactive mindset.
Experience working with salesforce or Priority, familiarity with additional BI tools such as Fabric or Power BI - an advantage.
This position is open to all candidates.
 
Show more...
הגשת מועמדות
עדכון קורות החיים לפני שליחה
8528532
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
1 ימים
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
We are looking for a professional and enthusiastic BI Developer to join our team!
Job Id: 25366
The BI Developer will be responsible for defining and building our data pipelines and analytics infrastructure. You will work closely with product managers, R&D engineers, and business stakeholders to support new initiatives and proactively improve our data solutions and AI agents. You will help create and manage data processes for actionable insights and operational needs.
If you are passionate about data modeling and data processing, with excellent visualization and analytical skills, this job is for you.
Key Responsibilities
As a BI Developer, you will be responsible for:
End-to-End Solutions: Design, build, and maintain easy-to-use data solutions. This includes orchestration, data modeling, dashboards, building data agents, monitoring, alerting, and experimentation models.
AI Integration: Actively use AI tools to accelerate development-from writing optimized SQL and automating repetitive tasks to building intelligent agents.
Collaboration: Work effectively with cross-department stakeholders to capture analytical requirements and translate them into technical reality.
Insights: Analyze complex datasets to provide useful, actionable recommendations that drive business value.
Requirements:
SQL & Python: Full proficiency in SQL (5+ years) manipulating large, complex, semi-structured datasets, and at least 3 years of practical Python experience.
dbt: Hands-on experience with dbt for data transformation (experience with dbt Core is a significant advantage).
Data Architecture & Ecosystems: Deep understanding of BI methodologies and experience with decoupled/distributed architectures and file-based systems (e.g., Athena, S3), and data warehouses (e.g., Snowflake, BigQuery, Redshift).
Data Visualization: At least 2 years of experience developing visualizations in Power BI (including DAX). Must have solid data storytelling skills and the ability to translate business processes into actionable dashboards.
AI Fluency: Demonstrated ability to leverage AI tools to speed up coding, automate workflows, and enhance data delivery.
Education: BSc in Computer Science, Mathematics, Industrial Engineering, or equivalent academic knowledge.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8597076
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Part Time and May Be Suitable For Students
Were looking for a Market Research Student to join our team and work closely with the Product department. You will operate at the intersection of Product, technology, market solutions, and Business Strategy. This role is about impact - not just ownership, but streamlining market research that helps us build the next gen of products.
This role is a part-time position with a scope of 10-20 flexible hours per week. The position is based in Tel Aviv.
This might be for you if:
You have a sharp eye for detail and can spot product nuances and value propositions.
You are a self-driven explorer who enjoys getting hands-on with new tools to understand their target audience and user experience.
You are comfortable diving deep into understanding data quality, coverage, and the competitive landscape of tech products.
You care about outcomes and are accountable for turning raw observations into precise, high-quality insights.
What youll be doing:
Hands-on Benchmarking: Actively exploring and deconstructing industry-leading solutions to identify best-in-class user journeys and synergy with complementary systems.
Research Operations: Owning the maintenance and accessibility of our internal research assets, ensuring the environment is optimized for the Product and R&D teams.
Precision Reporting: Delivering structured, accurate intelligence reports that influence our technical roadmap. You are responsible for the accuracy and depth of the final output.
Intelligence Augmentation: Leveraging AI tools to accelerate your work and synthesis, while maintaining human oversight to ensure every deliverable is precise.
Infrastructure Optimization: Managing the administrative side of our research ecosystem (procurement and access) to ensure the "lab" is always ready for the team.
Requirements:
2nd-year student in Industrial Engineering & Management or Information Systems (B.Sc.).
Accountability: You take full ownership of your deliverables. You understand that the Product team relies on your research to make strategic decisions, and you ensure your output is validated, precise, and professional.
GenAI Orientation: You use GenAI tools to conduct research and gather information
Product Sense: A strong natural curiosity about product logic, value propositions, and how different audiences interact with software.
Data Integrity: A meticulous approach to work; you take pride in the accuracy, data coverage analysis, and professional standard of your work.
Operational Reliability: Highly organized and responsible; able to handle the lifecycle of research resources and accounts independently.
Communication: Fluent/Native English is essential for analyzing global markets and producing professional documentation.
Bonus Points:
Experience with automation tools such as N8N or Zapier.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8595433
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
1 ימים
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
We are looking for a talented Data Engineer to join our analytics team in the Big Data Platform group.
Job Id: 25380
You will support our product and business data initiatives, expand our data warehouse, and optimize our data pipeline architecture with an AI first attitude.
The ideal candidate is experienced in leveraging AI tools as part of modern data pipeline development, enabling scalable solutions, accelerating delivery, and continuously exploring new approaches and technologies.
The right candidate is excited by the prospect of building the data architecture for the next generation of products and data initiatives.
This is a unique opportunity to join a team full of outstanding people making a big impact on us.
We work on multiple products in many domains to deliver truly innovative solutions in the Cyber Security and Big Data realm.
This role requires the ability to collaborate closely with both R&D teams and business stakeholders, to understand their needs and translate them into robust and scalable data solutions.
Key Responsibilities
Maintain and develop enterprise-grade Data Warehouse and Data Lake environments
Create data infrastructure for various R&D groups across the organization to support product development and optimization
Work with data experts to assist with technical data-related issues and support infrastructure needs
Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, redesigning infrastructure for scalability
Build and maintain robust ETL/ELT pipelines for data ingestion, transformation, and delivery across various systems
Incorporate AI-assisted tools into data pipeline design, development, and optimization to improve efficiency, scalability, and innovation
Requirements:
B.Sc. in Engineering or a related field
3+ years of experience as a Data Engineer working on production systems
Advanced SQL knowledge and experience with relational databases
Proven experience using Python
Hands-on experience building, optimizing, and automating data pipelines, architectures, and data sets
Experience in creating and maintaining ETL/ELT processes
Strong project management and organizational skills
Strong collaboration skills with both technical (R&D) and non-technical (business) teams
Experience using AI tools as part of the data engineering workflow, with a mindset of experimentation, working at scale, and exploring new technologies
Advantage: Azure data services, Databricks, EventHub, and Spark.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8597003
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
Location: Tel Aviv-Yafo
Job Type: More than one
Required Project Configuration Specialist
About the Role:
As a Project Configuration Specialist, you will be responsible for:
Configuration of our product for clients' needs using our configuration system
Setup our Digital Twin for new projects into our system
Solve technical configuration issues.
Work together with Project Managers and Customer Success to support the project and clients needs.
Ongoing project maintenance through the project life cycle.
Be part of creating the Teams tools and workflows internally and with the product/development team.
Requirements:
B.Sc. 4th-year student in Industrial Engineering and Management or Information Systems, or a graduate from the past year.
Strong analytical skills with a passion for data
Ability to manage multiple tasks at the same time
Excellent communication skills; ability to quickly and clearly articulate key issues and to listen and capture concerns of others
Able to demonstrate initiative and a proactive approach to daily tasks
Flexible attitude and problem-solving
Fast learner, adaptive, and thrive in a fast-paced environment
Excellent English, verbal and written
Immediate availability
Be a great person and a team player.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8566191
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
לפני 5 שעות
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
This role has been designed as Hybrid with an expectation that you will work on average 2 days per week from an office.

We are looking for a talented Data Engineer to help build and enhance the data platform that supports analytics, operations, and data-driven decision-making across the organization. You will work hands-on to develop scalable data pipelines, improve data models, ensure data quality, and contribute to the continuous evolution of our modern data ecosystem.

Youll collaborate closely with Senior Engineers, Analysts, Data Scientists, and stakeholders across the business to deliver reliable, well-structured, and well-governed data solutions.


What Youll Do:

Engineering & Delivery

Build, maintain, and optimize data pipelines for batch and streaming workloads.

Develop reliable data models and transformations to support analytics, reporting, and operational use cases.

Integrate new data sources, APIs, and event streams into the platform.

Implement data quality checks, testing, documentation, and monitoring.

Write clean, performant SQL and Python code.

Contribute to improving performance, scalability, and cost-efficiency across the data platform.

Collaboration & Teamwork

Work closely with senior engineers to implement architectural patterns and best practices.

Collaborate with analysts and data scientists to translate requirements into technical solutions.

Participate in code reviews, design discussions, and continuous improvement initiatives.

Help maintain clear documentation of data flows, models, and processes.

Platform & Process

Support the adoption and roll-out of new data tools, standards, and workflows.

Contribute to DataOps processes such as CI/CD, testing, and automation.

Assist in monitoring pipeline health and resolving data-related issues.
Requirements:
What Were Looking For

2-5+ years of experience as a Data Engineer or similar role.

Hands-on experience with Snowflake (mandatory)-including SQL, modeling, and basic optimization.

Experience with dbt (or similar)-model development, tests, documentation, and version control workflows.

Strong SQL skills for data modeling and analysis.

Proficiency with Python for pipeline development and automation.

Experience working with orchestration tools (Airflow, Dagster, Prefect, or equivalent).

Understanding of ETL/ELT design patterns, data lifecycle, and data modeling best practices.

Familiarity with cloud environments (AWS, GCP, or Azure).

Knowledge of data quality, observability, or monitoring concepts.

Good communication skills and the ability to collaborate with cross-functional teams.


Nice to Have:

Exposure to streaming/event technologies (Kafka, Kinesis, Pub/Sub).

Experience with data governance or cataloging tools.

Basic understanding of ML workflows or MLOps concepts.

Experience with infrastructure-as-code tools (Terraform, CloudFormation).

Familiarity with testing frameworks or data validation tools.

Additional Skills:

Cloud Architectures, Cross Domain Knowledge, Design Thinking, Development Fundamentals, DevOps, Distributed Computing, Microservices Fluency, Full Stack Development, Security-First Mindset, User Experience (UX).
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8598093
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
1 ימים
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
Required ML Data Engineer
Israel: Tel Aviv/ Hybrid (Israel)
R&D | Full Time | Job Id: 24792
Key Responsibilities
Your Impact & Responsibilities:
As a Data Engineer - AI Technologies, you will be responsible for building and operating the data foundation that enables our LLM and ML research: from ingestion and augmentation, through labeling and quality control, to efficient data delivery for training and evaluation.
You will:
Own data pipelines for LLM training and evaluation
Design, build and maintain scalable pipelines to ingest, transform and serve large-scale text, log, code and semi-structured data from multiple products and internal systems.
Drive data augmentation and synthetic data generation
Implement and operate pipelines for data augmentation (e.g., prompt-based generation, paraphrasing, negative sampling, multi-positive pairs) in close collaboration with ML Research Engineers.
Build tagging, labeling and annotation workflows
Support human-in-the-loop labeling, active learning loops and semi-automated tagging. Work with domain experts to implement tools, schemas and processes for consistent, high-quality annotations.
Ensure data quality, observability and governance
Define and monitor data quality checks (coverage, drift, anomalies, duplicates, PII), manage dataset versions, and maintain clear documentation and lineage for training and evaluation datasets.
Optimize training data flows for efficiency and cost
Design storage layouts and access patterns that reduce training time and cost (e.g., sharding, caching, streaming). Work with ML engineers to make sure the right data arrives at the right place, in the right format.
Build and maintain data infrastructure for LLM workloads
Work with cloud and platform teams to develop robust, production-grade infrastructure: data lakes / warehouses, feature stores, vector stores, and high-throughput data services used by training jobs and offline evaluation.
Collaborate closely with ML Research Engineers and security experts
Translate modeling and security requirements into concrete data tasks: dataset design, splits, sampling strategies, and evaluation data construction for specific security use.
Requirements:
3+ years of hands-on experience as a Data Engineer or ML/Data Engineer, ideally in a product or platform team.
Strong programming skills in Python and experience with at least one additional language commonly used for data / backend (e.g., SQL, Scala, or Java).
Solid experience building ETL / ELT pipelines and batch/stream processing using tools such as Spark, Beam, Flink, Kafka, Airflow, Argo, or similar.
Experience working with cloud data platforms (e.g., AWS, GCP, Azure) and modern data storage technologies (object stores, data warehouses, data lakes).
Good understanding of data modeling, schema design, partitioning strategies and performance optimization for large datasets.
Familiarity with ML / LLM workflows: train/validation/test splits, dataset versioning, and the basics of model training and evaluation (you dont need to be the primary model researcher, but you understand what the models need from the data).
Strong software engineering practices: version control, code review, testing, CI/CD, and documentation.

Ability to work independently and in collaboration with ML engineers, researchers and security experts, and to translate high-level requirements into concrete data engineering tasks. 
Nice to Have 
Experience supporting LLM or NLP workloads, including dataset construction for pre-training / fine-tuning, or retrieval-augmented generation (RAG) pipelines. 
Familiarity with ML tooling such as experiment tracking (e.g., Weights & Biases, MLflow) and ML-focused data tooling (feature stores, vector databases). 
Background in security / cyber domains (logs, alerts, incidents, SOC workflows) or other high-volume, high-variance data environments. 
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8597480
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
We're seeking an outstanding and passionate Data Platform Engineer to join our growing R&D team.

You will work in an energetic startup environment following Agile concepts and methodologies. Joining the company at this unique and exciting stage in our growth journey creates an exceptional opportunity to take part in shaping Finaloop's data infrastructure at the forefront of Fintech and AI.

What you'll do:

Design, build, and maintain scalable data pipelines and ETL processes for our financial data platform.

Develop and optimize data infrastructure to support real-time analytics and reporting.

Implement data governance, security, and privacy controls to ensure data quality and compliance.

Create and maintain documentation for data platforms and processes.

Collaborate with data scientists and analysts to deliver actionable insights to our customers.

Troubleshoot and resolve data infrastructure issues efficiently.

Monitor system performance and implement optimizations.

Stay current with emerging technologies and implement innovative solutions.
Requirements:
What you'll bring:

3+ years experience in data engineering or platform engineering roles.

Strong programming skills in Python and SQL.

Experience with orchestration platforms like Airflow/Dagster/Temporal.

Experience with MPPs like Snowflake/Redshift/Databricks.

Hands-on experience with cloud platforms (AWS) and their data services.

Understanding of data modeling, data warehousing, and data lake concepts.

Ability to optimize data infrastructure for performance and reliability.

Experience working with containerization (Docker) in Kubernetes environments.

Familiarity with CI/CD concepts.

Fluent in English, both written and verbal.

And it would be great if you have (optional):

Experience with big data processing frameworks (Apache Spark, Hadoop).

Experience with stream processing technologies (Flink, Kafka, Kinesis).

Knowledge of infrastructure as code (Terraform).

Experience building analytics platforms.

Experience building clickstream pipelines.

Familiarity with machine learning workflows and MLOps.

Experience working in a startup environment or fintech industry.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8573265
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
01/03/2026
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
Were looking for a Senior Data Architect to help shape how high-growth startups build and scale on AWS. In this role, youll design and deliver end-to-end data and analytics solutions - from architecture and pipelines to visualization and insights - guiding customers from concept through production. Youll work closely with startup founders, technical leaders, and account executives to create scalable, cost-efficient architectures that drive real business impact.
📍 Work location - hybrid from Tel Aviv
If you are interested in this opportunity, please submit your CV in English.
Responsibilities:
Design, develop, and implement data & analytics solutions to meet business requirements and create cost-efficient, highly available, and scalable customer solutions, including Well-Architected reviews and SoW.
Research and analyze current solutions and initiate improvement plans.
Collaborate with other engineers and stakeholders to ensure solutions are designed and developed according to best practices.
Lead workshops, POCs, and architecture reviews with startup customers, conferences, webinars, and more.
Stay up to date on Data Engineering and Analytics trends and contribute to internal enablement.
Frequent travels - locally (on-demand to meet with customers and partners and attend local events) and abroad (at least once a quarter).
Requirements:
3+ years of hands-on experience in AWS, including solution design, migration, and maintenance
2+ years in customer-facing technical roles (e.g., SRE, Cloud Architect, Customer Engineer)
Production experience with AWS infrastructure, data services, and real-time data processing
Proficiency in a wide range of AWS services (e.g., EC2, S3, RDS, Lambda, IAM, VPC, CloudFormation, DynamoDB)
Skilled in AWS analytics tools (Glue, Athena, Redshift, EMR, Kinesis, MSK, QuickSight, dbt)
Understanding of information security best practices
Strong verbal and written communication in English and local language
Ability to lead end-to-end technical engagements and work in fast-paced environments
AWS Solutions Architect - Associate certification
Experience with Iceberg- an advantage
Experience with Kubernetes, CI/CD, and DevOps tools - an advantage
Experience with ETL processes, data lakes, and pipelines - an advantage
Experience writing SOWs, HLDs, and effort estimates - an advantage
AWS Professional or Data Analytics/Data Engineer certifications - an advantage
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8565491
סגור
שירות זה פתוח ללקוחות VIP בלבד
סגור
דיווח על תוכן לא הולם או מפלה
מה השם שלך?
תיאור
שליחה
סגור
v נשלח
תודה על שיתוף הפעולה
מודים לך שלקחת חלק בשיפור התוכן שלנו :)
1 ימים
חברה חסויה
Location: Tel Aviv-Yafo
Job Type: Full Time and Hybrid work
Required Data Engineer
Israel: Tel Aviv/ Hybrid (Israel)
R&D | Full Time | Job Id: 25316
Why Join Us?
We are building next-generation GenAI security intelligence and SaaS Security Posture Management (SSPM) solutions that protect enterprises worldwide. If you enjoy turning complex security data into actionable insights and delivering end-to-end systems, this role is for you.
About the role:
You will own, build, and maintain our Pythonic data pipeline and enrichment system on top of PostgreSQL and BigQuery. This system powers security analytics, detections, and intelligence. A core part of your job will be to design and implement new components, improve reliability and performance, and ensure data quality and observability.
Key Responsibilities:
Own, build, and maintain production data pipelines and enrichment services using Python, PostgreSQL, and BigQuery.
Architect data systems end to end, including design, deployment, monitoring, and iterative improvement.
Analyze complex security datasets and SaaS telemetry to uncover risks, patterns, and opportunities.
Research emerging threat vectors and contribute to automated intelligence feeds and published reports.
Work across security domains such as SSPM, Shadow Integrations, DLP, and GenAI Protection.
Requirements:
4+ years in data-focused roles (engineering, analytics, science)
Strong SQL and Python skills
Experience with cloud platforms (GCP, AWS, Azure) and modern data warehouses (BigQuery, Databricks)
Proven ability to build data infrastructure from scratch
Ability to turn complex data into actionable insights
Fast learner with systematic problem-solving skills
Comfortable with technical research in unfamiliar domains
Independent and determined, with strong collaboration skills
BSc in Computer Science, Mathematics, Statistics, or related field
Excellent communication skills for technical and non-technical audiences.
This position is open to all candidates.
 
Show more...
הגשת מועמדותהגש מועמדות
עדכון קורות החיים לפני שליחה
עדכון קורות החיים לפני שליחה
8597109
סגור
שירות זה פתוח ללקוחות VIP בלבד