In this role, you will work closely with data scientists, analysts, and other cross-functional teams to understand their data needs and requirements. You will also be responsible for ensuring that data is easily accessible and can be used to support data-driven decision making.
Location: Ramat Gan, Hybrid Model
What will you do?
Design, build, and maintain data pipelines to extract, transform, and load data from various sources, including databases, APIs, and flat files
Enhance our data warehouse system to dynamically support multiple analytics use cases
Reimplement and productize scientific computational methods
Implement data governance policies and procedures to ensure data quality, security, and privacy
Collaborate with data scientists and other cross-functional teams to understand their data needs and requirements
Develop and maintain documentation for data pipelines, processes, and systems
Requirements: Bachelor's or Master's degree in a related field (e.g. computer science, data science, engineering, computational biology)
At least 5 years of experience with programming languages, specifically Python
Must have at least 3+ years of experience as a Data Engineer, ideally with experience in multiple data ecosystems
Proficiency in SQL and experience with database technologies (e.g. MySQL, PostgreSQL, Oracle)
Familiarity with data storage technologies (e.g. HDFS, NoSQL databases)
Experience with ETL tools (e.g. Apache Beam, Apache Spark)
Experience with orchestration tools (e.g. Apache Airflow, Dagster)
Experience with data warehousing technologies (ideally BigQuery)
Experience working with large and complex data sets
Experience working in a cloud environment
Strong problem-solving and communication skills
Familiarity with biotech or healthcare data - an advantage
This position is open to all candidates.