You will specialize in designing and building world class, scalable data architectures, ensuring reliable data flow and integration for groundbreaking biotechnological research. Your expertise in big data tools and pipelines will accelerate our ability to derive actionable insights from complex datasets, driving innovations in improving patients outcome and in delivering life savings treatment solutions.
In this role, you will work closely with data scientists, analysts, and other cross-functional teams to understand their data needs and requirements. You will also be responsible for ensuring that data is easily accessible and can be used to support data-driven decision making.
Location: Ramat Gan, Hybrid Model
What will you do?
Design, build, and maintain data pipelines to extract, transform, and load data from various sources, including databases, APIs, and flat files.
Enhance our data warehouse system to dynamically support multiple analytics use cases.
Reimplement and productize scientific computational methods.
Implement data governance policies and procedures to ensure data quality, security, and privacy.
Collaborate with data scientists and other cross-functional teams to understand their data needs and requirements.
Develop and maintain documentation for data pipelines, processes, and systems.
Requirements: We will only consider data engineers with strong coding skills with an extensive background in data orchestration, data warehousing and ETL tools.
Required qualifications:
Bachelor's or Master's degree in a related field (e.g. computer science, data science, engineering, computational biology).
At least 5 years of experience with programming languages, specifically Python.
Must have at least 3+ years of experience as a Data Engineer, ideally with experience in multiple data ecosystems.
Proficiency in SQL and experience with database technologies (e.g. MySQL, PostgreSQL, Oracle).
Familiarity with data storage technologies (e.g. HDFS, NoSQL databases).
Experience with ETL tools (e.g. Apache Beam, Apache Spark).
Experience with orchestration tools (e.g. Apache Airflow, Dagster).
Experience with data warehousing technologies (ideally BigQuery).
Experience working with large and complex data sets.
Experience working in a cloud environment.
Strong problem-solving and communication skills.
Familiarity with biotech or healthcare data - an advantage.
Desired personal traits:
You want to make an impact on humankind.
You prioritize We over I.
You enjoy getting things done and striving for excellence.
You collaborate effectively with people of diverse backgrounds and cultures.
You constantly challenge your own assumptions, pushing for continuous improvement.
You have a growth mindset.
You make decisions that favor the company, not yourself or your team.
You are candid, authentic, and transparent.
This position is open to all candidates.