ID: 31598
We are looking for driven Data Engineers who enjoy solving problems, who initiate solutions and discussions and who believe that any challenge can be scaled with the right mindset and tools.
As a Data Engineer, you will be in charge of the development, performance, quality, and scaling of our data pipelines, focusing especially on data integrity and quality. You will work independently and will also be responsible for making technical decisions within a team.
End-to-end ownership of data integrity and quality in our data pipelines.
Providing tools that improve Data Quality company-wide.
Providing self-organizing tools that help the analytics community discover data, assess quality, explore usage, and find peers with relevant expertise.
Experimenting with new tools and technologies to meet business requirements regarding performance, scaling, and data quality.
Rapidly developing next-generation scalable, flexible, and high-performance data pipelines.
Solving issues with data and data pipelines, prioritizing based on customer impact.
Acting as an intermediary for problems, with both technical and non-technical audiences.
Supplying to a high scale, complex, world renowned product and seeing real-time impact of your work on millions of travelers worldwide
Promote and drive impactful and innovative engineering solutions.
Technical, behavioral and interpersonal competence advancement via on-the-job opportunities, experimental projects, hackathons, conferences, and active community participation
Requirements: You have more than +3 years of professional experience as a Software Developer/Data Engineer and hold a degree in a quantitative field
You have hands-on experience with schema design and data modeling and working with ML scientists and ML engineers to provide production level ML solutions.
You have experience designing systems E2E and knowledge of basic concepts (lb, db, caching, NoSQL, etc)
You have built production data pipelines in the cloud, setting up data-lake and server-less solutions;
Proven knowledge of Python is required, JVM languages will be a plus
Knowledge of Snowflake, Hadoop, Hive, MySQL imports, Airflow, Oozie, Spark, CDC, Kafka.
Experience with Data Warehousing and ETL/ELT pipelines
Experience with ML monitoring and evaluation tools such as DBT.
You are excited by handling petabytes of data via Big data technologies.
Excellent communication skills - verbal and written
This position is open to all candidates.