Currently Hiring: Data Infra Scientist / Data Workflow Engineer
It’s our dream to create the ultimate iterative workflow for data science that talks SQL. Let’s break that down.
Building a great model as a data scientist happens in experimental iterations. That’s why so many of us live in Notebooks. The analytics data that you work on clearly does not fit a local machine, so it’s either working with samples, or spinning up complex data lakes. We believe there is so much development going on in the SQL space, that today it’s possible to create a single interactive & experimental environment for the data scientist to build models, while iterating on the full dataset.
We are building a workflow where the data scientist can perform any operation or basic model on a full SQL stored dataset, and get interactive results in a dataframe in a Notebook. Then, easily go back a few steps to any point in the DAG, apply some other operations or models again, and with a single click get the results back in a dataframe. From that point, have the full freedom to add any machine learning model on that.
To make this all happen, we have started our journey in the space of product analytics. Creating a standardized taxonomy and validation that carries through a whole pipeline that ends all the way in the notebook. We see product analytics as a stepping stone in a much larger ambition. To create the dream experimental modelling workflow for data science.
So, where is the product?
In a repository near you after summer. Our team is working hard to get the initial version ready. Then we will continue to develop in public.
That is a weird job title!
Well, to start we don’t really get excited about titles. But more importantly; you will be joining an early team of 10 analytics enthusiasts that are together building a great product. The lines between data science, data engineering, frontend and ops are just very blurry in our team, as everyone is taking multiple parts of that.
This role is at the intersection of data science and data engineering. What we are looking for in this role:
- You know your way in Python, Notebooks & SQL, but also get excited about data engineering tech like dbt.
- You have at least 5 years of working experience in a data science / data engineering role and ideally worked with analytics data.
- You have a master’s degree in data science, software engineering, math, physics or something related.
- You have a can-do attitude. Even when the problem seems huge, you're able to chop it up and solve the chunks
- You are fluent in English.
- This is a remote role in EU timezone with regular team offsites in The Netherlands.
What we offer
A role in a super driven and experienced team that is on a mission. Our team has over a decade of experience in analytics and has made it their core focus to build the best possible workflow for data science. Thanks to the support of Fly Ventures & LocalGlobe we can focus on the long-term and take the time to get the product just right. They were also the early backers of startups like Figma, Tessian and Robinhood.
Early days in the product, where you will be able to leave your mark in the open-source community for data science. Our goals are very ambitious and this will allow you to grow with the company.
Like what you have read and think you meet the mark? Then we'd love to hear from you.