(Senior) Data Engineer

Job description

Snappet is looking for a Data Engineer who will join our Data Science-Machine Learning team and take full ownership of turning proven algorithms and deploying them into machine learning models in production.


Improve the learning results of children: that’s what we do at Snappet. We develop smart software that allows teachers and children to get the best results. Our software is used in more than 3500 schools across Europe and now also in the United States and used by more than 300.000 pupils on a daily basis. We value the continuous improvement of our products to support all these schools and pupils.


As Data Engineer you play a crucial role in the continuous improvement of learning results, by designing solutions capable of analyzing massive data and drawing the right conclusions, and by improving algorithms and machine learning models serving hundreds of thousands of customers daily. Through data analysis pipeline design and implementation, you continuously proof the effects of improvements on learning results.


You collaborate closely with data scientists, developers, designers and product owners. You take full ownership and won't be put off by complex problems or massive amounts of data.


We believe in Agile and autonomy. We release fast and often. We are driven by the fact that we actively contribute to improving education. Our teams are empowered to work in the best way possible. We continuously improve our working environment. Besides the teamwork, development teams regularly experiment and learn new technologies and collaborate with others on this. We have monthly ‘pizza sessions’ where we get together and share knowledge on all things tech or fun.


The ideal candidate

  • Relevant higher education or university degree;
  • 2+ years’ work experience in a business environment;
  • Knowledge of and experience with data engineering best practices is crucial, including the ability to work quickly and independently, communicate well, and furthermore knowing how to devise working solutions based on proven machine learning models, and which also generalize towards future workloads;
  • The ability to deal with (big) data workloads is considered a given, like how to deal with daily data quality issues, big data scaling issues (like performance issues and memory constraint issues), etc.
  • Advanced knowledge of Python, SQL and Apache Spark;
  • Knowledge of and experience with AWS (especially S3, Athena, DynamoDB, RedShift)
  • Knowledge of and experience with Git for source code version-control. Experience with GitHub or Gitlab is a plus.
  • Knowledge of and experience with column-oriented data storage formats or Presto/Athena is a plus;
  • Knowledge of and experience with Docker (Compose). Knowledge of and experience with Kubernetes or Airflow is a plus;
  • Knowledge of machine learning and deep learning models is a plus;
  • Neural network frameworks such as PyTorch, Tensorflow, caffe2, Keras; are a plus;
  • Team player with flexible, collaborative, proactive and positive attitude;
  • Motivation for the mission to improve education from the perspective of a successful scale-up environment.

Other frameworks, languages and tools used by our teams include

  • R programming language would be a plus.

What we offer

  • Personal training budget.
  • Plenty of room for own input.
  • Working with experienced, friendly and motivated colleagues.
  • Flexible working hours.
  • Office right next to Utrecht central station.
  • An arcade cabinet and foosball table.