We are looking for self-motivated, creative thinkers, people that are flexible and enjoy working in teams. Our data engineering/ETL team is responsible for the development of daily ETL processes in which large amounts of behavioral data from consumer panels are imported in our data lake. The data is used by our software to provide insights into consumer behavior. The ETL processes need to evolve in order to deal with the increase in the size and complexity of the data and to cope with higher requirements with respect to data quality and throughput time. Our data engineering team is pragmatic and keen to apply the best tools for the job. We have wide experience with distributed systems such as Hadoop and Hive, in addition to in-memory distributed computation platforms like Spark. And we develop everything on Linux locally, manage the source code in Git, and run our workflows on AWS in the cloud.
The team has an open culture, works in an agile style and in close cooperation with software developers and colleagues from other disciplines, such as data scientists and client-facing solution managers. You will have the opportunity to develop yourself in areas like big data, cloud computing, data lake architecture and data orchestration.