We are seeking an experienced Data Engineer / Data Architect to design, build and maintain a modern, fully open-source Data Lakehouse in a cloud environment.
This is a hands-on role focused on scalable architecture, high-performance data pipelines, and production-grade engineering standards.
Key responsibilities
- Develop and maintain an open-source Data Lakehouse platform
- Build scalable end-to-end ELT pipelines
- Integrate data from databases, APIs, cloud platforms and streaming services
- Optimise queries and workflows for performance and efficiency
- Implement data quality monitoring, validation and testing
- Write clean, modular, production-ready code
- Support deployment, configuration and documentation
Required Skills
- Strong experience as a Data Engineer or Data Architect
- Deep knowledge of data warehouse / lakehouse architecture
- Excellent SQL
- Strong experience with dbt & spark
- Good knowledge of Python
- Experience with Airflow
- Experience with Kafka or RabbitMQ
- Understanding of Apache Iceberg or Delta Lake
- Experience with Kubernetes and Docker
- Strong grasp of ELT frameworks and relational databases
Solliciteren