We are looking for a Data Warehouse & Reporting Developer for a key client based in Belgium .
Main tasks
- Development and maintenance of a fully open-source Data Lakehouse.
- Design and development of data pipelines for scalable and reliable data workflows.
- Data integration from databases, APIs, data streaming services and cloud data platforms.
- Writing modular, testable and production-grade code.
- Document processes comprehensively to ensure seamless data pipeline management and troubleshooting.
- Assistance with deployment and configuration of the system and elaboration of test programs.
Specific Requirements
- Extensive hands-on experience as Data Engineer or Data Architect in modern cloud-based open-source data platform solutions and on data analytics tools.
- Excellent knowledge of data warehouse and/or data lakehouse design & architecture.
- Excellent knowledge of open-source, code-based data transformation tools such as dbt, Spark & Trino .
- Excellent knowledge of SQL and Python .
- Good knowledge of open-source orchestration tools such as Airflow, Dagster or Luigi .
- Experience with AI-powered assistants like Amazon Q.
- Good knowledge of event streaming platforms and message brokers like Kafka and RabbitMQ .
- Extensive experience in creating end-to-end data pipelines and the ELT framework .
Nice to have:
- Understanding of the principles behind storage protocols like Apache Iceberg or Delta Lake .
- Proficiency with Kubernetes and Docker/Podman .
- Good knowledge of online analytical data processing ( OLAP ) and data mining and modelling tools.
Language
Location
- Brussels, Belgium (On-site with teleworking)
Solliciteren