We are looking for a Data Warehouse & Reporting Developer for a key client based in Belgium. Main tasks Development and maintenance of a fully open-source Data Lakehouse. Design and development of data pipelines for scalable and reliable data workflows. Data integration from databases, APIs, data streaming services and cloud data platforms. Writing modular, testable and production-grade code. Document processes comprehensively to ensure seamless data pipeline management and troubleshooting. Assistance with deployment and configuration of the system and elaboration of test programs. Specific Requirements Extensive hands-on experience as Data Engineer or Data Architect in modern cloud-based open-source data platform solutions and on data analytics tools. Excellent knowledge of data warehouse and/or data lakehouse design & architecture. Excellent knowledge of open-source, code-based data transformation tools such as dbt, Spark & Trino. Excellent knowledge of SQL and Python. Good knowledge of open-source orchestration tools such as Airflow, Dagster or Luigi. Experience with AI-powered assistants like Amazon Q. Good knowledge of event streaming platforms and message brokers like Kafka and RabbitMQ. Extensive experience in creating end-to-end data pipelines and the ELT framework. Nice to have: Understanding of the principles behind storage protocols like Apache Iceberg or Delta Lake. Proficiency with Kubernetes and Docker/Podman. Good knowledge of online analytical data processing ( OLAP ) and data mining and modelling tools. Language English C1 Location Brussels, Belgium (On-site with teleworking)