We are seeking an experienced Data Engineer / Data Architect to design, build and maintain a modern, fully open-source Data Lakehouse in a cloud environment. This is a hands-on role focused on scalable architecture, high-performance data pipelines, and production-grade engineering standards. Key responsibilities Develop and maintain an open-source Data Lakehouse platform Build scalable end-to-end ELT pipelines Integrate data from databases, APIs, cloud platforms and streaming services Optimise queries and workflows for performance and efficiency Implement data quality monitoring, validation and testing Write clean, modular, production-ready code Support deployment, configuration and documentation Required Skills Strong experience as a Data Engineer or Data Architect Deep knowledge of data warehouse / lakehouse architecture Excellent SQL Strong experience with dbt & spark Good knowledge of Python Experience with Airflow Experience with Kafka or RabbitMQ Understanding of Apache Iceberg or Delta Lake Experience with Kubernetes and Docker Strong grasp of ELT frameworks and relational databases