We are seeking an experienced Data Engineer / Data Architect to design, build and maintain a modern, fully open-source Data Lakehouse in a cloud environment.
This is a hands-on role focused on scalable architecture, high-performance data pipelines, and production-grade engineering standards.
Key responsibilities
* Develop and maintain an open-source Data Lakehouse platform
* Build scalable end-to-end ELT pipelines
* Integrate data from databases, APIs, cloud platforms and streaming services
* Optimise queries and workflows for performance and efficiency
* Implement data quality monitoring, validation and testing
* Write clean, modular, production-ready code
* Support deployment, configuration and documentation
Required Skills
* Strong experience as a Data Engineer or Data Architect
* Deep knowledge of data warehouse / lakehouse architecture
* Excellent SQL
* Strong experience with dbt & spark
* Good knowledge of Python
* Experience with Airflow
* Experience with Kafka or RabbitMQ
* Understanding of Apache Iceberg or Delta Lake
* Experience with Kubernetes and Docker
* Strong grasp of ELT frameworks and relational databases