Onboarding data to Qlik Open Lakehouse
Qlik Open Lakehouse facilitates data movement from any supported source to Apache Iceberg tables. The Iceberg open table format offers interoperability between apps and data platforms, enabling data to be read from a variety of tools, reducing the need to duplicate data. You can mirror tables in your Qlik Open Lakehouse projects to Snowflake, and users can query data directly in Snowflake.
To onboard data to Qlik Open Lakehouse and store data in Iceberg format tables, you start by establishing a network integration to your AWS VPC. This integration enables Qlik to securely provision and manage resources within your cloud environment. Next, you create a lakehouse cluster to provide a compute environment for processing and storing your data. The final step of the preparation is to set up a data catalog connection to a centralized metadata repository.
You can create a data pipelines using Qlik Open Lakehouse projects when these components are configured. Data pipelines ingest your data and store it in Iceberg tables. You can access and query your Iceberg tables from a wide range of tools, including directly from Snowflake.
Qlik Open Lakehouse architecture
With Qlik Open Lakehouse you can create a scalable, high-performing lakehouse using the Iceberg open table format.
Setting up Qlik Open Lakehouse
Configure your cloud environment with a network integration and lakehouse cluster to run Qlik Open Lakehouse projects.
Create a Qlik Open Lakehouse project
Create a project to onboard data to Iceberg open table format storage, and optionally mirror tables to Snowflake without data duplication.