Skip to main content

Onboarding data to Qlik Open Lakehouse

Qlik Open Lakehouse facilitates data movement from any supported source to Apache Iceberg tables. The Iceberg open table format offers interoperability between apps and data platforms, enabling data to be read from a variety of tools, reducing the need to duplicate data. You can mirror tables in your Qlik Open Lakehouse projects to Snowflake, and users can query data directly in Snowflake.

To onboard data to Qlik Open Lakehouse and store data in Iceberg format tables, you start by establishing a network integration to your AWS VPC. This integration enables Qlik to securely provision and manage resources within your cloud environment. Next, you create a lakehouse cluster to provide a compute environment for processing and storing your data. The final step of the preparation is to set up a data catalog connection to a centralized metadata repository.

You can create a data pipelines using Qlik Open Lakehouse projects when these components are configured. Data pipelines ingest your data and store it in Iceberg tables. You can access and query your Iceberg tables from a wide range of tools, including directly from Snowflake.

Govnote-not-inNot supported in Qlik Cloud Government.
Govnote-not-inNot supported in Qlik Cloud Government - DoD.

Qlik Open Lakehouse architecture

With Qlik Open Lakehouse you can create a scalable, high-performing lakehouse using the Iceberg open table format.

Setting up Qlik Open Lakehouse

Configure your cloud environment with a network integration and lakehouse cluster to run Qlik Open Lakehouse projects.

Create a Qlik Open Lakehouse project

Create a project to onboard data to Iceberg open table format storage, and optionally mirror tables to Snowflake without data duplication.

Did this page help you?

If you find any issues with this page or its content – a typo, a missing step, or a technical error – please let us know!