The index is the main element of Qlik Big Data Index. Before you can index the data, you need to prepare the data using a schema, and connect to the data with a specific connector.
Indexlets are persisted data and symbol tables representing the big data. They are created during indexing, and are used when a request comes from a client to access the big data.
Traditionally, symbol tables contain distinct field values while data tables contain bit-stuffed references to the symbol tables. This provides a compact data model, but is not suited for calculations with big data tables.
In the indexlet concept of Qlik Big Data Index, symbol tables are created per indexlet table, with the data table represented by indices from symbols to records and vice versa. This bi-directional index makes it possible to navigate from any column value to all rows with that value, as well as from any row to find all the column values for that row. This allows for distributed computing of big data aggregations.
You can use the Qlik Big Data Index management console to configure the indexing cluster and create the index. The management console is available on the 8080 port.
You can also make the management console available on port 80 of your qlik-nginx-ingress-controller node external IP by enabling nginx ingress. Example of an address from kubectl get svc:
You need to prepare the data source files in Parquet format, and place them on a shared folder that can be accessed across all nodes.
You can use the Qlik Big Data Index management console to configure the indexing cluster and create the index.
You can also create the index by executing supplied shell scripts in the cluster.
You can append data from the data sources to the index cluster using the Data Append REST API...
Manage cluster settings and indexing services...