Big Data Chart

From QPR ProcessAnalyzer Wiki
Revision as of 15:38, 23 March 2022 by Ollvihe (talk | contribs) (Created page with "Big data chart is similar to the in-memory chart, except in the Big data chart backend processing is performed in the datasource where the eventlog data is stored. Depending o...")
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

Big data chart is similar to the in-memory chart, except in the Big data chart backend processing is performed in the datasource where the eventlog data is stored. Depending on the [[QPR_ProcessAnalyzer_Project_Workspace#Models|model type], processing is done in Snowflake (for models using Snowflake datatables) or in SQL Server (for models using Local datatables).

The Big data chart can be added to dashboard by selecting the second item from the tool palette (Bid Data Chart).

Visualization settings are mainly the same in the Big data chart comparing to the in-memory chart. On the other hand data selection, measures and dimensions work differently. The main differences are:

Following functionalities supported by the in-memory are not available in the Big data chart:

  • There are different set of analyzed objects available

Notes about performance [[QPR_ProcessAnalyzer_System_Architecture#Snowflake_Powered_Calculation||Snowflake powered calculation] will allow practically unlimited scaling when the amount of data and users increases. The Big data chart is the component to use in dashboards for the Snowflake models. In addition, the Big data chart can be used for model using local datatables. Note that the processing is then performed in SQL Server which is not optimal for analytics queries. There are still special usecases when the Big data chart is the best option for model using local datatables:

  • Eventlogs are filtered heavily so that the number of remaining cases and events are low (usually maximum of some thousands). Then processing may be done in the SQL Server without using the in-memory processing (which will require less memory)
  • If the model is not currently loaded in the memory, the fastest method is to use the Big data chart comparing to the in-memory chart, when also calculating the required time to load the model in-memory.