Big Data Chart

From QPR ProcessAnalyzer Wiki
Revision as of 15:41, 23 March 2022 by Ollvihe (talk | contribs)
Jump to navigation Jump to search

Big data chart is similar to the in-memory chart, except in the Big data chart backend processing is performed in the datasource where the eventlog data is stored. Depending on the [[QPR_ProcessAnalyzer_Project_Workspace#Models|model type], processing is done in Snowflake (for models using Snowflake datatables) or in SQL Server (for models using Local datatables).

The Big data chart can be added to dashboard by selecting the second item from the tool palette (Bid Data Chart).

Differences to in-memory chart

Visualization settings are mainly the same in the Big data chart comparing to the in-memory chart. On the other hand data selection, measures and dimensions work differently. The main differences are:

  • There are different set of analyzed objects, measures and dimensions available
  • Filtering can be done for each measure and dimension separately

Following functionalities supported by the in-memory are not available in the Big data chart: Presets, Group rows exceeding maximum, Analyzed objects sample size, Find root causes, and Business calendars. In addition, the following measure/dimension settings are not available: Round to decimals, Calculate measure for, Variable name, Custom aggregation expression, and Adjustment expression.

When to use Big Data Chart

[[QPR_ProcessAnalyzer_System_Architecture#Snowflake_Powered_Calculation||Snowflake powered calculation] will allow practically unlimited scaling when the amount of data and users increases. The Big data chart is the component to use in dashboards for the Snowflake models. In addition, the Big data chart can be used for model using local datatables. Note that the processing is then performed in SQL Server which is not optimal for analytics queries. There are still special usecases when the Big data chart is the best option for model using local datatables:

  • Eventlogs are filtered heavily so that the number of remaining cases and events are low (usually maximum of some thousands). Then processing may be done in the SQL Server without using the in-memory processing (which will require less memory)
  • If the model is not currently loaded in the memory, the fastest method is to use the Big data chart comparing to the in-memory chart, when also calculating the required time to load the model in-memory.