...
Interactive Cluster | Job Cluster |
---|---|
All the Spark jobs are submitted to a common cluster. The Virtual machines are shared across the jobs.
| Each Spark job is submitted to a new/dedicated cluster. The Virtual machines are not shared across the jobs
|
...
Semantic Model Scenario | Details | Recommended Cluster | |||
---|---|---|---|---|---|
Huge fact data processing (>300M row count) is required during the Full or Incremental semantic model processes, and the semantic model does not have other fact datasets with low data volume. | Level jobs processing time is three hours or more | Job | |||
The indexing time of the semantic model process job is less. | Indexing jobs time is 20% or less in the semantic model process time | Job | The semantic model has a large number of fact transformations. | The utilization of Virtual machines/Cores could exceed the instance pool maximum nodes or subscription quota. If there are ten fact transformations and the cluster is configured to use 20 worker nodes at max, the System might end up using 200 Virtual Machines (at max) simultaneously and result in Usage + quota limit errors. | Interactive |
Process jobs on low data volume. | Less usage of Azure Databricks | Interactive | |||
Wide semantic model having a huge number of dimensions or attributes. | Too many process jobs are underutilizing the Virtual machines | Interactive |
...