To use the Kyvos Copilot feature, you need to configure GenAI connection details through Kyvos Manager. The default configurations are auto-populated, and you can modify these settings as needed.
To configure the GenAI connection, perform the following steps:
On the navigation pane, click Kyvos and Ecosystem > GenAI Configuration. The page displays information about the GenAI Configuration connection details.
Enter details as:
Parameter/Field | Description |
---|---|
Enable | Select this checkbox to enable GenAI connection. This will enable the Kyvos Copilot fetaure. |
Connection Name | A unique name that identifies your GenAI connections. |
Provider | The name of the GenAI provider the system will use to generate output. |
URL | The URL of the provider-specific endpoint for generating output. |
Authentication Key | A unique key for authenticating and authorizing requests to the provider's endpoint. |
Usage | Select one of the following:
You can specify whether you want to use the feature for MDX calculations or MDX queries. If you want to create expressions for calculated measures or members, select MDX calculations. If you want to query on the semantic model, choose the MDX queries option. However:
|
Model | The name of the GenAI LLM model used to generate the output. |
Temperature | Configures temperature, controlling randomness. Lowering it results in less random completions. As the temperature approaches zero, the model becomes deterministic and repetitive. It is recommended to alter either this or top p, but not both.
|
Maximum length | Specifies the maximum number of tokens shared between the prompt and output, which varies by model. One token is roughly four characters for English text.
|
Top P | Controls diversity via nucleus sampling. If set to 0.5, half of all likelihood-weighted options are considered. It is recommended to adjust either this or the temperature parameter, but not both.
|
Frequency penalty | Specifies a number between -2.0 and 2.0. Positive values penalize new tokens based on their frequency in the existing text. This reduces the likelihood of the model repeating the same line verbatim.
|
Presence penalty | Specifies a number between -2.0 and 2.0. Positive values penalize new tokens based on their appearance in the text so far, thereby increasing the model's likelihood of discussing new topics.
|
Click Save.