Codeunit "AOAI Chat Completion Params"
Represents the Chat Completion parameters used by the API. See more details at https://aka.ms/AAlrz36.
Properties
Name | Value |
---|---|
Access | Public |
InherentEntitlements | X |
InherentPermissions | X |
Methods
GetTemperature
What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.
procedure GetTemperature(): Decimal
Returns
Type | Description |
---|---|
Decimal |
The sampling temperature being used. |
GetMaxTokens
Gets the maximum number of tokens allowed for the generated answer.
procedure GetMaxTokens(): Integer
Returns
Type | Description |
---|---|
Integer |
The maximum number of tokens allowed for the generated answer. |
Remarks
0 or less uses the API default.
GetMaxHistory
Gets the maximum number of messages to send back as the message history.
procedure GetMaxHistory(): Integer
Returns
Type | Description |
---|---|
Integer |
The maximum number of messages to send. |
GetPresencePenalty
Gets the presence penalty value.
procedure GetPresencePenalty(): Decimal
Returns
Type | Description |
---|---|
Decimal |
The presence penalty value. |
GetFrequencyPenalty
Gets the frequency penalty value.
procedure GetFrequencyPenalty(): Decimal
Returns
Type | Description |
---|---|
Decimal |
The frequency penalty value. |
SetTemperature
Sets the sampling temperature to use, between 0 and 2. A higher temperature increases the likelihood that the next most probable token will not be selected. When requesting structured data, set the temperature to 0. For human sounding speech, 0.7 is a typical value
procedure SetTemperature(NewTemperature: Decimal)
Parameters
Name | Type | Description |
---|---|---|
NewTemperature | Decimal |
The new sampling temperature to use. |
SetMaxTokens
Sets the maximum number of tokens allowed for the generated answer. The maximum number of tokens allowed for the generated answer. By default, the number of tokens the model can return will be (4096 - prompt tokens).
procedure SetMaxTokens(NewMaxTokens: Integer)
Parameters
Name | Type | Description |
---|---|---|
NewMaxTokens | Integer |
The new maximum number of tokens allowed for the generated answer. |
Remarks
If the prompt's tokens + max_tokens exceeds the model's context length, the generate request will return an error.
SetJsonMode
Sets if the model should return a valid JSON object as a chat completion.
procedure SetJsonMode(NewJsonMode: Boolean)
Parameters
Name | Type | Description |
---|---|---|
NewJsonMode | Boolean |
The new Json mode for the chat completion: true or false. |
Remarks
Default is false.
SetMaxHistory
Sets the maximum number of messages to send back as the message history.
procedure SetMaxHistory(NewMaxHistory: Integer)
Parameters
Name | Type | Description |
---|---|---|
NewMaxHistory | Integer |
The new maximum number of messages to send. |
Remarks
The default is 10 messages including the primary System Message.
SetPresencePenalty
Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.
procedure SetPresencePenalty(NewPresencePenalty: Decimal)
Parameters
Name | Type | Description |
---|---|---|
NewPresencePenalty | Decimal |
The new presence penalty value. |
SetFrequencyPenalty
Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.
procedure SetFrequencyPenalty(NewFrequencyPenalty: Decimal)
Parameters
Name | Type | Description |
---|---|---|
NewFrequencyPenalty | Decimal |
The new frequency penalty value. |