Codeunit "AOAI Chat Completion Params"

ID 7761
Namespace: System.AI

Represents the Chat Completion parameters used by the API. See more details at https://aka.ms/AAlrz36.

Properties

Name Value
Access Public
InherentEntitlements X
InherentPermissions X

Methods

GetTemperature

What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.

procedure GetTemperature(): Decimal

Returns

Type Description
Decimal

The sampling temperature being used.

GetMaxTokens

Gets the maximum number of tokens allowed for the generated answer.

procedure GetMaxTokens(): Integer

Returns

Type Description
Integer

The maximum number of tokens allowed for the generated answer.

Remarks

0 or less uses the API default.

GetMaxHistory

Gets the maximum number of messages to send back as the message history.

procedure GetMaxHistory(): Integer

Returns

Type Description
Integer

The maximum number of messages to send.

GetPresencePenalty

Gets the presence penalty value.

procedure GetPresencePenalty(): Decimal

Returns

Type Description
Decimal

The presence penalty value.

GetFrequencyPenalty

Gets the frequency penalty value.

procedure GetFrequencyPenalty(): Decimal

Returns

Type Description
Decimal

The frequency penalty value.

SetTemperature

Sets the sampling temperature to use, between 0 and 2. A higher temperature increases the likelihood that the next most probable token will not be selected. When requesting structured data, set the temperature to 0. For human sounding speech, 0.7 is a typical value

procedure SetTemperature(NewTemperature: Decimal)

Parameters

Name Type Description
NewTemperature Decimal

The new sampling temperature to use.

SetMaxTokens

Sets the maximum number of tokens allowed for the generated answer. The maximum number of tokens allowed for the generated answer. By default, the number of tokens the model can return will be (4096 - prompt tokens).

procedure SetMaxTokens(NewMaxTokens: Integer)

Parameters

Name Type Description
NewMaxTokens Integer

The new maximum number of tokens allowed for the generated answer.

Remarks

If the prompt's tokens + max_tokens exceeds the model's context length, the generate request will return an error.

SetJsonMode

Sets if the model should return a valid JSON object as a chat completion.

procedure SetJsonMode(NewJsonMode: Boolean)

Parameters

Name Type Description
NewJsonMode Boolean

The new Json mode for the chat completion: true or false.

Remarks

Default is false.

SetMaxHistory

Sets the maximum number of messages to send back as the message history.

procedure SetMaxHistory(NewMaxHistory: Integer)

Parameters

Name Type Description
NewMaxHistory Integer

The new maximum number of messages to send.

Remarks

The default is 10 messages including the primary System Message.

SetPresencePenalty

Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.

procedure SetPresencePenalty(NewPresencePenalty: Decimal)

Parameters

Name Type Description
NewPresencePenalty Decimal

The new presence penalty value.

SetFrequencyPenalty

Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.

procedure SetFrequencyPenalty(NewFrequencyPenalty: Decimal)

Parameters

Name Type Description
NewFrequencyPenalty Decimal

The new frequency penalty value.

See also