ModelOptions
Configuration options for model text generation behavior.
These parameters control aspects such as repetition penalties, randomness, prediction limits, and stopping criteria for language model output generation.
Constructors
Properties
Minimum probability ratio for a token to be considered. Filters out tokens whose probability is too small relative to the most likely token. Default: 0.0
The size of the context window used to generate the next token. Default: 4096
Maximum number of tokens to generate. -1 means unlimited (until a stop condition is met). Default: -1
How far back the model looks to prevent repetition. Default: 64. Set to 0 to disable, or -1 to use the full context window.
Strength of repetition penalty. Higher values (e.g., 1.5) penalize repeated tokens more strongly. Default: 1.1
Stop sequence(s) that end generation when encountered. If null, no stop condition is applied. Example: "AI assistant:"
Sampling temperature for generation. Higher values increase randomness; lower values make output more deterministic. Default: 0.8
Limits token sampling to the top K most likely tokens. Higher values increase output diversity. Default: 40
Top-p (nucleus) sampling parameter. Tokens are sampled from the smallest set whose cumulative probability exceeds top_p. Default: 0.9