Interface: OllamaTextGenerationSettings
Hierarchy
-
↳
OllamaTextGenerationSettings
Properties
format
• Optional
format: "json"
The format to return a response in. Currently the only accepted value is 'json'. Leave undefined to return a string.
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:101
maxGenerationTokens
• Optional
maxGenerationTokens: number
Specifies the maximum number of tokens (words, punctuation, parts of words) that the model can generate in a single response. It helps to control the length of the output.
Does nothing if the model does not support this setting.
Example: maxGenerationTokens: 1000
Inherited from
TextGenerationModelSettings.maxGenerationTokens
Defined in
packages/modelfusion/src/model-function/generate-text/TextGenerationModel.ts:28
mirostat
• Optional
mirostat: number
Enable Mirostat sampling for controlling perplexity. (default: 0, 0 = disabled, 1 = Mirostat, 2 = Mirostat 2.0)
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:22
mirostatEta
• Optional
mirostatEta: number
Influences how quickly the algorithm responds to feedback from the generated text. A lower learning rate will result in slower adjustments, while a higher learning rate will make the algorithm more responsive. (Default: 0.1)
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:29
mirostatTau
• Optional
mirostatTau: number
Controls the balance between coherence and diversity of the output. A lower value will result in more focused and coherent text. (Default: 5.0)
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:35
model
• model: string
The name of the model to use. For example, 'mistral'.
See
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:10
numGpu
• Optional
numGpu: number
The number of layers to send to the GPU(s). On macOS it defaults to 1 to enable metal support, 0 to disable.
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:47
numGqa
• Optional
numGqa: number
The number of GQA groups in the transformer layer. Required for some models, for example it is 8 for llama2:70b
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:41
numThreads
• Optional
numThreads: number
Sets the number of threads to use during computation. By default, Ollama will detect this for optimal performance. It is recommended to set this value to the number of physical CPU cores your system has (as opposed to the logical number of cores).
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:54
numberOfGenerations
• Optional
numberOfGenerations: number
Number of texts to generate.
Specifies the number of responses or completions the model should generate for a given prompt. This is useful when you need multiple different outputs or ideas for a single prompt. The model will generate 'n' distinct responses, each based on the same initial prompt. In a streaming model this will result in both responses streamed back in real time.
Does nothing if the model does not support this setting.
Example: numberOfGenerations: 3
// The model will produce 3 different responses.
Inherited from
TextGenerationModelSettings.numberOfGenerations
Defined in
packages/modelfusion/src/model-function/generate-text/TextGenerationModel.ts:55
observers
• Optional
observers: FunctionObserver
[]
Observers that are called when the model is used in run functions.
Inherited from
TextGenerationModelSettings.observers
Defined in
packages/modelfusion/src/model-function/Model.ts:8
repeatLastN
• Optional
repeatLastN: number
Sets how far back for the model to look back to prevent repetition. (Default: 64, 0 = disabled, -1 = num_ctx)
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:60
repeatPenalty
• Optional
repeatPenalty: number
Sets how strongly to penalize repetitions. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. (Default: 1.1)
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:67
seed
• Optional
seed: number
Sets the random number seed to use for generation. Setting this to a specific number will make the model generate the same text for the same prompt. (Default: 0)
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:74
stopSequences
• Optional
stopSequences: string
[]
Stop sequences to use. Stop sequences are an array of strings or a single string that the model will recognize as end-of-text indicators. The model stops generating more content when it encounters any of these strings. This is particularly useful in scripted or formatted text generation, where a specific end point is required. Stop sequences not included in the generated text.
Does nothing if the model does not support this setting.
Example: stopSequences: ['\n', 'END']
Inherited from
TextGenerationModelSettings.stopSequences
Defined in
packages/modelfusion/src/model-function/generate-text/TextGenerationModel.ts:41
temperature
• Optional
temperature: number
The temperature of the model. Increasing the temperature will make the model answer more creatively. (Default: 0.8)
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:16
template
• Optional
template: string
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:103
tfsZ
• Optional
tfsZ: number
Tail free sampling is used to reduce the impact of less probable tokens from the output. A higher value (e.g., 2.0) will reduce the impact more, while a value of 1.0 disables this setting. (default: 1)
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:81
topK
• Optional
topK: number
Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative. (Default: 40)
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:88
topP
• Optional
topP: number
Works together with top-k. A higher value (e.g., 0.95) will lead to more diverse text, while a lower value (e.g., 0.5) will generate more focused and conservative text. (Default: 0.9)
Defined in
packages/modelfusion/src/model-provider/ollama/OllamaTextGenerationSettings.ts:95
trimWhitespace
• Optional
trimWhitespace: boolean
When true, the leading and trailing white space and line terminator characters are removed from the generated text.
Default: true.
Inherited from
TextGenerationModelSettings.trimWhitespace
Defined in
packages/modelfusion/src/model-function/generate-text/TextGenerationModel.ts:63