78 lines
2.7 KiB
YAML
78 lines
2.7 KiB
YAML
# Default model (gpt-3.5-turbo, gpt-4, ggml-gpt4all-j...).
|
|
default-model: llamacorn-1.1b
|
|
# Text to append when using the -f flag.
|
|
format-text:
|
|
markdown: "Format the response as markdown without enclosing backticks."
|
|
json: "Format the response as json without enclosing backticks."
|
|
# List of predefined system messages that can be used as roles.
|
|
roles:
|
|
"default": []
|
|
# Example, a role called `shell`:
|
|
shell:
|
|
- you are a shell expert
|
|
- you do not explain anything
|
|
- you simply output one liners to solve the problems you're asked
|
|
- you do not provide any explanation whatsoever, ONLY the command
|
|
# Ask for the response to be formatted as markdown unless otherwise set.
|
|
format: false
|
|
# System role to use.
|
|
role: "default"
|
|
# Render output as raw text when connected to a TTY.
|
|
raw: false
|
|
# Quiet mode (hide the spinner while loading and stderr messages for success).
|
|
quiet: false
|
|
# Temperature (randomness) of results, from 0.0 to 2.0.
|
|
temp: 1.0
|
|
# TopP, an alternative to temperature that narrows response, from 0.0 to 1.0.
|
|
topp: 1.0
|
|
# TopK, only sample from the top K options for each subsequent token.
|
|
topk: 50
|
|
# Turn off the client-side limit on the size of the input into the model.
|
|
no-limit: false
|
|
# Wrap formatted output at specific width (default is 80)
|
|
word-wrap: 80
|
|
# Include the prompt from the arguments in the response.
|
|
include-prompt-args: false
|
|
# Include the prompt from the arguments and stdin, truncate stdin to specified number of lines.
|
|
include-prompt: 0
|
|
# Maximum number of times to retry API calls.
|
|
max-retries: 5
|
|
# Your desired level of fanciness.
|
|
fanciness: 10
|
|
# Text to show while generating.
|
|
status-text: Generating
|
|
# Theme to use in the forms. Valid units are: 'charm', 'catppuccin', 'dracula', and 'base16'
|
|
theme: base16
|
|
# Default character limit on input to model.
|
|
max-input-chars: 12250
|
|
# Maximum number of tokens in response.
|
|
# max-tokens: 100
|
|
# Aliases and endpoints for OpenAI compatible REST API.
|
|
apis:
|
|
# localai:
|
|
# # LocalAI setup instructions: https://github.com/go-skynet/LocalAI#example-use-gpt4all-j-model
|
|
# base-url: http://localhost:8080
|
|
# models:
|
|
# ggml-gpt4all-j:
|
|
# aliases: ["local", "4all"]
|
|
# max-input-chars: 12250
|
|
# fallback:
|
|
localai:
|
|
base-url: http://127.0.0.1:1337
|
|
models:
|
|
"deepseek-coder-1.3b":
|
|
aliases: ["ds-coder"]
|
|
max-input-chars: 16484
|
|
fallback: llamacorn-1.1b
|
|
"llamacorn-1.1b":
|
|
aliases: ["local", "llcorn"]
|
|
max-input-chars: 2048
|
|
fallback: "tinyllama-1.1b"
|
|
"llava-phi-3-mini-int4.gguf":
|
|
aliases: ["llava"]
|
|
max-input-chars: 2048
|
|
fallback: llamacorn-1.1b
|
|
"tinyllama-1.1b":
|
|
aliases: ["llama"]
|
|
max-input-chars: 2048
|
|
fallback:
|