dotfiles/dot_config/private_mods/mods.yml

78 lines
2.7 KiB
YAML

# Default model (gpt-3.5-turbo, gpt-4, ggml-gpt4all-j...).
default-model: llamacorn-1.1b
# Text to append when using the -f flag.
format-text:
markdown: "Format the response as markdown without enclosing backticks."
json: "Format the response as json without enclosing backticks."
# List of predefined system messages that can be used as roles.
roles:
"default": []
# Example, a role called `shell`:
shell:
- you are a shell expert
- you do not explain anything
- you simply output one liners to solve the problems you're asked
- you do not provide any explanation whatsoever, ONLY the command
# Ask for the response to be formatted as markdown unless otherwise set.
format: false
# System role to use.
role: "default"
# Render output as raw text when connected to a TTY.
raw: false
# Quiet mode (hide the spinner while loading and stderr messages for success).
quiet: false
# Temperature (randomness) of results, from 0.0 to 2.0.
temp: 1.0
# TopP, an alternative to temperature that narrows response, from 0.0 to 1.0.
topp: 1.0
# TopK, only sample from the top K options for each subsequent token.
topk: 50
# Turn off the client-side limit on the size of the input into the model.
no-limit: false
# Wrap formatted output at specific width (default is 80)
word-wrap: 80
# Include the prompt from the arguments in the response.
include-prompt-args: false
# Include the prompt from the arguments and stdin, truncate stdin to specified number of lines.
include-prompt: 0
# Maximum number of times to retry API calls.
max-retries: 5
# Your desired level of fanciness.
fanciness: 10
# Text to show while generating.
status-text: Generating
# Theme to use in the forms. Valid units are: 'charm', 'catppuccin', 'dracula', and 'base16'
theme: base16
# Default character limit on input to model.
max-input-chars: 12250
# Maximum number of tokens in response.
# max-tokens: 100
# Aliases and endpoints for OpenAI compatible REST API.
apis:
# localai:
# # LocalAI setup instructions: https://github.com/go-skynet/LocalAI#example-use-gpt4all-j-model
# base-url: http://localhost:8080
# models:
# ggml-gpt4all-j:
# aliases: ["local", "4all"]
# max-input-chars: 12250
# fallback:
localai:
base-url: http://127.0.0.1:1337
models:
"deepseek-coder-1.3b":
aliases: ["ds-coder"]
max-input-chars: 16484
fallback: llamacorn-1.1b
"llamacorn-1.1b":
aliases: ["local", "llcorn"]
max-input-chars: 2048
fallback: "tinyllama-1.1b"
"llava-phi-3-mini-int4.gguf":
aliases: ["llava"]
max-input-chars: 2048
fallback: llamacorn-1.1b
"tinyllama-1.1b":
aliases: ["llama"]
max-input-chars: 2048
fallback: