diff --git a/dot_config/private_mods/mods.yml b/dot_config/private_mods/mods.yml new file mode 100644 index 0000000..3b53ea1 --- /dev/null +++ b/dot_config/private_mods/mods.yml @@ -0,0 +1,78 @@ +# Default model (gpt-3.5-turbo, gpt-4, ggml-gpt4all-j...). +default-model: llamacorn-1.1b +# Text to append when using the -f flag. +format-text: + markdown: "Format the response as markdown without enclosing backticks." + json: "Format the response as json without enclosing backticks." +# List of predefined system messages that can be used as roles. +roles: + "default": [] + # Example, a role called `shell`: + shell: + - you are a shell expert + - you do not explain anything + - you simply output one liners to solve the problems you're asked + - you do not provide any explanation whatsoever, ONLY the command +# Ask for the response to be formatted as markdown unless otherwise set. +format: false +# System role to use. +role: "default" +# Render output as raw text when connected to a TTY. +raw: false +# Quiet mode (hide the spinner while loading and stderr messages for success). +quiet: false +# Temperature (randomness) of results, from 0.0 to 2.0. +temp: 1.0 +# TopP, an alternative to temperature that narrows response, from 0.0 to 1.0. +topp: 1.0 +# TopK, only sample from the top K options for each subsequent token. +topk: 50 +# Turn off the client-side limit on the size of the input into the model. +no-limit: false +# Wrap formatted output at specific width (default is 80) +word-wrap: 80 +# Include the prompt from the arguments in the response. +include-prompt-args: false +# Include the prompt from the arguments and stdin, truncate stdin to specified number of lines. +include-prompt: 0 +# Maximum number of times to retry API calls. +max-retries: 5 +# Your desired level of fanciness. +fanciness: 10 +# Text to show while generating. +status-text: Generating +# Theme to use in the forms. Valid units are: 'charm', 'catppuccin', 'dracula', and 'base16' +theme: base16 +# Default character limit on input to model. +max-input-chars: 12250 +# Maximum number of tokens in response. +# max-tokens: 100 +# Aliases and endpoints for OpenAI compatible REST API. +apis: + # localai: + # # LocalAI setup instructions: https://github.com/go-skynet/LocalAI#example-use-gpt4all-j-model + # base-url: http://localhost:8080 + # models: + # ggml-gpt4all-j: + # aliases: ["local", "4all"] + # max-input-chars: 12250 + # fallback: + localai: + base-url: http://127.0.0.1:1337 + models: + "deepseek-coder-1.3b": + aliases: ["ds-coder"] + max-input-chars: 16484 + fallback: llamacorn-1.1b + "llamacorn-1.1b": + aliases: ["local", "llcorn"] + max-input-chars: 2048 + fallback: "tinyllama-1.1b" + "llava-phi-3-mini-int4.gguf": + aliases: ["llava"] + max-input-chars: 2048 + fallback: llamacorn-1.1b + "tinyllama-1.1b": + aliases: ["llama"] + max-input-chars: 2048 + fallback: