forked from llm-workflow-engine/llm-workflow-engine
-
Notifications
You must be signed in to change notification settings - Fork 0
/
config.sample.yaml
125 lines (109 loc) · 4.47 KB
/
config.sample.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
# DO NOT SIMPLY COPY AND PASTE THIS CONFIGURATION FILE!
# It is a reference for the default configuration. The correct way to
# do it is to ONLY include the values in your configuration that you
# want to override.
# Configuration syntax is YAML, learn more at: https://yaml.org
##########################################################
# Backend settings.
##########################################################
# The backend the shell should use, one of: browser, api
backend: api
# The database connection string, in a format SQLAlchemy understands.
# DO NOT USE THE LINE AS IT IS WRITTEN BELOW, IT ONLY ILLUSTRATES THE DEFAULT LOCATION.
# database: sqlite:////home/[username]/.local/share/chatgpt-wrapper/profiles/default/storage.db
##########################################################
# Shell settings.
##########################################################
shell:
# NOTE: This setting is only available on the following backends: api
# Customize the prompt prefix, the following token are available:
# - $SYSTEM_MESSAGE_ALIAS: The current your system message alias
# - $USER: Logged in username
# - $MODEL: The LLM model name
# - $PRESET_OR_MODEL: The active preset if set, or the LLM model name
# - $NEWLINE: Insert a newline
# - $TEMPERATURE: The current temperature
# - $MAX_SUBMISSION_TOKENS: The maximum number of tokens in a submission
# - $CURRENT_CONVERSATION_TOKENS: The token count of the current conversation (only supported for chat models)
prompt_prefix: '($TEMPERATURE/$MAX_SUBMISSION_TOKENS/$CURRENT_CONVERSATION_TOKENS): $SYSTEM_MESSAGE_ALIAS$NEWLINE$USER@$PRESET_OR_MODEL'
# Full path to the file that stores a history of commands in the shell.
# Set to false to disable storing history.
history_file: /tmp/repl_history.log
##########################################################
# Browser settings.
##########################################################
# NOTE: These settings only work on the following backends: browser
browser:
# Enable debug mode (shows browser window if true, headless if false)
debug: false
# The browser to use, one of: firefox, chromium, webkit
provider: firefox
##########################################################
# Model settings.
##########################################################
model:
# The preset to use on startup.
# See '/help presets' for more information.
default_preset: null
# If true, responses will be streamed in real time (no markdown formatting).
# By default, streaming is enabled in the REPL for supported providers, and
# disabled otherwise.
streaming: true
# Configure aliases to switch between system messages, key is alias, value is message.
# NOTE: This setting only works on the following backend: api
# NOTE: This setting is only active for chat models.
# See '/help system_message' for more information.
system_message:
programmer: "You are an expert programmer, who responds to questions with brief examples in code."
##########################################################
# Chat settings.
##########################################################
chat:
# Chat logs.
log:
# Enable logging to a file.
enabled: false
# Full path to the log file.
filepath: chatgpt.log
##########################################################
# Log settings.
##########################################################
log:
# Console logging.
console:
# Message format.
format: '%(name)s - %(levelname)s - %(message)s'
# Message level (must be valid Python logging module level).
level: ERROR
##########################################################
# Plugin settings.
# See the README for more informaiton on plugins.
##########################################################
plugins:
enabled:
# Core plugins, see README for more details.
- echo
#- awesome
#- database
#- data_query
#- shell
#- zap
# Core provider plugins, see README for more details.
#- provider_ai21
#- provider_cohere
#- provider_huggingface_hub
#- provider_openai
##########################################################
# Debug settings.
##########################################################
debug:
# Debug logging.
log:
# Enable to log debug messages.
enabled: false
# Full path to the log file.
filepath: /tmp/chatgpt-debug.log
# Message format.
format: '%(name)s - %(asctime)s - %(levelname)s - %(message)s'
# Message level (must be valid Python logging module level).
level: DEBUG