-
Notifications
You must be signed in to change notification settings - Fork 47
/
Copy pathbridgeData_template.yaml
361 lines (294 loc) · 15 KB
/
bridgeData_template.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
## Common for all worker types
# !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!!
# See also the readme's "Suggested settings" section for recommended settings. !!!
# !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!! !!!
# The Horde URL. Do not change this unless you are using a custom Horde.
horde_url: "https://aihorde.net/api/"
# The API key identifies a unique user in the Horde
# Visit https://aihorde.net/register to create one before you can join
api_key: "0000000000"
# List of usernames whose prompts you want to prioritize.
# The owner's username is always included, so you don't need to add it here if you use the key specified in `api_key` for requests.
priority_usernames: []
# The maximum number of parallel jobs to run at the same time.
# Only high-end cards (e.g., 3080 or better) benefit from this setting.
# If you have a 20xx or earlier, or a xx60/xx70, do not change this setting from 1.
max_threads: 1
# Suggested values:
# 24GB+ VRAM: 1 (only increase if you have >32gb system RAM and are *not* offering Flux/Cascade)
# 12GB-16GB VRAM: 1
# 8GB-10GB VRAM: 1
# Number of extra requests to keep in the queue to start working as soon as a thread is available.
# Generally should be 1 or 2. Never set this higher than 2 if your max_threads is 2.
# Warning: Increasing this value directly increases system RAM usage significantly. Close unused programs if you increase this value.
# It does not significantly increase VRAM usage, but ~150mb of fixed VRAM per queue slot is reserved at all times.
queue_size: 1
# Suggested values:
# >64gb system RAM: 3
# 64gb system RAM: 2
# 32gb system RAM: 0 or 1
# 32gb system RAM and offering Flux/Cascade: 0
# <32gb system RAM: 0
# Number of jobs to pull per request and perform batched inference.
# More optimized than doing them one by one but slower.
# Ensure you can generate your max_batch at half your max_power.
max_batch: 1
# Suggested values:
# 24GB+ VRAM: 8 or higher
# 12GB-16GB VRAM: 4 or higher
# 8GB-10GB VRAM: 4 (no higher than 4)
# Run CLIP model (checking for potential CSAM or NSFW) on GPU instead of CPU.
# Enable this on cards with 12GB or more VRAM to increase job completion rate.
# ~1.2GB of VRAM overhead
safety_on_gpu: false
# Suggested values:
# 24GB+ VRAM: true
# 12GB-16GB VRAM: true (consider false if offering Cascade or Flux)
# 8GB-10GB VRAM: false
# Only pick up jobs where the user has the required kudos upfront.
# Excludes all anonymous accounts and registered accounts who haven't contributed.
require_upfront_kudos: false
# If true, the worker will clear maintenance mode on startup.
# Warning: While this can be convenient, being put into maintenance mode is a safety feature to prevent workers
# which are not working correctly from taking jobs. If you see this happening frequently, you should investigate
# your logs to determine the cause.
remove_maintenance_on_init: false
# If true, fewer log messages will be printed to the console.
# This is intended for workers running on the cloud or other headless environments.
# We *do not* recommend this setting for most users.
limited_console_messages: false
#######################################
## Dreamer (Stable Diffusion Worker) ##
#######################################
# Worker name for running a Dreamer instance.
# This must be unique across the entire horde! Do not use the default!
# If you get the error `Wrong credentials to submit as this worker.`, try changing this name, someone else is already using it.
dreamer_name: "An Awesome Dreamer"
# Max resolution (max pixels) supported.
# Formula: `64 * 64 * 8 * max_power` (total pixels)
# Examples:
# 8 = 512x512
# 18 = 768x768
# 32 = 1024x1024
# 50 = 1280x1280
max_power: 8
# Suggested values:
# 8GB-10GB VRAM: 32 (no higher than 32)
# 12GB-16GB VRAM: 32-64 (no higher than 64)
# 24GB+ VRAM: 64-128 (no higher than 128)
# Use more VRAM on average but reduce time spent loading models.
high_memory_mode: false
# Suggested values:
# 24GB+ VRAM: true
# 12GB-16GB VRAM: true (consider false if offering Cascade or Flux)
# Fill local queue much faster but may be penalized by the server if you cannot keep up with jobs.
high_performance_mode: false
# Suggested values:
# 24GB+ VRAM: true
# Fill local queue somewhat faster but may be penalized by the server if you cannot keep up with jobs.
# Overridden by high_performance_mode.
moderate_performance_mode: false
# Suggested values:
# 12GB-16GB VRAM: true
# Start processing the next job before the current job finishes post-processing.
# Reduces time between jobs but may cause crashes on low RAM or VRAM systems.
post_process_job_overlap: false
# Suggested values:
# 24GB+ VRAM: true
# Aggressively unload models from VRAM when not in use.
# Should be true for most workers with GPUs with less than 16GB of VRAM.
unload_models_from_vram_often: true
# Suggested values:
# 24GB+ VRAM: false
# 12GB-16GB VRAM: false
# 8GB-10GB VRAM: true
# Normally only one model will load off disk at a time. Set to true to load multiple models at once.
# This requires a very fast disk. You will see a sharp increase in disk usage, especially with SDXL/Cascade/Flux/other large models.
very_fast_disk_mode: false
# List of words to reject if they appear in the prompt.
blacklist: []
# Serve NSFW images if true.
nsfw: true
# Censor NSFW images if true.
censor_nsfw: false
# List of words to always censor, even if `nsfw` is true.
censorlist: []
# Accept jobs using a user-supplied image.
allow_img2img: true
# Accept jobs using a user-supplied image and an inpainting-specific model.
# Forced to false if `allow_img2img` is false.
allow_painting: true
# Allow user requests from behind VPNs.
# Note: The worker does not directly interact with user IPs - it only interacts with the StableHorde API.
allow_unsafe_ip: true
# Allow upscaling, facefixer, and other post-generation features.
allow_post_processing: true
# 8GB-10GB VRAM: false (if offering SDXL or Flux, otherwise true)
# Allow ControlNet jobs.
# Note: Additional RAM/VRAM overhead. Low VRAM cards (<6GB) should be cautious.
allow_controlnet: false
# Allow SDXL jobs with high memory add-ons like ControlNet or transparency.
# Note: Significant additional RAM/VRAM overhead. Medium VRAM cards (<12GB) should be cautious.
# Note that if this is true, allow_controlnet must also be true.
allow_sdxl_controlnet: false
# 16GB+ VRAM: true
# 8GB-10GB VRAM: false
# Allow LoRas to be used. Requires a fast internet connection.
# LoRas will be downloaded on demand. `max_lora_cache_size` controls how many gigabytes to keep downloaded.
# 5GB of preselected LoRas are always downloaded the first time you start the worker with this setting.
# Note that there can be a significant delay when downloading LoRas causing GPU downtime.
# If you set this, we strongly recommend setting `civitai_api_token` as well.
allow_lora: false
# If set, this worker will use this civitai API token when downloading any resources from civitai.
# This is required in order to provide LoRas/TIs (or other resources)
# which are marked as requiring a civitai token to download.
#
# If you set "models_to_load" to something like "top 5", it is necessary to provide this token,
# as many of the most popular models require it for the download. Downloading many models in bulk
# could also lead to issues if no token is required.
#
# You can get your civitai API Key from https://civitai.com/user/account (look for 'Add API Key')
#
# Remove the # from the line below and add your civitai API token to enable this feature.
# civitai_api_token:
# Delete any unknown LoRas from the loras folder when `download_models.py` is run.
# Warning: This option will delete any LoRas not in the model reference, including custom LoRas.
purge_loras_on_download: false
# Number of gigabytes of LoRas to keep cached. Minimum is 10GB.
max_lora_cache_size: 10
# Set to true if your worker is extraordinarily slow (below 0.1 mps/s).
# Users can choose to skip it when requesting generations, but job timeout and request expiry timeout are tripled.
extra_slow_worker: false
# Low-end cards or low performance: true
# Only pick up jobs requesting steps lower than the model's average steps.
# Useful for slower workers or if you don't want to serve requests with an extraordinary number of steps.
limit_max_steps: false
# Low-end cards or low performance: true
# Automatically determine the models with the highest queue and offer those.
dynamic_models: false # Currently unused in reGen
# Number of models to offer when `dynamic_models` is true.
number_of_dynamic_models: 0 # Currently unused in reGen
# Maximum number of models to download automatically for `dynamic_models`.
max_models_to_download: 10 # Currently unused in reGen
# Frequency (in seconds) to output worker summary stats, such as kudos per hour.
# Set to zero to disable stats output completely.
stats_output_frequency: 20
# Location where models are stored.
cache_home: "./models/"
# Location of the temp directory, also used for the model cache.
temp_dir: "./tmp" # Currently unused in reGen
# Always download models when required without prompting.
always_download: true # Currently unused in reGen
# Disable the terminal GUI, which displays information about the worker and the Horde.
disable_terminal_ui: false # Currently unused in reGen
# Obsolete
vram_to_leave_free: "80%" # Currently unused in reGen
# Target amount of system RAM to keep free.
# The worker only makes a best effort. Avoid using too much RAM with other programs.
ram_to_leave_free: "80%" # Currently unused in reGen
# Obsolete
disable_disk_cache: false # Currently unused in reGen
# Models to use.
# Instead of a model name, you may use any of the following magic constants:
# "ALL" - Load all possible models (over 1TB of space).
# "TOP n" - Load the top "N" most popular models (e.g., "top 5").
# "BOTTOM n" - Load the bottom "N" models (e.g., "bottom 5").
# "ALL SD15 MODELS" - All Stable Diffusion 1.5 models.
# "ALL SD21 MODELS" - All Stable Diffusion 2.0/2.1 models.
# "ALL SDXL MODELS" - All Stable Diffusion XL models.
# "ALL INPAINTING MODELS" - All models marked for inpainting.
# "ALL SFW MODELS" - All models marked as SFW.
# "ALL NSFW MODELS" - All models marked as NSFW.
#
# The official model reference (in JSON format) is at https://github.com/Haidra-Org/AI-Horde-image-model-reference/blob/main/stable_diffusion.json.
# The model name must match the name in the model reference or be a magic constant.
# If you use `TOP` or `ALL` load commands, you should set a `civitai_api_token`. See that configuration entry for more information.
#
# SD15 models are generally 1-2GB each, SD21 models are generally 5-37B each, and SDXL models are generally 7B each.
# Flux.1-Schnell fp8 (Compact) is 17GB, and Cascade is 14gb total.
#
#
# Important note: models_to_skip are only removed and other models are *not* added back in.
# If you set `TOP 10`, and add `Flux.1-Schnell fp8 (Compact)` to `models_to_skip`,
# you will only have 9 models loaded.
#
models_to_load:
- "top 2"
#- "ALL MODELS"
#- "TOP 3"
#- "ALL SFW"
#- "Flux.1-Schnell fp8 (Compact)"
#- "stable_diffusion"
#- "Anything Diffusion"
#- "stable_diffusion_inpainting" # requires `allow_painting` to be true
# <32gb RAM:
# Only SD15 models (set `load_large_models: false` and add SDXL, SD21, and all of the "Un-pruned" models to your `models_to_skip` below if using TOP/ALL commands)
# Be sure to set your queue_size to 0
# 32gb+ RAM, 8GB VRAM:
# SD15, SD21
# If running SDXL
# If running Flux, be sure to free up as much VRAM as possible before starting the worker.
# 32gb+ RAM, 10GB+ VRAM:
# All model types can be run.
# If you use a meta command, such as ALL or TOP n, you can allow very large models, such as cascade or flux to be included.
# By default, these models are excluded due to their large size.
# Set to true if have a 24GB card and want to include these models.
# Otherwise, I suggest including the models you know you can handle manually.
load_large_models: false
# Models to skip when `dynamic_models` is true or TOP n models are selected in models_to_load.
# Avoid loading models due to VRAM constraints, NSFW content, or other reasons.
# Models skipped are not replaced by other models, meaning TOP n will load fewer models. See the note above models_to_load.
models_to_skip:
- "pix2pix" # Not currently supported
- "SDXL_beta::stability.ai#6901" # Do not remove this, as this model would never work
- "A to Zovya RPG" # This model is known to cause problems with reGen
# - "Stable Cascade 1.0" # This a *very* VRAM intensive model
# - ALL NSFW MODELS
#- "stable_diffusion_inpainting" # Inpainting is generally quite heavy along with other models for smaller GPUs.
#- "ALL SD21" # Stable diffusion 2.1 has bigger memory requirements than 1.5, so if your card cannot lift, it, disable it
# - HASDX # Un-pruned SD1.5 model (5gb+)
# - Anygen # Un-pruned SD1.5 model (5gb+)
# - PFG # Un-pruned SD1.5 model (5gb+)
# - Poison # Un-pruned SD1.5 model (5gb+)
# - MoistMix # Un-pruned SD1.5 model (5gb+)
# - Laolei New Berry Protogen Mix # Un-pruned SD1.5 model (5gb+)
# Suppress speed warnings if jobs are taking too long.
# Note: If you are getting these messages, you are serving jobs much slower than ideal.
# Lower your max_power for more kudos/hr.
suppress_speed_warnings: false
# Exit if an unhandled fault occurs. Useful for setting up the worker as a system service.
exit_on_unhandled_faults: false
#########################
## Scribe (LLM Worker) ##
#########################
# Note: Scribe is not currently supported in reGen. This section is for future use.
# Use https://github.com/Haidra-Org/AI-Horde-Worker to run a Scribe worker.
# Worker name for running a Scribe worker.
# This must be unique across the entire horde! Do not use the default!
# If you get the error `Wrong credentials to submit as this worker.`, try changing this name, someone else is already using it.
scribe_name: "An Awesome Scribe"
# KoboldAI Client API URL.
kai_url: "http://localhost:5000"
# Max tokens to generate with this worker.
max_length: 80
# Max tokens to use from the prompt.
max_context_length: 1024
# Append the Horde alias behind the API key to the model advertised to the Horde.
# Prevents the model from being used from the shared pool but ensures no other worker can pretend to serve it.
branded_model: true
#########################
## Alchemist (Image Interrogation and Post-Processing)
#########################
# Note: Scribe is not currently supported in reGen. This section is for future use.
# Use https://github.com/Haidra-Org/AI-Horde-Worker to run an Alchemist worker.
# Worker name for running an Alchemist worker.
# This must be unique across the entire horde! Do not use the default!
# If you get the error `Wrong credentials to submit as this worker.`, try changing this name, someone else is already using it.
alchemist_name: "An Awesome Alchemist"
# Alchemy forms this worker can serve.
forms:
- "caption"
- "nsfw" # Uses CPU
# Heavier than the others, but rewards more kudos
- "interrogation"
- "post-process"