Skip to content

Commit f9a4b61

Browse files
authored
Merge pull request #206 from Dartvauder/dev
Dev
2 parents a5b35c7 + 7bb4ae3 commit f9a4b61

File tree

7 files changed

+882
-10
lines changed

7 files changed

+882
-10
lines changed

LaunchFile/app.py

+13-4
Original file line numberDiff line numberDiff line change
@@ -330,7 +330,8 @@ def load_translation(lang):
330330

331331
translations = {
332332
"EN": {},
333-
"RU": load_translation("ru")
333+
"RU": load_translation("ru"),
334+
"ZH": load_translation("zh")
334335
}
335336

336337
def _(text, lang="EN"):
@@ -10233,8 +10234,16 @@ def reload_interface():
1023310234
wiki_interface = gr.Interface(
1023410235
fn=get_wiki_content,
1023510236
inputs=[
10236-
gr.Textbox(label=_("Online Wiki", lang), value="https://github.com/Dartvauder/NeuroSandboxWebUI/wiki/EN‐Wiki" if lang == "EN" else "https://github.com/Dartvauder/NeuroSandboxWebUI/wiki/RU‐Wiki", interactive=False),
10237-
gr.Textbox(label=_("Local Wiki", lang), value="Wikies/WikiEN.md" if lang == "EN" else "Wikies/WikiRU.md", interactive=False)
10237+
gr.Textbox(label=_("Online Wiki", lang), value=(
10238+
"https://github.com/Dartvauder/NeuroSandboxWebUI/wiki/EN‐Wiki" if lang == "EN" else
10239+
"https://github.com/Dartvauder/NeuroSandboxWebUI/wiki/ZH‐Wiki" if lang == "ZH" else
10240+
"https://github.com/Dartvauder/NeuroSandboxWebUI/wiki/RU‐Wiki"
10241+
), interactive=False),
10242+
gr.Textbox(label=_("Local Wiki", lang), value=(
10243+
"Wikies/WikiEN.md" if lang == "EN" else
10244+
"Wikies/WikiZH.md" if lang == "ZH" else
10245+
"Wikies/WikiRU.md"
10246+
), interactive=False)
1023810247
],
1023910248
outputs=gr.HTML(label=_("Wiki Content", lang)),
1024010249
title=_("NeuroSandboxWebUI - Wiki", lang),
@@ -10289,7 +10298,7 @@ def reload_interface():
1028910298
settings_interface = gr.Interface(
1029010299
fn=settings_interface,
1029110300
inputs=[
10292-
gr.Radio(choices=["EN", "RU"], label=_("Language", lang), value=settings['language']),
10301+
gr.Radio(choices=["EN", "RU", "ZH"], label=_("Language", lang), value=settings['language']),
1029310302
gr.Radio(choices=["True", "False"], label=_("Share Mode", lang), value="False"),
1029410303
gr.Radio(choices=["True", "False"], label=_("Debug Mode", lang), value="False"),
1029510304
gr.Radio(choices=["True", "False"], label=_("Monitoring Mode", lang), value="False"),

README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -106,7 +106,7 @@ To work with the virtual environment through the terminal, run `./Venv.sh`
106106

107107
## Wiki
108108

109-
* https://github.com/Dartvauder/NeuroSandboxWebUI/wiki
109+
* https://github.com/Dartvauder/NeuroSandboxWebUI/wiki/EN‐Wiki
110110

111111
## Acknowledgment to developers
112112

@@ -142,7 +142,7 @@ First of all, I want to thank the developers of [PyCharm](https://www.jetbrains.
142142
* `pixeloe` - https://github.com/KohakuBlueleaf/PixelOE
143143
* `k-diffusion` - https://github.com/crowsonkb/k-diffusion
144144
* `open-parse` - https://github.com/Filimoa/open-parse
145-
* `AudioSR` - https://github.com/haoheliu/versatile_audio_super_resolution/
145+
* `AudioSR` - https://github.com/haoheliu/versatile_audio_super_resolution
146146

147147
## Third Party Licenses:
148148

Readmes/README_RU.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -142,7 +142,7 @@
142142
* `pixeloe` - https://github.com/KohakuBlueleaf/PixelOE
143143
* `k-diffusion` - https://github.com/crowsonkb/k-diffusion
144144
* `open-parse` - https://github.com/Filimoa/open-parse
145-
* `AudioSR` - https://github.com/haoheliu/versatile_audio_super_resolution/
145+
* `AudioSR` - https://github.com/haoheliu/versatile_audio_super_resolution
146146

147147
## Лицензии третьих сторон:
148148

Readmes/README_ZH.md

+240-1
Original file line numberDiff line numberDiff line change
@@ -1 +1,240 @@
1-
# Soon...
1+
## [功能](/#功能) | [依赖](/#必需依赖) | [系统要求](/#最低系统要求) | [安装](/#如何安装) | [Wiki](/#Wiki) | [致谢](/#致开发者的感谢) | [许可证](/#第三方许可证)
2+
3+
# ![主图](https://github.com/Dartvauder/NeuroSandboxWebUI/assets/140557322/4ea0d891-8979-45ad-b052-626c41ae991a)
4+
* 仍在进行中但已稳定!
5+
* [English](/README.md) | [Русский](/Readmes/README_RU.md) | 漢語
6+
7+
## 描述:
8+
9+
一个简单方便的界面,用于使用各种神经网络模型。您可以通过文本、语音和图像输入与LLM和Moondream2进行通信;使用StableDiffusion、Kandinsky、Flux、HunyuanDiT、Lumina-T2X、Kolors、AuraFlow、Würstchen、DeepFloydIF、PixArt和PlaygroundV2.5生成图像;使用ModelScope、ZeroScope 2、CogVideoX和Latte生成视频;使用StableFast3D、Shap-E、SV34D和Zero123Plus生成3D对象;使用StableAudioOpen、AudioCraft和AudioLDM 2生成音乐和音频;使用CoquiTTS、MMS和SunoBark进行文本到语音转换;使用OpenAI-Whisper和MMS进行语音到文本转换;使用Wav2Lip进行唇形同步;使用LivePortrait为图像添加动画;使用Roop进行换脸;使用Rembg移除背景;使用CodeFormer修复面部;使用PixelOE进行图像像素化;使用DDColor为图像上色;使用LibreTranslate和SeamlessM4Tv2进行文本翻译;使用Demucs和UVR进行音频文件分离;使用RVC进行语音转换。您还可以在图库中查看输出目录中的文件,下载LLM和StableDiffusion模型,在界面内更改应用程序设置并检查系统传感器。
10+
11+
项目目标 - 创建一个尽可能简单易用的神经网络模型应用程序
12+
13+
### 文本:<img width="1119" alt="1zh" src="https://github.com/user-attachments/assets/1cbda009-8230-4dc2-beb7-a77505e96d81">
14+
15+
### 图像:<img width="1127" alt="2zh" src="https://github.com/user-attachments/assets/151001bc-27b2-4561-84f1-7ec521ad972e">
16+
17+
### 视频:<img width="1117" alt="3zh" src="https://github.com/user-attachments/assets/70ebf95e-f82e-467a-a027-64ee917527cc">
18+
19+
### 3D:<img width="1121" alt="4zh" src="https://github.com/user-attachments/assets/6c78d5a9-3794-43af-b167-fb63a4102d83">
20+
21+
### 音频:<img width="1117" alt="5zh" src="https://github.com/user-attachments/assets/f87efe4d-095f-4e99-abfb-759801ff4f29">
22+
23+
### 额外功能:<img width="1116" alt="6zh" src="https://github.com/user-attachments/assets/f92da750-9ce6-4982-80e1-324c0ee749c3">
24+
25+
### 界面:<img width="1120" alt="7zh" src="https://github.com/user-attachments/assets/f444b0bb-cc7d-46fc-8eb6-a944e9269838">
26+
27+
## 功能:
28+
29+
* 通过install.bat(Windows)或install.sh(Linux)轻松安装
30+
* 您可以通过移动设备在本地主机(通过IPv4)或在线任何地方(通过Share)使用应用程序
31+
* 灵活且优化的界面(由Gradio提供)
32+
*`Install``Update`文件进行调试日志记录
33+
* 提供三种语言版本
34+
* 支持Transformers和llama.cpp模型(LLM)
35+
* 支持diffusers和safetensors模型(StableDiffusion)- txt2img、img2img、depth2img、marigold、pix2pix、controlnet、upscale(latent)、upscale(SUPIR)、refiner、inpaint、outpaint、gligen、diffedit、blip-diffusion、animatediff、hotshot-xl、video、ldm3d、sd3、cascade、t2i-ip-adapter、ip-adapter-faceid和riffusion标签
36+
* 支持stable-diffusion-cpp模型用于FLUX
37+
* 支持额外的图像生成模型:Kandinsky(txt2img、img2img、inpaint)、Flux(支持LoRA)、HunyuanDiT(txt2img、controlnet)、Lumina-T2X、Kolors(支持LoRA的txt2img、img2img、ip-adapter-plus)、AuraFlow(支持LoRA和AuraSR)、Würstchen、DeepFloydIF(txt2img、img2img、inpaint)、PixArt和PlaygroundV2.5
38+
* 支持使用Rembg、CodeFormer、PixelOE、DDColor、DownScale、格式转换器、换脸(Roop)和放大(Real-ESRGAN)模型进行图像、视频和音频的额外处理
39+
* 支持StableAudio
40+
* 支持AudioCraft(模型:musicgen、audiogen和magnet)
41+
* 支持AudioLDM 2(模型:audio和music)
42+
* 支持TTS和Whisper模型(用于LLM和TTS-STT)
43+
* 支持MMS进行文本到语音和语音到文本转换
44+
* 支持Lora、Textual inversion(embedding)、Vae、MagicPrompt、Img2img、Depth、Marigold、Pix2Pix、Controlnet、Upscalers(latent和SUPIR)、Refiner、Inpaint、Outpaint、GLIGEN、DiffEdit、BLIP-Diffusion、AnimateDiff、HotShot-XL、Videos、LDM3D、SD3、Cascade、T2I-IP-ADAPTER、IP-Adapter-FaceID和Riffusion模型(用于StableDiffusion)
45+
* 支持Multiband Diffusion模型(用于AudioCraft)
46+
* 支持LibreTranslate(本地API)和SeamlessM4Tv2进行语言翻译
47+
* 支持ModelScope、ZeroScope 2、CogVideoX和Latte进行视频生成
48+
* 支持SunoBark
49+
* 支持Demucs和UVR进行音频文件分离
50+
* 支持RVC进行语音转换
51+
* 支持StableFast3D、Shap-E、SV34D和Zero123Plus进行3D生成
52+
* 支持Wav2Lip
53+
* 支持LivePortrait为图像添加动画
54+
* 支持LLM的多模态(Moondream 2)、PDF解析(OpenParse)、TTS(CoquiTTS)、STT(Whisper)、LORA和网络搜索(使用DuckDuckGo)
55+
* 用于生成图像、视频和音频的元数据信息查看器
56+
* 界面内的模型设置
57+
* 在线和离线Wiki
58+
* 图库
59+
* 模型下载器(用于LLM和StableDiffusion)
60+
* 应用程序设置
61+
* 能够查看系统传感器
62+
63+
## 必需依赖:
64+
65+
* [Python](https://www.python.org/downloads/)(3.10.11)
66+
* [Git](https://git-scm.com/downloads)
67+
* [CUDA](https://developer.nvidia.com/cuda-downloads)(12.4)和[cuDNN](https://developer.nvidia.com/cudnn-downloads)(9.1)
68+
* [FFMPEG](https://ffmpeg.org/download.html)
69+
- C++编译器
70+
- Windows:[VisualStudio](https://visualstudio.microsoft.com/ru/)[VisualStudioCode](https://code.visualstudio.com)[Cmake](https://cmake.org)
71+
- Linux:[GCC](https://gcc.gnu.org/)[VisualStudioCode](https://code.visualstudio.com)[Cmake](https://cmake.org)
72+
73+
## 最低系统要求:
74+
75+
* 系统:Windows或Linux
76+
* GPU:6GB+或CPU:8核3.6GHZ
77+
* RAM:16GB+
78+
* 磁盘空间:20GB+
79+
* 需要互联网连接以下载模型和进行安装
80+
81+
## 如何安装:
82+
83+
### Windows
84+
85+
1) 首先安装所有[必需依赖](/#必需依赖)
86+
2) 在任意位置执行`Git clone https://github.com/Dartvauder/NeuroSandboxWebUI.git`
87+
3) 运行`Install.bat`并等待安装完成
88+
4) 安装完成后,运行`Start.bat`
89+
5) 选择文件版本并等待应用程序启动
90+
6) 现在您可以开始生成了!
91+
92+
要获取更新,请运行`Update.bat`
93+
要通过终端使用虚拟环境,请运行`Venv.bat`
94+
95+
### Linux
96+
97+
1) 首先安装所有[必需依赖](/#必需依赖)
98+
2) 在任意位置执行`Git clone https://github.com/Dartvauder/NeuroSandboxWebUI.git`
99+
3) 在终端中运行`./Install.sh`并等待所有依赖项安装完成
100+
4) 安装完成后,运行`./Start.sh`
101+
5) 等待应用程序启动
102+
6) 现在您可以开始生成了!
103+
104+
要获取更新,请运行`./Update.sh`
105+
要通过终端使用虚拟环境,请运行`./Venv.sh`
106+
107+
## Wiki
108+
109+
* https://github.com/Dartvauder/NeuroSandboxWebUI/wiki/ZH‐Wiki
110+
111+
## 致开发者的感谢
112+
113+
#### 非常感谢这些项目,因为正是通过他们的应用程序/库,我才能够创建我的应用程序:
114+
115+
首先,我要感谢[PyCharm](https://www.jetbrains.com/pycharm/)[GitHub](https://desktop.github.com)的开发者。借助他们的应用程序,我能够创建并分享我的代码
116+
117+
* `gradio` - https://github.com/gradio-app/gradio
118+
* `transformers` - https://github.com/huggingface/transformers
119+
* `tts` - https://github.com/coqui-ai/TTS
120+
* `openai-whisper` - https://github.com/openai/whisper
121+
* `torch` - https://github.com/pytorch/pytorch
122+
* `soundfile` - https://github.com/bastibe/python-soundfile
123+
* `cuda-python` - https://github.com/NVIDIA/cuda-python
124+
* `gitpython` - https://github.com/gitpython-developers/GitPython
125+
* `diffusers` - https://github.com/huggingface/diffusers
126+
* `llama.cpp-python` - https://github.com/abetlen/llama-cpp-python
127+
* `stable-diffusion-cpp-python` - https://github.com/william-murray1204/stable-diffusion-cpp-python
128+
* `audiocraft` - https://github.com/facebookresearch/audiocraft
129+
* `AudioLDM2` - https://github.com/haoheliu/AudioLDM2
130+
* `xformers` - https://github.com/facebookresearch/xformers
131+
* `demucs` - https://github.com/facebookresearch/demucs
132+
* `libretranslate` - https://github.com/LibreTranslate/LibreTranslate
133+
* `libretranslatepy` - https://github.com/argosopentech/LibreTranslate-py
134+
* `rembg` - https://github.com/danielgatis/rembg
135+
* `trimesh` - https://github.com/mikedh/trimesh
136+
* `suno-bark` - https://github.com/suno-ai/bark
137+
* `IP-Adapter` - https://github.com/tencent-ailab/IP-Adapter
138+
* `PyNanoInstantMeshes` - https://github.com/vork/PyNanoInstantMeshes
139+
* `CLIP` - https://github.com/openai/CLIP
140+
* `rvc-python` - https://github.com/daswer123/rvc-python
141+
* `audio-separator` - https://github.com/nomadkaraoke/python-audio-separator
142+
* `pixeloe` - https://github.com/KohakuBlueleaf/PixelOE
143+
* `k-diffusion` - https://github.com/crowsonkb/k-diffusion
144+
* `open-parse` - https://github.com/Filimoa/open-parse
145+
* `AudioSR` - https://github.com/haoheliu/versatile_audio_super_resolution
146+
147+
## 第三方许可证:
148+
149+
#### 许多模型都有自己的使用许可证。在使用之前,我建议您熟悉它们:
150+
151+
* [Transformers](https://github.com/huggingface/transformers/blob/main/LICENSE)
152+
* [llama.cpp](https://github.com/ggerganov/llama.cpp/blob/master/LICENSE)
153+
* [stable-diffusion.cpp](https://github.com/leejet/stable-diffusion.cpp/blob/master/LICENSE)
154+
* [CoquiTTS](https://coqui.ai/cpml)
155+
* [OpenAI-Whisper](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
156+
* [LibreTranslate](https://github.com/LibreTranslate/LibreTranslate/blob/main/LICENSE)
157+
* [Diffusers](https://github.com/huggingface/diffusers/blob/main/LICENSE)
158+
* [StableDiffusion1.5](https://huggingface.co/spaces/CompVis/stable-diffusion-license)
159+
* [StableDiffusion2](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/LICENSE.md)
160+
* [StableDiffusion3](https://huggingface.co/stabilityai/stable-diffusion-3-medium-diffusers/blob/main/LICENSE)
161+
* [StableDiffusionXL](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/LICENSE.md)
162+
* [StableCascade](https://huggingface.co/stabilityai/stable-cascade/blob/main/LICENSE)
163+
* [LatentDiffusionModel3D](https://huggingface.co/spaces/CompVis/stable-diffusion-license)
164+
* [StableVideoDiffusion](https://huggingface.co/stabilityai/stable-video-diffusion-img2vid-xt-1-1/blob/main/LICENSE)
165+
* [I2VGen-XL](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md)
166+
* [Rembg](https://github.com/danielgatis/rembg/blob/main/LICENSE.txt)
167+
* [Shap-E](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md)
168+
* [StableAudioOpen](https://huggingface.co/stabilityai/stable-audio-open-1.0/blob/main/LICENSE)
169+
* [AudioCraft](https://spdx.org/licenses/CC-BY-NC-4.0)
170+
* [AudioLDM2](https://spdx.org/licenses/CC-BY-NC-SA-4.0)
171+
* [Demucs](https://github.com/facebookresearch/demucs/blob/main/LICENSE)
172+
* [SunoBark](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md)
173+
* [Moondream2](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
174+
* [ZeroScope2](https://spdx.org/licenses/CC-BY-NC-4.0)
175+
* [GLIGEN](https://huggingface.co/spaces/CompVis/stable-diffusion-license)
176+
* [Wav2Lip](https://github.com/Rudrabha/Wav2Lip)
177+
* [Roop](https://github.com/s0md3v/roop/blob/main/LICENSE)
178+
* [CodeFormer](https://github.com/sczhou/CodeFormer/blob/master/LICENSE)
179+
* [ControlNet](https://github.com/lllyasviel/ControlNet/blob/main/LICENSE)
180+
* [AnimateDiff](https://github.com/guoyww/AnimateDiff/blob/main/LICENSE.txt)
181+
* [Pix2Pix](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md)
182+
* [Kandinsky 2.1; 2.2; 3](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
183+
* [Flux-schnell](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
184+
* [Flux-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md)
185+
* [HunyuanDiT](https://huggingface.co/Tencent-Hunyuan/HunyuanDiT/blob/main/LICENSE.txt)
186+
* [Lumina-T2X](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
187+
* [DeepFloydIF](https://huggingface.co/spaces/DeepFloyd/deepfloyd-if-license)
188+
* [PixArt](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/LICENSE.md)
189+
* [CogVideoX](https://huggingface.co/THUDM/CogVideoX-2b/blob/main/LICENSE)
190+
* [Latte](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
191+
* [Kolors](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
192+
* [AuraFlow](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
193+
* [Würstchen](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md)
194+
* [ModelScope](https://spdx.org/licenses/CC-BY-NC-4.0)
195+
* [StableFast3D](https://github.com/Stability-AI/stable-fast-3d/blob/main/LICENSE.md)
196+
* [SV34D](https://huggingface.co/stabilityai/sv4d/blob/main/LICENSE.md)
197+
* [Zero123Plus](https://huggingface.co/blog/open_rail)
198+
* [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN/blob/master/LICENSE)
199+
* [Refiner](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/LICENSE.md)
200+
* [PlaygroundV2.5](https://huggingface.co/playgroundai/playground-v2.5-1024px-aesthetic/blob/main/LICENSE.md)
201+
* [AuraSR](https://huggingface.co/fal/AuraSR/blob/main/LICENSE.md)
202+
* [IP-Adapter-FaceID](https://huggingface.co/h94/IP-Adapter-FaceID)
203+
* [T2I-IP-Adapter](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
204+
* [MMS](https://spdx.org/licenses/CC-BY-NC-4.0)
205+
* [SeamlessM4Tv2](https://spdx.org/licenses/CC-BY-NC-4.0)
206+
* [HotShot-XL](https://github.com/hotshotco/Hotshot-XL/blob/main/LICENSE)
207+
* [Riffusion](https://huggingface.co/spaces/CompVis/stable-diffusion-license)
208+
* [MozillaCommonVoice17](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/cc0-1.0.md)
209+
* [UVR-MDX](https://github.com/kuielab/mdx-net/blob/main/LICENSE)
210+
* [RVC](https://github.com/RVC-Project/Retrieval-based-Voice-Conversion-WebUI/blob/main/LICENSE)
211+
* [DDColor](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
212+
* [PixelOE](https://github.com/KohakuBlueleaf/PixelOE/blob/main/LICENSE)
213+
* [LivePortrait](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md)
214+
* [SUPIR](https://github.com/Fanghua-Yu/SUPIR/blob/master/LICENSE)
215+
* [MagicPrompt](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md)
216+
* [Marigold](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
217+
* [BLIP-Diffusion](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md)
218+
219+
#### 这些第三方仓库代码也在我的项目中使用:
220+
221+
* [Generative-Models for SV34D](https://github.com/Stability-AI/generative-models)
222+
* [CodeFormer for extras](https://github.com/sczhou/CodeFormer)
223+
* [Real-ESRGAN for upscale](https://github.com/xinntao/Real-ESRGAN)
224+
* [HotShot-XL for StableDiffusion](https://github.com/hotshotco/Hotshot-XL)
225+
* [Roop for extras](https://github.com/s0md3v/roop)
226+
* [StableFast3D for 3D](https://github.com/Stability-AI/stable-fast-3d)
227+
* [Riffusion for StableDiffusion](https://github.com/riffusion/riffusion-hobby)
228+
* [DDColor for extras](https://github.com/piddnad/DDColor)
229+
* [LivePortrait for video](https://github.com/KwaiVGI/LivePortrait)
230+
* [SUPIR for StableDiffusion](https://github.com/Fanghua-Yu/SUPIR)
231+
232+
## 捐赠
233+
234+
### *如果您喜欢我的项目并想要捐赠,这里有捐赠选项。非常感谢您的支持!*
235+
236+
* [!["给我买杯咖啡"](https://www.buymeacoffee.com/assets/img/custom_images/orange_img.png)](https://www.buymeacoffee.com/Dartvauder)
237+
238+
## Star历史
239+
240+
[![Star历史图表](https://api.star-history.com/svg?repos=Dartvauder/NeuroSandboxWebUI&type=Date)](https://star-history.com/#Dartvauder/NeuroSandboxWebUI&Date)

Start.bat

+1-1
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ echo Logging in to Hugging Face...
2525
huggingface-cli login --token %HF_TOKEN% --add-to-git-credential
2626

2727
cls
28-
echo Launching app.py...
28+
echo Launching NeuroSandboxWebUI...
2929
start /b py -c "import os, sys; sys.path.insert(0, os.path.join(os.path.dirname(__name__), 'LaunchFile')); import app"
3030

3131
call "%CURRENT_DIR%venv\Scripts\deactivate.bat"

Start.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ echo "Logging in to Hugging Face..."
2828
huggingface-cli login --token "$HF_TOKEN" --add-to-git-credential
2929

3030
clear
31-
echo "Launching app.py..."
31+
echo "Launching NeuroSandboxWebUI..."
3232
python -c "import os, sys; sys.path.insert(0, os.path.join('$(dirname "${BASH_SOURCE[0]}")', 'LaunchFile')); import app" &
3333

3434
deactivate

0 commit comments

Comments
 (0)