|
1 |
| -# Soon... |
| 1 | +## [功能](/#功能) | [依赖](/#必需依赖) | [系统要求](/#最低系统要求) | [安装](/#如何安装) | [Wiki](/#Wiki) | [致谢](/#致开发者的感谢) | [许可证](/#第三方许可证) |
| 2 | + |
| 3 | +#  |
| 4 | +* 仍在进行中但已稳定! |
| 5 | +* [English](/README.md) | [Русский](/Readmes/README_RU.md) | 漢語 |
| 6 | + |
| 7 | +## 描述: |
| 8 | + |
| 9 | +一个简单方便的界面,用于使用各种神经网络模型。您可以通过文本、语音和图像输入与LLM和Moondream2进行通信;使用StableDiffusion、Kandinsky、Flux、HunyuanDiT、Lumina-T2X、Kolors、AuraFlow、Würstchen、DeepFloydIF、PixArt和PlaygroundV2.5生成图像;使用ModelScope、ZeroScope 2、CogVideoX和Latte生成视频;使用StableFast3D、Shap-E、SV34D和Zero123Plus生成3D对象;使用StableAudioOpen、AudioCraft和AudioLDM 2生成音乐和音频;使用CoquiTTS、MMS和SunoBark进行文本到语音转换;使用OpenAI-Whisper和MMS进行语音到文本转换;使用Wav2Lip进行唇形同步;使用LivePortrait为图像添加动画;使用Roop进行换脸;使用Rembg移除背景;使用CodeFormer修复面部;使用PixelOE进行图像像素化;使用DDColor为图像上色;使用LibreTranslate和SeamlessM4Tv2进行文本翻译;使用Demucs和UVR进行音频文件分离;使用RVC进行语音转换。您还可以在图库中查看输出目录中的文件,下载LLM和StableDiffusion模型,在界面内更改应用程序设置并检查系统传感器。 |
| 10 | + |
| 11 | +项目目标 - 创建一个尽可能简单易用的神经网络模型应用程序 |
| 12 | + |
| 13 | +### 文本:<img width="1119" alt="1zh" src="https://github.com/user-attachments/assets/1cbda009-8230-4dc2-beb7-a77505e96d81"> |
| 14 | + |
| 15 | +### 图像:<img width="1127" alt="2zh" src="https://github.com/user-attachments/assets/151001bc-27b2-4561-84f1-7ec521ad972e"> |
| 16 | + |
| 17 | +### 视频:<img width="1117" alt="3zh" src="https://github.com/user-attachments/assets/70ebf95e-f82e-467a-a027-64ee917527cc"> |
| 18 | + |
| 19 | +### 3D:<img width="1121" alt="4zh" src="https://github.com/user-attachments/assets/6c78d5a9-3794-43af-b167-fb63a4102d83"> |
| 20 | + |
| 21 | +### 音频:<img width="1117" alt="5zh" src="https://github.com/user-attachments/assets/f87efe4d-095f-4e99-abfb-759801ff4f29"> |
| 22 | + |
| 23 | +### 额外功能:<img width="1116" alt="6zh" src="https://github.com/user-attachments/assets/f92da750-9ce6-4982-80e1-324c0ee749c3"> |
| 24 | + |
| 25 | +### 界面:<img width="1120" alt="7zh" src="https://github.com/user-attachments/assets/f444b0bb-cc7d-46fc-8eb6-a944e9269838"> |
| 26 | + |
| 27 | +## 功能: |
| 28 | + |
| 29 | +* 通过install.bat(Windows)或install.sh(Linux)轻松安装 |
| 30 | +* 您可以通过移动设备在本地主机(通过IPv4)或在线任何地方(通过Share)使用应用程序 |
| 31 | +* 灵活且优化的界面(由Gradio提供) |
| 32 | +* 从`Install`和`Update`文件进行调试日志记录 |
| 33 | +* 提供三种语言版本 |
| 34 | +* 支持Transformers和llama.cpp模型(LLM) |
| 35 | +* 支持diffusers和safetensors模型(StableDiffusion)- txt2img、img2img、depth2img、marigold、pix2pix、controlnet、upscale(latent)、upscale(SUPIR)、refiner、inpaint、outpaint、gligen、diffedit、blip-diffusion、animatediff、hotshot-xl、video、ldm3d、sd3、cascade、t2i-ip-adapter、ip-adapter-faceid和riffusion标签 |
| 36 | +* 支持stable-diffusion-cpp模型用于FLUX |
| 37 | +* 支持额外的图像生成模型:Kandinsky(txt2img、img2img、inpaint)、Flux(支持LoRA)、HunyuanDiT(txt2img、controlnet)、Lumina-T2X、Kolors(支持LoRA的txt2img、img2img、ip-adapter-plus)、AuraFlow(支持LoRA和AuraSR)、Würstchen、DeepFloydIF(txt2img、img2img、inpaint)、PixArt和PlaygroundV2.5 |
| 38 | +* 支持使用Rembg、CodeFormer、PixelOE、DDColor、DownScale、格式转换器、换脸(Roop)和放大(Real-ESRGAN)模型进行图像、视频和音频的额外处理 |
| 39 | +* 支持StableAudio |
| 40 | +* 支持AudioCraft(模型:musicgen、audiogen和magnet) |
| 41 | +* 支持AudioLDM 2(模型:audio和music) |
| 42 | +* 支持TTS和Whisper模型(用于LLM和TTS-STT) |
| 43 | +* 支持MMS进行文本到语音和语音到文本转换 |
| 44 | +* 支持Lora、Textual inversion(embedding)、Vae、MagicPrompt、Img2img、Depth、Marigold、Pix2Pix、Controlnet、Upscalers(latent和SUPIR)、Refiner、Inpaint、Outpaint、GLIGEN、DiffEdit、BLIP-Diffusion、AnimateDiff、HotShot-XL、Videos、LDM3D、SD3、Cascade、T2I-IP-ADAPTER、IP-Adapter-FaceID和Riffusion模型(用于StableDiffusion) |
| 45 | +* 支持Multiband Diffusion模型(用于AudioCraft) |
| 46 | +* 支持LibreTranslate(本地API)和SeamlessM4Tv2进行语言翻译 |
| 47 | +* 支持ModelScope、ZeroScope 2、CogVideoX和Latte进行视频生成 |
| 48 | +* 支持SunoBark |
| 49 | +* 支持Demucs和UVR进行音频文件分离 |
| 50 | +* 支持RVC进行语音转换 |
| 51 | +* 支持StableFast3D、Shap-E、SV34D和Zero123Plus进行3D生成 |
| 52 | +* 支持Wav2Lip |
| 53 | +* 支持LivePortrait为图像添加动画 |
| 54 | +* 支持LLM的多模态(Moondream 2)、PDF解析(OpenParse)、TTS(CoquiTTS)、STT(Whisper)、LORA和网络搜索(使用DuckDuckGo) |
| 55 | +* 用于生成图像、视频和音频的元数据信息查看器 |
| 56 | +* 界面内的模型设置 |
| 57 | +* 在线和离线Wiki |
| 58 | +* 图库 |
| 59 | +* 模型下载器(用于LLM和StableDiffusion) |
| 60 | +* 应用程序设置 |
| 61 | +* 能够查看系统传感器 |
| 62 | + |
| 63 | +## 必需依赖: |
| 64 | + |
| 65 | +* [Python](https://www.python.org/downloads/)(3.10.11) |
| 66 | +* [Git](https://git-scm.com/downloads) |
| 67 | +* [CUDA](https://developer.nvidia.com/cuda-downloads)(12.4)和[cuDNN](https://developer.nvidia.com/cudnn-downloads)(9.1) |
| 68 | +* [FFMPEG](https://ffmpeg.org/download.html) |
| 69 | +- C++编译器 |
| 70 | + - Windows:[VisualStudio](https://visualstudio.microsoft.com/ru/)、[VisualStudioCode](https://code.visualstudio.com)和[Cmake](https://cmake.org) |
| 71 | + - Linux:[GCC](https://gcc.gnu.org/)、[VisualStudioCode](https://code.visualstudio.com)和[Cmake](https://cmake.org) |
| 72 | + |
| 73 | +## 最低系统要求: |
| 74 | + |
| 75 | +* 系统:Windows或Linux |
| 76 | +* GPU:6GB+或CPU:8核3.6GHZ |
| 77 | +* RAM:16GB+ |
| 78 | +* 磁盘空间:20GB+ |
| 79 | +* 需要互联网连接以下载模型和进行安装 |
| 80 | + |
| 81 | +## 如何安装: |
| 82 | + |
| 83 | +### Windows |
| 84 | + |
| 85 | +1) 首先安装所有[必需依赖](/#必需依赖) |
| 86 | +2) 在任意位置执行`Git clone https://github.com/Dartvauder/NeuroSandboxWebUI.git` |
| 87 | +3) 运行`Install.bat`并等待安装完成 |
| 88 | +4) 安装完成后,运行`Start.bat` |
| 89 | +5) 选择文件版本并等待应用程序启动 |
| 90 | +6) 现在您可以开始生成了! |
| 91 | + |
| 92 | +要获取更新,请运行`Update.bat` |
| 93 | +要通过终端使用虚拟环境,请运行`Venv.bat` |
| 94 | + |
| 95 | +### Linux |
| 96 | + |
| 97 | +1) 首先安装所有[必需依赖](/#必需依赖) |
| 98 | +2) 在任意位置执行`Git clone https://github.com/Dartvauder/NeuroSandboxWebUI.git` |
| 99 | +3) 在终端中运行`./Install.sh`并等待所有依赖项安装完成 |
| 100 | +4) 安装完成后,运行`./Start.sh` |
| 101 | +5) 等待应用程序启动 |
| 102 | +6) 现在您可以开始生成了! |
| 103 | + |
| 104 | +要获取更新,请运行`./Update.sh` |
| 105 | +要通过终端使用虚拟环境,请运行`./Venv.sh` |
| 106 | + |
| 107 | +## Wiki |
| 108 | + |
| 109 | +* https://github.com/Dartvauder/NeuroSandboxWebUI/wiki/ZH‐Wiki |
| 110 | + |
| 111 | +## 致开发者的感谢 |
| 112 | + |
| 113 | +#### 非常感谢这些项目,因为正是通过他们的应用程序/库,我才能够创建我的应用程序: |
| 114 | + |
| 115 | +首先,我要感谢[PyCharm](https://www.jetbrains.com/pycharm/)和[GitHub](https://desktop.github.com)的开发者。借助他们的应用程序,我能够创建并分享我的代码 |
| 116 | + |
| 117 | +* `gradio` - https://github.com/gradio-app/gradio |
| 118 | +* `transformers` - https://github.com/huggingface/transformers |
| 119 | +* `tts` - https://github.com/coqui-ai/TTS |
| 120 | +* `openai-whisper` - https://github.com/openai/whisper |
| 121 | +* `torch` - https://github.com/pytorch/pytorch |
| 122 | +* `soundfile` - https://github.com/bastibe/python-soundfile |
| 123 | +* `cuda-python` - https://github.com/NVIDIA/cuda-python |
| 124 | +* `gitpython` - https://github.com/gitpython-developers/GitPython |
| 125 | +* `diffusers` - https://github.com/huggingface/diffusers |
| 126 | +* `llama.cpp-python` - https://github.com/abetlen/llama-cpp-python |
| 127 | +* `stable-diffusion-cpp-python` - https://github.com/william-murray1204/stable-diffusion-cpp-python |
| 128 | +* `audiocraft` - https://github.com/facebookresearch/audiocraft |
| 129 | +* `AudioLDM2` - https://github.com/haoheliu/AudioLDM2 |
| 130 | +* `xformers` - https://github.com/facebookresearch/xformers |
| 131 | +* `demucs` - https://github.com/facebookresearch/demucs |
| 132 | +* `libretranslate` - https://github.com/LibreTranslate/LibreTranslate |
| 133 | +* `libretranslatepy` - https://github.com/argosopentech/LibreTranslate-py |
| 134 | +* `rembg` - https://github.com/danielgatis/rembg |
| 135 | +* `trimesh` - https://github.com/mikedh/trimesh |
| 136 | +* `suno-bark` - https://github.com/suno-ai/bark |
| 137 | +* `IP-Adapter` - https://github.com/tencent-ailab/IP-Adapter |
| 138 | +* `PyNanoInstantMeshes` - https://github.com/vork/PyNanoInstantMeshes |
| 139 | +* `CLIP` - https://github.com/openai/CLIP |
| 140 | +* `rvc-python` - https://github.com/daswer123/rvc-python |
| 141 | +* `audio-separator` - https://github.com/nomadkaraoke/python-audio-separator |
| 142 | +* `pixeloe` - https://github.com/KohakuBlueleaf/PixelOE |
| 143 | +* `k-diffusion` - https://github.com/crowsonkb/k-diffusion |
| 144 | +* `open-parse` - https://github.com/Filimoa/open-parse |
| 145 | +* `AudioSR` - https://github.com/haoheliu/versatile_audio_super_resolution |
| 146 | + |
| 147 | +## 第三方许可证: |
| 148 | + |
| 149 | +#### 许多模型都有自己的使用许可证。在使用之前,我建议您熟悉它们: |
| 150 | + |
| 151 | +* [Transformers](https://github.com/huggingface/transformers/blob/main/LICENSE) |
| 152 | +* [llama.cpp](https://github.com/ggerganov/llama.cpp/blob/master/LICENSE) |
| 153 | +* [stable-diffusion.cpp](https://github.com/leejet/stable-diffusion.cpp/blob/master/LICENSE) |
| 154 | +* [CoquiTTS](https://coqui.ai/cpml) |
| 155 | +* [OpenAI-Whisper](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 156 | +* [LibreTranslate](https://github.com/LibreTranslate/LibreTranslate/blob/main/LICENSE) |
| 157 | +* [Diffusers](https://github.com/huggingface/diffusers/blob/main/LICENSE) |
| 158 | +* [StableDiffusion1.5](https://huggingface.co/spaces/CompVis/stable-diffusion-license) |
| 159 | +* [StableDiffusion2](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/LICENSE.md) |
| 160 | +* [StableDiffusion3](https://huggingface.co/stabilityai/stable-diffusion-3-medium-diffusers/blob/main/LICENSE) |
| 161 | +* [StableDiffusionXL](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/LICENSE.md) |
| 162 | +* [StableCascade](https://huggingface.co/stabilityai/stable-cascade/blob/main/LICENSE) |
| 163 | +* [LatentDiffusionModel3D](https://huggingface.co/spaces/CompVis/stable-diffusion-license) |
| 164 | +* [StableVideoDiffusion](https://huggingface.co/stabilityai/stable-video-diffusion-img2vid-xt-1-1/blob/main/LICENSE) |
| 165 | +* [I2VGen-XL](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md) |
| 166 | +* [Rembg](https://github.com/danielgatis/rembg/blob/main/LICENSE.txt) |
| 167 | +* [Shap-E](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md) |
| 168 | +* [StableAudioOpen](https://huggingface.co/stabilityai/stable-audio-open-1.0/blob/main/LICENSE) |
| 169 | +* [AudioCraft](https://spdx.org/licenses/CC-BY-NC-4.0) |
| 170 | +* [AudioLDM2](https://spdx.org/licenses/CC-BY-NC-SA-4.0) |
| 171 | +* [Demucs](https://github.com/facebookresearch/demucs/blob/main/LICENSE) |
| 172 | +* [SunoBark](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md) |
| 173 | +* [Moondream2](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 174 | +* [ZeroScope2](https://spdx.org/licenses/CC-BY-NC-4.0) |
| 175 | +* [GLIGEN](https://huggingface.co/spaces/CompVis/stable-diffusion-license) |
| 176 | +* [Wav2Lip](https://github.com/Rudrabha/Wav2Lip) |
| 177 | +* [Roop](https://github.com/s0md3v/roop/blob/main/LICENSE) |
| 178 | +* [CodeFormer](https://github.com/sczhou/CodeFormer/blob/master/LICENSE) |
| 179 | +* [ControlNet](https://github.com/lllyasviel/ControlNet/blob/main/LICENSE) |
| 180 | +* [AnimateDiff](https://github.com/guoyww/AnimateDiff/blob/main/LICENSE.txt) |
| 181 | +* [Pix2Pix](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md) |
| 182 | +* [Kandinsky 2.1; 2.2; 3](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 183 | +* [Flux-schnell](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 184 | +* [Flux-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md) |
| 185 | +* [HunyuanDiT](https://huggingface.co/Tencent-Hunyuan/HunyuanDiT/blob/main/LICENSE.txt) |
| 186 | +* [Lumina-T2X](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 187 | +* [DeepFloydIF](https://huggingface.co/spaces/DeepFloyd/deepfloyd-if-license) |
| 188 | +* [PixArt](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/LICENSE.md) |
| 189 | +* [CogVideoX](https://huggingface.co/THUDM/CogVideoX-2b/blob/main/LICENSE) |
| 190 | +* [Latte](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 191 | +* [Kolors](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 192 | +* [AuraFlow](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 193 | +* [Würstchen](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md) |
| 194 | +* [ModelScope](https://spdx.org/licenses/CC-BY-NC-4.0) |
| 195 | +* [StableFast3D](https://github.com/Stability-AI/stable-fast-3d/blob/main/LICENSE.md) |
| 196 | +* [SV34D](https://huggingface.co/stabilityai/sv4d/blob/main/LICENSE.md) |
| 197 | +* [Zero123Plus](https://huggingface.co/blog/open_rail) |
| 198 | +* [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN/blob/master/LICENSE) |
| 199 | +* [Refiner](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/LICENSE.md) |
| 200 | +* [PlaygroundV2.5](https://huggingface.co/playgroundai/playground-v2.5-1024px-aesthetic/blob/main/LICENSE.md) |
| 201 | +* [AuraSR](https://huggingface.co/fal/AuraSR/blob/main/LICENSE.md) |
| 202 | +* [IP-Adapter-FaceID](https://huggingface.co/h94/IP-Adapter-FaceID) |
| 203 | +* [T2I-IP-Adapter](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 204 | +* [MMS](https://spdx.org/licenses/CC-BY-NC-4.0) |
| 205 | +* [SeamlessM4Tv2](https://spdx.org/licenses/CC-BY-NC-4.0) |
| 206 | +* [HotShot-XL](https://github.com/hotshotco/Hotshot-XL/blob/main/LICENSE) |
| 207 | +* [Riffusion](https://huggingface.co/spaces/CompVis/stable-diffusion-license) |
| 208 | +* [MozillaCommonVoice17](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/cc0-1.0.md) |
| 209 | +* [UVR-MDX](https://github.com/kuielab/mdx-net/blob/main/LICENSE) |
| 210 | +* [RVC](https://github.com/RVC-Project/Retrieval-based-Voice-Conversion-WebUI/blob/main/LICENSE) |
| 211 | +* [DDColor](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 212 | +* [PixelOE](https://github.com/KohakuBlueleaf/PixelOE/blob/main/LICENSE) |
| 213 | +* [LivePortrait](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md) |
| 214 | +* [SUPIR](https://github.com/Fanghua-Yu/SUPIR/blob/master/LICENSE) |
| 215 | +* [MagicPrompt](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md) |
| 216 | +* [Marigold](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 217 | +* [BLIP-Diffusion](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
| 218 | + |
| 219 | +#### 这些第三方仓库代码也在我的项目中使用: |
| 220 | + |
| 221 | +* [Generative-Models for SV34D](https://github.com/Stability-AI/generative-models) |
| 222 | +* [CodeFormer for extras](https://github.com/sczhou/CodeFormer) |
| 223 | +* [Real-ESRGAN for upscale](https://github.com/xinntao/Real-ESRGAN) |
| 224 | +* [HotShot-XL for StableDiffusion](https://github.com/hotshotco/Hotshot-XL) |
| 225 | +* [Roop for extras](https://github.com/s0md3v/roop) |
| 226 | +* [StableFast3D for 3D](https://github.com/Stability-AI/stable-fast-3d) |
| 227 | +* [Riffusion for StableDiffusion](https://github.com/riffusion/riffusion-hobby) |
| 228 | +* [DDColor for extras](https://github.com/piddnad/DDColor) |
| 229 | +* [LivePortrait for video](https://github.com/KwaiVGI/LivePortrait) |
| 230 | +* [SUPIR for StableDiffusion](https://github.com/Fanghua-Yu/SUPIR) |
| 231 | + |
| 232 | +## 捐赠 |
| 233 | + |
| 234 | +### *如果您喜欢我的项目并想要捐赠,这里有捐赠选项。非常感谢您的支持!* |
| 235 | + |
| 236 | +* [](https://www.buymeacoffee.com/Dartvauder) |
| 237 | + |
| 238 | +## Star历史 |
| 239 | + |
| 240 | +[](https://star-history.com/#Dartvauder/NeuroSandboxWebUI&Date) |
0 commit comments