Skip to content

Commit

Permalink
models : change HF hosting from dataset to model
Browse files Browse the repository at this point in the history
  • Loading branch information
ggerganov committed Mar 22, 2023
1 parent 09e9068 commit 1beff6f
Show file tree
Hide file tree
Showing 7 changed files with 9 additions and 10 deletions.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -466,7 +466,7 @@ The original models are converted to a custom binary format. This allows to pack
You can download the converted models using the [models/download-ggml-model.sh](models/download-ggml-model.sh) script
or manually from here:
- https://huggingface.co/datasets/ggerganov/whisper.cpp
- https://huggingface.co/ggerganov/whisper.cpp
- https://ggml.ggerganov.com
For more details, see the conversion script [models/convert-pt-to-ggml.py](models/convert-pt-to-ggml.py) or the README
Expand Down
6 changes: 3 additions & 3 deletions bindings/go/examples/go-model-download/main.go
Original file line number Diff line number Diff line change
Expand Up @@ -17,9 +17,9 @@ import (
// CONSTANTS

const (
srcUrl = "https://huggingface.co/datasets/ggerganov/whisper.cpp/resolve/main" // The location of the models
srcExt = ".bin" // Filename extension
bufSize = 1024 * 64 // Size of the buffer used for downloading the model
srcUrl = "https://huggingface.co/ggerganov/whisper.cpp/resolve/main" // The location of the models
srcExt = ".bin" // Filename extension
bufSize = 1024 * 64 // Size of the buffer used for downloading the model
)

var (
Expand Down
2 changes: 1 addition & 1 deletion examples/talk/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ To run this, you will need a ggml GPT-2 model: [instructions](https://github.com
Alternatively, you can simply download the smallest ggml GPT-2 117M model (240 MB) like this:

```
wget --quiet --show-progress -O models/ggml-gpt-2-117M.bin https://huggingface.co/datasets/ggerganov/ggml/raw/main/ggml-model-gpt-2-117M.bin
wget --quiet --show-progress -O models/ggml-gpt-2-117M.bin https://huggingface.co/ggerganov/ggml/raw/main/ggml-model-gpt-2-117M.bin
```

## TTS
Expand Down
4 changes: 2 additions & 2 deletions models/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ using the [convert-pt-to-ggml.py](convert-pt-to-ggml.py) script. You can either
the `ggml` files yourself using the conversion script, or you can use the [download-ggml-model.sh](download-ggml-model.sh)
script to download the already converted models. Currently, they are hosted on the following locations:

- https://huggingface.co/datasets/ggerganov/whisper.cpp
- https://huggingface.co/ggerganov/whisper.cpp
- https://ggml.ggerganov.com

Sample usage:
Expand All @@ -23,7 +23,7 @@ You can now use it like this:

A third option to obtain the model files is to download them from Hugging Face:

https://huggingface.co/datasets/ggerganov/whisper.cpp/tree/main
https://huggingface.co/ggerganov/whisper.cpp/tree/main

## Available models

Expand Down
2 changes: 1 addition & 1 deletion models/download-ggml-model.cmd
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ if exist "ggml-%model%.bin" (
goto :eof
)

PowerShell -NoProfile -ExecutionPolicy Bypass -Command "Invoke-WebRequest -Uri https://huggingface.co/datasets/ggerganov/whisper.cpp/resolve/main/ggml-%model%.bin -OutFile ggml-%model%.bin"
PowerShell -NoProfile -ExecutionPolicy Bypass -Command "Invoke-WebRequest -Uri https://huggingface.co/ggerganov/whisper.cpp/resolve/main/ggml-%model%.bin -OutFile ggml-%model%.bin"

if %ERRORLEVEL% neq 0 (
echo Failed to download ggml model %model%
Expand Down
2 changes: 1 addition & 1 deletion models/download-ggml-model.sh
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
#src="https://ggml.ggerganov.com"
#pfx="ggml-model-whisper"

src="https://huggingface.co/datasets/ggerganov/whisper.cpp"
src="https://huggingface.co/ggerganov/whisper.cpp"
pfx="resolve/main/ggml"

# get the path of this script
Expand Down
1 change: 0 additions & 1 deletion whisper.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -631,7 +631,6 @@ struct whisper_context {
int64_t t_load_us = 0;
int64_t t_start_us = 0;


ggml_type wtype = ggml_type::GGML_TYPE_F16; // weight type (FP32 or FP16)

whisper_model model;
Expand Down

0 comments on commit 1beff6f

Please sign in to comment.