From f29cb24090555589be6ab2cf4821c4d9721c51b2 Mon Sep 17 00:00:00 2001 From: nateraw Date: Thu, 20 Oct 2022 03:16:06 +0000 Subject: [PATCH] :pushpin: pin release 0.6.0 and update readme --- README.md | 4 +--- stable_diffusion_videos/__init__.py | 2 +- 2 files changed, 2 insertions(+), 4 deletions(-) diff --git a/README.md b/README.md index b0ac2ac..3c1b6db 100644 --- a/README.md +++ b/README.md @@ -23,7 +23,7 @@ The app is built with [Gradio](https://gradio.app/), which allows you to interac 2. Generate videos using the "Videos" tab - Using the images you found from the step above, provide the prompts/seeds you recorded - - Set the `num_walk_steps` - for testing you can use a small number like 3 or 5, but to get great results you'll want to use something larger (60-200 steps). + - Set the `num_interpolation_steps` - for testing you can use a small number like 3 or 5, but to get great results you'll want to use something larger (60-200 steps). - You can set the `output_dir` to the directory you wish to save to ## Python Package @@ -51,7 +51,6 @@ import torch pipeline = StableDiffusionWalkPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4", - use_auth_token=True, torch_dtype=torch.float16, revision="fp16", scheduler=LMSDiscreteScheduler( @@ -83,7 +82,6 @@ import torch pipeline = StableDiffusionWalkPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4", - use_auth_token=True, torch_dtype=torch.float16, revision="fp16", scheduler=LMSDiscreteScheduler( diff --git a/stable_diffusion_videos/__init__.py b/stable_diffusion_videos/__init__.py index 8617b02..39fb9e0 100644 --- a/stable_diffusion_videos/__init__.py +++ b/stable_diffusion_videos/__init__.py @@ -114,4 +114,4 @@ def __dir__(): }, ) -__version__ = "0.5.3" +__version__ = "0.6.0"