diff --git a/README.md b/README.md index ef7bc38..94610f7 100644 --- a/README.md +++ b/README.md @@ -1,8 +1,9 @@ # stable-diffusion-videos Try it yourself in Colab: [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/nateraw/stable-diffusion-videos/blob/main/stable_diffusion_videos.ipynb) - + **Example** - morphing between "blueberry spaghetti" and "strawberry spaghetti" @@ -104,9 +105,10 @@ This work built off of [a script](https://gist.github.com/karpathy/00103b0037c5a You can file any issues/feature requests [here](https://github.com/nateraw/stable-diffusion-videos/issues) Enjoy 🤗 + + diff --git a/requirements.txt b/requirements.txt index 4e760fc..9ff6c72 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,10 +1,10 @@ transformers>=4.21.0 -diffusers==0.11.1 +diffusers scipy fire gradio librosa av<10.0.0 -realesrgan==0.2.5.0 +realesrgan protobuf==3.20.* -fsspec>=2023.4.0 \ No newline at end of file +fsspec>=2023.4.0 diff --git a/stable_diffusion_videos.ipynb b/stable_diffusion_videos.ipynb index 3524de0..0182739 100644 --- a/stable_diffusion_videos.ipynb +++ b/stable_diffusion_videos.ipynb @@ -7,7 +7,7 @@ "colab_type": "text" }, "source": [ - "\"Open" + "\"Open" ] }, { @@ -24,7 +24,7 @@ "\n", "If you like this notebook:\n", "- consider giving the [repo a star](https://github.com/nateraw/stable-diffusion-videos) ⭐️\n", - "- consider following me on Github [@nateraw](https://github.com/nateraw) \n", + "- consider following me on Github [@nateraw](https://github.com/nateraw)\n", "\n", "You can file any issues/feature requests [here](https://github.com/nateraw/stable-diffusion-videos/issues)\n", "\n", @@ -49,7 +49,7 @@ "outputs": [], "source": [ "%%capture\n", - "! pip install stable_diffusion_videos" + "! pip install git+https://github.com/nateraw/stable-diffusion-videos@nate-fix-breaking-changes" ] }, { @@ -82,7 +82,8 @@ "source": [ "import torch\n", "\n", - "from stable_diffusion_videos import StableDiffusionWalkPipeline, Interface\n", + "from stable_diffusion_videos.app import Interface\n", + "from stable_diffusion_videos.stable_diffusion_pipeline import StableDiffusionWalkPipeline\n", "\n", "pipeline = StableDiffusionWalkPipeline.from_pretrained(\n", " \"CompVis/stable-diffusion-v1-4\",\n", @@ -138,7 +139,7 @@ "\n", "2. Generate videos using the \"Videos\" tab\n", " - Using the images you found from the step above, provide the prompts/seeds you recorded\n", - " - Set the `num_interpolation_steps` - for testing you can use a small number like 3 or 5, but to get great results you'll want to use something larger (60-200 steps). \n", + " - Set the `num_interpolation_steps` - for testing you can use a small number like 3 or 5, but to get great results you'll want to use something larger (60-200 steps).\n", "\n", "💡 **Pro tip** - Click the link that looks like `https://.gradio.app` below , and you'll be able to view it in full screen." ] diff --git a/stable_diffusion_videos/__init__.py b/stable_diffusion_videos/__init__.py index 3552e10..6dc238c 100644 --- a/stable_diffusion_videos/__init__.py +++ b/stable_diffusion_videos/__init__.py @@ -118,4 +118,4 @@ def __dir__(): }, ) -__version__ = "0.8.1" +__version__ = "0.9.0" diff --git a/stable_diffusion_videos/app.py b/stable_diffusion_videos/app.py index 0fd0b0c..0cb797e 100644 --- a/stable_diffusion_videos/app.py +++ b/stable_diffusion_videos/app.py @@ -2,7 +2,7 @@ import gradio as gr -from stable_diffusion_videos import generate_images, generate_images_flax +from .image_generation import generate_images, generate_images_flax class Interface: diff --git a/stable_diffusion_videos/stable_diffusion_pipeline.py b/stable_diffusion_videos/stable_diffusion_pipeline.py index fb48122..8c41b89 100644 --- a/stable_diffusion_videos/stable_diffusion_pipeline.py +++ b/stable_diffusion_videos/stable_diffusion_pipeline.py @@ -9,7 +9,7 @@ import torch from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNet2DConditionModel -from diffusers.pipeline_utils import DiffusionPipeline +from diffusers import DiffusionPipeline from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import ( StableDiffusionSafetyChecker, diff --git a/stable_diffusion_videos/upsampling.py b/stable_diffusion_videos/upsampling.py index 299caf2..3e17c22 100644 --- a/stable_diffusion_videos/upsampling.py +++ b/stable_diffusion_videos/upsampling.py @@ -6,14 +6,6 @@ from torch import nn from diffusers.utils import logging -try: - from realesrgan import RealESRGANer - from basicsr.archs.rrdbnet_arch import RRDBNet -except ImportError as e: - raise ImportError( - "You tried to import realesrgan without having it installed properly. To install Real-ESRGAN, run:\n\n" - "pip install realesrgan" - ) logger = logging.get_logger(__name__) # pylint: disable=invalid-name