From 130e8502a3f30ddf1b4050223687e3c88dbdad37 Mon Sep 17 00:00:00 2001 From: Adam Letts Date: Sun, 10 Apr 2022 12:41:39 -0400 Subject: [PATCH] Update .ipynb title for v5.2 VR Mode --- Disco_Diffusion.ipynb | 7 ++++--- disco.py | 5 +++-- 2 files changed, 7 insertions(+), 5 deletions(-) diff --git a/Disco_Diffusion.ipynb b/Disco_Diffusion.ipynb index 179f64b..614a910 100644 --- a/Disco_Diffusion.ipynb +++ b/Disco_Diffusion.ipynb @@ -16,7 +16,7 @@ "id": "TitleTop" }, "source": [ - "# Disco Diffusion v5.1 - Now with Turbo\n", + "# Disco Diffusion v5.2 - Now with VR Mode\n", "\n", "In case of confusion, Disco is the name of this notebook edit. The diffusion model in use is Katherine Crowson's fine-tuned 512x512 model\n", "\n", @@ -2042,7 +2042,8 @@ "#@markdown Google made the VR180 Creator tool but subsequently stopped supporting it. It's available for download in a few places including https://www.patrickgrunwald.de/vr180-creator-download\n", "#@markdown The tool is not only good for stitching (videos and photos) but also for adding the correct metadata into existing videos, which is needed for services like YouTube to identify the format correctly.\n", "#@markdown Watching YouTube VR videos isn't necessarily the easiest depending on your headset. For instance Oculus have a dedicated media studio and store which makes the files easier to access on a Quest https://creator.oculus.com/manage/mediastudio/\n", - "#@markdown The command to get ffmpeg to concat your frames for each eye is in the form: ffmpeg -framerate 15 -i frame_%4d_l.png l.mp4 (repeat for r)\n", + "#@markdown \n", + "#@markdown The command to get ffmpeg to concat your frames for each eye is in the form: `ffmpeg -framerate 15 -i frame_%4d_l.png l.mp4` (repeat for r)\n", "\n", "vr_mode = False #@param {type:\"boolean\"}\n", "#@markdown `vr_eye_angle` is the y-axis rotation of the eyes towards the center\n", @@ -2717,7 +2718,7 @@ "ExtraSetTop" ], "machine_shape": "hm", - "name": "Disco Diffusion v5.1 [w/ Turbo]", + "name": "Disco Diffusion v5.2 [w/ VR Mode]", "private_outputs": true, "provenance": [], "include_colab_link": true diff --git a/disco.py b/disco.py index b00dd5e..9ccf510 100644 --- a/disco.py +++ b/disco.py @@ -12,7 +12,7 @@ # !! "id": "TitleTop" # !! }} """ -# Disco Diffusion v5.1 - Now with Turbo +# Disco Diffusion v5.2 - Now with VR Mode In case of confusion, Disco is the name of this notebook edit. The diffusion model in use is Katherine Crowson's fine-tuned 512x512 model @@ -1985,7 +1985,8 @@ frames_skip_steps = '60%' #@param ['40%', '50%', '60%', '70%', '80%'] {type: 'st #@markdown Google made the VR180 Creator tool but subsequently stopped supporting it. It's available for download in a few places including https://www.patrickgrunwald.de/vr180-creator-download #@markdown The tool is not only good for stitching (videos and photos) but also for adding the correct metadata into existing videos, which is needed for services like YouTube to identify the format correctly. #@markdown Watching YouTube VR videos isn't necessarily the easiest depending on your headset. For instance Oculus have a dedicated media studio and store which makes the files easier to access on a Quest https://creator.oculus.com/manage/mediastudio/ -#@markdown The command to get ffmpeg to concat your frames for each eye is in the form: ffmpeg -framerate 15 -i frame_%4d_l.png l.mp4 (repeat for r) +#@markdown +#@markdown The command to get ffmpeg to concat your frames for each eye is in the form: `ffmpeg -framerate 15 -i frame_%4d_l.png l.mp4` (repeat for r) vr_mode = False #@param {type:"boolean"} #@markdown `vr_eye_angle` is the y-axis rotation of the eyes towards the center