jbilcke-hf HF staff commited on
Commit
634ff01
·
1 Parent(s): ef3ecd5

add explanations about the architecture

Browse files
Files changed (1) hide show
  1. README.md +25 -0
README.md CHANGED
@@ -10,3 +10,28 @@ app_port: 8000
10
 
11
  Media server 📡
12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
 
11
  Media server 📡
12
 
13
+
14
+ The main code of the webtv is located inside the [media-server](https://huggingface.co/spaces/jbilcke-hf/media-server/tree/main) :
15
+
16
+ manual steps:
17
+ - human input to write a short paragraph describing a multi-shot video sequence
18
+ - manual submit it to GPT-4 to generate a list of video captions for each shot (the system instructions are extracts from a stable diffusion guide)
19
+ - commit the captions to the [playlist database](https://huggingface.co/spaces/jbilcke-hf/media-server/raw/main/database.json)
20
+
21
+ Inside the `media-server` space (generation process running in the background):
22
+ - for each prompt in the database
23
+ - generate a silent 3 seconds video clip with Zeroscope V2 576w (hosted on Hugging Face Spaces)
24
+ - upscale the clip with Zeroscope V2 XL (also a HF Space)
25
+ - perform frame interpolation with FILM (also a HF Space)
26
+ - storage in the Persistent Storage of the media-server Space
27
+
28
+ Inside the `media-server` space (streaming process running in the foreground):
29
+ - for each video file in the persistent storage folder
30
+ - add it to a new FFmpeg playlist (it's just a .txt file)
31
+ - broadcast it over the RTMP protocol using FFmpeg (in FLV format)
32
+ - diffusion of the stream using node-media-server
33
+
34
+ Inside the `AI-WebTV` space:
35
+ - display the stream using `mpegts.js`
36
+ - this doesn't work on iPhone, but now there is also a Twitch mirror
37
+