The Checkpoint - Issue #14
Stability Diffusion beta signup now open – Run, don't walk.
Compositional human-scene interaction synthesis with semantic control (COINS) – Given a 3D scene and a text instruction, COINS can generate realistic human interaction with that scene. Project page, Github.
Featured notebook – CogVideo
The code for the most powerful open source text-to-video model is now available. It requires a serious GPU, such as an A100, but is also available to try out with Hugging Face Spaces.
My attempt was... less successful:
Thanks for reading!
If you have anything you’d like to be featured or want to get in touch, give me a shout on Twitter or via email. Please also consider supporting me on Patreon so that I can spend more time creating content like this.