Composable-Diffusion approach released* Pixel Art Diffusion released* Latent Diffusion fine-tuning code* DiVAE paper released Text2Human released*
Explainability-Aided Image Classification and Generation released*
Multi-lingual CLIP* LAION OpenCLIP ViT-L/14* UniCL - unified image, caption and labels* CyCLIP - geometrically consistent CLIP*
* code released
Compositional Visual Generation with Composable Diffusion Models
— AK (@ak92501) June 6, 2022
abs: https://t.co/FEKYaDOlwf
project page: https://t.co/7NrvhKYJxt pic.twitter.com/09xRm0WhBi
#pixelartdiffusion pic.twitter.com/jLBg6jhjGi
— multimodal ai art (@multimodalart) June 3, 2022
Checking now @laion_ai's Latent Diffusion fine tuning code and two fine-tuned models https://t.co/3n8O5Ylj3G
— multimodal ai art (@multimodalart) June 7, 2022
ongo: fine-tuned for paintings (left) https://t.co/iO0ef21xp8
erlich: fine-tuned for generating logos (right) https://t.co/0l0Qy6mqup pic.twitter.com/ubpMXj7qiM
DiVAE : Photorealistic Images Synthesis with Denoising Diffusion Decoder
— AK (@ak92501) June 2, 2022
abs: https://t.co/vdgSJ42qpa
Training on ImageNet 256×256, model achieves state-of-the-art results and generates more photorealistic images pic.twitter.com/O61kTlXahq
Text2Human: Text-Driven Controllable Human Image Generation
— AK (@ak92501) June 1, 2022
abs: https://t.co/jfYdGa50h0
project page: https://t.co/YzAGqooe7D
github: https://t.co/VfyUXN1uoO pic.twitter.com/r2Cgo46EAN
Apple released a repo with three Colab notebooks:
— Doron Adler (@Norod78) June 6, 2022
1. Notebook for spatially conditioned image generation (uses VQGAN ImageNet)
2. Notebook for image editing (Based on StyleCLIP by @OPatashnik )
3. Notebook for image generation (Uses BigGAN)https://t.co/dUZ2sTZ28Z
@AccountForAI and I trained a better multilingual encoder aligned with openai clip vit-l/14 image encoder. https://t.co/xTgpUUWG9Z 1/6 pic.twitter.com/ag1SfCeJJj
— Romain Beaumont (@rom1504) June 2, 2022
Unified Contrastive Learning in Image-Text-Label Space
— AK (@ak92501) April 8, 2022
abs: https://t.co/nQxVKfYI4c
github: https://t.co/9hjKpBuh6i pic.twitter.com/AAF9gOQO2Y
A new approach for CLIP to make the embeddings geometrically consistent. Code and models released. Same as with UniCL, it looks promising but pre-trained models are currently way smaller than OpenAI or OpenCLIP to asses the approach improvements scale. Also worth keeping an eye on (hey that's what we're here for).
CLIP puts image and text representation in the same space but what if we enforce that this representations are geometrically consistent?
— multimodal ai art (@multimodalart) May 31, 2022
CyCLIP does that - and the paper finds initial improvements from CLIP, I'm curious to see how it would perform guiding image generation! https://t.co/V5X2FRFNUH
AIArt is a free and open source AI art course by John Whitaker. There are synchronous classes for the next few Saturdays 4 PM UTC on Twitch. All previous classes stay recorded and available on Google Colabs on the GitHub link