From ab9d3e50492b3e29f71ded9246e233065a80a36a Mon Sep 17 00:00:00 2001 From: HannahBenita <77296142+HannahBenita@users.noreply.github.com> Date: Fri, 1 Dec 2023 12:45:52 +0100 Subject: [PATCH] Update index.html --- index.html | 14 +++++++------- 1 file changed, 7 insertions(+), 7 deletions(-) diff --git a/index.html b/index.html index 80afda8..9f67a0e 100644 --- a/index.html +++ b/index.html @@ -246,7 +246,7 @@
Image Composition is a known limitation of Diffusion Models. Through evaluation of our new benchmark MCC-250 we show that multimodal prompting leads to more compositional robustness as judged by humans.
-Evaluating the alignment of prompt embeddings as well as generated images across multiple languagese we show that well aligned embeddings enable the transfer of multiligualism to downstream tasks even for task-specific monolingual training data.
@@ -259,7 +259,7 @@Attention Manipulation allows us to weight image and text tokens at inference time and guide their influence on the resulting generation.
-MultiFusion increases expressivness in composition through arbitrary and flexible promptin of image and text sequences.
-Negative prompting with images enables a more powerful supression than through text prompts.
-MultiFusion produces meaningful image variations without the need for inversion or renoising if the input image.
-