HanDiffuser proposes a diffusion-based architecture to generate images with realistic hands by injecting hand embeddings in the generative process. The model consists of two components: Text-to-Hand-Params and Text-Guided Hand-Params-to-Image, showcasing efficacy in generating high-quality hands.
CogView3 introduces relay diffusion to improve text-to-image generation efficiency and quality, outperforming existing models.
B2B module improves spatial control and attribute binding in T2I models.
HanDiffuserは、テキストからリアルな手の外観を生成します。
PromptCharm supports novice users in creating high-quality images by facilitating multi-modal prompt engineering and refinement.
Predicated Diffusion effectively addresses common challenges in text-based image generation, offering superior quality and fidelity compared to existing methods.
Improving the trade-off between textual and spatial grounding in image generation models through network rewiring.
Predicated Diffusion effectively addresses common challenges in text-based image generation, offering superior quality and fidelity to prompts.
SwiftBrush introduces an innovative image-free distillation scheme for one-step text-to-image generation, achieving high-quality results without the need for training image data.
Isolated Diffusion optimizes multi-concept text-to-image synthesis by isolating the synthesizing processes of different concepts.