Magic-Wan-T2IV V3.0

模型在 v2.0 的基础上,有以下几方面的提升 (Has the following improvements:):

  1. 更逼真的实物景致与人像肌肤,更好的提示词遵循能力。(Realistic scenes and skin, better prompt compliance)
  2. 更好的美学构图与人脸优化,对 NSFW 支持能力更好。(Aesthetic composition and facial optimization, better support for NSFW)
  3. 文生图与文生视频统一模型,支持较短长度的文生视频能力。(Unified for T2I and T2V, support short-length T2V capabilities)
  4. 支持英文及少量中文字体的渲染。(English and a limited of Chinese fonts Rendering)

本模型依旧为纯净底模,可用于微调或 LoRA 训练,也可叠加各种 Turbo / Distill / NSFW / SFW 的 LoRA 使用。

Pure base model, suitable for fine-tuning or LoRA training, and can also be combined with various Turbo / Distill / SFW / NSFW LoRAs.

感谢:@亮亮rayne@GuangyuanSD@kanttouchthis@QuantFunc 在本模型调制优化过程中给与技术、构图、美学、量化等方面的指导和支持。

感谢:@T8star-Aix@AI-KSK@黑鹤001@大麦茶看AI@joeMultimedia 等一众大佬对本模型的技术评测、使用介绍、社区推荐及进一步微调提升等方面的建议和支持。

基础工作流请参见(Base Workflow):V3-T2Image-Workflow.json, V3-T2Video-Workflow.json

Also on: Modelscope, Civitai

===================================================================================

Magic-Wan-Image V2.0

逼真肌肤、极致高清,适合人像与实景照片类图像的生成。细节丰富,支持 TTP 直放到 8M 像素。本模型为纯净底模,用户可根据个人喜好组合添加各种 NSFW / SFW 的 LoRA 模型。

Highly realistic skin texture and ultra-high resolution, suitable for generating Portrait and Real-world scenes. rich details and supports direct upscale to 8M pixels using TTP technology. This is pure base model, you can add all kind of NSFW / SFW LoRAs according to you like.

基础工作流请参见:V2-example-workflow-image.png

Base workflow pls refer to: V2-example-workflow-image.png

Also on: Modelscope, Civitai

===================================================================================

Magic-Wan-Image V1.0

本模型是一个实验模型,是 Wan2.2-T2V-14B 文生视频模型的混调版本,目的是能让广大 Wan 2.2 模型的爱好者,能像使用 Flux 一样,简单方便的使用 Wan2.2 T2V 模型来生成各种图片。Wan 2.2 模型擅长写实类图像的生成,同时兼顾多种风格,由于是视频模型演化而来,模型在生图能力的泛化性方面稍微弱一点。本模型最大的兼顾了模型的写实能力和风格变化,同时尽可能的体现更多的细节,基本上达到与 Flux.1-Dev 模型相当的创作力和表现力。模型的混调方法是将 Wan2.2-T2V-14B 模型的 High-Noise 和 Low-Noise 两部分分层进行不同权重比例的混合,再简单微调而成。目前是一个实验模型,可能还存在一些不足之处,欢迎大家试用并反馈信息,以便在未来版本改进。

This model is an experimental model. A mixed and finetuned version of the Wan2.2-T2V-14B text-to-video model, Let many enthusiasts of the Wan 2.2 model to easily use the Wan2.2 T2V model to generate various images, similar to use the Flux model. The Wan 2.2 model excels at generating realistic images while also accommodating various styles. However, since it evolved from a video model, its generative capabilities for raw images are slightly weaker. This model balances the realistic capabilities and style variations while striving to include more details, essentially achieving creativity and expressiveness comparable to the Flux.1-Dev model. The mixing method used for this model involves layering the High-Noise and Low-Noise parts of the Wan2.2-T2V-14B model and blending them with different weight ratios, followed by simple fine-tuning. Currently, it is an experimental model that may still have some shortcomings, and we welcome everyone to try it out and provide feedback for improvements in future versions.

使用方法:

请参见例图工作流:V2-example-workflow-image.png

model shift: 1.0 - 8.0,您可以自己尝试;

model cfg: 1.0 -4.0,您可以自己尝试;

inference steps: 20 -50, 您可以自己尝试;

sample / schduler: deis/simple 或 euler/beta,或任意组合,您可以自己尝试。

Also on: Modelscope, Civitai

The model under Apache 2.0.

Downloads last month
931
GGUF
Model size
14B params
Architecture
wan
Hardware compatibility
Log In to add your hardware

4-bit

5-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for wikeeyang/Magic-Wan-T2IV-V3

Finetuned
(35)
this model
Adapters
5 models