Ayase Models

Pre-downloaded model weights for Ayase modules that require non-HuggingFace downloads.

These models are hosted here because their original CDNs (dl.fbaipublicfiles.com, openaipublic.azureedge.net, GitHub releases) can be unreliable or blocked on some servers.

Structure

ayase-models/
β”œβ”€β”€ dover/                    # DOVER video quality assessment (ICCV 2023)
β”‚   β”œβ”€β”€ DOVER.pth             # 229 MB β€” S-Lab License 1.0
β”‚   └── convnext_tiny_1k_224_ema.pth  # 110 MB β€” MIT
β”œβ”€β”€ i2v_similarity/           # Image-to-Video similarity metrics
β”‚   β”œβ”€β”€ ViT-B-32.safetensors  # 338 MB β€” MIT (OpenAI CLIP)
β”‚   β”œβ”€β”€ dinov2_vitb14_pretrain.pth  # 331 MB β€” Apache 2.0 (Meta)
β”‚   └── alex.pth              # 6 KB β€” BSD-2 (LPIPS)
β”œβ”€β”€ advanced_flow/            # RAFT optical flow (ECCV 2020)
β”‚   β”œβ”€β”€ raft_large_C_T_SKHT_V2-ff5fadd5.pth  # 21 MB β€” BSD-3
β”‚   └── raft_small_C_T_V2-01064c6d.pth       # 3.9 MB β€” BSD-3
β”œβ”€β”€ fast_vqa/                 # FAST-VQA quality assessment (ECCV 2022)
β”‚   β”œβ”€β”€ FAST_VQA_3D_1_1.pth   # 121 MB β€” MIT
β”‚   β”œβ”€β”€ FAST_VQA_B_1_4.pth    # 121 MB β€” MIT
β”‚   └── FAST_VQA_M_1_4.pth    # 105 MB β€” MIT
β”œβ”€β”€ aesthetic_scoring/        # LAION aesthetic predictor
β”‚   └── sac+logos+ava1-l14-linearMSE.pth  # 3.5 MB β€” MIT
β”œβ”€β”€ video_memorability/       # Video memorability estimation
β”‚   └── dinov2_vits14_pretrain.pth  # 84 MB β€” Apache 2.0 (Meta)
β”œβ”€β”€ spectral/                 # Spectral analysis
β”‚   └── dinov2_vits14_pretrain.pth  # 84 MB β€” Apache 2.0 (Meta)
β”œβ”€β”€ trajan/                   # Point tracking (CoTracker2)
β”‚   └── cotracker2.pth        # 194 MB β€” Apache 2.0 (Meta)
β”œβ”€β”€ depth_map_quality/        # Monocular depth estimation
β”‚   β”œβ”€β”€ dpt_swin2_tiny_256.pt # 164 MB β€” MIT (Intel ISL MiDaS)
β”‚   └── midas_v21_small_256.pt # 82 MB β€” MIT (Intel ISL MiDaS)
β”œβ”€β”€ depth_consistency/        # Temporal depth consistency
β”‚   β”œβ”€β”€ dpt_swin2_tiny_256.pt # 164 MB β€” MIT
β”‚   └── midas_v21_small_256.pt # 82 MB β€” MIT
β”œβ”€β”€ motion_smoothness/        # RIFE motion smoothness
β”‚   └── flownet.pkl           # Motion interpolation network
β”œβ”€β”€ brightvq/                 # BrightRate / BrightVQ HDR no-reference quality
β”‚   β”œβ”€β”€ brightrate_brightvq.pt             # 161 MB β€” BrightRate regressor
β”‚   β”œβ”€β”€ CONTRIQUE_checkpoint25.tar         # 107 MB β€” CONTRIQUE feature extractor
β”‚   β”œβ”€β”€ frames_modelparameters.mat         # 8 KB β€” NIQE/HDR stats params
β”‚   β”œβ”€β”€ ViT-B-32.safetensors              # 338 MB β€” MIT (OpenAI CLIP)
β”‚   β”œβ”€β”€ ViT-L-14.safetensors              # 890 MB β€” MIT (OpenAI CLIP)
β”‚   β”œβ”€β”€ CLIP-IQA+_learned_prompts-603f3273.pth # 17 KB β€” CLIP-IQA+ prompts
β”‚   β”œβ”€β”€ CLIPIQA+_RN50_512-89f5d940.pth     # 309 KB β€” CLIP-IQA+ RN50 weights
β”‚   └── CLIPIQA+_ViTL14_512-e66488f2.pth   # 463 KB β€” CLIP-IQA+ ViT-L/14 weights
β”œβ”€β”€ rqvqa/                    # RQ-VQA rich quality-aware VQA (CVPR 2024 NTIRE)
β”‚   β”œβ”€β”€ LIQE.pt               # 337 MB β€” LIQE feature extractor
β”‚   └── Swin_b_384_in22k_SlowFast_Fast_LSVQ.pth  # 345 MB β€” Swin-B + SlowFast backbone
└── song_eval/                # SongEval song aesthetic evaluation
    └── model.safetensors     # 96 MB β€” Apache 2.0 (ASLP-lab)

Total: ~4.6 GB

Models hosted on HuggingFace Hub (not included here)

These models are downloaded directly via transformers / open_clip and work without re-hosting:

Module HF Model License
semantic_alignment openai/clip-vit-base-patch32 MIT
clip_temporal openai/clip-vit-base-patch32 MIT
captioning Salesforce/blip-image-captioning-base BSD-3
sd_reference stabilityai/stable-diffusion-xl-base-1.0 CreativeML Open RAIL++-M
action_recognition MCG-NJU/videomae-large-finetuned-kinetics CC-BY-NC-4.0
ocr_fidelity PaddleOCR (self-managed) Apache 2.0

License

Each subdirectory contains a LICENSE.md with attribution for the respective model weights. All models are redistributed under their original open-source licenses.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support