WAON Collection WAON: Large-Scale and High-Quality Japanese Image-Text Pair Dataset for Vision-Language Models • 4 items • Updated Mar 2 • 3
Marco-MoE Collection A suit of multilingual MoE models with highly-sparse architectures • 5 items • Updated 7 days ago • 14
view article Article Welcome Gemma 4: Frontier multimodal intelligence on device +5 13 days ago • 846
Sarashina2.2 Collection Large Language Models developed by SB Intuitions. Pretrained and instruction-tuned models are available in three sizes: 0.5B, 1B, and 3B. • 6 items • Updated Mar 5, 2025 • 9
Constructing Synthetic Instruction Datasets for Improving Reasoning in Domain-Specific LLMs: A Case Study in the Japanese Financial Domain Paper • 2603.01353 • Published Mar 2 • 3
Nemotron-Cascade 2: Post-Training LLMs with Cascade RL and Multi-Domain On-Policy Distillation Paper • 2603.19220 • Published 27 days ago • 66
InternVL-U: Democratizing Unified Multimodal Models for Understanding, Reasoning, Generation and Editing Paper • 2603.09877 • Published Mar 10 • 48
DEJIMA: A Novel Large-scale Japanese Dataset for Image Captioning and Visual Question Answering Paper • 2512.00773 • Published Nov 30, 2025 • 1
SkyReels-V4: Multi-modal Video-Audio Generation, Inpainting and Editing model Paper • 2602.21818 • Published Feb 25 • 55
Qwen3.5 Collection Qwen3.5 is Qwen's new model family including Qwen3.5 Small: 0.8B, 2B, 4B, 9B and Qwen3.5 Medium: 35B-A3B, 27B, 122B-A10B and 397B-A17B. • 25 items • Updated 11 days ago • 146