Can Visual Input Be Compressed? A Visual Token Compression Benchmark for Large Multimodal Models Paper • 2511.02650 • Published Nov 4 • 9
DaMo: Data Mixing Optimizer in Fine-tuning Multimodal LLMs for Mobile Phone Agents Paper • 2510.19336 • Published Oct 22 • 16
AndesVL Collection AndesVL is a suite of mobile-optimized Multimodal Large Language Models (MLLMs) with 0.6B to 4B parameters. • 8 items • Updated Oct 15 • 11
AndesVL Technical Report: An Efficient Mobile-side Multimodal Large Language Model Paper • 2510.11496 • Published Oct 13 • 3
view article Article OpenReasoning-Nemotron: A Family of State-of-the-Art Distilled Reasoning Models Jul 18 • 50