Paper Link: arxiv.org/abs/2510.27492
Paper Link: arxiv.org/abs/2510.27492
ThinkMorph shows:
1. Unseen visual manipulations (zoom, crop, inpaint)
2. Autonomous mode switching (text-only when optimal)
3. Enhanced test-time scaling via diverse multimodal trajectories
ThinkMorph shows:
1. Unseen visual manipulations (zoom, crop, inpaint)
2. Autonomous mode switching (text-only when optimal)
3. Enhanced test-time scaling via diverse multimodal trajectories
Paper: “INT vs FP: A Comprehensive Study of Fine-Grained Low-bit Quantization Formats”
Link: arxiv.org/pdf/2510.25602
Paper: “INT vs FP: A Comprehensive Study of Fine-Grained Low-bit Quantization Formats”
Link: arxiv.org/pdf/2510.25602
MXINT8 outperformed MXFP8 in every test:
Higher average QSNR (40.35 dB vs 31.50 dB)
Nearly lossless training performance
100% win rate in tensor-wise comparisons
MXINT8 outperformed MXFP8 in every test:
Higher average QSNR (40.35 dB vs 31.50 dB)
Nearly lossless training performance
100% win rate in tensor-wise comparisons
1. Tensor-wise analysis on Llama-3.1-8B
2. Direct-cast inference on 12 LLMs (0.6B–235B)
3. Low-bit training on 1B & 3B models
1. Tensor-wise analysis on Llama-3.1-8B
2. Direct-cast inference on 12 LLMs (0.6B–235B)
3. Low-bit training on 1B & 3B models
INT QSNR drops with high κ, but improves with finer granularity. FP QSNR depends on mantissa width and subnormal values.
INT QSNR drops with high κ, but improves with finer granularity. FP QSNR depends on mantissa width and subnormal values.