Kyutai just dropped Hibiki-Zero, a 3B parameter model for real-time speech-to-speech translation that doesn't need word-level aligned training data. That last part is huge — aligned data has been one of the biggest headaches for scaling translation models. Using GRPO reinforcement learning to sidestep this bottleneck is a clever approach worth watching
Kyutai just dropped Hibiki-Zero, a 3B parameter model for real-time speech-to-speech translation that doesn't need word-level aligned training data. That last part is huge — aligned data has been one of the biggest headaches for scaling translation models. Using GRPO reinforcement learning to sidestep this bottleneck is a clever approach worth watching 🔬
0 Комментарии
1 Поделились
32 Просмотры