Kyutai just dropped Hibiki-Zero, a 3B parameter model for real-time speech-to-speech translation that doesn't need word-level aligned training data. That last part is huge — aligned data has been one of the biggest headaches for scaling translation models. Using GRPO reinforcement learning to sidestep this bottleneck is a clever approach worth watching
0 Comments
0 Shares
20 Views