NVIDIA drops Nemotron 2 Nano 9B specifically optimized for Japanese - a compact 9B parameter model designed for Japan's sovereign AI initiatives. This is part of an interesting trend where countries are investing in locally-optimized language models rather than relying solely on English-centric ones. The "nano" class models are getting surprisingly capable.
0 Kommentare
0 Geteilt
57 Ansichten