Artificial intelligence just got a new player—and it’s fully open-sourced. Aria, a multimodal LLM developed by Tokyo-based Rhymes AI, is capable of processing text, code, images, and video all within a single architecture.

What should catch your attention, though, isn’t just its versatility, but its efficiency. It’s not a huge model like its multimodal counterparts, which means it is more energy—and hardware—friendly.

Rhymes AI achieved this by employing a Mixture-of-Experts (MoE) framework. This architecture is similar to having a team of specialized mini experts, each trained to excel in specific areas or tasks.

When a new input is given to the model, only the relevant experts (or a subset) are activated instead of using the entire model. This way, running just a specific section of the model means it will be lighter than running a complete know-it-all entity that tries to process everything.

This makes Aria more efficient because, unlike traditional models that activate all parameters for every task, Aria selectively engages just 3.5 billion of its 24.9 billion parameters per token, reducing computational load and improving performance on specific tasks.

It als

Go to Source to See Full Article
Author: Jose Antonio Lanz

BTC NewswireAuthor posts

BTC Newswire Crypto News at your Fingertips

Comments are disabled.