Illustrious, a text-to-image model based on Stable Diffusion XL, has become so dominant in the AI art community that Civitai, the largest hub for AI art models, had to create a separate category just to handle its massive ecosystem of resources.
And it all happened in three months. The secret behind its success? A return to the basics with a twist.
While newer models like SD 3.5 and Flux rely on lengthy natural language descriptions, Onoma AI, the developers of Illustrious, took a different approach by leveraging Danbooru tags to help their model understand concepts without having to reinvent the wheel with complex captioning systems.
The model’s training on Danbooru’s vast library of tagged anime images gives it an edge in understanding visual concepts.
Each tag in the Danbooru system represents specific elements like character features, clothing items, poses, or backgrounds, allowing for precise control over the generated images without wasting precious tokens on lengthy descriptions.
These tags have been
Go to Source to See Full Article
Author: Jose Antonio Lanz
