Granite-4.0-H-Small: a 32B-A9B MoE Mamba for high efficency
Damn! IBM is on the map. The American Qwen? I barely even knew IBM made LLMs, this is solid
www.ibm.com/new/announce...
Granite-4.0-H-Small: a 32B-A9B MoE Mamba for high efficency
View original thread
31
2
initial take: okay, it can work as a model. it doesn't have a fun personality, but that's kinda what you get with a 32b-a9b. fwiw it doesn't hallucinate too badly afaict
could be worth looking into if you're thinking about RL'ing LoRAs, it's probably malleable
could be worth looking into if you're thinking about RL'ing LoRAs, it's probably malleable
2
also: Mambas aren’t transformers, so this isn’t technically an LLM, whatever..
1