Arcee Trinity Mini: US-Trained Moe Model

https://www.arcee.ai/blog/the-trinity-manifesto?src=hn

Comments

halJordanDec 2, 2025, 1:50 AM
Looks like a less good version of qwen 30b3a which makes sense bc it is slightly smaller. If they can keep that effiency going into the large one it'll be sick.

Trinity Large [will be] a 420B parameter model with 13B active parameters. Just perfect for a large Ram pool @ q4.

davidsainezDec 2, 2025, 4:46 AM
Excited to put this through its paces. It seems most directly comparable to GPT-OSS-20B. Comparing their numbers on the Together API: Trinity Mini is slightly less expensive ($0.045/$0.15 v $0.05/$0.20) and seems to have better latency and throughput numbers.
BalinaresDec 2, 2025, 2:09 PM
Interesting. Always glad to see more open weight models.

I do appreciate that they openly acknowledge the areas where they followed DeepSeek's research. I wouldn't consider that a given for a US company.

Anyone tried these as a coding model yet?

htrpDec 2, 2025, 1:49 AM
Trinity Nano Preview: 6B parameter MoE (1B active, ~800M non-embedding), 56 layers, 128 experts with 8 active per token

Trinity Mini: 26B parameter MoE (3B active), fully post-trained reasoning model

They did pretraining on their own and are still training the large version on 2048 B300 GPUs

bitwizeDec 2, 2025, 1:39 AM
A moe model you say? How kawaii is it? uwu
ghcDec 2, 2025, 2:13 AM
Capitalization makes a surprising amount of difference here...
donwDec 2, 2025, 3:07 AM
Meccha at present, but it may reach sugoi levels with fine-tuning.
noxaDec 2, 2025, 1:40 AM
I hate that I laughed at this. Thanks ;)
ksynwaDec 2, 2025, 4:37 AM
> Trinity Large is currently training on 2048 B300 GPUs and will arrive in January 2026.

How long does the training take?

arthurcolleDec 2, 2025, 6:10 AM
Couple days or weeks usually. No one is doing 9 month training runs
trvzDec 2, 2025, 6:20 AM
Moe ≠ MoE
cachiusDec 2, 2025, 6:31 AM
?
azinman2Dec 2, 2025, 6:42 AM
The HN title uses incorrect capitalization.
rbanffyDec 2, 2025, 7:56 AM
I was eagerly waiting for the Larry and Curly models.
m4rtinkDec 2, 2025, 9:21 AM
^_-