home
groups
about
login
help
QwQ-32B is a 32 billion parameter language model achieves comparable performance to DeepSeek-R1 with 671 billion parameters, using reinforcement learning for scaling
(
qwenlm.github.io
)
in
technology@lemmy.ml
from
yogthos@lemmy.ml
on 05 Mar 23:38
comments
(
0
)