50b per token isn't very efficient... Wonder if we could make this 4: https://huggingface.co/inclusionAI/Ling-1T/blob/main/config.json#L22
Cody Cocks
AI & ML interests
Recent Activity
Organizations
We've released OpenSynth Battles, a benchmark dataset featuring generations from five large language models on shared prompts. Each prompt includes:
Responses from:
gpt-oss-120b, deepseek-v3.1-thinking, deepseek-v3.1-instruct, moonshotai/kimi-k2-instruct, and deepseek-r1-0528
Automated scoring by gpt-oss-120b
Useful for model comparison, automated evaluation research, and prompt-level performance analysis.
No data splits included.
🔗 https://huggingface.co/datasets/ccocks-deca/open-synth-battles
But here’s the flip side: once these tools start peeking into .env files, nodes, and source code, the risk of leaking critical keys or sensitive info becomes very real.
People who have such sensitive stuff should have the due diligence to pay for a model that fits their privacy needs.
Using some shady free API for building such stuff is a skill issue.
It was fine but got concatenated with the "the" :/ Fixed! Thanks!
Something big* coming
*big = biggest in the worldAnnnnnd... here it is! https://huggingface.co/deca-ai/3-alpha-ultra —the largest AI model in the world by
1. **No commercial use yet**: We're still working on Deca 2.5 (Proprietary), and releasing Deca 3 for commercial use right now would impact that. Once Deca 3.5 hits in early '26, we’ll be opening it up with a more permissive license.
2. **Built on existing models**: Deca 3 isn’t a ground-up creation—it’s a huge step forward, building on what’s already out there.
3. **It’s experimental**: As much as we’re hyped about its scale, it’s still in testing.
4. **DynaMoE architecture**: Run a (very) small part of the model with 64GB of RAM/VRAM (when quantized - quants coming soon), or the whole thing with 1TB. It’s that scalable.
5. **Not widely supported yet**: Frameworks like vLLM and Transformers aren’t compatible with Deca 3 at the moment, so until we drop the DynaMoE software (beta coming soon), it’s mostly just a concept.
We’re super excited to see what you do with it once the full setup’s ready. Hang tight, and stay tuned!
ETA is about 6-12 hours
Yes, why would it be on huggingface if it wasn't
This might be a better question for r/localllama
They are good.. For the price.. But if you can afford it, R1 or even Qwen3 are much better
It's not there yet. Sorry for the delay. We've had a mess with the vision (vit) parts. It can't even do simple OCR stuff. So it will be a while.. I would say August 10th-ish but really depends on how fast it works. Deca 2 (older version) is at deca-ai/2-mini and /2-pro
Yes! It has gone from 671B to 2.7T in slightly more than 2 weeks.
671B to 2.7T?!? I can barely run 235B models! (and that's with Q3 or Q2).
Regardless, as long as it performs well and does the job i suppose.
But i wonder if trimming the models or getting them to be more optimized in size vs performance shouldn't be a bigger push. Though i'm new to this scene so i could just be ignorant in how this is all done.
"Ultra" is the keywords here. We are going to release Pro and Mini models. Also because of DynaMoE architecture (publishing soon), you can run parts of the model on your setup. The downside is it isn't very well supported by many apps so everything will be using our code (until support improves).
We're on the verge of something monumental. Right now, we're in the final stages of testing, and we're about to drop a game-changing milestone in the open-source AI community. 🎉
In just two weeks, we've managed to almost 4x the size of the largest open-source LLM at that time (and we are still 2.6x bigger than the largest LLM). This is unprecedented and a testament to the power of collaboration, innovation, and the relentless pursuit of pushing AI to its limits.
The future of open-source AI is now. Stay tuned for the release – we’re just getting started.
- Model testing finishes: 24hrs from now
- Model gets uploaded: 30hrs from now
- Related code/inference stack gets published: 70-90hrs from now
Yes! It has gone from 671B to 2.7T in slightly more than 2 weeks. This Alpha variant is very crude and proof-of-concept so don't expect insane performance, but it is nice to see that open-source AI is catching up.
We're on the verge of something monumental. Right now, we're in the final stages of testing, and we're about to drop a game-changing milestone in the open-source AI community. 🎉
In just two weeks, we've managed to almost 4x the size of the largest open-source LLM at that time (and we are still 2.6x bigger than the largest LLM). This is unprecedented and a testament to the power of collaboration, innovation, and the relentless pursuit of pushing AI to its limits.
The future of open-source AI is now. Stay tuned for the release – we’re just getting started.
- Model testing finishes: 24hrs from now
- Model gets uploaded: 30hrs from now
- Related code/inference stack gets published: 70-90hrs from now
Please add Kimi k2!
Use a gguf?
is Mirage opensource?
You can do that here: https://huggingface.co/spaces/autotrain-projects/autotrain-advanced?duplicate=true
You will almost always need a GPU