Highlights
Dragon LLM introduces its new LLM Architecture. Built on a new hybrid GDN -Transformer that outperforms traditional architectures, it can power frugal, sovereign models that can be rapidly specialized on business data and use cases.
Dragon Architecture features :
- Very strong ability to remember past words in the sequence compared to other Hybrid approach, inspired by Hymba (NVIDIA)
- Ability to be used simultaneously by more users on equivalent hardware and better throughput on long-context scenario
- Extremely efficient learning It has been been validated at large scale by the training of a 3B model on 3.5T tokens. It achieves comparable performance against smolLM-3B-Base and Qwen3-4B-Base on ARC, HellaSwag, LAMBADA, and PIQA, while trained on 3-5 time less data.
Why is this important?
- Proves performance β same performance with 3β5Γ less data.
- Cut cost : more users can be served on the same hardware
- Ability to deploy in secure environment with constraint on the hardware (even on CPU)
- Scales better : higher throughput and strong long-context handling (Long documents, files, codes or contracts).
How has Dragon LLM achieved this? β’ By combining the best recent research papers on LLM architectures, cumulating gains across all processes, from deep layer optimization to attention head or kv cache management. β’ Agile Team able to adapt quickly and test new ideas extremely fast β’ Compute support by the EU Commission (euroHPC - JUPITER and Leonardo HPC)
What's next? The next step is to deliver foundation models for this architecture : β’ a 3B and 7B version of DragonBase trained on 10T+ tokens β’ Chat version of these models β’ Specialized versions for specific industry vertical such as Finance
If you want to know more and get updates on the project, follow us !
If you would like a comprehensive deep dive on the architecture : read our blog post
Model Overview
Model Benchmark
| Benchmarks | Dragon | Qwen3-4B | SmolLM3 |
|---|---|---|---|
| ARC Challenge | 50% | 51.28% | 52.56% |
| ARC Easy | 76.01% | 75.97% | 76.81% |
| HellaSwag | 71.73% | 54.46% | 75.2% |
| LAMBADA | 65.03% | 62.62% | 65.05% |
| PIQA | 79.11% | 77.86% | 78.84% |
| SWDE | 89.92% | 91.99% | 88.03% |
| FDA | 81.13% | 86.75% | 76.13% |
| Average | 73.27% | 71.56% | 73.23% |
All evaluations are performed using with lm-eval and few shot set to 0.
Limitations
This model is a foundation model, trained on large-scale general-purpose text corpora. It has not been fine-tuned for any specific downstream task. As such:
It may produce inaccurate or misleading information, particularly for factual or time-sensitive queries.
It has no understanding of truth or intent and may generate biased, toxic, or harmful content inherited from its training data.
It is not suitable for direct use in safety-critical or decision-making contexts (e.g., healthcare, finance, law) without additional alignment or validation.
The model does not perform well on tasks requiring domain-specific expertise, numerical precision, or structured reasoning unless further fine-tuned.
Long or complex prompts may lead to loss of coherence or hallucinations as context length grows.
Fine-tuning, prompt-engineering, or evaluation on downstream tasks is recommended before any production use.
Quickstart
Try it with:
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "DragonLLM/Dragon-3B-Base-alpha"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
dtype="auto",
device_map="auto",
trust_remote_code=True,
)
prompt = "Once upon a time, a valiant knight named Segurant set out on a quest to chase a dragon. He was"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
generated_ids = model.generate(
**inputs,
max_new_tokens=512,
)
print(tokenizer.decode(generated_ids[0], skip_special_tokens=True))
Setup
For better performance on GPU, we recommend using :
- flash-linear-attention: the Gated DeltaNet Triton kernels
Install withpip install flash-linear-attention
If you use NVIDIA GPU, you can improve performance with :
flash-attention:
Install withpip install flash-attn --no-build-isolationcausal-conv1d: a short convolution is used as part of the Gated DeltaNet layer
Install withpip install causal-conv1d(optional, recommended only for A100) flex-head-ha: computing attention with different head dimensions for qk and vo, used for differential attention
Install withpip install flex-head-fa --no-build-isolation
- Downloads last month
- 311
Dataset used to train DragonLLM/Dragon-3B-Base-alpha
Space using DragonLLM/Dragon-3B-Base-alpha 1
Evaluation results
- Test accuracy on AI2 ARC (Challenge)test set self-reported50.000
- Test accuracy on AI2 ARC (Easy)test set self-reported76.010
- Acc on HellaSwagvalidation set self-reported71.730
- Acc on LAMBADAtest set self-reported65.030
- Acc on PIQAvalidation set self-reported79.110
- Acc on SWDEtest set self-reported89.920
- Acc on FDAtest set self-reported81.130
