# Code Generation Optimization

Areal Boba 2 8B
Apache-2.0
AReaL is an asynchronous reinforcement learning training system developed by Ant Group, designed specifically for large inference models, supporting fast training and cutting-edge performance.
Large Language Model Transformers
A
inclusionAI
1,294
2
Acereason Nemotron 7B
Other
A math and code reasoning model trained through reinforcement learning, based on DeepSeek-R1-Distilled-Qwen-7B, excelling in mathematical and code reasoning tasks
Large Language Model Transformers
A
nvidia
4,278
10
SWE Agent LM 32B GGUF
Apache-2.0
SWE-agent-LM-32B is an open-source software engineering language model, fine-tuned based on Qwen/Qwen2.5-Coder-32B-Instruct, and is specifically designed for software engineering tasks.
Large Language Model Transformers English
S
Mungert
2,933
1
Falcon H1 34B Instruct GPTQ Int8
Other
Falcon-H1 is a high-performance hybrid architecture language model developed by TII, combining the strengths of Transformers and Mamba architectures, supporting English and multilingual tasks.
Large Language Model Transformers
F
tiiuae
105
3
Falcon H1 1.5B Base
Other
Falcon-H1 is a decoder-only causal model with a hybrid Transformers + Mamba architecture developed by TII, supporting English and multilingual tasks.
Large Language Model Transformers Supports Multiple Languages
F
tiiuae
454
2
GLM 4 9B 0414 GGUF
MIT
GLM-4-9B-0414 is a lightweight member of the GLM family with 9 billion parameters, excelling in mathematical reasoning and general tasks, providing an efficient solution for resource-constrained scenarios.
Large Language Model Supports Multiple Languages
G
unsloth
4,291
9
Granite 4.0 Tiny Preview
Apache-2.0
Granite-4-Tiny-Preview is a fine-grained Mixture of Experts (MoE) instruction-tuned model with 7 billion parameters, developed based on Granite-4.0-Tiny-Base-Preview, suitable for general instruction-following tasks.
Large Language Model Transformers
G
ibm-granite
7,906
108
Qwen2.5 Recursive Coder 14B Instruct
Apache-2.0
A 14B-parameter code generation and comprehension model based on the Qwen2.5 architecture, integrated through the Model Stock method by combining multiple specialized coding models
Large Language Model Transformers
Q
spacematt
39
2
Qwen2.5 CompositeFlow Coder 14B Instruct
Apache-2.0
A hybrid model based on Qwen2.5-Coder-14B-Instruct, merged using the mergekit tool with multiple specialized coding models
Large Language Model Transformers
Q
spacematt
31
3
Burtenshaw GemmaCoder3 12B GGUF
A quantized version based on burtenshaw/GemmaCoder3-12B, optimized for code generation tasks, supporting multiple quantization levels to adapt to different hardware requirements.
Large Language Model
B
bartowski
9,066
8
Z1 7B
MIT
Z1 is a large language model based on Qwen2.5-Coder-7B-Instruct, focusing on efficient reasoning through thought migration.
Large Language Model Transformers
Z
efficientscaling
125
18
Qwen2.5 Coder 32B YOYO
Qwen2.5-Coder-32B-instruct is an instruction fine-tuned version based on Qwen2.5-Coder-32B, specializing in code generation and understanding tasks.
Large Language Model Transformers
Q
YOYO-AI
14
2
YOYO O1 32B V2
Combines the top 32B reasoning and coding models from the open-source community, using the SCE merging technique with Qwen2.5-Coder-32B as the base model for fusion.
Large Language Model Transformers
Y
YOYO-AI
29
2
Qwq Coder 32B
A 32B-parameter large language model merged from the Qwen series, specializing in code generation and comprehension tasks
Large Language Model Transformers
Q
YOYO-AI
490
7
Gemma 3 4b It Codeforces SFT
This model is a fine-tuned version of google/gemma-3-4b-it on the codeforces-cots dataset, primarily used for code-related tasks.
Large Language Model Transformers
G
qgallouedec
48
3
Opencerebrum 1.0 7b SFT GGUF
Apache-2.0
OpenCerebrum-1.0-7b-SFT is a statically quantized model based on Locutusque/OpenCerebrum-1.0-7b-SFT, supporting multiple quantization versions, suitable for text generation and Q&A tasks in fields such as code generation, mathematics, chemistry, and biology.
Large Language Model Supports Multiple Languages
O
mradermacher
162
1
Qwen2.5 Coder 7B Instruct Uncensored
MIT
An optimized version based on Qwen2.5-Coder-7B-Instruct, focused on eliminating refusal behaviors
Large Language Model Transformers Supports Multiple Languages
Q
BlossomsAI
23
1
Qwen2.5 Coder 1.5B Instruct Abliterated GGUF
Apache-2.0
Llamacpp imatrix quantized version based on the Qwen2.5-Coder-1.5B-Instruct-abliterated model, suitable for code generation and text generation tasks.
Large Language Model English
Q
bartowski
729
1
Granite 3.0 8b Instruct
Apache-2.0
Granite-3.0-8B-Instruct is an 8-billion-parameter model fine-tuned from Granite-3.0-8B-Base, trained using a combination of various open-source instruction datasets and internally synthesized datasets.
Large Language Model Transformers
G
ibm-granite
24.29k
201
Powerlm 3b
Apache-2.0
PowerLM-3B is a small language model with 3 billion parameters, trained using the Power learning rate scheduler, and demonstrates excellent performance across multiple benchmarks including natural language multiple-choice, code generation, and mathematical reasoning.
Large Language Model Transformers
P
ibm-research
11.07k
20
API Pack Model
A large language model fine-tuned based on the CodeLlama-13b-hf architecture, specifically optimized for API call generation tasks
Large Language Model Transformers
A
apipack
16
1
Nxcode CQ 7B Orpo
Other
A code generation model fine-tuned on 100,000 high-quality ranked data points using reference-free monolithic preference optimization, based on the Qwen/CodeQwen1.5-7B model
Large Language Model Transformers Other
N
NTQAI
3,387
130
Mixtral AI CyberCoder 7b
Apache-2.0
This is a merged model focused on code generation and application development, created by fusing multiple high-quality models
Large Language Model Transformers Other
M
LeroyDyer
46
4
Opencodeinterpreter DS 33B GGUF
OpenCodeInterpreter is an open-source code generation system that enhances code generation capabilities by integrating code execution and iterative optimization functions
Large Language Model Supports Multiple Languages
O
LoneStriker
278
28
Deepmagic Coder 7b Alt
Other
DeepMagic-Coder-7b is a merged version of the DeepSeek-Coder and Magicoder models, focusing on code generation and programming tasks.
Large Language Model Transformers
D
rombodawg
61
7
Tinymistral 6x248M
Apache-2.0
TinyMistral-6x248M is a Mixture of Experts system that integrates 6 TinyMistral variants using the LazyMergekit tool, pre-trained on the nampdn-ai/mini-peS2o dataset
Large Language Model Transformers
T
M4-ai
51
14
Everyone Coder 33b Base
Other
The EveryoneLLM series models are created by the community and serve the community. This is a specialized coding model fine-tuned based on deepseek-coder-33b-base.
Large Language Model Transformers
E
rombodawg
15
19
Phixtral 2x2 8
MIT
phixtral-2x2_8 is the first Mixture of Experts (MoE) model built upon two microsoft/phi-2 models, outperforming each individual expert model.
Large Language Model Transformers Supports Multiple Languages
P
mlabonne
178
148
Airoboros M 7b 3.1.2
Apache-2.0
Instruction-tuned model based on Mistral-7B, focusing on following complex instructions rather than daily chat, using llama-2 chat format
Large Language Model Transformers
A
jondurbin
127
48
Codellama 13b Oasst Sft V10
A version fine-tuned by Open-Assistant based on Meta's CodeLlama 13B large language model, supporting English, with a new RoPE Theta value (1e6 instead of 1e4).
Large Language Model Transformers English
C
OpenAssistant
159
69
Phind CodeLlama 34B Python V1
A large language model fine-tuned based on CodeLlama-34B-Python, achieving 69.5% pass@1 on HumanEval, surpassing GPT-4's performance
Large Language Model Transformers
P
Phind
878
253
Gpt Neo 125M Code Clippy Dedup 2048
A 125M-parameter code generation model based on the GPT-Neo architecture, fine-tuned on the CodeClippy dataset, supporting sequence generation up to 2048 tokens
Large Language Model
G
flax-community
1,528
14
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase