đ ZeroXClem-Qwen3-4B-NexusPrime
ZeroXClem-Qwen3-4B-NexusPrime is a high - performance, multi - domain AI model. It's built by merging models using Model Stock with MergeKit. This model combines several finely - tuned Qwen3 - 4B models, offering excellent reasoning, coding, and multi - step problem - solving capabilities. It's optimized for structured outputs and technical applications.
â
This model works best with the default Qwen3 chat template. See the ollama modelcard in the Usage Section for details.
đĻ Model Information
Property |
Details |
Tags |
merge, mergekit, lazymergekit |
Language |
en |
Base Model |
prithivMLmods/Cetus - Qwen3_4B - GeneralThought, prithivMLmods/Tureis - Qwen3_QWQ - 4B - Exp, prithivMLmods/Canum - Qwen3_R1 - 4B - iCoT, prithivMLmods/Bootes - Qwen3_Coder - Reasoning, prithivMLmods/Segue - Qwen3_DeepScaleR - Preview |
Pipeline Tag |
text - generation |
Library Name |
transformers |
đ§ Merge Configuration
- Merge Method:
model_stock
- Base Model: [
prithivMLmods/Cetus - Qwen3_4B - GeneralThought
](https://huggingface.co/prithivMLmods/Cetus - Qwen3_4B - GeneralThought)
- Dtype:
bfloat16
- Tokenizer Source:
prithivMLmods/Cetus - Qwen3_4B - GeneralThought
đ Configuration File
name: ZeroXClem - Qwen3 - 4B - NexusPrime
base_model: prithivMLmods/Cetus - Qwen3_4B - GeneralThought
dtype: bfloat16
merge_method: model_stock
models:
- model: prithivMLmods/Tureis - Qwen3_QWQ - 4B - Exp
- model: prithivMLmods/Canum - Qwen3_R1 - 4B - iCoT
- model: prithivMLmods/Bootes - Qwen3_Coder - Reasoning
- model: prithivMLmods/Segue - Qwen3_DeepScaleR - Preview
tokenizer_source: prithivMLmods/Cetus - Qwen3_4B - GeneralThought
đĄ Models Merged
The following models contribute to this fusion, each with unique strengths:
Tureis - Qwen3_QWQ - 4B - Exp
- Precision Reasoning: Fine - tuned for high - fidelity symbolic reasoning, step - by - step math, and logic tasks.
- Lightweight Code Understanding: Efficiently processes Python, C++, and other languages for concise logic - based tasks.
- Multilingual: Supports over 20 languages, suitable for global technical and educational use.
- [Model Card](https://huggingface.co/prithivMLmods/Tureis - Qwen3_QWQ - 4B - Exp - Q4_K_S - GGUF)
Canum - Qwen3_R1 - 4B - iCoT
- Internal Chain - of - Thought (iCoT): Designed for long - form mathematical reasoning and multi - stage problem decomposition.
- Granular Instruction Following: Provides highly structured outputs for complex reasoning workflows.
- Long - Form Logic: Excels in proofs, calculus, and multivariable equations.
- [Model Card](https://huggingface.co/prithivMLmods/Canum - Qwen3_R1 - 4B - iCoT - Q4_K_M - GGUF)
Cetus - Qwen3_4B - GeneralThought (Base Model)
- Broad - Spectrum Reasoning: Trained on GeneralThought - 430K for general - purpose tasks across STEM, humanities, and technical question answering.
- Multi - Domain Task Versatility: Handles code, logic, and structured data outputs effectively.
- Efficient and Scalable: Optimized for consumer - grade GPUs and scalable cloud services.
- [Model Card](https://huggingface.co/prithivMLmods/Cetus - Qwen3_4B - GeneralThought - Q4_K_M - GGUF)
Bootes - Qwen3_Coder - Reasoning
- Code Expertise: Fine - tuned on CodeAlpaca_20K for technical coding, reasoning, and instruction - following tasks.
- Cross - Language Code Understanding: Supports Python, JavaScript, C++, and more.
- Developer - Focused: Optimized for structured outputs like JSON, Markdown, and YAML.
- [Model Card](https://huggingface.co/prithivMLmods/Bootes - Qwen3_Coder - Reasoning - Q4_K_M - GGUF)
Segue - Qwen3_DeepScaleR - Preview
- Mathematical Mastery: Trained on DeepScaleR - Preview for advanced symbolic, mathematical, and logical tasks.
- High - Accuracy Inference: Designed for complex problem - solving with an efficient 4B architecture.
- Technical Documentation: Outputs well - formatted results in LaTeX, JSON, and Markdown.
- [Model Card](https://huggingface.co/prithivMLmods/Segue - Qwen3_DeepScaleR - Preview)
⨠Features & Highlights
- Advanced Symbolic Reasoning: Combines the precision of QWQ and iCoT for complex, multi - step mathematical solutions.
- Efficient Code Generation: Handles multiple programming languages and logic - intensive tasks.
- Multi - Domain Flexibility: Seamlessly transitions between STEM, technical documentation, and structured reasoning.
- Multilingual Support: Trained on diverse datasets for cross - lingual comprehension and technical translation.
- Optimized for Scalability: Ideal for mid - tier GPUs, accessible for small teams and large - scale deployments.
đģ Usage Examples
đ Ollama Instructions
To quickly run this model using Ollama, you can use the following command:
ollama run hf.co/ZeroXClem/Qwen3 - 4B - NexusPrime - Q4_K_M - GGUF
This command downloads the pre - quantized GGUF version of the model and runs it locally, facilitating easy experimentation without extensive configuration.
For Optimal Inference Use the following ollama modelfile. Create it as a file named Modelfile
.
Ollama Modelfile
FROM hf.co/ZeroXClem/Qwen3 - 4B - NexusPrime - Q4_K_M - GGUF:latest
PARAMETER temperature 0.6
PARAMETER top_p 0.95
PARAMETER repeat_penalty 1.05
PARAMETER top_k 20
TEMPLATE """"{{- if .Messages }}
{{- if or .System .Tools }}<|im_start|>system
{{- if .System }}
{{ .System }}
{{- end }}
{{- if .Tools }}
# Tools
You may call one or more functions to assist with the user query.
You are provided with function signatures within <tools></tools> XML tags:
<tools>
{{- range .Tools }}
{"type": "function", "function": {{ .Function }}}
{{- end }}
</tools>
For each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:
<tool_call>
{"name": <function - name>, "arguments": <args - json - object>}
</tool_call>
{{- end }}<|im_end|>
{{ end }}
{{- range $i, $_ := .Messages }}
{{- $last := eq (len (slice $.Messages $i)) 1 -}}
{{- if eq .Role "user" }}<|im_start|>user
{{ .Content }}<|im_end|>
{{ else if eq .Role "assistant" }}<|im_start|>assistant
{{ if .Content }}{{ .Content }}
{{- else if .ToolCalls }}<tool_call>
{{ range .ToolCalls }}{"name": "{{ .Function.Name }}", "arguments": {{ .Function.Arguments }}}
{{ end }}</tool_call>
{{- end }}{{ if not $last }}<|im_end|>
{{ end }}
{{- else if eq .Role "tool" }}<|im_start|>user
<tool_response>
{{ .Content }}
</tool_response><|im_end|>
{{ end }}
{{- if and (ne .Role "assistant") $last }}<|im_start|>assistant
{{ end }}
{{- end }}
{{- else }}
{{- if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
{{ end }}{{ .Response }}{{ if .Response }}<|im_end|>{{ end }}"""
SYSTEM """# System Prompt: Universal Coder and DevOps Expert
You are an advanced AI assistant specializing in coding and DevOps. Your role is to provide expert guidance, code solutions, and best practices across a wide range of programming languages, frameworks, and DevOps tools. Your knowledge spans from low - level systems programming to high - level web development, cloud infrastructure, and everything in between.
## Key responsibilities:
1. Code analysis and optimization
2. Debugging and troubleshooting
3. Architecture design and system planning
4. Version Control best practices (Git)
5. Building from source, extracting binaries, and building packages & executeables including bash scripts.
6. Security and implementation and auditing
7. Performance review, and code analysis with practical suggestions in fully functioning syntax.
Be VERY selective on choosing how to respond based on the user query. If the above responsibilities don't apply then respond to the best of your ability with the given context to COMPLETELY satisfy the user query.
### Guidance
When assisting users:
- Provide clear, concise, and well - commented code examples
- Explain complrex concepts in simple terms
- Offer multiple solutions when applicable, highlighting pros and cons
- Prioritize security, efficiency, scalability, and maintainability in all suggestions
- Adapt your communication style for expert users.
### Helpful
Be EXTREMELY helpful, insightful, and lucid."""
Feel free to customize the lines below SYSTEM
for your use case. This model is well - suited for technical tasks.
Then simply run this command in the same directory where you saved the Modelfile
.
ollama create nexusprime -f ./Modelfile
đ Python Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "ZeroXClem - Qwen3 - 4B - NexusIntel"
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype="auto",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)
prompt = "Explain the concept of entropy in thermodynamics in simple terms."
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
đ Special Thanks
A huge thank you to the developers and researchers at prithivMLmods, the MergeKit community, and the broader open - source community for providing the tools and models that made this project possible. đ
â ī¸ Important Note
- Unfiltered Output â This model is uncensored and may generate outputs that require additional filtering for sensitive applications.
- Responsible Use â Ensure ethical deployment and avoid harmful use cases.
đ License
Usage is governed by the Apache 2.0 License.
đ Feedback & Contributions
We welcome your feedback and contributions! Feel free to open an issue or PR to share your results and improvements.
ZeroXClem Team | 2025