🚀 Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B
This repository contains the full - precision source code in "safe tensors" format, which can be used to generate GGUFs, GPTQ, EXL2, AWQ, HQQ, and other formats. The source code can also be used directly.
"V1.01" has been modified to address some issues related to non - stop/overly long generation and/or repeated "end paragraph" problems. I'm also keeping the original quantizations because the difference in creative generation between the two versions is quite significant. I'm not saying that the "reg" version is better than "v1.01"; they are just different, and in my opinion, you should have the option to choose between them.
The "GGUF" link at the bottom of the page leads to a repository that contains both V1.01 and "reg" quantizations.
⚠️ Important Note
If you plan to create GGUF quantizations, it is recommended to create the master file in float32 ("f32") and then perform quantization from this file due to the float 32 components/models in this merge. (Source files will be uploaded when the parameter count is shown in the upper left.)
💡 Usage Tip
Links to GGUFs are provided below.
✨ Features
High - Quality Settings and Optimal Operation
If you're going to use this model (source, GGUF, or a different quantization), please review the following document for critical parameter, sampler, and advanced sampler settings (for multiple AI/LLM apps).
This is a "Class 3/4" model (settings will enhance operation). For all settings used for this model (including specifics for its "class"), including example generations and an advanced settings guide (which often addresses any model issues), including methods to improve model performance for all use cases, as well as chat, role - play, and other use cases (especially for use cases beyond the model's design), please see:
[ https://huggingface.co/DavidAU/Maximizing - Model - Performance - All - Quants - Types - And - Full - Precision - by - Samplers_Parameters ]
The reason is that regardless of the "model class", this document will detail methods to enhance operations. For Class 3/4 models, the default settings (parameters, samplers, advanced samplers) must be set correctly for specific use cases. Some AI/LLM apps do not have consistent default settings, which results in sub - par model operation. Similarly, for Class 3/4 models (which operate somewhat to very differently from standard models), additional samplers and advanced sampler settings are required to "smooth out" operation and/or allow full operation for use cases the model was not designed for.
Bonus for All Models
This document also details parameters, samplers, and advanced samplers that can be used for any model, from any repository - all quantizations, and of course, source code operation too - to enhance the operation of any model.
[ https://huggingface.co/DavidAU/Maximizing - Model - Performance - All - Quants - Types - And - Full - Precision - by - Samplers_Parameters ]
💡 Usage Tip
I strongly suggest you also visit the DavidAU GGUF (below) repository for more details on using this model, especially if it is "Class 3" or "Class 4", to get maximum performance from the model.
📚 Documentation
For full information about this model, including:
- Details about this model and its use cases.
- Context limits
- Special usage notes/settings.
- Any models used to create this model.
- Templates used to access/use this model.
- Example generations
- GGUF quantizations of this model
Please go to:
[ https://huggingface.co/DavidAU/Llama - 3.1 - 1 - million - cxt - Dark - Planet - 8B - GGUF ]
📦 Model Information
Property |
Details |
Library Name |
transformers |
Tags |
mergekit, merge |
Base Models |
Sao10K/L3 - 8B - Stheno - v3.2, NeverSleep/Llama - 3 - Lumimaid - 8B - v0.1 - OAS, Hastagaras/Jamet - 8B - L3 - MK.V - Blackroot, nvidia/Llama - 3.1 - Nemotron - 8B - UltraLong - 1M - Instruct, DavidAU/Llama - 3.1 - 1million - ctx - Dark - Planet - 8B |