E

Excalibur 7b DPO

Developed by InferenceIllusionist
Excalibur-7b-DPO is a large language model based on the Excalibur-7b foundation model, fine-tuned with Direct Preference Optimization (DPO), focusing on improving dialogue quality and performance in visual application scenarios.
Downloads 22
Release Time : 3/28/2024

Model Overview

This model was fine-tuned using the Intel/orca_dpo_pairs dataset with DPO to enhance the response quality of the original model, particularly in visual application scenarios. The fine-tuned model exhibits more conversational and comprehensive responses, showing improvements across multiple benchmarks.

Model Features

DPO fine-tuning optimization
Fine-tuned using Direct Preference Optimization (DPO), significantly improving dialogue quality and response comprehensiveness.
Enhanced visual applications
Specifically optimized for performance in visual application scenarios, supporting image understanding and description.
Multi-format support
Supports ChatML and Alpaca prompt formats, adaptable to various application scenarios.
Quantized versions available
Offers weighted and static quantized versions to meet different hardware requirements.

Model Capabilities

Text generation
Visual scene understanding
Multi-turn dialogue
Knowledge Q&A
Reasoning tasks

Use Cases

Visual applications
Image caption generation
Generates detailed descriptions based on input images
Requires additional mmproj file support
Dialogue systems
Intelligent assistant
Builds more natural and fluent conversational assistants
Significant improvement in dialogue quality after fine-tuning
Educational applications
Knowledge Q&A
Answers various knowledge-based questions
Performs well on benchmarks like the AI2 Reasoning Challenge
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase