đ GOAT-70B-Storytelling Model
The GOAT-70B-Storytelling model is trained by the GOAT.AI lab as the core model for an autonomous story - writing agent. It can generate high - quality, cohesive, and captivating narratives, such as books, novels, and movie scripts.
đ Quick Start
Self - Hosted via transformers
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "GOAT-AI/GOAT-70B-Storytelling"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.bfloat16
)
Using with GOAT - Storytelling - Agent
from goat_storytelling_agent.storytelling_agent import StoryAgent
backend_uri =
writer = StoryAgent(backend_uri, form='novel')
novel_scenes = writer.generate_story('treasure hunt in a jungle')
⨠Features
- High - Quality Narrative Generation: Generate books, novels, movie scripts, etc. with high quality, cohesion, and captivation.
- Flexible Input Utilization: Utilize inputs like plot outlines, character profiles, and their interrelationships to generate stories.
đĻ Installation
The README does not provide specific installation steps, so this section is skipped.
đģ Usage Examples
Basic Usage
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "GOAT-AI/GOAT-70B-Storytelling"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.bfloat16
)
Advanced Usage
from goat_storytelling_agent.storytelling_agent import StoryAgent
backend_uri =
writer = StoryAgent(backend_uri, form='novel')
novel_scenes = writer.generate_story('treasure hunt in a jungle')
đ Documentation
Model Description
Property |
Details |
Model Type |
LLaMA 2 70B |
License |
llama2 |
Context window length |
4096 tokens |
Training Details
Training was performed on a GPU cluster of 64xH100s. FSDP ZeRO - 3 sharding is employed for efficient training. We instruction finetune on a dataset of 18K examples for one epoch with batch size of 336, AdamW optimizer with learning rate 1e - 5.
Learn More
Uses
The main purpose of GOAT - 70B - Storytelling is to generate books, novels, movie scripts, etc. as an agent in coping with our GOAT - Storytelling - Agent. It is specifically designed for storywriters.
đ§ Technical Details
Training was carried out on a GPU cluster of 64xH100s. To ensure efficient training, FSDP ZeRO - 3 sharding was adopted. The model was instruction finetuned on a dataset of 18K examples for one epoch, with a batch size of 336. The AdamW optimizer was used with a learning rate of 1e - 5.
đ License
GOAT - 70B - Storytelling model is based on [Meta's LLaMA - 2 - 70b - hf](https://huggingface.co/meta - llama/Llama - 2 - 70b - hf), and uses its own datasets. The model weights are available under the LLAMA - 2 license.
â ī¸ Important Note
The GOAT - 70B - Storytelling model can produce factually incorrect output and should not be relied on to deliver factually accurate information. Therefore, the model could possibly generate wrong, biased, or otherwise offensive outputs.
Detailed results can be found [here](https://huggingface.co/datasets/open - llm - leaderboard/details_GOAT - AI__GOAT - 70B - Storytelling)
Metric |
Value |
Avg. |
67.38 |
AI2 Reasoning Challenge (25 - Shot) |
68.77 |
HellaSwag (10 - Shot) |
87.74 |
MMLU (5 - Shot) |
69.92 |
TruthfulQA (0 - shot) |
53.53 |
Winogrande (5 - shot) |
83.50 |
GSM8k (5 - shot) |
40.79 |
