đ Palmyra-Med, a powerful LLM designed for healthcare
Palmyra-Med is a powerful large language model (LLM) specifically designed to meet the needs of the healthcare industry. It offers high - quality performance in biomedical tasks and is a reliable tool for healthcare - related research.
đ Quick Start
Use with transformers
You can run conversational inference using the Transformers Auto classes with the generate()
function. Here is an example:
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_id = "Writer/Palmyra-Med-70B"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.float16,
device_map="auto",
attn_implementation="flash_attention_2",
)
messages = [
{
"role": "system",
"content": "You are a highly knowledgeable and experienced expert in the healthcare and biomedical field, possessing extensive medical knowledge and practical expertise.",
},
{
"role": "user",
"content": "Does danzhi Xiaoyao San ameliorate depressive-like behavior by shifting toward serotonin via the downregulation of hippocampal indoleamine 2,3-dioxygenase?",
},
]
input_ids = tokenizer.apply_chat_template(
messages, tokenize=True, add_generation_prompt=True, return_tensors="pt"
)
gen_conf = {
"max_new_tokens": 256,
"eos_token_id": [tokenizer.eos_token_id, tokenizer.convert_tokens_to_ids("<|eot_id|>")],
"temperature": 0.0,
"top_p": 0.9,
}
with torch.inference_mode():
output_id = model.generate(input_ids, **gen_conf)
output_text = tokenizer.decode(output_id[0][input_ids.shape[1] :])
print(output_text)
⨠Features
Model Description
- Developed by: Writer
- Language(s) (NLP): English
- License: Writer open model license
- Finetuned from model: Palmyra-X-003
- Context window: 8192
Model Details
Palmyra-Med is a model built by Writer specifically for the healthcare industry. It is the leading LLM on biomedical benchmarks, with an average score of 85.87%, outperforming GPT - 4, claude Opus, Gemini, and the Med - PaLM - 2 base model, as well as a medically trained human test - taker.
Resources and Technical Documentation
Specialized for Biomedical Applications
Palmyra-Med-70B is carefully designed to meet the unique linguistic and knowledge demands of the medical and life sciences sectors. It has been fine - tuned on a large collection of high - quality biomedical data, ensuring it can understand and generate text with precise domain - specific accuracy and fluency.
Our system integrates the DPO dataset, a well - crafted fine - tuning recipe, and a custom diverse medical instruction dataset, making it highly capable of handling the specific needs of this field. Key components of our training pipeline include:
- Policy Optimization: Utilizing Direct Preference Optimization to enhance the model's performance. DPO.
- Fine - tuning dataset: Custom Medical Instruct dataset (Writer in - house build)
Evaluation Results
Palmyra-Med-70b outperforms larger models like GPT - 4, Gemini, and Med - PaLM - 2 across 9 diverse biomedical datasets, achieving state - of - the - art results with an average score of 85.9% despite having fewer parameters. Its strong performance in tasks like Clinical KG, Medical Genetics, and PubMedQA shows its effective understanding of biomedical knowledge.
Performance on Biomedical Benchmarks

Medical Use Cases
Palmyra-Med-70b is excellent at analyzing and summarizing complex clinical notes, EHR data, and discharge summaries, extracting key information to generate concise, structured summaries. It helps enhance clinical decision - making by performing advanced clinical entity recognition, identifying key medical concepts such as diseases, symptoms, medications, procedures, and anatomical structures from unstructured text.
By leveraging its deep understanding of medical terminology, the model improves information retrieval, data analysis, and knowledge discovery from EHRs, research articles, and other biomedical sources. These capabilities support applications like clinical decision support, pharmacovigilance, and medical research.
đ Documentation
Intended Use
Intended Use Cases: Palmyra-Med-70b is intended for non - commercial and research use in English. Instruction - tuned models are for assistant - like chat, while pretrained models can be adapted for various natural language generation tasks.
Out - of - scope: Use in any way that violates applicable laws or regulations (including trade compliance laws). Use in any other way prohibited by Writer's Acceptable Use Policy and the Writer open model license. Use in languages other than English.
Note: Developers may fine - tune Palmyra-Med-70b models for languages beyond English as long as they comply with the Writer open model license and the Acceptable Use Policy.
Watermarks: All models built by Writer.com contain watermarks to detect and prevent misuse and illegal use.
Bias, Risks, and Limitations
Palmyra-Med-70b, despite using high - quality data, may contain inaccuracies, biases, or misalignments and has not been rigorously evaluated in clinical trials or real - world healthcare settings.
It is advised not to use the model for direct patient care, clinical decision support, or professional medical purposes. Instead, its use should be limited to research by qualified individuals who understand its limitations. Palmyra-Med-70b should not replace professional medical judgment, and adapting it for medical use would require extensive additional work, including thorough testing, guideline alignment, bias mitigation, human oversight, and regulatory compliance. Always consult a qualified healthcare provider for personal medical needs.
Citation and Related Information
To cite this model:
@misc{Palmyra-Med-70B,
author = {Writer Engineering team},
title = {{Palmyra-Med-70b: A powerful LLM designed for healthcare}},
howpublished = {\url{https://dev.writer.com}},
year = 2024,
month = June
}
Contact: Hello@writer.com
đ License
This model is released under the Writer open model license.