Open-source medical large language model adapted from Llama 2 to the medical domain.
7b
70b
39.6K Pulls Updated 11 months ago
Updated 12 months ago
12 months ago
a917f4ff62ba · 4.1GB
model
archllama
·
parameters6.74B
·
quantizationQ4_K_M
4.1GB
params
{"num_ctx":2048,"stop":["\u003c|im_start|\u003e","\u003c|im_end|\u003e"]}
74B
template
<|im_start|>system
{{ .System }}<|im_end|>
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assi
106B
system
A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful,
155B
Readme
Meditron is a large language model adapted from Llama 2 to the medical domain through training on a corpus of medical data, papers and guidelines. It outperforms Llama 2, GPT 3.5 and Flan-PaLM on many medical reasoning tasks.
Potential use cases include:
- Medical exam question answering
- Supporting differential diagnosis
- Disease information (symptoms, cause, treatment) query
- General health information query
Example prompts
What are the symptoms of the common cold?
What causes the seasonal flu?
What medication would be prescribed for a headache?
References
MEDITRON-70B: Scaling Medical Pretraining for Large Language Models