CXR-LLaVA: Multimodal Large Language Model for Interpreting Chest X-ray Images
Paper • 2310.18341 • Published • 2
How to use ECOFRI/CXR-LLAVA-v2 with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("feature-extraction", model="ECOFRI/CXR-LLAVA-v2", trust_remote_code=True) # Load model directly
from transformers import AutoModel
model = AutoModel.from_pretrained("ECOFRI/CXR-LLAVA-v2", trust_remote_code=True, dtype="auto")https://github.com/ECOFRI/CXR_LLaVA
CXR LLaVA is an innovative open-source, multimodal large language model specifically designed for generating radiologic reports from chest X-ray images.
| Version | Input CXR resolution | Channels | Vision Encoder | Base LLM | Weight |
|---|---|---|---|---|---|
| v1.0 | 512x512 | RGB | RN50 | LLAMA2-13B-CHAT | Deprecated |
| v2.0 (Latest) | 512x512 | Grayscale | ViT-L/16 | LLAMA2-7B-CHAT | Link |