Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co./docs/hub/model-cards#model-card-metadata)

Model Card

Parrot is a multi-language and multi-modal large language model capable of achieving excellent performance. For a comprehensive introduction, please refer to Parrot Paper and Parrot GitHub.

Model Details

Performance

Usage

Below is a code snippet to run Parrot with multimodal inputs. For additional usage instructions, including inference wrapper and Gradio UI, please refer to Parrot GitHub.

pip install torch==2.1.2 transformers==4.43.2 pillow==10.3.0
import torch
from PIL import Image
from transformers import AutoModelForCausalLM

Citation

If you find Parrot useful, please cite the paper

@article{sun2024parrot,
  title={Parrot: Multilingual Visual Instruction Tuning},
  author={Sun, Hai-Long and Zhou, Da-Wei and Li, Yang and Lu, Shiyin and Yi, Chao and Chen, Qing-Guo and Xu, Zhao and Luo, Weihua and Zhang, Kaifu and Zhan, De-Chuan and others},
  journal={arXiv preprint arXiv:2406.02539},
  year={2024}
}

License

The project is licensed under Apache License Version 2.0 and is restricted to uses that comply with the license agreements of Qwen and Clip.

Downloads last month
57
Safetensors
Model size
14.9B params
Tensor type
BF16
·
Inference API
Unable to determine this model’s pipeline type. Check the docs .