--- inference: false library_name: transformers --- # VW-LMM Model Card This repo contains the weights of VW-LMM-Mistral-7b proposed in paper "Multi-modal Auto-regressive Modeling via Visual Words" For specific usage and chat templates, please refer to our project repo https://github.com/pengts/VW-LMM ## Model details **Model type:** VW-LMM is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture. Base LLM: [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) **paper:** https://arxiv.org/abs/2403.07720 **code:** https://github.com/pengts/VW-LMM ## License [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) license. ## Citation If you find our paper and code useful in your research, please consider giving a star :star: and citation :pencil:. ```BibTeX @misc{peng2024multimodal, title={Multi-modal Auto-regressive Modeling via Visual Words}, author={Tianshuo Peng and Zuchao Li and Lefei Zhang and Hai Zhao and Ping Wang and Bo Du}, year={2024}, eprint={2403.07720}, archivePrefix={arXiv}, primaryClass={cs.CV} } ```