Link model to paper
Browse filesThis PR improves the discoverability of the model by linking it to the paper: https://huggingface.co/papers/2407.07895
README.md
CHANGED
|
@@ -46,4 +46,18 @@ Use the code below to evaluate the model.
|
|
| 46 |
Please first edit /path/to/ckpt to the path of checkpoint, /path/to/images to the path of "interleave_data" in scripts/interleave/eval_all.sh and then run
|
| 47 |
```bash
|
| 48 |
bash scripts/interleave/eval_all.sh
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 49 |
```
|
|
|
|
| 46 |
Please first edit /path/to/ckpt to the path of checkpoint, /path/to/images to the path of "interleave_data" in scripts/interleave/eval_all.sh and then run
|
| 47 |
```bash
|
| 48 |
bash scripts/interleave/eval_all.sh
|
| 49 |
+
```
|
| 50 |
+
|
| 51 |
+
## Bibtex citation
|
| 52 |
+
|
| 53 |
+
```bibtex
|
| 54 |
+
@misc{li2024llavanextinterleavetacklingmultiimagevideo,
|
| 55 |
+
title={LLaVA-NeXT-Interleave: Tackling Multi-image, Video, and 3D in Large Multimodal Models},
|
| 56 |
+
author={Feng Li and Renrui Zhang and Hao Zhang and Yuanhan Zhang and Bo Li and Wei Li and Zejun Ma and Chunyuan Li},
|
| 57 |
+
year={2024},
|
| 58 |
+
eprint={2407.07895},
|
| 59 |
+
archivePrefix={arXiv},
|
| 60 |
+
primaryClass={cs.CV},
|
| 61 |
+
url={https://arxiv.org/abs/2407.07895},
|
| 62 |
+
}
|
| 63 |
```
|