Welcome to our GitHub repository! This repository is based on the ideas introduced in
To use the code in this repository, clone the repo and create a conda environment using:
conda env create --file=environment.yaml
conda activate sloth
Our data can be found on HuggingFace.
Go to vqa_generation
python main.py artifacts_dir=<artifacts_dir>
Comments:
- There are many configuration options inside
vqa_generation/confing/conf.yamlso be sure to check it before you start. - Api-keys for GPT and/or claude are needed to run the code.
- Additional installation is required -
pip install vqa_generation/requirements.txt - You will need to clone and install the llava repo to work with the stand-alone blind filtering (you can switch to any other model with minimal changes)
If you are interested in checking how our efficient eval method works in practice, please check this notebook.
Please check our notebooks.
If you find LiveXiv useful for your research and applications, please cite using this BibTeX:
@misc{shabtay2024livexivmultimodallive,
title={LiveXiv -- A Multi-Modal Live Benchmark Based on Arxiv Papers Content},
author={Nimrod Shabtay and Felipe Maia Polo and Sivan Doveh and Wei Lin and M. Jehanzeb Mirza and Leshem Chosen and Mikhail Yurochkin and Yuekai Sun and Assaf Arbelle and Leonid Karlinsky and Raja Giryes},
year={2024},
eprint={2410.10783},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2410.10783},
}