SUPIR/llava/eval
2024-01-25 22:42:59 +08:00
..
table 20240125 2024-01-25 22:42:59 +08:00
webpage 20240125 2024-01-25 22:42:59 +08:00
eval_gpt_review.py 20240125 2024-01-25 22:42:59 +08:00
eval_gpt_review_bench.py 20240125 2024-01-25 22:42:59 +08:00
eval_gpt_review_visual.py 20240125 2024-01-25 22:42:59 +08:00
eval_pope.py 20240125 2024-01-25 22:42:59 +08:00
eval_science_qa.py 20240125 2024-01-25 22:42:59 +08:00
eval_science_qa_gpt4.py 20240125 2024-01-25 22:42:59 +08:00
eval_science_qa_gpt4_requery.py 20240125 2024-01-25 22:42:59 +08:00
eval_textvqa.py 20240125 2024-01-25 22:42:59 +08:00
generate_webpage_data_from_table.py 20240125 2024-01-25 22:42:59 +08:00
m4c_evaluator.py 20240125 2024-01-25 22:42:59 +08:00
model_qa.py 20240125 2024-01-25 22:42:59 +08:00
model_vqa.py 20240125 2024-01-25 22:42:59 +08:00
model_vqa_loader.py 20240125 2024-01-25 22:42:59 +08:00
model_vqa_mmbench.py 20240125 2024-01-25 22:42:59 +08:00
model_vqa_science.py 20240125 2024-01-25 22:42:59 +08:00
qa_baseline_gpt35.py 20240125 2024-01-25 22:42:59 +08:00
run_llava.py 20240125 2024-01-25 22:42:59 +08:00
summarize_gpt_review.py 20240125 2024-01-25 22:42:59 +08:00