table
|
20240125
|
2024-01-25 22:42:59 +08:00 |
webpage
|
20240125
|
2024-01-25 22:42:59 +08:00 |
eval_gpt_review.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
eval_gpt_review_bench.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
eval_gpt_review_visual.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
eval_pope.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
eval_science_qa.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
eval_science_qa_gpt4.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
eval_science_qa_gpt4_requery.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
eval_textvqa.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
generate_webpage_data_from_table.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
m4c_evaluator.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
model_qa.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
model_vqa.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
model_vqa_loader.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
model_vqa_mmbench.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
model_vqa_science.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
qa_baseline_gpt35.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
run_llava.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |
summarize_gpt_review.py
|
20240125
|
2024-01-25 22:42:59 +08:00 |