| | import sys |
| | from vlmeval import * |
| | from vlmeval.dataset import SUPPORTED_DATASETS |
| | FAIL_MSG = 'Failed to obtain answer via API.' |
| |
|
| | root = sys.argv[1] |
| | if root[-1] in '/\\': |
| | root = root[:-1] |
| |
|
| | model_name = root.split('/')[-1] |
| |
|
| | for d in SUPPORTED_DATASETS: |
| | fname = f'{model_name}_{d}.xlsx' |
| | pth = osp.join(root, fname) |
| | if osp.exists(pth): |
| | data = load(pth) |
| | |
| | assert 'prediction' in data |
| | data['prediction'] = [str(x) for x in data['prediction']] |
| | fail = [FAIL_MSG in x for x in data['prediction']] |
| | if sum(fail): |
| | nfail = sum(fail) |
| | ntot = len(fail) |
| | print(f'Model {model_name} x Dataset {d}: {nfail} out of {ntot} failed. {nfail / ntot * 100: .2f}%. ') |
| |
|
| | eval_files = ls(root, match=f'{model_name}_{d}_') |
| | eval_files = [x for x in eval_files if listinstr([f'{d}_openai', f'{d}_gpt'], x) and x.endswith('.xlsx')] |
| |
|
| | if len(eval_files) == 0: |
| | print(f'Model {model_name} x Dataset {d} openai missing') |
| | continue |
| | |
| | assert len(eval_files) == 1 |
| | eval_file = eval_files[0] |
| | data = load(eval_file) |
| | |
| | if 'MMVet' in d: |
| | bad = [x for x in data['log'] if 'All 5 retries failed.' in str(x)] |
| | if len(bad): |
| | print(f'Model {model_name} x Dataset {d} Evaluation: {len(bad)} out of {len(data)} failed.') |
| | elif 'MathVista' in d: |
| | bad = [x for x in data['res'] if FAIL_MSG in str(x)] |
| | if len(bad): |
| | print(f'Model {model_name} x Dataset {d} Evaluation: {len(bad)} out of {len(data)} failed.') |
| | |
| | elif d == 'LLaVABench': |
| | sub = data[data['gpt4_score'] == -1] |
| | sub = sub[sub['gpt4_score'] == -1] |
| | if len(sub): |
| | print(f'Model {model_name} x Dataset {d} Evaluation: {len(sub)} out of {len(data)} failed.') |
| | else: |
| | bad = [x for x in data['log'] if FAIL_MSG in str(x)] |
| | if len(bad): |
| | print(f'Model {model_name} x Dataset {d} Evaluation: {len(bad)} out of {len(data)} failed.') |
| | |