mirror of
https://github.com/OpenBMB/MiniCPM-V.git
synced 2026-02-04 17:59:18 +08:00
50 lines
1.5 KiB
Python
50 lines
1.5 KiB
Python
import os
|
|
from ...smp import load_env
|
|
|
|
INTERNAL = os.environ.get('INTERNAL', 0)
|
|
|
|
|
|
def build_judge(**kwargs):
|
|
from ...api import OpenAIWrapper, SiliconFlowAPI
|
|
model = kwargs.pop('model', None)
|
|
kwargs.pop('nproc', None)
|
|
load_env()
|
|
LOCAL_LLM = os.environ.get('LOCAL_LLM', None)
|
|
if LOCAL_LLM is None:
|
|
model_map = {
|
|
'gpt-4-turbo': 'gpt-4-1106-preview',
|
|
'gpt-4-0613': 'gpt-4-0613',
|
|
'gpt-4-0125': 'gpt-4-0125-preview',
|
|
'gpt-4-0409': 'gpt-4-turbo-2024-04-09',
|
|
'chatgpt-1106': 'gpt-3.5-turbo-1106',
|
|
'chatgpt-0125': 'gpt-3.5-turbo-0125',
|
|
'gpt-4o': 'gpt-4o-2024-05-13',
|
|
'gpt-4o-0806': 'gpt-4o-2024-08-06',
|
|
'gpt-4o-mini': 'gpt-4o-mini-2024-07-18',
|
|
'qwen-7b': 'Qwen/Qwen2.5-7B-Instruct',
|
|
'qwen-72b': 'Qwen/Qwen2.5-72B-Instruct',
|
|
'deepseek': 'deepseek-ai/DeepSeek-V2.5',
|
|
}
|
|
model_version = model_map[model]
|
|
else:
|
|
model_version = LOCAL_LLM
|
|
|
|
if model in ['qwen-7b', 'qwen-72b', 'deepseek']:
|
|
model = SiliconFlowAPI(model_version, **kwargs)
|
|
else:
|
|
model = OpenAIWrapper(model_version, **kwargs)
|
|
return model
|
|
|
|
|
|
DEBUG_MESSAGE = """
|
|
To debug the OpenAI API, you can try the following scripts in python:
|
|
```python
|
|
from vlmeval.api import OpenAIWrapper
|
|
model = OpenAIWrapper('gpt-4o', verbose=True)
|
|
msgs = [dict(type='text', value='Hello!')]
|
|
code, answer, resp = model.generate_inner(msgs)
|
|
print(code, answer, resp)
|
|
```
|
|
You cam see the specific error if the API call fails.
|
|
"""
|