Files
MiniCPM-o/docs/xinference_infer.md

2.1 KiB

about xinference

xinference is a unified inference platform that provides a unified interface for different inference engines. It supports LLM, text generation, image generation, and more.but it's not bigger than swift too much.

xinference install

pip install "xinference[all]"

quick start

  1. start xinference
xinference
  1. start the web ui.
  2. Search for "MiniCPM-Llama3-V-2_5" in the search box. alt text
  3. find and click the MiniCPM-Llama3-V-2_5 button.
  4. follow the config and launch the model.
Model engine : Transformers
model format : pytorch
Model size   : 8
quantization : none
N-GPU        : auto
Replica      : 1
  1. after first click the launch button,xinference will download the model from huggingface. we should click the webui button. alt text
  2. upload the image and chatting with the MiniCPM-Llama3-V-2_5

local MiniCPM-Llama3-V-2_5 launch

  1. start xinference
xinference
  1. start the web ui.
  2. To register a new model, follow these steps: the settings highlighted in red are fixed and cannot be changed, whereas others are customizable according to your needs. Complete the process by clicking the 'Register Model' button. alt text alt text
  3. After completing the model registration, proceed to 'Custom Models' and locate the model you just registered.
  4. follow the config and launch the model.
Model engine : Transformers
model format : pytorch
Model size   : 8
quantization : none
N-GPU        : auto
Replica      : 1
  1. after first click the launch button,xinference will download the model from huggingface. we should click the chat button. alt text
  2. upload the image and chatting with the MiniCPM-Llama3-V-2_5

FAQ

  1. Why can't the sixth step open the WebUI? maybe your firewall or mac os to prevent the web to open.