This is the int4 quantized version of MiniCPM-V 2.6.Runni... | This is the int4 quantized version of MiniCPM-V 2.6.Runni...
This is the int4 quantized version of MiniCPM-V 2.6.
Running with int4 version would use lower GPU memory (about 7GB).

Usage
Inference using Huggingface transformers on NVIDIA GPUs. Requirements tested on python 3.10: