[2024.05.24] We release the MiniCPM-Llama3-V 2.5 gguf, wh... | [2024.05.24] We release the MiniCPM-Llama3-V 2.5 gguf, wh...
[2024.05.24] We release the MiniCPM-Llama3-V 2.5 gguf, which supports llama.cpp inference and provides a 6~8 token/s smooth decoding on mobile phones. Try it now!