docker run -it --rm ghcr.io/purton-tech/mpt-7b-chat
It's a big download due to the model size i.e. 5GB. The model is quantized and runs via the ggml tensor library. https://ggml.ai/.
docker run -it --rm ghcr.io/purton-tech/mpt-7b-chat
It's a big download due to the model size i.e. 5GB. The model is quantized and runs via the ggml tensor library. https://ggml.ai/.