Run MiniCPM4 with CPU only

2025-06-08
This is a guide for serving the MiniCPM4 0.5b model with CPU only(in my laptop WSL Ubuntu 24.04). The model is served by llama.cpp, and I break down it for several steps. Download the model from hf huggingface-cli download openbmb/MiniCPM4-0.5B Install(and compile) the llama.cpp # build llama.cpp git clone https://github.com/ggml-org/llama.cpp.git sudo apt install python3-dev build-essential cmake libcurl4-openssl-dev cmake -B build cmake --build build --config Release # prepare llama tools uv venv --python=3. Continue reading

Call MCP Server(stdio) directly in the shell

2025-05-19
stdio is one of the transport mechanisms MCP(Model Context Protocol) natively supports. Actually stdio is the only transport mechanism Claude Desktop supports currently(as in May, 2025). This article will show you how to call a MCP server in the shell, without mcp dev or any third party tools, only with echo > or copying/pasting JSONRPC message directly. First, let’s write a very simple MCP server(get_time.py) to get the current time: Continue reading