Description
This plugin provides llama.cpp integration for the WordPress AI Client. It lets WordPress sites use large language models running via a llama.cpp server for text generation and other AI capabilities.
llama.cpp exposes an OpenAI-compatible API, and this provider uses that API to communicate with any GGUF model loaded into your llama.cpp server.
Features:
- Text generation with any llama.cpp-loaded model
- Automatic model discovery from your llama.cpp server
- Function calling support
- Structured output (JSON mode) support
- Settings page to configure the server URL (default: http://127.0.0.1:8080)
- Works without an API key for local instances
Requirements:
- PHP 7.4 or higher
- WordPress AI Client plugin must be installed and activated
- llama.cpp server running locally or on a remote host
Getting Started
What do I need before using this plugin?
You need a running llama.cpp server (local or remote) and the WordPress AI Client plugin installed and activated.
How do I install llama.cpp?
On macOS run brew install llama.cpp. For other platforms, see the official docs at https://llama-cpp.com/download/
Where do I get a model?
Download a GGUF model from Hugging Face. TinyLlama 1.1B (Q4_K_M, ~636 MB) is a good starting point for testing.
How do I start the server?
Run llama-server --models-dir ~/models. The server starts on http://127.0.0.1:8080 by default.
How do I connect WordPress to my llama.cpp server?
Go to Settings > llama.cpp and enter your server URL. Leave it blank to use the default (http://127.0.0.1:8080).
For full step-by-step setup instructions, see the Getting Started Guide.
Installation
- Ensure the WordPress AI Client plugin is installed and activated.
- Upload the plugin files to
/wp-content/plugins/ai-provider-for-llamacpp/. - Activate the plugin through the ‘Plugins’ menu in WordPress.
- Go to Settings > llama.cpp to configure the server URL.
FAQ
-
Do I need an API key?
-
No. For local llama.cpp instances, no API key is needed. The plugin automatically registers an empty key as a fallback.
-
What is the default server URL?
-
http://127.0.0.1:8080 — the default address used by
llama-server. -
How do I change the server URL?
-
Go to Settings > llama.cpp and enter your server’s base URL in the “Server URL” field.
Reviews
There are no reviews for this plugin.
Contributors & Developers
“AI Provider for llama.cpp” is open source software. The following people have contributed to this plugin.
ContributorsTranslate “AI Provider for llama.cpp” into your language.
Interested in development?
Browse the code, check out the SVN repository, or subscribe to the development log by RSS.
Changelog
0.0.3
- Added support for the WordPress AI plugin — hooks into
wpai_has_ai_credentialsso the AI plugin recognises llama.cpp as a configured provider without requiring an API key
0.0.1
- Initial release
- Text generation with llama.cpp OpenAI-compatible API
- Settings page for server URL configuration
0.0.2
- Added documentation in the docs/ directory
- Fixed minor issues and small bugs
- Updated plugin icon and banner images
