Skip to main content

Overview

AI Puffer uses a model. Add your provider credentials, sync models, then choose the provider and model inside each module.
When you purchase AI Puffer Pro, you get access to additional plugin features. API usage is separate. You still need your own provider account and billing setup for OpenAI, Google, Claude, or any other provider you connect. AI Puffer does not include API credits.
https://mintcdn.com/aipuffer-a96fe641/65Dxat_5K2P6C8ev/images/logos/providers/openai.svg?fit=max&auto=format&n=65Dxat_5K2P6C8ev&q=85&s=2d699fd830ab01b49e3e70a7743e2a43

OpenAI

Text, images, embeddings, speech, web, and realtime.
https://mintcdn.com/aipuffer-a96fe641/65Dxat_5K2P6C8ev/images/logos/providers/google.svg?fit=max&auto=format&n=65Dxat_5K2P6C8ev&q=85&s=a3d22c630b1174f3208368eb26c1c496

Google

Gemini, images, video, embeddings, TTS, and grounding.
https://mintcdn.com/aipuffer-a96fe641/65Dxat_5K2P6C8ev/images/logos/providers/claude.svg?fit=max&auto=format&n=65Dxat_5K2P6C8ev&q=85&s=36368555e8aed1b5f2cef66819d09625

Claude

Text, web search, and supported image analysis.
https://mintcdn.com/aipuffer-a96fe641/65Dxat_5K2P6C8ev/images/logos/providers/openrouter.svg?fit=max&auto=format&n=65Dxat_5K2P6C8ev&q=85&s=d1f5fd38d03c2add42f26eacd3173455

OpenRouter

Access many text, image, embedding, and web-capable models.
https://mintcdn.com/aipuffer-a96fe641/65Dxat_5K2P6C8ev/images/logos/providers/azure.svg?fit=max&auto=format&n=65Dxat_5K2P6C8ev&q=85&s=ce5f475072ef2858a2ed72ff542da5f4

Azure

Azure OpenAI deployments for text, images, embeddings, and speech.
https://mintcdn.com/aipuffer-a96fe641/65Dxat_5K2P6C8ev/images/logos/providers/deepseek.svg?fit=max&auto=format&n=65Dxat_5K2P6C8ev&q=85&s=9061a37597ffbfb787eab8d88cc5d432

DeepSeek

Text generation for chat, writing, forms, and automations.
https://mintcdn.com/aipuffer-a96fe641/65Dxat_5K2P6C8ev/images/logos/providers/ollama.svg?fit=max&auto=format&n=65Dxat_5K2P6C8ev&q=85&s=6e6dcc958f8fec88dc4469962d919ea9

Ollama

Local or self-hosted text and embedding models.
AI providers power text, image, audio, , and retrieval features.
ProviderTextImagesEmbeddingsAudioVideoWebRealtime
OpenAITTS + STT-
GoogleTTS-
Claude-----
OpenRouter---
AzureSTT---
DeepSeek------
Ollama-----
means supported. - means not supported in the current provider strategy. AI Puffer > Settings is the central area for managing your connections to different AI providers.
Settings

OpenAI

OpenAI supports the widest set of AI Puffer features.
  1. Open AI Puffer > Settings > AI.
  2. Select OpenAI.
  3. Paste your OpenAI API key.
  4. Keep the default base URL unless you use a compatible custom endpoint.
  5. Open Advanced and click Sync Models.
  6. Select the model you want to use by default.
For knowledge features, OpenAI embedding models are available in embedding model selectors. OpenAI Vector Stores are handled in Knowledge Base.
OpenAI API key settings

Google

Google provides Gemini, embeddings, image/video models, text to speech, and grounding.
  1. Open AI Puffer > Settings > AI.
  2. Select Google.
  3. Paste your Google API key.
  4. Open Advanced and click Sync Models.
  5. Select the default model.
  6. Review Google safety settings if you need to change blocking thresholds.
Google image and video model settings are configured in Images.
Google safety settings can block some responses. Change them only when the default behavior is too restrictive for your site.
Google API key settings

Claude

Claude is available for text workflows, web search, and supported image analysis models.
  1. Open AI Puffer > Settings > AI.
  2. Select Claude.
  3. Paste your Anthropic API key.
  4. Open Advanced and click Sync Models.
  5. Select the default model.
Claude does not provide embeddings through the current AI Puffer provider strategy.
Claude API key settings

OpenRouter

OpenRouter gives access to models available in your OpenRouter account.
  1. Open AI Puffer > Settings > AI.
  2. Select OpenRouter.
  3. Paste your OpenRouter API key.
  4. Open Advanced and click Sync Models.
  5. Select the default model.
Some capabilities depend on the selected model.
OpenRouter support depends on the model you select. Check the model capabilities before using image, embedding, or web features.
OpenRouter API key settings

Azure

Azure uses your Azure OpenAI .
  1. Open AI Puffer > Settings > AI.
  2. Select Azure.
  3. Paste your Azure API key.
  4. Enter the Azure endpoint URL for your resource.
  5. Open Advanced.
  6. Check the API versions if your Azure resource requires different versions.
  7. Click Sync Models to load deployments.
  8. Select the deployment to use as the default model.
Azure uses deployment names. If a deployment does not appear, confirm it exists in Azure and that the endpoint and API key belong to the same resource.
Azure API key settings

DeepSeek

DeepSeek is available for text workflows.
  1. Open AI Puffer > Settings > AI.
  2. Select DeepSeek.
  3. Paste your DeepSeek API key.
  4. Open Advanced and click Sync Models.
  5. Select the default model.
DeepSeek embeddings are not supported by the current provider strategy. Use OpenAI, Google, Azure, OpenRouter, or Ollama for embedding workflows.
DeepSeek API key settings

Ollama

Ollama connects local or self-hosted models.
  1. Install Ollama from ollama.com/download.
  2. Run Ollama on the computer or server you want to use as the AI server.
  3. Pull a model:
ollama pull llama3
  1. Open AI Puffer > Settings > AI.
  2. Select Ollama.
  3. Enter the Ollama base URL. The default is http://localhost:11434.
  4. Click Sync Models.
  5. Select the default model.
Install notes:
SystemNotes
macOSDownload the macOS ZIP, move Ollama to Applications, then run the app.
WindowsRun the Windows installer. Ollama runs in the background after installation.
LinuxRun curl -fsSL https://ollama.com/install.sh | sh. The installer sets up Ollama as a service.
You can pull more than one model. AI Puffer shows synced Ollama models after Sync Models runs.
If WordPress and Ollama are on different servers, do not use localhost unless Ollama is running on the same server as WordPress. Use the reachable server URL instead.
Do not expose Ollama publicly without access controls. Anyone who can reach the Ollama server can send model requests to it.
Ollama API key settings

Troubleshooting

The provider account is out of credits, has reached a spend limit, or is blocked by billing settings.Check these items:
  1. Open the provider billing page.
  2. Check credits, usage limits, and monthly spend limits.
  3. Add credits or update billing if needed.
  4. If you just changed billing, wait a few minutes and try again.
For OpenAI, see error codes.
The saved key is wrong, old, revoked, copied with extra spaces, or belongs to a different account or organization.Check these items:
  1. Create or copy a fresh API key from the provider dashboard.
  2. Paste it again in AI Puffer > Settings > AI.
  3. Save the setting.
  4. Sync models again.
For OpenAI, see Incorrect API key provided.
AI Puffer syncs models when you switch providers, but the list can be out of date.Check these items:
  1. Open AI Puffer > Settings > AI.
  2. Select the provider.
  3. Open Advanced.
  4. Click Sync Models.
  5. Open Settings > Others.
  6. Click Clear Model Cache.
  7. Click Clear Transients.
  8. Click Sync All.
  9. Return to Settings > AI and check the model list again.
Settings Others model cache buttons
If the model still does not appear, confirm the provider account has access to that model.
Some OpenAI models require a verified API organization.Check these items:
  1. Open OpenAI Platform settings.
  2. Go to Organization > General.
  3. Complete organization verification.
  4. Wait up to 30 minutes.
  5. Generate a new API key if the error continues.
  6. Make sure AI Puffer is using a key from the verified organization.
For OpenAI, see API Organization Verification.
A WordPress security plugin or firewall may be blocking AI Puffer settings requests.Check these items:
  1. Check your security plugin or firewall logs.
  2. Whitelist AI Puffer admin requests if your tool supports allow rules.
  3. If using Wordfence or a similar firewall, switch to learning mode.
  4. Save AI Puffer settings and use the affected module a few times.
  5. Switch the firewall back to normal mode after it learns the requests.