How to use FuelR
A full-width walkthrough — from connecting OpenAI, Gemini, or Claude to deploying a production chat widget.
Use the sidebar to jump between infrastructure, prompts, datasets, widgets, and runtime logs.
Platform Overview
FuelR is a Prompt Management & LLM Operations platform designed for teams that need to build, test, and ship AI features with confidence. Rather than scattering prompts across notebooks or Slack threads, FuelR gives every prompt a structured home with version history, dataset-level testing, and a one-click deployment path.
Prompt Playground
Design, iterate, and run prompts against any connected LLM provider in a sandboxed environment.
Dataset Evaluations
Upload CSV test sets and benchmark prompt versions across hundreds of real inputs.
Embeddable Widgets
Turn any validated prompt into a live chat widget with a single embed snippet.
Runtime Logs
Inspect every prompt execution — inputs, outputs, token counts, and cost — in real time.
Multi-Provider Support
Connect OpenAI, Anthropic, and Google Gemini. Switch providers per run without re-wiring anything.
Usage Dashboard
Track request volume, token spend, and model-level cost trends over time.
The Core Workflow
Everything in FuelR follows one intentional path. Complete each step and you will have a production-grade AI feature running by the end of this guide.
LLM Settings
Before any prompt can run, FuelR needs at least one LLM provider API key. Navigate to Settings → LLM Settings in the left sidebar. This is a one-time setup — keys are encrypted at rest and only the last four characters are ever shown in the UI.
ℹ Info
Adding a Provider
Select your provider
Paste your API key
Save and verify
Set a default (optional)
| Provider | Models Available | API Key Format |
|---|---|---|
| OpenAI | GPT-4o, GPT-4, GPT-3.5-Turbo | sk-… |
| Anthropic | Claude 3.5 Sonnet, Claude 3 Opus | sk-ant-… |
| Google Gemini | Gemini 1.5 Pro, Gemini Flash | AI… |
Prompts
The Prompts module is the core of FuelR. Every widget, evaluation, and log entry traces back to a prompt version. Navigate to Prompts in the sidebar and click New Prompt to get started.
Prompts come in two types:
Multi-turn conversations with System, User, and Assistant message roles. Best for conversational AI and support bots.
Single-turn completions. Best for summarization, classification, extraction, and other single-shot tasks.
Playground
The Playground is a sandboxed environment where you write and iterate on a prompt before it touches production traffic. It surfaces your variables as live input fields so you can test different inputs instantly.
System message
User message
Variable value
Expected output
✅ Tip
Variables
Variables let you parameterize a prompt so the same template works across different inputs. Wrap any placeholder in double curly braces — FuelR automatically detects and renders an input field for each one.
// In your prompt text:
Summarize the following {{ document_type }} in {{ language }}:
{{ content }}
// FuelR detects three variables:
// → document_type, language, contentWhen you hit Run, each variable value is substituted into the prompt before it is sent to the LLM. The runtime log records the fully resolved prompt so you can always see exactly what was sent.
Versioning
Every time you click Save, FuelR creates an immutable new version of the prompt. You are always prompted for a commit message — treat these like Git commit messages.
ℹ Info
Datasets
A single Playground run proves a prompt works on one input. A dataset proves it works across the distribution of real inputs your users send. Navigate to Datasets → New Dataset to create one.
Each dataset is a collection of test cases. A test case is a pair of:
- →input— the value injected into the prompt's primary variable
- →expected_output (optional) — ideal response, used as a reference during review
CSV Upload
The fastest way to build a dataset at scale. Prepare a .csv file with an input column and an optional expected_output column:
input,expected_output
"Charged twice, want refund","{'issue_category':'billing','severity':'high'}"
"App crashes on iOS 17","{'issue_category':'technical','severity':'critical'}"
"How do I change my email?","{'issue_category':'account','severity':'low'}"| Max test cases | 1,000 per dataset |
| Max file size | 5 MB |
| Max text per cell | 10,000 characters |
FuelR validates the CSV before saving and surfaces any row-level errors inline — you can see exactly which rows fail and why without losing your other work.
Manual Entry
Use the row-by-row editor for small datasets (under ~20 cases) or when you want fine-grained control over each test case without preparing a spreadsheet. Both methods produce identical datasets that can be used in evaluations.
✅ Tip
billing edge-cases v1-regression) to make them easy to find when running evaluations across many prompt versions.Widgets
A Widget is a self-contained, embeddable chat interface backed by one of your prompt versions. Once a prompt has been validated in the Playground and tested against a dataset, a Widget is the path to getting it in front of real users — with no custom backend work required.
Creating a Widget
Open Widgets → New Widget
Name your widget
Select a prompt
Pin to a specific version
Configure and create
⚠ Note
Embedding the Widget
After creation, FuelR generates a <script> tag. Add it to any HTML page — the widget loads asynchronously and does not block your page rendering:
<!-- FuelR Widget -->
<script
src="https://your-furlr-domain.com/widget.js"
data-widget-id="YOUR_WIDGET_ID"
defer
></script>ℹ Info
script-src and connect-src:Content-Security-Policy: script-src 'self' https://your-furlr-domain.com;
connect-src 'self' https://your-backend-domain.com;Runtime Logs
Every prompt execution — from the Playground, an evaluation run, or a live widget — produces a log entry. The Runtime Logs page is your real-time observatory for everything happening in your account. Navigate to Runtime Logs in the sidebar.
Logs are grouped by date for easier navigation. Use the search bar to filter by any text in the prompt or response — useful for finding a specific ticket text or isolating runs that produced a particular error string.
Log Fields
| Field | Description |
|---|---|
| Timestamp | Exact time the request was made |
| Prompt name | Which prompt and version was used |
| Provider / Model | Which LLM handled the request |
| Input tokens | Tokens consumed by the prompt + injected variables |
| Output tokens | Tokens in the model's response |
| Cost | Estimated dollar cost for the request |
| Prompt sent | Full resolved prompt with variables substituted |
| Response received | Raw model output |
Click any row to open a side panel. Toggle between the Prompt and Response views to verify that variables were substituted correctly and that the output matches expectations.
Debugging with Logs
Runtime Logs are your primary debugging tool. Here is the workflow when something goes wrong in production:
Identify the bad response
Search in Runtime Logs
Open the log entry
Fix in the Playground
Promote the widget
Dashboard
The Dashboardgives you a high-level view of your account's activity. Navigate to Dashboard in the sidebar.
Count of all prompt executions across the Playground, evaluations, and widgets
Input and output token totals broken down by model and provider
Estimated spend over time — identify which prompts are expensive to run
Use the Dashboard alongside Runtime Logs to understand both high-level spend trends and ground-level prompt quality. When cost spikes, drill into logs to find which prompt or widget is responsible.
Ready to build?
Connect your first LLM provider, create a prompt, and have a production widget running in under 30 minutes.
Last updated March 2026 · FuelR Platform Documentation