|overview
Getting started guide

How to use FuelR

A full-width walkthrough — from connecting OpenAI, Gemini, or Claude to deploying a production chat widget.

Use the sidebar to jump between infrastructure, prompts, datasets, widgets, and runtime logs.

Platform Overview

FuelR is a Prompt Management & LLM Operations platform designed for teams that need to build, test, and ship AI features with confidence. Rather than scattering prompts across notebooks or Slack threads, FuelR gives every prompt a structured home with version history, dataset-level testing, and a one-click deployment path.

Prompt Playground

Design, iterate, and run prompts against any connected LLM provider in a sandboxed environment.

Dataset Evaluations

Upload CSV test sets and benchmark prompt versions across hundreds of real inputs.

Embeddable Widgets

Turn any validated prompt into a live chat widget with a single embed snippet.

Runtime Logs

Inspect every prompt execution — inputs, outputs, token counts, and cost — in real time.

Multi-Provider Support

Connect OpenAI, Anthropic, and Google Gemini. Switch providers per run without re-wiring anything.

Usage Dashboard

Track request volume, token spend, and model-level cost trends over time.

The Core Workflow

Everything in FuelR follows one intentional path. Complete each step and you will have a production-grade AI feature running by the end of this guide.

Connect Provider
Build Prompt
Test with Dataset
Deploy Widget
Monitor Logs

LLM Settings

Before any prompt can run, FuelR needs at least one LLM provider API key. Navigate to Settings → LLM Settings in the left sidebar. This is a one-time setup — keys are encrypted at rest and only the last four characters are ever shown in the UI.

ℹ Info

API keys are stored encrypted. FuelR only displays the last four characters in the table. You can connect all three supported providers and switch between them per run.

Adding a Provider

1

Select your provider

Choose from the dropdown: OpenAI, Anthropic, or Google Gemini.
2

Paste your API key

Enter the full API key. It is encrypted before being stored in the database.
3

Save and verify

Click Save. The row will appear in the provider table with the truncated key visible.
4

Set a default (optional)

Mark one provider as default. The Playground pre-selects it for every new prompt run.
Supported Providers
ProviderModels AvailableAPI Key Format
OpenAIGPT-4o, GPT-4, GPT-3.5-Turbosk-…
AnthropicClaude 3.5 Sonnet, Claude 3 Opussk-ant-…
Google GeminiGemini 1.5 Pro, Gemini FlashAI…

Prompts

The Prompts module is the core of FuelR. Every widget, evaluation, and log entry traces back to a prompt version. Navigate to Prompts in the sidebar and click New Prompt to get started.

Prompts come in two types:

💬Chat

Multi-turn conversations with System, User, and Assistant message roles. Best for conversational AI and support bots.

📝Text

Single-turn completions. Best for summarization, classification, extraction, and other single-shot tasks.

Playground

The Playground is a sandboxed environment where you write and iterate on a prompt before it touches production traffic. It surfaces your variables as live input fields so you can test different inputs instantly.

Example — Ticket Triage Prompt (Chat type)

System message

You are a support ticket parser. Extract the following fields and return ONLY valid JSON: - issue_category: "billing" | "technical" | "account" | "other" - severity: "low" | "medium" | "high" | "critical" - requested_action: short string - sentiment: "positive" | "neutral" | "negative" | "frustrated"

User message

{{ticket_text}}

Variable value

"I was charged twice this month and no one is responding to my emails. I want a refund immediately."

Expected output

{ "issue_category": "billing", "severity": "high", "requested_action": "issue a refund", "sentiment": "frustrated" }

✅ Tip

Run the prompt against several edge cases — very short inputs, non-English text, ambiguous requests — before saving a version. Iteration in the Playground is free and fast.

Variables

Variables let you parameterize a prompt so the same template works across different inputs. Wrap any placeholder in double curly braces — FuelR automatically detects and renders an input field for each one.

// In your prompt text:
Summarize the following {{ document_type }} in {{ language }}:

{{ content }}

// FuelR detects three variables:
// → document_type, language, content

When you hit Run, each variable value is substituted into the prompt before it is sent to the LLM. The runtime log records the fully resolved prompt so you can always see exactly what was sent.

Versioning

Every time you click Save, FuelR creates an immutable new version of the prompt. You are always prompted for a commit message — treat these like Git commit messages.

v1Initial working version
v2Add JSON format constraint
v3Handle edge case: empty ticket
v4Stable — passes 300-case datasetcurrent

ℹ Info

Widgets pin to a specific version. Saving a new version does not automatically update any deployed widget — you promote when you are ready.

Datasets

A single Playground run proves a prompt works on one input. A dataset proves it works across the distribution of real inputs your users send. Navigate to Datasets → New Dataset to create one.

Each dataset is a collection of test cases. A test case is a pair of:

  • input— the value injected into the prompt's primary variable
  • expected_output (optional) — ideal response, used as a reference during review

CSV Upload

The fastest way to build a dataset at scale. Prepare a .csv file with an input column and an optional expected_output column:

input,expected_output
"Charged twice, want refund","{'issue_category':'billing','severity':'high'}"
"App crashes on iOS 17","{'issue_category':'technical','severity':'critical'}"
"How do I change my email?","{'issue_category':'account','severity':'low'}"
Dataset Limits
Max test cases1,000 per dataset
Max file size5 MB
Max text per cell10,000 characters

FuelR validates the CSV before saving and surfaces any row-level errors inline — you can see exactly which rows fail and why without losing your other work.

Manual Entry

Use the row-by-row editor for small datasets (under ~20 cases) or when you want fine-grained control over each test case without preparing a spreadsheet. Both methods produce identical datasets that can be used in evaluations.

✅ Tip

Add tags to datasets (e.g. billing edge-cases v1-regression) to make them easy to find when running evaluations across many prompt versions.

Widgets

A Widget is a self-contained, embeddable chat interface backed by one of your prompt versions. Once a prompt has been validated in the Playground and tested against a dataset, a Widget is the path to getting it in front of real users — with no custom backend work required.

Creating a Widget

1

Open Widgets → New Widget

Navigate to Widgets in the left sidebar and click New Widget.
2

Name your widget

Give it an internal label, e.g. Support Bot — Billing Tier. This name is only visible to your team.
3

Select a prompt

Use the typeahead search to find the prompt by name. Only prompts you have created will appear.
4

Pin to a specific version

Choose the version you validated. Pinning to a specific version (not “latest”) means your widget won't change behavior when you iterate on the prompt.
5

Configure and create

Set a welcome message and any other options, then click Create Widget.

⚠ Note

Always pin to a specific version. If you pin to “latest” or leave version selection at default, updating the prompt could silently change live widget behavior without a deliberate deployment step.

Embedding the Widget

After creation, FuelR generates a <script> tag. Add it to any HTML page — the widget loads asynchronously and does not block your page rendering:

<!-- FuelR Widget -->
<script
  src="https://your-furlr-domain.com/widget.js"
  data-widget-id="YOUR_WIDGET_ID"
  defer
></script>

ℹ Info

CSP note: If your host page enforces a Content-Security-Policy, add the FuelR domain to script-src and connect-src:
Content-Security-Policy: script-src 'self' https://your-furlr-domain.com; connect-src 'self' https://your-backend-domain.com;

Runtime Logs

Every prompt execution — from the Playground, an evaluation run, or a live widget — produces a log entry. The Runtime Logs page is your real-time observatory for everything happening in your account. Navigate to Runtime Logs in the sidebar.

Logs are grouped by date for easier navigation. Use the search bar to filter by any text in the prompt or response — useful for finding a specific ticket text or isolating runs that produced a particular error string.

Log Fields

FieldDescription
TimestampExact time the request was made
Prompt nameWhich prompt and version was used
Provider / ModelWhich LLM handled the request
Input tokensTokens consumed by the prompt + injected variables
Output tokensTokens in the model's response
CostEstimated dollar cost for the request
Prompt sentFull resolved prompt with variables substituted
Response receivedRaw model output

Click any row to open a side panel. Toggle between the Prompt and Response views to verify that variables were substituted correctly and that the output matches expectations.

Debugging with Logs

Runtime Logs are your primary debugging tool. Here is the workflow when something goes wrong in production:

1

Identify the bad response

Collect a keyword or phrase from the problematic output or the user's input.
2

Search in Runtime Logs

Paste the keyword into the search bar. FuelR filters across all prompt and response text.
3

Open the log entry

Click the matching row. In the Prompt tab, verify that variables were injected correctly. In the Response tab, see exactly what the model returned.
4

Fix in the Playground

Reproduce the issue using the logged input as a variable value. Refine the prompt, save a new version, and run the dataset again.
5

Promote the widget

Once the new version passes testing, update the widget to point to the new version. The change is immediate.

Dashboard

The Dashboardgives you a high-level view of your account's activity. Navigate to Dashboard in the sidebar.

Total Requests

Count of all prompt executions across the Playground, evaluations, and widgets

Token Usage

Input and output token totals broken down by model and provider

Cost Trends

Estimated spend over time — identify which prompts are expensive to run

Use the Dashboard alongside Runtime Logs to understand both high-level spend trends and ground-level prompt quality. When cost spikes, drill into logs to find which prompt or widget is responsible.

Ready to build?

Connect your first LLM provider, create a prompt, and have a production widget running in under 30 minutes.

Last updated March 2026 · FuelR Platform Documentation