Fiddler Debugging Assistant

Updated on Feb 26, 2026

The Fiddler Debugging Assistant enables you to leverage large language models (LLMs) to reach new levels of developer productivity. This built-in chat feature provides additional insights by combining your preferred AI model with Fiddler's powerful traffic analysis capabilities.

Prerequisites

  • The latest version of Fiddler Everywhere.
  • A Fiddler Everywhere Pro or higher subscription tier.
  • An API key with access to one of the supported language models. Fiddler Everywhere currently supports the following providers:
    • OpenAI
    • Anthropic
    • Azure OpenAI
    • Google Gemini

Requests are processed using your own API key and billed directly to your account with the LLM provider. Data usage and retention follow the provider’s Terms of Service and Privacy Policy.

Configuring the Debugging Assistant

The Debugging Assistant is accessible through the Ask Assistant button in the Fiddler toolbar. To use it, you must provide a valid API key for one of the supported AI providers.

Select your preferred LLM provider and set its API key through Settings > Assistant.

Configuring the debugging assistant API key

Once all properties are set, use the Test Connection button to verify that the connection is properly established.

Testing and verifying the LLM connection

Configuration Details

  • When setting the model, you can choose from the dropdown list of available models or enter a custom model name if it's not listed.

  • When setting the Azure target URI, note that this must be a complete URL containing the endpoint, deployment name, and API version. You can find this target URI in the deployment details page in Azure AI Foundry.

Getting Started with a Free Gemini API Key

Google provides a free‑tier Gemini API key through Google AI Studio that lets you use the Debugging Assistant without a paid LLM provider subscription; however, you still need a Fiddler Everywhere Pro (or higher) subscription to access the Debugging Assistant.

Obtaining a Free Gemini API Key

  1. Navigate to https://aistudio.google.com/api-keys.
  2. Sign in with your Google account.
  3. Click Create API key and follow the prompts.
  4. Copy the generated API key.

Configuring Gemini in Fiddler Everywhere

  1. Open Fiddler Everywhere and go to Settings > Assistant.
  2. Select Google Gemini as the provider.
  3. Paste your API key from Google AI Studio.
  4. Select a supported Gemini model from the dropdown list.
  5. Click Test Connection to confirm the setup.

Supported Gemini Models and Free-Tier Limits

Fiddler Everywhere supports the following Gemini models. Of these, only two are currently available under the Google AI Studio free tier:

ModelFree-Tier Available
Gemini 2.5 ProNo
Gemini 2.5 FlashYes
Gemini 2.5 Flash LiteYes
Gemini 2.0 FlashNo
Gemini 2.0 Flash LiteNo

For the latest model availability and rate limit information, see the Google AI Studio rate limits page.

Google may use data submitted through the free-tier API to improve its models. Avoid sharing sensitive, confidential, or personally identifiable information when using the free Gemini API key. Review Google's Terms of Service and Privacy Policy before use.

Using the Debugging Assistant

To use the Debugging Assistant:

  1. Open the Fiddler Everywhere application.

  2. Click the Ask Assistant button.

  3. Ask questions in natural language about Fiddler or the captured traffic data.

    Debugging assistant chat overview

Use the New Chat button to start a fresh conversation and clear the current chat history from the context.

Limitations

When working with the Debugging Assistant, consider the following limitations:

  • The current version of the Debugging Assistant does not have access to captured traffic in live sessions or saved snapshots. To analyze specific captured sessions, you need to copy and paste the relevant traffic details into the chat.

  • Most LLMs (like Claude Sonnet, GPT-5, Gemini Pro) work only with preexisting training data or information provided in the conversation. Internet connectivity features (such as web search) depend on the selected LLM model.

Debugging Assistant Access Policies

Fiddler Everywhere provides managed application policies through its Enterprise tier. License administrators can use these policies to control access to the Fiddler Debugging Assistant for licensed users.

For more information about configuring enterprise policies, see the Managed App Configuration article.

Windows

IT teams managing Windows systems can apply app configuration keys using their preferred administrative tooling by setting values in the following registry path:

txt
HKEY_CURRENT_USER\SOFTWARE\Policies\Progress\Fiddler Everywhere
Key NameDescriptionValue TypeValue Example
DisableAssistantEnables or disables the Debugging AssistantDWORD-32 (hexadecimal)1
DefaultAssistantSettingsSets the default settings (LLM provider, API key, and model) for the Debugging AssistantREG_SZ (string)See the JSON structure here
DisableAssistantSettingsUpdateEnables or disables the option to update the Debugging Assistant settingsDWORD-32 (hexadecimal)1

macOS

IT teams managing macOS systems can apply app configuration using their preferred device management solution (such as Jamf, Intune, or similar) by setting the following keys:

Key NameDescriptionValue TypeValue Example
DisableAssistantEnables or disables the Debugging Assistantinteger1
DefaultAssistantSettingsSets the default settings (LLM provider, API key, and model) for the Debugging AssistantStringSee the JSON structure here
DisableAssistantSettingsUpdateEnables or disables the option to update the Debugging Assistant settingsinteger1

The DefaultAssistantSettings policy expects a JSON object that contains the following properties:

  • provider - Sets the LLM provider. Supports the following values (case-sensitive):
txt
openai
anthropic
azure_openai
google_gemini
  • providerApiKey - Sets the API key for the selected provider.
  • model - Sets a specific model from the selected provider. Available when the provider key is set to openai,anthropic, or google_gemini.
  • azureUri - Sets the Azure OpenAI resource URI . Available only when the provider key is set to azure_openai.

For more information on using managed application configurations, see the Managed Application Policies article.

Configuring the DefaultAssistantSettings Policy

The DefaultAssistantSettings policy expects a JSON object that contains the following properties:

  • provider - Sets the LLM provider. Supports the following case-sensitive values:
txt
openai
anthropic
azure_openai
google_gemini
  • providerApiKey - Sets the API key for the selected provider.
  • model - Sets a specific model from the selected provider. Available when the provider key is set to openai, anthropic, or google_gemini.
  • azureUri - Sets the Azure OpenAI resource URI. Available only when the provider key is set to azure_openai.

Example JSON for setting Anthropic using provider, providerApiKey, and model:

JSON
{
    "provider": "anthropic",
    "providerApiKey": "my-api-key",
    "model": "claude-sonnet-4-20250514"
}

Example JSON for setting Azure using provider, providerApiKey, and azureUri:

JSON
{
    "provider": "azure_openai",
    "providerApiKey": "my-api-key",
    "azureUri": "https://<your-azure-app-endpoint>/openai/deployments/gpt-4.1-mini/chat/completions?api-version=2025-01-01-preview"
}