Full control over your AI applications

Governance and FinOps for your AI agents and applications: Keep track of costs at all times with Token Control and manage access to your AI models efficiently and securely.

Seamless integration with leading AI providers

We support all common AI models and thus enable you to flexibly connect your existing AI solution.

AI costs always in mind

feature-analytics
Features

Managing your AI budgets & limits

Manage the AI budgets of your user groups and users centrally. Define budget limits for departments, applications, projects, or individual users. Distribute your AI budgets according to the needs of your company or organization and avoid unplanned costs.

Features

Monitor & analyze your AI costs

Monitor your AI costs and the budget status of your users clearly and in real time. Get monthly cost reports for your internal billing processes. Analyze your AI consumption based on your user groups, users, and budgets.

feature-connection
feature-connection
Features

Plug-and-play integration with existing AI solutions

Integrate Token Control into your AI solution with our transparent plug-and-play approach. It's simple and requires no additional development effort. And of course, it also works with white duck PrivateGPT Chat.

suad
“With Token Control, you avoid hidden costs caused by more and more complex agent systems.”

Suad Wolgram, Product Manager, white duck GmbH

AI governance under control

Features

Organization of your business entities

Manage your business entities centrally by clearly structuring user groups, departments, cost centers, or solutions such as AI agents, chats, and applications. Manage access rights and budgets in a targeted manner to ensure transparency, efficiency, and control over your AI models.

feature-connection
feature-analytics
Features

API key management for your AI solutions

Create and manage dedicated API keys for your AI solutions to control access individually and ensure security. Enable targeted use by specific applications, projects, or user groups and maintain an overview of resource control for your AI models.

Features

Control access to your LLM models

Monitor and manage access to your LLM models with clearly defined TPM (tokens per minute) limits per application. Set individual usage limits for applications, teams, or projects to ensure regulated resource usage and guarantee optimal performance of your AI models.

feature-connection

Start now

SaaS

For a quick and easy start to your AI management.

Fully managed solution

Integrate with your existing AI solution

Pay-as-you-Go

Data residency in Germany

Start now

Privat

For your AI solution with complete data sovereignty.

Deployed in your own Azure environment

Isolated data storage

Tailored to your individual needs

Operated by us

Start now

Partners

For your market as our partner.

Dedicated multi-tenant environment

Your individual branding

Integrated with your solutions

Tailored to your business model

Start now
Schattdecor SE

With NOVA—our customized AI assistant—we can use AI securely and in compliance with data protection regulations within our company. Thanks to its seamless integration into our corporate environment, intuitive operation, and tailored responses. Working with white duck is always professional and straightforward, meaning short communication channels, quick feedback, and a strong team with diverse AI and cloud native expertise.

University of Applied Sciences Rosenheim

white duck GmbH as our innovation partner!
They advise us, provide us with some concepts and a turnkey solution with full cost control for our computer science, business informatics and artificial intelligence students. This enables us to adapt teaching concepts and also provide students with lightweight programming interfaces. The first projects for industrial partners are already underway.

FAQ

Which AI models are supported?

keyboard_arrow_down

Token Control supports all common LLM models, including:

Azure AI Foundry:
All common OpenAI Modelle, Microsoft Modelle (Phi, model-router), Mistral AI, Meta, DeepSeek, xAI, Nvidia as well as Hugging Face and open source models.

(Azure) OpenAI:
GPT-5 (Chat, Mini, Nano), GPT-OSS, GPT-4.1, GPT-4o, 03-mini, o1, o1-mini, GPT-Image-1, Text-embedding-3

Google Gemini:
Gemini 2.5 (Pro, Flash, Flash-lite)

Mistral AI:
Medium 3.1, Magistral Medium 1.1, Codestral 2508, 3B, Medium 3

Meta Llama:
Llama 4 (Maverick, Scout, Behemoth)

In addition, Token Control supports all open-source LLM models with an OpenAI-compliant API.

How is integration into existing AI solutions implemented?

keyboard_arrow_down

Token Control can be easily integrated into your existing AI solutions and enables simple, efficient implementation—without any additional development effort.

Seamless integration:
All requests sent to your AI models are processed by Token Control and forwarded unchanged. This preserves the full functionality of your existing systems.

No code changes required:
AI solutions that already use Azure OpenAI / Azure AI Foundry or other supported models can be operated directly with Token Control without any changes to the code. This saves valuable time and resources and enables quick and easy implementation.

Flexible model management and scaling:
Token Control supports the scaling of AI models and the management of multiple model deployments. This allows you to run different models in parallel and respond flexibly to the requirements of your applications, projects, or teams.

Can I integrate Token Control into third-party solutions?

keyboard_arrow_down

Token Control can be easily integrated into third-party solutions. Thanks to its plug-and-play architecture, it can be quickly and easily integrated into existing AI chats, chatbots, and tools such as GitHub Copilot Chat. You can seamlessly integrate Token Control into your existing systems without any complex adjustments or additional development work.

Integration is achieved through the use of dedicated API keys, which ensure a secure and controlled connection between Token Control and your third-party solutions. These API keys allow you to precisely control access to your AI models and efficiently monitor usage. This gives you full control over your AI resources at all times and enables you to implement governance and cost policies in a targeted manner.

Support for web search or deep search?

keyboard_arrow_down

Token Control offers comprehensive support for grounding and deep search through seamless integration with leading AI agent frameworks such as FLOCK (our open-source AI agent framework), Langgraph, LangChain, Semantic Kernel, and Autogen. These frameworks enable you to connect your AI models to external data sources, delivering more accurate and contextually relevant results.

Support for web search is a key part of our roadmap and has the highest priority. If you have specific requirements or use cases, please contact us to discuss your needs and work together to develop the optimal solution.

Is RAG (Retrieval-Augmented Generation) and Grounding supported?

keyboard_arrow_down

Token Control fully and directly supports RAG (Retrieval-Augmented Generation) and Grounding “out-of-the-box.” This allows you to seamlessly connect your AI models to external knowledge sources or databases to generate more accurate and context-aware responses. Thanks to its simple and efficient implementation, you can use these technologies without additional development effort, significantly increasing the performance of your AI solutions.

Is administration possible via Microsoft Entra ID groups?

keyboard_arrow_down

Token Control supports management via Microsoft Entra ID groups (formerly Azure AD). You can use M365 and security groups to map API keys and business identities such as user groups, departments, or cost centers. This enables centralized and structured control of access to your AI resources based on existing organizational structures.

Is the integration with an API gateway possible?

keyboard_arrow_down

Token Control can be easily integrated with an API gateway (e.g., Azure API Management). While Token Control solves organizational challenges such as the management and reporting of API keys, user groups, departments, or cost centers, the API gateway handles technical aspects such as routing and load balancing. Together, they offer a comprehensive solution that addresses both organizational and technical hurdles.

Does Token Control offer an MCP server?

keyboard_arrow_down

An MCP server for detailed analyses and comprehensive reporting is currently in planning and is a central component of our roadmap. With this MCP server, we want to offer you expanded options in the future with the help of AI for overseeing, evaluating, and optimizing your AI usage. This feature is not yet available, but we are working hard to implement it. If you have specific requirements or requests, please contact us—we look forward to incorporating your needs into the development process.

Is Token Control GDPR compliant?

keyboard_arrow_down

Token Control is fully GDPR compliant and was developed with a clear focus on data protection and data security. All data is processed and hosted exclusively within the EU, using the proven infrastructure of Microsoft Azure. By using Azure's EU Data Boundary, we ensure that all data flows and storage comply with the strict requirements of the European General Data Protection Regulation (GDPR).

Your partner for Microsoft Azure & AI

Microsoft

Microsoft Solution Partner

Protect

More than 13 years on the market

European Union Circular Flag

With ❤️ from Germany