AI-ToolLab: LLM API Access

Overview

Members of Goethe University can obtain OpenAI-compatible API access to various Large Language Models (LLMs) via the AI-ToolLab.

This service enables experimental access to AI technologies in a protected environment and is provided by studiumdigitale.


Further Information about the Offer

General Information

Endpoints

We use LiteLLM as an API gateway to provide an OpenAI-compatible interface.
The endpoints are as follows:

Azure-compatible:
https://litellm.s.studiumdigitale.uni-frankfurt.de/

OpenAI-compatible:
https://litellm.s.studiumdigitale.uni-frankfurt.de/v1/

Authentication

For API access, you need an individual API key, which will be provided to you after the request.

You can use this key in combination with the endpoint URL in your API requests.

Available Models

The API currently provides access to 65 different models from three hosting categories.

In principle, the offer includes a mix of commercial and open-source models, optimized in various sizes and for different use cases.

The following models are currently available:

🇪🇺 Azure OpenAI (EU Data Zone)

Hosting: Microsoft Azure in the EU
Costs: Paid (budget-based)

The GU uses the Azure OpenAI Service in the EU Data Zone to provide access to commercial models from OpenAI. These models are generally powerful and offer a wide range of functions.

Models:

gpt-5
gpt-5-chat
gpt-5-mini
gpt-5-nano
GPT-5 model series for demanding or simple, yet very fast execution
gpt-4.1
gpt-4.1-mini
gpt-4.1-nano
GPT-4.1 model series
gpt-4o
gpt-4o-mini
GPT-4o model series
o3
o3-mini
o3 model series
codex-miniCoding model
text-embedding-3-large
text-embedding-3-small
Embedding models
🇩🇪 GWDG/KissKI

Hosting: in Germany at the GWDG via the KissKI project and their service chat-ai.academiccloud.de
Costs: Free under fair use conditions

Through the cooperation with GWDG and KissKI, we can offer a variety of open-source and commercial models. These models are hosted locally in Germany and offer high availability.
We have no influence on the availability and number of models, as these are provided by the GWDG.
If a model is not available, we cannot guarantee that it will be available again.

Overview of the GWDG models: GWDG website

Models:

openai-gpt-oss-120bRecommended: Open source model from OpenAI
mistral-large-3-675b-instruct-2512Recommended: large new Mistral model
gemma-3-27b-it
medgemma-27b-it
Google’s Gemma 3, including a model adapted for medical scenarios
glm-4.7
devstral-2-123b-instruct-2512
qwen2.5-coder-32b-instruct
Coding models
internvl3.5-30b-a3bVision model
llama-3.1-sauerkrautlm-70b-instruct
llama-3.3-70b-instruct
meta-llama-3.1-8b-instruct
Llama models
qwen2.5-vl-72b-instruct
qwen3-235b-a22b
qwen3-30b-a3b-instruct-2507
qwen3-30b-a3b-thinking-2507
qwen3-32b
qwen3-coder-30b-a3b-instruct
qwen3-omni-30b-a3b-instruct
qwen3-vl-30b-a3b-instruct
qwq-32b
Various models from the Qwen2 and Qwen3 series with different focuses (reasoning, thinking, vision, etc.)
teuken-7b-instruct-research
deepseek-r1-distill-llama-70b

apertus-70b-instruct-2509
Other available models

🏛️ Goethe University / studiumdigitale

Hosting: Locally on a server at studiumdigitale at the GU
Costs: Free

Chat models:

llama3.1:8bLatest Llama version
llama3:8bProven all-purpose model
llama2:7b Small text model
mistral:7bCompact language model
codellama:7bCode-specialized
ollama_defaultStandard model (Llama 3.1 8B)

Embedding models:

all-minilm:33m
bge-large:335m
bge-m3:567m
granite-embedding:278m
mxbai-embed-large:335m
nomic-embed-text:v1.5
paraphrase-multilingual:278m
snowflake-arctic-embed2:568m
snowflake-arctic-embed:335m

We update the list depending on availability, demand, and our capabilities.

Instructions: Retrieving current model information

Since the available models change regularly, it is recommended to retrieve the current information directly via the API.

For the following queries, you must have a basic technical knowledge of how to use APIs

/models – Short model overview

Content: Simple list of all available model IDs
Usage: Quick overview of which models are available

Request via CURL:

Example response:

/model/info – Detailed model information

Content: Complete information on all models including hosting and description
Usage: Decision support for model selection

Request via CURL:

Example response (abbreviated):

Important information from the API response:

  • model_name: The exact name for API calls
  • description: Short description of the model
  • hosted_by: Hosting provider and location
  • input_cost_per_token/output_cost_per_token: Cost structure (0 = free)
Note: Hosting locations and data security

⚠️ Important note:
Pay special attention to the hosted_by-Feld, der model/info Route as there are three different hosting locations:

  • Azure OpenAI Service in EU Data Zone – Microsoft Azure (EU)
  • KissKI/GWDG in Göttingen – Germany (GWDG)
  • studiumdigitale, Goethe University Frankfurt – Local (Germany)

Depending on the hosting location, different data protection regulations and security guidelines apply. Choose the model that is suitable for your application according to the sensitivity of your data.

Note: Budget for the use of paid models

A starting budget of €50 is currently provided for each API key issued. The allocation of API keys is still in the testing phase. Further rules for budget usage (e.g., top-ups) and an option to view the remaining budget will follow shortly.

The use of GWDG and GU models is free of charge; only the Azure OpenAI models are subject to a fee.

Usage

You can use the API access like a standard OpenAI API access. For this you need your individual API key and the endpoint.

Usage in OpenAI-compatible tools and WebApps

If you are using an OpenAI-compatible tool, and this tool offers the possibility to configure an API key and an OpenAI Proxy/Server/Endpoint, you can use your individual API key and the endpoint. Try both endpoints (with and without /v1/ at the end) to see which one works.

Make sure that you only use the API key with trusted applications, as this grants access to your LLM API account.

!

Usage in different programming languages

You can use the API in any programming language that supports HTTP requests. Here is an example of using the API in Python with the requests library:

Python with requests-Library
Python with openai-Library

with other libraries

For examples of using the API in other programming languages or libraries, please visit the LiteLLM documentation.


Support

If you have any problems or questions, write an email to the AI-ToolLabs team.