Skip to content

Service description

Our service enables you to use the provided models as Shared Models. The term Shared Models refers to models that are used communally by all clients. Through the shared hosting of our LLMs, we enable a large number of users to cost-effectively access these powerful models and utilize them for their specific applications. There will be no data exchange between projects or even between different customers. To prevent misuse and service overuse we enforce rate-limits. Read more about rate-limits in Rate Limits with AI.

STACKIT AI Model Serving provides two APIs to interact with our service:

  • STACKIT AI Model Serving API (Product API): This API allows you to manage your authentication tokens and get information about the shared LLMs. To get started with the Control Pane API, please refer to Manage auth tokens and have a look at the API examples there.
  • Inference API (OpenAI Compatible API): This API is used to send requests to and receive responses from the hosted models. To use the inference API, you will need to authenticate yourself with an auth token. For more information, please refer to our Getting started with shared models guide.