Skip to content

FAQ

We want to give our customers the information they need to get the most from our STACKIT AI Model Serving. This FAQ section answers common questions. This helps you quickly find solutions and improve your experience. We encourage you to check these FAQ before contacting our support team, as you might find your answer here.

  • Errors

    Why does my authentication token (aka API-key) not work?

    STACKIT AI Model Serving provides an OpenAI-compatible API. Therefore, the service integrates well with most OpenAI-compatible clients. To use models provided by STACKIT AI Model Serving instead of OpenAI, the following configurations must be adjusted accordingly:

    • API base URL: https://api.openai-compat.model-serving.eu01.onstackit.cloud/v1
    • API key / Authentication token / Secret key: STACKIT AI Model Serving Auth Token (refer to Getting Started with the STACKIT Portal UI to create a STACKIT AI Model Serving Auth Token in the STACKIT Portal UI)

    How can I resolve a "404 Not Found" error from the API?

    This error occurs in case a requested resource cannot be found. Most likely this is due to an incorrect “model” parameter in the request body.
    Beware that all our models are exclusive to their model type (e.g., chat, embedding). This means a chat model cannot be used to compute embeddings, or vice versa.
    Refer to the Getting Started with Shared Models documentation for a list of available models and their types.

    My request result in a "LengthFinishReasonError", especially when working with structured output.

    This problem can be solved by adjusting thefrequency_penaltyparameter. A value of 0.7 or higher has proven to be sufficient.

  • General information

    Which clients can be used with the STACKIT AI Model Serving?

    The public IP address is randomly assigned based on current availabilities, manual selection is not possible.

    Where does my data go?

    We do not store any customer data from the requests. Your data belongs solely to you and is not stored or used by us. We do not train any models using your data.

    What data is used to train the LLMs?

    We serve open-source models only. These models are publicly available on Hugging Face, accompanied by their individual model cards, which provide information on, for example, training data, training procedure, model architecture, and many more details. We do not train these models with any data, nor do we store any customer data.

    Which models are offered?

    With the STACKIT AI Model Serving, we aim to provide state-of-the-art LLMs for our customers. The offered models are selected carefully. An up-to-date table of shared models can be found on Getting Started with Shared Models. We focus on the best OpenAI open-source models while keeping a stable portfolio. If you require additional models that exceed the shared models offering, please create a service request in the STACKIT Help Center.

    I need a specific model. Can you serve it for me?

    In case our model portfolio does not cover your requirements, please create a service request in the STACKIT Help Center; we are happy to hear about your requirements and find a solution that covers your needs.

    Can I use multiple models with a single authentication token? / Do I need different authentication tokens for different model types (e.g. embedding-models, chat-models)?

    An authentication token, known as the STACKIT AI Model Serving Auth Token, is valid for all shared models, including all types of models. A single STACKIT AI Model Serving Auth Token can be used to embed, test, and answer questions in the chat completions API.

    Why does Nextcloud Assistant respond after approx. 5min?

    Nextcloud Assistant works with background tasks in Nextcloud. By default these tasks get picked up every 5 minutes. Refer to the Official Nextcloud documentation for performance improvements.