Skip to content

AddCustomModelRequestBody

Configuration for registering a custom model. Defines the model's capabilities, endpoint, and default parameters.

Example Usage

typescript
import { AddCustomModelRequestBody } from "@meetkai/mka1/models/operations";

let value: AddCustomModelRequestBody = {
  name: "<value>",
  baseUrl: "https://definitive-responsibility.info/",
};

Fields

FieldTypeRequiredDescription
namestring✔️Unique name for the custom model. Must be between 1 and 255 characters. This name is used to reference the model in API requests.
baseUrlstring✔️Base URL of the custom model API endpoint. Must be a valid URL (maximum 512 characters). This is where API requests for this model will be sent.
apiKeystringOptional API key for authenticating with the custom model endpoint. Maximum 512 characters. Include if the custom model requires authentication.
supportsChatbooleanWhether this model supports chat completions (conversational interactions). Set to true if the model can handle multi-turn conversations.
supportsCompletionbooleanWhether this model supports text completions (single-turn text generation). Set to true if the model can generate text from a prompt.
supportsEmbeddingbooleanWhether this model supports text embeddings (vector representations). Set to true if the model can generate semantic embeddings.
supportsImagebooleanWhether this model supports image generation or image understanding. Set to true if the model can work with images.
supportsTranscriptionbooleanWhether this model supports audio transcription (speech-to-text). Set to true if the model can transcribe audio.
supportsSpeechbooleanWhether this model supports speech synthesis (text-to-speech). Set to true if the model can generate audio from text.
maxTokensnumberDefault maximum number of tokens the model can generate in a single response. Must be at least 1. This limits the length of generated outputs.
contextWindownumberThe context window size of the model (total tokens including prompt and completion). Must be at least 1. This determines how much text the model can process at once.
temperaturenumberDefault sampling temperature for the model (0-2). Higher values (e.g., 0.8) make output more random, lower values (e.g., 0.2) make it more deterministic.
topPnumberDefault nucleus sampling parameter (0-1). Alternative to temperature. Only tokens with cumulative probability up to topP are considered.
frequencyPenaltynumberDefault frequency penalty (-2 to 2). Positive values reduce repetition of tokens based on their frequency in the generated text.
presencePenaltynumberDefault presence penalty (-2 to 2). Positive values encourage the model to talk about new topics by penalizing tokens that have appeared.
rpmnumberRate limit in requests per minute for this model. Must be at least 1. Used for throttling requests to prevent overloading the custom model endpoint.