Supported LLMs

AI Studio supports a variety of LLMs, enabling you to choose the right model for a given task. Whenever possible we provision redundant resources spanning different cloud regions and providers to maximize reliability. For an overview of how AI studio interacts with LLMs please review the section on Prompt AI Behaviors.

In order to insure proper resource allocation, some LLMs require you to contact Quiq Support to be enabled.

LLMDescriptionExternal API ReferenceExample FunctionsContact Support?
gpt-35-turbo-1106ChatCompletion style built by OpenAI, available through both OpenAI and Azure cloud offerings. Fast given the level of accuracy.

- completion maps to /choices[0]["message"]["content"] in the response
LinkLink No
gpt-35-turbo-0125ChatCompletion style built by OpenAI, available through both OpenAI and Azure cloud offerings. Fast given the level of accuracy.

- completion maps to /choices[0]["message"]["content"] in the response
Link Link No
gpt-4o-0513ChatCompletion style built by OpenAI, available through both OpenAI and Azure cloud offerings. Slower but generally more accurate than 35-turbo models.

- completion maps to /choices[0]["message"]["content"] in the response
LinkLinkYes
gpt-4o-0806ChatCompletion style built by OpenAI, available through both OpenAI and Azure cloud offerings. Slower but generally more accurate than 35-turbo models.

- completion maps to /choices[0]["message"]["content"] in the response
Link Link Yes
gpt-4o-mini-0718ChatCompletion style built by OpenAI, available through both OpenAI and Azure cloud offerings. Slower but generally more accurate than 35-turbo models.

- completion maps to /choices[0]["message"]["content"] in the response
Link Link No
llama3-8b-instruct-v1The 8B parameter variant of the Llama 3 Chat family of LLMs built by Meta. Hosted on AWS Bedrock.

- completion maps to /generation in the response
Link Link Yes
llama3-70b-instruct-v1The 70B parameter variant of the Llama 3 Chat family of LLMs built by Meta. Hosted on AWS Bedrock.

- completion maps to /generation in the response
Link Link Yes
llama3-1-8b-instruct-v1The 8B parameter variant of the Llama 3 Chat family of LLMs built by Meta. Hosted on AWS Bedrock.

- completion maps to /generation in the response
Link Link Yes
llama3-1-70b-instruct-v1The 70B parameter variant of the Llama 3 Chat family of LLMs built by Meta. Hosted on AWS Bedrock.

- completion maps to /generation in the response
Link Link Yes
llama3-2-1b-instructThe 1B parameter variant of the Llama3 Chat family of LLMs built by Meta. Hosted on AWS Bedrock.

- completion maps to /generation in the response
Link Link Yes
llama3-2-3b-instructThe 3B parameter variant of the Llama3 Chat family of the LLMs built by Meta. Hosted on AWS Bedrock.

- completion maps to /generation in the response
Link Link Yes
llama3-2-11b-instructThe 11B parameter variant of the Llama3 Chat family of the LLMs built by Meta. Hosted on AWS Bedrock.

- completion maps to /generation in the response
Link Link Yes
llama3-2-90b-instructThe 90B parameter variant of the Llama3 Chat family of the LLMs built by Meta. hosted on AWS Bedrock and GCP.

- completion maps to /generation in the response
Link Link Yes
claude-3-haikuClaude V3 Haiku model by Anthropic - hosted on AWS Bedrock.

- completion maps to /content/text in the response
Link Link No
claude-3-5-haikuClaude V3.5 Haiku model by Anthropic - hosted on AWS Bedrock.

- completion maps to /content/text in the response
Link Link Yes
claude-3-sonnetClaude V3 Sonnet model by Anthropic - hosted on AWS Bedrock.

- completion maps to /content/text in the response
Link Link Yes
claude-3-5-sonnetClaude V3.5 Sonnet model by Anthropic - hosted on AWS Bedrock.

- completion maps to /content/text in the response
Link Link Yes
claude-3-5-sonnet-v2Claude V3.5 V2 Sonnet model by Anthropic - hosted on AWS Bedrock.

- completion maps to /content/text in the response
Link Link Yes
gemini-1.5-proGoogles Gemini 1.5 Pro model which tends to be slower but generally more accurate.

- completion maps to /candidates/0/content/parts/0/text
Link Link Yes
gemini-1.5-pro-002Googles Gemini 1.5 Pro 002 model which tends to be faster but generally more accurate than the Pro model.

- completion maps to /candidates/0/content/parts/0/text
Link Link Yes
gemini-1.5-flashGoogles Gemini 1.5 Flash model which tends to be faster than the pro version.

- completion maps to /candidates/0/content/parts/0/text
Link Link Yes
gemini-1.5-flash-002Googles Gemini 1.5 Flash 002 model which tends to be faster than the 1.5 Flash version.

- completion maps to /candidates/0/content/parts/0/text
Link Link No

Notes

  • When building prompt payloads in the Function Editor you can assume that the request object should generally mimic the request payload of the underlying LLM according to its docs, but the framework may override certain parameters. See the example functions
  • Some LLM request payloads specify the model in the payload. AI studio will implicitly set/override these parts of your prompt objects
  • When an LLM is invoked on behalf of a resource transformation we may override settings such as temperature to produce the most deterministic result

We continually uptake new LLMs and have preliminary support for additional models not listed in this documentation. Contact Quiq support or your account rep if you have specific LLM requirements

Custom LLMs

Custom LLMs enable you to communicate with your own LLM resources rather than Quiq's managed resources. For example, if you've created your own fine-tuned ChatCompletion model hosted by OpenAI, you can configure AI studio to use your fine-tuned resource. Configuring a Custom LLM requires the following information:

  • The URL where the LLM is hosted
  • Appropriate credentials enabling AI Studio to securely access the LLM
  • The family/style of LLM resource (for purposes of matching the API contract)

In order to register a Custom LLM, contact Quiq Support.

Embeddings Models

The following table lists the supported models for embedding your AI Resources for purposes of semantic search.

ModelDescriptionContact Support?
text-ada-002-embeddingBuilt by OpenAI, available through both OpenAI and Azure cloud offeringsNo
titan-embed-text-v1Built by Amazon and and available through AWSYes
text-embedding-3-smallBuilt by OpenAI, available through both OpenAI and Azure cloud offeringsYes
text-embedding-3-largeBuilt by OpenAI, available through both OpenAI and Azure cloud offeringsYes