LLM connections

In order to start using the LLM Mesh, an administrator first needs to define connections to LLM.

There are two kinds of connections to LLMs:

  • Hosted LLM APIs

  • Locally-running LLMs, using HuggingFace models running on GPUs.

Hosted LLM APIs

The LLM Mesh provides support for a vast number of LLM API providers in order to maximize your options for choosing your preferred LLM provider.

Anthropic

The Anthropic connection provides connection to Anthropic text models. You will need a Anthropic API key.

The Claude, Claude-instant, Claude 2 and Claude 3 models are supported.

AWS Bedrock

The Bedrock connection provides connection to Bedrock text models. You will need:

  • An AWS account with Bedrock access enabled

  • An existing S3 connection with credentials properly setup.

The Bedrock connection provides access to the following Bedrock models:

  • The Anthropic Claude models family (v1, v2)

  • The AI21 Labs Jurassic 2 models family

  • The Cohere Command, Cohere Command Light, and Cohere Embed models

  • The AWS Titan G1 models family

  • The Meta Llama2 Chat model

Text completion, chat completion, and text embedding models are supported.

AWS SageMaker LLM

The SageMaker LLM connection allows connecting to some completion and summarization models deployed as SageMaker endpoints. You will need an existing SageMaker connection.

The following models have builtin handling modes:

  • The Anthropic Claude models family (v1, v2)

  • The AI21 Labs Jurassic 2 models family

  • The Cohere Command and Cohere Command Light models

  • The LLama2 family (v1, v2, v3)

  • Hugging Face models

Limited support for some other models and endpoints is provided through configuration of a custom query template.

Azure OpenAI

The Azure OpenAI connection provides connection to Azure OpenAI text models. You will need:

  • An Azure account with Azure OpenAI enabled

  • A deployed Azure OpenAI service

  • One or several Azure OpenAI model deployments

  • An Azure OpenAI API key

You will need to declare each Azure OpenAI model deployment, as well as the underlying model that is being deployed (for the purpose of cost computation).

Text completion, chat completion, and text embedding models are supported.

As of October 2023, Azure OpenAI Terms and Conditions indicate that Azure will not retain your data for enhancing its models.

Cohere

The Cohere connection provides connection to Cohere text models. You will need a Cohere API key.

The command and command-light models are supported.

Databricks Mosaic AI (previously MosaicML)

The Databricks Mosaic AI connection provides connection to Databricks Foundation Model APIs. You will need an existing Databricks Model Deployment connection.

The Lama2 (chat completion), Mixtral (chat completion), MPT (text completion) and BGE (text embedding) models are supported.

Google Vertex Generative AI

The Google Vertex LLM connection provides connection to Vertex PaLM text models. You will need either a service account key, or OAuth credentials.

The Chat Bison and Gemini Pro models are supported.

OpenAI

The OpenAI connection provides connection to OpenAI text models (GPT 3.5 Turbo, GPT 4). You will need an OpenAI API key (not to be confused with a ChatGPT account). You will be able to select which OpenAI models are allowed

The OpenAI connection supports both text completion and embedding. Only “standard” OpenAI models are supported.

As of October 2023, OpenAI Terms and Conditions indicate that OpenAI will not retain your data for enhancing its models.

Locally-running HuggingFace models

See Running HuggingFace models