Store Chat Messages & State Without Managing Infrastructure.Check Out DialogueDB
Skip to content

Meta

When using Meta models via AWS Bedrock, llm-exe will make POST requests to https://bedrock-runtime.us-west-2.amazonaws.com/model/{MODEL_ID}/invoke.

Setup

Meta LLama

ts
const llm = useLlm("amazon:meta.chat.v1", {
    model: "llama3-8b-instruct-v1:0", // This is the model id from Bedrock
});

LLama-Specific Options

In addition to the generic options, the following options are available for Meta LLama models on Bedrock.

OptionTypeDefaultDescription
modelstringThe Bedrock model id. Must be specified. See AWS Bedrock Docs
temperaturenumberundefinedMaps to temperature.
maxTokensnumber2048Maps to max_gen_len. See AWS Bedrock Docs
topPnumberundefinedMaps to top_p. See AWS Bedrock Docs
awsRegionstringundefinedAWS Region. Can be set via AWS_REGION environment variable
awsSecretKeystringundefinedAWS Secret Key. Can be set via AWS_SECRET_ACCESS_KEY environment variable
awsAccessKeystringundefinedAWS Access Key. Can be set via AWS_ACCESS_KEY_ID environment variable