Skip to content

Conversation

subhaviv
Copy link
Contributor

Added Mistral 2 Large and Llama 3.1 models on Amazon Bedrock

Copy link
Collaborator

@yifanmai yifanmai left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Left a minor comment; looks good otherwise. Thanks!

- name: amazon/llama3-1-405b-instruct-v1:0
model_name: meta/amazon-llama3-1-405b-instruct-v1:0
tokenizer_name: huggingface/gpt2
max_sequence_length: 2000
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Does the Bedrock version have a length of 2000 tokens? Most other versions of Llama 3.1 have a longer sequence length of 128K tokens.

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Discussed over email - 2000 tokens is the correct length as per the Bedrock docs.

@yifanmai yifanmai merged commit 2b1fad5 into stanford-crfm:main Oct 28, 2024
8 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants