posts/gemma/ #4
Replies: 1 comment 2 replies
-
1 input_ids = tokenizer("I want to move").input_ids
----> 3 hiddens = p.embedding[input_ids]
4 # p.embedding.shape = (256000, 2048)
5 # input_ids.shape = (5,)
6 # hiddens.shape = (5, 2048)
NameError: name 'p' is not defined Could you please, add more code for define |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
posts/gemma/
Transformer-based LLMs seem mysterious, but they don’t need to. In this post, we’ll walk through a modern transformer LLM, Google’s Gemma, providing bare-bones PyTorch code and some intuition for why each step is there. If you’re a programmer and casual ML enthusiast, this is written for you.
https://graphcore-research.github.io/posts/gemma/?utm_source=substack&utm_medium=email
Beta Was this translation helpful? Give feedback.
All reactions