Closed
Description
🚀 The feature, motivation and pitch
This task is to enable DeepSeek R1 Distill Llama 8B on ExecuTorch. That way, people can run these models in a mobile app, locally, without talking to the server.
In theory, ExecuTorch already supports Llama 3.1 8B architecture anyway, so it should just work out of the box (https://github.com/pytorch/executorch/blob/main/examples/models/llama/README.md)
Please document (and make necessary changes) on how to run DeepSeek R1 Distill Llama 8B e2e via ExecuTorch on iOS and Android.
Update 1:
Was able to verify that export works as such: #7981 (comment)
Update 2:
Currently looking into tokenizers
Alternatives
No response
Additional context
No response
RFC (Optional)
No response
cc @cccclai @helunwencser @dvorjackz @byjlw
Metadata
Metadata
Assignees
Labels
Type
Projects
Status
Done