Skip to content

Reproduce/enable DeepSeek R1 Distill Llama 8B #7981

Closed
@mergennachin

Description

@mergennachin

🚀 The feature, motivation and pitch

This task is to enable DeepSeek R1 Distill Llama 8B on ExecuTorch. That way, people can run these models in a mobile app, locally, without talking to the server.

In theory, ExecuTorch already supports Llama 3.1 8B architecture anyway, so it should just work out of the box (https://github.com/pytorch/executorch/blob/main/examples/models/llama/README.md)

Please document (and make necessary changes) on how to run DeepSeek R1 Distill Llama 8B e2e via ExecuTorch on iOS and Android.

Update 1:

Was able to verify that export works as such: #7981 (comment)

Update 2:

Currently looking into tokenizers

Alternatives

No response

Additional context

No response

RFC (Optional)

No response

cc @cccclai @helunwencser @dvorjackz @byjlw

Metadata

Metadata

Assignees

Labels

good first issueGood for newcomersmodule: llmIssues related to LLM examples and apps, and to the extensions/llm/ codemodule: user experienceIssues related to reducing friction for userstriagedThis issue has been looked at a team member, and triaged and prioritized into an appropriate module

Type

No type

Projects

Status

Done

Milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions