Llama3 and Llama2 are ExecuTorch compatible#34101
Llama3 and Llama2 are ExecuTorch compatible#34101ArthurZucker merged 1 commit intohuggingface:mainfrom
Conversation
752ce90 to
402c2cd
Compare
7b7fcec to
ecfc5a9
Compare
ecfc5a9 to
1ccb858
Compare
|
Verified on |
|
@ArthurZucker I see the original tokenizer for |
|
@ArthurZucker do you mind reviewing this PR? |
|
Hey sorry for being late! |
|
We convert them using |
|
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
Llama3_1b and Llama2_7b are ExecuTorch compatible Co-authored-by: Guang Yang <guangyang@fb.com>
What does this PR do?
Llama 2&# is compatible with
ExecuTorch.Note that LLama2&3 in ExecuTorch repo has been fully optimized for SOTA perf using its own model definition and optimization. You can read details in https://github.com/pytorch/executorch/tree/main/examples/models/llama2. The work here is to make the Llama model compatible with
ExecuTorchusing HuggingFace's model definition.Additional Test in
ExecuTorchRunning
Llama-3.2-1BE2E:cmake-out/examples/models/llama2/llama_main --tokenizer_path=tokenizer_llama3_1b.model --model_path=llama3_1b.pte --prompt="My name is"Before submitting
Pull Request section?
to it if that's the case. Llama is ExecuTorch compatible #32505
Who can review?
Anyone in the community is free to review the PR once the tests have passed. Feel free to tag
members/contributors who may be interested in your PR.
@ArthurZucker