You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Instructions of how to enable this model for ExecuTorch:
Export the model to ExportIR. For LLM, to run with performance, typically you will need to export the model with cache. Llama3 and Llama2 are ExecuTorch compatible #34101 is a reference of how to export and validate the model. Note that you may run into some export issue and it may require fixes in the modeling code.
(Optional) Congrats! Once you complete step 1-3, you will be able to run the model on a host machine. Now if you would to go further like making the model faster, smaller, cheaper for your model use-case, you can create more complicated recipes with quantizations and delegations for different HW accelerators. You can find more tutorials on our website, for example to optimize and run the model with Core ML on Apple’s platform: https://pytorch.org/executorch/stable/build-run-coreml.html
Feature request
Feature request
Enable SmolLM to "Export to ExecuTorch" workflow.
Instructions
Instructions of how to enable this model for ExecuTorch:
(Optional) Congrats! Once you complete step 1-3, you will be able to run the model on a host machine. Now if you would to go further like making the model faster, smaller, cheaper for your model use-case, you can create more complicated recipes with quantizations and delegations for different HW accelerators. You can find more tutorials on our website, for example to optimize and run the model with Core ML on Apple’s platform: https://pytorch.org/executorch/stable/build-run-coreml.html
Motivation
See details in #32253
Your contribution
TBD
The text was updated successfully, but these errors were encountered: