Using Hugging Face model card name in export_llama #8872
Labels
module: llm
Issues related to LLM examples and apps, and to the extensions/llm/ code
triaged
This issue has been looked at a team member, and triaged and prioritized into an appropriate module
🚀 The feature, motivation and pitch
Currently, user need to manually download hugging face safetensors, convert to llama_transformer format, and load the checkpoint and config for the export and inference.
It would be great to directly download and cache (don't have to load it again) the converted checkpoints, and do the inference. Similar to what mlx does:
Alternatives
No response
Additional context
No response
RFC (Optional)
No response
cc @mergennachin @cccclai @helunwencser @jackzhxng
The text was updated successfully, but these errors were encountered: