-
Notifications
You must be signed in to change notification settings - Fork 431
How to OPT model with PowerInfer? #234
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Due to limited bandwidth, this part of the model support hasn't been merged to main branch yet. We plan to release this part recently, and we will release the OPT related code as soon as possible. Stay tuned. |
Hi, I want to know when the OPT model and its related code are expected to be released? |
|
@wuooo339 @Ryuukinn55 Hi everyone! Our code for the OPT model has been officially released. Our predictor is now available on HuggingFace: https://huggingface.co/PowerInfer/OPT-7B-predictor. For other model sizes, such as 13B or larger, we will release the predictors soon, within the next few days. You can convert the model from the original version at https://huggingface.co/facebook/opt-6.7b using the
For any other questions, please feel free to ask! |
@YixinSong-e llm_load_gpu_split_with_budget: error: activation files under '/share-data/wzk-1/model/powerinfer/activation' not found |
sorry ,I found the activation in the predictor file.And the model after convertion should be put in the same file. |
Uh oh!
There was an error while loading. Please reload this page.
Prerequisites
Before submitting your question, please ensure the following:
Question Details
I have read your article of PowerInfer and I have seen you use OPT-30B to compare with llama.cpp.But I can not read any information about OPT in this READMD?
Additional Context
I want to test OPT model using PowerInfer.So I might need your help?
I am from Harbin Institude of Technology studying HPC(high performance computing).And I am trying different offloading strategies recently.
The text was updated successfully, but these errors were encountered: