-
Notifications
You must be signed in to change notification settings - Fork 115
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
medium based on llama 3.1 #226
Comments
Hi, we are working on it actively right now. Looking forward to sharing good news soon! |
@jeffrey-fong would it possible to release the datasets or atleast lora adapters? |
Heavily awaiting llama3.1-8b-functionary-medium-128k🤩 |
Hi @xdevfaheem , @themrzmaster we have just released our new model: meetkai/functionary-small-v3.1 that is based on: meta-llama/Meta-Llama-3.1-8B-Instruct |
Hi @xdevfaheem , @themrzmaster we have also released our 70b model: functionary-medium-v3.1 |
That's awesome! Great work @khai-meetkai. TQSM |
@khai-meetkai Thanks a lot for the great work! Any timeline for functionary-medium-v3.1/3.2 quantized with AWQ? |
I tried to develop an AWQ quantized version of functionary-medium-v3.2 using AutoAWQ's quantized scripts (full GPU, GPU-CPU offload is bugged with the transformers version used for Llama 3.1). Unfortunately, it seems that I may have done something wrong as it is not performing very well, or at least way below my expectations compared to the AWQ 3.0 version. I used AutoAWQ's quantization script as follows (I manually edited config.json to replace FunctionaryForCausalLM by LlamaForCausalLM in architectures) :
|
would be really nice to have a functionary version of llama 3.1 70b/8b!
The text was updated successfully, but these errors were encountered: