You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hey, first of all, amazing work. Thanks for building an open llama model. As the title suggests, I would like to know if this model could be compatible with llamacpp
Thanks
The text was updated successfully, but these errors were encountered:
Amazing, then we can take advantage of all the tools in llama.cpp. Especially the M1 gpu inference. I get 88ms per token on guanaco 13b on my MacBook. which is quite fast considering how big the model is.
Hey, first of all, amazing work. Thanks for building an open llama model. As the title suggests, I would like to know if this model could be compatible with llamacpp
Thanks
The text was updated successfully, but these errors were encountered: