Skip to content

compatibility with llama.cpp #38

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
nivibilla opened this issue Jun 6, 2023 · 2 comments
Closed

compatibility with llama.cpp #38

nivibilla opened this issue Jun 6, 2023 · 2 comments

Comments

@nivibilla
Copy link

Hey, first of all, amazing work. Thanks for building an open llama model. As the title suggests, I would like to know if this model could be compatible with llamacpp

Thanks

@codesoap
Copy link

codesoap commented Jun 6, 2023

Works flawlessly with OpenAlpaca and I strongly assume this means OpenLLaMa will work as well.

Check out ggml-org/llama.cpp#1291.

@nivibilla
Copy link
Author

Amazing, then we can take advantage of all the tools in llama.cpp. Especially the M1 gpu inference. I get 88ms per token on guanaco 13b on my MacBook. which is quite fast considering how big the model is.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants