Inference code for facebook LLaMA models with Wrapyfi support
Why do you think that https://github.com/tloen/llama-int8 is a good alternative to wrapyfi-examples_llama
Inference code for facebook LLaMA models with Wrapyfi support
Why do you think that https://github.com/tloen/llama-int8 is a good alternative to wrapyfi-examples_llama