@opny721

4 Followers
14 Following
7 Posts
Pretty wild night with LLMs. You don't need Gitlab Copilot, there is offline running solutions ready: https://continue.dev/ - alternative to GPT4ALL https://ollama.ai/ - check the models! and heartwarming full-c++ implementation of LLaMA : LLaMA.cpp https://github.com/ggerganov/llama.cpp (works on OpenCL and Radeon GPUs). #llm #largelanguagemodel #largelanguagemodels #llama #llama2
Continue

Continue

Mistral 7B

Grouped query attention and sliding window attention increase efficiency and performance of LLMs. Source of training data not disclosed, but the model is released under a permissible license.

My summary on HFPapers: https://huggingface.co/papers/2310.06825#6527a2be0ef49cfb784b936f

arxiv: https://arxiv.org/abs/2310.06825

#paper #newpaper #llm #arxiv

Paper page - Mistral 7B

Join the discussion on this paper page