Quoted By:
A general dedicated to the discussion and development of local language models.
AIcg niggers need to go back.
Previous Threads: >>97923309 & >>97918783
►News
>(12/15) 3x prompt processing speedup in llama.cpp for Mixtral
>(12/14) Oobabooga adds Mixtral support for Nvidia and CPU-only
>(12/13) Theoretical possibility of 1Bit quantization of MoE layers https://github.com/ggerganov/llama.cpp/issues/4445#issuecomment-1854380383
>(12/13) Koboldcpp merged Mixtral support
>(12/13) Mixtral support added for llama.cpp
>(12/12) Mistral 7B v0.2 release
>(12/08) Mistral released 8x7B MoE model & Together.ai released StripedHyena-7B SSM
>(12/06) Oobabooga adds QuIP# support
>(11/29) DeepSeek-67B released
>(11/28) Yuan 2 released (2B,51B,102B) & Qwen-72B released
>(11/09) Open LLM Leaderboard adds 3 new benchmarks
>(11/05) EXL2 adds 8-bit cache for decreased VRAM usage
>(11/04) koboldcpp-1.48.1 adds Context Shifting for less/no reprocessing (not Smart Context)
>(11/02) Yi-34B model released (200k context also available): https://huggingface.co/01-ai/Yi-34B
►FAQ: https://rentry.org/er2qd
►Links: https://rentry.org/LocalModelsLinks
►Getting Started
https://rentry.org/MixtralForRetards
https://rentry.org/llama-mini-guide
https://rentry.org/8-step-llm-guide
https://rentry.org/llama_v2_sillytavern
https://rentry.org/lmg-spoonfeed-guide
►Further Learning
https://rentry.org/machine-learning-roadmap
https://rentry.org/llm-training
https://rentry.org/LocalModelsPapers
►Benchmarks
General Purpose: https://hf.co/spaces/HuggingFaceH4/open_llm_leaderboard
Programming: https://hf.co/spaces/bigcode/bigcode-models-leaderboard
Censorbench: https://codeberg.org/jts2323/censorbench
►ERP/RP Datasets
https://rentry.org/qib8f
https://rentry.org/ashh2
►Alpha Calculator
https://desmos.com/calculator/ffngla98yc
►Text Gen. UI
https://github.com/oobabooga/text-generation-webui
https://github.com/LostRuins/koboldcpp
https://github.com/lmg-anon/mikupad
https://github.com/turboderp/exui
AIcg niggers need to go back.
Previous Threads: >>97923309 & >>97918783
►News
>(12/15) 3x prompt processing speedup in llama.cpp for Mixtral
>(12/14) Oobabooga adds Mixtral support for Nvidia and CPU-only
>(12/13) Theoretical possibility of 1Bit quantization of MoE layers https://github.com/ggerganov/llama.cpp/issues/4445#issuecomment-1854380383
>(12/13) Koboldcpp merged Mixtral support
>(12/13) Mixtral support added for llama.cpp
>(12/12) Mistral 7B v0.2 release
>(12/08) Mistral released 8x7B MoE model & Together.ai released StripedHyena-7B SSM
>(12/06) Oobabooga adds QuIP# support
>(11/29) DeepSeek-67B released
>(11/28) Yuan 2 released (2B,51B,102B) & Qwen-72B released
>(11/09) Open LLM Leaderboard adds 3 new benchmarks
>(11/05) EXL2 adds 8-bit cache for decreased VRAM usage
>(11/04) koboldcpp-1.48.1 adds Context Shifting for less/no reprocessing (not Smart Context)
>(11/02) Yi-34B model released (200k context also available): https://huggingface.co/01-ai/Yi-34B
►FAQ: https://rentry.org/er2qd
►Links: https://rentry.org/LocalModelsLinks
►Getting Started
https://rentry.org/MixtralForRetards
https://rentry.org/llama-mini-guide
https://rentry.org/8-step-llm-guide
https://rentry.org/llama_v2_sillytavern
https://rentry.org/lmg-spoonfeed-guide
►Further Learning
https://rentry.org/machine-learning-roadmap
https://rentry.org/llm-training
https://rentry.org/LocalModelsPapers
►Benchmarks
General Purpose: https://hf.co/spaces/HuggingFaceH4/open_llm_leaderboard
Programming: https://hf.co/spaces/bigcode/bigcode-models-leaderboard
Censorbench: https://codeberg.org/jts2323/censorbench
►ERP/RP Datasets
https://rentry.org/qib8f
https://rentry.org/ashh2
►Alpha Calculator
https://desmos.com/calculator/ffngla98yc
►Text Gen. UI
https://github.com/oobabooga/text-generation-webui
https://github.com/LostRuins/koboldcpp
https://github.com/lmg-anon/mikupad
https://github.com/turboderp/exui