fosstodon.org is one of the many independent Mastodon servers you can use to participate in the fediverse.
Fosstodon is an invite only Mastodon instance that is open to those who are interested in technology; particularly free & open source software. If you wish to join, contact us for an invite.

Administered by:

Server stats:

9.8K
active users

#koboldcpp

0 posts0 participants0 posts today
RainI want to do a good write up in my README for PixelPolygot as one of my last touches but I need the damn <a class="hashtag" href="https://melonbread.dev/tag/rocm" rel="nofollow noopener noreferrer" target="_blank">#rocm</a> fork of <a class="hashtag" href="https://melonbread.dev/tag/koboldcpp" rel="nofollow noopener noreferrer" target="_blank">#KoboldCpp</a> to update so I can do some more testing with Qwen2.5-VL locally. Like it works with vulkan on the main branch but way slower than rocm. :notlikeblob:
Reverend Elvis<p>DeepSeek and the big step forward<br>Yesterday, with very little effort, I managed to get a working <a href="https://social.undeadnetwork.de/tags/DeepSeek" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>DeepSeek</span></a> model running on my computer. It felt like I was a super-rich tech oligarch, but I just have a simple casual gaming computer with a cheap 8 GB graphics card for playing a game or editing videos. A lot of people have a machine like that; it's below average. And it took me half an hour. Hauke explains here how anyone can do the same. <a href="https://tube.tchncs.de/w/cuNhJ5tRzdv4WRE2oZjCv3" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">tube.tchncs.de/w/cuNhJ5tRzdv4W</span><span class="invisible">RE2oZjCv3</span></a></p><p>You only need the DeepSeek-R1-Distill-Qwen model (32 billion parameters, approx. 19 GB)<a href="https://huggingface.co/mradermacher/DeepSeek-R1-Distill-Qwen-32B-Uncensored-GGUF" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">huggingface.co/mradermacher/De</span><span class="invisible">epSeek-R1-Distill-Qwen-32B-Uncensored-GGUF</span></a>, which is a very slimmed-down version that will also run on a home computer. For institutes, educational institutions and companies, there are also more powerful models, right up to the full model, which then really requires enormous resources. And you need a program like <a href="https://social.undeadnetwork.de/tags/KoboldCpp" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>KoboldCpp</span></a> <a href="https://github.com/LostRuins/koboldcpp" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="">github.com/LostRuins/koboldcpp</span><span class="invisible"></span></a> which loads the AI model into the graphics card and, for a weak card like mine, offloads the rest to the CPU. Furthermore, a chat GUI is provided via http://localhost:5001. As I said, if you've ever used a console, it's really not rocket science.</p><p>There will certainly be loud voices that want to ban DeepSeek. Racists and capitalists will shout that the Chinese cannot be trusted. But it is open-source. Under an MIT license. I also ran it offline. It runs 100% locally.</p><p>What you could do with it! If smart people continue to train the models. Now everyone has an AI and the tech bros can pack up! Thanks Deepseek and thanks China. A huge gift for the open-source community and maybe for humanity.</p><p>PS. And once again, the conspiracy theorists were right, claiming that commercial AIs are only so large and resource-hungry so that they are reserved exclusively for the rich and the people remain customers. That's over now.</p><p><a href="https://word.undead-network.de/2025/02/10/deepseek-and-the-big-step-forward/" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">word.undead-network.de/2025/02</span><span class="invisible">/10/deepseek-and-the-big-step-forward/</span></a><br><a href="https://social.undeadnetwork.de/tags/ai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ai</span></a> <a href="https://social.undeadnetwork.de/tags/china" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>china</span></a> <a href="https://social.undeadnetwork.de/tags/deepseek" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>deepseek</span></a> <a href="https://social.undeadnetwork.de/tags/FOSS" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>FOSS</span></a> <a href="https://social.undeadnetwork.de/tags/ki" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ki</span></a> <a href="https://social.undeadnetwork.de/tags/openai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>openai</span></a> <a href="https://social.undeadnetwork.de/tags/opensourceAI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>opensourceAI</span></a> <a href="https://social.undeadnetwork.de/tags/opensoure" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>opensoure</span></a> <a href="https://social.undeadnetwork.de/tags/oss" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>oss</span></a> <a href="https://social.undeadnetwork.de/tags/techbros" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>techbros</span></a></p>
Reverend Elvis<p>DeepSeek und der große Sprung nach vorn<br>Mir ist es gestern mit sehr kleinem Aufwand gelungen, ein funktionierendes <a href="https://social.undeadnetwork.de/tags/DeepSeek" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>DeepSeek</span></a> Model auf meinen Rechner zum Laufen zu bekommen. Hat sich angefühlt, als wäre ich jetzt superreicher Techoligarch, aber ich habe nur einen simplen Casual Gamer Rechner mit einer billigen 8 GB Grafikkarte, um mal ein Spiel zu spielen oder Videos zu bearbeiten. So einen Rechner haben viele Menschen, es ist unterer Durchschnitt. Und es hat mich eine halbe Stunde gekostet.&nbsp; Wie das für jeden nachzumachen ist, erklärt Hauke hier. <a href="https://tube.tchncs.de/w/cuNhJ5tRzdv4WRE2oZjCv3" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">tube.tchncs.de/w/cuNhJ5tRzdv4W</span><span class="invisible">RE2oZjCv3</span></a></p><p>Man benötigt nur das Model DeepSeek-R1-Distill-Qwen (32 Mrd. Parameter, ca. 19 GB) <a href="https://huggingface.co/mradermacher/DeepSeek-R1-Distill-Qwen-32B-Uncensored-GGUF" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">huggingface.co/mradermacher/De</span><span class="invisible">epSeek-R1-Distill-Qwen-32B-Uncensored-GGUF</span></a> das ist eine sehr abgespeckte Version, die eben auch auf einem Home Rechner läuft. Für Institute, Bildungseinrichtungen und Firmen gibt es auch stärkere Modelle bis zum vollen Modell, welches dann wirklich enorme Ressourcen benötigt. und man braucht ein Programm wie <a href="https://social.undeadnetwork.de/tags/KoboldCpp" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>KoboldCpp</span></a> <a href="https://github.com/LostRuins/koboldcpp" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="">github.com/LostRuins/koboldcpp</span><span class="invisible"></span></a> welche das KI-Model in die Grafikkarte lädt und bei einer schwachen Karte, wie bei mir, den Rest auf die CPU auslagert. Des Weiteren wird eine Chat GUI über http://localhost:5001 zur verfügung gestellt. Wie gesagt, wenn man schon mal eine Konsole benutzt hat, ist das echt kein Hexenwerk.</p><p>Es wird ganz sicher laute Stimmen geben, die DeepSeek verbieten wollen. Rassisten und Kapitalisten die rufen werden das man den Schlitzau... den Chinesen nicht trauen kann. Nur es ist Open-Source. Unter MIT-Lizenz. Ich habe es außerdem offline betrieben. Es läuft 100% lokal.</p><p>Was man damit alles machen könnte! Wenn kluge Menschen die Modelle weiter trainieren. So haben jetzt alle eine KI und die Techbros können einpacken! Danke Deepseek und Danke China. Für die Open-Source-Community ein riesengroßes Geschenk und vielleicht auch für die Menschheit.</p><p>PS. und einmal mehr hatten die Aluhüte recht, die behaupteten, dass die kommerziellen KIs nur so groß und Ressourcen hungrig sind, damit sie exklusiv den Reichen vorbehalten sind und das Volk Kunde bleibt. Das ist jetzt vorbei.</p><p><a href="https://word.undead-network.de/2025/02/10/deepseek-und-der-grosse-sprung-nach-vorn/" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">word.undead-network.de/2025/02</span><span class="invisible">/10/deepseek-und-der-grosse-sprung-nach-vorn/</span></a><br><a href="https://social.undeadnetwork.de/tags/ai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ai</span></a> <a href="https://social.undeadnetwork.de/tags/china" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>china</span></a> <a href="https://social.undeadnetwork.de/tags/deepseek" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>deepseek</span></a> <a href="https://social.undeadnetwork.de/tags/FOSS" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>FOSS</span></a> <a href="https://social.undeadnetwork.de/tags/ki" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ki</span></a> <a href="https://social.undeadnetwork.de/tags/openai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>openai</span></a> <a href="https://social.undeadnetwork.de/tags/opensourceAI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>opensourceAI</span></a> <a href="https://social.undeadnetwork.de/tags/opensoure" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>opensoure</span></a> <a href="https://social.undeadnetwork.de/tags/oss" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>oss</span></a> <a href="https://social.undeadnetwork.de/tags/techbros" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>techbros</span></a></p>
uvok<p>Tried a roleplay with an LLM which more or less turned into a story. I actually quite like it. Could have more action, though </p><p>Not sure if chat mode was the right choice. <br>Also, still no idea how to do multi character chats.</p><p><a href="https://nc.uvokchee.de/s/7deWiWPAmB63K5E" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">nc.uvokchee.de/s/7deWiWPAmB63K</span><span class="invisible">5E</span></a></p><p><a href="https://furry.engineer/tags/koboldcpp" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>koboldcpp</span></a> <a href="https://furry.engineer/tags/koboldai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>koboldai</span></a> <a href="https://furry.engineer/tags/llm" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>llm</span></a></p>
uvok<p>Yeah, this is what a mean by "the model gets confused".<br>Apparently KoboldAI/cpp can't distinguish the stop words anymore.</p><p>The initial prompt was</p><pre><code>User: This is a roleplay between two animal characters, which I will define later.<br>User: Please generate and describe an animal character, which you will play. Choose a mammal. Please prefix your messages by a newline and your characters name, like so: '<br>Bob: waves "Hi, how are you?"'.<br></code></pre><p><a href="https://furry.engineer/tags/koboldai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>koboldai</span></a> <a href="https://furry.engineer/tags/koboldcpp" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>koboldcpp</span></a></p>
uvok<p>And apparently, you can't have two characters as "you"...</p><p><a href="https://furry.engineer/tags/koboldai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>koboldai</span></a> <a href="https://furry.engineer/tags/koboldcpp" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>koboldcpp</span></a></p>
uvok<p>Hm, I'm unsure whether I should instruct the model to reply with their characters name prefixed.<br>I have no idea how much "automagic" the chat mode has inbuilt.<br>I actually wanna get away from only seeing "KoboldAI" and "user".</p><p>Also, changing "my name" in the chat mode settings seems to confuse the model?</p><p><a href="https://furry.engineer/tags/koboldai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>koboldai</span></a> <a href="https://furry.engineer/tags/koboldcpp" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>koboldcpp</span></a></p>
uvok<p><a href="https://furry.engineer/tags/koboldcpp" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>koboldcpp</span></a> has a nice interface.<br>I kind wish they had a better explanation what's the difference between "chat" and "KoboldGPT chat".</p><p>Also, I'm simply <em>overwhelmed</em> by the amount of models. Each of them needs to be treated / prompted differently, it seems.</p><p>And then I still have problems with the model generating my character's actions / speech occasionally, and I don't know whether that's due to my insufficient prompting, or due to this "end token" (e.g. &lt;|im_end|&gt;) not being configured correctly, which seems to differ for some models again.</p><p><a href="https://furry.engineer/tags/llm" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>llm</span></a> <a href="https://furry.engineer/tags/ai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ai</span></a></p>
Ryan He (帳號已遷移到其他站台)<p>把 <a href="https://pastwind.top/tags/koboldcpp" rel="nofollow noopener noreferrer" target="_blank">#koboldcpp</a> 以前景模式執行,打開 <a href="https://aihorde.net/" rel="nofollow noopener noreferrer" target="_blank">AI Horde</a><span> 的任務分派程式,很明顯可以感覺到早上的使用者明顯比下午多。然後 RP 跟 ERP 的比例大約是 1:4。<br><br></span><a href="https://pastwind.top/tags/LLM" rel="nofollow noopener noreferrer" target="_blank">#LLM</a></p>
Habr<p>Как запустить Mixtral на своём компьютере</p><p>Всякий раз, когда выходит новая хорошая ИИ модель, Хабр наполняется вопросами "Как нам её попробовать" и неправильными ответами, будто нужно платить за какие-то сервисы или иметь железа на сто лямов. Поэтому я вновь напишу инструкцию, как запустить новейший mixtral-8x7 на обычных средних компьютерах.</p><p><a href="https://habr.com/ru/articles/781702/" rel="nofollow noopener noreferrer" target="_blank"><span class="invisible">https://</span><span class="">habr.com/ru/articles/781702/</span><span class="invisible"></span></a></p><p><a href="https://zhub.link/tags/LLM" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLM</span></a> <a href="https://zhub.link/tags/Mixtral" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Mixtral</span></a> <a href="https://zhub.link/tags/KoboldCPP" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>KoboldCPP</span></a> <a href="https://zhub.link/tags/GGUF" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>GGUF</span></a> #18+</p>