Yesterday, with very little effort, I managed to get a working #DeepSeek model running on my computer. It felt like I was a super-rich tech oligarch, but I just have a simple casual gaming computer with a cheap 8 GB graphics card for playing a game or editing videos. A lot of people have a machine like that; it’s below average. And it took me half an hour. Hauke explains here how anyone can do the same. https://tube.tchncs.de/w/cuNhJ5tRzdv4WRE2oZjCv3
You only need the DeepSeek-R1-Distill-Qwen model (32 billion parameters, approx. 19 GB)https://huggingface.co/mradermacher/DeepSeek-R1-Distill-Qwen-32B-Uncensored-GGUF, which is a very slimmed-down version that will also run on a home computer. For institutes, educational institutions and companies, there are also more powerful models, right up to the full model, which then really requires enormous resources. And you need a program like #KoboldCpp https://github.com/LostRuins/koboldcpp which loads the AI model into the graphics card and, for a weak card like mine, offloads the rest to the CPU. Furthermore, a chat GUI is provided via http://localhost:5001. As I said, if you’ve ever used a console, it’s really not rocket science.
There will certainly be loud voices that want to ban DeepSeek. Racists and capitalists will shout that the Chinese cannot be trusted. But it is open-source. Under an MIT license. I also ran it offline. It runs 100% locally.
What you could do with it! If smart people continue to train the models. Now everyone has an AI and the tech bros can pack up! Thanks Deepseek and thanks China. A huge gift for the open-source community and maybe for humanity.
PS. And once again, the conspiracy theorists were right, claiming that commercial AIs are only so large and resource-hungry so that they are reserved exclusively for the rich and the people remain customers. That’s over now.