qwen
md5: a68c7396933d3069db2807412f50bc3c
🔍
as someone who is retarded and has a shitter laptop that can't run any llm loccally, is this my best option for something that's free? or should i just kms
>>106117339 (OP)I'm sure you can run the small ones...
Why is the logo of every AI company a 6-pointed star?
>>106117381hell no i got a pentium siver in this lenovo shit ideapad i can barely install ollama
>>106117422also 4gb memory so tinyllama and phi 3 chug like goddamn thomas
>>106117458blocked at my work unfortunately but i get around that by using shit r1 distills on WebLLM
>>106117406It's either goatse or that jewish swastika star. OpenAI's logo is clearly an asshole.
>>106117339 (OP)Kimi k2 is also decent and the web chat is free.
>>106117339 (OP)Why does it have to be free? Just make an account at any of the 999999 LLM API companies and top it up with $20 (enough to fuck around and get an idea of what your usage is likely to be).
>>106117555But Kimi needs registration, Qwen doesn't.
Diddling don rapes little kids.
>>106117558why? im stingy as fuck and dont want to grind dumbass api sorry buddy
>>106117728Bro just throw twenty bucks their way, you can't be that cheap??
surely you can run qwen3 0.6b
>>106117781Also qwen 3 4b at q4 in 2.5 gigs
>>106117339 (OP)depending on how much ram you have you can run it locally. I have 32gb on a mini pc and I run quantized qwen3:30b with ollama. with 8gb you should be able to run qwen3:7b.
create an openrouter account, minimally fund it, only use the free models.
>>106118389this or litellm seems like the only solution here - and are they still trying to get wllama out of demo? that might be another avenue as well
>>106117339 (OP)I have a 3090 and run qwen-coder-14b, it eats 18gb of vram and still sucks ass
I like Qwen better than DeepSeek in my limited experience
Is DeepSeek actually better tho?
>>106117339 (OP) can't you just get a laptop/pc that have have avx512? ollama supports that well now and it is not too slow to run some models with that, that is probably the cheapest way to run something local.
>>106117339 (OP)Nothing you can run on a laptop is worth running.
>want to run local models
>have an amd gpu
>>106121647>>have an amd gpuso has lotsa video ram, and can run better models. lmao even. my gemma3 27b go brrrr
Qwen 3 A3B 30B Instruct and Thinking on llama.cpp at IQ4 NL quantization is the best I've found yet; with no GPU, 6 core i7, 16 GB of RAM. It runs fine, but lacks some knowledge, especially pop culture. Good at reasoning, good at non role-play chat, good at returning JSON for ad-hoc function calling from shell scripts. Local is getting better. Also, wait to see what OpenAI releases for local soon. If it's meant for mobile or edge devices, it will probably work on no-GPU laptops too.
Some don't know this, but A3B means 3 billion active parameters. It runs that fast, but has intelligence closer to a 30 billion parameter model. MoE is working good for those of us without GPU, at least on Linux with llama.cpp it is.