zlacker

[parent] [thread] 10 comments
1. LTL_FT+(OP)[view] [source] 2026-01-26 07:03:37
It sounds like you don’t need immediate llm responses and can batch process your data nightly? Have you considered running a local llm? May not need to pay for api calls. Today’s local models are quite good. I started off with cpu and even that was fine for my pipelines.
replies(4): >>queenk+t9 >>kreetx+xj >>ydu1a2+NF >>ok_orc+2K1
2. queenk+t9[view] [source] 2026-01-26 08:44:17
>>LTL_FT+(OP)
Agreed, I'm pretty amazed at what I'm able to do locally just with an AMD 6700XT and 32GB of RAM. It's slow, but if you've got all night...
3. kreetx+xj[view] [source] 2026-01-26 10:20:51
>>LTL_FT+(OP)
Though haven't done any extensive testing then I personally could easily get by with current local models. The only reason I don't is that the hosted ones all have free tiers.
4. ydu1a2+NF[view] [source] 2026-01-26 13:15:24
>>LTL_FT+(OP)
Can you suggest any good llms for cpu?
replies(2): >>R_D_Ol+R71 >>LTL_FT+xf9
◧◩
5. R_D_Ol+R71[view] [source] [discussion] 2026-01-26 15:41:08
>>ydu1a2+NF
Following.
replies(2): >>LTL_FT+ng9 >>Aerbil+X6k
6. ok_orc+2K1[view] [source] 2026-01-26 18:16:02
>>LTL_FT+(OP)
I haven't thought about that, but really want to dig in more now. Any places you recommend starting?
replies(1): >>LTL_FT+bg9
◧◩
7. LTL_FT+xf9[view] [source] [discussion] 2026-01-28 17:11:15
>>ydu1a2+NF
I started off using gpt-oss-120b on cpu. It uses about 60-65gb of memory or so but my workstation has 128gb of ram. If I had less ram, I would start off with the gpt-oss-20b model and go from there. Look for MoE models as they are more efficient to run.
◧◩
8. LTL_FT+bg9[view] [source] [discussion] 2026-01-28 17:13:26
>>ok_orc+2K1
I started off using gpt-oss-120b on cpu. It uses about 60-65gb of memory or so but my workstation has 128gb of ram. If I had less ram, I would start off with the gpt-oss-20b model and go from there. Look for MoE models as they are more efficient to run.

My old threadripper pro was seeing about 15tps, which was quite acceptable for the background tasks I was running.

◧◩◪
9. LTL_FT+ng9[view] [source] [discussion] 2026-01-28 17:14:00
>>R_D_Ol+R71
I started off using gpt-oss-120b on cpu. It uses about 60-65gb of memory or so but my workstation has 128gb of ram. If I had less ram, I would start off with the gpt-oss-20b model and go from there. Look for MoE models as they are more efficient to run.
◧◩◪
10. Aerbil+X6k[view] [source] [discussion] 2026-01-31 19:36:16
>>R_D_Ol+R71
Hey Olivaw, saw a comment of yours asking about planners. Wanted to reply but it’s expired. Check out bullet journalling.
replies(1): >>R_D_Ol+1ps
◧◩◪◨
11. R_D_Ol+1ps[view] [source] [discussion] 2026-02-03 15:04:38
>>Aerbil+X6k
Thanks for the reply!

Bullet journaling is neat, but I'm far too whacky with my notes to stick to that kind of structure.

I have various other structures I implement, but they're just hodge podges of things.

[go to top]