I’ve never really followed X (nor Twitter), Bluesky, Instagram, TikTok, etc. so I basically live under a rock. Sometimes I ask dumb questions to try to understand people a little better. Apologies if my questions inadvertently offend anyone. I mean no harm.

  • 0 Posts
  • 89 Comments
Joined 1 year ago
cake
Cake day: May 3rd, 2025

help-circle




  • Gemma4:26b is also worth trying. I find it runs much faster on my hardware.

    Edit: Qwen3.6:35B might be the sweet spot. It’s bigger than the 27B, but actually more lightweight when running. TIL the 27B is not a MoE model; it’s a dense model. The 35B is a MoE model with only 3B active params.

    So far, I think Qwen3.6:35B might be giving me better results than Gemma4:26B. It’s a bit slower than Gemma4:26B, but definitely faster than Qwen3.6:27B.



  • Do you not use it enough because yet get bad results? I discovered that, no matter how smart the LLM might be, its first attempt is never its best work. Tell it to review its work (or its plan, if using planning mode). If it makes any changes, tell it to review its work again. Repeat until there are no more changes.

    (You don’t actually have to do this repetition manually; just tell the AI to do it in a loop. I recommend making it into a SKILL.md so you don’t have to explain the loop every time.)

    With these loops, I get better results AND burn lots of tokens. (Yes, it feels strange that excessive token consumption is actually considered a good thing)