Cryptography nerd

Fediverse accounts;
@[email protected] (main)
@[email protected]
@[email protected]

Lemmy moderation account: @[email protected] - [email protected]

@[email protected]

Bluesky: natanael.bsky.social

  • 64 Posts
  • 165 Comments
Joined 1 year ago
cake
Cake day: January 18th, 2025

help-circle
  • This gets at my own personal perspective of using LLMs to respond - it’s not just about not putting effort into understanding and responding yourself, rather it is about making yourself a proxy to a tool I could use myself, and doing so *without even having a better understanding of how to use the tool to answer my question*, and still thinking you’re somehow made a positive contribution, that is the most disrespectful.

    If you genuinely thought the LLM could help me then you should be explaining your process to me for how to use it and validate responses, or else at least you should ask me for more info and explain how you think it’s responses could help if you really do think you’re better at operating it.

    Imagine doing the same in a workshop, and taking a powertool to an object before you even bothered figuring out what the other person wanted. Or trying to be helpful by asking questions on your behalf to other departments, but messing up the context and thus repeatedly producing useless answers that you have to put time into refuting.



  • It’s actually kinda easy. Neural networks are just weirder than usual logic gate circuits. You can program them just the same and insert explicit controlled logic and deterministic behavior. To somebody who don’t know the details of LLM training, they wouldn’t be able to tell much of a difference. It will be packaged as a bundle of node weights and work with the same interfaces and all.

    The reason that doesn’t work well if you try to insert strict logic into a traditional LLM despite the node properties being well known is because of how intricately interwoven and mutually dependent all the different parts of the network is (that’s why it’s a LARGE language model). You can’t just arbitrarily edit anything or insert more nodes or replace logic, you don’t know what you might break. It’s easier to place inserted logic outside of the LLM network and train the model to interact with it (“tool use”).



















  • Even as a Linux desktop it would mostly just be interesting for devs and people doing relatively lightweight 3D design work (especially because it will take a while before other distros support it), I don’t see it competing against regular desktops.

    Any company who depend on their employees having a decent GPU will likely want to be able to upgrade/reconfigure new orders at will, and will prefer a tower, and they will prefer the quick repairability of a tower. Those who don’t are increasingly ok with using mini PCs.