• thevoidzero@lemmy.world
    link
    fedilink
    arrow-up
    2
    ·
    9 hours ago

    Yeah but the people who made it like that probably understand whether to trust it to write code or not. The AI Tony wrote, he knows what it does best and he trusts it to write his code. Just because it’s AI doesn’t mean it’s LLM. Like I trust the errors compilers give me even if I didn’t write them because it’s good. And I trust my scripts to do things that I wrote them for, specifically since I tested them. Same with the AI you yourself made, you’d test it, and you’d know the design principles.

    • vrighter@discuss.tchncs.de
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      7 hours ago

      an ai is not a script. You can know what a script does. neural networks don’t work that way. You train them, and hope you picked the right dataset for it to hopefully learn what you want it to learn. You can’t test it. You can know that it works sometimes but you also know that it will also not work sometimes and there’sjacksjit you can do about it. A couple of gigabytes of floating point numbers is not decipherable to anyone.

      • thevoidzero@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        8 minutes ago

        And who says AI means neural network? That’s what we use, doesn’t mean that’s the only AI possible to write. There are a lot of different models, neural network is popular right now because it can learn from data without anyone having to teach it actual logic. An AI written by fictional character can be a deterministic kind with very similar logic to humans that you can inspect and write and give weights to things.