• 3 Posts
  • 591 Comments
Joined 2 years ago
cake
Cake day: August 18th, 2023

help-circle





  • BatmanAoDtoProgrammingLLMS Are Not Fun
    link
    fedilink
    arrow-up
    12
    arrow-down
    1
    ·
    17 days ago

    I think you’re misunderstanding that paragraph. It’s specifically explaining how LLMs are not like humans, and one way is that you can’t “nurture growth” in them the way you can for a human. That’s not analogous to refining your nvim config and habits.




  • BatmanAoDtoProgrammer HumorRust bad Jai good
    link
    fedilink
    arrow-up
    5
    ·
    1 month ago

    At one point the user linked to a rust-lang forum thread from 2016-2019 as evidence that Jai has “some of the tools to make the code language agnostic” or something like that. The thread started with a discussion of array-of-struct vs struct-of-array data layouts, which of course has nothing to do with making code “language agnostic.” The user also mentioned the coding influencer lunduke multiple times. So I think they are simply misinformed on a lot of points, and I doubt they’re in the closed beta for Jai.

    (I read some of the comments simply because I had the same question you did. And, as it happens, the last post in the forum thread I mentioned was written by me, which was a funny surprise.)




  • BatmanAoDtoProgrammingLLM's hallucinating or taking our jobs?
    link
    fedilink
    arrow-up
    3
    arrow-down
    2
    ·
    1 month ago

    making the same mistakes

    This is key, and I feel like a lot of people arguing about “hallucinations” don’t recognize it. Human memory is extremely fallible; we “hallucinate” wrong information all the time. If you’ve ever forgotten the name of a method, or whether that method even exists in the API you’re using, and started typing it out to see if your autocompleter recognizes it, you’ve just “hallucinated” in the same way an LLM would. The solution isn’t to require programmers to have perfect memory, but to have easily-searchable reference information (e.g. the ability to actually read or search through a class’s method signatures) and tight feedback loops (e.g. the autocompleter and other LSP/IDE features).





  • BatmanAoDtoProgrammer HumorDo you like (AI) clocks?
    link
    fedilink
    arrow-up
    10
    arrow-down
    1
    ·
    edit-2
    2 months ago

    Thanks for sharing this! I really think that when people see LLM failures and say that such failures demonstrate how fundamentally different LLMs are from human cognition, they tend to overlook how humans actually do exhibit remarkably similar failures modes. Obviously dementia isn’t really analogous to generating text while lacking the ability to “see” a rendering based on that text. But it’s still pretty interesting that whatever feedback loops did get corrupted in these patients led to such a variety of failure modes.

    As an example of what I’m talking about, I appreciated and generally agreed with this recent Octomind post, but I disagree with the list of problems that “wouldn’t trip up a human dev”; these are all things I’ve seen real humans do, or could imagine a human doing.