• TheGrandNagus@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    1 month ago

    LLMs are an interesting tool to fuck around with, but I see things that are hilariously wrong often enough to know that they should not be used for anything serious. Shit, they probably shouldn’t be used for most things that are not serious either.

    It’s a shame that by applying the same “AI” naming to a whole host of different technologies, LLMs being limited in usability - yet hyped to the moon - is hurting other more impressive advancements.

    For example, speech synthesis is improving so much right now, which has been great for my sister who relies on screen reader software.

    Being able to recognise speech in loud environments, or removing background noice from recordings is improving loads too.

    My friend is involved in making a mod for a Fallout 4, and there was an outreach for people recording voice lines - she says that there are some recordings of dubious quality that would’ve been unusable before that can now be used without issue thanks to AI denoising algorithms. That is genuinely useful!

    As is things like pattern/image analysis which appears very promising in medical analysis.

    All of these get branded as “AI”. A layperson might not realise that they are completely different branches of technology, and then therefore reject useful applications of “AI” tech, because they’ve learned not to trust anything branded as AI, due to being let down by LLMs.

    • snooggums@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      LLMs are like a multitool, they can do lots of easy things mostly fine as long as it is not complicated and doesn’t need to be exactly right. But they are being promoted as a whole toolkit as if they are able to be used to do the same work as effectively as a hammer, power drill, table saw, vise, and wrench.

      • sugar_in_your_tea@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        Exactly! LLMs are useful when used properly, and terrible when not used properly, like any other tool. Here are some things they’re great at:

        • writer’s block - get something relevant on the page to get ideas flowing
        • narrowing down keywords for an unfamiliar topic
        • getting a quick intro to an unfamiliar topic
        • looking up facts you’re having trouble remembering (i.e. you’ll know it when you see it)

        Some things it’s terrible at:

        • deep research - verify everything an LLM generated of accuracy is at all important
        • creating important documents/code
        • anything else where correctness is paramount

        I use LLMs a handful of times a week, and pretty much only when I’m stuck and need a kick in a new (hopefully right) direction.

        • snooggums@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          1 month ago
          • narrowing down keywords for an unfamiliar topic
          • getting a quick intro to an unfamiliar topic
          • looking up facts you’re having trouble remembering (i.e. you’ll know it when you see it)

          I used to be able to use Google and other search engines to do these things before they went to shit in the pursuit of AI integration.

    • Punkie@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      I’d compare LLMs to a junior executive. Probably gets the basic stuff right, but check and verify for anything important or complicated. Break tasks down into easier steps.

      • zbyte64@awful.systems
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        1 month ago

        A junior developer actually learns from doing the job, an LLM only learns when they update the training corpus and develop an updated model.

          • zbyte64@awful.systems
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 month ago

            Why would you ever yell at an employee unless you’re bad at managing people? And you think you can manage an LLM better because it doesn’t complain when you’re obviously wrong?