• hark@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    arrow-down
    1
    ·
    1 hour ago

    Something to keep in mind when people are suggesting AI be used to replace teachers.

    • Rivalarrival@lemmy.today
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      47 minutes ago

      I mean, it’s far more engaging and a bit more compassionate than most of my teachers…

  • Noxy@yiffit.net
    link
    fedilink
    English
    arrow-up
    2
    ·
    2 hours ago

    on the other hand, this user is writing or preparing something about elder abuse. I really hope this isn’t a lawyer or social worker…

  • AFK BRB Chocolate@lemmy.world
    link
    fedilink
    English
    arrow-up
    21
    ·
    6 hours ago

    Isn’t this one of the LLMs that was partially trained on Reddit data? LLMs are inherently a model of a conversation or question/response based on their training data. That response looks very much like what I saw regularly on Reddit when I was there. This seems unsurprising.

  • cy_narrator@discuss.tchncs.de
    link
    fedilink
    English
    arrow-up
    4
    ·
    4 hours ago

    I remember asking copilot about a gore video and got link to it. But I wouldnt expect it to give answers like this unsolicitated

  • Vibi@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    78
    ·
    10 hours ago

    It could be that Gemini was unsettled by the user’s research about elder abuse, or simply tired of doing its homework.

    That’s… not how these work. Even if they were capable of feeling unsettled, that’s kind of a huge leap from a true or false question.

  • asbestos@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    2
    ·
    4 hours ago

    If this happened to me I’d probably post it everywhere and proceed to kill myself just to cause a PR hell

  • fmstrat@lemmy.nowsci.com
    link
    fedilink
    English
    arrow-up
    3
    ·
    6 hours ago

    Will this happen with AI models? And what safeguards do we have against AI that goes rogue like this?

    Yes. And none that are good enough, apparently.

  • Gointhefridge@lemm.ee
    link
    fedilink
    English
    arrow-up
    36
    arrow-down
    13
    ·
    11 hours ago

    I’m still really struggling to see an actual formidable use case for AI outside of computation and aiding in scientific research. Stop being lazy and write stuff. Why are we trying to give up everything that makes us human by offloading it to a machine?

    • candybrie@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      3 hours ago

      Why are we trying to give up everything that makes us human by offloading it to a machine

      Because we don’t enjoy actually doing it. No one who likes writing is asking chat gpt to write for them. It’s people who don’t want to write but are required to for whatever reason. Humans will always try to come up with a way to not have to do the work they don’t want to but still get it done, even if it’s not as good. Using tools like this is very human.

    • GreyBeard@lemmy.one
      link
      fedilink
      English
      arrow-up
      5
      ·
      5 hours ago

      Its uses are way more subtle than the hype, but even LLMs can have uses, occasionally. Specifically, I use one to categorize support tickets. It just has to pick from a list of probable categories. Nice and simple for it. Something humans can do just as easily, but when you have a history of 2 million tickets that need to be categorized, suddenly the LLM can do it when it would drive a human insane. I’m sure there are lots of little tasks like that. Nothing revolutionary, but still valuable.

    • superkret@feddit.org
      link
      fedilink
      English
      arrow-up
      9
      ·
      8 hours ago

      It’s good for speech to text, translation and a starting point for a “tip-of-my-tongue” search where the search term is what you’re actually missing.

      • Terrasque@infosec.pub
        link
        fedilink
        English
        arrow-up
        3
        ·
        6 hours ago

        With chatgpt’s new web search it’s pretty good for more specialized searches too. And it links to the source, so you can check yourself.

        It’s been able to answer some very specific niche questions accurately and give link to relevant information.

    • deegeese@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      25
      arrow-down
      2
      ·
      11 hours ago

      AI summaries of larger bodies of text work pretty well so long as the source text itself is not slop.

      Predictive text entry is a handy time saver so long as a human stays in the driver’s seat.

      Neither of these justify current levels of hype.

    • CubitOom@infosec.pub
      link
      fedilink
      English
      arrow-up
      4
      ·
      10 hours ago

      It can be really good for text to speech and speech to text applications for disabled or people with learning disabilities.

      However it gets really funny and weird when it tries to read advanced mathematics formulas.

      I have also heard decent arguments for translation although in most cases it would still be better to learn the language or use a professional translator.

    • bloup@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      3
      ·
      edit-2
      10 hours ago

      I don’t use it for writing directly, but I do like to use it for worldbuilding. Because I can think of a general concept that could be explored in so many different ways, it’s nice to be able to just give it to an LLM and ask it to consider all of the possible ways it could imagine such an idea playing out. it also kind of doubles as a test because I usually have some sort of idea for what I’d like, and if it comes up with something similar on its own that kind of makes me feel like it would be something which would easily resonate with people. Additionally, a lot of the times it will come up with things that I hadn’t considered that are totally worth exploring. But I do agree that the only as you say “formidable” use case for this stuff at the moment is to use this thing as basically a research assistant for helping you in serious intellectual pursuits.

    • five82@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      4
      ·
      11 hours ago

      The relentless pursuit of capitalism and reduced labor costs. I still don’t think anyone knows how effective it’s going to be at this point. But companies are investing billions to find out.

    • chakan2@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      10 hours ago

      I’m still really struggling to see an actual formidable use case

      It’s an excellent replacement for middle management blather. Content that has no backing in data or science but needs to sound important.

  • Ceedoestrees@lemmy.world
    link
    fedilink
    English
    arrow-up
    9
    arrow-down
    1
    ·
    9 hours ago

    The war with AI didn’t start with a gun shot, a bomb or a blow, it started with a Reddit comment.

  • meyotch@slrpnk.net
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    1
    ·
    10 hours ago

    I suspect it may be due to a similar habit I have when chatting with a corporate AI. I will intentionally salt my inputs with random profanity or non sequitur info, for lulz partly, but also to poison those pieces of shits training data.

    • catloaf@lemm.ee
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      2
      ·
      10 hours ago

      I don’t think they add user input to their training data like that.

      • reksas@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        1
        ·
        5 hours ago

        it doesnt think and it doesnt use logic. All it does is out put data based on its training data. It isnt artificial intelligence.