Black Mirror creator unafraid of AI because it’s “boring”::Charlie Brooker doesn’t think AI is taking his job any time soon because it only produces trash

  • SkyeStarfall@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    120
    arrow-down
    11
    ·
    1 year ago

    The thing with AI, is that it mostly only produces trash now.

    But look back to 5 years ago, what were people saying about AI? Hell, many thought that the kind of art that AI can make today would be impossible for it to create! …And then it suddenly did. We’ll, it wasn’t actually suddenly, and the people in the space probably saw it coming, but still.

    The point is, we keep getting better at creating AIs that do stuff we thought were impossible a few years ago, stuff that we said would show true intelligence if an AI can do them. And yet, every time some new impressive AI gets developed, people say it sucks, is boring, is far from good enough, etc. While it slowly, every time, creeps on closer to us, replacing a few jobs here and there in the fringes. Sure, it’s not true intelligence, and it still doesn’t beat humans, but, it beats most, at demand, and what happens when inevitably better AIs get created?

    Maybe we’re in for another decades long AI winter… or maybe we’re not, and plenty more AI revolutions are just around the corner. I think AIs current capabilities are frighteningly good, and not something I expected to happen this soon. And the last decade or so has seen massive progress in this area, who’s to say where the current path stops?

    • Telodzrum@lemmy.world
      link
      fedilink
      English
      arrow-up
      81
      arrow-down
      31
      ·
      1 year ago

      Nah, nah to all of it. LLM is a parlor trick and not a very good one. If we are ever able to make a general artificial intelligence, that’s an entirely different story. But text prediction on steroids doesn’t move the needle.

      • fsmacolyte@lemmy.world
        link
        fedilink
        English
        arrow-up
        29
        arrow-down
        12
        ·
        edit-2
        1 year ago

        The best ones can literally write pretty good code, and explain any concept on the Internet to you that you ask them to. If you don’t understand a specific thing about their explanation, they can add onto their explanation, and they can respond in the style you want (explain as if I’m ten, explain as if I’m an undergrad, etc).

        I use it literally every day for work in a somewhat niche field. I don’t really agree that it’s a “parlor trick”.

        • state_electrician@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          29
          arrow-down
          5
          ·
          1 year ago

          LLMs are awful for facts, because they don’t understand what facts are. You should never rely on them if you require factual correctness.

          They are OK for text summation, formatting and just making shit up. For summation a human with experience still produces nicer output, because they understand the content and don’t just look at words. As for making shit up you will get the statistically most likely output, so it’s usually trite and boring. I think the progress is amazing, but there are still so many problems to be solved.

          Right now I use them for boiler plate stuff, like writing a text with some parameters and then I polish it. For code I find them quite useless, because with an IDE I can write boiler plate just as fast as when I polish the prompts until the LLM delivers useful stuff. And with the IDE I don’t get references to methods or entire libraries that just don’t exist.

          • banneryear1868@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 year ago

            Right now I use them for boiler plate stuff, like writing a text with some parameters and then I polish it

            It’s actually great for dnd to produce NPC dialogue or names on the fly. We also tried using it to calculate area of effect spells, ie “how many average sized humans in armor with swords could fit in a circle with a diameter of 30ft.” We were rolling with it before someone pointed out that it didn’t calculate the area of a circle correctly, however it got the rest more or less accurate. So we don’t use it for that anymore, and it’s funny how what often appears to be the simplest component of a question is the thing it most often gets wrong.

          • darth_helmet@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            1 year ago

            People are also kind of shit at facts. There are so many facts, and many of them aren’t practical for every person who needs to assess a fact’s accuracy to do so. But it isn’t structurally impossible to mimic how humans learn how to gauge truthfulness, we just have to be prepared for the idea that it will be bound by the limitations of language, as well as the risk inherent in trusting data that it has not independently verified.

        • Blóðbók@slrpnk.net
          link
          fedilink
          English
          arrow-up
          12
          arrow-down
          3
          ·
          1 year ago

          I use LLMs for having things explained to me, too… but if you want to know how much salt to pour in that soup, try asking it about something niche and complicated you already know the answer to.

          They can be useful in figuring out the correct terminology so that you can find the answer on your own, or for pointing some very very obvious mistakes in your understandings (but it will still miss most of them).

          Please don’t use those things as answer machines.

          • fsmacolyte@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            I’m going to use those things as answer machines and you can’t stop me.

            Jokes aside, I always validate what chatbots tell me, not even just important things. I use GPT-4 for work and 90% of the time it can show me how to use very specific functions in complex ways, but yesterday (for the first time in awhile) it made up a function that didn’t exist. To its credit, I said, “Are you sure about [function]?” and it said, “I’m sorry, I got confused. That function doesn’t exist. However, look into X, Y, Z for further resources” and I did and they were the correct things to look into.

            • Blóðbók@slrpnk.net
              link
              fedilink
              English
              arrow-up
              2
              ·
              edit-2
              1 year ago

              If you press it the same way again (“are you sure the function doesn’t exist?”), there is a high chance it will “rectify” its rectification.

        • Telodzrum@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          6
          ·
          1 year ago

          No they can’t. Your phrasing is misleading. It’s a Chinese Room test output and nothing more. I had an Encarta CD that could do rudimentary version of this in 1995. That was more impressive, tbh.

          • aubertlone@lemmy.world
            link
            fedilink
            English
            arrow-up
            5
            arrow-down
            2
            ·
            1 year ago

            If you’re really comparing LLM’s to your Encarta cd from 1995 and saying the Encarta CD was the superior experience…

            I’m afraid there’s not much left for us to discuss… Our views are too far apart.

      • GnuLinuxDude@lemmy.ml
        link
        fedilink
        English
        arrow-up
        16
        ·
        1 year ago

        Sam Altman (Creator of the freakish retina scanning based Worldcoin) would agree, it seems. The current path for LLMs and GPT seems to be in something of a bind, because to seriously improve upon what it currently does it needs to do something different, not more of the same. And figuring out something different could be very hard. https://www.wired.com/story/openai-ceo-sam-altman-the-age-of-giant-ai-models-is-already-over/

        At least that’s what I understand of it.

        • TheWiseAlaundo@lemmy.whynotdrs.org
          link
          fedilink
          English
          arrow-up
          13
          arrow-down
          1
          ·
          edit-2
          1 year ago

          He’s not saying “AI is done, there’s nothing else to do, we’ve hit the limit”, he’s saying “bigger models don’t necessarily yield better results like we had initially anticipated”

          Sam recently went before congress and advocated for limiting model sizes as a means of regulation, because, at the time, he believed bigger would generally always mean better outputs. What we’re seeing now is that if a model is too large it will have trouble producing truthful output, which is super important to us humans.

          And honestly, I don’t think anyone should be shocked by this. Our own human brains have different sections that control different aspects of our lives. Why would an AI brain be different?

          • gregoryw3@lemmy.ml
            link
            fedilink
            English
            arrow-up
            3
            ·
            1 year ago

            Future of AI is definitely going towards Manager/Agent model. It allows for an AI to handle all the tasks without keeping it to one model or method. We’re already seeing this with ChatGPT using Mathematica for math questions. Soon we can see art AI using different models and methods based on text input.

          • Browning@lemmings.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            I gather that this is partly because data sizes haven’t been going up with model sizes. That is likely to change soon as synthetic data starts to overtake organic data in both quantity and quality.

      • ChaoticNeutralCzech@feddit.de
        link
        fedilink
        English
        arrow-up
        17
        arrow-down
        6
        ·
        1 year ago

        In humans, abstract thinking developed hand in hand with language. So despite their limitations, I think that at least early AGI will include an LLM in some way.

        • IonAddis@lemmy.world
          link
          fedilink
          English
          arrow-up
          14
          arrow-down
          2
          ·
          1 year ago

          I’ve been having a lot of vague thoughts about the unconscious bits of our brains and body, in regards to LLMs. The parts of our brains/neurons that started evolving back in simple animals as basically super primitive ways to process visual/audio/whatever input.

          Our brains do a LOT of signal processing and filtering that never reaches conscious thought, that we can’t even reach with our conscious thought if we tried, but which is necessary for our squishy body-things to take in input from our environment and turn it into something useful instead of drowning in a screeching eye-searing tangled mess of chaotic sensory input all the time.

          LLMs strike me as that sort of low-level input processing, the pattern-recognition and filtering. I think true generalized AI would have to be built on pieces like this–probably a lot of them. Ways to pluck patterns out of complex but repeated input. Like, this stuff definitely isn’t self-aware, but could eventually end up as some sort of processing library for something else far down the line.

          Now might be a good time to pick up Peter Watts’ sci-fi book Blindsight. He doesn’t exactly write about AI in it, but he does write about a creature that responds to input but isn’t exactly conscious like you or I.

          • ChaoticNeutralCzech@feddit.de
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 year ago

            some sort of processing library for something else far down the line

            This is what I meant.

            pick up Peter Watts’ sci-fi book Blindsight

            I just got the EPUB, thanks. Looking forward to reading it.

      • Clent@lemmy.world
        link
        fedilink
        English
        arrow-up
        15
        arrow-down
        11
        ·
        1 year ago

        Parlor trick is a perfect description.

        People don’t get that these things aren’t anymore intelligent than their smartphones predicting the next word. The main difference is instead of a couple words it has thousands to choose from.

        Half of the trick is how it uses the prompt to decided what words to start with.

        • Zeth0s@lemmy.world
          link
          fedilink
          English
          arrow-up
          13
          arrow-down
          1
          ·
          edit-2
          1 year ago

          That is not how it works. Your smartphone has all the dictionary available, same as LLM. It is simply something very different. People super confidently discussing about AI on lemmy are the real hallucinating parrots

          • Clent@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            12
            ·
            1 year ago

            There is an inverse relationship between the intelligence of a person and their amazement at what these large language models can produce.

              • Clent@lemmy.world
                link
                fedilink
                English
                arrow-up
                5
                arrow-down
                2
                ·
                1 year ago

                The ability to generate plausible language was a lack of compute power. The actual programs running the LLM is not complicate.

                The model that is produced is complex.

                Its training required compute power that was not previously available but the math/code behind these systems is not complex. They are resource intensive. There is a difference that a layperson often cannot comprehend.

                • Honytawk@lemmy.zip
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  1 year ago

                  So what is it now?

                  Are LLMs more intelligent than your smartphone, or do they need a lot more computer power to produce the same thing as your smartphone?

            • Zeth0s@lemmy.world
              link
              fedilink
              English
              arrow-up
              4
              arrow-down
              2
              ·
              1 year ago

              I heard the same for people who downvote on lemmy when notified about being an exemplification of the dunning Kruger effect

          • Clent@lemmy.world
            link
            fedilink
            English
            arrow-up
            8
            arrow-down
            1
            ·
            1 year ago

            Yes. I’ve used them. I have used it beyond the point of it hallucinating.

            I am also a software engineer and have deeper understanding of how these systems work than your average user.

            The software community tends to approach these things with more caution than the general population. The media overblows the capabilities of these systems.

            A more concrete example is autonomous vehicles which were promised for decades and even now with a form of them on the road, they are still closer to remote controlled vehicles than the intelligent self contained systems we have been promised.

            The difference between predictive text on a smart phone and predictive text of an LLM is my smart phone is predicting what I am likely to type next based on things i have typed in the past, while the LLM is predicting what comes next based on a larger body of work from source pulled from all across the internet. The LLM is then tuned by humans. This tuning step is under reported.

            The LLM is unable to determine the truth of its own output. I would argue that is a key to claiming intelligence but determining what intelligence means is itself a philosophical question up for debate.

            • banneryear1868@lemmy.world
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              1
              ·
              1 year ago

              The LLM is unable to determine the truth of its own output. I would argue that is a key to claiming intelligence but determining what intelligence means is itself a philosophical question up for debate.

              Yeah exactly and a great way to see this is by asking it to produce two viewpoints about the same subject, a negative and positive review of something you’re familiar with is perfect. It produces this hilarious “critic” type jargon but you can tell it doesn’t actually understand. Coincidentally, it’s drawing from a lot of text where the original human author(s) might not understand either and are merely themselves re-producing a jargon-heavy text for an assignment by their employer or academic institution. If AI can so accurately replicate some academic paper that probably didn’t need to be written for anything other than to meet publishing standards for tenured professors, then that’s really a reflection on the source material. Since LLM can only create something based on existing input, almost all the criticisms of it, are criticisms that can apply to it’s source material.

          • banneryear1868@lemmy.world
            link
            fedilink
            English
            arrow-up
            6
            arrow-down
            1
            ·
            1 year ago

            It’s not really “intelligent” though, as in it’s not thinking about what it’s doing. What AI will do very well is reproduce jargon, and if it’s jargon that we associate with intelligence then it appears intelligent. Academic papers for instance it can do a very convincing job because that format is so repetitive and jargon heavy.

            You can do an experiment by asking it to produce a positive review of something niche and academic you’re familiar with, then ask it to produce a negative review of the same subject. It will produce convincing dialogue for either scenario, but it does not know which is more true/accurate, and it will come across as a student writing about something they didn’t do the reading for.

            The “question if humans are [intelligent]” is the more relevant thing here. We’re constantly expected to communicate with thoughtlessly reproduced jargon, and many of us can do this very well in a way that gives the impression of intelligent thought. The fact AI can do this, and that people are concerned about how intelligent it appears, is more a reflection on how derivative our notions of intelligence can be in these settings.

          • rambaroo@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            3
            ·
            1 year ago

            The fact that you believe an LLM is “intelligent” tells me you have no clue how they work and your comments on the matter can be ignored.

      • aubertlone@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        2
        ·
        1 year ago

        Nah, nah to your understanding of LLM’s

        No it’s not true intelligence. Yes, it makes humans much faster at their work

        It has really sped up my work, especially when coding in unfamiliar languages.

        It’s silly to compare it to a parlor trick or text prediction.

      • Honytawk@lemmy.zip
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        1 year ago

        LLM’s are like an interface to allow computers to talk to humans.

        They are a necessary step in order to create general AI, because a general AI that can’t generate text wouldn’t be able to convey what they learned.

    • TwilightVulpine@lemmy.world
      link
      fedilink
      English
      arrow-up
      18
      arrow-down
      2
      ·
      1 year ago

      By its nature, Large Language Models won’t ever be truly innovative, after all they rely on expected patterns. But a lot of the media that we consume is also made to appeal to patterns that we expect: genres, tropes, usual messages. AI could replace a lot of it and frankly, that’s scary to think in a world where we need to work to earn our living.

      Truly groundbreaking art may not be what people usually seek, it’s often something they don’t even know they want until they experience it, or they might even fail to appreciate it. But it likely won’t be automated unless AI achieves full consciousness, but if it does we will have a much more complicated situation in our hands than “we can command AI to make art better than we can do ourselves”.

      Still, getting paranoid over the uncertain latter won’t help us with the former that is just around the corner.

      • KevonLooney@lemm.ee
        link
        fedilink
        English
        arrow-up
        7
        ·
        1 year ago

        Good points.

        One problem with replacing everything with AI that people don’t think about: middle managers will start to be replaced too. There’s no way to ask a LLM “why did you do that”? Fewer people will need to be managed.

        • TwilightVulpine@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          ·
          1 year ago

          It seems unwise to replace managers with LLMs because LLMs don’t understand the real world implications of their responses, they don’t have awareness of the real world, they simply give you often used language patterns, which can be innacurate or biased based on flawed human data. But it would be a great way for sketchy human executives to offload responsibility for unethical actions and feign objectivity or uninvolvement, so I don’t doubt they will try.

          Even if we imagine a perfect AI that does takes into account every objective fact and philosophical argument, that still leaves the question of how will the people who get replaced in all these intellectual, artistic and service jobs will make a living. That’s not an answer that technology will give us, that will a nasty political situation.

          • KevonLooney@lemm.ee
            link
            fedilink
            English
            arrow-up
            5
            ·
            1 year ago

            No, you misunderstood. The managers are fired because there’s fewer people to manage.

            • TwilightVulpine@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              ·
              1 year ago

              That makes sense too. Overall, a lot of people’s jobs are threatened, but I don’t think “learn AI” is going to cut it this time. Not for all these people.

          • Honytawk@lemmy.zip
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            LLMs don’t understand the real world implications of their responses

            LLMs don’t, but specialised AI trained for that specific purpose would.

      • aesthelete@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        edit-2
        1 year ago

        Truly groundbreaking art may not be what people usually seek, it’s often something they don’t even know they want until they experience it, or they might even fail to appreciate it.

        Everyone in these threads likes to talk about being impressed by these llm or not being impressed by them as being some sort of intelligence test. I think of it more as a test of a person’s sense of creativity.

        It spits out a lot of passable text very easily, but as you’re saying here its creativity is essentially nil. Even its “hallucinations” are just versions of things it borrowed from elsewhere injected slightly to wildly out of context in order to satisfy a prompt.

        I tried to play a generative AI RPG builder game online and it came up with scenarios so boring I can’t imagine playing it for longer than ten minutes.

        I also find the same with generated content in other video games. At its best it’s passable and that’s about it. No man’s sky has infinite worlds full of weird ligar creatures and after you’ve visited a couple dozen worlds they’re pretty much all the same.

        • Honytawk@lemmy.zip
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          1
          ·
          1 year ago

          And who is to say that we humans don’t process creativity exactly the same way? By borrowing from things we encounter.

          Even the earliest creative expats of humans was just things we saw in nature, which we drew on cave walls.

          We humans just have more experience since we existed longer, so the line feels a lot more blurred.

          I also encountered games made by humans that were so boring I couldn’t manage more than 10 minutes.

          • aesthelete@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            1 year ago

            And who is to say that we humans don’t process creativity exactly the same way? By borrowing from things we encounter.

            That’s part of it, but it’s definitely not all of it.

            There’s more creativity in the average prompt than there is in any response I’ve ever seen from ChatGPT.

            If creativity were as simple as mashing a few things together as you’re saying, ChatGPT would be there already because that’s obviously what it’s doing.

            I also encountered games made by humans that were so boring I couldn’t manage more than 10 minutes.

            Me too, but that’s an indictment of a single creator or team’s idea that was boring, not an indictment of a system. This thing was basically a framework with the llm being the central “creator” at the center. It would find the most boring aspects of the prompts and lean into them. This is of course a subjective assessment, but I’d argue that it’s not an uninformed one.

        • MycoPete@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          I also find the same with generated content in other video games. At its best it’s passable and that’s about it.

          Minecraft would like to have a word with you…

          • aesthelete@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            Minecraft isn’t generating new animals or narrative. Landscape generation is relatively straightforward from an algorithm / computation perspective. If it started generating its own models or characters or character dialogue I suspect it would very quickly fall into the territory of what I’m talking about.

            There’s just a feeling of emptiness to me that’s pervasive in games with main parts of narrative or gameplay that are randomly generated.

    • ezchili@iusearchlinux.fyi
      link
      fedilink
      English
      arrow-up
      21
      arrow-down
      7
      ·
      edit-2
      1 year ago

      I think the breakthroughs in AI have largely happened now as we’re reaching a slowndown and an adoption phase

      The research has been stagnating. Video with temporal consistency doesn’t want to come, voice is still perceptibly non-human, openai is assembling 5 models in a trenchcoat to make gpt do images and it passing as progress, …

      Companies and people are adopting what is already there for new applications, it’s getting more common to see neural network models in lots of solutions where the tech adds good value and is applicable, but the models aren’t breaking new grounds like in 2021 anymore

      The only new fundamental developments i can recall in the core technology is the push for smaller models trainable on way less data and that can be specialized for certain applications. Far away from the shock we all got when AI suddenly learned to draw a picture from a prompt

        • havocpants@lemm.ee
          link
          fedilink
          English
          arrow-up
          7
          arrow-down
          3
          ·
          1 year ago

          We should be happy that we might still have a few more years left before AI renders us all obsolete.

          Wow, this is some spectacular hyperbole!

            • aesthelete@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              ·
              1 year ago

              It’s ridiculously easy to recreate almost anything on there at a similar or sometimes even better level of quality

              And ridiculously difficult to copyright any of it because it was generated.

              • Honytawk@lemmy.zip
                link
                fedilink
                English
                arrow-up
                2
                ·
                1 year ago

                Yes, AI doesn’t work with copyright.

                And since AI is here to stay, we better replace our failed copyright system with something proper. Disney be damned.

                • aesthelete@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  1 year ago

                  we better replace our failed copyright system with something proper. Disney be damned.

                  I’d like that? But if you’re expecting the “we” in here to be the current people in their current power structures I suspect you’ll be waiting an awfully long time for that result.

                • aesthelete@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  arrow-down
                  1
                  ·
                  edit-2
                  1 year ago

                  The interesting question left is: Will static art survive at all? Will the future still have static movies or will everybody just generate their personalized dynamic entertainment on demand?

                  Lol this reminds me of when Kramer from Seinfeld asks if we’ll still be using napkins in the year 2000 or if this “mouth vacuum” thing is for real.

                  There’s already been court cases suggesting that AI art isn’t copyrightable.

                  The AI art I’ve seen so far is about as compelling as random crap from deviant art. The only difference being at least the starving artists on there know how many fingers are on a hand.

      • SkyeStarfall@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        10
        arrow-down
        1
        ·
        1 year ago

        I want to note that everything you talk about is happening on the scales of months to single years. That’s incredibly rapid pace, and also too short of a timeframe to determine true research trends.

        Usually research is considered rapid if there is meaningful progression within a few years, and more realistically about a decade or so. I mean, take something like real time ray tracing, for comparison.

        When I’m talking about the future of AI, I’m thinking like 10-20 years. We simply don’t know enough about what is possible to say what will happen by then.

  • Flying Squid@lemmy.world
    link
    fedilink
    English
    arrow-up
    59
    arrow-down
    3
    ·
    1 year ago

    Movie and TV executives don’t care about boring. Reality shows are boring. They just care if they make money.

        • currycourier@lemmy.world
          link
          fedilink
          English
          arrow-up
          13
          arrow-down
          1
          ·
          1 year ago

          ChatGPT isn’t the entirety of AI, AI research has been going on much longer than ChatGPT has been around

            • NoMoreCocaine@lemmy.world
              link
              fedilink
              English
              arrow-up
              3
              ·
              1 year ago

              Yes. Honestly it’s crazy how much people read into ChatGPT, when in practice it’s effectively just a dice roller that depends in incredibly big dataset to guess what’s the most likely word to come next.

              There’s been some research about this, the fact that people are assigning intelligence into things that ML does. Because it doesn’t compute for us that something can appear to make sense without actually having any intelligence. To humans, the appearance of the intelligence is enough to assume intelligence - even if it’s just a result of a complicated dice roller.

        • matter@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          1 year ago

          AI has been being developed for 50 years and the best we can do so far is a dunning-kruger sim. Sure, who knows what it “can do” at some point, but I wouldn’t hold my breath.

      • danque@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        2
        ·
        edit-2
        1 year ago

        Depends on the ai though. With koboldcpp you can make memories for the ai to come back with. Even text personalities (like bitchy and sassy responses) when using tavernai together with kobold.

      • jandar_fett@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        1
        ·
        1 year ago

        This. You have to baby it and then if you want it to do something different you have to tell it a hundred times in a hundred different ways before it stops producing the same stuff with the same structure with slight differences. It is a nightmare.

      • Flying Squid@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        I agree, but at some point it will advance to the level where it can write boring, predictable scripts.

        • kameecoding@lemmy.world
          link
          fedilink
          English
          arrow-up
          7
          ·
          1 year ago

          I mean it’s literally never “computer bad” / “technology bad” it’s always Humans bad, using this technology this time.

      • Pregnenolone@lemmy.world
        link
        fedilink
        English
        arrow-up
        9
        arrow-down
        1
        ·
        1 year ago

        I thought this season was way better than the season before it. I was glad they went with something different like the horror theme. The season prior was a shit show of boring tropes.

      • Honytawk@lemmy.zip
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        They just ran out of ideas because they used all the good ones in the first seasons.

    • ItsMeSpez@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 year ago

      I just recently started rewatching some of the older episodes and I realized that “Be RIght Back” was inadvertently an LLM episode. Having a computer absorb the online presence of a loved one to allow you to talk with them after they’ve passed is honestly something that seems within reach for these models.

      • AVincentInSpace
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 year ago

        Fun fact, that’s what Replika was originally designed for, before they realized they could make more money marketing it as a therapist and/or erotic roleplay partner.

  • homoludens@feddit.de
    link
    fedilink
    English
    arrow-up
    56
    arrow-down
    2
    ·
    1 year ago

    Bold of him to assume that companies would not just publish the trash - and that people would not watch it anyway.

      • drislands@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        ·
        1 year ago

        100% correct. The IRC channel I hang out in has a bot utilizing ChatGPT and it does a summary of the most recent conversations when someone joins.

        Sometimes, it does a great job! It impresses me how well it’s able to summarize multiple ongoing conversations in a succinct way.

        …and often times, it gets shit quite wrong. Not the actual topics, those it’s good at – but it is outright terrible at correctly indicating who actually said what.

        Granted this is all to be expected – it’s an LLM, not really AI.

      • banneryear1868@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        They wouldn’t have AI produce the whole show like that, it’s like feeding it a context to create dialogue within set parameters.

          • banneryear1868@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            1
            ·
            1 year ago

            Overall yea, but there’s a thing about how the more niche and specialized content is to a person’s interests, the more they’re willing to sacrifice on quality. So I think what will happen, because this will also reduce production costs so much (in theory), is we’ll get these incredibly specific shows made for smaller and smaller target audiences. I’m hoping this ends up generating some hilarious content that just seems absurd to people who aren’t targeted. Instead of catering to universal human experiences it will be like, “a show centered around a support group for people who love black licorice, and the challenges they face in their relationships with people who hate black licorice.”

    • banneryear1868@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      Yeah as if there already isn’t complete trash, presumably it will just be cheaper and easier to produce, so expect more ubiquitous and niche trash!

  • deafboy@lemmy.world
    link
    fedilink
    English
    arrow-up
    52
    arrow-down
    5
    ·
    1 year ago

    “I was frightened a second ago; now I’m bored because this is so derivative.” - Me, while watching some of the Black Mirror episodes, proudly made by fellow humans.

  • thefloweracidic@lemmy.world
    link
    fedilink
    English
    arrow-up
    38
    arrow-down
    1
    ·
    1 year ago

    I’m not to worried about AI. Isn’t the next iteration of GPT closed source? Technology is made best as a research or passion project, but once profits become the focus everything goes down hill. That and when you consider the global supply chain required to manufacture the chips that AI depends on, well things aren’t looking too great in that department.

    Tl;DR humans will shit all over the prospect of scary intelligent AI well before we get there.

    • darth_helmet@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      17
      ·
      1 year ago

      “Open”AI is entirely proprietary and closed-source.

      Meta’s Llama series are kind of open source, but don’t publish the weights and so can’t really be reproduced with full accuracy without a ton of manual effort.

      These and many other companies in the hype-space are using the same published research from a few years ago, which is why they have similar qualities.

  • quindraco@lemm.ee
    link
    fedilink
    English
    arrow-up
    28
    arrow-down
    1
    ·
    1 year ago

    Did he not watch the latest season? Fuck, one episode was literally devoid of scifi entirely. Latest season only had one good ep in it.

      • MeatsOfRage@lemmy.world
        link
        fedilink
        English
        arrow-up
        26
        arrow-down
        2
        ·
        1 year ago

        Eh, I never bought this. The show has always been wildly uneven. The first season was very strong with two great episodes but the first episode is not great. The second season is pretty bad overall with the exception of Be Right Back. White Christmas was a good one off. The first Netflix season had some really strong showings with Nosedive, Shut Up And Dance and my personal favorite of the whole series San Junipero. I even thought season 4 was pretty decent overall. Also, banderanatch was actually pretty cool, especially if you really dug into all the paths.

        The last 2 seasons, 5 & 6 were pretty bad overall with only one good episode each and some particularly bad ones. But honestly Charlie is probably just running out of ideas and I can’t really blame him at this point. I suspect he’s just trying different things. Sometimes big swings work and sometimes you just get Mazey Day.

        • P03 Locke@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          8
          arrow-down
          1
          ·
          1 year ago

          I really, really could not get behind the premise that the PM of the UK would fuck a pig to save a princess.

          In the real world, it would have been “we do not make deals with terrorists” and that would be that, ratings be damned. Like holy shit, did he start with a completely unrealistic premise.

          • dimeslime@lemmy.ca
            link
            fedilink
            English
            arrow-up
            7
            arrow-down
            3
            ·
            edit-2
            1 year ago

            Edit: complete fabrication be here.

            The “do horrible thing to save someone” story, yeah, maybe.

            But the pig thing had been going around for years about the then/former prime minister. Some kind of hazing at eton or wherever those posh folks go. The episode was just a reference to that.

        • linearchaos@lemmy.world
          link
          fedilink
          English
          arrow-up
          7
          ·
          edit-2
          1 year ago

          I think he’s trying to steal the lightning from the episodes that resonated well with the audiences to make new episodes and it’s just not hitting quite as hard.

          The old formula was amazing stories meets Twilight zone meets tales from the crypt we’re 8 out of 10 times the protagonist gets right fucked for just trying to be a good person.

          It feels to me like the take away from san junipero was that people like happy endings so they’re trying to apply it to everything

          The reason you got Maisy day is because he got access to Miley Cyrus and had to write something that she would be happy with.

        • dangblingus@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          Nah. Even season 3 there’s a distinct lack of “okay this is supremely fucked up” that you got in series 1 and 2. There’s some fantastic episodes in the first couple netflix seasons, but they dialed back the dread factor in favor of exploring new sci fi concepts. Not complaining, but the og channel 4 episodes are way darker.

        • aesthelete@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          my personal favorite of the whole series San Junipero.

          Me too. It’s literally the only episode I rewatched. 🙌

        • Cheesus@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          It comes down to what you like about the show. Do you like the darker sides of technology that raises questions? Seasons 1 and 2 are great. Do you like sci-fi happy endings with a better production value? The Netflix seasons are better.

      • kratoz29@lemm.ee
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 year ago

        I thought it was a Netflix original 😂

        Where did this tv show aired then if not Netflix?

      • Echo Dot@feddit.uk
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 year ago

        He does a series called yearly wipe. I’m pretty sure it’s only ever shown on UK television but it’s definitely worth watching.

        • dangblingus@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          1 year ago

          The yearly wipe hasn’t really been a thing for the last 5 years or so since he’s been focusing on Black Mirror/Netflix. It’s given more opportunity to Philomena Cunk to become a full fledged character. But I fucking love How TV Ruined Your Life and Screenwipe!

  • egeres@lemmy.world
    link
    fedilink
    English
    arrow-up
    14
    arrow-down
    1
    ·
    edit-2
    1 year ago

    Maybe the 5th episode of the 6th season was written by an AI and they were playing some 4D chess game all along with our minds, because otherwise, I wonder how such fucking trash got the green light to be produced 🤗

    Edit: Typo

    • gronjo45@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      1 year ago

      This. I think the only one I really thought was good was the Aaron Paul one where they went into space… I might be someone neo-ludditish but that movie shows some true terrors of those who want to eradicate technologies and the individuals associated with them. Cold ending…

      • egeres@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 year ago

        By far my favorite episode of this season, it felt like a refreshing scifi 50’s comic, it felt like reading something new from asimov. The retro-aesthetic was a nice artistic decision to tell us that tech doesn’t have to be super advanced to tell a good story. On top of this, they subverted my expectations at least 3 times:

        spoiler

        First, when the guy who draws sees the wife of aaron, I immediately though the story would be that she cheats on him and they both play mindgames on aaron who eventually looses his family. But no, she does feel something about the other guy but to my surprise, never cheats on aaron

        spoiler

        When the guy started to paint the house I though that “of course, he paints the wife naked because they have sex, and then aaron discovers this”. Indeed it happens, but interestingly, not because the other astronaut had sex with aaron’s wife

        spoiler

        By the end it was veeery clear to me that the other guy will either kill aaron, or trap him in some way to take control of him and live his life. It was obvious to me that the other astronaut was going to eject aaron from the ship an cast him away in space to then report that aaron had gone missing on space. I was extreeemely confident about this in the scene where the door is taking long to be opened, but no. Actually, yes, the other guy fucks aaron, but by killing his family so that he learns to value what he has, I found that quite unexpected and interesting

        What I didn’t get from the episode was what they were tying to tell with the child, I’m not sure what is that meant to communicate

  • quams69@lemmy.world
    link
    fedilink
    English
    arrow-up
    14
    arrow-down
    4
    ·
    1 year ago

    Lmao black mirror feels like it was written by an AI so quite a statement

  • psmgx@lemmy.world
    link
    fedilink
    English
    arrow-up
    14
    arrow-down
    6
    ·
    1 year ago

    It’s only producing trash now. Already there is a decent jump in quality from GPT-3 to 4, and it’s only gonna get better.

    Plus it can do a lot of heavy lifting – tell it to make 20 scripts with different prompts and then a single writer or team can Whittle them down. That’s how a lot of scripts end up in production anyways, but now you ain’t gotta deal with writers and can make rapid, drastic changes

    • Monkeytennis@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      1
      ·
      1 year ago

      I also find the “just look how bad the hands are heh heh heh” thing so dumb … it’s going to learn how to draw hands pretty quickly

    • Evotech@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      The problem atm it’s that chat gpt has pretty terrible memory. It couldn’t write a coherent show if it wanted to

    • olutukko@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      I’m here worried about the fact that it’s going to take some jobs, but leave some. So there will be huge unequality for a while until ai cna actually do all jobs

    • steltek@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      There’s more to “AI” than ChatGPT. Deepfakes, propaganda swarms, precise tracking of people online across pseudonyms/handles. The power available to malicious organizations and governments is absolutely terrifying. Any social media that doesn’t also have AI-based countermeasures is vulnerable.

    • hedgehog@ttrpg.network
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      1 year ago

      Both possibilities can be concerning, but that doesn’t mean every discussion of one of them necessarily has to include discussion of the other.

  • ILikeBoobies@lemmy.ca
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    2
    ·
    1 year ago

    Most of telly is trash already, if it’s cheap enough for entry then it can saturate the market and there will be no need for the expensive “good” writers

  • gregorum@lemm.ee
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    9
    ·
    1 year ago

    Didn’t he just make an episode of Black Mirror depicting the opposite?

    • ZeroCool@feddit.ch
      link
      fedilink
      English
      arrow-up
      15
      ·
      edit-2
      1 year ago

      I don’t think so. He says he isn’t afraid of AI replacing creative jobs because it’s incapable of originality and as a result, boring. The episode you’re referencing didn’t depict the opposite. There was a quantum computer that was only capable of producing a show that recreated data from the protagonist’s life in real time. It was always limited by actual events as they played out. That episode seems fairly consistent with his views.