Teachers describe a deterioration in behaviour and attitudes that has proved to be fertile terrain for misogynistic influencers

“As soon as I mention feminism, you can feel the shift in the room; they’re shuffling in their seats.” Mike Nicholson holds workshops with teenage boys about the challenges of impending manhood. Standing up for the sisterhood, it seems, is the last thing on their minds.

When Nicholson says he is a feminist himself, “I can see them look at me, like, ‘I used to like you.’”

Once Nicholson, whose programme is called Progressive Masculinity, unpacks the fact that feminism means equal rights and opportunities for women, many of the boys with whom he works are won over.

“A lot of it is bred from misunderstanding and how the word is smeared,” he says.

But he is battling against what he calls a “dominance-based model” of masculinity. “These old-fashioned, regressive ideas are having a renaissance, through your masculinity influencers – your grifters, like Andrew Tate.”

  • Bull205@sh.itjust.works
    link
    fedilink
    arrow-up
    55
    arrow-down
    6
    ·
    10 months ago

    YouTube Algorithms, facebook Algorithms, etc. make them all publishers responsible for their content.

      • MrMcGasion@lemmy.world
        link
        fedilink
        arrow-up
        10
        ·
        edit-2
        10 months ago

        I forget who I heard it from, but some bigger YouTuber mentioned that when talking to someone at YouTube about “the algorithm” and the person who worked at YouTube suggested rather than always thinking about it being the algorithm that drives what’s popular, that it’s the users who engage with that content. In the “line goes up” capitalist mindset, the algorithms at these companies are really just designed around engagement, and keeping people hooked. The “algorithm” is just what it thinks the audience wants.

        And while I think a lot of us would like to think better of ourselves, I think we all have a strong tendency to engage with ragebait, and “shocking” content. Which wouldn’t necessarily be a bad trait in a pre-internet world. But in the world where the shareholders always need the line to go up infinitely, all of our media gets filled with the garbage that makes the line go up the most.

        In the short term, we can all try more to engage less with the kind of content, showing the algorithms that we don’t actually want that content.

        In the long term, we should probably de-couple our media from the infinite-growth investor-first capitalism that has formerly-respected publications writing articles about what 5 random people said on Twitter that they can ragebait people into engaging with.

        • yamanii@lemmy.world
          link
          fedilink
          arrow-up
          2
          ·
          10 months ago

          Yes people like stuff that’s not good for them, violence focused “journalistic” shows were all the rage during the early millenium since they did get a big viewership, but nowadays they are mostly over with only a few left, we should demand change from those that have the power to do it.

      • Bull205@sh.itjust.works
        link
        fedilink
        arrow-up
        10
        ·
        10 months ago

        Yes but…I try ink media literacy is something that isn’t necessarily intuitive. It can and should be taught in elementary and secondary schools.

      • GoodbyeBlueMonday@startrek.website
        link
        fedilink
        English
        arrow-up
        1
        ·
        10 months ago

        True, but imagine if we gave everyone an automatic weapon and told people they need to be responsible for what they choose to shoot. True, but we probably shouldn’t have given out so many weapons.

        It’s a terrible metaphor, but there’s an intersection between personal, collective, corporate, and technological responsibility that we need to consider, and it’s hard to articulate in a few sentences. IMHO we’re all in an ouroboros of thought and action, internally and externally.

      • AVincentInSpace
        link
        fedilink
        English
        arrow-up
        11
        arrow-down
        1
        ·
        edit-2
        10 months ago

        Not necessarily. If it only applies to sites with algorithmic feeds (i.e. specifically ones that serve individualized streams to each user based on what they specifically have liked in the past), companies who choose to be in control of what content they show are held to account and smaller platforms are safe.

        • maynarkh@feddit.nl
          link
          fedilink
          arrow-up
          7
          ·
          10 months ago

          Or do what the EU did with the DMA.

          Write a law specifically for megacorps and only megacorps. It’s possible and it works.

          • AVincentInSpace
            link
            fedilink
            English
            arrow-up
            2
            ·
            10 months ago

            Not if algorithmic feeds are defined as ones that show individualized feeds to each user, like I said in my comment

                • Herbal Gamer@sh.itjust.works
                  link
                  fedilink
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  10 months ago

                  Yeah, the definition in the law would have to be based on what constitutes an algorithm. That’s what I meant, doofus

                  • AVincentInSpace
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    ·
                    edit-2
                    10 months ago

                    What specific algorithms the law applies to does not have to be all pieces of code that could be conceivably classed as an “algorithm”. The law can use a different word if it makes you feel better