https://archive.ph/nbF2s

  • JayTwo [any]
    ·
    edit-2
    7 months ago

    In one of the forums for the niche hobby I'm into, Google snippets have already been causing chaos for years.

    Eg: "You're wrong because Google says that I need to do this". Well Google is wrong and doing that is the entire reason why you're having so many problems.
    When investigating how the snippet was made it's either from a review or forum comment made by a newbie that somehow got traction or often someone saying DON'T do it that way, but their algorithm doesn't pick up the nuance and gets it twisted.
    More recently it's been through taking snippets of entirely AI pages that write absolute gibberish which sound impressive to people with only a passing familiarity.

    This is gonna make things sooooooo much worse.

    • InevitableSwing [none/use name]
      hexagon
      ·
      7 months ago

      I wonder if this will happen: "To make a chocolate milkshake: 1. Thank you kind stranger!..." And then news will break that Google is claiming a malformed algorithm caused their AI to suck up the entirety of Reddit. And after that Google will be forced to admit "Oh, oops!" the wonky algorithm caused Google to suck up ginormous amounts of data from 10,000s of sites on the net. Then they'll say they're "untraining" which is another big lie. All they'll do as fast as they can is smooth out plagiarism so they can have deniability.

      I wouldn't be surprised if a few years from now Google's legal team is at the supreme court claiming something ridiculous. Some of the best legal minds in the US are pushing the bullshit idea that AI cannot plagiarize because it doesn't know what plagiarism actually is. And the GOP majority seems to love the idea.

    • blobjim [he/him]
      ·
      edit-2
      7 months ago

      Three quarters of that is idiots not only believing what they see on the internet, but believing the first thing they see at the top of search results, without context.

      We should be using computers as a means of interfacing with things generated by other humans, not expecting some simple algorithm to read our minds.