• go $fsck yourself@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    1
    ·
    edit-2
    4 days ago

    Ah yes, the “splurdge” part of the article, a word everyone knows as a very technical term that’s used for filling in for an inability to articulate an actual line of logic. Instead of logic, just explain that “I don’t like the conclusion of the article, so it must be wrong somehow even though I can’t explain why.”

    It’s also important for people to chime in that an article must be wrong because it just “feels wrong”. Of course don’t actually provide any reasoning for it, because why would that be necessary?

    Don’t forget how useful it is to ask a question, completely ignoring that the article addresses it. Don’t even bring it up in the questioning. When someone points that out, then the best strategy is to lash out at them, because they were such a big meanie by pointing out the obvious problem of not reading the article.

    Lemmy communities are all about feelings, not information!

    Oh, also I took a screenshot of your comment because I knew you were going to edit it.

    • Paragone@lemmy.ca
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      1 day ago

      There is a “convenient” presumption in the article, and in what I’m replying-to:

      That audio is uploaded to big-tech, after someone speaks near their phone…

      Why TF would they need to do that, if they can cram a “hey google” audio-to-logic routine into mere kilobytes, within their neuro-DSP??

      The Active Listening system mentioned in the article needs to:

      1. have the onboard ai listen for voices,

      2. differentiate significant items ( for whatever values of “significance” they want to manipulate )

      3. upload … what, a YAML file, … of spoken keywords, & maybe an identification of if it was the owner or not who spoke them?

      Something like that…

      1. NOT react to vocal bots, or other idiocy ( a recording repeating something endlessly ought be immediately-ignored, as … from the Active Listening manipulation-system’s perspective … mere jamming )

      The “we aren’t seeing audio being uploaded” is a fucking red-herring, & any competent geek nowadays, who understands that the neuro-DSP chips in our phones can accomplish stupidly capable term-recognition on near-zero code, when in active-sleep ( or whatever that energy-state is called, screen off, but things still running in the background ), and that’s all they need.

      You know how in Google AdWords you pay for specific keywords?

      Guess what: the same system could easily be implimented within that subsystem which says “Sorry, the microphone is turned off”, when you say “Hey Google” after turning off the mic…

      it’s ALWAYS listening, it uses tiny power, & when it matches, then it activates other subsystems…

      Selling that ads can be show on phones/tablets that someone recently spoke a keyword for, would be idiot-simple for Google or Apple to impliment, & profitable-as-hell to boot.

      IF the “Hey Google” audio prompt can work in a specific energy-configuration of a device, THEN this could too, as it’s the same subsystem.

      https://github.com/KoljaB/RealtimeSTT

      as an example of something related to what I’m talking about, but not identical…