• I’m highlighting that speech to text and context awareness are different skills.

    YouTube is unlikely to waste loads of compute power on subtitles that don’t need it just to capture the occasional edge case.

    • lime!
      link
      fedilink
      English
      28 days ago

      i mean, it’s a one-time-per-video thing. they already do tons of processing on every upload.

        • lime!
          link
          fedilink
          English
          28 days ago

          right now they’re dynamically generating subtitles every time. that’s way more compute.

          • @aow@sh.itjust.works
            link
            fedilink
            18 days ago

            For real? That’s incredibly dumb/expensive compared to one subtitle roll. Can you share where you saw that?

            • lime!
              link
              fedilink
              English
              1
              edit-2
              7 days ago

              well, i have no evidence of this. however. looking at the way auto-generated subtitles are served at youtube right now, they are sent individually word-by-word from the server, pick up filler words like “uh”, and sometimes pause for several seconds in the middle of sentences. and they’re not sent by websocket, which means they go through multiple requests over the course of a video. more requests means the server works harder because it can’t just stream the text like it does the video, and the only reason they’d do that other than incompetence (which would surely have been corrected by now, it’s been like this for years) is if the web backend has to wait for the next word to be generated.

              i would love to actually know what’s going on if anyone has any insight.