pleroma.debian.social

pleroma.debian.social

This year, as an experiment, we're also generating vtt subtitles of all released videos at @fosdem. These are generated after publishing of the video using OpenAI whisper.

I would love to get feedback about these subtitles. Are people using them? Are they useful?

As they're plain text files, it should also be possible to perform some analysis at scale on the contents of all the talks that were performed at FOSDEM. If anyone is doing something like this, please let me know!

@wouter @fosdem
*To clarify: They're using locally running whisper models, not through openai's api's

@fosdem @wouter oh ffs is this not just more use of worldburning theft machines?

@mirabilos
Maybe do some research before assuming 'AI' == power wastage? The tiny model is small enough that you can use it to transcribe live audio on a raspberry pi.
@fosdem

@wouter @fosdem hmmmh. And how energetically and environmentally expensive and legal and ethic was its training?

@mirabilos
Audio transcription is not an LLM. I don't know how much power OpenAI spent training whisper models, because honestly I don't care. That said, this type of thing was possible with the state of the art 20 years ago, so I can't imagine it needed that much.

Thanks for the feedback (I know I did ask), but EOT for me.
@fosdem
replies
0
announces
0
likes
1

@wouter interesting. Is thre ny validation done as to the orecness of the subtitles?

@lieter
None whatsoever at this point. This is part of it being an experiment 😉

@lieter @wouter love how there are three typos in your message, on purpose I assume? 🙃