cross-posted from: https://lemmy.ca/post/37011397
The popular open-source VLC video player was demonstrated on the floor of CES 2025 with automatic AI subtitling and translation, generated locally and offline in real time. Parent organization VideoLAN shared a video on Tuesday in which president Jean-Baptiste Kempf shows off the new feature, which uses open-source AI models to generate subtitles for videos in several languages.
The nice thing is, now at least this can be used with live tv from other countries and languages.
Think you want to watch Japanese tv or Korean channels with out bothering about downloading, searching and syncing subtitles
I prefer watching Mexican football announcers, and it would be nice to know what they’re saying. Though that might actually detract from the experience.
GOOOOOOAAAAAAAAALLLLLLLLLL
This might be one of the few times I’ve seen AI being useful and not just slapped on something for marketing purposes.
And not to do evil shit
When we getting amd’s fsr upscaling and frame-gen? Also would subtitles make more sense to use the jellyfin approach.
Will it be possible to export these AI subs?
Imagine the possibilities!
The technology is nowhere near being good though. On synthetic tests, on the data it was trained and tweeked on, maybe, I don’t know.
I corun an event when we invite speakers from all over the world, and we tried every way to generate subtitles, all of them run on the level of YouTube autogenerated ones. It’s better than nothing, but you can’t rely on it really.No, but I think it would be super helpful to synchronize subtitles that are not aligned to the video.
You were not able to test it yet calling it nowhere near good 🤦🏻
Like how should you know?!
Relax, they didn’t write a new way of doing magic, they integrated a solution from the market.
I don’t know what the new BMW car they introduce this year is capable of, but I know for a fact it can’t fly.
Hes maaaa, hes what? Hes maaaaaa!
This sounds like a great thing for deaf people and just in general, but I don’t think AI will ever replace anime fansub makers who have no problem throwing a wall of text on screen for a split second just to explain an obscure untranslatable pun.
That still happens? Maybe wanna share your groups? ;)
Bless those subbers. I love those walls of text.
It’s unlikely to even replace good subtitles, fan or not. It’s just a nice thing to have for a lot of content though.
I have family members who can’t really understand spoken English because it’s a bit fast, and can’t read English subtitles again, because again, too fast for them.
Sometimes you download a movie and all the Estonian subtitles are for an older release and they desynchronize. Sometimes you can barely even find synchronized English subtitles, so even that doesn’t work.
This seems like a godsend, honestly.
Funnily enough, of all the streaming services, I’m again going to have to commend Apple TV+ here. Their shit has Estonian subtitles. Netflix, Prime, etc, do not. Meaning if I’m watching with a family member who doesn’t understand English well, I’ll watch Apple TV+ with a subscription, and everything else is going to be pirated for subtitles. So I don’t bother subscribing anymore. We’re a tiny country, but for some reason Apple of all companies has chosen to acknowledge us. Meanwhile, I was setting up an Xbox for someone a few years ago, and Estonia just… straight up doesn’t exist. I’m not talking about language support - you literally couldn’t pick it as your LOCATION.
For all their faults, Apple knows accessibility. Good job Timmy.
They are like the * in any Terry Pratchett (GNU) novel, sometimes a funny joke can have a little more spice added to make it even funnier
Translator’s note: keikaku means plan
No such comment yet? I’ll be the first then.
Oh no, AI bad, next thing they add is cryptocurrency mining!
ai for accessibility is nowhere near the same thing as crypto mining
Lol, that is his opinion as well, this was sarcasm, i’m 99% sure Yours does not seem so sarcastic…
But it’s burning amazon forests for capitalist greed!
/s
I will be impressed only when it can get through a single episode of Still Game without making a dozen mistakes
What’s important is that this is running on your machine locally, offline, without any cloud services. It runs directly inside the executable
YES, thank you JB
Justin Bieber?
Jack Black?
James Brown?
Finally, some good fucking AI
Finally some good AI fucking 🤭
I was just thinking, this is exactly what AI should be used for. Pattern recognition, full stop.
Yup, and if it isn’t perfect that is ok as long as it is close enough.
Like getting name spellings wrong or mixing homophones is fine because it isn’t trying to be factually accurate.
I’d like to see this fix the most annoying part about subtitles, timing. find transcript/any subs on the Internet and have the AI align it with the audio properly.
YES! I can’t stand when subtitles are misaligned to the video. If this AI tool could help with that, it would be super useful.
Problem ist that now people will say that they don’t get to create accurate subtitles because VLC is doing the job for them.
Accessibility might suffer from that, because all subtitles are now just “good enough”
Regular old live broadcast closed captioning is pretty much ‘good enough’ and that is the standard I’m comparing to.
Actual subtitles created ahead of time should be perfect because they have the time to double check.
Or they can get OK ones with this tool, and fix the errors. Might save a lot of time
Honestly though? If your audio is even half decent you’ll get like 95% accuracy. Considering a lot of media just wouldn’t have anything, that is a pretty fair trade off to me
From experience AI translation is still garbage, specially for languages like Chinese, Japanese, and Korean , but if it only subtitles in the actual language such creating English subtitles for English then it is probably fine.
That’s probably more due to lack of training than anything else. Existing models are mostly made by American companies and trained on English-language material. Naturally, the further you get from the model, the worse the result.
It is not the lack of training material that is the issue, it doesn’t understand context and cultural references. Someone commented here that crunchyroll AI subtitles translated Asura Hall a name to asshole.
English it’s been great for me yes
I imagine it would be not-exactly-simple-but-not- complicated to add a “threshold” feature. If Ai is less than X% certain, it can request human clarification.
Edit: Derp. I forgot about the “real time” part. Still, as others have said, even a single botched word would still work well enough with context.
That defeats the purpose of doing it in real time as it would introduce a delay.
Derp. You’re right, I’ve added an edit to my comment.
Yeah it’s pretty wonderful To see how far auto generated transcription/captioning has become over the last couple of years. A wonderful victory for many communities with various disabilities.
And yet they turned down having thumbnails for seeking because it would be too resource intensive. 😐
I mean, it would. For example Jellyfin implements it, but it does so by extracting the pictures ahead of time and saving them. It takes days to do this for my library.
Video decoding is resource intensive. We’re used to it, we have hardware acceleration for some of it, but spewing something around 52 million pixels every second from a highly compressed data source is not cheap. I’m not sure how both compare, but small LLM models are not that costly to run if you don’t factor their creation in.
It is useful for internet streams though, not really for local or lan video.
Now I want some AR glasses that display subtitles above someone’s head when they talk à la Cyberpunk that also auto-translates. Of course, it has to be done entirely locally.
I guess we have most of the ingredients to make this happen. Software-wise we’re there, hardware wise I’m still waiting for AR glasses I can replace my normal glasses with (that I wear 24/7 except for sleep). I’d accept having to carry a spare in a charging case so I swap them out once a day or something but other than that I want them to be close enough in terms of weight and comfort to my regular glasses and just give me AR like overlaid GPS, notifications, etc, and indeed instant translation with subtitles would be a function that I could see having a massive impact on civilization tbh.
soon
Breaking news: “WW3 starts over an insult due to a mistranslated phrase at the G7 summit. We will be nuked in 37 seconds. Fuck like rabbits, it’s all we can do. Now over to Robert with traffic.”
It’d be incredible for deaf people being able to read captions for spoken conversations and to have the other person’s glasses translate from ASL to English.
Honestly I’d be a bit shocked if the AI ASL -> English doesn’t exist already, there’s so much training data available, the Deaf community loves video for obvious reasons.
I think we’re closer with hardware than software. the xreal/rokid category of hmds are comfortable enough to wear all day, and I don’t mind a cable running from behind my ear under a clothes layer to a phone or mini PC in my pocket. Unfortunately you still need to byo cameras to get the overlays appearing in the correct points in space, but cameras are cheap, I suspect these glasses will grow some cameras in the next couple of iterations.
I believe you can put prescription lenses in most AR glasses out there, but I suppose the battery is a concern…
I’m in the same boat, I gotta wear my glasses 24/7.
I hope Mozilla can benefit of a good local translation engine that could come out of it as well.
They technically already do with Project Bergamot.
I know they do, but it’s lacking so many languages.
Haven’t watched the video yet, but it makes a lot of sense that you could train an AI using already subtitled movies and their audio. There are times when official subtitles paraphrase the speech to make it easier to read quickly, so I wonder how that would work. There’s also just a lot of voice recognition everywhere nowadays, so maybe that’s all they need?