Whisper and its many forms

I just tested one SRT translation by redirecting SubtitleEdit to use DeepSeek V3 endpoints. The result looks better than DeepL. However it was rather slow speed. Might that be because SubtitleEdit is not using optimised calls, or should I expect that Deepseek API calls are slow (copmapred to DeepL)?
 
I just tested one SRT translation by redirecting SubtitleEdit to use DeepSeek V3 endpoints. The result looks better than DeepL. However it was rather slow speed. Might that be because SubtitleEdit is not using optimised calls, or should I expect that Deepseek API calls are slow (copmapred to DeepL)?
Hmm. This sounds interesting. I'd like to have a play.

Can I assume you redirected to the DeepSeek end point by changing the ChatGPT Auto-translate settings in SubtitleEdit?

But how did you get the DeepSeek model as an option available to select?
 
I played around with deepseek and VSM and have no idea how it works for you. For me, half the subtitle is not translated and the other half is empty.
Therefore, I wrote my own little script which does the exact same thing. I also tried one which works with chunks instead of each line seperately. This massively improves the translation quality (due to context) and speed. But for some reason, it bugs out sometimes and does not translate lines or whole chunks. Retries don't work either. Haven't figured out what the issue is. If anyone has an idea how to solve it, I would be very interested.

The script works as follows (both versions, chunk one is bugged as mentioned):

- pip3 install openai
- enter your API Key
- specify the path where all your subtitles are
- adjust the prompt if you are not happy with the one I provided
- run it
 

Attachments

  • Like
Reactions: mei2
I just tested one SRT translation by redirecting SubtitleEdit to use DeepSeek V3 endpoints. The result looks better than DeepL. However it was rather slow speed. Might that be because SubtitleEdit is not using optimised calls, or should I expect that Deepseek API calls are slow (copmapred to DeepL)?
They are really slow. One line of sub takes generelly 1.5s.
 
Last edited:
I played around with deepseek and VSM and have no idea how it works for you. For me, half the subtitle is not translated and the other half is empty.
Therefore, I wrote my own little script which does the exact same thing. I also tried one which works with chunks instead of each line seperately.

I've found this repo to do a reasonable job for gpt:

I haven''t tested it yet with Deepseek --the endpoints and models are in the coonfig file.
 
I've found this repo to do a reasonable job for gpt:

I haven''t tested it yet with Deepseek --the endpoints and models are in the coonfig file.
Need to try it but deepseek usually has an issue with sending the whole file at once. Token max is reached pretty soon.
 
I played around with deepseek and VSM and have no idea how it works for you. For me, half the subtitle is not translated and the other half is empty.
Therefore, I wrote my own little script which does the exact same thing.
I queued four files to translate with VSM yesterday. Two failed and the other two were much slower than previous times.
Maybe there is or was an issue at DeepSeek rather than at our end?

And it's great you've written your scripts. I'm going to try them.
I was planning to do the same myself but hadn't got around to it. A simple script is much more elegant than running an application just to send API calls.
I'll let you know if I have the same issue with the chunks version.