Hey all,

Some might remember this from about 9 months ago. I’ve been running it with zero maintenance since then, but saw there were some new updates that could be leveraged.

What has changed?

  • Jellyfin is supported (in addition to Plex and Tautulli)
  • Moved away from whisper.cpp to stable-ts and faster-whisper (faster-whisper can support Nvidia GPUs)
  • Significant refactoring of the code to make it easier to read and for others to add ‘integrations’ or webhooks
  • Renamed the webhook from webhook to plex/tautulli/jellyfin
  • New environment variables for additional control

What is this?

This will transcribe your personal media on a Plex or Jellyfin server to create subtitles (.srt). It is currently reliant on webhooks from Jellyfin, Plex, or Tautulli. This uses stable-ts and faster-whisper which can use both Nvidia GPUs and CPUs.

How do I run it?

I recommend reading through the documentation at: McCloudS/subgen: Autogenerate subtitles using OpenAI Whisper Model via Jellyfin, Plex, and Tautulli (github.com) , but quick and dirty, pull mccloud/subgen from Dockerhub, configure Tautulli/Plex/Jellyfin webhooks, and map your media volumes to match Plex/Jellyfin identically.

What can I do?

I’d love any feedback or PRs to update any of the code or the instructions. Also interested to hear if anyone can get GPU transcoding to work. I have a Tesla T4 in the mail to try it out soon.

  • PoundKitchenB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Suhweeet!!! English only or will it handle other languages and translation too, Spanish to English?

    • McCloudOPB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      It can only translate into English, but the source audio can be a foreign language.

      • PoundKitchenB
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Great, that’s what I need!

        I see a Docker pull in my future.