Subtitle Sphere Logo

Subtitle Sphere

A product of Tandis 24 Design Lab

Subtitle Sphere Sub Logo

Frequently Asked Questions

1. What is Subtitle Sphere?

Subtitle Sphere is a powerful AI transcription, translation, subtitling, and narration tool designed for desktop use. It supports over 128 languages for translation and offers AI-powered transcription for video and audio files in 58 languages. Users can generate SRT files from text, create subtitles, and utilize AI-powered voice narration in multiple languages. As a standalone application, Subtitle Sphere works offline and requires no subscription, ensuring user privacy while providing unlimited transcription, translation, subtitling, and narration capabilities without relying on cloud services. However, since most of the processing (transcription and subtitling) happens offline on your computer, it can take up system memory, and processing times may be slower depending on your device’s specifications. This is a trade-off between privacy and speed.

2. Does it require an internet connection?

No, Subtitle Sphere is a standalone application that works offline once downloaded. However, an internet connection is required for the translation, SRT generation from plain text, and text-to-speech features.

3. How do I download the software?

To download the software, go to the download page, accept the End User Agreement, and you will be provided with a link. Since the file is hosted on Google Drive and is larger than 100MB, you will need to manually accept the download. While the file is safe and virus-free, please note that Google cannot scan files larger than 100MB.

4. What languages are supported?

Subtitle Sphere supports 128 languages for translation and 58 languages for transcription and AI-powered narration. For a complete list of supported languages, please check our Languages page.

5. Is there a subscription fee?

No, Subtitle Sphere is completely free and does not require any subscription fees. If you are charged for using our service or come across other websites selling it, please report such incidents through our contact page, as they are illegal.

6. How do I install Subtitle Sphere?

After downloading, visit our instructions page to complete the installation process.

7. Can I use Subtitle Sphere on multiple devices?

Yes, you can install Subtitle Sphere on multiple devices as long as each device meets the system requirements.

8. What if I encounter issues during installation?

If you encounter any issues during installation, please visit our contact page. You can reach us by email, send us a direct message on Instagram, or leave a comment on our YouTube page.

9. How can I provide feedback or report a bug?

We welcome your feedback! Please use our contact page to send us your comments or report any bugs you encounter.

10. Is my data stored anywhere?

No, Subtitle Sphere does not collect or store any user data. The software is standalone and operates offline on your computer, ensuring your privacy. However, for translation and AI narration, it utilizes Google Translate and Google Text-to-Speech services, which require an internet connection and transmit text to these services. Users should consult the privacy and data collection policies of these third-party services, as Subtitle Sphere is not responsible for their data collection practices, as outlined in the End User License Agreement (EULA). Because most of the process is done offline on your machine, memory usage may increase, and processing could be slower depending on your system’s capabilities—a trade-off for enhanced privacy.

11. Are there any limitations on usage?

Subtitle Sphere does not impose any limitations on the number of transcriptions, translations, or narrations you can perform. There are also no restrictions on data usage for transcription, ensuring you can work freely. However, for services such as translation and AI narration, which rely on Google Translate and Google gTTS, users should consult those third-party services for their data usage policies and limitations. Subtitle Sphere is not responsible for any personal data usage that exceeds the limits set by these external services.

12. How do I update the software?

Since this is a beta version and still evolving, you can stay informed about updates by following our social media channels. Keep an eye out for the latest improvements and features!

13. What operating systems are supported?

Subtitle Sphere is compatible with macOS, and Windows(coming soon) operating systems.

14. Can I customize subtitle styles?

Yes, Subtitle Sphere allows you to customize the styles, colors, and fonts of your subtitles to match your preferences. However, the font type can only be customized for languages that use the English alphabet. For languages with special characters, a specific font is required to ensure proper display of those characters.

15. Where can I find additional resources or tutorials?

Visit our Instructions page for guides, videos, and other helpful resources to maximize your use of Subtitle Sphere. Follow us on Instagram and YouTube for the latest updates and tips! Links to our social channels can be found on our contact page.

16. Which models are used in this software?

For transcription, the software uses Python's OpenAI Whisper library (free version), which works offline and is integrated directly into the application. For translation, it utilizes the Python's Google Translate via the deep-translator Python module, which requires an internet connection. For text-to-speech, the software relies on Python's gTTS (Google Text-to-Speech) library, which also requires an internet connection.

17. What types of files are supported?

Subtitle Sphere supports various file types. For inputs, it accepts video formats like .mov and .mp4, audio formats like .wav, as well as text and .srt files. The outputs can be either .mp4 or .srt files, depending on your needs.

18. How accurate are the transcription and translation services?

The accuracy of transcription depends on the quality of the audio, the language being transcribed, and the chosen Whisper model (tiny, base, small, medium, or large). The large model offers the highest accuracy, while the tiny model has the lowest but still works well for languages like English and French, which have extensive training data. However, larger models are more computationally intensive, leading to a higher load on your computer. Translation accuracy also depends on the chosen language, with commonly spoken languages generally having higher precision.

19. How long does translation, transcription, and subtitling take?

The time required for translation, transcription, and subtitling depends on several factors, including the length of the video or audio, the file size and quality, and the computational power of your device. More powerful systems will process tasks faster. Selecting larger Whisper models for transcription (like medium or large) will increase both accuracy and processing time. Since the majority of the processing happens offline on your machine, this can affect your system’s memory usage and slow down the process, especially with larger files. Translation speeds are also influenced by internet connection quality.

20. Are there any limits to the file types or sizes?

There is no strict limit on file size, but larger files will take longer to process. Currently, Subtitle Sphere supports the following file types for input: .mov, .mp4, mp3, .wav, .txt, and .srt. If your file is larger or in a different format, it may require more time or conversion to one of the supported types. We are working to expand support for additional formats in future updates.

21. Does Subtitle Sphere support multilingual content in the same file?

Yes, Subtitle Sphere supports multilingual transcription with Whisper. By selecting "Auto" for the source language, the model can detect and transcribe multiple languages in the same file. For best results, use the medium or large model. Note that Whisper may prioritize one language if there's a dominant accent, even when multiple languages are spoken. You can also translate the transcription into any target language after processing.

22. What is Whisper, and how does it work in Subtitle Sphere?

Whisper is an AI-powered speech recognition system developed by OpenAI, used in Subtitle Sphere for transcription and subtitling. It operates offline, ensuring privacy by processing audio directly on your computer. Whisper offers five models—Tiny, Base, Small, Medium, and Large—which vary in speed and accuracy. For clear audio in languages like English and French, the smaller Tiny and Base models are sufficient and faster. For multilingual content or complex audio, the Medium and Large models offer better accuracy but take longer to process. Users should note that larger models reduce the risk of "hallucinations" (incorrect transcriptions), especially with less common languages or difficult audio. You can find more information about Whisper on OpenAI's GitHub page.

23. Why am I getting this error: "Error during translation: invalid literal for int() with base 10: ‘’"?

If you're seeing the error message, "Error during translation: invalid literal for int() with base 10: ‘’", it usually means there is an extra line at the end of your SRT file. An SRT file typically consists of a subtitle ID, a time stamp, and the subtitle text. Each subtitle block should be followed by a blank line, including the last one in the file. However, sometimes there may be additional blank lines or spaces after the final blank line. To resolve the issue, ensure there is exactly one blank line after the last subtitle, and no more. Once this is corrected, click "Update Subtitles" and try translating again. If the issue persists, please don't hesitate to contact us.

24. Why was the transcribing process interrupted and returned to the main menu when using the Whisper-Google Fusion option?

If the transcribing process was interrupted and the main menu appeared, it likely means there was either a disruption in your internet connection or you have reached Google’s Speech Recognition usage limit. Please check your internet connection and try again in a few minutes. Since you are using a third-party service (Google Speech Recognition), it is recommended to review their quota limits and ensure you stay within their usage guidelines.

25. What is the Whisper-Google Fusion transcribing option?

The Whisper-Google Fusion is a proprietary transcription method developed by Tandis 24 Design Lab. Without going into too much detail, it combines the strengths of OpenAI’s Whisper transcription with Google Speech Recognition. This approach is particularly useful for less common languages, where smaller Whisper models may struggle with accuracy, and larger models can be computationally demanding. The Whisper-Google Fusion model offers a faster and more accurate transcription by balancing these two technologies. The slider bar allows you to adjust the accuracy of the transcription—after a few attempts, you'll get the hang of it and be able to fine-tune it for better results. We apologize for not providing a more detailed explanation, as this is a proprietary method.

26. Why with Google video/audio transcription option, I only get a long text and no time stamps?

The Google video/audio transcription option utilized in Subtitle Sphere is based on the free version of Google Speech Recognition. While this tool excels at producing coherent and high-quality text transcriptions, it unfortunately does not provide timestamps. If your primary need is accurate text without timestamps, Google Speech Recognition is a great choice and can serve as a reliable benchmark for your content. However, for users requiring timestamped transcriptions, we recommend leveraging Subtitle Sphere's features with the Whisper or Whisper-Google Fusion models.