Hi everyone!
A few days ago I released Whishper, a new version of a project I’ve been working for about a year now.
It’s a self-hosted audio transcription suite, you can transcribe audio to text, generate subtitles, translate subtitles and edit them all from one UI and 100% locally (it even works offline).
I hope you like it, check out the website for self-hosting instructions: https://whishper.net
Would love to deploy this, but unfortunately I’m running server equipment that apparently doesn’t support MongoDB 5 (Error message MongoDB 5.0+ requires a CPU with AVX support, and your current system does not appear to have that!). Tried deploying with both 4.4.18 and 4.4.6 and can’t get it to work. If anybody has some recommendations, I’d appreciate hearing them!
Edit: Changed my proxmox environment processor to host, fixed my issue.
I’m glad you were able to solve the problem, I add the comment I made to another user with the same problem:
Didn’t know about this problem. I’ll try to add a MariaDB alternative database option soon.
I’ve been looking for a tool to do this for YEARS, my god! Years!!! ❤️❤️
Oh, awesome! Does it do speaker detection? That’s been one of my main gripes with Whisper.
Unfortunately, not yet. Whisper per se is not able to do that. Currently, there are few viable solutions for integration, and I’m looking at this one, but all current solutions I know about need GPU for this.
This is excellent timing for me. I was just taking a break from working on setting up whisper.cpp with a web front end to transcribe interviews. This is a much nicer package than I ever had a chance of pulling together. Nice work!
Just tried this out but couldn’t get it to work until downgrading mongo to 4.4.6 because my NAS doesn’t ha``ve AVX support. But then, mongo stays unhealthy. No idea why.
Didn’t know about this problem. I’ll try to add a MariaDB alternative database option soon to solve this.
Even this is an good sound to text converter and a good ai transcription service