feat: tweak whisper to use local whisper.cpp instance#224
feat: tweak whisper to use local whisper.cpp instance#224hezirel wants to merge 1 commit intoRobitx:mainfrom
Conversation
|
@hezirel Sorry, I've been overrun at my day job (I hope to finish the current project at the end of November, so there is a light at the end of that tunnel..). ^ this was removed by mistake or on purpose? |
|
hi @Robitx ! Thanks for checking out my PR, and no need to excuse yourself, you are doing so much just as i is :) I've removed the language parameter on purpose due to the differences in requests parameters for the /inference endpoint of the whisper.cpp backend You can check out the full server implementation and accepted requests parameters of whisper.cpp server This was just a quick and dirty PR on my part to make this work. I mainly changed the audio encoding parameters to match the ones expected by whisper.cpp and prevent duplicate conversion of the audio file as to save resource. I think there is some structural work to be done to integrate this backend efficiently into the plugin, i just wanted to throw the idea out here about using whisper.cpp I'm free to discuss different options if you need it and again, thanks for all the work you did, you really augmented my workflow with this plugin, much appreciated ❤️ |
I made some changes to the whisper.lua file
Now by building whisper.cpp
Starting the server using the command
./server -m models/ggml-distil-large-v3.bin --convertAnd set your config.whisper.endpoint = "http://127.0.0.1:8080/inference"
You can now use fully local whisper, it runs so well on silicon it's incredible