Wrote a tiny new LLM plugin this morning: llm-whisper-api, which lets you do this (if you have an OpenAI API key configured already)
-
Wrote a tiny new LLM plugin this morning: llm-whisper-api, which lets you do this (if you have an OpenAI API key configured already):
llm install llm-whisper-api
llm whisper-api myfile.mp3 > transcript.txtllm-whisper-api
I wanted to run an experiment through the [OpenAI Whisper API](https://platform.openai.com/docs/guides/speech-to-text) this morning so I knocked up a _very_ quick plugin for [LLM](https://llm.datasette.io/) that provides the following interface: llm install …
(simonwillison.net)
-
@simon Now we just also need a plugin for the groq whisper api, which can be used for free!
-
@ame I got Claude to port my Whisper API plugin to use Groq instead! It seems to work - I've only released it as an alpha though as I've not yet added automated tests or manually QAd all of the options https://github.com/simonw/llm-groq-whisper
-
@simon Amazing, that works for me!
Thank you!I think it's pretty easy to get an LLM to write code to have the plugin stitch the json response to an SRT file too
-
Simon Willisonreplied to Ame last edited by [email protected]
@ame I got Claude to make an artifact for that instead, looks like it might work OK https://gistpreview.github.io/?e29943852f371f638c9a3ae1dcc4784e
Claude transcript: https://gist.github.com/simonw/49b52ce2a7b5796edf4e0e2e2152db41
-
Simon Willisonreplied to Simon Willison last edited by
@ame let me know if it works, I'm not sure how best to test it!