Overview
Currently, we pass each chunk of transcript to Gemini via the Vertex API. This is fine but slow and expensive.
We should distil our own version of an equivalent PaLM2 model so we can host this on our own cloud instance.
This work should follow on from the equivalent issue in the GenAI semantic enricher project.
Requirements
Overview
Currently, we pass each chunk of transcript to Gemini via the Vertex API. This is fine but slow and expensive.
We should distil our own version of an equivalent PaLM2 model so we can host this on our own cloud instance.
This work should follow on from the equivalent issue in the GenAI semantic enricher project.
Requirements