-
Notifications
You must be signed in to change notification settings - Fork 119
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FEAT] Support RecursiveRules (in reverse) for OverlapRefiner #150
Comments
Hey @Sankgreall! 😊 Firstly, thank you for creating the issue; I am positively amazed at the due diligence put in this issue and the blog post as well~ This is awesome! I'd like to boost this blog from Chonkie's main channels as well, if you don't mind 🫶🚀 Some comments I had on this enhancement are as follows (in no particular order):
Happy to discuss more on this over on our Discord channel! Hit me up, @bhavnicksm and we can chat on one of the Voice Channels as well~ Thanks! 😊 |
Thanks for the reply, some excellent points, and would be greatly honoured if you wanted to circulate the blog! With regards to latency, I just ran some tests. My existing transcript was about 15,000 tokens and I was making chunks with a token limit of 600. I increased the transcript size by copying and pasting it many times, and then reran the operation with some timers. Here are the results (just for the refining). On my device at least, I don't notice a significant issue.
On your last bullet, as we've chatted about now, the goal with this refiner is to return the minimum prefix size necessary to establish speaker context for the LLM. The way I've implemented this is through leveraging the This is because we want to provide the LLM with maximum useful information (i.e., who is speaking), whilst being aware that more tokens increases the chance of model decoherency. The implications for this approach also go well beyond transcripts. My grander plans for this would be for the refiner to be capable of scrolling past I've been working on a more complete class today. Would be happy to raise a PR? |
Hey @Sankgreall! 😄 PRs welcome 🤗 It would be great to merge the HeirarchialOverlapRefinery functionality with the current OverlapRefinery and add an additional signature parameter Let me know if you have any issues or discussion points; happy to connect! Thanks |
📋 Quick Check
💡 Feature Description
OverlapRefinery
is limited to taking a fixed token/word prefix, but would be much more powerful if we could apply recursive rules with a minimum token length in reverse.🛠️ Implementation Approach
Commentary on the issue here, and a sample implementation: https://sankgreall.medium.com/intelligent-transcript-chunking-63bfca62a5ad. Pasting the relevant code below as well.
🎯 Why is this needed?
Extends the capability of
OverlapRefinery
so that it can consider document/chunk structure when making prefix decisions.Will do my best to submit a PR in the future, but raising as a feature request in case there is any comment or feedback.
The text was updated successfully, but these errors were encountered: