Skip to content

How to use AzureOpenAITextEmbeddingGenerationService.UpsertBatchAsync #9846

Discussion options

You must be logged in to vote

@ordinaryAndConfident when considering a chunking strategy it's important to strike a balance on size. If you create chunks that are too small they may be less likely to be found when doing vector searches. If the chunks are too big, they can increase token usage when passed to the LLM for context. The right size will also depend on the type of data that you want to generate embeddings for, so experimenting with different sizes is important.

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@ordinaryAndConfident
Comment options

Answer selected by sophialagerkranspandey
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants