Skip to content

Using LocalAI embedding, allow for control of batch size with models. #283

@knot-split

Description

@knot-split

Describe the feature you'd like to request

Perhaps an environment variable that either sets a batch size for the embedding or else defaults to the model's config for batch size.

Describe the solution you'd like

We have tried setting a dozen different places for adjusting the physical_batch_size in various configs, either in the ccb data or else in localai or even in our model yaml files, but context chat always seems to hit the same limit no matter what we do: batch size: 512
Describe alternatives you've considered

Perhaps not using localai embedding and instead the built-in embedding from ccb doesn't have this problem, but I'm not just yet ready to try it, since it took a month for me to get the ccb+local ai on one machine and nc aio on the other.

Metadata

Metadata

Assignees

No one assigned

    Type

    No type
    No fields configured for issues without a type.

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions