Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[ENH] Added batch_size as parameter to SentenceTransformerEmbeddingFu… #2759

Closed
wants to merge 1 commit into from

Conversation

cowolff
Copy link

@cowolff cowolff commented Sep 3, 2024

…nction class

Description of changes

Added the hyperparameter batch_size to the SentenceTransformerEmbeddingFunction class in order to have more control over memory requirements when it comes to deploying large sentence transformers.

Test plan

  • Tests pass locally with pytest for python. Local installation and usage also worked.

Documentation Changes

Added docstring for the new hyperparameter

…nction class

## Description of changes

Added the hyperparameter batch_size to the SentenceTransformerEmbeddingFunction class in order to have more control over memory requirements when it comes to deploying large sentence transformer.

## Test plan
- [x] Tests pass locally with `pytest` for python. Local installation and usage also worked.

## Documentation Changes
Added docstring for the new hyperparameter
Copy link

github-actions bot commented Sep 3, 2024

Reviewer Checklist

Please leverage this checklist to ensure your code review is thorough before approving

Testing, Bugs, Errors, Logs, Documentation

  • Can you think of any use case in which the code does not behave as intended? Have they been tested?
  • Can you think of any inputs or external events that could break the code? Is user input validated and safe? Have they been tested?
  • If appropriate, are there adequate property based tests?
  • If appropriate, are there adequate unit tests?
  • Should any logging, debugging, tracing information be added or removed?
  • Are error messages user-friendly?
  • Have all documentation changes needed been made?
  • Have all non-obvious changes been commented?

System Compatibility

  • Are there any potential impacts on other parts of the system or backward compatibility?
  • Does this change intersect with any items on our roadmap, and if so, is there a plan for fitting them together?

Quality

  • Is this code of a unexpectedly high quality (Readability, Modularity, Intuitiveness)

Copy link
Contributor

@tazarov tazarov left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM with a minor nit.

@cowolff Thanks for this. Reading on the effects of batch_size (huggingface/transformers#2401), I feel this is a somewhat contentious topic, but having explicit is better DX (also Zen of Python)

@@ -17,6 +17,7 @@ def __init__(
model_name: str = "all-MiniLM-L6-v2",
device: str = "cpu",
normalize_embeddings: bool = False,
batch_size: int = 32,
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can we make this Optional[int] = 32

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Also, technically, this is already supported by the kwargs (not that we pass them to the encode method), but perhaps making it explicit is arguably a better DX.

@jeffchuber jeffchuber mentioned this pull request Sep 16, 2024
@jeffchuber
Copy link
Contributor

Our underlying impl has changed and so this PR is not landable as is.

That being said - we'd still like to add this functionality and that is now tracked in this issue.

@jeffchuber jeffchuber closed this Sep 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants