Skip to content

Commit

Permalink
Merge pull request #328 from markstur/wrapped_sentence_transformer
Browse files Browse the repository at this point in the history
Wrapped sentence transformer
  • Loading branch information
evaline-ju committed Mar 1, 2024
2 parents 2ab8abd + 1a9ad04 commit 8ca93c3
Show file tree
Hide file tree
Showing 5 changed files with 389 additions and 252 deletions.
16 changes: 16 additions & 0 deletions caikit_nlp/config/config.yml
Original file line number Diff line number Diff line change
Expand Up @@ -36,5 +36,21 @@ training_data_limit:
6655831b-960a-4dc5-8df4-867026e2cd41:
add_model_name_here: 10000

# Config used only in EmbeddingModule. Set here or use env vars like EMBEDDING_RETRIES=32
embedding:
# Number of times to retry on error. Most deployments should use 0 retries.
retries: 0
# Batch size for encode() if <= 0 or invalid, the sentence-transformers default is used
batch_size: 0
# Attempt to optimize with PyTorch compile()
pt2_compile: false
# Use IPEX optimize. Works best when used with autocast (bfloat16) below.
ipex: false
# Use autocast in encode with its default dtype (bfloat16)
autocast: false
# For testing, set device to "mps" on MacOS or "xpu" for IPEX GPU.
# Otherwise, the default does automatic checks for cuda GPU (else cpu).
device: ""

runtime:
library: caikit_nlp

0 comments on commit 8ca93c3

Please sign in to comment.