Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for infly/inf-retriever-v1-1.5b #514

Open
2 tasks done
Hirro029 opened this issue Mar 14, 2025 · 0 comments
Open
2 tasks done

Support for infly/inf-retriever-v1-1.5b #514

Hirro029 opened this issue Mar 14, 2025 · 0 comments

Comments

@Hirro029
Copy link

Hirro029 commented Mar 14, 2025

TEI currently does not seem to support infly's latest embedding model: infly/inf-retriever-v1-1.5b
When I try to load the model
I get this error:

Model description

root@dify:/opt/text-embeddings-inference# text-embeddings-router --model-id infly/inf-retriever-v1-1.5b --port 8080
2025-03-14T07:05:12.969694Z INFO text_embeddings_router: router/src/main.rs:175: Args { model_id: "inf**/-****--1.5b", revision: None, tokenization_workers: None, dtype: None, pooling: None, max_concurrent_requests: 512, max_batch_tokens: 16384, max_batch_requests: None, max_client_batch_size: 32, auto_truncate: false, default_prompt_name: None, default_prompt: None, hf_api_token: None, hostname: "0.0.0.0", port: 8080, uds_path: "/tmp/text-embeddings-inference-server", huggingface_hub_cache: None, payload_limit: 2000000, api_key: None, json_output: false, otlp_endpoint: None, otlp_service_name: "text-embeddings-inference.server", cors_allow_origin: None }
2025-03-14T07:05:12.970142Z INFO hf_hub: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/hf-hub-0.3.2/src/lib.rs:55: Token file not found "/root/.cache/huggingface/token"
2025-03-14T07:05:13.417820Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:20: Starting download
2025-03-14T07:05:13.417857Z INFO download_artifacts:download_pool_config: text_embeddings_core::download: core/src/download.rs:53: Downloading 1_Pooling/config.json
2025-03-14T07:05:13.417984Z INFO download_artifacts:download_new_st_config: text_embeddings_core::download: core/src/download.rs:77: Downloading config_sentence_transformers.json
2025-03-14T07:05:13.418028Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:40: Downloading config.json
2025-03-14T07:05:13.418067Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:43: Downloading tokenizer.json
2025-03-14T07:05:13.418109Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:47: Model artifacts downloaded in 294.183µs
2025-03-14T07:05:13.830632Z WARN tokenizers::tokenizer::serialization: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/tokenizers-0.19.1/src/tokenizer/serialization.rs:159: Warning: Token '<|endoftext|>' was expected to have ID '151643' but was given ID 'None'
2025-03-14T07:05:13.830686Z WARN tokenizers::tokenizer::serialization: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/tokenizers-0.19.1/src/tokenizer/serialization.rs:159: Warning: Token '<|im_start|>' was expected to have ID '151644' but was given ID 'None'
2025-03-14T07:05:13.830696Z WARN tokenizers::tokenizer::serialization: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/tokenizers-0.19.1/src/tokenizer/serialization.rs:159: Warning: Token '<|im_end|>' was expected to have ID '151645' but was given ID 'None'
2025-03-14T07:05:13.835713Z INFO text_embeddings_router: router/src/lib.rs:188: Maximum number of tokens per request: 32768
2025-03-14T07:05:13.836218Z INFO text_embeddings_core::tokenization: core/src/tokenization.rs:28: Starting 4 tokenization workers
2025-03-14T07:05:14.130077Z INFO text_embeddings_router: router/src/lib.rs:230: Starting model backend
2025-03-14T07:05:14.130145Z INFO text_embeddings_backend: backends/src/lib.rs:486: Downloading model.safetensors
2025-03-14T07:57:02.508011Z INFO text_embeddings_backend: backends/src/lib.rs:370: Model weights downloaded in 3108.370651683s
2025-03-14T07:57:02.532664Z ERROR text_embeddings_backend: backends/src/lib.rs:381: Could not start Candle backend: Could not start backend: Model is not supported

Caused by:
invalid type: null, expected usize
Error: Could not create backend

Caused by:
Could not start backend: Could not start a suitable backend

Open source status

  • The model implementation is available
  • The model weights are available

Provide useful links for the implementation

https://huggingface.co/infly/inf-retriever-v1-1.5b

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant