You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
TEI currently does not seem to support infly's latest embedding model: infly/inf-retriever-v1-1.5b
When I try to load the model
I get this error:
Model description
root@dify:/opt/text-embeddings-inference# text-embeddings-router --model-id infly/inf-retriever-v1-1.5b --port 8080
2025-03-14T07:05:12.969694Z INFO text_embeddings_router: router/src/main.rs:175: Args { model_id: "inf**/-****--1.5b", revision: None, tokenization_workers: None, dtype: None, pooling: None, max_concurrent_requests: 512, max_batch_tokens: 16384, max_batch_requests: None, max_client_batch_size: 32, auto_truncate: false, default_prompt_name: None, default_prompt: None, hf_api_token: None, hostname: "0.0.0.0", port: 8080, uds_path: "/tmp/text-embeddings-inference-server", huggingface_hub_cache: None, payload_limit: 2000000, api_key: None, json_output: false, otlp_endpoint: None, otlp_service_name: "text-embeddings-inference.server", cors_allow_origin: None }
2025-03-14T07:05:12.970142Z INFO hf_hub: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/hf-hub-0.3.2/src/lib.rs:55: Token file not found "/root/.cache/huggingface/token"
2025-03-14T07:05:13.417820Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:20: Starting download
2025-03-14T07:05:13.417857Z INFO download_artifacts:download_pool_config: text_embeddings_core::download: core/src/download.rs:53: Downloading 1_Pooling/config.json
2025-03-14T07:05:13.417984Z INFO download_artifacts:download_new_st_config: text_embeddings_core::download: core/src/download.rs:77: Downloading config_sentence_transformers.json
2025-03-14T07:05:13.418028Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:40: Downloading config.json
2025-03-14T07:05:13.418067Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:43: Downloading tokenizer.json
2025-03-14T07:05:13.418109Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:47: Model artifacts downloaded in 294.183µs
2025-03-14T07:05:13.830632Z WARN tokenizers::tokenizer::serialization: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/tokenizers-0.19.1/src/tokenizer/serialization.rs:159: Warning: Token '<|endoftext|>' was expected to have ID '151643' but was given ID 'None'
2025-03-14T07:05:13.830686Z WARN tokenizers::tokenizer::serialization: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/tokenizers-0.19.1/src/tokenizer/serialization.rs:159: Warning: Token '<|im_start|>' was expected to have ID '151644' but was given ID 'None'
2025-03-14T07:05:13.830696Z WARN tokenizers::tokenizer::serialization: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/tokenizers-0.19.1/src/tokenizer/serialization.rs:159: Warning: Token '<|im_end|>' was expected to have ID '151645' but was given ID 'None'
2025-03-14T07:05:13.835713Z INFO text_embeddings_router: router/src/lib.rs:188: Maximum number of tokens per request: 32768
2025-03-14T07:05:13.836218Z INFO text_embeddings_core::tokenization: core/src/tokenization.rs:28: Starting 4 tokenization workers
2025-03-14T07:05:14.130077Z INFO text_embeddings_router: router/src/lib.rs:230: Starting model backend
2025-03-14T07:05:14.130145Z INFO text_embeddings_backend: backends/src/lib.rs:486: Downloading model.safetensors
2025-03-14T07:57:02.508011Z INFO text_embeddings_backend: backends/src/lib.rs:370: Model weights downloaded in 3108.370651683s
2025-03-14T07:57:02.532664Z ERROR text_embeddings_backend: backends/src/lib.rs:381: Could not start Candle backend: Could not start backend: Model is not supported
Caused by:
invalid type: null, expected usize
Error: Could not create backend
Caused by:
Could not start backend: Could not start a suitable backend
TEI currently does not seem to support infly's latest embedding model: infly/inf-retriever-v1-1.5b
When I try to load the model
I get this error:
Model description
root@dify:/opt/text-embeddings-inference# text-embeddings-router --model-id infly/inf-retriever-v1-1.5b --port 8080
2025-03-14T07:05:12.969694Z INFO text_embeddings_router: router/src/main.rs:175: Args { model_id: "inf**/-****--1.5b", revision: None, tokenization_workers: None, dtype: None, pooling: None, max_concurrent_requests: 512, max_batch_tokens: 16384, max_batch_requests: None, max_client_batch_size: 32, auto_truncate: false, default_prompt_name: None, default_prompt: None, hf_api_token: None, hostname: "0.0.0.0", port: 8080, uds_path: "/tmp/text-embeddings-inference-server", huggingface_hub_cache: None, payload_limit: 2000000, api_key: None, json_output: false, otlp_endpoint: None, otlp_service_name: "text-embeddings-inference.server", cors_allow_origin: None }
2025-03-14T07:05:12.970142Z INFO hf_hub: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/hf-hub-0.3.2/src/lib.rs:55: Token file not found "/root/.cache/huggingface/token"
2025-03-14T07:05:13.417820Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:20: Starting download
2025-03-14T07:05:13.417857Z INFO download_artifacts:download_pool_config: text_embeddings_core::download: core/src/download.rs:53: Downloading
1_Pooling/config.json
2025-03-14T07:05:13.417984Z INFO download_artifacts:download_new_st_config: text_embeddings_core::download: core/src/download.rs:77: Downloading
config_sentence_transformers.json
2025-03-14T07:05:13.418028Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:40: Downloading
config.json
2025-03-14T07:05:13.418067Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:43: Downloading
tokenizer.json
2025-03-14T07:05:13.418109Z INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:47: Model artifacts downloaded in 294.183µs
2025-03-14T07:05:13.830632Z WARN tokenizers::tokenizer::serialization: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/tokenizers-0.19.1/src/tokenizer/serialization.rs:159: Warning: Token '<|endoftext|>' was expected to have ID '151643' but was given ID 'None'
2025-03-14T07:05:13.830686Z WARN tokenizers::tokenizer::serialization: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/tokenizers-0.19.1/src/tokenizer/serialization.rs:159: Warning: Token '<|im_start|>' was expected to have ID '151644' but was given ID 'None'
2025-03-14T07:05:13.830696Z WARN tokenizers::tokenizer::serialization: /root/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/tokenizers-0.19.1/src/tokenizer/serialization.rs:159: Warning: Token '<|im_end|>' was expected to have ID '151645' but was given ID 'None'
2025-03-14T07:05:13.835713Z INFO text_embeddings_router: router/src/lib.rs:188: Maximum number of tokens per request: 32768
2025-03-14T07:05:13.836218Z INFO text_embeddings_core::tokenization: core/src/tokenization.rs:28: Starting 4 tokenization workers
2025-03-14T07:05:14.130077Z INFO text_embeddings_router: router/src/lib.rs:230: Starting model backend
2025-03-14T07:05:14.130145Z INFO text_embeddings_backend: backends/src/lib.rs:486: Downloading
model.safetensors
2025-03-14T07:57:02.508011Z INFO text_embeddings_backend: backends/src/lib.rs:370: Model weights downloaded in 3108.370651683s
2025-03-14T07:57:02.532664Z ERROR text_embeddings_backend: backends/src/lib.rs:381: Could not start Candle backend: Could not start backend: Model is not supported
Caused by:
invalid type: null, expected usize
Error: Could not create backend
Caused by:
Could not start backend: Could not start a suitable backend
Open source status
Provide useful links for the implementation
https://huggingface.co/infly/inf-retriever-v1-1.5b
The text was updated successfully, but these errors were encountered: