-
Notifications
You must be signed in to change notification settings - Fork 2
Expand file tree
/
Copy pathtei_timeout.log
More file actions
25 lines (25 loc) · 6.11 KB
/
tei_timeout.log
File metadata and controls
25 lines (25 loc) · 6.11 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
[2m2025-05-07T19:34:32.481255Z[0m [32m INFO[0m [2mtext_embeddings_router[0m[2m:[0m [2mrouter/src/main.rs[0m[2m:[0m[2m189:[0m Args { model_id: "Ali****-***/***-*****-*.**-*****uct", revision: None, tokenization_workers: None, dtype: None, pooling: None, max_concurrent_requests: 512, max_batch_tokens: 16384, max_batch_requests: None, max_client_batch_size: 32, auto_truncate: true, default_prompt_name: None, default_prompt: None, hf_api_token: None, hf_token: None, hostname: "0.0.0.0", port: 19998, uds_path: "/tmp/text-embeddings-inference-server", huggingface_hub_cache: None, payload_limit: 2000000, api_key: None, json_output: false, disable_spans: false, otlp_endpoint: None, otlp_service_name: "text-embeddings-inference.server", prometheus_port: 9000, cors_allow_origin: None }
[2m2025-05-07T19:34:32.486659Z[0m [32m INFO[0m [1mdownload_artifacts[0m[2m:[0m [2mtext_embeddings_core::download[0m[2m:[0m [2mcore/src/download.rs[0m[2m:[0m[2m20:[0m Starting download
[2m2025-05-07T19:34:32.486675Z[0m [32m INFO[0m [1mdownload_artifacts[0m[2m:[0m[1mdownload_pool_config[0m[2m:[0m [2mtext_embeddings_core::download[0m[2m:[0m [2mcore/src/download.rs[0m[2m:[0m[2m53:[0m Downloading `1_Pooling/config.json`
[2m2025-05-07T19:34:32.487341Z[0m [32m INFO[0m [1mdownload_artifacts[0m[2m:[0m[1mdownload_new_st_config[0m[2m:[0m [2mtext_embeddings_core::download[0m[2m:[0m [2mcore/src/download.rs[0m[2m:[0m[2m77:[0m Downloading `config_sentence_transformers.json`
[2m2025-05-07T19:34:32.487367Z[0m [32m INFO[0m [1mdownload_artifacts[0m[2m:[0m [2mtext_embeddings_core::download[0m[2m:[0m [2mcore/src/download.rs[0m[2m:[0m[2m40:[0m Downloading `config.json`
[2m2025-05-07T19:34:32.487388Z[0m [32m INFO[0m [1mdownload_artifacts[0m[2m:[0m [2mtext_embeddings_core::download[0m[2m:[0m [2mcore/src/download.rs[0m[2m:[0m[2m43:[0m Downloading `tokenizer.json`
[2m2025-05-07T19:34:32.487520Z[0m [32m INFO[0m [1mdownload_artifacts[0m[2m:[0m [2mtext_embeddings_core::download[0m[2m:[0m [2mcore/src/download.rs[0m[2m:[0m[2m47:[0m Model artifacts downloaded in 861.667µs
[2m2025-05-07T19:34:32.585305Z[0m [32m INFO[0m [2mtext_embeddings_router[0m[2m:[0m [2mrouter/src/lib.rs[0m[2m:[0m[2m193:[0m Maximum number of tokens per request: 32768
[2m2025-05-07T19:34:32.585400Z[0m [32m INFO[0m [2mtext_embeddings_core::tokenization[0m[2m:[0m [2mcore/src/tokenization.rs[0m[2m:[0m[2m38:[0m Starting 14 tokenization workers
[2m2025-05-07T19:34:32.671605Z[0m [32m INFO[0m [2mtext_embeddings_router[0m[2m:[0m [2mrouter/src/lib.rs[0m[2m:[0m[2m235:[0m Starting model backend
[2m2025-05-07T19:34:32.671903Z[0m [32m INFO[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m534:[0m Downloading `model.onnx`
[2m2025-05-07T19:34:33.079916Z[0m [33m WARN[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m538:[0m Could not download `model.onnx`: request error: HTTP status client error (404 Not Found) for url (https://huggingface.co/Alibaba-NLP/gte-Qwen2-1.5B-instruct/resolve/main/model.onnx)
[2m2025-05-07T19:34:33.079938Z[0m [32m INFO[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m539:[0m Downloading `onnx/model.onnx`
[2m2025-05-07T19:34:33.489241Z[0m [33m WARN[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m543:[0m Could not download `onnx/model.onnx`: request error: HTTP status client error (404 Not Found) for url (https://huggingface.co/Alibaba-NLP/gte-Qwen2-1.5B-instruct/resolve/main/onnx/model.onnx)
[2m2025-05-07T19:34:33.489254Z[0m [32m INFO[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m548:[0m Downloading `model.onnx_data`
[2m2025-05-07T19:34:33.753610Z[0m [33m WARN[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m552:[0m Could not download `model.onnx_data`: request error: HTTP status client error (404 Not Found) for url (https://huggingface.co/Alibaba-NLP/gte-Qwen2-1.5B-instruct/resolve/main/model.onnx_data)
[2m2025-05-07T19:34:33.753637Z[0m [32m INFO[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m553:[0m Downloading `onnx/model.onnx_data`
[2m2025-05-07T19:34:34.513003Z[0m [33m WARN[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m557:[0m Could not download `onnx/model.onnx_data`: request error: HTTP status client error (404 Not Found) for url (https://huggingface.co/Alibaba-NLP/gte-Qwen2-1.5B-instruct/resolve/main/onnx/model.onnx_data)
[2m2025-05-07T19:34:34.513013Z[0m [31mERROR[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m346:[0m Model ONNX files not found in the repository
[2m2025-05-07T19:34:34.513294Z[0m [31mERROR[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m358:[0m Could not start ORT backend: Could not start backend: Pooling last_token is not supported for this backend. Use `candle` backend instead.
[2m2025-05-07T19:34:34.513515Z[0m [32m INFO[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m493:[0m Downloading `model.safetensors`
[2m2025-05-07T19:34:35.332343Z[0m [33m WARN[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m496:[0m Could not download `model.safetensors`: request error: HTTP status client error (404 Not Found) for url (https://huggingface.co/Alibaba-NLP/gte-Qwen2-1.5B-instruct/resolve/main/model.safetensors)
[2m2025-05-07T19:34:35.332358Z[0m [32m INFO[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m501:[0m Downloading `model.safetensors.index.json`
[2m2025-05-07T19:34:35.333710Z[0m [32m INFO[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m523:[0m Downloading `model-00002-of-00002.safetensors`
[2m2025-05-07T19:36:48.542283Z[0m [32m INFO[0m [2mtext_embeddings_backend[0m[2m:[0m [2mbackends/src/lib.rs[0m[2m:[0m[2m523:[0m Downloading `model-00001-of-00002.safetensors`