Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

cannot find tensor cls.predictions.decoder.weight #548

Open
2 of 4 tasks
jetnet opened this issue Mar 29, 2025 · 0 comments
Open
2 of 4 tasks

cannot find tensor cls.predictions.decoder.weight #548

jetnet opened this issue Mar 29, 2025 · 0 comments

Comments

@jetnet
Copy link

jetnet commented Mar 29, 2025

System Info

System Info:

# Ubuntu 24.04.2 LTS
# podman version 4.9.3

podman run --device nvidia.com/gpu=all --rm --gpus all nvidia/cuda:12.8.1-base-ubuntu22.04 nvidia-smi -L
GPU 0: NVIDIA L40 (UUID: GPU-xxx)
GPU 1: NVIDIA L40 (UUID: GPU-xxx)

Information

  • Docker
  • The CLI directly

Tasks

  • An officially supported command
  • My own modifications

Reproduction

Trying to start the model:

image="ghcr.io/huggingface/text-embeddings-inference:89-1.6"

model=google-bert/bert-base-german-dbmdz-uncased
name=german-dbmdz-uncased

podman run -d \
  --device nvidia.com/gpu=all \
  -e CUDA_VISIBLE_DEVICES=0,1 \
  -v $modeldir:/data \
  -p 18080:80 \
  --name $name $image \
  --model-id $model --pooling splade

Error:

ERROR text_embeddings_backend: backends/src/lib.rs:388:
Could not start Candle backend:
Could not start backend: cannot find tensor cls.predictions.decoder.weight

Expected behavior

container up and running

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant