You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The tutorial doesn't work end to end and the container keep on crashing with "The node was low on resource: ephemeral-storage. Threshold quantity: 10120387530, available: 9166736Ki. Container llm was using 57900024Ki, request is 0, has larger consumption of ephemeral-storage"
Hi here @raushan2016 thanks for flagging, indeed that's because the path to be mounted should be /tmp, I realised I didn't update that when updating the samples but just updated the container URI, and for the TGI DLCs we're mounting /tmp as the HF_HOME i.e. where the model weights are downloaded, so on, we should be mounting /tmp instead of /data, I'll create a PR to update the mounts. Sorry for the inconvenience if any!
ghcr.io/huggingface/text-generation-inference:2.0.4
result in to successful run. Below is the disk usage for the same.NOTE: There might be other sample also impacted with the above change. Since we don't have any automated gates for the validation.
The text was updated successfully, but these errors were encountered: