You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Found that the output token sequence cannot match exactly between llama-tokenize and AutoTokenizer for models like Meta-Llama-3-8B-Instruct, internlm2_5-7b-chat.
Name and Version
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 8 CUDA devices:
Device 0: NVIDIA A100-SXM4-80GB, compute capability 8.0, VMM: yes
Device 1: NVIDIA A100-SXM4-80GB, compute capability 8.0, VMM: yes
Device 2: NVIDIA A100-SXM4-80GB, compute capability 8.0, VMM: yes
Device 3: NVIDIA A100-SXM4-80GB, compute capability 8.0, VMM: yes
Device 4: NVIDIA A100-SXM4-80GB, compute capability 8.0, VMM: yes
Device 5: NVIDIA A100-SXM4-80GB, compute capability 8.0, VMM: yes
Device 6: NVIDIA A100-SXM4-80GB, compute capability 8.0, VMM: yes
Device 7: NVIDIA A100-SXM4-80GB, compute capability 8.0, VMM: yes
version: 4354 (0e70ba6)
built with cc (GCC) 9.3.1 20200408 (Red Hat 9.3.1-2) for x86_64-redhat-linux
Operating systems
Linux
GGML backends
CUDA
Hardware
NVIDIA A100-SXM4-80GB
Models
Meta-Llama-3-8B-Instruct
Problem description & steps to reproduce
Found that the output token sequence cannot match exactly between
llama-tokenize
andAutoTokenizer
for models likeMeta-Llama-3-8B-Instruct
,internlm2_5-7b-chat
.reproduce
llama-tokenize
results
Meta-Llama-3-8B-Instruct
internlm2_5-7b-chat
First Bad Commit
No response
Relevant log output
The text was updated successfully, but these errors were encountered: