Skip to content

Commit

Permalink
remove prints
Browse files Browse the repository at this point in the history
  • Loading branch information
ArthurZucker committed Jul 11, 2024
1 parent 7e83218 commit 1639ae2
Show file tree
Hide file tree
Showing 3 changed files with 0 additions and 3 deletions.
1 change: 0 additions & 1 deletion tokenizers/src/tokenizer/added_vocabulary.rs
Original file line number Diff line number Diff line change
Expand Up @@ -487,7 +487,6 @@ impl AddedVocabulary {
Ok(self.split_with_indices(sequence, &self.split_normalized_trie))
})
.expect("AddedVocabulary bad split");
println!("after pretokenized.split {:?}", pretokenized);
// ["I read a book", " <s>", "Hey"] -> ["▁I read a book", "▁ <s>", "▁Hey"]
// ["▁I read a book", "▁ <s>", "▁Hey"] -> [.., "▁ ", "<s>", "▁Hey"]

Expand Down
1 change: 0 additions & 1 deletion tokenizers/src/tokenizer/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -943,7 +943,6 @@ where
pretokenized: P,
) -> Result<PreTokenizedString> {
let mut pretokenized: PreTokenizedString = pretokenized.into();
println!("do pretokenized converted: {:?}", pretokenized);
if let Some(ref pretok) = self.pre_tokenizer {
pretok.pre_tokenize(&mut pretokenized)?;
}
Expand Down
1 change: 0 additions & 1 deletion tokenizers/src/tokenizer/pre_tokenizer.rs
Original file line number Diff line number Diff line change
Expand Up @@ -166,7 +166,6 @@ impl PreTokenizedString {
if let Some(converter) = offset_converter {
offsets = converter.convert(offsets).unwrap_or(offsets);
}
println!("into encode, {:?}, {:?}", normalized.clone(), token.value);
(
token.id,
token.value,
Expand Down

0 comments on commit 1639ae2

Please sign in to comment.