summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--sci-libs/tokenizers/tokenizers-0.20.1.ebuild23
1 files changed, 13 insertions, 10 deletions
diff --git a/sci-libs/tokenizers/tokenizers-0.20.1.ebuild b/sci-libs/tokenizers/tokenizers-0.20.1.ebuild
index f2172b1cfd5a..f5e839834cd8 100644
--- a/sci-libs/tokenizers/tokenizers-0.20.1.ebuild
+++ b/sci-libs/tokenizers/tokenizers-0.20.1.ebuild
@@ -325,7 +325,6 @@ src_test() {
cd ../bindings/python
local EPYTEST_DESELECT=(
"tests/bindings/test_tokenizer.py::TestTokenizer::test_encode_formats"
- "tests/documentation/test_tutorial_train_from_iterators.py::TestTrainFromIterators::test_datasets"
"tests/bindings/test_encoding.py::TestEncoding::test_sequence_ids"
"tests/bindings/test_encoding.py::TestEncoding::test_n_sequences"
"tests/bindings/test_encoding.py::TestEncoding::test_word_to_tokens"
@@ -339,28 +338,32 @@ src_test() {
"tests/bindings/test_encoding.py::TestEncoding::test_invalid_truncate_direction"
"tests/bindings/test_models.py::TestBPE::test_instantiate"
"tests/bindings/test_models.py::TestWordLevel::test_instantiate"
+ "tests/bindings/test_models.py::TestWordPiece::test_instantiate"
"tests/bindings/test_processors.py::TestByteLevelProcessing::test_processing"
- "tests/bindings/test_trainers.py::TestUnigram::test_train"
- "tests/documentation/test_pipeline.py::TestPipeline::test_pipeline"
- "tests/documentation/test_pipeline.py::TestPipeline::test_bert_example"
- "tests/implementations/test_char_bpe.py::TestCharBPETokenizer::test_basic_encode"
- "tests/implementations/test_char_bpe.py::TestCharBPETokenizer::test_lowercase"
- "tests/implementations/test_char_bpe.py::TestCharBPETokenizer::test_decoding"
- "tests/implementations/test_char_bpe.py::TestCharBPETokenizer::test_multiprocessing_with_parallelism"
- "tests/test_serialization.py::TestSerialization::test_full_serialization_albert"
- "tests/test_serialization.py::TestSerialization::test_str_big"
"tests/bindings/test_tokenizer.py::TestTokenizer::test_encode_add_special_tokens"
"tests/bindings/test_tokenizer.py::TestTokenizer::test_from_pretrained"
"tests/bindings/test_tokenizer.py::TestTokenizer::test_from_pretrained_revision"
"tests/bindings/test_tokenizer.py::TestTokenizer::test_encode_special_tokens"
"tests/bindings/test_tokenizer.py::TestTokenizer::test_splitting"
+ "tests/bindings/test_trainers.py::TestUnigram::test_continuing_prefix_trainer_mistmatch"
+ "tests/bindings/test_trainers.py::TestUnigram::test_train"
+ "tests/documentation/test_pipeline.py::TestPipeline::test_pipeline"
+ "tests/documentation/test_pipeline.py::TestPipeline::test_bert_example"
"tests/documentation/test_quicktour.py::TestQuicktour::test_quicktour"
+ "tests/documentation/test_tutorial_train_from_iterators.py::TestTrainFromIterators::test_datasets"
+ "tests/documentation/test_tutorial_train_from_iterators.py::TestTrainFromIterators::test_gzip"
"tests/implementations/test_bert_wordpiece.py::TestBertWordPieceTokenizer::test_basic_encode"
"tests/implementations/test_bert_wordpiece.py::TestBertWordPieceTokenizer::test_multiprocessing_with_parallelism"
"tests/implementations/test_byte_level_bpe.py::TestByteLevelBPE::test_basic_encode"
"tests/implementations/test_byte_level_bpe.py::TestByteLevelBPE::test_add_prefix_space"
"tests/implementations/test_byte_level_bpe.py::TestByteLevelBPE::test_lowerspace"
"tests/implementations/test_byte_level_bpe.py::TestByteLevelBPE::test_multiprocessing_with_parallelism"
+ "tests/implementations/test_char_bpe.py::TestCharBPETokenizer::test_basic_encode"
+ "tests/implementations/test_char_bpe.py::TestCharBPETokenizer::test_lowercase"
+ "tests/implementations/test_char_bpe.py::TestCharBPETokenizer::test_decoding"
+ "tests/implementations/test_char_bpe.py::TestCharBPETokenizer::test_multiprocessing_with_parallelism"
+ "tests/test_serialization.py::TestSerialization::test_full_serialization_albert"
+ "tests/test_serialization.py::TestSerialization::test_str_big"
)
local -x EPYTEST_IGNORE=(benches/)
distutils-r1_src_test