ModelInputΒΆ
-
class
pytext.models.pair_classification_model.
ModelInput
Bases:
ModelInput
All Attributes (including base classes)
- tokens1: TokenTensorizer.Config = TokenTensorizer.Config(column=
'text1'
)- tokens2: TokenTensorizer.Config = TokenTensorizer.Config(column=
'text2'
)- labels: LabelTensorizer.Config = LabelTensorizer.Config()
Default JSON
{
"tokens1": {
"is_input": true,
"column": "text1",
"tokenizer": {
"Tokenizer": {
"split_regex": "\\s+",
"lowercase": true,
"use_byte_offsets": false
}
},
"add_bos_token": false,
"add_eos_token": false,
"use_eos_token_for_bos": false,
"max_seq_len": null,
"vocab": {
"build_from_data": true,
"size_from_data": 0,
"min_counts": 0,
"vocab_files": []
},
"vocab_file_delimiter": " "
},
"tokens2": {
"is_input": true,
"column": "text2",
"tokenizer": {
"Tokenizer": {
"split_regex": "\\s+",
"lowercase": true,
"use_byte_offsets": false
}
},
"add_bos_token": false,
"add_eos_token": false,
"use_eos_token_for_bos": false,
"max_seq_len": null,
"vocab": {
"build_from_data": true,
"size_from_data": 0,
"min_counts": 0,
"vocab_files": []
},
"vocab_file_delimiter": " "
},
"labels": {
"LabelTensorizer": {
"is_input": false,
"column": "label",
"allow_unknown": false,
"pad_in_vocab": false,
"label_vocab": null,
"label_vocab_file": null,
"add_labels": null
}
}
}