File tree 1 file changed +4
-0
lines changed
src/transformers/models/mt5
1 file changed +4
-0
lines changed Original file line number Diff line number Diff line change @@ -60,6 +60,8 @@ class MT5Config(PretrainedConfig):
60
60
testing).
61
61
feed_forward_proj (:obj:`string`, `optional`, defaults to :obj:`"gated-gelu"`):
62
62
Type of feed forward layer to be used. Should be one of :obj:`"relu"` or :obj:`"gated-gelu"`.
63
+ use_cache (:obj:`bool`, `optional`, defaults to :obj:`True`):
64
+ Whether or not the model should return the last key/values attentions (not used by all models).
63
65
"""
64
66
model_type = "mt5"
65
67
keys_to_ignore_at_inference = ["past_key_values" ]
@@ -79,6 +81,7 @@ def __init__(
79
81
initializer_factor = 1.0 ,
80
82
feed_forward_proj = "gated-gelu" ,
81
83
is_encoder_decoder = True ,
84
+ use_cache = True ,
82
85
tokenizer_class = "T5Tokenizer" ,
83
86
tie_word_embeddings = False ,
84
87
pad_token_id = 0 ,
@@ -109,6 +112,7 @@ def __init__(
109
112
self .layer_norm_epsilon = layer_norm_epsilon
110
113
self .initializer_factor = initializer_factor
111
114
self .feed_forward_proj = feed_forward_proj
115
+ self .use_cache = use_cache
112
116
113
117
@property
114
118
def hidden_size (self ):
You can’t perform that action at this time.
0 commit comments