diff --git a/Models/EncoderModels.py b/Models/EncoderModels.py index fe665aa..0a2f6b7 100644 --- a/Models/EncoderModels.py +++ b/Models/EncoderModels.py @@ -234,7 +234,7 @@ def __init__(self,embedding_dim, the last dense layer, for example Relu or sigmoid, default is None. """ - super(Annotator,self).__init__() + super(AnnotatorGRU,self).__init__() self.return_attent_weights=return_attent_weights