diff --git a/model_custom.py b/model_custom.py index 4e245e2e28abe67869477f075dda4c7e4da269c2..96cfc64035ba8c8c40d79b94b81246c05cc54005 100644 --- a/model_custom.py +++ b/model_custom.py @@ -101,7 +101,6 @@ class Model_Common_Transformer(nn.Module): d_model=self.embedding_dim) def forward(self, seq, charge): - print(seq.shape) meta_ohe = torch.nn.functional.one_hot(charge - 1, self.charge_max).float() seq_emb = torch.nn.functional.one_hot(seq, self.nb_aa).float() emb = self.pos_embedding(self.emb(seq_emb))