From ad33c6053050dc07acc6897f80fa89b2e4745438 Mon Sep 17 00:00:00 2001
From: lschneider <leo.schneider@univ-lyon1.fr>
Date: Thu, 19 Sep 2024 15:15:15 +0200
Subject: [PATCH] test cossim

---
 model_custom.py | 3 +--
 1 file changed, 1 insertion(+), 2 deletions(-)

diff --git a/model_custom.py b/model_custom.py
index 32fe630..a7947f5 100644
--- a/model_custom.py
+++ b/model_custom.py
@@ -25,7 +25,6 @@ class PositionalEncoding(nn.Module):
         pe = torch.zeros(max_len, 1, d_model)
         pe[:, 0, 0::2] = torch.sin(position * div_term)
         pe[:, 0, 1::2] = torch.cos(position * div_term)
-        print(d_model, max_len)
         self.register_buffer('pe', pe)
 
     def forward(self, x):
@@ -106,7 +105,7 @@ class Model_Common_Transformer(nn.Module):
     def forward(self, seq, charge):
         meta_ohe = torch.nn.functional.one_hot(charge - 1, self.charge_max).float()
         seq_emb = torch.nn.functional.one_hot(seq, self.nb_aa).float()
-        print(self.emb(seq_emb).size())
+        print(seq_emb.size())
         emb = self.pos_embedding(self.emb(seq_emb))
         meta_enc = self.meta_enc(meta_ohe)
 
-- 
GitLab