isoformer-anonymous commited on
Commit
f529e8a
1 Parent(s): 0b3aec7

Update modeling_isoformer.py

Browse files
Files changed (1) hide show
  1. modeling_isoformer.py +2 -24
modeling_isoformer.py CHANGED
@@ -54,30 +54,8 @@ class Isoformer(PreTrainedModel):
54
  )
55
  self.config = config
56
 
57
- # self.enformer_config = EnformerConfig(
58
- # dim=config.enformer_dim,
59
- # depth=config.enformer_depth,
60
- # heads=config.enformer_heads,
61
- # output_heads=dict(
62
- # human=1,
63
- # mouse=1 # TODO CHANGE
64
- # ),
65
- # target_length=config.enformer_target_length, # 896,
66
- # attn_dim_key=config.enformer_attn_dim_key,
67
- # dropout_rate=0.4,
68
- # attn_dropout=0.05,
69
- # pos_dropout=0.01,
70
- # use_checkpointing=config.enformer_use_checkpointing,
71
- # use_convnext=config.enformer_use_convnext,
72
- # num_downsamples=config.enformer_num_downsamples,
73
- # # genetic sequence is downsampled 2 ** 7 == 128x in default Enformer - can be changed for higher resolution
74
- # dim_divisible_by=config.enformer_dim_divisible_by,
75
- # use_tf_gamma=False,
76
- # )
77
-
78
- self.esm_model = EsmForMaskedLM(self.esm_config) # protein encoder
79
- self.nt_model = NTForMaskedLM(self.nt_config) # rna encoder
80
- #self.enformer_model = Enformer(self.enformer_config) # dna encoder
81
  self.enformer_model = Enformer.from_pretrained("EleutherAI/enformer-official-rough")
82
 
83
  self.cross_attention_layer_rna = MultiHeadAttention(
 
54
  )
55
  self.config = config
56
 
57
+ self.esm_model = EsmForMaskedLM(self.esm_config)
58
+ self.nt_model = NTForMaskedLM(self.nt_config)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
59
  self.enformer_model = Enformer.from_pretrained("EleutherAI/enformer-official-rough")
60
 
61
  self.cross_attention_layer_rna = MultiHeadAttention(