Skip to content

Commit

Permalink
disable templates
Browse files Browse the repository at this point in the history
  • Loading branch information
ardagoreci committed Aug 13, 2024
1 parent bbdb23d commit 41ad3df
Show file tree
Hide file tree
Showing 3 changed files with 18 additions and 18 deletions.
4 changes: 2 additions & 2 deletions configs/data/erebor.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,7 @@ config:
data_module:
use_small_bfd: false
data_loaders:
batch_size: 1
batch_size: 2
num_workers: 16
pin_memory: true

Expand Down Expand Up @@ -220,7 +220,7 @@ config:
- "between_segment_residues"
- "deletion_matrix"
- "no_recycling_iters"
use_templates: true
use_templates: false
use_template_torsion_angles: false
template_features:
- "template_all_atom_positions"
Expand Down
4 changes: 2 additions & 2 deletions configs/model/alphafold3.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -114,7 +114,7 @@ model:
pairformer_stack:
c_s: ${model.model.c_token}
c_z: ${model.model.c_pair}
no_blocks: 48
no_blocks: 24 # 48
c_hidden_mul: ${model.model.c_hidden_tri_mul}
c_hidden_pair_attn: ${model.model.c_hidden_pair_attn}
no_heads_tri_attn: ${model.model.no_heads_tri_attn}
Expand All @@ -139,7 +139,7 @@ model:
atom_attention_n_keys: 128
atom_decoder_blocks: 3
atom_decoder_heads: 16
token_transformer_blocks: 24 # 24
token_transformer_blocks: 12 # 24
token_transformer_heads: 16
sd_data: 16.0
s_max: 160.0
Expand Down
28 changes: 14 additions & 14 deletions src/models/model.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,9 +22,9 @@ def __init__(self, config):
**self.config["input_embedder"]
)

self.template_embedder = TemplateEmbedder(
**self.config["template_embedder"]
)
# self.template_embedder = TemplateEmbedder(
# **self.config["template_embedder"]
# )

self.msa_module = MSAModule(
**self.config["msa_module"]
Expand Down Expand Up @@ -107,17 +107,17 @@ def run_trunk(
del s_prev, z_prev, s_init, z_init

# Embed the templates
z = add(z,
self.template_embedder(
feats,
z,
pair_mask=pair_mask,
chunk_size=self.globals.chunk_size,
use_deepspeed_evo_attention=self.globals.use_deepspeed_evo_attention,
inplace_safe=inplace_safe
),
inplace=inplace_safe
)
# z = add(z,
# self.template_embedder(
# feats,
# z,
# pair_mask=pair_mask,
# chunk_size=self.globals.chunk_size,
# use_deepspeed_evo_attention=self.globals.use_deepspeed_evo_attention,
# inplace_safe=inplace_safe
# ),
# inplace=inplace_safe
# )

# Process the MSA
z = add(z,
Expand Down

0 comments on commit 41ad3df

Please sign in to comment.