Skip to content

Commit

Permalink
feat: support classification
Browse files Browse the repository at this point in the history
feat: support classification
  • Loading branch information
n0w0f authored Nov 26, 2024
2 parents 3e0e8de + 2d5d901 commit 866a0e2
Show file tree
Hide file tree
Showing 69 changed files with 2,100 additions and 362 deletions.
33 changes: 33 additions & 0 deletions conf/bandgap.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,33 @@


hydra:
job:
name: bandgap
run:
dir: ${hydra:runtime.cwd}/outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.override_dirname}

# launcher:
# _target_: hydra_plugins.hydra_submitit_launcher.submitit_launcher.SlurmLauncher
# submitit_folder: ${hydra.sweep.dir}/.submitit/%j
# timeout_min: 3600
# mem_gb: 160
# nodes: 1
# #gpus_per_task: 1
# gres: gpu:1
# #gpus_per_node: 2
# name: ${hydra.job.name}
# partition: 'gpu'
# additional_parameters:
# nodelist: 'gpu[008,013-017]'
# tasks_per_node: 1

defaults:
- model: none
# - override hydra/launcher: submitit_slurm

runs:
- name: benchmark_run
tasks: [benchmark]
53 changes: 31 additions & 22 deletions conf/benchmark.yaml
Original file line number Diff line number Diff line change
@@ -1,24 +1,33 @@
hydra:
job:
name: benchmark
run:
dir: ${hydra:runtime.cwd}/outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.override_dirname}



defaults:
- model: none



runs:


- name: benchmark_run
tasks: [benchmark]



hydra:
job:
name: benchmark
run:
dir: ${hydra:runtime.cwd}/outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.override_dirname}

# launcher:
# _target_: hydra_plugins.hydra_submitit_launcher.submitit_launcher.SlurmLauncher
# submitit_folder: ${hydra.sweep.dir}/.submitit/%j
# timeout_min: 3600
# mem_gb: 160
# nodes: 1
# #gpus_per_task: 1
# gres: gpu:1
# #gpus_per_node: 2
# name: ${hydra.job.name}
# partition: 'gpu'
# additional_parameters:
# nodelist: 'gpu[008,013-017]'
# tasks_per_node: 1

defaults:
- model: none
# - override hydra/launcher: submitit_slurm

runs:
- name: benchmark_run
tasks: [benchmark]
19 changes: 19 additions & 0 deletions conf/bg/atoms.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,19 @@
# @package _global_
model:
representation: atom_sequences
dataset: "bandgap"
dataset_type: matbench
special_num_token: False
checkpoint: n0w0f/MatText-atom-seq-2m
logging:
wandb_project: revision-bg

finetune:
model_name: revision-bg
context_length: 32
training_arguments:
per_device_train_batch_size: 1024
path:
pretrained_checkpoint: n0w0f/MatText-atom-seq-2m


17 changes: 17 additions & 0 deletions conf/bg/atoms_params.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
# @package _global_
model:
representation: atom_sequences_plusplus
dataset: "bandgap"
dataset_type: matbench
special_num_token: False
checkpoint: n0w0f/MatText-atom-seq-plusplus-2m
logging:
wandb_project: revision-bg

finetune:
model_name: revision-bg
context_length: 32
training_arguments:
per_device_train_batch_size: 1024


17 changes: 17 additions & 0 deletions conf/bg/cifp1.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
# @package _global_
model:
representation: cif_p1
dataset: "bandgap"
dataset_type: matbench
special_num_token: False
checkpoint: n0w0f/MatText-cifp1-2m
logging:
wandb_project: revision-bg

finetune:
model_name: revision-bg
context_length: 1024
training_arguments:
per_device_train_batch_size: 128
path:
pretrained_checkpoint: n0w0f/MatText-cifp1-2m
17 changes: 17 additions & 0 deletions conf/bg/cifpsym.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
# @package _global_
model:
representation: cif_symmetrized
dataset: "bandgap"
dataset_type: matbench
special_num_token: False
checkpoint: n0w0f/MatText-cifsymmetrized-2m
logging:
wandb_project: revision-bg

finetune:
model_name: revision-bg
context_length: 1024
training_arguments:
per_device_train_batch_size: 64
path:
pretrained_checkpoint: n0w0f/MatText-cifsymmetrized-2m
17 changes: 17 additions & 0 deletions conf/bg/composition.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
# @package _global_
model:
representation: composition
dataset: "bandgap"
dataset_type: matbench
special_num_token: False
checkpoint: n0w0f/MatText-composition-2m
logging:
wandb_project: revision-bg

finetune:
model_name: revision-bg
context_length: 32
training_arguments:
per_device_train_batch_size: 1024


16 changes: 16 additions & 0 deletions conf/bg/crystal_llm.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,16 @@
# @package _global_
model:
representation: crystal_text_llm
dataset: "bandgap"
dataset_type: matbench
special_num_token: False
checkpoint: /home/so87pot/n0w0f/structllm_ckpt/alpaca_ckpt/checkpoint-393000
logging:
wandb_project: revision-bg

finetune:
model_name: revision-bg
context_length: 512
training_arguments:
per_device_train_batch_size: 256

17 changes: 17 additions & 0 deletions conf/bg/local_env.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
# @package _global_
model:
representation: local_env
dataset: "bandgap"
dataset_type: matbench
special_num_token: False
checkpoint: /home/so87pot/n0w0f/structllm_ckpt/santiago_ckpt_rt/checkpoint-95000
logging:
wandb_project: revision-bg

finetune:
model_name: revision-bg
context_length: 512
training_arguments:
per_device_train_batch_size: 256
path:
pretrained_checkpoint: /home/so87pot/n0w0f/structllm_ckpt/santiago_ckpt_rt/checkpoint-95000
17 changes: 17 additions & 0 deletions conf/bg/slices.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
# @package _global_
model:
representation: slices
dataset: "bandgap"
dataset_type: matbench
special_num_token: False
checkpoint: n0w0f/MatText-slices-2m
logging:
wandb_project: revision-bg

finetune:
model_name: revision-bg
context_length: 512
training_arguments:
per_device_train_batch_size: 256
path:
pretrained_checkpoint: n0w0f/MatText-slices-2m
17 changes: 17 additions & 0 deletions conf/bg/zmatrix.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
# @package _global_
model:
representation: zmatrix
dataset: "bandgap"
dataset_type: matbench
special_num_token: False
checkpoint: n0w0f/MatText-zmatrix-2m
logging:
wandb_project: revision-bg

finetune:
model_name: revision-bg
context_length: 512
training_arguments:
per_device_train_batch_size: 256
path:
pretrained_checkpoint: n0w0f/MatText-zmatrix-2m
13 changes: 13 additions & 0 deletions conf/bg2m/atoms.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# @package _global_
model:
representation: atoms_params
logging:
wandb_project: 2m_intel_ft

finetune:
model_name: 2m_intel_ft
context_length: 32
training_arguments:
per_device_train_batch_size: 1024
path:
pretrained_checkpoint: /work/so87pot/mattext/megaloop/checkpoints/checkpoints/atoms_params_pt_30k_atoms/checkpoint-1000
13 changes: 13 additions & 0 deletions conf/bg2m/atoms_params.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# @package _global_
model:
representation: atoms_params
logging:
wandb_project: 2m_intel_ft

finetune:
model_name: 2m_intel_ft
context_length: 32
training_arguments:
per_device_train_batch_size: 1024
path:
pretrained_checkpoint: /work/so87pot/mattext/megaloop/checkpoints/checkpoints/atoms_params_pt_30k_atoms/checkpoint-1000
13 changes: 13 additions & 0 deletions conf/bg2m/cifp1.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# @package _global_
model:
representation: cif_p1
logging:
wandb_project: 2m_intel_ft

finetune:
model_name: 2m_intel_ft
context_length: 1024
training_arguments:
per_device_train_batch_size: 32
path:
pretrained_checkpoint: /work/so87pot/mattext/megaloop2/checkpoints/checkpoints/cif_p1_pt_30k_rt_2/checkpoint-46000
13 changes: 13 additions & 0 deletions conf/bg2m/cifsymmetrized.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# @package _global_
model:
representation: cif_symmetrized
logging:
wandb_project: 2m_intel_ft

finetune:
model_name: 2m_intel_ft
context_length: 1024
training_arguments:
per_device_train_batch_size: 32
path:
pretrained_checkpoint: /work/so87pot/mattext/megaloop2/checkpoints/checkpoints/cif_symmetrized_pt_30k_rt/checkpoint-45000
13 changes: 13 additions & 0 deletions conf/bg2m/composition.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# @package _global_
model:
representation: composition
logging:
wandb_project: 2m_intel_ft

finetune:
model_name: 2m_intel_ft
context_length: 32
training_arguments:
per_device_train_batch_size: 1024
path:
pretrained_checkpoint: /work/so87pot/mattext/megaloop2/checkpoints/checkpoints/composition_pt_30k_rt/checkpoint-1000
13 changes: 13 additions & 0 deletions conf/bg2m/crystal_llm.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# @package _global_
model:
representation: crystal_llm_rep
logging:
wandb_project: 2m_intel_ft

finetune:
model_name: 2m_intel_ft
context_length: 512
training_arguments:
per_device_train_batch_size: 64
path:
pretrained_checkpoint: /work/so87pot/mattext/megaloop2/checkpoints/checkpoints/crystal_llm_rep_pt_30k_rt/checkpoint-11000
13 changes: 13 additions & 0 deletions conf/bg2m/local_env.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# @package _global_
model:
representation: zmatrix
logging:
wandb_project: 2m_intel_ft

finetune:
model_name: 2m_intel_ft
context_length: 512
training_arguments:
per_device_train_batch_size: 64
path:
pretrained_checkpoint: /work/so87pot/mattext/megaloop/checkpoints/checkpoints/atoms_params_pt_30k_atoms/checkpoint-1000
13 changes: 13 additions & 0 deletions conf/bg2m/slice.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# @package _global_
model:
representation: slice
logging:
wandb_project: 2m_intel_ft

finetune:
model_name: 2m_intel_ft
context_length: 512
training_arguments:
per_device_train_batch_size: 64
path:
pretrained_checkpoint: /work/so87pot/mattext/megaloop2/checkpoints/checkpoints/slice_pt_30k_rt/checkpoint-23000
13 changes: 13 additions & 0 deletions conf/bg2m/zmatrix.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# @package _global_
model:
representation: zmatrix
logging:
wandb_project: 2m_intel_ft

finetune:
model_name: 2m_intel_ft
context_length: 512
training_arguments:
per_device_train_batch_size: 64
path:
pretrained_checkpoint: /work/so87pot/mattext/megaloop/checkpoints/checkpoints/atoms_params_pt_30k_atoms/checkpoint-1000
Loading

0 comments on commit 866a0e2

Please sign in to comment.