| | import os |
| | import time |
| | import torch |
| | import torch.nn as nn |
| | import torch.backends.cudnn as cudnn |
| | import torch.optim |
| | import torchvision.transforms as transforms |
| | import torchvision.datasets as datasets |
| | import torchvision.models as models |
| | from torch.utils.data import DataLoader |
| | import wandb |
| | import numpy as np |
| | from pathlib import Path |
| | from ptflops import get_model_complexity_info |
| | from timm.scheduler.cosine_lr import CosineLRScheduler |
| | from timm.scheduler.step_lr import StepLRScheduler |
| | from robustbench import benchmark |
| |
|
| | from omegaconf import DictConfig, OmegaConf |
| | import hydra |
| | from hydra.utils import get_original_cwd |
| | import logging |
| | from apex import amp |
| | from robustarch.utils import configure_optimizers, get_datasets, pad_str, save_checkpoint, make_linear_schedule |
| | from robustarch.adv_train import train, test_natural, test_pgd |
| |
|
| | from robustarch.models.model_torch import TorchModel |
| | from robustarch.models.model import NormalizedConfigurableModel |
| |
|
| | log = logging.getLogger(__name__) |
| |
|
| | @hydra.main(config_path="../configs", config_name="main", version_base="1.2") |
| | def main(cfg: DictConfig): |
| | |
| | device = torch.device(cfg.train_test.device) |
| | cwd = get_original_cwd() |
| | cudnn.benchmark = True |
| |
|
| | |
| | log.info(f"\n{pad_str(' ARGUMENTS ')}\n{OmegaConf.to_yaml(cfg)}\n{pad_str('')}") |
| |
|
| | |
| | log.info(f"=> creating model:") |
| | assert cfg.model.model_source in ["torch", "local", "madry"], f"{cfg.model.model_source} not supported" |
| | if cfg.model.model_source == "torch": |
| | model = TorchModel(cfg.model.arch, cfg.dataset.mean, cfg.dataset.std, **cfg.model.kwargs) |
| | elif cfg.model.model_source == "local": |
| | model = NormalizedConfigurableModel(cfg.dataset.mean, cfg.dataset.std, **hydra.utils.instantiate(cfg.model.kwargs)) |
| | else: |
| | raise NotImplementedError |
| | model.to(device) |
| | log.info(model) |
| |
|
| | |
| | log.info(pad_str(" Model Complexity Info ")) |
| | macs, params = get_model_complexity_info(model, (3, cfg.train_test.crop_size, cfg.train_test.crop_size), as_strings=True, |
| | print_per_layer_stat=False, verbose=False) |
| | log.info(f"{'Computational complexity: ':<30} {macs:<8}") |
| | log.info(f"{'Number of parameters: ':<30} {params:<8}") |
| |
|
| | |
| | criterion = nn.CrossEntropyLoss().to(device) |
| |
|
| | param_groups = configure_optimizers(model) |
| | optimizer_name = cfg.train_test.optim.lower() |
| | if optimizer_name == "sgd": |
| | optimizer = torch.optim.SGD(params=param_groups, lr=cfg.train_test.lr, momentum=cfg.train_test.momentum, weight_decay=cfg.train_test.weight_decay) |
| | elif optimizer_name == "adamw": |
| | raise NotImplementedError |
| |
|
| | |
| | if cfg.train_test.half and not cfg.train_test.evaluate: |
| | model, optimizer = amp.initialize(model, optimizer, opt_level="O1") |
| | model = nn.DataParallel(model) |
| |
|
| | |
| | if cfg.train_test.resume: |
| | prev_model_dir = os.path.normpath(Path(cwd) / cfg.train_test.resume) |
| | if os.path.isfile(prev_model_dir): |
| | log.info(f"=> loading checkpoint '{prev_model_dir}'") |
| | ckpt = torch.load(prev_model_dir) |
| | start_epoch = ckpt["epoch"] |
| | best_prec1 = ckpt["best_prec1"] |
| | model.load_state_dict(ckpt["state_dict"]) |
| | optimizer.load_state_dict(ckpt["optimizer"]) |
| | else: |
| | log.info(f"=> no checkpoint found at '{cfg.train_test.resume}'") |
| | start_epoch = cfg.train_test.start_epoch |
| |
|
| | |
| | train_loader, test_loader = get_datasets(cfg) |
| |
|
| | |
| | if cfg.train_test.evaluate: |
| | if cfg.attack.test.name == "pgd": |
| | log.info(pad_str(" Performing PGD Attacks ")) |
| | test_pgd(model, test_loader, criterion, device, cfg, log) |
| | test_natural(model, test_loader, criterion, device, cfg, log) |
| | elif cfg.attack.test.name == "aa": |
| | model.eval() |
| | test_aug = list() |
| | if cfg.train_test.mode == "at_pgd": |
| | if cfg.dataset.dataset == "imagenet": |
| | test_aug.append(transforms.Resize(256)) |
| | elif cfg.dataset.dataset == "cifar10": |
| | test_aug.append(transforms.Resize(cfg.train_test.crop_size)) |
| | test_aug.extend([ |
| | transforms.CenterCrop(cfg.train_test.crop_size), |
| | transforms.ToTensor(), |
| | ]) |
| | test_transform = transforms.Compose(test_aug) |
| | log.info(benchmark( |
| | model, |
| | dataset=cfg.dataset.dataset, |
| | data_dir=cfg.dataset.data_dir, |
| | device=device, |
| | batch_size=cfg.attack.test.batch_size, |
| | eps=cfg.attack.test.eps / 255., |
| | preprocessing=test_transform, |
| | n_examples=cfg.attack.test.n_examples, |
| | threat_model="Linf" if cfg.attack.test.norm == "linf" else cfg.attack.test.norm |
| | )) |
| |
|
| | return |
| |
|
| | |
| | if cfg.visualization.tool == "wandb": |
| | log.info(f"=> Visualization with wandb") |
| | wandb.init(project=cfg.visualization.project, entity=cfg.visualization.entity, resume=True) |
| | wandb.run.name = cfg.name |
| |
|
| | |
| | model_dir = Path(cwd) / cfg.train_test.model_dir |
| | if cfg.train_test.mode == "fat": |
| | model_dir = model_dir / cfg.train_test.phase |
| | model_dir.mkdir(parents=True, exist_ok=True) |
| |
|
| | |
| | eps_schedule = None |
| | if cfg.train_test.mode == "fat": |
| | lr_schedule = lambda t: np.interp(t, cfg.train_test.lr_epochs, cfg.train_test.lr_values) |
| | total_epochs = cfg.train_test.end_epoch |
| | else: |
| | total_epochs = cfg.train_test.end_epoch - cfg.train_test.start_epoch |
| | if cfg.train_test.schedule.lower() == "cosine": |
| | lr_schedule = CosineLRScheduler( |
| | optimizer, |
| | t_initial=total_epochs, |
| | lr_min=cfg.train_test.min_lr, |
| | warmup_lr_init=cfg.train_test.warmup_lr, |
| | warmup_t=cfg.train_test.warmup_epochs, |
| | cycle_mul=cfg.train_test.lr_cycle_mul, |
| | cycle_decay=cfg.train_test.lr_cycle_decay, |
| | cycle_limit=cfg.train_test.lr_cycle_limit, |
| | ) |
| | elif cfg.train_test.schedule.lower() == "step": |
| | lr_schedule = StepLRScheduler( |
| | optimizer, |
| | decay_t=cfg.train_test.decay_t, |
| | decay_rate=cfg.train_test.decay_rate, |
| | warmup_t=cfg.train_test.warmup_epochs, |
| | warmup_lr_init=cfg.train_test.warmup_lr, |
| | ) |
| | total_epochs = total_epochs + cfg.train_test.cooldown_epochs |
| |
|
| | |
| | if cfg.attack.train.eps_schedule and cfg.attack.train.eps_schedule.lower() == "linear" and cfg.attack.train.eps_schedule_epochs: |
| | eps_schedule = make_linear_schedule(1.0 * cfg.attack.train.eps / cfg.dataset.max_color_value, cfg.attack.train.eps_schedule_epochs, cfg.attack.train.zero_eps_epochs) |
| |
|
| | |
| | best_prec1 = 0. |
| | for epoch in range(start_epoch, total_epochs): |
| | |
| | train(model, train_loader, optimizer, criterion, lr_schedule, epoch, device, cfg, log, eps_schedule) |
| |
|
| | |
| | prec1 = test_natural(model, test_loader, criterion, device, cfg, log) |
| |
|
| | if cfg.visualization.tool == "wandb": |
| | wandb.log({"Test natural accuracy": prec1}, step=epoch) |
| |
|
| | |
| | is_best = prec1 > best_prec1 |
| | best_prec1 = max(prec1, best_prec1) |
| | model_info = dict( |
| | tag=cfg.name, |
| | best_prec1=best_prec1, |
| | epoch=epoch + 1, |
| | state_dict=model.state_dict(), |
| | optimizer=optimizer.state_dict() |
| | ) |
| | save_checkpoint(model_info, model_dir, is_best) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|