Commit 6e9d5cb0 authored by mohammad's avatar mohammad Committed by Deepak Narayanan
Browse files

Move args writer to the beginning of training

parent 8a6e56b8
Loading
Loading
Loading
Loading
+3 −4
Original line number Diff line number Diff line
@@ -79,8 +79,6 @@ def initialize_megatron(extra_args_provider=None, args_defaults={},
        # Autoresume.
        _init_autoresume()
        
        # Write arguments to tensorboard.
        _write_args_to_tensorboard()
        # No continuation function
        return None
        
@@ -154,13 +152,14 @@ def _set_random_seed(seed_):
        raise ValueError('Seed ({}) should be a positive integer.'.format(seed))


def _write_args_to_tensorboard():
def write_args_to_tensorboard():
    """Write arguments to tensorboard."""
    args = get_args()
    writer = get_tensorboard_writer()
    if writer:
        for arg in vars(args):
            writer.add_text(arg, str(getattr(args, arg)))
            writer.add_text(arg, str(getattr(args, arg)),
                            global_step=args.iteration)


def _initialize_mem_buffs():
+4 −0
Original line number Diff line number Diff line
@@ -41,6 +41,7 @@ from megatron.checkpointing import save_checkpoint
from megatron.fp16 import FP16_Module
from megatron.fp16 import FP16_Optimizer
from megatron.initialize import initialize_megatron
from megatron.initialize import write_args_to_tensorboard
from megatron.learning_rates import AnnealingLR
from megatron.model import DistributedDataParallel as LocalDDP
from megatron.model import get_params_for_weight_decay_optimization
@@ -811,6 +812,9 @@ def train(forward_step_func, model, optimizer, lr_scheduler,
    args = get_args()
    timers = get_timers()

    # Write args to tensorboard
    write_args_to_tensorboard()

    # Turn on training mode which enables dropout.
    model.train()