Checkpoint clean fn useable stand alone

pull/910/head
Ross Wightman 3 years ago
parent b544ad4d3f
commit 38804c721b

@ -20,7 +20,7 @@ parser.add_argument('--checkpoint', default='', type=str, metavar='PATH',
help='path to latest checkpoint (default: none)') help='path to latest checkpoint (default: none)')
parser.add_argument('--output', default='', type=str, metavar='PATH', parser.add_argument('--output', default='', type=str, metavar='PATH',
help='output path') help='output path')
parser.add_argument('--use-ema', dest='use_ema', action='store_true', parser.add_argument('--no-use-ema', dest='no_use_ema', action='store_true',
help='use ema version of weights if present') help='use ema version of weights if present')
parser.add_argument('--clean-aux-bn', dest='clean_aux_bn', action='store_true', parser.add_argument('--clean-aux-bn', dest='clean_aux_bn', action='store_true',
help='remove auxiliary batch norm layers (from SplitBN training) from checkpoint') help='remove auxiliary batch norm layers (from SplitBN training) from checkpoint')
@ -35,19 +35,23 @@ def main():
print("Error: Output filename ({}) already exists.".format(args.output)) print("Error: Output filename ({}) already exists.".format(args.output))
exit(1) exit(1)
clean_checkpoint(args.checkpoint, args.output, not args.no_use_ema, args.clean_aux_bn)
def clean_checkpoint(checkpoint, output='', use_ema=True, clean_aux_bn=False):
# Load an existing checkpoint to CPU, strip everything but the state_dict and re-save # Load an existing checkpoint to CPU, strip everything but the state_dict and re-save
if args.checkpoint and os.path.isfile(args.checkpoint): if checkpoint and os.path.isfile(checkpoint):
print("=> Loading checkpoint '{}'".format(args.checkpoint)) print("=> Loading checkpoint '{}'".format(checkpoint))
state_dict = load_state_dict(args.checkpoint, use_ema=args.use_ema) state_dict = load_state_dict(checkpoint, use_ema=use_ema)
new_state_dict = {} new_state_dict = {}
for k, v in state_dict.items(): for k, v in state_dict.items():
if args.clean_aux_bn and 'aux_bn' in k: if clean_aux_bn and 'aux_bn' in k:
# If all aux_bn keys are removed, the SplitBN layers will end up as normal and # If all aux_bn keys are removed, the SplitBN layers will end up as normal and
# load with the unmodified model using BatchNorm2d. # load with the unmodified model using BatchNorm2d.
continue continue
name = k[7:] if k.startswith('module') else k name = k[7:] if k.startswith('module') else k
new_state_dict[name] = v new_state_dict[name] = v
print("=> Loaded state_dict from '{}'".format(args.checkpoint)) print("=> Loaded state_dict from '{}'".format(checkpoint))
try: try:
torch.save(new_state_dict, _TEMP_NAME, _use_new_zipfile_serialization=False) torch.save(new_state_dict, _TEMP_NAME, _use_new_zipfile_serialization=False)
@ -57,17 +61,19 @@ def main():
with open(_TEMP_NAME, 'rb') as f: with open(_TEMP_NAME, 'rb') as f:
sha_hash = hashlib.sha256(f.read()).hexdigest() sha_hash = hashlib.sha256(f.read()).hexdigest()
if args.output: if output:
checkpoint_root, checkpoint_base = os.path.split(args.output) checkpoint_root, checkpoint_base = os.path.split(output)
checkpoint_base = os.path.splitext(checkpoint_base)[0] checkpoint_base = os.path.splitext(checkpoint_base)[0]
else: else:
checkpoint_root = '' checkpoint_root = ''
checkpoint_base = os.path.splitext(args.checkpoint)[0] checkpoint_base = os.path.splitext(checkpoint)[0]
final_filename = '-'.join([checkpoint_base, sha_hash[:8]]) + '.pth' final_filename = '-'.join([checkpoint_base, sha_hash[:8]]) + '.pth'
shutil.move(_TEMP_NAME, os.path.join(checkpoint_root, final_filename)) shutil.move(_TEMP_NAME, os.path.join(checkpoint_root, final_filename))
print("=> Saved state_dict to '{}, SHA256: {}'".format(final_filename, sha_hash)) print("=> Saved state_dict to '{}, SHA256: {}'".format(final_filename, sha_hash))
return final_filename
else: else:
print("Error: Checkpoint ({}) doesn't exist".format(args.checkpoint)) print("Error: Checkpoint ({}) doesn't exist".format(checkpoint))
return ''
if __name__ == '__main__': if __name__ == '__main__':

Loading…
Cancel
Save