https://github.com/freewym/espresso
Raw File
Tip revision: 660facf088ded9f084cc1a24a1f00f64ce5f6918 authored by freewym on 20 July 2023, 23:05:26 UTC
allows dictionary files w/o the counts column; rename task's
Tip revision: 660facf
test_average_checkpoints.py
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.

import collections
import os
import shutil
import tempfile
import unittest

import numpy as np
import torch
from scripts.average_checkpoints import average_checkpoints
from torch import nn


class ModelWithSharedParameter(nn.Module):
    def __init__(self):
        super(ModelWithSharedParameter, self).__init__()
        self.embedding = nn.Embedding(1000, 200)
        self.FC1 = nn.Linear(200, 200)
        self.FC2 = nn.Linear(200, 200)
        # tie weight in FC2 to FC1
        self.FC2.weight = nn.Parameter(self.FC1.weight)
        self.FC2.bias = nn.Parameter(self.FC1.bias)

        self.relu = nn.ReLU()

    def forward(self, input):
        return self.FC2(self.ReLU(self.FC1(input))) + self.FC1(input)


class TestAverageCheckpoints(unittest.TestCase):
    def test_average_checkpoints(self):
        params_0 = collections.OrderedDict(
            [
                ("a", torch.DoubleTensor([100.0])),
                ("b", torch.FloatTensor([[1.0, 2.0, 3.0], [4.0, 5.0, 6.0]])),
                ("c", torch.IntTensor([7, 8, 9])),
            ]
        )
        params_1 = collections.OrderedDict(
            [
                ("a", torch.DoubleTensor([1.0])),
                ("b", torch.FloatTensor([[1.0, 1.0, 1.0], [1.0, 1.0, 1.0]])),
                ("c", torch.IntTensor([2, 2, 2])),
            ]
        )
        params_avg = collections.OrderedDict(
            [
                ("a", torch.DoubleTensor([50.5])),
                ("b", torch.FloatTensor([[1.0, 1.5, 2.0], [2.5, 3.0, 3.5]])),
                # We expect truncation for integer division
                ("c", torch.IntTensor([4, 5, 5])),
            ]
        )

        fd_0, path_0 = tempfile.mkstemp()
        fd_1, path_1 = tempfile.mkstemp()
        torch.save(collections.OrderedDict([("model", params_0)]), path_0)
        torch.save(collections.OrderedDict([("model", params_1)]), path_1)

        output = average_checkpoints([path_0, path_1])["model"]

        os.close(fd_0)
        os.remove(path_0)
        os.close(fd_1)
        os.remove(path_1)

        for (k_expected, v_expected), (k_out, v_out) in zip(
            params_avg.items(), output.items()
        ):
            self.assertEqual(
                k_expected,
                k_out,
                "Key mismatch - expected {} but found {}. "
                "(Expected list of keys: {} vs actual list of keys: {})".format(
                    k_expected, k_out, params_avg.keys(), output.keys()
                ),
            )
            np.testing.assert_allclose(
                v_expected.numpy(),
                v_out.numpy(),
                err_msg="Tensor value mismatch for key {}".format(k_expected),
            )

    def test_average_checkpoints_with_shared_parameters(self):
        def _construct_model_with_shared_parameters(path, value):
            m = ModelWithSharedParameter()
            nn.init.constant_(m.FC1.weight, value)
            torch.save({"model": m.state_dict()}, path)
            return m

        tmpdir = tempfile.mkdtemp()
        paths = []
        path = os.path.join(tmpdir, "m1.pt")
        m1 = _construct_model_with_shared_parameters(path, 1.0)
        paths.append(path)

        path = os.path.join(tmpdir, "m2.pt")
        m2 = _construct_model_with_shared_parameters(path, 2.0)
        paths.append(path)

        path = os.path.join(tmpdir, "m3.pt")
        m3 = _construct_model_with_shared_parameters(path, 3.0)
        paths.append(path)

        new_model = average_checkpoints(paths)
        self.assertTrue(
            torch.equal(
                new_model["model"]["embedding.weight"],
                (m1.embedding.weight + m2.embedding.weight + m3.embedding.weight) / 3.0,
            )
        )

        self.assertTrue(
            torch.equal(
                new_model["model"]["FC1.weight"],
                (m1.FC1.weight + m2.FC1.weight + m3.FC1.weight) / 3.0,
            )
        )

        self.assertTrue(
            torch.equal(
                new_model["model"]["FC2.weight"],
                (m1.FC2.weight + m2.FC2.weight + m3.FC2.weight) / 3.0,
            )
        )
        shutil.rmtree(tmpdir)


if __name__ == "__main__":
    unittest.main()
back to top