Skip to content

Commit

Permalink
Add Upsample Layer (#2255)
Browse files Browse the repository at this point in the history
* Add skeleton for upsampling layer

* Add upsample cudnn backend support

* Change import and tiny fix

* Add dnn lib implementation to upsample layer

* Fix order of parameters

* Add upsample layer unit test

* Add distconv support

* Add distconv unit test and minor tweaks

* Update layer to fix build issues

* Update copyright year

* Fix unit tests

* Throw error and skip tests on ROCm systems

* clang format

* Fix distconv test

* Apply Tom's suggestions

Co-authored-by: Tom Benson <[email protected]>

* General cleanup and properly serialize state

* Fix CPU build and general cleanup

* Update include/lbann/layers/transform/upsample.hpp

Co-authored-by: Tal Ben-Nun <[email protected]>

* Tal's suggestions + cleanup and formatting

---------

Co-authored-by: Tom Benson <[email protected]>
Co-authored-by: Tal Ben-Nun <[email protected]>
  • Loading branch information
3 people authored Jun 28, 2024
1 parent 602de5b commit c920d33
Show file tree
Hide file tree
Showing 17 changed files with 1,699 additions and 1 deletion.
2 changes: 1 addition & 1 deletion ci_test/common_python/tools.py
Original file line number Diff line number Diff line change
Expand Up @@ -835,7 +835,7 @@ def test_func(cluster, dirname, weekly):
)
assert_success(return_code, stderr_log_file)
if post_test_func is not None:
post_test_func(lbann, weekly)
post_test_func(lbann, weekly, **_kwargs)
return {
'return_code': return_code,
'work_dir': work_dir,
Expand Down
288 changes: 288 additions & 0 deletions ci_test/unit_tests/test_unit_layer_upsample.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,288 @@
import functools
import math
import operator
import os
import os.path
import sys
import numpy as np
import lbann.contrib.args

# Bamboo utilities
current_file = os.path.realpath(__file__)
current_dir = os.path.dirname(current_file)
sys.path.insert(0, os.path.join(os.path.dirname(current_dir), "common_python"))
import tools

# ==============================================
# Objects for Python data reader
# ==============================================
# Note: The Python data reader imports this file as a module and calls
# the functions below to ingest data.


def make_random_array(shape, seed):
"""Hacked function to generate a random array.
NumPy's RNG produces different values with different NumPy
versions. This function is helpful when array values must be
identical across all runs, e.g. when checking against precomputed
metric values.
Args:
shape (Iterable of int): Array dimensions
seed (int): Parameter for RNG. Must be non-zero.
Returns:
numpy.ndarray: Array of `np.float32`. Values will be in
[-0.5,0.5).
"""
size = functools.reduce(operator.mul, shape)
eps = np.finfo(np.float32).eps
x = (seed / np.linspace(math.sqrt(eps), 0.1, size)) % 1 - 0.5
return x.reshape(shape).astype(np.float32)


# Data
_num_samples = 64
_sample_dims = [6, 11, 7]
_sample_dims_3d = [2, 3, 11, 7]
_sample_size = functools.reduce(operator.mul, _sample_dims)
_samples = make_random_array([_num_samples] + _sample_dims, 7)


# Sample access functions
def get_sample(index):
return _samples[index, :].reshape(-1)


def num_samples():
return _num_samples


def sample_dims():
return (_sample_size,)


# ==============================================
# Setup LBANN experiment
# ==============================================


def setup_experiment(lbann, weekly):
"""Construct LBANN experiment.
Args:
lbann (module): Module for LBANN Python frontend
"""
mini_batch_size = num_samples() // 2
trainer = lbann.Trainer(mini_batch_size)
model = construct_model(lbann)
data_reader = construct_data_reader(lbann)
optimizer = lbann.NoOptimizer()
return (
trainer,
model,
data_reader,
optimizer,
None,
) # Don't request any specific number of nodes


upsample_configs = []

# 3x3 upsampling
for mode in ["nearest"]:
upsample_configs.append(
{
"name": "3x3 {} upsample".format(mode),
"scale_factors": (3, 3),
"upsample_mode": mode,
}
)

# 2x4 upsampling
for mode in ["nearest"]:
upsample_configs.append(
{
"name": "2x4 {} upsample".format(mode),
"scale_factors": (2, 4),
"upsample_mode": mode,
}
)

# 2x2x2 3D upsampling
for mode in ["nearest"]:
upsample_configs.append(
{
"name": "2x2x2 {} upsample".format(mode),
"scale_factors": (2, 2, 2),
"upsample_mode": mode,
}
)


def construct_model(lbann):
"""Construct LBANN model.
Args:
lbann (module): Module for LBANN Python frontend
"""

# Input data
# Note: Sum with a weights layer so that gradient checking will
# verify that error signals are correct.
x_weights = lbann.Weights(
optimizer=lbann.SGD(),
initializer=lbann.ConstantInitializer(value=0.0),
name="input_weights",
)
x = lbann.Sum(
lbann.Reshape(lbann.Input(data_field="samples"), dims=_sample_dims),
lbann.WeightsLayer(weights=x_weights, dims=_sample_dims),
)
x_lbann = x

# Objects for LBANN model
obj = []
metrics = []
callbacks = []

# ------------------------------------------
# Upsample
# ------------------------------------------

for u in upsample_configs:
uname = u["name"].split(" ")[0]

# Apply upsampling
x = x_lbann
if len(u["scale_factors"]) == 3:
x = lbann.Reshape(x, dims=_sample_dims_3d)
x = lbann.Identity(x, name=f"in_{uname}")

y = lbann.Upsample(
x,
num_dims=len(u["scale_factors"]),
has_vectors=True,
scale_factors=u["scale_factors"],
upsample_mode=u["upsample_mode"],
)
y = lbann.Identity(y, name=f"out_{uname}")
z = lbann.L2Norm2(y)

obj.append(z)

# Save the inputs and outputs to check later.
callbacks.append(
lbann.CallbackDumpOutputs(
layers=f"in_{uname} out_{uname}", directory="outputs"
)
)

# ------------------------------------------
# Gradient checking
# ------------------------------------------

callbacks.append(lbann.CallbackCheckGradients(error_on_failure=True))

# ------------------------------------------
# Construct model
# ------------------------------------------

num_epochs = 0
return lbann.Model(
num_epochs,
layers=lbann.traverse_layer_graph(x_lbann),
objective_function=obj,
metrics=metrics,
callbacks=callbacks,
)


def construct_data_reader(lbann):
"""Construct Protobuf message for Python data reader.
The Python data reader will import the current Python file to
access the sample access functions.
Args:
lbann (module): Module for LBANN Python frontend
"""

# Note: The training data reader should be removed when
# https://github.com/LLNL/lbann/issues/1098 is resolved.
message = lbann.reader_pb2.DataReader()
message.reader.extend(
[
tools.create_python_data_reader(
lbann, current_file, "get_sample", "num_samples", "sample_dims", "train"
)
]
)
message.reader.extend(
[
tools.create_python_data_reader(
lbann, current_file, "get_sample", "num_samples", "sample_dims", "test"
)
]
)
return message


# ==============================================
# Setup PyTest
# ==============================================


def check_output(lbann, weekly, **kwargs):
for u in upsample_configs:
uname = u["name"].split(" ")[0]
in_data = np.loadtxt(
os.path.join(
kwargs["work_dir"],
"outputs",
"trainer0",
"model0",
f"sgd.testing.epoch.0.step.0_in_{uname}_output0.csv",
),
delimiter=",",
)
out_data = np.loadtxt(
os.path.join(
kwargs["work_dir"],
"outputs",
"trainer0",
"model0",
f"sgd.testing.epoch.0.step.0_out_{uname}_output0.csv",
),
delimiter=",",
)

ndims = len(u["scale_factors"])
upsampled_data = in_data.copy().reshape(
[-1] + (_sample_dims if ndims == 2 else _sample_dims_3d)
)
for i, scale_fac in enumerate(u["scale_factors"]):
if u["upsample_mode"] == "nearest":
upsampled_data = upsampled_data.repeat(scale_fac, axis=i + 2)

assert np.allclose(upsampled_data.ravel(), out_data.ravel())


# Runtime parameters/arguments
environment = lbann.contrib.args.get_distconv_environment()
environment["LBANN_KEEP_ERROR_SIGNALS"] = 1

# Create test functions that can interact with PyTest
# Note: Create test name by removing ".py" from file name
_test_name = os.path.splitext(os.path.basename(current_file))[0]
for _test_func in tools.create_tests(
setup_experiment,
_test_name,
post_test_func=check_output,
skip_clusters=["tioga", "corona"],
):
globals()[_test_func.__name__] = _test_func
Loading

0 comments on commit c920d33

Please sign in to comment.