Skip to content

Commit e13d34c

Browse files
committed
Preparing for release v0.100.0 of LBANN. Updated the release notes to
include accomplishments since release v0.99.0.
1 parent 2cb48ba commit e13d34c

File tree

2 files changed

+97
-1
lines changed

2 files changed

+97
-1
lines changed

CMakeLists.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ endif ()
4848
#
4949

5050
set(LBANN_VERSION_MAJOR 0)
51-
set(LBANN_VERSION_MINOR 99)
51+
set(LBANN_VERSION_MINOR 100)
5252
set(LBANN_VERSION_PATCH 0)
5353

5454
set(LBANN_VERSION "${LBANN_VERSION_MAJOR}.${LBANN_VERSION_MINOR}.${LBANN_VERSION_PATCH}")

ReleaseNotes.txt

Lines changed: 96 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -21,6 +21,102 @@ Bug fixes:
2121

2222
Retired features:
2323

24+
============================== Release Notes: v0.100 ==============================
25+
Support for new network structures:
26+
- 3D molecular generation models for Metal Organic Frameworks from the CoRE MOF Database.
27+
- 3D CosmoFlow Model
28+
- DenseNet
29+
- ATOM LSTM model
30+
- RAS state classifier
31+
- node2vec
32+
- Transformer and other attention-based models
33+
- ExaGAN (formerly CosmoGAN)
34+
- MaCC ICF surrogate model
35+
36+
Applications:
37+
- Created a directory of example applications, deprecating the "model zoo" directory
38+
39+
Support for new layers:
40+
- Embedding layer
41+
- Distributed embedding layer
42+
- Channel-wise scale/bias layer
43+
- Entry-wise scale/bias layer
44+
- Gated-Recurrent Units (GRU)
45+
- Entry-wise batchnorm
46+
- Argmax, Argmin, and one-hot layers
47+
- Layer norm
48+
- Deconvolution layer (transposed convolution)
49+
- Layers for channel-wise operations (channel-wise fully-connected, channel-wise softmax, channel-wise scale/bias, instance norm)
50+
- Matrix multiply layer
51+
52+
Python front-end:
53+
- Can now configure contrib launcher with environment variables
54+
- Added NERSC compute center
55+
- Per-layer specification of compute device (CPU or GPU)
56+
- Option to write custom batch scripts with Python front-end
57+
58+
Performance optimizations:
59+
- Parallelized Python data reader with "multiprocessing" module
60+
- Fuse batchnorm stats allreduces in FP/BP.
61+
- Tuned concatenate and slice layer
62+
- Dynamically allocate and free memory for layer error signals (halves LBANN's memory footprint)
63+
64+
Model portability & usability:
65+
- Bamboo tests for individual layers
66+
67+
Internal features:
68+
- Added support for DistConv features (distributed, generalized,
69+
parallel convolution)
70+
- Added support for NVSHMEM 1.0 API (used in distributed embedding
71+
layer and DistConv halo exchange)
72+
- Support for multiple data types per model (per-layer)
73+
- Support for per-layer mixed-precision weight training and inference,
74+
includes per-weight object and objective function mixed-precision.
75+
- Improved how and when the RNGs are initialized
76+
- Callback to dump images to TensorBoard
77+
- Callback to save model weights (useful to export to PyTorch)
78+
- Callback to save top K models (LTFB)
79+
- Improved run-to-run reproducibility by initializing weights in alphabetical order
80+
- Moved models from model_zoo directory to applications directory
81+
- Cleanup and refactoring of callbacks and layer instantiation
82+
- Grouped batchnorm statistics
83+
- Callback to print model description
84+
- Refactored trainer and training-state out of the model class
85+
- Support for transposing data in matrix multiply layers
86+
- Added DiHydrogen tensor and DistConv library
87+
- Added parallel strategy to layer class to support DistConv
88+
- LBANN inference mode supports loading models from multiple directories
89+
- Cleanup of checkpoint and restart logic
90+
91+
I/O & data readers:
92+
- Added in-memory data store that caches samples in CPU memory. It can be loaded
93+
during the first epoch or preloaded
94+
- Added new "transform" data preprocessing ingestion pipeline
95+
- Added sample list format for specifying data sets
96+
- Introduced data coordinator that manages data readers and extracts them from
97+
the input layers
98+
- Data store is able to checkpoint / spill it's contents to local disk
99+
- Data reader for SMILE strings
100+
101+
Build system:
102+
- Hydrogen 1.3.4
103+
- Aluminum 0.3.3
104+
- Improved documentation on read the docs (RTD)
105+
- Robust support for using Spack as a build system around CMake
106+
- Identified compute centers for specifying build and run dependencies
107+
- Added Catch2-based tests
108+
109+
Bug fixes:
110+
- Fixed path resolution for dump weights, save model, and checkpoint callbacks
111+
- Added mutexes for preloading the data store
112+
- Fixed the LTFB exchange to include all ADAM optimizer state
113+
- Fixed the mapping of I/O RNGs to I/O processing threads to ensure
114+
consistent and correct multi-threaded performance
115+
116+
Retired features:
117+
- moving MNIST data reader is replaced by python data reader
118+
- ASCII data reader is deprecated
119+
24120
============================== Release Notes: v0.99 ==============================
25121
Support for new training algorithms:
26122
- Improvements to LTFB infrastructure (including transfer of SGD and Adam hyperparameters)

0 commit comments

Comments
 (0)