@@ -21,6 +21,102 @@ Bug fixes:
2121
2222Retired features:
2323
24+ ============================== Release Notes: v0.100 ==============================
25+ Support for new network structures:
26+ - 3D molecular generation models for Metal Organic Frameworks from the CoRE MOF Database.
27+ - 3D CosmoFlow Model
28+ - DenseNet
29+ - ATOM LSTM model
30+ - RAS state classifier
31+ - node2vec
32+ - Transformer and other attention-based models
33+ - ExaGAN (formerly CosmoGAN)
34+ - MaCC ICF surrogate model
35+
36+ Applications:
37+ - Created a directory of example applications, deprecating the "model zoo" directory
38+
39+ Support for new layers:
40+ - Embedding layer
41+ - Distributed embedding layer
42+ - Channel-wise scale/bias layer
43+ - Entry-wise scale/bias layer
44+ - Gated-Recurrent Units (GRU)
45+ - Entry-wise batchnorm
46+ - Argmax, Argmin, and one-hot layers
47+ - Layer norm
48+ - Deconvolution layer (transposed convolution)
49+ - Layers for channel-wise operations (channel-wise fully-connected, channel-wise softmax, channel-wise scale/bias, instance norm)
50+ - Matrix multiply layer
51+
52+ Python front-end:
53+ - Can now configure contrib launcher with environment variables
54+ - Added NERSC compute center
55+ - Per-layer specification of compute device (CPU or GPU)
56+ - Option to write custom batch scripts with Python front-end
57+
58+ Performance optimizations:
59+ - Parallelized Python data reader with "multiprocessing" module
60+ - Fuse batchnorm stats allreduces in FP/BP.
61+ - Tuned concatenate and slice layer
62+ - Dynamically allocate and free memory for layer error signals (halves LBANN's memory footprint)
63+
64+ Model portability & usability:
65+ - Bamboo tests for individual layers
66+
67+ Internal features:
68+ - Added support for DistConv features (distributed, generalized,
69+ parallel convolution)
70+ - Added support for NVSHMEM 1.0 API (used in distributed embedding
71+ layer and DistConv halo exchange)
72+ - Support for multiple data types per model (per-layer)
73+ - Support for per-layer mixed-precision weight training and inference,
74+ includes per-weight object and objective function mixed-precision.
75+ - Improved how and when the RNGs are initialized
76+ - Callback to dump images to TensorBoard
77+ - Callback to save model weights (useful to export to PyTorch)
78+ - Callback to save top K models (LTFB)
79+ - Improved run-to-run reproducibility by initializing weights in alphabetical order
80+ - Moved models from model_zoo directory to applications directory
81+ - Cleanup and refactoring of callbacks and layer instantiation
82+ - Grouped batchnorm statistics
83+ - Callback to print model description
84+ - Refactored trainer and training-state out of the model class
85+ - Support for transposing data in matrix multiply layers
86+ - Added DiHydrogen tensor and DistConv library
87+ - Added parallel strategy to layer class to support DistConv
88+ - LBANN inference mode supports loading models from multiple directories
89+ - Cleanup of checkpoint and restart logic
90+
91+ I/O & data readers:
92+ - Added in-memory data store that caches samples in CPU memory. It can be loaded
93+ during the first epoch or preloaded
94+ - Added new "transform" data preprocessing ingestion pipeline
95+ - Added sample list format for specifying data sets
96+ - Introduced data coordinator that manages data readers and extracts them from
97+ the input layers
98+ - Data store is able to checkpoint / spill it's contents to local disk
99+ - Data reader for SMILE strings
100+
101+ Build system:
102+ - Hydrogen 1.3.4
103+ - Aluminum 0.3.3
104+ - Improved documentation on read the docs (RTD)
105+ - Robust support for using Spack as a build system around CMake
106+ - Identified compute centers for specifying build and run dependencies
107+ - Added Catch2-based tests
108+
109+ Bug fixes:
110+ - Fixed path resolution for dump weights, save model, and checkpoint callbacks
111+ - Added mutexes for preloading the data store
112+ - Fixed the LTFB exchange to include all ADAM optimizer state
113+ - Fixed the mapping of I/O RNGs to I/O processing threads to ensure
114+ consistent and correct multi-threaded performance
115+
116+ Retired features:
117+ - moving MNIST data reader is replaced by python data reader
118+ - ASCII data reader is deprecated
119+
24120============================== Release Notes: v0.99 ==============================
25121Support for new training algorithms:
26122 - Improvements to LTFB infrastructure (including transfer of SGD and Adam hyperparameters)
0 commit comments