All notable changes to MONAI are documented in this file.
The format is based on Keep a Changelog and this project adheres to Semantic Versioning.
- renamed model's
n_classes
tonum_classes
0.6.0 - 2021-07-08
- 10 new transforms, a masked loss wrapper, and a
NetAdapter
for transfer learning - APIs to load networks and pre-trained weights from Clara Train Medical Model ARchives (MMARs)
- Base metric and cumulative metric APIs, 4 new regression metrics
- Initial CSV dataset support
- Decollating mini-batch as the default first postprocessing step, Migrating your v0.5 code to v0.6 wiki shows how to adapt to the breaking changes
- Initial backward compatibility support via
monai.utils.deprecated
- Attention-based vision modules and
UNETR
for segmentation - Generic module loaders and Gaussian mixture models using the PyTorch JIT compilation
- Inverse of image patch sampling transforms
- Network block utilities
get_[norm, act, dropout, pool]_layer
unpack_items
mode forapply_transform
andCompose
- New event
INNER_ITERATION_STARTED
in the deepgrow interactive workflow set_data
API for cache-based datasets to dynamically update the dataset content- Fully compatible with PyTorch 1.9
--disttests
and--min
options forruntests.sh
- Initial support of pre-merge tests with Nvidia Blossom system
- Base Docker image upgraded to
nvcr.io/nvidia/pytorch:21.06-py3
fromnvcr.io/nvidia/pytorch:21.04-py3
- Optionally depend on PyTorch-Ignite v0.4.5 instead of v0.4.4
- Unified the demo, tutorial, testing data to the project shared drive, and
Project-MONAI/MONAI-extra-test-data
- Unified the terms:
post_transform
is renamed topostprocessing
,pre_transform
is renamed topreprocessing
- Unified the postprocessing transforms and event handlers to accept the "channel-first" data format
evenly_divisible_all_gather
andstring_list_all_gather
moved tomonai.utils.dist
- Support of 'batched' input for postprocessing transforms and event handlers
TorchVisionFullyConvModel
set_visible_devices
utility functionSegmentationSaver
andTransformsInverter
handlers
- Issue of handling big-endian image headers
- Multi-thread issue for non-random transforms in the cache-based datasets
- Persistent dataset issue when multiple processes sharing a non-exist cache location
- Typing issue with Numpy 1.21.0
- Loading checkpoint with both
model
andoptmizier
usingCheckpointLoader
whenstrict_shape=False
SplitChannel
has different behaviour depending on numpy/torch inputs- Transform pickling issue caused by the Lambda functions
- Issue of filtering by name in
generate_param_groups
- Inconsistencies in the return value types of
class_activation_maps
- Various docstring typos
- Various usability enhancements in
monai.transforms
0.5.3 - 2021-05-28
- Project default branch renamed to
dev
frommaster
- Base Docker image upgraded to
nvcr.io/nvidia/pytorch:21.04-py3
fromnvcr.io/nvidia/pytorch:21.02-py3
- Enhanced type checks for the
iteration_metric
handler - Enhanced
PersistentDataset
to usetempfile
during caching computation - Enhanced various info/error messages
- Enhanced performance of
RandAffine
- Enhanced performance of
SmartCacheDataset
- Optionally requires
cucim
when the platform isLinux
- Default
device
ofTestTimeAugmentation
changed tocpu
- Download utilities now provide better default parameters
- Duplicated
key_transforms
in the patch-based transforms - A multi-GPU issue in
ClassificationSaver
- A default
meta_data
issue inSpacingD
- Dataset caching issue with the persistent data loader workers
- A memory issue in
permutohedral_cuda
- Dictionary key issue in
CopyItemsd
box_start
andbox_end
parameters for deepgrowSpatialCropForegroundd
- Tissue mask array transpose issue in
MaskedInferenceWSIDataset
- Various type hint errors
- Various docstring typos
- Support of
to_tensor
anddevice
arguments forTransformInverter
- Slicing options with SpatialCrop
- Class name alias for the networks for backward compatibility
k_divisible
option for CropForegroundmap_items
option forCompose
- Warnings of
inf
andnan
for surface distance computation - A
print_log
flag to the image savers - Basic testing pipelines for Python 3.9
0.5.0 - 2021-04-09
- Overview document for feature highlights in v0.5.0
- Invertible spatial transforms
InvertibleTransform
base APIs- Batch inverse and decollating APIs
- Inverse of
Compose
- Batch inverse event handling
- Test-time augmentation as an application
- Initial support of learning-based image registration:
- Bending energy, LNCC, and global mutual information loss
- Fully convolutional architectures
- Dense displacement field, dense velocity field computation
- Warping with high-order interpolation with C++/CUDA implementations
- Deepgrow modules for interactive segmentation:
- Workflows with simulations of clicks
- Distance-based transforms for guidance signals
- Digital pathology support:
- Efficient whole slide imaging IO and sampling with Nvidia cuCIM and SmartCache
- FROC measurements for lesion
- Probabilistic post-processing for lesion detection
- TorchVision classification model adaptor for fully convolutional analysis
- 12 new transforms, grid patch dataset,
ThreadDataLoader
, EfficientNets B0-B7 - 4 iteration events for the engine for finer control of workflows
- New C++/CUDA extensions:
- Conditional random field
- Fast bilateral filtering using the permutohedral lattice
- Metrics summary reporting and saving APIs
- DiceCELoss, DiceFocalLoss, a multi-scale wrapper for segmentation loss computation
- Data loading utilities:
decollate_batch
PadListDataCollate
with inverse support
- Support of slicing syntax for
Dataset
- Initial Torchscript support for the loss modules
- Learning rate finder
- Allow for missing keys in the dictionary-based transforms
- Support of checkpoint loading for transfer learning
- Various summary and plotting utilities for Jupyter notebooks
- Contributor Covenant Code of Conduct
- Major CI/CD enhancements covering the tutorial repository
- Fully compatible with PyTorch 1.8
- Initial nightly CI/CD pipelines using Nvidia Blossom Infrastructure
- Enhanced
list_data_collate
error handling - Unified iteration metric APIs
densenet*
extensions are renamed toDenseNet*
se_res*
network extensions are renamed toSERes*
- Transform base APIs are rearranged into
compose
,inverse
, andtransform
_do_transform
flag for the random augmentations is unified viaRandomizableTransform
- Decoupled post-processing steps, e.g.
softmax
,to_onehot_y
, from the metrics computations - Moved the distributed samplers to
monai.data.samplers
frommonai.data.utils
- Engine's data loaders now accept generic iterables as input
- Workflows now accept additional custom events and state properties
- Various type hints according to Numpy 1.20
- Refactored testing utility
runtests.sh
to have--unittest
and--net
(integration tests) options - Base Docker image upgraded to
nvcr.io/nvidia/pytorch:21.02-py3
fromnvcr.io/nvidia/pytorch:20.10-py3
- Docker images are now built with self-hosted environments
- Primary contact email updated to
monai.contact@gmail.com
- Now using GitHub Discussions as the primary communication forum
- Compatibility tests for PyTorch 1.5.x
- Format specific loaders, e.g.
LoadNifti
,NiftiDataset
- Assert statements from non-test files
from module import *
statements, addressed flake8 F403
- Uses American English spelling for code, as per PyTorch
- Code coverage now takes multiprocessing runs into account
- SmartCache with initial shuffling
ConvertToMultiChannelBasedOnBratsClasses
now supports channel-first inputs- Checkpoint handler to save with non-root permissions
- Fixed an issue for exiting the distributed unit tests
- Unified
DynUNet
to have single tensor output w/o deep supervision SegmentationSaver
now supports user-specified data types and asqueeze_end_dims
flag- Fixed
*Saver
event handlers output filenames with adata_root_dir
option - Load image functions now ensure little-endian
- Fixed the test runner to support regex-based test case matching
- Usability issues in the event handlers
0.4.0 - 2020-12-15
- Overview document for feature highlights in v0.4.0
- Torchscript support for the net modules
- New networks and layers:
- Discrete Gaussian kernels
- Hilbert transform and envelope detection
- Swish and mish activation
- Acti-norm-dropout block
- Upsampling layer
- Autoencoder, Variational autoencoder
- FCNet
- Support of initialisation from pretrained weights for densenet, senet, multichannel AHNet
- Layer-wise learning rate API
- New model metrics and event handlers based on occlusion sensitivity, confusion matrix, surface distance
- CAM/GradCAM/GradCAM++
- File format-agnostic image loader APIs with Nibabel, ITK readers
- Enhancements for dataset partition, cross-validation APIs
- New data APIs:
- LMDB-based caching dataset
- Cache-N-transforms dataset
- Iterable dataset
- Patch dataset
- Weekly PyPI release
- Fully compatible with PyTorch 1.7
- CI/CD enhancements:
- Skipping, speed up, fail fast, timed, quick tests
- Distributed training tests
- Performance profiling utilities
- New tutorials and demos:
- Autoencoder, VAE tutorial
- Cross-validation demo
- Model interpretability tutorial
- COVID-19 Lung CT segmentation challenge open-source baseline
- Threadbuffer demo
- Dataset partitioning tutorial
- Layer-wise learning rate demo
- MONAI Bootcamp 2020
- Base Docker image upgraded to
nvcr.io/nvidia/pytorch:20.10-py3
fromnvcr.io/nvidia/pytorch:20.08-py3
monai.apps.CVDecathlonDataset
is extended to a genericmonai.apps.CrossValidation
with andataset_cls
option- Cache dataset now requires a
monai.transforms.Compose
instance as the transform argument - Model checkpoint file name extensions changed from
.pth
to.pt
- Readers'
get_spatial_shape
returns a numpy array instead of list - Decoupled postprocessing steps such as
sigmoid
,to_onehot_y
,mutually_exclusive
,logit_thresh
from metrics and event handlers, the postprocessing steps should be used before calling the metrics methods ConfusionMatrixMetric
andDiceMetric
computation now returns an additionalnot_nans
flag to indicate valid resultsUpSample
optionalmode
now supports"deconv"
,"nontrainable"
,"pixelshuffle"
;interp_mode
is only used whenmode
is"nontrainable"
SegResNet
optionalupsample_mode
now supports"deconv"
,"nontrainable"
,"pixelshuffle"
monai.transforms.Compose
class inheritsmonai.transforms.Transform
- In
Rotate
,Rotated
,RandRotate
,RandRotated
transforms, theangle
related parameters are interpreted as angles in radians instead of degrees. SplitChannel
andSplitChanneld
moved fromtransforms.post
totransforms.utility
- Support of PyTorch 1.4
- Enhanced loss functions for stability and flexibility
- Sliding window inference memory and device issues
- Revised transforms:
- Normalize intensity datatype and normalizer types
- Padding modes for zoom
- Crop returns coordinates
- Select items transform
- Weighted patch sampling
- Option to keep aspect ratio for zoom
- Various CI/CD issues
0.3.0 - 2020-10-02
- Overview document for feature highlights in v0.3.0
- Automatic mixed precision support
- Multi-node, multi-GPU data parallel model training support
- 3 new evaluation metric functions
- 11 new network layers and blocks
- 6 new network architectures
- 14 new transforms, including an I/O adaptor
- Cross validation module for
DecathlonDataset
- Smart Cache module in dataset
monai.optimizers
modulemonai.csrc
module- Experimental feature of ImageReader using ITK, Nibabel, Numpy, Pillow (PIL Fork)
- Experimental feature of differentiable image resampling in C++/CUDA
- Ensemble evaluator module
- GAN trainer module
- Initial cross-platform CI environment for C++/CUDA code
- Code style enforcement now includes isort and clang-format
- Progress bar with tqdm
- Now fully compatible with PyTorch 1.6
- Base Docker image upgraded to
nvcr.io/nvidia/pytorch:20.08-py3
fromnvcr.io/nvidia/pytorch:20.03-py3
- Code contributions now require signing off on the Developer Certificate of Origin (DCO)
- Major work in type hinting finished
- Remote datasets migrated to Open Data on AWS
- Optionally depend on PyTorch-Ignite v0.4.2 instead of v0.3.0
- Optionally depend on torchvision, ITK
- Enhanced CI tests with 8 new testing environments
MONAI/examples
folder (relocated intoProject-MONAI/tutorials
)MONAI/research
folder (relocated toProject-MONAI/research-contributions
)
dense_patch_slices
incorrect indexing- Data type issue in
GeneralizedWassersteinDiceLoss
ZipDataset
return value inconsistenciessliding_window_inference
indexing anddevice
issues- importing monai modules may cause namespace pollution
- Random data splits issue in
DecathlonDataset
- Issue of randomising a
Compose
transform - Various issues in function type hints
- Typos in docstring and documentation
PersistentDataset
issue with existing file folder- Filename issue in the output writers
0.2.0 - 2020-07-02
- Overview document for feature highlights in v0.2.0
- Type hints and static type analysis support
MONAI/research
foldermonai.engine.workflow
APIs for supervised trainingmonai.inferers
APIs for validation and inference- 7 new tutorials and examples
- 3 new loss functions
- 4 new event handlers
- 8 new layers, blocks, and networks
- 12 new transforms, including post-processing transforms
monai.apps.datasets
APIs, includingMedNISTDataset
andDecathlonDataset
- Persistent caching,
ZipDataset
, andArrayDataset
inmonai.data
- Cross-platform CI tests supporting multiple Python versions
- Optional import mechanism
- Experimental features for third-party transforms integration
For more details please visit the project wiki
- Core modules now require numpy >= 1.17
- Categorized
monai.transforms
modules into crop and pad, intensity, IO, post-processing, spatial, and utility. - Most transforms are now implemented with PyTorch native APIs
- Code style enforcement and automated formatting workflows now use autopep8 and black
- Base Docker image upgraded to
nvcr.io/nvidia/pytorch:20.03-py3
fromnvcr.io/nvidia/pytorch:19.10-py3
- Enhanced local testing tools
- Documentation website domain changed to https://docs.monai.io
- Support of Python < 3.6
- Automatic installation of optional dependencies including pytorch-ignite, nibabel, tensorboard, pillow, scipy, scikit-image
- Various issues in type and argument names consistency
- Various issues in docstring and documentation site
- Various issues in unit and integration tests
- Various issues in examples and notebooks
0.1.0 - 2020-04-17
- Public alpha source code release under the Apache 2.0 license (highlights)
- Various tutorials and examples
- Medical image classification and segmentation workflows
- Spacing/orientation-aware preprocessing with CPU/GPU and caching
- Flexible workflows with PyTorch Ignite and Lightning
- Various GitHub Actions
- CI/CD pipelines via self-hosted runners
- Documentation publishing via readthedocs.org
- PyPI package publishing
- Contributing guidelines
- A project logo and badges