This repository contains a generator for Keras (using a channel last configuration, eg with TensorFlow backend), capable of handling 2D and 3D arrays, and augmenting the data using affine transforms and elastics deformations while being based on keras.utils.Sequence for efficient and safe multiprocessing.
The requirements for the generator and augmentation are the following: numpy, matplotlib, scipy, and keras (only for the generator part)
The demo requires on top: SimpleITK and pydicom
In order to use mulitprocessing, you need to specify it in your call to fit_generator:
model.fit_generator(generator=generator,
use_multiprocessing=True,
workers=8)
use_multiprocessing should be True, otherwise it defaults to False and you'll be using only one thread. The number of workers (ie thread) is also to be choosen (1 is the default value), if you have no idea pick something between your CPU core count and thread count.
-
new_generator.py contains the generator class (it's built for generating images and their corresponding segmentation mask, but can be adapted to only work on the images without mask)
-
image_augmentation.py contains the image augmentation functions (which can act on either image, or image,mask )
-
demo.py contains a demo script loading dicom files and deforming them
-
demo contains demo images and output from the demo script
is used to supply the generator with a lot of useful informations regarding the data to be generated. The following list explains the parameters useful for data generation:
- dataset: the name of the dataset used
- augmentation: a tuple of booleans values selecting the augmentations techniques to apply in the following order: random_transform, deform_grid, deform_pixel
- random_deform: a dictionnary containing the parameters to be passed to random_transform
- only: to return only the "im" or the "mask", useful for training autoencoders, set it to None otherwise
- e_deform_g: a dictionnary containing the parameters to be passed to the chosen elastic deformation function, the deformation is computed from a grid (see section below on the augmentation techniques)
- e_deform_p: a dictionnary containing the parameters to be passed to the chosen elastic deformation function, the deformation is computed independtly, pixel-wise (see section below on the augmentation techniques)
- shape: the input shape of the network, ie the shape of the images including the channel
- ReadFunction: a function to read the images, which given an ID and if it should query the corresponding image or mask, returns a npy array (the ID can be any object, eg a string identifying the file or a list with the filename and the coordiantes of a patch to take from it and so on...)
- PreProcessing: a preprocessing function, which takes as input the image, the size of the desired output image and, optionnally if the image is a mask or not (in order to apply different preprocessing) and return the preprocessed image
- savefolder: the path of the folder to which the examples are going to be saved
The file contains 2 augmentation "techniques": - elastic deformation with 2 different methods: deform_pixel on a pixel wide basis and deform_grid on a grid basis. - random_transform, which provides most of the keras image augmentation techniques.
These 3 functions take as input X (the image), Y (an optionnal mask), and some keyed parameters. They also work both on 2D and 3D images. They depend on numpy and scipy.ndimage Elastic deformation is quite slow for 3D images, one could try to tune the order of the splines used for the different interpolations.
The following parameters can be supplied in the random_transform dictionnary:
- rotation_range_alpha = angle in degrees (0 to 180), produces a range in which to uniformly pick the rotation in x.
- rotation_range_beta = angle in degrees (0 to 180), produces a range in which to uniformly pick the rotation in y.
- rotation_range_gamma = angle in degrees (0 to 180), produces a range in which to uniformly pick the rotation in z.
- width_shift_range: fraction of total width, produces a range in which to uniformly pick the shift.
- height_shift_range: fraction of total height, produces a range in which to uniformly pick the shift.
- depth_shift_range: fraction of total depth, produces a range in which to uniformly pick the shift.
- zoom_range = factor of zoom. A zoom factor per axis will be randomly picked in the range [a, b].
- horizontal_flip: boolean, whether to randomly flip images horizontally.
- vertical_flip: boolean, whether to randomly flip images vertically.
- z_flip: boolean, whether to randomly flip images along the z axis.
for grid based deformation, name the dictionnary "e_deform_g":
- sigma = standard deviation of the normal distribution
- points = number of points of the each side of the square grid
for pixel based deformation, name the dictionnary "e_deform_p":
- alpha = scaling factor the deformation
- sigma = smooting factor
The function save_image can be used to save the predicted images, see the save_predict function in demo.py
A few examples are present in the demo folder extracted from the 3D-IRCADb 01 dataset (https://www.ircad.fr/research/3d-ircadb-01/)
'augmentation': [1, 0, 1], 'random_deform': {'rotation_range_alpha': 20, 'width_shift_range': 0.1, 'height_shift_range': 0.1}, 'e_deform': {'points': 3, 'sigma': 10}
'augmentation': [1, 1, 0], 'random_deform': {'height_shift_range': 0.1, 'width_shift_range': 0.1, 'rotation_range_alpha': 20}, 'e_deform': {'sigma': 3, 'alpha': 10}