- avalanche.benchmarks.classic.SplitFMNIST(n_experiences: int, *, first_batch_with_half_classes: bool = False, return_task_id=False, seed: typing.Optional[int] = None, fixed_class_order: typing.Optional[typing.Sequence[int]] = None, shuffle: bool = True, train_transform: typing.Optional[typing.Any] = Compose( ToTensor() Normalize(mean=(0.286, ), std=(0.353, )) ), eval_transform: typing.Optional[typing.Any] = Compose( ToTensor() Normalize(mean=(0.286, ), std=(0.353, )) ), dataset_root: typing.Optional[typing.Union[str, pathlib.Path]] = None)
Creates a CL benchmark using the Fashion MNIST dataset.
If the dataset is not present in the computer, this method will automatically download and store it.
The returned benchmark will return experiences containing all patterns of a subset of classes, which means that each class is only seen “once”. This is one of the most common scenarios in the Continual Learning literature. Common names used in literature to describe this kind of scenario are “Class Incremental”, “New Classes”, etc. By default, an equal amount of classes will be assigned to each experience.
This generator doesn’t force a choice on the availability of task labels, a choice that is left to the user (see the return_task_id parameter for more info on task labels).
The benchmark instance returned by this method will have two fields, train_stream and test_stream, which can be iterated to obtain training and test
Experience. Each Experience contains the dataset and the associated task label.
The benchmark API is quite simple and is uniform across all benchmark generators. It is recommended to check the tutorial of the “benchmark” API, which contains usage examples ranging from “basic” to “advanced”.
n_experiences – The number of experiences in the current benchmark. If the first experience is a “pretraining” step and it contains half of the classes. The value of this parameter should be a divisor of 10 if first_task_with_half_classes if false, a divisor of 5 otherwise.
first_batch_with_half_classes – A boolean value that indicates if a first pretraining batch containing half of the classes should be used. If it’s True, a pretraining batch with half of the classes (5 for cifar100) is used. If this parameter is False no pretraining task will be used, and the dataset is simply split into a the number of experiences defined by the parameter n_experiences. Default to False.
return_task_id – if True, a progressive task id is returned for every experience. If False, all experiences will have a task ID of 0.
seed – A valid int used to initialize the random number generator. Can be None.
fixed_class_order – A list of class IDs used to define the class order. If None, value of
seedwill be used to define the class order. If non-None,
seedparameter will be ignored. Defaults to None.
shuffle – If true, the class order in the incremental experiences is randomly shuffled. Default to false.
train_transform – The transformation to apply to the training data, e.g. a random crop, a normalization or a concatenation of different transformations (see torchvision.transform documentation for a comprehensive list of possible transformations). If no transformation is passed, the default train transformation will be used.
eval_transform – The transformation to apply to the test data, e.g. a random crop, a normalization or a concatenation of different transformations (see torchvision.transform documentation for a comprehensive list of possible transformations). If no transformation is passed, the default test transformation will be used.
dataset_root – The root path of the dataset. Defaults to None, which means that the default location for ‘fashionmnist’ will be used.
A properly initialized