Downloading datasets with fast ai
Creation of a Dataloaders object
bears = DataBlock( blocks=(ImageBlock, CategoryBlock), get_items=get_image_files, splitter=RandomSplitter(valid_pct=0.2, seed=42), get_y=parent_label, item_tfms=Resize(128))
blocks=(ImageBlock, CategoryBlock)
get_items=get_image_files
splitter=RandomSplitter(valid_pct=0.2, seed=42),
get_y=parent_label
item_tfms=Resize(128))
The above is a simple item transformation on the images to resize them to 128x122 pixels.
The model is not fed a singular image at a time, but tehy are batched together in an array(knowns as a tensor)
dls = bears.dataloaders(path)
dls.valid.show_batch(max_n=4, nrows=1)
Methods for resizing
bears = bears.new(item_tfms=Resize(128, ResizeMethod.Squish))
bears = bears.new(item_tfms=Resize(128, ResizeMethod.Pad, pad_mode='zeros'))
Dealing with images of different sizes
Instead of squishing images down or creating wasted space for computation you can instead randomly select an area of the image and crop to that part.
Different features will be highlighted because of this.
bears = bears.new(item_tfms=RandomResizedCrop(128, min_scale=0.3)) dls.train.show_batch(max_n=4, nrows=1, unique=True)
Data augmenentation
Data augmentation is almost like creating a more robust model by stressing it. You augment your images in order to improve their ability to pick out the correct features and not find workarounds.
Some examples include
- Rotation
- Flipping
- Perspective warping
- Brightness changes
- Contrast Changes
bears = bears.new(item_tfms=Resize(128), batch_tfms=aug_transforms(mult=2)) dls.train.show_batch(max_n=4, nrows=1, unique=True)
Fuller code example
bears = bears.new( item_tfms=RandomResizedCrop(224, min_scale=0.5), batch_tfms=aug_transforms()) dls = bears.dataloaders(path) learn = vision_learner(dls, resnet18, metrics=error_rate) learn.fine_tune(4)
This will now train the model to attempt to categorize grizzly bears, black bears and teddy bears.
Interpretation
After you have trained the model and gotten it to clasify the images you can have a look at how it is doing. We can get a look at what it is doing when predicting and then infer something about hwo it works from the results.
interp = ClassificationInterpretation.from_learner(learn) interp.plot_confusion_matrix()
You can look at where the greatest amount of loss is occuring.
interp.plot_top_losses(5, nrows=1)
Cleaning up your data set
After looking at the data with highest loss, you might then realise that some of the images are corrupt or need to be re-labelled. Fastai has additional tools to help in this process.
cleaner = ImageClassifierCleaner(learn) cleaner
When we ‘delete’ these images we are just apssing them to anothe array. So we have to actually delete them.
for idx in cleaner.delete(): cleaner.fns[idx].unlink()
Or move them to a different category or label.
for idx,cat in cleaner.change(): shutil.move(str(cleaner.fns[idx]), path/cat)