%matplotlib inline
from d2l import torch as d2l
from torch import nn
import torch
import torchvision
import os14.2 Fine-Tuning
In earlier chapters, we discussed how to train models on the Fashion-MNIST training dataset with only 60000 images. We also described ImageNet, the most widely used large-scale image dataset in academia, which has more than 10 million images and 1000 objects. However, the size of the dataset that we usually encounter is between those of the two datasets.
Suppose that we want to recognize different types of chairs from images, and then recommend purchase links to users. One possible method is to first identify 100 common chairs, take 1000 images of different angles for each chair, and then train a classification model on the collected image dataset. Although this chair dataset may be larger than the Fashion-MNIST dataset, the number of examples is still less than one-tenth of that in ImageNet. This may lead to overfitting of complicated models that are suitable for ImageNet on this chair dataset. Besides, due to the limited amount of training examples, the accuracy of the trained model may not meet practical requirements.
In order to address the above problems, an obvious solution is to collect more data. However, collecting and labeling data can take a lot of time and money. For example, in order to collect the ImageNet dataset, researchers have spent millions of dollars from research funding. Although the current data collection cost has been significantly reduced, this cost still cannot be ignored.
Another solution is to apply transfer learning to transfer the knowledge learned from the source dataset to the target dataset. For example, although most of the images in the ImageNet dataset have nothing to do with chairs, the model trained on this dataset may extract more general image features, which can help identify edges, textures, shapes, and object composition. These similar features may also be effective for recognizing chairs.
14.2.1 Steps
In this section, we will introduce a common technique in transfer learning: fine-tuning. As shown in Figure 14.2.1, fine-tuning consists of the following four steps:
- Pretrain a neural network model, i.e., the source model, on a source dataset (e.g., the ImageNet dataset).
- Create a new neural network model, i.e., the target model. This copies all model designs and their parameters on the source model except the output layer. We assume that these model parameters contain the knowledge learned from the source dataset and this knowledge will also be applicable to the target dataset. We also assume that the output layer of the source model is closely related to the labels of the source dataset; thus it is not used in the target model.
- Add an output layer to the target model, whose number of outputs is the number of categories in the target dataset. Then randomly initialize the model parameters of this layer.
- Train the target model on the target dataset, such as a chair dataset. The output layer will be trained from scratch, while the parameters of all the other layers are fine-tuned based on the parameters of the source model.
When target datasets are much smaller than source datasets, fine-tuning helps to improve models’ generalization ability.
14.2.2 Hot Dog Recognition
Let’s demonstrate fine-tuning via a concrete case: hot dog recognition. We will fine-tune a ResNet model on a small dataset, which was pretrained on the ImageNet dataset. This small dataset consists of thousands of images with and without hot dogs. We will use the fine-tuned model to recognize hot dogs from images.
%matplotlib inline
from d2l import jax as d2l
import tensorflow as tf
import numpy as np
import os%matplotlib inline
from d2l import mxnet as d2l
from mxnet import gluon, init, np, npx
from mxnet.gluon import nn
import os
npx.set_np()14.2.2.1 Reading the Dataset
The hot dog dataset we use was taken from online images. This dataset consists of 1400 positive-class images containing hot dogs, and as many negative-class images containing other foods. 1000 images of both classes are used for training and the rest are for testing.
After unzipping the downloaded dataset, we obtain two folders hotdog/train and hotdog/test. Both folders have hotdog and not-hotdog subfolders, either of which contains images of the corresponding class.
d2l.DATA_HUB['hotdog'] = (d2l.DATA_URL + 'hotdog.zip',
'fba480ffa8aa7e0febbb511d181409f899b9baa5')
data_dir = d2l.download_extract('hotdog')We create two instances to read all the image files in the training and testing datasets, respectively.
train_imgs = torchvision.datasets.ImageFolder(os.path.join(data_dir, 'train'))
test_imgs = torchvision.datasets.ImageFolder(os.path.join(data_dir, 'test'))# Load images as (PIL.Image, label) lists for compatibility with show_images
from PIL import Image as _PILImage
import pathlib
def _load_image_folder(path):
"""Load images from a directory with class subfolders, returning
a list of (PIL.Image, class_index) tuples."""
path = pathlib.Path(path)
class_names = sorted([p.name for p in path.iterdir() if p.is_dir()])
class_to_idx = {c: i for i, c in enumerate(class_names)}
items = []
for cls in class_names:
for img_path in sorted((path / cls).iterdir()):
try:
img = _PILImage.open(str(img_path)).convert('RGB')
items.append((img, class_to_idx[cls]))
except Exception:
continue
return items
train_imgs = _load_image_folder(os.path.join(data_dir, 'train'))
test_imgs = _load_image_folder(os.path.join(data_dir, 'test'))train_imgs = gluon.data.vision.ImageFolderDataset(
os.path.join(data_dir, 'train'))
test_imgs = gluon.data.vision.ImageFolderDataset(
os.path.join(data_dir, 'test'))The first 8 positive examples and the last 8 negative images are shown below. As you can see, the images vary in size and aspect ratio.
hotdogs = [train_imgs[i][0] for i in range(8)]
not_hotdogs = [train_imgs[-i - 1][0] for i in range(8)]
d2l.show_images(hotdogs + not_hotdogs, 2, 8, scale=1.4);
During training, we first crop a random area of random size and random aspect ratio from the image, and then scale this area to a \(224 \times 224\) input image. During testing, we scale both the height and width of an image to 256 pixels, and then crop a central \(224 \times 224\) area as input. In addition, for the three RGB (red, green, and blue) color channels we standardize their values channel by channel. Concretely, the mean value of a channel is subtracted from each value of that channel and then the result is divided by the standard deviation of that channel.
# Specify the means and standard deviations of the three RGB channels to
# standardize each channel
normalize = torchvision.transforms.Normalize(
[0.485, 0.456, 0.406], [0.229, 0.224, 0.225])
train_augs = torchvision.transforms.Compose([
torchvision.transforms.RandomResizedCrop(224),
torchvision.transforms.RandomHorizontalFlip(),
torchvision.transforms.ToTensor(),
normalize])
test_augs = torchvision.transforms.Compose([
torchvision.transforms.Resize([256, 256]),
torchvision.transforms.CenterCrop(224),
torchvision.transforms.ToTensor(),
normalize])# Image preprocessing using tf.keras for the JAX tab
IMG_SIZE = 224
IMAGENET_MEAN = np.array([0.485, 0.456, 0.406], dtype='float32')
IMAGENET_STD = np.array([0.229, 0.224, 0.225], dtype='float32')
def _imagenet_normalize(x):
return (x - IMAGENET_MEAN) / IMAGENET_STD
train_preprocess = tf.keras.Sequential([
tf.keras.layers.RandomCrop(IMG_SIZE, IMG_SIZE),
tf.keras.layers.RandomFlip('horizontal'),
tf.keras.layers.Rescaling(1./255),
tf.keras.layers.Lambda(_imagenet_normalize),
])
test_preprocess = tf.keras.Sequential([
tf.keras.layers.Resizing(256, 256),
tf.keras.layers.CenterCrop(IMG_SIZE, IMG_SIZE),
tf.keras.layers.Rescaling(1./255),
tf.keras.layers.Lambda(_imagenet_normalize),
])# Specify the means and standard deviations of the three RGB channels to
# standardize each channel
normalize = gluon.data.vision.transforms.Normalize(
[0.485, 0.456, 0.406], [0.229, 0.224, 0.225])
train_augs = gluon.data.vision.transforms.Compose([
gluon.data.vision.transforms.RandomResizedCrop(224),
gluon.data.vision.transforms.RandomFlipLeftRight(),
gluon.data.vision.transforms.ToTensor(),
normalize])
test_augs = gluon.data.vision.transforms.Compose([
gluon.data.vision.transforms.Resize(256),
gluon.data.vision.transforms.CenterCrop(224),
gluon.data.vision.transforms.ToTensor(),
normalize])14.2.2.2 Defining and Initializing the Model
We use ResNet-18, which was pretrained on the ImageNet dataset, as the source model. Here, we specify pretrained=True to automatically download the pretrained model parameters. If this model is used for the first time, Internet connection is required for download.
pretrained_net = torchvision.models.resnet18(
weights=torchvision.models.ResNet18_Weights.DEFAULT)# Load a pretrained ResNet50 from tf.keras.applications
pretrained_net = tf.keras.applications.ResNet50(weights='imagenet')pretrained_net = gluon.model_zoo.vision.resnet18_v2(pretrained=True)The pretrained source model instance contains a number of feature layers and an output layer fc. The main purpose of this division is to facilitate the fine-tuning of model parameters of all layers but the output layer. The member variable fc of source model is given below.
The pretrained source model from tf.keras.applications contains a number of feature layers and an output layer. We will use the pretrained weights for transfer learning. The final layer of the source model is shown below.
The pretrained source model instance contains two member variables: features and output. The former contains all layers of the model except the output layer, and the latter is the output layer of the model. The main purpose of this division is to facilitate the fine-tuning of model parameters of all layers but the output layer. The member variable output of source model is shown below.
pretrained_net.fcLinear(in_features=512, out_features=1000, bias=True)
pretrained_net.layers[-1]pretrained_net.outputDense(512 -> 1000, linear)
As a fully connected layer, it transforms ResNet’s final global average pooling outputs into 1000 class outputs of the ImageNet dataset. We then construct a new neural network as the target model. It is defined in the same way as the pretrained source model except that its number of outputs in the final layer is set to the number of classes in the target dataset (rather than 1000).
In the code below, the model parameters before the output layer of the target model instance finetune_net are initialized to model parameters of the corresponding layers from the source model. Since these model parameters were obtained via pretraining on ImageNet, they are effective. Therefore, we can only use a small learning rate to fine-tune such pretrained parameters. In contrast, model parameters in the output layer are randomly initialized and generally require a larger learning rate to be learned from scratch. Letting the base learning rate be \(\eta\), a learning rate of \(10\eta\) will be used to iterate the model parameters in the output layer.
finetune_net = torchvision.models.resnet18(
weights=torchvision.models.ResNet18_Weights.DEFAULT)
finetune_net.fc = nn.Linear(finetune_net.fc.in_features, 2)
nn.init.xavier_uniform_(finetune_net.fc.weight);# Build a fine-tuning model using TF/Keras: pretrained ResNet50 base + new head
base_model = tf.keras.applications.ResNet50(weights='imagenet',
include_top=False,
input_shape=(IMG_SIZE, IMG_SIZE, 3))
base_model.trainable = False # Freeze pretrained layers
finetune_net = tf.keras.Sequential([
base_model,
tf.keras.layers.GlobalAveragePooling2D(),
tf.keras.layers.Dense(2, kernel_initializer='glorot_uniform'),
])finetune_net = gluon.model_zoo.vision.resnet18_v2(classes=2)
finetune_net.features = pretrained_net.features
finetune_net.output.initialize(init.Xavier())
# The model parameters in the output layer will be iterated using a learning
# rate ten times greater
finetune_net.output.collect_params().setattr('lr_mult', 10)14.2.2.3 Fine-Tuning the Model
First, we define a training function train_fine_tuning that uses fine-tuning so it can be called multiple times.
# If `param_group=True`, the model parameters in the output layer will be
# updated using a learning rate ten times greater
def train_fine_tuning(net, learning_rate, batch_size=128, num_epochs=5,
param_group=True):
train_iter = torch.utils.data.DataLoader(torchvision.datasets.ImageFolder(
os.path.join(data_dir, 'train'), transform=train_augs),
batch_size=batch_size, shuffle=True)
test_iter = torch.utils.data.DataLoader(torchvision.datasets.ImageFolder(
os.path.join(data_dir, 'test'), transform=test_augs),
batch_size=batch_size)
devices = d2l.try_all_gpus()
loss = nn.CrossEntropyLoss(reduction="none")
if param_group:
params_1x = [param for name, param in net.named_parameters()
if name not in ["fc.weight", "fc.bias"]]
trainer = torch.optim.SGD([{'params': params_1x},
{'params': net.fc.parameters(),
'lr': learning_rate * 10}],
lr=learning_rate, weight_decay=0.001)
else:
trainer = torch.optim.SGD(net.parameters(), lr=learning_rate,
weight_decay=0.001)
d2l.train_ch13(net, train_iter, test_iter, loss, trainer, num_epochs,
devices)def _make_tf_dataset(img_dir, preprocess, batch_size, shuffle=False):
"""Create a tf.data.Dataset from an image folder directory."""
ds = tf.keras.utils.image_dataset_from_directory(
img_dir, label_mode='int', image_size=(256, 256),
batch_size=None, shuffle=shuffle)
ds = ds.map(lambda x, y: (preprocess(x, training=shuffle), y),
num_parallel_calls=tf.data.AUTOTUNE)
ds = ds.batch(batch_size).prefetch(tf.data.AUTOTUNE)
return ds
def train_fine_tuning(net, learning_rate, batch_size=128, num_epochs=5,
param_group=True):
train_ds = _make_tf_dataset(os.path.join(data_dir, 'train'),
train_preprocess, batch_size, shuffle=True)
test_ds = _make_tf_dataset(os.path.join(data_dir, 'test'),
test_preprocess, batch_size, shuffle=False)
# Compile with differential learning rates if param_group
if param_group:
# Use a smaller lr for the pretrained base, 10x for the new head
optimizer = tf.keras.optimizers.SGD(learning_rate=learning_rate * 10,
momentum=0.9)
# Freeze base so only head trains at the higher lr
net.layers[0].trainable = False
else:
optimizer = tf.keras.optimizers.SGD(learning_rate=learning_rate,
momentum=0.9)
net.compile(optimizer=optimizer,
loss=tf.keras.losses.SparseCategoricalCrossentropy(
from_logits=True),
metrics=['accuracy'])
history = net.fit(train_ds, validation_data=test_ds, epochs=num_epochs,
verbose=1)
train_loss = history.history['loss'][-1]
train_acc = history.history['accuracy'][-1]
test_acc = history.history['val_accuracy'][-1]
print(f'loss {train_loss:.3f}, train acc {train_acc:.3f}, '
f'test acc {test_acc:.3f}')def train_fine_tuning(net, learning_rate, batch_size=128, num_epochs=5):
train_iter = gluon.data.DataLoader(
train_imgs.transform_first(train_augs), batch_size, shuffle=True)
test_iter = gluon.data.DataLoader(
test_imgs.transform_first(test_augs), batch_size)
devices = d2l.try_all_gpus()
net.collect_params().reset_ctx(devices)
net.hybridize()
loss = gluon.loss.SoftmaxCrossEntropyLoss()
trainer = gluon.Trainer(net.collect_params(), 'sgd', {
'learning_rate': learning_rate, 'wd': 0.001})
d2l.train_ch13(net, train_iter, test_iter, loss, trainer, num_epochs,
devices)We set the base learning rate to a small value in order to fine-tune the model parameters obtained via pretraining. Based on the previous settings, we will train the output layer parameters of the target model from scratch using a learning rate ten times greater.
train_fine_tuning(finetune_net, 5e-5)loss 0.279, train acc 0.905, test acc 0.945
1578.4 examples/sec on [device(type='cuda', index=0)]
train_fine_tuning(finetune_net, 5e-5)train_fine_tuning(finetune_net, 0.01)loss 0.218, train acc 0.931, test acc 0.868
131.5 examples/sec on [gpu(0)]
For comparison, we define an identical model, but initialize all of its model parameters to random values. Since the entire model needs to be trained from scratch, we can use a larger learning rate.
scratch_net = torchvision.models.resnet18()
scratch_net.fc = nn.Linear(scratch_net.fc.in_features, 2)
train_fine_tuning(scratch_net, 5e-4, param_group=False)loss 0.378, train acc 0.838, test acc 0.840
1924.0 examples/sec on [device(type='cuda', index=0)]
# Train from scratch: same architecture but with random weights
scratch_base = tf.keras.applications.ResNet50(weights=None, include_top=False,
input_shape=(IMG_SIZE, IMG_SIZE, 3))
scratch_net = tf.keras.Sequential([
scratch_base,
tf.keras.layers.GlobalAveragePooling2D(),
tf.keras.layers.Dense(2, kernel_initializer='glorot_uniform'),
])
train_fine_tuning(scratch_net, 5e-4, param_group=False)scratch_net = gluon.model_zoo.vision.resnet18_v2(classes=2)
scratch_net.initialize(init=init.Xavier())
train_fine_tuning(scratch_net, 0.1)loss 0.345, train acc 0.852, test acc 0.835
133.7 examples/sec on [gpu(0)]
As we can see, the fine-tuned model tends to perform better for the same epoch because its initial parameter values are more effective.
14.2.3 Summary
- Transfer learning transfers knowledge learned from the source dataset to the target dataset. Fine-tuning is a common technique for transfer learning.
- The target model copies all model designs with their parameters from the source model except the output layer, and fine-tunes these parameters based on the target dataset. In contrast, the output layer of the target model needs to be trained from scratch.
- Generally, fine-tuning parameters uses a smaller learning rate, while training the output layer from scratch can use a larger learning rate.
14.2.4 Exercises
- Keep increasing the learning rate of
finetune_net. How does the accuracy of the model change? - Further adjust hyperparameters of
finetune_netandscratch_netin the comparative experiment. Do they still differ in accuracy? - Set the parameters before the output layer of
finetune_netto those of the source model and do not update them during training. How does the accuracy of the model change? You can use the following code.
for param in finetune_net.parameters():
param.requires_grad = Falsefinetune_net.layers[0].trainable = Falsefinetune_net.features.collect_params().setattr('grad_req', 'null')- In fact, there is a “hotdog” class in the
ImageNetdataset. Its corresponding weight parameter in the output layer can be obtained via the following code. How can we leverage this weight parameter?
weight = pretrained_net.fc.weight
hotdog_w = torch.split(weight.data, 1, dim=0)[934]
hotdog_w.shapetorch.Size([1, 512])
weight = pretrained_net.layers[-1].get_weights()[0] # Shape: (2048, 1000)
hotdog_w = weight[:, 934]
hotdog_w.shapeweight = pretrained_net.output.weight
hotdog_w = np.split(weight.data(), 1000, axis=0)[713]
hotdog_w.shape(1, 512)