Skip to main content

Gen-SIS: Generative Self-augmentation Improves Self-supervised Learning

Preprint

Varun Belagali*1, Srikar Yellapragada*1, Alexandros Graikos1, Saarthak Kapse1,
Zilinghan Li2, Tarak Nath Nandi2,3, Ravi K Madduri2,3,
Prateek Prasanna1, Joel Saltz1, Dimitris Samaras1

1Stony Brook University, 2Argonne National Laboratory, 3University of Chicago

TL;DR

Self-supervised learning (SSL) learns visual representations by training an image encoder to maximize similarity between features of different views of the same image. While SoTA SSL algorithms rely on hand-crafted augmentations (cropping, color jittering), recently, generative diffusion models have been shown to improve SSL by providing a wider range of generative data augmentations [1]. The downside is that this usually requires pre-training the diffusion model on large-scale image-text datasets, which might not be available for many specialized domains like histopathology.

To solve this, we introduce Gen-SIS, a diffusion-based augmentation technique trained exclusively on unlabeled image data. We train an initial SSL encoder using only hand-crafted augmentations. We then use this encoder to train an embedding-conditioned diffusion model. The embedding-conditioned diffusion model can synthesize diverse views of an image as well as interpolate between images in the encoder latent space. We show that these self-augmentations, i.e. generative augmentations based on the vanilla SSL encoder embeddings, facilitate the training of a stronger SSL encoder which we complement with our novel distantanglement pretext task. We showcase Gen-SIS's effectiveness by demonstrating performance improvements in both natural images and digital histopathology.

Method

We propose two types of self-augmentations:

  • Generative Augmentations: In the generative augmentation setting, a synthetic image is generated using a single real image as the source. This involves first extracting an embedding ee from the source image using the conditioning-encoder, and then guiding the image generation process with that embedding to create a synthetic image Is = E-LDM(z,e). Generative augmentations introduce novel variations in the shape, size, and position of objects, as well as changes in the background, while preserving the semantic content of the objects in the image.
  • Interpolated Augmentations: An interesting property of diffusion models is their ability to generate an image that partially resembles each source image when conditioned on embeddings interpolated from two sources. We leverage this property to produce an interpolated synthetic image from two real source images, which we use to perform a new pretext task during the SSL training. With embeddings e1 and e2 representing the two source images (I1, I2), and an interpolation ratio α, we compute an interpolated embedding eint using spherical linear interpolation (SLERP), eint = SLERP(e1, e2, α). We choose SLERP over linear interpolation since high-dimensional vectors are concentrated near the surface of the unit sphere. This interpolated embedding serves as the conditioning to generate the synthetic interpolated image, Iint = E-LDM(eint).

Training

We use DINO as our vanilla SSL method. To integrate generative augmentations into SSL, we use the real image and a corresponding synthetic image as an input pair for the SSL pretext task. We also apply hand-crafted augmentations to both real and synthetic images.

Since the interpolated image contains components of both source images, we propose a disentanglement task where the network learns to separate the distinct features of each source image used in the interpolation.

PyTorch-style pseudocode for the disentanglement pretext task
# Parameters:
# gs, gt: student and teacher networks
# tps, tpt: student and teacher temperatures
# c: center
# alpha: interpolation ratio

# Load an image from the dataset
for img_1 in loader:
# Read secondary source image -- Another random image from the dataset
img_2 = ReadImage(secondary(img_1))
# Read interpolated image of primary and secondary source image
# Can be generated on-the-fly but we pre-generate to reduce training time
img_int = ReadInterpImage(img_1, img_2, alpha)

# Apply vanilla DINO augmentation to get a view of the primary image
img_1_view = vanilla_augment(img_1)
# Apply vanilla DINO augmentation to get a view of secondary image
img_2_view = vanilla_augment(img_2)
# Apply vanilla DINO augmentation to get a view of the interpolated image
img_int_view = vanilla_augment(img_int)

# Get student output for interpolated image
stud_int = gs(img_int_view)
# Get teacher output for primary and secondary images
teach_1 = gt(img_1_view).detach()
teach_2 = gt(img_2_view).detach()

# Student sharpening
stud_int = softmax(stud_int / tps, dim=1)
# Entanglement of teacher output
teach_ent = alpha * teach_1 + (1-alpha) * teach_2
# Teacher sharpening and centering
teach_ent = softmax((teach_ent - c) / tpt, dim=1)

# Compute disentanglement loss
disentanglement_loss = - (teach_ent * log(stud_int)).sum(dim=1).mean()

Results

For our enhanced SSL training, we improve DINO with the Gen-SIS framework and call the method Gen-DINO. In Gen-DINO, we pre-train the ViT-S/16 model with generative and interpolated augmentations.