U-net
The U-Net is a convolutional neural network designed for image segmentation, featuring a U-shaped architecture. It consists of an encoder (contracting path) to capture context and a decoder (expanding path) for precise localization. Skip connections bridge the encoder and decoder, ensuring spatial information is preserved.
Object Segmentation on ClincDB
For object segmentation, we are using U-Net model trained on ClincDB data. For this example, we are showing benchmarking of DL Backtrace (Default Mode), DL Backtrace (Contrastive Positive & Negative modes) & GradCam
Input
Prompt* String
*
number
(minimum: 0, maximum: 1)
Default: 0.7
*
number
(minimum: 0, maximum: 1)
Default: 0.7
*
integer
(maximum: 1)
Default: 0.7
Label:
Input:
Output
Object segmentation model prediction (ClinicDB)
DL Backtrace Explainability (ClincDB)
DL Backtrace has 2 modes - Default mode & Contrastive mode. In default model, the relevance is aggregated where as in contrastive mode, relevance can be both positive and negative.
Default Mode:
Contrastive Mode:
Positive Relevance:
Negative Relevance:
GradCAM Explainability (ClinicDB)
U-Net is a powerful deep learning architecture primarily designed for semantic segmentation tasks. It was introduced in the paper titled "U-Net: Convolutional Networks for Biomedical Image Segmentation" and has since become a standard model for various segmentation applications due to its efficiency and effectiveness in handling limited annotated data.Architecture OverviewThe U-Net architecture is characterized by its distinctive U-shaped structure, which consists of two main paths: the contracting path (encoder) and the expansive path (decoder).
Contracting Path (Encoder)
- The contracting path follows a typical convolutional network structure.
- It consists of repeated applications of:
- Two 3×33×3 convolutional layers (without padding).
- Each convolution is followed by a ReLU activation function.
- A 2×22×2 max pooling operation with a stride of 2 for downsampling.
- At each downsampling step, the number of feature channels is doubled, allowing the network to capture increasingly abstract representations of the input image.
Expansive Path (Decoder)
- The expansive path aims to upsample the feature maps and recover spatial information lost during downsampling.
- Each step in this path includes:
- An upsampling operation followed by a 2×22×2 convolution (often referred to as "up-convolution") that reduces the number of feature channels by half.
- A concatenation with the corresponding cropped feature map from the contracting path, which helps retain spatial details.
- Two 3×33×3 convolutional layers followed by ReLU activations.
Skip Connections
- One of the key innovations of U-Net is the use of skip connections between the encoder and decoder. These connections concatenate feature maps from the encoder to those in the decoder at corresponding levels, preserving spatial context and improving gradient flow during training. This design helps mitigate issues related to vanishing gradients and enhances segmentation accuracy.
Final Layer
- The final layer typically consists of a 1×11×1 convolution that maps each feature vector to the desired number of classes, providing a pixel-wise classification output.
Applications
U-Net has been widely adopted across various fields beyond medical imaging, including:
- Satellite image analysis
- Biological image segmentation
- Object detection in autonomous driving
- General image segmentation tasks
Its ability to perform well with limited training data makes it particularly valuable in domains where annotated datasets are scarce.
Summary
In summary, U-Net's unique architecture, which combines a contracting path for feature extraction with an expansive path for precise localization through skip connections, allows it to excel in semantic segmentation tasks. Its design not only enhances performance but also ensures that it can be trained effectively even with smaller datasets.
Is Explainability critical for your 'AI' solutions?
Schedule a demo with our team to understand how AryaXAI can make your mission-critical 'AI' acceptable and aligned with all your stakeholders.