Jiankun-chen / CVCMFFNet-master

We present a novel complex-valued convolutional and multi-feature fusion network (CVCMFF Net) specifically for building semantic segmentation of InSAR images. It can effectively segment the layover, shadow, and background on both the simulated InSAR building images and the real airborne InSAR images. The segmentation performance of CVCMFF Net is significantly improved compared with other state-of-the-art networks. The comparative experiments are link to https://github.com/Jiankun-chen/building-semantic-segmentation-of-InSAR-images

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

CVCMFF Net master


About the CVCMFF Net

We present a novel complex-valued convolutional and multi-feature fusion network (CVCMFF Net) specifically for building semantic segmentation of InSAR images. This CVCMFF Net not only learns from the complex-valued SAR images but also takes into account multi-scale and multi-channel feature fusion. It can effectively segment the layover, shadow, and background on both the simulated InSAR building images and the real airborne InSAR images.

About the Dataset

The simulated InSAR building dataset contains 312 simulated SAR image pairs generated from 39 different building models. Each building model is simulated at 8 viewing-angles. The sample number is 270 of the train set and is 42 of the test set. Each simulated InSAR sample contains three channels: master SAR image, slave SAR image, and interferometric phase image. This dataset serves the CVCMFF Net for building semantic segmentation of InSAR images.

For more information, please refer to the article: https://ieeexplore.ieee.org/document/9397870

Configuration of the ground truth label is: 0 = shadow; 1 = ground or single scattering; 2 = layover or secondary scattering.

You can get more information about the simulated InSAR building dataset from: https://ieee-dataport.org/documents/simulated-insar-building-dataset-cvcmff-net

Description of Source File

CVCMFFNet -- _init_.py : Functions loaded into the CVCMFF Net library
          -- count_para.py: Calculation of network's FLOPs and number of parameters
          -- evaluation.py: Calculation of loss, accuarcy, and IoU
          -- GetData.py: Read input image for placeholder
          -- helpers.py: Set up and design the picture to be displayed in tensorboard
          -- inference.py: Backbone of the CVCMFF Net
          -- inputs.py: Initialize placeholder
          -- layers.py: Upsampling function
          -- training.py: Scheduling the optimizer

  Data    -- Training: train set
             -- Images:
                -- real1: Real channel of master SAR image
                -- real2: Real channel of slave SAR image
                -- imag1: Imaginary channel of master SAR image
                -- imag2: Imaginary channel of slave SAR image
                -- ang: Interferometric phase image
             -- Labels: GT (one-hot)
          -- Test1: Simulated InSAR building images for test and "label" is blank.
          -- Test2: Real airborne InSAR images for test and "label" is blank.

  Output  -- model:
             -- Training: Output tf events in train
             -- Test: Output tf events in test
             checkpoint files -> data-index-meta

  Outtag:   Predicted segmentation pattern

  train.py: The main function to execute CVCMFF Net (train and test)

Requirements

GPU MEMORY >= 8GB

tensorflow-gpu >= 1.14

imageio >= 2.6.1

scipy >= 1.2.1

Train and Test

Please run train.py and set the parameter SAVE_INTERVAL in line 34. It is the number of iteration intervals to save the model during training. When the specified SAVE_INTERVAL is reached, the network will automatically extract samples for testing and synchronously output the current network's performance on both the train set and the test set. So we can monitor the network status dynamically. We can also view the curves and predicted patterns on tensorboard.

The curves (orange for TRAIN and blue for TEST) of OA and LOSS versus training epoch are:

The curves (orange for TRAIN and blue for TEST) of Iou_class0, Iou_class1, and Iou_class2 versus training epoch are:

Experimental Results

The experimental curve are in the folder ./Curve: accuracy.svg; loss.svg; iou-class0.svg; iou-class1.svg; iou-class2.svg;

The Predicted segmentation pattern are in folders: ./outtag/result1-> simulated InSAR building images; ./outtag/result2-> real airborne InSAR images;

Segmentation results on the simulated InSAR building dataset:

Segmentation results on the real airborne InSAR images:

The segmentation performance of CVCMFF Net is significantly improved compared with those of other state-of-the-art networks, referring to: https://github.com/Jiankun-chen/building-semantic-segmentation-of-InSAR-images

Feature Visualization

You can visualize feature maps in train.py by turning on line 50 while commenting out line 52, turning on line 110 while commenting out line 111, and turning on line 139-170.

Model Evaluation

When the model reaches an absolute accuracy, additional evaluation indexes are needed to evaluate the model:

  • Floating-point operations (FLOPs), which can be understood as a calculation amount required for forward propagation and reflects the hardware’s performance requirements such as GPU;
  • Number of parameters describe model complexity;
  • Model size reflects the amount of memory occupied;
  • Model speed is used to measure the time required for a model to train or test a single image;
  • Training time evaluates the convergence speed of the model in practical engineering applications.

The above indicators have an essential reference for model implementation and are shown in Table:

Author: Jiankun Chen

This version was updated on 4/18/2021 0:44:21 PM

About

We present a novel complex-valued convolutional and multi-feature fusion network (CVCMFF Net) specifically for building semantic segmentation of InSAR images. It can effectively segment the layover, shadow, and background on both the simulated InSAR building images and the real airborne InSAR images. The segmentation performance of CVCMFF Net is significantly improved compared with other state-of-the-art networks. The comparative experiments are link to https://github.com/Jiankun-chen/building-semantic-segmentation-of-InSAR-images


Languages

Language:Python 100.0%