Now showing 1 - 2 of 2
  • Publication
    Multi-modal image acquisition for AI-based bulky waste sorting (incl. terahertz synthetic aperture radar)
    ( 2023)
    Cibiraite-Lukenskiene, Dovile
    ;
    Gundacker, Dominik
    ;
    Bihler, M.
    ;
    Heizmann, M.
    ;
    ; ;
    Roming, Lukas
    ;
    ; ;
    This work presents the results of the initial acquisition of a multi-modal dataset that will be utilized to train and test a neural network for wood sorting. The aim of the project is to improve wood recycling from bulky waste by using four complementary sensing systems: visual, infrared, terahertz, and thermography. The four systems were combined to capture 57 multi-modal images of bulky waste samples moving on the conveyor belt at a speed of 10 cm/s. Early fusion results on THz show 0.77 accuracy, whereas the best multi-modal data fusion accuracy equals 0.921.
  • Publication
    Multi-sensor data fusion using deep learning for bulky waste image classification
    ( 2023)
    Bihler, Manuel
    ;
    Roming, Lukas
    ;
    Jiang, Yifan
    ;
    Afifi, Ahmed J.
    ;
    ;
    Cibiraite-Lukenskiene, Dovile
    ;
    Lorenz, Sandra
    ;
    Gloaguen, Richard
    ;
    ;
    Heizmann, Michael
    Deep learning techniques are commonly utilized to tackle various computer vision problems, including recognition, segmentation, and classification from RGB images. With the availability of a diverse range of sensors, industry-specific datasets are acquired to address specific challenges. These collected datasets have varied modalities, indicating that the images possess distinct channel numbers and pixel values that have different interpretations. Implementing deep learning methods to attain optimal outcomes on such multimodal data is a complicated procedure. To enhance the performance of classification tasks in this scenario, one feasible approach is to employ a data fusion technique. Data fusion aims to use all the available information from all sensors and integrate them to obtain an optimal outcome. This paper investigates early fusion, intermediate fusion, and late fusion in deep learning models for bulky waste image classification. For training and evaluation of the models, a multimodal dataset is used. The dataset consists of RGB, hyperspectral Near Infrared (NIR), Thermography, and Terahertz images of bulky waste. The results of this work show that multimodal sensor fusion can enhance classification accuracy compared to a single-sensor approach for the used dataset. Hereby, late fusion performed the best with an accuracy of 0.921 compared to intermediate and early fusion, on our test data.