A pruned VGG19 model subjected to Axial Coronal Sagittal (ACS) convolutions and a custom VGG16 model are benchmarked to predict 3D fabric descriptors from a set of 2D images. The data used for training and testing are extracted from a set of 600 3D biphase microstructures created numerically. Fabric descriptors calculated from the 3D microstructures constitute the ground truth, while the input data are obtained by slicing the 3D microstructures in each direction of space at regular intervals. The computational cost to train the custom ACS-VGG19 model increases linearly with p (the number of images extracted in each direction of space), and increasing p does not improve the performance of the model - or only does so marginally. The best performing ACS-VGG19 model provides a MAPE of 2 to 5% for the means of aggregate size, aspect ratios and solidity, but cannot be used to estimate orientations. The custom VGG16 yields a MAPE of 2% or less for the means of aggregate size, distance to nearest neighbor, aspect ratios and solidity. The MAPE is less than 3% for the mean roundness, and in the range of 5-7% for the aggregate volume fraction and the mean diagonal components of the orientation matrix. Increasing p improves the performance of the custom VGG16 model, but becomes cost ineffective beyond 3 images per direction. For both models, the aggregate volume fraction is predicted with less accuracy than higher order descriptors, which is attributed to the bias given by the loss function towards highly-correlated descriptors. Both models perform better to predict means than standard deviations, which are noisy quantities. The custom VGG16 model performs better than the pruned version of the ACS-VGG19 model, likely because it contains 3 times (p = 1) to 28 times (p = 10) less parameters than the ACS-VGG19 model, allowing better and faster cnvergence, with less data. The custom VGG16 model predicts the second and third invariants of the orientation matrix with a MAPE of 2.8% and 8.9%, respectively, which suggests that the model can predict orientation descriptors regardless of the orientation of the input images.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11026488PMC
http://dx.doi.org/10.1038/s41598-024-59554-xDOI Listing

Publication Analysis

Top Keywords

custom vgg16
20
vgg16 model
16
acs-vgg19 model
16
model
11
fabric descriptors
8
direction space
8
mape aggregate
8
aggregate size
8
aspect ratios
8
ratios solidity
8

Similar Publications

Smart farming is a hot research area for experts globally to fulfill the soaring demand for food. Automated approaches, based on convolutional neural networks (CNN), for crop disease identification, weed classification, and monitoring have substantially helped increase crop yields. Plant diseases and pests are posing a significant danger to the health of plants, thus causing a reduction in crop production.

View Article and Find Full Text PDF

A Combined CNN-LSTM Network for Ship Classification on SAR Images.

Sensors (Basel)

December 2024

ENSTA Bretagne, Lab-STICC, UMR CNRS 6285, 29806 Brest, France.

Satellite SAR (synthetic aperture radar) imagery offers global coverage and all-weather recording capabilities, making it valuable for applications like remote sensing and maritime surveillance. However, its use in machine learning-based automatic target classification faces challenges, including the limited availability of SAR target training samples and the inherent constraints of SAR images, which provide less detailed features compared to natural images. These issues hinder the effective training of convolutional neural networks (CNNs) and complicate the transfer learning process due to the distinct imaging mechanisms of SAR and natural images.

View Article and Find Full Text PDF

With the advancement of artificial intelligence technology, unmanned boats utilizing deep learning models have shown significant potential in water surface garbage classification. This study employs Convolutional Neural Network (CNN) to extract features of water surface floating objects and constructs the VGG16-15 model based on the VGG-16 architecture, capable of identifying 15 common types of water surface floatables. A garbage classification dataset was curated to obtain 5707 images belonging to 15 categories, which were then split into training and validation sets in a 4:1 ratio.

View Article and Find Full Text PDF

Accurate classification of logos is a challenging task in image recognition due to variations in logo size, orientation, and background complexity. Deep learning models, such as VGG16, have demonstrated promising results in handling such tasks. However, their performance is highly dependent on optimal hyperparameter settings, whose fine-tuning is both labor-intensive and time-consuming.

View Article and Find Full Text PDF

EA-CNN: Enhanced attention-CNN with explainable AI for fruit and vegetable classification.

Heliyon

December 2024

Future Technology Research Center, National Yunlin University of Science and Technology, 123 University Road, Section 3, Douliu, Yunlin, 64002, Taiwan, ROC.

The quality of vegetables and fruits are judged by their visual features. Misclassification of fruits and vegetables lead to a financial loss. To prevent the loss, superstores need to classify fruits and vegetables in terms of size, color and shape.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!