Have a personal or library account? Click to login
Deep Learning-Based Computer Vision Is Not Yet the Answer to Taphonomic Equifinality in Bone Surface Modifications Cover

Deep Learning-Based Computer Vision Is Not Yet the Answer to Taphonomic Equifinality in Bone Surface Modifications

Open Access
|Dec 2024

Figures & Tables

Table 1

Description of the DL-based CV algorithms experimented with in the present study. This includes a summary of the number of parameters (millions) in each algorithm if trained from scratch, the number of convolutional layers (Nº Conv), whether Batch Normalisation (B.N.) layers are included in the algorithm, and whether pretrained weights are available in TensorFlow for the purpose of transfer learning. The DS columns indicate which of the three datasets in this study were originally used to train algorithms by Abellá, Baquedano & Domínguez-Rodrigo (2022) (DS1), Domínguez-Rodrigo et al., (2020) (DS2), Pizarro-Monzo et al. (2023) (DS3). The reference list refers to the original publication of each of the datasets. * This number includes all of the convolutional layers included within the inception modules – large blocks of multiple convolutional layers. ** A precise number cannot be reported as the architecture works differently from the others.

ALGORITHMPARAMS.Nº CONVB.N.PRETRAINEDDSREFERENCES
123
Jason1≈ 3 Mil.8FalseFalsexBrownlee, 2017
Jason2≈ 3 Mil.8TrueFalsexxBrownlee, 2017
VGG16≈ 15 Mil.13FalseTruexSimonyan & Zisserman, 2015
DenseNet201≈ 18 Mil.201TrueTruexxxHuang et al., 2017
VGG19≈ 20 Mil.16FalseTruexxSimonyan & Zisserman, 2015
InceptionV3≈ 22 Mil.96 *TrueTruexSzegedy et al., 2015
ResNet50≈ 24 Mil.49TrueTruexxxHe et al., 2016
Alexnet≈ 35 Mil.5TrueFalsexKrizhevsky et al., 2012
EfficientNetB7≈ 64 Mil.**TrueTruexTan and Le, 2020
jcaa-7-1-145-g1.png
Figure 1

Examples of ideal train/validation learning curves. These curves were obtained from a neural network trained on a toy dataset, alongside examples of both underfitting and overfitting neural networks.

Table 2

Descriptive statistics of the different metrics obtained from analysed images. Descriptive statistics of the different metrics extracted from images from the different classes present in DS1 and DS3; DS2 was excluded from the table as it is contained within DS3. Metrics include the Laplacian of Gaussian (LoG) variance, percentage of image presenting detectable features using Canny Edge Detection (CED), Fast Fourier Transform (FFT) magnitudes, the percentage of images presenting adequate levels of contrast, and the percentage of images presenting complications due to the presence of specularities (Spec.). Descriptive statistics report the central tendency followed by 95% confidence intervals constructed using distribution quantiles. CM = Cut Mark, Croc. = Crocodylian tooth score, TM = Carnivoran Tooth Mark, Tmp = Trampling.

SAMPLELOG VARIANCEFFT MAGNITUDECED (%)CONTRAST (%)SPEC. (%)
DS1-CM20.6 +/– [13.8, 55.6]9.1 +/– [0.6, 24.1]33.4 +/– [16.3, 56.7]45.011.3
DS1-Croc.71.0 +/– [17.8, 230.8]19.0 +/– [5.7, 30.7]49.9 +/– [26.8, 65.7]95.758.7
DS3-CM22.1 +/– [14.1, 70.2]8.7 +/– [0.7, 22.9]35.7 +/– [16.6, 66.3]29.69.8
DS3-TM41.9 +/– [13.6, 133.3]16.2 +/– [–0.2, 28.0]48.3 +/– [14.6, 68.2]80.962.1
DS3-Tmp44.1 +/– [10.2, 378.4]9.7 +/– [–6.7, 93.2]45.2 +/– [2.7, 87.8]47.340.0
jcaa-7-1-145-g2.jpg
Figure 2

Examples of images displaying exceptionally poor quality. Examples of photographs of cut marks from DS1 and DS3 presenting especially poor image quality, with a considerable portion of pixels out-of-focus towards the image border. Sobel gradient maps in the right-hand panels highlight these features with sharp changes in gradient being clearly visible in the centre of each image, while gradients towards the edges present a high degree of out-of-focus blur with almost no detectable features.

jcaa-7-1-145-g3.jpg
Figure 3

Examples of photographs presenting specular reflections. Examples of photographs of tooth marks from DS1 and DS3, presenting areas of abnormally intense brightness in certain pixels as a product of specular reflections. Panels on the right present pixels where these abnormalities have been detected.

Table 3

Evaluation metrics when predicting the different types of BSMs using the models trained on each dataset and evaluated on the test set. Note that a high accuracy value does not necessarily imply positive performance, as evidenced by the (imbalanced) DS1 dataset.

DS1DS2DS3
Precision0.460.770.88
Recall0.500.690.87
F10.480.660.88
Accuracy0.920.860.91
Table 4

Error rates (in %) when predicting the different types of BSMs using the models trained on each dataset and evaluated on the test sets. Error rates are reported as the RMSE of the labels.

DS1DS2DS3
Tooth Score15.347.95
Trampling55.3022.35
Cut Mark5.2910.147.90
Crocodile24.41
Overall8.6315.139.77
Table 5

Confusion matrix obtained when evaluating a Jason2 model trained on the DS1 test set. Note that the confusion matrix presents a true positive rate of 0; the algorithm classifies all samples as cut marks regardless of whether they are or not.

PREDICTED
CROCODILECUT MARK
TrueCrocodile013
Cut Mark0146
jcaa-7-1-145-g4.png
Figure 4

Empirical learning curves for neural networks. Learning curves obtained from the best performing convolutional neural network architectures on each of the datasets; Jason2 for DS1, VGG16 for DS2, and DenseNet201 for DS3.

Table 6

Confusion matrix obtained when evaluating VGG16 on the test set of DS2 and DenseNet201 on the test set of DS3.

PREDICTED DS2PREDICTED DS3
CUT MARKSCORETRAMPLINGCUT MARKSCORETRAMPLING
TrueCut Mark13411116326
Score228071263
Trampling113411133
jcaa-7-1-145-g5.jpg
Figure 5

Grad-CAM results displaying suboptimal detection of features. Grad-CAM results for a selection of images displaying particularly poor identification of relevant features for BSM classification. The lighter shades of yellow highlight areas where the CNN are identifying notable features. Darker areas leaning more towards blue indicate areas that are not of interest to the CNN when identifying each type of BSM.

DOI: https://doi.org/10.5334/jcaa.145 | Journal eISSN: 2514-8362
Language: English
Submitted on: Dec 20, 2023
Accepted on: Oct 10, 2024
Published on: Dec 10, 2024
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2024 Lloyd Austin Courtenay, Nicolas Vanderesse, Luc Doyon, Antoine Souron, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.