Machine learning may significantly support forensic medicine, particularly in age estimation through medical imaging soon. This technology offers great potential for supporting decisions especially when age documentation is missing or disputed.
Objective
This study investigates the potential of generative models for forensic age estimation. The focus is on addressing the challenges of interpretability and generalizability commonly faced by traditional discriminative models.
Methods
We applied a family of generative models to postmortem computed tomography (PMCT) scans, focusing on the ossification of the medial clavicular epiphysis for age prediction. The latent space representations from these models were analyzed for their ability to predict age accurately and interpretably across different datasets.
Results
While the methods did not perform as well as discriminative state of the art approaches using German Working Group for Forensic Age Diagnostics (AGFAD) guidelines, the variational autoencoders were able to learn a meaningful latent space. The age of the participants could even be visualized within a two-dimensional projection. Additionally, the re-use of the learned space led to high performance on a smaller dataset collected from a forensic center.
Conclusion
The consideration of “soft factors”, such as explainability in addition to absolute performance remains crucial for bringing machine learning methods into forensic practice. Depending on the set-up, generative models might be attractive for assessing the reasoning within models and sharing information between datasets.
Hinweise
×
Scan QR code & read article online
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Introduction
Similar to other parts of our society, machine learning has emerged as a popular tool within different areas of forensic medicine and will soon fuel more and more research and practice niches of our disciplines. Given the rapid advances, the technology has the potential to crucially enhance the accuracy of findings from the perspective of the law and support the complex processes resulting from regulatory issues. By now, exemplified surveys, such as the detection of causes of death [3] or gender and age identification [10] have highlighted the potential areas within legal medicine and the inherent limitations of the technology.
One application considered in this article is the technical support of forensic age estimation, a critical task that assists legal decisions by estimating an individual’s minimum age primarily based on medical imagery combined with a physical examination according to the recent guidelines of the Study Group on Forensic Age Diagnostics (AGFAD) of the German Society of Legal Medicine [15].
Forensic age estimation is essential in legal and immigration contexts, especially when documentation of an individual’s biological age is unavailable or disputed. Among the anatomical markers, the medial clavicular epiphysis (MCE) is frequently assessed due to its predictable ossification stages, which continue into early adulthood allowing discriminative landmarks of the 18 years and 21 years of age thresholds [19]. This makes the MCE particularly valuable for distinguishing between minors and adults. Imaging techniques such as computed tomography (CT) scans are commonly used in this process, providing clear visualization of the MCE developmental stage and providing a reliable basis for forensic age estimation [14].
Just within 2024, at least two published contributions utilizing deep learning models for continuous age assessment based on CT scans have shown the potentials and its scientific importance. Wesp et al. [18] utilized a well-established convolutional neural network (CNN) based on the publicly available ResNet-18 and trained on 4400 CT scans. Their approach included isolating regions of interest (ROI) around the MCE cartilage and employing this model to effectively perform segmentation tasks. With this architecture, they reported a mean absolute error (MAE) of 1.65 years (± 0.53 years) on testing data derived from the same source as their training data. Similarly, Qiu et al. [13] demonstrated additional evidence for the efficacy of CNNs and regression techniques, utilizing 1049 CT scans. This study also incorporated ROI isolation and the application of the U‑Net architecture for precise segmentation. They achieved a MAE of 1.73 years (± 1.28 years). Especially given the fact that these examinations are normally only part of a full assessment and could be further enriched, the obtained predictions appear quite impressive.
Despite these advances, the application of these proposed discriminative machine learning models in forensic medicine is accompanied by significant challenges. One primary issue is the interpretability of these models, often criticized as “black boxes” due to their opaque decision-making processes. Utilizing a notion of probability, the published discriminative models learn to predict the age given the data: \(p(\mathrm{Age}| \text{Image})\). While they excel at this task, they are not optimized for “explaining” the reasoning. Accordingly, interpretability is limited and commonly conducted through post hoc analyses. This bears dangers, as the generalizability of machine learning models is frequently overestimated and may lead to overconfidence in their performance on unseen data.
Given these challenges, this article explores the proposition of using so-called generative models for forensic age estimation tasks as an alternative to the discriminative models previously published [18]. The generative models learn the underlying distribution of the data rather than focusing only on high-dimensional decision boundaries. This shift offers several advantages [8]: the family of methods can produce synthetic samples, which can be employed to test and evaluate the robustness of the approach by effectively acting as a diagnostic tool. Moreover, some of these methods, which we will focus on in this work, provide access to their lower dimensional latent space, which contains the learned representation of the complex image modalities. By analyzing those representations, one may easily compare the similarity between hardly comparable entities. A natural application is the identification of outliers, thereby signalling potential distribution shifts and ensuring model reliability. Finally, the latent representations learned by generative models can serve as valuable inputs for subsequent tasks, such as the regression of age, which may then make methods explainable and verifiable.
This work aims to examine the potential of generative models and their representations in addressing the inherent challenges of machine learning applications in forensic age estimation using CT scans. The utilization of deep learning as a powerful feature extractor for learning representations might provide a pathway towards more interpretable and reliable application of machine learning. While this work faces certain challenges in the data basis, such as a limited sample size, it remains a promising feasibility study with significant potential. The limited data availability is primarily due to natural mortality: young individuals, the focus of this study, rarely pass away at such an age, making large datasets in this age range naturally difficult to obtain. Additionally, the study was conducted using a slice thickness of 3 mm, which, while approaching the threshold of optimal conditions for precise forensic age estimation, still provides valuable insights. Despite these factors, the work demonstrates the substantial potential of generative artificial intelligence (AI) in advancing forensic age assessment and offers a strong foundation for future research in this area.
Methods
Data
To evaluate the effectiveness and robustness of generative models for forensic age estimation, we utilized two distinct datasets that reflect the diversity of data sources encountered in real-world scenarios.
The first dataset, used for training and validation, comprised a total of 486 postmortem computed tomography (PMCT) scans collected between 2010 and 2017 by the Office of the Medical Investigator (OMI) of the University of New Mexico [17], a centralized, university-based institution operating as a state-level medicolegal death investigation agency. These data were selected because the OMI provides a dataset characterized by standardized imaging protocols and a broad diversity of cases. The availability of a curated, anonymized dataset enabled efficient research without additional ethical and logistical hurdles. Furthermore, the diversity of cases ensures the development of a model that can be applied across various demographic and geographic contexts. To obtain access to the dataset, a formal research request was submitted, which included the specific research objectives, hypotheses and potential benefits of the study, ensuring that the data would be used for legitimate scientific purposes. The OMI research committee reviewed and approved our research request and a written confirmation of approval was given. Although all personal identifiers were fully anonymized, the dataset still included comprehensive metadata about the cases, such as age, sex and cause of death. The dataset included scans from both male and female individuals aged between 10 and 30 years who passed away under various circumstances, including natural causes, accidents and violent deaths. The individuals underwent PMCT scanning as part of routine for forensic procedures to help determine the cause of death. The age range was specifically selected due to its association with distinct skeletal development patterns. Only individuals without known traumatic bone fractures affecting MCE were included. Additionally, only patients who had undergone whole-body PMCT scans as part of the forensic investigation were considered. The scans, stored in the well-established DICOM format, were obtained using a Philips Brilliance Big Bore CT Scanner (Philips, Amsterdam, The Netherlands). The technical parameters were standardized, featuring a slice thickness of 3.0 mm, a scan speed between 0.567 s and 0.933 s, a current strength ranging from 104 mA to 254 mA, and a voltage setting of 120 kV.
The second dataset of 40 PMCT scans for testing was provided by the Institute of Legal Medicine at the University Medical Center Hamburg-Eppendorf (UKE). These scans were obtained using a Philips Incisive 128 CT Scanner (Philips, Amsterdam, The Netherlands). The technical specifications included a slice thickness of 0.67–1.0 mm, a scan speed of 0.3–0.6 s, a voltage setting of 120 kV and a current strength of 104–399 mA. As the scans in the second dataset were recorded as part of routine forensic procedures and included individuals outside the relevant age range (10–30 years), we manually filtered the UKE scans to retain only those within this age range, excluding all others and then anonymize all scans technically before investigating them. Subsequently, we harmonized the datasets by subsampling the second dataset with a higher resolution to the slice thickness of the training dataset.
We focused on the assessment of the MCE for age estimation as this region provides a reliable indicator of skeletal maturity [14]. For the validation of the approaches, a human rater (AC), who has a solid foundation in the annotation of image data, annotated corresponding masks of the MCE on the PMCT scans. To ensure the accuracy of the annotations, random samples of the results were reviewed by a clinician (AJW), who confirmed the reliability of the annotations. To increase the data volume and facilitate the segmentation process, we virtually split the PMCT and the corresponding mask data vertically, along the midline through the sternum, treating each side of the MCE as an independent sample. For the model input, only the MCE regions were specifically masked. Further preprocessing included random cropping for data augmentation to prevent overfitting by limiting the model’s options to remember samples despite the limited amount of data. Additionally, the image intensities were normalized based on Hounsfield units to highlight bone contrasts and minimize irrelevant information. Those masks defined the ROIs used as input for the generative models.
Architecture of generative models
In selecting the model architecture, we considered various paradigms of generative models. For the advantages of an explicit latent space, we focused on models providing an explicit density. Given their popularity and extensive literature regarding attractive properties for their latent spaces, we chose variational autoencoders (VAEs) directly approximating this density for this study. These models from the autoencoding framework share the usage of two distinctive components: their encoder component maps data points from a high-dimensional data space into a low-dimensional space. The decoder component is trained to map these latent embeddings back to the high-dimensional space. As all data must run through, the model is forced to “compress” the spatial information within the images in a far smaller latent space. Crucially, VAEs treat the vectors from the latent space as proper random variables. Any vector within the latent space, despite being created from a sample, randomly sampled or otherwise created, could be converted back to an image. A more rigid mathematical formulation and details regarding the training by maximizing the evidence lower bound can be found in Kingma et al. [7].
While VAEs have proven to be powerful in knowledge extraction, there is no guarantee regarding the precise information stored in the latent representation and their accessibility for other tasks. Accordingly, a body of literature focuses on additional training objectives for VAEs to separate the underlying factors of variation into variables with semantic significance, a concept known as disentangled learning. To assess the impact of these learning objectives, we first consider a standard VAE as the baseline model. Additionally, we explored two more specialized VAE variants: ArVAE [12] which enforces a structured relationship between the latent space and specific continuous attributes, allowing for controlled manipulation of these attributes. Lastly, we employed AttriVAE [2] which incorporates the regression task into the training. These training strategies may enable a more interpretable representation, focusing on the influence of these attributes on the data.
The general architecture of the used deep-learning model is described in Fig. 1 and [7]. The same architecture was used for all three previously named training objectives. The encoder utilized 3D convolutional layers with increasing channel numbers (16–256), followed by batch normalization [6], an activation using Leaky ReLU [9] and dropout [16]. The weights are initialized using Kaiming initialization [5] except the final layer with a sigmoid activation layer and initialized according to the Xavier initialization [4]. The decoder mirrors this structure with transposed convolutional layers.
Fig. 1
General variational autoencoders (VAE) architecture of the deep-learning model used
×
Training and evaluation
Only the first dataset from Mexican PMCT scans was used for training and validation of the models. Hyperparameters such as the learning rate were fine-tuned iteratively by utilizing 30% of this dataset for validation purposes. From the results of the first experiments, we chose 128 as dimensionality of the latent space. The dimensionality refers to the number of features that the model uses to represent the data in its compressed form. The latent space is a lower dimensional representation of the input data, where the model learns the most important patterns and structures. The dimensionality was determined through an iterative process based on model evaluation. The dimension was selected after testing several candidate values and assessing the corresponding performance metrics. Specifically, the quality of the generated outputs and the stability of the learning curve were closely monitored. The dimensionality of 128 was chosen as it provided the best balance between effective data compression and the preservation of important features, as evidenced by the improved model performance and the convergence of the training process.
We used the Adam optimizer with a learning rate of 1e‑4 and a batch size of 16 as an optimization technique through gradient descent. The learning rate was dynamically reduced on missing improvement on the training dataset. Additionally, we stopped the training early once overfitting became evident due to missing improvements on the validation dataset. All models were trained until convergence of their loss. The second dataset from UKE scans was exclusively reserved for the final testing purposes.
We analyzed the resulting latent spaces of the different learning objectives both qualitatively and quantitatively. Depending on the set-up, we reduced the dimension of the latent space further by projecting it into a visualizable two-dimensional space. For this dimensionality reduction, we applied a principal component analysis (PCA) as a simple baseline with a focus on the global structure of the latent space. Additionally, we employed uniform manifold approximation and projection (UMAP) in its supervised formulation [11] as a non-linear alternative with tuneable control for preserving both local and global structures.
For the qualitative analysis, we directly assessed the quality of the reconstructions and the obtained embeddings. By coloring the projected embeddings given the age associated with the corresponding sample, we assessed whether this information was implicitly encoded within the learned representations.
For the quantitative evaluation, we focused on the predictive performance and its generalizability across the datasets when the embeddings rather than the original samples were used for age prediction. Again, we conducted the calculation both for the original embedding and the representation with reduced dimensionality to quantify the trade-off for using an interpretable decision. As regression algorithms, we chose three different parametric and non-parametric methods: a simple linear regression, a k-nearest neighbor algorithm and a boosted ensemble of decision trees. For all comparisons, we used the mean absolute difference between prediction and ground truth. Unlike in the training of the models with a fixed validation dataset, we used a 10-fold cross-validation to obtain an estimate regarding the uncertainty within the prediction quality.
As an additional quantitative evaluation, we conducted a classification task exclusively on the limited data from the UKE to assess the usability of the learned representation for solving the related but not identical task of predicting whether a person is older than 21 years. Given the small number of samples, solving such a task without a pretrained feature extractor appears particularly challenging. Therefore, we applied a simple and interpretable support vector machine with a linear kernel to enforce utilization of the latent space derived from the significantly larger training dataset. In this case, we did not conduct hyperparameter optimization and used a grouped five-fold cross-validation to avoid overfitting.
Results
Dataset characteristics
A total of 684 PMCT scans were identified from both data sources. Of these 29% (n = 196) were excluded due to missing age information, duplication and scans of unsuitable body parts, where the MCE was obscured due to trauma, resulting in a final dataset of 488 scans or 976 samples.
The demographic distribution included 70% males (n = 341) and 30% females (n = 147) (Table 1). The age groups were categorized as follows: 10–14 years (n = 65), 15–17 years (n = 107), 18–21 years (n = 107), and 22–30 years (n = 209). The 22–30 years group represented 43% of the dataset, while the 10–14 years group was the smallest, comprising 13%. The training data consisted of 80 samples from the UKE, while the remaining 896 samples were randomly split into 627 for training (70%) and 269 for validation (30%).
Table 1
Distribution of the scans across different age groups and genders in both data sources
Age group
Training/validation dataset
Testing dataset
Total (n)
Male (n)
Female (n)
Total (n)
Male (n)
Female (n)
10–14 years
64
40
24
1
1
0
15–17 years
101
76
25
6
2
4
18–21 years
101
71
30
6
3
3
22–30 years
182
130
52
27
18
9
Total
448
317
131
40
24
16
Quantitative results
In our quantitative analysis, the architecture incorporating all training objectives demonstrated effective learning of embeddings and successfully reconstructed content within the masked regions. To evaluate the visual quality of these reconstructions, Fig. 2 presents five samples from the test set from different individuals alongside their original counterparts.
Fig. 2
The original input and reconstructed approximations from the corresponding latent space of five samples from the test set. VAE variational autoencoder, ArVAE attribute-regularized VAE, AttriVAE attribute-interpreter VAE
×
On direct comparison, the models showed the expected blurry effect well-documented for VAEs. Especially when assessing the MCE, the visual reconstruction alone could not be used for an assessment without the original ground truth due to this effect. While other generative models might provide a “crisper” representation, the shape itself was well-preserved and could be overlayed to ensure proper identification and parsing of the clavicle. Noticeably, the impact of additional training objectives is clearly discernible in the ArVAE and AttriVAE models. Embedding additional information during training appeared to significantly support the reconstruction quality on the test set.
In terms of the latent space for direct and interpretable age estimations as visualized in Fig. 3, the difference between the methods where only minor. When using a PCA for a projection focusing on the global structure of the semantic space, age did not appear to represent the primary axis of variances identified by this method; however, using a method adjustable for the influence of global and local properties issued far better results. In all cases, UMAP created projections that could, depending on the task of interest, be easily used by humans to assess the rough age of a person within the visualization. The actual effect of the training objective was only of minor influence in this case.
Fig. 3
Projected embeddings both on the validation and test dataset in one graph. Each sample is colored according to the age groups of the subjects, the color scale on the right shows the age distribution for this study ranging from 10 to 30 years of age. VAE variational autoencoder, ArVAE attribute-regularized VAE, AttriVAE attribute-interpreter VAE, PCA principal component analysis, UMAP uniform manifold approximation and projection
×
Qualitative results
The analysis of the results visualized in Fig. 4, indicates several noteworthy patterns concerning the generative models’ efficacy in forensic age estimation tasks phrased as regression task. The performance across the datasets, dimensionality reduction through UMAP and the utilized regression techniques reveal the interplay between those different components.
Fig. 4
Mean absolute error (MAE) obtained on the different datasets. For the training datasets (left), the error bars indicate the variation across k results obtained through k‑fold cross-validation. In contrast, for the test dataset (right) only a single test set was evaluated. VAE variational autoencoder, ArVAE attribute-regularized VAE, AttriVAE attribute-interpreter VAE
×
In terms of absolute performance, the difference between the algorithms is relatively low and falls well behind the standard deviation of results on different folds of the cross-validation. Given the learned structure of the latent space, the complex ensemble models with their demand for more data do not perform better than a simple linear regression. The trend remains between the training and test dataset: predictive performance using latent representations differed not significantly between both set-ups and might indicate robust findings and the linear regression as an attractive tool for the task.
Reducing the latent space dimensionality using UMAP into the interpretable 2D domain caused a performance drop across all models and regression algorithms. Nevertheless, the embeddings retained predictive power, demonstrating their potential utility in enabling fully interpretable downstream analyses.
In the classification task of determining whether a person was above 21 years old, the state simple vector machine with the linear kernel model was able to well utilize the “pretrained” semantic space. The best results were obtained on the ArVAE and are depicted in Fig. 5. Despite the smaller number of individuals below the age of 21 years in the dataset, the model achieved a precision of 76% and a recall of 73%. For individuals reaching 21 years or older, the model demonstrated higher performance, with a precision of 90% and a recall of 91%. Overall, the model achieved an accuracy of discrimination of 86%. The individuals rated as 21 years of age or older by AI but had not reached the age cut-off in reality included two individuals (18 years old and 19 years old) where both sides were predicted to be over 21 years old, as well as a 16-year-old and a 19-year-old, where only one side of the MCE was predicted to be over 21 years old.
Fig. 5
Confusion matrix obtained during the fivefold cross-validation when utilizing the learned semantic space for the classification on data of the University Medical Center Hamburg-Eppendorf
×
Discussion
Understanding complex and unstructured data, such as medical images and time series, commonly necessitates reasoning based on simpler representations than the original raw data. Regardless of whether CNNs, transformers, or one of the other architectures are employed, the success of deep learning, shown in the last years, results primarily from their power as feature extractors. Unlike before, we no longer require a human understanding of the features and modelling of the modality and problem by data itself for research; however, once the structured features are extracted, the large ensemble of possible machine learning algorithms with their unique advantages and disadvantages are commonly equally powerful and still outperform deep learning on structured data [1].
Some generative methods, with their emphasis on learning the hidden data distributions, enable utilization of these capabilities regarding the “understanding” of the data in terms of its latent underlying factors. Accordingly, they promise advantages that extend beyond mere performance metrics. Although we quantitatively observed that they might not have the same degree of predictive accuracy compared to discriminative models, generative approaches with access to the underlying latent space commonly offer a more tweakable and interpretable look into the model’s reasoning processes. This enables the verification of different parts of the data pipeline like “understanding the data” and “learning a classifier”, which are essential when creating trustworthy medical utilities. In this study, we demonstrated the potential chances in legal medicine: relatively independent of the selected training architecture, all the employed training objectives learned a latent space allowing solving tasks on complex images within an easily visualizable and interpretable 2D space. For a first view on the expected age range given a CT scan of the clavicle, the pipeline appears suitable and useful. Like popular foundation models often trained through self-supervised learning on far larger datasets, the latent space additionally enables the effective re-use of information even in paradigms with limited data. In the task conducted for evaluation, data from 40 patients were sufficient to train an understandable classifier with an accuracy of 86%; however, there were six false positive decisions (Fig. 5), which would be unacceptable in a real-case scenario of medical age assessment due to the potential legal consequences.
The ethical use of artificial intelligence in medicine and especially in sensible areas like forensic age estimation necessitates other considerations beyond pure performance metrics. Our findings demonstrate the feasibility of designing a system that accommodates these non-functional requirements while maintaining reasonable performance levels. The probabilistic nature enabled additional actions by scientists and clinicians providing insights into the powerful “black boxes”. As the representations could always be converted back into masks visualized on top of the original data, a doctor could always check whether the model examined the clavicle correctly. If there are observable challenges, any classification conducted on the faulty representations must be considered with care. The clinician could even start modifying the representations to conduct hypothetical “what would be if” questions. Finally, they could easily consider the similarity between the samples. Guided by computers or, in the case of the projection, by a simple glance, they could compare those samples which were analyzed before, or which have a known age level. In summary, by examining the latent space of these representations, one could build a supportive tool helping humans to detect outliers and identify similar or dissimilar samples, enhancing the overall transparency and reliability of the system.
In a clinical setting, this could mean that a clinician, while analyzing PMCT scans, inspects the model’s latent representation to ensure that key bones, such as the MCE, are correctly identified. If the model misclassifies an image, the doctor could adjust the representation in real time, significantly reducing misclassification. This approach could be a steppingstone toward a more advanced goal. The aim is to develop a fully automated model that functions independently of human supervision, minimizing the risk of errors introduced by manual adjustments. In the future, the model could ideally perform with high accuracy and reliability, without the need for human intervention.
Nevertheless, some limitations and future directions of our studies must be acknowledged and should be considered in future research. The relatively small amount of data available rendered the use of some advanced deep learning architectures impractical and might not enable the best possible reconstruction quality. The small sample size in the test dataset, particularly in the 18–21 years age group, where only 12 scans were available, is a critical limitation. This is especially because 21 years marks a significant legal threshold in German law for the age of majority. Accurately addressing this threshold requires visualizing advanced ossification stages of the MCE. This study was not designed for immediate integration into the AGFAD guidelines, as it deliberately excluded physical and dental examinations as well as hand and intraoral X‑rays for the sake of simplification. Instead, its primary aim was to demonstrate the potential of generative AI in forensic age assessment by focusing solely on the analysis of the MCE.
While the current study’s performance, measured by metrics such as MAE and accuracy, is influenced by factors like the small data size and the 3 mm slice thickness, these challenges also highlight valuable areas for future improvement. The 3 mm slice thickness, while close to the threshold for optimal precision in forensic age estimation, still offers a valid starting point. Future studies can build on these findings by utilizing larger datasets, potentially from multicenter studies and by employing advanced imaging techniques with thinner slice thickness, which should further enhance the model’s performance. Additionally, the current pipeline does not inherently include the localization of the clavicle. While the use of a well-established segmentation tool provided a level of trust for the system, it also imposes an upper limit on achievable performance within the segmentation task. Joint training, which could further optimize all components of the entire system, should accordingly be evaluated in future research. Finally, additional work is required to balance the different training objectives: balancing the reconstruction quality and the quality of the learned latent space requires considerations beyond mere technology but a more extensive discussion with other scientists. Depending on the set-up of interest, the fine tuning of the corresponding hyperparameters may customize the flexible generative models more specifically for the forensic set-ups of interest.
Finally, post hoc analyses present a promising opportunity to further enhance the transparency of the models, particularly the discriminative ones used in this study. These methods can provide valuable insights into how the model processes the input data, helping to understand the decision-making process. An in-depth discussion of these post hoc methods could significantly contribute to a deeper understanding of the model’s reasoning. By providing greater transparency, these approaches not only improve the model’s trustworthiness but also create opportunities for refining its performance through enhanced interpretability.
Conclusion
This study highlights the utility and promise of generative models, more specifically of variational autoencoders, for forensic age estimation even with datasets of limited size. The proposed methodologies present a compelling case for rethinking forensic assessments through a lens that balances accuracy with transparency, thereby fostering greater trust in AI applications. As this domain evolves, ongoing research and interdisciplinary collaboration will be essential to fully realize the potential of generative models, ensuring their effective integration into forensic practices that are both technologically advanced and ethically sound.
Funding
The work was conducted as part of a Masters degree of Anastasia Chernysheva without external funding.
Declarations
Conflict of interest
A. Chernysheva, C. Gundler, A.J. Wiederhold, E. Jopp-van Well, A. Heinemann and B. Ondruschka declare that they have no competing interests.
All procedures performed in studies involving human participants or on human tissue were in accordance with the ethical standards of the institutional and/or national research committee and with the 1975 Helsinki declaration and its later amendments or comparable ethical standards. Data collection and pseudonymization has been approved by the ethics committee of the Hamburg Chamber of Physicians (reference 2020-10353-BO-ff) according to the guidelines from the central ethics committee of the Federal Medical Association. Anonymized data were used according to the Hamburg Hospital Act HmbHKG § 12.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Der Einsatz künstlicher Intelligenz (KI) in der Rechtsmedizin ist absehbar, KI-Anwendungen könnten im Ermittlungsverfahren oder in foro bald eine zentrale Rolle in Entscheidungsprozessen einnehmen. Dann werden insbesondere Transparenz …
Similar to other parts of our society, machine learning has emerged as a popular tool within different areas of forensic medicine and will soon fuel more and more research and practice niches of our disciplines. Given the rapid advances, the …
Artificial intelligence application has gained popularity in the last decade. Its application is implemented into multiple industries including the health sector; however, discipline-specific artificial intelligence applications are not widely …
Fragen zu Implementierung und Nutzbarkeit von künstlicher Intelligenz (KI) spielen eine immer größere Rolle in der Forensischen Altersdiagnostik bei Lebenden, insbesondere im Rahmen forensisch-radiologischer Ansätze. Bis dato liegen bereits …