Abstract
Background: Autosomal dominant nonsyndromic hearing loss (ADNSHL) is highly heterogeneous, with more than 64 genes implicated in its etiology. This complexity limits the diagnostic power of clinical examinations and audiometry alone, while existing computational approaches have achieved only moderate accuracy and often lack interpretability. As precision medicine increasingly emphasizes genotype-phenotype correlations, there is a recognized need for diagnostic tools that provide clinicians with transparent, interpretable outputs.
Objective: This study aimed to develop and evaluate the AudioGene Translational Dashboard, an interpretable clinical informatics tool that integrates machine learning models and interactive visualizations to enhance genotype-phenotype correlations and support diagnostic decision-making in ADNSHL.
Methods: We developed the AudioGene Translational Dashboard, integrating 2 machine learning models (AudioGene version 4 and AudioGene version 9.1) with 6 interactive visualization tools. AudioGene version 4 uses a multi-instance support vector machine classifier for patients with multiple audiograms, while AudioGene version 9.1 combines adaptive boosting, k-nearest neighbors, random forest models, and logistic regression for patients with a single audiogram. Visualizations include audiometric profile plots, audioprofile surfaces, clustering analyses, and data distribution charts designed to facilitate clinical interpretation.
Results: The AudioGene Translational Dashboard was developed to address the “70/30” phenomenon, indicating a 74% likelihood that the causative gene is among the top 3 predicted genes, thereby providing clinicians with a clear confidence indicator (“green flag”) or a caution alert (“red flag”) during diagnosis. While this level of performance is well suited for hypothesis generation, the remaining uncertainty underscores the need for interpretive context in clinical decision-making. Visualization tools enhanced clinicians’ ability to interpret and correlate phenotypic data with predicted genetic outcomes, improving diagnostic confidence and interpretability.
Conclusions: The AudioGene Translational Dashboard advances clinical informatics in genetic diagnosis of ADNSHL by integrating explainable artificial intelligence with interactive visualizations, enhancing clinical interpretability and diagnostic accuracy. This approach facilitates informed clinical decision-making, highlights the translational potential of genotype-phenotype computational models, and supports precision medicine in hearing loss diagnostics. Future enhancements will target improving class balance and incorporating additional user-customizable features to further optimize clinical applicability.
doi:10.2196/85212
Keywords
Introduction
Background
Autosomal dominant nonsyndromic hearing loss (ADNSHL) presents a significant genetic diagnostic challenge due to its underlying heterogeneity—more than 64 genes are implicated in its etiology []. Because of this complex genetic landscape, computational tools designed to correlate audiogram profiles (commonly called audioprofiles) with specific genes have achieved only moderate success [,]. One such tool, which we developed approximately 15 years ago, is AudioGene. AudioGene uses numerous machine learning (ML) approaches to improve diagnostic precision [,,]. These approaches include semisupervised support vector machines (SVMs), ensemble models, and hyper-tuning methods. However, challenges such as data imbalance and class sparsity continue to restrict the accuracy of these models.
Precision medicine harnesses information about the genome of an individual, environment, and lifestyle to guide medical care. With heterogeneous disorders such as ADNSHL, genetic variant interpretation can be challenging, complicating the diagnostic process and impacting patient care. Computational tools may improve the precision and reliability of genetic assessments by capitalizing on genotype-phenotype associations [,].
Current diagnostic methods for ADNSHL largely rely on clinical examination and audiometry, which do not provide sufficient resolution for the complex genetic landscape of ADNSHL [,]. However, with the availability of ML and artificial intelligence–driven approaches, there has been a shift toward integrating computational and visualization tools with genetic diagnostics to improve accuracy and predictive power [,].
To address these challenges, we have developed the AudioGene Translational Dashboard with the goal of enhancing both the accuracy and interpretability of genetic predictions. A feature of the AudioGene Translational Dashboard is the “70/30” phenomenon: by integrating the results from both models on a training dataset comprising 3189 audiograms from 1445 patients, we observed that the correct disease-causing gene was predicted within the top 3 predictions 74% of the time, with incorrect predictions accounting for the remaining 26%, hence “70/30.” This observation signals to health care providers when they can have confidence in the top predictions, serving as a “green flag” or “red flag” in the diagnostic process. Having a true positive rate of 70% is beneficial from a research perspective; however, for a diagnostic tool, the remaining 30% represents some risk that necessitates additional interpretative context. By providing this context, the AudioGene Translational Dashboard enables health care providers to weigh their confidence in the predictions, supporting more informed diagnostic decisions.
The AudioGene Translational Dashboard was introduced into the AudioGene toolset to increase transparency into the “black box” underlying the models by providing explainable artificial intelligence (XAI) to enhance model interpretability and utility in clinical settings, in line with trends in precision medicine that emphasize the importance of genotype-phenotype associations in improving diagnostic outcomes [,].
Related Works
Early attempts to map audiometric phenotypes to their underlying genotypes were spearheaded by AudioGene version 4 (AG4), a semisupervised multi-instance SVM that treats the collection of audiograms for a single patient as a “bag” and ranks loci according to pair-wise–coupled probability estimates [,]. Building on this foundation, AudioGene version 9.1 (AG9.1) introduced selective intraensemble data partitioning: training examples are first divided by gene-specific data volume, patient age, and audiogram shape, then modeled with a committee of k-nearest neighbor (KNN), adaptive boosting, and random forest subclassifiers, whose outputs are fused by logistic regression. AG9.1 offers a top-3 accuracy of 77.8%, with a precision of 0.51 and a recall of 0.56, at the cost of introducing a more complex model. We report top-3 accuracy rather than top-1 accuracy because, in the context of gene prioritization for validation sequencing, the cost of excluding the true causative gene is higher than the cost of evaluating a small number of candidate genes. In addition, the top-3 threshold represents a practical trade-off between high confidence in predictions and an acceptable loss of significance when selecting genes for sequence-based validation [,,,]. Both frameworks have improved locus-ranking accuracy for the 23 well-curated ADNSHL genes that account for roughly three-quarters of cases in populations of European ancestry []. Nonetheless, their predictions can still be difficult to interpret when class imbalance, sparse age coverage, or atypical audiogram morphologies are present.
Complementary to algorithmic advances, domain-specific visualization has been welcomed as a potentially beneficial tool for clinical use. Audioprofile surfaces (APS) plot 3D trajectories of frequency-specific threshold shift over time, revealing gene-characteristic progression patterns that are not obvious in 2D audiograms []. Circle-based genome views (eg, Circos [version 0.69-10; Krzywinski, Canada’s Michael Smith Genome Sciences Center] enable high-density comparison of structural variation or copy number events [], while integrative genome browsers such as Integrative Genomics Viewer (version 2.19.7; UC San Diego and Broad Institute of MIT and Harvard) allow rapid inspection of read evidence at candidate loci []. More recent health care dashboards use fuzzy logic overlays and interactive filtering to expose outliers or low-confidence regions directly to end users []. Despite these advances, few systems combine genotype-prediction engines with audiogram-aware visual contexts; therefore, clinicians must cross-reference separate tools, a workflow that can erode trust in algorithmic suggestions and slow decision-making []. These limitations reflect shortcomings in how models communicate their reasoning and how results are presented to end users.
Accordingly, the literature reveals two unmet needs:
- Model transparency—while ensemble and semisupervised approaches improve predictive accuracy, they do not inherently communicate why a particular gene is ranked highly, especially when training data are imbalanced or noisy.
- Unified, clinician-friendly interfaces—existing genomic viewers excel at sequence-level detail but lack phenotype-specific visualization; conversely, stand-alone audiogram tools rarely link observed hearing profiles back to the underlying variant evidence.
The AudioGene Translational Dashboard addresses these gaps by (1) merging the complementary strengths of AG4 and AG9.1 and (2) embedding 6 interactive visual modules—APS, 2D audioprofiles, uniform manifold approximation and projection (UMAP) cluster projections, gene count bar charts, region-of-origin pie charts, and age distribution plots—around the model output. This hybrid XAI-driven design supports clinicians in validating or questioning the algorithm’s “70/30” confidence observation and thus advances the state of practice in ADNSHL diagnostics.
Methods
Tool Overview
The AudioGene Translational Dashboard integrates 2 ML models to increase diagnostic accuracy for ADNSHL:
- AG4—it is a multi-instance classifier designed for patients with multiple audiograms that uses a semisupervised SVM and ranks loci based on modified SVM probability outputs []. It was developed using the Waikato Environment for Knowledge Analysis (version 3.7.2; WekaIO Inc) platform [].
- AG9.1—it is a single-instance classifier for patients with only 1 audiogram, developed using the scikit-learn library in Python. It comprises multiple submodels: 3 KNNs, 6 adaptive boosting models, 2 random forest models, and a logistic regression module for combining outputs [,,].
The AudioGene Translational Dashboard interface provides six distinct visualization tools for health care providers and researchers to interactively assess genetic data and model predictions in real time: (1) audioprofile, a 2D plot displaying the average hearing loss (in dB) over 10 frequencies (125 Hz to 8000 Hz) for each age group, allowing comparison with a patient’s hearing loss over time (); (2) APS, a 3D surface plot depicting gene-specific hearing loss progression over time across frequencies, illustrating age-related changes in decibel loss () []; (3) a region-of-origin pie chart, which displays the geographic origin distribution (eg, Dutch, German, and Chinese) for audiograms associated with each gene; (4) a count bar chart, which illustrates the audiogram count for each gene, highlighting class imbalance challenges (); (5) spatial analysis and clustering, which shows the cluster position of each gene and a 3D plot of audiograms (clusters are created using the k-means algorithm to partition data into 23 gene-specific groups []; the 3D plot compresses 11 features [age plus 10 frequencies] into 3 dimensions using the UMAP method () []); and (2) 6) an age distribution scatter plot, which shows the age distribution for each gene within the training dataset, providing context for the predictive model outputs.

The first 3 visualization tools were developed to compare patient-specific data to average thresholds for each of the 23 ADNSHL-associated genes. This comparison allows patient audiograms to be contextualized for each gene. The audioprofile visualization shows how a patient’s audiogram compares to the expected audiograms associated with each gene.
The APS adds time as the third axis to provide a 3D rendering of gene-specific audiometric thresholds over time. Audiometric thresholds are represented as a 3D plane, depicting the expected dB loss over time (in years) at each frequency, thereby enabling comparisons between a patient’s hearing thresholds and gene-specific expectations.
The spatial analysis and clustering tool uses a bar chart to show the distribution of each prediction among 23 different clusters. These clusters are created using k-means clustering []. Additionally, a 3D plot visualizes the audiograms within our data that have a confirmed genetic diagnosis, compressing 11 features into 3 dimensions using UMAP []. This feature allows users to interact with the bar chart, highlighting corresponding clusters in the 3D plot (). Users can compare their patient’s audiogram, represented by a large red dot, with others in the cluster, facilitating the identification of similar audiograms and associated genetic diagnoses.
By using these 3 visualizations, we aim to either enhance or reduce confidence in the predictions. For example, if the model ranks the COCH gene second among the top 3 predictions, health care providers can analyze the APS and spatial analysis tools to determine the degree of correlation with patterns typically associated with the COCH gene, potentially increasing confidence in that diagnosis.
The last 3 visualization tools provide context for the data in the AudioGene dataset. The count bar chart highlights significant class imbalance, showcasing the challenge the model encounters in predicting smaller classes due to underrepresentation. The region-of-origin pie chart and age distribution scatter plot provide additional context about data distribution, allowing health care providers to understand model limitations and adjust diagnostic strategies accordingly.
The AudioGene Translational Dashboard integrates into the workflow of a clinician as a secondary validation layer. For example, within our clinical workflow, clinicians, genetic counselors, and bioinformaticians review the results of a clinical genetic test, including patient history, family structure, audiograms, and identified variants in hearing loss genes. This team can then inspect a patient’s audiometric data relative to the landscape of audiometric data across all genes and patients, considering variance within a gene, rarity or abundance of cases within a cluster, and distance to genetically validated cases.
System Design
The system was designed using the SERN (SQL, Express.js, React.js, and Node.js) stack, which uses a client-server architecture where computationally intensive tasks are performed by the server, deployed in a Docker (version 28.5.1; Docker Inc) container [,]. The client is supported by React (version 18.2.0; React Foundation), a JavaScript library facilitating user interactions []. Data preprocessing used linear interpolation and extrapolation for missing values. The same methods for handling missing values were applied in the ML models [,,].
The Pandas library in Python was used for data manipulation and analysis [], and visualization libraries such as Plotly were used to create interactive graphs and plots [].
For more information, please refer to the master’s thesis by DeSollar [] and the GitLab repository.
Ethical Considerations
This study was reviewed and approved by the University of Iowa Institutional Review Board (199701065). The institutional review board granted a waiver of informed consent under US federal regulation 45 CFR 46.116(f) (also known as the “Common Rule”), because, although audiograms were originally collected in clinical and research settings, the dataset used for this study was fully deidentified prior to analysis []. All procedures adhered to the ethical standards of the institutional and national research committees and to the 1964 Declaration of Helsinki and its later amendments. This paper does not contain any individual’s data in any form, including individual details, images, or videos. No compensation was provided to participants, as this study involved secondary analysis of a fully deidentified existing dataset and no participants were directly recruited or enrolled.
Results
Introduction to AudioGene Translational Dashboard
The AudioGene Translational Dashboard combines advanced ML models with several visualization tools to create a platform that facilitates the prioritization of ADNSHL-associated genes in genetic testing results. Using patient data, the AudioGene Translational Dashboard generates gene rankings and enables auditory scientists and health care providers to explore these predictions interactively through various visualization tools. Gene ranking in phenotype-genotype associations can aid in the interpretation of complex genetic data, thereby providing greater context and confidence in diagnostic decisions [,].
The “70/30” phenomenon serves as an indicator for health care providers, providing them with the necessary context through visualizations to assess the reliability of the predictions. When the top 3 predictions include the correct gene, health care providers can have greater confidence in proceeding with targeted genetic testing.
Case Studies and Clinical Implications
Several case studies have been carried out to demonstrate the effectiveness of the AudioGene Translational Dashboard in the diagnosis of specific genetic types of ADNSHL [].
Case 1: MYO7A Gene—Patient 1 (ID 5)
In this case study, we analyzed the results from a patient diagnosed with MYO7A-related hearing loss. Our dataset included 2 audiograms, which were predicted by AG4 to be associated with MYO7A-related, EYA4-related, or WFS1-related hearing loss ().

Examining these predictions relative to the patient’s audiograms, the following observations can be made: (1) MYO7A’s audioprofile is similar in the low-to-mid frequencies but diverges in the high frequencies (); (2) EYA4 displays a comparable shape, but the patient’s thresholds are consistently lower than typical values (); and (3) WFS1 shares some similarities in the low-to-mid frequencies but diverges in the high frequencies, as observed with MYO7A ().
The audioprofiles for the 3 genes all show moderate to moderately severe hearing loss thresholds, with either close similarity (<5 dB) at several frequencies or similar shapes. Therefore, we can conclude that the correct gene is likely captured within the top 3 predictions.
Case 2: MYO6 Gene—Patient 2 (ID 2)
In our second case, we explored the results from a patient (ID 2) previously diagnosed with MYO6-related hearing loss. Our dataset contains only 1 audiogram, with gene predictions shown in by AG9.1.

None of the top 3 candidate genes (WFS1, TECTA, or MYO7A) display an audioprofile that closely aligns with the patient’s thresholds (). This mismatch strongly suggests that the true causative gene (MYO6) does not appear among the model’s top 3 predictions for this patient.
From the clustering interface, we observe that the genes closest to the patient’s audiogram by the KNN metric (WFS1, EYA4, and KCNQ4) also fail to match the patient’s observed audioprofile in any convincing way. Furthermore, KCNQ4, GSDME, and POU4F3, which are noted to have multiple data points near the patient’s cluster, likewise show audioprofiles inconsistent with the patient’s hearing loss. These factors combine to produce inconclusive predictions in this case. When we then integrated these data with the genetic data—which identified no genetic variants in WFS1, TECTA, or MYO7A and confirmed a known variant in MYO6—we verified that the correct gene was not found among the model’s top 3 predictions. This outcome highlights how conflicting audioprofiles and clustering results can indicate that a prediction should be viewed with caution.
Case 3: WFS1 Gene—Patient 3 (ID 13)
In our third case, the patient (ID 13) was diagnosed with WFS1-related hearing loss and, based on 3 audiograms, was predicted by AG4 to have TECTA-related, WFS1-related, or COL11A2-related hearing loss ().

When examining these predictions in relation to the patient’s audiograms, the following conclusions emerge regarding why WFS1 is likely the correct gene and is captured within the top 3 predictions. First, the audioprofile for TECTA () shows some similarities; however, it does not fully capture the nuanced relationship between low-frequency and high-frequency thresholds observed in the patient’s data. Second, COL11A2 () also exhibits differences that diverge from the patient’s pattern. Finally, WFS1 () demonstrates an especially close match to the patient’s audiometric profile, particularly in the way it mirrors better hearing at the high frequencies relative to the low frequencies. Although one might contend that TECTA or COL11A2 could also be considered candidates based on partial pattern matches, the overall evidence—supported by the 3D clustering in —reinforces that WFS1 provides the best fit.
Thus, whether one emphasizes the possibility of TECTA or COL11A2 as contenders, the integrated data confirm that the correct gene, WFS1, is indeed within the top 3 predictions. This close alignment between the patient’s audiometric data and the WFS1 reference profile, combined with supporting clustering analysis, enhances confidence in the diagnostic utility of the AudioGene Translational Dashboard.
Discussion
Principal Findings
These studies demonstrate how the model can raise or lower confidence in variant interpretation based on whether the correct genetic cause of ADNSHL appears among the top 3 predicted genes. Cases 1 and 3 illustrate scenarios in which the model successfully includes the causative gene in its top predictions and closely matches the patient’s audiometric data, thereby justifying a higher level of trust in the result. In contrast, case 2 underscores how mismatched audioprofiles and inconclusive clustering can reveal when the actual gene of interest is likely missing from the top 3 predictions. The interactive visualizations of the AudioGene Translational Dashboard, such as the APS and spatial analysis tools, remain valuable in identifying gene-specific patterns that align with clinical observations [].
However, there are important limitations of the AudioGene Translational Dashboard, especially concerning smaller gene classes. The sparsity of data and the lower accuracy of models in these categories can make the AudioGene Translational Dashboard and phenotypic predictions less reliable. However, by presenting visualizations of the data distribution and class imbalance, these limitations become more apparent, allowing data interpretation to be adjusted accordingly [,].
Conclusions
The AudioGene Translational Dashboard represents an advancement in the field of genetic diagnostics for ADNSHL. By integrating advanced ML algorithms with interactive visualization tools, the AudioGene Translational Dashboard enhances health care providers’ ability to interpret genetic data and make more informed diagnostic decisions.
A central feature of the AudioGene Translational Dashboard is the “70/30” phenomenon, which provides health care providers with critical context for confidence in genetic predictions. When the top 3 predictions are likely to contain the correct gene, the tool serves as a “green flag” for health care providers, increasing diagnostic confidence. Conversely, it alerts health care providers when predictions may be less reliable, serving as a “red flag” and prompting further investigation.
The AudioGene Translational Dashboard is an example of XAI in clinical settings, offering a context-driven method with increased transparency for the diagnosis of ADNSHL. Future developments will focus on incorporating custom model building, enhancing class imbalance functionality, and implementing user suggestions. The AudioGene Translational Dashboard not only advances genetic diagnostics for hearing loss but also serves as an example of a hybrid ML system.
Funding
This research was supported by the National Institutes of Health and the National Institute on Deafness and Other Communication Disorders through grants DC002842, DC012049, and DC017955. These funding sources provided financial support for the development and testing of the AudioGene Translational Dashboard tool, contributing to advancements in machine learning and visualization methodologies for the diagnosis of autosomal dominant nonsyndromic hearing loss.
Data Availability
The datasets generated or analyzed during this study are not publicly available but are available from the corresponding author on reasonable request. The source code for the AudioGene Translational Dashboard is publicly available [] under the GNU General Public License version 3.0 or later.
Authors' Contributions
BD led the design and development of the AudioGene Translational Dashboard, implemented the machine learning models, and drafted the manuscript. NS contributed to the integration of visualization tools and assisted with manuscript preparation. DW and AMO were responsible for data acquisition and preprocessing and contributed to tool validation. KTAB and HA provided expertise in genetic diagnostics and guided the tool’s clinical relevance. MS supported statistical analysis and interpretation of the results. RJHS and TB provided project oversight, secured funding, and critically revised the manuscript. TC supervised the software engineering components and contributed to system architecture design. All authors reviewed and approved the final manuscript.
Conflicts of Interest
None declared.
References
- Walls WD, Azaiez H, Smith RJ. Hereditary Hearing Loss Homepage. URL: https://hereditaryhearingloss.org [Accessed 2026-03-28]
- Taylor KR, Deluca AP, Shearer AE, et al. AudioGene: predicting hearing loss genotypes from phenotypes to guide genetic screening. Hum Mutat. Apr 2013;34(4):539-545. [CrossRef] [Medline]
- DeSollar BR. AGTD - The AudioGene Translational Dashboard: a hybrid machine learning and visualization interface for genetic diagnosis of autosomal dominant non-syndromic hearing loss [Master’s thesis]. University of Iowa; 2024. URL: https://iro.uiowa.edu/esploro/outputs/graduate/9984647256502771 [Accessed 2026-03-28]
- Ryan S. Machine learning prediction of genetic hearing loss via selective intraensemble data partitioning [Master’s thesis]. University of Iowa; 2024. URL: https://iro.uiowa.edu/esploro/outputs/graduate/Machine-learning-prediction-of-genetic-hearing/9984647557802771 [Accessed 2026-03-28]
- Nwakama CC. AudioGene 9.0: novel ensemble machine learning classification of 23 classes of autosomal non-syndromic hearing loss (deafness) [Master’s thesis]. University of Iowa; 2021. URL: https://iro.uiowa.edu/view/pdfCoverPage?instCode=01IOWA_INST&filePid=13841170450002771&download=true [Accessed 2026-03-28]
- Gunning D, Stefik M, Choi J, Miller T, Stumpf S, Yang GZ. XAI-Explainable artificial intelligence. Sci Robot. Dec 18, 2019;4(37):eaay7120. [CrossRef] [Medline]
- Smith RJH, Bale JF Jr, White KR. Sensorineural hearing loss in children. Lancet. Mar 2005;365(9462):879-890. [CrossRef] [Medline]
- Venkatesh MD, Moorchung N, Puri B. Genetics of non syndromic hearing loss. Med J Armed Forces India. Oct 2015;71(4):363-368. [CrossRef] [Medline]
- API reference—Pandas 1.5.3 documentation. Pandas. URL: https://pandas.pydata.org/pandas-docs/version/1.5/reference/index.html [Accessed 2026-03-28]
- Albarrak AM. Improving the trustworthiness of interactive visualization tools for healthcare data through a medical fuzzy expert system. Diagnostics (Basel). May 13, 2023;13(10):1733. [CrossRef] [Medline]
- Frank E, Hall MA, Witten IH. The WEKA workbench. In: Witten IH, Frank E, Hall MA, Pal CJ, editors. Data Mining: Practical Machine Learning Tools and Techniques. 4th ed. Morgan Kaufmann; 2016.
- Deza E, Deza MM. Encyclopedia of Distances. Springer; 2009.
- Krzywinski M, Schein J, Birol I, et al. Circos: an information aesthetic for comparative genomics. Genome Res. Sep 2009;19(9):1639-1645. [CrossRef] [Medline]
- Cover T, Hart P. Nearest neighbor pattern classification. IEEE Trans Inform Theory. 1967;13(1):21-27. [CrossRef]
- MacQueen J. Some methods for classification and analysis of multivariate observations. In: Berkeley Symposium on Mathematical Statistics and Probability. University of California Press; 1967:281-297.
- Thorvaldsdóttir H, Robinson JT, Mesirov JP. Integrative Genomics Viewer (IGV): high-performance genomics data visualization and exploration. Brief Bioinform. Mar 2013;14(2):178-192. [CrossRef] [Medline]
- Wu TF, Lin CJ, Weng RC. Probability estimates for multi-class classification by pairwise coupling. J Mach Learn Res. 2004;5:975-1005. URL: https://www.jmlr.org/papers/volume5/wu04a/wu04a.pdf [Accessed 2026-03-28]
- Weininger O, Warnecke A, Lesinski-Schiedat A, Lenarz T, Stolle S. Computational analysis based on audioprofiles: a new possibility for patient stratification in office-based otology. Audiol Res. Sep 2, 2019;9(2):230. [CrossRef] [Medline]
- Merkel D. Docker: lightweight Linux containers for consistent development and deployment. Linux J. 2014;2014(239):2. [CrossRef]
- Taylor KR, Booth KT, Azaiez H, et al. Audioprofile surfaces: the 21st century audiogram. Ann Otol Rhinol Laryngol. May 2016;125(5):361-368. [CrossRef] [Medline]
- McInnes L, Healy J, Saul N, Großberger L. UMAP: Uniform Manifold Approximation and Projection. J Open Source Softw. 2018;3(29):861. [CrossRef]
- Docker Compose. Docker Docs. URL: https://docs.docker.com/compose [Accessed 2026-03-28]
- React. URL: https://reactjs.org [Accessed 2026-03-28]
- Plotly JavaScript open source graphing library. Plotly. URL: https://plotly.com/javascript [Accessed 2026-03-28]
- 45 CFR §46.116 - General requirements for informed consent. Code of Federal Regulations. 2018. URL: https://www.ecfr.gov/current/title-45/subtitle-A/subchapter-A/part-46/subpart-A/section-46.116 [Accessed 2026-04-10]
- Schaefer N. AudioGene. URL: https://research-git.uiowa.edu/morl/audiogene/website/AudioGene [Accessed 2026-04-07]
Abbreviations
| ADNSHL: autosomal dominant nonsyndromic hearing loss |
| AG4: AudioGene version 4 |
| AG9.1: AudioGene version 9.1 |
| APS: audioprofile surface |
| KNN: k-nearest neighbor |
| ML: machine learning |
| SERN: SQL, Express.js, React.js, and Node.js |
| SVM: support vector machine |
| UMAP: uniform manifold approximation and projection |
| XAI: explainable artificial intelligence |
Edited by Sean Hacking; submitted 03.Oct.2025; peer-reviewed by Hao Wu, Sunny, Chi Lik Au, Wesley Goar; final revised version received 12.Mar.2026; accepted 12.Mar.2026; published 14.Apr.2026.
Copyright© Benjamin DeSollar, Nathan Schaefer, Daniel Walls, Amanda M Odell, Kevin T A Booth, Hela Azaiez, Michael Schnieders, Richard J H Smith, Terry Braun, Thomas Casavant. Originally published in JMIR Bioinformatics and Biotechnology (https://bioinform.jmir.org), 14.Apr.2026.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Bioinformatics and Biotechnology, is properly cited. The complete bibliographic information, a link to the original publication on https://bioinform.jmir.org/, as well as this copyright and license information must be included.

