Abstract
Our sense of hearing is mediated by cochlear hair cells, of which there are two types organized in one row of inner hair cells and three rows of outer hair cells. Each cochlea contains 5–15 thousand terminally differentiated hair cells, and their survival is essential for hearing as they do not regenerate after insult. It is often desirable in hearing research to quantify the number of hair cells within cochlear samples, in both pathological conditions, and in response to treatment. Machine learning can be used to automate the quantification process but requires a vast and diverse dataset for effective training. In this study, we present a large collection of annotated cochlear hair-cell datasets, labeled with commonly used hair-cell markers and imaged using various fluorescence microscopy techniques. The collection includes samples from mouse, rat, guinea pig, pig, primate, and human cochlear tissue, from normal conditions and following in-vivo and in-vitro ototoxic drug application. The dataset includes over 107,000 hair cells which have been identified and annotated as either inner or outer hair cells. This dataset is the result of a collaborative effort from multiple laboratories and has been carefully curated to represent a variety of imaging techniques. With suggested usage parameters and a well-described annotation procedure, this collection can facilitate the development of generalizable cochlear hair-cell detection models or serve as a starting point for fine-tuning models for other analysis tasks. By providing this dataset, we aim to give other hearing research groups the opportunity to develop their own tools with which to analyze cochlear imaging data more fully, accurately, and with greater ease.
| Original language | English |
|---|---|
| Article number | 416 |
| Journal | Scientific data |
| Volume | 11 |
| Issue number | 1 |
| DOIs | |
| State | Published - Dec 2024 |
Bibliographical note
Publisher Copyright:© The Author(s) 2024.
Funding
We would like to extend our sincere gratitude to the laboratories that generously offered to contribute data for our study, even though, for various reasons, we were unable to incorporate it. Your willingness to collaborate and share your expertise is greatly appreciated, and we are pleased to note that your views are aligned with the most recent NIH policies and guidelines on data sharing. This work was supported by: R01DC020190, R01DC017166 and R01DC017166-04S1 to AAI; T32 DC000038 to CJB; R01DC018827 to GB and JEB; R01DC016365, R21DC020312 and N00014-18-1-2716 to BJW; ZIA DC-000079 to LLC; R01DC000188 and P50DC015857 to MCL; NIDCD DIR DC000096 to AB; R01DC021325 to ACV; R01DC014712 to MAR; DBR and UM were supported by the David F. and Margaret T. Grohne Family Foundation, Core Grant application NCI CCSG (CA014195), R01DC021075 and the CZI Imaging Scientist Award https://doi.org/10.37921/694870itnyzk from the Chan Zuckerberg Initiative DAF, an advised fund of Silicon Valley Community Foundation (funder https://doi.org/10.13039/100014989).
| Funders | Funder number |
|---|---|
| Chan Zuckerberg Initiative | |
| Silicon Valley Community Foundation | |
| David F. and Margaret T. Grohne Family Foundation | |
| National Institutes of Health (NIH) | ZIA DC-000079, T32 DC000038, R21DC020312, P50DC015857, R01DC000188, N00014-18-1-2716, R01DC016365, R01DC017166, R01DC020190, R01DC018827 |
| National Institutes of Health (NIH) | |
| National Childhood Cancer Registry – National Cancer Institute | R01DC021075, CA014195 |
| National Childhood Cancer Registry – National Cancer Institute | |
| National Institute on Deafness and Other Communication Disorders | R01DC014712, R01DC021325, DIR DC000096 |
| National Institute on Deafness and Other Communication Disorders |
ASJC Scopus subject areas
- Statistics and Probability
- Information Systems
- Education
- Computer Science Applications
- Statistics, Probability and Uncertainty
- Library and Information Sciences