TY - JOUR
T1 - Explainable multi-task learning for multi-modality biological data analysis
AU - Tang, Xin
AU - Zhang, Jiawei
AU - He, Yichun
AU - Zhang, Xinhe
AU - Lin, Zuwan
AU - Partarrieu, Sebastian
AU - Hanna, Emma Bou
AU - Ren, Zhaolin
AU - Shen, Hao
AU - Yang, Yuhong
AU - Wang, Xiao
AU - Li, Na
AU - Ding, Jie
AU - Liu, Jia
N1 - Publisher Copyright:
© 2023, The Author(s).
PY - 2023/12
Y1 - 2023/12
N2 - Current biotechnologies can simultaneously measure multiple high-dimensional modalities (e.g., RNA, DNA accessibility, and protein) from the same cells. A combination of different analytical tasks (e.g., multi-modal integration and cross-modal analysis) is required to comprehensively understand such data, inferring how gene regulation drives biological diversity and functions. However, current analytical methods are designed to perform a single task, only providing a partial picture of the multi-modal data. Here, we present UnitedNet, an explainable multi-task deep neural network capable of integrating different tasks to analyze single-cell multi-modality data. Applied to various multi-modality datasets (e.g., Patch-seq, multiome ATAC + gene expression, and spatial transcriptomics), UnitedNet demonstrates similar or better accuracy in multi-modal integration and cross-modal prediction compared with state-of-the-art methods. Moreover, by dissecting the trained UnitedNet with the explainable machine learning algorithm, we can directly quantify the relationship between gene expression and other modalities with cell-type specificity. UnitedNet is a comprehensive end-to-end framework that could be broadly applicable to single-cell multi-modality biology. This framework has the potential to facilitate the discovery of cell-type-specific regulation kinetics across transcriptomics and other modalities.
AB - Current biotechnologies can simultaneously measure multiple high-dimensional modalities (e.g., RNA, DNA accessibility, and protein) from the same cells. A combination of different analytical tasks (e.g., multi-modal integration and cross-modal analysis) is required to comprehensively understand such data, inferring how gene regulation drives biological diversity and functions. However, current analytical methods are designed to perform a single task, only providing a partial picture of the multi-modal data. Here, we present UnitedNet, an explainable multi-task deep neural network capable of integrating different tasks to analyze single-cell multi-modality data. Applied to various multi-modality datasets (e.g., Patch-seq, multiome ATAC + gene expression, and spatial transcriptomics), UnitedNet demonstrates similar or better accuracy in multi-modal integration and cross-modal prediction compared with state-of-the-art methods. Moreover, by dissecting the trained UnitedNet with the explainable machine learning algorithm, we can directly quantify the relationship between gene expression and other modalities with cell-type specificity. UnitedNet is a comprehensive end-to-end framework that could be broadly applicable to single-cell multi-modality biology. This framework has the potential to facilitate the discovery of cell-type-specific regulation kinetics across transcriptomics and other modalities.
UR - http://www.scopus.com/inward/record.url?scp=85159546081&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85159546081&partnerID=8YFLogxK
U2 - 10.1038/s41467-023-37477-x
DO - 10.1038/s41467-023-37477-x
M3 - Article
C2 - 37137905
AN - SCOPUS:85159546081
SN - 2041-1723
VL - 14
JO - Nature Communications
JF - Nature Communications
IS - 1
M1 - 2546
ER -