TY - CHAP
T1 - Data standardization in the omics field
AU - Rugamika, Chimusa Emile
AU - Kumuthini, Judit
AU - Zass, Lyndon
AU - Chaouch, Melek
AU - Gill, Zoe
AU - Ras, Verena
AU - Mungloo-Dilmohamud, Zahra
AU - Sathan, Dassen
AU - Ghoorah, Anisah W.
AU - Fadlelmola, Faisal
AU - Fields, Christopher
AU - Van Horn, John
AU - Radouani, Fouzia
AU - Konopko, Melissa
AU - Baichoo, Shakuntala
PY - 2023/1
Y1 - 2023/1
N2 - In the past decade, the decreased cost of advanced high-throughput technologies has revolutionized biomedical sciences in terms of data volume and diversity. To handle the sheer volumes of sequencing data, quantitative techniques such as machine learning have been employed to handle and find meaning in these data. The need for the integration of complex and multidimensional datasets poses one of the grand challenges of modern bioinformatics. Integrating data from various sources to create larger datasets can allow for greater knowledge transfer and reuse following publication, whether data are submitted to a public repository or shared directly. Standardized procedures, data formats, and comprehensive quality management considerations are the cornerstones of data integration. Combining data from multiple sources can expand the knowledge of a subject. This chapter discusses the importance of incorporating data standardization and good data governance practices in the biomedical sciences. The chapter also describes existing standardization resources and efforts, as well as the challenges related to these practices, emphasizing the critical role of standardization in the omics era. The discussion has been supplemented with practical examples from different “omics” fields.
AB - In the past decade, the decreased cost of advanced high-throughput technologies has revolutionized biomedical sciences in terms of data volume and diversity. To handle the sheer volumes of sequencing data, quantitative techniques such as machine learning have been employed to handle and find meaning in these data. The need for the integration of complex and multidimensional datasets poses one of the grand challenges of modern bioinformatics. Integrating data from various sources to create larger datasets can allow for greater knowledge transfer and reuse following publication, whether data are submitted to a public repository or shared directly. Standardized procedures, data formats, and comprehensive quality management considerations are the cornerstones of data integration. Combining data from multiple sources can expand the knowledge of a subject. This chapter discusses the importance of incorporating data standardization and good data governance practices in the biomedical sciences. The chapter also describes existing standardization resources and efforts, as well as the challenges related to these practices, emphasizing the critical role of standardization in the omics era. The discussion has been supplemented with practical examples from different “omics” fields.
UR - https://www.elsevier.com/books/genomic-data-sharing/mccormick/978-0-12-819803-2
U2 - 10.1016/B978-0-12-819803-2.00008-0
DO - 10.1016/B978-0-12-819803-2.00008-0
M3 - Chapter
SN - 9780128198032
SP - 137
EP - 155
BT - Genomic Data Sharing: Case Studies, Challenges, and Opportunities for Precision Medicine
A2 - Mccormick, Jennifer
A2 - Pathak, Jyotishman
PB - Elsevier
ER -