Science and Science Communication in the Age of Large Language Models
Artificial Intelligence (AI) is emerging as a transformative force in scientific research and communication, with seemingly limitless potential applications. Recent advances in large language models (LLMs) have particularly encouraged researchers to explore their applications in their own work.
Fields such as Medicine and Chemistry have already embraced LLMs in various applications 1-3. This includes using them for accelerating drug discovery, optimising vaccine design, even making mind-reading a tangible possibility 1. In biotechnology, researchers are actively using LLMs to generate protein sequences with predictable functions across diverse families, demonstrating considerable potential for protein design and engineering 4. There are even ongoing studies exploring the automation of the entire scientific processes using LLMs, including everything from design of research to planning and execution of lab experiments 5. In social science, researchers have been inspired to use LLMs to improve agent-based models, simulate human populations taking surveys, and even replacing human participants in social experiments 6-8. The Impressive ability of LLMs to answer questions has also inspired researchers to use them for creating expert-domain chatbots, in diverse areas such as climate change, finance, physics, material science, and clinical knowledge 9-13.
This rapid proliferation of AI tools in scientific research and communication, however, has raised significant concerns among scientific communities. They caution against various risks, including the environmental footprint, the disruption of nuanced and contextual understanding, contributions to an AI-driven infodemic, and the undermining of human interaction14. There are also concerns about researchers becoming vulnerable to illusions of understanding, where they may believe they comprehend more about the world than they actually do. This is followed by epistemic risks that may arise if these tools dominate the production of scientific knowledge15. These and other concerns have encouraged new efforts aimed at developing ethical and policy guidance for the use of AI within scientific research and communication16.
In this symposium, we invite contributions that engage critically and empirically with the question of how LLMs might transform the production of scientific knowledge and science communication. What are the risks and long-term challenges associated with integrating these technologies into the daily practices of researchers and science communicators? And how can we envision new ways to design, develop, and deploy LLMs in a responsible way that best aligns with the aspirations of scientists and science communicators?
This event is organised by ANU Responsible Innovation Lab, and supported by ANU Institute for Climate, Energy & Disaster Solutions, Australian National Centre for Public Awareness of Science.
This free event will be held both in-person.
If you would like to share your ongoing project on this topic, please contact Ehsan Nabavi (Ehsan.Nabavi@anu.edu.au).
October 4, 2024 |9:30am-12:30pm
Friday 4th October 2024, 9:30am-12:30pm AEST
Innovation Space, Level 2 ANU Birch Building 35 Science Rd, Acton ACT 2601, Australia
REFERENCES
1 Tang, J., LeBel, A., Jain, S. & Huth, A. G. Semantic reconstruction of continuous language from non-invasive brain recordings. Nature Neuroscience, 1-9 (2023).
2 Pan, J. Large language model for molecular chemistry. Nature Computational Science 3, 5-5 (2023). https://doi.org:10.1038/s43588-023-00399-1
3 Li, S. et al. CodonBERT: Large Language Models for mRNA Design and Optimization. bioRxiv, 2023.2009. 2009.556981 (2023).
4 Madani, A. et al. Large language models generate functional protein sequences across diverse families. Nature Biotechnology 41, 1099-1106 (2023). https://doi.org:10.1038/s41587-022-01618-2
5 Boiko, D. A., MacKnight, R. & Gomes, G. Emergent autonomous scientific research capabilities of large language models. arXiv preprint arXiv:2304.05332 (2023).
6 Aher, G., Arriaga, R. I. & Kalai, A. T. Using large language models to simulate multiple humans. arXiv preprint arXiv:2208.10264 (2022).
7 Ziems, C. et al. Can Large Language Models Transform Computational Social Science? arXiv preprint arXiv:2305.03514 (2023).
8 Bail, C. A. Can Generative AI Improve Social Science? (2023). https://doi.org:10.31235/osf.io/rwtzs
9 Caruccio, L. et al. Can ChatGPT provide intelligent diagnoses? A comparative study between predictive models and ChatGPT to define a new medical diagnostic bot. Expert Systems with Applications 235, 121186 (2024).
10 Vaghefi, S. A. et al. Chatclimate: Grounding conversational AI in climate science. (2023).
11 Wu, S. et al. Bloomberggpt: A large language model for finance. arXiv preprint arXiv:2303.17564 (2023).
12 Singhal, K. et al. Large language models encode clinical knowledge. Nature 620, 172-180 (2023). https://doi.org:10.1038/s41586-023-06291-2
13 Xie, T. et al. DARWIN Series: Domain Specific Large Language Models for Natural Science. arXiv preprint arXiv:2308.13565 (2023).
14 Nabavi, E. et al. Potential Benefits and Dangers of Using Large Language Models for Advancing Sustainability Science and Communication. Authorea Preprints (2024).
15 Messeri, L. & Crockett, M. Artificial intelligence and illusions of understanding in scientific research. Nature 627, 49-58 (2024).
16 Resnik, D. B. & Hosseini, M. The ethics of using artificial intelligence in scientific research: new guidance needed for a new tool. AI and Ethics, 1-23 (2024).