Heng Ji

 

Heng Ji

 

 

 

 

Talk: Science-Inspired AI

Unlike machines, human scientists are inherently “multilingual,” seamlessly navigating diverse modalities—from natural language and scientific figures in literature to complex scientific data such as molecular structures and cellular profiles in knowledge bases. Moreover, their reasoning process is deeply reflective and deliberate; they “think before talk”, consistently applying critical thinking to generate new hypotheses. In this talk, I will discuss how AI algorithms can be designed by drawing inspiration from the scientific discovery process itself. For example, recent advances in block chemistry involve the manual design of drugs and materials by decomposing molecules into graph substructures—i.e., functional modules—and reassembling them into new molecules with desired functions. However, the process of discovering and manufacturing functional molecules has remained highly artisanal, slow, and expensive. Most importantly, there are many instances of known commercial drugs or materials that have well-documented functional limitations that have remained unaddressed. Inspired by scientists who frequently “code-switch”, we aim to teach computers to speak two complementary languages: one that represents molecular subgraph structures indicative of specific functions, and another that describes these functions in natural language, through a function-infused and synthesis-friendly modular chemical language model (mCLM). In experiments on 430 FDA-approved drugs, we find mCLM significantly improved 5 out of 6 chemical functions critical to determining drug potentials. More importantly, mCLM can reason on multiple functions and improve the FDA-rejected drugs (“fallen angels”) over multiple iterations to greatly improve their shortcomings. Preliminary animal testing results further underscore the promise of this approach.

Workshop Home Page

Return to the TAU-UIUC Workshop Home Page

Home→