News
11/14/22 Our international collaboration won the DOE INCITE award
Our ultimate goal is to contribute to advancing AI from narrow to “broad” (general) while ensuring AI Safety
and alignment with human values, and contribute towards advances in other fields (healthcare, biomedical
sciences, and others) via developing generic, powerful large-scale models pretrained in a self-supervised
manner on broad variety of datasets. Such models can serve as a foundation of transferable knowledge
and can be used in a broad variety of applications ("downstream tasks") due to their drastically improved
generalization abilities as compared to prior state-of-art in the field of AI. More specifically, building on
recent successes in this area, we plan to train large-scale neural network models called Transformers, which
recently demonstrated impressive performance in language modeling and image processing; we plan to
evaluate their scaling with increasing model and pretraining dataset size, as well as the amount of compute
available. Such models, also known as "foundation models", appear to improve their generalization and
few-shot learning abilities with scale. We plan to investigate this trend in more detail, and identify the
most promising approaches to scaling the architecture, and datasets. Next, we plan to extend these models
to handle a much wider range of modalities beyond text and images, as well as various machine-learning
tasks, and expand them towards adaptive, continually learning systems. Finally, we plan to use the obtained
foundation models for predictive modeling in several applications such as healthcare and brain imaging. As
an outcome of this project, we will obtain highly transferable multi-modal models that we will make publicly
available. Hence we also see this project as a step towards the democratization of large multi-modal models
for the broader research community. The public availability of these models would allow researchers to
investigate both strengths and weaknesses of large multi-modal models, and further improve their transfer
capabilities to be used widely across various scientific domains.