NSF requires disclosure of AI tool usage in proposal preparation. Ensure you disclose the use of FindGrants' AI drafting in your application.
NSF
Many advanced technologies, from self-driving cars to processing medical images, rely on machine-learning to succeed. Technologies based on deep learning or deep neural networks have proven to be especially effective at learning from vast quantities of data, and yet our theoretical understanding of these tools has lagged behind. As we rely more on neural-network systems in our daily lives, it becomes even more important that we can guarantee their safety and reliability. One failure mode of current systems is that they can be confidently incorrect, and this can cause real-world harm. It would be better if our systems "know what they don't know" by maintaining an internal representation of their own uncertainty. In recent years, progress has been made in the mathematical study of "calibration" and "multicalibration," which establish a mathematical framework for uncertainty, probability, and fairness in problems having to do with categorical prediction such as classifiers and recommender systems. The goal of this project is to port these ideas into perceptual domains such as image or video processing. This project will result in the creation of new general-purpose neural networks for image-processing based on a new mathematical principle of learning "calibrated representations" of images, resulting in general-purpose systems that effectively "know what they don't know." This will enable more robust and reliable image-processing applications across wide sectors of research and technology. Representation-learning is a challenging area of machine learning (ML) in which the goal is not to solve any particular task, but to learn – from unlabeled and minimally-structured data – to form a representation or embedding vector of the data that will turn out to be useful, robust, and generalizeable in a variety of downstream tasks. Recent progress in Self-Supervised Learning (SSL) has produced embedding models that begin to rival task-specific models in areas like vision and language-processing. However, SSL is an area where empirical results have consistently outpaced theoretical understanding, making some SSL models susceptible to surprising failure-modes. The next generation of representation-learning methods must build on the success of current methods while establishing firmer theoretical foundations. This project advances such a foundation in terms of probability and uncertainty quantification. The key idea is to draw on the recent development of "multicalibration," which provides theoretical guarantees for trustworthy and fair classifiers. However, multicalibration in its existing form is only applicable to supervised learning tasks where labels or regression targets are known. First, this project extends multicalibration to weakly-structured but unlabeled data, where it gives rise to a constrained optimization objective. This then naturally leads to a set of self-consistency constraints on the outputs of a representation-learning or embedding model. These self-consistency constraints closely resemble the kinds of heuristic learning objectives that have been empirically successful in SSL. This project therefore aims to design and train embedding models with new self-consistency constraints derived from (multi)calibration. The theoretical contribution will be to place SSL methods on more solid theoretical footing, namely by establishing a connection to probabilistic inference and representations of uncertainty. The practical contribution will be to create and release more trustworthy and robust embedding models to serve as a foundation for general downstream visual tasks. This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
Up to $174K
2027-08-31
Detailed requirements not yet analyzed
Have the NOFO? Paste it below for AI-powered requirement analysis.
One-time $49 fee · Includes AI drafting + templates + PDF export
Research Infrastructure: National Geophysical Facility (NGF): Advancing Earth Science Capabilities through Innovation - EAR Scope
NSF — up to $26.6M
AmLight: The Next Frontier Towards Discovery in the Americas and Africa
NSF — up to $9M
CREST Phase II Center for Complex Materials Design
NSF — up to $7.5M
EPSCoR CREST Phase I: Center for Energy Technologies
NSF — up to $7.5M
EPSCoR CREST Phase I: Center for Post-Transcriptional Regulation
NSF — up to $7.5M
EPSCoR CREST Phase I: Center for Semiconductors Research
NSF — up to $7.5M