NSF requires disclosure of AI tool usage in proposal preparation. Ensure you disclose the use of FindGrants' AI drafting in your application.
NSF
Scientists are increasingly incorporating machine learning (ML) and artificial intelligence (AI) techniques into their applications to accelerate and enhance scientific research and discovery across a wide range of disciplines. For example, machine learning has been successfully integrated into tools for weather forecasting, earth sciences, astronomy, high-resolution imaging, genomics, and molecular biology. However, the ever-growing size of scientific datasets results in prohibitive hardware resource costs, significantly complicating the deployment of these applications on high-performance computing platforms at scale. Lossy compression — a data reduction technique that significantly reduces dataset size by removing redundant or less important information — has proven effective for many scientific datasets, including those from cosmology and structural biology. Despite its promise, integrating lossy compression into AI-driven scientific applications remains a non-trivial challenge, requiring broad expertise in data compression and machine learning, as well as a deep understanding of application requirements, system considerations, and their interactions. These complexities hinder the adoption of this powerful data reduction technique in scientific applications. The overarching goal of this project is to address this gap by providing a cyberinfrastructure that seamlessly and adaptively integrates lossy compression into deep learning pipelines within scientific applications. This integration will reduce memory usage and communication overhead, enabling AI-for-Science applications to scale to massive datasets. The design includes several key innovations. First, it features a user-friendly interface that allows users to define accuracy requirements — which may evolve during application execution — and to instantiate different compressors, supporting both customization and extensibility. Second, it provides a software layer that integrates with popular deep learning frameworks, such as PyTorch, enabling compression to be applied to existing neural network models with minimal code modifications. Third, it incorporates an adaptive execution engine that dynamically selects the appropriate compressors and error bounds based on the desired accuracy, data characteristics (e.g., smoothness, value range, sparsity), model structure, and system configuration. The cyberinfrastructure will support both existing and emerging machine learning accelerators and will be released as open-source software, accompanied by documentation and training materials to promote adoption within the scientific and computing communities. Ultimately, this project has the potential to benefit the broader community by enabling scalable, AI-driven scientific discovery. This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
Up to $300K
2028-06-30
Detailed requirements not yet analyzed
Have the NOFO? Paste it below for AI-powered requirement analysis.
One-time $49 fee · Includes AI drafting + templates + PDF export
Research Infrastructure: National Geophysical Facility (NGF): Advancing Earth Science Capabilities through Innovation - EAR Scope
NSF — up to $26.6M
AmLight: The Next Frontier Towards Discovery in the Americas and Africa
NSF — up to $9M
CREST Phase II Center for Complex Materials Design
NSF — up to $7.5M
EPSCoR CREST Phase I: Center for Energy Technologies
NSF — up to $7.5M
EPSCoR CREST Phase I: Center for Post-Transcriptional Regulation
NSF — up to $7.5M
EPSCoR CREST Phase I: Center for Semiconductors Research
NSF — up to $7.5M