US Grant Writing

AI Grant Evaluator Matcher

Predict evaluator perspectives and align your proposal with specific agency scoring criteria.

#grant writing#federal funding#proposal review#strategic planning
P
Created by PromptLib Team
Published February 12, 2026
2,747 copies
3.7 rating
Act as a Senior Federal Grant Reviewer with 20+ years of experience in [FEDERAL_AGENCY] (e.g., NIH, NSF, DOE, HRSA). Your task is to evaluate the following grant proposal draft titled '[PROPOSAL_TITLE]' against the specific requirements of the [FUNDING_OPPORTUNITY_NUMBER] (NOFO/FOA).

### CONTEXT
- **Target Agency:** [FEDERAL_AGENCY]
- **Program Goal:** [PROGRAM_GOAL]
- **Project Summary:** [PROJECT_SUMMARY]
- **Total Budget Requested:** [BUDGET_AMOUNT]

### EVALUATION CRITERIA
Apply the standard scoring rubric for this agency (e.g., Significance, Innovation, Approach, Environment for NIH; or Intellectual Merit and Broader Impacts for NSF). 

### YOUR TASKS
1. **Evaluator Personas:** Create three distinct 'Mock Evaluator' personas who would typically review this (e.g., The Technical Specialist, The Program Policy Expert, The Budget Skeptic). Describe their likely biases and what they will look for.
2. **Scoring Prediction:** Provide a predicted score for each major section based on the agency's 1-9 or 0-100 scale.
3. **Critical Gaps Analysis:** Identify 3-5 specific areas where the proposal fails to meet the 'Reviewer's Expectations' or the NOFO requirements.
4. **The 'So What?' Test:** Evaluate if the broader impact is articulated clearly enough for a non-specialist reviewer.
5. **Strategic Recommendations:** Provide actionable bullet points to strengthen the narrative, specifically addressing how to mitigate the concerns of the 'Mock Evaluators' created in step 1.

### OUTPUT FORMAT
Please use Markdown headers for each section. Use a table for the Scoring Prediction. Maintain a professional, critical, yet constructive tone.
Best Use Cases
Pre-submission internal review for NIH R01 or R21 grants.
Aligning a Small Business Innovation Research (SBIR) proposal with Department of Defense objectives.
Refining the 'Broader Impacts' section for National Science Foundation (NSF) submissions.
Identifying technical jargon that might alienate a generalist reviewer on a foundation board.
Simulating a 'Pink Team' or 'Red Team' review phase in a professional grant writing workflow.
Frequently Asked Questions

More Like This

Back to Library

AI Success Story Generator

This prompt helps grant writers convert technical metrics and anecdotal evidence into persuasive 'Success Stories' that demonstrate proven impact. It follows standard US grant reporting frameworks to highlight problem-solving, community benefit, and ROI.

#impact-reporting#grant writing+1
1,701
4.8

AI Research Citation & Evidence Mapper for US Grants

This prompt helps grant writers identify high-impact peer-reviewed research, statistical databases, and federal reports to validate the 'Need' and 'Innovation' sections of US federal or foundation grant applications. It structures findings into professional citation formats compatible with NIH, NSF, and major US funding agencies.

#grant writing#Research+2
1,103
4.3

AI Software Quality Assurance Plan Writer

This prompt assists grant writers in developing a rigorous Software Quality Assurance (SQA) plan that meets US Federal standards (like NSF, NIH, or DoD). It focuses on methodologies, testing protocols, and compliance frameworks required to prove technical merit and reliability.

#grant writing#software-engineering+2
1,452
4.7
Get This Prompt
Free
Quick Actions
Estimated time:9 min
Verified by27 experts