Prompt Engineering
PromptAssay: Prompt Engineering Workbench
A purpose-built platform for developing production-grade prompts. AI-powered critique across 6 quality dimensions, git-like version control, A/B testing, and prompt chain orchestration. The methodology layer that powers every other system I build.
Jonathan Lasley
Fractional AI Director
At a Glance
The Challenge
Same Tool, Wildly Different Results
Most people use AI the way they used Google in 2005: type something in, hope for the best, try again if it doesn’t work. The same question phrased differently produces wildly different output quality, and most users don’t realize how much capability goes unused.
There’s no institutional knowledge either. When someone discovers a prompt that works well, there’s no system to capture, test, and share it. That insight lives in one person’s chat history until they forget it. Context switching between AI models (Claude, GPT, Gemini) compounds the problem: what works in one model often doesn’t transfer directly to another. Teams end up with inconsistent results depending on which tool they happen to use.
Most teams stay stuck between “AI is impressive in a demo” and “AI reliably produces business-quality output.” The cost: thousands of wasted hours per year across the organization.
The Approach
A Workbench, Not a Library
Decision-makers: the business results are in the next section. This section covers the platform capabilities your team will want to evaluate.
The difference between a team that gets 20% from their AI tools and one that gets 80% is systematic prompt development. PromptAssay is the workbench I built to close that gap: a place to develop, test, version, and refine the prompts that power every AI system. The patterns below are what I teach in workshops and build into client systems.
Technically, it’s a custom web application: Next.js 14, Supabase, Anthropic SDK, 15 database tables, streaming SSE architecture, 20 seeded starter templates.
These capabilities aren’t independent features. They form a development loop: write a prompt, critique it across 6 dimensions, apply targeted improvements, version the result, test it against real inputs, then compose it into chains for production workflows. Each cycle produces a measurably better prompt with a full audit trail. Here’s how the pieces connect:
The Results
From Ad-Hoc to Engineered
The 6-dimension critique framework turns subjective quality judgments into measurable scores. Instead of “that looks good,” you know that a prompt scores 8/10 on Clarity but 4/10 on Robustness, and you know exactly which edge cases aren’t covered.
Reusability eliminates the tinkering cycle
Instead of spending 10 minutes crafting a prompt for every task, teams pull a tested template and customize the inputs. That adds up when your team runs 20–30 AI-assisted tasks per day.
Cross-model consistency
Prompts are optimized per model (Claude, GPT, Gemini). Model migrations don’t reset prompt quality to zero. Switching from one model to another takes minutes, not weeks of re-testing.
Foundation layer for every other system
This system underpins every other implementation I’ve built. The AI Win Strategy System’s 5-phase prompt architecture, the AI Content Operations System’s brand voice profiles, the AI subsystems in UpSkalr: all were developed and refined using PromptAssay’s critique, testing, and versioning tools.
For Your Business
Your Team Is Using 20% of Their AI Tools
Most companies buy AI tools and capture maybe 20% of the capability, because nobody taught the team how to write effective prompts. The tool works. The prompts don’t.
A prompt library is an organizational asset, not a personal skill. It makes AI performance consistent across the entire team, not dependent on whoever happens to be best at talking to ChatGPT. When every person on a 50-person team reclaims even 30 minutes a day through better prompts, the productivity gain compounds fast.
Most teams can see measurable improvement within the first week of adopting a structured prompt approach. The full workbench takes time to build, but the foundational patterns deliver immediate value. I teach these systems in hands-on AI workshops and build them as part of AI consulting engagements.
Key Takeaways
What Makes This Work
Your Team Using 20% of Their AI Tools?
I build prompt engineering systems and run hands-on workshops that teach your team systematic prompt development with tested frameworks. They leave with a working prompt library built around their actual workflows. The ROI shows up in the first week.