The Six Frequencies Every Prompt Must Carry (Or the AI Fills In the Blanks)
Table of Contents
The Core Thesis
Every prompt you write is a signal. That signal carries information across 6 distinct frequency bands. If any band is missing, the LLM reconstructs it from its training distribution — producing artifacts that look like answers but are fabrications. This is identical to aliasing in signal processing: when you undersample a waveform, you get phantom frequencies that were never in the original.
My sinc-prompt specification defines these 6 bands based on my empirical research across 1 million simulations and 275 production observations. They are not arbitrary categories. They are the 6 information dimensions that an LLM requires to reconstruct your intent without guessing.
The Six Bands Defined
Band 0: PERSONA (Weight: 12.1%)
Who should answer this question? Not just a role label — the expertise domain, the perspective, the communication style, the assumed knowledge level. When PERSONA is missing, the model defaults to "helpful AI assistant" — the most generic, least useful voice in its repertoire.
Signal example: "You are a principal security engineer at a FAANG company reviewing code for production deployment. You communicate findings as severity-ranked issues with remediation steps. You assume the reader is a senior developer."
Absence cost: The model guesses expertise level, tone, and assumed reader knowledge. Each guess introduces error.
Band 1: CONTEXT (Weight: 9.8%)
What situation are we in? What has happened before this prompt? What is the environment, the project, the constraints of the situation? CONTEXT provides the background the model needs to scope its response correctly.
Signal example: "We are in sprint 14 of a 16-sprint migration from a monolith to microservices. 8 of 12 services are deployed. The remaining 4 share a database that cannot be split until all are migrated. Production traffic: 2,400 requests per second."
Absence cost: The model invents a context from training averages. Your specific situation is lost.
Band 2: DATA (Weight: 6.3%)
What specific inputs, numbers, facts, references, and measurements does the model need? DATA is the factual foundation that prevents the model from generating plausible-sounding but invented statistics.
Signal example: "Current database: PostgreSQL 14. Rows: 180M in the largest table. Indexes: 23, of which 7 are unused (verified via pg_stat_user_indexes). Average query time: 340ms. P99: 4.7 seconds."
Absence cost: The model invents numbers. This is the most visible form of hallucination and the easiest to prevent.
Band 3: CONSTRAINTS (Weight: 42.7%)
What rules, boundaries, prohibitions, requirements, and limitations apply? CONSTRAINTS is the most important band — nearly half of output quality depends on it. This is the band that most people skip entirely.
Signal example: "Never suggest adding hardware. Budget for tooling changes: $0. Maximum acceptable query time: 500ms for p95. All changes must be reversible within 1 hour. Do not modify any table with active writes during business hours (8am-8pm EST). Do not recommend ORM changes."
Absence cost: The model has no boundaries. Every possible approach is equally valid. It picks the statistically average recommendation, which is rarely your best option.
Band 4: FORMAT (Weight: 26.3%)
What shape should the output take? Sections, tables, code blocks, numbered steps, paragraphs, JSON? FORMAT eliminates an entire dimension of model guessing and ensures consistent output structure.
Signal example: "Return a numbered list of exactly 5 optimizations. Each item: (1) one-line description, (2) exact SQL or configuration change, (3) expected improvement percentage, (4) risk level (LOW/MEDIUM/HIGH), (5) estimated implementation time."
Absence cost: The model chooses a random output structure. Different runs produce different formats. Downstream parsing breaks.
Band 5: TASK (Weight: 2.8%)
What is the actual objective? What do you want the model to produce? TASK is the only band most people include — and it accounts for the least quality impact because it merely states the destination without specifying how to get there.
Signal example: "Identify the 5 highest-impact PostgreSQL performance optimizations for our production database and provide implementation-ready fixes."
Absence cost: Minimal, since most people do state their task. The problem is that they state only the task.
What Happens When Bands Are Missing
The Nyquist-Shannon sampling theorem states that to perfectly reconstruct a signal, you must sample at twice its highest frequency. For a 6-band prompt signal, you need at minimum 6 samples — one per band.
When you provide only the TASK band (1 of 6), you are sampling at 1/6 the Nyquist rate. The reconstruction is guaranteed to alias. This is not a risk. It is a mathematical certainty.
| Bands Provided | Sampling Rate | Reconstruction Quality | Aliasing Level |
|---|---|---|---|
| 1 of 6 | 16.7% Nyquist | Catastrophic | 83% fabrication |
| 2 of 6 | 33.3% Nyquist | Poor | 67% fabrication |
| 3 of 6 | 50% Nyquist | Marginal | 50% fabrication |
| 4 of 6 | 66.7% Nyquist | Acceptable | 33% fabrication |
| 5 of 6 | 83.3% Nyquist | Good | 17% fabrication |
| 6 of 6 | 100% Nyquist | Clean | <5% fabrication |
Aliasing: The Technical Term for Hallucination
In signal processing, aliasing produces phantom frequencies — signals that appear real but are artifacts of undersampling. In LLM prompting, aliasing produces phantom facts — outputs that appear authoritative but are artifacts of missing specification bands.
The word "hallucination" is a projection from human cognition. The correct term is specification aliasing. Understanding this distinction changes how you fix the problem: you do not need to fix the model (anti-hallucination research). You need to fix the input (add the missing bands).
Band Interaction Effects
Bands do not contribute independently. My research found significant interaction effects:
- CONSTRAINTS + FORMAT together contribute more than their individual weights suggest. When both are present, the model's probability space is collapsed on two dimensions simultaneously, producing a multiplicative quality improvement.
- PERSONA + CONSTRAINTS interact to set the expertise level at which constraints are interpreted. A "junior developer" persona with "production-grade" constraints produces different output than a "senior architect" with the same constraints.
- CONTEXT + DATA together eliminate the model's need to infer the situation. Without both, the model must decide what is relevant — with both, relevance is predetermined.
The Optimal Token Allocation
Given a token budget of 200-230 tokens (the optimal range I identified), the allocation should match the quality weights:
| Band | Quality Weight | Token Allocation | Target Tokens |
|---|---|---|---|
| CONSTRAINTS | 42.7% | 42-45% | 85-100 tokens |
| FORMAT | 26.3% | 25-28% | 50-60 tokens |
| PERSONA | 12.1% | 10-12% | 20-25 tokens |
| CONTEXT | 9.8% | 10-12% | 20-25 tokens |
| DATA | 6.3% | 8-10% | 15-20 tokens |
| TASK | 2.8% | 3-5% | 5-10 tokens |
Notice the inversion from common practice: people spend 80% of their prompt on TASK and 0% on CONSTRAINTS. The optimal allocation is 3-5% on TASK and 42-45% on CONSTRAINTS.
Practical Application
I built the sinc-LLM tool at sincllm.com to automatically decompose any raw prompt into 6 bands, compute the SNR, identify missing bands, and generate the structured prompt. You do not need to memorize these weights or manually construct JSON. Paste your raw prompt, get a 6-band signal back.
The math does not care whether you believe in my framework. A 6-band prompt produces better output than a 1-band prompt because 6 samples reconstruct a signal better than 1 sample. This is the Nyquist theorem, and it has been proven correct for 75 years.
Transform any prompt into 6 Nyquist-compliant bands
Try sinc-LLM FreeOr install: pip install sinc-llm