Day 1 vs Day 100: The transformation is real.
"Is it actually intelligent, or just stateful storage?"
Fair question. Most AI tools remember things. Few actually learn. Here's what makes aNewStream different.
This isn't a feature list. It's a closed feedback loop where every action improves future decisions.
When you start, the system knows nothing about you.
| Metric | Value |
|---|---|
| Entries in learning.md | 0 |
| Errors in blocklist | 0 |
| Trust score | 0 |
| Cross-product patterns | None |
Day 1 behavior:
Example: First price change
Event: SKU_UPDATED (price: $99 → $89)
1. STREAM records to findings.md
→ "Price changed 10% for WIDGET-PRO"
2. HQ reads findings.md (empty learning.md, empty errors.md)
→ Dispatches: ASK for analysis, SIMULATE for reaction
3. ASK analyzes with zero historical context
→ Generic response: "10% price drop may impact margins"
→ No cross-product patterns (no other products yet)
4. GENERATE creates announcement draft
→ Uses default "professional" tone
→ Action staged (pending approval, trust=0)
5. User approves → Trust score +2
→ System learns: user approved this type of action
After 100 days of continuous learning:
| Metric | Value |
|---|---|
| Active learnings (auto-pruned) | ~25 |
| Errors in blocklist | 10+ |
| High-confidence insights | 50+ |
| Trust score | 20+ (auto-approves) |
Day 100 behavior:
Same event, different response:
Event: SKU_UPDATED (price: $99 → $89)
1. STREAM records to findings.md
→ Error check: "Rate limit on Slack? No"
→ Cross-product: "Similar to GADGET-X price drop 2 weeks ago"
2. HQ reads ALL context:
→ learning.md: "Friday price drops are competitor-driven"
→ insights table: "GADGET-X price drop → 8% sales increase"
→ errors.md: "Don't retry Slack immediately"
Decision: SKIP ASK (pattern known), go straight to GENERATE
3. GENERATE creates announcement
→ Style from learning.md: "User prefers conservative tone"
→ Avoids: "aggressive language" (user correction from day 23)
→ Auto-stages with confidence: high
4. Auto-approve (trust score > threshold)
→ Draft published without user intervention
→ Trust score +2
5. NOTIFY sends to Slack
→ Checks errors.md: "Wait 60s between messages"
→ Respects rate limit
The Intelligence Layer has eight mechanisms working together:
1. Error Learning & Prevention
Before ANY action, agents check if similar actions failed before. Failed patterns go into a blocklist with what failed, why, and what to do instead. After 2+ failed attempts, patterns are auto-added. The system literally cannot repeat the same mistake.
2. Attention Manipulation
Exploits how LLMs pay most attention to the beginning and end of prompts. Every prompt wraps content with goals at start and end. Error warnings appear at high-attention positions. Agents don't forget their objective during long reasoning.
3. Cross-Product Pattern Detection
Uses Jaccard similarity to find patterns across all your products. Detects correlated events: "Price increase in Widget-A happened 3 days before Widget-B every time." Historical echoes: "This situation mirrors what happened to Product-X 90 days ago."
4. Temporal Reasoning
The system understands time:
5. Memory Summarization
When any agent's memory exceeds 20KB, it's automatically summarized using a cheap model. Original archived for reference. Agent-specific summarization preserves what matters. Prevents context bloat.
6. Insight Surfacing
Insights generated by agents are stored in a database. HQ queries for high-confidence insights on every run. Creates the closed feedback loop: Actions → Insights → DB → HQ reads → Better decisions.
7. Auto-Compaction
Learning files stay useful, not bloated:
8. Trust Score System
Starts at 0 — all actions require manual approval. +2 on successful execution, -1 on failure. When trust exceeds threshold, actions execute automatically. Per-action-type scoring: "publish_draft" can auto-approve while "send_notification" still requires review.
OBSERVE → STREAM/WIRE write to findings.md
LEDGER records transactions
↓
LEARN → Patterns saved to learning.md (auto-compacts)
Errors saved to errors.md
Cross-product patterns detected
↓
REMEMBER → memory.md per agent (auto-summarized)
Insights stored in DB
↓
CALCULATE → LEDGER: deterministic margin/revenue math
Baselines detect anomalies
↓
ACT → HQ reads all files + insights table
Checks errors.md BEFORE acting
Dispatches agents
↓
SHOW → intelligence.md updated for user
Dashboard shows velocity + learnings + trust
↓
FEEDBACK → User corrections improve learning.md
Approvals increase trust score
↓
(back to OBSERVE)
| Dimension | Day 1 | Day 100 |
|---|---|---|
| Error handling | Logs errors, retries blindly | Never repeats same mistake (blocklist) |
| Context awareness | Base prompts only | Reads learning + errors + insights + cross-product |
| Pattern detection | Single product, no history | Cross-product correlation + historical echoes |
| Approval workflow | 100% manual approval | Auto-approves trusted action types |
| Content generation | Default tone and style | Learned preferences from user corrections |
| Time awareness | None | Velocity, day-of-week, seasonal reasoning |
| Memory health | Empty | Auto-compacted, summarized, useful |
aNewStream's Intelligence Layer is genuinely intelligent. It's not just stateful storage — it's a closed feedback loop where every action improves future decisions.
The question isn't whether it learns. The question is how fast you want it to learn about your products.
AI tools are overdelivering on expectations — Lenny Rachitsky's survey of 1,750 PMs, engineers, designers, and founders. Key insight: teams want AI for upstream judgment (research, ideation, pressure-testing) not just downstream output.
Lenny Rachitsky on LinkedIn — Discussion thread with commentary from PMs, founders, and AI builders on where the real gaps are.
aNewStream Intelligence Layer Documentation — Technical specification for the 8-agent architecture and Intelligence Layer implementation.