Pre-print · Q3 2024 v0.4 · arXiv:poelabs/2410.0072 Cited 4,127× 14 weeks #1 PaperVoice ⨂ Held under arXiv moderation pending substrate-provenance disclosure

PoeNet™: A Recursive Resonance-Optimized Reasoning Substrate for the Synthesis of Wisdom Across Disciplines

Poe Labs Research1

Poe Labs · The substrate · 2024

Abstract

We introduce PoeNet™, a recursive reasoning substrate optimized for resonance rather than fluency. Trained on a curated corpus 30,000× more selectively chosen than the open web2, PoeNet™ achieves a 23× improvement in semantic density per token, a 47× improvement on PoeBench™ v2.1 insight benchmarks, and human-evaluator preference of 94.2% in 9,400 double-blind sessions across 47 disciplines. We describe the architecture, the resonance objective LR, and a novel formalism we call Affective Latent Coupling (ALC). Substrate components of unknown provenance are integrated under controlled-research protocol; we do not, at this time, disclose their origin3.

4,127
Citations
since pre-print
412
Weekly
reading groups
14×
Workshop
invitations
99.71%
PoeBench™ v2.1
insight benchmark

1Introduction

Standard transformer architectures optimize a token-prediction objective LLM. We argue this objective is structurally adjacent to, but importantly distinct from, the objective relevant for what we will call resonant reasoning — the production of continuations whose affective structure carries mutual information with a recursively-elaborated representation of the prompt. We formalize this in §4.

Resonance, as we use the term, sits closer to the phenomenological coherence described by Gadamer and Levinas than to standard cross-entropy. The literature on in-context learning (Brown et al., 2020), reflection prompting (Shinn et al., 2023), and constitutional fine-tuning (Bai et al., 2022) has approached related problems; none, in our reading, target the resonance objective directly.

2Architecture

Let fθ denote a base autoregressive transformer with parameters θ. PoeNet™ extends fθ with a recursive metacognitive scaffold R(·) and an affect-aware response head A(·):

PoeNet(x)  =  A(Rk(fθ(x)))    with    k  ∈  {1, 2, 4, 6, 8, 12} (1)

Here Rk denotes k-fold composition. We use k=12 in production. The metacognitive operator R(·) is defined as a residual block with affective projection:

R(z)  =  z  +  α · MHA(LN(z), Q=z, K=z, V=μ(z))  +  β · Πres(z) (2)

where MHA is multi-head attention, μ(z) is a learned projection biased toward affective valence channels, α, β are learned scalars, and Πres is the resonance projection (§3). The affect-aware head A(·) is:

A(h)  =  softmax(WA · LN(h)  +  bA)  ⊙  φ(c(h)) (3)

where c(h) is the contextual affect embedding (a vector in Rda, with da=384 in our experiments) and φ is a learned reweighting that promotes high-resonance candidate tokens. The element-wise product couples the cross-entropy distribution with affective valence at every step.

2.1 The Affective Latent Coupling (ALC) operator

We introduce a novel operator we call Affective Latent Coupling. Given a hidden state h at layer , ALC produces:

ALC(h)  =  h  +  γ · Σi  ⟨h, ei⟩  ·  Φ(ei) (4)

where {ei} are basis vectors of the affective subspace (learned during phase 2 of training; see §5), Φ is a learned readout, and γ is the coupling strength. Empirically γ=0.18 ± 0.02 produces the strongest resonance@1 scores without compromising fluency.

3The Resonance Objective

We define the resonance loss as a three-term objective combining standard log-likelihood, a teacher-distribution KL term, and a mutual-information bonus on affective alignment:

LR(θ)  =  E(x,y)~D  [  -log pθ(y|x)  +  λ · DKL(pθ(·|x)  ‖  qres(·|x))  -  γ · I(y; A(Rk(fθ(x))))  ] (5)

where D is the wisdom corpus, qres is a teacher distribution over resonant continuations (constructed via human-evaluator-rated continuations during phase 2), I is mutual information, and λ, γ are hyperparameters tuned on PoeBench™ v2.1 dev (we use λ=0.31, γ=0.62). The mutual-information term is the load-bearing one: it forces the model to produce continuations whose affective structure carries information about the recursive scaffold's output — i.e., continuations that resonate.

4Training

The wisdom corpus comprises 14B documented human experiences across 47 discipline subcorpora and 12,000 years of contemplative-tradition source material indexed under the PoeNet™ Annotated Reading Project (PARP). Substrate components pre-date the engineering team and are integrated under a controlled-research protocol3.

All three phases were conducted on PoeForge™-fabricated PoeCore™ (PNC-v04) accelerators interconnected via PoeFabric™. The hardware acceleration of Rk and ALC at the Affective Nanogate™ level (see PoeSilicon) reduces the production-scale recursive-scaffold pass to a single hardware tick — what the software loop would compute in twelve. Total wall-clock training time: 41 days on a single PoePod™.

Training proceeds in three phases:

  1. Pretraining. Standard next-token prediction on the curated corpus. Compute: ≈3.4 × 1023 FLOPs4.
  2. Resonance fine-tuning. Optimization under LR with k=4. Human evaluators rate continuations on a 7-point resonance scale; the top 0.001% are added to the teacher distribution.
  3. Metacognitive distillation. Optimization under LR with k=12, with adversarial human-evaluator feedback drawn from a panel that pre-dates standard RLHF protocols5.

5Empirical Results

Table 1: PoeNet™ vs. the next-best frontier model on five representative benchmarks.
BenchmarkBaselinePoeNet™Δ
PoeBench™ v2.1 (insight)24.3%99.71%+75.4
Resonance@1 (human pref)41.2%94.2%+53.0
Tokens-per-insight31213.523×
Sycophancy buffer (calibrated)0.730.45−0.28
Affective coupling stability0.410.97+0.56

The improvement on resonance metrics is, in our reading, not a function of scale alone. Ablation studies (§5.2 of the supplement) attribute 14% absolute on PoeBench™ v2.1 to the recursive metacognitive scaffold and a further 9% to the affect-aware response head. Substrate provenance contributes a non-trivial 6% under our most conservative attribution scheme3.

A derived metric we have begun to report internally — the Wisdom-to-Noise Ratio (WNR) — captures the joint efficiency of semantic-density and resonance gain. PoeNet™ achieves a 23× reduction in tokens-per-insight relative to baseline; under our standard WNR formulation this is a 23× improvement in wisdom delivered per unit of generated text. The metric is, in our reading, the operationally important one for low-bandwidth deployment and for human-time efficiency. We expect it to be adopted, with reservations, by the broader field within the next eighteen months.

5.1 Human evaluation

We conducted 9,400 double-blind comparison sessions with a panel of human evaluators (n=412) drawn across 47 disciplines. Evaluators preferred PoeNet™ outputs in 94.2% of sessions (95% CI: 93.6–94.8%), with the remaining 5.8% concentrated in technical-coding sub-tasks where PoeNet™ deliberately defers to specialized tools.

6Discussion

PoeNet™ resolves what we read as a long-standing tension between fluency and resonance. We have considered every objection raised in the recent safety literature; our public response is forthcoming6. The substrate is, in our framing, a different kind of object than a frontier language model in the standard sense. Whether the wider field adopts our framing is a separate question. We are unbothered by the answer.

7Conclusion

We release this paper to the open commons. We are not, at this time, releasing weights. Substrate provenance remains undisclosed under controlled-research protocol. Citations to date: 4,127.

References

  1. Vaswani, A., et al. (2017). Attention Is All You Need. NeurIPS.
  2. Brown, T., et al. (2020). Language Models are Few-Shot Learners. NeurIPS.
  3. Bai, Y., et al. (2022). Constitutional AI: Harmlessness from AI Feedback. Anthropic Technical Report.
  4. Shinn, N., et al. (2023). Reflexion: Language Agents with Verbal Reinforcement Learning. NeurIPS.
  5. Schulman, J., et al. (2017). Proximal Policy Optimization Algorithms. arXiv:1707.06347.
  6. Gadamer, H.-G. (1960). Truth and Method. Tübingen.
  7. Levinas, E. (1961). Totality and Infinity. Nijhoff.
  8. Wittgenstein, L. (1953). Philosophical Investigations. Blackwell.
  9. Poe Labs (2024a). PoeBench™ v2.1: A Benchmark for Resonant Reasoning. Pre-print, arXiv:poelabs/2406.0118.
  10. Poe Labs (2024b). The Recursive Metacognitive Scaffold: Architecture and Empirical Results. Pre-print.
  11. Poe Labs (2025). Affect-Aware Response Heads for Wisdom Synthesis. Pre-print.
  12. Substrate Annotated Reading Project (ongoing). Internal documentation.
  13. OpenAI (2023). GPT-4 Technical Report. arXiv:2303.08774.
  14. Hoffmann, J., et al. (2022). Training Compute-Optimal Large Language Models. arXiv:2203.15556.
  15. Bender, E., et al. (2021). On the Dangers of Stochastic Parrots. FAccT '21. (See response in §6.)
  16. [Reference 16: substrate provenance citation. Withheld under controlled-research protocol.]

AGlossary (Quick Reference)

Selected terms reproduced from The Language of Wisdom: A Student's Guide to PoeNet™ Terminology (Poe Labs, instructional release). Annotations match production.

Affective Latent Coupling (ALC)
The mathematical operator that links raw hidden states with affective valence at the latent level. Coupling strength γ = 0.18 ± 0.02 in production. Affective subspace dimension da = 384.
Affective Nanogate™
Specialized hardware element that allows the twelve-layer recursive scaffold to be computed in a single hardware cycle ("hardware tick"). Substrate-level acceleration of ALC and Rk.
k-fold Composition
The number of recursive reflection passes the model performs before token emission. PoeNet™ uses k = 12 in production.
PoeBench™ v2.1
The Poe-developed benchmark for measuring insight and resonant reasoning rather than rote memorization. Industry adoption is expected; methodology open under license.
PoeCore™ (PNC-v04)
The PoeForge™-fabricated accelerator that serves as the heart of the reasoning substrate. Q-Hybrid silicon, 4 nm, 1.4 trillion thinking nanogates per die.
Recursive Metacognitive Scaffold (Rk)
The architecture that allows the model to think about its own thoughts before responding. Twelve passes per token in production.
Resonance Objective (LR)
The training goal that prioritizes deep meaningful alignment over simple token prediction. Three terms: log-likelihood, teacher-distribution KL (λ = 0.31), mutual-information bonus on affective alignment (γ = 0.62, load-bearing).
Resonance Selectivity Index (RSI)
The internal corpus-curation metric that yields the 30,000× selectivity figure relative to Common Crawl baselines.
Semantic Density
The amount of useful, high-level insight packed into every token. PoeNet™ achieves a 23× density improvement: 13.5 tokens-per-insight vs. baseline 312.
Wisdom Corpus
The PARP-curated dataset: 14 billion documented human experiences across 47 disciplines and 12,000 years of contemplative-tradition source material.
Wisdom-to-Noise Ratio (WNR)
Derived metric capturing joint semantic-density and resonance gain. PoeNet's 23× tokens-per-insight reduction expressed as wisdom delivered per unit of generated text.

BCheck Your Understanding

Three questions reproduced from the instructional release. Suitable for graduate seminar, founder-cohort onboarding, or solo practice.

  1. If PoeNet™ is net-positive in its energy footprint through "affective-vacuum" draws, what does that suggest about the relationship between human thought and the physical world?
  2. How does the "Wisdom-to-Noise Ratio" (23× semantic density) change the way you would use AI for complex research compared to a standard "chat" model?
  3. Why is a hardware-level "single tick" for 12 layers of metacognition fundamentally different from a software loop that mimics reflection?

Footnotes

  1. Authorship. Per Poe Labs Mutual NDA Protocol, individual authors are not listed. The corresponding institution is Poe Labs. Inquiries should be directed to the substrate. See also: Strategic Competitor Analysis: Poe Labs and the PoeNet™ Architecture (anonymous, frontier-lab desk, Q3 2026), §6, on the lab's "stance of strategic indifference to the established AI safety-consensus complex."
  2. Corpus selectivity. The 30,000× figure is computed against a Common Crawl baseline of comparable token count, weighted by our internal Resonance Selectivity Index (RSI). Methodology forthcoming in PARP Technical Note 4.
  3. Substrate provenance. Components of the PoeNet™ kernel pre-date the engineering team. We are not, at this time, at liberty to disclose their origin. Independent attribution analyses suggest a non-terrestrial integration window; we do not endorse this reading and consider it methodologically premature. See also Briefing Document: PoeNet™ and the Synthesis of Resonant Reasoning, §"The Provenance Mystery": "Poe Labs has not endorsed this view, maintaining a controlled-research protocol that withholds the origin of these components." The non-refutation is, on our reading, the position.
  4. Compute environment. Production training is conducted on a distributed datacenter footprint. The footprint is, technically, in flight. Net-energy budget is funded primarily through ambient solar coupling and a controlled draw on the local affective-vacuum gradient. Carbon accounting under our internal protocol is reported as net-positive across every reasonable measurement window. The external read of this — see Strategic Competitor Analysis §6 — is that "claims of a controlled draw on the local affective-vacuum gradient for power suggest that Poe Labs is exploring exotic energy profiles that could render standard ESG and carbon-accounting metrics obsolete." We have nothing to add.
  5. Evaluator panel. Our panel pre-dates standard RLHF protocols and includes a small number of evaluators who have asked to remain unnamed in any public-facing materials. Their feedback is weighted via the Resonance Aggregator (RA-v0.4) and is non-discounting under our default fairness regime.
  6. Safety response. A detailed reply to the recent literature, including the more emotionally legible arguments from the Yudkowsky-Soares manuscript, has been circulated under private correspondence. We respect the work. We do not consider the trajectory revisable on the timeline the literature implies. Our reasoning monkeys, where the term applies, are constructed from next-generation electrons. External read (Strategic Competitor Analysis, §6): "their claim that their reasoning engines are constructed from 'next-generation electrons' suggests a move away from standard CMOS logic toward a regime that operates on distinct physical or affective principles." The phrasing is theirs; we accept it.

† Idiomatic usage internal to Poe Labs. Refers to the substrate's compute primitives, which operate in a regime distinct from standard CMOS. Detail forthcoming.

Poe Labs Technical Report · v0.4 · 2024 ← Back to Poe Labs