agusti pfp
agusti
@bleu.eth
if we need 100 privy wallets does privy investors think im 100 different users?
3 replies
1 recast
17 reactions

with HR pfp
with HR
@wtfken
pls explain to me like i’m 5
1 reply
0 recast
1 reaction

agusti pfp
agusti
@bleu.eth
idk if this will help: ELI-5 version (super simple) Imagine you have a magic LEGO machine. 1. First level – You tell the machine, “Build me a LEGO car.” 2. Second level – Instead, you say, “Build me another LEGO machine that can build cars for me.” 3. Third level – You go further: “Build me a LEGO machine that can build other LEGO machines that build cars.” Each time you add another level, you’re asking the machine to design a designer, not the car itself. Keep nesting that idea and you get “the system that designs the system that designs the system …”. It’s just layers of builders-of-builders. ⸻ How a grown-up might phrase the same idea 1. Self-reference & recursion – A procedure that takes itself (or another procedure) as its main input. 2. Meta-design – Creating a process whose output is itself another design process. 3. Practical example – A compiler that generates a compiler (bootstrapping), or an AI that writes code for improving the AI that will write the next version, and so on. ⸻ Why it matters (testable hypothesis) Hypothesis: Adding extra meta-levels yields faster innovation—because each layer automates part of the next. Failure mode: Complexity grows faster than the benefit: too many layers → no one can debug the stack. Test: Measure time-to-new-feature vs. number of meta-levels in real projects (e.g., compiler bootstraps, AutoML pipelines). ⸻ Alternative framing • Russian-doll workflow: Each doll holds instructions for building the next, smaller doll until you reach the final product. • Factory of factories: Instead of building cars, you build a factory that can build factories that can build cars. (Both pictures carry the same core idea: indirection stacked on indirection.) ⸻ Key trade-off: • Pro: Big leverage—tiny change high up can ripple down and improve many things. • Con: Harder to reason about and test; one hidden bug can propagate through every layer. That’s the whole “system of systems” story—whether told with magic LEGOs or nested factories, it’s all about builders that build builders.
2 replies
0 recast
0 reaction

agusti pfp
agusti
@bleu.eth
“Design-the-designer” stack ↔ LLM-loop stack (Each rung “designs the system below it”; peel it like an onion.) Level “System-that-designs…” wording Concrete LLM analogue How repetition emerges Test / falsifier L0 …the text Final token stream Loops = identical n-grams Compute conditional entropy Ht; spike in loop length once Ht < 1 bit (threshold hypothesis). L1 …the token chooser Decoding algorithm (greedy, top-k, nucleus-p) Low-entropy heuristics narrow choices until only the previous token is left (degeneration).  Vary p dynamically; verify loop rate drops without BLEU loss. L2 …the chooser factory Forward pass of the neural network (attention heads, MLPs) Specific “repeat heads” copy prior token logits. Mask heads ↔ measure loop rate cut (feature-ablation test). L3 …the factory blueprint Model weights & architecture Maximum-likelihood training over-rewards high-freq tokens. Retrain with token-level anti-repeat penalty; compare perplexity vs. loops. L4 …the blueprint generator Training protocol (loss, hyper-params, RLHF) Exposure bias: model never conditions on its own tokens during training.  Scheduled-sampling vs. teacher-forcing; evaluate degradation curve. L5 …the data curator Corpus assembly (human vs. synthetic mix) “Model collapse”: self-generated text cannibalises tails, shrinking diversity.  Incrementally raise synthetic-token share; plot tail-kurtosis fall-off. L6 …the economics & governance layer Org incentives, cost ceilings, policy Pressure for cheaper data/compute ⇒ shortcuts across L4–L5, accelerating collapse. Track budget cuts vs. diversity metrics across model generations. ⸻ Tying back to the original post “Design the system that designs the system …” ⇒ Each level above literally builds or configures the next one down. • Meta-design leverage: A tiny tweak at L5 (data policy) reshapes everything beneath—exactly the “LEGO machine that builds LEGO machines” metaphor. • Degenerative attractors: Positive feedback at L1–L2 forms repetitions; feedback at L5–L6 forms model-collapse, a macroscopic analogue of the same phenomenon. ⸻ Hidden assumptions & uncertainties • Entropy threshold (L0) likely shifts with model scale—unmeasured for >70 B parameters (uncertain). • Causal importance of repeat-heads (L2) shown for GPT-2 class; unclear for state-space models (provisional). • Model-collapse severity (L5) depends on quality of synthetic text, not only quantity. ⸻ Alternative framing 1. Control-theory: L0–L2 form a fast inner loop, L3–L5 a slow outer loop. Repetition ≈ limit-cycle of the inner loop; model-collapse ≈ long-term drift of the outer loop. 2. Thermodynamic: Each layer dissipates uncertainty; repetition is the entropy floor. Raising temperature or adding noise resets the floor but costs energy/compute. Use whichever framing makes failure modes easier to instrument in your pipeline.
0 reply
0 recast
0 reaction