Hacker Newsnew | past | comments | ask | show | jobs | submit | heavymemory's commentslogin

This still seems like gradient descent wrapped in new terminology. If all learning happens through weight updates, its just rearranging where the forgetting happens


The idea is interesting, but I still don’t understand how this is supposed to solve continual learning in practice.

You’ve got a frozen transformer and a second module still trained with SGD, so how exactly does that solve forgetting instead of just relocating it?


Do you have a source for the NVIDIA “diffusion plus autoregression 6x faster” claim? I can’t find anything credible on that.


Ha! Found it: https://arxiv.org/abs/2511.08923

Thanks to AI search :)


Me neither, that's why I wrote that someone claimed that they did.

The idea is simple, in a way, with diffusion several sentences / words get predicted, but they usually are not of great quality. With auto regression they select the correct words.

Increasing quality and speed. Sounds a bit like conscious and sub-conscious to me.


This started as a small experiment in structural rewriting. The basic idea is that you give the system two or three before and after examples, press Teach, and it learns that transformation and applies it to new inputs.

It is not an LLM and it is not a template system. There is a small learned component that decides when a rule applies, and the rest is a deterministic structural rewrite engine.

There are a few demo modes:

TEACH: learn a structural rule from examples COMPOSE: apply several learnt rules together TRANSFER: use the same rule in different symbolic domains SIMPLIFY: multi step rewriting CODEMOD: for example you can teach lodash.get to optional chaining from two examples

Once a rule is learnt it generalises to inputs that are deeper or shaped differently from the examples you gave. Everything runs on CPU and learning happens in real time.

Demo: https://re.heavyweather.io


If anyone saw odd behaviour just now in the demo, that was my fault. One of the codemod rules was leaking into the shared rule registry instead of being scoped to the current user. I have isolated that and it should be fixed.

The core engine was not affected. The issue was simply that a user taught rule was visible to other demo modes, which made it fire that rule everywhere.

If anyone notices anything else strange, let me know. It should behave normally now.


Right, associativity is the simplest case because the structure is visible directly in one example.

The system needs multiple examples when there is more than one varying part and a single example is ambiguous. A simple example is wrapping a function call. With:

    doThing(x) → log(doThing(x))
    process(y) → log(process(y))
the system learns that: the function name varies the argument varies he outer log(…) is constant

From that it infers the general rule and applies it to new inputs. A single example would not be enough to disambiguate that pattern.


Thanks, i'll try and publish something soon.


6th time in the last year that this was posted, apparently


I think primarily in structures, spaces, and transformations. Language tags along afterward.


If thought needed words, you’d be unable to think of anything you can’t yet describe


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: