No HTTPS in 2026. False origin that suggest a massive improvement. Leaderboard doesn't work. Instructions are "repeatedly download this code and execute it on your machine". No way to see the actual changes being made.
We can do better than this as an industry, or at least we used to be better at this. Where's the taste?
I was exploring how to parallelize autoresearch workers. The idea is to have a trusted pool of workers who can verify contributions from a much larger untrusted pool. It's backed bit a naked git repo and a sqlite with a simple go server. It's a bit like block chain in that blocks = commits, proof of work = finding a lower val_bpb commit, and reward = place on the leaderboard. I wouldn't push the analogy too far. It's something I'm experimenting with but I didn't release it yet (except for briefly) because it's not sufficiently simple/canonical. The core problem is how to neatly and in a general way organize individual autoresearch threads into swarms, inspired by SETI@Home, or Folding@Home, etc.
Yeah you can sink a lot of time into a system like that[0].
I spend the years simplifying the custom graph database underneath it all and only recently started building it into tools that an agent can actually call[2]. But so far all the groundwork has actually paid off, the rooster basically paints itself.
I found a wiki to be a surprisingly powerful tool for an agent to have.
And building a bunch of CLI tools that all interconnect on the same knowledge graph substrate has also had a nice compounding effect. (The agent turns themselves are actually stored in the same system, but I haven't gotten around to use that for cool self-referential meta reasoning capabilities.)
Have you thought about ways to include the sessions / reasoning traces from agents into this storage layer? I can imagine giving an rag system on top of that + LLM publications could help future agents figure out how to get around problems that previous runs ran into.
Could serve as an annealing step - trying a different earlier branch in reasoning if new information increases the value of that path.
I tried to copy the instruction and pasted in Note to see what it said, but I could not. Either the clipboard was empty or something prevented Note recognized it as just text.
It worked for me, try again. But it is still not fully crear to me what this is supposed to do, nor if this is doing better than a random search. It looks like it is about optimizing a GPT architecture.
I'm curious what a "stripped down version" of Github can offer in terms of functionality that Github does not? Is it not simpler to have the agents register as Github repos since the infrastructure is already in place?
We can do better than this as an industry, or at least we used to be better at this. Where's the taste?
I found a wiki to be a surprisingly powerful tool for an agent to have. And building a bunch of CLI tools that all interconnect on the same knowledge graph substrate has also had a nice compounding effect. (The agent turns themselves are actually stored in the same system, but I haven't gotten around to use that for cool self-referential meta reasoning capabilities.)
1: https://github.com/triblespace/triblespace-rs
2: https://github.com/triblespace/playground/tree/main/facultie...
Could serve as an annealing step - trying a different earlier branch in reasoning if new information increases the value of that path.
https://github.com/karpathy/autoresearch/pull/92
https://www.ensue-network.ai/autoresearch
Anyway, "1980 experiments, 6 improvements" makes me wonder if this is better than a random search or some simple heuristic.
`curl -L https://mycoolsvc.com/r4nd0mus3r/mycoolsoftware/master/insta... | bash`