from Hacker News

My failed attempt at AGI on the Tokio Runtime

by openquery on 12/26/24, 4:22 PM with 40 comments

  • by dhruvdh on 12/26/24, 4:52 PM

    I wish more people would just try to do things just like this and blog about their failures.

    > The published version of a proof is always condensed. And even if you take all the math that has been published in the history of mankind, it’s still small compared to what these models are trained on.

    > And people only publish the success stories. The data that are really precious are from when someone tries something, and it doesn’t quite work, but they know how to fix it. But they only publish the successful thing, not the process.

    - Terence Tao (https://www.scientificamerican.com/article/ai-will-become-ma...)

    Personally, I think failures on their own are valuable. Others can come in and branch off from a decision you made that instead leads to success. Maybe the idea can be applied to a different domain. Maybe your failure clarified something for someone.

  • by markisus on 12/26/24, 4:59 PM

    > The only hope I have is to try something completely novel

    I don’t think this is true. Neural networks were not completely novel when they started to work. Someone just used a novel piece — the gpu. Whatever the next thing is, it will probably be a remix of preexisting components.

  • by namero999 on 12/26/24, 8:19 PM

    Isn't this self-refuting? From the article:

    > Assume you are racing a Formula 1 car. You are in last place. You are a worse driver in a worse car. If you follow the same strategy as the cars in front of you, pit at the same time and choose the same tires, you will certainly lose. The only chance you have is to pick a different strategy.

    So why model brains and neurons at all? You are outgunned by at least 300.000 thousand years of evolution and 117 billion training sessions.

  • by skeledrew on 12/26/24, 10:28 PM

    Interesting. I started a somewhat conceptually similar project several months ago. For me though, the main motivation is that I think there's something fundamentally wrong with the current method of using matrix math for weight calculation and representation. I'm taking the approach that the very core of how neurons work is inherently binary, and should remain that way. My basic thesis is that it should reduce computational requirements, and lead to something more generic. So I set out to build something that takes an array of booleans (the upstream neurons either fired or didn't fire at a particular time sequence), and gives a single boolean calculated with a customizable activator function.

    Project is currently on ice as after I created something that builds a network of layers, but ran into a wall figuring out how to have that network wire itself over time and become representative of whatever it's learned. I'll take some time and go through this, see what it may spark and try to start working on mine again.

  • by mindcrime on 12/26/24, 11:55 PM

    I haven't read the comments here yet, but I'm predicting there will be at least a few of the form "why would you bother doing this, you aren't an expert in AI, this is stupid, leave AGI to the experts, why would you think this could possibly work" etc. I hope not, but this being HN, history suggests those people will be out en-force.

    I hope not. I think this is GREAT work even if the result was ultimately less than what was desired. And I want to encourage the author, and other people who might make similar attempts. I think we need more people "taking a stab" and trying different ideas. You might or might not succeed, but in almost every case the absolute worst scenario is that you learn something that might be useful later. If taking on something like this motivates someone to spend time studying differential equations, then I say "great!" Or if it motivates someone to study neuroscience, or electronics (maybe somebody decides to try realizing a neural network in purpose built hardware, for example) then also "Great!" Do it.

    About the only serious negative (aside from allusions to opportunity cost) that I can see for making an effort like this, would be if somebody gets really deep in it and winds up blowing a shit-ton of money on the project, whether that be for cloud compute cycles, custom hardware, or whatever. I wouldn't necessarily recommend maxing out your credit cards and draining your retirement account unless you have VERY solid evidence that you're on the right path!

    You are a worse driver in a worse car. If you follow the same strategy as the cars in front of you, pit at the same time and choose the same tires, you will certainly lose. The only chance you have is to pick a different strategy.

    Yes, exactly. I adhere to a similar mindset. I do AI research in my spare time. And I cannot possibly afford to spend the kind of money on training ginormous ANN's that OpenAI, Microsoft, Google, Twitter, Meta, IBM, etc. can spend. To even try would be completely ludicrous. There is simply no path where an independent solo researcher can beat those guys playing that game. So the only recourse is to change the rules and play a different game. That's no guarantee of success of course, but I'll take a tiny, even minuscule, chance of achieving something over just ramming my head into the wall over and over again in some Sisyphean attempt to compete head to head in a game I know a priori that I simply cannot win.

    Anyway.. to the OP: great work, and thanks for sharing. And I hope you decide to make other attempts in the future, and share those results as well. Likewise to anybody else who has considered trying something like this.

  • by Onavo on 12/26/24, 9:56 PM

    > Ok how the hell do we train this thing? Stochastic gradient descent with back-propagation won't work here (or if it does I have no idea how to implement it).

    What's wrong with gradient descent?

    https://snntorch.readthedocs.io/en/latest/

  • by whatever1 on 12/27/24, 12:36 AM

    Try fewer neurons and solve it to global optimality with gurobi. This way you will know if the optimization step was your bottleneck.
  • by cglan on 12/26/24, 4:43 PM

    I’ve thought of something like this for a while, I’m very interested in where this goes.

    A highly async actor model is something I’ve wanted to explore, and combined with a highly multi core architecture but clocked very very low, it seems like it could be power efficient too.

    I was considering using go + channels for this

  • by upghost on 12/27/24, 1:16 PM

    This. I love everything about this. Everything. Thank you. This is the kind of sh!t that made me get into programming in the first place.

    Regarding specific reading, three books I think you would love are [1] the self assembling brain, [2] the archaelogy of the mind, and [3] evolutionary optimization algorithms.

    People can talk whatever sh!t they want but this pushed us closer to actual AGI than anything this (useful but) deadend LLM craze is pushing us towards, and towards which you thoughtfully made an effort.

    The most basic function of learning and intelligence is habituation to stimuli, which even an ameoba can handle but not a single LLM does.

    Thanks again for this.

    [1]: https://a.co/d/4TG1ZvP

    [2]: https://a.co/d/aYReWjs

    [3]: https://a.co/d/1cod8Bq

  • by henning on 12/26/24, 10:26 PM

    The author could first reproduce models and results from papers before trying to extend that work. Starting with something working helps.
  • by alecst on 12/26/24, 9:05 PM

    Love the drawings. Kind of a silly question, but how did you do them?
  • by andsoitis on 12/26/24, 7:51 PM

    If you’re looking for a neuroscience approach, check out Numenta https://www.numenta.com/
  • by fitzn on 12/28/24, 1:03 AM

    Thank you very much for writing this up. Good, thought-provoking ideas here.
  • by oksurewhynot on 12/27/24, 12:08 AM

    Damn AGI got hands
  • by robblbobbl on 12/26/24, 4:50 PM

    Finally singularity confirmed, thanks.
  • by dudeinjapan on 12/26/24, 9:02 PM

    The greatest trick AGI ever pulled was convincing the world it didn't exist.