by montebicyclelo on 5/17/25, 9:31 AM with 35 comments
by evrimoztamur on 5/17/25, 12:00 PM
I firmly believe that differentiable logic CA is the winner, in particular because it extracts the logic directly, and thus leads to generalize-able programs as opposed to staying stuck in matrix multiplication land.
by constantcrying on 5/17/25, 11:22 AM
But I think the paper fails to answer the most important question. It alleges that this isn't a statistical model: "it is not a statistical model that predicts the most likely next state based on all the examples it has been trained on.
We observe that it learns to use its attention mechanism to compute 3x3 convolutions — 3x3 convolutions are a common way to implement the Game of Life, since it can be used to count the neighbours of a cell, which is used to decide whether the cell lives or dies."
But it is never actually shown that this is the case. It later on isn't even alleged that this is true, rather the metric they use is that it gives the correct answers often enough, as a test for convergence and not that the net has converged to values which give the correct algorithm.
But there is no guarantee that it actually has learned the game. There are still learned parameters and the paper doesn't investigate if these parameters actually have converged to something where the Net is actually just a computation of the algorithm. The most interesting question is left unanswered.
by Dwedit on 5/17/25, 1:10 PM
by Nopoint2 on 5/17/25, 12:47 PM
by wrs on 5/17/25, 3:45 PM
by bonzini on 5/17/25, 11:18 AM
by eapriv on 5/17/25, 11:23 AM
by amelius on 5/17/25, 12:49 PM
by xchip on 5/17/25, 12:53 PM