Epicycles All the Way Down (2025)
Posted by surprisetalk 4 days ago
Comments
Comment by edo_cat 19 hours ago
> Right now we seem stuck with Ptolemaic astronomy, scholastically adding epicycles upon epicycles, without making the leap to hit the inverse-square law.
This is a great analogy but just isn’t what happened at all. There is no evidence medieval astronomers added epicycles. Copernicus added epicycles to his heliocentric model - and this was a reason his model was criticised was because it was too complicated!
It’s still good analogy, but in reality each planet required a hand tuned; equant, deferent, epicycle and sometimes 1 epicyclet..
Also surely the great logical leap was Kepler’s elliptical orbits which broke free of the perfect circle constraint?
> Reason may be employed in two ways to establish a point: firstly, for the purpose of furnishing sufficient proof of some principle [...]. Reason is employed in another way, not as furnishing a sufficient proof of a principle, but as confirming an already established principle, by showing the congruity of its results, as in astronomy the theory of eccentrics and epicycles is considered as established, because thereby the sensible appearances of the heavenly movements can be explained; not, however, as if this proof were sufficient, forasmuch as some other theory might explain them.
Thomas Aquinas (dumbass Scholastic)
Comment by throwaway210426 22 hours ago
Comment by suddenlybananas 21 hours ago
Comment by throwaway210426 21 hours ago
Comment by OutOfHere 21 hours ago
As for chess, although an LLM knows the rules of chess, it is not expected to have been trained on many optimal chess games. As such, is it fair to gauge its skill in chess, especially without showing it generated images of its candidate moves? Even if representational and training limitations were addressed, we know that LLMs are architecturally crippled in that they have no neural memory beyond their context. Imagine a next-gen LLM that if presented with a chess puzzle would first update its internal weights for playing optimal chess via a simulation of a billion games, and then return to address the puzzle you gave it. Even with the current arch, it could equivalently create a fork of itself for the same purpose, a new trained model in effect, but the rushing human's desire for wanting the answer immediately comes in the way.
Comment by suddenlybananas 20 hours ago
Well, it's read every book ever written on chess so you would expect it to be at least half-way decent.
Comment by gus_massa 18 hours ago
* 2025 https://www.youtube.com/playlist?list=PLBRObSmbZluRddpWxbM_r...
* 2026 https://www.youtube.com/playlist?list=PLBRObSmbZluQwBIvxyiWf...
I recommend to watch at least the last game of each list, that has the final game with the bots that play the best.
My takeaway:
Most chatbots know openings very well, the problem start when one of them makes an unexpected (legal or ilegal) move. Some models just copy moves from old games that make no sense in this game, and other models continue playing (almost) correctly. In particular ChatCPT was very bad in 2025 but very good in 2026.
(When a chatbot makes an ilegal move, most of the times he just follow the bot instructions. I think it's bad because it confuses the other chatbot that may interpret the incorrect move in a different way. Let's say if white moves the rook form a1 to a8 jumping over a pawn in a4, he may left the pawn in position but black may interpret that magically there is no pawn in a4. Anyway, he is in the show business, not in the let-s-get-a-nobel business, and weird games are more fun to cast.)
Comment by ogogmad 21 hours ago
Comment by DarkNova6 21 hours ago
If anything, I see greater verticality of specialized software that is using LLMs at their core, but with much aid and technology around it to really make the most out of it.
Comment by ogogmad 20 hours ago
> This was solved by GPT-5.4 Pro (prompted by Price)
See the discussion here: https://www.erdosproblems.com/forum/thread/1196
Comment by FrustratedMonky 21 hours ago
Why do these distinctions matter?
is it an LLM, or symbolic, or a combo, or a dozen technologies stitched together. Who cares. It is all automation. It is all artificial.
Comment by DarkNova6 20 hours ago
In the context of evolving LLM this is the crucial distinction.
Comment by layer8 17 hours ago
Comment by FrustratedMonky 16 hours ago
But 'purely LLM' also isn't a definition for generalizable.
Comment by suddenlybananas 20 hours ago