Quote from: The Minsky Moment on November 26, 2025, 10:35:36 AMI think you're being unreasonably reductionist. At the end of the day, your brain has no idea what losses are either, it's just your neurons firing.Quote from: Tamas on November 26, 2025, 08:59:19 AMQuote from: crazy canuck on November 26, 2025, 07:01:20 AMQuote from: Tamas on November 26, 2025, 06:50:44 AMAnyways, we have been ignoring a key question:
How can LLMs be made to play computer games?
I was thinking: if the UI of something like a Paradox game could be "translated" into a format processable by these "AIs" (so, text, I guess) then they would be able to learn optimal strategies from online sources, wouldn't they?
Yes, but it would not be able to distinguish good strategies from bad. So the same problems exist and are likely worse as AI inevitably gets trained on AI slop.
Would they, though? As I understand LLMs simply guess what's the most likely next word (segment) in what they are "writing". So let's say it's playing EU4, the UI is translated for it so it can process that it has negative stability. Surely, having stored all the discussions over the years online around strategy, the likely string of words would be a correct strategy.
The likely string of words would be a correct strategy if the training data contains good optimal strategies and the model has been given feedback that leads us to give higher probability weighting to those choices. Looking at DG's CHAT GTP output, it appears that the model has been fed the Paradox forums and reddit threads on HoI4 as it parrots some of the buzzwords you'd find there.
The model "knows" that certain symbols in the game correspond to certain words like "manpower", "losses," and resources. It can spit back out those numbers and associate it with those characters. Then it can make a guess (i.e. assign probability distros) as what a likely response would be to a combat screen showing zero "losses" based on what people have said on the forums, reddit etc. about similar situations: "insane / bugged / encirclement abuse / nukes + AI collapse / paradrop cheese". CHAT GPT has no idea what "losses" are or what "paradrop cheese" is, but its algorithm tells it that in the context of discussions about the symbol "HoI 4" there is some probabilistic relationship between those two phrases. The model has no idea what an optimal strategy is, or what a strategy is, or what Hearts of Iron is. It is just making probabilistic connections between different symbolic representations in accordance with its algorithm.
Quote from: Richard Hakluyt on November 26, 2025, 07:28:54 AMI'm arguing for the right to call for a trial by jury, not mandatory trial by jury. For most offences it will just be trial by magistrate, as it is now. But if someone is in a politically sensitive trial (such as those silly buggers with the orange powder at stonehenge or the "we support Palestine Action" crowd, I think it is only right that they can call for a trial by jury.
Quote from: Tamas on November 26, 2025, 08:59:19 AMQuote from: crazy canuck on November 26, 2025, 07:01:20 AMQuote from: Tamas on November 26, 2025, 06:50:44 AMAnyways, we have been ignoring a key question:
How can LLMs be made to play computer games?
I was thinking: if the UI of something like a Paradox game could be "translated" into a format processable by these "AIs" (so, text, I guess) then they would be able to learn optimal strategies from online sources, wouldn't they?
Yes, but it would not be able to distinguish good strategies from bad. So the same problems exist and are likely worse as AI inevitably gets trained on AI slop.
Would they, though? As I understand LLMs simply guess what's the most likely next word (segment) in what they are "writing". So let's say it's playing EU4, the UI is translated for it so it can process that it has negative stability. Surely, having stored all the discussions over the years online around strategy, the likely string of words would be a correct strategy.
Quote from: Tamas on November 26, 2025, 08:59:19 AMQuote from: crazy canuck on November 26, 2025, 07:01:20 AMQuote from: Tamas on November 26, 2025, 06:50:44 AMAnyways, we have been ignoring a key question:
How can LLMs be made to play computer games?
I was thinking: if the UI of something like a Paradox game could be "translated" into a format processable by these "AIs" (so, text, I guess) then they would be able to learn optimal strategies from online sources, wouldn't they?
Yes, but it would not be able to distinguish good strategies from bad. So the same problems exist and are likely worse as AI inevitably gets trained on AI slop.
Would they, though? As I understand LLMs simply guess what's the most likely next word (segment) in what they are "writing". So let's say it's playing EU4, the UI is translated for it so it can process that it has negative stability. Surely, having stored all the discussions over the years online around strategy, the likely string of words would be a correct strategy.
Quote from: DGuller on November 26, 2025, 12:22:23 AMYou are asserting that an LLM cannot generalize beyond its training data.
Quote from: Richard Hakluyt on November 26, 2025, 09:20:42 AMQuote from: Sheilbh on November 26, 2025, 07:47:20 AMTotally separately but the OBR published its update before the budget by mistake. Apparently due to a "technical error".
I still think we should abolish the OBR in general.
But on this - people need to resign/be fired. We've got the City making moves on it but also the Guardian live blogging what the Chancellor is going to announce as they work through the OBR report before they've even stood up in the Commons. It's clown-ish.
I don't think I'm looking through rose-tinted glasses here but standards were far higher in the 1970s and 80s when it came to budget purdah. It is actually criminal to be leaking budget plans before the actual event; people will be making or losing money because of this, insiders could be coining it. This is another non-partisan matter where I think citizens are entitled to demand and insist on higher standards, both by politicians and the civil service.
Page created in 0.022 seconds with 11 queries.