• 0 Posts
  • 42 Comments
Joined 2 years ago
cake
Cake day: July 8th, 2023

help-circle



  • Did you read the article, or the actual research paper? They present a mathematical proof that any hypothetical method of training an AI that produces an algorithm that performs better than random chance could also be used to solve a known intractible problem, which is impossible with all known current methods. This means that any algorithm we can produce that works by training an AI would run in exponential time or worse.

    The paper authors point out that this also has severe implications for current AI, too–since the current AI-by-learning method that underpins all LLMs is fundamentally NP-hard and can’t run in polynomial time, “the sample-and-time requirements grow non-polynomially (e.g. exponentially or worse) in n.” They present a thought experiment of an AI that handles a 15-minute conversation, assuming 60 words are spoken per minute (keep in mind the average is roughly 160). The resources this AI would require to process this would be 60*15 = 900. The authors then conclude:

    “Now the AI needs to learn to respond appropriately to conversations of this size (and not just to short prompts). Since resource requirements for AI-by-Learning grow exponentially or worse, let us take a simple exponential function O(2n ) as our proxy of the order of magnitude of resources needed as a function of n. 2^900 ∼ 10^270 is already unimaginably larger than the number of atoms in the universe (∼10^81 ). Imagine us sampling this super-astronomical space of possible situations using so-called ‘Big Data’. Even if we grant that billions of trillions (10 21 ) of relevant data samples could be generated (or scraped) and stored, then this is still but a miniscule proportion of the order of magnitude of samples needed to solve the learning problem for even moderate size n.”

    That’s why LLMs are a dead end.


  • When IT folks say devs don’t know about hardware, they’re usually talking about the forest-level overview in my experience. Stuff like how the software being developed integrates into an existing environment and how to optimize code to fit within the bounds of reality–it may be practical to dump a database directly into memory when it’s a 500 MB testing dataset on your local workstation, but it’s insane to do that with a 500+ GB database in production environment. Similarly, a program may run fine when it’s using a NVMe SSD, but lots of environments even today still depend on arrays of traditional electromechanical hard drives because they offer the most capacity per dollar, and aren’t as prone to suddenly tombstoning when it dies like flash media. Suddenly, once the program is in production, it turns out that same program’s making a bunch of random I/O calls that could be optimized into a more sequential request or batched together into a single transaction, and now it runs like dogshit and drags down every other VM, container, or service sharing that array with it. That’s not accounting for the real dumb shit I’ve read about, like “dev hard coded their local IP address and it breaks in production because of NAT” or “program crashes because it doesn’t account for network latency.”

    Game dev is unique because you’re explicitly targeting a single known platform (for consoles) or targeting for an extremely wide range of performance specs (for PC), and hitting an acceptable level of performance pre-release is (somewhat) mandatory, so this kind of mindfulness is drilled into devs much more heavily than business software dev is, especially in-house dev. Business development is almost entirely focused on “does it run without failing catastrophically” and almost everything else–performance, security, cleanliness, resource optimization–is given bare lip service at best.


  • I gave up on it for now when the questline involving the NPC learning to write broke, and then I started crashing to desktop (without any logs anywhere, either in the Buffout directory or even in Windows’ Event Viewer) every time I left the Swan or fast traveled directly to it, even though traveling to another point literally fifty feet south worked just fine. And since there’s no logs describing the crash, I have no idea how to fix it.

    I could probably fix it by uninstalling and re-downloading it again, but I have a goddamn data cap that my roommate already blows through every month with the fucking massive updates Fallout 76 has taken to pushing out, I have zero desire to download 60 GB of data (30 GB base game + 30 GB FOLON) every fucking time I sneeze wrong and make the game start crashing again. =|


  • It’s pretty okay. If you like the gameplay loop of scavenging parts to maintain and upgrade your car, and don’t mind the roguelite elements, it’s pretty fun, and it does a good job of creating tension–there’s been multiple occasions where I wanted to loot more but I was out of time and likely to die if I stayed much longer.

    The world building is immaculate, but IMO unfortunately the plot doesn’t really pay off, and the ending isn’t… super satisfying. It does enough to drive you along (no pun intended). The best part of the game is easily the soundtrack, and the best song in the soundtrack is easily The Freeze.


  • Are you talking about the chibi models vs. more realistic models? I think that was an artifact of an FF trope left over from the NES era where the world sprites were limited to one tile due to NES hardware limitations while the battle sprites were more detailed 1x2 tiles, and this was kept all the way up to FF6 where they finally used the same sprite for world and battles.

    I have no clue why they went back to using different/less detailed models for world exploration in FF7 (if I had to guess they were unfamiliar with the PSX hardware and the chibi models used fewer polygons), but that go a long way to explain why the FMVs sometimes used different models–IIRC, the FMVs with chibi models played directly from the field, and the ones with more detailed models had some kind of scene transition into them, or otherwise were used for major plot beats. It’s good they abandoned this entirely with FF8 onwards, though.






  • Yup, I delivered pizza for the Hut around the same time. Big ol’ map of the area divided into sectors, each order listed which sector the address was in. I’d write directions on the back of the order slip, and go off into the night with nothing but a flashlight. First day I got a lecture by the manager on how to navigate by address and tell which side of the street a house was on, I learned more about navigating that day than in the entire rest of my life.

    Sometimes I miss those days and wish I could be 19 and driving my tiny Honda Civic through the highlands again, listening to video game songs downloaded from OCRemix on my little MP3 player plugged into the car audio with a tape adapter.






  • Eccitaze@yiffit.nettoGames@lemmy.worldThe Sega Dreamcast
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 year ago

    I’m no game designer or coder so I’m just going off what I read on Wikipedia, but… Apparently the Saturn was a mostly 2D focused system, so it had a processor that could do warping and manipulation of sprites. So when it drew a “polygon” it was really drawing together a bunch of sprites and manipulating them.

    …yeah.


  • Eccitaze@yiffit.nettoGames@lemmy.worldThe Sega Dreamcast
    link
    fedilink
    English
    arrow-up
    8
    ·
    1 year ago

    Nobody wanted to develop for it because it had an insanely complex architecture (3x 32-bit processors and dual CPUs that shared a bus and couldn’t access RAM at the same time), and developers in the 90s were unaccustomed to multi-core programming. It also used quadrilaterals for the baseline polygon instead of triangles. All this was made worse by poor development tools around launch, leaving most coders stuck using raw assembly language until Sega wrote custom libraries.

    Sega also never really had a killer app for it like Mario 64 was for the N64, or FF7 was for the PlayStation. They were developing a game called Sonic XTreme, but it wound up getting canceled.


  • The problem is that there’s no incentive for employees to stay beyond a few years. Why spend months or years training someone if they leave after the second year?

    But then you have to question why employees aren’t loyal any longer, and that’s because pensions and benefits have eroded, and your pay doesn’t keep up as you stay longer at a company. Why stay at a company for 20, 30, or 40 years when you can come out way ahead financially by hopping jobs every 2-4 years?