Do you have a dumb question that you're kind of embarrassed to ask in the main thread? Is there something you're just not sure about?
This is your opportunity to ask questions. No question too simple or too silly.
Culture war topics are accepted, and proposals for a better intro post are appreciated.

Jump in the discussion.
No email address required.
Notes -
So how did LLMs impressed you this week? My case - I had lost original 3mf, but had OrcaSlicer temp folder, which ironically orca can't open. I ask codex-cli to try and reconstruct a proper 3mf. For my astonishment is did on the second try.
Finally got around to test progress on image generation again.
My go-to test is creating an entire fake Instagram influencer from scratch. That nicely tests consistency between images, spacial understanding of the scene, prompt following on minute details, ect. It keeps me up to date on what the problems are when you fake photos of people (or fake entire people in general). I mostly create women, because that's more fun to me and also because this tests model censorship more effectively - the commercial models are a lot more touchy when creating women than when they create men.
The main result of my most recent session is particularly funny: Nano Banana 2 is another significant step forward on photo-realism, but it is exceedingly difficult to get it to produce images of conventionally beautiful people from scratch. Getting just a portrait of a woman that is above a 7 requires a lot of coaxing. If the major focus of the prompt is on some other detail, it will generate the most mid women you've ever seen. Nano Banana 1 was perfectly happy to just spit out 10s. You start the prompt with "photo-realistic full body shot of an attractive female college student..." and you could focus on scene, clothes, body position, camera equipment, ect., and it only needed minor coaxing for some body types and poses (as long as you kept it SFW). But Nano Banana 2 will often simply ignores instructions that coax other models towards conventional beauty. I wonder why. Peak body positivity seems long past. Did earlier models train predominantly on pictures of influencers on social media (because they post so much), and now photos of the rest of humanity have a more proportional ratio in the training data? Or are they trying to stop me, in particular, from creating and monetizing an Instagram e-thot? (I'm not, of course, I've lost interest in image generation, again, very quickly).
Other than that: prompt following is truly impressive now. You can pick scene, clothes, and body positions (either by describing them or supplying reference photos), and it will usually one-shot them down to the correct head tilt angle. Consistency (same person in different images) requires a bit of care, or ideally tons of reference images. We're not completely out of the uncanny valley for faces created completely from scratch, but this is where I notice the most progress (Nano Banana 1 makes beautiful people, but they look like influencers with the filters maxed out in the best case, and like very good paintings in the median case). Around 1% of images still have extra limbs or other easy tells.
Oh, and making images that help explain a technical concept is still hilariously bad. A straight rip-off of an existing image with a liberal dose of detail errors is the best you can expect. Ah, factual correctness in every detail... the old nemesis of AI still lives on.
More options
Context Copy link
I recently had sonnet make it through an entire session without mixing up VBIL and VBILX. I'm going to call the improvement.
More options
Context Copy link
A lot of the newest hotness has been a little too automated for my tastes, and haven't had much free time, so mostly screwing around with older configs.
Successes :
Failures :
Try build123 for cad cam
More options
Context Copy link
More options
Context Copy link
Not sure if this counts, but I can't get over the fact that Hollywood wordcells, notorious for their poor understanding of science and technology, somehow ended up being right about how computers (will) work:
I've noticed this too. I've also noticed that the Enterprise's computer says "Acknowledged" instead of "Wow, that command is absolutely chef's kiss and has real Starfleet energy, I'll get on that right away. While we're here, tell me what thoughts you have on Deck Seven?"
I see you haven't watched any of thr new Treks, then.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
They didn't, but mostly because I have been too busy to use them for anything I did not expect them to do by default. Whenever a new and exciting model launches, I stress test it extensively, but for at least a year, the models are good enough for my personal and professional needs. Last time I saw a massive improvement in quality that unlocked entirely new use-cases that blew me away was o3, otherwise I tend to feel slightly impressed.
From memory, GPT-2, 3 and 4, then whatever Claude just came out then, then o1 (from seeing others use it), then R1/o3. Native image gen with a variant of 4o. Those standout. Everything else falls under slightly better in ways that don't stick out.
But I am happy enough with them being good for research or editing my writing, or generating images. If they get significantly better in a manner that is glaringly obvious in normal use, I'm close to worrying (much harder) about losing my job.
More options
Context Copy link
Agent was successfully able to submit a public record request via an online portal, given sufficient detail and just a little nudging.
The power this potentially unlocks is quite sizeable, actually.
More options
Context Copy link
Using github copilot, GPT 5.4 seems pretty solid. Far more capable than 5.2, and more robust than 5.3. The only downside is that it seems a tad slower, but that I accept given how much more it ends up doing.
I've been using it privately as well as at work, and right now my main complaint is that it tends to be a little too eager to write more code, when a little less logic would keep the overall codebase a lot more maintainable. But maybe I'll yet be able to evangelize the LLM until it believes in the gospel of clean code.
More options
Context Copy link
More options
Context Copy link