14 comments

  • proc0 a day ago ago

    Right, I think it's a limitation of deep learning. Transformers and gigantic leaps in scaling have allowed AI models to reach impressive capabilities, but at its core it's limited by its training data. This is not how actual intelligence works. People don't need terabytes of information to learn how to communicate and reason.

    For this reason the current AI trend will have a bigger impact on creative tasks, rather than critical technical ones. It's already great for generating art quickly, at least for the concepting phase, and other creative assets that can afford to be generic. Solving technical problems, on the other hand, requires reasoning beyond what can be extracted from training data.

    We'll need a new paradigm of AI in order to have a chance at creating models that properly reason. Even without detail knowledge of the brain, we can safely speculate that the reason and language areas are extremely efficient compare to cutting edge LLMs, which means there are algorithms more complex and efficient than simple artificial neural connections that just sum weights with a bias.

  • eschneider a day ago ago

    Yes. You're expecting too much. Generative AI models don't "understand" your problem, they don't even "understand" how to program. They're just fitting whatever data they've seen to your input.

    • gwoolhurme 17 hours ago ago

      To his defense, that is how it’s marketed. That this new model can reason.

      • loveparade 12 hours ago ago

        Yeah, but "reason" is not a well-defined term. It means different things to different people in different contexts. It's just marketing speech. You can easily argue that all ML models, even those from 50 years ago, can reason to some extent.

        • gwoolhurme 11 hours ago ago

          Fully agree that’s kind of my point though. It’s a very tall order for some people. Like the OP

  • segmondy 6 hours ago ago

    It's a limitation of your approach. If you have an idea what you are trying to do, a good detail might produce the correct code. But if it doesn't, then your next prompt will not be asking for it. But explaining to the AI how it went wrong, suggesting and guiding it towards a different path. It will often help you resolve the issue. If you have no domain knowledge or just keep regenerating responses in home that one will be correct, you are going to be wasting time and money.

  • mergisi 9 hours ago ago

    AI can be frustrating when it falls short on complex tasks, especially with long or intricate code. To get the most out of it, you need to break down big problems into smaller pieces and choose the right AI for your needs—like OpenAI o1, Claude Sonnet. Also, leveraging dev tools like Cursor AI can help enhance productivity. For tasks like SQL generation, specialized tools like https://ai2sql.io/ work great. AI isn't perfect for everything, but when used selectively, it can still be super helpful.

  • jprete a day ago ago

    It never occurred to me before that chatbot randomness might have the same reward structure as a slot machine, but apparently it does. OpenAI got you to spend all your credits on seven attempts at this one problem. I'm not saying you're addicted, but I wonder about the people who absolutely gush over it.

  • resource0x a day ago ago

    It's time for "How many programmers does it take to screw in a lightbulb using AI" jokes.

  • muzani 21 hours ago ago

    It might be just an OpenAI thing where they ramp up the power a week after the demo to get more subscriptions, then bring it down gradually. The forums now happily gaslight you into thinking it's a conspiracy theory, despite the evidence. It's easy to catch - just share an amazing input/output response with a friend, then try the exact same thing a month later.

    It's one of the arguments for using open source AI even though it's still a little behind - at least when you're running it on your own system, you know if you're the problem.

  • solardev a day ago ago

    Does Claude work any better for you?

    • moomoo11 19 hours ago ago

      I have tried Claude and I find that it is more "to the point", but still suffers from giving the wrong/unusable answers.

      Both claude and chatgpt are good for rote tasks, but claude is definitely more succinct.

  • b20000 20 hours ago ago

    these models are search engines with some interpolation thrown in

  • rvz 21 hours ago ago

    > When people hype up that the AI solved something for them, I wonder were they lazy like me working with something complex, or were they lazy and didn't even try on something simple?

    The truth they won't tell you is that they have likely invested in that AI tool and they are hyping it up with their VC friends that 'It works' even when they know it doesn't.

    Each time I talk to the AI bros about these limitations, they retort to their whataboutisms with 'But humans hallucinate too!', 'The human brain is the same as an LLM' nonsense excuses.

    LLMs do not 'understand' your problems nor can they reason about them. O1 is no different and instead of buying into the scam and prompting endlessly with garbage results and attempting to replace your co-worker, actual programmers can write the plan for the code themselves and solve it. Especially for unseen code or new changing syntax for it.

    Whenever I see someone promoting another AI tool, I always see who invested and 9/10 of the time it is funded by VCs and ex-FAANG engineers yet again on the snake oil grift. (And they know it but will never admit it.)