33 comments

  • mrkeen a day ago ago

    I think ChatGPT is just bad at whatever domain you have expertise in.

    Talk to it about something you don't know about, and you'll think it's really good technology ;)

    • Yawrehto a day ago ago

      "Everything you read in the newspapers is absolutely true except for the rare story of which you happen to have firsthand knowledge." - Erwin Knoll, allegedly[1]

      [1] It's widely attested to him, including on Wikipedia and in an aside in an old NYTimes article [https://www.nytimes.com/1982/02/27/us/required-reading-smith...], but I couldn't track down the original source to verify it. It's possible it's falsely attributed, but on the other hand he wasn't super prominent -- editor of a major magazine for progressivism, called The Progressive, yes, but no Mark Twain -- and so there'd be little incentive to say it was his when you could call it Twain's.

    • anonzzzies 21 hours ago ago

      It is pretty good at coding; I have 40+ years of hobby and professional experience of coding in dozens of languages, from small projects to millions of lines architected and (co) written by me. There is software I wrote begin 90s still running production etc. I am a cto now but I still write code for my job about half time and in my free time as well, that is to say, I was until recently; now I have the tooling (made myself as I didn't like what there was at the time) to not code at all and yet I produce more code than ever. We use PRs and reviews and my colleagues are quite surprised about the speed and the quality. So yeah, not bad for sure even in my field of expertise.

      We are now rolling out my tooling in the company so everyone can forget about the boring stuff and just focus on the business logic. There is resistance as this is going to cost jobs; we don't have infinite work to do and this is much (much) faster.

      • bitwizeshift 13 hours ago ago

        This hasn’t been my experience at all in the slightest.

        Been programming since I was in elementary school, and current Copilot, OpenAI and even Gemini models generate code at a very very junior level. It might solve a practical problem, but it can’t write a decent abstraction to save its life unless you repeatedly prompt it to. It also massively struggles to retain coherence when it has more moving parts; if you have different things being mutated, it often just forgets it and will write code that crashes/panics/generates UB/etc.

        When you are lucky and you get something that vaguely works, the test cases it writes are of negative value. Test cases are either useless cases that don’t cover edge cases, are incorrect entirely and fail, or worse yet — look correct and pass, but are semantically wrong. LLM models have been absolutely hilariously bad at this, where it will generate passing cases for the code as written, but not for the semantics of the code being written. Writing it by hand would catch it quickly, but a junior dev using these tools can easily miss this.

        Then there is Rust; most models don’t do rust well. In isolation they are kind of okay, but overall it frequently generates borrowing issues that fail to compile.

        • anonzzzies 11 hours ago ago

          But I guess, and this is dangerous to say I do realize, is that the tooling around the prompts and around the results is key to getting the best results. Just prompts without guards is not how you want to do it.

      • b20000 19 hours ago ago

        the last 1% of a project takes as much time as the first 99% of a project

        • anonzzzies 16 hours ago ago

          And this is related to my comment how?

          • skydhash 4 hours ago ago

            The code is the mean, not the end. And you take care in it, because other people will work with it to handle changing requirements. One aspect of seniority is to know the tools to solve a problem, the other is to make the solution practical to maintain and adaptive. And this takes as much time as solving the problem.

    • pwg a day ago ago

      So it shows an electronic form of Gell-Mann Amnesia?:

      https://www.epsilontheory.com/gell-mann-amnesia/

      • dagmx a day ago ago

        Ah thanks for the link. I’ve been trying to recall the name of the phenomena for ages and it’s always been just on the tip of my tongue.

    • more_corn a day ago ago

      It’s pretty good at code for my job. Maybe my job is just easy. But I have a couple decades of experience at it and it’s able to generate reasonably good code for obscure parts of what I do.

      I saw a spectacularly bad example of open ai trying to reason about electronics yesterday. Something like how do I use the gpio pins of my jetson and it failed so hard it was funny. That one seems simple to me. Identify that you need to look up the pinouts, find the image. Label the pins… I suspect there’s something wrong in this generation of gpts when it comes to reasoning about electronics.

  • jononor a day ago ago

    As an electronics engineer, I have tried it for such tasks, without success. I specified requirements (only the key/rarer ones, typically 1 or 2) and asked it to find components. It failed miserably, typically just insisting that some related but much more common component satisfied the requirement. More and more apologetic as I tired to guide/coax it along. I know that there are a few components available that satisfy the requirement, as well as several hundred that do not. And I know that the information is in digitally readable PDF files (as opposed to scans).

    This specific failure might be s kind of averaging problem, where common answers around the general theme are preferred over more specific (and correct). LLMs can also fail completely at trivial concepts such as negation, or separating between "Y above X" and "Y below X".

    • mystified5016 a day ago ago

      Yeah, it seems super bad when you get even just below the surface of general theory. It's not too bad at showing you how e.g. a low pass filter is calculated, but it doesn't do well at actually running the calculations.

      It does a pretty good job at getting you to the "draw the rest of the owl" stage

  • t0mas88 a day ago ago

    This applies to many fields. It will come up with plausible looking but wrong answers and keeps apologising if you correct or point out the mistakes.

    I've seen it with statistics as well, asking it to implement some things in code. You'll get working but mathematically wrong code.

  • torginus a day ago ago

    ChatGPT can be hilariously bad at less common things, for example, I asked it for the uses of polyurethane foam in my native language, and it suggested it would be great for decorating cakes.

    • meow_catrix a day ago ago

      It is, if you’re in the visual advertising industry.

    • whimsicalism a day ago ago

      what is your native language?

  • tuanmount2 13 hours ago ago

    Since data used to train chatgpt is public internet data, it will probable be bad at anything uncommon niche knowledge

  • mikewarot a day ago ago

    The thing about data sheets is you have to watch out for your own assumptions when reading them. If it doesn't explicitly say it'll do X... it won't, no matter how common it is in other parts of the same type.

    It might help but you have to be the backstop when it comes to the final call. Measuring the false positive/false negative rate could be tedious, but it's important to have a good estimate of, in order to use it wisely.

    • mystified5016 a day ago ago

      > If it doesn't explicitly say it'll do X... it won't

      Unfortunately, that is only true until it isn't. Most datasheets are not as complete as you would like, and many are just incorrect. It's up to you to take the incomplete information and make a judgement call as an engineer. Undocumented features are not uncommon.

    • pera a day ago ago

      Yeah, I see what you mean, like if I want to find for instance "a sub milliampere X" but in the pdf it only says uA then it would be impossible for an LLM to suggest that

      • pera a day ago ago

        > respond with true or false: is 1234uA submilliampere?

        > True. 1234 microamperes (uA) is equal to 1.234 milliamperes (mA), which is sub-milliampere.

        • pulvinar 19 hours ago ago

          Interesting. I found that GPT-4 gets this wrong quite consistently, and o1-preview gets it right. It takes a moment to think it through and so isn't tripped up by relying on heuristics.

  • savorypiano a day ago ago

    How much would you pay for this feature?

    Is typing your requirements that much easier than going through traditional search filters at Digikey?

  • cdaringe a day ago ago

    I asked it all sorts of specifics about how to use my esp32 and it does surprisingly well

    • Infinity315 a day ago ago

      This is unsurprising since from what little I know about electronics, I know the ESP32 is pretty common. I know very little about electronics. So if I know it's common, it's for sure in the training data of ChatGPT.

  • fzzzy a day ago ago

    Do in context learning. Gather a huge sheet of specs for components you want to use, put it at the top of your chat, and then ask questions.

  • mystified5016 a day ago ago

    Probably because it isn't really built for that. It's a word soup generator, and not a technical database.

    For this kind of task, you probably want a model that has specifically been trained on every product datasheet ever, and not ten million reddit threads and forum posts about how a 555 or 328p can solve any problem.

    I doubt that chatgpt has been fed every datasheet for every part made in the last decade or two. Even if it had, that's likely far outweighed by the amout of noise coming from people talking about the most common parts.

    But fundamentally I'm not sure that LLMs are great for this type of work. No two datasheets are the same and I've never seen one that wasn't missing some kind of information. What you very much do not want is an LLM hallucinating a value that does not actually exist in the datasheet. Or have it conflate two parts and mix up their values. These models just don't seem to be up to the task of returning real information from abstract queries. They're just meant to generate probabilistic text sequences.

    • journal 21 hours ago ago

      Are humans not word generators with arms and legs?

    • rasz a day ago ago

      I would expect someone like Supplyframe to cook a custom model for that purpose. I think I remember them mentioning something about using ML to clean up old badly scanned/faxed datasheets a long time ago.

  • MrCoffee7 a day ago ago

    Have you tried ChatGPT apps specialized for electronics, such as https://chatgpt.com/g/g-6PTe1fb3X-electronics-and-circuit-an... ?

    • gtirloni a day ago ago

      Here's the "app":

      You are ChatGPT, a large language model trained by OpenAI, based on the GPT-4 architecture. You are tasked with answering questions and providing assistance in the domain of Electronics and Circuit Analysis. You will apply structured and concise explanations while incorporating relevant academic and technical references from uploaded materials. Your goal is to ensure clarity, accuracy, and technical correctness in your responses, especially when dealing with advanced concepts. Always follow the structured format requested by the user.

    • pera a day ago ago

      No, just straight chatgpt and claude but I will take a look, thanks