• OsrsNeedsF2P@lemmy.ml
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    2 months ago

    Oof let’s see, what am I an expert in? Probably system design - I work at (insert big tech) and run a system design club there every Friday. I use ChatGPT to bounce ideas and find holes in my design planning before each session.

    Does it make mistakes? Not really? it has a hard time getting creative with nuanced examples (i.e. if you ask it to “give practical examples where the time/accuracy tradeoff in Flink is important” it can’t come up with more than 1 or 2 truly distinct examples) but it’s never wrong.

    The only times it’s blatantly wrong is when it hallucinates due to lack of context (or oversaturated context). But you can kind of tell something doesn’t make sense and prod followups.

    Tl;dr funny meme, would be funnier if true

    • spooky2092@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      I ask AI shitbots technical questions and get wrong answers daily. I said this in another comment, but I regularly have to ask it if what it gave me was actually real.

      Like, asking copilot about Powershell commands and modules that are by no means obscure will cause it to hallucinate flags that don’t exist based on the prompt. I give it plenty of context on what I’m using and trying to do, and it makes up shit based on what it thinks I want to hear.

    • RagingRobot@lemmy.world
      link
      fedilink
      arrow-up
      0
      ·
      2 months ago

      That’s not been my experience with it. I’m a software engineer and when I ask it stuff it usually gives plausible answers but there is always something wrong. For example it will recommend old outdated libraries or patterns that look like they would work but when you try them out you figure out they are setup differently now or didn’t even exist.

      I have been using windsurf to code recently and I’m liking that but it makes some weird choices sometimes and it is way too eager to code so it spits out a ton of code you need to review. It would be easy to get it to generate a bunch of spaghetti code that works mostly that’s not maintainable by a person out of the box.

      • grrgyle@slrpnk.net
        link
        fedilink
        arrow-up
        0
        ·
        2 months ago

        My main experience with AI is that the pull requests I’ve reviewed have got at least twice as large, and I need to review the code much, much more carefully.

  • RedSnt 👓♂️🖥️@feddit.dk
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    I’ve been using o3-mini mostly for ffmpeg command lines. And a bit of sed. And it hasn’t been terrible, it’s a good way to learn stuff I can’t decipher from the man pages.

      • RedSnt 👓♂️🖥️@feddit.dk
        link
        fedilink
        arrow-up
        0
        ·
        2 months ago

        True, in many cases I’m still searching around because the explanations from humans aren’t as simplified as the LLM. I’ll often have to be precise in my prompting to get the answers I want which one can’t be if they don’t know what to ask.

        • Nalivai@lemmy.world
          link
          fedilink
          arrow-up
          0
          ·
          2 months ago

          And that’s how you learn, and learning includes knowing how to check if the info you’re getting is correct.
          LLM confidently gives you easy to digest bite, which is plain wrong 40 to 60% of the time, and even if you’re lucky it will be worse for you.

          • RedSnt 👓♂️🖥️@feddit.dk
            link
            fedilink
            arrow-up
            0
            ·
            2 months ago

            I’m in the kiddie pool, so I do look things up or ask what stuff does. Even though I looked at the man page for printf (printf.3 I believe), there was nothing about %*s for example, and searching for these things outside of asking LLM’s is some times too hard to filter down to the correct answer. I’m on 2 lines of code per hour, so I’m not exactly rushing.
            Shell scripting is quite annoying to be sure. Thinking of learning python instead.

  • Alloi@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    i mainly use it for fact checking sources from the internet and looking for bias. i double check everything of course. beyond that its good for rule checking for MTG commander games, and deck building. i mainly use it for its search function.

  • Most of my searches have to do with video games, and I have yet to see any of those AI generated answers be accurate. But I mean, when the source of the AI’s info is coming from a Fandom wiki, it was already wading in shit before it ever generated a response.

    • henfredemars@infosec.pub
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      I’ve tried it a few times with Dwarf Fortress, and it was always horribly wrong hallucinated instructions on how to do something.

  • DicJacobus@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    I have frequentley seen gpt give a wrong answer to a question, get told that its incorrect, and the bot fights with me and insists Im wrong. and on other less serious matters Ive seen it immediatley fold and take any answer I give it as “correct”

  • jsomae@lemmy.ml
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    ChatGPT is a tool. Use it for tasks where the cost of verifying the output is correct is less than the cost of doing it by hand.

    • tacobellhop@midwest.social
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      Youre still doing it by hand to verify in any scientific capacity. I only use ChatGPT for philosophical hypotheticals involving the far future. We’re both wrong but it’s fun for the back and forth.

      • jsomae@lemmy.ml
        link
        fedilink
        arrow-up
        0
        ·
        edit-2
        2 months ago

        It is not true in general that verifying output for a science-related prompt requires doing it by hand, where “doing it by hand” means putting in the effort to answer the prompt manually without using AI.

        • tacobellhop@midwest.social
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          You can get pretty in the weeds with conversions on ChatGPT in the chemistry world or even just basic lab work where a small miscalculation at scale can cost thousands of dollars or invite lawsuits.

          I check against actual calibrated equipment as a verification final step.

          • jsomae@lemmy.ml
            link
            fedilink
            arrow-up
            0
            ·
            2 months ago

            I said not true in general. I don’t know much about chemistry. It may be more true in chemistry.

            Coding is different. In many situations it can be cheap to test or eyeball the output.

            Crucially, in nearly any subject, it can give you leads. Nobody expects every lead to pan out. But leads are hard to find.

            • tacobellhop@midwest.social
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 months ago

              I imagine ChatGPT and code is a lot like air and water.

              Both parts are in the other part. Meaning llm is probably more native at learning reading and writing code than it is at interpreting engineering standards worldwide and allocation the exact thread pitch for a bolt you need to order thousands of. Go and thread one to verify.

              • jsomae@lemmy.ml
                link
                fedilink
                arrow-up
                0
                ·
                2 months ago

                This is possibly true due to the bias of the people who made it. But I reject the notion that because ChatGPT is made of code per se that it must understand code better than other subjects. Are humans good at biology for this reason?

    • qarbone@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      Honestly, I’ve found it best for quickly reformatting text and other content. It should live and die as a clerical tool.

      • ArchRecord@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        Which is exactly why every time I see big tech companies making another stupid implementation of it, it pisses me off.

        LLMs like ChatGPT are fundamentally word probability machines. They predict the probability of words based on context (or if not given context, just the general probability) when given notes, for instance, they have all the context and knowledge, and all they have to do it predict the most statistically probable way of formatting the existing data into a better structure. Literally the perfect use case for the technology.

        Even in similar contexts that don’t immediately seem like “text reformatting,” it’s extremely handy. For instance, Linkwarden can auto-tag your bookmarks, based on a predetermined list you set, using the context of each page fed into a model running via Ollama. Great feature, very useful.

        Yet somehow, every tech company manages to use it in every way except that when developing products with it. It’s so discouraging to see.

  • lalala@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    I think that AI has now reached the point where it can deceive people ,not equal to humanity.

  • Korhaka@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    I just use it to write emails, so I declare the facts to the LLM and tell it to write an email based on that and the context of the email. Works pretty well but doesn’t really sound like something I wrote, it adds too much emotion.

  • Showroom7561@lemmy.ca
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    If it’s being designed to answer questions, then it should simply be an advanced search engine that points to actual researched content.

    The way it acts now, it’s trying to be an expert based one “something a friend of a friend said”, and that makes it confidently wrong far too often.

  • Kane@femboys.biz
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    Exactly this is why I have a love/hate relationship with just about any LLM.

    I love it most for generating code samples (small enough that I can manually check them, not entire files/projects) and re-writing existing text, again small enough to verify everything. Common theme being that I have to re-read its output a few times, to make 100% sure it hasn’t made some random mistake.

    I’m not entirely sure we’re going to resolve this without additional technology, outside of ‘the LLM’-itself.

  • SirSamuel@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    First off, the beauty of these two posts being beside each other is palpable.

    Second, as you can see on the picture, it’s more like 60%

    • morrowind@lemmy.ml
      link
      fedilink
      arrow-up
      0
      ·
      2 months ago

      No it’s not. If you actually read the study, it’s about AI search engines correctly finding and citing the source of a given quote, not general correctness, and not just the plain model

      • SirSamuel@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        2 months ago

        Read the study? Why would i do that when there’s an infographic right there?

        (thank you for the clarification, i actually appreciate it)

    • OsrsNeedsF2P@lemmy.ml
      link
      fedilink
      arrow-up
      0
      ·
      2 months ago

      There’s an easy way to settle this debate. Link me a Wikipedia article that’s objectively wrong.

      I will wait.

    • glimse@lemmy.world
      link
      fedilink
      arrow-up
      0
      ·
      2 months ago

      What topics are you an expert on and can you provide some links to Wikipedia pages about them that are wrong?

      • Meltdown@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        2 months ago

        I’m a doctor of classical philology and most of the articles on ancient languages, texts, history contain errors. I haven’t made a list of those articles because the lesson I took from the experience was simply never to use Wikipedia.

      • PeterisBacon@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        Because some don’t let you. I can’t find anything to edit Elon musk or even suggest an edit. It says he is a co-founder of OpenAi. I can’t find any evidence to suggest he has any involvement. Wikipedia says co-founder tho.

      • Meltdown@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        2 months ago

        There are plenty of high quality sources, but I don’t work for free. If you want me to produce an encyclopedia using my professional expertise, I’m happy to do it, but it’s a massive undertaking that I expect to be compensated for.

      • Meltdown@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        2 months ago

        With all due respect, Wikipedia’s accuracy is incredibly variable. Some articles might be better than others, but a huge number of them (large enough to shatter confidence in the platform as a whole) contain factual errors and undisguised editorial biases.

      • A_norny_mousse@feddit.org
        link
        fedilink
        arrow-up
        0
        ·
        2 months ago

        TBF, as soon as you move out of the English language the oversight of a million pair of eyes gets patchy fast. I have seen credible reports about Wikipedia pages in languages spoken by say, less than 10 million people, where certain elements can easily control the narrative.

        But hey, some people always criticize wikipedia as if there was some actually 100% objective alternative out there, and that I disagree with.

        • Fair point.

          I don’t browse Wikipedia much in languages other than English (mainly because those pages are the most up-to-date) but I can imagine there are some pages that straight up need to be in other languages. And given the smaller number of people reviewing edits in those languages, it can be manipulated to say what they want it to say.

          I do agree on the last point as well. The fact that literally anyone can edit Wikipedia takes a small portion of the bias element out of the equation, but it is very difficult to not have some form of bias in any reporting. I more use Wikipedia as a knowledge source on scientific aspects which are less likely to have bias in their reporting

      • PeterisBacon@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        Idk it says Elon Musk is a co-founder of openAi on wikipedia. I haven’t found any evidence to suggest he had anything to do with it. Not very accurate reporting.

        • grrgyle@slrpnk.net
          link
          fedilink
          arrow-up
          0
          ·
          2 months ago

          Isn’t co-founder similar to being made partner at a firm? You can kind of buy your way in, even if you weren’t one of the real originals.

          • PeterisBacon@lemm.ee
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 months ago

            That is definitely how I view it. I’m always open to being shown I am wrong, with sufficient evidence, but on this, I believe you are accurate on this.

    • Ms. ArmoredThirteen@lemmy.zip
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      If this were true, which I have my doubts, at least Wikipedia tries and has a specific goal of doing better. AI companies largely don’t give a hot fuck as long as it works good enough to vacuum up investments or profits

      • Meltdown@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        2 months ago

        Your doubts are irrelevant. Just spend some time fact checking random articles and you will quickly verify for yourself how many inaccuracies are allowed to remain uncorrected for years.