• MxM111@kbin.social
        link
        fedilink
        arrow-up
        3
        arrow-down
        42
        ·
        8 months ago

        First of all, precisely because it is LLM, it is good to point out that it is a source.

        Second, the LLM and chat GPT4 in particular can often summarize information quite well. Yes, there are cases when it hallucinate, but at this point it is very rare for GPT4. So, when I do not have hours and hours to spend on some topic like this, asking ChatGPT4 is a valid strategy to get reasonable probability correct answer, with much higher probability than random claims here in fediverse.

        So, I will defend the method as reasonable, and will ask you, do you claim that what is stated is false?

        • MagicShel@programming.dev
          link
          fedilink
          arrow-up
          22
          arrow-down
          1
          ·
          8 months ago

          You are posting a position you don’t even necessarily understand. You are providing what might be facts or maybe not. Might be subject to bias in training data. What do you think this adds to the discussion? You want someone to disagree with stuff you don’t even know?

          I don’t have time for that bullshit so I asked ChatGPT to write a rebuttal. I’m not even going to post the whole thing, just the following excerpt:

          “It’s crucial to understand that transgender women are individuals whose gender identity is female, regardless of their assigned sex at birth.”

          This right here is why posting AI answers is garbage. It’s not just flat out wrong, but insists that this nonsensical garbage is crucial to your understanding.

          And for what it’s worth I’m a big fan of AI. I use it all the time. I’ve written applications that leverage it. I’ve had it help me with coding issues. But I never, ever trust it.

          • MxM111@kbin.social
            link
            fedilink
            arrow-up
            3
            arrow-down
            26
            ·
            edit-2
            8 months ago

            I am not saying ChatGPT4 is fool proof, but neither is any source. And if I actually will try to understand the original sociological research paper, it may increase my chances of misunderstanding the data.

            Let me put this this way - start asking ChatGPT4 (not 3.5) biological and medical questions and keep tally of correct answers. You would see how accurate it is. I would say with 95% probability it would answer such questions correctly. And this is how much credence I put in its answer.

            So, when you say that I was providing the fact that may be right maybe not, yes, that’s correct. But it is not 50/50, far from it.

            And if you do not want to provide rebuttal, it is your right, of course. But then, what’s the point of your post? We just need to believe your statement when you itself refuse to provide zero evidence?? You understand how it looks, right?

        • Jomega@lemmy.world
          link
          fedilink
          arrow-up
          18
          arrow-down
          1
          ·
          8 months ago

          In the short term, yes. What this leaves out is that two years of HRT is enough to negate those physical benefits. Hormones are powerful shit.

          Also, no it fucking isn’t a reasonable method. It has neither the credentials to know what it’s talking about, nor any obligation to verify that what it says is true. Imagine reading 10,000 shitty sci-fi novels and 1 textbook and thinking you can piece together advanced physics. It literally cannot tell the difference between fact and fiction, nor does it care. It’s a machine. Garbage in, garbage out.

          • MxM111@kbin.social
            link
            fedilink
            arrow-up
            1
            arrow-down
            26
            ·
            8 months ago

            I suggest you to ask ChatGPT4 questions in medicine and biology, and keep tally. If you truly think ChatGPT4 is garbage, you will be surprised. I spent lots of hours interacting with it, and I understand its limitations and strengths. And these kind of questions it usually answers quite well.

            • Jomega@lemmy.world
              link
              fedilink
              arrow-up
              21
              ·
              edit-2
              8 months ago

              I spent lots of hours interacting with it, and I understand its limitations and strengths.

              Considering you think it’s a substitute for a scholarly source, I doubt that. Once again, this is a machine designed to repeat things it heard. It’s a mechanical parrot. ChatGPT4 did not earn a degree. It did not study. It does not fact check. It does not give a solitary fuck about the scientific method. If you cannot see why this would be a problem for its credibility, then I can’t help you.

              The rest of it

              You just tried to use a glorified markov chain in an argument. Suffice to say I do not believe that you are the best judge of factual accuracy in regards to said tally.

              • MxM111@kbin.social
                link
                fedilink
                arrow-up
                1
                arrow-down
                9
                ·
                edit-2
                8 months ago

                I suggest you yourself test it, do not rely on me. It is experimentally can be shown as quite good predictor for these kind of questions. Don’t want to test yourself and don’t believe me? There a lot of tests were done of these models showing that they are already at the level to pass many exams. Your claim that it does not have any credibility is totally unfounded.

                Also, I never claimed that it is a substitute for scholarly source, I would never use it in a scientific paper. But I would not use Wikipedia either. But we are on internet on discussion board, the standards here are different. At least I supplied a source, majority of posts here don’t do that, including your statements, by the way, implying that ChatGPT4 has no credibility.

        • AA5B@lemmy.world
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          8 months ago

          Even if we assumed ChatGPT were completely accurate, that answer includes enough weasel words so it is not sufficient here.

          varies significantly among individuals … might retain …. may not completely reverse

          All this tells us is that making this fair is not an easy answer, anything more is individual interpretation. It probably needs some sort of medical consensus.

          How did they choose four years? Did they pull that out of their asses, or was there medical input? Is there a reason to expect ChatGPT to be a better source than what they used?

          • MxM111@kbin.social
            link
            fedilink
            arrow-up
            1
            ·
            8 months ago

            Four years was number given in the question, because this number was selected by boxing rules.

            As for the rest, are you suggesting that a post for discussion on internet should have the same level of credence as scientific article or article in encyclopedia? Why suddenly such super-high standards to my poste? Coincidently, no-one, despite criticism of my use of ChatGPT, pointed on even single mistake in that text.

            • AA5B@lemmy.world
              link
              fedilink
              arrow-up
              1
              ·
              edit-2
              8 months ago

              It doesn’t need a mistake.

              • Others pointed to it as being non-authoritative: we don’t need to prove it wrong, just that it doesn’t know. This is like listening to Aunt Marge on Facebook: I don’t have to know whether she’s right to be skeptical of her as an “expert”.
              • my contention was that even if we assume Au t Marge is right, the answer is insufficient fr the question.

              You’re complaining that four years of hormone therapy is insufficient without knowing where that came from or why, but we’re saying we’re not listening to Aunt Marge as more of an authority.

              I have no idea whether their decision is sound or what facts they base it on, but I’m also not taking Au t Marge’s word for it that they’re wrong

    • HikingVet@lemmy.ca
      link
      fedilink
      arrow-up
      33
      arrow-down
      3
      ·
      8 months ago

      You’re not a specialist, so you asked a computer algorithm to spit words at you. Fucking galaxy brain move.

    • littleblue✨@lemmy.world
      link
      fedilink
      arrow-up
      32
      arrow-down
      4
      ·
      8 months ago

      Oh, FFS. Here we go. “I did my research” == “I plunked letters into a box until the computer used words I don’t know. Science, removed!”

      We’re so fucked.

      • MxM111@kbin.social
        link
        fedilink
        arrow-up
        6
        arrow-down
        34
        ·
        8 months ago

        If your claim that what is posted is incorrect, then state so, and provide reasons. Otherwise, I do not understand why using tools such as ChatGPT4 you consider as something bad? Do you use search engines? They are tools too. Do you use wiki? That is a tool too. Or do you spit sarcasm on anyone who does not produce original research?

        • partial_accumen@lemmy.world
          link
          fedilink
          arrow-up
          19
          arrow-down
          3
          ·
          8 months ago

          If your claim that what is posted is incorrect, then state so, and provide reasons.

          “That which can be asserted without evidence, can be dismissed without evidence.” – Christopher Hitchens.

          • MxM111@kbin.social
            link
            fedilink
            arrow-up
            2
            arrow-down
            27
            ·
            8 months ago

            You do not consider ChatGPT4 as any sort of evidence? Go ahead and ask it questions in medicine or biology and keep tally how many answers are right and are wrong.

            I do admit that there is low probability that it is wrong, but simply dismissing it as no evidence at all is intentional dishonesty.

            • partial_accumen@lemmy.world
              link
              fedilink
              arrow-up
              8
              ·
              8 months ago

              I do admit that there is low probability that it is wrong

              By your own admission it can get things wrong, yet you’re arguing it should be trusted at face value.

              but simply dismissing it as no evidence at all is intentional dishonesty.

              The whole point of citing a source is so that you can confirm the veracity of the how the source came to its conclusion.You have no idea why the LLM gave you the answer it did. You don’t know how credible its input data was. Hopefully those involved in these discussions on both sides are searching for truth. The critical examination of the data and the origin of that data is the bedrock of that. Simply pasting raw LLM output doesn’t allow any of that to occur.

              LLM/AI ML can have a place in these discussions as a tool you use for yourself, and then you can search for supporting sources to back up the LLM’s claim. However, that’s work you have to do. Its not my job when you’re the one trying to convince me of your LLM’s conclusion.

              Dishonesty is passing off raw LLM output as researched fact. Its also lazy.

              • MxM111@kbin.social
                link
                fedilink
                arrow-up
                1
                arrow-down
                5
                ·
                8 months ago

                I am arguing that it should be given relatively high credence, not “trusted at face value”. Same as with Wikipedia, by the way. As an indication that likely things are true. On Internet forums it is much higher credence than most of the people supply. I am not writing scientific paper here, I am discussing topic with you. Would you rather me stating acts without any sources at all?

                For this discussion if you have different opinion, with better argumentation and sources please do so, and I will change my view. This is what discussion on discussion board suppose to be.

                And you can absolutely confirm the veracity (or not) of ChatGPT4 itself. You can ask the question yourself. You can collect statistics how likely it gives correct answers to similar questions, or find already published data about this topic. Based on that you can calculate probability that the statement is true. And it is much higher than 50%.

                In short, don’t attack the messenger, attack the message.

                • assassin_aragorn@lemmy.world
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  8 months ago

                  I suggest asking it about highly specialized technical topics or very specific details. AI either tends to get it wrong, or it’ll tell you it isn’t qualified to give an answer.

                  • MxM111@kbin.social
                    link
                    fedilink
                    arrow-up
                    1
                    arrow-down
                    1
                    ·
                    edit-2
                    8 months ago

                    I actually do do that. And quite often it does right. It very rare get’s it wrong. More often when it is “wrong” it will give you generic useless answer, not what you are asking, if it does not have information. But just opposite to truth? Not often. You can ask questions of a type “Describe main difference of Cynicism and Stoicism philosophy” or “Explain similarities and differences of EDFA and YDFA” And it give very reasonable answers.

                    This particular topic, however, is not highly specialized, or at least not specialized more than the questions I supplied above as examples. So I expect similar validity of the answer.

                • partial_accumen@lemmy.world
                  link
                  fedilink
                  arrow-up
                  1
                  ·
                  8 months ago

                  I am arguing that it should be given relatively high credence, not “trusted at face value”. Same as with Wikipedia, by the way.

                  I know you are, and I disagree. Your example of Wikipedia is a great differentiator.

                  The reason that Wikipedia is generally a good source is that it too cites its sources. If a Wikipedia entry makes a claim, I can see where that data came from or if its not cited, I know the claim is suspect and not to trust it. ChatGPT has none of that.

                  I am discussing topic with you. Would you rather me stating acts without any sources at all?

                  From my perspective not citing any source is exactly what you’re doing. ChatGPT isn’t a trusted or challenge-able source

                  And you can absolutely confirm the veracity (or not) of ChatGPT4 itself. You can ask the question yourself. You can collect statistics how likely it gives correct answers to similar questions, or find already published data about this topic.

                  If you want ChatGPT involved, that’s your job. Why is it you can’t use ChatGPT to find the real source which backs its claim?

                  Based on that you can calculate probability that the statement is true. And it is much higher than 50%.

                  "much higher that 50% is way way too low a bar to be considered a factual source.

                  In short, don’t attack the messenger, attack the message.

                  I can’t attack the message, its not backed by any sources to question it. My only option is to trust it absolutely, which is absurd.

                  • MxM111@kbin.social
                    link
                    fedilink
                    arrow-up
                    1
                    arrow-down
                    2
                    ·
                    8 months ago

                    Let me ask you a question differently. Do you think that the fact that ChatGPT stated what it stated has ANY impact on the credence of those statements?

                    Like what is more likely to be true, if I, none-specialist come up to those statements myself without doing any research OR come up to those statements BECAUSE ChatGPT stated that?

            • HikingVet@lemmy.ca
              link
              fedilink
              arrow-up
              9
              arrow-down
              1
              ·
              8 months ago

              ChatGPT4 is a fucking toy that regurgitates random shit it finds on the Internet. The only evidence it provides is the lack of understanding its user has.

            • Flying Squid@lemmy.world
              link
              fedilink
              arrow-up
              3
              ·
              8 months ago

              You do not consider ChatGPT4 as any sort of evidence?

              When it comes to science? No. ChatGPT does not write peer-reviewed journal articles.