• PoisonedPrisonPanda@discuss.tchncs.de
    link
    fedilink
    arrow-up
    19
    arrow-down
    3
    ·
    edit-2
    1 年前

    The whole system should get ready for the 21st century.

    Most of the scientists arent great writers. It does not make sense to still force them to be a good writer.

    Let be fishes be good at swimming instead of climbing trees.

    In a modern world where basically EVERYTHING is specialized and no generalist is alive anymore we should make use of language tools.

    Hell Chatgpt writes an introduction which is fun to read instead or my overcomplicated bullshit that I would have brought up

    Edit: the comment was not related to the OP but to a general chatgpt discussion.

    • sab@kbin.social
      link
      fedilink
      arrow-up
      12
      arrow-down
      4
      ·
      1 年前

      One important thing is that you have potential. ChatGDP will write something alright-ish, but it’s literally impossible for it to move beyond that. It doesn’t have the power of creativity.

      Writing is painful, but it also helps us think clearer about our work and contribution. I think it’s an important part of the process of doing science, no matter which field. And one gets better at it with training.

        • Vilian@lemmy.ca
          link
          fedilink
          arrow-up
          7
          ·
          1 年前

          a test for creativity seriously that work? also after scraping the entire of internet of course someone could think that, ask any programmer and they gonna explain that the IA don’t create anything, it can’t even do basic msth because it don’t gave logic in that,maybe one day, but not with chatgpt of today

          • FaceDeer@kbin.social
            link
            fedilink
            arrow-up
            2
            ·
            1 年前

            Yes, a test for creativity. If you’re going to say something “doesn’t have the power of creativity” then it behooves you to accept the notion that creativity is measurable.

          • boredazfcuk@mastodon.social
            link
            fedilink
            arrow-up
            0
            ·
            1 年前

            @Vilian @FaceDeer I agree. I’m no programmer but do a fair bit of Linux/powershell/bash scripting. Virtually all the code that ChatGPT gives me is wrong. You tell it the errors, and it gives you a modified script with errors, point out those errors and it’s go back to its first answer. The only thing it is useful for is writing lots of basic code, really quickly. I can just copy/paste then start debugging.

            • FaceDeer@kbin.social
              link
              fedilink
              arrow-up
              1
              ·
              1 年前

              I am a programmer and I’ve found ChatGPT to be able to produce plenty of good, useful code. I haven’t encountered the problems you’re describing in correcting its errors, perhaps you’re not prompting it well.

                • FaceDeer@kbin.social
                  link
                  fedilink
                  arrow-up
                  0
                  ·
                  1 年前

                  So you’re telling me that the code it generated for me wasn’t good and useful, and that when I told it to correct errors it actually did introduce new errors and restore old ones, contrary to what I just said? Guess all that stuff I got done using its help didn’t actually get done after all and I’m descending ever deeper into a world of delusion, thinking my projects are finished and working when in fact they aren’t.

                  Obviously if you’re trying to get it to use APIs from after 2021 that’s not going to work. It also won’t bake you a cake if you ask it to. Use tools for the tasks they’re good for, don’t use them for things you know they can’t do.

                  • boredazfcuk@mastodon.social
                    link
                    fedilink
                    arrow-up
                    0
                    ·
                    1 年前

                    @FaceDeer @floofloof @henfredemars @PoisonedPrisonPanda @sab @Vilian the 500 char limit made me pick the first failure that spring to mind. Maybe you forget, but AI wasn’t trained on “good” data. It was trained “all” data and large amounts of that is plain wrong. People with problems pasting blocks of code and responses correcting a single line. ChatGPT isn’t smart enough to merge those into a single block of working code.

          • sab@kbin.social
            link
            fedilink
            arrow-up
            1
            arrow-down
            1
            ·
            1 年前

            This is the key - it does not create, it can only copy. Which is good enough to fool us - there’s enough stuff to copy out there that you can spend your whole life copying other people and nobody will ever notice you’re not actually creating anything new. What’s more, you’ll probably come across as pretty clever. But you’re not creating anything new.

            For me, this poses an existential threat to academia. It might halt development in the field without researchers even noticing: Their words look fine, as if they had thought it through, and they of course read it to make sure it’s logically consistent. However, the creative force is gone. Nothing new will come under the sun - the kind of new thoughts that can only be made by creative humans thinking new thoughts that have never been put on paper before.

            If we give up that, what’s even the point of doing science in the first place.

            • FaceDeer@kbin.social
              link
              fedilink
              arrow-up
              1
              ·
              1 年前

              This is the key - it does not create, it can only copy.

              I have asked ChatGPT to write poetry on subjects that I know with great certainty have never had poems written about them.

              You can of course shuffle around the meanings of “create” and “copy” to try to accommodate that, but eventually you end up with a “copying” process that’s so flexible and malleable that it might as well be creativity. It’s not like what comes out of human brains isn’t based on stuff that went into them earlier either.

            • Sinnerman@kbin.social
              link
              fedilink
              arrow-up
              1
              ·
              1 年前

              There’s a difference between:

              1. Using ChatGPT to help write parts of the text in the same way you’d use a grammar- or spell-checker (e.g. if English isn’t your first language) after you’ve finished the experiments

              2. Using ChatGPT to write a paper without even doing any experiments

              Clearly the second is academic misconduct. The first one is a lot more defensible.

              • sab@kbin.social
                link
                fedilink
                arrow-up
                2
                arrow-down
                2
                ·
                1 年前

                Yes, absolutely. But I still think it has its dangers.

                Using it to write the introduction doesn’t change the substance of the paper, yet it does provide the framework for how the reader interprets it, and also often decides whether it’ll be read at all.

                Maybe worse, I find that it’s oftem in the painful writing and rewriting of the introduction and conclusion that I truly understand my own contribution - I’ve done the analysis and all that, but in forcing myself to think about the relevance for the field and the reader I also bring myself to better understand what the paper means in a deeper sense. I believe this kind of deep thinking at the end of the process is incredibly valuable, and it’s what I’m afraid we might be losing with AI.

      • PoisonedPrisonPanda@discuss.tchncs.de
        link
        fedilink
        arrow-up
        0
        arrow-down
        1
        ·
        1 年前

        I dont need it to be beyond that? It does what I told it. And if I am creative enough to get my preferred output its great. I have still to decide if Ill use it.

        Its a tool which can be used by people and helps with work.

        I think it’s an important part of the process of doing science, no matter which field. And one gets better at it with training

        Sorry but this expression is probably a similar one when paper writting shifted to digital only format or when the typewriter was introduced.

        Boomer tell me the same with printed paper. “oNlY whEn ItS PriNtED yOu cAn rEaD pRoberly”

        Thats bullshit its just fear of the something new and convenience of routine.

        Nothing personal against you. I welcome any tool that helps me.

        • bane_killgrind@lemmy.ml
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 年前

          So I’m not sure it’s helping you.

          You would refrain from doing the work of organizing the concept in your head into a clearly communicable explanation of the concept.

    • henfredemars@infosec.pub
      link
      fedilink
      English
      arrow-up
      3
      ·
      edit-2
      1 年前

      It’s interesting that you write this because the last place I worked focused on unspecializing by having almost everyone do every job.

      In fact, they relocated across the country to save on building costs, and instead of hiring actual technical writers and office staff, they pushed the extra work down on their engineers because it’s more profitable to bill for the engineering time.

      I spent much of my job editing papers and I’m not even good at it while getting paid to do embedded design. It was weird. It was basically fraud but walking the fine line of technically legal.

      I observed this happening multiple times throughout my career. Sometimes, inefficiency is the point in this case driven by capitalists and market forces.

    • stockRot@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      1 年前

      … Did you read the article? Language tools like grammarly and deepL are in use by scientists today. Copying+pasting the output of chatGPT without ever looking at it, or even using a language tool to publish thoughts that were never in your head to begin with, is the actual concern

      • Did you read the article?

        I for sure didnt.

        Thanks for highlighting that.

        I was carried away by having the discussions at my university with my peers in mind.

        Copying+pasting the output of chatGPT without ever looking at it, or even using a language tool to publish thoughts that were never in your head to begin with, is the actual concern

        Nevertheless I dont understand why this is a concern.

        The scientific standards existed decades if not already at least a century.

        Those discussions are putting chatgpt in a bad light. However the fact that our scientific system was eroded and made a mockery of before the introduction of chatgpt is not highlighted.

        There are still plagiarizations around and nobody cares. Mostly because of political sensitivity.

        However science has failed to repel “bad actors” (intentional or unintentional) from the scene.

        I dont know when. And why. But publisher have for sure something to do with it.

    • monobot@lemmygrad.ml
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      1 年前

      I agree, I have no problem with people guiding chatgpt to help them write something they want and they checked it.

      Generating bunch of articles even they didn’t read is something else.