Guardian investigation finds almost 7,000 proven cases of cheating – and experts says these are tip of the iceberg

Thousands of university students in the UK have been caught misusing ChatGPT and other artificial intelligence tools in recent years, while traditional forms of plagiarism show a marked decline, a Guardian investigation can reveal.

A survey of academic integrity violations found almost 7,000 proven cases of cheating using AI tools in 2023-24, equivalent to 5.1 for every 1,000 students. That was up from 1.6 cases per 1,000 in 2022-23.

Figures up to May suggest that number will increase again this year to about 7.5 proven cases per 1,000 students – but recorded cases represent only the tip of the iceberg, according to experts.

The data highlights a rapidly evolving challenge for universities: trying to adapt assessment methods to the advent of technologies such as ChatGPT and other AI-powered writing tools.

    • kescusay@lemmy.world
      link
      fedilink
      English
      arrow-up
      46
      arrow-down
      4
      ·
      1 day ago

      Why fight against it? Because some of these students will be going into jobs that are life-or-death levels of importance and won’t know how to do what they’re hired to do.

      There’s nothing wrong with using a large language model to check your essay for errors and clumsy phrasing. There’s a lot wrong with trying to make it do your homework for you. If you graduate with a degree indicating you know your field, and you don’t actually know your field, you and everyone you work with are going to have a bad time.

      • Cosmonauticus@lemmy.world
        link
        fedilink
        English
        arrow-up
        31
        arrow-down
        5
        ·
        1 day ago

        It’s almost like we shouldn’t value the importance of just passing the exam/ writing the paper and revamp our entire approach to teaching

          • CanadaPlus@lemmy.sdf.org
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            20 hours ago

            Yup. Although “avoid relying on writing and especially subjective long-form writing” is much more practicable.

        • aceshigh@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          1 day ago

          Ditto. A license just means you can pass a test. It doesn’t say anything more than that. That’s why you’re always advised to get 2nd opinions.

        • Deestan@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          edit-2
          1 day ago

          This is what I have been hoping LLMs would provoke since the beginning.

          Testing understanding via asking students to parrot textbooks with changed wording was always a shitty method, and one that de-incentivizes deep learning:

          It allows for teachers that do not understand their field beyond a superficial level to teach, and to evaluate. What happens when a student given the test question “Give an intuitive description of an orbit in your own words” answers by describing orbital mechanics in a relative frame instead of a global frame, when the textbook only mentions global frame? They demonstrate understanding beyond the material which is excellent but all they do is risk being marked down by a teacher who can’t see the connection.

          A student who only memorized the words and has the ability to rearrange them a bit, gets full marks no risk.

      • grte@lemmy.ca
        link
        fedilink
        English
        arrow-up
        15
        ·
        edit-2
        1 day ago

        Maybe we can get back to workplaces training their employees for the job they want them to do rather than putting the entire cost of training on 18 year oldish high school graduates for a job market that might not even be great (or exist) once they graduate post-secondary.

        • amorpheus@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          edit-2
          1 day ago

          Realistically, any job that you hire someone with a degree for will require training - the question is how much they’ll understand and at what level the training needs to start.

      • CanadaPlus@lemmy.sdf.org
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        21 hours ago

        If ChatGPT can consistently do the needed activities without raising eyebrows, it’s kind of a redundant job anyway. If they’re posing problems that have little to do with the job, it’s a shit course.

        • kescusay@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          20 hours ago

          That’s an odd thing to say. For one thing, there are plenty of physical activities that one could get a reasonable description of from ChatGPT, but if you can’t actually do them or understand the steps, you’re gonna have a bad time.

          Example: I’ve never seen any evidence that ChatGPT can properly clean and sterilize beakers in an autoclave for a chemical engineering laboratory, even if it can describe the process. If you turned in homework cribbed from ChatGPT and don’t actually know how to do it, your future lab partners aren’t going to be happy that you passed your course by letting ChatGPT do all the work on paper.

          There’s also the issue that ChatGPT is frequently wrong. The whole point here is that these cheaters are getting caught because their papers have all the hallmarks of having been written by a large language model, and don’t show any comprehension of the study material by the student.

          And finally, if you’re cheating to get a degree in a field you don’t actually want to know anything about… Why?

          • CanadaPlus@lemmy.sdf.org
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            1
            ·
            edit-2
            19 hours ago

            Well then make the students clean and sterilise beakers. It only kills assignments in written form with limited supervision, the way I assume calculators and traditional algorithms changed a bunch of things before my time. Lots of professors and educators in general are making exactly that kind of change.

            That means more lab time, it’s true, but it’s not the first time universities have had to expand their facilities. And at the other end think of all the savings on teaching writing.

            And finally, if you’re cheating to get a degree in a field you don’t actually want to know anything about… Why?

            It’s kind of unrelated, but money and prestige, probably. Caring about knowing for it’s own sake is depressingly rare.

    • Ech@lemm.ee
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 day ago

      Everyone should care about the efficacy of the education system. It’s the foundation of a useful society, and essential for its betterment. This method of cheating and its rate of adoption is a major blow to a whole generation of students that will learn to have others think for them instead of leaning to think for themselves.

    • Doorbook@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      3
      ·
      1 day ago

      I think universities care because they depend heavily on reputation. Companies will higher more graduate from a higher standard university.

      I think if no solution comes quickly or univirsties doesn’t adapt, this will render the whole echo system useless.

    • plantfanatic@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      11
      ·
      1 day ago

      I remember during school that translators were frowned upon, and now guess what, they are accepted as a normal tool.

      Ai is going through the exact same phase. Peoples memories are horrifically short.

        • plantfanatic@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          10
          ·
          edit-2
          1 day ago

          Googling answers at the time is no different, this just makes it easier. Anything can be shined in a bad light. Google has answers, translators do, calculators do, tutors do, it’s all the same in the end. They can also provide incorrect results too, so they are all essentially the same when finding the metric YOU choose.

          • I Cast Fist@programming.dev
            link
            fedilink
            English
            arrow-up
            5
            arrow-down
            1
            ·
            edit-2
            1 day ago

            The metric is “how much does this tool does the assignment for you”. A translator does zero (I’m assuming a translator at classes, though any “translate this” assignment coupled with google translate or similars would make this fit in with AI). Computers made the writing and copy-pasting part easier and faster, but they didn’t do the assignment. Google made the “find the correct stuff” easy, you no longer needed to manually look thru several books or ask people where the answer would be, but you still had to find it. Still, if an assignment asked for something that wasn’t perfectly answered on some page on the internet, like some random equation or specific programming code, you’d still have to work out something.

            With AI, you just throw the prompt and copy the result. “A patient arrived and has complained about severe chest pain. What procedures should be executed in which order?” - Don’t know, don’t care, AI wrote something, so it must be true. Copypaste, send, done.
            That leads to the point where, if the people who are supposed to actually know stuff only pretend to know thanks to AI covering their asses, why should anyone bother with them? Skip the literal middleman and just ask an AI the same thing they’d ask and be done with it. “AI doc, I feel like my heart is burning, what do I do?” - Is the answer right or wrong? Don’t know, don’t care, the person who should didn’t, either.