What are pros and cons of doing this? What impact it will have on the personality / mind of the person down the line after say 10 yrs?

  • AnarchistArtificer@slrpnk.net
    link
    fedilink
    English
    arrow-up
    29
    ·
    9 months ago

    A friend of mine is a French teacher, and I was discussing with her an idea for how to incorporate Chat-GPT into the curriculum. Specifically, her idea was to explore its limitations as a tool, by having a lesson in the computer suite where students actively try to answer GCSE (exams for 15/16 year olds) French questions using Chat-GPT, and then peer mark them, with the goal of “catching out” their peers.

    The logic was that when she was learning French in school, Google translate was still fairly new, and whilst many of her teachers desperately tried to ignore Google Translate, one teacher took the time to look at how one should (and shouldn’t) use this new tool. She said that it was useful to actually be able to evaluate the limitations of online translators, rather than just saying they’re always wrong and should never be used.

    We tried out a few examples to see whether her idea with Chat-GPT had merit and we found that it was pretty easy to generate errors that’d be hard to spot if you’re a student looking for a quick solution. Stuff like “I can’t answer that because I’m a large language model” or whatever, but in French.

    • JungleJim@sh.itjust.works
      link
      fedilink
      arrow-up
      18
      ·
      9 months ago

      That’s a great teacher. Refusing to teach a technology only leads to poor use. Even if one thinks it’s a poor technology, teach THAT instead of just black boxing the topic. The bottle is open, the genie is out. Better to teach how to make legally airtight wishes than to ban wishmaking.

  • u/lukmly013 💾 (lemmy.sdf.org)@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    23
    ·
    edit-2
    9 months ago

    It’s just AI chatbot, I don’t see how it would be dangerous.

    And I am also pretty sure a 16 year old knows to expect inaccurate results from it, unless they’ve been living restricted from the outside world until now.

    The only negative thing I see from it so far is kids using it to create essays, but it’s not like there wasn’t a countless number of them available on the internet before. It was just easier to detect as you could search up the text and see if you can find it online.

    Anyway, for just playing around it gets boring after 15 minutes.
    Why don’t you try?

    • LWD@lemm.ee
      link
      fedilink
      arrow-up
      10
      ·
      edit-2
      9 months ago

      Something that appears more human is more likely to elicit them sending their private data. And that data is then sold, obviously without consent, and used however the buyers feel.

      Instead of being scared to share information with it, you will volunteer your data…

      – Vladimir Prelovac, CEO of Kagi AI and Search

      Remember Replika, the AI chatbot that sexually harassed minors and SA victims, and (allegedly) repeated the contents of other people’s messages verbatim?

      It might not be as mind-rotting as TikTok but it’s not good.

  • lemontree@lemm.ee
    link
    fedilink
    arrow-up
    21
    ·
    9 months ago

    I would go back a few years and ask: Should i let a 16 year old use search engines?

    Probably not too different

    • PrinceWith999Enemies@lemmy.world
      link
      fedilink
      arrow-up
      9
      ·
      9 months ago

      That’s exactly my perspective.

      I came of age with the birth of the web. I was using systems like Usenet, gopher, wais, and that sort of thing. I was very much into the whole cypherpunk, “information wants to be free” philosophy that thought that the more information people had, the more they could talk to each other, the better the world would be.

      Boy, was I wrong.

      But you can’t put the genie back into the bottle. So now, in addition to having NPR online, we have kids eating tide pods and getting recruited into fascist ideologies. And of course it’s not just kids. It’s tough to see how the anti-vax movement or QAnon could have grown without the internet (which obviously has search engines as a major driver of traffic).

      I think you’re better off teaching critical thinking, and even demonstrating the failings of ChatGPT by showing them how bad it is at answering questions. There’s plenty of resources you can find that should give you a starting point. Ironically, you can find them using a search engine.

      • rufus@discuss.tchncs.de
        link
        fedilink
        arrow-up
        3
        ·
        edit-2
        9 months ago

        I think that’s a good take on things.

        Ultimately it still holds true. Information does want to be free. You just can’t mix that with misinformation, have everything on the same level and a general audience completely oblivious to the fact and uneducated.

        Things have changed. Back in those times it was a small elite on the internet. People who could afford computers and an internet connection and make some use out of it. You needed some amount of intelligence because you had to put some effort in to get online, learn about the tools because that wasn’t easy or provided to you. So you’d generally be at least somewhat intelligent if you ended up on the internet. And that’s beneficial when it comes to receiving unfiltered information. Combined with the fact that there were comparatively more academics and students, because that was the origin of the internet.

        And it wasn’t that common to push your agenda there or advertise for your skewed political views in the way people do it nowadays. Due to the nature of the internet and the amount of people there, it wasn’t worth the effort. You’d be better off focusing somewhere else where you could influence more people. So the dynamics were just different due to history and circumstances.

        Things have changed. Nowadays everyone is online all the time. It’s the place to influence people and make money. And that’s the other part of the problem. The actual people, connecting them and providing information to them (or to each other) isn’t what’s most of the internet is about, anymore. Motivations are gathering data about people and selling them, making people become addicted to your platform so they spend more time there and you can make more money. Everyone is competing for attention. And bad, emotional stories are what works best. Giving people the “simple truths” they seek instead of an intellectual and nuanced view. Factuality just gets in the way of all of that.

        I sometimes like to compare that to the Age of Reason / Enlightenment. Back then it was monarchs, bad dynamics and missing education. Now it’s big tech companies, bad dynamics and insufficient education. People need to get emancipated, educated and leave the current “immature state of ignorance” (to quote Kant.)

        Information and education are key. And the internet, algorithms and AI are just tools. They can be used for progress, or to enslave us. At least the internet has the potential (and was build) to connect people and provide a level playing field to everyone. But it can be used for a variety of different things. And choosing the right things isn’t something that can be solved by technology alone.

  • amio@kbin.social
    link
    fedilink
    arrow-up
    14
    ·
    9 months ago

    It’s not even a good idea to let quite a lot of adults use ChatGPT. People don’t know how it works, don’t treat the answers with anything close to appropriate skepticism, and often ask about things they don’t have the knowledge/skills to verify. And anything it tells you, you likely will need to verify.

    It’s quite unlikely to affect their personality, but it might make them believe a bunch of weird shit that some unknowable, undebuggable computer program hallucinated up. If you’ve done an uncommonly great job with their critical thinking skills, great. If not, better get started. That is not specific to “AI” though.

    • NoiseColor@startrek.website
      link
      fedilink
      arrow-up
      4
      ·
      9 months ago

      People don’t know how TV works and we are hardly gonna tell people not to use it.

      As long as people are aware that some responses might be made up it should be fine for anyone to use it.

  • Saigonauticon@voltage.vn
    link
    fedilink
    English
    arrow-up
    12
    ·
    9 months ago

    I think it would be a bad idea to do otherwise. Children need to learn about useful tools, and the shortcomings of those tools.

    16 year old me would have had a great time getting an AI to teach me things that my teachers in school did not have expertise in. Sure, it would be wrong some of the time, but so were my teachers at that age. It would have given me such a head start on university!

  • AlwaysNowNeverNotMe@kbin.social
    link
    fedilink
    arrow-up
    12
    ·
    9 months ago

    The context of the word “let” is interesting here.

    I would recommend a collaborative approach, it’s not as if they can’t use it because you tell them no. They don’t need a credit card or a driver’s license or even a computer.

  • TheEntity@kbin.social
    link
    fedilink
    arrow-up
    11
    ·
    9 months ago

    You cannot let or forbid a 16yo to use stuff. You can only decide whether they will do it in the open or in hiding. Personally I’d rather have them talk to me about it than hide it from me.

  • AFK BRB Chocolate@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    ·
    9 months ago

    I agree with those saying you can’t/shouldn’t forbid it. As someone in computer science, the important thing to me is to make sure they understand what those LLMs are and aren’t. Specifically, the ‘M’ in “LLM” is for “model” - they’re a detailed model of what a conversation should look like, especially what a response to a question should look like. But looking right is different from being correct. You can ask one for a mathematical proof and it will give you one that looks right, but it probably won’t be.

    The other thing I’d try to get them to understand is that the leaning part of school is much more important than the grade part, especially if they’re going to go on to college. They could use an LLM to help them create a term paper, but if they didn’t learn anything it’s going to catch up to them and cause problems down the road.

    • Scrubbles@poptalk.scrubbles.tech
      link
      fedilink
      English
      arrow-up
      7
      ·
      9 months ago

      Yeah having an open and honest conversation seems like the best thing, but that requires Parent here also understanding it, but that’s a good time for both of them to look into it and learn more.

      The biggest thing is going to be something along the lines of “I know you’re going to want to use this for homework, but I want to ask you to please not just use it as a way to get the answer. This very well may be a tool that helps you understand problems better, we all learn differently and maybe this can help explain confusing questions to you in a way that you can understand, but if you just ask for the answer you won’t learn anything”.

      Being honest with them on why you don’t want them to just plain use it will go a long way. Teens are (in some ways) smart, they know if you’re just forbidding it that there’s probably a reason but that you don’t want to explain it, and so they’ll rebel and use it more. Being honest and explaining your reasoning will usually sit longer with them. Sure, they’ll probably say “What is X?”, but maybe they’ll adjust it to say “Can you show me how to solve for X?”

      • FaceDeer@fedia.io
        link
        fedilink
        arrow-up
        4
        ·
        9 months ago

        Yeah, I would heartily recommend LLMs being used in education as a tutor and “homework buddy”. I find their interactive nature to be really useful for learning stuff - I’m always able to ask “wait, what did that bit mean?” or “Walk me through this part.” The LLM isn’t always right, but with that back-and-forth it’s quick to catch errors.

  • 𝘋𝘪𝘳𝘬@lemmy.ml
    link
    fedilink
    arrow-up
    9
    ·
    9 months ago

    To use as a tool? Yes.
    To use as a friend? No.

    A person using a tool for a longer time will become better in using said tool.

    • driving_crooner@lemmy.eco.br
      link
      fedilink
      arrow-up
      3
      ·
      9 months ago

      I use ChatGPT for help in my MBA in actuarial sciences everyday. I always start with “pretend you are an statistics/probability/finances professor and I’m and advanced student. We are working on {topic} and I need help with {thing I didn’t fully understood in class}” l. It have been fantastic. Even fkg Terence Tao is using ChatGPT as help in the most advanced mathematics is out there.

  • yokonzo@lemmy.world
    link
    fedilink
    arrow-up
    7
    ·
    9 months ago

    I’ll be honest, if I got it at 16, I would fuck around with it for a few weeks and then get bored

  • CanadaPlus@lemmy.sdf.org
    link
    fedilink
    arrow-up
    7
    ·
    9 months ago

    At 16 they should be just as capable of understanding the limitations as anyone. Just be sure to explain that it has no interest in truth, but only writing convincingly.

    I doubt it will have personality impacts. The one thing that could be an issue is if they use it as a replacement for real human friends.

  • PeepinGoodArgs@reddthat.com
    link
    fedilink
    English
    arrow-up
    6
    ·
    9 months ago

    Have you asked ChatGPT? Jk lol

    Honestly, whatever they use ChatGPT for is probably fine. If you feel like they’re going to cheat on their homework or something, you can just ask them to do a small sample in front of you. Plus, it’s not like ChatGPT is going away, no matter how much the NYT and Disney complain. Best bet is for them to get familiar with the technology now.

    Also, there’s literally no way to the long-term effects of AI. I strongly suspect that if people use it as a crutch, it will create intellectually and creatively stunted people. But it’s not like we don’t have that now…

  • wathek@discuss.online
    link
    fedilink
    arrow-up
    5
    ·
    edit-2
    9 months ago

    ChatGPT is overly safe in terms of personality and the worldview it presents when asked. it’s a great tool to learn, more so than a teacher because you can freely ask it very specific questions in your own words and it will give an understandable answer. I think it’s actually a perfect tool for someone that age. Once the topics get too advanced, the results become less reliable though.

    It doesnt make things up anymore as much as it used to. It still does sometimes with topics that are less commonly discussed in the dataset it’s trained on (this is similar with websearch). It will however confidently claim that it’s answer is correct sometimes. As long as you understand that it’s not always correct and have the sense to verify things that seem off, you’ll be fine.

    You’ll get the best results from the paid GPT4 subscription (20 dollars a month), which i would recommend.

    The only real risk i see is overreliance on it. I notice this in myself too, it’s almost like i forgot googling things is an option, so when i’m stuck rather than trying another approaxh, i just keep throwing prompts at GPT-4 until i give up and find the solution elsewhere, often within minutes. The way things are going, classic web search is becoming obsolete (unreliable result because of AI written content and fake news) while AI actively tries to be unbiased.

    tldr: Yes, it’s extremely useful, make sure they don’t forget how to do things without chatgpt too.