• LemmysMum@lemmy.world
    link
    fedilink
    arrow-up
    2
    arrow-down
    3
    ·
    edit-2
    1 year ago

    The ai does have perceptions, fed into by us as inputs. I give the ai my perceptions, the ai creates a facsimile, and I adjust the perceptions I feed into the ai until I receive an output that meets the needs of my requirements, no different from doing it myself except I didn’t need to read all the books, and learn all the lessons myself. I still tailor the end product, just not to the same micro scale that we needed to traditionally.

    • Phanatik@kbin.social
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      You can’t feed it perceptions no more than you can feed me your perceptions. You give it text and the quality of the output is determined by how the LLM has been trained to understand that text. If by feeding it perceptions, you mean by what it’s trained on, I have to remind you that the reality GPT is trained on is the one dictated by the internet with all of its biases. The internet is not a reflection of reality, it’s how many people escape from reality and share information. It’s highly subject to survivorship bias. If the information doesn’t appear on the internet, GPT is unaware of it.

      To give an example, if GPT gives you a bad output and you tell it that it’s a bad output, it will apologise. This seems smart but it’s not really. It doesn’t actually feel remorse, it’s giving a predetermined response based on what it’s understood by your text.

      • LemmysMum@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        edit-2
        1 year ago

        We’re not talking about perceptions as in making an AI literally perceive anything. I can feed you prompts and ideas of my own and get an output no different than if I was using AI tools, the difference being ai tools have already gathered the collective knowledge you’d get from say doing a course in photoshop, taking an art class, reading an encyclopaedia or a novel, going to school for music theory, etc.

        • Phanatik@kbin.social
          link
          fedilink
          arrow-up
          1
          ·
          1 year ago

          I get that part but I think what gets taken more seriously is how 'human" the responses seem which is a testament to how good the LLM model is. But that’s set dressing when GPT has been known to give incorrect, outdated or contradictory answers. Not always but unless you know what kind of answer to expect, you have to verify what it’s telling you which means you’ll be spending half the time fact-checking the LLM.

          • LemmysMum@lemmy.world
            link
            fedilink
            arrow-up
            1
            ·
            edit-2
            1 year ago

            Exactly, how is the end result not that of the user if they need to craft and modify and adjust and manipulate the prompts inputs and outputs of ai to produce something new or coherent?

            It’s just a tool. A tool that will improve access to human knowledge and improve each individuals ability to create and produce more complex works with less effort. Each of which will feed back into the algorithm expanding the knowledge and capacity of ai and human ingenuity.