Feel like we’ve got a lot of tech savvy people here seems like a good place to ask. Basically as a dumb guy that reads the news it seems like everyone that lost their mind (and savings) on crypto just pivoted to AI. In addition to that you’ve got all these people invested in AI companies running around with flashlights under their chins like “bro this is so scary how good we made this thing”. Seems like bullshit.

I’ve seen people generating bits of programming with it which seems useful but idk man. Coming from CNC I don’t think I’d just send it with some chatgpt code. Is it all hype? Is there something actually useful under there?

  • thelastknowngod@lemm.ee
    link
    fedilink
    arrow-up
    16
    arrow-down
    1
    ·
    1 year ago

    Same. 5 minutes after installing Copilot I literally said out loud, “Well… I’m never turning this off.”

    It’s one of the nicest software releases in years. And it’s instantly useful too… No real adjustment period at all.

    • GarlicBender@lemmy.ml
      link
      fedilink
      arrow-up
      7
      ·
      1 year ago

      I tried it for a couple months and it was alright but eventually it got too frustrating. I did love how well it did some really repetitive things. But rarely did it actually get anything complex 100% right. In computing, “almost right” is wrong. But because it was so close, it was hard to spot the mistakes.

      There were cases where my IDE knew the right answer but Copilot did not. Realizing that Copilot was messing up my IDE enhancements to produce code I was painfully babysitting, I cancelled it.

      • sLLiK@lemmy.ml
        link
        fedilink
        arrow-up
        5
        ·
        edit-2
        1 year ago

        This is the most insidious conundrum related to AI usage. At the end of the day, a LLM’s top priority is to ensure that your question is answered in a way that satisfies that model. The accuracy of its answers are a secondary concern. If forced to choose between making up BS so it can have a response that looks right versus admitting it doesn’t have enough information to answer, it can and often will choose the former. Thus the “hallucination” problem was born.

        The chance of getting your answer lightly sprinkled with made up stuff is disturbingly high. This transfers the cognitive load of the AI user from “what is the answer” to “I must repeatedly go verify everything in this answer because I can’t trust it”.

        Not an insurmountable obstacle, and they will likely solve it sooner rather than later, but AI right now is arguably the perfect extension of the modern internet - take absolutely everything you read with at least a grain of salt… and keep a pile of salt cubes close by.