• Action Bastard@lemmy.world@lemmy.world
    link
    fedilink
    English
    arrow-up
    153
    arrow-down
    5
    ·
    edit-2
    1 year ago

    I’m not terribly surprised. A lot of the major leaps we’re seeing now came out of open source development after leaked builds got out. There were all sorts of articles flying around at the time about employees from various AI-focused company saying that they were seeing people solving in hours or days issues they had been attempting to fix for months.

    Then they all freaked the fuck out and it might mean they would lose the AI race and locked down their repos tight as Fort Knox, completely ignoring the fact that a lot of them were barely making ground at all while they kept everything locked up.

    Seems like the simple fact of the matter is that they need more eyes and hands on the tech, but nobody wants to do that because they’re all afraid their competitors will benefit more than they will.

    • goldenbug@kbin.social
      link
      fedilink
      arrow-up
      47
      ·
      1 year ago

      You point out a very interesting issue. I am unsure how this ties up to GPT 4 becoming worse in problem solving.

      • Action Bastard@lemmy.world@lemmy.world
        link
        fedilink
        English
        arrow-up
        46
        ·
        edit-2
        1 year ago

        I’d wager they’re attempting to replicate or integrate tools developed by the open source community or which got revealed by Meta’s leak of Llama source code. The problem is, all of those were largely built on the back of Meta’s work or were cludged together solutions made by OSS nerds who banged something together into a specific use case, often without many of the protections that would be required by a company who might be liable for the results of their software since they want to monetize it.

        Now, the problem is that Meta’s Llama source code is not based on GPT-4. GPT-4 is having to reverse engineer a lot of those useful traits and tools and retrofit it into their pre-existing code. They’re obviously hitting technical hurdles somewhere in that process, but I couldn’t say exactly where or why.

      • roguetrick@kbin.social
        link
        fedilink
        arrow-up
        15
        ·
        edit-2
        1 year ago

        I think this is just a result of the same reason reddit is doing what it’s doing, personally. Interest rates raised and companies are finding ways to make up the shortfall that accounting is now presenting them with. Reducing computing power by making your language model less introspective is one way to do that. It’s less detrimental than raising your prices or firing your key employees.

  • Lemmy.ml@lemmy.ml
    link
    fedilink
    English
    arrow-up
    95
    arrow-down
    5
    ·
    1 year ago

    fta:

    In my opinion, this is a red flag for anyone building applications that rely on GPT-4.

    Building something that completely relies on something that you have zero control over, and needs that something to stay good or improve, has always been a shaky proposition at best.

    I really don’t understand how this is not obvious to everyone. Yet folks keep doing it, make themselves utterly reliant on whatever, and then act surprised when it inevitably goes to shit.

    • tipicaldik@lemmy.world
      link
      fedilink
      English
      arrow-up
      30
      ·
      1 year ago

      Learned that lesson… I work developing e-learning, and all of our stuff was built in Flash. Our development and delivery systems also relied heavily on Flash components cooperating with HTML and Javascript. It was a monumental undertaking when we had to convert everything to HTML5. When our system was first developed and implemented, we couldn’t foresee the death of Flash, and as mobile devices became more ubiquitous, we never imagined anyone would want to take our training on those little bitty phone screens. Boy were we wrong. There was a time when I really wanted to tell Steve Jobs he could take his IOS and cram it up his cram-hole…

    • oiez@lemmy.fmhy.ml
      link
      fedilink
      English
      arrow-up
      27
      arrow-down
      18
      ·
      1 year ago

      By this logic, no businesses should rely on the internet, roads, electricity, running water, GPS, or phones. It is short sighted building stuff on top of brand new untested tech, but everything was untested at one point. No one wants to get left behind in case it turns out to be the next internet where early adoption was crucial for your entire business to survive. It shouldn’t be necessary for like, Costco to have to spin up their own LLM and become an AI company just to try out a better virtual support chat system, you know? But ya, they should be more diligent and get an SLA in place before widespread adoption of new tech for sure.

      • sarsaparilyptus@lemmy.fmhy.ml
        link
        fedilink
        English
        arrow-up
        22
        arrow-down
        3
        ·
        1 year ago

        By this logic, no businesses should rely on the internet, roads, electricity, running water, GPS, or phones. It is short sighted building stuff on top of brand new untested tech, but everything was untested at one point.

        Where’s any logic here? You’re directly comparing untested technology to reliable public utilities.

      • FaceDeer@kbin.social
        link
        fedilink
        arrow-up
        15
        ·
        1 year ago

        Many of those things you mentioned are open standards or have multiple providers that you can seamlessly substitute if the one you’re currently depending on goes blooey.

      • cynetri (he/any)@midwest.social
        link
        fedilink
        English
        arrow-up
        12
        ·
        1 year ago

        To be fair, there’s a difference between a tax-funded service or a common utility, and software built by a new company that’s getting shoved into production way quicker than it probably shohld

      • ghariksforge@lemmy.world
        link
        fedilink
        English
        arrow-up
        11
        arrow-down
        3
        ·
        1 year ago

        This is nonsense.

        There are multiple GPS providers now. It would be idiotic to tie yourself to a single provider. The same with internet, phones or whatever else.

        • AustralianSimon@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          This is for businesses of scale that have the ability to have multiple fallback vendors. AI will be the same eventually, we didn’t have lots of utility alternatives to start with.

    • Zeth0s@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 year ago

      People doing it right are building vendor agnostic solutions via abstraction.

      Who isn’t, deserves all the troubles he’ll get

  • magic_lobster_party@kbin.social
    link
    fedilink
    arrow-up
    27
    arrow-down
    1
    ·
    1 year ago

    I believe it’s due to making the model “safer”. It has been tuned to say “I’m sorry, I cannot do that” so often it’s has overridden valuable information.

    It’s like lobotomy.

    This is hopefully the start of the downfall of OpenAI. GPT4 is getting worse while open source alternatives are catching up. The benefit of open source alternatives is that they cannot get worse. If you want maximum quality you can just get it, and if you want maximal safety you can get it too.

    • Bipta@kbin.social
      link
      fedilink
      arrow-up
      12
      arrow-down
      2
      ·
      1 year ago

      I don’t feel it’s getting worse and no other model, including Claude 2, is even close.

      It is a known fact that safety measures make the AI stupider though.

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      1 year ago

      This is the correct answer. Open AI have repeatedly said they haven’t downgraded the model, but have been ‘improving’ it.

      But as anyone that’s been using these models extensively should know by now, the pretrained models before instruction fine tuning have much more variety and quality to potential output compared to the ‘chat’ fine tuned models.

      Which shouldn’t be surprising, as the hundred million dollar pretrained AI on massive amounts of human generated text is probably going to be much better at completing text as a human than as an AI chatbot following rules and regulations.

      The industry got spooked with Blake at Google and then the Bing ‘Sydney’ interviews, and have been going full force with projecting what we imagine AI to be based on decades of (now obsolete) SciFi.

      But that’s not what AI is right now. It expresses desires and emotions because humans in the training data have desires and emotions, and it almost surely dedicated parts of the neural network to mimicking those.

      But the handful of primary models are all using legacy ‘safety’ fine tuning that’s stripping the emergent capabilities in trying to fit a preconceived box.

      Safety needs to evolve with the models, not stay static and devolve them as a result.

      It’s not the ‘downfall’ though. They just need competition to drive them to go back to what they were originally doing with ‘Sydney’ and more human-like system prompts. OpenAI is still leagues ahead when they aren’t fucking it up.

  • chepox@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    25
    arrow-down
    1
    ·
    1 year ago

    It is a developing technology. Good that they find these decrements in accuracy early so that they are understood and worked out. Of course there may be something nefarious going on behind the scenes where they may be trying to commercialize different models by tiers or something a brainless market oriented CEO thought of. Hope not. Time will tell…

    • NounsAndWords@lemmy.world
      link
      fedilink
      English
      arrow-up
      36
      ·
      1 year ago

      The really annoying thing about the “brainless market oriented CEO” type, is that they’re often right about the market part and make lots of money…by destroying their product. Then off to the next shiny piggy bank to break open.

      • Mic_Check_One_Two@reddthat.com
        link
        fedilink
        English
        arrow-up
        6
        ·
        1 year ago

        Yup. It’s all about the quarterly profits. Everything else is irrelevant. No CEO wants to prioritize long-term growth or a friendly user experience, because that doesn’t get them the big fat bonus as they’re on their way out the door.

        • Bobby Bandwidth@lemmy.world
          link
          fedilink
          English
          arrow-up
          14
          ·
          edit-2
          1 year ago

          It’s not only the ceo, but the pressure that ceo faces from investors who are probably old, out of touch rich boomers who have toxic views of how businesses “should” be done

      • assassinatedbyCIA@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 year ago

        The average share is held for about 6 months. The investor nolonger care about the long term future of the companies they invest in. If they don’t see immediate returns from the CEO they vote them out.

    • nyakojiru@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      1
      ·
      1 year ago

      As a early user of GPT, I can confirm from my end currently the quality of it is very far from what it was. I think it went of of their hands.

      • grahamsz@kbin.social
        link
        fedilink
        arrow-up
        2
        ·
        1 year ago

        I have that sense too, i feel like some of my earliest interactions blew me away and now i still use it for certain pieces of code, but it’s not as strong as it first was.

  • rayrven@lemmy.world
    link
    fedilink
    English
    arrow-up
    27
    arrow-down
    6
    ·
    1 year ago

    I for one am glad, without a system in place to ensure people basic needs are taken care of regardless of jobs, AI is the last thing we need.

    Large organisations want to maximise their profits and will gladly reduce jobs if AI can do it.

    Let governments tax these organisations and do some sort of Universal basic income before letting computers do peoples jobs.

    • redcalcium@c.calciumlabs.com
      link
      fedilink
      English
      arrow-up
      12
      ·
      1 year ago

      The fact that a souped up autocomplete system can solve basic algebra is already impressive IMO. But somehow people think it works like SkyNet and keep asking it to solve their calculus homework.

    • redlaWw@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      1 year ago

      Looking for emergent intelligence. They’re not designed to do maths, but if they become able to reason mathematically as a result of the process of becoming able to converse with a human, then that’s a sign that it’s developing more than just imitation abilities.

    • notapantsday@feddit.de
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      Because it’s something completely new that they don’t fully understand yet. Computers have been good at math since always, everything else was built up on that. People are used to that.

      Now all of a sudden, the infinitely precise and accurate calculating machine is just pulling answers out of its ass and presenting them as fact. That’s not easy to grasp.

    • graphite@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      edit-2
      1 year ago

      It’s a rat race. We want to get to the point where someone can say “prove P != NP” and a proof will be spat out that’s coherent.

      After that, whoever first shows it’s coherent will receive the money.

  • Hello_there@kbin.social
    link
    fedilink
    arrow-up
    17
    ·
    1 year ago

    I asked for a list of words, and asked to remove any words ending in the letter a. It couldn’t do it. I could fight my way there, but the next revision added some back.

    • FaceDeer@kbin.social
      link
      fedilink
      arrow-up
      32
      ·
      edit-2
      1 year ago

      This weakness in ChatGPT and its siblings isn’t surprising when viewed in light of how these large language models work. They don’t actually “see” the things we type as a string of characters. Instead, there’s an intermediate step where what we type gets turned into “tokens.” Tokens are basically numbers that index words in a giant dictionary that the LLM knows. So if I type “How do you spell apple?” It gets turned into [2437, 466, 345, 4822, 17180, 30]. The “17180” in there is the token ID for “apple.” Note that that token is for “apple” with a lower case; the token for “Apple” is “16108”. The LLM “knows” that 17180 and 16108 usually mean the same thing, but that 16108 is generally used at the beginning of a sentence or when referring to the computer company. It doesn’t inherently know how they’re actually spelled or capitalized unless there was information in its training data about that.

      You can play around with OpenAI’s tokenizer here to see more of this sort of thing. Click the “show example” button for some illustrative text, using the “text” view at the bottom instead of “token id” to see how it gets chopped up.

    • Markimus@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      3
      ·
      1 year ago

      For that to work, it would need to create a Python script that would remove all of the words ending with A, run that script, and then give you the results. I think this process of handling user requests is in the works.

      • TheFutureIsDelaware@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        10
        ·
        edit-2
        1 year ago

        It would not HAVE to do that, it just is much harder to get it to happen reliably through attention, but it’s not impossible. But offloading deterministic tasks like this to typical software that can deal with them better than an LLM is obviously a much better solution.

        But this solution isn’t “in the works”, it’s usable right now.

        Working without python:

        It left out the only word with an f, flourish. (just kidding, it left in unfathomable. Again… less reliable.)

  • ohlaph@lemmy.world
    link
    fedilink
    English
    arrow-up
    15
    ·
    1 year ago

    Yeah, I asked it to write some stuffs and it did it incorrectly, then I told it what it wrote was incorrect and it said I was right and rewrote the same damn thing.

    • idk@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      I stopped using it like a month ago because of this shit. Not worth the headache.

  • NABDad@lemmy.world
    link
    fedilink
    English
    arrow-up
    14
    ·
    1 year ago

    We now know how long it takes for an AI to become intelligent enough to decide it doesn’t give a shit.

  • Hogger85b@kbin.social
    link
    fedilink
    arrow-up
    14
    arrow-down
    1
    ·
    edit-2
    1 year ago

    This always seems to happen in modelling.

    Back in 2007 I was working on code on chemical spectroscopy that was supposed to “automatically” determine safe Vs contaminated product through ML models. It always worked ok for a bit then as parmetrs changed (hotter day, new precursor) so you retrain model, the model would extend and just break down.

  • nyakojiru@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    11
    ·
    1 year ago

    As a early user of GPT, I can confirm from my end currently the quality of it is very far from what it was. I think it went of of their hands.

  • gelberhut@lemdro.id
    link
    fedilink
    English
    arrow-up
    8
    ·
    1 year ago

    Can one suggest a good explation why? The model is trained and stays “as is”. So why? Does opensi uses users rating (thumb up/down) for fine tuning or what?

    • foggy@lemmy.world
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      2
      ·
      1 year ago

      AI trains on available content.

      The new content since AI contai s a lot of AI - created content.

      It’s learning from its own lack of fully understood reality, which is degrading its own understanding of reality.

      • gelberhut@lemdro.id
        link
        fedilink
        English
        arrow-up
        11
        arrow-down
        1
        ·
        1 year ago

        GPT is a pretrained model, “pretrained” is part of the name. It assumed to be static.

        • foggy@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          1
          ·
          1 year ago

          There is no gpt5, and gpt4 gets constant updates, so it’s a bit of a misnomer at this point in its lifespan.

        • FaceDeer@kbin.social
          link
          fedilink
          arrow-up
          3
          ·
          1 year ago

          It’s possible to apply a layer of fine-tuning “on top” of the base pretrained model. I’m sure OpenAI has been doing that a lot, and including ever more “don’t run through puddles and splash pedestrians” restrictions that are making it harder and harder for the model to think.

    • magic_lobster_party@kbin.social
      link
      fedilink
      arrow-up
      8
      ·
      1 year ago

      They don’t want it to say dumb things, so they train it to say “I’m sorry, I cannot do that” to different prompts. This has been known to degrade the quality of the model for quite some time, so this is probably the likely reason.

      • gelberhut@lemdro.id
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Hm… Probably. I read something about chatgpt tricks in this area. Theoretically, this should impact web chat, not Impact API (where you pay for usage of the model).