I asked GPT4 to refactor a simple, working python script for my smart lights… and it completely butchered the code and apologized mid-generation.

No amount of pleading or correction would get it to function as it did just a week or two ago.

It is so over.

  • Aidan@lemm.ee
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    1 year ago

    Gpt4 is not good at writing code. I think it’s because it has a lower token limit. Ask Gpt 4 to write out detailed specs for the code you want, then copy and paste that into a Gpt-3.5 session and ask it to write the code

    And if it gets cut off, paste in the last line it output successfully and ask it to continue with the line following that one. Then just copy and paste the blocks together

  • ultranaut@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    I noticed this today working on some bash scripts. Compared to a few weeks ago it’s become noticeably dumber, but also faster.

    • twelvewings@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 year ago

      Oh great, so now I have to pay on top of my Plus membership?

      A couple of prompts already cost be $0.03. I could easily run up hundreds of dollars if I used it as much as I have in the past.

      Do you have any tips or preferring settings when using the Playground to code?

      • fossilesque@mander.xyz
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        1 year ago

        It does not add up that much unless you’re tokenizing a metric ton of content. It’s a few extra bucks and you can set a rate limit. I have rarely hit a 5 dollar mark with consistent writing. Limit it to a fiver and see how it goes. As for the settings, it is really model specific. There are a ton of guides out there, though for effective instructing. I believe Wolfram Alpha has a few. Their formatting is nice: https://medium.com/machine-minds/chatgpt-python-machine-learning-prompts-abefc544412c Playground pricing is what they use for commercial API use. It has to be affordable.

      • Mango@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        It actually works out cheaper in most cases to use the playground over actually paying for chat gpt plus

  • GutterPunch@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 year ago

    It’s moderately good at in-line commenting functions and creating full function doc comments for the specific language / documentation format you need, but its code generation abilities are still not game-changing. Getting it to generate anything longer than a few helper functions is a test of patience.

    • twelvewings@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      0
      arrow-down
      1
      ·
      edit-2
      1 year ago

      This wasn’t always this case. I had zero Python experience a month ago, and managed to make a 300 line Python script that checks credit card validation, and has a beautiful UI. This would be impossible today.

      • mycroft@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        They had to make it too dumb to draw Disney Characters… you think I’m joking, try getting it to render a disney character in SVG or javascript…

        • twelvewings@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          0
          arrow-down
          1
          ·
          1 year ago

          Considering how long I’ve been using Python, and how it looked when I started, it is to me. And here is the ancient one I was previously using:

  • vegivamp@feddit.nl
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 year ago

    Microsoft bought it. They’re not going to let their paying userbase of millions of coders evaporate…

    • KonaKoder@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Microsoft wants to own tools crucial to the mainstream of software development. They also want to own the cloud infrastructure on which those tools depend. Today, they might lose dimes on every LLM call. In five years, they’ll make a penny on orders of magnitude more calls. Microsoft has many flaws, including cloud capacity, but they aren’t short-sighted about investment. (I used to work in DevDiv and Azure Machine Learning.)

      • Madrigal@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Good and logical decisions are plausible. However, expecting Microsoft to make consistent decisions and be able to work as a single cohesive team, now that’s delusional.

  • NotAPenguin@kbin.social
    link
    fedilink
    arrow-up
    0
    arrow-down
    19
    ·
    1 year ago

    People were saying the exact same thing a few weeks ago, and have been ever since it came out basically.

    You having issues with one prompt or one conversation doesn’t mean it’s dumb now.

    Still working fine for me.