• Phen
    link
    fedilink
    arrow-up
    22
    arrow-down
    1
    ·
    2 days ago

    Reading the paper, AI did a lot better than I would expect. It showed experienced devs working on a familiar code base got 19% slower. It’s telling that they thought they had been more productive, but the result was not that bad tbh.

    I wish we had similar research for experienced devs on unfamiliar code bases, or for inexperienced devs, but those would probably be much harder to measure.

    • cley_faye@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      3 hours ago

      but the result was not that bad tbh

      Using a tool that lowers your productivity by 1/5 instead of not using it at all is “not that bad” to you? A tool that costs an awful lot to run, requires heavy security compromises, too?

      On the other parts… experienced devs on unfamiliar code base is a common task. They just get familiar with the code base in a few weeks. And they can base their familiarity on actual code, not hallucinated summaries which would be meaningless.

      Inexperienced devs are not going to improve with something else doing the job for them, and LLM without guidance is not able to do any substantial work without deviating into garbage first, then broken garbage second. We already have data on that too.

    • staircase@programming.dev
      link
      fedilink
      arrow-up
      16
      arrow-down
      1
      ·
      1 day ago

      I don’t understand your point. How is it good that the developers thought they were faster? Does that imply anything at all in LLMs’ favour? IMO that makes the situation worse because we’re not only fighting inefficiency, but delusion.

      20% slower is substantial. Imagine the effect on the economy if 20% of all output was discarded (or more accurately, spent using electricity).

      • Phen
        link
        fedilink
        arrow-up
        7
        ·
        1 day ago

        I’m not saying it’s good, I’m saying I expected it to be even worse.