• mavu@discuss.tchncs.de
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    Well, good luck with that. Software development is a shit show already anyway. You can find me in my Gardening business in 2027.

    • Rooster326@programming.dev
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      1 month ago

      Good Luck. When the economy finally bottoms out the first budget to go is always the gardening budget.

      You can find me in my plumbing business in 2028.

      I already deal with shit daily so it’s what we in biz call a horizontal promotion.

  • hoshikarakitaridia@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    Literally not how any of this works. You don’t let AI check your work, at best you use AI and check it’s work, and at worst you have to do everything by hand anyway.

    • UnderpantsWeevil@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      1 month ago

      You don’t let AI check your work

      From a game dev perspective, user Q&A is often annoying and repetitive labor. Endlessly criss-crossing terran hitting different buttons to make sure you don’t snag a corner or click objects in a sequence that triggers a state freeze. Hooking a PS controller to Roomba logic and having a digital tool rapidly rerun routes and explore button combos over and over, looking for failed states, is significantly better for you than hoping an overworked team of dummy players can recreate the failed state by tripping into it manually.

        • Grimy@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          1 month ago

          If it does the job better, who the fuck cares. No one actually cares about how you feel about the tech. Cry me a river.

          • _stranger_@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            1 month ago

            The problem is that if it doesn’t do a better job, no one left in charge will even know enough to give a shit, so quality will go down.

        • UnderpantsWeevil@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 month ago

          I mean, as a branding exercise, every form of sophisticated automation is getting the “AI” label.

          Past that, advanced pathing algorithms are what Q&A systems need to validate all possible actions within a space. That’s the bread-and-butter of AI. Its also generally how you’d describe simulated end-users on a test system.

          • subignition@fedia.io
            link
            fedilink
            arrow-up
            0
            ·
            1 month ago

            I mean, as a branding exercise, every form of sophisticated automation is getting the “AI” label. The article is specifically talking about generative AI. I think we need to find new terminology to describe the kind of automation that was colloquially referred to as AI before chatgpt et al. came into existence.

            The important distinction, I think, is that these things are still purpose-built and (mostly) explainable. When you have a bunch of nails, you design a hammer. An “AI bot” QA tester the way Booty describes in the article isn’t going to be an advanced algorithm that carries out specific tests. That exists already and has for years. He’s asking for something that will figure out specific tests that are worth doing when given a vague or nonexistent test plan, most likely. You need a human, or an actual AGI, for something on that level, not generative AI. And explicitly with generative AI, as pertains to Square Enix’s initiative in the article, there are the typical huge risks of verifiability and hallucination. However unpleasant you may think a QA worker’s job is now, I guarantee you it will be even more unpleasant when the job consists of fact-checking AI bug reports all day instead of actually doing the testing.

    • ThePowerOfGeek@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      1 month ago

      “Well it works for unit testing, so just extend that out to all testing! Problem solved!” -Senior Management, probably

      Who am I kidding. They have no idea what unit testing is.

    • III@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      That’s why they want 70% QA from AI. Because right now their games are only 10% QA tested.

  • tal@lemmy.today
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    1 month ago

    Hmm. While I don’t know what their QA workflow is, my own experience is that working with QA people to design a QA procedure for a given feature tends to require familiarity with the feature in the context of real-world knowledge and possible problems, and that human-validating a feature isn’t usually something done at massive scale, where you’d get a lot of benefit from heavy automation.

    It’s possible that one might be able to use LLMs to help write test code — reliability and security considerations there are normally less-critical than in front-line code. Worst case is getting a false positive, and if you can get more test cases covered, I imagine that might pay off.

    Square does an MMO, among their other stuff. If they can train a model to produce AI-driven characters that act sufficiently like human players, where they can theoretically log training data from human players, that might be sufficient to populate an MMO “experimental” deployment so that they can see if anything breaks prior to moving code to production.

    “Because I would love to be able to start up 10,000 instances of a game in the cloud, so there’s 10,000 copies of the game running, deploy an AI bot to spend all night testing that game, then in the morning we get a report. Because that would be transformational.”

    I think that the problem is that you’re likely going to need more-advanced AI than an LLM, if you want them to just explore and try out new features.

    One former Respawn employee who worked in a senior QA role told Business Insider that he believes one of the reasons he was among 100 colleagues laid off this past spring is because AI was reviewing and summarising feedback from play testers, a job he usually did.

    We can do a reasonable job of summarizing human language with LLMs today. I think that that might be a viable application.

    • [deleted]@piefed.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      1 month ago

      Worst case is getting a false positive, and if you can get more test cases covered, I imagine that might pay off.

      False positives during testing are a huge time sink. QA has to replicate and explain away each false report and the faster AI ‘completes’ tasks the faster the flood of false reports come in.

      There is plenty of non-AI automation that can be used intentionally to do tedious repetitive tasks already where they only increase work if they aren’t set up right.