The Nvidia NV1 was released in 1995, it was the first GPU with 3D capabilities for PC… form there we know how things went by.

Now it’s 2023, so let’s make some “retro futuristic” prediction… what would you think about a AI board, open source driver, open API as Vulkan which you can buy to power the AI for your videogames? It would make sense to you? Which price range it should be?

What’s supposed to do for your games… well, that’s depend on videogames. The quickiest example I can think of is having endless discussion with your NPC in your average, single player, Fantasy RPG.

For example, the videogame load your 4~5 companions with the psychology/behaviors: they are fixated with the main quest goal (like you talk with fanatic people, this to make sure the game the main quest is as much stable as possible) but you can “break them” by making attempt to reveal some truths (for example, breaking the fourth wall), and if you go for this path, the game warns that you’re probably going to lock out the main quest (like in Morrowind when you kill essential NPC)

  • RightHandOfIkaros@lemmy.world
    link
    fedilink
    arrow-up
    6
    ·
    1 year ago

    This just sounds like putting a second CPU on a PCIe board. I can’t see this being a benefit for games because developers would never go through the pain of programming AI with advanced enough behaviours to even need a secondary CPU.

    • thepianistfroggollum@lemmynsfw.com
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      1 year ago

      Programming AI is actually super easy, unless you decided to create your own foundation model. Even then, you would have data scientists building it, not devs.

      Plenty of FMs and LLMs already exist that would be up to the task.

      • RightHandOfIkaros@lemmy.world
        link
        fedilink
        arrow-up
        1
        arrow-down
        1
        ·
        1 year ago

        Programming AI with behaviour complex enough to need a second CPU would be hard. Syncing its output with the primary CPU could be a problem.

        LLMs would not be useful for anything except maybe generating new dialogue, but it would need a lot of restraints to prevent the end user from breaking it. For the purposes of dialogue and story telling, most developers would opt to just pre-program dialogue like they always have.

        Again, this sounds like a useless PC part that pretty much no game developer would ever take advantage of.

        • thepianistfroggollum@lemmynsfw.com
          link
          fedilink
          arrow-up
          1
          arrow-down
          1
          ·
          1 year ago

          You don’t need an LLM for this. You just need a FM that you fine tune, and you’d be surprised at how little computing power is actually required.

          For our uses (which are similar to what OP wants), it takes longer for us to do an OCR scan on the documents our AI works with than for Sagemaker to do it’s thing on a rather small instance.

          And, devs would just be implementing API calls, so it wouldn’t be a big deal to make the switch.

    • howrar@lemmy.ca
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      1 year ago

      Why wouldn’t they? It’s a lot easier to write out intricate backstories for each character/location independently than it is to build decision trees for every possible combination of decisions that the player makes. That’s basically what current LLMs allow for.

  • Omega_Jimes@lemmy.ca
    link
    fedilink
    arrow-up
    2
    ·
    1 year ago

    Yeah, so, dedicated hardware like that rarely ever pans out. I mean, graphics cards did, but there’s not much of a market for gaming sound cards or physX cards anymore. I imagine that the specific type of AI that will be useful for this will eventually just be improved and made efficient enough that it’ll be done by processors that already exist in your system.

  • Throwaway@lemm.ee
    link
    fedilink
    arrow-up
    4
    arrow-down
    2
    ·
    1 year ago

    Wouldn’t that just be a GPU? That’s literally what all our AIs run on. Just a ton of tiny little processors running in parallel.

  • norske@lemmynsfw.com
    link
    fedilink
    arrow-up
    2
    ·
    1 year ago

    If the board provided enough benefit to outweigh the cost? Sure I might be talked into it.

    Reminiscent of PhysX boards when they were a thing for 30 seconds. It’s all about the return on investment for me.

  • squid@feddit.uk
    link
    fedilink
    arrow-up
    2
    ·
    1 year ago

    Game publishers won’t want direct ai in games, losses them too much control, also they can’t use the excuse of its always online so NPCs have ai powered language. With how things look as everything is becoming subscription I doubt well be getting powerful ai on a single board to put into pci-e my prediction is more Aline to we won’t have gaming PCs, GPUs will be price hiked and anyone wanting to game will be on a subscription service

  • solariplex@slrpnk.net
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    I mean, that kind of board has existed for a while. They’re usually called AI-accelerator boards, IIRC

  • BCsven@lemmy.ca
    link
    fedilink
    arrow-up
    1
    ·
    edit-2
    1 year ago

    Is the 1995 and first 3d accurate? we were using 3d CAD tools in the range of 1991-1995 before Nvidia. Edit: seems S3 and Creative Labs had some earlier CAD cards, prices too high for general PC use till voodoo cards in 95

  • BetaDoggo_@lemmy.world
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    If this were to ever become mainstream this would likely be incorporated into the GPU for cost reasons. Small machine learning acceleration boards already exist but their uses are limited because of limited memory. Google has larger ones available but they’re cloud only.

    Currently I don’t see many uses in gaming other than upscaling.