Site icon Philip Conrod

Co-Developing Unreal Engine Video Games using Local Ollama LLM Models with the Latest v1.0.66 Beta Version of the Betide NeoStack AI Unreal Plugin

Back in February I started testing the NeoStack AI Plugin by BETIDE STUDIO for AI assisted Unreal Engine game development.   The NeoStack AI plugin is still in Beta, and I am currently testing the latest beta version (v1.0.66).

In addition to using OpenRouter to run free LLM models, NeoStack AI recently added the capability to run my free local LLM models via Ollama running on my GeForce RTX-3080.

I am really excited to use their new Ollama Local LLM model capability so I can take advantage of the free processing power of my offline NVIDIA GeForce RTX-3080 Graphics card.

As a reminder, the key features of the NeoStack AI are:

I’ll continue to keep you update on both my cloud and local LLM testing with NeoStack AI for Unreal Engine.

 

Exit mobile version