![]() |
LLM for Unity
v3.0.0
Create characters in Unity with LLMs!
|
LLM for Unity enables seamless integration of Large Language Models (LLMs) within the Unity engine.
It allows to create intelligent AI characters that players can interact with for an immersive experience.
The package includes a Retrieval-Augmented Generation (RAG) system for semantic search across your data, which can be used to enhance the character's knowledge.
The LLM backend, LlamaLib, is built on top of the awesome llama.cpp library and provided as a standalone C++/C# library.
๐งช Tested on Unity: 2021 LTS, 2022 LTS, 2023, Unity 6
๐ฆ Upcoming Releases
For business inquiries you can reach out at hello.nosp@m.@und.nosp@m.ream..nosp@m.ai.
Contact hello.nosp@m.@und.nosp@m.ream..nosp@m.ai to add your project!
Method 1: Install using the asset store
Add to My AssetsWindow > Package ManagerPackages: My Assets option from the drop-downLLM for Unity package, click Download and then ImportMethod 2: Install using the GitHub repo:
Window > Package Manager+ button and select Add package from git URLhttps://github.com/undreamai/LLMUnity.git and click Add
First you will setup the LLM for your game:
Add Component and select the LLM script.Download Model button (~GBs).Load model button (see LLM model management).Then you can setup each of your characters as follows:
Add Component and select the LLMAgent script.System Prompt.LLM field if you have more than one LLM GameObjects.In your script you can then use it as follows:
You can also specify a function to call when the model reply has been completed:
To stop the chat without waiting for its completion you can use:
That's it! Your AI character is ready to chat! โจ
For mobile apps you can use models with up to 1-2 billion parameters ("Tiny models" in the LLM model manager).
Larger models will typically not work due to the limited mobile hardware.
iOS iOS can be built with the default player settings.
Android On Android you need to specify the IL2CPP scripting backend and the ARM64 as the target architecture in the player settings.
These settings can be accessed from the Edit > Project Settings menu within the Player > Other Settings section.

Since mobile app sizes are typically small, you can download the LLM model the first time the app launches. This functionality is enabled with the Download on Build option. In your project you can wait until the model download is complete with:
You can also receive calls the download progress during the model download:
This is useful to present e.g. a progress bar. The MobileDemo demonstrates an example application for Android / iOS.
To restrict the output of the LLM you can use a grammar, read more here.
The grammar can edited directly in the Grammar field of the LLMAgent or saved in a gbnf / json schema file and loaded with the Load Grammar button (Advanced options).
For instance to receive replies in json format you can use the json.gbnf grammar.
Alternatively you can set the grammar directly in your script:
For function calling you can define similarly a grammar that allows only the function names as output, and then call the respective function.
You can look into the FunctionCalling sample for an example implementation.
The chat history of a LLMAgent is retained in the chat variable that is a list of ChatMessage objects.
The ChatMessage is a class that defines the role of the message and the content.
The list contains alternating messages with the player prompt and the AI reply.
You can modify the chat history and then set it to your LLMAgent GameObject:
To add new messages you can do:
To automatically save / load your chat history, you can specify the Save parameter of the LLMAgent to the filename (or relative path) of your choice. The chat history is saved in the persistentDataPath folder of Unity as a json object.
To manually save your chat history, you can use:
and to load the history:
The last argument of the Chat function is a boolean that specifies whether to add the message to the history (default: true):
You can use a remote server to carry out the processing and implement characters that interact with it.
Create the server
To create the server:
LLM script as described aboveRemote option of the LLM and optionally configure the server port and API keyAlternatively you can use a server binary for easier deployment:
servers folder selected and start the server by running the command copied from above.Create the characters
Create a second project with the game characters using the LLMAgent script as described above. Enable the Remote option and configure the host with the IP address (starting with "http://") and port / API key of the server.
The Embeddings function can be used to obtain the emdeddings of a phrase:
A detailed documentation on function level can be found here:
LLM for Unity implements a super-fast similarity search functionality with a Retrieval-Augmented Generation (RAG) system.
It is based on the LLM embeddings, and the Approximate Nearest Neighbors (ANN) search from the usearch library.
Semantic search works as follows.
Building the data You provide text inputs (a phrase, paragraph, document) to add to the data.
Each input is split into chunks (optional) and encoded into embeddings with a LLM.
Searching You can then search for a query text input.
The input is again encoded and the most similar text inputs or chunks in the data are retrieved.
To use semantic serch:
Add Component and select the RAG script.SimpleSearch is a simple brute-force search, whileDBSearch is a fast ANN method that should be preferred in most cases.Alternatively, you can create the RAG from code (where llm is your LLM):
In your script you can then use it as follows :unicorn::
You can also add / search text inputs for groups of data e.g. for a specific character or scene:
{.cs} rag.Save("rag.zip");
{.cs} await rag.Load("rag.zip");
{.cs} string message = "How is the weather?"; (string[] similarPhrases, float[] distances) = await rag.Search(message, 3);
string prompt = "Answer the user query based on the provided data. "; prompt += $"User query: {message} "; prompt += $"Data: "; foreach (string similarPhrase in similarPhrases) prompt += $" - {similarPhrase}";
_ = llmAgent.Chat(prompt, HandleReply, ReplyCompleted); ```
The RAG sample includes an example RAG implementation as well as an example RAG-LLM integration.
That's all :sparkles:!
LLM for Unity includes a built-in model manager for easy model handling.
The model manager allows to load or download LLMs and can be found as part of the LLM GameObject:

You can download models with the Download model button.
LLM for Unity includes different state of the art models built-in for different model sizes, quantised with the Q4_K_M method.
Alternative models can be downloaded from HuggingFace in .gguf format.
You can download a model locally and load it with the Load model button, or copy the URL in the Download model > Custom URL field to directly download it.
If a HuggingFace model does not provide a gguf file, it can be converted to gguf with this online converter.
Models added in the model manager are copied to the game during the building process.
You can omit a model by deselecting the "Build" checkbox.
To remove the model (but not delete it from disk) you can click the bin button.
The path and URL of models can be diplayed in the expanded view of the model manager with the >> button:

You can create lighter builds by selecting the Download on Build option.
The models will be downloaded the first time the game starts instead of bundled in the build.
If you have loaded a model locally you need to set its URL through the expanded view, otherwise it will be copied in the build.
โ Before using any model make sure you check their license โ
The Samples~ folder contains several examples of interaction ๐ค:

To install a sample:
Window > Package ManagerLLM for Unity Package. From the Samples Tab, click Import next to the sample you want to install.The samples can be run with the Scene.unity scene they contain inside their folder.
In the scene, select the LLM GameObject and specify the LLM of your choice (see LLM model management).
Save the scene, run and enjoy!
The license of LLM for Unity is Apache 2.0 (LICENSE.md) and uses third-party software with MIT and Apache licenses. Some models included in the asset define their own license terms, please review them before using each model. Third-party licenses can be found in the (Third Party Notices.md).