Skip to content

IhateCreatingUserNames2/LLMUnityRemoteAPI

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Connect Remote API Servers in LLMUNity https://github.com/undreamai/LLMUnity

In this example im using MiniMax API for as LLM server. 

Setup process:
Host a NodeServer
Configure endpoints /completion and /template
Configure LLmCharacter.cs In Unity Inspector 
	Toggle REMOTE
	Add Host = Your NodeJS Server Host 
            port: 443 

image

In this example, im using Render.com to Host the NodeServer
In this Example App.js has 4 Endpoints, /chatbot, /chat, /completion, /template
LLMUnity uses /completion and /template
In app.js there is a RAG feature using OpenAI Embeddings and PineCone Vector Database. C
In App_no_rag.js there is no RAG Feature. Only endpoints. 

***** This example also hosts a index.html with a /chatbot endpoint, where u can talk with the LLM(test it)

*********** BE AWARE THAT THIS EXAMPLE CODE IS NOT VERY GOOD AND MAY HAVE SOME ISSUES with the /template endpoint *****

About

How to build your Own API and Use it With LLMUnity as Remote LLM SERVER

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published