Skip to main content
At the heart of Inworld Runtime are primitives: foundational components like Large Language Models (LLMs), Speech-to-Text (STT), Text-to-Speech (TTS), and Text Embedders. These are the “raw ingredients” of any AI-powered application. Think of them as a library of high-performance AI modules, designed to abstract away the complexities of working with various providers, models, and hardware - allowing you to build on a consistent, provider-agnostic foundation. These primitives are then used within Inworld’s graph system to power various nodes. Each primitive type can be configured through Unreal Engine’s Project Settings interface, allowing you to:
  • Configure multiple named instances of each primitive type (LLM, STT, TTS, Text Embedder)
  • Set up provider-specific configurations (Inworld, OpenAI, Anthropic, Google, etc.)
  • Define reusable configurations that can be referenced by nodes in your graphs
  • Easily access and modify your primitives across your project

Accessing Primitives Configuration

To configure primitives in your project:
  1. Open Edit > Project Settings from the main menu
  2. Navigate to Plugins > Inworld in the left sidebar
  3. Scroll down to the Primitives section
Here you’ll find configuration maps for each primitive type:
  • LLM Creation Config - Configure Large Language Models
  • STT Creation Config - Configure Speech-to-Text models
  • TTS Creation Config - Configure Text-to-Speech models
  • Text Embedder Creation Config - Configure text embedding models

LLM

LLMs are powerful models that can be used to understand and generate text and other content. To configure an LLM:
  1. In the LLM Creation Config map, click the + button to add a new entry
  2. Set a descriptive name for this configuration. This name will be used for selecting this configuration in your graph.
  3. Select either Local or Remote for the Compute Host
    • Remote: this means that the models will be run by cloud servers.
    If a model or service provider is not available in the dropdown, you can add additional options under the LLM section of the Inworld Runtime Settings.
    • Local: this means that the models will run locally
      • Local Compute Host: Choose CPU or GPU
      • Model: Path to the local model.
  4. Your configuration will now be available for selection in LLM powered nodes throughout your graphs.

Adding models or providers

All service providers and models listed under Chat Completion are supported. If a model or service provider is not available in the Remote LLM Creation Config dropdown, you can add additional options under the LLM section of the Inworld Runtime Settings.
  1. In the Remote LLMProviders list, click the + button to add any additional service providers you want to use.
  2. In the Remote LLMModels list, click the + button to add any additional models you want to use.

TTS

Text-to-Speech converts text into audio. To configure TTS:
  1. In the TTS Creation Config map, click the + button to add a new entry
  2. Set a descriptive name for this configuration. This name will be used for selecting this configuration in your graph.
  3. Select either Local or Remote for the Compute Host
    • Remote: this means that the models will be run by cloud servers.
      • Provider: Select INWORLD or ELEVEN_LABS
      • Configure provider-specific settings like model selection and voice parameters.
    • Local: this means that the models will run locally
      • Local Compute Host: Choose CPU or GPU
      • Model: Path to the local model.
      • Speech Synthesis Config: Adjust parameters like sample rate, temperature, and speaking rate.
  4. Your configuration will now be available for selection in TTS nodes throughout your graphs.

Using ElevenLabs

To use ElevenLabs voices, you’ll need to set up your ElevenLabs API key in Portal.
  1. Go to Portal
  2. In your workspace, go to Settings > API Keys.
  3. Enter your ElevenLabs API key in the Eleven Labs API Key field
  4. Now when you select Provider = ElevenLabs in the TTS node, your ElevenLabs API key will be used.

STT

Speech-to-Text converts audio into text. To configure STT:
  1. In the STT Creation Config map, click the + button to add a new entry
  2. Set a descriptive name for this configuration. This name will be used for selecting this configuration in your graph.
  3. Select either Local or Remote for the Compute Host
    • Remote: this means that the models will be run by cloud servers.
    • Local: this means that the models will run locally
      • Local Compute Host: Choose CPU or GPU
      • Model: Path to the local model.
  4. Your configuration will now be available for selection in STT nodes throughout your graphs.

Text Embedder

Text Embedders convert text into numerical vectors for semantic operations, and powers features like intent detection and knowledge retrieval. To configure embedders:
  1. In the Text Embedder Creation Config map, click the + button to add a new entry
  2. Set a descriptive name for this configuration. This name will be used for selecting this configuration in your graph.
  3. Select either Local or Remote for the Compute Host
    • Remote: this means that the models will be run by cloud servers.
      • Provider: Select from available providers
      • Model: Choose the embedding model
    • Local: this means that the models will run locally
      • Local Compute Host: Choose CPU or GPU
      • Model: Path to the local model.
  4. Your configuration will now be available for selection in embeddings powered nodes throughout your graphs.