Install OpenRAG with the automatic installer script
To quickly install and test OpenRAG's core features, try the quickstart.
The installer script installs uv, Docker or Podman, Docker Compose, and OpenRAG.
Then, it installs and runs OpenRAG with uvx.
When you install OpenRAG with the installer script, you will use the Terminal User Interface (TUI) to configure and manage your OpenRAG deployment.
This installation method is best for testing OpenRAG by running it outside of a Python project. For other installation methods, see Select an installation method.
Prerequisites
- For Microsoft Windows, you must use the Windows Subsystem for Linux (WSL). See Install OpenRAG on Windows before proceeding.
- Install Python version 3.13 or later.
-
Gather the credentials and connection details for your preferred model providers. You must have access to at least one language model and one embedding model. If a provider offers both types, you can use the same provider for both models. If a provider offers only one type, you must select two providers.
-
OpenAI: Create an OpenAI API key.
-
Anthropic: Create an Anthropic API key. Anthropic provides language models only; you must select an additional provider for embeddings.
-
IBM watsonx.ai: Get your watsonx.ai API endpoint, IBM project ID, and IBM API key from your watsonx deployment.
-
Ollama: Deploy an Ollama instance and models locally, in the cloud, or on a remote server. Then, get your Ollama server's base URL and the names of the models that you want to use.
infoOpenRAG isn't guaranteed to be compatible with all models that are available through Ollama. For example, some models might produce unexpected results, such as JSON-formatted output instead of natural language responses, and some models aren't appropriate for the types of tasks that OpenRAG performs, such as those that generate media.
The OpenRAG team recommends the following models when using Ollama as your model provider:
-
Language models:
gpt-oss:20bormistral-nemo:12b.If you choose
gpt-oss:20b, consider using Ollama Cloud or running Ollama on a remote machine because this model requires at least 16GB of RAM. -
Embedding models:
nomic-embed-text:latest,mxbai-embed-large:latest, orembeddinggemma:latest.
You can experiment with other models, but if you encounter issues that you are unable to resolve through other RAG best practices (like context filters and prompt engineering), try switching to one of the recommended models. You can submit an OpenRAG GitHub issue to request support for specific models.
-
-
-
Optional: Install GPU support with an NVIDIA GPU, CUDA support, and compatible NVIDIA drivers on the OpenRAG host machine. If you don't have GPU capabilities, OpenRAG provides an alternate CPU-only deployment.
Run the installer script
-
Create a directory to store your OpenRAG configuration files and data, and then change to that directory:
mkdir openrag-workspace
cd openrag-workspace -
Get and run the installer script:
curl -fsSL https://docs.openr.ag/files/run_openrag_with_prereqs.sh | bashThe installer script installs OpenRAG with
uvxin the directory where you run the script. -
Wait while the installer script prepares your environment and installs OpenRAG. You might be prompted to install certain dependencies if they aren't already present in your environment.
The entire process can take a few minutes. Once the environment is ready, the OpenRAG TUI starts.

Because the installer script uses
uvx, it creates a cached, ephemeral environment in your localuvcache, and your OpenRAG configuration files and data are stored separately from theuvcache. Clearing the cache doesn't delete your entire OpenRAG installation, only the temporary TUI environment. After clearing the cache, runuvx openragto access the TUI and continue with your preserved configuration and data.
If you encounter errors during installation, see Troubleshoot OpenRAG.
Set up OpenRAG with the TUI
When you install OpenRAG with the installer script, you manage the OpenRAG services with the TUI. The TUI guides you through the initial configuration process before you start the OpenRAG services.
Your configuration values are stored in an OpenRAG .env file that is created automatically at ~/.openrag/tui.
If OpenRAG detects an existing .env file in this directory, then the TUI can populate those values automatically during setup and onboarding.
Container definitions are stored in the docker-compose files in the same directory as the OpenRAG .env file.
You can use either Basic Setup or Advanced Setup to configure OpenRAG. This choice determines how OpenRAG authenticates with your deployment's OpenSearch instance, and it controls user access to documents stored in your OpenSearch knowledge base:
-
No-auth mode: If you select Basic Setup in the TUI, or your OpenRAG
.envfile doesn't include OAuth credentials, then the OpenRAG OpenSearch instance runs in no-auth mode.This mode uses one anonymous JWT token for OpenSearch authentication. There is no differentiation between users; all users that access your OpenRAG instance can access all documents uploaded to your knowledge base.
-
OAuth mode: If you select Advanced Setup in the TUI, or your OpenRAG
.envfile includes OAuth credentials, then the OpenRAG OpenSearch instance runs in OAuth mode.This mode uses a unique JWT token for each OpenRAG user, and each document is tagged with user ownership. Documents are filtered by user owner; users see only the documents that they uploaded or have access to through their cloud storage accounts.
To enable OAuth mode after initial setup, see Ingest files with OAuth connectors.
You must use Advanced Setup if you want to use OAuth connectors to upload documents from cloud storage.
If OpenRAG detects OAuth credentials during setup, it recommends Advanced Setup in the TUI.
- Basic setup
- Advanced setup
-
In the TUI, select Basic Setup.
-
Enter administrator passwords for the OpenRAG OpenSearch and Langflow services.
The OpenSearch password is required, and a secure password is automatically generated if you don't provide one manually.
The Langflow password is recommended but optional. If the Langflow password is empty, the Langflow server starts without authentication enabled. For more information, see Langflow settings.
You can click Generate Password to create a Langflow password and username automatically.
-
Optional: Under API Keys, enter your model provider credentials, or leave these fields empty if you want to configure model provider credentials during the application onboarding process.
There is no material difference between providing these values now or during the application onboarding process. If you provide a credential now, it can be populated automatically during the application onboarding process if you enable the Use environment API key option.
OpenRAG's core functionality requires access to language and embedding models. By default, OpenRAG uses OpenAI models. If you aren't sure which models or providers to use, you must provide an OpenAI API key to use OpenRAG's default model configuration.
-
Optional: Under Others, edit the knowledge base paths if you don't want to use the default paths:
- Documents Paths: One or more paths to directories are where OpenRAG looks for documents to ingest.
- OpenSearch Data Path: Specify the path where you want OpenRAG to create your OpenSearch index.
-
Click Save Configuration.
Your passwords and API keys, if provided, are stored in the OpenRAG
.envfile at~/.openrag/tui. If you modified any credentials that were pulled from an existing.envfile, those values are updated in the.envfile. -
Click Start OpenRAG to start the OpenRAG services.
This process can take some time while OpenRAG pulls and runs the container images. If all services start successfully, the TUI prints a confirmation message:
Services started successfully
Command completed successfully -
Click Close, and then click Launch OpenRAG or navigate to
localhost:3000in your browser. -
Continue with the application onboarding process.
-
In the TUI, select Advanced Setup.
-
Enter administrator passwords for the OpenRAG OpenSearch and Langflow services.
The OpenSearch password is required, and a secure password is automatically generated if you don't provide one manually.
The Langflow password is recommended but optional. If the Langflow password is empty, the Langflow server starts without authentication enabled. For more information, see Langflow settings.
You can click Generate Password to create a Langflow password and username automatically.
-
Optional: Under API Keys, enter your model provider credentials, or leave the OpenAI, Anthropic, Ollama, and IBM watsonx.ai fields empty if you want to configure model provider credentials during the application onboarding process.
There is no material difference between providing these values now or during the application onboarding process. If you provide a credential now, it can be populated automatically during the application onboarding process if you enable the Use environment API key option.
OpenRAG's core functionality requires access to language and embedding models. By default, OpenRAG uses OpenAI models. If you aren't sure which models or providers to use, you must provide an OpenAI API key to use OpenRAG's default model configuration.
-
Recommended: To upload documents from external storage, such as Google Drive, add the required OAuth credentials for the connectors that you want to use under API Keys. These settings can be populated automatically if OpenRAG detects these credentials in an OpenRAG
.envfile at~/.openrag/tui.- Google: Provide your Google OAuth Client ID and Google OAuth Client Secret. You can generate these in the Google Cloud Console. For more information, see the Google OAuth client documentation.
- Microsoft: For the Microsoft OAuth Client ID and Microsoft OAuth Client Secret, provide Azure application registration credentials for SharePoint and OneDrive. For more information, see the Microsoft Graph OAuth client documentation.
- Amazon: Provide your AWS Access Key ID and AWS Secret Access Key with access to your S3 instance. For more information, see the AWS documentation on Configuring access to AWS applications.
You can manage OAuth credentials later, but it is recommended to configure them during initial set up.
-
Register the redirect URIs shown in the TUI in your OAuth provider. These are the URLs your OAuth provider will use to redirect users back to OpenRAG after they sign in.
-
Optional: Under Others, you can edit the following settings if needed:
- Documents Paths: Use the default path or provide one or more paths to directories are where OpenRAG looks for documents to ingest in to your knowledge base.
- OpenSearch Data Path: Specify the path where you want OpenRAG to create your OpenSearch index.
- Langflow Public URL (
LANGFLOW_PUBLIC_URL) : Sets the base address to access the Langflow web interface. This is where users interact with flows in a browser. - Webhook Base URL (
WEBHOOK_BASE_URL): If applicable, set the base address for your OAuth connector endpoints. If set, the OAuth connector webhook URLs are constructed asWEBHOOK_BASE_URL/connectors/${provider}/webhook.
-
Click Save Configuration.
Your passwords, API key, and OAuth credentials, if provided, are stored in the OpenRAG
.envfile at~/.openrag/tui. If you modified any credentials that were pulled from an existing.envfile, those values are updated in the.envfile. -
Click Start OpenRAG to start the OpenRAG services.
This process can take some time while OpenRAG pulls and runs the container images. If all services start successfully, the TUI prints a confirmation message:
Services started successfully
Command completed successfully -
Click Close, and then click Launch OpenRAG or navigate to
localhost:3000in your browser. -
If you enabled OAuth connectors, you must sign in to your OAuth provider before being redirected to your OpenRAG instance.
-
Continue with the application onboarding process.
Complete the application onboarding process
The first time you start the OpenRAG application, you must complete the application onboarding process to select language and embedding models that are essential for OpenRAG features like the Chat.
Some of these variables, such as the embedding models, can be changed seamlessly after onboarding. Others are immutable and require you to destroy and recreate the OpenRAG containers. For more information, see the OpenRAG environment variables reference.
You can use different providers for your language model and embedding model, such as Anthropic for the language model and OpenAI for the embedding model. Additionally, you can set multiple embedding models.
You only need to complete onboarding for your preferred providers.
- Anthropic
- IBM watsonx.ai
- Ollama
- OpenAI (default)
Anthropic doesn't provide embedding models. If you select Anthropic for your language model, you must select a different provider for the embedding model.
-
Enter your Anthropic API key, or enable Use environment API key to pull the key from your OpenRAG
.envfile. -
Under Advanced settings, select the language model that you want to use.
-
Click Complete.
-
Select a provider for embeddings, provide the required information, and then select the embedding model you want to use. For information about another provider's credentials and settings, see the instructions for that provider.
-
Click Complete.
After you configure the embedding model, OpenRAG uses your credentials and models to ingest some initial documents. This tests the connection, and it allows you to ask OpenRAG about itself in the Chat. If there is a problem with the model configuration, an error occurs and you are redirected back to the application onboarding screen. Verify that the credential is valid and has access to the selected model, and then click Complete to retry ingestion.
-
Continue through the overview slides for a brief introduction to OpenRAG, or click Skip overview. The overview demonstrates some basic functionality that is covered in the quickstart and in other parts of the OpenRAG documentation.
-
For watsonx.ai API Endpoint, select the base URL for your watsonx.ai model deployment.
-
Enter your watsonx.ai deployment's project ID and API key.
You can enable Use environment API key to pull the key from your OpenRAG
.envfile. -
Under Advanced settings, select the language model that you want to use.
-
Click Complete.
-
Select a provider for embeddings, provide the required information, and then select the embedding model you want to use. For information about another provider's credentials and settings, see the instructions for that provider.
-
Click Complete.
After you configure the embedding model, OpenRAG uses your credentials and models to ingest some initial documents. This tests the connection, and it allows you to ask OpenRAG about itself in the Chat. If there is a problem with the model configuration, an error occurs and you are redirected back to the application onboarding screen. Verify that the credentials are valid and have access to the selected model, and then click Complete to retry ingestion.
-
Continue through the overview slides for a brief introduction to OpenRAG, or click Skip overview. The overview demonstrates some basic functionality that is covered in the quickstart and in other parts of the OpenRAG documentation.
Using Ollama as your language and embedding model provider offers greater flexibility and configuration options for hosting models. However, it requires additional setup because Ollama isn't included with OpenRAG. You must deploy Ollama separately if you want to use Ollama as a model provider.
OpenRAG isn't guaranteed to be compatible with all models that are available through Ollama. For example, some models might produce unexpected results, such as JSON-formatted output instead of natural language responses, and some models aren't appropriate for the types of tasks that OpenRAG performs, such as those that generate media.
The OpenRAG team recommends the following models when using Ollama as your model provider:
-
Language models:
gpt-oss:20bormistral-nemo:12b.If you choose
gpt-oss:20b, consider using Ollama Cloud or running Ollama on a remote machine because this model requires at least 16GB of RAM. -
Embedding models:
nomic-embed-text:latest,mxbai-embed-large:latest, orembeddinggemma:latest.
You can experiment with other models, but if you encounter issues that you are unable to resolve through other RAG best practices (like context filters and prompt engineering), try switching to one of the recommended models. You can submit an OpenRAG GitHub issue to request support for specific models.
-
Install Ollama locally or on a remote server, or run models in Ollama Cloud.
If you are running a remote server, it must be accessible from your OpenRAG deployment.
-
In the OpenRAG onboarding dialog, enter your Ollama server's base URL:
- Local Ollama server: Enter your Ollama server's base URL and port. The default Ollama server address is
http://localhost:11434. - Ollama Cloud: Because Ollama Cloud models run at the same address as a local Ollama server and automatically offload to Ollama's cloud service, you can use the same base URL and port as you would for a local Ollama server. The default address is
http://localhost:11434. - Remote server: Enter your remote Ollama server's base URL and port, such as
http://your-remote-server:11434.
- Local Ollama server: Enter your Ollama server's base URL and port. The default Ollama server address is
-
Select the language model that your Ollama server is running.
If your server isn't running any language models, you must either deploy a language model on your Ollama server, or use another provider for the language model.
Language model and embedding model selections are independent. You can use the same or different servers for each model.
To use different providers for each model, you must configure both providers, and select the relevant model for each provider.
-
Click Complete.
-
Select a provider for embeddings, provide the required information, and then select the embedding model you want to use. For information about another provider's credentials and settings, see the instructions for that provider.
-
Click Complete.
After you configure the embedding model, OpenRAG uses your credentials and models to ingest some initial documents. This tests the connection, and it allows you to ask OpenRAG about itself in the Chat. If there is a problem with the model configuration, an error occurs and you are redirected back to the application onboarding screen. Verify that the server address is valid, and that the selected model is running on the server. Then, click Complete to retry ingestion.
-
Continue through the overview slides for a brief introduction to OpenRAG, or click Skip overview. The overview demonstrates some basic functionality that is covered in the quickstart and in other parts of the OpenRAG documentation.
-
Enter your OpenAI API key, or enable Use environment API key to pull the key from your OpenRAG
.envfile. -
Under Advanced settings, select the language model that you want to use.
-
Click Complete.
-
Select a provider for embeddings, provide the required information, and then select the embedding model you want to use. For information about another provider's credentials and settings, see the instructions for that provider.
-
Click Complete.
After you configure the embedding model, OpenRAG uses your credentials and models to ingest some initial documents. This tests the connection, and it allows you to ask OpenRAG about itself in the Chat. If there is a problem with the model configuration, an error occurs and you are redirected back to the application onboarding screen. Verify that the credential is valid and has access to the selected model, and then click Complete to retry ingestion.
-
Continue through the overview slides for a brief introduction to OpenRAG, or click Skip overview. The overview demonstrates some basic functionality that is covered in the quickstart and in other parts of the OpenRAG documentation.
Next steps
- Try some of OpenRAG's core features in the quickstart.
- Learn how to manage OpenRAG services.
- Upload documents, and then use the Chat to explore your data.