---
title: OpenAI
---
To use Open Interpreter with a model from OpenAI, simply run:
```bash Terminal
interpreter
```
```python Python
from interpreter import interpreter
interpreter.chat()
```
This will default to `gpt-4`, which is the most capable publicly available model for code interpretation (Open Interpreter was designed to be used with `gpt-4`).
Trouble accessing `gpt-4`? Read our [gpt-4 setup
article](/language-model-setup/hosted-models/gpt-4-setup).
To run a specific model from OpenAI, set the `model` flag:
```bash Terminal
interpreter --model gpt-3.5-turbo
```
```python Python
from interpreter import interpreter
interpreter.llm.model = "gpt-3.5-turbo"
interpreter.chat()
```
# Supported Models
We support any model on [OpenAI's models page:](https://platform.openai.com/docs/models/)
```bash Terminal
interpreter --model gpt-4
interpreter --model gpt-4-32k
interpreter --model gpt-3.5-turbo
interpreter --model gpt-3.5-turbo-16k
```
```python Python
interpreter.llm.model = "gpt-4"
interpreter.llm.model = "gpt-4-32k"
interpreter.llm.model = "gpt-3.5-turbo"
interpreter.llm.model = "gpt-3.5-turbo-16k"
```
# Required Environment Variables
Set the following environment variables [(click here to learn how)](https://chat.openai.com/share/1062cdd8-62a1-4aa8-8ec9-eca45645971a) to use these models.
| Environment Variable | Description | Where to Find |
| -------------------- | ---------------------------------------------------- | ------------------------------------------------------------------- |
| `OPENAI_API_KEY` | The API key for authenticating to OpenAI's services. | [OpenAI Account Page](https://platform.openai.com/account/api-keys) |