Skip to content

Commit a790803

Browse files
authored
Falcon-180B // Any 🤗 model via --model meta/llama
1 parent 3e8416a commit a790803

File tree

1 file changed

+23
-5
lines changed

1 file changed

+23
-5
lines changed

‎README.md

Lines changed: 23 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -151,6 +151,20 @@ print(interpreter.system_message)
151151

152152
### Change the Model
153153

154+
For `gpt-3.5-turbo`, use fast mode:
155+
156+
```shell
157+
interpreter --fast
158+
```
159+
160+
In Python, you will need to set the model manually:
161+
162+
```python
163+
interpreter.model = "gpt-3.5-turbo"
164+
```
165+
166+
### Running Open Interpreter locally
167+
154168
ⓘ **Issues running locally?** Read our new [GPU setup guide](/docs/GPU.md) and [Windows setup guide](/docs/WINDOWS.md).
155169

156170
You can run `interpreter` in local mode from the command line to use `Code Llama`:
@@ -159,16 +173,20 @@ You can run `interpreter` in local mode from the command line to use `Code Llama
159173
interpreter --local
160174
```
161175

162-
For `gpt-3.5-turbo`, use fast mode:
176+
Or run any HugginFace model **locally** by using its repo ID (e.g. "tiiuae/falcon-180B"):
163177

164178
```shell
165-
interpreter --fast
179+
interpreter --model tiiuae/falcon-180B
166180
```
167181

168-
In Python, you will need to set the model manually:
182+
#### Local model params
169183

170-
```python
171-
interpreter.model = "gpt-3.5-turbo"
184+
You can easily modify the `max_tokens` and `context_window` (in tokens) of locally running models.
185+
186+
Smaller context windows will use less RAM, so we recommend trying a shorter window if GPU is failing.
187+
188+
```shell
189+
interpreter --max_tokens 2000 --context_window 16000
172190
```
173191

174192
### Azure Support

0 commit comments

Comments
 (0)