You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: source/en/user_guide/internnav/quick_start/installation.md
+5-8Lines changed: 5 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -13,10 +13,7 @@
13
13
14
14
# Installation Guide
15
15
16
-
😄 Don’t worry — both [Quick Installation](#quick-installation) and [Dataset Preparation](#dataset-preparation) are beginner-friendly.
17
-
```
18
-
Detailed technical report will be released in about two weeks.
19
-
```
16
+
😄 Don’t worry — both [Quick Installation](#quick-installation) and [Dataset Preparation](#dataset-preparation) are beginner-friendly. Technical report of InternVLA-N1 is available at [page](https://internrobotics.github.io/internvla-n1.github.io/).
20
17
21
18
## Prerequisites
22
19
@@ -256,7 +253,7 @@ If you need to train or evaluate models on [Habitat](#optional-habitat-environme
256
253
257
254
#### Prerequisite
258
255
- Python 3.9
259
-
- Pytorch 2.1.2
256
+
- Pytorch 2.6.0
260
257
- CUDA 12.4
261
258
- GPU: NVIDIA A100 or higher (optional for VLA training)
262
259
@@ -353,7 +350,7 @@ InternNav/
353
350
│ │ ├── model-00001-of-00004.safetensors
354
351
│ │ ├── config.json
355
352
│ │ └── ...
356
-
│ depth_anything_v2_vits.pth
353
+
│ ├── depth_anything_v2_vits.pth
357
354
│ ├── r2r
358
355
│ │ ├── fine_tuned
359
356
│ │ └── zero_shot
@@ -365,7 +362,7 @@ InternNav/
365
362
Currently the gradio demo is only available in **habitat** environment. Replace the 'model_path' variable in 'vln_ray_backend.py' with the path of InternVLA-N1 checkpoint.
Find the IP address of the node allocated by Slurm. Then change the BACKEND_URL in the gradio client (navigation_ui.py) to the server's IP address. Start the gradio.
Note that it's better to run the Gradio client on a machine with a graphical user interface (GUI) but ensure there is proper network connectivity between the client and the server. Download the gradio scene assets from [huggingface](https://huggingface.co/datasets/InternRobotics/Scene-N1) and extract them into the `scene_assets` directory of the client. Then open a browser and enter the Gradio address (such as http://0.0.0.0:5700). We can see the interface as shown below.
Click the 'Start Navigation Simulation' button to send a VLN request to the backend. The backend will submit a task to ray server and simulate the VLN task with InternVLA-N1 models. Wait about 1 minutes, the VLN task will be finished and return a result video. We can see the result video in the gradio like this.
374
+
Click the 'Start Navigation Simulation' button to send a VLN request to the backend. The backend will submit a task to ray server and simulate the VLN task with InternVLA-N1 models. Wait about 2 minutes, the VLN task will be finished and return a result video. We can see the result video in the gradio like this.
The evaluation results will be saved in the `eval_results.log` file in the output_dir of the config file. The whole evaluation process takes about 10 hours at RTX-4090 graphics platform.
0 commit comments