Skip to content

Commit 89484ef

Browse files
authored
docs: update distributed_inferencing.md
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
1 parent a9757fb commit 89484ef

File tree

1 file changed

+17
-6
lines changed

1 file changed

+17
-6
lines changed

docs/content/docs/features/distributed_inferencing.md

Lines changed: 17 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -122,12 +122,6 @@ The server logs should indicate that new workers are being discovered.
122122

123123
![output](https://github.com/mudler/LocalAI/assets/2420543/8ca277cf-c208-4562-8929-808b2324b584)
124124

125-
## Notes
126-
127-
- If running in p2p mode with container images, make sure you start the container with `--net host` or `network_mode: host` in the docker-compose file.
128-
- Only a single model is supported currently.
129-
- Ensure the server detects new workers before starting inference. Currently, additional workers cannot be added once inference has begun.
130-
- For more details on the implementation, refer to [LocalAI pull request #2343](https://github.com/mudler/LocalAI/pull/2343)
131125

132126
## Environment Variables
133127

@@ -138,3 +132,20 @@ There are options that can be tweaked or parameters that can be set using enviro
138132
| **LOCALAI_P2P_DISABLE_DHT** | Set to "true" to disable DHT and enable p2p layer to be local only (mDNS) |
139133
| **LOCALAI_P2P_DISABLE_LIMITS** | Set to "true" to disable connection limits and resources management |
140134
| **LOCALAI_P2P_TOKEN** | Set the token for the p2p network |
135+
136+
## Architecture
137+
138+
LocalAI uses https://github.com/libp2p/go-libp2p under the hood, the same project powering IPFS. Differently from other frameworks, LocalAI uses peer2peer without a single master server, but rather it uses sub/gossip and ledger functionalities to achieve consensus across different peers.
139+
140+
[EdgeVPN](https://github.com/mudler/edgevpn) is used as a library to establish the network and expose the ledger functionality under a shared token to ease out automatic discovery and have separated, private peer2peer networks.
141+
142+
The weights are split proportional to the memory when running into worker mode, when in federation mode each request is split to every node which have to load the model fully.
143+
144+
## Notes
145+
146+
- If running in p2p mode with container images, make sure you start the container with `--net host` or `network_mode: host` in the docker-compose file.
147+
- Only a single model is supported currently.
148+
- Ensure the server detects new workers before starting inference. Currently, additional workers cannot be added once inference has begun.
149+
- For more details on the implementation, refer to [LocalAI pull request #2343](https://github.com/mudler/LocalAI/pull/2343)
150+
151+

0 commit comments

Comments
 (0)