Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

introduce ilab profile set doc #130

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open

Conversation

cdoern
Copy link
Contributor

@cdoern cdoern commented Sep 3, 2024

No description provided.

Signed-off-by: Charlie Doern <cdoern@redhat.com>
1. User needs to set:
a. Model Path for SDG, Eval, Serving, Training
b. GPUs for SDG, Eval, Serving, Training
c. Training Config Per-GPU (based on vram)
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
c. Training Config Per-GPU (based on vram)
c. Training Config Per-GPU (based on vRAM)

Comment on lines +6 to +7
a. Model Path for SDG, Eval, Serving, Training
b. GPUs for SDG, Eval, Serving, Training
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Serving just for vLLM, or vLLM and Llama CPP?


### Workflow

The user will run `ilab profile set`
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Will this be required before running other ilab commands? If not, what is the path for a user that does not do this?


The user will run `ilab profile set`
Alongside the various model paths and GPU amounts, this command will
set the train profile for the following scenarions:
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
set the train profile for the following scenarions:
set the train profile for the following scenarios:

Comment on lines +18 to +21
2. single consumer GPU
3. multi consumer GPU
4. single server GPU
5. multi server GPU
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Are terms like server and consumer familiar to users? I myself don't fully understand what this means.

5. multi server GPU
6. MacOS (once MPS support is added)

There is also a Choose for me option which reads the Nvidia cards on the system and assigns a cfg+train profile based off the amount of vRAM
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
There is also a Choose for me option which reads the Nvidia cards on the system and assigns a cfg+train profile based off the amount of vRAM
There is also a "Choose for me" option which reads the Nvidia cards on the system and assigns a cfg+train profile based off the amount of vRAM

What if I don't have a Nvidia card?

Comment on lines +33 to +38
[1] CPU
[2] Single Consumer GPU
[3] Single Server GPU
[4] Multi Consumer GPU
[5] Multi Server GPU
[6] MacOS
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is the assumption here systems 0-5 are Linux?

@nathan-weinberg
Copy link
Member

@cdoern what is going on with this?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants