Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

deprecated-mig-v2 #1208

Merged
merged 1 commit into from
Oct 31, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
18 changes: 13 additions & 5 deletions docs/Researcher/cli-reference/new-cli/runai_config_set.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,14 +6,22 @@ Set configuration values
runai config set [flags]
```

### Examples

```
runai config set --status-timeout-duration 5s
runai config set --status-timeout-duration 300ms
```

### Options

```
--auth-url string set the authorization URL; most likely the same as the control plane URL
--cp-url string set the control plane URL
-h, --help help for set
--interactive enable set interactive mode (enabled|disabled)
--output string set the default output type
--auth-url string set the authorization URL; most likely the same as the control plane URL
--cp-url string set the control plane URL
-h, --help help for set
--interactive enable set interactive mode (enabled|disabled)
--output string set the default output type
--status-timeout-duration string set cluster status call timeout duration value, the default is 3 second ("3s")
```

### Options inherited from parent commands
Expand Down
1 change: 1 addition & 0 deletions docs/Researcher/cli-reference/new-cli/runai_mpi.md
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@ alias for mpi management

* [runai](runai.md) - Run:ai Command-line Interface
* [runai mpi attach](runai_mpi_attach.md) - attach to a running container in a mpi training job
* [runai mpi bash](runai_mpi_bash.md) - open a bash shell in a training mpi job
* [runai mpi delete](runai_mpi_delete.md) - delete mpi training workload
* [runai mpi describe](runai_mpi_describe.md) - describe mpi training
* [runai mpi exec](runai_mpi_exec.md) - execute a command in a training mpi job
Expand Down
45 changes: 45 additions & 0 deletions docs/Researcher/cli-reference/new-cli/runai_mpi_bash.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
## runai mpi bash

open a bash shell in a training mpi job

```
runai mpi bash [WORKLOAD_NAME] [flags]
```

### Examples

```
# Open a bash shell in the training mpi's main worker
runai training mpi bash mpi-01

# Open a bash shell in a specific training mpi worker
runai training mpi bash mpi-01 --pod mpi-01-worker-1
```

### Options

```
-c, --container string Container name for log extraction
-h, --help help for bash
--pod string Workload pod ID for log extraction, default: master (0-0)
--pod-running-timeout duration Pod check for running state timeout.
-p, --project string Specify the project to which the command applies. By default, commands apply to the default project. To change the default project use ‘runai config project <project name>’
-i, --stdin Pass stdin to the container
-t, --tty Stdin is a TTY
--wait-timeout duration Timeout for waiting for workload to be ready for log streaming
```

### Options inherited from parent commands

```
--config-file string config file name; can be set by environment variable RUNAI_CLI_CONFIG_FILE (default "config.json")
--config-path string config path; can be set by environment variable RUNAI_CLI_CONFIG_PATH (default "~/.runai/")
-d, --debug enable debug mode
-q, --quiet enable quiet mode, suppress all output except error messages
--verbose enable verbose mode
```

### SEE ALSO

* [runai mpi](runai_mpi.md) - alias for mpi management

2 changes: 1 addition & 1 deletion docs/Researcher/cli-reference/new-cli/runai_mpi_submit.md
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ runai training mpi submit <name> -p <project_name> -i gcr.io/run-ai-demo/quickst
--master-environment stringArray Set master environment variables in the container
--master-extended-resource stringArray Request access to an extended resource. Use the format: resource_name=quantity
--master-no-pvcs Do not mount any persistent volumes in the master pod
--mig-profile string MIG profile to allocate for the job (1g.5gb, 2g.10gb, 3g.20gb, 4g.20gb, 7g.40gb)
--mig-profile string [Deprecated] MIG profile to allocate for the job (1g.5gb, 2g.10gb, 3g.20gb, 4g.20gb, 7g.40gb)
--name-prefix string Set defined prefix for the workload name and add index as a suffix
--new-pvc stringArray Mount a persistent volume, create it if it does not exist. Use the format: claimname=CLAIM_NAME,storageclass=STORAGE_CLASS,size=SIZE,path=PATH,accessmode-rwo,accessmode-rom,accessmode-rwm,ro,ephemeral
--nfs stringArray NFS storage details. Use the format: path=PATH,server=SERVER,mountpath=MOUNT_PATH,readwrite
Expand Down
1 change: 1 addition & 0 deletions docs/Researcher/cli-reference/new-cli/runai_pytorch.md
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@ alias for pytorch management

* [runai](runai.md) - Run:ai Command-line Interface
* [runai pytorch attach](runai_pytorch_attach.md) - attach to a running container in a pytorch training job
* [runai pytorch bash](runai_pytorch_bash.md) - open a bash shell in a training pytorch job
* [runai pytorch delete](runai_pytorch_delete.md) - delete pytorch training workload
* [runai pytorch describe](runai_pytorch_describe.md) - describe pytorch training
* [runai pytorch exec](runai_pytorch_exec.md) - execute a command in a training pytorch job
Expand Down
45 changes: 45 additions & 0 deletions docs/Researcher/cli-reference/new-cli/runai_pytorch_bash.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
## runai pytorch bash

open a bash shell in a training pytorch job

```
runai pytorch bash [WORKLOAD_NAME] [flags]
```

### Examples

```
# Open a bash shell in the training pytorch's main worker
runai training pytorch bash pytorch-01

# Open a bash shell in a specific training pytorch worker
runai training pytorch bash pytorch-01 --pod pytorch-01-worker-1
```

### Options

```
-c, --container string Container name for log extraction
-h, --help help for bash
--pod string Workload pod ID for log extraction, default: master (0-0)
--pod-running-timeout duration Pod check for running state timeout.
-p, --project string Specify the project to which the command applies. By default, commands apply to the default project. To change the default project use ‘runai config project <project name>’
-i, --stdin Pass stdin to the container
-t, --tty Stdin is a TTY
--wait-timeout duration Timeout for waiting for workload to be ready for log streaming
```

### Options inherited from parent commands

```
--config-file string config file name; can be set by environment variable RUNAI_CLI_CONFIG_FILE (default "config.json")
--config-path string config path; can be set by environment variable RUNAI_CLI_CONFIG_PATH (default "~/.runai/")
-d, --debug enable debug mode
-q, --quiet enable quiet mode, suppress all output except error messages
--verbose enable verbose mode
```

### SEE ALSO

* [runai pytorch](runai_pytorch.md) - alias for pytorch management

Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ runai training pytorch submit <name> -p <project_name> -i gcr.io/run-ai-demo/qui
--master-gpu-portion-request float GPU portion request (between 0 and 1, e.g. 0.5, 0.2)
--master-no-pvcs Do not mount any persistent volumes in the master pod
--max-replicas int32 Maximum number of replicas for an elastic PyTorch job
--mig-profile string MIG profile to allocate for the job (1g.5gb, 2g.10gb, 3g.20gb, 4g.20gb, 7g.40gb)
--mig-profile string [Deprecated] MIG profile to allocate for the job (1g.5gb, 2g.10gb, 3g.20gb, 4g.20gb, 7g.40gb)
--min-replicas int32 Minimum number of replicas for an elastic PyTorch job
--name-prefix string Set defined prefix for the workload name and add index as a suffix
--new-pvc stringArray Mount a persistent volume, create it if it does not exist. Use the format: claimname=CLAIM_NAME,storageclass=STORAGE_CLASS,size=SIZE,path=PATH,accessmode-rwo,accessmode-rom,accessmode-rwm,ro,ephemeral
Expand Down
2 changes: 1 addition & 1 deletion docs/Researcher/cli-reference/new-cli/runai_submit.md
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ runai submit [flags]
--large-shm Request large /dev/shm device to mount
--memory string CPU memory to allocate for the job (e.g. 1G, 500M)
--memory-limit string CPU memory limit to allocate for the job (e.g. 1G, 500M)
--mig-profile string MIG profile to allocate for the job (1g.5gb, 2g.10gb, 3g.20gb, 4g.20gb, 7g.40gb)
--mig-profile string [Deprecated] MIG profile to allocate for the job (1g.5gb, 2g.10gb, 3g.20gb, 4g.20gb, 7g.40gb)
--new-pvc stringArray Mount a persistent volume, create it if it does not exist. Use the format: claimname=CLAIM_NAME,storageclass=STORAGE_CLASS,size=SIZE,path=PATH,accessmode-rwo,accessmode-rom,accessmode-rwm,ro,ephemeral
--node-pools stringArray List of node pools to use for scheduling the job, ordered by priority
--node-type string Enforce node type affinity by setting a node-type label
Expand Down
1 change: 1 addition & 0 deletions docs/Researcher/cli-reference/new-cli/runai_tensorflow.md
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@ alias for tensorflow management

* [runai](runai.md) - Run:ai Command-line Interface
* [runai tensorflow attach](runai_tensorflow_attach.md) - attach to a running container in a tf training job
* [runai tensorflow bash](runai_tensorflow_bash.md) - open a bash shell in a training tf job
* [runai tensorflow delete](runai_tensorflow_delete.md) - delete tf training workload
* [runai tensorflow describe](runai_tensorflow_describe.md) - describe tf training
* [runai tensorflow exec](runai_tensorflow_exec.md) - execute a command in a training tf job
Expand Down
45 changes: 45 additions & 0 deletions docs/Researcher/cli-reference/new-cli/runai_tensorflow_bash.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
## runai tensorflow bash

open a bash shell in a training tf job

```
runai tensorflow bash [WORKLOAD_NAME] [flags]
```

### Examples

```
# Open a bash shell in the training tf's main worker
runai training tf bash tf-01

# Open a bash shell in a specific training tf worker
runai training tf bash tf-01 --pod tf-01-worker-1
```

### Options

```
-c, --container string Container name for log extraction
-h, --help help for bash
--pod string Workload pod ID for log extraction, default: master (0-0)
--pod-running-timeout duration Pod check for running state timeout.
-p, --project string Specify the project to which the command applies. By default, commands apply to the default project. To change the default project use ‘runai config project <project name>’
-i, --stdin Pass stdin to the container
-t, --tty Stdin is a TTY
--wait-timeout duration Timeout for waiting for workload to be ready for log streaming
```

### Options inherited from parent commands

```
--config-file string config file name; can be set by environment variable RUNAI_CLI_CONFIG_FILE (default "config.json")
--config-path string config path; can be set by environment variable RUNAI_CLI_CONFIG_PATH (default "~/.runai/")
-d, --debug enable debug mode
-q, --quiet enable quiet mode, suppress all output except error messages
--verbose enable verbose mode
```

### SEE ALSO

* [runai tensorflow](runai_tensorflow.md) - alias for tensorflow management

Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ runai training tf submit <name> -p <project_name> -i gcr.io/run-ai-demo/quicksta
--master-gpu-portion-request float GPU portion request (between 0 and 1, e.g. 0.5, 0.2)
--master-no-pvcs Do not mount any persistent volumes in the master pod
--max-replicas int32 Maximum number of replicas for an elastic PyTorch job
--mig-profile string MIG profile to allocate for the job (1g.5gb, 2g.10gb, 3g.20gb, 4g.20gb, 7g.40gb)
--mig-profile string [Deprecated] MIG profile to allocate for the job (1g.5gb, 2g.10gb, 3g.20gb, 4g.20gb, 7g.40gb)
--min-replicas int32 Minimum number of replicas for an elastic PyTorch job
--name-prefix string Set defined prefix for the workload name and add index as a suffix
--new-pvc stringArray Mount a persistent volume, create it if it does not exist. Use the format: claimname=CLAIM_NAME,storageclass=STORAGE_CLASS,size=SIZE,path=PATH,accessmode-rwo,accessmode-rom,accessmode-rwm,ro,ephemeral
Expand Down
3 changes: 2 additions & 1 deletion docs/Researcher/cli-reference/new-cli/runai_training.md
Original file line number Diff line number Diff line change
Expand Up @@ -22,10 +22,11 @@ training management

* [runai](runai.md) - Run:ai Command-line Interface
* [runai training attach](runai_training_attach.md) - attach to a running container in a standard training job
* [runai training bash](runai_training_bash.md) - open a bash shell in a training standard job
* [runai training delete](runai_training_delete.md) - delete standard training workload
* [runai training describe](runai_training_describe.md) - describe standard training
* [runai training exec](runai_training_exec.md) - execute a command in a training standard job
* [runai training list](runai_training_list.md) - list training
* [runai training list](runai_training_list.md) - list all training frameworks
* [runai training logs](runai_training_logs.md) - view logs of a standard training job
* [runai training mpi](runai_training_mpi.md) - mpi management
* [runai training port-forward](runai_training_port-forward.md) - forward one or more local ports to a standard training job
Expand Down
45 changes: 45 additions & 0 deletions docs/Researcher/cli-reference/new-cli/runai_training_bash.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
## runai training bash

open a bash shell in a training standard job

```
runai training bash [WORKLOAD_NAME] [flags]
```

### Examples

```
# Open a bash shell in the training standard's main worker
runai training standard bash standard-01

# Open a bash shell in a specific training standard worker
runai training standard bash standard-01 --pod standard-01-worker-1
```

### Options

```
-c, --container string Container name for log extraction
-h, --help help for bash
--pod string Workload pod ID for log extraction, default: master (0-0)
--pod-running-timeout duration Pod check for running state timeout.
-p, --project string Specify the project to which the command applies. By default, commands apply to the default project. To change the default project use ‘runai config project <project name>’
-i, --stdin Pass stdin to the container
-t, --tty Stdin is a TTY
--wait-timeout duration Timeout for waiting for workload to be ready for log streaming
```

### Options inherited from parent commands

```
--config-file string config file name; can be set by environment variable RUNAI_CLI_CONFIG_FILE (default "config.json")
--config-path string config path; can be set by environment variable RUNAI_CLI_CONFIG_PATH (default "~/.runai/")
-d, --debug enable debug mode
-q, --quiet enable quiet mode, suppress all output except error messages
--verbose enable verbose mode
```

### SEE ALSO

* [runai training](runai_training.md) - training management

Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
## runai training list

list training
list all training frameworks

```
runai training list [flags]
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@ mpi management

* [runai training](runai_training.md) - training management
* [runai training mpi attach](runai_training_mpi_attach.md) - attach to a running container in a mpi training job
* [runai training mpi bash](runai_training_mpi_bash.md) - open a bash shell in a training mpi job
* [runai training mpi delete](runai_training_mpi_delete.md) - delete mpi training workload
* [runai training mpi describe](runai_training_mpi_describe.md) - describe mpi training
* [runai training mpi exec](runai_training_mpi_exec.md) - execute a command in a training mpi job
Expand Down
45 changes: 45 additions & 0 deletions docs/Researcher/cli-reference/new-cli/runai_training_mpi_bash.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
## runai training mpi bash

open a bash shell in a training mpi job

```
runai training mpi bash [WORKLOAD_NAME] [flags]
```

### Examples

```
# Open a bash shell in the training mpi's main worker
runai training mpi bash mpi-01

# Open a bash shell in a specific training mpi worker
runai training mpi bash mpi-01 --pod mpi-01-worker-1
```

### Options

```
-c, --container string Container name for log extraction
-h, --help help for bash
--pod string Workload pod ID for log extraction, default: master (0-0)
--pod-running-timeout duration Pod check for running state timeout.
-p, --project string Specify the project to which the command applies. By default, commands apply to the default project. To change the default project use ‘runai config project <project name>’
-i, --stdin Pass stdin to the container
-t, --tty Stdin is a TTY
--wait-timeout duration Timeout for waiting for workload to be ready for log streaming
```

### Options inherited from parent commands

```
--config-file string config file name; can be set by environment variable RUNAI_CLI_CONFIG_FILE (default "config.json")
--config-path string config path; can be set by environment variable RUNAI_CLI_CONFIG_PATH (default "~/.runai/")
-d, --debug enable debug mode
-q, --quiet enable quiet mode, suppress all output except error messages
--verbose enable verbose mode
```

### SEE ALSO

* [runai training mpi](runai_training_mpi.md) - mpi management

Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ runai training mpi submit <name> -p <project_name> -i gcr.io/run-ai-demo/quickst
--master-environment stringArray Set master environment variables in the container
--master-extended-resource stringArray Request access to an extended resource. Use the format: resource_name=quantity
--master-no-pvcs Do not mount any persistent volumes in the master pod
--mig-profile string MIG profile to allocate for the job (1g.5gb, 2g.10gb, 3g.20gb, 4g.20gb, 7g.40gb)
--mig-profile string [Deprecated] MIG profile to allocate for the job (1g.5gb, 2g.10gb, 3g.20gb, 4g.20gb, 7g.40gb)
--name-prefix string Set defined prefix for the workload name and add index as a suffix
--new-pvc stringArray Mount a persistent volume, create it if it does not exist. Use the format: claimname=CLAIM_NAME,storageclass=STORAGE_CLASS,size=SIZE,path=PATH,accessmode-rwo,accessmode-rom,accessmode-rwm,ro,ephemeral
--nfs stringArray NFS storage details. Use the format: path=PATH,server=SERVER,mountpath=MOUNT_PATH,readwrite
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@ pytorch management

* [runai training](runai_training.md) - training management
* [runai training pytorch attach](runai_training_pytorch_attach.md) - attach to a running container in a pytorch training job
* [runai training pytorch bash](runai_training_pytorch_bash.md) - open a bash shell in a training pytorch job
* [runai training pytorch delete](runai_training_pytorch_delete.md) - delete pytorch training workload
* [runai training pytorch describe](runai_training_pytorch_describe.md) - describe pytorch training
* [runai training pytorch exec](runai_training_pytorch_exec.md) - execute a command in a training pytorch job
Expand Down
Loading
Loading