You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
-6Lines changed: 0 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -196,12 +196,6 @@ The [Device plugins operator README](cmd/operator/README.md) gives the installat
196
196
197
197
The [Device plugins Operator for OpenShift](https://github.com/intel/intel-technology-enabling-for-openshift) gives the installation and usage details for the operator available on [Red Hat OpenShift Container Platform](https://catalog.redhat.com/software/operators/detail/61e9f2d7b9cdd99018fc5736).
198
198
199
-
## XeLink XPU Manager Sidecar
200
-
201
-
To support interconnected GPUs in Kubernetes, XeLink sidecar is needed.
202
-
203
-
The [XeLink XPU Manager sidecar README](cmd/xpumanager_sidecar/README.md) gives information how the sidecar functions and how to use it.
204
-
205
199
## Intel GPU Level-Zero sidecar
206
200
207
201
Sidecar uses Level-Zero API to provide additional GPU information for the GPU plugin that it cannot get through sysfs interfaces.
Copy file name to clipboardExpand all lines: cmd/gpu_plugin/README.md
+3-13Lines changed: 3 additions & 13 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -47,20 +47,17 @@ Intel GPU plugin may register four node resources to the Kubernetes cluster:
47
47
| gpu.intel.com/xe | GPU instance running new `xe` KMD |
48
48
| gpu.intel.com/xe_monitoring | Monitoring resource for the new `xe` KMD devices |
49
49
50
-
While GPU plugin basic operations support nodes having both (`i915` and `xe`) KMDs on the same node, its resource management (=GAS) does not, for that node needs to have only one of the KMDs present.
51
-
52
50
For workloads on different KMDs, see [KMD and UMD](#kmd-and-umd).
53
51
54
52
## Modes and Configuration Options
55
53
56
54
| Flag | Argument | Default | Meaning |
57
55
|:---- |:-------- |:------- |:------- |
58
56
| -enable-monitoring | - | disabled | Enable '*_monitoring' resource that provides access to all Intel GPU devices on the node, [see use](./monitoring.md)|
| -health-management | - | disabled | Enable health management by requesting data from oneAPI/Level-Zero interface. Requires [GPU Level-Zero](../gpu_levelzero/) sidecar. See [health management](#health-management)|
61
58
| -wsl | - | disabled | Adapt plugin to run in the WSL environment. Requires [GPU Level-Zero](../gpu_levelzero/) sidecar. |
62
59
| -shared-dev-num | int | 1 | Number of containers that can share the same GPU device |
63
-
| -allocation-policy | string | none | 3 possible values: balanced, packed, none. For shared-dev-num > 1: _balanced_ mode spreads workloads among GPU devices, _packed_ mode fills one GPU fully before moving to next, and _none_ selects first available device from kubelet. Default is _none_. Allocation policy does not have an effect when resource manager is enabled. |
60
+
| -allocation-policy | string | none | 3 possible values: balanced, packed, none. For shared-dev-num > 1: _balanced_ mode spreads workloads among GPU devices, _packed_ mode fills one GPU fully before moving to next, and _none_ selects first available device from kubelet. Default is _none_. |
64
61
65
62
The plugin also accepts a number of other arguments (common to all plugins) related to logging.
66
63
Please use the -h option to see the complete list of logging related options.
@@ -75,9 +72,6 @@ Intel GPU-plugin supports a few different operation modes. Depending on the work
75
72
|:---- |:-------- |:------- |:------- |
76
73
| shared-dev-num == 1 | No, 1 container per GPU | Workloads using all GPU capacity, e.g. AI training | Yes |
77
74
| shared-dev-num > 1 | Yes, >1 containers per GPU | (Batch) workloads using only part of GPU resources, e.g. inference, media transcode/analytics, or CPU bound GPU workloads | No |
78
-
| shared-dev-num > 1 && resource-management | Depends on resource requests | Any. For requirements and usage, see [fractional resource management](./fractional.md)| Yes. 1000 millicores = exclusive GPU usage. See note below. |
79
-
80
-
> **Note**: Exclusive GPU usage with >=1000 millicores requires that also *all other GPU containers* specify (non-zero) millicores resource usage.
GPU plugin can be installed with the Intel Device Plugin Operator. It allows configuring GPU plugin's parameters without kustomizing the deployment files. The general installation is described in the [install documentation](../operator/README.md#installation). For configuring the GPU Custom Resource (CR), see the [configuration options](#modes-and-configuration-options) and [operation modes](#operation-modes-for-different-workload-types).
124
118
125
-
### Install alongside with GPU Aware Scheduling (deprecated)
126
-
127
-
GPU plugin can be installed alongside with GPU Aware Scheduling (GAS). It allows scheduling Pods which e.g. request only partial use of a GPU. The installation is described in [fractional resources](./fractional.md) page.
128
-
129
119
### Verify Plugin Installation
130
120
131
121
You can verify that the plugin has been installed on the expected nodes by searching for the relevant
@@ -212,9 +202,9 @@ Furthermore, the deployments `securityContext` must be configured with appropria
212
202
213
203
More info: https://kubernetes.io/blog/2021/11/09/non-root-containers-and-devices/
214
204
215
-
### Labels created by GPU plugin
205
+
### Labels created for Intel GPUs via NFD
216
206
217
-
If installed with NFD and started with resource-management, plugin will export a set of labels for the node. For detailed info, see [labeling documentation](./labels.md).
207
+
When NFD's NodeFeatureRules for Intel GPUs are installed, nodes are labeled with a variaty of GPU specific labels. For detailed info, see [labeling documentation](./labels.md).
0 commit comments