File: typos.patch

package info (click to toggle)
nvidia-cuda-toolkit 11.2.2-3%2Bdeb11u3
  • links: PTS, VCS
  • area: non-free
  • in suites: bullseye
  • size: 12,794,176 kB
  • sloc: ansic: 148,865; cpp: 50,458; javascript: 21,680; xml: 11,978; python: 4,582; makefile: 1,808; sh: 553; sql: 187; perl: 153
file content (547 lines) | stat: -rw-r--r-- 33,587 bytes parent folder | download
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
Description: fix spelling-error-in-manpage and other typos
Author: Andreas Beckmann <anbe@debian.org>
Author: Graham Inggs <graham@nerve.org.za>

--- a/nvidia-cuda/cuda_cudart/include/cuda.h
+++ b/nvidia-cuda/cuda_cudart/include/cuda.h
@@ -783,7 +783,7 @@ typedef enum CUcomputemode_enum {
  * Memory advise values
  */
 typedef enum CUmem_advise_enum {
-    CU_MEM_ADVISE_SET_READ_MOSTLY          = 1, /**< Data will mostly be read and only occassionally be written to */
+    CU_MEM_ADVISE_SET_READ_MOSTLY          = 1, /**< Data will mostly be read and only occasionally be written to */
     CU_MEM_ADVISE_UNSET_READ_MOSTLY        = 2, /**< Undo the effect of ::CU_MEM_ADVISE_SET_READ_MOSTLY */
     CU_MEM_ADVISE_SET_PREFERRED_LOCATION   = 3, /**< Set the preferred location for the data as the specified device */
     CU_MEM_ADVISE_UNSET_PREFERRED_LOCATION = 4, /**< Clear the preferred location for the data */
@@ -792,7 +792,7 @@ typedef enum CUmem_advise_enum {
 } CUmem_advise;
 
 typedef enum CUmem_range_attribute_enum {
-    CU_MEM_RANGE_ATTRIBUTE_READ_MOSTLY            = 1, /**< Whether the range will mostly be read and only occassionally be written to */
+    CU_MEM_RANGE_ATTRIBUTE_READ_MOSTLY            = 1, /**< Whether the range will mostly be read and only occasionally be written to */
     CU_MEM_RANGE_ATTRIBUTE_PREFERRED_LOCATION     = 2, /**< The preferred location of the range */
     CU_MEM_RANGE_ATTRIBUTE_ACCESSED_BY            = 3, /**< Memory range has ::CU_MEM_ADVISE_SET_ACCESSED_BY set for specified device */
     CU_MEM_RANGE_ATTRIBUTE_LAST_PREFETCH_LOCATION = 4  /**< The last location to which the range was prefetched */
@@ -3074,7 +3074,7 @@ CUresult CUDAAPI cuDeviceGet(CUdevice *d
 CUresult CUDAAPI cuDeviceGetCount(int *count);
 
 /**
- * \brief Returns an identifer string for the device
+ * \brief Returns an identifier string for the device
  *
  * Returns an ASCII string identifying the device \p dev in the NULL-terminated
  * string pointed to by \p name. \p len specifies the maximum length of the
@@ -3385,7 +3385,7 @@ CUresult CUDAAPI cuDeviceGetTexture1DLin
  *   supports native atomic operations.
  * - ::CU_DEVICE_ATTRIBUTE_SINGLE_TO_DOUBLE_PRECISION_PERF_RATIO: Ratio of single precision performance
  *   (in floating-point operations per second) to double precision performance.
- * - ::CU_DEVICE_ATTRIBUTE_PAGEABLE_MEMORY_ACCESS: Device suppports coherently accessing
+ * - ::CU_DEVICE_ATTRIBUTE_PAGEABLE_MEMORY_ACCESS: Device supports coherently accessing
  *   pageable memory without calling cudaHostRegister on it.
  * - ::CU_DEVICE_ATTRIBUTE_CONCURRENT_MANAGED_ACCESS: Device can coherently access managed memory
  *   concurrently with the CPU.
@@ -3393,7 +3393,7 @@ CUresult CUDAAPI cuDeviceGetTexture1DLin
  * - ::CU_DEVICE_ATTRIBUTE_CAN_USE_HOST_POINTER_FOR_REGISTERED_MEM: Device can access host registered
  *   memory at the same virtual address as the CPU.
  * -  ::CU_DEVICE_ATTRIBUTE_MAX_SHARED_MEMORY_PER_BLOCK_OPTIN: The maximum per block shared memory size
- *    suported on this device. This is the maximum value that can be opted into when using the cuFuncSetAttribute() call.
+ *    supported on this device. This is the maximum value that can be opted into when using the cuFuncSetAttribute() call.
  *    For more details see ::CU_FUNC_ATTRIBUTE_MAX_DYNAMIC_SHARED_SIZE_BYTES
  * - ::CU_DEVICE_ATTRIBUTE_PAGEABLE_MEMORY_ACCESS_USES_HOST_PAGE_TABLES: Device accesses pageable memory via the host's
  *   page tables.
@@ -3455,7 +3455,7 @@ CUresult CUDAAPI cuDeviceGetAttribute(in
  *
  * At least one of these flags must be set, failing which the API
  * returns ::CUDA_ERROR_INVALID_VALUE. Both the flags are orthogonal
- * to one another: a developer may set both these flags that allows to
+ * to one another: a developer may set both these flags that allows one to
  * set both wait and signal specific attributes in the same \p nvSciSyncAttrList.
  *
  * \param nvSciSyncAttrList     - Return NvSciSync attributes supported.
@@ -3564,7 +3564,7 @@ __CUDA_DEPRECATED CUresult CUDAAPI cuDev
  *
  * \deprecated
  *
- * This function was deprecated as of CUDA 5.0 and its functionality superceded
+ * This function was deprecated as of CUDA 5.0 and its functionality superseded
  * by ::cuDeviceGetAttribute().
  *
  * Returns in \p *major and \p *minor the major and minor revision numbers that
@@ -3704,7 +3704,7 @@ CUresult CUDAAPI cuDevicePrimaryCtxRelea
 /**
  * \brief Set flags for the primary context
  *
- * Sets the flags for the primary context on the device overwriting perviously
+ * Sets the flags for the primary context on the device overwriting previously
  * set ones.
  *
  * The three LSBs of the \p flags parameter can be used to control how the OS
@@ -5663,7 +5663,7 @@ CUresult CUDAAPI cuMemHostGetFlags(unsig
  * ::cuStreamAttachMemAsync will be required to enable access on such devices.
  *
  * If the association is later changed via ::cuStreamAttachMemAsync to
- * a single stream, the default association as specifed during ::cuMemAllocManaged
+ * a single stream, the default association as specified during ::cuMemAllocManaged
  * is restored when that stream is destroyed. For __managed__ variables, the
  * default association is always ::CU_MEM_ATTACH_GLOBAL. Note that destroying a
  * stream is an asynchronous operation, and as a result, the change to default
@@ -8848,13 +8848,13 @@ CUresult CUDAAPI cuMemAddressFree(CUdevi
 * \brief Create a CUDA memory handle representing a memory allocation of a given size described by the given properties
 *
 * This creates a memory allocation on the target device specified through the
-* \p prop strcuture. The created allocation will not have any device or host
+* \p prop structure. The created allocation will not have any device or host
 * mappings. The generic memory \p handle for the allocation can be
 * mapped to the address space of calling process via ::cuMemMap. This handle
 * cannot be transmitted directly to other processes (see
 * ::cuMemExportToShareableHandle).  On Windows, the caller must also pass
 * an LPSECURITYATTRIBUTE in \p prop to be associated with this handle which
-* limits or allows access to this handle for a recepient process (see
+* limits or allows access to this handle for a recipient process (see
 * ::CUmemAllocationProp::win32HandleMetaData for more).  The \p size of this
 * allocation must be a multiple of the the value given via
 * ::cuMemGetAllocationGranularity with the ::CU_MEM_ALLOC_GRANULARITY_MINIMUM
@@ -8892,7 +8892,7 @@ CUresult CUDAAPI cuMemCreate(CUmemGeneri
 * are unmapped and when all outstanding references to the handle (including it's
 * shareable counterparts) are also released. The generic memory handle can be
 * freed when there are still outstanding mappings made with this handle. Each
-* time a recepient process imports a shareable handle, it needs to pair it with
+* time a recipient process imports a shareable handle, it needs to pair it with
 * ::cuMemRelease for the handle to be freed.  If \p handle is not a valid handle
 * the behavior is undefined. 
 *
@@ -12210,7 +12210,7 @@ CUresult CUDAAPI cuLaunchCooperativeKern
  * All kernels launched must be identical with respect to the compiled code. Note that
  * any __device__, __constant__ or __managed__ variables present in the module that owns
  * the kernel launched on each device, are independently instantiated on every device.
- * It is the application's responsiblity to ensure these variables are initialized and
+ * It is the application's responsibility to ensure these variables are initialized and
  * used appropriately.
  *
  * The size of the grids as specified in blocks, the size of the blocks themselves
--- a/nvidia-cuda/cuda_cudart/include/cuda_runtime.h
+++ b/nvidia-cuda/cuda_cudart/include/cuda_runtime.h
@@ -171,8 +171,8 @@ struct  __device_builtin__ __nv_lambda_p
  * \p stream specifies a stream the invocation is associated to.
  *
  * \param func        - Device function symbol
- * \param gridDim     - Grid dimentions
- * \param blockDim    - Block dimentions
+ * \param gridDim     - Grid dimensions
+ * \param blockDim    - Block dimensions
  * \param args        - Arguments
  * \param sharedMem   - Shared memory (defaults to 0)
  * \param stream      - Stream identifier (defaults to NULL)
@@ -237,8 +237,8 @@ static __inline__ __host__ cudaError_t c
  * \p stream specifies a stream the invocation is associated to.
  *
  * \param func        - Device function symbol
- * \param gridDim     - Grid dimentions
- * \param blockDim    - Block dimentions
+ * \param gridDim     - Grid dimensions
+ * \param blockDim    - Block dimensions
  * \param args        - Arguments
  * \param sharedMem   - Shared memory (defaults to 0)
  * \param stream      - Stream identifier (defaults to NULL)
@@ -420,7 +420,7 @@ static __inline__ __host__ cudaError_t c
  * ::cudaStreamAttachMemAsync will be required to enable access on such devices.
  *
  * If the association is later changed via ::cudaStreamAttachMemAsync to
- * a single stream, the default association, as specifed during ::cudaMallocManaged,
+ * a single stream, the default association, as specified during ::cudaMallocManaged,
  * is restored when that stream is destroyed. For __managed__ variables, the
  * default association is always ::cudaMemAttachGlobal. Note that destroying a
  * stream is an asynchronous operation, and as a result, the change to default
@@ -1772,7 +1772,7 @@ static __inline__ __host__ cudaError_t c
  * streaming multiprocessor for the device function.
  *
  * \param numBlocks       - Returned occupancy
- * \param func            - Kernel function for which occupancy is calulated
+ * \param func            - Kernel function for which occupancy is calculated
  * \param blockSize       - Block size the kernel is intended to be launched with
  * \param dynamicSMemSize - Per-block dynamic shared memory usage intended, in bytes
  *
@@ -1823,7 +1823,7 @@ static __inline__ __host__ cudaError_t c
  *   section of the Maxwell tuning guide.
  *
  * \param numBlocks       - Returned occupancy
- * \param func            - Kernel function for which occupancy is calulated
+ * \param func            - Kernel function for which occupancy is calculated
  * \param blockSize       - Block size the kernel is intended to be launched with
  * \param dynamicSMemSize - Per-block dynamic shared memory usage intended, in bytes
  * \param flags           - Requested behavior for the occupancy calculator
@@ -2206,7 +2206,7 @@ static __inline__ __host__ cudaError_t c
 }
 
 /**
- * \brief Returns grid and block size that achived maximum potential occupancy for a device function with the specified flags
+ * \brief Returns grid and block size that achieved maximum potential occupancy for a device function with the specified flags
  *
  * Returns in \p *minGridSize and \p *blocksize a suggested grid /
  * block size pair that achieves the best potential occupancy
--- a/nvidia-cuda/cuda_cudart/include/cuda_runtime_api.h
+++ b/nvidia-cuda/cuda_cudart/include/cuda_runtime_api.h
@@ -409,7 +409,7 @@ extern __host__ cudaError_t CUDARTAPI cu
  * - ::cudaLimitMallocHeapSize: size in bytes of the heap used by the
  *   ::malloc() and ::free() device system calls;
  * - ::cudaLimitDevRuntimeSyncDepth: maximum grid depth at which a
- *   thread can isssue the device runtime call ::cudaDeviceSynchronize()
+ *   thread can issue the device runtime call ::cudaDeviceSynchronize()
  *   to wait on child grid launches to complete.
  * - ::cudaLimitDevRuntimePendingLaunchCount: maximum number of outstanding
  *   device runtime launches.
@@ -902,7 +902,7 @@ extern __host__ cudaError_t CUDARTAPI cu
 /**
  * \brief Attempts to close memory mapped with cudaIpcOpenMemHandle
  * 
- * Decrements the reference count of the memory returnd by ::cudaIpcOpenMemHandle by 1.
+ * Decrements the reference count of the memory returned by ::cudaIpcOpenMemHandle by 1.
  * When the reference count reaches 0, this API unmaps the memory. The original allocation
  * in the exporting process as well as imported mappings in other processes
  * will be unaffected.
@@ -1830,7 +1830,7 @@ extern __host__ __cudart_builtin__ cudaE
  *
  * At least one of these flags must be set, failing which the API
  * returns ::cudaErrorInvalidValue. Both the flags are orthogonal
- * to one another: a developer may set both these flags that allows to
+ * to one another: a developer may set both these flags that allows one to
  * set both wait and signal specific attributes in the same \p nvSciSyncAttrList.
  *
  * \param nvSciSyncAttrList     - Return NvSciSync attributes supported.
@@ -3762,8 +3762,8 @@ extern __host__ cudaError_t CUDARTAPI cu
  * \p stream specifies a stream the invocation is associated to.
  *
  * \param func        - Device function symbol
- * \param gridDim     - Grid dimentions
- * \param blockDim    - Block dimentions
+ * \param gridDim     - Grid dimensions
+ * \param blockDim    - Block dimensions
  * \param args        - Arguments
  * \param sharedMem   - Shared memory
  * \param stream      - Stream identifier
@@ -3821,8 +3821,8 @@ extern __host__ cudaError_t CUDARTAPI cu
  * \p stream specifies a stream the invocation is associated to.
  *
  * \param func        - Device function symbol
- * \param gridDim     - Grid dimentions
- * \param blockDim    - Block dimentions
+ * \param gridDim     - Grid dimensions
+ * \param blockDim    - Block dimensions
  * \param args        - Arguments
  * \param sharedMem   - Shared memory
  * \param stream      - Stream identifier
@@ -3862,7 +3862,7 @@ extern __host__ cudaError_t CUDARTAPI cu
  *
  * The same kernel must be launched on all devices. Note that any __device__ or __constant__
  * variables are independently instantiated on every device. It is the application's
- * responsiblity to ensure these variables are initialized and used appropriately.
+ * responsibility to ensure these variables are initialized and used appropriately.
  *
  * The size of the grids as specified in blocks, the size of the blocks themselves and the
  * amount of shared memory used by each thread block must also match across all launched kernels.
@@ -4410,7 +4410,7 @@ extern __host__ __cudart_builtin__ cudaE
  * ::cudaStreamAttachMemAsync will be required to enable access on such devices.
  *
  * If the association is later changed via ::cudaStreamAttachMemAsync to
- * a single stream, the default association, as specifed during ::cudaMallocManaged,
+ * a single stream, the default association, as specified during ::cudaMallocManaged,
  * is restored when that stream is destroyed. For __managed__ variables, the
  * default association is always ::cudaMemAttachGlobal. Note that destroying a
  * stream is an asynchronous operation, and as a result, the change to default
@@ -5379,7 +5379,7 @@ struct cudaMemcpy3DParms {
 };
 \endcode
  *
- * ::cudaMemcpy3D() copies data betwen two 3D objects. The source and
+ * ::cudaMemcpy3D() copies data between two 3D objects. The source and
  * destination objects may be in either host memory, device memory, or a CUDA
  * array. The source, destination, extent, and kind of copy performed is
  * specified by the ::cudaMemcpy3DParms struct which should be initialized to
@@ -5515,7 +5515,7 @@ struct cudaMemcpy3DParms {
 };
 \endcode
  *
- * ::cudaMemcpy3DAsync() copies data betwen two 3D objects. The source and
+ * ::cudaMemcpy3DAsync() copies data between two 3D objects. The source and
  * destination objects may be in either host memory, device memory, or a CUDA
  * array. The source, destination, extent, and kind of copy performed is
  * specified by the ::cudaMemcpy3DParms struct which should be initialized to
--- a/nvidia-cuda/cuda_cupti/extras/CUPTI/include/cupti_callbacks.h
+++ b/nvidia-cuda/cuda_cupti/extras/CUPTI/include/cupti_callbacks.h
@@ -590,7 +590,7 @@ CUptiResult CUPTIAPI cuptiSupportedDomai
  * \param subscriber Returns handle to initialize subscriber
  * \param callback The callback function
  * \param userdata A pointer to user data. This data will be passed to
- * the callback function via the \p userdata paramater.
+ * the callback function via the \p userdata parameter.
  *
  * \retval CUPTI_SUCCESS on success
  * \retval CUPTI_ERROR_NOT_INITIALIZED if unable to initialize CUPTI
--- a/nvidia-cuda/libcublas/include/cublasXt.h
+++ b/nvidia-cuda/libcublas/include/cublasXt.h
@@ -73,7 +73,7 @@ cublasStatus_t CUBLASWINAPI cublasXtMaxB
 /* This routine selects the Gpus that the user want to use for CUBLAS-XT */
 cublasStatus_t CUBLASWINAPI cublasXtDeviceSelect(cublasXtHandle_t handle, int nbDevices, int deviceId[]);
 
-/* This routine allows to change the dimension of the tiles ( blockDim x blockDim ) */
+/* This routine allows one to change the dimension of the tiles ( blockDim x blockDim ) */
 cublasStatus_t CUBLASWINAPI cublasXtSetBlockDim(cublasXtHandle_t handle, int blockDim);
 cublasStatus_t CUBLASWINAPI cublasXtGetBlockDim(cublasXtHandle_t handle, int *blockDim);
 
@@ -81,7 +81,7 @@ typedef enum {
     CUBLASXT_PINNING_DISABLED   = 0,  
     CUBLASXT_PINNING_ENABLED    = 1        
 } cublasXtPinnedMemMode_t;
-/* This routine allows to CUBLAS-XT to pin the Host memory if it find out that some of the matrix passed
+/* This routine allows one to CUBLAS-XT to pin the Host memory if it find out that some of the matrix passed
    are not pinned : Pinning/Unpinning the Host memory is still a costly operation
    It is better if the user controls the memory on its own (by pinning/unpinning oly when necessary)
 */
--- a/nvidia-cuda/cuda_cudart/include/driver_types.h
+++ b/nvidia-cuda/cuda_cudart/include/driver_types.h
@@ -1599,7 +1599,7 @@ enum __device_builtin__ cudaLimit
  */
 enum __device_builtin__ cudaMemoryAdvise
 {
-    cudaMemAdviseSetReadMostly          = 1, /**< Data will mostly be read and only occassionally be written to */
+    cudaMemAdviseSetReadMostly          = 1, /**< Data will mostly be read and only occasionally be written to */
     cudaMemAdviseUnsetReadMostly        = 2, /**< Undo the effect of ::cudaMemAdviseSetReadMostly */
     cudaMemAdviseSetPreferredLocation   = 3, /**< Set the preferred location for the data as the specified device */
     cudaMemAdviseUnsetPreferredLocation = 4, /**< Clear the preferred location for the data */
@@ -1612,7 +1612,7 @@ enum __device_builtin__ cudaMemoryAdvise
  */
 enum __device_builtin__ cudaMemRangeAttribute
 {
-    cudaMemRangeAttributeReadMostly           = 1, /**< Whether the range will mostly be read and only occassionally be written to */
+    cudaMemRangeAttributeReadMostly           = 1, /**< Whether the range will mostly be read and only occasionally be written to */
     cudaMemRangeAttributePreferredLocation    = 2, /**< The preferred location of the range */
     cudaMemRangeAttributeAccessedBy           = 3, /**< Memory range has ::cudaMemAdviseSetAccessedBy set for specified device */
     cudaMemRangeAttributeLastPrefetchLocation = 4  /**< The last location to which the range was prefetched */
@@ -2384,8 +2384,8 @@ enum __device_builtin__ cudaCGScope {
 struct __device_builtin__ cudaLaunchParams
 {
     void *func;          /**< Device function symbol */
-    dim3 gridDim;        /**< Grid dimentions */
-    dim3 blockDim;       /**< Block dimentions */
+    dim3 gridDim;        /**< Grid dimensions */
+    dim3 blockDim;       /**< Block dimensions */
     void **args;         /**< Arguments */
     size_t sharedMem;    /**< Shared memory */
     cudaStream_t stream; /**< Stream identifier */
--- a/nvidia-cuda/cuda_nvml_dev/include/nvml.h
+++ b/nvidia-cuda/cuda_nvml_dev/include/nvml.h
@@ -725,7 +725,7 @@ typedef enum nvmlPStates_enum
 /**
  * GPU Operation Mode
  *
- * GOM allows to reduce power usage and optimize GPU throughput by disabling GPU features.
+ * GOM allows one to reduce power usage and optimize GPU throughput by disabling GPU features.
  *
  * Each GOM is designed to meet specific user needs.
  */
@@ -781,7 +781,7 @@ typedef enum nvmlReturn_enum
     NVML_ERROR_IN_USE = 19,                  //!< An operation cannot be performed because the GPU is currently in use
     NVML_ERROR_MEMORY = 20,                  //!< Insufficient memory
     NVML_ERROR_NO_DATA = 21,                 //!< No data
-    NVML_ERROR_VGPU_ECC_NOT_SUPPORTED = 22,  //!< The requested vgpu operation is not available on target device, becasue ECC is enabled
+    NVML_ERROR_VGPU_ECC_NOT_SUPPORTED = 22,  //!< The requested vgpu operation is not available on target device, because ECC is enabled
     NVML_ERROR_INSUFFICIENT_RESOURCES = 23,  //!< Ran out of critical resources, other than memory
     NVML_ERROR_UNKNOWN = 999                 //!< An internal driver error occurred
 } nvmlReturn_t;
@@ -1930,7 +1930,7 @@ nvmlReturn_t DECLDIR nvmlSystemGetNVMLVe
  *
  * For all products.
  *
- * The CUDA driver version returned will be retreived from the currently installed version of CUDA.
+ * The CUDA driver version returned will be retrieved from the currently installed version of CUDA.
  * If the cuda library is not found, this function will return a known supported version number.
  *
  * @param cudaDriverVersion                    Reference in which to return the version identifier
@@ -4084,7 +4084,7 @@ nvmlReturn_t DECLDIR nvmlDeviceGetEncode
  * array elememt count is passed in \a sessionCount, and \a sessionCount is used to return the number of sessions
  * written to the buffer.
  *
- * If the supplied buffer is not large enough to accomodate the active session array, the function returns
+ * If the supplied buffer is not large enough to accommodate the active session array, the function returns
  * NVML_ERROR_INSUFFICIENT_SIZE, with the element count of nvmlEncoderSessionInfo_t array required in \a sessionCount.
  * To query the number of active encoder sessions, call this function with *sessionCount = 0.  The code will return
  * NVML_SUCCESS with number of active encoder sessions updated in *sessionCount.
@@ -4132,7 +4132,7 @@ nvmlReturn_t DECLDIR nvmlDeviceGetDecode
 * For Maxwell &tm; or newer fully supported devices.
 *
 * @param device                            The identifier of the target device
-* @param fbcStats                          Reference to nvmlFBCStats_t structure contianing NvFBC stats
+* @param fbcStats                          Reference to nvmlFBCStats_t structure containing NvFBC stats
 *
 * @return
 *         - \ref NVML_SUCCESS                  if \a fbcStats is fetched
@@ -5900,7 +5900,7 @@ nvmlReturn_t DECLDIR nvmlDeviceGetProces
  * pointed at by \a vgpuTypeIds. The element count of nvmlVgpuTypeId_t array is passed in \a vgpuCount, and \a vgpuCount
  * is used to return the number of vGPU types written to the buffer.
  *
- * If the supplied buffer is not large enough to accomodate the vGPU type array, the function returns
+ * If the supplied buffer is not large enough to accommodate the vGPU type array, the function returns
  * NVML_ERROR_INSUFFICIENT_SIZE, with the element count of nvmlVgpuTypeId_t array required in \a vgpuCount.
  * To query the number of vGPU types supported for the GPU, call this function with *vgpuCount = 0.
  * The code will return NVML_ERROR_INSUFFICIENT_SIZE, or NVML_SUCCESS if no vGPU types are supported.
@@ -5929,7 +5929,7 @@ nvmlReturn_t DECLDIR nvmlDeviceGetSuppor
  * can concurrently run on a device.  For example, if only one vGPU type is allowed at a time on a device, then the creatable
  * list will be restricted to whatever vGPU type is already running on the device.
  *
- * If the supplied buffer is not large enough to accomodate the vGPU type array, the function returns
+ * If the supplied buffer is not large enough to accommodate the vGPU type array, the function returns
  * NVML_ERROR_INSUFFICIENT_SIZE, with the element count of nvmlVgpuTypeId_t array required in \a vgpuCount.
  * To query the number of vGPU types createable for the GPU, call this function with *vgpuCount = 0.
  * The code will return NVML_ERROR_INSUFFICIENT_SIZE, or NVML_SUCCESS if no vGPU types are creatable.
@@ -6011,7 +6011,7 @@ nvmlReturn_t DECLDIR nvmlVgpuTypeGetGpuI
  *
  * @param vgpuTypeId               Handle to vGPU type
  * @param deviceID                 Device ID and vendor ID of the device contained in single 32 bit value
- * @param subsystemID              Subsytem ID and subsytem vendor ID of the device contained in single 32 bit value
+ * @param subsystemID              Subsystem ID and subsystem vendor ID of the device contained in single 32 bit value
  *
  * @return
  *         - \ref NVML_SUCCESS                 successful completion
@@ -6153,7 +6153,7 @@ nvmlReturn_t DECLDIR nvmlVgpuTypeGetMaxI
  * array elememt count is passed in \a vgpuCount, and \a vgpuCount is used to return the number of vGPU instances
  * written to the buffer.
  *
- * If the supplied buffer is not large enough to accomodate the vGPU instance array, the function returns
+ * If the supplied buffer is not large enough to accommodate the vGPU instance array, the function returns
  * NVML_ERROR_INSUFFICIENT_SIZE, with the element count of nvmlVgpuInstance_t array required in \a vgpuCount.
  * To query the number of active vGPU instances, call this function with *vgpuCount = 0.  The code will return
  * NVML_ERROR_INSUFFICIENT_SIZE, or NVML_SUCCESS if no vGPU Types are supported.
@@ -6352,7 +6352,7 @@ nvmlReturn_t DECLDIR nvmlVgpuInstanceGet
  * @param encoderCapacity          Reference to an unsigned int for the encoder capacity
  *
  * @return
- *         - \ref NVML_SUCCESS                 if \a encoderCapacity has been retrived
+ *         - \ref NVML_SUCCESS                 if \a encoderCapacity has been retrieved
  *         - \ref NVML_ERROR_UNINITIALIZED     if the library has not been successfully initialized
  *         - \ref NVML_ERROR_INVALID_ARGUMENT  if \a vgpuInstance is 0, or \a encoderQueryType is invalid
  *         - \ref NVML_ERROR_NOT_FOUND         if \a vgpuInstance does not match a valid active vGPU instance on the system
@@ -6405,7 +6405,7 @@ nvmlReturn_t DECLDIR nvmlVgpuInstanceGet
  * array element count is passed in \a sessionCount, and \a sessionCount is used to return the number of sessions
  * written to the buffer.
  *
- * If the supplied buffer is not large enough to accomodate the active session array, the function returns
+ * If the supplied buffer is not large enough to accommodate the active session array, the function returns
  * NVML_ERROR_INSUFFICIENT_SIZE, with the element count of nvmlEncoderSessionInfo_t array required in \a sessionCount.
  * To query the number of active encoder sessions, call this function with *sessionCount = 0. The code will return
  * NVML_SUCCESS with number of active encoder sessions updated in *sessionCount.
@@ -6435,7 +6435,7 @@ nvmlReturn_t DECLDIR nvmlVgpuInstanceGet
 * For Maxwell &tm; or newer fully supported devices.
 *
 * @param vgpuInstance                      Identifier of the target vGPU instance
-* @param fbcStats                          Reference to nvmlFBCStats_t structure contianing NvFBC stats
+* @param fbcStats                          Reference to nvmlFBCStats_t structure containing NvFBC stats
 *
 * @return
 *         - \ref NVML_SUCCESS                  if \a fbcStats is fetched
@@ -6506,7 +6506,7 @@ typedef struct nvmlVgpuMetadata_st
     char                     guestDriverVersion[NVML_SYSTEM_DRIVER_VERSION_BUFFER_SIZE]; //!< Version of driver installed in guest
     char                     hostDriverVersion[NVML_SYSTEM_DRIVER_VERSION_BUFFER_SIZE];  //!< Version of driver installed in host
     unsigned int             reserved[6];                                                //!< Reserved for internal use
-    unsigned int             vgpuVirtualizationCaps;                                     //!< vGPU virtualizaion capabilities bitfileld
+    unsigned int             vgpuVirtualizationCaps;                                     //!< vGPU virtualization capabilities bitfield
     unsigned int             guestVgpuVersion;                                           //!< vGPU version of guest driver
     unsigned int             opaqueDataSize;                                             //!< Size of opaque data field in bytes
     char                     opaqueData[4];                                              //!< Opaque data
@@ -6520,7 +6520,7 @@ typedef struct nvmlVgpuPgpuMetadata_st
     unsigned int            version;                                                    //!< Current version of the structure
     unsigned int            revision;                                                   //!< Current revision of the structure
     char                    hostDriverVersion[NVML_SYSTEM_DRIVER_VERSION_BUFFER_SIZE];  //!< Host driver version
-    unsigned int            pgpuVirtualizationCaps;                                     //!< Pgpu virtualizaion capabilities bitfileld
+    unsigned int            pgpuVirtualizationCaps;                                     //!< Pgpu virtualization capabilities bitfield
     unsigned int            reserved[5];                                                //!< Reserved for internal use
     nvmlVgpuVersion_t       hostSupportedVgpuRange;                                     //!< vGPU version range supported by host driver
     unsigned int            opaqueDataSize;                                             //!< Size of opaque data field in bytes
@@ -6618,7 +6618,7 @@ nvmlReturn_t DECLDIR nvmlDeviceGetVgpuMe
  *
  * The caller passes in a buffer via \a compatibilityInfo, into which a compatibility information structure is written. The
  * structure defines the states in which the vGPU / VM may be booted on the physical GPU. If the vGPU / VM compatibility
- * with the physical GPU is limited, a limit code indicates the factor limiting compability.
+ * with the physical GPU is limited, a limit code indicates the factor limiting compatibility.
  * (see \ref nvmlVgpuPgpuCompatibilityLimitCode_t for details).
  *
  * Note: vGPU compatibility does not take into account dynamic capacity conditions that may limit a system's ability to
--- a/nvidia-cuda/cuda_cupti/extras/CUPTI/samples/openacc_trace/openacc_trace.cpp
+++ b/nvidia-cuda/cuda_cupti/extras/CUPTI/samples/openacc_trace/openacc_trace.cpp
@@ -108,7 +108,7 @@ void finalize()
 }
 
 // acc_register_library is defined by the OpenACC tools interface
-// and allows to register this library with the OpenACC runtime.
+// and allows one to register this library with the OpenACC runtime.
 
 extern "C" void
 acc_register_library(void *profRegister, void *profUnregister, void *profLookup)
--- a/nvidia-cuda/libcublas/include/cublasLt.h
+++ b/nvidia-cuda/libcublas/include/cublasLt.h
@@ -746,7 +746,7 @@ cublasLtMatrixTransformDescGetAttribute(
     size_t sizeInBytes,
     size_t *sizeWritten);
 
-/** For computation with complex numbers, this enum allows to apply the Gauss Complexity reduction algorithm
+/** For computation with complex numbers, this enum allows one to apply the Gauss Complexity reduction algorithm
  */
 typedef enum {
     CUBLASLT_3M_MODE_DISALLOWED = 0,
--- a/nvidia-cuda/libcusolver/include/cusolverMg.h
+++ b/nvidia-cuda/libcusolver/include/cusolverMg.h
@@ -92,7 +92,7 @@ cusolverStatus_t CUSOLVERAPI cusolverMgD
 
 /**
  * \brief Allocates resources related to the shared memory device grid.
- * \param[out] grid the opaque data strcuture that holds the grid
+ * \param[out] grid the opaque data structure that holds the grid
  * \param[in] numRowDevices number of devices in the row
  * \param[in] numColDevices number of devices in the column
  * \param[in] deviceId This array of size height * width stores the
@@ -109,7 +109,7 @@ cusolverStatus_t CUSOLVERAPI cusolverMgC
 
 /**
  * \brief Releases the allocated resources related to the distributed grid.
- * \param[in] grid the opaque data strcuture that holds the distributed grid
+ * \param[in] grid the opaque data structure that holds the distributed grid
  * \returns the status code
  */
 cusolverStatus_t CUSOLVERAPI cusolverMgDestroyGrid(
@@ -117,7 +117,7 @@ cusolverStatus_t CUSOLVERAPI cusolverMgD
 
 /**
  * \brief Allocates resources related to the distributed matrix descriptor.
- * \param[out] desc the opaque data strcuture that holds the descriptor
+ * \param[out] desc the opaque data structure that holds the descriptor
  * \param[in] numRows number of total rows
  * \param[in] numCols number of total columns
  * \param[in] rowBlockSize row block size
@@ -137,7 +137,7 @@ cusolverStatus_t CUSOLVERAPI cusolverMgC
 
 /**
  * \brief Releases the allocated resources related to the distributed matrix descriptor.
- * \param[in] desc the opaque data strcuture that holds the descriptor
+ * \param[in] desc the opaque data structure that holds the descriptor
  * \returns the status code
  */
 cusolverStatus_t CUSOLVERAPI cusolverMgDestroyMatrixDesc(
--- a/nvidia-cuda/cuda_cudart/include/cuda_egl_interop.h
+++ b/nvidia-cuda/cuda_cudart/include/cuda_egl_interop.h
@@ -360,7 +360,7 @@ extern __host__ cudaError_t CUDARTAPI cu
  *
  * Disconnect CUDA as a consumer to EGLStreamKHR.
  *
- * \param conn            - Conection to disconnect.
+ * \param conn            - Connection to disconnect.
  *
  * \return
  * ::cudaSuccess,
@@ -458,7 +458,7 @@ extern __host__ cudaError_t CUDARTAPI cu
  *
  * Disconnect CUDA as a producer to EGLStreamKHR.
  *
- * \param conn            - Conection to disconnect.
+ * \param conn            - Connection to disconnect.
  *
  * \return
  * ::cudaSuccess,