@@ -5869,7 +5869,7 @@ void ggml_init_cublas() {
5869
5869
}
5870
5870
} else {
5871
5871
g_cudaMemPools[id] = nullptr ;
5872
- fprintf (stderr, " , CUDA memory pool is not supported (cant load default pool)\n " );
5872
+ fprintf (stderr, " , CUDA memory pool is not supported (can't load default pool)\n " );
5873
5873
}
5874
5874
// test alloc/dealoc
5875
5875
if (err == cudaSuccess) {
@@ -5887,6 +5887,8 @@ void ggml_init_cublas() {
5887
5887
fprintf (stderr, " , CUDA memory pool is not supported (allocation failed)\n " );
5888
5888
}
5889
5889
}
5890
+ } else {
5891
+ fprintf (stderr, " , CUDA memory pool is not supported\n " );
5890
5892
}
5891
5893
#endif
5892
5894
g_tensor_split[id] = total_vram;
@@ -5915,6 +5917,7 @@ void ggml_init_cublas() {
5915
5917
fprintf (stderr,
5916
5918
" Warning: Device %d doesnt support CUDA memory pool, skipping pool access config\n " ,
5917
5919
id);
5920
+ continue ;
5918
5921
}
5919
5922
5920
5923
cudaMemAccessDesc desc_device = {};
@@ -5923,13 +5926,13 @@ void ggml_init_cublas() {
5923
5926
desc_device.flags = cudaMemAccessFlagsProtReadWrite;
5924
5927
cudaError_t err = cudaMemPoolSetAccess (main_device_pool, &desc_device, 1 /* numDescs */ );
5925
5928
if (err != cudaSuccess) {
5926
- fprintf (stderr, " Cant give access for main device memory pool to device %d\n " , id);
5929
+ fprintf (stderr, " Can't give access for main device memory pool to device %d\n " , id);
5927
5930
}
5928
5931
cudaMemPool_t mempool;
5929
5932
CUDA_CHECK (cudaDeviceGetDefaultMemPool (&mempool, id));
5930
5933
err = cudaMemPoolSetAccess (mempool, &desc_main_device, 1 /* numDescs */ );
5931
5934
if (err != cudaSuccess) {
5932
- fprintf (stderr, " Cant give access for device %d memory pool to main device \n " , id);
5935
+ fprintf (stderr, " Can't give access for device %d memory pool to main device \n " , id);
5933
5936
}
5934
5937
}
5935
5938
} else {
0 commit comments