@@ -265,20 +265,20 @@ static ggml_cuda_device_info ggml_cuda_init() {
265265 id, prop.name , prop.major , prop.minor , device_vmm ? " yes" : " no" );
266266 std::string device_name (prop.name );
267267 if (device_name == " NVIDIA GeForce MX450" ) {
268- affected_devices .push_back ({ id, device_name });
268+ turing_devices_without_mma .push_back ({ id, device_name });
269269 } else if (device_name == " NVIDIA GeForce MX550" ) {
270- affected_devices .push_back ({ id, device_name });
270+ turing_devices_without_mma .push_back ({ id, device_name });
271271 } else if (device_name.substr (0 , 21 ) == " NVIDIA GeForce GTX 16" ) {
272- affected_devices .push_back ({ id, device_name });
272+ turing_devices_without_mma .push_back ({ id, device_name });
273273 }
274274#endif // defined(GGML_USE_HIP)
275275 }
276276
277277 if (ggml_cuda_highest_compiled_arch (GGML_CUDA_CC_TURING) >= GGML_CUDA_CC_TURING && !turing_devices_without_mma.empty ()) {
278278 GGML_LOG_INFO (" The following devices will have suboptimal performance due to a lack of tensor cores:\n " );
279- for (size_t affected_id = 0 ; affected_id < affected_devices .size (); affected_id ++) {
279+ for (size_t device_pos = 0 ; device_pos < turing_devices_without_mma .size (); device_pos ++) {
280280 GGML_LOG_INFO (
281- " Device %d: %s\n " , affected_devices[affected_id ].first , affected_devices[affected_id ].second .c_str ());
281+ " Device %d: %s\n " , turing_devices_without_mma[device_pos ].first , turing_devices_without_mma[device_pos ].second .c_str ());
282282 }
283283 GGML_LOG_INFO (
284284 " Consider compiling with CMAKE_CUDA_ARCHITECTURES=61-virtual;80-virtual and DGGML_CUDA_FORCE_MMQ to force the use of the Pascal code for Turing.\n " );
0 commit comments