File tree Expand file tree Collapse file tree 1 file changed +0
-16
lines changed Expand file tree Collapse file tree 1 file changed +0
-16
lines changed Original file line number Diff line number Diff line change @@ -262,22 +262,6 @@ static bool fp16_mma_hardware_available(const int cc) {
262262 (GGML_CUDA_CC_IS_MTHREADS (cc) && cc >= GGML_CUDA_CC_QY2);
263263}
264264
265- static bool bf16_mma_hardware_available (const int cc) {
266- return (GGML_CUDA_CC_IS_NVIDIA (cc) && cc >= GGML_CUDA_CC_AMPERE) || GGML_CUDA_CC_IS_CDNA (cc) || cc >= GGML_CUDA_CC_RDNA3;
267- }
268-
269- static bool fp32_mma_hardware_available (const int cc) {
270- return GGML_CUDA_CC_IS_CDNA (cc);
271- }
272-
273- static bool bf16_mma_hardware_available (const int cc) {
274- return (GGML_CUDA_CC_IS_NVIDIA (cc) && cc >= GGML_CUDA_CC_AMPERE) || GGML_CUDA_CC_IS_CDNA (cc) || cc >= GGML_CUDA_CC_RDNA3;
275- }
276-
277- static bool fp32_mma_hardware_available (const int cc) {
278- return GGML_CUDA_CC_IS_CDNA (cc);
279- }
280-
281265// Volta technically had FP16 tensor cores but they work very differently compared to Turing and later.
282266static bool new_mma_available (const int cc) {
283267 return GGML_CUDA_CC_IS_NVIDIA (cc) && ggml_cuda_highest_compiled_arch (cc) >= GGML_CUDA_CC_TURING;
You can’t perform that action at this time.
0 commit comments