@@ -225,11 +225,11 @@ struct TestbedImpl {
225
225
//
226
226
// Methods
227
227
//
228
- template <class , class , class > class convert_fp8_to_fp16_name ;
228
+ template <class , class > class convert_fp8_to_fp16_name ;
229
229
230
- template <typename SrcT, typename DstT, typename Runner >
230
+ template <typename SrcT, typename DstT>
231
231
void convert_fp8_to_fp16 (const SrcT* d_src, DstT* d_dst, size_t size) {
232
- cutlasscompat::get_default_queue ().parallel_for <convert_fp8_to_fp16_name<SrcT, DstT, Runner >>(size, [=](auto indx) {
232
+ cutlasscompat::get_default_queue ().parallel_for <convert_fp8_to_fp16_name<SrcT, DstT>>(size, [=](auto indx) {
233
233
d_dst[indx] = static_cast <DstT>(d_src[indx]);
234
234
}).wait ();
235
235
}
@@ -241,7 +241,7 @@ struct TestbedImpl {
241
241
using outType = cute::conditional_t <is_fp8_v<Tin>, half_t , Tin>;
242
242
if constexpr (is_fp8_v<Tin>) {
243
243
cutlass::DeviceAllocation<outType> out (in.size ());
244
- convert_fp8_to_fp16<Tin, outType, TestbedImpl >(in.get (), out.get (), in.size ());
244
+ convert_fp8_to_fp16<Tin, outType>(in.get (), out.get (), in.size ());
245
245
return out;
246
246
} else {
247
247
return in;
@@ -625,7 +625,7 @@ struct TestbedImpl {
625
625
626
626
#if !defined(SYCL_EXT_ONEAPI_WORK_GROUP_SCRATCH_MEMORY)
627
627
using namespace cutlasscompat ::experimental;
628
- auto event = launch<cutlass::device_kernel<FlashAttention>, FlashAttention >(
628
+ auto event = launch<cutlass::device_kernel<FlashAttention>>(
629
629
launch_policy{sycl_grid, sycl_block, local_mem_size{static_cast <std::size_t >(smem_size)},
630
630
kernel_properties{sycl_exp::sub_group_size<FlashAttention::DispatchPolicy::SubgroupSize>}},
631
631
params);
@@ -680,8 +680,9 @@ template <
680
680
typename FlashAttention
681
681
>
682
682
struct Testbed3x {
683
- using TestBedImpl = typename detail::TestbedImpl<FlashAttention>;
684
- TestBedImpl impl_;
683
+ // using TestBedImp = typename detail::TestbedImpl<FlashAttention>;
684
+ // TestBedImp impl_;
685
+ detail::TestbedImpl<FlashAttention> impl_;
685
686
686
687
//
687
688
// Methods
0 commit comments