@@ -51,18 +51,38 @@ static int ColsFromArg(int arg) { return (arg % kRows); }
5151 BENCHMARK (BM_##DEVICE##_##FUNC##_##TYPE)->Range (4 << 10 , 1 << 20 );
5252
5353BM_UNARY (cpu, Floor, float , DT_FLOAT);
54+ #if GOOGLE_CUDA
5455BM_UNARY (gpu, Floor, float , DT_FLOAT);
56+ #endif // GOOGLE_CUDA
57+ #ifdef TENSORFLOW_USE_SYCL
58+ BM_UNARY (sycl, Floor, float , DT_FLOAT);
59+ #endif // TENSORFLOW_USE_SYCL
60+
5561BM_UNARY (cpu, Floor, double , DT_DOUBLE);
62+ #if GOOGLE_CUDA
5663BM_UNARY (gpu, Floor, double , DT_DOUBLE);
64+ #endif // GOOGLE_CUDA
65+ #ifdef TENSORFLOW_USE_SYCL
66+ BM_UNARY (sycl, Floor, double , DT_DOUBLE);
67+ #endif // TENSORFLOW_USE_SYCL
68+
5769BM_UNARY (cpu, Conj, std::complex <float >, DT_COMPLEX64);
70+ #if GOOGLE_CUDA
5871BM_UNARY (gpu, Conj, std::complex <float >, DT_COMPLEX64);
72+ #endif // GOOGLE_CUDA
5973BM_UNARY (cpu, Conj, std::complex <double >, DT_COMPLEX128);
74+ #if GOOGLE_CUDA
6075BM_UNARY (gpu, Conj, std::complex <double >, DT_COMPLEX128);
76+ #endif // GOOGLE_CUDA
6177
6278BM_UNARY (cpu, Rint, double , DT_DOUBLE);
79+ #if GOOGLE_CUDA
6380BM_UNARY (gpu, Rint, double , DT_DOUBLE);
81+ #endif // GOOGLE_CUDA
6482BM_UNARY (cpu, Rint, float , DT_FLOAT);
83+ #if GOOGLE_CUDA
6584BM_UNARY (gpu, Rint, float , DT_FLOAT);
85+ #endif // GOOGLE_CUDA
6686
6787// data func scalar.
6888static Graph* BinaryScalar (int num, const string& func) {
@@ -90,9 +110,20 @@ static Graph* BinaryScalar(int num, const string& func) {
90110 ->Arg(1048576 );
91111
92112BM_BINARY_SCALAR (cpu, Less);
113+ #if GOOGLE_CUDA
93114BM_BINARY_SCALAR (gpu, Less);
115+ #endif // GOOGLE_CUDA
116+ #ifdef TENSORFLOW_USE_SYCL
117+ BM_BINARY_SCALAR (sycl, Less);
118+ #endif // TENSORFLOW_USE_SYCL
119+
94120BM_BINARY_SCALAR (cpu, Add);
121+ #if GOOGLE_CUDA
95122BM_BINARY_SCALAR (gpu, Add);
123+ #endif // GOOGLE_CUDA
124+ #ifdef TENSORFLOW_USE_SYCL
125+ BM_BINARY_SCALAR (sycl, Add);
126+ #endif // TENSORFLOW_USE_SYCL
96127#undef BM_BINARY_SCALAR
97128
98129template <class T >
@@ -130,9 +161,13 @@ static Graph* BiasAdd(int rows, int cols, DataType type) {
130161
131162using Eigen::half;
132163BM_BIAS_ADD_ALL (cpu, float , DT_FLOAT);
164+ #if GOOGLE_CUDA
133165BM_BIAS_ADD_ALL (gpu, float , DT_FLOAT);
166+ #endif // GOOGLE_CUDA
134167BM_BIAS_ADD_ALL (cpu, half, DT_HALF);
168+ #if GOOGLE_CUDA
135169BM_BIAS_ADD_ALL (gpu, half, DT_HALF);
170+ #endif // GOOGLE_CUDA
136171#undef BM_BIAS_ADD_ALL
137172#undef BM_BIAS_ADD
138173
@@ -180,12 +215,18 @@ static Graph* BiasAddGrad(int rows, int cols, int channels, DataType type,
180215 BM_BIAS_ADD_GRAD (DEVICE, FORMAT, C_TYPE, TF_TYPE, 4096 , 4096 , 1 );
181216
182217using Eigen::half;
218+ #if GOOGLE_CUDA
183219BM_BIAS_ADD_GRAD_ALL (gpu, NCHW, float , DT_FLOAT);
184220BM_BIAS_ADD_GRAD_ALL (gpu, NCHW, half, DT_HALF);
221+ #endif // GOOGLE_CUDA
185222BM_BIAS_ADD_GRAD_ALL (cpu, NHWC, float , DT_FLOAT);
223+ #if GOOGLE_CUDA
186224BM_BIAS_ADD_GRAD_ALL (gpu, NHWC, float , DT_FLOAT);
225+ #endif // GOOGLE_CUDA
187226BM_BIAS_ADD_GRAD_ALL (cpu, NHWC, half, DT_HALF);
227+ #if GOOGLE_CUDA
188228BM_BIAS_ADD_GRAD_ALL (gpu, NHWC, half, DT_HALF);
229+ #endif // GOOGLE_CUDA
189230#undef BM_BIAS_ADD_GRAD_ALL
190231#undef BM_BIAS_ADD_GRAD
191232
@@ -223,7 +264,12 @@ static Graph* BcastAdd(int rows, int cols, int dim) {
223264 BM_BCAST_ADD_ROW (DEVICE, 2048 , 512 ); \
224265 BM_BCAST_ADD_ROW (DEVICE, 4096 , 512 );
225266BM_BCAST_ADD_ROW_ALL (cpu);
267+ #if GOOGLE_CUDA
226268BM_BCAST_ADD_ROW_ALL (gpu);
269+ #endif // GOOGLE_CUDA
270+ #ifdef TENSORFLOW_USE_SYCL
271+ BM_BCAST_ADD_ROW_ALL (sycl);
272+ #endif // TENSORFLOW_USE_SYCL
227273#undef BM_BCAST_ADD_ROW_ALL
228274#undef BM_BCAST_ADD_ROW
229275
@@ -244,7 +290,12 @@ BM_BCAST_ADD_ROW_ALL(gpu);
244290 BM_BCAST_ADD_COL (DEVICE, 2048 , 512 ); \
245291 BM_BCAST_ADD_COL (DEVICE, 4096 , 512 );
246292BM_BCAST_ADD_COL_ALL (cpu);
293+ #if GOOGLE_CUDA
247294BM_BCAST_ADD_COL_ALL (gpu);
295+ #endif // GOOGLE_CUDA
296+ #ifdef TENSORFLOW_USE_SYCL
297+ BM_BCAST_ADD_COL_ALL (sycl);
298+ #endif // TENSORFLOW_USE_SYCL
248299#undef BM_BCAST_ADD_COL_ALL
249300#undef BM_BCAST_ADD_COL
250301
0 commit comments