@@ -36,7 +36,8 @@ if(CMAKE_BUILD_TESTS)
36
36
include (CTest)
37
37
38
38
if (ENABLE_CUDA)
39
- # 1a. Check the PyTorch model runs on a CUDA device and its outputs meet expectations
39
+ # 1a. Check the PyTorch model runs on a CUDA device and its outputs meet
40
+ # expectations
40
41
add_test (NAME simplenet
41
42
COMMAND ${Python_EXECUTABLE} ${PROJECT_SOURCE_DIR} /simplenet.py
42
43
--device_type cuda)
@@ -49,16 +50,16 @@ if(CMAKE_BUILD_TESTS)
49
50
cuda --filepath ${PROJECT_BINARY_DIR}
50
51
WORKING_DIRECTORY ${PROJECT_BINARY_DIR} )
51
52
52
- # 3a. Check the model can be loaded from file and run on two CUDA devices in Python and that its
53
- # outputs meet expectations
53
+ # 3a. Check the model can be loaded from file and run on two CUDA devices
54
+ # in Python and that its outputs meet expectations
54
55
add_test (
55
56
NAME multigpu_infer_python
56
57
COMMAND ${Python_EXECUTABLE}
57
58
${PROJECT_SOURCE_DIR} /multigpu_infer_python.py --device_type cuda
58
59
WORKING_DIRECTORY ${PROJECT_BINARY_DIR} )
59
60
60
- # 4a. Check the model can be loaded from file and run on two CUDA devices in Fortran and that
61
- # its outputs meet expectations
61
+ # 4a. Check the model can be loaded from file and run on two CUDA devices
62
+ # in Fortran and that its outputs meet expectations
62
63
add_test (
63
64
NAME multigpu_infer_fortran
64
65
COMMAND multigpu_infer_fortran cuda
@@ -71,7 +72,8 @@ if(CMAKE_BUILD_TESTS)
71
72
endif ()
72
73
73
74
if (ENABLE_XPU)
74
- # 1b. Check the PyTorch model runs on an XPU device and its outputs meet expectations
75
+ # 1b. Check the PyTorch model runs on an XPU device and its outputs meet
76
+ # expectations
75
77
add_test (NAME simplenet
76
78
COMMAND ${Python_EXECUTABLE} ${PROJECT_SOURCE_DIR} /simplenet.py
77
79
--device_type xpu)
@@ -84,16 +86,16 @@ if(CMAKE_BUILD_TESTS)
84
86
xpu --filepath ${PROJECT_BINARY_DIR}
85
87
WORKING_DIRECTORY ${PROJECT_BINARY_DIR} )
86
88
87
- # 3b. Check the model can be loaded from file and run on two XPU devices in Python and that its
88
- # outputs meet expectations
89
+ # 3b. Check the model can be loaded from file and run on two XPU devices
90
+ # in Python and that its outputs meet expectations
89
91
add_test (
90
92
NAME multigpu_infer_python
91
93
COMMAND ${Python_EXECUTABLE}
92
94
${PROJECT_SOURCE_DIR} /multigpu_infer_python.py --device_type xpu
93
95
WORKING_DIRECTORY ${PROJECT_BINARY_DIR} )
94
96
95
- # 4b. Check the model can be loaded from file and run on two XPU devices in Fortran and that
96
- # its outputs meet expectations
97
+ # 4b. Check the model can be loaded from file and run on two XPU devices
98
+ # in Fortran and that its outputs meet expectations
97
99
add_test (
98
100
NAME multigpu_infer_fortran
99
101
COMMAND multigpu_infer_fortran xpu
@@ -106,7 +108,8 @@ if(CMAKE_BUILD_TESTS)
106
108
endif ()
107
109
108
110
if (ENABLE_MPS)
109
- # 1c. Check the PyTorch model runs on an MPS device and its outputs meet expectations
111
+ # 1c. Check the PyTorch model runs on an MPS device and its outputs meet
112
+ # expectations
110
113
add_test (NAME simplenet
111
114
COMMAND ${Python_EXECUTABLE} ${PROJECT_SOURCE_DIR} /simplenet.py
112
115
--device_type mps)
0 commit comments