Skip to content
Projects
Groups
Snippets
Help
This project
Loading...
Sign in / Register
Toggle navigation
T
tic
Overview
Overview
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
wenyuanbo
tic
Commits
2f2170f4
Commit
2f2170f4
authored
Oct 27, 2017
by
masahi
Committed by
Tianqi Chen
Oct 26, 2017
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
add helpful message to topi test (#592)
parent
20144de2
Hide whitespace changes
Inline
Side-by-side
Showing
13 changed files
with
22 additions
and
0 deletions
+22
-0
topi/tests/python/test_topi_broadcast.py
+2
-0
topi/tests/python/test_topi_conv2d_hwcn.py
+1
-0
topi/tests/python/test_topi_conv2d_nchw.py
+1
-0
topi/tests/python/test_topi_conv2d_transpose_nchw.py
+1
-0
topi/tests/python/test_topi_dense.py
+1
-0
topi/tests/python/test_topi_depthwise_conv2d.py
+2
-0
topi/tests/python/test_topi_depthwise_conv2d_back_input.py
+1
-0
topi/tests/python/test_topi_depthwise_conv2d_back_weight.py
+1
-0
topi/tests/python/test_topi_pooling.py
+2
-0
topi/tests/python/test_topi_reduce.py
+1
-0
topi/tests/python/test_topi_relu.py
+1
-0
topi/tests/python/test_topi_softmax.py
+2
-0
topi/tests/python/test_topi_transform.py
+6
-0
No files found.
topi/tests/python/test_topi_broadcast.py
View file @
2f2170f4
...
...
@@ -12,6 +12,7 @@ def verify_broadcast_to_ele(in_shape, out_shape):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_broadcast
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
@@ -52,6 +53,7 @@ def verify_broadcast_binary_ele(lhs_shape, rhs_shape, typ="add"):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_broadcast
(
C
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
topi/tests/python/test_topi_conv2d_hwcn.py
View file @
2f2170f4
...
...
@@ -34,6 +34,7 @@ def verify_conv2d_hwcn(batch, in_channel, in_size, num_filter, kernel, stride, p
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
ctx
=
tvm
.
context
(
device
,
0
)
a
=
tvm
.
nd
.
array
(
a_np
,
ctx
)
w
=
tvm
.
nd
.
array
(
w_np
,
ctx
)
...
...
topi/tests/python/test_topi_conv2d_nchw.py
View file @
2f2170f4
...
...
@@ -33,6 +33,7 @@ def verify_conv2d_nchw(batch, in_channel, in_size, num_filter, kernel, stride, p
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s1
=
topi
.
generic
.
schedule_conv2d_nchw
([
B
])
s2
=
topi
.
generic
.
schedule_conv2d_nchw
([
C
])
...
...
topi/tests/python/test_topi_conv2d_transpose_nchw.py
View file @
2f2170f4
...
...
@@ -32,6 +32,7 @@ def verify_conv2d_transpose_nchw(batch, in_channel, in_size, num_filter, kernel,
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s1
=
topi
.
generic
.
schedule_conv2d_transpose_nchw
([
B
])
s2
=
topi
.
generic
.
schedule_conv2d_transpose_nchw
([
C
])
...
...
topi/tests/python/test_topi_dense.py
View file @
2f2170f4
...
...
@@ -32,6 +32,7 @@ def verify_dense(batch, in_dim, out_dim, use_bias=True):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_dense
(
D
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
topi/tests/python/test_topi_depthwise_conv2d.py
View file @
2f2170f4
...
...
@@ -27,6 +27,7 @@ def depthwise_conv2d_with_workload_nchw(batch, in_channel, in_height, channel_mu
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
# schedule
s1
=
topi
.
generic
.
schedule_depthwise_conv2d_nchw
(
DepthwiseConv2d
)
...
...
@@ -111,6 +112,7 @@ def depthwise_conv2d_with_workload_nhwc(batch, in_channel, in_height, channel_mu
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s1
=
topi
.
generic
.
schedule_depthwise_conv2d_nhwc
(
DepthwiseConv2d
)
...
...
topi/tests/python/test_topi_depthwise_conv2d_back_input.py
View file @
2f2170f4
...
...
@@ -35,6 +35,7 @@ def verify_depthwise_conv2d_back_input(batch, in_channel, in_h, channel_multipli
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
ctx
=
tvm
.
context
(
device
,
0
)
# build the kernel
f
=
tvm
.
build
(
schedule
,
[
Filter
,
Out_grad
,
In_grad
],
device
)
...
...
topi/tests/python/test_topi_depthwise_conv2d_back_weight.py
View file @
2f2170f4
...
...
@@ -35,6 +35,7 @@ def verify_depthwise_conv2d_back_weight(batch, in_channel, in_h, channel_multipl
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
ctx
=
tvm
.
context
(
device
,
0
)
# build the kernel
f
=
tvm
.
build
(
schedule
,
[
Input
,
Out_grad
,
Weight_grad
],
device
)
...
...
topi/tests/python/test_topi_pooling.py
View file @
2f2170f4
...
...
@@ -35,6 +35,7 @@ def verify_pool(n, ic, ih, kh, sh, padding, pool_type):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_pool
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
@@ -70,6 +71,7 @@ def verify_global_pool(n, c, h, w, pool_type):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_global_pool
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
topi/tests/python/test_topi_reduce.py
View file @
2f2170f4
...
...
@@ -50,6 +50,7 @@ def verify_reduce_map_ele(in_shape, axis, keepdims, type="sum"):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_reduce
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
topi/tests/python/test_topi_relu.py
View file @
2f2170f4
...
...
@@ -16,6 +16,7 @@ def verify_relu(m, n):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_elemwise
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
topi/tests/python/test_topi_softmax.py
View file @
2f2170f4
...
...
@@ -20,6 +20,7 @@ def verify_softmax(m, n):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_softmax
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
@@ -50,6 +51,7 @@ def verify_log_softmax(m, n):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_softmax
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
topi/tests/python/test_topi_transform.py
View file @
2f2170f4
...
...
@@ -10,6 +10,7 @@ def verify_expand_dims(in_shape, out_shape, axis, num_newaxis):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_broadcast
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
@@ -32,6 +33,7 @@ def verify_tranpose(in_shape, axes):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_injective
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
@@ -54,6 +56,7 @@ def verify_reshape(src_shape, dst_shape):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_injective
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
@@ -76,6 +79,7 @@ def verify_squeeze(src_shape, axis):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_injective
(
B
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
@@ -103,6 +107,7 @@ def verify_concatenate(shapes, axis):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_injective
(
out_tensor
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
@@ -125,6 +130,7 @@ def verify_split(src_shape, indices_or_sections, axis):
if
not
tvm
.
module
.
enabled
(
device
):
print
(
"Skip because
%
s is not enabled"
%
device
)
return
print
(
"Running on target:
%
s"
%
device
)
with
tvm
.
target
.
create
(
device
):
s
=
topi
.
generic
.
schedule_injective
(
tensor_l
)
ctx
=
tvm
.
context
(
device
,
0
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment