Skip to content
Projects
Groups
Snippets
Help
This project
Loading...
Sign in / Register
Toggle navigation
T
tic
Overview
Overview
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
wenyuanbo
tic
Commits
836cf13a
Commit
836cf13a
authored
Oct 04, 2018
by
Pariksheet Pinjari
Committed by
Tianqi Chen
Oct 03, 2018
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Correction in documentation (#1810)
parent
9afde69b
Hide whitespace changes
Inline
Side-by-side
Showing
7 changed files
with
14 additions
and
14 deletions
+14
-14
tests/python/contrib/test_cblas.py
+1
-1
tests/python/contrib/test_cublas.py
+1
-1
tests/python/contrib/test_mps.py
+2
-2
tests/python/contrib/test_nnpack.py
+4
-4
tests/python/contrib/test_random.py
+3
-3
tests/python/contrib/test_rocblas.py
+1
-1
topi/include/topi/nn.h
+2
-2
No files found.
tests/python/contrib/test_cblas.py
View file @
836cf13a
...
@@ -18,7 +18,7 @@ def test_matmul_add():
...
@@ -18,7 +18,7 @@ def test_matmul_add():
print
(
"skip because
%
s is not enabled..."
%
target
)
print
(
"skip because
%
s is not enabled..."
%
target
)
return
return
if
not
tvm
.
get_global_func
(
"tvm.contrib.cblas.matmul"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.cblas.matmul"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
cpu
(
0
)
ctx
=
tvm
.
cpu
(
0
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
D
,
bias
],
target
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
D
,
bias
],
target
)
...
...
tests/python/contrib/test_cublas.py
View file @
836cf13a
...
@@ -16,7 +16,7 @@ def test_matmul_add():
...
@@ -16,7 +16,7 @@ def test_matmul_add():
print
(
"skip because
%
s is not enabled..."
%
target
)
print
(
"skip because
%
s is not enabled..."
%
target
)
return
return
if
not
tvm
.
get_global_func
(
"tvm.contrib.cublas.matmul"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.cublas.matmul"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
gpu
(
0
)
ctx
=
tvm
.
gpu
(
0
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
C
],
target
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
C
],
target
)
...
...
tests/python/contrib/test_mps.py
View file @
836cf13a
...
@@ -33,7 +33,7 @@ def test_matmul():
...
@@ -33,7 +33,7 @@ def test_matmul():
def
verify
(
A
,
B
,
D
,
s
,
target
=
"metal"
):
def
verify
(
A
,
B
,
D
,
s
,
target
=
"metal"
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.mps.matmul"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.mps.matmul"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
metal
(
0
)
ctx
=
tvm
.
metal
(
0
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
D
],
"metal"
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
D
],
"metal"
)
...
@@ -64,7 +64,7 @@ def test_conv2d():
...
@@ -64,7 +64,7 @@ def test_conv2d():
def
verify
(
A
,
B
,
C
,
target
=
"llvm"
):
def
verify
(
A
,
B
,
C
,
target
=
"llvm"
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.mps.conv2d"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.mps.conv2d"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
metal
(
0
)
ctx
=
tvm
.
metal
(
0
)
f
=
tvm
.
build
(
s1
,
[
A
,
B
,
C
],
"metal"
)
f
=
tvm
.
build
(
s1
,
[
A
,
B
,
C
],
"metal"
)
...
...
tests/python/contrib/test_nnpack.py
View file @
836cf13a
...
@@ -19,7 +19,7 @@ def test_fully_connected_output():
...
@@ -19,7 +19,7 @@ def test_fully_connected_output():
print
(
"skip because
%
s is not enabled..."
%
target
)
print
(
"skip because
%
s is not enabled..."
%
target
)
return
return
if
not
tvm
.
get_global_func
(
"tvm.contrib.nnpack.fully_connected_output"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.nnpack.fully_connected_output"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
cpu
(
0
)
ctx
=
tvm
.
cpu
(
0
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
D
,
bias
],
target
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
D
,
bias
],
target
)
...
@@ -49,7 +49,7 @@ def test_fully_connected_inference():
...
@@ -49,7 +49,7 @@ def test_fully_connected_inference():
print
(
"skip because
%
s is not enabled..."
%
target
)
print
(
"skip because
%
s is not enabled..."
%
target
)
return
return
if
not
tvm
.
get_global_func
(
"tvm.contrib.nnpack.fully_connected_inference"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.nnpack.fully_connected_inference"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
cpu
(
0
)
ctx
=
tvm
.
cpu
(
0
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
D
,
bias
],
target
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
D
,
bias
],
target
)
...
@@ -128,7 +128,7 @@ def test_convolution_inference():
...
@@ -128,7 +128,7 @@ def test_convolution_inference():
print
(
"skip because
%
s is not enabled..."
%
target
)
print
(
"skip because
%
s is not enabled..."
%
target
)
return
return
if
not
tvm
.
get_global_func
(
"tvm.contrib.nnpack.fully_connected_inference"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.nnpack.fully_connected_inference"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
cpu
(
0
)
ctx
=
tvm
.
cpu
(
0
)
f
=
tvm
.
build
(
s
,
[
data
,
kernel
,
bias
,
output
],
target
)
f
=
tvm
.
build
(
s
,
[
data
,
kernel
,
bias
,
output
],
target
)
...
@@ -173,7 +173,7 @@ def test_convolution_output():
...
@@ -173,7 +173,7 @@ def test_convolution_output():
print
(
"skip because
%
s is not enabled..."
%
target
)
print
(
"skip because
%
s is not enabled..."
%
target
)
return
return
if
not
tvm
.
get_global_func
(
"tvm.contrib.nnpack.fully_connected_inference"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.nnpack.fully_connected_inference"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
cpu
(
0
)
ctx
=
tvm
.
cpu
(
0
)
f
=
tvm
.
build
(
s
,
[
data
,
kernel
,
bias
,
output
],
target
)
f
=
tvm
.
build
(
s
,
[
data
,
kernel
,
bias
,
output
],
target
)
...
...
tests/python/contrib/test_random.py
View file @
836cf13a
...
@@ -13,7 +13,7 @@ def test_randint():
...
@@ -13,7 +13,7 @@ def test_randint():
print
(
"skip because
%
s is not enabled..."
%
target
)
print
(
"skip because
%
s is not enabled..."
%
target
)
return
return
if
not
tvm
.
get_global_func
(
"tvm.contrib.random.randint"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.random.randint"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
cpu
(
0
)
ctx
=
tvm
.
cpu
(
0
)
f
=
tvm
.
build
(
s
,
[
A
],
target
)
f
=
tvm
.
build
(
s
,
[
A
],
target
)
...
@@ -37,7 +37,7 @@ def test_uniform():
...
@@ -37,7 +37,7 @@ def test_uniform():
print
(
"skip because
%
s is not enabled..."
%
target
)
print
(
"skip because
%
s is not enabled..."
%
target
)
return
return
if
not
tvm
.
get_global_func
(
"tvm.contrib.random.uniform"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.random.uniform"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
cpu
(
0
)
ctx
=
tvm
.
cpu
(
0
)
f
=
tvm
.
build
(
s
,
[
A
],
target
)
f
=
tvm
.
build
(
s
,
[
A
],
target
)
...
@@ -61,7 +61,7 @@ def test_normal():
...
@@ -61,7 +61,7 @@ def test_normal():
print
(
"skip because
%
s is not enabled..."
%
target
)
print
(
"skip because
%
s is not enabled..."
%
target
)
return
return
if
not
tvm
.
get_global_func
(
"tvm.contrib.random.normal"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.random.normal"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
cpu
(
0
)
ctx
=
tvm
.
cpu
(
0
)
f
=
tvm
.
build
(
s
,
[
A
],
target
)
f
=
tvm
.
build
(
s
,
[
A
],
target
)
...
...
tests/python/contrib/test_rocblas.py
View file @
836cf13a
...
@@ -16,7 +16,7 @@ def test_matmul_add():
...
@@ -16,7 +16,7 @@ def test_matmul_add():
print
(
"skip because
%
s is not enabled..."
%
target
)
print
(
"skip because
%
s is not enabled..."
%
target
)
return
return
if
not
tvm
.
get_global_func
(
"tvm.contrib.rocblas.matmul"
,
True
):
if
not
tvm
.
get_global_func
(
"tvm.contrib.rocblas.matmul"
,
True
):
print
(
"skip because extern function is not avalable"
)
print
(
"skip because extern function is not ava
i
lable"
)
return
return
ctx
=
tvm
.
rocm
(
0
)
ctx
=
tvm
.
rocm
(
0
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
C
],
target
)
f
=
tvm
.
build
(
s
,
[
A
,
B
,
C
],
target
)
...
...
topi/include/topi/nn.h
View file @
836cf13a
...
@@ -64,7 +64,7 @@ inline tvm::Tensor relu(const tvm::Tensor& t,
...
@@ -64,7 +64,7 @@ inline tvm::Tensor relu(const tvm::Tensor& t,
* \param name The name of the operation
* \param name The name of the operation
* \param tag The tag to mark the operation
* \param tag The tag to mark the operation
*
*
* \return A Tensor whose op member is the relu operation
* \return A Tensor whose op member is the
leaky
relu operation
*/
*/
inline
tvm
::
Tensor
leaky_relu
(
const
tvm
::
Tensor
&
t
,
inline
tvm
::
Tensor
leaky_relu
(
const
tvm
::
Tensor
&
t
,
double
alpha
=
0
.
1
,
double
alpha
=
0
.
1
,
...
@@ -90,7 +90,7 @@ inline tvm::Tensor leaky_relu(const tvm::Tensor& t,
...
@@ -90,7 +90,7 @@ inline tvm::Tensor leaky_relu(const tvm::Tensor& t,
* \param name The name of the operation
* \param name The name of the operation
* \param tag The tag to mark the operation
* \param tag The tag to mark the operation
*
*
* \return A Tensor whose op member is the relu operation
* \return A Tensor whose op member is the
parametric
relu operation
*/
*/
inline
tvm
::
Tensor
prelu
(
const
tvm
::
Tensor
&
x
,
inline
tvm
::
Tensor
prelu
(
const
tvm
::
Tensor
&
x
,
const
tvm
::
Tensor
&
slope
,
const
tvm
::
Tensor
&
slope
,
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment