Skip to content
Projects
Groups
Snippets
Help
This project
Loading...
Sign in / Register
Toggle navigation
T
tic
Overview
Overview
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
wenyuanbo
tic
Commits
c6339730
Commit
c6339730
authored
Mar 29, 2019
by
Haichen Shen
Committed by
Tianqi Chen
Mar 30, 2019
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
[Tutorial] Cache the test data in tutorial (#2923)
parent
8eef1565
Show whitespace changes
Inline
Side-by-side
Showing
25 changed files
with
170 additions
and
252 deletions
+170
-252
nnvm/tests/python/frontend/coreml/model_zoo/__init__.py
+1
-1
nnvm/tests/python/frontend/darknet/test_forward.py
+1
-1
nnvm/tutorials/deploy_model_on_mali_gpu.py
+7
-6
nnvm/tutorials/deploy_model_on_rasp.py
+7
-6
nnvm/tutorials/deploy_ssd_mxnet.py
+10
-14
nnvm/tutorials/from_coreml.py
+8
-20
nnvm/tutorials/from_darknet.py
+11
-13
nnvm/tutorials/from_mxnet.py
+8
-7
nnvm/tutorials/from_mxnet_to_webgl.py
+6
-8
nnvm/tutorials/from_onnx.py
+5
-17
nnvm/tutorials/from_tensorflow.py
+16
-16
nnvm/tutorials/nlp/from_darknet_rnn.py
+6
-8
nnvm/tutorials/nlp/keras_s2s_translate.py
+5
-5
python/tvm/contrib/download.py
+3
-0
python/tvm/relay/testing/tf.py
+1
-1
tests/python/frontend/coreml/model_zoo/__init__.py
+2
-1
tutorials/frontend/deploy_model_on_rasp.py
+7
-6
tutorials/frontend/deploy_ssd_gluoncv.py
+3
-2
tutorials/frontend/from_caffe2.py
+6
-20
tutorials/frontend/from_coreml.py
+8
-20
tutorials/frontend/from_keras.py
+8
-20
tutorials/frontend/from_mxnet.py
+7
-6
tutorials/frontend/from_onnx.py
+5
-17
tutorials/frontend/from_tensorflow.py
+16
-16
tutorials/frontend/from_tflite.py
+13
-21
No files found.
nnvm/tests/python/frontend/coreml/model_zoo/__init__.py
View file @
c6339730
...
@@ -18,7 +18,7 @@ def get_resnet50():
...
@@ -18,7 +18,7 @@ def get_resnet50():
def
get_cat_image
():
def
get_cat_image
():
url
=
'https://gist.githubusercontent.com/zhreshold/bcda4716699ac97ea44f791c24310193/raw/fa7ef0e9c9a5daea686d6473a62aacd1a5885849/cat.png'
url
=
'https://gist.githubusercontent.com/zhreshold/bcda4716699ac97ea44f791c24310193/raw/fa7ef0e9c9a5daea686d6473a62aacd1a5885849/cat.png'
dst
=
'cat.png'
dst
=
'cat.png'
real_dst
=
download_testdata
(
url
,
dst
,
module
=
'
coreml
'
)
real_dst
=
download_testdata
(
url
,
dst
,
module
=
'
data
'
)
img
=
Image
.
open
(
real_dst
)
.
resize
((
224
,
224
))
img
=
Image
.
open
(
real_dst
)
.
resize
((
224
,
224
))
img
=
np
.
transpose
(
img
,
(
2
,
0
,
1
))[
np
.
newaxis
,
:]
img
=
np
.
transpose
(
img
,
(
2
,
0
,
1
))[
np
.
newaxis
,
:]
return
np
.
asarray
(
img
)
return
np
.
asarray
(
img
)
nnvm/tests/python/frontend/darknet/test_forward.py
View file @
c6339730
...
@@ -103,7 +103,7 @@ def test_forward(net, build_dtype='float32'):
...
@@ -103,7 +103,7 @@ def test_forward(net, build_dtype='float32'):
test_image
=
'dog.jpg'
test_image
=
'dog.jpg'
img_url
=
'https://github.com/siju-samuel/darknet/blob/master/data/'
+
test_image
+
'?raw=true'
img_url
=
'https://github.com/siju-samuel/darknet/blob/master/data/'
+
test_image
+
'?raw=true'
img_path
=
download_testdata
(
img_url
,
test_image
,
module
=
'da
rknet
'
)
img_path
=
download_testdata
(
img_url
,
test_image
,
module
=
'da
ta
'
)
img
=
LIB
.
letterbox_image
(
LIB
.
load_image_color
(
img_path
.
encode
(
'utf-8'
),
0
,
0
),
net
.
w
,
net
.
h
)
img
=
LIB
.
letterbox_image
(
LIB
.
load_image_color
(
img_path
.
encode
(
'utf-8'
),
0
,
0
),
net
.
w
,
net
.
h
)
darknet_output
=
get_darknet_output
(
net
,
img
)
darknet_output
=
get_darknet_output
(
net
,
img
)
batch_size
=
1
batch_size
=
1
...
...
nnvm/tutorials/deploy_model_on_mali_gpu.py
View file @
c6339730
...
@@ -15,6 +15,7 @@ import nnvm.compiler
...
@@ -15,6 +15,7 @@ import nnvm.compiler
import
nnvm.testing
import
nnvm.testing
from
tvm
import
rpc
from
tvm
import
rpc
from
tvm.contrib
import
util
,
graph_runtime
as
runtime
from
tvm.contrib
import
util
,
graph_runtime
as
runtime
from
tvm.contrib.download
import
download_testdata
######################################################################
######################################################################
# Build TVM Runtime on Device
# Build TVM Runtime on Device
...
@@ -81,7 +82,6 @@ from tvm.contrib import util, graph_runtime as runtime
...
@@ -81,7 +82,6 @@ from tvm.contrib import util, graph_runtime as runtime
# You can found more details about this part at tutorial :ref:`tutorial-from-mxnet`.
# You can found more details about this part at tutorial :ref:`tutorial-from-mxnet`.
from
mxnet.gluon.model_zoo.vision
import
get_model
from
mxnet.gluon.model_zoo.vision
import
get_model
from
mxnet.gluon.utils
import
download
from
PIL
import
Image
from
PIL
import
Image
import
numpy
as
np
import
numpy
as
np
...
@@ -92,8 +92,9 @@ block = get_model('resnet18_v1', pretrained=True)
...
@@ -92,8 +92,9 @@ block = get_model('resnet18_v1', pretrained=True)
# In order to test our model, here we download an image of cat and
# In order to test our model, here we download an image of cat and
# transform its format.
# transform its format.
img_name
=
'cat.png'
img_name
=
'cat.png'
download
(
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
,
img_name
)
img_path
=
download_testdata
(
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
,
image
=
Image
.
open
(
img_name
)
.
resize
((
224
,
224
))
img_name
,
module
=
'data'
)
image
=
Image
.
open
(
img_path
)
.
resize
((
224
,
224
))
def
transform_image
(
image
):
def
transform_image
(
image
):
image
=
np
.
array
(
image
)
-
np
.
array
([
123.
,
117.
,
104.
])
image
=
np
.
array
(
image
)
-
np
.
array
([
123.
,
117.
,
104.
])
...
@@ -112,9 +113,9 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
...
@@ -112,9 +113,9 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'imagenet1000_clsid_to_human.txt'
])
'imagenet1000_clsid_to_human.txt'
])
synset_name
=
'
synset
.txt'
synset_name
=
'
imagenet1000_clsid_to_human
.txt'
download
(
synset_url
,
synset_name
)
synset_path
=
download_testdata
(
synset_url
,
synset_name
,
module
=
'data'
)
with
open
(
synset_
name
)
as
f
:
with
open
(
synset_
path
)
as
f
:
synset
=
eval
(
f
.
read
())
synset
=
eval
(
f
.
read
())
######################################################################
######################################################################
...
...
nnvm/tutorials/deploy_model_on_rasp.py
View file @
c6339730
...
@@ -14,6 +14,7 @@ import nnvm.compiler
...
@@ -14,6 +14,7 @@ import nnvm.compiler
import
nnvm.testing
import
nnvm.testing
from
tvm
import
rpc
from
tvm
import
rpc
from
tvm.contrib
import
util
,
graph_runtime
as
runtime
from
tvm.contrib
import
util
,
graph_runtime
as
runtime
from
tvm.contrib.download
import
download_testdata
######################################################################
######################################################################
# .. _build-tvm-runtime-on-device:
# .. _build-tvm-runtime-on-device:
...
@@ -78,7 +79,6 @@ from tvm.contrib import util, graph_runtime as runtime
...
@@ -78,7 +79,6 @@ from tvm.contrib import util, graph_runtime as runtime
# You can found more details about this part at tutorial :ref:`tutorial-from-mxnet`.
# You can found more details about this part at tutorial :ref:`tutorial-from-mxnet`.
from
mxnet.gluon.model_zoo.vision
import
get_model
from
mxnet.gluon.model_zoo.vision
import
get_model
from
mxnet.gluon.utils
import
download
from
PIL
import
Image
from
PIL
import
Image
import
numpy
as
np
import
numpy
as
np
...
@@ -89,8 +89,9 @@ block = get_model('resnet18_v1', pretrained=True)
...
@@ -89,8 +89,9 @@ block = get_model('resnet18_v1', pretrained=True)
# In order to test our model, here we download an image of cat and
# In order to test our model, here we download an image of cat and
# transform its format.
# transform its format.
img_name
=
'cat.png'
img_name
=
'cat.png'
download
(
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
,
img_name
)
img_path
=
download_testdata
(
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
,
image
=
Image
.
open
(
img_name
)
.
resize
((
224
,
224
))
img_name
,
module
=
'data'
)
image
=
Image
.
open
(
img_path
)
.
resize
((
224
,
224
))
def
transform_image
(
image
):
def
transform_image
(
image
):
image
=
np
.
array
(
image
)
-
np
.
array
([
123.
,
117.
,
104.
])
image
=
np
.
array
(
image
)
-
np
.
array
([
123.
,
117.
,
104.
])
...
@@ -108,9 +109,9 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
...
@@ -108,9 +109,9 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'imagenet1000_clsid_to_human.txt'
])
'imagenet1000_clsid_to_human.txt'
])
synset_name
=
'
synset
.txt'
synset_name
=
'
imagenet1000_clsid_to_human
.txt'
download
(
synset_url
,
synset_name
)
synset_path
=
download_testdata
(
synset_url
,
synset_name
,
module
=
'data'
)
with
open
(
synset_
name
)
as
f
:
with
open
(
synset_
path
)
as
f
:
synset
=
eval
(
f
.
read
())
synset
=
eval
(
f
.
read
())
######################################################################
######################################################################
...
...
nnvm/tutorials/deploy_ssd_mxnet.py
View file @
c6339730
...
@@ -18,7 +18,7 @@ import numpy as np
...
@@ -18,7 +18,7 @@ import numpy as np
from
nnvm
import
compiler
from
nnvm
import
compiler
from
nnvm.frontend
import
from_mxnet
from
nnvm.frontend
import
from_mxnet
from
tvm
import
relay
from
tvm
import
relay
from
tvm.contrib.download
import
download
from
tvm.contrib.download
import
download
_testdata
from
tvm.contrib
import
graph_runtime
from
tvm.contrib
import
graph_runtime
from
mxnet.model
import
load_checkpoint
from
mxnet.model
import
load_checkpoint
...
@@ -65,28 +65,24 @@ inference_symbol_folder = \
...
@@ -65,28 +65,24 @@ inference_symbol_folder = \
inference_symbol_url
=
"https://gist.github.com/kevinthesun/c1904e900848df4548ce5dfb18c719c7/"
\
inference_symbol_url
=
"https://gist.github.com/kevinthesun/c1904e900848df4548ce5dfb18c719c7/"
\
"archive/a28c4856c827fe766aa3da0e35bad41d44f0fb26.zip"
"archive/a28c4856c827fe766aa3da0e35bad41d44f0fb26.zip"
dir
=
"ssd_model"
model_file_path
=
download_testdata
(
model_url
,
model_file
,
module
=
[
"mxnet"
,
"ssd_model"
])
if
not
os
.
path
.
exists
(
dir
):
inference_symbol_path
=
download_testdata
(
inference_symbol_url
,
"inference_model.zip"
,
os
.
makedirs
(
dir
)
module
=
[
"mxnet"
,
"ssd_model"
])
model_file_path
=
"
%
s/
%
s"
%
(
dir
,
model_file
)
test_image_path
=
download_testdata
(
image_url
,
test_image
,
module
=
"data"
)
test_image_path
=
"
%
s/
%
s"
%
(
dir
,
test_image
)
model_dir
=
os
.
path
.
dirname
(
model_file_path
)
inference_symbol_path
=
"
%
s/inference_model.zip"
%
dir
download
(
model_url
,
model_file_path
)
download
(
image_url
,
test_image_path
)
download
(
inference_symbol_url
,
inference_symbol_path
)
zip_ref
=
zipfile
.
ZipFile
(
model_file_path
,
'r'
)
zip_ref
=
zipfile
.
ZipFile
(
model_file_path
,
'r'
)
zip_ref
.
extractall
(
dir
)
zip_ref
.
extractall
(
model_
dir
)
zip_ref
.
close
()
zip_ref
.
close
()
zip_ref
=
zipfile
.
ZipFile
(
inference_symbol_path
)
zip_ref
=
zipfile
.
ZipFile
(
inference_symbol_path
)
zip_ref
.
extractall
(
dir
)
zip_ref
.
extractall
(
model_
dir
)
zip_ref
.
close
()
zip_ref
.
close
()
######################################################################
######################################################################
# Convert and compile model with NNVM or Relay for CPU.
# Convert and compile model with NNVM or Relay for CPU.
sym
=
mx
.
sym
.
load
(
"
%
s/
%
s/ssd_resnet50_inference.json"
%
(
dir
,
inference_symbol_folder
))
sym
=
mx
.
sym
.
load
(
"
%
s/
%
s/ssd_resnet50_inference.json"
%
(
model_
dir
,
inference_symbol_folder
))
_
,
arg_params
,
aux_params
=
load_checkpoint
(
"
%
s/
%
s"
%
(
dir
,
model_name
),
0
)
_
,
arg_params
,
aux_params
=
load_checkpoint
(
"
%
s/
%
s"
%
(
model_
dir
,
model_name
),
0
)
import
argparse
import
argparse
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
...
...
nnvm/tutorials/from_coreml.py
View file @
c6339730
...
@@ -21,19 +21,7 @@ import tvm
...
@@ -21,19 +21,7 @@ import tvm
import
coremltools
as
cm
import
coremltools
as
cm
import
numpy
as
np
import
numpy
as
np
from
PIL
import
Image
from
PIL
import
Image
from
tvm.contrib.download
import
download_testdata
def
download
(
url
,
path
,
overwrite
=
False
):
import
os
if
os
.
path
.
isfile
(
path
)
and
not
overwrite
:
print
(
'File {} existed, skip.'
.
format
(
path
))
return
print
(
'Downloading from url {} to {}'
.
format
(
url
,
path
))
try
:
import
urllib.request
urllib
.
request
.
urlretrieve
(
url
,
path
)
except
:
import
urllib
urllib
.
urlretrieve
(
url
,
path
)
######################################################################
######################################################################
# Load pretrained CoreML model
# Load pretrained CoreML model
...
@@ -42,9 +30,9 @@ def download(url, path, overwrite=False):
...
@@ -42,9 +30,9 @@ def download(url, path, overwrite=False):
# provided by apple in this example
# provided by apple in this example
model_url
=
'https://docs-assets.developer.apple.com/coreml/models/MobileNet.mlmodel'
model_url
=
'https://docs-assets.developer.apple.com/coreml/models/MobileNet.mlmodel'
model_file
=
'mobilenet.mlmodel'
model_file
=
'mobilenet.mlmodel'
download
(
model_url
,
model_file
)
model_path
=
download_testdata
(
model_url
,
model_file
,
module
=
'coreml'
)
# now you mobilenet.mlmodel on disk
# now you mobilenet.mlmodel on disk
mlmodel
=
cm
.
models
.
MLModel
(
model_
file
)
mlmodel
=
cm
.
models
.
MLModel
(
model_
path
)
# we can load the graph as NNVM compatible model
# we can load the graph as NNVM compatible model
sym
,
params
=
nnvm
.
frontend
.
from_coreml
(
mlmodel
)
sym
,
params
=
nnvm
.
frontend
.
from_coreml
(
mlmodel
)
...
@@ -54,8 +42,8 @@ sym, params = nnvm.frontend.from_coreml(mlmodel)
...
@@ -54,8 +42,8 @@ sym, params = nnvm.frontend.from_coreml(mlmodel)
# A single cat dominates the examples!
# A single cat dominates the examples!
from
PIL
import
Image
from
PIL
import
Image
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
download
(
img_url
,
'cat.png
'
)
img_path
=
download_testdata
(
img_url
,
'cat.png'
,
module
=
'data
'
)
img
=
Image
.
open
(
'cat.png'
)
.
resize
((
224
,
224
))
img
=
Image
.
open
(
img_path
)
.
resize
((
224
,
224
))
#x = np.transpose(img, (2, 0, 1))[np.newaxis, :]
#x = np.transpose(img, (2, 0, 1))[np.newaxis, :]
image
=
np
.
asarray
(
img
)
image
=
np
.
asarray
(
img
)
image
=
image
.
transpose
((
2
,
0
,
1
))
image
=
image
.
transpose
((
2
,
0
,
1
))
...
@@ -95,8 +83,8 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
...
@@ -95,8 +83,8 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'imagenet1000_clsid_to_human.txt'
])
'imagenet1000_clsid_to_human.txt'
])
synset_name
=
'
synset
.txt'
synset_name
=
'
imagenet1000_clsid_to_human
.txt'
download
(
synset_url
,
synset_name
)
synset_path
=
download_testdata
(
synset_url
,
synset_name
,
module
=
'data'
)
with
open
(
synset_
name
)
as
f
:
with
open
(
synset_
path
)
as
f
:
synset
=
eval
(
f
.
read
())
synset
=
eval
(
f
.
read
())
print
(
'Top-1 id'
,
top1
,
'class name'
,
synset
[
top1
])
print
(
'Top-1 id'
,
top1
,
'class name'
,
synset
[
top1
])
nnvm/tutorials/from_darknet.py
View file @
c6339730
...
@@ -25,7 +25,7 @@ import tvm
...
@@ -25,7 +25,7 @@ import tvm
import
sys
import
sys
from
ctypes
import
*
from
ctypes
import
*
from
tvm.contrib.download
import
download
from
tvm.contrib.download
import
download
_testdata
from
nnvm.testing.darknet
import
__darknetffi__
from
nnvm.testing.darknet
import
__darknetffi__
# Model name
# Model name
...
@@ -41,8 +41,8 @@ REPO_URL = 'https://github.com/siju-samuel/darknet/blob/master/'
...
@@ -41,8 +41,8 @@ REPO_URL = 'https://github.com/siju-samuel/darknet/blob/master/'
CFG_URL
=
REPO_URL
+
'cfg/'
+
CFG_NAME
+
'?raw=true'
CFG_URL
=
REPO_URL
+
'cfg/'
+
CFG_NAME
+
'?raw=true'
WEIGHTS_URL
=
'https://pjreddie.com/media/files/'
+
WEIGHTS_NAME
WEIGHTS_URL
=
'https://pjreddie.com/media/files/'
+
WEIGHTS_NAME
download
(
CFG_URL
,
CFG_NAME
)
cfg_path
=
download_testdata
(
CFG_URL
,
CFG_NAME
,
module
=
"darknet"
)
download
(
WEIGHTS_URL
,
WEIGHTS_NAME
)
weights_path
=
download_testdata
(
WEIGHTS_URL
,
WEIGHTS_NAME
,
module
=
"darknet"
)
# Download and Load darknet library
# Download and Load darknet library
if
sys
.
platform
in
[
'linux'
,
'linux2'
]:
if
sys
.
platform
in
[
'linux'
,
'linux2'
]:
...
@@ -55,12 +55,10 @@ else:
...
@@ -55,12 +55,10 @@ else:
err
=
"Darknet lib is not supported on {} platform"
.
format
(
sys
.
platform
)
err
=
"Darknet lib is not supported on {} platform"
.
format
(
sys
.
platform
)
raise
NotImplementedError
(
err
)
raise
NotImplementedError
(
err
)
download
(
DARKNET_URL
,
DARKNET_LIB
)
lib_path
=
download_testdata
(
DARKNET_URL
,
DARKNET_LIB
,
module
=
"darknet"
)
DARKNET_LIB
=
__darknetffi__
.
dlopen
(
'./'
+
DARKNET_LIB
)
DARKNET_LIB
=
__darknetffi__
.
dlopen
(
lib_path
)
cfg
=
"./"
+
str
(
CFG_NAME
)
net
=
DARKNET_LIB
.
load_network
(
cfg_path
.
encode
(
'utf-8'
),
weights_path
.
encode
(
'utf-8'
),
0
)
weights
=
"./"
+
str
(
WEIGHTS_NAME
)
net
=
DARKNET_LIB
.
load_network
(
cfg
.
encode
(
'utf-8'
),
weights
.
encode
(
'utf-8'
),
0
)
dtype
=
'float32'
dtype
=
'float32'
batch_size
=
1
batch_size
=
1
...
@@ -88,9 +86,9 @@ test_image = 'dog.jpg'
...
@@ -88,9 +86,9 @@ test_image = 'dog.jpg'
print
(
"Loading the test image..."
)
print
(
"Loading the test image..."
)
img_url
=
'https://github.com/siju-samuel/darknet/blob/master/data/'
+
\
img_url
=
'https://github.com/siju-samuel/darknet/blob/master/data/'
+
\
test_image
+
'?raw=true'
test_image
+
'?raw=true'
download
(
img_url
,
test_image
)
img_path
=
download_testdata
(
img_url
,
test_image
,
"data"
)
data
=
nnvm
.
testing
.
darknet
.
load_image
(
test_image
,
netw
,
neth
)
data
=
nnvm
.
testing
.
darknet
.
load_image
(
img_path
,
netw
,
neth
)
######################################################################
######################################################################
# Execute on TVM Runtime
# Execute on TVM Runtime
# ----------------------
# ----------------------
...
@@ -150,10 +148,10 @@ coco_name = 'coco.names'
...
@@ -150,10 +148,10 @@ coco_name = 'coco.names'
coco_url
=
'https://github.com/siju-samuel/darknet/blob/master/data/'
+
coco_name
+
'?raw=true'
coco_url
=
'https://github.com/siju-samuel/darknet/blob/master/data/'
+
coco_name
+
'?raw=true'
font_name
=
'arial.ttf'
font_name
=
'arial.ttf'
font_url
=
'https://github.com/siju-samuel/darknet/blob/master/data/'
+
font_name
+
'?raw=true'
font_url
=
'https://github.com/siju-samuel/darknet/blob/master/data/'
+
font_name
+
'?raw=true'
download
(
coco_url
,
coco_name
)
coco_path
=
download_testdata
(
coco_url
,
coco_name
,
module
=
'data'
)
download
(
font_url
,
font_name
)
font_path
=
download_testdata
(
font_url
,
font_name
,
module
=
'data'
)
with
open
(
coco_
name
)
as
f
:
with
open
(
coco_
path
)
as
f
:
content
=
f
.
readlines
()
content
=
f
.
readlines
()
names
=
[
x
.
strip
()
for
x
in
content
]
names
=
[
x
.
strip
()
for
x
in
content
]
...
...
nnvm/tutorials/from_mxnet.py
View file @
c6339730
...
@@ -20,30 +20,31 @@ https://mxnet.incubator.apache.org/versions/master/install/index.html
...
@@ -20,30 +20,31 @@ https://mxnet.incubator.apache.org/versions/master/install/index.html
"""
"""
# some standard imports
# some standard imports
import
mxnet
as
mx
import
mxnet
as
mx
import
numpy
as
np
import
nnvm
import
nnvm
import
tvm
import
tvm
import
numpy
as
np
from
tvm.contrib.download
import
download_testdata
######################################################################
######################################################################
# Download Resnet18 model from Gluon Model Zoo
# Download Resnet18 model from Gluon Model Zoo
# ---------------------------------------------
# ---------------------------------------------
# In this section, we download a pretrained imagenet model and classify an image.
# In this section, we download a pretrained imagenet model and classify an image.
from
mxnet.gluon.model_zoo.vision
import
get_model
from
mxnet.gluon.model_zoo.vision
import
get_model
from
mxnet.gluon.utils
import
download
from
PIL
import
Image
from
PIL
import
Image
from
matplotlib
import
pyplot
as
plt
from
matplotlib
import
pyplot
as
plt
block
=
get_model
(
'resnet18_v1'
,
pretrained
=
True
)
block
=
get_model
(
'resnet18_v1'
,
pretrained
=
True
)
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
img_name
=
'cat.png'
img_name
=
'cat.png'
synset_url
=
''
.
join
([
'https://gist.githubusercontent.com/zhreshold/'
,
synset_url
=
''
.
join
([
'https://gist.githubusercontent.com/zhreshold/'
,
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'imagenet1000_clsid_to_human.txt'
])
'imagenet1000_clsid_to_human.txt'
])
synset_name
=
'
synset
.txt'
synset_name
=
'
imagenet1000_clsid_to_human
.txt'
download
(
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
,
img_name
)
img_path
=
download_testdata
(
img_url
,
img_name
,
module
=
'data'
)
download
(
synset_url
,
synset_name
)
synset_path
=
download_testdata
(
synset_url
,
synset_name
,
module
=
'data'
)
with
open
(
synset_
name
)
as
f
:
with
open
(
synset_
path
)
as
f
:
synset
=
eval
(
f
.
read
())
synset
=
eval
(
f
.
read
())
image
=
Image
.
open
(
img_
name
)
.
resize
((
224
,
224
))
image
=
Image
.
open
(
img_
path
)
.
resize
((
224
,
224
))
plt
.
imshow
(
image
)
plt
.
imshow
(
image
)
plt
.
show
()
plt
.
show
()
...
...
nnvm/tutorials/from_mxnet_to_webgl.py
View file @
c6339730
...
@@ -33,6 +33,7 @@ from __future__ import print_function
...
@@ -33,6 +33,7 @@ from __future__ import print_function
import
numpy
as
np
import
numpy
as
np
import
tvm
import
tvm
from
tvm.contrib.download
import
download_testdata
import
nnvm.compiler
import
nnvm.compiler
import
nnvm.testing
import
nnvm.testing
...
@@ -106,16 +107,14 @@ def download_synset():
...
@@ -106,16 +107,14 @@ def download_synset():
print
(
"Downloading synset..."
)
print
(
"Downloading synset..."
)
from
mxnet
import
gluon
url
=
"https://gist.githubusercontent.com/zhreshold/"
+
\
url
=
"https://gist.githubusercontent.com/zhreshold/"
+
\
"4d0b62f3d01426887599d4f7ede23ee5/raw/"
+
\
"4d0b62f3d01426887599d4f7ede23ee5/raw/"
+
\
"596b27d23537e5a1b5751d2b0481ef172f58b539/"
+
\
"596b27d23537e5a1b5751d2b0481ef172f58b539/"
+
\
"imagenet1000_clsid_to_human.txt"
"imagenet1000_clsid_to_human.txt"
file_name
=
"
synset
.txt"
file_name
=
"
imagenet1000_clsid_to_human
.txt"
gluon
.
utils
.
download
(
url
,
file_name
)
file_path
=
download_testdata
(
url
,
file_name
,
module
=
'data'
)
with
open
(
file_
name
)
as
f
:
with
open
(
file_
path
)
as
f
:
synset
=
eval
(
f
.
read
())
synset
=
eval
(
f
.
read
())
print
(
"- Synset downloaded!"
)
print
(
"- Synset downloaded!"
)
...
@@ -144,14 +143,13 @@ def download_image():
...
@@ -144,14 +143,13 @@ def download_image():
print
(
"Downloading cat image..."
)
print
(
"Downloading cat image..."
)
from
matplotlib
import
pyplot
as
plt
from
matplotlib
import
pyplot
as
plt
from
mxnet
import
gluon
from
PIL
import
Image
from
PIL
import
Image
url
=
"https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true"
url
=
"https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true"
img_name
=
"cat.png"
img_name
=
"cat.png"
gluon
.
utils
.
download
(
url
,
img_name
)
img_path
=
download_testdata
(
url
,
img_name
,
module
=
'data'
)
image
=
Image
.
open
(
img_
name
)
.
resize
((
224
,
224
))
image
=
Image
.
open
(
img_
path
)
.
resize
((
224
,
224
))
print
(
"- Cat image downloaded!"
)
print
(
"- Cat image downloaded!"
)
...
...
nnvm/tutorials/from_onnx.py
View file @
c6339730
...
@@ -18,22 +18,10 @@ https://github.com/onnx/onnx
...
@@ -18,22 +18,10 @@ https://github.com/onnx/onnx
"""
"""
import
nnvm
import
nnvm
import
tvm
import
tvm
from
tvm.contrib.download
import
download_testdata
import
onnx
import
onnx
import
numpy
as
np
import
numpy
as
np
def
download
(
url
,
path
,
overwrite
=
False
):
import
os
if
os
.
path
.
isfile
(
path
)
and
not
overwrite
:
print
(
'File {} existed, skip.'
.
format
(
path
))
return
print
(
'Downloading from url {} to {}'
.
format
(
url
,
path
))
try
:
import
urllib.request
urllib
.
request
.
urlretrieve
(
url
,
path
)
except
:
import
urllib
urllib
.
urlretrieve
(
url
,
path
)
######################################################################
######################################################################
# Load pretrained ONNX model
# Load pretrained ONNX model
# ---------------------------------------------
# ---------------------------------------------
...
@@ -44,9 +32,9 @@ model_url = ''.join(['https://gist.github.com/zhreshold/',
...
@@ -44,9 +32,9 @@ model_url = ''.join(['https://gist.github.com/zhreshold/',
'bcda4716699ac97ea44f791c24310193/raw/'
,
'bcda4716699ac97ea44f791c24310193/raw/'
,
'93672b029103648953c4e5ad3ac3aadf346a4cdc/'
,
'93672b029103648953c4e5ad3ac3aadf346a4cdc/'
,
'super_resolution_0.2.onnx'
])
'super_resolution_0.2.onnx'
])
download
(
model_url
,
'super_resolution.onnx'
,
True
)
model_path
=
download_testdata
(
model_url
,
'super_resolution.onnx'
,
module
=
'onnx'
)
# now you have super_resolution.onnx on disk
# now you have super_resolution.onnx on disk
onnx_model
=
onnx
.
load_model
(
'super_resolution.onnx'
)
onnx_model
=
onnx
.
load_model
(
model_path
)
# we can load the graph as NNVM compatible model
# we can load the graph as NNVM compatible model
sym
,
params
=
nnvm
.
frontend
.
from_onnx
(
onnx_model
)
sym
,
params
=
nnvm
.
frontend
.
from_onnx
(
onnx_model
)
...
@@ -56,8 +44,8 @@ sym, params = nnvm.frontend.from_onnx(onnx_model)
...
@@ -56,8 +44,8 @@ sym, params = nnvm.frontend.from_onnx(onnx_model)
# A single cat dominates the examples!
# A single cat dominates the examples!
from
PIL
import
Image
from
PIL
import
Image
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
download
(
img_url
,
'cat.png
'
)
img_path
=
download_testdata
(
img_url
,
'cat.png'
,
module
=
'data
'
)
img
=
Image
.
open
(
'cat.png'
)
.
resize
((
224
,
224
))
img
=
Image
.
open
(
img_path
)
.
resize
((
224
,
224
))
img_ycbcr
=
img
.
convert
(
"YCbCr"
)
# convert to YCbCr
img_ycbcr
=
img
.
convert
(
"YCbCr"
)
# convert to YCbCr
img_y
,
img_cb
,
img_cr
=
img_ycbcr
.
split
()
img_y
,
img_cb
,
img_cr
=
img_ycbcr
.
split
()
x
=
np
.
array
(
img_y
)[
np
.
newaxis
,
np
.
newaxis
,
:,
:]
x
=
np
.
array
(
img_y
)[
np
.
newaxis
,
np
.
newaxis
,
:,
:]
...
...
nnvm/tutorials/from_tensorflow.py
View file @
c6339730
...
@@ -52,8 +52,8 @@ map_proto = 'imagenet_2012_challenge_label_map_proto.pbtxt'
...
@@ -52,8 +52,8 @@ map_proto = 'imagenet_2012_challenge_label_map_proto.pbtxt'
map_proto_url
=
os
.
path
.
join
(
repo_base
,
map_proto
)
map_proto_url
=
os
.
path
.
join
(
repo_base
,
map_proto
)
# Human readable text for labels
# Human readable text for labels
lab
le
_map
=
'imagenet_synset_to_human_label_map.txt'
lab
el
_map
=
'imagenet_synset_to_human_label_map.txt'
lab
le_map_url
=
os
.
path
.
join
(
repo_base
,
lable
_map
)
lab
el_map_url
=
os
.
path
.
join
(
repo_base
,
label
_map
)
# Target settings
# Target settings
# Use these commented settings to build for cuda.
# Use these commented settings to build for cuda.
...
@@ -70,19 +70,19 @@ ctx = tvm.cpu(0)
...
@@ -70,19 +70,19 @@ ctx = tvm.cpu(0)
# Download required files
# Download required files
# -----------------------
# -----------------------
# Download files listed above.
# Download files listed above.
from
mxnet.gluon.utils
import
download
from
tvm.contrib.download
import
download_testdata
download
(
image_url
,
img_name
)
img_path
=
download_testdata
(
image_url
,
img_name
,
module
=
'data'
)
download
(
model_url
,
model_name
)
model_path
=
download_testdata
(
model_url
,
model_name
,
module
=
[
'tf'
,
'InceptionV1'
]
)
download
(
map_proto_url
,
map_proto
)
map_proto_path
=
download_testdata
(
map_proto_url
,
map_proto
,
module
=
'data'
)
download
(
lable_map_url
,
lable_map
)
label_path
=
download_testdata
(
label_map_url
,
label_map
,
module
=
'data'
)
######################################################################
######################################################################
# Import model
# Import model
# ------------
# ------------
# Creates tensorflow graph definition from protobuf file.
# Creates tensorflow graph definition from protobuf file.
with
tf
.
gfile
.
FastGFile
(
os
.
path
.
join
(
"./"
,
model_name
)
,
'rb'
)
as
f
:
with
tf
.
gfile
.
FastGFile
(
model_path
,
'rb'
)
as
f
:
graph_def
=
tf
.
GraphDef
()
graph_def
=
tf
.
GraphDef
()
graph_def
.
ParseFromString
(
f
.
read
())
graph_def
.
ParseFromString
(
f
.
read
())
graph
=
tf
.
import_graph_def
(
graph_def
,
name
=
''
)
graph
=
tf
.
import_graph_def
(
graph_def
,
name
=
''
)
...
@@ -103,7 +103,7 @@ with tf.gfile.FastGFile(os.path.join("./", model_name), 'rb') as f:
...
@@ -103,7 +103,7 @@ with tf.gfile.FastGFile(os.path.join("./", model_name), 'rb') as f:
#
#
from
PIL
import
Image
from
PIL
import
Image
image
=
Image
.
open
(
img_
name
)
.
resize
((
299
,
299
))
image
=
Image
.
open
(
img_
path
)
.
resize
((
299
,
299
))
x
=
np
.
array
(
image
)
x
=
np
.
array
(
image
)
...
@@ -117,7 +117,7 @@ x = np.array(image)
...
@@ -117,7 +117,7 @@ x = np.array(image)
# params: params converted from tensorflow params (tensor protobuf).
# params: params converted from tensorflow params (tensor protobuf).
sym
,
params
=
nnvm
.
frontend
.
from_tensorflow
(
graph_def
,
layout
=
layout
)
sym
,
params
=
nnvm
.
frontend
.
from_tensorflow
(
graph_def
,
layout
=
layout
)
print
(
"Tensorflow protobuf imported as nnvm graph"
)
print
(
"Tensorflow protobuf imported as nnvm graph"
)
######################################################################
######################################################################
# NNVM Compilation
# NNVM Compilation
# ----------------
# ----------------
...
@@ -157,8 +157,8 @@ predictions = tvm_output.asnumpy()
...
@@ -157,8 +157,8 @@ predictions = tvm_output.asnumpy()
predictions
=
np
.
squeeze
(
predictions
)
predictions
=
np
.
squeeze
(
predictions
)
# Creates node ID --> English string lookup.
# Creates node ID --> English string lookup.
node_lookup
=
tf_testing
.
NodeLookup
(
label_lookup_path
=
os
.
path
.
join
(
"./"
,
map_proto
)
,
node_lookup
=
tf_testing
.
NodeLookup
(
label_lookup_path
=
map_proto_path
,
uid_lookup_path
=
os
.
path
.
join
(
"./"
,
lable_map
)
)
uid_lookup_path
=
label_path
)
# Print top 5 predictions from TVM output.
# Print top 5 predictions from TVM output.
top_k
=
predictions
.
argsort
()[
-
5
:][::
-
1
]
top_k
=
predictions
.
argsort
()[
-
5
:][::
-
1
]
...
@@ -175,7 +175,7 @@ for node_id in top_k:
...
@@ -175,7 +175,7 @@ for node_id in top_k:
def
create_graph
():
def
create_graph
():
"""Creates a graph from saved GraphDef file and returns a saver."""
"""Creates a graph from saved GraphDef file and returns a saver."""
# Creates graph from saved graph_def.pb.
# Creates graph from saved graph_def.pb.
with
tf
.
gfile
.
FastGFile
(
model_
name
,
'rb'
)
as
f
:
with
tf
.
gfile
.
FastGFile
(
model_
path
,
'rb'
)
as
f
:
graph_def
=
tf
.
GraphDef
()
graph_def
=
tf
.
GraphDef
()
graph_def
.
ParseFromString
(
f
.
read
())
graph_def
.
ParseFromString
(
f
.
read
())
graph
=
tf
.
import_graph_def
(
graph_def
,
name
=
''
)
graph
=
tf
.
import_graph_def
(
graph_def
,
name
=
''
)
...
@@ -209,8 +209,8 @@ def run_inference_on_image(image):
...
@@ -209,8 +209,8 @@ def run_inference_on_image(image):
predictions
=
np
.
squeeze
(
predictions
)
predictions
=
np
.
squeeze
(
predictions
)
# Creates node ID --> English string lookup.
# Creates node ID --> English string lookup.
node_lookup
=
tf_testing
.
NodeLookup
(
label_lookup_path
=
os
.
path
.
join
(
"./"
,
map_proto
)
,
node_lookup
=
tf_testing
.
NodeLookup
(
label_lookup_path
=
map_proto_path
,
uid_lookup_path
=
os
.
path
.
join
(
"./"
,
lable_map
)
)
uid_lookup_path
=
label_path
)
# Print top 5 predictions from tensorflow.
# Print top 5 predictions from tensorflow.
top_k
=
predictions
.
argsort
()[
-
5
:][::
-
1
]
top_k
=
predictions
.
argsort
()[
-
5
:][::
-
1
]
...
@@ -220,4 +220,4 @@ def run_inference_on_image(image):
...
@@ -220,4 +220,4 @@ def run_inference_on_image(image):
score
=
predictions
[
node_id
]
score
=
predictions
[
node_id
]
print
(
'
%
s (score =
%.5
f)'
%
(
human_string
,
score
))
print
(
'
%
s (score =
%.5
f)'
%
(
human_string
,
score
))
run_inference_on_image
(
img_name
)
run_inference_on_image
(
img_path
)
nnvm/tutorials/nlp/from_darknet_rnn.py
View file @
c6339730
...
@@ -22,9 +22,9 @@ by the script.
...
@@ -22,9 +22,9 @@ by the script.
"""
"""
import
random
import
random
import
numpy
as
np
import
numpy
as
np
from
mxnet.gluon.utils
import
download
import
tvm
import
tvm
from
tvm.contrib
import
graph_runtime
from
tvm.contrib
import
graph_runtime
from
tvm.contrib.download
import
download_testdata
from
nnvm.testing.darknet
import
__darknetffi__
from
nnvm.testing.darknet
import
__darknetffi__
import
nnvm
import
nnvm
import
nnvm.frontend.darknet
import
nnvm.frontend.darknet
...
@@ -49,17 +49,15 @@ REPO_URL = 'https://github.com/dmlc/web-data/blob/master/darknet/'
...
@@ -49,17 +49,15 @@ REPO_URL = 'https://github.com/dmlc/web-data/blob/master/darknet/'
CFG_URL
=
REPO_URL
+
'cfg/'
+
CFG_NAME
+
'?raw=true'
CFG_URL
=
REPO_URL
+
'cfg/'
+
CFG_NAME
+
'?raw=true'
WEIGHTS_URL
=
REPO_URL
+
'weights/'
+
WEIGHTS_NAME
+
'?raw=true'
WEIGHTS_URL
=
REPO_URL
+
'weights/'
+
WEIGHTS_NAME
+
'?raw=true'
download
(
CFG_URL
,
CFG_NAME
)
cfg_path
=
download_testdata
(
CFG_URL
,
CFG_NAME
,
module
=
'darknet'
)
download
(
WEIGHTS_URL
,
WEIGHTS_NAME
)
weights_path
=
download_testdata
(
WEIGHTS_URL
,
WEIGHTS_NAME
,
module
=
'darknet'
)
# Download and Load darknet library
# Download and Load darknet library
DARKNET_LIB
=
'libdarknet.so'
DARKNET_LIB
=
'libdarknet.so'
DARKNET_URL
=
REPO_URL
+
'lib/'
+
DARKNET_LIB
+
'?raw=true'
DARKNET_URL
=
REPO_URL
+
'lib/'
+
DARKNET_LIB
+
'?raw=true'
download
(
DARKNET_URL
,
DARKNET_LIB
)
lib_path
=
download_testdata
(
DARKNET_URL
,
DARKNET_LIB
,
module
=
'darknet'
)
DARKNET_LIB
=
__darknetffi__
.
dlopen
(
'./'
+
DARKNET_LIB
)
DARKNET_LIB
=
__darknetffi__
.
dlopen
(
lib_path
)
cfg
=
"./"
+
str
(
CFG_NAME
)
net
=
DARKNET_LIB
.
load_network
(
cfg_path
.
encode
(
'utf-8'
),
weights_path
.
encode
(
'utf-8'
),
0
)
weights
=
"./"
+
str
(
WEIGHTS_NAME
)
net
=
DARKNET_LIB
.
load_network
(
cfg
.
encode
(
'utf-8'
),
weights
.
encode
(
'utf-8'
),
0
)
dtype
=
'float32'
dtype
=
'float32'
batch_size
=
1
batch_size
=
1
...
...
nnvm/tutorials/nlp/keras_s2s_translate.py
View file @
c6339730
...
@@ -59,9 +59,9 @@ model_url = os.path.join(repo_base, model_file)
...
@@ -59,9 +59,9 @@ model_url = os.path.join(repo_base, model_file)
data_url
=
os
.
path
.
join
(
repo_base
,
data_file
)
data_url
=
os
.
path
.
join
(
repo_base
,
data_file
)
# Download files listed below.
# Download files listed below.
from
mxnet.gluon.utils
import
download
from
tvm.contrib.download
import
download_testdata
download
(
model_url
,
model_file
)
model_path
=
download_testdata
(
model_url
,
model_file
,
module
=
'keras'
)
d
ownload
(
data_url
,
model_file
)
d
ata_path
=
download_testdata
(
data_url
,
data_file
,
module
=
'data'
)
latent_dim
=
256
# Latent dimensionality of the encoding space.
latent_dim
=
256
# Latent dimensionality of the encoding space.
test_samples
=
10000
# Number of samples used for testing.
test_samples
=
10000
# Number of samples used for testing.
...
@@ -76,7 +76,7 @@ input_texts = []
...
@@ -76,7 +76,7 @@ input_texts = []
target_texts
=
[]
target_texts
=
[]
input_characters
=
set
()
input_characters
=
set
()
target_characters
=
set
()
target_characters
=
set
()
with
open
(
data_
file
,
'r'
,
encoding
=
'utf-8'
)
as
f
:
with
open
(
data_
path
,
'r'
,
encoding
=
'utf-8'
)
as
f
:
lines
=
f
.
read
()
.
split
(
'
\n
'
)
lines
=
f
.
read
()
.
split
(
'
\n
'
)
test_samples
=
min
(
test_samples
,
len
(
lines
))
test_samples
=
min
(
test_samples
,
len
(
lines
))
max_encoder_seq_length
=
0
max_encoder_seq_length
=
0
...
@@ -112,7 +112,7 @@ reverse_target_char_index = dict(
...
@@ -112,7 +112,7 @@ reverse_target_char_index = dict(
# Load Keras Model
# Load Keras Model
# ----------------
# ----------------
# Restore the model and construct the encoder and decoder.
# Restore the model and construct the encoder and decoder.
model
=
load_model
(
model_
file
)
model
=
load_model
(
model_
path
)
encoder_inputs
=
model
.
input
[
0
]
# input_1
encoder_inputs
=
model
.
input
[
0
]
# input_1
encoder_outputs
,
state_h_enc
,
state_c_enc
=
model
.
layers
[
2
]
.
output
# lstm_1
encoder_outputs
,
state_h_enc
,
state_c_enc
=
model
.
layers
[
2
]
.
output
# lstm_1
...
...
python/tvm/contrib/download.py
View file @
c6339730
...
@@ -28,6 +28,9 @@ def download(url, path, overwrite=False, size_compare=False, verbose=1, retries=
...
@@ -28,6 +28,9 @@ def download(url, path, overwrite=False, size_compare=False, verbose=1, retries=
verbose: int, optional
verbose: int, optional
Verbose level
Verbose level
retries: int, optional
Number of time to retry download, default at 3.
"""
"""
if
sys
.
version_info
>=
(
3
,):
if
sys
.
version_info
>=
(
3
,):
import
urllib.request
as
urllib2
import
urllib.request
as
urllib2
...
...
python/tvm/relay/testing/tf.py
View file @
c6339730
...
@@ -331,7 +331,7 @@ def get_workload_ptb():
...
@@ -331,7 +331,7 @@ def get_workload_ptb():
ptb_model_file
=
'RNN/ptb/ptb_model_with_lstmblockcell.pb'
ptb_model_file
=
'RNN/ptb/ptb_model_with_lstmblockcell.pb'
import
tarfile
import
tarfile
file_path
=
download_testdata
(
sample_url
,
sample_data_file
,
module
=
[
'
tf
'
,
'ptb_data'
])
file_path
=
download_testdata
(
sample_url
,
sample_data_file
,
module
=
[
'
data
'
,
'ptb_data'
])
dir_path
=
os
.
path
.
dirname
(
file_path
)
dir_path
=
os
.
path
.
dirname
(
file_path
)
t
=
tarfile
.
open
(
file_path
,
'r'
)
t
=
tarfile
.
open
(
file_path
,
'r'
)
t
.
extractall
(
dir_path
)
t
.
extractall
(
dir_path
)
...
...
tests/python/frontend/coreml/model_zoo/__init__.py
View file @
c6339730
...
@@ -18,7 +18,7 @@ def get_resnet50():
...
@@ -18,7 +18,7 @@ def get_resnet50():
def
get_cat_image
():
def
get_cat_image
():
url
=
'https://gist.githubusercontent.com/zhreshold/bcda4716699ac97ea44f791c24310193/raw/fa7ef0e9c9a5daea686d6473a62aacd1a5885849/cat.png'
url
=
'https://gist.githubusercontent.com/zhreshold/bcda4716699ac97ea44f791c24310193/raw/fa7ef0e9c9a5daea686d6473a62aacd1a5885849/cat.png'
dst
=
'cat.png'
dst
=
'cat.png'
real_dst
=
download_testdata
(
url
,
dst
,
module
=
'
coreml
'
)
real_dst
=
download_testdata
(
url
,
dst
,
module
=
'
data
'
)
img
=
Image
.
open
(
real_dst
)
.
resize
((
224
,
224
))
img
=
Image
.
open
(
real_dst
)
.
resize
((
224
,
224
))
img
=
np
.
transpose
(
img
,
(
2
,
0
,
1
))[
np
.
newaxis
,
:]
img
=
np
.
transpose
(
img
,
(
2
,
0
,
1
))[
np
.
newaxis
,
:]
return
np
.
asarray
(
img
)
return
np
.
asarray
(
img
)
\ No newline at end of file
tutorials/frontend/deploy_model_on_rasp.py
View file @
c6339730
...
@@ -14,6 +14,7 @@ import tvm
...
@@ -14,6 +14,7 @@ import tvm
import
tvm.relay
as
relay
import
tvm.relay
as
relay
from
tvm
import
rpc
from
tvm
import
rpc
from
tvm.contrib
import
util
,
graph_runtime
as
runtime
from
tvm.contrib
import
util
,
graph_runtime
as
runtime
from
tvm.contrib.download
import
download_testdata
######################################################################
######################################################################
# .. _build-tvm-runtime-on-device:
# .. _build-tvm-runtime-on-device:
...
@@ -82,7 +83,6 @@ from tvm.contrib import util, graph_runtime as runtime
...
@@ -82,7 +83,6 @@ from tvm.contrib import util, graph_runtime as runtime
# You can found more details about this part at tutorial :ref:`tutorial-from-mxnet`.
# You can found more details about this part at tutorial :ref:`tutorial-from-mxnet`.
from
mxnet.gluon.model_zoo.vision
import
get_model
from
mxnet.gluon.model_zoo.vision
import
get_model
from
mxnet.gluon.utils
import
download
from
PIL
import
Image
from
PIL
import
Image
import
numpy
as
np
import
numpy
as
np
...
@@ -92,9 +92,10 @@ block = get_model('resnet18_v1', pretrained=True)
...
@@ -92,9 +92,10 @@ block = get_model('resnet18_v1', pretrained=True)
######################################################################
######################################################################
# In order to test our model, here we download an image of cat and
# In order to test our model, here we download an image of cat and
# transform its format.
# transform its format.
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
img_name
=
'cat.png'
img_name
=
'cat.png'
download
(
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
,
img_name
)
img_path
=
download_testdata
(
img_url
,
img_name
,
module
=
'data'
)
image
=
Image
.
open
(
img_
name
)
.
resize
((
224
,
224
))
image
=
Image
.
open
(
img_
path
)
.
resize
((
224
,
224
))
def
transform_image
(
image
):
def
transform_image
(
image
):
image
=
np
.
array
(
image
)
-
np
.
array
([
123.
,
117.
,
104.
])
image
=
np
.
array
(
image
)
-
np
.
array
([
123.
,
117.
,
104.
])
...
@@ -112,9 +113,9 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
...
@@ -112,9 +113,9 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'imagenet1000_clsid_to_human.txt'
])
'imagenet1000_clsid_to_human.txt'
])
synset_name
=
'
synset
.txt'
synset_name
=
'
imagenet1000_clsid_to_human
.txt'
download
(
synset_url
,
synset_name
)
synset_path
=
download_testdata
(
synset_url
,
synset_name
,
module
=
'data'
)
with
open
(
synset_
name
)
as
f
:
with
open
(
synset_
path
)
as
f
:
synset
=
eval
(
f
.
read
())
synset
=
eval
(
f
.
read
())
######################################################################
######################################################################
...
...
tutorials/frontend/deploy_ssd_gluoncv.py
View file @
c6339730
...
@@ -12,6 +12,7 @@ from matplotlib import pyplot as plt
...
@@ -12,6 +12,7 @@ from matplotlib import pyplot as plt
from
tvm.relay.testing.config
import
ctx_list
from
tvm.relay.testing.config
import
ctx_list
from
tvm
import
relay
from
tvm
import
relay
from
tvm.contrib
import
graph_runtime
from
tvm.contrib
import
graph_runtime
from
tvm.contrib.download
import
download_testdata
from
gluoncv
import
model_zoo
,
data
,
utils
from
gluoncv
import
model_zoo
,
data
,
utils
...
@@ -49,9 +50,9 @@ target_list = ctx_list()
...
@@ -49,9 +50,9 @@ target_list = ctx_list()
######################################################################
######################################################################
# Download and pre-process demo image
# Download and pre-process demo image
im_fname
=
utils
.
download
(
'https://github.com/dmlc/web-data/blob/master/'
+
im_fname
=
download_testdata
(
'https://github.com/dmlc/web-data/blob/master/'
+
'gluoncv/detection/street_small.jpg?raw=true'
,
'gluoncv/detection/street_small.jpg?raw=true'
,
path
=
'street_small.jpg
'
)
'street_small.jpg'
,
module
=
'data
'
)
x
,
img
=
data
.
transforms
.
presets
.
ssd
.
load_test
(
im_fname
,
short
=
512
)
x
,
img
=
data
.
transforms
.
presets
.
ssd
.
load_test
(
im_fname
,
short
=
512
)
######################################################################
######################################################################
...
...
tutorials/frontend/from_caffe2.py
View file @
c6339730
...
@@ -19,21 +19,6 @@ A quick solution is to install via conda
...
@@ -19,21 +19,6 @@ A quick solution is to install via conda
or please refer to official site
or please refer to official site
https://caffe2.ai/docs/getting-started.html
https://caffe2.ai/docs/getting-started.html
"""
"""
######################################################################
# Utils for downloading files
# ----------------------------
def
download
(
url
,
path
,
overwrite
=
False
):
import
os
if
os
.
path
.
isfile
(
path
)
and
not
overwrite
:
print
(
'File {} exists, skip.'
.
format
(
path
))
return
print
(
'Downloading from url {} to {}'
.
format
(
url
,
path
))
try
:
import
urllib.request
urllib
.
request
.
urlretrieve
(
url
,
path
)
except
:
import
urllib
urllib
.
urlretrieve
(
url
,
path
)
######################################################################
######################################################################
# Load pretrained Caffe2 model
# Load pretrained Caffe2 model
...
@@ -52,12 +37,13 @@ resnet50 = Model('resnet50')
...
@@ -52,12 +37,13 @@ resnet50 = Model('resnet50')
# Load a test image
# Load a test image
# ------------------
# ------------------
# A single cat dominates the examples!
# A single cat dominates the examples!
from
tvm.contrib.download
import
download_testdata
from
PIL
import
Image
from
PIL
import
Image
from
matplotlib
import
pyplot
as
plt
from
matplotlib
import
pyplot
as
plt
import
numpy
as
np
import
numpy
as
np
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
download
(
img_url
,
'cat.png
'
)
img_path
=
download_testdata
(
img_url
,
'cat.png'
,
module
=
'data
'
)
img
=
Image
.
open
(
'cat.png'
)
.
resize
((
224
,
224
))
img
=
Image
.
open
(
img_path
)
.
resize
((
224
,
224
))
plt
.
imshow
(
img
)
plt
.
imshow
(
img
)
plt
.
show
()
plt
.
show
()
# input preprocess
# input preprocess
...
@@ -118,9 +104,9 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
...
@@ -118,9 +104,9 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'imagenet1000_clsid_to_human.txt'
])
'imagenet1000_clsid_to_human.txt'
])
synset_name
=
'
synset
.txt'
synset_name
=
'
imagenet1000_clsid_to_human
.txt'
download
(
synset_url
,
synset_name
)
synset_path
=
download_testdata
(
synset_url
,
synset_name
,
module
=
'data'
)
with
open
(
synset_
name
)
as
f
:
with
open
(
synset_
path
)
as
f
:
synset
=
eval
(
f
.
read
())
synset
=
eval
(
f
.
read
())
print
(
'Relay top-1 id: {}, class name: {}'
.
format
(
top1_tvm
,
synset
[
top1_tvm
]))
print
(
'Relay top-1 id: {}, class name: {}'
.
format
(
top1_tvm
,
synset
[
top1_tvm
]))
# confirm correctness with caffe2 output
# confirm correctness with caffe2 output
...
...
tutorials/frontend/from_coreml.py
View file @
c6339730
...
@@ -19,23 +19,11 @@ https://github.com/apple/coremltools
...
@@ -19,23 +19,11 @@ https://github.com/apple/coremltools
"""
"""
import
tvm
import
tvm
import
tvm.relay
as
relay
import
tvm.relay
as
relay
from
tvm.contrib.download
import
download_testdata
import
coremltools
as
cm
import
coremltools
as
cm
import
numpy
as
np
import
numpy
as
np
from
PIL
import
Image
from
PIL
import
Image
def
download
(
url
,
path
,
overwrite
=
False
):
import
os
if
os
.
path
.
isfile
(
path
)
and
not
overwrite
:
print
(
'File {} existed, skip.'
.
format
(
path
))
return
print
(
'Downloading from url {} to {}'
.
format
(
url
,
path
))
try
:
import
urllib.request
urllib
.
request
.
urlretrieve
(
url
,
path
)
except
:
import
urllib
urllib
.
urlretrieve
(
url
,
path
)
######################################################################
######################################################################
# Load pretrained CoreML model
# Load pretrained CoreML model
# ----------------------------
# ----------------------------
...
@@ -43,17 +31,17 @@ def download(url, path, overwrite=False):
...
@@ -43,17 +31,17 @@ def download(url, path, overwrite=False):
# provided by apple in this example
# provided by apple in this example
model_url
=
'https://docs-assets.developer.apple.com/coreml/models/MobileNet.mlmodel'
model_url
=
'https://docs-assets.developer.apple.com/coreml/models/MobileNet.mlmodel'
model_file
=
'mobilenet.mlmodel'
model_file
=
'mobilenet.mlmodel'
download
(
model_url
,
model_file
)
model_path
=
download_testdata
(
model_url
,
model_file
,
module
=
'coreml'
)
# Now you have mobilenet.mlmodel on disk
# Now you have mobilenet.mlmodel on disk
mlmodel
=
cm
.
models
.
MLModel
(
model_
file
)
mlmodel
=
cm
.
models
.
MLModel
(
model_
path
)
######################################################################
######################################################################
# Load a test image
# Load a test image
# ------------------
# ------------------
# A single cat dominates the examples!
# A single cat dominates the examples!
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
download
(
img_url
,
'cat.png
'
)
img_path
=
download_testdata
(
img_url
,
'cat.png'
,
module
=
'data
'
)
img
=
Image
.
open
(
'cat.png'
)
.
resize
((
224
,
224
))
img
=
Image
.
open
(
img_path
)
.
resize
((
224
,
224
))
x
=
np
.
transpose
(
img
,
(
2
,
0
,
1
))[
np
.
newaxis
,
:]
x
=
np
.
transpose
(
img
,
(
2
,
0
,
1
))[
np
.
newaxis
,
:]
######################################################################
######################################################################
...
@@ -94,8 +82,8 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
...
@@ -94,8 +82,8 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'imagenet1000_clsid_to_human.txt'
])
'imagenet1000_clsid_to_human.txt'
])
synset_name
=
'
synset
.txt'
synset_name
=
'
imagenet1000_clsid_to_human
.txt'
download
(
synset_url
,
synset_name
)
synset_path
=
download_testdata
(
synset_url
,
synset_name
,
module
=
'data'
)
with
open
(
synset_
name
)
as
f
:
with
open
(
synset_
path
)
as
f
:
synset
=
eval
(
f
.
read
())
synset
=
eval
(
f
.
read
())
print
(
'Top-1 id'
,
top1
,
'class name'
,
synset
[
top1
])
print
(
'Top-1 id'
,
top1
,
'class name'
,
synset
[
top1
])
tutorials/frontend/from_keras.py
View file @
c6339730
...
@@ -20,22 +20,10 @@ https://keras.io/#installation
...
@@ -20,22 +20,10 @@ https://keras.io/#installation
"""
"""
import
tvm
import
tvm
import
tvm.relay
as
relay
import
tvm.relay
as
relay
from
tvm.contrib.download
import
download_testdata
import
keras
import
keras
import
numpy
as
np
import
numpy
as
np
def
download
(
url
,
path
,
overwrite
=
False
):
import
os
if
os
.
path
.
isfile
(
path
)
and
not
overwrite
:
print
(
'File {} exists, skip.'
.
format
(
path
))
return
print
(
'Downloading from url {} to {}'
.
format
(
url
,
path
))
try
:
import
urllib.request
urllib
.
request
.
urlretrieve
(
url
,
path
)
except
:
import
urllib
urllib
.
urlretrieve
(
url
,
path
)
######################################################################
######################################################################
# Load pretrained keras model
# Load pretrained keras model
# ----------------------------
# ----------------------------
...
@@ -43,10 +31,10 @@ def download(url, path, overwrite=False):
...
@@ -43,10 +31,10 @@ def download(url, path, overwrite=False):
weights_url
=
''
.
join
([
'https://github.com/fchollet/deep-learning-models/releases/'
,
weights_url
=
''
.
join
([
'https://github.com/fchollet/deep-learning-models/releases/'
,
'download/v0.2/resnet50_weights_tf_dim_ordering_tf_kernels.h5'
])
'download/v0.2/resnet50_weights_tf_dim_ordering_tf_kernels.h5'
])
weights_file
=
'resnet50_weights.h5'
weights_file
=
'resnet50_weights.h5'
download
(
weights_url
,
weights_file
)
weights_path
=
download_testdata
(
weights_url
,
weights_file
,
module
=
'keras'
)
keras_resnet50
=
keras
.
applications
.
resnet50
.
ResNet50
(
include_top
=
True
,
weights
=
None
,
keras_resnet50
=
keras
.
applications
.
resnet50
.
ResNet50
(
include_top
=
True
,
weights
=
None
,
input_shape
=
(
224
,
224
,
3
),
classes
=
1000
)
input_shape
=
(
224
,
224
,
3
),
classes
=
1000
)
keras_resnet50
.
load_weights
(
'resnet50_weights.h5'
)
keras_resnet50
.
load_weights
(
weights_path
)
######################################################################
######################################################################
# Load a test image
# Load a test image
...
@@ -56,8 +44,8 @@ from PIL import Image
...
@@ -56,8 +44,8 @@ from PIL import Image
from
matplotlib
import
pyplot
as
plt
from
matplotlib
import
pyplot
as
plt
from
keras.applications.resnet50
import
preprocess_input
from
keras.applications.resnet50
import
preprocess_input
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
download
(
img_url
,
'cat.png
'
)
img_path
=
download_testdata
(
img_url
,
'cat.png'
,
module
=
'data
'
)
img
=
Image
.
open
(
'cat.png'
)
.
resize
((
224
,
224
))
img
=
Image
.
open
(
img_path
)
.
resize
((
224
,
224
))
plt
.
imshow
(
img
)
plt
.
imshow
(
img
)
plt
.
show
()
plt
.
show
()
# input preprocess
# input preprocess
...
@@ -92,9 +80,9 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
...
@@ -92,9 +80,9 @@ synset_url = ''.join(['https://gist.githubusercontent.com/zhreshold/',
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'imagenet1000_clsid_to_human.txt'
])
'imagenet1000_clsid_to_human.txt'
])
synset_name
=
'
synset
.txt'
synset_name
=
'
imagenet1000_clsid_to_human
.txt'
download
(
synset_url
,
synset_name
)
synset_path
=
download_testdata
(
synset_url
,
synset_name
,
module
=
'data'
)
with
open
(
synset_
name
)
as
f
:
with
open
(
synset_
path
)
as
f
:
synset
=
eval
(
f
.
read
())
synset
=
eval
(
f
.
read
())
print
(
'Relay top-1 id: {}, class name: {}'
.
format
(
top1_tvm
,
synset
[
top1_tvm
]))
print
(
'Relay top-1 id: {}, class name: {}'
.
format
(
top1_tvm
,
synset
[
top1_tvm
]))
# confirm correctness with keras output
# confirm correctness with keras output
...
...
tutorials/frontend/from_mxnet.py
View file @
c6339730
...
@@ -29,22 +29,23 @@ import numpy as np
...
@@ -29,22 +29,23 @@ import numpy as np
# Download Resnet18 model from Gluon Model Zoo
# Download Resnet18 model from Gluon Model Zoo
# ---------------------------------------------
# ---------------------------------------------
# In this section, we download a pretrained imagenet model and classify an image.
# In this section, we download a pretrained imagenet model and classify an image.
from
tvm.contrib.download
import
download_testdata
from
mxnet.gluon.model_zoo.vision
import
get_model
from
mxnet.gluon.model_zoo.vision
import
get_model
from
mxnet.gluon.utils
import
download
from
PIL
import
Image
from
PIL
import
Image
from
matplotlib
import
pyplot
as
plt
from
matplotlib
import
pyplot
as
plt
block
=
get_model
(
'resnet18_v1'
,
pretrained
=
True
)
block
=
get_model
(
'resnet18_v1'
,
pretrained
=
True
)
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
img_name
=
'cat.png'
img_name
=
'cat.png'
synset_url
=
''
.
join
([
'https://gist.githubusercontent.com/zhreshold/'
,
synset_url
=
''
.
join
([
'https://gist.githubusercontent.com/zhreshold/'
,
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'4d0b62f3d01426887599d4f7ede23ee5/raw/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'596b27d23537e5a1b5751d2b0481ef172f58b539/'
,
'imagenet1000_clsid_to_human.txt'
])
'imagenet1000_clsid_to_human.txt'
])
synset_name
=
'
synset
.txt'
synset_name
=
'
imagenet1000_clsid_to_human
.txt'
download
(
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
,
img_name
)
img_path
=
download_testdata
(
img_url
,
'cat.png'
,
module
=
'data'
)
download
(
synset_url
,
synset_name
)
synset_path
=
download_testdata
(
synset_url
,
synset_name
,
module
=
'data'
)
with
open
(
synset_
name
)
as
f
:
with
open
(
synset_
path
)
as
f
:
synset
=
eval
(
f
.
read
())
synset
=
eval
(
f
.
read
())
image
=
Image
.
open
(
img_
name
)
.
resize
((
224
,
224
))
image
=
Image
.
open
(
img_
path
)
.
resize
((
224
,
224
))
plt
.
imshow
(
image
)
plt
.
imshow
(
image
)
plt
.
show
()
plt
.
show
()
...
...
tutorials/frontend/from_onnx.py
View file @
c6339730
...
@@ -20,19 +20,7 @@ import onnx
...
@@ -20,19 +20,7 @@ import onnx
import
numpy
as
np
import
numpy
as
np
import
tvm
import
tvm
import
tvm.relay
as
relay
import
tvm.relay
as
relay
from
tvm.contrib.download
import
download_testdata
def
download
(
url
,
path
,
overwrite
=
False
):
import
os
if
os
.
path
.
isfile
(
path
)
and
not
overwrite
:
print
(
'File {} existed, skip.'
.
format
(
path
))
return
print
(
'Downloading from url {} to {}'
.
format
(
url
,
path
))
try
:
import
urllib.request
urllib
.
request
.
urlretrieve
(
url
,
path
)
except
:
import
urllib
urllib
.
urlretrieve
(
url
,
path
)
######################################################################
######################################################################
# Load pretrained ONNX model
# Load pretrained ONNX model
...
@@ -44,9 +32,9 @@ model_url = ''.join(['https://gist.github.com/zhreshold/',
...
@@ -44,9 +32,9 @@ model_url = ''.join(['https://gist.github.com/zhreshold/',
'bcda4716699ac97ea44f791c24310193/raw/'
,
'bcda4716699ac97ea44f791c24310193/raw/'
,
'93672b029103648953c4e5ad3ac3aadf346a4cdc/'
,
'93672b029103648953c4e5ad3ac3aadf346a4cdc/'
,
'super_resolution_0.2.onnx'
])
'super_resolution_0.2.onnx'
])
download
(
model_url
,
'super_resolution.onnx'
,
False
)
model_path
=
download_testdata
(
model_url
,
'super_resolution.onnx'
,
module
=
'onnx'
)
# now you have super_resolution.onnx on disk
# now you have super_resolution.onnx on disk
onnx_model
=
onnx
.
load
(
'super_resolution.onnx'
)
onnx_model
=
onnx
.
load
(
model_path
)
######################################################################
######################################################################
# Load a test image
# Load a test image
...
@@ -54,8 +42,8 @@ onnx_model = onnx.load('super_resolution.onnx')
...
@@ -54,8 +42,8 @@ onnx_model = onnx.load('super_resolution.onnx')
# A single cat dominates the examples!
# A single cat dominates the examples!
from
PIL
import
Image
from
PIL
import
Image
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
img_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
download
(
img_url
,
'cat.png
'
)
img_path
=
download_testdata
(
img_url
,
'cat.png'
,
module
=
'data
'
)
img
=
Image
.
open
(
'cat.png'
)
.
resize
((
224
,
224
))
img
=
Image
.
open
(
img_path
)
.
resize
((
224
,
224
))
img_ycbcr
=
img
.
convert
(
"YCbCr"
)
# convert to YCbCr
img_ycbcr
=
img
.
convert
(
"YCbCr"
)
# convert to YCbCr
img_y
,
img_cb
,
img_cr
=
img_ycbcr
.
split
()
img_y
,
img_cb
,
img_cr
=
img_ycbcr
.
split
()
x
=
np
.
array
(
img_y
)[
np
.
newaxis
,
np
.
newaxis
,
:,
:]
x
=
np
.
array
(
img_y
)[
np
.
newaxis
,
np
.
newaxis
,
:,
:]
...
...
tutorials/frontend/from_tensorflow.py
View file @
c6339730
...
@@ -43,8 +43,8 @@ map_proto = 'imagenet_2012_challenge_label_map_proto.pbtxt'
...
@@ -43,8 +43,8 @@ map_proto = 'imagenet_2012_challenge_label_map_proto.pbtxt'
map_proto_url
=
os
.
path
.
join
(
repo_base
,
map_proto
)
map_proto_url
=
os
.
path
.
join
(
repo_base
,
map_proto
)
# Human readable text for labels
# Human readable text for labels
lab
le
_map
=
'imagenet_synset_to_human_label_map.txt'
lab
el
_map
=
'imagenet_synset_to_human_label_map.txt'
lab
le_map_url
=
os
.
path
.
join
(
repo_base
,
lable
_map
)
lab
el_map_url
=
os
.
path
.
join
(
repo_base
,
label
_map
)
# Target settings
# Target settings
# Use these commented settings to build for cuda.
# Use these commented settings to build for cuda.
...
@@ -61,19 +61,19 @@ ctx = tvm.cpu(0)
...
@@ -61,19 +61,19 @@ ctx = tvm.cpu(0)
# Download required files
# Download required files
# -----------------------
# -----------------------
# Download files listed above.
# Download files listed above.
from
mxnet.gluon.utils
import
download
from
tvm.contrib.download
import
download_testdata
download
(
image_url
,
img_name
)
img_path
=
download_testdata
(
image_url
,
img_name
,
module
=
'data'
)
download
(
model_url
,
model_name
)
model_path
=
download_testdata
(
model_url
,
model_name
,
module
=
[
'tf'
,
'InceptionV1'
]
)
download
(
map_proto_url
,
map_proto
)
map_proto_path
=
download_testdata
(
map_proto_url
,
map_proto
,
module
=
'data'
)
download
(
lable_map_url
,
lable_map
)
label_path
=
download_testdata
(
label_map_url
,
label_map
,
module
=
'data'
)
######################################################################
######################################################################
# Import model
# Import model
# ------------
# ------------
# Creates tensorflow graph definition from protobuf file.
# Creates tensorflow graph definition from protobuf file.
with
tf
.
gfile
.
FastGFile
(
os
.
path
.
join
(
"./"
,
model_name
)
,
'rb'
)
as
f
:
with
tf
.
gfile
.
FastGFile
(
model_path
,
'rb'
)
as
f
:
graph_def
=
tf
.
GraphDef
()
graph_def
=
tf
.
GraphDef
()
graph_def
.
ParseFromString
(
f
.
read
())
graph_def
.
ParseFromString
(
f
.
read
())
graph
=
tf
.
import_graph_def
(
graph_def
,
name
=
''
)
graph
=
tf
.
import_graph_def
(
graph_def
,
name
=
''
)
...
@@ -94,7 +94,7 @@ with tf.gfile.FastGFile(os.path.join("./", model_name), 'rb') as f:
...
@@ -94,7 +94,7 @@ with tf.gfile.FastGFile(os.path.join("./", model_name), 'rb') as f:
#
#
from
PIL
import
Image
from
PIL
import
Image
image
=
Image
.
open
(
img_
name
)
.
resize
((
299
,
299
))
image
=
Image
.
open
(
img_
path
)
.
resize
((
299
,
299
))
x
=
np
.
array
(
image
)
x
=
np
.
array
(
image
)
...
@@ -110,7 +110,7 @@ shape_dict = {'DecodeJpeg/contents': x.shape}
...
@@ -110,7 +110,7 @@ shape_dict = {'DecodeJpeg/contents': x.shape}
dtype_dict
=
{
'DecodeJpeg/contents'
:
'uint8'
}
dtype_dict
=
{
'DecodeJpeg/contents'
:
'uint8'
}
sym
,
params
=
relay
.
frontend
.
from_tensorflow
(
graph_def
,
layout
=
layout
,
shape
=
shape_dict
)
sym
,
params
=
relay
.
frontend
.
from_tensorflow
(
graph_def
,
layout
=
layout
,
shape
=
shape_dict
)
print
(
"Tensorflow protobuf imported to relay frontend."
)
print
(
"Tensorflow protobuf imported to relay frontend."
)
######################################################################
######################################################################
# Relay Build
# Relay Build
# -----------
# -----------
...
@@ -148,8 +148,8 @@ predictions = tvm_output.asnumpy()
...
@@ -148,8 +148,8 @@ predictions = tvm_output.asnumpy()
predictions
=
np
.
squeeze
(
predictions
)
predictions
=
np
.
squeeze
(
predictions
)
# Creates node ID --> English string lookup.
# Creates node ID --> English string lookup.
node_lookup
=
tf_testing
.
NodeLookup
(
label_lookup_path
=
os
.
path
.
join
(
"./"
,
map_proto
)
,
node_lookup
=
tf_testing
.
NodeLookup
(
label_lookup_path
=
map_proto_path
,
uid_lookup_path
=
os
.
path
.
join
(
"./"
,
lable_map
)
)
uid_lookup_path
=
label_path
)
# Print top 5 predictions from TVM output.
# Print top 5 predictions from TVM output.
top_k
=
predictions
.
argsort
()[
-
5
:][::
-
1
]
top_k
=
predictions
.
argsort
()[
-
5
:][::
-
1
]
...
@@ -166,7 +166,7 @@ for node_id in top_k:
...
@@ -166,7 +166,7 @@ for node_id in top_k:
def
create_graph
():
def
create_graph
():
"""Creates a graph from saved GraphDef file and returns a saver."""
"""Creates a graph from saved GraphDef file and returns a saver."""
# Creates graph from saved graph_def.pb.
# Creates graph from saved graph_def.pb.
with
tf
.
gfile
.
FastGFile
(
model_
name
,
'rb'
)
as
f
:
with
tf
.
gfile
.
FastGFile
(
model_
path
,
'rb'
)
as
f
:
graph_def
=
tf
.
GraphDef
()
graph_def
=
tf
.
GraphDef
()
graph_def
.
ParseFromString
(
f
.
read
())
graph_def
.
ParseFromString
(
f
.
read
())
graph
=
tf
.
import_graph_def
(
graph_def
,
name
=
''
)
graph
=
tf
.
import_graph_def
(
graph_def
,
name
=
''
)
...
@@ -200,8 +200,8 @@ def run_inference_on_image(image):
...
@@ -200,8 +200,8 @@ def run_inference_on_image(image):
predictions
=
np
.
squeeze
(
predictions
)
predictions
=
np
.
squeeze
(
predictions
)
# Creates node ID --> English string lookup.
# Creates node ID --> English string lookup.
node_lookup
=
tf_testing
.
NodeLookup
(
label_lookup_path
=
os
.
path
.
join
(
"./"
,
map_proto
)
,
node_lookup
=
tf_testing
.
NodeLookup
(
label_lookup_path
=
map_proto_path
,
uid_lookup_path
=
os
.
path
.
join
(
"./"
,
lable_map
)
)
uid_lookup_path
=
label_path
)
# Print top 5 predictions from tensorflow.
# Print top 5 predictions from tensorflow.
top_k
=
predictions
.
argsort
()[
-
5
:][::
-
1
]
top_k
=
predictions
.
argsort
()[
-
5
:][::
-
1
]
...
@@ -211,4 +211,4 @@ def run_inference_on_image(image):
...
@@ -211,4 +211,4 @@ def run_inference_on_image(image):
score
=
predictions
[
node_id
]
score
=
predictions
[
node_id
]
print
(
'
%
s (score =
%.5
f)'
%
(
human_string
,
score
))
print
(
'
%
s (score =
%.5
f)'
%
(
human_string
,
score
))
run_inference_on_image
(
img_name
)
run_inference_on_image
(
img_path
)
tutorials/frontend/from_tflite.py
View file @
c6339730
...
@@ -52,25 +52,14 @@ Below you can find an example for how to compile TFLite model using TVM.
...
@@ -52,25 +52,14 @@ Below you can find an example for how to compile TFLite model using TVM.
######################################################################
######################################################################
# Utils for downloading and extracting zip files
# Utils for downloading and extracting zip files
# ---------------------------------------------
# ---------------------------------------------
import
os
def
download
(
url
,
path
,
overwrite
=
False
):
import
os
if
os
.
path
.
isfile
(
path
)
and
not
overwrite
:
print
(
'File {} existed, skip.'
.
format
(
path
))
return
print
(
'Downloading from url {} to {}'
.
format
(
url
,
path
))
try
:
import
urllib.request
urllib
.
request
.
urlretrieve
(
url
,
path
)
except
:
import
urllib
urllib
.
urlretrieve
(
url
,
path
)
def
extract
(
path
):
def
extract
(
path
):
import
tarfile
import
tarfile
if
path
.
endswith
(
"tgz"
)
or
path
.
endswith
(
"gz"
):
if
path
.
endswith
(
"tgz"
)
or
path
.
endswith
(
"gz"
):
dir_path
=
os
.
path
.
dirname
(
path
)
tar
=
tarfile
.
open
(
path
)
tar
=
tarfile
.
open
(
path
)
tar
.
extractall
()
tar
.
extractall
(
path
=
dir_path
)
tar
.
close
()
tar
.
close
()
else
:
else
:
raise
RuntimeError
(
'Could not decompress the file: '
+
path
)
raise
RuntimeError
(
'Could not decompress the file: '
+
path
)
...
@@ -80,14 +69,17 @@ def extract(path):
...
@@ -80,14 +69,17 @@ def extract(path):
# Load pretrained TFLite model
# Load pretrained TFLite model
# ---------------------------------------------
# ---------------------------------------------
# we load mobilenet V1 TFLite model provided by Google
# we load mobilenet V1 TFLite model provided by Google
from
tvm.contrib.download
import
download_testdata
model_url
=
"http://download.tensorflow.org/models/mobilenet_v1_2018_08_02/mobilenet_v1_1.0_224.tgz"
model_url
=
"http://download.tensorflow.org/models/mobilenet_v1_2018_08_02/mobilenet_v1_1.0_224.tgz"
# we download model tar file and extract, finally get mobilenet_v1_1.0_224.tflite
# we download model tar file and extract, finally get mobilenet_v1_1.0_224.tflite
download
(
model_url
,
"mobilenet_v1_1.0_224.tgz"
,
False
)
model_path
=
download_testdata
(
model_url
,
"mobilenet_v1_1.0_224.tgz"
,
module
=
[
'tf'
,
'official'
])
extract
(
"mobilenet_v1_1.0_224.tgz"
)
model_dir
=
os
.
path
.
dirname
(
model_path
)
extract
(
model_path
)
# now we have mobilenet_v1_1.0_224.tflite on disk and open it
# now we have mobilenet_v1_1.0_224.tflite on disk and open it
tflite_model_file
=
"mobilenet_v1_1.0_224.tflite"
tflite_model_file
=
os
.
path
.
join
(
model_dir
,
"mobilenet_v1_1.0_224.tflite"
)
tflite_model_buf
=
open
(
tflite_model_file
,
"rb"
)
.
read
()
tflite_model_buf
=
open
(
tflite_model_file
,
"rb"
)
.
read
()
# get TFLite model from buffer
# get TFLite model from buffer
...
@@ -103,8 +95,8 @@ from matplotlib import pyplot as plt
...
@@ -103,8 +95,8 @@ from matplotlib import pyplot as plt
import
numpy
as
np
import
numpy
as
np
image_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
image_url
=
'https://github.com/dmlc/mxnet.js/blob/master/data/cat.png?raw=true'
download
(
image_url
,
'cat.png
'
)
image_path
=
download_testdata
(
image_url
,
'cat.png'
,
module
=
'data
'
)
resized_image
=
Image
.
open
(
'cat.png'
)
.
resize
((
224
,
224
))
resized_image
=
Image
.
open
(
image_path
)
.
resize
((
224
,
224
))
plt
.
imshow
(
resized_image
)
plt
.
imshow
(
resized_image
)
plt
.
show
()
plt
.
show
()
image_data
=
np
.
asarray
(
resized_image
)
.
astype
(
"float32"
)
image_data
=
np
.
asarray
(
resized_image
)
.
astype
(
"float32"
)
...
@@ -179,11 +171,11 @@ label_file_url = ''.join(['https://raw.githubusercontent.com/',
...
@@ -179,11 +171,11 @@ label_file_url = ''.join(['https://raw.githubusercontent.com/',
'app/src/main/assets/'
,
'app/src/main/assets/'
,
'labels_mobilenet_quant_v1_224.txt'
])
'labels_mobilenet_quant_v1_224.txt'
])
label_file
=
"labels_mobilenet_quant_v1_224.txt"
label_file
=
"labels_mobilenet_quant_v1_224.txt"
download
(
label_file_url
,
label_file
)
label_path
=
download_testdata
(
label_file_url
,
label_file
,
module
=
'data'
)
# map id to 1001 classes
# map id to 1001 classes
labels
=
dict
()
labels
=
dict
()
with
open
(
label_
file
)
as
f
:
with
open
(
label_
path
)
as
f
:
for
id
,
line
in
enumerate
(
f
):
for
id
,
line
in
enumerate
(
f
):
labels
[
id
]
=
line
labels
[
id
]
=
line
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment