- 06 Jan, 2020 2 commits
-
-
* [CONV] Asymmetic padding * fix lint error * update for legalize, rocm and cudnn * add more test cases * change more symmetric padding * change conv2d winograd tests according orginal cases * remove 'alter_op_layout.h' header in bitserial.cc
optima2005 committed -
* Support empty tensor * Fix schedule * Refactor * Minor fix * Fix pylint * Merge cpp and python is_empty_shape
Yao Wang committed
-
- 03 Jan, 2020 1 commit
-
-
* add onnx resize converter * update frontends * updating topi * adding onnx resize tests * fixed NHWC test by casting size dtype to int32 * fix tests * fix lint * update existing test cases * fix tensorflow frontend * fix lint * remove NHWC stuff * update topi resize test for half_pixel * update doc * fix doc * remove onnx resize bits
masahi committed
-
- 01 Jan, 2020 1 commit
-
-
* [FRONTEND][TF] Add conv3d * fix high rtol
optima2005 committed
-
- 27 Dec, 2019 1 commit
-
-
* [TOPI] add 3D upsampling Op. * fix lint issues * change align_corners to coordinate_transformation_mode * fix resize3d half_pixel * make a simple function and clean up trilinear_resize3d_python * fix doc
optima2005 committed
-
- 24 Dec, 2019 1 commit
-
-
* Added tvm function stencil for subpixel operations to topi. * Topi subpixel operators added and tested. * Added subpixel attrs. * Added depth_to_space relay attributes. * depth_to_space fully working. * Fixed NHWC shape bug. * SpaceToDepth in and all tests passing. * lint fixes. * Added string include * Fixed topi formatting. * Added DCR/CDR mode to depthtospace operator.
Josh Fromm committed
-
- 23 Dec, 2019 1 commit
-
-
* [Relay] add max_pool3d in relay and TF converter * fix comments
Yong Wu committed
-
- 18 Dec, 2019 1 commit
-
-
Alex Gladkov committed
-
- 12 Dec, 2019 1 commit
-
-
* [TOPI] implement pool3d op * use PoolInferCorrectLayout for both 2d and 3d pooling * unify MakeMaxPool and MakeAvgPool
optima2005 committed
-
- 04 Dec, 2019 1 commit
-
-
* implement conv3d op * add back missed conv2d_output_shape by mistake * fix typo and docs, add topi test * rebase to master and merge 2d/3d unification * use cudnn.conv_forward
optima2005 committed
-
- 03 Dec, 2019 1 commit
-
-
abergeron committed
-
- 21 Nov, 2019 1 commit
-
-
* [TOPI] Fix flaky testcase for floor div * avoid check at 0.0
Yizhi Liu committed
-
- 18 Nov, 2019 1 commit
-
-
* Add tf FloorMod * Add floor_div/mod into topi and relay * Add to rst * Fix test
Yao Wang committed
-
- 13 Nov, 2019 1 commit
-
-
* Support Proposal operator on CPU. * PyLint space issue * PyLint space issue * Pylint singleton-comparison issue
Zhao Wu committed
-
- 06 Nov, 2019 1 commit
-
-
* fix winograd * move get padding after kernel transform
Cody Hao Yu committed
-
- 30 Oct, 2019 1 commit
-
-
* Add support for Any op * Support ONNX frontend * Add doc * Add to relay docs * Dummy change to retrigger CI
Jon Soifer committed
-
- 28 Oct, 2019 2 commits
-
-
* :add scale2 for upsample * update unit test for upsampling * support latest upsample op for multiple frontend * fix lint * fix lint * fix lint * fix lint * update scale description and rebase
Xingyu Zhou committed -
Tianqi Chen committed
-
- 24 Oct, 2019 1 commit
-
-
* support conv2d HWCN in AutoTVM and Relay * fix lint * fix comments and unit tests
Cody Hao Yu committed
-
- 11 Oct, 2019 1 commit
-
-
* overload half operators for cuda codegen * add float16 te test_op_level1 * fix test_op_level1.py * fix lint * disable fp16 test if gpu does not support * disable fp16 test if gpu does not support * bypass float16 test if gpu does not support float16
Xingyu Zhou committed
-
- 10 Oct, 2019 1 commit
-
-
* Add FIFO buffer op to enable explicit computation re-use in convolution * Add a test * Add end-to-end test with 1D convolution * Add a stub in MXNet frontend * Address reviewer comments * Add back stub for MXNet frontend
Philip Hyunsu Cho committed
-
- 02 Oct, 2019 1 commit
-
-
Umang Yadav committed
-
- 22 Sep, 2019 1 commit
-
-
* add expr `isnan` * move to intrinsic * doc & add to topi * fix error from ci
Huang, Guangtai committed
-
- 20 Sep, 2019 1 commit
-
-
MXNet pad is described at: https://mxnet.incubator.apache.org/api/python/symbol/symbol.html#mxnet.symbol.pad Add support for parameter 'None' in MXNet slice operator. MXNet 'slice' is described at https://mxnet.incubator.apache.org/api/python/symbol/symbol.html#mxnet.symbol.slice Add support for MXNet cos, sin, arctan MXNet 'cos' is described at https://mxnet.incubator.apache.org/api/python/symbol/symbol.html#mxnet.symbol.cos MXNet 'sin' is described at https://mxnet.incubator.apache.org/api/python/symbol/symbol.html#mxnet.symbol.sin MXNet arctan is descirbed at https://mxnet.incubator.apache.org/api/python/symbol/symbol.html#mxnet.symbol.arctan Add support for MXNet 1D Convolution and 1D Deconvolution MXNet convolution is described at: https://mxnet.incubator.apache.org/api/python/symbol/symbol.html#mxnet.symbol.Convolution MXNet Deconvolution is described at: https://mxnet.incubator.apache.org/api/python/symbol/symbol.html#mxnet.symbol.Deconvolution
Alex Gladkov committed
-
- 19 Sep, 2019 1 commit
-
-
* add proper scheduling for dense on CUDA * add fallback config and fix unit test * fix corner cases * refactoring * fix bias and add testcase * let fusion happen
Cody Hao Yu committed
-
- 16 Sep, 2019 1 commit
-
-
* [TOPI] operator support: logical_and, logical_or, logical_not * [TOPI] operator support: logical_and, logical_or, logical_not * [TOPI] fix test cases for operator support: logical_and, logical_or, logical_not * [TOPI] fix test cases for operator support: logical_not
Neo Chien committed
-
- 09 Sep, 2019 1 commit
-
-
* add more ops * stop vectorization for erf * x * cleanup * fix * add whitelist for vectorizable intrin * add tf converter * fix dense * fix * add missing intrin * fix mxnet frontend * fix nvptx
Haichen Shen committed
-
- 08 Sep, 2019 1 commit
-
-
雾雨魔理沙 committed
-
- 01 Sep, 2019 1 commit
-
-
* init shape func in interpreter and vm compiler * Update interpreter * fix * lint * lint * fix * remove hack * update * fix * fix * update * address comments & update for shape_of * fix lint * update * fix hybrid * lint * fix bug & add take shape func * lint * lint * update * fix flaky test * add todo
Haichen Shen committed
-
- 22 Aug, 2019 2 commits
-
-
* Add one-hot to Relay * topi implementation * Working * add topi test * Add TF test * Fix check * fix linting issues * fix documentation * Fix documentation * Add support for on_value, off_value, axis, dtype * Add full support for axis * Fix compute and update test_forward * Move on_value and off_value to inputs * Add topi test * Update tests * Update docs * Fix style * re-enable tests * Add one_hot to mxnet converter
Jon Soifer committed -
Josh Fromm committed
-
- 06 Aug, 2019 1 commit
-
-
* add build gcn tutorial * add transpose operator for square sparse matrices * remove extra files * change loop tag * comply with lint * comply with lint -- line too long * comply with lint * lint check * lint check * lint check * apply marisa and theirry's reviews
Yulun Yao committed
-
- 01 Aug, 2019 1 commit
-
-
The patch adds support for Tensorflow operators log1p and cos Tensorflow log1p is described at https://www.tensorflow.org/api_docs/python/tf/math/log1p Tensorflow cos is described at https://www.tensorflow.org/api_docs/python/tf/math/cos Tensorflow sin is described at https://www.tensorflow.org/api_docs/python/tf/math/sin
alexgl-github committed
-
- 31 Jul, 2019 1 commit
-
-
* [TOPI][CUDA] schedule for group_conv2d * Fix #flops
Wuwei Lin committed
-
- 30 Jul, 2019 1 commit
-
-
* Fix traverse_inline not inline zero input op properly * Add where to python and set tag to broadcast * Fix inline * test * fix test target * fix
Wuwei Lin committed
-
- 28 Jul, 2019 1 commit
-
-
Balint Cristian committed
-
- 26 Jul, 2019 1 commit
-
-
* [TOPI][CUDA] Schedule for pool_grad * Relay test * Fix fused op * doc * Remove set scope local
Wuwei Lin committed
-
- 25 Jul, 2019 1 commit
-
-
Balint Cristian committed
-
- 24 Jul, 2019 1 commit
-
-
Wuwei Lin committed
-
- 23 Jul, 2019 1 commit
-
-
internally and externally, interested in replacing standard dense layers with block-sparse matrix multiplication layers. The motivations are generally: higher performance (due to reduction in FLOPs, memory bandwidth/cache footprint), enabling larger models (e.g. fitting more layers in a given memory budget). Some public work along these lines: * https://openai.com/blog/block-sparse-gpu-kernels/ * https://openai.com/blog/sparse-transformer/ * https://arxiv.org/abs/1802.08435 * https://arxiv.org/abs/1711.02782 Various groups have been able to successfully train models with reasonable levels of sparsity (90%+) with marginal accuracy changes, which suggests substantial speedups are possible (as this implies a >10x reduction in FLOPs). It is fairly straightforward to realize these theoretical speedups, see e.g. TVM benchmarks for Intel CPUs in https://gist.github.com/ajtulloch/e65f90487bceb8848128e8db582fe902, and CUDA results in https://github.com/openai/blocksparse, etc. * https://github.com/openai/blocksparse (CUDA) * https://software.intel.com/en-us/mkl-developer-reference-c-mkl-bsrmm (MKL BSRM) * https://docs.scipy.org/doc/scipy-0.14.0/reference/generated/scipy.sparse.bsr_matrix.html (SCIPY BSR representation) This is extracted from an internal patch we've been using internally. There are various extensions possible (int8/fp16/bf16, CUDA/other GPU architectures), but this is a reasonable starting point. This needs more thorough unit test coverage however. We follow the conventions established by scipy.sparse.bsr_matrix and other libraries, see the unit tests for details. For folks interested in experimenting with scheduling/AutoTVM etc, https://gist.github.com/ajtulloch/e65f90487bceb8848128e8db582fe902 is a useful starting point.
Andrew Tulloch committed
-