- 23 Mar, 2020 1 commit
-
-
Neo Chien committed
-
- 22 Mar, 2020 5 commits
-
-
- This allows to emit vectorized loads/stores for CUDA math intrinsics. - A few intrinsics should be lowered as CUDAMath not CUDAFastMath ones. - Fixed the code block identation.
Wei Pan committed -
Haichen Shen committed
-
* Fix up the final pieces * Tweak build.rs
Jared Roesch committed -
Tianqi Chen committed
-
* [DOCS] include a tarball of docs during deployment * [DOCS] Add a short security faq
Tianqi Chen committed
-
- 21 Mar, 2020 3 commits
-
-
* [KERAS]Conv3d support added * Keras conv3d testcase added
Samuel committed -
* Update relay docs * any -> py:func * make clean
Zhi committed -
Yagna Srinath Reddy Battula committed
-
- 20 Mar, 2020 8 commits
-
-
* add Atan/Atan2 op * fix bug and testing
Andrew Liu committed -
masahi committed
-
* bind constant to subgraphs * con -> constant
Zhi committed -
* fix CompilerAttrs * retrigger ci
Zhi committed -
* [TOPI][OP] Use Thrust sort for argsort and topk The current GPU sort implementation (odd-even transposition sort) is too slow when the number of elements is large. This PR introduces Thrust implementation of sort which is much faster. Note that this change requires CMake 3.8 or later since we have to use nvcc to compile a thrust code. * cmake: make CUDA optional * allow .cu file to be into the repository * pylint fix and cleanup * require cmake 3.8 only when thrust is enabled * fix nvcc compiler error when passing -pthread * add missing include * add USE_THRUST option in config.cmake * retrigger CI * retrigger CI
MORITA Kazutaka committed -
* Temporary fix to the stack overflow issue in autotvm task extraction * fix lint * fix graph tuner test
Haichen Shen committed -
yongfeng-nv committed
-
As part of the unified IR refactor. This PR refactors the target codegen to use IRModule containing tir::PrimFuncs. In order to break the refactor into several steps without breaking the codebase, we built an conversion pass to convert Array<LoweredFunc> into IRModule. The follow-up refactors will gradually move the passes covered by IRModule up until we cover all the passes. Then we can remove the additional redundant concepts such as LoweredFunc.
Tianqi Chen committed
-
- 19 Mar, 2020 4 commits
-
-
* fix minor lint issue * add conv3d and adaptive avg pool3d conversion with test * fix max pool handling * add batch norm 3d test * add resnet 3d test * add more conv3d test * clean up batch norm test * add note on disabling inception v3 test * add more tests * add more tests * fix names
masahi committed -
* [DOC] Add doc for Relay op strategy * update * address more comments * update * update
Haichen Shen committed -
* [Relay][Frontend][ONNX] operator support: NonZero * update * Solve the build fail * solve the build fail * Replace ctx_list with tvm.cpu()
Neo Chien committed -
* [ConvertLayout] Support QNN ops. * Changing layouts to C. * Fixing dilation. * Empty commit. Co-authored-by: Ubuntu <ubuntu@ip-172-31-53-55.us-west-2.compute.internal>
Animesh Jain committed
-
- 18 Mar, 2020 7 commits
-
-
Tianqi Chen committed
-
* add missing upcast to avgpool * add avg pool test
masahi committed -
Change-Id: I015b2c66a50b64d0eb2e9efe336f6c18ea1fdc67
lhutton1 committed -
Samuel committed
-
* [CODEGEN][OPENCL] Explicitly cast min/max operands * retrigger CI
MORITA Kazutaka committed -
Zhi committed
-
Zhi committed
-
- 17 Mar, 2020 4 commits
-
-
Tianqi Chen committed
-
Tianqi Chen committed
-
* refactor relay python * revert relay/ir/*.py to relay * Address comments * remove direct access to analysis and transform namespace
Zhi committed -
* update docs for dilation 2d * dilation2d compute * dilation2d register * dilation2d rel compute * dilation2d strategy * dilation2d attrs * dilation2d generic schedule * dilation2d tf frontend support * dilation2d tf frontend test case * dilation2d test cases * pylint fixes * add exception for cuda target * Update docstring * Update docstring * change rates to dilations * removed unused param * merge master * Update nn.py * Update nn.py
Mahesh Ambule committed
-
- 16 Mar, 2020 4 commits
-
-
* [TFLITE]DepthToSpace and SpaceToDepth op parser support * DepthToSpace and SpaceToDepth testcases * Review comments fixed
Samuel committed -
Samuel committed
-
* Use dummy func when no lowered_funcs exists in Relay mod * Dummy func -> CSourceModule with empty code str * Added comments describing the empty CSouceModule * Always import external modules w/o assertions * Use CSourceModule as a fallback for LLVMModule * Changed cond for target == llvm * Create an empty LLVM module w/o using dummy func * Avoid using IR str concat to create LLVM module * Improved comments for codegen.LLVMModuleCreate * Satisfy the linter for LLVMModuleCreate
Ruizhe Zhao committed -
Yizhi Liu committed
-
- 15 Mar, 2020 4 commits
-
-
Tianqi Chen committed
-
* add stub for nd impl * refactored indices compute * refactored divide step * remove unused variables, add doc * fix lint * add relay op def * add python registration * refactor topi test * update relay tests, but test result is weird * workaround for weird bug * add relay adaptive pool 3d test * add topi tests * update doc for 3d * typo fix * fix lint * add more tests including NDHWC
masahi committed -
No changes seem to be needed to _fused_batch_norm. It just works.
ojotoxy committed -
* [TESTS] Triage the testcases to fit the naming convention of the new namespaces * Remove multiple usage of system lib to avoid test problems
Tianqi Chen committed
-