- 23 Apr, 2020 1 commit
-
-
This PR removes ir_pass(old style pass functions) in favor of analysis/transform(new style pass manager).
Tianqi Chen committed
-
- 22 Apr, 2020 1 commit
-
-
Haichen Shen committed
-
- 21 Apr, 2020 1 commit
-
-
The legacy Simplify/CanonicalSimplify are now a thin wrapper around the Analyzer. This PR removes these functions and migrated every place that requires simplification to enforce Analyzer creation. The new API would encourage more Analyzer sharing and potentially enable context-aware analyzer-based simplification.
Tianqi Chen committed
-
- 17 Apr, 2020 1 commit
-
-
Samuel committed
-
- 16 Apr, 2020 1 commit
-
-
Samuel committed
-
- 15 Apr, 2020 1 commit
-
-
[Runtime][Relay][Cleanup] Clean up for memory pass to enable heterogenous execution support. (#5324) * Cleanup type pack and unpack for tuples. * Clean up the memory_pass using common helpers * Clean up memory.cc * Refactor pass * Add doc strings * Fix CPPlint * Fix PyLint * Fix * Apply suggestions from code review Co-Authored-By: Zhi <5145158+zhiics@users.noreply.github.com> * Fix typo Co-authored-by: Zhi <5145158+zhiics@users.noreply.github.com>
Jared Roesch committed
-
- 14 Apr, 2020 1 commit
-
-
* [RELAY][PYTORCH]isNan, isinf, isfinite, ceil, clamp, round ops * Review comments
Samuel committed
-
- 10 Apr, 2020 1 commit
-
-
* Use runtime::String * move string to tvm namespace * add const char* constructor * implicit cast from std::string
Zhi committed
-
- 07 Apr, 2020 1 commit
-
-
* add fast erf * doc * lint * fix * fix indent
Haichen Shen committed
-
- 06 Apr, 2020 1 commit
-
-
Haichen Shen committed
-
- 05 Apr, 2020 1 commit
-
-
* Functional conv3d winograd working. * Formatted python code. * registered conv3d winograd compute and started adding relay without_weight_transform operator. * Add topi testing for conv3d winograd. * Format file. * small tweak to unrolling to prevent build sticking. * Refactoring convolution ops in relay. * Refactored relay convolutions. * Bug fixes. * Fixed static bug in convolution. * Added conv3d alter op layout and related support. * Bug fixes and testing done. * Fix a few autotvm bugs. * Drop silly debug print. * Removed debug_skip_region. * Add variant of conv3d_winograd that doesn't transform depth. * initial infrastructure done for depthless conv. * Fix no_depth schedule bugs. * automatic topi switching between depth and depthless winograd. * Fixed bug in schedule. * lint fixes. * Removed indents in convolution.cc * missed a few indents oops. * fixed flop count. * One more small tweak. * Change kernel pack inner axes order. * Style changes. * Comment fixes.
Josh Fromm committed
-
- 02 Apr, 2020 1 commit
-
-
* [REFACTOR][TIR] Introduce ExprDeepEqual, Remove IRDeepCompare This PR introduces ExprDeepEqual which reuses the StructuralEqual infra. We migrated the usecases of ir_pass::Equal to ExprDeepEqual and StructuralEqual. * Address comments
Tianqi Chen committed
-
- 29 Mar, 2020 1 commit
-
-
Zhi committed
-
- 24 Mar, 2020 1 commit
-
-
* [Torch] Add support for max_pool1d * add test * fix line-too-long * remove wrapper class
Wang Yucheng committed
-
- 23 Mar, 2020 3 commits
-
-
* relay Node::make to constructor * patternwildcard * Address comments
Zhi committed -
* isfinite doc update * isfinit expr * isfinit expr * isfinite schedule reg * isfinite python binding * isfinite python binding * relay register isfinite * isfinite type relation * intrin isfinite * topi isfinite * testcase topi isfinite * tf frontend isfinite * tf frontend isfinite testcase * test case relay isfinite * small fixes * test forward tf isfinite * test cases injective for cuda * remove float16 test case * add support for isinf * remove unwanted import * fix conflict
Mahesh Ambule committed -
* first cut unravel_index * merge fixes * change rates to dilations * unravel_index op relay, topi, mxnet, tf * doc changes * small changes * remove empty unravel and argwhere attrs * remove empty unravel and argwhere attrs
Mahesh Ambule committed
-
- 20 Mar, 2020 2 commits
- 19 Mar, 2020 1 commit
-
-
* [ConvertLayout] Support QNN ops. * Changing layouts to C. * Fixing dilation. * Empty commit. Co-authored-by: Ubuntu <ubuntu@ip-172-31-53-55.us-west-2.compute.internal>
Animesh Jain committed
-
- 17 Mar, 2020 2 commits
-
-
* refactor relay python * revert relay/ir/*.py to relay * Address comments * remove direct access to analysis and transform namespace
Zhi committed -
* update docs for dilation 2d * dilation2d compute * dilation2d register * dilation2d rel compute * dilation2d strategy * dilation2d attrs * dilation2d generic schedule * dilation2d tf frontend support * dilation2d tf frontend test case * dilation2d test cases * pylint fixes * add exception for cuda target * Update docstring * Update docstring * change rates to dilations * removed unused param * merge master * Update nn.py * Update nn.py
Mahesh Ambule committed
-
- 15 Mar, 2020 1 commit
-
-
* add stub for nd impl * refactored indices compute * refactored divide step * remove unused variables, add doc * fix lint * add relay op def * add python registration * refactor topi test * update relay tests, but test result is weird * workaround for weird bug * add relay adaptive pool 3d test * add topi tests * update doc for 3d * typo fix * fix lint * add more tests including NDHWC
masahi committed
-
- 12 Mar, 2020 1 commit
-
-
* [refactor][relay pass] Separate analysis and transform passes into different subfolders * remove pass folder
Zhi committed
-
- 11 Mar, 2020 1 commit
-
-
* Add relay operation relay.op.tan. * Update tan implementation in TVM. * Update tests. * Add shape function for tan. * Add missing main test to python/frontend/tensorflow/test_forward. * Revert, back to sin/cos. * Revert "Revert, back to sin/cos." This reverts commit 4da5b503b921585ba9d80944b29136142b575c40. * Fix implementation of tan in cuda. Do not support tan for float16. Simplify topi/tests/python/test_topi_math. Add testing for tan with float32 and float64. Finally implement tan as sin/cos in llvm.
notoraptor committed
-
- 10 Mar, 2020 1 commit
-
- 06 Mar, 2020 1 commit
-
-
* Add relay operation relay.op.tan. * Update tan implementation in TVM. * Update tests. * Add shape function for tan. * Add missing main test to python/frontend/tensorflow/test_forward. * Revert, back to sin/cos. * Revert "Revert, back to sin/cos." This reverts commit 4da5b503b921585ba9d80944b29136142b575c40. * Fix implementation of tan in cuda. Do not support tan for float16. Simplify topi/tests/python/test_topi_math. Add testing for tan with float32 and float64. Try again to implement tan as sin/cos in llvm.
Yao Wang committed
-
- 01 Mar, 2020 1 commit
-
-
* [Relay][FastMath] Relay pass to use fast exp/tanh * Adding required_pass to the tests. * FastMath test changes.
Animesh Jain committed
-
- 27 Feb, 2020 1 commit
-
-
* move contrib * lint * address comment * address comment
Cody Yu committed
-
- 24 Feb, 2020 1 commit
-
-
* relay op strategy fix lint bitpack strategy bitserial_dense (#6) * update strategy * address comments fix a few topi test Dense strategy (#5) * dense * add biforst; remove comments * address comment Refactor x86 conv2d_NCHWc (#4) * Refactor x86 conv2d * Add x86 depthwise_conv2d_NCHWc * Add back topi x86 conv2d_nchw * Merge x86 conv2d_nchw and conv2d_NCHWc * Minor fix for x86 conv2d fix more strategy Add x86 conv2d_NCHWc_int8 strategy (#8) * Add x86 conv2d_NCHWc_int8 strategy * Remove contrib_conv2d_nchwc_int8 * Fix generic conv2d_NCHWc for int8 * Fix topi arm_cpu conv2d_NCHWc_int8 update x86 conv2d enable specify relay ops to be tuned for autotvm add cuda conv2d strategy add conv2d strategy for rocm add conv2d strategy for hls add conv2d strategy for arm cpu add conv2d strategy for mali add conv2d strategy for bifrost add conv2d strategy for intel graphics clean up and fix lint remove template keys from autotvm remove 2 in the func name address comments fix * fix bugs * lint * address comments * add name to op implement * Modify topi tests (#9) * Add pooling, reorg, softmax and vision * Add lrn * fix topi test * fix more topi test * lint * address comments * x * fix more tests & bugs * Modify more tests (#10) * Modify tests for bitserial_conv2d, bitserial_dense, bitserial_conv2d_rasp and bnn * Minor fix * More minor fix * fix more test * try to update vta using strategy * fix cpptest * x * fix rebase err * Fix two tests (#11) * change autotvm log format * lint * minor fix * try fix vta test * fix rebase err * tweak * tmp hack for vta pass * fix tutorial * fix * fix more tutorials * fix vta tutorial * minor * address comments * fix * address comments * fix cpptest * fix docs * change data structure name and api * address comments * lint * fix rebase err * updates * fix winograd test * fix doc * rebase * upgrade tophub version number * fix bug * re-enable vta tsim test after tophub is upgraded * fix vta test to use the correct args so the config can be found in tophub Co-authored-by: Yao Wang <kevinthesunwy@gmail.com>
Haichen Shen committed
-
- 11 Feb, 2020 1 commit
-
-
hlu1 committed
-
- 06 Feb, 2020 1 commit
-
-
* Add bitwise ops to topi * Add the bitwise ops to relay.
abergeron committed
-
- 23 Jan, 2020 1 commit
-
-
Alexander Pivovarov committed
-
- 22 Jan, 2020 1 commit
-
-
Alexander Pivovarov committed
-
- 21 Jan, 2020 1 commit
-
-
Bring up namespace te -- Tensor expression language DSL.
Tianqi Chen committed
-
- 20 Jan, 2020 1 commit
-
-
* [REFACTOR][TYPE] Finish move all types to IR. - Move definition of Ref and TensorType to ir - Move type_functor.h to public header. - Rename RefType -> RelayRefType for clarity. * Add atol
Tianqi Chen committed
-
- 19 Jan, 2020 2 commits
-
-
This PR moves the codegen related code into the target folder, as they are target specific functionalities. We also adopt the term "compiler driver" in common compiler infra such as rust, GHC and clang. As a result, build_module is moved into the driver folder.
Tianqi Chen committed -
TIR is the new namespace for low-level IR for tensor-level optimizations and loop transformations. This PR establishes the namespace and files. - lowered_func.h,buffer.h,data_layout.h -> tir/buffer.h,tir/data_layout.h,tir/lowered_func.h - ir.h -> tir/expr.h, tir/stmt.h - ir_functor_ext.h -> tir/expr_functor.h, tir/stmt_functor.h
Tianqi Chen committed
-
- 17 Jan, 2020 2 commits
-
-
Move the conversion extensions to the specific class definitions so that we longer need to include packed_func_ext.
Tianqi Chen committed -
- Remove operator bool from base object ref macro - Raitionale: operator bool can be dangerous for sub-classes that also overloads other operators(e.g. ==). - If bool is still needed, use explicit operator bool. - Use absolute include when necessary - Move type related util to data_type - Isolate stackvm code from compiler
Tianqi Chen committed
-