1. 10 Apr, 2020 5 commits
  2. 09 Apr, 2020 1 commit
  3. 08 Apr, 2020 3 commits
    • [BUGFIX][IR] Fix String SEqual (#5275) · ea063888
      * fix String SEqual
      
      * retrigger ci
      Zhi committed
    • [RELAY][BYOC] Add support for composite functions in BYOC (#5261) · d2de35eb
      * [RELAY] Add 'check' functions to MergeComposite
      
      Currently, MergeComposite can only perform structural
      matches. This patch introduces the ability to specify
      a 'check' function alongside the pattern which can include
      custom logic to determine whether an extracted pattern
      should be merged.
      
      For example, if you only want to merge 'NHWC' convolutions,
      you can specify a 'check' function which queries the
      data_layout value of the extracted pattern (see the test).
      
      Change-Id: I9337ce39f10997051a286d888be38ed0d410d340
      
      * [RELAY] Reformat merge_composite.cc
      
      Run clang-format on merge_composite.cc
      
      Change-Id: I1736bff798cc6d93e57519b08ab3362869098779
      
      * [RELAY][BYOC] Support composite functions in AnnotateTarget
      
      This patch introduces support to annotate composite functions
      in the AnnotateTarget pass. In order for a composite function
      to be annotated, you should name it according to the style:
      
      {codegen}.{name}
      eg. dnnl.add_relu
      
      Change-Id: I74d6c0b506153d866f6d1feb203b32dad59f2871
      mbaret committed
  4. 07 Apr, 2020 7 commits
    • [RUNTIME] Implement TVMDSOOp(TensorFlow custom op) for TVM runtime (#4459) · 53a4ad35
      * Add implementation of TVMDSOOp
      
      * feat: Update cmake script to work with c++11 and in-repo build
      
      * feat: Use libtvm as oplib dependency
      
      * fix: Add missing link dependency to libtvm
      
      * feat: Update tf tvmdso op by review comments
      
      * fix: Update with pr comments
      
      * fix: Fix lint
      
      * feat: Add test script and fix gpu shape
      
      * feat: Add test script and fix gpu shape
      
      * fix: Conditional build tftvm op for gpu
      
      * fix: Conditional build tftvm op for gpu
      
      * fix: Fix pylint of tf_op module.py
      
      * fix: Fix pylint of tf_op module.py
      
      * feat: Conditional enable gpu test for tftvm op
      
      * feat: Conditional enable gpu test for tftvm op
      
      * feat: Add tf_tvmdsoop test script as an app test
      
      * fix: Fix gpu/cpu enabled check on tvm in test script
      
      * fix: Make tf tvmdso op test script runnable with pytest
      
      * remove unused test script test_tfop_module.py
      
      * fix: Remove pushd & popd in tfdsoop test script
      
      * fix: Upgrade tftvmop use python3 to find TensorFlow
      
      * fix: Upgrade tftvmop use python3 to find TensorFlow
      
      * fix: Change target_link_options to target_link_libraries
      
      * fix: Add tftvmop build script's c++ option
      
      * fix: Add tvm library path to tf op test library path
      
      * fix: Debug ci build for tftvm dso op
      
      * fix: Fix cmake error and skip tfop test
      
      * fix: Fix typo and indentation issues
      
      * feat: Use TF list input op def
      
      * fix: Fix style and unexpected changes
      
      Co-authored-by: baoxinqi <baoxinqi@4paradigm.com>
      Co-authored-by: Chen Dihao <chendihao@4paradigm.com>
      Co-authored-by: wrongtest <wrongtest@4paradigm.com>
      tobe committed
    • [uTVM][Runtime] Introduce Virtual Memory Allocator to CRT (#5124) · e11a6092
      * initial crt_memory and memory leak fix in graph_runtime
      
      Change-Id: I0f79f909a04d1c677aabb80f202f0612c5ce7f2a
      
      * fix memory leak
      
      Change-Id: I37104c09e28112b1974fa2b064c809d0a8d686c3
      
      * clean up
      
      Change-Id: I039b12015a1d56c8f4120867cd5a5292da34f3e3
      
      * implement vrealloc
      
      Change-Id: I35800470bcbfcf96652494f359711cb4c2d34398
      
      * allocate from stack memory for most of the variables
      
      Change-Id: I72071289843fff4031c0df8796868a0b9fbc57ee
      
      * allocate from stack memory for all of the variables
      
      Change-Id: I32dba85ac1660c77f51c2d0d8ab6436ed0c01c74
      
      * lint
      
      Change-Id: If12cd240685d7791fc60bc0cfb66389cdc186b73
      
      * lint
      
      Change-Id: I7c9d90c11b60b8edda2427ebd189ebe535af2100
      
      * facilitate the growth of TVM_CRT_MAX_NDIM
      
      Change-Id: I939fa43027a5c7529c5c7c6bd8d6e6beb91b7581
      
      * extend test coverage of vmalloc
      
      Change-Id: Ie4ff6b64fdfe6810836cf8fd44dace82a20c4581
      
      * lint
      
      Change-Id: Ibf3c06619ef296df5c49f3945cb6428777781d69
      
      * move logging.h to src
      
      * fix an error in macOS
      
      * remove logging.h
      
      * use cflags for gcc
      
      * fix compilation error
      Liangfu Chen committed
    • [Relay][OP] Add fast_erf implementation (#5241) · f5b02fdb
      * add fast erf
      
      * doc
      
      * lint
      
      * fix
      
      * fix indent
      Haichen Shen committed
    • [TFLITE]Hard Swish & MobilnetV3 model testing (#5239) · 608e9458
      * [TFLITE]Hard Swish & MobilnetV3 model testing
      
      * CI Failure addressed
      Samuel committed
  5. 06 Apr, 2020 5 commits
  6. 05 Apr, 2020 3 commits
    • [Relay][Topi][AutoTVM] Winograd support for Conv3D (#5186) · 02eb1833
      * Functional conv3d winograd working.
      
      * Formatted python code.
      
      * registered conv3d winograd compute and started adding relay without_weight_transform operator.
      
      * Add topi testing for conv3d winograd.
      
      * Format file.
      
      * small tweak to unrolling to prevent build sticking.
      
      * Refactoring convolution ops in relay.
      
      * Refactored relay convolutions.
      
      * Bug fixes.
      
      * Fixed static bug in convolution.
      
      * Added conv3d alter op layout and related support.
      
      * Bug fixes and testing done.
      
      * Fix a few autotvm bugs.
      
      * Drop silly debug print.
      
      * Removed debug_skip_region.
      
      * Add variant of conv3d_winograd that doesn't transform depth.
      
      * initial infrastructure done for depthless conv.
      
      * Fix no_depth schedule bugs.
      
      * automatic topi switching between depth and depthless winograd.
      
      * Fixed bug in schedule.
      
      * lint fixes.
      
      * Removed indents in convolution.cc
      
      * missed a few indents oops.
      
      * fixed flop count.
      
      * One more small tweak.
      
      * Change kernel pack inner axes order.
      
      * Style changes.
      
      * Comment fixes.
      Josh Fromm committed
    • [Relay][ADT]Static Tensor Array (#5103) · b5352ee2
      * Add other static tensor array ops
      
      * Add tensor array get data
      
      * Minor refactor
      
      * Fix pylint
      
      * Update docstring
      
      * Make get data more generic
      
      * Improve test
      
      * Improve split test
      
      * Improve get data
      
      * Minor fix
      
      * Further improvement for static shape
      
      * Improve shape parsing
      
      * Unify get_static_name
      Yao Wang committed
    • [REFACTOR][TIR] Migrate all low-level passes to the Pass Manager. (#5233) · e63e08fe
      * [REFACTOR][TIR] Migrate all low-level passes to the Pass Manager.
      
      This PR migrates the tvm.lower to return IRModule of PrimFuncs
      instead of the LoweredFuncs.
      
      * Remove LoweredFunc.
      Tianqi Chen committed
  7. 04 Apr, 2020 2 commits
  8. 03 Apr, 2020 5 commits
    • [REFACTOR][TIR] Migrate most of low-level build to use the Pass Manager. (#5225) · 75e936e1
      * [REFACTOR][TIR] Migrate most of low-level build to use the Pass Manager.
      
      - SplitHostDevice
      - ThreadSync
      - BindDevice
      - LowerThreadAllreduce
      - Provide a temp fix for printing IRModule with PrimFunc before the formal text printer.
      
      * Address comments, fix tests.
      
      * Fix relay tests
      
      * Explicit move
      Tianqi Chen committed
    • [RELAY] Non-recursive Graph Vistor and Rewriter (#4886) · 7de8a539
      * First pass a defining a non-recursive Graph Vistor and Rewriter
      
      autoformat
      
      remove a currently empty test until testing is solidfied
      
      * Make CalcDep from Dead Code Elimination non-recursive
      
      * Partially working, not passing all tests yet
      
      passes tests when disabling GetExprRefCount, I think I have a bug in visit counting
      
      fix GetExprRefCount
      
      Fix a subtle bug with nested recursive/non-recursive scopes
      
      * Refactor
      
      * improve comments
      
      * respond to review comments on comments
      
      * Fix a problem with default recursion for dataflow nodes
      
      mark DataflowVisitor methods as override
      
      * implement ScopeMutator
      
      * convert forward_rewrite to ScopeMutator, remove DataflowMutator
      
      * rewrite ExprRewriter and convert fast_math to use it
      
      * switch BiasAddSimplifier to ExprRewriter
      
      fix a clang warning
      
      fix cpp lint
      
      fix doc param error
      
      * respond to review comments
      
      * fix a typo in the iterative looping
      
      * add a regression test for GetExprRefCount issue
      
      * Normalize naming
      
      * fix lint
      
      * First pass a defining a non-recursive Graph Vistor and Rewriter
      
      autoformat
      
      remove a currently empty test until testing is solidfied
      
      * Make CalcDep from Dead Code Elimination non-recursive
      
      * Partially working, not passing all tests yet
      
      passes tests when disabling GetExprRefCount, I think I have a bug in visit counting
      
      fix GetExprRefCount
      
      Fix a subtle bug with nested recursive/non-recursive scopes
      
      * Refactor
      
      * improve comments
      
      * respond to review comments on comments
      
      * Fix a problem with default recursion for dataflow nodes
      
      mark DataflowVisitor methods as override
      
      * implement ScopeMutator
      
      * convert forward_rewrite to ScopeMutator, remove DataflowMutator
      
      * rewrite ExprRewriter and convert fast_math to use it
      
      * switch BiasAddSimplifier to ExprRewriter
      
      fix a clang warning
      
      fix cpp lint
      
      fix doc param error
      
      * respond to review comments
      
      * fix a typo in the iterative looping
      
      * add a regression test for GetExprRefCount issue
      
      * Normalize naming
      
      * fix lint
      
      * respond to review comments
      Matthew Brookhart committed
    • [KERAS]Upsample3d & ZeroPadding3d op (#5125) · b796c13c
      * [KERAS]upsampling3d and zeropadding3d op
      
      * [KERAS]upsampling3d and zeropadding3d test case
      
      * Review comments updated
      Samuel committed
    • [CodeGen][CUDA] Fix bugs (#5209) · 316ce055
      - Support vectorized casts
      
      - It is incorrect to extract elements from int8x4 with
      
         0x000000ff & (x >> i * 8)
      
        as this value is of type int in C/C++. If this expression
        is used for sign extensions, the sign bit will be wrong.
        Simply use C style casts instead and sign bits will just work.
      
      Signed-off-by: Wei Pan <weip@nvidia.com>
      Wei Pan committed
  9. 02 Apr, 2020 9 commits