1. 12 Mar, 2020 1 commit
  2. 11 Mar, 2020 2 commits
    • Conv3D ONNX support and conv3D_ncdhw x86 schedules (#4949) · d56829ea
      * Support 3d Convolution with the ONNX frontend
      
      * add unit tests for conv3d in onnx frontend
      
      respond to PR formatting requests
      
      add x86 schedules to conv3d ncdhw test
      
      fix a doc string format issue
      
      refactor for changed upsream API
      
      * first attempt at conv3d autotuning
      
      add default schedule for conv3d_ncdhw
      
      fill in autotvm integration
      
      add a fallback for invalid schedules
      
      fix fallback
      
      fix reduction order to get simd working correctly
      Matthew Brookhart committed
    • [topi][relay] new PR to re-add tan to TVM (#5025) · 45ee7b5f
      * Add relay operation relay.op.tan.
      
      * Update tan implementation in TVM.
      
      * Update tests.
      
      * Add shape function for tan.
      
      * Add missing main test to python/frontend/tensorflow/test_forward.
      
      * Revert, back to sin/cos.
      
      * Revert "Revert, back to sin/cos."
      
      This reverts commit 4da5b503b921585ba9d80944b29136142b575c40.
      
      * Fix implementation of tan in cuda. Do not support tan for float16.
      
      Simplify topi/tests/python/test_topi_math. Add testing for tan with float32 and float64.
      
      Finally implement tan as sin/cos in llvm.
      notoraptor committed
  3. 10 Mar, 2020 1 commit
  4. 06 Mar, 2020 1 commit
    • [topi][relay] add operation tan to TVM (#4938) · d992468d
      * Add relay operation relay.op.tan.
      
      * Update tan implementation in TVM.
      
      * Update tests.
      
      * Add shape function for tan.
      
      * Add missing main test to python/frontend/tensorflow/test_forward.
      
      * Revert, back to sin/cos.
      
      * Revert "Revert, back to sin/cos."
      
      This reverts commit 4da5b503b921585ba9d80944b29136142b575c40.
      
      * Fix implementation of tan in cuda. Do not support tan for float16.
      
      Simplify topi/tests/python/test_topi_math. Add testing for tan with float32 and float64.
      
      Try again to implement tan as sin/cos in llvm.
      Yao Wang committed
  5. 01 Mar, 2020 2 commits
  6. 27 Feb, 2020 1 commit
  7. 25 Feb, 2020 1 commit
  8. 24 Feb, 2020 1 commit
    • [Relay][AutoTVM] Relay op strategy (#4644) · 623dd208
      * relay op strategy
      
      fix lint
      
      bitpack strategy
      
      bitserial_dense (#6)
      
      * update strategy
      
      * address comments
      
      fix a few topi test
      
      Dense strategy (#5)
      
      * dense
      
      * add biforst; remove comments
      
      * address comment
      
      Refactor x86 conv2d_NCHWc (#4)
      
      * Refactor x86 conv2d
      
      * Add x86 depthwise_conv2d_NCHWc
      
      * Add back topi x86 conv2d_nchw
      
      * Merge x86 conv2d_nchw and conv2d_NCHWc
      
      * Minor fix for x86 conv2d
      
      fix more strategy
      
      Add x86 conv2d_NCHWc_int8 strategy (#8)
      
      * Add x86 conv2d_NCHWc_int8 strategy
      
      * Remove contrib_conv2d_nchwc_int8
      
      * Fix generic conv2d_NCHWc for int8
      
      * Fix topi arm_cpu conv2d_NCHWc_int8
      
      update x86 conv2d
      
      enable specify relay ops to be tuned for autotvm
      
      add cuda conv2d strategy
      
      add conv2d strategy for rocm
      
      add conv2d strategy for hls
      
      add conv2d strategy for arm cpu
      
      add conv2d strategy for mali
      
      add conv2d strategy for bifrost
      
      add conv2d strategy for intel graphics
      
      clean up and fix lint
      
      remove template keys from autotvm
      
      remove 2 in the func name
      
      address comments
      
      fix
      
      * fix bugs
      
      * lint
      
      * address comments
      
      * add name to op implement
      
      * Modify topi tests (#9)
      
      * Add pooling, reorg, softmax and vision
      
      * Add lrn
      
      * fix topi test
      
      * fix more topi test
      
      * lint
      
      * address comments
      
      * x
      
      * fix more tests & bugs
      
      * Modify more tests (#10)
      
      * Modify tests for bitserial_conv2d, bitserial_dense, bitserial_conv2d_rasp and bnn
      
      * Minor fix
      
      * More minor fix
      
      * fix more test
      
      * try to update vta using strategy
      
      * fix cpptest
      
      * x
      
      * fix rebase err
      
      * Fix two tests (#11)
      
      * change autotvm log format
      
      * lint
      
      * minor fix
      
      * try fix vta test
      
      * fix rebase err
      
      * tweak
      
      * tmp hack for vta pass
      
      * fix tutorial
      
      * fix
      
      * fix more tutorials
      
      * fix vta tutorial
      
      * minor
      
      * address comments
      
      * fix
      
      * address comments
      
      * fix cpptest
      
      * fix docs
      
      * change data structure name and api
      
      * address comments
      
      * lint
      
      * fix rebase err
      
      * updates
      
      * fix winograd test
      
      * fix doc
      
      * rebase
      
      * upgrade tophub version number
      
      * fix bug
      
      * re-enable vta tsim test after tophub is upgraded
      
      * fix vta test to use the correct args so the config can be found in tophub
      
      Co-authored-by: Yao Wang <kevinthesunwy@gmail.com>
      Haichen Shen committed
  9. 21 Feb, 2020 2 commits
  10. 20 Feb, 2020 1 commit
  11. 17 Feb, 2020 1 commit
  12. 16 Feb, 2020 1 commit
    • [CodeGen][CUDA] Fix issues in cuda codegen (#4876) · d50ba721
      - Do not emit __shared__ etc. as part of type for casting
      
      - Fix fp16 reduction kernels with compiler errors:
      
        "no operator "+" matches these operands, volatile half + volatile half
      
        This patch inserts casts to remove volatile type qualifier following
        volatile loads (fp16 only). CUDA fp16 library headers should add
        volatile member functions.
      
      - Update have_fp16 to include compute 6.1 GPUs, which do support fp16,
        although their fp16 throughput is low. Updated tests.
      
      Signed-off-by: Wei Pan <weip@nvidia.com>
      wpan11nv committed
  13. 14 Feb, 2020 2 commits
  14. 13 Feb, 2020 2 commits
  15. 12 Feb, 2020 1 commit
    • [REFACTOR][PY][API-CHANGE] establish tvm.ir, migrate corresponding files (#4862) · a5661611
      * [REFACTOR][PY][API-CHANGE] establish tvm.ir, migrate corresponding relay files.
      
      This PR establishes tvm.ir and migrates the corresponding relay
      files into the new folder.
      
      API Change:
      - relay.Module -> tvm.IRModule
      
      * Update with ADT
      
      * Migrate transform
      
      * address comments
      
      * Migrate module
      
      * Migrate json_compact
      
      * Migrate attrs
      
      * Move LoweredFunc to stmt temporarily
      
      * temp migrate container
      
      * Finish migrate container
      Tianqi Chen committed
  16. 11 Feb, 2020 1 commit
  17. 10 Feb, 2020 1 commit
  18. 09 Feb, 2020 1 commit
  19. 07 Feb, 2020 2 commits
  20. 06 Feb, 2020 1 commit
  21. 05 Feb, 2020 3 commits
    • [Relay] Conv2D padding representation (#4787) · 5ea4f0d5
      * enforce 4-way padding
      
      * add util with get_pad_tuple
      
      * delete unnecessary arguments
      
      * fix lint
      
      * add container.Array case
      
      * fix cudnn conv2d asymmetric padding logic
      
      * rename get_pad_tuple to get_pad_tuple2d
      
      * revert change for topi/python/topi/nn/conv2d.py
      
      * add get_pad_tuple2d for several contrib conv2d ops
      
      * add get_pad_tuple2d for all conv2d ops
      Xingyu Zhou committed
    • [REFACTOR][PY] Establish tvm.runtime (#4818) · fc7dd6d7
      * [REFACTOR][PY] Establish tvm.runtime
      
      This PR establishes the tvm.runtime namespace that contains the core runtime data structures.
      The top-level API are kept inact for now via re-exporting.
      
      We will followup later to cleanup some of the top-level APIs.
      
      * Fix ndarray name
      Tianqi Chen committed
    • [REFACTOR][PY] tvm._ffi (#4813) · f9b46c43
      * [REFACTOR][PY] tvm._ffi
      
      - Remove from __future__ import absolute_import in the related files as they are no longer needed if the code only runs in python3
      - Remove reverse dependency of _ctypes _cython to object_generic.
      - function.py -> packed_func.py
      - Function -> PackedFunc
      - all registry related logics goes to tvm._ffi.registry
      - Use absolute references for FFI related calls.
        - tvm._ffi.register_object
        - tvm._ffi.register_func
        - tvm._ffi.get_global_func
      
      * Move get global func to the ffi side
      Tianqi Chen committed
  22. 04 Feb, 2020 1 commit
  23. 03 Feb, 2020 1 commit
  24. 01 Feb, 2020 1 commit
  25. 31 Jan, 2020 1 commit
  26. 24 Jan, 2020 2 commits
  27. 22 Jan, 2020 2 commits
  28. 21 Jan, 2020 1 commit
  29. 20 Jan, 2020 1 commit
  30. 19 Jan, 2020 1 commit