1. 20 Mar, 2020 1 commit
  2. 18 Mar, 2020 1 commit
  3. 13 Mar, 2020 1 commit
  4. 12 Mar, 2020 1 commit
  5. 27 Feb, 2020 1 commit
  6. 24 Feb, 2020 1 commit
    • [Relay][AutoTVM] Relay op strategy (#4644) · 623dd208
      * relay op strategy
      
      fix lint
      
      bitpack strategy
      
      bitserial_dense (#6)
      
      * update strategy
      
      * address comments
      
      fix a few topi test
      
      Dense strategy (#5)
      
      * dense
      
      * add biforst; remove comments
      
      * address comment
      
      Refactor x86 conv2d_NCHWc (#4)
      
      * Refactor x86 conv2d
      
      * Add x86 depthwise_conv2d_NCHWc
      
      * Add back topi x86 conv2d_nchw
      
      * Merge x86 conv2d_nchw and conv2d_NCHWc
      
      * Minor fix for x86 conv2d
      
      fix more strategy
      
      Add x86 conv2d_NCHWc_int8 strategy (#8)
      
      * Add x86 conv2d_NCHWc_int8 strategy
      
      * Remove contrib_conv2d_nchwc_int8
      
      * Fix generic conv2d_NCHWc for int8
      
      * Fix topi arm_cpu conv2d_NCHWc_int8
      
      update x86 conv2d
      
      enable specify relay ops to be tuned for autotvm
      
      add cuda conv2d strategy
      
      add conv2d strategy for rocm
      
      add conv2d strategy for hls
      
      add conv2d strategy for arm cpu
      
      add conv2d strategy for mali
      
      add conv2d strategy for bifrost
      
      add conv2d strategy for intel graphics
      
      clean up and fix lint
      
      remove template keys from autotvm
      
      remove 2 in the func name
      
      address comments
      
      fix
      
      * fix bugs
      
      * lint
      
      * address comments
      
      * add name to op implement
      
      * Modify topi tests (#9)
      
      * Add pooling, reorg, softmax and vision
      
      * Add lrn
      
      * fix topi test
      
      * fix more topi test
      
      * lint
      
      * address comments
      
      * x
      
      * fix more tests & bugs
      
      * Modify more tests (#10)
      
      * Modify tests for bitserial_conv2d, bitserial_dense, bitserial_conv2d_rasp and bnn
      
      * Minor fix
      
      * More minor fix
      
      * fix more test
      
      * try to update vta using strategy
      
      * fix cpptest
      
      * x
      
      * fix rebase err
      
      * Fix two tests (#11)
      
      * change autotvm log format
      
      * lint
      
      * minor fix
      
      * try fix vta test
      
      * fix rebase err
      
      * tweak
      
      * tmp hack for vta pass
      
      * fix tutorial
      
      * fix
      
      * fix more tutorials
      
      * fix vta tutorial
      
      * minor
      
      * address comments
      
      * fix
      
      * address comments
      
      * fix cpptest
      
      * fix docs
      
      * change data structure name and api
      
      * address comments
      
      * lint
      
      * fix rebase err
      
      * updates
      
      * fix winograd test
      
      * fix doc
      
      * rebase
      
      * upgrade tophub version number
      
      * fix bug
      
      * re-enable vta tsim test after tophub is upgraded
      
      * fix vta test to use the correct args so the config can be found in tophub
      
      Co-authored-by: Yao Wang <kevinthesunwy@gmail.com>
      Haichen Shen committed
  7. 20 Feb, 2020 1 commit
  8. 18 Feb, 2020 2 commits
  9. 15 Feb, 2020 1 commit
  10. 14 Feb, 2020 1 commit
  11. 13 Feb, 2020 2 commits
  12. 12 Feb, 2020 1 commit
    • [REFACTOR][PY][API-CHANGE] establish tvm.ir, migrate corresponding files (#4862) · a5661611
      * [REFACTOR][PY][API-CHANGE] establish tvm.ir, migrate corresponding relay files.
      
      This PR establishes tvm.ir and migrates the corresponding relay
      files into the new folder.
      
      API Change:
      - relay.Module -> tvm.IRModule
      
      * Update with ADT
      
      * Migrate transform
      
      * address comments
      
      * Migrate module
      
      * Migrate json_compact
      
      * Migrate attrs
      
      * Move LoweredFunc to stmt temporarily
      
      * temp migrate container
      
      * Finish migrate container
      Tianqi Chen committed
  13. 09 Feb, 2020 1 commit
  14. 04 Feb, 2020 1 commit
  15. 29 Jan, 2020 1 commit
  16. 28 Jan, 2020 1 commit
  17. 15 Jan, 2020 2 commits
  18. 11 Jan, 2020 1 commit
    • [Relay][TOPI]Fix meaning of conv2d_transpose output_padding parameter (#4318) · dcf7fbf1
      * Add output_padding to generic
      
      * Add output_padding to the reference impl
      
      * Add output_padding to arm_cpu
      
      * Add output_padding to the test
      
      * Add output_padding for cuda
      
      * Add output_padding for x86
      
      * Make use of the new output_padding argument in Relay
      
      * Adjust conv2d_transpose Relay test
      
      * Fix lint errors
      
      * Fix the VTA declaration of conv2d_transpose
      
      * support for output padding in conv2d transpose
      
      * some output padding will break IR pass
      
      * Fix new conv2d_transpose test
      
      * Update tophub
      
      * Fix conv1d output_padding too.
      
      * Fix the conv1d_transpose reference function.
      
      * Fix the cuda impl
      
      * fix the topi test for conv1d
      
      * Update the versions in tophub.py
      
      Co-authored-by: Thierry Moreau <tmoreau@octoml.ai>
      abergeron committed
  19. 10 Jan, 2020 1 commit
  20. 09 Jan, 2020 1 commit
  21. 27 Dec, 2019 1 commit
  22. 26 Dec, 2019 1 commit
  23. 22 Dec, 2019 1 commit
  24. 18 Dec, 2019 1 commit
  25. 16 Dec, 2019 1 commit
  26. 26 Nov, 2019 1 commit
  27. 21 Nov, 2019 1 commit
  28. 19 Nov, 2019 1 commit
  29. 16 Nov, 2019 1 commit
  30. 15 Nov, 2019 1 commit
  31. 11 Nov, 2019 1 commit
    • Add More Shape Functions (#4179) · 62521453
      * Add shape functions
      
      * Fix get_const_tuple
      
      * Fix cpplint
      
      * Fix pylint
      
      * Fix pylint
      
      * rebase and fix
      
      * Check Any for infer type
      
      * Fix expand_dim shape func for zero rank input
      
      * Fix pooling infer type
      
      * Address comment
      
      * Register layout transform attr
      Yao Wang committed
  32. 07 Nov, 2019 1 commit
    • [AutoTVM] Add batch_matmul to tunable operations (#4242) · 14a5a358
      * Batch matmul tuning running but with errors.
      
      * Default x86 schedule as good as before.
      
      * Code Cleanup
      
      * Remove unused argument.
      
      * improved template documentation.
      
      * Silly lint fix
      
      * Removed leftover comment.
      
      * Moved cfg declaration to schedule for batch_matmul
      
      * Moved x86 dense cfg declaration to schedule.
      
      * lint fix
      
      * Removed duplicate cfg declaration in dense.
      
      * Reverted changes to dense.
      Josh Fromm committed
  33. 29 Oct, 2019 1 commit
  34. 24 Oct, 2019 1 commit
  35. 22 Oct, 2019 1 commit
  36. 03 Oct, 2019 1 commit
  37. 01 Oct, 2019 1 commit