1. 02 Apr, 2020 1 commit
  2. 23 Mar, 2020 1 commit
  3. 27 Feb, 2020 1 commit
  4. 24 Feb, 2020 1 commit
    • [Relay][AutoTVM] Relay op strategy (#4644) · 623dd208
      * relay op strategy
      
      fix lint
      
      bitpack strategy
      
      bitserial_dense (#6)
      
      * update strategy
      
      * address comments
      
      fix a few topi test
      
      Dense strategy (#5)
      
      * dense
      
      * add biforst; remove comments
      
      * address comment
      
      Refactor x86 conv2d_NCHWc (#4)
      
      * Refactor x86 conv2d
      
      * Add x86 depthwise_conv2d_NCHWc
      
      * Add back topi x86 conv2d_nchw
      
      * Merge x86 conv2d_nchw and conv2d_NCHWc
      
      * Minor fix for x86 conv2d
      
      fix more strategy
      
      Add x86 conv2d_NCHWc_int8 strategy (#8)
      
      * Add x86 conv2d_NCHWc_int8 strategy
      
      * Remove contrib_conv2d_nchwc_int8
      
      * Fix generic conv2d_NCHWc for int8
      
      * Fix topi arm_cpu conv2d_NCHWc_int8
      
      update x86 conv2d
      
      enable specify relay ops to be tuned for autotvm
      
      add cuda conv2d strategy
      
      add conv2d strategy for rocm
      
      add conv2d strategy for hls
      
      add conv2d strategy for arm cpu
      
      add conv2d strategy for mali
      
      add conv2d strategy for bifrost
      
      add conv2d strategy for intel graphics
      
      clean up and fix lint
      
      remove template keys from autotvm
      
      remove 2 in the func name
      
      address comments
      
      fix
      
      * fix bugs
      
      * lint
      
      * address comments
      
      * add name to op implement
      
      * Modify topi tests (#9)
      
      * Add pooling, reorg, softmax and vision
      
      * Add lrn
      
      * fix topi test
      
      * fix more topi test
      
      * lint
      
      * address comments
      
      * x
      
      * fix more tests & bugs
      
      * Modify more tests (#10)
      
      * Modify tests for bitserial_conv2d, bitserial_dense, bitserial_conv2d_rasp and bnn
      
      * Minor fix
      
      * More minor fix
      
      * fix more test
      
      * try to update vta using strategy
      
      * fix cpptest
      
      * x
      
      * fix rebase err
      
      * Fix two tests (#11)
      
      * change autotvm log format
      
      * lint
      
      * minor fix
      
      * try fix vta test
      
      * fix rebase err
      
      * tweak
      
      * tmp hack for vta pass
      
      * fix tutorial
      
      * fix
      
      * fix more tutorials
      
      * fix vta tutorial
      
      * minor
      
      * address comments
      
      * fix
      
      * address comments
      
      * fix cpptest
      
      * fix docs
      
      * change data structure name and api
      
      * address comments
      
      * lint
      
      * fix rebase err
      
      * updates
      
      * fix winograd test
      
      * fix doc
      
      * rebase
      
      * upgrade tophub version number
      
      * fix bug
      
      * re-enable vta tsim test after tophub is upgraded
      
      * fix vta test to use the correct args so the config can be found in tophub
      
      Co-authored-by: Yao Wang <kevinthesunwy@gmail.com>
      Haichen Shen committed
  5. 14 Feb, 2020 1 commit
  6. 09 Feb, 2020 1 commit
  7. 04 Jan, 2020 1 commit
  8. 05 Dec, 2019 1 commit
  9. 03 Dec, 2019 1 commit
  10. 11 Nov, 2019 1 commit
    • Add More Shape Functions (#4179) · 62521453
      * Add shape functions
      
      * Fix get_const_tuple
      
      * Fix cpplint
      
      * Fix pylint
      
      * Fix pylint
      
      * rebase and fix
      
      * Check Any for infer type
      
      * Fix expand_dim shape func for zero rank input
      
      * Fix pooling infer type
      
      * Address comment
      
      * Register layout transform attr
      Yao Wang committed
  11. 01 Oct, 2019 1 commit
    • [TOPI]Add op argwhere (#3994) · fa4d3ec6
      * Add op argwhere
      
      * Move shape func to _algorithm.py
      
      * Add lint rule
      
      * Raise exception if rank is not supportted
      
      * move argwhere to transform
      
      * Add argwhere example
      
      * Fix lint
      
      * Add 1-d support
      
      * cleanup
      
      * Add more dtype support
      
      * CR comment
      
      * Improve error message
      
      * Docs
      
      * raise exception
      Wei Chen committed
  12. 27 Sep, 2019 1 commit
  13. 08 Sep, 2019 1 commit
  14. 01 Sep, 2019 1 commit
    • [Relay][Any] Add shape func for dynamic shape (#3606) · eef35a57
      * init shape func in interpreter and vm compiler
      
      * Update interpreter
      
      * fix
      
      * lint
      
      * lint
      
      * fix
      
      * remove hack
      
      * update
      
      * fix
      
      * fix
      
      * update
      
      * address comments & update for shape_of
      
      * fix lint
      
      * update
      
      * fix hybrid
      
      * lint
      
      * fix bug & add take shape func
      
      * lint
      
      * lint
      
      * update
      
      * fix flaky test
      
      * add todo
      Haichen Shen committed
  15. 22 Aug, 2019 1 commit
    • [TOPI][Relay][TensorFlow] Add OneHot operator (#3781) · 554df211
      * Add one-hot to Relay
      
      * topi implementation
      
      * Working
      
      * add topi test
      
      * Add TF test
      
      * Fix check
      
      * fix linting issues
      
      * fix documentation
      
      * Fix documentation
      
      * Add support for on_value, off_value, axis, dtype
      
      * Add full support for axis
      
      * Fix compute and update test_forward
      
      * Move on_value and off_value to inputs
      
      * Add topi test
      
      * Update tests
      
      * Update docs
      
      * Fix style
      
      * re-enable tests
      
      * Add one_hot to mxnet converter
      Jon Soifer committed
  16. 23 Jul, 2019 1 commit
    • {relay,topi}.reinterpret support (#3599) · 2ed31b24
      = Motivation
      
      It's useful to expose the tvm::reinterpret functionality to Relay/TOPI users, as
      this allows them to build (fused) operators leveraging the bitwise
      reinterpretation of an operator. An example is approximate transcendental
      functions, which can be implemented similar to:
      
      ```.py
          def C(x):
              return relay.expr.const(x, "float32")
      
          def approx_exp(x):
              x = relay.minimum(relay.maximum(x, C(-88.0)), C(88.0))
              x = C(127.0) + x * C(1.44269504)
              xf = relay.floor(x)
              i = relay.cast(xf, "int32")
              x = x - xf
              Y = C(0.99992522) + x * (C(0.69583354) + x * (C(0.22606716) + x * C(0.078024523)))
              exponent = relay.left_shift(i, relay.expr.const(23, "int32"))
              exponent = relay.reinterpret(exponent, "float32")
              return exponent * Y
      
          def approx_sigmoid(x):
              # <2.0e-5 absolute error over [-5, 5]
              y = approx_exp(x)
              return y / (y + C(1.0))
      
          def approx_tanh(x):
              # <4.0e-5 absolute error over [-5, 5]
              x = x * C(2.0)
              y = approx_exp(x)
              return (y - C(1.0)) / (y + C(1.0))
      ```
      
      See unit tests for implementations of these approximate transendentals.
      Andrew Tulloch committed
  17. 28 Jun, 2019 1 commit
    • [RELAY] [OP] [MXNet Frontend] Add sequence_mask (#3437) · 8ef22176
      * Add sequence_mask
      
      use exactly the same arguments as mxnet
      
      fix
      
      * fix lint
      
      * fix lint
      
      * add mxnet conversion + relay
      
      * update
      
      * update doc
      
      * fix pylint
      
      * fix doc
      
      * address comment
      
      * try to address comments
      
      * try to enable shape check for valid_length
      
      * fix
      
      * try to fix
      
      * fix bug
      
      * try to fix
      
      * address comment
      
      * address comment
      Xingjian Shi committed
  18. 17 May, 2019 1 commit
  19. 08 Apr, 2019 1 commit
    • [HEADER] Add Header to Comply with ASF Release Policy (#2982) · cffb4fba
      * [HEADER] ASF header dir=include
      
      * [HEADER] ASF Header dir=src
      
      * [HEADER] ASF Header -dir=python
      
      * [HEADER] ASF header dir=topi
      
      * [HEADER] ASF Header dir=nnvm
      
      * [HEADER] ASF Header -dir=tutorials
      
      * [HEADER] ASF Header dir=tests
      
      * [HEADER] ASF Header -dir=docker
      
      * fix whitespace
      
      * [HEADER] ASF Header -dir=jvm
      
      * [HEADER] ASF Header -dir=web
      
      * [HEADER] ASF Header --dir=apps
      
      * [HEADER] ASF Header --dir=vta
      
      * [HEADER] ASF Header -dir=go
      
      * temp
      
      * [HEADER] ASF Header --dir=rust
      
      * [HEADER] Add ASF Header --dir=cmake
      
      * [HEADER] ASF Header --dir=docs
      
      * [HEADER] Header for Jenkinsfile
      
      * [HEADER] ASF Header to toml and md
      
      * [HEADER] ASF Header to gradle
      
      * Finalize rat cleanup
      
      * Fix permission
      
      * Fix java test
      
      * temporary remove nnvm onnx test
      Tianqi Chen committed
  20. 02 Apr, 2019 1 commit
  21. 13 Mar, 2019 1 commit
  22. 11 Mar, 2019 1 commit
  23. 06 Mar, 2019 1 commit
  24. 23 Feb, 2019 1 commit
  25. 04 Feb, 2019 1 commit
  26. 18 Dec, 2018 1 commit
  27. 04 Dec, 2018 1 commit
  28. 30 Nov, 2018 1 commit
    • [Relay] Alter Op Layout (#2150) · 2a5656bf
      * [RELAY] Finish alter op pass
      
      * [RELAY] AlterOpLayout Pass
      
      * fix broadcast operators
      
      * fix broadcast operators
      
      * fix broadcast operators
      
      * Support concatenate
      
      * address comments
      
      * address comments
      
      * add comments
      
      * rebase
      Lianmin Zheng committed
  29. 29 Nov, 2018 1 commit
  30. 28 Nov, 2018 1 commit
  31. 27 Nov, 2018 1 commit
  32. 26 Nov, 2018 1 commit
  33. 25 Nov, 2018 3 commits
  34. 20 Nov, 2018 1 commit
  35. 13 Nov, 2018 1 commit