Skip to content
This repository was archived by the owner on Jan 3, 2023. It is now read-only.

upstream merge for quantization scripts changes #562

Open
wants to merge 681 commits into
base: master
Choose a base branch
from

Conversation

ashokei
Copy link
Contributor

@ashokei ashokei commented Mar 6, 2019

Description

(Brief description on what this PR is about)

Checklist

Essentials

Please feel free to remove inapplicable items for your PR.

  • The PR title starts with [MXNET-$JIRA_ID], where $JIRA_ID refers to the relevant JIRA issue created (except PRs with tiny changes)
  • Changes are complete (i.e. I finished coding on this PR)
  • All changes have test coverage:
  • Unit tests are added for small changes to verify correctness (e.g. adding a new operator)
  • Nightly tests are added for complicated/long-running ones (e.g. changing distributed kvstore)
  • Build tests will be added for build configuration changes (e.g. adding a new build option with NCCL)
  • Code is well-documented:
  • For user-facing API changes, API doc string has been updated.
  • For new C++ functions in header files, their functionalities and arguments are documented.
  • For new examples, README.md is added to explain the what the example does, the source of the dataset, expected performance on test set and reference to the original paper if applicable
  • Check the API doc at http://mxnet-ci-doc.s3-accelerate.dualstack.amazonaws.com/PR-$PR_ID/$BUILD_ID/index.html
  • To the my best knowledge, examples are either not affected by this change, or have been fixed to be compatible with this change

Changes

  • Feature1, tests, (and when applicable, API doc)
  • Feature2, tests, (and when applicable, API doc)

Comments

  • If this change is a backward incompatible change, why must this change be made.
  • Interesting edge cases to note here

piyushghai and others added 30 commits January 14, 2019 18:14
* Adding Scala Demo to be run as a part of Nightly CI

* Addressed PR feedback : making a profile to fetch nightly jars only on CI

* Changed name from scalacidemo to scala_ci_demo

* Synchronized the scala-demo and java-demo for nightly CI runs

* Pruned the maven command to simply maven install

* changed running from ./.sh to bash .sh to be consistent
* fix ssd quantization script error

* update readme for ssd

* move quantized SSD instructions from quantization/README.md to ssd/README.md

* update ssd readme and accuracy

* update readme for SSD-vGG16
In current implemenation, we override signal handler regardless if MXNET_USE_SIGNAL_HANDLER=1.
This breaks caller process behavior and cause process exit unexpectedly.
The example use case is libmxnet.so is loadded into java process via JNI or JNA. JVM will crash
due to SegfaultLogger.

In this PR, we will not register SegfaultLogger if there is a signal handler registered.
* fix the wrong argument

* fix broken link
* update sn_code

* update sn_code
…310)

* [MXNET-703] Install CUDA 10 compatible cmake

This works around a CUDA 10 cmake issue documented here:
clab/dynet#1457

This fix is temporary; once an updated cmake package is published to Ubuntu's
package repo it may be reverted.

* [MXNET-703] Update to TensorRT 5 ONNX IR 3. Fix inference bugs.

* [MXNET-703] Describe onnx opsets and major version
* fix the order of error term's operands

* address comments
* add mkldnn slice

* fix lint

* fix lint

* mv SliceEx to matrix_op.cc

* fix lint

* optimize dispatch_mode

* retrigger ci

* fix indent
* fix bug in nag optimizer

```
grad += wd * weight
mom[:] += grad
grad[:] += self.momentum * mom
weight[:] += -lr * grad
```
This will minus wd*weight twice, but in`state = momentum * state + grad + wd * weight   weight = weight - (lr * (grad + momentum * state)) ` only minus once.

* fix bug in nag test

fix bug in nag test

* rewrite nag test

* rewrite nag

* fix nag with in-place operations

* fix nag with in-place operations
* Update c_api_profile.cc

* Update c_api_profile.cc
* updated java dependency

* update to duplicated java cpu

* java gpu update

* Updated java dependency version information
* update scripts and tutorial

* add the static test for scala package

* kill publish test

* fix build issue

* address comments
- fix 0-based index output to 1-based index

close #13786
* add an optional error_msg in engine on_complete callbcak

* use dmlc::Error struct to make error population extendable
* fix doc build

* Revert "Temporarily disable website testing (#13887)"

This reverts commit 9d42812.
* Moves seed_aug parameter to ImageRecParserParam and re-seeds RNG before each augmentation to guarantee reproducibilit

* Update image record iterator tests to check the whole iterator not only first image
ptrendx and others added 30 commits March 8, 2019 10:57
* Optimize NMS part 2

* Guarding ldg intrinsics
* add MKL-DNN quantized innerproduct

* initial qfc with mkldnn

* Add MKL-DNN quantized_fully_connected

* refactor params order for fullyconnected

* update quantized_fully_connected unittest, force data to uint8 type temporary

* change mkl based quantized fully_connected to FCompute

* add check data type for mkldnn quantized_fc

* add fuse requantize and dequantize for mkldnn quantized fullyconnected

* add env setting for enable/disable fuse requantize/dequantize for quantize fullyconnected

* fix requantize scaling error

* add fallback when input data is int8

* fix mkl quantized fullyconnected index error

* update quantized fc test cases

* add subgraph node for mkldnn fullyconnected

* fix compiling and lint error

* clean and refactor code

* enable quantized_fc for imagenet

* cleanup code

* Fix StorageType error for non-mkldnn path

* fix pylint

* reverse BUILD_TAG for MKL IGEMM ut, remove IGEMM qfc check

* rename variables and refactor codes according to comments

* add subgraph qfc tests and fix shape error

* remove fuse_requantize and change fuse_dequantize to enable_float_output.

* change to use mxnet::Tuple and update tests

* update description in file header

* update input0 type check for quantized FullyConnected

* fix conflit of mkl/test_subgraph.py

* retrigger CI

* retrigger CI due to hang
* add pos_weight for SigmoidBinaryCrossEntropyLoss in gluon.loss

* Update loss.py

* add test

add test

* set the default value of pos_weight to be 1

* fix unittest

* set N be a random number

* fix issues

* test without random number

* test with random N

* fix

* fix errors

* fix errors

* fix order

* Update loss.py

* Update loss.py

* fix pylint

* default pos_weight=None

* add broadcast_mul and fix pylint

* fix unittest

* Update loss.py

* Update loss.py

* Update loss.py
- `symbolic-node/type.jl`
- `symbolic-node/show.jl`
- `symbolic-node/arithmetic.jl`
- `symbolic-node/io.jl`
- `symbolic-node/array.jl`
- `symbolic-node/op.jl`
- `symbolic-node/autodiff.jl`

See also: #14001
* [clojure-package][wip] add `->nd-vec` function in `ndarray.clj`

* WIP
* Unit tests need to be added

* [clojure-package][ndarray] add unit tests for `->nd-vec` util fn
* fix engine crash in shutdown phase

* fix lint

* Revert "Bypass ThreadedEngine in test_operator_gpu.py:test_convolution_multiple_streams. (#14338)"

This reverts commit d6eafca.
- in order to match Julia `Base.clamp` interface

- depwarn for `mx.clip` included
* add backgroud class in box_nms

* add unittest

* trigger CI
* Change the paths of windows tests to match new AMIs

* Re-enable windows jenkins steps

* Disables test_operator_gpu.test_bulking test
* [Clojure] Helper function for n-dim vector to ndarray

* More tests, specs and rename method

* Address comments

* Allow every number type
* fix doc

* Update symbol.py

* Retrigger CI
This reverts commit a9b3886.
* Add dockerfiles for python3

* Fix

* Fix the dockerfile names

* Update the README with python3 images

* Parameterize the mxnet version

* Fix typo

* Reduce the number of docker layers

* Combine build and Test, run builds in parallel. Use variable names instead of numbers

* minor fix

* Update README for build command

* Bug Fix: Script should fail if any background process returns non-zero code

* Changes needed for Benchmarking, 4 new tags, change python to python-dev

* Minor typos

* Minor Rearrangement - sometimes tagging fails

* Move error checking to a method

* Push BAI images to dockerhub

* Create ~/temp dir

* Bug Fix

* Fix merge issues

I don't know if this is tha main merge issue the PR is complaining about on mxnet, but maybe it will help
* Reenable test_operator_gpu.py:test_bulking.

* Add temporary debug output.

* Add test_gluon_gpu.py:test_bulking to test spawn approach.

* Reorder testing to see test_gluon_gpu.py:test_bulking result.

* Change test_operator_gpu.py:test_bulking to use spawn approach.

* Remove diagnostic output and revert windows ci changes.
* update rnn for fp16

* fix typo in test

* fix tests

* fix tests

* fix gpu tests

* Update test_gluon_rnn.py

* Update test_gluon_rnn.py

* trigger

* try removing checks for unix
* Fix relative difference scala

* Increase number of cases for scala arange test

* Add cases where arange produces NDArray of [0]

* Remote whitespace
* Add test exposing issue with conv dgrad algo 3 for some cudnn's.

* Add test temporarily to tests run with tensorrt CI build (cuda10, cudnn7.4.2)

* Relax tol of new test.

* Fix for problematic conv dgrad algo 3 for some cuDNNs.

* Add algo exclusion term to cudnnFind result processing.

* Revert "Add test temporarily to tests run with tensorrt CI build (cuda10, cudnn7.4.2)"

This reverts commit 1cb743bd2313ca5d1845dbdda6cccd9dcaa5d30c.

* Trigger CI.

* Add link to cuDNN release notes.

* Trigger CI.
* add Docs update for MXNet Java

* fixed based on comments

* Update docs/install/java_setup.md

Co-Authored-By: lanking520 <[email protected]>

* Update docs/tutorials/java/mxnet_java_on_intellij.md

Co-Authored-By: lanking520 <[email protected]>
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.