This repository was archived by the owner on Jan 3, 2023. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 3
upstream merge for quantization scripts changes #562
Open
ashokei
wants to merge
681
commits into
master
Choose a base branch
from
ashoke/upstream_merge030519
base: master
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
* Adding Scala Demo to be run as a part of Nightly CI * Addressed PR feedback : making a profile to fetch nightly jars only on CI * Changed name from scalacidemo to scala_ci_demo * Synchronized the scala-demo and java-demo for nightly CI runs * Pruned the maven command to simply maven install * changed running from ./.sh to bash .sh to be consistent
* fix ssd quantization script error * update readme for ssd * move quantized SSD instructions from quantization/README.md to ssd/README.md * update ssd readme and accuracy * update readme for SSD-vGG16
In current implemenation, we override signal handler regardless if MXNET_USE_SIGNAL_HANDLER=1. This breaks caller process behavior and cause process exit unexpectedly. The example use case is libmxnet.so is loadded into java process via JNI or JNA. JVM will crash due to SegfaultLogger. In this PR, we will not register SegfaultLogger if there is a signal handler registered.
* fix the wrong argument * fix broken link
* update sn_code * update sn_code
…310) * [MXNET-703] Install CUDA 10 compatible cmake This works around a CUDA 10 cmake issue documented here: clab/dynet#1457 This fix is temporary; once an updated cmake package is published to Ubuntu's package repo it may be reverted. * [MXNET-703] Update to TensorRT 5 ONNX IR 3. Fix inference bugs. * [MXNET-703] Describe onnx opsets and major version
* fix the order of error term's operands * address comments
* add mkldnn slice * fix lint * fix lint * mv SliceEx to matrix_op.cc * fix lint * optimize dispatch_mode * retrigger ci * fix indent
* fix bug in nag optimizer ``` grad += wd * weight mom[:] += grad grad[:] += self.momentum * mom weight[:] += -lr * grad ``` This will minus wd*weight twice, but in`state = momentum * state + grad + wd * weight weight = weight - (lr * (grad + momentum * state)) ` only minus once. * fix bug in nag test fix bug in nag test * rewrite nag test * rewrite nag * fix nag with in-place operations * fix nag with in-place operations
* Update c_api_profile.cc * Update c_api_profile.cc
* updated java dependency * update to duplicated java cpu * java gpu update * Updated java dependency version information
* update scripts and tutorial * add the static test for scala package * kill publish test * fix build issue * address comments
- fix 0-based index output to 1-based index close #13786
* add an optional error_msg in engine on_complete callbcak * use dmlc::Error struct to make error population extendable
* fix doc build * Revert "Temporarily disable website testing (#13887)" This reverts commit 9d42812.
* Moves seed_aug parameter to ImageRecParserParam and re-seeds RNG before each augmentation to guarantee reproducibilit * Update image record iterator tests to check the whole iterator not only first image
* Optimize NMS part 2 * Guarding ldg intrinsics
* add MKL-DNN quantized innerproduct * initial qfc with mkldnn * Add MKL-DNN quantized_fully_connected * refactor params order for fullyconnected * update quantized_fully_connected unittest, force data to uint8 type temporary * change mkl based quantized fully_connected to FCompute * add check data type for mkldnn quantized_fc * add fuse requantize and dequantize for mkldnn quantized fullyconnected * add env setting for enable/disable fuse requantize/dequantize for quantize fullyconnected * fix requantize scaling error * add fallback when input data is int8 * fix mkl quantized fullyconnected index error * update quantized fc test cases * add subgraph node for mkldnn fullyconnected * fix compiling and lint error * clean and refactor code * enable quantized_fc for imagenet * cleanup code * Fix StorageType error for non-mkldnn path * fix pylint * reverse BUILD_TAG for MKL IGEMM ut, remove IGEMM qfc check * rename variables and refactor codes according to comments * add subgraph qfc tests and fix shape error * remove fuse_requantize and change fuse_dequantize to enable_float_output. * change to use mxnet::Tuple and update tests * update description in file header * update input0 type check for quantized FullyConnected * fix conflit of mkl/test_subgraph.py * retrigger CI * retrigger CI due to hang
* add pos_weight for SigmoidBinaryCrossEntropyLoss in gluon.loss * Update loss.py * add test add test * set the default value of pos_weight to be 1 * fix unittest * set N be a random number * fix issues * test without random number * test with random N * fix * fix errors * fix errors * fix order * Update loss.py * Update loss.py * fix pylint * default pos_weight=None * add broadcast_mul and fix pylint * fix unittest * Update loss.py * Update loss.py * Update loss.py
- `symbolic-node/type.jl` - `symbolic-node/show.jl` - `symbolic-node/arithmetic.jl` - `symbolic-node/io.jl` - `symbolic-node/array.jl` - `symbolic-node/op.jl` - `symbolic-node/autodiff.jl` See also: #14001
* nan comparison * fix relu grad
* [clojure-package][wip] add `->nd-vec` function in `ndarray.clj` * WIP * Unit tests need to be added * [clojure-package][ndarray] add unit tests for `->nd-vec` util fn
* fix engine crash in shutdown phase * fix lint * Revert "Bypass ThreadedEngine in test_operator_gpu.py:test_convolution_multiple_streams. (#14338)" This reverts commit d6eafca.
- in order to match Julia `Base.clamp` interface - depwarn for `mx.clip` included
* add backgroud class in box_nms * add unittest * trigger CI
* Change the paths of windows tests to match new AMIs * Re-enable windows jenkins steps * Disables test_operator_gpu.test_bulking test
* [Clojure] Helper function for n-dim vector to ndarray * More tests, specs and rename method * Address comments * Allow every number type
…to fix-inner-product
* fix doc * Update symbol.py * Retrigger CI
This reverts commit a9b3886.
* Add dockerfiles for python3 * Fix * Fix the dockerfile names * Update the README with python3 images * Parameterize the mxnet version * Fix typo * Reduce the number of docker layers * Combine build and Test, run builds in parallel. Use variable names instead of numbers * minor fix * Update README for build command * Bug Fix: Script should fail if any background process returns non-zero code * Changes needed for Benchmarking, 4 new tags, change python to python-dev * Minor typos * Minor Rearrangement - sometimes tagging fails * Move error checking to a method * Push BAI images to dockerhub * Create ~/temp dir * Bug Fix * Fix merge issues I don't know if this is tha main merge issue the PR is complaining about on mxnet, but maybe it will help
* Reenable test_operator_gpu.py:test_bulking. * Add temporary debug output. * Add test_gluon_gpu.py:test_bulking to test spawn approach. * Reorder testing to see test_gluon_gpu.py:test_bulking result. * Change test_operator_gpu.py:test_bulking to use spawn approach. * Remove diagnostic output and revert windows ci changes.
* update rnn for fp16 * fix typo in test * fix tests * fix tests * fix gpu tests * Update test_gluon_rnn.py * Update test_gluon_rnn.py * trigger * try removing checks for unix
* Fix relative difference scala * Increase number of cases for scala arange test * Add cases where arange produces NDArray of [0] * Remote whitespace
* Add test exposing issue with conv dgrad algo 3 for some cudnn's. * Add test temporarily to tests run with tensorrt CI build (cuda10, cudnn7.4.2) * Relax tol of new test. * Fix for problematic conv dgrad algo 3 for some cuDNNs. * Add algo exclusion term to cudnnFind result processing. * Revert "Add test temporarily to tests run with tensorrt CI build (cuda10, cudnn7.4.2)" This reverts commit 1cb743bd2313ca5d1845dbdda6cccd9dcaa5d30c. * Trigger CI. * Add link to cuDNN release notes. * Trigger CI.
* add Docs update for MXNet Java * fixed based on comments * Update docs/install/java_setup.md Co-Authored-By: lanking520 <[email protected]> * Update docs/tutorials/java/mxnet_java_on_intellij.md Co-Authored-By: lanking520 <[email protected]>
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description
(Brief description on what this PR is about)
Checklist
Essentials
Please feel free to remove inapplicable items for your PR.
Changes
Comments