- 12 Oct, 2018 5 commits
-
-
Adam Procter authored
-
Robert Kimball authored
* update test to verify all header files are complete, meaning they include what they use. * disable
-
Ayan Moitra authored
* Project initialization commit * Added unit tests for 3D tensors for argmax * Refactored reduce to be used by argmax argmin. argmax argmin still has some issues. WIP * [WIP]First working version of ArgMax ArgMin * added reduce buffer for the cudnn api calls * added reduce buffer for the cudnn api calls * Further modifications. Using rvalues to pass enums to build reduce method * more unit tests added * Incorporate Fenglei's comments * Incorporating Chris's first set of comments * small change to test file * Resolving clang issue that was causing argmin test to fail * Incorporate Chris's comments * clang format issue
-
Artur Wojcik authored
Signed-off-by: Artur Wojcik <artur.wojcik@intel.com>
-
Amy Zhuang authored
-
- 11 Oct, 2018 3 commits
-
-
Nick Korovaiko authored
-
Nick Korovaiko authored
-
Robert Kimball authored
* updated unit tests * remove debug comments
-
- 10 Oct, 2018 3 commits
-
-
Fenglei authored
-
Fenglei authored
* update onehot * clang * fix bugs * format * add output_datatype_size to hash * typo * hash
-
Nick Korovaiko authored
* reshape sinking working on mnist_conv * forgot to add reshape_sinking files * refactoring of binary case * Quantize/Dequantize case, fix add case, add assert * address bob and scott's feedback * debug * fix a bug where reshapes are removed too early
-
- 09 Oct, 2018 4 commits
-
-
Adam Procter authored
-
Robert Kimball authored
-
Chris Sullivan authored
* add find algorithm for convolution without extra padding * Use cudnnFind* or cudnnGet* depending on tuning param boolean. Add select function to search the perf results of the cudnn queries. * Formatting. * Algo search no longer binary, now it is either off, a heuristic search (cudnnGet*) or an explicit search (cudnnFind*). * Formatting. * switch to explicit. * Throw if no suitable cudnn algo found. * Formatting * Remove comment.
-
Jayaram Bobba authored
* Added a bounds check for mkldnn layout descriptor creation * Added dims check
-
- 08 Oct, 2018 5 commits
-
-
Adam Procter authored
-
Robert Kimball authored
-
Chris Sullivan authored
* Add pad with fill operator using the outward-in index pattern. * Remove static pad and rename build_pad_dynamic -> build_pad. Update maxpool 1d padding. * Formatting. * Split build_pad_dynamic into build_pad and build_pad_fill. * Add test coverage for fixed bug in op::Pad for gpu.
-
Jayaram Bobba authored
* Reshape optimizations for when unit-sized dimensions are added/removed from tensors * Added unit tests for eliminating squeeze and expand_dims operations * Bug fix to expand dims layout * Style fix
-
Jayaram Bobba authored
* Check output shape when setting memory layout for slice op. * Miscellaneous fusion and other optimizations for inception-resnetv2 - ConvBias Batchnorm folding - ConvBias Affine folding - Check if MKLDNN can slice a given layout and select layouts appropriately * Fixed unit test and bug in conv bias pattern * Addressed PR feedback * Addressed PR feedback
-
- 06 Oct, 2018 2 commits
-
-
gcwenger authored
* Eliminated two warnings introduced in #1459 * Removed unnecessary call to reserve_workspace.
-
VINOD KUMAR DEVARAMPATI authored
* added constant folding for dequantize * modified as per review comments
-
- 05 Oct, 2018 13 commits
-
-
gcwenger authored
* LRN WIP * Explicit lambda captures. * Switched to Ayan's new caching routine. * Remove commented out lrn from manifest. * Fixed clang 3.9 error. * Corrected lrn hash. Only call cudnnSetLRNDescriptor once. * Simplified lrn hash. Removed redundant parameters. No longer passing CUDNN_LRN_CROSS_CHANNEL_DIM1 as parameter because it's the only choice for cudnnLRNCrossChannelForward.
-
Jaikrishnan Menon authored
-
Scott Cyphers authored
* More op doc, fix formatting * sqrt, tan * Formatting.
-
Robert Kimball authored
-
Robert Kimball authored
* address klocwork issue * move class init * more klocwork * more klocwork * more klocwork * comment on where the magic number is from * address review comments * address review comments
-
Chris Sullivan authored
* Add op::Sigmoid to nvgpu. * Bring rnn fusion and concat passes over into GPU from IA. This is a temporary move until generalization and gpu specification can occur. * Add LSTM fusion and cudnn inference kernel. Next need recurrent fusion and layer fusion. * Formatting * Removed unecessary extra output from LSTM op (rnn with seq. length = 1, so y = hy). * Add RNN fusion of LSTM cells within a recurrent layer. * Formatting. * Add fusion across RNN layers. * Formatting. * Add algebraic simplification. * Added rnn fusion tests. * Updated conditional on LSTM fusion to better distinguish bound nodes as ht vs xt. * Formatting. * Removed print statements. * Formatting. * Committing missing file. * Remove concat inputs pass and mkldnn references. * fix cmake paths * conflict resolution with merge from master. * remove explicit lstm op support. bare LSTM ops are converted to RNN ops for emission. * Formatting. * Use NGRAPH_ASSERT. Formatting of intel copyright. * Add check on the feature size (shape) of the recurrent (hidden) input and cell state, to ensure they are the same size. * fix wrong rnn header * Formatting. * Add back lstm op to dispatch table. * Added RNN test which shows cudnn rnn kernel is not producing correct results. * With update to AlgSimpl. to simplify concat-reshape-slice, the check modifed in this commit needed to be relaxed. * Bug fix in parameter tensor packing. * Alias third output element of RNN for cell state (bug fix). * Resolve numerical correctness issue with negative values in RNN (bug fix). Add minimal test to evaluate LSTM and compare with values calculated by hand. * Add tensor parameter sizes to kernel hash as they are kernel-specific. * Add 2 layer lstm fusion test against by-hand solution. * Export param concatenation to graph for cudnn kernel at both the single rnn layer and multi-layer. * Formatting. * Finishing touches after merge: add support for macro expansed dispatch via op_tbl. * Simplify macro support for gpu ops. * Add CUDNN_VERSION >= 7200 defguards for RNN fusion. Need to decide how to notify user of increased performance with >= 7200. * Revert lstm_analytic test to explicitly copy data to tensor params. * Removed namespace arg from NGRAPH_GPU_OP. * Refactored macros to different header so op_tbl only contains op list. * Defguard on cudnn_descriptor<cudnnRNNDataDescriptor_t>. * doubles -> floats * Reorg. pass asserts, prepare to replace with non-throwing pass failures. * Remove Lstm op and replace it with Rnn. * Format * Utilize RETURN_IF_FALSE in rnn pass to avoid any RT asserts. Note that falling back to raw (no passes) graph for 2rnn_3lstm json from mxnet models results in a double free inside of the memory layout pass. Appears to be a bug in Reshape pass through. * Removed print statements. Add check on input data and recurrent data. * Don't reuse memory for non-destructive ops. * Add back Rnn test. * Formatting. * Clean up comments. * Update test per review comments.
-
Adam Procter authored
* Add some asserts to make sure we don't overshoot certain iterators in the reference kernels * Add missing assertion.hpp include
-
dmyershov authored
IntelGPU backend: Broadcast bug fix: (output_shape.at(0) == 1) doesn't mean that it is scalar (#1754)
-
Chris Sullivan authored
* global stats fix * Formatting.
-
Robert Kimball authored
* address klocwork number overflow issue * one more issue
-
Robert Kimball authored
-
Robert Kimball authored
-
Adam Procter authored
* Adapt Tensor class to have partial shapes * Add PartialShapes to Input, Output, Function, Node classes * Terminological cleanup
-
- 04 Oct, 2018 5 commits
-
-
Nishant Patel authored
* Add conv+bias * Add test case for QuantizedConv2DWithBiasAndRelu and address feedback
-
Robert Kimball authored
-
Fenglei authored
* add a test failed on gpu, pass on cpu * fixed bug * get datatype size * add descript for test * update comment * update comments and name
-
Nick Korovaiko authored
* show types in visualize_tree * fix a warning * address Bob's feedback
-
Robert Kimball authored
-