Commit 660ae5ac authored by Leona C's avatar Leona C Committed by Scott Cyphers

Update doc to 0.28 (#4166)

* Update sitemap to not use a page title

* Document Release Notes for 0.28.0; update section on Distributed Training
Co-authored-by: 's avatarScott Cyphers <diyessi@users.noreply.github.com>
parent 19a7d710
......@@ -73,11 +73,11 @@ author = 'Intel Corporation'
# built documents.
#
# The short X.Y version.
version = '0.27'
version = '0.28'
# The Documentation full version, including alpha/beta/rc tags. Some features
# available in the latest code will not necessarily be documented first
release = '0.27.1'
release = '0.28.0'
# The language for content autogenerated by Sphinx. Refer to documentation
# for a list of supported languages.
......@@ -143,6 +143,7 @@ html_sidebars = {
]
}
html_last_updated_fmt= ''
# -- Options for HTMLHelp output ------------------------------------------
......
......@@ -43,6 +43,6 @@ Outputs
C++ Interface
=============
.. doxygenclass:: ngraph::op::Constant
.. doxygenclass:: ngraph::op::v0::Constant
:project: ngraph
:members:
......@@ -4,7 +4,7 @@
Parameter
#########
.. code-block: cpp
.. code-block:: cpp
Parameter // A function parameter.
......@@ -38,7 +38,9 @@ Outputs
| ``output`` | ``element_type`` | ``shape`` |
+------------+------------------+------------+
A ``Parameter`` produces the value of the tensor passed to the function in the position of the parameter in the function's arguments. The passed tensor must have the element type and shape specified by the parameter.
A ``Parameter`` produces the value of the tensor passed to the function
in the position of the parameter in the function's arguments. The passed
tensor must have the element type and shape specified by the parameter.
Backprop
========
......@@ -51,6 +53,6 @@ Backprop
C++ Interface
=============
.. doxygenclass:: ngraph::op::Parameter
.. doxygenclass:: ngraph::op::v0::Parameter
:project: ngraph
:members:
......@@ -45,6 +45,6 @@ Mathematical Definition
C++ Interface
=============
.. doxygenclass:: ngraph::op::Result
.. doxygenclass:: ngraph::op::v0::Result
:project: ngraph
:members:
......@@ -21,19 +21,15 @@ We are pleased to announce the release of version |version|.
Core updates for |version|
--------------------------
+ New ops
+ Provenance improvements from 0.25.1
+ More dynamic shape ops
+ More informative errors
Latest documentation updates
----------------------------
+ Additional details on quantization
+ Index updates
+ API updates
+ Dynamic Shapes and APIs
+ Provenance
+ Add linkages and overview for quantization APIs
+ New ngraph.ai themed illustrations
.. important:: Pre-releases (``-rc-0.*``) have newer features, and are less stable.
......@@ -42,6 +38,16 @@ Latest documentation updates
Changelog on Previous Releases
==============================
0.27.1
+ Fixes broken serializer for Sum and Product
+ New ops
+ Provenance improvements from 0.25.1
+ More dynamic shape ops
+ More informative errors
+ Additional details on quantization
+ Index updates
+ API updates
+ All ops support ``Output<Node>`` arguments
+ Additional ops
+ ONNX handling unknown domains
......
......@@ -5,10 +5,14 @@
Distributed Training
####################
.. important:: Distributed training for CPU backend is not supported. Distributed
training support is provided only with the Intel® Nervana™ Neural Network Processor
for Training (NNP-T).
.. toctree::
:maxdepth: 1
overview.rst
data_ingest.rst
:orphan:
.. training/overview.rst:
.. _overview:
......@@ -5,28 +7,3 @@
Basic Concepts
==============
.. important:: Distributed training is not officially supported as of version
|version|; however, some configuration options have worked for nGraph
devices in testing environments.
Data scientists with locally-scalable rack or cloud-based resources will likely
find it worthwhile to experiment with different modes or variations of
distributed training. Deployments using nGraph Library with supported backends
can be configured to train with data parallelism and will soon work with model
parallelism. Distributing workloads is increasingly important, as more data and
bigger models mean the ability to :doc:`../core/constructing-graphs/distribute-train`
work with larger and larger datasets, or to work with models having many layers
that aren't designed to fit to a single device.
Distributed training with data parallelism splits the data and each worker
node has the same model; during each iteration, the gradients are aggregated
across all workers with an op that performs "allreduce", and applied to update
the weights.
Using multiple machines helps to scale and speed up deep learning. With large
mini-batch training, one could train ResNet-50 with Imagenet-1k data to the
*Top 5* classifier in minutes using thousands of CPU nodes. See
`arxiv.org/abs/1709.05011`_.
.. _arxiv.org/abs/1709.05011: https://arxiv.org/format/1709.05011
\ No newline at end of file
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment