summaryrefslogtreecommitdiff
path: root/docs/source/auto_examples/plot_optim_OTreg.rst
diff options
context:
space:
mode:
authorRémi Flamary <remi.flamary@gmail.com>2020-04-21 17:48:37 +0200
committerGitHub <noreply@github.com>2020-04-21 17:48:37 +0200
commita303cc6b483d3cd958c399621e22e40574bcbbc8 (patch)
treedea049cb692020462da8f00d9e117f93b839bb55 /docs/source/auto_examples/plot_optim_OTreg.rst
parent0b2d808aaebb1cab60a272ea7901d5f77df43a9f (diff)
[MRG] Actually run sphinx-gallery (#146)
* generate gallery * remove mock * add sklearn to requirermnt?txt for example * remove latex from fgw example * add networks for graph example * remove all * add requirement.txt rtd * rtd debug * update readme * eradthedoc with redirection * add conf rtd
Diffstat (limited to 'docs/source/auto_examples/plot_optim_OTreg.rst')
-rw-r--r--docs/source/auto_examples/plot_optim_OTreg.rst593
1 files changed, 0 insertions, 593 deletions
diff --git a/docs/source/auto_examples/plot_optim_OTreg.rst b/docs/source/auto_examples/plot_optim_OTreg.rst
deleted file mode 100644
index cd5bcf5..0000000
--- a/docs/source/auto_examples/plot_optim_OTreg.rst
+++ /dev/null
@@ -1,593 +0,0 @@
-.. only:: html
-
- .. note::
- :class: sphx-glr-download-link-note
-
- Click :ref:`here <sphx_glr_download_auto_examples_plot_optim_OTreg.py>` to download the full example code
- .. rst-class:: sphx-glr-example-title
-
- .. _sphx_glr_auto_examples_plot_optim_OTreg.py:
-
-
-==================================
-Regularized OT with generic solver
-==================================
-
-Illustrates the use of the generic solver for regularized OT with
-user-designed regularization term. It uses Conditional gradient as in [6] and
-generalized Conditional Gradient as proposed in [5][7].
-
-
-[5] N. Courty; R. Flamary; D. Tuia; A. Rakotomamonjy, Optimal Transport for
-Domain Adaptation, in IEEE Transactions on Pattern Analysis and Machine
-Intelligence , vol.PP, no.99, pp.1-1.
-
-[6] Ferradans, S., Papadakis, N., Peyré, G., & Aujol, J. F. (2014).
-Regularized discrete optimal transport. SIAM Journal on Imaging Sciences,
-7(3), 1853-1882.
-
-[7] Rakotomamonjy, A., Flamary, R., & Courty, N. (2015). Generalized
-conditional gradient: analysis of convergence and applications.
-arXiv preprint arXiv:1510.06567.
-
-
-
-
-
-.. code-block:: default
-
-
- import numpy as np
- import matplotlib.pylab as pl
- import ot
- import ot.plot
-
-
-
-
-
-
-
-
-Generate data
--------------
-
-
-.. code-block:: default
-
-
- n = 100 # nb bins
-
- # bin positions
- x = np.arange(n, dtype=np.float64)
-
- # Gaussian distributions
- a = ot.datasets.make_1D_gauss(n, m=20, s=5) # m= mean, s= std
- b = ot.datasets.make_1D_gauss(n, m=60, s=10)
-
- # loss matrix
- M = ot.dist(x.reshape((n, 1)), x.reshape((n, 1)))
- M /= M.max()
-
-
-
-
-
-
-
-
-Solve EMD
----------
-
-
-.. code-block:: default
-
-
- G0 = ot.emd(a, b, M)
-
- pl.figure(3, figsize=(5, 5))
- ot.plot.plot1D_mat(a, b, G0, 'OT matrix G0')
-
-
-
-
-.. image:: /auto_examples/images/sphx_glr_plot_optim_OTreg_001.png
- :class: sphx-glr-single-img
-
-
-
-
-
-Solve EMD with Frobenius norm regularization
---------------------------------------------
-
-
-.. code-block:: default
-
-
-
- def f(G):
- return 0.5 * np.sum(G**2)
-
-
- def df(G):
- return G
-
-
- reg = 1e-1
-
- Gl2 = ot.optim.cg(a, b, M, reg, f, df, verbose=True)
-
- pl.figure(3)
- ot.plot.plot1D_mat(a, b, Gl2, 'OT matrix Frob. reg')
-
-
-
-
-.. image:: /auto_examples/images/sphx_glr_plot_optim_OTreg_002.png
- :class: sphx-glr-single-img
-
-
-.. rst-class:: sphx-glr-script-out
-
- Out:
-
- .. code-block:: none
-
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 0|1.760578e-01|0.000000e+00|0.000000e+00
- 1|1.669467e-01|5.457501e-02|9.111116e-03
- 2|1.665639e-01|2.298130e-03|3.827855e-04
- 3|1.664378e-01|7.572776e-04|1.260396e-04
- 4|1.664077e-01|1.811855e-04|3.015066e-05
- 5|1.663912e-01|9.936787e-05|1.653393e-05
- 6|1.663852e-01|3.555826e-05|5.916369e-06
- 7|1.663814e-01|2.305693e-05|3.836245e-06
- 8|1.663785e-01|1.760450e-05|2.929009e-06
- 9|1.663767e-01|1.078011e-05|1.793559e-06
- 10|1.663751e-01|9.525192e-06|1.584755e-06
- 11|1.663737e-01|8.396466e-06|1.396951e-06
- 12|1.663727e-01|6.086938e-06|1.012700e-06
- 13|1.663720e-01|4.042609e-06|6.725769e-07
- 14|1.663713e-01|4.160914e-06|6.922568e-07
- 15|1.663707e-01|3.823502e-06|6.361187e-07
- 16|1.663702e-01|3.022440e-06|5.028438e-07
- 17|1.663697e-01|3.181249e-06|5.292634e-07
- 18|1.663692e-01|2.698532e-06|4.489527e-07
- 19|1.663687e-01|3.258253e-06|5.420712e-07
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 20|1.663682e-01|2.741118e-06|4.560349e-07
- 21|1.663678e-01|2.624135e-06|4.365715e-07
- 22|1.663673e-01|2.645179e-06|4.400714e-07
- 23|1.663670e-01|1.957237e-06|3.256196e-07
- 24|1.663666e-01|2.261541e-06|3.762450e-07
- 25|1.663663e-01|1.851305e-06|3.079948e-07
- 26|1.663660e-01|1.942296e-06|3.231320e-07
- 27|1.663657e-01|2.092896e-06|3.481860e-07
- 28|1.663653e-01|1.924361e-06|3.201470e-07
- 29|1.663651e-01|1.625455e-06|2.704189e-07
- 30|1.663648e-01|1.641123e-06|2.730250e-07
- 31|1.663645e-01|1.566666e-06|2.606377e-07
- 32|1.663643e-01|1.338514e-06|2.226810e-07
- 33|1.663641e-01|1.222711e-06|2.034152e-07
- 34|1.663639e-01|1.221805e-06|2.032642e-07
- 35|1.663637e-01|1.440781e-06|2.396935e-07
- 36|1.663634e-01|1.520091e-06|2.528875e-07
- 37|1.663632e-01|1.288193e-06|2.143080e-07
- 38|1.663630e-01|1.123055e-06|1.868347e-07
- 39|1.663628e-01|1.024487e-06|1.704365e-07
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 40|1.663627e-01|1.079606e-06|1.796061e-07
- 41|1.663625e-01|1.172093e-06|1.949922e-07
- 42|1.663623e-01|1.047880e-06|1.743277e-07
- 43|1.663621e-01|1.010577e-06|1.681217e-07
- 44|1.663619e-01|1.064438e-06|1.770820e-07
- 45|1.663618e-01|9.882375e-07|1.644049e-07
- 46|1.663616e-01|8.532647e-07|1.419505e-07
- 47|1.663615e-01|9.930189e-07|1.652001e-07
- 48|1.663613e-01|8.728955e-07|1.452161e-07
- 49|1.663612e-01|9.524214e-07|1.584459e-07
- 50|1.663610e-01|9.088418e-07|1.511958e-07
- 51|1.663609e-01|7.639430e-07|1.270902e-07
- 52|1.663608e-01|6.662611e-07|1.108397e-07
- 53|1.663607e-01|7.133700e-07|1.186767e-07
- 54|1.663605e-01|7.648141e-07|1.272349e-07
- 55|1.663604e-01|6.557516e-07|1.090911e-07
- 56|1.663603e-01|7.304213e-07|1.215131e-07
- 57|1.663602e-01|6.353809e-07|1.057021e-07
- 58|1.663601e-01|7.968279e-07|1.325603e-07
- 59|1.663600e-01|6.367159e-07|1.059240e-07
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 60|1.663599e-01|5.610790e-07|9.334102e-08
- 61|1.663598e-01|5.787466e-07|9.628015e-08
- 62|1.663596e-01|6.937777e-07|1.154166e-07
- 63|1.663596e-01|5.599432e-07|9.315190e-08
- 64|1.663595e-01|5.813048e-07|9.670555e-08
- 65|1.663594e-01|5.724600e-07|9.523409e-08
- 66|1.663593e-01|6.081892e-07|1.011779e-07
- 67|1.663592e-01|5.948732e-07|9.896260e-08
- 68|1.663591e-01|4.941833e-07|8.221188e-08
- 69|1.663590e-01|5.213739e-07|8.673523e-08
- 70|1.663589e-01|5.127355e-07|8.529811e-08
- 71|1.663588e-01|4.349251e-07|7.235363e-08
- 72|1.663588e-01|5.007084e-07|8.329722e-08
- 73|1.663587e-01|4.880265e-07|8.118744e-08
- 74|1.663586e-01|4.931950e-07|8.204723e-08
- 75|1.663585e-01|4.981309e-07|8.286832e-08
- 76|1.663584e-01|3.952959e-07|6.576082e-08
- 77|1.663584e-01|4.544857e-07|7.560750e-08
- 78|1.663583e-01|4.237579e-07|7.049564e-08
- 79|1.663582e-01|4.382386e-07|7.290460e-08
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 80|1.663582e-01|3.646051e-07|6.065503e-08
- 81|1.663581e-01|4.197994e-07|6.983702e-08
- 82|1.663580e-01|4.072764e-07|6.775370e-08
- 83|1.663580e-01|3.994645e-07|6.645410e-08
- 84|1.663579e-01|4.842721e-07|8.056248e-08
- 85|1.663578e-01|3.276486e-07|5.450691e-08
- 86|1.663578e-01|3.737346e-07|6.217366e-08
- 87|1.663577e-01|4.282043e-07|7.123508e-08
- 88|1.663576e-01|4.020937e-07|6.689135e-08
- 89|1.663576e-01|3.431951e-07|5.709310e-08
- 90|1.663575e-01|3.052335e-07|5.077789e-08
- 91|1.663575e-01|3.500538e-07|5.823407e-08
- 92|1.663574e-01|3.063176e-07|5.095821e-08
- 93|1.663573e-01|3.576367e-07|5.949549e-08
- 94|1.663573e-01|3.224681e-07|5.364492e-08
- 95|1.663572e-01|3.673221e-07|6.110670e-08
- 96|1.663572e-01|3.635561e-07|6.048017e-08
- 97|1.663571e-01|3.527236e-07|5.867807e-08
- 98|1.663571e-01|2.788548e-07|4.638946e-08
- 99|1.663570e-01|2.727141e-07|4.536791e-08
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 100|1.663570e-01|3.127278e-07|5.202445e-08
- 101|1.663569e-01|2.637504e-07|4.387670e-08
- 102|1.663569e-01|2.922750e-07|4.862195e-08
- 103|1.663568e-01|3.076454e-07|5.117891e-08
- 104|1.663568e-01|2.911509e-07|4.843492e-08
- 105|1.663567e-01|2.403398e-07|3.998215e-08
- 106|1.663567e-01|2.439790e-07|4.058755e-08
- 107|1.663567e-01|2.634542e-07|4.382735e-08
- 108|1.663566e-01|2.452203e-07|4.079401e-08
- 109|1.663566e-01|2.852991e-07|4.746137e-08
- 110|1.663565e-01|2.165490e-07|3.602434e-08
- 111|1.663565e-01|2.450250e-07|4.076149e-08
- 112|1.663564e-01|2.685294e-07|4.467159e-08
- 113|1.663564e-01|2.821800e-07|4.694245e-08
- 114|1.663564e-01|2.237390e-07|3.722040e-08
- 115|1.663563e-01|1.992842e-07|3.315219e-08
- 116|1.663563e-01|2.166739e-07|3.604506e-08
- 117|1.663563e-01|2.086064e-07|3.470297e-08
- 118|1.663562e-01|2.435945e-07|4.052346e-08
- 119|1.663562e-01|2.292497e-07|3.813711e-08
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 120|1.663561e-01|2.366209e-07|3.936334e-08
- 121|1.663561e-01|2.138746e-07|3.557935e-08
- 122|1.663561e-01|2.009637e-07|3.343153e-08
- 123|1.663560e-01|2.386258e-07|3.969683e-08
- 124|1.663560e-01|1.927442e-07|3.206415e-08
- 125|1.663560e-01|2.081681e-07|3.463000e-08
- 126|1.663559e-01|1.759123e-07|2.926406e-08
- 127|1.663559e-01|1.890771e-07|3.145409e-08
- 128|1.663559e-01|1.971315e-07|3.279398e-08
- 129|1.663558e-01|2.101983e-07|3.496771e-08
- 130|1.663558e-01|2.035645e-07|3.386414e-08
- 131|1.663558e-01|1.984492e-07|3.301317e-08
- 132|1.663557e-01|1.849064e-07|3.076024e-08
- 133|1.663557e-01|1.795703e-07|2.987255e-08
- 134|1.663557e-01|1.624087e-07|2.701762e-08
- 135|1.663557e-01|1.689557e-07|2.810673e-08
- 136|1.663556e-01|1.644308e-07|2.735399e-08
- 137|1.663556e-01|1.618007e-07|2.691644e-08
- 138|1.663556e-01|1.483013e-07|2.467075e-08
- 139|1.663555e-01|1.708771e-07|2.842636e-08
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 140|1.663555e-01|2.013847e-07|3.350146e-08
- 141|1.663555e-01|1.721217e-07|2.863339e-08
- 142|1.663554e-01|2.027911e-07|3.373540e-08
- 143|1.663554e-01|1.764565e-07|2.935449e-08
- 144|1.663554e-01|1.677151e-07|2.790030e-08
- 145|1.663554e-01|1.351982e-07|2.249094e-08
- 146|1.663553e-01|1.423360e-07|2.367836e-08
- 147|1.663553e-01|1.541112e-07|2.563722e-08
- 148|1.663553e-01|1.491601e-07|2.481358e-08
- 149|1.663553e-01|1.466407e-07|2.439446e-08
- 150|1.663552e-01|1.801524e-07|2.996929e-08
- 151|1.663552e-01|1.714107e-07|2.851507e-08
- 152|1.663552e-01|1.491257e-07|2.480784e-08
- 153|1.663552e-01|1.513799e-07|2.518282e-08
- 154|1.663551e-01|1.354539e-07|2.253345e-08
- 155|1.663551e-01|1.233818e-07|2.052519e-08
- 156|1.663551e-01|1.576219e-07|2.622121e-08
- 157|1.663551e-01|1.452791e-07|2.416792e-08
- 158|1.663550e-01|1.262867e-07|2.100843e-08
- 159|1.663550e-01|1.316379e-07|2.189863e-08
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 160|1.663550e-01|1.295447e-07|2.155041e-08
- 161|1.663550e-01|1.283286e-07|2.134810e-08
- 162|1.663550e-01|1.569222e-07|2.610479e-08
- 163|1.663549e-01|1.172942e-07|1.951247e-08
- 164|1.663549e-01|1.399809e-07|2.328651e-08
- 165|1.663549e-01|1.229432e-07|2.045221e-08
- 166|1.663549e-01|1.326191e-07|2.206184e-08
- 167|1.663548e-01|1.209694e-07|2.012384e-08
- 168|1.663548e-01|1.372136e-07|2.282614e-08
- 169|1.663548e-01|1.338395e-07|2.226484e-08
- 170|1.663548e-01|1.416497e-07|2.356410e-08
- 171|1.663548e-01|1.298576e-07|2.160242e-08
- 172|1.663547e-01|1.190590e-07|1.980603e-08
- 173|1.663547e-01|1.167083e-07|1.941497e-08
- 174|1.663547e-01|1.069425e-07|1.779038e-08
- 175|1.663547e-01|1.217780e-07|2.025834e-08
- 176|1.663547e-01|1.140754e-07|1.897697e-08
- 177|1.663546e-01|1.160707e-07|1.930890e-08
- 178|1.663546e-01|1.101798e-07|1.832892e-08
- 179|1.663546e-01|1.114904e-07|1.854694e-08
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 180|1.663546e-01|1.064022e-07|1.770049e-08
- 181|1.663546e-01|9.258231e-08|1.540149e-08
- 182|1.663546e-01|1.213120e-07|2.018080e-08
- 183|1.663545e-01|1.164296e-07|1.936859e-08
- 184|1.663545e-01|1.188762e-07|1.977559e-08
- 185|1.663545e-01|9.394153e-08|1.562760e-08
- 186|1.663545e-01|1.028656e-07|1.711216e-08
- 187|1.663545e-01|1.115348e-07|1.855431e-08
- 188|1.663544e-01|9.768310e-08|1.625002e-08
- 189|1.663544e-01|1.021806e-07|1.699820e-08
- 190|1.663544e-01|1.086303e-07|1.807113e-08
- 191|1.663544e-01|9.879008e-08|1.643416e-08
- 192|1.663544e-01|1.050210e-07|1.747071e-08
- 193|1.663544e-01|1.002463e-07|1.667641e-08
- 194|1.663543e-01|1.062747e-07|1.767925e-08
- 195|1.663543e-01|9.348538e-08|1.555170e-08
- 196|1.663543e-01|7.992512e-08|1.329589e-08
- 197|1.663543e-01|9.558020e-08|1.590018e-08
- 198|1.663543e-01|9.993772e-08|1.662507e-08
- 199|1.663543e-01|8.588499e-08|1.428734e-08
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 200|1.663543e-01|8.737134e-08|1.453459e-08
-
-
-
-
-Solve EMD with entropic regularization
---------------------------------------
-
-
-.. code-block:: default
-
-
-
- def f(G):
- return np.sum(G * np.log(G))
-
-
- def df(G):
- return np.log(G) + 1.
-
-
- reg = 1e-3
-
- Ge = ot.optim.cg(a, b, M, reg, f, df, verbose=True)
-
- pl.figure(4, figsize=(5, 5))
- ot.plot.plot1D_mat(a, b, Ge, 'OT matrix Entrop. reg')
-
-
-
-
-.. image:: /auto_examples/images/sphx_glr_plot_optim_OTreg_003.png
- :class: sphx-glr-single-img
-
-
-.. rst-class:: sphx-glr-script-out
-
- Out:
-
- .. code-block:: none
-
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 0|1.692289e-01|0.000000e+00|0.000000e+00
- 1|1.617643e-01|4.614437e-02|7.464513e-03
- 2|1.612639e-01|3.102965e-03|5.003963e-04
- 3|1.611291e-01|8.371098e-04|1.348827e-04
- 4|1.610468e-01|5.110558e-04|8.230389e-05
- 5|1.610198e-01|1.672927e-04|2.693743e-05
- 6|1.610130e-01|4.232417e-05|6.814742e-06
- 7|1.610090e-01|2.513455e-05|4.046887e-06
- 8|1.610002e-01|5.443507e-05|8.764057e-06
- 9|1.609996e-01|3.657071e-06|5.887869e-07
- 10|1.609948e-01|2.998735e-05|4.827807e-06
- 11|1.609695e-01|1.569217e-04|2.525962e-05
- 12|1.609533e-01|1.010779e-04|1.626881e-05
- 13|1.609520e-01|8.043897e-06|1.294681e-06
- 14|1.609465e-01|3.415246e-05|5.496718e-06
- 15|1.609386e-01|4.898605e-05|7.883745e-06
- 16|1.609324e-01|3.837052e-05|6.175060e-06
- 17|1.609298e-01|1.617826e-05|2.603564e-06
- 18|1.609184e-01|7.080015e-05|1.139305e-05
- 19|1.609083e-01|6.273206e-05|1.009411e-05
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 20|1.608988e-01|5.940805e-05|9.558681e-06
- 21|1.608853e-01|8.380030e-05|1.348223e-05
- 22|1.608844e-01|5.185045e-06|8.341930e-07
- 23|1.608824e-01|1.279113e-05|2.057868e-06
- 24|1.608819e-01|3.156821e-06|5.078753e-07
- 25|1.608783e-01|2.205746e-05|3.548567e-06
- 26|1.608764e-01|1.189894e-05|1.914259e-06
- 27|1.608755e-01|5.474607e-06|8.807303e-07
- 28|1.608737e-01|1.144227e-05|1.840760e-06
- 29|1.608676e-01|3.775335e-05|6.073291e-06
- 30|1.608638e-01|2.348020e-05|3.777116e-06
- 31|1.608627e-01|6.863136e-06|1.104023e-06
- 32|1.608529e-01|6.110230e-05|9.828482e-06
- 33|1.608487e-01|2.641106e-05|4.248184e-06
- 34|1.608409e-01|4.823638e-05|7.758383e-06
- 35|1.608373e-01|2.256641e-05|3.629519e-06
- 36|1.608338e-01|2.132444e-05|3.429691e-06
- 37|1.608310e-01|1.786649e-05|2.873484e-06
- 38|1.608260e-01|3.103848e-05|4.991794e-06
- 39|1.608206e-01|3.321265e-05|5.341279e-06
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 40|1.608201e-01|3.054747e-06|4.912648e-07
- 41|1.608195e-01|4.198335e-06|6.751739e-07
- 42|1.608193e-01|8.458736e-07|1.360328e-07
- 43|1.608159e-01|2.153759e-05|3.463587e-06
- 44|1.608115e-01|2.738314e-05|4.403523e-06
- 45|1.608108e-01|3.960032e-06|6.368161e-07
- 46|1.608081e-01|1.675447e-05|2.694254e-06
- 47|1.608072e-01|5.976340e-06|9.610383e-07
- 48|1.608046e-01|1.604130e-05|2.579515e-06
- 49|1.608020e-01|1.617036e-05|2.600226e-06
- 50|1.608014e-01|3.957795e-06|6.364188e-07
- 51|1.608011e-01|1.292411e-06|2.078211e-07
- 52|1.607998e-01|8.431795e-06|1.355831e-06
- 53|1.607964e-01|2.127054e-05|3.420225e-06
- 54|1.607947e-01|1.021878e-05|1.643126e-06
- 55|1.607947e-01|3.560621e-07|5.725288e-08
- 56|1.607900e-01|2.929781e-05|4.710793e-06
- 57|1.607890e-01|5.740229e-06|9.229659e-07
- 58|1.607858e-01|2.039550e-05|3.279306e-06
- 59|1.607836e-01|1.319545e-05|2.121612e-06
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 60|1.607826e-01|6.378947e-06|1.025624e-06
- 61|1.607808e-01|1.145102e-05|1.841105e-06
- 62|1.607776e-01|1.941743e-05|3.121889e-06
- 63|1.607743e-01|2.087422e-05|3.356037e-06
- 64|1.607741e-01|1.310249e-06|2.106541e-07
- 65|1.607738e-01|1.682752e-06|2.705425e-07
- 66|1.607691e-01|2.913936e-05|4.684709e-06
- 67|1.607671e-01|1.288855e-05|2.072055e-06
- 68|1.607654e-01|1.002448e-05|1.611590e-06
- 69|1.607641e-01|8.209492e-06|1.319792e-06
- 70|1.607632e-01|5.588467e-06|8.984199e-07
- 71|1.607619e-01|8.050388e-06|1.294196e-06
- 72|1.607618e-01|9.417493e-07|1.513973e-07
- 73|1.607598e-01|1.210509e-05|1.946012e-06
- 74|1.607591e-01|4.392914e-06|7.062009e-07
- 75|1.607579e-01|7.759587e-06|1.247415e-06
- 76|1.607574e-01|2.760280e-06|4.437356e-07
- 77|1.607556e-01|1.146469e-05|1.843012e-06
- 78|1.607550e-01|3.689456e-06|5.930984e-07
- 79|1.607550e-01|4.065631e-08|6.535705e-09
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 80|1.607539e-01|6.555681e-06|1.053852e-06
- 81|1.607528e-01|7.177470e-06|1.153798e-06
- 82|1.607527e-01|5.306068e-07|8.529648e-08
- 83|1.607514e-01|7.816045e-06|1.256440e-06
- 84|1.607511e-01|2.301970e-06|3.700442e-07
- 85|1.607504e-01|4.281072e-06|6.881840e-07
- 86|1.607503e-01|7.821886e-07|1.257370e-07
- 87|1.607480e-01|1.403013e-05|2.255315e-06
- 88|1.607480e-01|1.169298e-08|1.879624e-09
- 89|1.607473e-01|4.235982e-06|6.809227e-07
- 90|1.607470e-01|1.717105e-06|2.760195e-07
- 91|1.607470e-01|6.148402e-09|9.883374e-10
-
-
-
-
-Solve EMD with Frobenius norm + entropic regularization
--------------------------------------------------------
-
-
-.. code-block:: default
-
-
-
- def f(G):
- return 0.5 * np.sum(G**2)
-
-
- def df(G):
- return G
-
-
- reg1 = 1e-3
- reg2 = 1e-1
-
- Gel2 = ot.optim.gcg(a, b, M, reg1, reg2, f, df, verbose=True)
-
- pl.figure(5, figsize=(5, 5))
- ot.plot.plot1D_mat(a, b, Gel2, 'OT entropic + matrix Frob. reg')
- pl.show()
-
-
-
-.. image:: /auto_examples/images/sphx_glr_plot_optim_OTreg_004.png
- :class: sphx-glr-single-img
-
-
-.. rst-class:: sphx-glr-script-out
-
- Out:
-
- .. code-block:: none
-
- It. |Loss |Relative loss|Absolute loss
- ------------------------------------------------
- 0|1.693084e-01|0.000000e+00|0.000000e+00
- 1|1.610202e-01|5.147342e-02|8.288260e-03
- 2|1.609508e-01|4.309685e-04|6.936474e-05
- 3|1.609484e-01|1.524885e-05|2.454278e-06
- 4|1.609477e-01|3.863641e-06|6.218444e-07
- 5|1.609475e-01|1.433633e-06|2.307397e-07
- 6|1.609474e-01|6.332412e-07|1.019185e-07
- 7|1.609474e-01|2.950826e-07|4.749276e-08
- 8|1.609473e-01|1.508393e-07|2.427718e-08
- 9|1.609473e-01|7.859971e-08|1.265041e-08
- 10|1.609473e-01|4.337432e-08|6.980981e-09
- /home/rflamary/PYTHON/POT/examples/plot_optim_OTreg.py:129: UserWarning: Matplotlib is currently using agg, which is a non-GUI backend, so cannot show the figure.
- pl.show()
-
-
-
-
-
-.. rst-class:: sphx-glr-timing
-
- **Total running time of the script:** ( 0 minutes 0.985 seconds)
-
-
-.. _sphx_glr_download_auto_examples_plot_optim_OTreg.py:
-
-
-.. only :: html
-
- .. container:: sphx-glr-footer
- :class: sphx-glr-footer-example
-
-
-
- .. container:: sphx-glr-download sphx-glr-download-python
-
- :download:`Download Python source code: plot_optim_OTreg.py <plot_optim_OTreg.py>`
-
-
-
- .. container:: sphx-glr-download sphx-glr-download-jupyter
-
- :download:`Download Jupyter notebook: plot_optim_OTreg.ipynb <plot_optim_OTreg.ipynb>`
-
-
-.. only:: html
-
- .. rst-class:: sphx-glr-signature
-
- `Gallery generated by Sphinx-Gallery <https://sphinx-gallery.github.io>`_