summaryrefslogtreecommitdiff
path: root/test/test_da.py
diff options
context:
space:
mode:
Diffstat (limited to 'test/test_da.py')
-rw-r--r--test/test_da.py307
1 files changed, 156 insertions, 151 deletions
diff --git a/test/test_da.py b/test/test_da.py
index 9f2bb50..4bf0ab1 100644
--- a/test/test_da.py
+++ b/test/test_da.py
@@ -19,7 +19,32 @@ except ImportError:
nosklearn = True
-def test_sinkhorn_lpl1_transport_class():
+def test_class_jax_tf():
+ backends = []
+ from ot.backend import jax, tf
+ if jax:
+ backends.append(ot.backend.JaxBackend())
+ if tf:
+ backends.append(ot.backend.TensorflowBackend())
+
+ for nx in backends:
+ ns = 150
+ nt = 200
+
+ Xs, ys = make_data_classif('3gauss', ns)
+ Xt, yt = make_data_classif('3gauss2', nt)
+
+ Xs, ys, Xt, yt = nx.from_numpy(Xs, ys, Xt, yt)
+
+ otda = ot.da.SinkhornLpl1Transport()
+
+ with pytest.raises(TypeError):
+ otda.fit(Xs=Xs, ys=ys, Xt=Xt)
+
+
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+def test_sinkhorn_lpl1_transport_class(nx):
"""test_sinkhorn_transport
"""
@@ -29,6 +54,8 @@ def test_sinkhorn_lpl1_transport_class():
Xs, ys = make_data_classif('3gauss', ns)
Xt, yt = make_data_classif('3gauss2', nt)
+ Xs, ys, Xt, yt = nx.from_numpy(Xs, ys, Xt, yt)
+
otda = ot.da.SinkhornLpl1Transport()
# test its computed
@@ -44,15 +71,15 @@ def test_sinkhorn_lpl1_transport_class():
mu_s = unif(ns)
mu_t = unif(nt)
assert_allclose(
- np.sum(otda.coupling_, axis=0), mu_t, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=0)), mu_t, rtol=1e-3, atol=1e-3)
assert_allclose(
- np.sum(otda.coupling_, axis=1), mu_s, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=1)), mu_s, rtol=1e-3, atol=1e-3)
# test transform
transp_Xs = otda.transform(Xs=Xs)
assert_equal(transp_Xs.shape, Xs.shape)
- Xs_new, _ = make_data_classif('3gauss', ns + 1)
+ Xs_new = nx.from_numpy(make_data_classif('3gauss', ns + 1)[0])
transp_Xs_new = otda.transform(Xs_new)
# check that the oos method is working
@@ -62,7 +89,7 @@ def test_sinkhorn_lpl1_transport_class():
transp_Xt = otda.inverse_transform(Xt=Xt)
assert_equal(transp_Xt.shape, Xt.shape)
- Xt_new, _ = make_data_classif('3gauss2', nt + 1)
+ Xt_new = nx.from_numpy(make_data_classif('3gauss2', nt + 1)[0])
transp_Xt_new = otda.inverse_transform(Xt=Xt_new)
# check that the oos method is working
@@ -85,24 +112,26 @@ def test_sinkhorn_lpl1_transport_class():
# test unsupervised vs semi-supervised mode
otda_unsup = ot.da.SinkhornLpl1Transport()
otda_unsup.fit(Xs=Xs, ys=ys, Xt=Xt)
- n_unsup = np.sum(otda_unsup.cost_)
+ n_unsup = nx.sum(otda_unsup.cost_)
otda_semi = ot.da.SinkhornLpl1Transport()
otda_semi.fit(Xs=Xs, ys=ys, Xt=Xt, yt=yt)
assert_equal(otda_semi.cost_.shape, ((Xs.shape[0], Xt.shape[0])))
- n_semisup = np.sum(otda_semi.cost_)
+ n_semisup = nx.sum(otda_semi.cost_)
# check that the cost matrix norms are indeed different
assert n_unsup != n_semisup, "semisupervised mode not working"
# check that the coupling forbids mass transport between labeled source
# and labeled target samples
- mass_semi = np.sum(
+ mass_semi = nx.sum(
otda_semi.coupling_[otda_semi.cost_ == otda_semi.limit_max])
assert mass_semi == 0, "semisupervised mode not working"
-def test_sinkhorn_l1l2_transport_class():
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+def test_sinkhorn_l1l2_transport_class(nx):
"""test_sinkhorn_transport
"""
@@ -112,6 +141,8 @@ def test_sinkhorn_l1l2_transport_class():
Xs, ys = make_data_classif('3gauss', ns)
Xt, yt = make_data_classif('3gauss2', nt)
+ Xs, ys, Xt, yt = nx.from_numpy(Xs, ys, Xt, yt)
+
otda = ot.da.SinkhornL1l2Transport()
# test its computed
@@ -128,15 +159,15 @@ def test_sinkhorn_l1l2_transport_class():
mu_s = unif(ns)
mu_t = unif(nt)
assert_allclose(
- np.sum(otda.coupling_, axis=0), mu_t, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=0)), mu_t, rtol=1e-3, atol=1e-3)
assert_allclose(
- np.sum(otda.coupling_, axis=1), mu_s, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=1)), mu_s, rtol=1e-3, atol=1e-3)
# test transform
transp_Xs = otda.transform(Xs=Xs)
assert_equal(transp_Xs.shape, Xs.shape)
- Xs_new, _ = make_data_classif('3gauss', ns + 1)
+ Xs_new = nx.from_numpy(make_data_classif('3gauss', ns + 1)[0])
transp_Xs_new = otda.transform(Xs_new)
# check that the oos method is working
@@ -156,7 +187,7 @@ def test_sinkhorn_l1l2_transport_class():
assert_equal(transp_ys.shape[0], ys.shape[0])
assert_equal(transp_ys.shape[1], len(np.unique(yt)))
- Xt_new, _ = make_data_classif('3gauss2', nt + 1)
+ Xt_new = nx.from_numpy(make_data_classif('3gauss2', nt + 1)[0])
transp_Xt_new = otda.inverse_transform(Xt=Xt_new)
# check that the oos method is working
@@ -169,22 +200,22 @@ def test_sinkhorn_l1l2_transport_class():
# test unsupervised vs semi-supervised mode
otda_unsup = ot.da.SinkhornL1l2Transport()
otda_unsup.fit(Xs=Xs, ys=ys, Xt=Xt)
- n_unsup = np.sum(otda_unsup.cost_)
+ n_unsup = nx.sum(otda_unsup.cost_)
otda_semi = ot.da.SinkhornL1l2Transport()
otda_semi.fit(Xs=Xs, ys=ys, Xt=Xt, yt=yt)
assert_equal(otda_semi.cost_.shape, ((Xs.shape[0], Xt.shape[0])))
- n_semisup = np.sum(otda_semi.cost_)
+ n_semisup = nx.sum(otda_semi.cost_)
# check that the cost matrix norms are indeed different
assert n_unsup != n_semisup, "semisupervised mode not working"
# check that the coupling forbids mass transport between labeled source
# and labeled target samples
- mass_semi = np.sum(
+ mass_semi = nx.sum(
otda_semi.coupling_[otda_semi.cost_ == otda_semi.limit_max])
mass_semi = otda_semi.coupling_[otda_semi.cost_ == otda_semi.limit_max]
- assert_allclose(mass_semi, np.zeros_like(mass_semi),
+ assert_allclose(nx.to_numpy(mass_semi), np.zeros(list(mass_semi.shape)),
rtol=1e-9, atol=1e-9)
# check everything runs well with log=True
@@ -193,7 +224,9 @@ def test_sinkhorn_l1l2_transport_class():
assert len(otda.log_.keys()) != 0
-def test_sinkhorn_transport_class():
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+def test_sinkhorn_transport_class(nx):
"""test_sinkhorn_transport
"""
@@ -203,6 +236,8 @@ def test_sinkhorn_transport_class():
Xs, ys = make_data_classif('3gauss', ns)
Xt, yt = make_data_classif('3gauss2', nt)
+ Xs, ys, Xt, yt = nx.from_numpy(Xs, ys, Xt, yt)
+
otda = ot.da.SinkhornTransport()
# test its computed
@@ -219,15 +254,15 @@ def test_sinkhorn_transport_class():
mu_s = unif(ns)
mu_t = unif(nt)
assert_allclose(
- np.sum(otda.coupling_, axis=0), mu_t, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=0)), mu_t, rtol=1e-3, atol=1e-3)
assert_allclose(
- np.sum(otda.coupling_, axis=1), mu_s, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=1)), mu_s, rtol=1e-3, atol=1e-3)
# test transform
transp_Xs = otda.transform(Xs=Xs)
assert_equal(transp_Xs.shape, Xs.shape)
- Xs_new, _ = make_data_classif('3gauss', ns + 1)
+ Xs_new = nx.from_numpy(make_data_classif('3gauss', ns + 1)[0])
transp_Xs_new = otda.transform(Xs_new)
# check that the oos method is working
@@ -247,7 +282,7 @@ def test_sinkhorn_transport_class():
assert_equal(transp_ys.shape[0], ys.shape[0])
assert_equal(transp_ys.shape[1], len(np.unique(yt)))
- Xt_new, _ = make_data_classif('3gauss2', nt + 1)
+ Xt_new = nx.from_numpy(make_data_classif('3gauss2', nt + 1)[0])
transp_Xt_new = otda.inverse_transform(Xt=Xt_new)
# check that the oos method is working
@@ -260,19 +295,19 @@ def test_sinkhorn_transport_class():
# test unsupervised vs semi-supervised mode
otda_unsup = ot.da.SinkhornTransport()
otda_unsup.fit(Xs=Xs, Xt=Xt)
- n_unsup = np.sum(otda_unsup.cost_)
+ n_unsup = nx.sum(otda_unsup.cost_)
otda_semi = ot.da.SinkhornTransport()
otda_semi.fit(Xs=Xs, ys=ys, Xt=Xt, yt=yt)
assert_equal(otda_semi.cost_.shape, ((Xs.shape[0], Xt.shape[0])))
- n_semisup = np.sum(otda_semi.cost_)
+ n_semisup = nx.sum(otda_semi.cost_)
# check that the cost matrix norms are indeed different
assert n_unsup != n_semisup, "semisupervised mode not working"
# check that the coupling forbids mass transport between labeled source
# and labeled target samples
- mass_semi = np.sum(
+ mass_semi = nx.sum(
otda_semi.coupling_[otda_semi.cost_ == otda_semi.limit_max])
assert mass_semi == 0, "semisupervised mode not working"
@@ -282,7 +317,9 @@ def test_sinkhorn_transport_class():
assert len(otda.log_.keys()) != 0
-def test_unbalanced_sinkhorn_transport_class():
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+def test_unbalanced_sinkhorn_transport_class(nx):
"""test_sinkhorn_transport
"""
@@ -292,6 +329,8 @@ def test_unbalanced_sinkhorn_transport_class():
Xs, ys = make_data_classif('3gauss', ns)
Xt, yt = make_data_classif('3gauss2', nt)
+ Xs, ys, Xt, yt = nx.from_numpy(Xs, ys, Xt, yt)
+
otda = ot.da.UnbalancedSinkhornTransport()
# test its computed
@@ -318,7 +357,7 @@ def test_unbalanced_sinkhorn_transport_class():
assert_equal(transp_ys.shape[0], ys.shape[0])
assert_equal(transp_ys.shape[1], len(np.unique(yt)))
- Xs_new, _ = make_data_classif('3gauss', ns + 1)
+ Xs_new = nx.from_numpy(make_data_classif('3gauss', ns + 1)[0])
transp_Xs_new = otda.transform(Xs_new)
# check that the oos method is working
@@ -328,7 +367,7 @@ def test_unbalanced_sinkhorn_transport_class():
transp_Xt = otda.inverse_transform(Xt=Xt)
assert_equal(transp_Xt.shape, Xt.shape)
- Xt_new, _ = make_data_classif('3gauss2', nt + 1)
+ Xt_new = nx.from_numpy(make_data_classif('3gauss2', nt + 1)[0])
transp_Xt_new = otda.inverse_transform(Xt=Xt_new)
# check that the oos method is working
@@ -341,12 +380,12 @@ def test_unbalanced_sinkhorn_transport_class():
# test unsupervised vs semi-supervised mode
otda_unsup = ot.da.SinkhornTransport()
otda_unsup.fit(Xs=Xs, Xt=Xt)
- n_unsup = np.sum(otda_unsup.cost_)
+ n_unsup = nx.sum(otda_unsup.cost_)
otda_semi = ot.da.SinkhornTransport()
otda_semi.fit(Xs=Xs, ys=ys, Xt=Xt, yt=yt)
assert_equal(otda_semi.cost_.shape, ((Xs.shape[0], Xt.shape[0])))
- n_semisup = np.sum(otda_semi.cost_)
+ n_semisup = nx.sum(otda_semi.cost_)
# check that the cost matrix norms are indeed different
assert n_unsup != n_semisup, "semisupervised mode not working"
@@ -357,7 +396,9 @@ def test_unbalanced_sinkhorn_transport_class():
assert len(otda.log_.keys()) != 0
-def test_emd_transport_class():
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+def test_emd_transport_class(nx):
"""test_sinkhorn_transport
"""
@@ -367,6 +408,8 @@ def test_emd_transport_class():
Xs, ys = make_data_classif('3gauss', ns)
Xt, yt = make_data_classif('3gauss2', nt)
+ Xs, ys, Xt, yt = nx.from_numpy(Xs, ys, Xt, yt)
+
otda = ot.da.EMDTransport()
# test its computed
@@ -382,15 +425,15 @@ def test_emd_transport_class():
mu_s = unif(ns)
mu_t = unif(nt)
assert_allclose(
- np.sum(otda.coupling_, axis=0), mu_t, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=0)), mu_t, rtol=1e-3, atol=1e-3)
assert_allclose(
- np.sum(otda.coupling_, axis=1), mu_s, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=1)), mu_s, rtol=1e-3, atol=1e-3)
# test transform
transp_Xs = otda.transform(Xs=Xs)
assert_equal(transp_Xs.shape, Xs.shape)
- Xs_new, _ = make_data_classif('3gauss', ns + 1)
+ Xs_new = nx.from_numpy(make_data_classif('3gauss', ns + 1)[0])
transp_Xs_new = otda.transform(Xs_new)
# check that the oos method is working
@@ -410,7 +453,7 @@ def test_emd_transport_class():
assert_equal(transp_ys.shape[0], ys.shape[0])
assert_equal(transp_ys.shape[1], len(np.unique(yt)))
- Xt_new, _ = make_data_classif('3gauss2', nt + 1)
+ Xt_new = nx.from_numpy(make_data_classif('3gauss2', nt + 1)[0])
transp_Xt_new = otda.inverse_transform(Xt=Xt_new)
# check that the oos method is working
@@ -423,28 +466,32 @@ def test_emd_transport_class():
# test unsupervised vs semi-supervised mode
otda_unsup = ot.da.EMDTransport()
otda_unsup.fit(Xs=Xs, ys=ys, Xt=Xt)
- n_unsup = np.sum(otda_unsup.cost_)
+ n_unsup = nx.sum(otda_unsup.cost_)
otda_semi = ot.da.EMDTransport()
otda_semi.fit(Xs=Xs, ys=ys, Xt=Xt, yt=yt)
assert_equal(otda_semi.cost_.shape, ((Xs.shape[0], Xt.shape[0])))
- n_semisup = np.sum(otda_semi.cost_)
+ n_semisup = nx.sum(otda_semi.cost_)
# check that the cost matrix norms are indeed different
assert n_unsup != n_semisup, "semisupervised mode not working"
# check that the coupling forbids mass transport between labeled source
# and labeled target samples
- mass_semi = np.sum(
+ mass_semi = nx.sum(
otda_semi.coupling_[otda_semi.cost_ == otda_semi.limit_max])
mass_semi = otda_semi.coupling_[otda_semi.cost_ == otda_semi.limit_max]
# we need to use a small tolerance here, otherwise the test breaks
- assert_allclose(mass_semi, np.zeros_like(mass_semi),
+ assert_allclose(nx.to_numpy(mass_semi), np.zeros(list(mass_semi.shape)),
rtol=1e-2, atol=1e-2)
-def test_mapping_transport_class():
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+@pytest.mark.parametrize("kernel", ["linear", "gaussian"])
+@pytest.mark.parametrize("bias", ["unbiased", "biased"])
+def test_mapping_transport_class(nx, kernel, bias):
"""test_mapping_transport
"""
@@ -455,101 +502,29 @@ def test_mapping_transport_class():
Xt, yt = make_data_classif('3gauss2', nt)
Xs_new, _ = make_data_classif('3gauss', ns + 1)
- ##########################################################################
- # kernel == linear mapping tests
- ##########################################################################
+ Xs, Xt, Xs_new = nx.from_numpy(Xs, Xt, Xs_new)
- # check computation and dimensions if bias == False
- otda = ot.da.MappingTransport(kernel="linear", bias=False)
+ # Mapping tests
+ bias = bias == "biased"
+ otda = ot.da.MappingTransport(kernel=kernel, bias=bias)
otda.fit(Xs=Xs, Xt=Xt)
assert hasattr(otda, "coupling_")
assert hasattr(otda, "mapping_")
assert hasattr(otda, "log_")
assert_equal(otda.coupling_.shape, ((Xs.shape[0], Xt.shape[0])))
- assert_equal(otda.mapping_.shape, ((Xs.shape[1], Xt.shape[1])))
+ S = Xs.shape[0] if kernel == "gaussian" else Xs.shape[1] # if linear
+ if bias:
+ S += 1
+ assert_equal(otda.mapping_.shape, ((S, Xt.shape[1])))
# test margin constraints
mu_s = unif(ns)
mu_t = unif(nt)
assert_allclose(
- np.sum(otda.coupling_, axis=0), mu_t, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=0)), mu_t, rtol=1e-3, atol=1e-3)
assert_allclose(
- np.sum(otda.coupling_, axis=1), mu_s, rtol=1e-3, atol=1e-3)
-
- # test transform
- transp_Xs = otda.transform(Xs=Xs)
- assert_equal(transp_Xs.shape, Xs.shape)
-
- transp_Xs_new = otda.transform(Xs_new)
-
- # check that the oos method is working
- assert_equal(transp_Xs_new.shape, Xs_new.shape)
-
- # check computation and dimensions if bias == True
- otda = ot.da.MappingTransport(kernel="linear", bias=True)
- otda.fit(Xs=Xs, Xt=Xt)
- assert_equal(otda.coupling_.shape, ((Xs.shape[0], Xt.shape[0])))
- assert_equal(otda.mapping_.shape, ((Xs.shape[1] + 1, Xt.shape[1])))
-
- # test margin constraints
- mu_s = unif(ns)
- mu_t = unif(nt)
- assert_allclose(
- np.sum(otda.coupling_, axis=0), mu_t, rtol=1e-3, atol=1e-3)
- assert_allclose(
- np.sum(otda.coupling_, axis=1), mu_s, rtol=1e-3, atol=1e-3)
-
- # test transform
- transp_Xs = otda.transform(Xs=Xs)
- assert_equal(transp_Xs.shape, Xs.shape)
-
- transp_Xs_new = otda.transform(Xs_new)
-
- # check that the oos method is working
- assert_equal(transp_Xs_new.shape, Xs_new.shape)
-
- ##########################################################################
- # kernel == gaussian mapping tests
- ##########################################################################
-
- # check computation and dimensions if bias == False
- otda = ot.da.MappingTransport(kernel="gaussian", bias=False)
- otda.fit(Xs=Xs, Xt=Xt)
-
- assert_equal(otda.coupling_.shape, ((Xs.shape[0], Xt.shape[0])))
- assert_equal(otda.mapping_.shape, ((Xs.shape[0], Xt.shape[1])))
-
- # test margin constraints
- mu_s = unif(ns)
- mu_t = unif(nt)
- assert_allclose(
- np.sum(otda.coupling_, axis=0), mu_t, rtol=1e-3, atol=1e-3)
- assert_allclose(
- np.sum(otda.coupling_, axis=1), mu_s, rtol=1e-3, atol=1e-3)
-
- # test transform
- transp_Xs = otda.transform(Xs=Xs)
- assert_equal(transp_Xs.shape, Xs.shape)
-
- transp_Xs_new = otda.transform(Xs_new)
-
- # check that the oos method is working
- assert_equal(transp_Xs_new.shape, Xs_new.shape)
-
- # check computation and dimensions if bias == True
- otda = ot.da.MappingTransport(kernel="gaussian", bias=True)
- otda.fit(Xs=Xs, Xt=Xt)
- assert_equal(otda.coupling_.shape, ((Xs.shape[0], Xt.shape[0])))
- assert_equal(otda.mapping_.shape, ((Xs.shape[0] + 1, Xt.shape[1])))
-
- # test margin constraints
- mu_s = unif(ns)
- mu_t = unif(nt)
- assert_allclose(
- np.sum(otda.coupling_, axis=0), mu_t, rtol=1e-3, atol=1e-3)
- assert_allclose(
- np.sum(otda.coupling_, axis=1), mu_s, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=1)), mu_s, rtol=1e-3, atol=1e-3)
# test transform
transp_Xs = otda.transform(Xs=Xs)
@@ -561,29 +536,39 @@ def test_mapping_transport_class():
assert_equal(transp_Xs_new.shape, Xs_new.shape)
# check everything runs well with log=True
- otda = ot.da.MappingTransport(kernel="gaussian", log=True)
+ otda = ot.da.MappingTransport(kernel=kernel, bias=bias, log=True)
otda.fit(Xs=Xs, Xt=Xt)
assert len(otda.log_.keys()) != 0
+
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+def test_mapping_transport_class_specific_seed(nx):
# check that it does not crash when derphi is very close to 0
+ ns = 20
+ nt = 30
np.random.seed(39)
Xs, ys = make_data_classif('3gauss', ns)
Xt, yt = make_data_classif('3gauss2', nt)
otda = ot.da.MappingTransport(kernel="gaussian", bias=False)
- otda.fit(Xs=Xs, Xt=Xt)
+ otda.fit(Xs=nx.from_numpy(Xs), Xt=nx.from_numpy(Xt))
np.random.seed(None)
-def test_linear_mapping():
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+def test_linear_mapping(nx):
ns = 150
nt = 200
Xs, ys = make_data_classif('3gauss', ns)
Xt, yt = make_data_classif('3gauss2', nt)
- A, b = ot.da.OT_mapping_linear(Xs, Xt)
+ Xsb, Xtb = nx.from_numpy(Xs, Xt)
- Xst = Xs.dot(A) + b
+ A, b = ot.da.OT_mapping_linear(Xsb, Xtb)
+
+ Xst = nx.to_numpy(nx.dot(Xsb, A) + b)
Ct = np.cov(Xt.T)
Cst = np.cov(Xst.T)
@@ -591,22 +576,26 @@ def test_linear_mapping():
np.testing.assert_allclose(Ct, Cst, rtol=1e-2, atol=1e-2)
-def test_linear_mapping_class():
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+def test_linear_mapping_class(nx):
ns = 150
nt = 200
Xs, ys = make_data_classif('3gauss', ns)
Xt, yt = make_data_classif('3gauss2', nt)
+ Xsb, Xtb = nx.from_numpy(Xs, Xt)
+
otmap = ot.da.LinearTransport()
- otmap.fit(Xs=Xs, Xt=Xt)
+ otmap.fit(Xs=Xsb, Xt=Xtb)
assert hasattr(otmap, "A_")
assert hasattr(otmap, "B_")
assert hasattr(otmap, "A1_")
assert hasattr(otmap, "B1_")
- Xst = otmap.transform(Xs=Xs)
+ Xst = nx.to_numpy(otmap.transform(Xs=Xsb))
Ct = np.cov(Xt.T)
Cst = np.cov(Xst.T)
@@ -614,7 +603,9 @@ def test_linear_mapping_class():
np.testing.assert_allclose(Ct, Cst, rtol=1e-2, atol=1e-2)
-def test_jcpot_transport_class():
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+def test_jcpot_transport_class(nx):
"""test_jcpot_transport
"""
@@ -627,6 +618,8 @@ def test_jcpot_transport_class():
Xt, yt = make_data_classif('3gauss2', nt)
+ Xs1, ys1, Xs2, ys2, Xt, yt = nx.from_numpy(Xs1, ys1, Xs2, ys2, Xt, yt)
+
Xs = [Xs1, Xs2]
ys = [ys1, ys2]
@@ -649,19 +642,24 @@ def test_jcpot_transport_class():
for i in range(len(Xs)):
# test margin constraints w.r.t. uniform target weights for each coupling matrix
assert_allclose(
- np.sum(otda.coupling_[i], axis=0), mu_t, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_[i], axis=0)), mu_t, rtol=1e-3, atol=1e-3)
# test margin constraints w.r.t. modified source weights for each source domain
assert_allclose(
- np.dot(otda.log_['D1'][i], np.sum(otda.coupling_[i], axis=1)), otda.proportions_, rtol=1e-3,
- atol=1e-3)
+ nx.to_numpy(
+ nx.dot(otda.log_['D1'][i], nx.sum(otda.coupling_[i], axis=1))
+ ),
+ nx.to_numpy(otda.proportions_),
+ rtol=1e-3,
+ atol=1e-3
+ )
# test transform
transp_Xs = otda.transform(Xs=Xs)
[assert_equal(x.shape, y.shape) for x, y in zip(transp_Xs, Xs)]
- Xs_new, _ = make_data_classif('3gauss', ns1 + 1)
+ Xs_new = nx.from_numpy(make_data_classif('3gauss', ns1 + 1)[0])
transp_Xs_new = otda.transform(Xs_new)
# check that the oos method is working
@@ -670,15 +668,16 @@ def test_jcpot_transport_class():
# check label propagation
transp_yt = otda.transform_labels(ys)
assert_equal(transp_yt.shape[0], yt.shape[0])
- assert_equal(transp_yt.shape[1], len(np.unique(ys)))
+ assert_equal(transp_yt.shape[1], len(np.unique(nx.to_numpy(*ys))))
# check inverse label propagation
transp_ys = otda.inverse_transform_labels(yt)
- [assert_equal(x.shape[0], y.shape[0]) for x, y in zip(transp_ys, ys)]
- [assert_equal(x.shape[1], len(np.unique(y))) for x, y in zip(transp_ys, ys)]
+ for x, y in zip(transp_ys, ys):
+ assert_equal(x.shape[0], y.shape[0])
+ assert_equal(x.shape[1], len(np.unique(nx.to_numpy(y))))
-def test_jcpot_barycenter():
+def test_jcpot_barycenter(nx):
"""test_jcpot_barycenter
"""
@@ -695,19 +694,23 @@ def test_jcpot_barycenter():
Xs1, ys1 = make_data_classif('2gauss_prop', ns1, nz=sigma, p=ps1)
Xs2, ys2 = make_data_classif('2gauss_prop', ns2, nz=sigma, p=ps2)
- Xt, yt = make_data_classif('2gauss_prop', nt, nz=sigma, p=pt)
+ Xt, _ = make_data_classif('2gauss_prop', nt, nz=sigma, p=pt)
- Xs = [Xs1, Xs2]
- ys = [ys1, ys2]
+ Xs1b, ys1b, Xs2b, ys2b, Xtb = nx.from_numpy(Xs1, ys1, Xs2, ys2, Xt)
- prop = ot.bregman.jcpot_barycenter(Xs, ys, Xt, reg=.5, metric='sqeuclidean',
+ Xsb = [Xs1b, Xs2b]
+ ysb = [ys1b, ys2b]
+
+ prop = ot.bregman.jcpot_barycenter(Xsb, ysb, Xtb, reg=.5, metric='sqeuclidean',
numItermax=10000, stopThr=1e-9, verbose=False, log=False)
- np.testing.assert_allclose(prop, [1 - pt, pt], rtol=1e-3, atol=1e-3)
+ np.testing.assert_allclose(nx.to_numpy(prop), [1 - pt, pt], rtol=1e-3, atol=1e-3)
@pytest.mark.skipif(nosklearn, reason="No sklearn available")
-def test_emd_laplace_class():
+@pytest.skip_backend("jax")
+@pytest.skip_backend("tf")
+def test_emd_laplace_class(nx):
"""test_emd_laplace_transport
"""
ns = 150
@@ -716,6 +719,8 @@ def test_emd_laplace_class():
Xs, ys = make_data_classif('3gauss', ns)
Xt, yt = make_data_classif('3gauss2', nt)
+ Xs, ys, Xt, yt = nx.from_numpy(Xs, ys, Xt, yt)
+
otda = ot.da.EMDLaplaceTransport(reg_lap=0.01, max_iter=1000, tol=1e-9, verbose=False, log=True)
# test its computed
@@ -732,15 +737,15 @@ def test_emd_laplace_class():
mu_t = unif(nt)
assert_allclose(
- np.sum(otda.coupling_, axis=0), mu_t, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=0)), mu_t, rtol=1e-3, atol=1e-3)
assert_allclose(
- np.sum(otda.coupling_, axis=1), mu_s, rtol=1e-3, atol=1e-3)
+ nx.to_numpy(nx.sum(otda.coupling_, axis=1)), mu_s, rtol=1e-3, atol=1e-3)
# test transform
transp_Xs = otda.transform(Xs=Xs)
[assert_equal(x.shape, y.shape) for x, y in zip(transp_Xs, Xs)]
- Xs_new, _ = make_data_classif('3gauss', ns + 1)
+ Xs_new = nx.from_numpy(make_data_classif('3gauss', ns + 1)[0])
transp_Xs_new = otda.transform(Xs_new)
# check that the oos method is working
@@ -750,7 +755,7 @@ def test_emd_laplace_class():
transp_Xt = otda.inverse_transform(Xt=Xt)
assert_equal(transp_Xt.shape, Xt.shape)
- Xt_new, _ = make_data_classif('3gauss2', nt + 1)
+ Xt_new = nx.from_numpy(make_data_classif('3gauss2', nt + 1)[0])
transp_Xt_new = otda.inverse_transform(Xt=Xt_new)
# check that the oos method is working
@@ -763,9 +768,9 @@ def test_emd_laplace_class():
# check label propagation
transp_yt = otda.transform_labels(ys)
assert_equal(transp_yt.shape[0], yt.shape[0])
- assert_equal(transp_yt.shape[1], len(np.unique(ys)))
+ assert_equal(transp_yt.shape[1], len(np.unique(nx.to_numpy(ys))))
# check inverse label propagation
transp_ys = otda.inverse_transform_labels(yt)
assert_equal(transp_ys.shape[0], ys.shape[0])
- assert_equal(transp_ys.shape[1], len(np.unique(yt)))
+ assert_equal(transp_ys.shape[1], len(np.unique(nx.to_numpy(yt))))