From f0c12fbdce04d09bf13b141d549e5e385c64caad Mon Sep 17 00:00:00 2001 From: Hind-M Date: Tue, 1 Jun 2021 18:39:31 +0200 Subject: First version allowing to fetch remote datasets --- src/python/CMakeLists.txt | 2 ++ 1 file changed, 2 insertions(+) (limited to 'src/python/CMakeLists.txt') diff --git a/src/python/CMakeLists.txt b/src/python/CMakeLists.txt index 98f2b85f..6f117588 100644 --- a/src/python/CMakeLists.txt +++ b/src/python/CMakeLists.txt @@ -542,6 +542,8 @@ if(PYTHONINTERP_FOUND) add_gudhi_py_test(test_dtm_rips_complex) endif() + # Fetch remote datasets + add_gudhi_py_test(test_remote_datasets) # Set missing or not modules set(GUDHI_MODULES ${GUDHI_MODULES} "python" CACHE INTERNAL "GUDHI_MODULES") -- cgit v1.2.3 From 8749199e00c0ed1c32b8e0198a65797de3ad192a Mon Sep 17 00:00:00 2001 From: Hind-M Date: Mon, 27 Sep 2021 15:19:25 +0200 Subject: Add option in cmake to enable or not the inclusion of datasets fetching test (disabled by default) --- src/cmake/modules/GUDHI_modules.cmake | 11 ++++++----- src/python/CMakeLists.txt | 4 +++- 2 files changed, 9 insertions(+), 6 deletions(-) (limited to 'src/python/CMakeLists.txt') diff --git a/src/cmake/modules/GUDHI_modules.cmake b/src/cmake/modules/GUDHI_modules.cmake index ccaf1ac5..9cc1a8f5 100644 --- a/src/cmake/modules/GUDHI_modules.cmake +++ b/src/cmake/modules/GUDHI_modules.cmake @@ -17,11 +17,12 @@ function(add_gudhi_module file_path) endfunction(add_gudhi_module) -option(WITH_GUDHI_BENCHMARK "Activate/desactivate benchmark compilation" OFF) -option(WITH_GUDHI_EXAMPLE "Activate/desactivate examples compilation and installation" OFF) -option(WITH_GUDHI_PYTHON "Activate/desactivate python module compilation and installation" ON) -option(WITH_GUDHI_TEST "Activate/desactivate examples compilation and installation" ON) -option(WITH_GUDHI_UTILITIES "Activate/desactivate utilities compilation and installation" ON) +option(WITH_GUDHI_BENCHMARK "Activate/deactivate benchmark compilation" OFF) +option(WITH_GUDHI_EXAMPLE "Activate/deactivate examples compilation and installation" OFF) +option(WITH_NETWORK "Activate/deactivate datasets fetching test which uses the Internet" OFF) +option(WITH_GUDHI_PYTHON "Activate/deactivate python module compilation and installation" ON) +option(WITH_GUDHI_TEST "Activate/deactivate examples compilation and installation" ON) +option(WITH_GUDHI_UTILITIES "Activate/deactivate utilities compilation and installation" ON) if (WITH_GUDHI_BENCHMARK) set(GUDHI_SUB_DIRECTORIES "${GUDHI_SUB_DIRECTORIES};benchmark") diff --git a/src/python/CMakeLists.txt b/src/python/CMakeLists.txt index 6f117588..6c8dfe32 100644 --- a/src/python/CMakeLists.txt +++ b/src/python/CMakeLists.txt @@ -543,7 +543,9 @@ if(PYTHONINTERP_FOUND) endif() # Fetch remote datasets - add_gudhi_py_test(test_remote_datasets) + if(WITH_NETWORK) + add_gudhi_py_test(test_remote_datasets) + endif() # Set missing or not modules set(GUDHI_MODULES ${GUDHI_MODULES} "python" CACHE INTERNAL "GUDHI_MODULES") -- cgit v1.2.3 From c9d6439fb9a6e65d7aa9f18bce675de65e901a0d Mon Sep 17 00:00:00 2001 From: Hind-M Date: Mon, 25 Oct 2021 11:43:09 +0200 Subject: Rename WITH_NETWORK option to WITH_GUDHI_REMOTE_TEST --- .appveyor.yml | 2 +- .circleci/config.yml | 2 +- .github/for_maintainers/tests_strategy.md | 2 +- src/cmake/modules/GUDHI_modules.cmake | 2 +- src/common/doc/installation.h | 4 ++-- src/python/CMakeLists.txt | 2 +- 6 files changed, 7 insertions(+), 7 deletions(-) (limited to 'src/python/CMakeLists.txt') diff --git a/.appveyor.yml b/.appveyor.yml index 521ec42d..ee6067e0 100644 --- a/.appveyor.yml +++ b/.appveyor.yml @@ -30,7 +30,7 @@ environment: CMAKE_FLAGS: -DWITH_GUDHI_EXAMPLE=OFF -DWITH_GUDHI_TEST=OFF -DWITH_GUDHI_UTILITIES=OFF -DWITH_GUDHI_PYTHON=ON - target: PythonTestsWithNetwork - CMAKE_FLAGS: -DWITH_GUDHI_EXAMPLE=OFF -DWITH_GUDHI_TEST=ON -DWITH_NETWORK=ON -DWITH_GUDHI_UTILITIES=OFF -DWITH_GUDHI_PYTHON=ON + CMAKE_FLAGS: -DWITH_GUDHI_EXAMPLE=OFF -DWITH_GUDHI_TEST=ON -DWITH_GUDHI_REMOTE_TEST=ON -DWITH_GUDHI_UTILITIES=OFF -DWITH_GUDHI_PYTHON=ON cache: diff --git a/.circleci/config.yml b/.circleci/config.yml index 85e42f8a..262e124b 100644 --- a/.circleci/config.yml +++ b/.circleci/config.yml @@ -89,7 +89,7 @@ jobs: git submodule update mkdir build cd build - cmake -DCMAKE_BUILD_TYPE=Release -DWITH_GUDHI_EXAMPLE=OFF -DWITH_GUDHI_UTILITIES=OFF -DWITH_GUDHI_PYTHON=ON -DPython_ADDITIONAL_VERSIONS=3 -DWITH_GUDHI_TEST=ON -DWITH_NETWORK=ON .. + cmake -DCMAKE_BUILD_TYPE=Release -DWITH_GUDHI_EXAMPLE=OFF -DWITH_GUDHI_UTILITIES=OFF -DWITH_GUDHI_PYTHON=ON -DPython_ADDITIONAL_VERSIONS=3 -DWITH_GUDHI_TEST=ON -DWITH_GUDHI_REMOTE_TEST=ON .. cd src/python python3 setup.py build_ext --inplace ctest --output-on-failure diff --git a/.github/for_maintainers/tests_strategy.md b/.github/for_maintainers/tests_strategy.md index 8fd7ac0d..610e1749 100644 --- a/.github/for_maintainers/tests_strategy.md +++ b/.github/for_maintainers/tests_strategy.md @@ -15,7 +15,7 @@ As all the third parties are already installed (thanks to docker), the compilati * utils (C++) * doxygen (C++ documentation that is available in the artefacts) * python (including documentation and code coverage that are available in the artefacts) -* python_tests_with_network (includes previous python with WITH_NETWORK option enabled which adds datasets fetching test) +* python_tests_with_network (includes previous python with WITH_GUDHI_REMOTE_TEST option enabled which adds datasets fetching test) (cf. `.circleci/config.yml`) diff --git a/src/cmake/modules/GUDHI_modules.cmake b/src/cmake/modules/GUDHI_modules.cmake index 9cc1a8f5..7cdce307 100644 --- a/src/cmake/modules/GUDHI_modules.cmake +++ b/src/cmake/modules/GUDHI_modules.cmake @@ -19,7 +19,7 @@ endfunction(add_gudhi_module) option(WITH_GUDHI_BENCHMARK "Activate/deactivate benchmark compilation" OFF) option(WITH_GUDHI_EXAMPLE "Activate/deactivate examples compilation and installation" OFF) -option(WITH_NETWORK "Activate/deactivate datasets fetching test which uses the Internet" OFF) +option(WITH_GUDHI_REMOTE_TEST "Activate/deactivate datasets fetching test which uses the Internet" OFF) option(WITH_GUDHI_PYTHON "Activate/deactivate python module compilation and installation" ON) option(WITH_GUDHI_TEST "Activate/deactivate examples compilation and installation" ON) option(WITH_GUDHI_UTILITIES "Activate/deactivate utilities compilation and installation" ON) diff --git a/src/common/doc/installation.h b/src/common/doc/installation.h index 72d4b1e5..b0fbdf20 100644 --- a/src/common/doc/installation.h +++ b/src/common/doc/installation.h @@ -40,8 +40,8 @@ make \endverbatim * `make test` is using Ctest (CMake test driver * program). If some of the tests are failing, please send us the result of the following command: * \verbatim ctest --output-on-failure \endverbatim - * Testing fetching datasets feature requires the use of the internet and is disabled by default. If you want to include this test, set WITH_NETWORK to ON when building in the previous step (note that this test is included in the python module): - * \verbatim cmake -DCMAKE_BUILD_TYPE=Release -DWITH_GUDHI_TEST=ON -DWITH_NETWORK=ON --DWITH_GUDHI_PYTHON=ON .. \endverbatim + * Testing fetching datasets feature requires the use of the internet and is disabled by default. If you want to include this test, set WITH_GUDHI_REMOTE_TEST to ON when building in the previous step (note that this test is included in the python module): + * \verbatim cmake -DCMAKE_BUILD_TYPE=Release -DWITH_GUDHI_TEST=ON -DWITH_GUDHI_REMOTE_TEST=ON --DWITH_GUDHI_PYTHON=ON .. \endverbatim * * \subsection documentationgeneration Documentation * To generate the documentation, Doxygen is required. diff --git a/src/python/CMakeLists.txt b/src/python/CMakeLists.txt index 6c8dfe32..ddb5c9c2 100644 --- a/src/python/CMakeLists.txt +++ b/src/python/CMakeLists.txt @@ -543,7 +543,7 @@ if(PYTHONINTERP_FOUND) endif() # Fetch remote datasets - if(WITH_NETWORK) + if(WITH_GUDHI_REMOTE_TEST) add_gudhi_py_test(test_remote_datasets) endif() -- cgit v1.2.3 From a00ce1990b112aa34f72e5504ae0cfe14f11e292 Mon Sep 17 00:00:00 2001 From: albert-github Date: Sun, 22 May 2022 17:58:57 +0200 Subject: Spelling corrections A number of spelling corrections as reported by the codespell (see: https://github.com/codespell-project/codespell) program and lists. Some remarks: - not considered are grammatical errors - not considered are names in the code although there are a number that could be improved (like `childs` -> `children`) - in the documentation it could be made clearer what are variables and what is running text (e.g. by placing variables in running text between backticks) - some comments are in the French language, I think it would be better to have them in the English (United States version). --- src/Alpha_complex/include/gudhi/Alpha_complex.h | 2 +- .../include/gudhi/Bitmap_cubical_complex.h | 2 +- .../include/gudhi/Bitmap_cubical_complex_base.h | 2 +- src/Bottleneck_distance/include/gudhi/Neighbors_finder.h | 2 +- src/Cech_complex/include/gudhi/Miniball.hpp | 2 +- src/Collapse/example/edge_collapse_conserve_persistence.cpp | 2 +- src/Collapse/include/gudhi/Flag_complex_edge_collapser.h | 2 +- .../distance_matrix_edge_collapse_rips_persistence.cpp | 2 +- .../utilities/point_cloud_edge_collapse_rips_persistence.cpp | 2 +- src/Contraction/doc/so3.svg | 2 +- src/Contraction/example/Garland_heckbert/Error_quadric.h | 2 +- src/Contraction/include/gudhi/Edge_contraction.h | 8 ++++---- src/Contraction/include/gudhi/Skeleton_blocker_contractor.h | 6 +++--- .../Coxeter_triangulation/Cell_complex/Hasse_diagram_cell.h | 6 +++--- .../include/gudhi/Functions/Function_affine_plane_in_Rd.h | 4 ++-- .../include/gudhi/Permutahedral_representation/Size_range.h | 2 +- src/GudhUI/todo.txt | 2 +- src/GudhUI/utils/Critical_points.h | 2 +- src/GudhUI/utils/Edge_contractor.h | 2 +- src/GudhUI/utils/Furthest_point_epsilon_net.h | 4 ++-- src/GudhUI/utils/K_nearest_builder.h | 2 +- src/GudhUI/utils/Lloyd_builder.h | 2 +- src/GudhUI/utils/Vertex_collapsor.h | 2 +- src/Nerve_GIC/utilities/km.py.COPYRIGHT | 2 +- .../include/gudhi/Persistence_intervals.h | 4 ++-- .../benchmark/performance_rips_persistence.cpp | 2 +- .../example/custom_persistence_sort.cpp | 2 +- .../example/persistence_from_simple_simplex_tree.cpp | 2 +- .../example/rips_multifield_persistence.cpp | 2 +- .../include/gudhi/Persistent_cohomology.h | 2 +- .../example_one_skeleton_rips_from_correlation_matrix.cpp | 2 +- src/Simplex_tree/example/graph_expansion_with_blocker.cpp | 2 +- src/Simplex_tree/example/simple_simplex_tree.cpp | 2 +- src/Simplex_tree/include/gudhi/Simplex_tree.h | 4 ++-- .../gudhi/Simplex_tree/Simplex_tree_node_explicit_storage.h | 2 +- .../test/simplex_tree_graph_expansion_unit_test.cpp | 4 ++-- src/Simplex_tree/test/simplex_tree_unit_test.cpp | 4 ++-- .../gudhi/Skeleton_blocker/Skeleton_blocker_simplex.h | 2 +- .../gudhi/Skeleton_blocker/Skeleton_blocker_sub_complex.h | 8 ++++---- .../include/gudhi/Skeleton_blocker/internal/Trie.h | 2 +- .../iterators/Skeleton_blockers_triangles_iterators.h | 4 ++-- .../include/gudhi/Skeleton_blocker_complex.h | 12 ++++++------ src/Tangential_complex/include/gudhi/Tangential_complex.h | 2 +- src/cmake/modules/FindTBB.cmake | 6 +++--- src/cmake/modules/GUDHI_modules.cmake | 4 ++-- src/cmake/modules/GUDHI_options.cmake | 10 +++++----- src/cmake/modules/GUDHI_third_party_libraries.cmake | 2 +- src/common/include/gudhi/reader_utils.h | 2 +- src/common/include/gudhi/writing_persistence_to_file.h | 4 ++-- src/python/CMakeLists.txt | 6 +++--- ...agram_persistence_from_correlation_matrix_file_example.py | 2 +- src/python/gudhi/hera/wasserstein.cc | 2 +- src/python/gudhi/persistence_graphical_tools.py | 2 +- src/python/gudhi/wasserstein/barycenter.py | 6 +++--- src/python/test/test_simplex_tree.py | 2 +- src/python/test/test_subsampling.py | 4 ++-- 56 files changed, 91 insertions(+), 91 deletions(-) (limited to 'src/python/CMakeLists.txt') diff --git a/src/Alpha_complex/include/gudhi/Alpha_complex.h b/src/Alpha_complex/include/gudhi/Alpha_complex.h index 028ec9bb..b1a9407b 100644 --- a/src/Alpha_complex/include/gudhi/Alpha_complex.h +++ b/src/Alpha_complex/include/gudhi/Alpha_complex.h @@ -69,7 +69,7 @@ template struct Is_Epeck_D> { static const bool val * \ingroup alpha_complex * * \details - * The data structure is constructing a CGAL Delaunay triangulation (for more informations on CGAL Delaunay + * The data structure is constructing a CGAL Delaunay triangulation (for more information on CGAL Delaunay * triangulation, please refer to the corresponding chapter in page http://doc.cgal.org/latest/Triangulation/) from a * range of points or from an OFF file (cf. Points_off_reader). * diff --git a/src/Bitmap_cubical_complex/include/gudhi/Bitmap_cubical_complex.h b/src/Bitmap_cubical_complex/include/gudhi/Bitmap_cubical_complex.h index aa255ec2..51f6a273 100644 --- a/src/Bitmap_cubical_complex/include/gudhi/Bitmap_cubical_complex.h +++ b/src/Bitmap_cubical_complex/include/gudhi/Bitmap_cubical_complex.h @@ -237,7 +237,7 @@ class Bitmap_cubical_complex : public T { * Filtration_simplex_iterator class provides an iterator though the whole structure in the order of filtration. * Secondary criteria for filtration are: * (1) Dimension of a cube (lower dimensional comes first). - * (2) Position in the data structure (the ones that are earlies in the data structure comes first). + * (2) Position in the data structure (the ones that are earliest in the data structure comes first). **/ class Filtration_simplex_range; diff --git a/src/Bitmap_cubical_complex/include/gudhi/Bitmap_cubical_complex_base.h b/src/Bitmap_cubical_complex/include/gudhi/Bitmap_cubical_complex_base.h index f8f80ded..bafe7981 100644 --- a/src/Bitmap_cubical_complex/include/gudhi/Bitmap_cubical_complex_base.h +++ b/src/Bitmap_cubical_complex/include/gudhi/Bitmap_cubical_complex_base.h @@ -43,7 +43,7 @@ namespace cubical_complex { * Each cell is represented by a single * bit (in case of black and white bitmaps, or by a single element of a type T * (here T is a filtration type of a bitmap, typically a double). - * All the informations needed for homology and + * All the information needed for homology and * persistent homology computations (like dimension of a cell, boundary and * coboundary elements of a cell, are then obtained from the * position of the element in C. diff --git a/src/Bottleneck_distance/include/gudhi/Neighbors_finder.h b/src/Bottleneck_distance/include/gudhi/Neighbors_finder.h index c65e6082..1d56f0b4 100644 --- a/src/Bottleneck_distance/include/gudhi/Neighbors_finder.h +++ b/src/Bottleneck_distance/include/gudhi/Neighbors_finder.h @@ -86,7 +86,7 @@ class Neighbors_finder { }; /** \internal \brief data structure used to find any point (including projections) in V near to a query point from U - * (which can be a projection) in a layered graph layer given as parmeter. + * (which can be a projection) in a layered graph layer given as parameter. * * V points have to be added manually using their index and before the first pull. A neighbor pulled is automatically * removed. diff --git a/src/Cech_complex/include/gudhi/Miniball.hpp b/src/Cech_complex/include/gudhi/Miniball.hpp index ce6cbb5b..55387a8a 100644 --- a/src/Cech_complex/include/gudhi/Miniball.hpp +++ b/src/Cech_complex/include/gudhi/Miniball.hpp @@ -1,4 +1,4 @@ -// Copright (C) 1999-2013, Bernd Gaertner +// Copyright (C) 1999-2013, Bernd Gaertner // $Rev: 3581 $ // // This program is free software: you can redistribute it and/or modify diff --git a/src/Collapse/example/edge_collapse_conserve_persistence.cpp b/src/Collapse/example/edge_collapse_conserve_persistence.cpp index b2c55e7a..19960597 100644 --- a/src/Collapse/example/edge_collapse_conserve_persistence.cpp +++ b/src/Collapse/example/edge_collapse_conserve_persistence.cpp @@ -103,7 +103,7 @@ int main(int argc, char* argv[]) { Gudhi::Euclidean_distance()); if (num_edges(proximity_graph) <= 0) { - std::cerr << "Total number of egdes are zero." << std::endl; + std::cerr << "Total number of edges is zero." << std::endl; exit(-1); } diff --git a/src/Collapse/include/gudhi/Flag_complex_edge_collapser.h b/src/Collapse/include/gudhi/Flag_complex_edge_collapser.h index c823901f..d0b3fe4a 100644 --- a/src/Collapse/include/gudhi/Flag_complex_edge_collapser.h +++ b/src/Collapse/include/gudhi/Flag_complex_edge_collapser.h @@ -53,7 +53,7 @@ struct Flag_complex_edge_collapser { #ifdef GUDHI_COLLAPSE_USE_DENSE_ARRAY // Minimal matrix interface // Using this matrix generally helps performance, but the memory use may be excessive for a very sparse graph - // (and in extreme cases the constant initialization of the matrix may start to dominate the runnning time). + // (and in extreme cases the constant initialization of the matrix may start to dominate the running time). // Are there cases where the matrix is too big but a hash table would help? std::vector neighbors_data; void init_neighbors_dense(){ diff --git a/src/Collapse/utilities/distance_matrix_edge_collapse_rips_persistence.cpp b/src/Collapse/utilities/distance_matrix_edge_collapse_rips_persistence.cpp index 11ee5871..38efb9e6 100644 --- a/src/Collapse/utilities/distance_matrix_edge_collapse_rips_persistence.cpp +++ b/src/Collapse/utilities/distance_matrix_edge_collapse_rips_persistence.cpp @@ -45,7 +45,7 @@ int main(int argc, char* argv[]) { min_persistence); Distance_matrix distances = Gudhi::read_lower_triangular_matrix_from_csv_file(csv_matrix_file); - std::cout << "Read the distance matrix succesfully, of size: " << distances.size() << std::endl; + std::cout << "Read the distance matrix successfully, of size: " << distances.size() << std::endl; Proximity_graph proximity_graph = Gudhi::compute_proximity_graph(boost::irange((size_t)0, distances.size()), diff --git a/src/Collapse/utilities/point_cloud_edge_collapse_rips_persistence.cpp b/src/Collapse/utilities/point_cloud_edge_collapse_rips_persistence.cpp index 0eea742c..d8f42ab6 100644 --- a/src/Collapse/utilities/point_cloud_edge_collapse_rips_persistence.cpp +++ b/src/Collapse/utilities/point_cloud_edge_collapse_rips_persistence.cpp @@ -77,7 +77,7 @@ int main(int argc, char* argv[]) { Gudhi::Euclidean_distance()); if (num_edges(proximity_graph) <= 0) { - std::cerr << "Total number of egdes are zero." << std::endl; + std::cerr << "Total number of edges is zero." << std::endl; exit(-1); } diff --git a/src/Contraction/doc/so3.svg b/src/Contraction/doc/so3.svg index adea3f38..f10cab98 100644 --- a/src/Contraction/doc/so3.svg +++ b/src/Contraction/doc/so3.svg @@ -177,7 +177,7 @@ x="309.4176" y="300.58682" id="tspan4515-4" - style="text-align:center;text-anchor:middle">Rips complex built uppon these pointsRips complex built upon these points class Error_quadric { * Quadric corresponding to the L2 distance to the plane. * * According to the notation of Garland Heckbert, they - * denote a quadric symetric matrix as : + * denote a quadric symmetric matrix as : * Q = [ q11 q12 q13 q14] * [ q12 q22 q23 q24] * [ q13 q23 q33 q34] diff --git a/src/Contraction/include/gudhi/Edge_contraction.h b/src/Contraction/include/gudhi/Edge_contraction.h index 58d627c2..0b43c3b3 100644 --- a/src/Contraction/include/gudhi/Edge_contraction.h +++ b/src/Contraction/include/gudhi/Edge_contraction.h @@ -46,7 +46,7 @@ the operations needed for edge contraction algorithms have polynomial complexity Therefore, the simplification can be done without enumerating the set of simplices that is often non tracktable in high-dimension and is then very efficient (sub-linear with regards to the number of simplices in practice). -A typical application of this package is homology group computation. It is illustrated in the next figure where a Rips complex is built uppon a set of high-dimensional points and +A typical application of this package is homology group computation. It is illustrated in the next figure where a Rips complex is built upon a set of high-dimensional points and simplified with edge contractions. It has initially a big number of simplices (around 20 millions) but simplifying it to a much reduced form with only 15 vertices (and 714 simplices) takes only few seconds on a desktop machine (see the example bellow). One can then compute homology group with a simplicial complex having very few simplices instead of running the homology algorithm on the much bigger initial set of @@ -65,7 +65,7 @@ This class design is policy based and heavily inspired from the similar edge col Four policies can be customized in this package: \li Cost_policy: specify how much cost an edge contraction of a given edge. The edge with lowest cost is iteratively picked and contracted if valid. -\li Valid_contraction_policy: specify if a given edge contraction is valid. For instance, this policy can check the link condition which ensures that the homotopy type is preserved afer the edge contraction. +\li Valid_contraction_policy: specify if a given edge contraction is valid. For instance, this policy can check the link condition which ensures that the homotopy type is preserved after the edge contraction. \li Placement_policy: every time an edge is contracted, its points are merge to one point specified by this policy. This may be the middle of the edge of some more sophisticated point such as the minimum of a cost as in \cite Garland. @@ -92,7 +92,7 @@ Despite this package is able to deal with \a arbitrary simplicial complexes (any it is still \a 65% times faster than the CGAL package which is focused on 2-manifold. The main reason is that few blockers appears during the simplification and hence, the algorithm only have to deal with the graph and not higher-dimensional simplices -(in this case triangles). However, we recall that higher-dimensional simplices are \a implicitely +(in this case triangles). However, we recall that higher-dimensional simplices are \a implicitly stored in the \ref skbl data-structure. Hence, one has to store simplices in an external map if some information needs to be associated with them (information that could be a filtration value or an orientation for instance). @@ -153,7 +153,7 @@ void build_rips(ComplexType& complex, double offset){ int main (int argc, char *argv[]) { if (argc!=3){ - std::cerr << "Usage "< { std::size_t id = 0; - // xxx do a parralel for + // xxx do a parallel for for (auto edge : complex_.edge_range()) { complex_[edge].index() = id++; Profile const& profile = create_profile(edge); @@ -474,7 +474,7 @@ typename GeometricSimplifiableComplex::Vertex_handle> { } void update_changed_edges() { - // xxx do a parralel for + // xxx do a parallel for DBG("update edges"); // sequential loop @@ -530,7 +530,7 @@ typename GeometricSimplifiableComplex::Vertex_handle> { // by definition of a blocker // todo uniqument utile pour la link condition - // laisser a l'utilisateur ? booleen update_heap_on_removed_blocker? + // laisser a l'utilisateur ? boolean update_heap_on_removed_blocker? Simplex blocker_copy(*blocker); for (auto x = blocker_copy.begin(); x != blocker_copy.end(); ++x) { for (auto y = x; ++y != blocker_copy.end();) { diff --git a/src/Coxeter_triangulation/include/gudhi/Coxeter_triangulation/Cell_complex/Hasse_diagram_cell.h b/src/Coxeter_triangulation/include/gudhi/Coxeter_triangulation/Cell_complex/Hasse_diagram_cell.h index 59e9a350..9b57da3c 100644 --- a/src/Coxeter_triangulation/include/gudhi/Coxeter_triangulation/Cell_complex/Hasse_diagram_cell.h +++ b/src/Coxeter_triangulation/include/gudhi/Coxeter_triangulation/Cell_complex/Hasse_diagram_cell.h @@ -95,7 +95,7 @@ class Hasse_diagram_cell { deleted_(false) {} /** - * Construcor of a cell of dimension dim having given additional information. + * Constructor of a cell of dimension dim having given additional information. **/ Hasse_diagram_cell(Additional_information ai, int dim) : dimension(dim), additional_info(ai), position(0), deleted_(false) {} @@ -125,7 +125,7 @@ class Hasse_diagram_cell { inline Additional_information& get_additional_information() { return this->additional_info; } /** - * Procedure to retrive position of the cell in the structure. It is used in + * Procedure to retrieve the position of the cell in the structure. It is used in * the implementation of Hasse diagram and set by it. Note that removal of * cell and subsequent call of clean_up_the_structure will change those * positions. @@ -186,7 +186,7 @@ class Hasse_diagram_cell { friend std::ostream& operator<<( std::ostream& out, const Hasse_diagram_cell& c) { // cout << "position : " << c.position << ", dimension : " << c.dimension << ", filtration: " << c.filtration << ", - // size of boudary : " << c.boundary.size() << "\n"; + // size of boundary : " << c.boundary.size() << "\n"; out << c.position << " " << c.dimension << " " << c.filtration << std::endl; for (std::size_t bd = 0; bd != c.boundary.size(); ++bd) { // do not write out the cells that has been deleted diff --git a/src/Coxeter_triangulation/include/gudhi/Functions/Function_affine_plane_in_Rd.h b/src/Coxeter_triangulation/include/gudhi/Functions/Function_affine_plane_in_Rd.h index b29f0906..dc6f5f90 100644 --- a/src/Coxeter_triangulation/include/gudhi/Functions/Function_affine_plane_in_Rd.h +++ b/src/Coxeter_triangulation/include/gudhi/Functions/Function_affine_plane_in_Rd.h @@ -51,7 +51,7 @@ struct Function_affine_plane_in_Rd { * plane in the d-dimensional Euclidean space. * * @param[in] normal_matrix A normal matrix of the affine plane. The number of rows should - * correspond to the ambient dimension, the number of columns should corespond to + * correspond to the ambient dimension, the number of columns should correspond to * the size of the normal basis (codimension). * @param[in] offset The offset vector of the affine plane. * The dimension of the vector should be the ambient dimension of the manifold. @@ -66,7 +66,7 @@ struct Function_affine_plane_in_Rd { * plane in the d-dimensional Euclidean space that passes through origin. * * @param[in] normal_matrix A normal matrix of the affine plane. The number of rows should - * correspond to the ambient dimension, the number of columns should corespond to + * correspond to the ambient dimension, the number of columns should correspond to * the size of the normal basis (codimension). */ Function_affine_plane_in_Rd(const Eigen::MatrixXd& normal_matrix) diff --git a/src/Coxeter_triangulation/include/gudhi/Permutahedral_representation/Size_range.h b/src/Coxeter_triangulation/include/gudhi/Permutahedral_representation/Size_range.h index c43effc8..6b137744 100644 --- a/src/Coxeter_triangulation/include/gudhi/Permutahedral_representation/Size_range.h +++ b/src/Coxeter_triangulation/include/gudhi/Permutahedral_representation/Size_range.h @@ -19,7 +19,7 @@ namespace Gudhi { namespace coxeter_triangulation { -/** \brief Auxillary iterator class for sizes of parts in an ordered set partition. +/** \brief Auxiliary iterator class for sizes of parts in an ordered set partition. */ template class Size_iterator diff --git a/src/GudhUI/todo.txt b/src/GudhUI/todo.txt index 19d99a77..e59d06d4 100644 --- a/src/GudhUI/todo.txt +++ b/src/GudhUI/todo.txt @@ -18,5 +18,5 @@ x faire le lien MainWindow - Model -- bug -x bug ordre contraction -> just that first vertex placement dont work great +x bug ordre contraction -> just that first vertex placement doesn't work great x pb construction rips diff --git a/src/GudhUI/utils/Critical_points.h b/src/GudhUI/utils/Critical_points.h index 97e58737..65695434 100644 --- a/src/GudhUI/utils/Critical_points.h +++ b/src/GudhUI/utils/Critical_points.h @@ -103,7 +103,7 @@ template class Critical_points { // reduced to one point -> contractible return 1; else - // we dont know + // we don't know return 2; } diff --git a/src/GudhUI/utils/Edge_contractor.h b/src/GudhUI/utils/Edge_contractor.h index 0707b186..a71d0742 100644 --- a/src/GudhUI/utils/Edge_contractor.h +++ b/src/GudhUI/utils/Edge_contractor.h @@ -65,7 +65,7 @@ template class Edge_contractor { /** * @brief Modify complex to be the expansion of the k-nearest neighbor - * symetric graph. + * symmetric graph. */ Edge_contractor(SkBlComplex& complex, unsigned num_contractions) : complex_(complex), num_contractions_(num_contractions) { diff --git a/src/GudhUI/utils/Furthest_point_epsilon_net.h b/src/GudhUI/utils/Furthest_point_epsilon_net.h index 6eb71071..195d0014 100644 --- a/src/GudhUI/utils/Furthest_point_epsilon_net.h +++ b/src/GudhUI/utils/Furthest_point_epsilon_net.h @@ -27,7 +27,7 @@ template class Furthest_point_epsilon_net { /** * Let V be the set of vertices. - * Initially v0 is one arbitrarly vertex and the set V0 is {v0}. + * Initially v0 is one, arbitrary, vertex and the set V0 is {v0}. * Then Vk is computed as follows. * First we compute the vertex pk that is the furthest from Vk * then Vk = Vk \cup pk. @@ -54,7 +54,7 @@ template class Furthest_point_epsilon_net { /** * @brief Modify complex to be the expansion of the k-nearest neighbor - * symetric graph. + * symmetric graph. */ Furthest_point_epsilon_net(SkBlComplex& complex) : complex_(complex) { diff --git a/src/GudhUI/utils/K_nearest_builder.h b/src/GudhUI/utils/K_nearest_builder.h index 34483e58..454b2587 100644 --- a/src/GudhUI/utils/K_nearest_builder.h +++ b/src/GudhUI/utils/K_nearest_builder.h @@ -41,7 +41,7 @@ template class K_nearest_builder { public: /** * @brief Modify complex to be the expansion of the k-nearest neighbor - * symetric graph. + * symmetric graph. */ K_nearest_builder(SkBlComplex& complex, unsigned k) : complex_(complex) { complex.keep_only_vertices(); diff --git a/src/GudhUI/utils/Lloyd_builder.h b/src/GudhUI/utils/Lloyd_builder.h index c042564f..57e3dc0f 100644 --- a/src/GudhUI/utils/Lloyd_builder.h +++ b/src/GudhUI/utils/Lloyd_builder.h @@ -27,7 +27,7 @@ template class Lloyd_builder { /** * @brief Modify complex to be the expansion of the k-nearest neighbor - * symetric graph. + * symmetric graph. */ Lloyd_builder(SkBlComplex& complex, unsigned num_iterations) : complex_(complex), dim(-1) { if (!complex_.empty()) { diff --git a/src/GudhUI/utils/Vertex_collapsor.h b/src/GudhUI/utils/Vertex_collapsor.h index 030e4bb0..b1c48efd 100644 --- a/src/GudhUI/utils/Vertex_collapsor.h +++ b/src/GudhUI/utils/Vertex_collapsor.h @@ -31,7 +31,7 @@ template class Vertex_collapsor { /** * @brief Modify complex to be the expansion of the k-nearest neighbor - * symetric graph. + * symmetric graph. */ Vertex_collapsor(SkBlComplex& complex, size_t num_collapses) : complex_(complex), num_collapses_(num_collapses) { diff --git a/src/Nerve_GIC/utilities/km.py.COPYRIGHT b/src/Nerve_GIC/utilities/km.py.COPYRIGHT index bef7b121..5358d287 100644 --- a/src/Nerve_GIC/utilities/km.py.COPYRIGHT +++ b/src/Nerve_GIC/utilities/km.py.COPYRIGHT @@ -1,7 +1,7 @@ km.py is a fork of https://github.com/MLWave/kepler-mapper. Only the visualization part has been kept (Mapper part has been removed). -This file has te following Copyright : +This file has the following Copyright : The MIT License (MIT) diff --git a/src/Persistence_representations/include/gudhi/Persistence_intervals.h b/src/Persistence_representations/include/gudhi/Persistence_intervals.h index a6c1d6f0..f4324cb2 100644 --- a/src/Persistence_representations/include/gudhi/Persistence_intervals.h +++ b/src/Persistence_representations/include/gudhi/Persistence_intervals.h @@ -109,7 +109,7 @@ class Persistence_intervals { std::vector cumulative_histogram_of_lengths(size_t number_of_bins = 10) const; /** - * In this procedure we assume that each barcode is a characteristic function of a hight equal to its length. The + * In this procedure we assume that each barcode is a characteristic function of a height equal to its length. The *persistence diagram is a sum of such a functions. The procedure below construct a function being a * sum of the characteristic functions of persistence intervals. The first two parameters are the range in which the *function is to be computed and the last parameter is the number of bins in @@ -207,7 +207,7 @@ class Persistence_intervals { /** * This is a simple function projecting the persistence intervals to a real number. The function we use here is a sum *of squared lengths of intervals. It can be naturally interpreted as - * sum of step function, where the step hight it equal to the length of the interval. + * sum of step function, where the step height it equal to the length of the interval. * At the moment this function is not tested, since it is quite likely to be changed in the future. Given this, when *using it, keep in mind that it * will be most likely changed in the next versions. diff --git a/src/Persistent_cohomology/benchmark/performance_rips_persistence.cpp b/src/Persistent_cohomology/benchmark/performance_rips_persistence.cpp index 030b072a..3bec8830 100644 --- a/src/Persistent_cohomology/benchmark/performance_rips_persistence.cpp +++ b/src/Persistent_cohomology/benchmark/performance_rips_persistence.cpp @@ -49,7 +49,7 @@ void timing_persistence(FilteredComplex & cpx * with a Hasse diagram. The Hasse diagram represents explicitly all * codimension 1 incidence relations in the complex, and hence leads to * a faster computation of persistence because boundaries are precomputed. - * Hovewer, the simplex tree may be constructed directly from a point cloud and + * However, the simplex tree may be constructed directly from a point cloud and * is more compact. * We compute persistent homology with coefficient fields Z/2Z and Z/1223Z. * We present also timings for the computation of multi-field persistent diff --git a/src/Persistent_cohomology/example/custom_persistence_sort.cpp b/src/Persistent_cohomology/example/custom_persistence_sort.cpp index 410cd987..bba0b2f7 100644 --- a/src/Persistent_cohomology/example/custom_persistence_sort.cpp +++ b/src/Persistent_cohomology/example/custom_persistence_sort.cpp @@ -33,7 +33,7 @@ using Persistent_cohomology = Gudhi::persistent_cohomology::Persistent_cohomolog Gudhi::persistent_cohomology::Field_Zp >; std::vector random_points() { - // Instanciate a random point generator + // Instantiate a random point generator CGAL::Random rng(0); // Generate "points_number" random points in a vector diff --git a/src/Persistent_cohomology/example/persistence_from_simple_simplex_tree.cpp b/src/Persistent_cohomology/example/persistence_from_simple_simplex_tree.cpp index bffaabdd..3da6771e 100644 --- a/src/Persistent_cohomology/example/persistence_from_simple_simplex_tree.cpp +++ b/src/Persistent_cohomology/example/persistence_from_simple_simplex_tree.cpp @@ -95,7 +95,7 @@ int main(int argc, char * const argv[]) { SimplexVector = {9, 10, 11}; st.insert_simplex_and_subfaces(SimplexVector, 0.3); - // ++ NINETH + // ++ NINTH std::clog << " - INSERT (2,10,12)" << std::endl; SimplexVector = {2, 10, 12}; st.insert_simplex_and_subfaces(SimplexVector, 0.3); diff --git a/src/Persistent_cohomology/example/rips_multifield_persistence.cpp b/src/Persistent_cohomology/example/rips_multifield_persistence.cpp index 2edf5bc4..d154bcde 100644 --- a/src/Persistent_cohomology/example/rips_multifield_persistence.cpp +++ b/src/Persistent_cohomology/example/rips_multifield_persistence.cpp @@ -104,7 +104,7 @@ void program_options(int argc, char * argv[] ("min-field-charac,p", po::value(&min_p)->default_value(2), "Minimal characteristic p of the coefficient field Z/pZ.") ("max-field-charac,q", po::value(&max_p)->default_value(1223), - "Minimial characteristic q of the coefficient field Z/pZ.") + "Minimal characteristic q of the coefficient field Z/pZ.") ("min-persistence,m", po::value(&min_persistence), "Minimal lifetime of homology feature to be recorded. Default is 0"); diff --git a/src/Persistent_cohomology/include/gudhi/Persistent_cohomology.h b/src/Persistent_cohomology/include/gudhi/Persistent_cohomology.h index d428e497..2301a66b 100644 --- a/src/Persistent_cohomology/include/gudhi/Persistent_cohomology.h +++ b/src/Persistent_cohomology/include/gudhi/Persistent_cohomology.h @@ -211,7 +211,7 @@ class Persistent_cohomology { /** \brief Update the cohomology groups under the insertion of an edge. * * The 0-homology is maintained with a simple Union-Find data structure, which - * explains the existance of a specific function of edge insertions. */ + * explains the existence of a specific function of edge insertions. */ void update_cohomology_groups_edge(Simplex_handle sigma) { Simplex_handle u, v; boost::tie(u, v) = cpx_->endpoints(sigma); diff --git a/src/Rips_complex/example/example_one_skeleton_rips_from_correlation_matrix.cpp b/src/Rips_complex/example/example_one_skeleton_rips_from_correlation_matrix.cpp index 3d2ba54f..3811d1f1 100644 --- a/src/Rips_complex/example/example_one_skeleton_rips_from_correlation_matrix.cpp +++ b/src/Rips_complex/example/example_one_skeleton_rips_from_correlation_matrix.cpp @@ -40,7 +40,7 @@ int main() { throw "The input matrix is not a correlation matrix. The program will now terminate.\n"; } correlations[i][j] = 1 - correlations[i][j]; - // Here we make sure that we will get the treshold value equal to maximal + // Here we make sure that we will get the threshold value equal to maximal // distance in the matrix. if (correlations[i][j] > threshold) threshold = correlations[i][j]; } diff --git a/src/Simplex_tree/example/graph_expansion_with_blocker.cpp b/src/Simplex_tree/example/graph_expansion_with_blocker.cpp index df52bf43..eef8b665 100644 --- a/src/Simplex_tree/example/graph_expansion_with_blocker.cpp +++ b/src/Simplex_tree/example/graph_expansion_with_blocker.cpp @@ -42,7 +42,7 @@ int main(int argc, char* const argv[]) { std::clog << vertex << ", "; } std::clog << "] ( " << stree.filtration(sh); - // User can re-assign a new filtration value directly in the blocker (default is the maximal value of boudaries) + // User can re-assign a new filtration value directly in the blocker (default is the maximal value of boundaries) stree.assign_filtration(sh, stree.filtration(sh) + 1.); std::clog << " + 1. ) = " << result << std::endl; diff --git a/src/Simplex_tree/example/simple_simplex_tree.cpp b/src/Simplex_tree/example/simple_simplex_tree.cpp index e8bec596..965711da 100644 --- a/src/Simplex_tree/example/simple_simplex_tree.cpp +++ b/src/Simplex_tree/example/simple_simplex_tree.cpp @@ -129,7 +129,7 @@ int main(int argc, char* const argv[]) { std::clog << " - 3 NOT INSERTED" << std::endl; } - // ++ NINETH + // ++ NINTH std::clog << " * INSERT (3,0)" << std::endl; typeVectorVertex ninethSimplexVector = {3, 0}; returnValue = simplexTree.insert_simplex(ninethSimplexVector, Filtration_value(SECOND_FILTRATION_VALUE)); diff --git a/src/Simplex_tree/include/gudhi/Simplex_tree.h b/src/Simplex_tree/include/gudhi/Simplex_tree.h index 34bc5ace..629a1f9c 100644 --- a/src/Simplex_tree/include/gudhi/Simplex_tree.h +++ b/src/Simplex_tree/include/gudhi/Simplex_tree.h @@ -965,7 +965,7 @@ class Simplex_tree { // If we reached the end of the vertices, and the simplex has more vertices than the given simplex // => we found a coface - // Add a coface if we wan't the star or if the number of vertices of the current simplex matches with nbVertices + // Add a coface if we want the star or if the number of vertices of the current simplex matches with nbVertices bool addCoface = (star || curr_nbVertices == nbVertices); if (addCoface) cofaces.push_back(simplex); @@ -1491,7 +1491,7 @@ class Simplex_tree { int sh_dimension = dimension(sh); if (sh_dimension >= dimension_) - // Stop browsing as soon as the dimension is reached, no need to go furter + // Stop browsing as soon as the dimension is reached, no need to go further return false; new_dimension = (std::max)(new_dimension, sh_dimension); } diff --git a/src/Simplex_tree/include/gudhi/Simplex_tree/Simplex_tree_node_explicit_storage.h b/src/Simplex_tree/include/gudhi/Simplex_tree/Simplex_tree_node_explicit_storage.h index ae140859..ad53710c 100644 --- a/src/Simplex_tree/include/gudhi/Simplex_tree/Simplex_tree_node_explicit_storage.h +++ b/src/Simplex_tree/include/gudhi/Simplex_tree/Simplex_tree_node_explicit_storage.h @@ -24,7 +24,7 @@ namespace Gudhi { * \brief Node of a simplex tree with filtration value * and simplex key. * - * It stores explicitely its own filtration value and its own Simplex_key. + * It stores explicitly its own filtration value and its own Simplex_key. */ template struct Simplex_tree_node_explicit_storage : SimplexTree::Filtration_simplex_base, SimplexTree::Key_simplex_base { diff --git a/src/Simplex_tree/test/simplex_tree_graph_expansion_unit_test.cpp b/src/Simplex_tree/test/simplex_tree_graph_expansion_unit_test.cpp index 6d63d8ae..54e23204 100644 --- a/src/Simplex_tree/test/simplex_tree_graph_expansion_unit_test.cpp +++ b/src/Simplex_tree/test/simplex_tree_graph_expansion_unit_test.cpp @@ -93,7 +93,7 @@ BOOST_AUTO_TEST_CASE_TEMPLATE(simplex_tree_expansion_with_blockers_3, typeST, li std::clog << vertex << ", "; } std::clog << "] ( " << simplex_tree.filtration(sh); - // User can re-assign a new filtration value directly in the blocker (default is the maximal value of boudaries) + // User can re-assign a new filtration value directly in the blocker (default is the maximal value of boundaries) simplex_tree.assign_filtration(sh, simplex_tree.filtration(sh) + 1.); std::clog << " + 1. ) = " << result << std::endl; @@ -160,7 +160,7 @@ BOOST_AUTO_TEST_CASE_TEMPLATE(simplex_tree_expansion_with_blockers_2, typeST, li std::clog << vertex << ", "; } std::clog << "] ( " << simplex_tree.filtration(sh); - // User can re-assign a new filtration value directly in the blocker (default is the maximal value of boudaries) + // User can re-assign a new filtration value directly in the blocker (default is the maximal value of boundaries) simplex_tree.assign_filtration(sh, simplex_tree.filtration(sh) + 1.); std::clog << " + 1. ) = " << result << std::endl; diff --git a/src/Simplex_tree/test/simplex_tree_unit_test.cpp b/src/Simplex_tree/test/simplex_tree_unit_test.cpp index b18e2ec4..79bb5a93 100644 --- a/src/Simplex_tree/test/simplex_tree_unit_test.cpp +++ b/src/Simplex_tree/test/simplex_tree_unit_test.cpp @@ -287,7 +287,7 @@ BOOST_AUTO_TEST_CASE_TEMPLATE(simplex_tree_insertion, typeST, list_of_tested_var set_and_test_simplex_tree_dim_fil(st, eighthSimplexVector.size(), eighthSimplex.second); BOOST_CHECK(st.num_vertices() == (size_t) 4); - // ++ NINETH + // ++ NINTH std::clog << " - INSERT (3,0)" << std::endl; typeVectorVertex ninethSimplexVector{3, 0}; BOOST_CHECK(ninethSimplexVector.size() == 2); @@ -361,7 +361,7 @@ BOOST_AUTO_TEST_CASE_TEMPLATE(simplex_tree_insertion, typeST, list_of_tested_var test_simplex_tree_contains(st, seventhSimplex, 8); // (2,1,0) -> 8 std::clog << "simplex_tree_insertion - eighth - 3" << std::endl; test_simplex_tree_contains(st, eighthSimplex, 3); // (3) -> 3 - std::clog << "simplex_tree_insertion - nineth - 7" << std::endl; + std::clog << "simplex_tree_insertion - ninth - 7" << std::endl; test_simplex_tree_contains(st, ninethSimplex, 7); // (3,0) -> 7 // Display the Simplex_tree - Can not be done in the middle of 2 inserts diff --git a/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/Skeleton_blocker_simplex.h b/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/Skeleton_blocker_simplex.h index 12fe6469..d83c0ab3 100644 --- a/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/Skeleton_blocker_simplex.h +++ b/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/Skeleton_blocker_simplex.h @@ -134,7 +134,7 @@ class Skeleton_blocker_simplex { } /** - * Substracts a from the simplex. + * Subtracts a from the simplex. */ void difference(const Skeleton_blocker_simplex & a) { std::vector v; diff --git a/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/Skeleton_blocker_sub_complex.h b/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/Skeleton_blocker_sub_complex.h index 4c48ff31..5abd64d7 100644 --- a/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/Skeleton_blocker_sub_complex.h +++ b/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/Skeleton_blocker_sub_complex.h @@ -76,8 +76,8 @@ class Skeleton_blocker_sub_complex : public ComplexType { public: /** * Add a vertex 'global' of K to L. When added to L, this vertex will receive - * another number, addresses(global), its local adress. - * return the adress where the vertex lay on L. + * another number, addresses(global), its local address. + * return the address where the vertex lay on L. * The vertex corresponding to 'global' must not be already present * in the complex. */ @@ -174,7 +174,7 @@ class Skeleton_blocker_sub_complex : public ComplexType { // /** // * Allocates a simplex in L corresponding to the simplex s in K - // * with its local adresses and returns an AddressSimplex. + // * with its local addresses and returns an AddressSimplex. // */ // boost::optional get_address(const Root_simplex_handle & s) const; @@ -226,7 +226,7 @@ bool proper_face_in_union( } // Remark: this function should be friend in order to leave get_adresses private -// however doing so seemes currently not possible due to a visual studio bug c2668 +// however doing so seems currently not possible due to a visual studio bug c2668 // "the compiler does not support partial ordering of template functions as specified in the C++ Standard" // http://www.serkey.com/error-c2668-ambiguous-call-to-overloaded-function-bb45ft.html diff --git a/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/internal/Trie.h b/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/internal/Trie.h index a43fa034..18ae6a92 100644 --- a/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/internal/Trie.h +++ b/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/internal/Trie.h @@ -107,7 +107,7 @@ struct Trie { } /** - * Goes to the root in the trie to consitute simplex + * Goes to the root in the trie to constitute simplex */ void add_vertices_up_to_the_root(Simplex& res) const { res.add_vertex(v); diff --git a/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/iterators/Skeleton_blockers_triangles_iterators.h b/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/iterators/Skeleton_blockers_triangles_iterators.h index 37c0b4d3..2c49a1b8 100644 --- a/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/iterators/Skeleton_blockers_triangles_iterators.h +++ b/src/Skeleton_blocker/include/gudhi/Skeleton_blocker/iterators/Skeleton_blockers_triangles_iterators.h @@ -21,7 +21,7 @@ namespace skeleton_blocker { /** * \brief Iterator over the triangles that are * adjacent to a vertex of the simplicial complex. - * \remark Will be removed soon -> dont look + * \remark Will be removed soon -> don't look */ template class Triangle_around_vertex_iterator : public boost::iterator_facade @@ -95,7 +95,7 @@ class Triangle_around_vertex_iterator : public boost::iterator_facade /** * \brief Iterator over the triangles of the * simplicial complex. - * \remark Will be removed soon -> dont look + * \remark Will be removed soon -> don't look * */ template diff --git a/src/Skeleton_blocker/include/gudhi/Skeleton_blocker_complex.h b/src/Skeleton_blocker/include/gudhi/Skeleton_blocker_complex.h index 031bcb9c..8ceaa480 100644 --- a/src/Skeleton_blocker/include/gudhi/Skeleton_blocker_complex.h +++ b/src/Skeleton_blocker/include/gudhi/Skeleton_blocker_complex.h @@ -438,7 +438,7 @@ class Skeleton_blocker_complex { } /** - * return the id of a vertex of adress local present in the graph + * return the id of a vertex of address local present in the graph */ Root_vertex_handle get_id(Vertex_handle local) const { assert(0 <= local.vertex && local.vertex < boost::num_vertices(skeleton)); @@ -740,7 +740,7 @@ class Skeleton_blocker_complex { * complex to the smallest flag complex that contains it. */ void remove_blockers() { - // Desallocate the blockers + // Deallocate the blockers while (!blocker_map_.empty()) { delete_blocker(blocker_map_.begin()->second); } @@ -764,8 +764,8 @@ class Skeleton_blocker_complex { public: /** - * Removes the simplex s from the set of blockers - * and desallocate s. + * Removes the simplex sigma from the set of blockers + * and deallocate sigma. */ void delete_blocker(Blocker_handle sigma) { if (visitor) @@ -960,7 +960,7 @@ class Skeleton_blocker_complex { } /* - * @brief returnrs true iff the complex is empty. + * @brief returns true iff the complex is empty. */ bool empty() const { return num_vertices() == 0; @@ -1043,7 +1043,7 @@ class Skeleton_blocker_complex { if (num_vertices() == 1) return true; for (auto vi : vertex_range()) { - // xxx todo faire une methode bool is_in_blocker(Vertex_handle) + // xxx todo create a method: bool is_in_blocker(Vertex_handle) if (blocker_map_.find(vi) == blocker_map_.end()) { // no blocker passes through the vertex, we just need to // check if the current vertex is linked to all others vertices of the complex diff --git a/src/Tangential_complex/include/gudhi/Tangential_complex.h b/src/Tangential_complex/include/gudhi/Tangential_complex.h index f3491f91..cc424810 100644 --- a/src/Tangential_complex/include/gudhi/Tangential_complex.h +++ b/src/Tangential_complex/include/gudhi/Tangential_complex.h @@ -1152,7 +1152,7 @@ class Tangential_complex { #ifdef GUDHI_TC_VERY_VERBOSE std::cerr << "Inserted " << num_inserted_points << " points / " << num_attempts_to_insert_points - << " attemps to compute the star\n"; + << " attempts to compute the star\n"; #endif update_star(i); diff --git a/src/cmake/modules/FindTBB.cmake b/src/cmake/modules/FindTBB.cmake index 13f4d929..e6c42dc7 100644 --- a/src/cmake/modules/FindTBB.cmake +++ b/src/cmake/modules/FindTBB.cmake @@ -34,7 +34,7 @@ # # GvdB: Mac OS X distribution places libraries directly in lib directory. # -# For backwards compatibility, you may explicitely set the CMake variables TBB_ARCHITECTURE and TBB_COMPILER. +# For backwards compatibility, you may explicitly set the CMake variables TBB_ARCHITECTURE and TBB_COMPILER. # TBB_ARCHITECTURE [ ia32 | em64t | itanium ] # which architecture to use # TBB_COMPILER e.g. vc9 or cc3.2.3_libc2.3.2_kernel2.4.21 or cc4.0.1_os10.4.9 @@ -54,8 +54,8 @@ # TBB_MALLOC_DEBUG_LIBRARY, the TBB debug malloc library # TBB_FOUND, If false, don't try to use TBB. # TBB_INTERFACE_VERSION, as defined in tbb/tbb_stddef.h -# TBB_MALLOCPROXY_DEBUG_LIBRARY, the TBB debug malloc_proxy library (not included in TBB_LIBRARIES since it's optionnal) -# TBB_MALLOCPROXY_RELEASE_LIBRARY, the TBB release malloc_proxy library (not included in TBB_LIBRARIES since it's optionnal) +# TBB_MALLOCPROXY_DEBUG_LIBRARY, the TBB debug malloc_proxy library (not included in TBB_LIBRARIES since it's optional) +# TBB_MALLOCPROXY_RELEASE_LIBRARY, the TBB release malloc_proxy library (not included in TBB_LIBRARIES since it's optional) include(CheckCXXSourceCompiles) diff --git a/src/cmake/modules/GUDHI_modules.cmake b/src/cmake/modules/GUDHI_modules.cmake index 13248f7e..ec1f756b 100644 --- a/src/cmake/modules/GUDHI_modules.cmake +++ b/src/cmake/modules/GUDHI_modules.cmake @@ -2,7 +2,7 @@ set(GUDHI_MODULES_FULL_LIST "") function(add_gudhi_module file_path) - option("WITH_MODULE_GUDHI_${file_path}" "Activate/desactivate ${file_path} compilation and installation" ON) + option("WITH_MODULE_GUDHI_${file_path}" "Activate/deactivate ${file_path} compilation and installation" ON) if (WITH_MODULE_GUDHI_${file_path}) set(GUDHI_MODULES ${GUDHI_MODULES} ${file_path} CACHE INTERNAL "GUDHI_MODULES") else() @@ -10,7 +10,7 @@ function(add_gudhi_module file_path) endif() # Required by user_version set(GUDHI_MODULES_FULL_LIST ${GUDHI_MODULES_FULL_LIST} ${file_path} PARENT_SCOPE) - # Include module headers is independant - You may ask for no Alpha complex module but Python interface i.e. + # Include module headers is independent - You may ask for no Alpha complex module but Python interface i.e. if(IS_DIRECTORY ${CMAKE_SOURCE_DIR}/src/${file_path}/include/) include_directories(src/${file_path}/include/) endif() diff --git a/src/cmake/modules/GUDHI_options.cmake b/src/cmake/modules/GUDHI_options.cmake index 3cd0a489..bffb3ffc 100644 --- a/src/cmake/modules/GUDHI_options.cmake +++ b/src/cmake/modules/GUDHI_options.cmake @@ -1,5 +1,5 @@ -option(WITH_GUDHI_BENCHMARK "Activate/desactivate benchmark compilation" OFF) -option(WITH_GUDHI_EXAMPLE "Activate/desactivate examples compilation and installation" OFF) -option(WITH_GUDHI_PYTHON "Activate/desactivate python module compilation and installation" ON) -option(WITH_GUDHI_TEST "Activate/desactivate examples compilation and installation" ON) -option(WITH_GUDHI_UTILITIES "Activate/desactivate utilities compilation and installation" ON) +option(WITH_GUDHI_BENCHMARK "Activate/deactivate benchmark compilation" OFF) +option(WITH_GUDHI_EXAMPLE "Activate/deactivate examples compilation and installation" OFF) +option(WITH_GUDHI_PYTHON "Activate/deactivate python module compilation and installation" ON) +option(WITH_GUDHI_TEST "Activate/deactivate examples compilation and installation" ON) +option(WITH_GUDHI_UTILITIES "Activate/deactivate utilities compilation and installation" ON) diff --git a/src/cmake/modules/GUDHI_third_party_libraries.cmake b/src/cmake/modules/GUDHI_third_party_libraries.cmake index 6a94d1f5..6ba822ad 100644 --- a/src/cmake/modules/GUDHI_third_party_libraries.cmake +++ b/src/cmake/modules/GUDHI_third_party_libraries.cmake @@ -174,7 +174,7 @@ if (WITH_GUDHI_PYTHON) message(FATAL_ERROR "ERROR: GUDHI_PYTHON_PATH is not valid.") endif(NOT GUDHI_PYTHON_PATH) - option(WITH_GUDHI_PYTHON_RUNTIME_LIBRARY_DIRS "Build with setting runtime_library_dirs. Usefull when setting rpath is not allowed" ON) + option(WITH_GUDHI_PYTHON_RUNTIME_LIBRARY_DIRS "Build with setting runtime_library_dirs. Useful when setting rpath is not allowed" ON) if(PYTHONINTERP_FOUND AND CYTHON_FOUND) if(SPHINX_FOUND) diff --git a/src/common/include/gudhi/reader_utils.h b/src/common/include/gudhi/reader_utils.h index 29d5423d..a7d82541 100644 --- a/src/common/include/gudhi/reader_utils.h +++ b/src/common/include/gudhi/reader_utils.h @@ -231,7 +231,7 @@ std::vector> read_lower_triangular_matrix_from_csv std::string line; - // the first line is emtpy, so we ignore it: + // the first line is empty, so we ignore it: std::getline(in, line); std::vector values_in_this_line; result.push_back(values_in_this_line); diff --git a/src/common/include/gudhi/writing_persistence_to_file.h b/src/common/include/gudhi/writing_persistence_to_file.h index 2e36b831..3a0df1a8 100644 --- a/src/common/include/gudhi/writing_persistence_to_file.h +++ b/src/common/include/gudhi/writing_persistence_to_file.h @@ -48,7 +48,7 @@ class Persistence_interval_common { : birth_(birth), death_(death), dimension_(dim), arith_element_(field) {} /** - * Operator to compare two persistence pairs. During the comparision all the + * Operator to compare two persistence pairs. During the comparison all the * fields: birth, death, dimensiona and arith_element_ are taken into account * and they all have to be equal for two pairs to be equal. **/ @@ -65,7 +65,7 @@ class Persistence_interval_common { /** * Operator to compare objects of a type Persistence_interval_common. * One intervals is smaller than the other if it has lower persistence. - * Note that this operator do not take Arith_element into account when doing comparisions. + * Note that this operator do not take Arith_element into account when doing comparisons. **/ bool operator<(const Persistence_interval_common& i2) const { return fabs(this->death_ - this->birth_) < fabs(i2.death_ - i2.birth_); diff --git a/src/python/CMakeLists.txt b/src/python/CMakeLists.txt index 54221151..af0b6115 100644 --- a/src/python/CMakeLists.txt +++ b/src/python/CMakeLists.txt @@ -329,9 +329,9 @@ if(PYTHONINTERP_FOUND) if(NOT CGAL_WITH_EIGEN3_VERSION VERSION_LESS 5.1.0) set (GUDHI_SPHINX_MESSAGE "Generating API documentation with Sphinx in ${CMAKE_CURRENT_BINARY_DIR}/sphinx/") # User warning - Sphinx is a static pages generator, and configured to work fine with user_version - # Images and biblio warnings because not found on developper version + # Images and biblio warnings because not found on developer version if (GUDHI_PYTHON_PATH STREQUAL "src/python") - set (GUDHI_SPHINX_MESSAGE "${GUDHI_SPHINX_MESSAGE} \n WARNING : Sphinx is configured for user version, you run it on developper version. Images and biblio will miss") + set (GUDHI_SPHINX_MESSAGE "${GUDHI_SPHINX_MESSAGE} \n WARNING : Sphinx is configured for user version, you run it on developer version. Images and biblio will miss") endif() # sphinx target requires gudhi.so, because conf.py reads gudhi version from it add_custom_target(sphinx @@ -484,7 +484,7 @@ if(PYTHONINTERP_FOUND) add_gudhi_py_test(test_euclidean_witness_complex) # Datasets generators - add_gudhi_py_test(test_datasets_generators) # TODO separate full python datasets generators in another test file independant from CGAL ? + add_gudhi_py_test(test_datasets_generators) # TODO separate full python datasets generators in another test file independent from CGAL ? endif (NOT CGAL_WITH_EIGEN3_VERSION VERSION_LESS 4.11.0) diff --git a/src/python/example/rips_complex_diagram_persistence_from_correlation_matrix_file_example.py b/src/python/example/rips_complex_diagram_persistence_from_correlation_matrix_file_example.py index ea2eb7e1..0b35dbc5 100755 --- a/src/python/example/rips_complex_diagram_persistence_from_correlation_matrix_file_example.py +++ b/src/python/example/rips_complex_diagram_persistence_from_correlation_matrix_file_example.py @@ -40,7 +40,7 @@ parser.add_argument( args = parser.parse_args() if not (-1.0 < args.min_edge_correlation < 1.0): - print("Wrong value of the treshold corelation (should be between -1 and 1).") + print("Wrong value of the threshold corelation (should be between -1 and 1).") sys.exit(1) print("#####################################################################") diff --git a/src/python/gudhi/hera/wasserstein.cc b/src/python/gudhi/hera/wasserstein.cc index 1a21f02f..fa0cf8aa 100644 --- a/src/python/gudhi/hera/wasserstein.cc +++ b/src/python/gudhi/hera/wasserstein.cc @@ -29,7 +29,7 @@ double wasserstein_distance( if(std::isinf(internal_p)) internal_p = hera::get_infinity(); params.internal_p = internal_p; params.delta = delta; - // The extra parameters are purposedly not exposed for now. + // The extra parameters are purposely not exposed for now. return hera::wasserstein_dist(diag1, diag2, params); } diff --git a/src/python/gudhi/persistence_graphical_tools.py b/src/python/gudhi/persistence_graphical_tools.py index 7ed11360..21275cdd 100644 --- a/src/python/gudhi/persistence_graphical_tools.py +++ b/src/python/gudhi/persistence_graphical_tools.py @@ -332,7 +332,7 @@ def plot_persistence_diagram( axes.plot([axis_start, axis_end], [infinity, infinity], linewidth=1.0, color="k", alpha=alpha) # Infinity label yt = axes.get_yticks() - yt = yt[np.where(yt < axis_end)] # to avoid ploting ticklabel higher than infinity + yt = yt[np.where(yt < axis_end)] # to avoid plotting ticklabel higher than infinity yt = np.append(yt, infinity) ytl = ["%.3f" % e for e in yt] # to avoid float precision error ytl[-1] = r"$+\infty$" diff --git a/src/python/gudhi/wasserstein/barycenter.py b/src/python/gudhi/wasserstein/barycenter.py index d67bcde7..bb6e641e 100644 --- a/src/python/gudhi/wasserstein/barycenter.py +++ b/src/python/gudhi/wasserstein/barycenter.py @@ -37,7 +37,7 @@ def lagrangian_barycenter(pdiagset, init=None, verbose=False): :param init: The initial value for barycenter estimate. If ``None``, init is made on a random diagram from the dataset. Otherwise, it can be an ``int`` (then initialization is made on ``pdiagset[init]``) - or a `(n x 2)` ``numpy.array`` enconding a persistence diagram with `n` points. + or a `(n x 2)` ``numpy.array`` encoding a persistence diagram with `n` points. :type init: ``int``, or (n x 2) ``np.array`` :param verbose: if ``True``, returns additional information about the barycenter. :type verbose: boolean @@ -45,7 +45,7 @@ def lagrangian_barycenter(pdiagset, init=None, verbose=False): (local minimum of the energy function). If ``pdiagset`` is empty, returns ``None``. If verbose, returns a couple ``(Y, log)`` where ``Y`` is the barycenter estimate, - and ``log`` is a ``dict`` that contains additional informations: + and ``log`` is a ``dict`` that contains additional information: - `"groupings"`, a list of list of pairs ``(i,j)``. Namely, ``G[k] = [...(i, j)...]``, where ``(i,j)`` indicates that `pdiagset[k][i]`` is matched to ``Y[j]`` if ``i = -1`` or ``j = -1``, it means they represent the diagonal. @@ -73,7 +73,7 @@ def lagrangian_barycenter(pdiagset, init=None, verbose=False): nb_iter = 0 - converged = False # stoping criterion + converged = False # stopping criterion while not converged: nb_iter += 1 K = len(Y) # current nb of points in Y (some might be on diagonal) diff --git a/src/python/test/test_simplex_tree.py b/src/python/test/test_simplex_tree.py index 688f4fd6..ba8c455f 100755 --- a/src/python/test/test_simplex_tree.py +++ b/src/python/test/test_simplex_tree.py @@ -528,7 +528,7 @@ def test_expansion_with_blocker(): def blocker(simplex): try: - # Block all simplices that countains vertex 6 + # Block all simplices that contains vertex 6 simplex.index(6) print(simplex, ' is blocked') return True diff --git a/src/python/test/test_subsampling.py b/src/python/test/test_subsampling.py index 4019852e..3431f372 100755 --- a/src/python/test/test_subsampling.py +++ b/src/python/test/test_subsampling.py @@ -91,7 +91,7 @@ def test_simple_choose_n_farthest_points_randomed(): assert gudhi.choose_n_farthest_points(points=[], nb_points=1) == [] assert gudhi.choose_n_farthest_points(points=point_set, nb_points=0) == [] - # Go furter than point set on purpose + # Go further than point set on purpose for iter in range(1, 10): sub_set = gudhi.choose_n_farthest_points(points=point_set, nb_points=iter) for sub in sub_set: @@ -117,7 +117,7 @@ def test_simple_pick_n_random_points(): assert gudhi.pick_n_random_points(points=[], nb_points=1) == [] assert gudhi.pick_n_random_points(points=point_set, nb_points=0) == [] - # Go furter than point set on purpose + # Go further than point set on purpose for iter in range(1, 10): sub_set = gudhi.pick_n_random_points(points=point_set, nb_points=iter) for sub in sub_set: -- cgit v1.2.3