|  | // Ceres Solver - A fast non-linear least squares minimizer | 
|  | // Copyright 2022 Google Inc. All rights reserved. | 
|  | // http://ceres-solver.org/ | 
|  | // | 
|  | // Redistribution and use in source and binary forms, with or without | 
|  | // modification, are permitted provided that the following conditions are met: | 
|  | // | 
|  | // * Redistributions of source code must retain the above copyright notice, | 
|  | //   this list of conditions and the following disclaimer. | 
|  | // * Redistributions in binary form must reproduce the above copyright notice, | 
|  | //   this list of conditions and the following disclaimer in the documentation | 
|  | //   and/or other materials provided with the distribution. | 
|  | // * Neither the name of Google Inc. nor the names of its contributors may be | 
|  | //   used to endorse or promote products derived from this software without | 
|  | //   specific prior written permission. | 
|  | // | 
|  | // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | 
|  | // AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | 
|  | // IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | 
|  | // ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE | 
|  | // LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR | 
|  | // CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF | 
|  | // SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS | 
|  | // INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN | 
|  | // CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) | 
|  | // ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE | 
|  | // POSSIBILITY OF SUCH DAMAGE. | 
|  | // | 
|  | // Author: joydeepb@cs.utexas.edu (Joydeep Biswas) | 
|  |  | 
|  | #include "ceres/cuda_sparse_matrix.h" | 
|  |  | 
|  | #include <string> | 
|  |  | 
|  | #include "ceres/block_sparse_matrix.h" | 
|  | #include "ceres/casts.h" | 
|  | #include "ceres/cuda_vector.h" | 
|  | #include "ceres/internal/config.h" | 
|  | #include "ceres/internal/eigen.h" | 
|  | #include "ceres/linear_least_squares_problems.h" | 
|  | #include "ceres/triplet_sparse_matrix.h" | 
|  | #include "glog/logging.h" | 
|  | #include "gtest/gtest.h" | 
|  |  | 
|  | namespace ceres { | 
|  | namespace internal { | 
|  |  | 
|  | #ifndef CERES_NO_CUDA | 
|  |  | 
|  | class CudaSparseMatrixTest : public ::testing::Test { | 
|  | protected: | 
|  | void SetUp() final { | 
|  | std::string message; | 
|  | CHECK(context_.InitCuda(&message)) | 
|  | << "InitCuda() failed because: " << message; | 
|  | std::unique_ptr<LinearLeastSquaresProblem> problem = | 
|  | CreateLinearLeastSquaresProblemFromId(2); | 
|  | CHECK(problem != nullptr); | 
|  | A_.reset(down_cast<BlockSparseMatrix*>(problem->A.release())); | 
|  | CHECK(A_ != nullptr); | 
|  | CHECK(problem->b != nullptr); | 
|  | CHECK(problem->x != nullptr); | 
|  | b_.resize(A_->num_rows()); | 
|  | for (int i = 0; i < A_->num_rows(); ++i) { | 
|  | b_[i] = problem->b[i]; | 
|  | } | 
|  | x_.resize(A_->num_cols()); | 
|  | for (int i = 0; i < A_->num_cols(); ++i) { | 
|  | x_[i] = problem->x[i]; | 
|  | } | 
|  | CHECK_EQ(A_->num_rows(), b_.rows()); | 
|  | CHECK_EQ(A_->num_cols(), x_.rows()); | 
|  | } | 
|  |  | 
|  | std::unique_ptr<BlockSparseMatrix> A_; | 
|  | Vector x_; | 
|  | Vector b_; | 
|  | ContextImpl context_; | 
|  | }; | 
|  |  | 
|  | TEST_F(CudaSparseMatrixTest, RightMultiplyAndAccumulate) { | 
|  | std::string message; | 
|  | CompressedRowSparseMatrix A_crs( | 
|  | A_->num_rows(), A_->num_cols(), A_->num_nonzeros()); | 
|  | A_->ToCompressedRowSparseMatrix(&A_crs); | 
|  | CudaSparseMatrix A_gpu(&context_, A_crs); | 
|  | CudaVector x_gpu(&context_, A_gpu.num_cols()); | 
|  | CudaVector res_gpu(&context_, A_gpu.num_rows()); | 
|  | x_gpu.CopyFromCpu(x_); | 
|  |  | 
|  | const Vector minus_b = -b_; | 
|  | // res = -b | 
|  | res_gpu.CopyFromCpu(minus_b); | 
|  | // res += A * x | 
|  | A_gpu.RightMultiplyAndAccumulate(x_gpu, &res_gpu); | 
|  |  | 
|  | Vector res; | 
|  | res_gpu.CopyTo(&res); | 
|  |  | 
|  | Vector res_expected = minus_b; | 
|  | A_->RightMultiplyAndAccumulate(x_.data(), res_expected.data()); | 
|  |  | 
|  | EXPECT_LE((res - res_expected).norm(), | 
|  | std::numeric_limits<double>::epsilon() * 1e3); | 
|  | } | 
|  |  | 
|  | TEST(CudaSparseMatrix, CopyValuesFromCpu) { | 
|  | // A1: | 
|  | // [ 1 1 0 0 | 
|  | //   0 1 1 0] | 
|  | // A2: | 
|  | // [ 1 2 0 0 | 
|  | //   0 3 4 0] | 
|  | // b: [1 2 3 4]' | 
|  | // A1 * b = [3 5]' | 
|  | // A2 * b = [5 18]' | 
|  | TripletSparseMatrix A1(2, 4, {0, 0, 1, 1}, {0, 1, 1, 2}, {1, 1, 1, 1}); | 
|  | TripletSparseMatrix A2(2, 4, {0, 0, 1, 1}, {0, 1, 1, 2}, {1, 2, 3, 4}); | 
|  | Vector b(4); | 
|  | b << 1, 2, 3, 4; | 
|  |  | 
|  | ContextImpl context; | 
|  | std::string message; | 
|  | CHECK(context.InitCuda(&message)) << "InitCuda() failed because: " << message; | 
|  | auto A1_crs = CompressedRowSparseMatrix::FromTripletSparseMatrix(A1); | 
|  | CudaSparseMatrix A_gpu(&context, *A1_crs); | 
|  | CudaVector b_gpu(&context, A1.num_cols()); | 
|  | CudaVector x_gpu(&context, A1.num_rows()); | 
|  | b_gpu.CopyFromCpu(b); | 
|  | x_gpu.SetZero(); | 
|  |  | 
|  | Vector x_expected(2); | 
|  | x_expected << 3, 5; | 
|  | A_gpu.RightMultiplyAndAccumulate(b_gpu, &x_gpu); | 
|  | Vector x_computed; | 
|  | x_gpu.CopyTo(&x_computed); | 
|  | EXPECT_EQ(x_computed, x_expected); | 
|  |  | 
|  | auto A2_crs = CompressedRowSparseMatrix::FromTripletSparseMatrix(A2); | 
|  | A_gpu.CopyValuesFromCpu(*A2_crs); | 
|  | x_gpu.SetZero(); | 
|  | x_expected << 5, 18; | 
|  | A_gpu.RightMultiplyAndAccumulate(b_gpu, &x_gpu); | 
|  | x_gpu.CopyTo(&x_computed); | 
|  | EXPECT_EQ(x_computed, x_expected); | 
|  | } | 
|  |  | 
|  | TEST(CudaSparseMatrix, RightMultiplyAndAccumulate) { | 
|  | // A: | 
|  | // [ 1 2 0 0 | 
|  | //   0 3 4 0] | 
|  | // b: [1 2 3 4]' | 
|  | // A * b = [5 18]' | 
|  | TripletSparseMatrix A(2, 4, {0, 0, 1, 1}, {0, 1, 1, 2}, {1, 2, 3, 4}); | 
|  | Vector b(4); | 
|  | b << 1, 2, 3, 4; | 
|  | Vector x_expected(2); | 
|  | x_expected << 5, 18; | 
|  |  | 
|  | ContextImpl context; | 
|  | std::string message; | 
|  | CHECK(context.InitCuda(&message)) << "InitCuda() failed because: " << message; | 
|  | auto A_crs = CompressedRowSparseMatrix::FromTripletSparseMatrix(A); | 
|  | CudaSparseMatrix A_gpu(&context, *A_crs); | 
|  | CudaVector b_gpu(&context, A.num_cols()); | 
|  | CudaVector x_gpu(&context, A.num_rows()); | 
|  | b_gpu.CopyFromCpu(b); | 
|  | x_gpu.SetZero(); | 
|  |  | 
|  | A_gpu.RightMultiplyAndAccumulate(b_gpu, &x_gpu); | 
|  |  | 
|  | Vector x_computed; | 
|  | x_gpu.CopyTo(&x_computed); | 
|  |  | 
|  | EXPECT_EQ(x_computed, x_expected); | 
|  | } | 
|  |  | 
|  | TEST(CudaSparseMatrix, LeftMultiplyAndAccumulate) { | 
|  | // A: | 
|  | // [ 1 2 0 0 | 
|  | //   0 3 4 0] | 
|  | // b: [1 2]' | 
|  | // A'* b = [1 8 8 0]' | 
|  | TripletSparseMatrix A(2, 4, {0, 0, 1, 1}, {0, 1, 1, 2}, {1, 2, 3, 4}); | 
|  | Vector b(2); | 
|  | b << 1, 2; | 
|  | Vector x_expected(4); | 
|  | x_expected << 1, 8, 8, 0; | 
|  |  | 
|  | ContextImpl context; | 
|  | std::string message; | 
|  | CHECK(context.InitCuda(&message)) << "InitCuda() failed because: " << message; | 
|  | auto A_crs = CompressedRowSparseMatrix::FromTripletSparseMatrix(A); | 
|  | CudaSparseMatrix A_gpu(&context, *A_crs); | 
|  | CudaVector b_gpu(&context, A.num_rows()); | 
|  | CudaVector x_gpu(&context, A.num_cols()); | 
|  | b_gpu.CopyFromCpu(b); | 
|  | x_gpu.SetZero(); | 
|  |  | 
|  | A_gpu.LeftMultiplyAndAccumulate(b_gpu, &x_gpu); | 
|  |  | 
|  | Vector x_computed; | 
|  | x_gpu.CopyTo(&x_computed); | 
|  |  | 
|  | EXPECT_EQ(x_computed, x_expected); | 
|  | } | 
|  |  | 
|  | // If there are numerical errors due to synchronization issues, they will show | 
|  | // up when testing with large matrices, since each operation will take | 
|  | // significant time, thus hopefully revealing any potential synchronization | 
|  | // issues. | 
|  | TEST(CudaSparseMatrix, LargeMultiplyAndAccumulate) { | 
|  | // Create a large NxN matrix A that has the following structure: | 
|  | // In row i, only columns i and i+1 are non-zero. | 
|  | // A_{i, i} = A_{i, i+1} = 1. | 
|  | // There will be 2 * N - 1 non-zero elements in A. | 
|  | // X = [1:N] | 
|  | // Right multiply test: | 
|  | // b = A * X | 
|  | // Left multiply test: | 
|  | // b = A' * X | 
|  |  | 
|  | const int N = 10 * 1000 * 1000; | 
|  | const int num_non_zeros = 2 * N - 1; | 
|  | std::vector<int> row_indices(num_non_zeros); | 
|  | std::vector<int> col_indices(num_non_zeros); | 
|  | std::vector<double> values(num_non_zeros); | 
|  |  | 
|  | for (int i = 0; i < N; ++i) { | 
|  | row_indices[2 * i] = i; | 
|  | col_indices[2 * i] = i; | 
|  | values[2 * i] = 1.0; | 
|  | if (i + 1 < N) { | 
|  | col_indices[2 * i + 1] = i + 1; | 
|  | row_indices[2 * i + 1] = i; | 
|  | values[2 * i + 1] = 1; | 
|  | } | 
|  | } | 
|  | TripletSparseMatrix A(N, N, row_indices, col_indices, values); | 
|  | Vector x(N); | 
|  | for (int i = 0; i < N; ++i) { | 
|  | x[i] = i + 1; | 
|  | } | 
|  |  | 
|  | ContextImpl context; | 
|  | std::string message; | 
|  | CHECK(context.InitCuda(&message)) << "InitCuda() failed because: " << message; | 
|  | auto A_crs = CompressedRowSparseMatrix::FromTripletSparseMatrix(A); | 
|  | CudaSparseMatrix A_gpu(&context, *A_crs); | 
|  | CudaVector b_gpu(&context, N); | 
|  | CudaVector x_gpu(&context, N); | 
|  | x_gpu.CopyFromCpu(x); | 
|  |  | 
|  | // First check RightMultiply. | 
|  | { | 
|  | b_gpu.SetZero(); | 
|  | A_gpu.RightMultiplyAndAccumulate(x_gpu, &b_gpu); | 
|  | Vector b_computed; | 
|  | b_gpu.CopyTo(&b_computed); | 
|  | for (int i = 0; i < N; ++i) { | 
|  | if (i + 1 < N) { | 
|  | EXPECT_EQ(b_computed[i], 2 * (i + 1) + 1); | 
|  | } else { | 
|  | EXPECT_EQ(b_computed[i], i + 1); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | // Next check LeftMultiply. | 
|  | { | 
|  | b_gpu.SetZero(); | 
|  | A_gpu.LeftMultiplyAndAccumulate(x_gpu, &b_gpu); | 
|  | Vector b_computed; | 
|  | b_gpu.CopyTo(&b_computed); | 
|  | for (int i = 0; i < N; ++i) { | 
|  | if (i > 0) { | 
|  | EXPECT_EQ(b_computed[i], 2 * (i + 1) - 1); | 
|  | } else { | 
|  | EXPECT_EQ(b_computed[i], i + 1); | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | #endif  // CERES_NO_CUDA | 
|  |  | 
|  | }  // namespace internal | 
|  | }  // namespace ceres |