Skip to main content

BRIEF RESEARCH REPORT article

Front. Neurosci., 19 May 2023
Sec. Visual Neuroscience
This article is part of the Research Topic Neuroscience-driven Visual Representation View all 12 articles

3D shape reconstruction with a multiple-constraint estimation approach

\r\nXia Chen,,Xia Chen1,2,3Zhan-Li Sun,
Zhan-Li Sun4,5*Ying ZhangYing Zhang4
  • 1School of Information and Computer, Anhui Agricultural University, Hefei, China
  • 2Key Laboratory of Intelligent Computing and Signal Processing of Ministry of Education, Institute of Physical Science and Information Technology, Anhui University, Hefei, China
  • 3Anhui Provincial Key Laboratory of Multimodal Cognitive Computation, Anhui University, Hefei, China
  • 4School of Electrical Engineering and Automation, Anhui University, Hefei, China
  • 5Information Materials and Intelligent Sensing Laboratory of Anhui Province, Anhui University, Hefei, China

In this study, a multiple-constraint estimation algorithm is presented to estimate the 3D shape of a 2D image sequence. Given the training data, a sparse representation model with an elastic net, i.e., l1−norm and l2−norm constraints, is devised to extract the shape bases. In the sparse model, the l1−norm and l2−norm constraints are enforced to regulate the sparsity and scale of coefficients, respectively. After obtaining the shape bases, a penalized least-square model is formulated to estimate 3D shape and motion, by considering the orthogonal constraint of the transformation matrix, and the similarity constraint between the 2D observations and the shape bases. Moreover, an Augmented Lagrange Multipliers (ALM) iterative algorithm is adopted to solve the optimization of the proposed approach. Experimental results on the well-known CMU image sequences demonstrate the effectiveness and feasibility of the proposed model.

1. Introduction

As an important component of computer vision, 3D shape reconstruction has been widely used in many applications (Li et al., 2016, 2018; Adamkiewicz et al., 2022; Chiang et al., 2022; Fombona-Pascual et al., 2022; Jang et al., 2022; Lu et al., 2022; Nian et al., 2022a,b; Wang et al., 2022; Wen et al., 2022). Among the various 3D shape reconstruction methods, non-rigid structure from motion (NRSFM) offers a technique to simultaneously recover the 3D structures and motions of an object, by using the 2D landmarks in a series of images (Graßhof and Brandt, 2022; Kumar and Van Gool, 2022; Song et al., 2022). Nevertheless, NRSFM is still an underconstrained and challenging issue because of lacking any prior knowledge of 3D structure deformation.

To alleviate the uncertainty, the various constraints are exploited constantly. Bregler et al. (2000), proposed a low-rank constraint-based approach to decompose the observation matrix into a motion factor and a shape basis. In order to reduce the number of the unknown variables proposed by Bregler et al. (2000), a point trajectory approach was presented by Akhter et al. (2010) by using the predefined bases of discrete cosine transform (DCT). However, the high-frequency deformation cannot be reconstructed well via this trajectory representation because of the low-rank constraint. Gotardo and Martinez (2011) modeled a smoothly deforming 3D shape as a single point moving along a smooth time trajectory within a linear shape space. In addition to the low-rank constraint, the higher frequency DCT was adopted to capture the high-frequency deformation.

For the low-rank constraint methods, it is difficult to determine the optimal number of shape bases or trajectory bases. To solve this problem, a Procrustean normal distribution (PND) model was presented by Lee et al. (2013) to separate the motion and deformation components strictly, without any additional constraints or prior knowledge. The experimental results demonstrate the performance of PND. Subsequently, the Procrustean Markov Process (PMP) algorithm was proposed by Lee et al. (2014), by combing in a first-order Markov model representing the smoothness between two adjacent frames with PND. Lee et al. (2016) reported a consensus of non-rigid reconstruction (CNR) approach to estimate 3D shapes based on local patches. However, the reconstruction performance of these methods may degrade significantly when the number of images becomes small, especially for a single image.

Referring to the active shape model (Cootes et al., 1995), a limb length constraint-based approach was presented by Wang et al. (2014) to estimate the 3D shape of an object from a single 2D image, by solving a l1−norm minimization problem. Zhou et al. (2013) proposed a sparse representation-based convex relaxation approach (CRA) to guarantee global optimality. The shape bases were extracted from a given training data by using a sparse representation model. The corresponding coefficients were obtained by adopting a convex relaxation assumption. A prominent advantage of CRA is that the algorithm can deal with a single image.

To further enhance the performance of the CRA algorithm, a multiple-constraint-based estimation approach is proposed to estimate the 3D shape of a 2D image sequence. Inspired by Zhang and Xing (2017), a dictionary learning model with l1−norm and l2−norm, i.e., elastic net, is constructed to extract more effective shape bases from a given training set. Referring to (Cheng et al., 2015), a penalized least-square model is constructed to estimate 3D shape and motion, by considering the orthogonal constraint of the transformation matrix and the similarity constraint between the 2D observations and the shape bases. In addition, an augmented Lagrange multipliers (ALM) iterative algorithm is developed to optimize the reconstruction model. The effectiveness and feasibility of the proposed algorithm are verified on the well-known CMU image sequences.

The rest of this article is organized as follows. A detailed description of the designed MCM-RR approach is introduced in Section 2. In Section 3, we report the experimental results. Finally, the article is concluded in Section 4.

2. Methods

According to the shape-space model by Zhou et al. (2013), the unknown 3D shape S ∈ ℝp is constructed as a linear combination of a few shape bases Bi3×p, i.e.,

S=i=1KciRiBi,    (1)

where p and K are the numbers of feature points and shape bases, respectively. The parameter ci and Ri3×3 denote the coefficient and rotation matrix, respectively. In terms of the weak-perspective projection model, the corresponding 2D observations are modeled as a matrix W ∈ ℝp,

W=i=1KMiBi.    (2)

The matrix Mi2×3 can be represented as

Mi=ciR~i,    (3)

where R~i2×3 is the first two rows of Ri. Combining the orthogonal constraint, the matrix Mi satisfies

MiMiT=ci2I2,    (4)

where I22×2 is an identity matrix. The 3D shape, i.e., z−coordinates, and the motion parameters ci and Ri, are estimated by utilizing the observations W, i.e., the (x, y) coordinates of feature points.

In the proposed method, the shape bases B ∈ ℝ3K×p are extracted via a sparse model with the elastic net constraint. The B is the stacking of Bi(i = 1, ..., K). The matrix M are solved by a penalized least-square model. Given M, the parameters ci and Ri are derived via refinement decompose (Zhou et al., 2013). After obtaining ci, Ri and Bi, the unknown 3D shape can be computed via (1). The pseudocode of the proposed algorithm is summarized in 1. The pseudocode of the proposed algorithm is summarized in Algorithm 1.

ALGORITHM 1
www.frontiersin.org

Algorithm 1. Pseudocode of the MCM-RR algorithm.

2.1. Extraction of shape bases via a sparse model with elastic net constraint

For a given 3D training set A ∈ ℝ3p×F, i.e., the (x, y, z) coordinates of feature points of training images, the shape bases N ∈ ℝ3p×K and the coefficient matrix X ∈ ℝK×F can be obtained from the following sparse model:

minN1,,NK 12A-NXF2+λ(τX1+(1-τ)X22)s.t.NiF1,               Xij0,i[1,K],j[1,F],    (5)

where F and τ are the number of frames and a weight coefficient, respectively. The Ni3p×1 is the i-th column of N. The linear combination of l1−norm and l2−norm, called elastic net constraint, are enforced to constraint the sparsity of coefficients X as well as scale. The parameter λ is a trade-off parameter between the reconstruction error and the elastic net constraint.

For (5), we first compute the partial differentials of X and N, i.e.,

X=(Nt)T(A-NtX)+λ(τIKF+2(1-τ)X),    (6)
N=(A-N(Xt+1)T)  (Xt+1)T,    (7)

where IKF is a K × F identity matrix. Thereafter, X and N can be updated alternately as

Xt+1=Xt-ϕ1X,    (8)
Nt+1=Nt-ϕ2N,    (9)

where ϕ1 and ϕ2 are the step size of ∂X and ∂N, respectively. After convergence, the shape bases B can be obtained by a re-arrangement of N.

2.2. 3D shape estimation via a penalized least-square model with similarity constraint

In terms of (2), the proposed penalized least-square model, including a relaxed orthogonality constraint (Zhou et al., 2013) and a similarity constraint (Cheng et al., 2015) can be formulated as

minM~,Z 12W-ZB~F2+αi=1KMi2+β2ZD22  s.t.  M~=Z,    (10)

where Z ∈ ℝ2×3K is an auxiliary variable and M~=[M1,,MK], B~=[B1T,,BKT]T. The parameters α and β are used to weight the two regularization terms. The diagonal matrix D ∈ ℝ3K×3K is represented as

D=(D̄I3).    (11)

For the diagonal similarity matrix D̄K×K, the diagonal element di is computed as

di=exp(W-ΠBi2γ2),    (12)

where Π = [1, 0, 0;0, 1, 0], γ2 is the parameter of an exponential function.

With the ALM iterative algorithm, the penalized least-square model (10) can be reformulated as

L=12W-ZB~F2+αi=1KMi2+Y,M~-Z        +β2ZD22+μ2M~-ZF2,    (13)

where Y and μ are a dual variable and a weight of penalty term, respectively. In (13), there are four unknown variables M~, Z, Y, and μ. The solutions can be solved by the alternating direction method of multipliers (ADMM).

First, the optimal M~ at the (t + 1)th iteration can be formulated as

M~t+1=arg minM~i=1K12Mi-PitF2+αμMi2,    (14)

where Pit is the ith column-triple of Zt-1μYt. According to the proximal problem (Zhou et al., 2013), Mit+1 can be computed as

Mit+1=Udiag(Σ-αμPl1(Σμα))VT,i[1,K],    (15)

where UΣVT=svd(Pit). The operation Pl1(·) denotes the projection of a vector to the unit l1−norm ball (Zhou et al., 2013).

Similarity, the optimal Z at the (t + 1)th iteration can be formulated as

Zt+1=arg minZ12W-ZB~F2+Yt,M~t+1-Z                  +β2ZD22+μ2M~t+1-ZF2.    (16)

We compute the one-order partial derivative of (16) with respect to Z and set it as zero. Thereafter, Zt+1 can be given by

Zt+1=(WB~T+μM~t+1+Yt)(B~B~T+μI+βDDT)-1.    (17)

Afterward, the optimal Y at the (t + 1)th iteration can be computed as

Yt+1=Yt+μ(M~t+1-Zt+1).    (18)

Given a weight τ, the coefficient μ at the (t + 1)th iteration can be given by

μt+1={2μt,   if   δ1>τδ2,μt/2,   if   δ2>τδ1,    (19)

where

δ1=M~t+1-Zt+1FZtF,δ2=Zt+1-ZtFZtF.    (20)

The iterations are repeated until

δ1<ε    &    δ2<ε,    (21)

where ε is a small threshold value. After obtaining Mi, the unknown 3D shape can be reconstructed by refinement reconstruction (Zhou et al., 2013).

In the refinement reconstruction, we assume that the rotation matrices of each shape base are equal, denoted as R̄. Thereafter, ci and R̄ can be estimated by the following rotation synchronization model

minc,R̄   i=1kMi-ciR̄F2s.t.   R̄R̄T=I2,,    (22)

which can be solved via the alternating minimization (Zhou et al., 2013). Finally, the 3D shape S can be estimated after Mi is obtained.

3. Experimental results

3.1. Experimental comparison of different algorithms

The performance evaluation of the proposed 3D shape reconstruction model (denoted as MCM-RR) is carried out on eight motion categories (walk, run, jump, climb, box, dance, sit, and basketball) from the CMU motion capture dataset (Zhou et al., 2013). Figure 1 shows one frame of those eight categories.

FIGURE 1
www.frontiersin.org

Figure 1. One frame of those eight categories.

In the experiments, the performance of several state-of-the-art 3D shape estimation methods are used to compare with the presented approach, including PND2 (Lee et al., 2013), CNR (Lee et al., 2016), PMP (Lee et al., 2014), and CRA (Zhou et al., 2013).

Mean error ξ of 3D shapes is calculated as the performance indicator to measure the estimation results:

ξ=1Ft=1FS~t-StF2,    (23)

where S~t3×p and St3×p are the reconstructed 3D structure and real 3D structure of tth frame, respectively.

Table 1 displays the mean and standard deviation (μ±σ) of reconstruction errors ξ of eight motion categories for the five methods, respectively. The best results are highlighted in red, whereas the second best is in blue.

TABLE 1
www.frontiersin.org

Table 1. Mean and standard deviation (μ±σ) of the 3D reconstruction errors ξ of eight motion categories for five methods.

Table 1 shows the estimation errors of the last two methods are clearly less than that of the first triple algorithms. Among eight categories, the mean reconstruction errors of MCM-RR are the lowest compared to CRA. Moreover, the standard deviations of MCM-RR are less than that of CRA among most categories. Therefore, compared to CRA, both accuracy and robustness are effectively improved for the proposed method.

Compared to CRA, the 3D reconstruction error decreased the percentage ξp(%) of MCM-RR can be computed as

ξp=ξCRA-ξMCM-RRξCRA×100%.    (24)

From Table 2, we can see that the mean reconstruction errors of MCM-RR decreased about 5.48%∽21.93% compared to CRA. Thus, MCM-RR has a better 3D reconstruction performance than CRA for the eight motion categories.

TABLE 2
www.frontiersin.org

Table 2. Corresponding 3D reconstruction error decreasing percentage ξp(%) of MCM-RR compared to CRA for eight motion categories.

Take one frame of Jump as an example. Figure 2 displays a comparison of reconstructed shapes between MCM-RR and the other methods from three different viewpoints. From Figure 2, we can see that compared to other methods, most estimated shapes of MCM-RR are closer to real points than that of the other methods.

FIGURE 2
www.frontiersin.org

Figure 2. Comparisons of estimated shapes for single frame of Jump between MCM-RR and other methods from three different viewpoints. The symbol “°” denotes the observed real points, whereas “+” denotes reconstructed points.

3.2. Ablation experiment

In order to verify the feasibility of the proposed two strategies, the elastic net (denoted as CRA-EN) and similarity constraint (denoted as CRA-SC) are separately applied to the original algorithm CRA. Table 3 displays the mean and standard deviation (μ±σ) of 3D reconstruction errors ξ of eight motion categories for the four methods, respectively. Compared to CRA, both the elastic net and similarity constraint can decrease the 3D reconstruction errors. Therefore, the 3D reconstruction performance can be effectively improved once the two methods are simultaneously designed into CRA.

TABLE 3
www.frontiersin.org

Table 3. Mean and standard deviation (μ±σ) of the 3D reconstruction errors ξ of eight motion categories for four methods.

4. Conclusion

In this study, a multiple-constraint algorithm is devised to estimate the 3D shape of a 2D image sequence. Experimental results on the well-known CMU datasets demonstrated that the proposed methods have higher accuracies and more robustness. Compared with CRA, the 3D reconstruction error is decreased by at least 5.48%.

Data availability statement

The datasets used in this article is from a public datesets, and it can be found in the CMU Graphics Lab Motion Capture Database.

Author contributions

XC proposed the initial research idea, conducted the experiments, and wrote the manuscript. Z-LS supervised the work and advised the entire research process. YZ collected the dataset, analyzed the formal, and revised the manuscript. All authors reviewed and approved the final manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 61972002), the University Natural Science Research Project of Anhui Province (No. KJ2021A0180), Natural Science Foundation of Anhui Agricultural University (No. K2148001), Research Talents Stable Project of Anhui Agricultural University (No. rc482004), Key Laboratory of Intelligent Computing & Signal Processing, Ministry of Education (Anhui University) (No. 2020A002), Anhui Provincial Key Laboratory of Multimodal Cognitive Computation (Anhui University) (No. MMC202004), and the Anhui Provincial Natural Science Foundation (No. 2108085MC96).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Adamkiewicz, M., Chen, T., Caccavale, A., Gardner, R., Culbertson, P., Bohg, J., et al. (2022). Vision-only robot navigation in a neural radiance world. IEEE Robot. Automat. Lett. 7, 4606–4613. doi: 10.1109/LRA.2022.3150497

CrossRef Full Text | Google Scholar

Akhter, I., Sheikh, Y., Khan, S., and Kanade, T. (2010). Trajectory space: a dual representation for nonrigid structure from motion. IEEE Trans. Pattern Anal. Mach. Intell. 33, 1442–1456. doi: 10.1109/TPAMI.2010.201

PubMed Abstract | CrossRef Full Text | Google Scholar

Bregler, C., Hertzmann, A., and Biermann, H. (2000). “Recovering non-rigid 3d shape from image streams,” in Proceedings IEEE Conference on Computer Vision and Pattern Recognition (Hilton Head, SC: IEEE), 690–696.

Google Scholar

Cheng, J., Yin, F., Wong, D. W. K., Tao, D., and Liu, J. (2015). Sparse dissimilarity-constrained coding for glaucoma screening. IEEE Trans. Biomed. Eng. 62, 1395–1403. doi: 10.1109/TBME.2015.2389234

PubMed Abstract | CrossRef Full Text | Google Scholar

Chiang, F.-K., Shang, X., and Qiao, L. (2022). Augmented reality in vocational training: a systematic review of research and applications. Comput. Hum. Behav. 129, 107125. doi: 10.1016/j.chb.2021.107125

CrossRef Full Text | Google Scholar

Cootes, T. F., Taylor, C. J., Cooper, D. H., and Graham, J. (1995). Active shape models-their training and application. Comput. Vis. Image Understand. 61, 38–59.

Google Scholar

Fombona-Pascual, A., Fombona, J., and Vicente, R. (2022). Augmented reality, a review of a way to represent and manipulate 3d chemical structures. J. Chem. Inform. Model. 62, 1863–1872. doi: 10.1021/acs.jcim.1c01255

PubMed Abstract | CrossRef Full Text | Google Scholar

Gotardo, P. F., and Martinez, A. M. (2011). Computing smooth time trajectories for camera and deformable shape in structure from motion with occlusion. IEEE Trans. Pattern Anal. Mach. Intell. 33, 2051–2065. doi: 10.1109/TPAMI.2011.50

PubMed Abstract | CrossRef Full Text | Google Scholar

Graßhof, S., and Brandt, S. S. (2022). “Tensor-based non-rigid structure from motion,” in Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition (Waikoloa, HI: IEEE), 3011–3020.

Google Scholar

Jang, H., Sedaghat, S., Athertya, J. S., Moazamian, D., Carl, M., Ma, Y., et al. (2022). Feasibility of ultrashort echo time quantitative susceptibility mapping with a 3d cones trajectory in the human brain. Front. Neurosci. 16, 1033801. doi: 10.3389/fnins.2022.1033801

PubMed Abstract | CrossRef Full Text | Google Scholar

Kumar, S., and Van Gool, L. (2022). “Organic priors in non-rigid structure from motion,” in Proceedings of the European Conference on Computer Vision (Springer), 71–88.

Google Scholar

Lee, M., Cho, J., Choi, C.-H., and Oh, S. (2013). “Procrustean normal distribution for non-rigid structure from motion,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Portland, OR: IEEE), 1280–1287.

PubMed Abstract | Google Scholar

Lee, M., Cho, J., and Oh, S. (2016). “Consensus of non-rigid reconstructions,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Las Vegas, NV: IEEE), 4670–4678.

Google Scholar

Lee, M., Choi, C.-H., and Oh, S. (2014). “A procrustean markov process for non-rigid structure recovery,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Columbus, OH: IEEE), 1550–1557.

Google Scholar

Li, T., Cheng, B., Ni, B., Liu, G., and Yan, S. (2016). Multitask low-rank affinity graph for image segmentation and image annotation. ACM Trans. Intell. Syst. Technol. 7, 1–18. doi: 10.1145/2856058

CrossRef Full Text | Google Scholar

Li, T., Wang, Y., Hong, R., Wang, M., and Wu, X. (2018). PDisVPL: probabilistic discriminative visual part learning for image classification. IEEE MultiMedia 25, 34–45. doi: 10.1109/MMUL.2018.2873499

PubMed Abstract | CrossRef Full Text | Google Scholar

Lu, W., Li, Z., Li, Y., Li, J., Chen, Z., Feng, Y., et al. (2022). A deep learning model for three-dimensional nystagmus detection and its preliminary application. Front. Neurosci. 16, 930028. doi: 10.3389/fnins.2022.930028

PubMed Abstract | CrossRef Full Text | Google Scholar

Nian, F., Li, T., Bao, B.-K., and Xu, C. (2022a). Relative coordinates constraint for face alignment. Neurocomputing 395, 119–127. doi: 10.1016/j.neucom.2017.12.071

CrossRef Full Text | Google Scholar

Nian, F., Sun, J., Jiang, D., Zhang, J., Li, T., and Lu, W. (2022b). Predicting dose-volume histogram of organ-at-risk using spatial geometric-encoding network for esophageal treatment planning. J. Ambient Intell. Smart Environ. 14, 25–37. doi: 10.3233/AIS-210084

CrossRef Full Text | Google Scholar

Song, J., Patel, M., Jasour, A., and Ghaffari, M. (2022). A closed-form uncertainty propagation in non-rigid structure from motion. IEEE Robot. Automat. Lett. 7, 6479–6486. doi: 10.1109/LRA.2022.3173733

CrossRef Full Text | Google Scholar

Wang, C., Wang, Y., Lin, Z., Yuille, A. L., and Gao, W. (2014). “Robust estimation of 3d human poses from a single image,” in Proceedings of the IEEE conference on Computer Vision and Pattern Recognition (Columbus, OH: IEEE), 2361–2368.

Google Scholar

Wang, T., Chen, B., Zhang, Z., Li, H., and Zhang, M. (2022). Applications of machine vision in agricultural robot navigation: a review. Comput. Electron. Agric. 198, 107085. doi: 10.1016/j.compag.2022.107085

CrossRef Full Text | Google Scholar

Wen, X., Zhou, J., Liu, Y.-S., Su, H., Dong, Z., and Han, Z. (2022). “3d shape reconstruction from 2d images with disentangled attribute flow,” in Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition (New Orleans, LA: IEEE), 3803–3813.

Google Scholar

Zhang, S., and Xing, W. (2017). “Object tracking with adaptive elastic net regression,” in Proceedings of the IEEE International Conference on Image Processing (Honolulu, HI: IEEE), 2597–2601.

Google Scholar

Zhou, X., Zhu, M., Leonardos, S., and Daniilidis, K. (2013). Sparse representation for 3d shape estimation: a convex relaxation approach. IEEE Trans. Pattern Anal. Mach. Intell. 39, 1648–1661. doi: 10.1109/TPAMI.2016.2605097

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: non-rigid structure from motion, elastic net, similarity constraint, Augmented Lagrange multipliers, 3D reconstruction

Citation: Chen X, Sun Z-L and Zhang Y (2023) 3D shape reconstruction with a multiple-constraint estimation approach. Front. Neurosci. 17:1191574. doi: 10.3389/fnins.2023.1191574

Received: 22 March 2023; Accepted: 17 April 2023;
Published: 19 May 2023.

Edited by:

Fudong Nian, Hefei University, China

Reviewed by:

Shengsheng Qian, Chinese Academy of Sciences (CAS), China
Sisi You, Nanjing University of Posts and Telecommunications, China
Fan Qi, Tianjin University of Technology, China

Copyright © 2023 Chen, Sun and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Zhan-Li Sun, zhlsun2006@126.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.