0 SNES Function norm 10.2781 0 KSP Residual norm 10.1755 1 KSP Residual norm 3.122 2 KSP Residual norm 1.98807 3 KSP Residual norm 0.523639 Linear solve converged due to CONVERGED_RTOL iterations 3 1 SNES Function norm 0.770366 0 KSP Residual norm 0.523639 1 KSP Residual norm 0.255653 2 KSP Residual norm 0.154552 3 KSP Residual norm 0.0968739 4 KSP Residual norm 0.0351983 Linear solve converged due to CONVERGED_RTOL iterations 4 2 SNES Function norm 0.0513487 0 KSP Residual norm 0.0351983 1 KSP Residual norm 0.0230681 2 KSP Residual norm 0.0106977 3 KSP Residual norm 0.00654554 4 KSP Residual norm 0.00404898 5 KSP Residual norm 0.00160906 Linear solve converged due to CONVERGED_RTOL iterations 5 3 SNES Function norm 0.00230473 0 KSP Residual norm 0.00160906 1 KSP Residual norm 0.000904566 2 KSP Residual norm 0.000549715 3 KSP Residual norm 0.000324851 4 KSP Residual norm 0.000168742 5 KSP Residual norm 7.79697e-05 Linear solve converged due to CONVERGED_RTOL iterations 5 4 SNES Function norm 0.000115341 0 KSP Residual norm 7.79697e-05 1 KSP Residual norm 4.57685e-05 2 KSP Residual norm 2.20533e-05 3 KSP Residual norm 1.44913e-05 4 KSP Residual norm 8.96527e-06 5 KSP Residual norm 4.83859e-06 Linear solve converged due to CONVERGED_RTOL iterations 5 5 SNES Function norm 6.79388e-06 0 KSP Residual norm 4.83859e-06 1 KSP Residual norm 2.77576e-06 2 KSP Residual norm 1.72657e-06 3 KSP Residual norm 1.03361e-06 4 KSP Residual norm 5.0434e-07 5 KSP Residual norm 2.71081e-07 Linear solve converged due to CONVERGED_RTOL iterations 5 6 SNES Function norm 3.65805e-07 0 KSP Residual norm 2.71081e-07 1 KSP Residual norm 1.67289e-07 2 KSP Residual norm 7.86136e-08 3 KSP Residual norm 4.9281e-08 4 KSP Residual norm 2.85897e-08 5 KSP Residual norm 1.65072e-08 Linear solve converged due to CONVERGED_RTOL iterations 5 7 SNES Function norm 2.17462e-08 L_2 Error: 0.000629252 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 7 SNES Object: 4 MPI processes type: newtonls maximum iterations=50, maximum function evaluations=10000 tolerances: relative=1e-08, absolute=1e-50, solution=1e-08 total number of linear solver iterations=32 total number of function evaluations=8 norm schedule ALWAYS SNESLineSearch Object: 4 MPI processes type: bt interpolation: cubic alpha=1.000000e-04 maxlambda=1.000000e+00, minlambda=1.000000e-12 tolerances: relative=1.000000e-08, absolute=1.000000e-15, lambda=1.000000e-08 maximum iterations=40 KSP Object: 4 MPI processes type: gmres restart=100, using classical (unmodified) Gram-Schmidt orthogonalization with no iterative refinement happy breakdown tolerance=1e-30 maximum iterations=10000, initial guess is zero tolerances: relative=0.1, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test PC Object: 4 MPI processes type: hpddm levels: 2 Neumann matrix attached? TRUE shared subdomain KSP between SLEPc and PETSc? FALSE coarse correction: DEFLATED on process #0, value (+ threshold if available) for selecting deflation vectors: 0 (0.1) grid and operator complexities: 1.0039 1.00069 KSP Object: (pc_hpddm_levels_1_) 4 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning not checking for convergence PC Object: (pc_hpddm_levels_1_) 4 MPI processes type: shell no name linear system matrix, which is also used to construct the preconditioner: Mat Object: 4 MPI processes type: mpiaij rows=513, cols=513 total: nonzeros=4345, allocated nonzeros=4345 total number of mallocs used during MatSetValues calls=0 not using I-node (on process 0) routines PC Object: (pc_hpddm_levels_1_) 4 MPI processes type: bjacobi number of blocks = 4 Local solver information for first block is in the following KSP and PC objects on rank 0: Use -pc_hpddm_levels_1_ksp_view ::ascii_info_detail to display information for all blocks KSP Object: (pc_hpddm_levels_1_sub_) 1 MPI process type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning not checking for convergence PC Object: (pc_hpddm_levels_1_sub_) 1 MPI process type: lu out-of-place factorization tolerance for zero pivot 2.22045e-14 matrix ordering: nd factor fill ratio given 5., needed 1.9272 Factored matrix: Mat Object: (pc_hpddm_levels_1_sub_) 1 MPI process type: seqaij rows=109, cols=109 package used to perform factorization: petsc total: nonzeros=1509, allocated nonzeros=1509 not using I-node routines linear system matrix, which is also used to construct the preconditioner: Mat Object: (pc_hpddm_levels_1_sub_) 1 MPI process type: seqaij rows=109, cols=109 total: nonzeros=783, allocated nonzeros=783 total number of mallocs used during MatSetValues calls=0 not using I-node routines linear system matrix, which is also used to construct the preconditioner: Mat Object: 4 MPI processes type: mpiaij rows=513, cols=513 total: nonzeros=4345, allocated nonzeros=4345 total number of mallocs used during MatSetValues calls=0 not using I-node (on process 0) routines KSP Object: (pc_hpddm_coarse_) 2 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning not checking for convergence PC Object: (pc_hpddm_coarse_) 2 MPI processes type: redundant First (color=0) of 2 PCs follows KSP Object: (pc_hpddm_coarse_redundant_) 1 MPI process type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning not checking for convergence PC Object: (pc_hpddm_coarse_redundant_) 1 MPI process type: cholesky out-of-place factorization tolerance for zero pivot 2.22045e-14 matrix ordering: natural factor fill ratio given 5., needed 1. Factored matrix: Mat Object: (pc_hpddm_coarse_redundant_) 1 MPI process type: seqsbaij rows=2, cols=2 package used to perform factorization: petsc total: nonzeros=3, allocated nonzeros=3 linear system matrix, which is also used to construct the preconditioner: Mat Object: 1 MPI process type: seqsbaij rows=2, cols=2 total: nonzeros=3, allocated nonzeros=3 total number of mallocs used during MatSetValues calls=0 linear system matrix, which is also used to construct the preconditioner: Mat Object: (pc_hpddm_coarse_) 2 MPI processes type: mpisbaij rows=2, cols=2 total: nonzeros=3, allocated nonzeros=3 total number of mallocs used during MatSetValues calls=0 linear system matrix, which is also used to construct the preconditioner: Mat Object: 4 MPI processes type: mpiaij rows=513, cols=513 total: nonzeros=4345, allocated nonzeros=4345 total number of mallocs used during MatSetValues calls=0 not using I-node (on process 0) routines