lid velocity = 0.000106281, prandtl # = 1., grashof # = 1. 0 SNES Function norm 1.036007954337e-02 0 KSP Residual norm 4.888657920112e-01 1 KSP Residual norm 3.141627616286e-01 2 KSP Residual norm 2.358812378525e-01 3 KSP Residual norm 1.259045437573e-01 4 KSP Residual norm 7.408575133258e-02 5 KSP Residual norm 4.159658325629e-02 6 KSP Residual norm 2.181861584091e-02 7 KSP Residual norm 1.076244407256e-02 8 KSP Residual norm 5.242924144023e-03 9 KSP Residual norm 2.588726461570e-03 10 KSP Residual norm 1.400700093556e-03 11 KSP Residual norm 7.759158101290e-04 12 KSP Residual norm 4.655744671687e-04 13 KSP Residual norm 3.244240900196e-04 14 KSP Residual norm 2.327144448481e-04 15 KSP Residual norm 1.500951563875e-04 16 KSP Residual norm 8.327955132177e-05 17 KSP Residual norm 4.672468799190e-05 18 KSP Residual norm 2.889321996959e-05 19 KSP Residual norm 1.903514112532e-05 20 KSP Residual norm 1.298123631775e-05 21 KSP Residual norm 8.775705379722e-06 22 KSP Residual norm 5.330175482523e-06 23 KSP Residual norm 2.869775929854e-06 1 SNES Function norm 1.370527153606e-06 0 KSP Residual norm 1.177060567864e-05 1 KSP Residual norm 6.928128579778e-06 2 KSP Residual norm 4.304601930959e-06 3 KSP Residual norm 2.454875323346e-06 4 KSP Residual norm 1.512790232898e-06 5 KSP Residual norm 1.001915525185e-06 6 KSP Residual norm 7.387623107528e-07 7 KSP Residual norm 4.953711915641e-07 8 KSP Residual norm 3.144821641519e-07 9 KSP Residual norm 2.134819078071e-07 10 KSP Residual norm 1.518551249827e-07 11 KSP Residual norm 1.014185303569e-07 12 KSP Residual norm 5.608673008996e-08 13 KSP Residual norm 3.030519067648e-08 14 KSP Residual norm 1.786947998191e-08 15 KSP Residual norm 1.177816744649e-08 16 KSP Residual norm 8.997335866987e-09 17 KSP Residual norm 6.356106556680e-09 18 KSP Residual norm 4.147037399694e-09 19 KSP Residual norm 2.711154831563e-09 20 KSP Residual norm 1.750865415567e-09 21 KSP Residual norm 1.127681613823e-09 22 KSP Residual norm 7.783034009647e-10 23 KSP Residual norm 5.047971816405e-10 24 KSP Residual norm 3.137090395920e-10 25 KSP Residual norm 2.182629803057e-10 26 KSP Residual norm 1.265836293994e-10 27 KSP Residual norm 7.756916170515e-11 2 SNES Function norm 2.339734806142e-11 SNES Object: 4 MPI processes type: newtonls maximum iterations=50, maximum function evaluations=10000 tolerances: relative=1e-08, absolute=1e-50, solution=1e-08 total number of linear solver iterations=50 total number of function evaluations=3 norm schedule ALWAYS Jacobian is built using colored finite differences on a DMDA SNESLineSearch Object: 4 MPI processes type: bt interpolation: cubic alpha=1.000000e-04 maxlambda=1.000000e+00, minlambda=1.000000e-12 tolerances: relative=1.000000e-08, absolute=1.000000e-15, lambda=1.000000e-08 maximum iterations=40 KSP Object: 4 MPI processes type: gmres restart=30, using classical (unmodified) Gram-Schmidt orthogonalization with no iterative refinement happy breakdown tolerance=1e-30 maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test PC Object: 4 MPI processes type: ml type is MULTIPLICATIVE, levels=5 cycles=v Cycles per PCApply=1 Using externally compute Galerkin coarse grid matrices Coarse grid solver -- level 0 ------------------------------- KSP Object: (mg_coarse_) 4 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning not checking for convergence PC Object: (mg_coarse_) 4 MPI processes type: redundant First (color=0) of 4 PCs follows KSP Object: (mg_coarse_redundant_) 1 MPI process type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning not checking for convergence PC Object: (mg_coarse_redundant_) 1 MPI process type: lu out-of-place factorization tolerance for zero pivot 2.22045e-14 using diagonal shift on blocks to prevent zero pivot [INBLOCKS] matrix ordering: nd factor fill ratio given 5., needed 1. Factored matrix: Mat Object: (mg_coarse_redundant_) 1 MPI process type: seqaij rows=16, cols=16, bs=4 package used to perform factorization: petsc total: nonzeros=256, allocated nonzeros=256 using I-node routines: found 4 nodes, limit used is 5 linear system matrix, which is also used to construct the preconditioner: Mat Object: 1 MPI process type: seqaij rows=16, cols=16, bs=4 total: nonzeros=256, allocated nonzeros=256 total number of mallocs used during MatSetValues calls=0 using I-node routines: found 4 nodes, limit used is 5 linear system matrix, which is also used to construct the preconditioner: Mat Object: 4 MPI processes type: mpiaij rows=16, cols=16, bs=4 total: nonzeros=256, allocated nonzeros=256 total number of mallocs used during MatSetValues calls=0 using I-node (on process 0) routines: found 1 nodes, limit used is 5 Down solver (pre-smoother) on level 1 ------------------------------- KSP Object: (mg_levels_1_) 4 MPI processes type: richardson damping factor=1. maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning not checking for convergence PC Object: (mg_levels_1_) 4 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix, which is also used to construct the preconditioner: Mat Object: 4 MPI processes type: mpiaij rows=88, cols=88, bs=4 total: nonzeros=4960, allocated nonzeros=4960 total number of mallocs used during MatSetValues calls=0 using I-node (on process 0) routines: found 4 nodes, limit used is 5 Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 2 ------------------------------- KSP Object: (mg_levels_2_) 4 MPI processes type: richardson damping factor=1. maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning not checking for convergence PC Object: (mg_levels_2_) 4 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix, which is also used to construct the preconditioner: Mat Object: 4 MPI processes type: mpiaij rows=764, cols=764, bs=4 total: nonzeros=35672, allocated nonzeros=35672 total number of mallocs used during MatSetValues calls=0 using I-node (on process 0) routines: found 50 nodes, limit used is 5 Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 3 ------------------------------- KSP Object: (mg_levels_3_) 4 MPI processes type: richardson damping factor=1. maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning not checking for convergence PC Object: (mg_levels_3_) 4 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix, which is also used to construct the preconditioner: Mat Object: 4 MPI processes type: mpiaij rows=7156, cols=7156, bs=4 total: nonzeros=259236, allocated nonzeros=259236 total number of mallocs used during MatSetValues calls=0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 4 ------------------------------- KSP Object: (mg_levels_4_) 4 MPI processes type: richardson damping factor=1. maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning not checking for convergence PC Object: (mg_levels_4_) 4 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix, which is also used to construct the preconditioner: Mat Object: 4 MPI processes type: mpiaij rows=37636, cols=37636, bs=4 total: nonzeros=746512, allocated nonzeros=746512 total number of mallocs used during MatSetValues calls=0 Up solver (post-smoother) same as down solver (pre-smoother) linear system matrix, which is also used to construct the preconditioner: Mat Object: 4 MPI processes type: mpiaij rows=37636, cols=37636, bs=4 total: nonzeros=746512, allocated nonzeros=746512 total number of mallocs used during MatSetValues calls=0 Number of SNES iterations = 2