• About The Regularized Singularity

The Regularized Singularity

~ The Eyes of a citizen; the voice of the silent

The Regularized Singularity

Category Archives: Uncategorized

Designing New Schemes Based On Van Leer’s Ideas

11 Saturday Jan 2014

Posted by Bill Rider in Uncategorized

≈ 1 Comment

Given the broader spectrum of variables to choose from new schemes with exceptional properties can be developed.  One can blend volume-averaged, point, moment and derivatives together.  These schemes are all compact and have exceptional numerical properties.  We introduced them earlier as the extensions of the basic interpolations employed by Van Leer, but using the same principles.

On thing to recognize immediately is that the famous PPM scheme is a cell-centered implementation of the basic approach associated with Scheme V.  The difference is the construction of edge-centered data from cell-centered quantities.  The way that the original PPM overcame the issue was to use higher order approximations for the cell-edge quantities, specifically fourth-order edges for what is an intrinsically third-order scheme.  This allows the approximation to asymptotically approach fourth-order accuracy in the limit of vanishing Courant number.

More recently fifth, sixth and seventh order approximations have been used with PPM to good effect.  These have been applied to the use of PPM in global climate model dynamic cores as well as by Colella.  The author used this approach in his extrema-preserving methodology where the edge values are utilized to produce useful approximations in regions with smooth extrema.  The bottom line is that this approach has proven to be enormously flexible in practice due to the choice of the edge values coupled with a cell-average preserving reconstructing polynomial.  It is reasonable to have faith in the ability for other schemes presented next to be used in a similarly flexible manner.  Indeed the construction of approximations via polynomial reconstruction is highly amenable to such extensions as the geometric picture offered greatly aids the subsequent method development and testing.

One route to nonlinear stability are ENO methods that offer adaptive stencil selection.  These methods are based upon the choice of the “smoothest” stencil defined in some sense.  Another maxim from the ENO literature is the need to bias the stencils toward being upstream-centered.  Put differently, Van Leer’s paper presents a set of upstream-centered approximation that should be favored over other approximations.  Given this approach we might define an ENO scheme based on Scheme V as predominantly using that scheme, but choosing other more upwind or downwind stencils if the function being approximated is insufficiently smooth.

For a positive characteristic velocity an upwind stencil could define a parabola, by u^n_{j-1}, u^n_{j-1/2} and u^n_{j}.  This gives a parabola of

P(\xi) = u_j^n + (-\frac{5}{4}u_j^n - \frac{1}{4} u_{j-1}^n + \frac{3}{2} u^n_{j+1/2})\xi + (-\frac{9}{4} u_j^n + \frac{3}{4} u_{j-1}^n + \frac{3}{2} u_{j+1/2}^n)(\xi^2 -\frac{1}{12}).

This produces a semi-discrete scheme with third-order accuracy and a dispersion relation of \imath \theta + \frac{1}{18} \theta^4 + \frac{1}{270}\imath\theta^5+ O(\theta^6) .

A second stencil to choose from resulting in a third-order method is defined by u^n_{j-1}, u^n_{j+1/2} and u^n_{j}. This gives a parabola of

P(\xi) = u_j^n + (\frac{5}{2}u_j^n + \frac{1}{2} u_{j-1}^n - 3 u^n_{j-1/2})\xi + (\frac{3}{2} u_j^n + \frac{3}{2} u_{j-1}^n - 3 u_{j-1/2}^n)(\xi^2 -\frac{1}{12}).

This produces a semi-discrete scheme with third-order accuracy and a dispersion relation of \imath \theta - \frac{5}{72} \theta^4 + \frac{1}{270}\imath\theta^5+ O(\theta^6) .

The downwind stencil defined by  u^n_{j+1}, u^n_{j+1/2} and u^n_{j} ends up not being useful because it results in a scheme with centered support.  This gives a parabola of

P(\xi) = u_j^n + (3 u^n_{j-1/2}-\frac{5}{2}u_j^n + \frac{1}{2} u_{j+1}^n)\xi + (\frac{3}{2} u_j^n + \frac{3}{2} u_{j+1}^n - 3 u_{j-1/2}^n)(\xi^2 -\frac{1}{12}).

The updates become effectively decoupled because of cancelation of terms and the order of accuracy drops to second-order.  This produces a semi-discrete scheme with third-order accuracy and a dispersion relation of \imath \theta - \frac{1}{24} \imath\theta^3 + \frac{1}{1920}\imath\theta^5+ O(\theta^7) .

The same thing happens for a stencil of latex u^n_{j+1}$, u^n_{j-1/2} and u^n_{j}.

Now we can move to describing the performance of high-order methods based on Van Leer’s ideas, but extended to slightly larger support.  All these methods will, at most, reference their nearest neighbor cells.  The first is rather obvious and gives fifth-order accuracy.  The reconstruction is defined by cell-averages, u_j^n, u_{j-1}^n, u_{j+1}^n and cell-edge values, u_{j-1/2}^n, u_{j+1/2}^n.  This defines a quartic polynomial,

P(\xi) = u_j^n + \frac{1}{8}(10 u_{j+1/2}^n-10 u_{j-1/2}^n +u_{j-1}^n - u_{j+1}^n)\xi

+ \frac{1}{8} (30 u_{j-1/2}^{n} + 30 u_{j+1/2}^{n} -u_{j-1}^{n}-58 u_j^n -u_{j+1}^{n}) (\xi^2-\frac{1}{12})

+ \frac{1}{2}(u_{j+1}^n-u_{j-1}^n - 2u_{j+1/2}^n+2u_{j-1/2}^n)\xi^3

+\frac{1}{12} (-30 u_{j-1/2}^n - 30u_{j+1/2}^n +5u_{j-1}^n+50 u_{j}^n +5u_{j+1}^n)(\xi^4-\frac{1}{80}).

This produces a semi-discrete scheme with fifth-order accuracy and a dispersion relation of \imath \theta + \frac{1}{900} \theta^6 + \frac{31}{31500}\imath\theta^7+ O(\theta^8) .  This is exceedingly accurate with the dispersion relation peaking at \theta=\pi with an error of only 2.5%.

The next method was discussed and recommended by Phil Roe at the JRV symposium and a conversation at another meeting.  It uses the cell-average u_j^n, the edge values u_{j\pm1/2}^n, and the edge-first-derivatives, s_{j\pm1/2}^n.It is also similar to the PQM method proposed for climate studies.  The PQM method is implemented much like PPM in that the edge variables are approximated in terms of the cell-centered degrees of freedom.  Here we will describe these variables as independent degrees of freedom.  The polynomial reconstruction defined by these conditions is

P(\xi) = u_j^n + \frac{1}{4}(6 u_{j+1/2}^n-6 u_{j-1/2}^n -s_{j-1/2}^n - s_{j+1/2}^n)\xi

+ \frac{1}{4} (3 s_{j-1/2}^{n} - 3 s_{j+1/2}^{n} +30_{j-1/2}^{n}-60 u_j^n +30u_{j+1/2}^{n}) (\xi^2-\frac{1}{12})

+ (s_{j+1/2}^n+s_{j-1/2}^n - 2u_{j+1/2}^n+2u_{j-1/2}^n)\xi^3

+\frac{1}{2} (-30 u_{j-1/2}^n - 30u_{j+1/2}^n +5 s_{j-1/2}^n+60 u_{j}^n +5 s_{j+1/2}^n)(\xi^4-\frac{1}{80}).

This produces a semi-discrete scheme with fifth-order accuracy and a dispersion relation of \imath \theta + \frac{1}{7200} \theta^6 + \frac{1}{42000}\imath\theta^7+ O(\theta^8) .  This is exceedingly accurate with the dispersion relation peaking at $latex\theta=\pi$ with an error of only 1.2%.

We will close with a couple of schemes that are more speculative, but use the degrees of freedom differently than before.  Both will be based on a cubic polynomial reconstruction and involve the first moment of the solution along with either edge-values or edge-slopes.  Aside from the use of the first moment it is similar to the schemes above.  The reconstruction for the moment, cell-average and cell-edge values is

P(\xi) = u_j^n + \frac{3}{2}(20 m_{1,j}^n- u_{j+1/2}^n- u_{j-1/2}^n)\xi

\frac{1}{2}(u_{j-1}^n+ u_{j+1}^n - 2 u_j^n)(\xi^2 -\frac{1}{12}) .

+ (10 u_{j+1/2}^n-10 u_{j-1/2}^n - 120 m_{1,j}^n)\xi^3 .

As previous experience would indicate the use of the moment has remarkable properties, with this method producing the same error as the quartic reconstruction based on edge-values and slopes plus the cell-average, but one order lower reconstruction.

I will close by presenting three schemes based on cubic reconstructions, but utilizing the first moments.  Two will achieve fifth-order accuracy, and the last will suffer from cancellation effects that lower the accuracy to fourth-order.  These methods are unlike the discontinuous Galerkin methods in that their support will go beyond a single cell.  The first of the three will use the three cell-centered variables, and the moment of the central cell , giving a polynomial,

P(\xi) = u_j^n+\frac{3}{44}(200 m_{1,j}^n - u_{j+1}^n + u_{j-1}^n)

3(u_{j-1/2}^n+ u_{j+1/2}^n - 2 u_j^n)(\xi^2 -\frac{1}{12})

+ \frac{5}{11} (u_{j+1}^n - u_{j-1}^n - 24 m_{1,j}^n) \xi^3 .

This produces a semi-discrete scheme with fifth-order accuracy and a dispersion relation of \imath \theta + \frac{11}{7200} \theta^6 + \frac{1157}{3024000}\imath\theta^7+ O(\theta^8) .  This is exceedingly accurate with the dispersion relation peaking at $latex\theta=\pi$ with an error of only 3.1%.

The next method uses the cell-center values of the cell and it upwind neighbor and the moments in each of those cells.  The reconstructing polynomial is

P(\xi) = u_j^n+\frac{1}{8}(78 m_{1,j}^n -18 m_{1,j-1}^n -+3 u_{j}^n -3 u_{j-1}^n)

\frac{1}{4}(66 m_{1,j-1}^n+114 m_{1,j}^n -15 u_j^n+15 u_{j-1}^n)(\xi^2 -\frac{1}{12})

+ \frac{1}{2} (5 u_{j-1}^n - 5 u_{j}^n + 15 m_{1,j}^n +15 m_{1,j-1}^n) \xi^3 .

This produces a semi-discrete scheme with fifth-order accuracy and a dispersion relation of \imath \theta - \frac{1}{1350} \theta^6 + \frac{1}{23625}\imath\theta^7+ O(\theta^8) .  This is  accurate with the dispersion relation peaking at $latex\theta=\pi$ with an error about 12%.

Our final method uses the cell-centered value of the cell and its downwind neighbor along with the first moment in each of those cells.  The reconstruction takes the form,

P(\xi) = u_j^n+\frac{1}{8}(78 m_{1,j}^n -18 m_{1,j+1}^n -+3 u_{j+1}^n -3 u_{j}^n)

\frac{1}{4}(-66 m_{1,j+1}^n+114 m_{1,j}^n -15 u_j^n+15 u_{j+1}^n)(\xi^2 -\frac{1}{12})

+ \frac{1}{2} (5 u_{j}^n - 5 u_{j+1}^n + 15 m_{1,j}^n +15 m_{1,j+1}^n) \xi^3 .

We note that this stencil produces cancellation that lowers its accuracy by one order.  This produces a semi-discrete scheme with fifth-order accuracy and a dispersion relation of \imath \theta - \frac{1}{1080} \imath\theta^5 + \frac{1}{54432}\imath\theta^7+ O(\theta^8) .  This is  accurate with the dispersion relation peaking at $latex\theta=\pi$ with an error of 10%. The last two methods may not be particularly interesting in and of themselves, but may be useful as alternative stencils for ENO-type methods.

Van Leer’s 1977 paper – Paper IV in the Quest for the Ultimate

06 Monday Jan 2014

Posted by Bill Rider in Uncategorized

≈ 1 Comment

Some of these options were profitably framed a long time ago in one of Van Leer’s papers (the fourth of his series published in 1977).  In that paper he described two methods that came into widespread use, the second-order accurate method that is lives on as the origin of much of the limiter work in the 1980’s, and is the most common modern approach used in aerodynamic codes, unstructured codes, and for remap in ALE codes.  He also introduced the base method that became the PPM method.  Ironically, these two methods were the worst of the six methods introduced.  Van Leer labeled them as schemes 1 and 4.    Van Leer’s scheme 1 is called the “piecewise linear method” and the scheme 4 is the “piecewise parabolic method (PPM).”  The polynomial approximation used in scheme 1 uses the integral average of cell value, and the derived gradient (or slope) to define the linear approximation.  Other linear approximations could be developed using biased approximation for the slope, but these methods would not be considered upstream centered.  The PPM approximation is built defining a parabola that reconstructs the integral of a cell and its two neighbors.  The more famous PPM method uses a polynomial matching the cell average, and high-order approximations to the values at the edge of the cell. 

Two other classes of methods were provided.  One of these is based on the moments of a solution (the zeroth moment is the cell average value).  These methods are equivalent to the discontinuous Galerkin (DG) method, which is the focus of copious research today, and one of the key approaches being explored to deal with the upcoming changes in computing architecture.    This is enabled by the scheme being developed to carry extra degrees of freedom in each computational cell (i.e., the moments, a single moment gives a linear basis, a second moment a quadratic basis).  Van Leer labeled them as schemes 3 and 6.  Scheme 6 was not described in details, but has been developed as the PPB (piecewise parabolic Boltzmann) scheme by Paul Woodward, and as a quadratic finite element method with the DG research.  The polynomial derivation will match the cell average and the first and/or second moment of the distribution.

It is notable that the bane of DG methods is the development of nonlinear limiters that do not destroy the increases in accuracy fidelity that the basic (linear) DG methods provide.  Thus far it is the development of effective limiters for DG that has most greatly “limited” their adoption by the community more broadly.  This is coupled to the next issue related to stability.

The methods in Van Leer’s paper used characteristic information to enhance their stability resulting in a Courant condition that was not restrictive.  Most DG methods developed today rely upon method-of-lines approaches, which means the spatial part of the operator is discretized, and the time operator is done using an ODE method.  This approach when using explicit methods results in a very restrictive Courant condition.  A challenge is to produce methods of this sort will greater stability.  Selectively reintroducing the characteristic information may be an answer.

The third class of approaches in Van Leer’s paper is significantly under-developed and deserves attention.  Moreover, I believe it can be profitably being extended to encompass even broader design principles.  As briefly commented on above these methods utilize point values of the quantity being integrated and derivative values.  In order to capture shocks, the integral of the quantity (i.e., the cell-centered value) is also integrated using the familiar control volume form.  Thus these methods can be described as combinations of finite volume and finite difference methods.    Van Leer’s Scheme 2 uses the integral average and a cell-centered derivative to define a linear polynomial.  The difference between this method and Scheme 1 is that Scheme 2’s derivative is evolved in time using the derivative of the PDE.  Scheme 5 is different.  The parabola is defined by the integral average and the edge values.  The difference between this method and the classic PPM method is that the edge values are evolved in time by the continuous version of the PDE.  By continuous version of the PDE I mean the strong differentiable form of the PDE and not the weak form used to integrate the integral average values or moments.

There is no reason, that this approach can’t be taken even further to include moments as well.  This provides the capability to design methods with tailored performance, both in terms of accuracy and memory use.  The key is to provide the capability to introduce appropriate nonlinear stability mechanisms of equivalent impact to those common with Van Leer’s piecewise linear or parabolic methods.  For example, a cubic polynomial could be derived using the integral average, the first moment and the cell-edge values.  A suggestion made by Phil Roe is to use the integral average with the cell-edge and cell-edge-derivative values to define a quartic polynomial and thus a fifth-order method.

Stability and accuracy analysis for these schemes becomes necessarily complex.  Using tools such as Mathematica renders the problem tractable.  I will note that Van Leer analyzed the method without such help, and this makes the analysis contained in his paper all the more masterful.

 

The paper’s abstract and introduction

06 Monday Jan 2014

Posted by Bill Rider in Uncategorized

≈ Leave a comment

A Basis for Reconsidering Remap Methods

William J. Rider

Sandia National Laboratories, Albuquerque 

Abstract

Methods for discretizing remap are based on algorithms developed for hyperbolic conservation laws.  Since its introduction in 1977 Van Leer’s monotonicity-preserving piecewise linear method and its extensions have been ubiquitous in remap. In that 1977 paper Van Leer introduced another five algorithms, which largely have not been used for remap despite the observation that the piecewise linear method had the least favorable properties. This adoption parallels the algorithmic choices in other related fields.  Two factors have led to the lack of attraction to the five algorithms: the simplicity and effectiveness of the piecewise linear method, and complications in practical implementation of the other methods. Plainly, stated Van Leer’s piecewise linear method enabled ALE methods to move forward by providing a high resolution, monotonicity-preserving remap.  As a cell-centered scheme, the extension to remap was straightforward.  Several factors may be conspiring to reconsider these methods anew: computing architectures are more favorable towards floating point intensive methods, lacking data movement, and 30 years of experience in devising nonlinear stability mechanisms (e.g., limiters).  In particular, one of the method blends characteristics of finite volume and finite difference methods together in an ingenious manner that has exceptional numerical properties, and should be considered a viable alternative to the ubiquitous piecewise linear method.

 Introduction

In the numerical solution of hyperbolic PDEs, a series of papers by Bram Van Leer is considered seminal.  It was a five paper series with the august title “Towards the Ultimate Differencing Scheme…” In the fourth paper of the series a whole set of different methods were proposed (six in total).  Each of the methods is based on piecewise polynomial interpolation using the degrees of freedom to define the polynomials.   A first order polynomial leads to a second-order method, and a second-order polynomial leads to a third order method because of the integration involved in the solution.  Furthermore, each method developed was fundamentally upwind biased.  The broad class of methods can be described as being upstream centered, which is a centered basis for the polynomial with an upwind bias for information propagation.  This class of methods is well known to be advantageous mathematically.

Subsequently most of the development of numerical methods was based upon two of these methods.  The starkest aspect of this thread is that the two base methods were actually the least attractive from their fundamental numerical properties (accuracy, dissipation and propagation or phase error).  The reason for this is related to the more elaborate techniques necessary for nonlinear stability of solutions and practical use.  These techniques go by the name of “limiters” that serve to keep solutions physical and properly bounded. 

The relatively unexplored methods are interesting for several reasons.  Some of the basic design principles are quite different.  The favored methods are all based on integrating over control volumes, implying the use of conservation form (as essentially dictated to the community by Lax-Wendroff’s famous theorem).  A couple of the methods use other integrated quantities such as moments of a solution where the first moment is the integral of the quantity multiplied by a linear function, the second moment, the integral of the quantity multiplied by a quadratic function and so on.  The evolution equations for the moments is derived using the same procedure, and naturally invokes the weak form of the PDE.  The other form uses the differential form of the PDEs and involves introducing point quantities, such as the solution at a point, or derivatives of the solution at a point.  One of the key aspects of this development is to make sure the quantities are all dependent upon each other so that degrees-of-freedom are well coupled.

The present day interest should be driven by the needs of modern computers that are already far different than the previous generations, with massive and pervasive levels of parallelism.  New computers will greatly penalize memory references and communication compared to floating point operations.  In other words, once you’ve got something in your computer memory, you will be greatly rewarded for doing lots of floating point operations.  This character will offer benefits to methods having their data local and high-order accuracy requiring many operations.  The key for all these methods is to develop robust and sensible stability enhancing techniques that do no undermine the fundamental properties of the methods they are applied to. 

As note, by far, the greatest amount of effort for solving hyperbolic PDEs* has gone into cell-centered schemes due to their intrinsic simplicity and connection to existing computational approaches.  This work has been enormously successful over the past thirty some odd years.  Starting in the early 1970’s key concepts were added to the repertoire of scientist that allowed one to simultaneously achieve the physically appealing results provided by first-order method such as upwind differencing with the accuracy of second-order methods such as Lax-Wendroff.  These techniques were probably first introduced in the form of flux-corrected transport (FCT) by Jay Boris (and later in collaboration with David Book), but also independently by Bram Van Leer in his “Toward the Ultimate” series of papers.  There is also the compelling case of a third invention of these methods by Kolgan in the Soviet Union.  Kolgan died before receiving any attention and his invention was only recently uncovered.  A fourth stream of development looks similar in the passage of time is due to the Israeli mathematician, Ami Harten, who figures prominently in our tale later.

These methods are known as monotonicity preserving where the numerical solution of the PDE does not produce any new minimum or maximum values.  This is a property of the simplest form of the equation used extensively in the development of numerical methods.   Upwind differencing is the archetypical monotonicity-preserving linear scheme, and Lax-Friedrichs is another example.  Monotone schemes are first-order accurate.  The work of Boris, Van Leer and Kolgan all sought to produce monotonicity preserving methods that were not limited to first-order accuracy.

All three developments were made with knowledge of Godunov’s theorem that disallows the possibility of developing a linear method that at once second-order accurate and monotone.   The key to the theorem is the word linear.  The key advancement made was to introduce nonlinearity into the method even in the case of solving a linear PDE.  What is a linear scheme?  A linear method uses the same finite difference stencil (or formula) everywhere, while a nonlinear scheme adapts the finite difference scheme to the solution itself.  In other words, the scheme detects the nature of the solution and adapts the finite difference scheme to optimally solve it. 

At a detail level, the developments were much different, but the conceptual framing is nearly identical.  FCT uses a sequential process where the limiter adds back as little of the first-order monotone dissipation as possible to still produce monotone results.  Both Van Leer and Kolgan developed second-order extensions of Godunov’s method**.

In the field of hyperbolic PDEs nonlinear scheme changed everything?  In many fields of computational science this development was like an event horizon.  These nonlinear differencing schemes were an enabling technology for many numerical simulations, and resulted in an order of magnitude or more increase in the effective accuracy of simulations.

From this foundation there was a flurry of research activity in the 1980’s and 1990’s developing, extending and using this broad class of methods.  Examples abound in the literature and the success of these approaches in undeniable.  Several of the ore successful methods from the line of work, are the piecewise parabolic method (PPM), and the weighted essentially non-oscillatory (WENO) methods, which dominate modern use of these methods.  Both of these approaches have the appeal of utilizing elements of method design that are higher order accurate and produce results of significantly higher fidelity than the original nonlinear methods.  For example, PPM dominates the field of astrophysical simulation and WENO dominates mathematical research and related simulations.  Both methods are based on developing a numerical solution for variables all defined at the center of control volumes.

A clear downside to this approach is the need to communicate the information of many control volumes to update the values for any other control volume.  This communication is becoming increasingly problematic on modern computing platforms because of the cost of communication relative to floating-point operations, or memory access.  In order to get the most out of modern computers both communication of information, and memory access must be carefully managed.  Relatively speaking, floating point operations are “free,” which more accurately is that they are not the rate limiting aspect of computing.   Therefore we can stand to have algorithms that are very floating point intensive as long as they have relatively low communication intensity, and memory requirements.  The key is striking the right sort of balance.  Because of this issue it is time to rethink the design of algorithms in a major way.  Recent results at a conference stunned me by showing (for the first time I can remember), new computing architectures actually providing worse performance.  The new chips were better for many things, but not scientific computing.  It has the potential to be a true “wake-up call.”

Some of Van Leer’s methods have received attention and extension, most notably the PPM method developed by Colella and Woodward.  This method is extensively used in astrophysics, but not significantly in remap methods although a very close variant to Van Leer’s approach (scheme IV) as been adopted by ALEGRA.    The PPM method can be thought of as a reasonable cell-centered approach to scheme V where the edge-centered variables are replaced with temporary high-order variables. Perhaps the most extensive development of this method was undertaken by Eymann in conjuction with Roe.  This extension provided a number of advances that may prove useful in the setting of remap.  More direct implementations of scheme V have been followed by Guinot, and Popov (with Norman).  These approaches highlight the positive aspects of the method as we will reiterate below, but thus far have not achieved widespread adoption.  Zeng’s hybrid finite-difference-finite volume method is yet another closely related scheme that may include extensions of the approach important for practical adoption.

Discontinuous Galerkin methods have received the most attention although the direct descendents of Van Leer’s approach are far more limited with the line of work arising from Cockburn and Shu being widest in adoption.  Woodward has developed Van Leer’s sixth method in detail despite its sparseness of detail in the 1977 paper.

*Hyperbolic PDE methods can also be described as solving an advection equation, convection or remapping values from one grid to another.  Where I work, remap is the most common term.  Remap can also be done purely geometrically, and the advection is applied using a mathematical transformation to achieve equivalence.

**Godunov’s work was done largely in isolation to the progress made in the West.  He produced the theorem mentioned above, which is fundamental, and a unique numerical method for solving systems of hyperbolic PDEs.  This method is a fairly pure extension of first-order upwinding to systems.  Godunov used the Riemann problem to provide upwinding to the system, a truly inspired choice.  The Riemann problem is the analytical solution to the pure initial value problem of two semi-infinite states interacting.   In this solution we must note that the number of independent modes of transport, or waves is equal to the number of equations in a system.  The magnitude of the wave’s speed can vary significantly and differ in direction.   This muddies the meaning of upwind in the classical sense.  The Riemann solution is used to sort this out in a physically appealing way.

Review of the analysis of Van Leer’s Six Schemes

05 Sunday Jan 2014

Posted by Bill Rider in Uncategorized

≈ 1 Comment

Van Leer’s analysis of these methods is a tour-de-force especially considering the sort of resources available to him at that time.  He did not have use of tools such as Mathematica upon which I rely.  Nonetheless, the modern tools do allow more flexibility in analysis, and exploration of extensions of the methods without undue effort.  This is not to say that the analysis is particularly easy as Mathematica analysis goes, it does require considerable skill, experience and attention.  My hope for this rearticulation of the results is for clarity with a substantial appreciation for the skill expressed in the original exposition.

Here we briefly review the six schemes presented by Van Leer in that 1977 paper in terms of their basic structure, and numerical properties.  We will utilize Von Neumann analysis for both the fully discrete and semi-discrete schemes.  In Von Neumann analysis the variable on the mesh is replaced by the Fourier transform,

u_j = \exp(\imath j \theta) or u_j^n = A^n\exp(\imath j \theta)

with \theta being the angle and A being the amplification factor.

The fully discrete analysis was the vehicle of presention for Van Leer, and the semi-discrete analysis will be additional detail.   The fully discrete schemes all fufill the characteristic condition, they recover the exact solution as the CFL number goes to one, which also demarks its’ stability limit.  The semi-discrete analysis only provides the character of the spatial scheme.

Scheme I. The Piecewise Linear Method.

This method uses a piecewise linear reconstruction

P(\xi) = P_0 + P_1 \xi with \xi = (x-x_j)/\Delta x, P_0 = u_j^n with slopes computed from local data P_1 = \frac{1}{2} (u_{j+1} - u_{j-1}).

The full update is

u^{n+1}_{j} = u^{n}_{j} - \nu (u^{n+1/2}_{j+1/2} - u^{n+1/2}_{j-1/2})

with u^{n+1/2}_{j+1/2} = P(1/2-1/2\nu) and \nu=1 \Delta t/\Delta x is the CFL number.

This method is second-order accurate as verified by expanding the method function in a Taylor series.  This can also be verified by expanding the Fourier transformed mesh function in a Taylor series expansion around \theta = 0.

amp1

Scheme I amplification surface plot, shows large error at the mesh scale

The amplification factor is A =1-\frac{1}{8} \left(\nu-2 \nu^2+2\nu^3-\nu^4 \right)\theta^4+O\left(\theta^6\right)

The phase error is \phi = 1+\frac{1}{12}(1-2\nu+3\nu^2)\theta^2 + O(\theta^4)  this is the leading order error (being one order higher due to how the phase error is written).  A key aspect of this scheme has zero phase error at CFL numbers of one-half and one.

phase1

As with the amplification, scheme I has large errors at the mesh scale.

The semi-discrete form is u_{j+1/2} - u_{j-1/2} = \imath \theta + \frac{1}{12}\imath \theta^3 + \frac{1}{8}\theta^4+ O(\theta^5), which is useful if the scheme is used with a method-of-lines time integration.

imag-S1

Dispersion error for Scheme I is large past pi/2

Scheme II. The Piecewise Linear Method with Evolving Slopes.

This method evolves both the variable and its first derivative and requires the analysis of a system of variables.  The scalar analysis of stability is replaced by the analysis of a matrix determined by the behavior of its eigenvalues.  One eigenvalue will define the behavior of the scheme, while the second eigenvalue has be referred to as spurious.  More recently it has been observed that the “spurious’’ eigenvalue describes the subcell behavior at the mesh scale.

The method uses the same reconstruction as Scheme I, but evolves both P_0 = u_j^n and P_1 = s_j^n  in time.  The update is then two equations, one for the cell-centered data,

u^{n+1}_{j} = u^{n}_{j} - \nu (u^{n+1/2}_{j+1/2} - u^{n+1/2}_{j-1/2}),

which is identical to Scheme I and a second for the slopes

s^{n+1}_j = u_j^n+\frac{1}{2} s_j^n-u_{j-1}^n - \frac{1}{2} s_{j-1}^n - \nu(s_j^n - s_{j-1}^n) ,

where the usual s_j^n as the initial slope is replaced by the first expression on the right hand side of the equation driving a coupling of the variable and the slope in the update.

The slope update is unusual and causes significant error as the CFL number goes to zero.  This method does not have an obvious method of lines form due to the nature of this update.

The amplification factor is A =1-\frac{1}{8} \left(\nu^4-2 \nu^3+2 \nu^2\right)\theta^4+O\left(\theta^6\right).

amp2

The phase error is \phi = 1+\frac{1}{12}(1-3\nu+2\nu^2)\theta^2 + O(\theta^4)  and is the leading order error as with all of the first three methods. Again, this method does not lend itself to a well-defined semi-discrete method.

phase2

Scheme III. The Piecewise Linear Method with Evolving First Moment.

This method uses a reconstruction that is similar to piecewise linear, but instead of slope (first derivative), the linear term is proportional to the moment of the solution,

P(\xi) = P_0 + P_1 \xi

again P_0=u_j^n, but P_1 = m_1 = \int_{-1/2}^{1/2} P(\xi)\xi d\xi.  Here, as with Scheme II, we evolve both P_0, and P_1, but the evolution equation for P_1 is much different, and naturally yields a coupled system.  This also provides a direct path to writing the scheme in a semi-discrete form.

The update for P_0 is the same as the previous two schemes,

u^{n+1}_{j} = u^{n}_{j} - \nu (u^{n+1/2}_{j+1/2} - u^{n+1/2}_{j-1/2}).

The update for P_1 is based on the weak form of the PDE and uses the definition of the first moment,

\int_{-1/2}^{1/2} \xi(\partial_t u + \partial_x u)d\xi =0, \partial_t m - \int_{-1/2}^{1/2} \partial_x \xi u d\xi + [u(1/2)+u(-1/2)] = 0.

Discretely this equation can be evaluated for the scalar wave equation to

m_j^{n+1} = m_j^n - \frac{1}{2}\nu (u^{n+1/2}_{j-1/2}+u^{n+1/2}_{j+1/2}) + \frac{1}{6}(u_j^n + 4 u_j^{n+1/2} + u_j^{n+1}), where u^{n+1/2}_{j} = u^{n}_{j} - \frac{1}{2}\nu (u^{n+1/2}_{j+1/2} - u^{n+1/2}_{j-1/2}).

The amplification factor is A =1-\frac{1}{72} \left(\nu-2 \nu^2+2 \nu^3-\nu^4\right)\theta^4+O\left(\theta^6\right).

amp3

The phase error is \phi = 1+\frac{1}{540}(2-5\nu+5\nu^3-2\nu^4)\theta^4 + O(\theta^4)  now the amplification error is leading and the method is confirmed as being third-order. This scheme is obviously more accurate than the previous two methods.

phase3

The semi-discrete form follows from the standard updates for both the variable and its moment without any problem.  The update has an accuracy of second order with small coefficients, u_{j+1/2} - u_{j-1/2} = \imath \theta + \frac{1}{72} \theta^4 + \frac{1}{540}\imath \theta^5+ O(\theta^6), which the keen observer will notice being third-order accurate instead of the expected second order.  Van Leer made a significant note of this.

imag-S3real-S3

Scheme IV. The Piecewise Parabolic Method.

This method uses a piecewise parabolic reconstruction

P(\xi) = P_0 + P_1 \xi + P_2(\xi^2 - \frac{1}{12}),

and uses the variables and its two nearest neighbors to determine the fit.  The fit is defined as recovering the integral average in all three cells.  The coefficients are then simply defined by P_0=u_j^n, P_1=\frac{1}{2}(u_{j+1}^n - u_{j-1}^n), and P_2=\frac{1}{2}(u_{j+1}^n - 2u_j^n +u_{j-1}^n) . This is a third-order method.

The update is simple based on the evaluation of edge and time-centered values that need to be considered slightly differently for third-order methods,

\bar{u}_{j+1/2} = \int_{j+1/2}^{j+1/2-\nu} P(\xi) d\xi = P_0 + P_1(\frac{1}{2} - \frac{\nu}{2}) + P_2 (\frac{1}{2} - \frac{\nu}{2} + \frac{\nu^2}{3}).  The update then is simply the standard conservative variable form,

u_j^{n+1} = u_j^n -\nu(\bar{u}_{j+1/2} - \bar{u}_{j-1/2}) .

The amplification factor is A =1-\frac{1}{24} \left(\nu^2-2 \nu^3+\nu^4 \right)\theta^4+O\left(\theta^6\right)

amp4

The phase error is \phi = 1-\frac{1}{60}(2-5\nu+5\nu^3-2\nu^4)\theta^4 + O(\theta^6)  this is the leading order error (being one order higher due to how the phase error is written).  A key aspect of this scheme has zero phase error at CFL numbers of one-half and one.

phase4

The semi-discrete form is u_{j+1/2} - u_{j-1/2} = \imath \theta + \frac{1}{12}\theta^4 - \frac{1}{30}\imath\theta^5 + O(\theta^6), which is useful if the scheme is used with a method-of-lines time integration.

imag-S4real-S4

Scheme V. The Piecewise Parabolic Edge-Based Method.

This method uses a piecewise parabolic reconstruction

P(\xi) = P_0 + P_1 \xi + P_2(\xi^2 - \frac{1}{12}),

and uses the variables and its two edge values determining the fit.  The fit recovers the integral of the cell j and the value at the edges j\pm1/2.  The coefficients are then simply defined by P_0=u_j^n, P_1 =u_{j+1/2}^n - u_{j-1/2}^n , and P_2 = 3(u_{j+1/2}^n-2u_j^n+u_{j-1/2}^n) . This is a third-order method.

This method is distinctly different than Scheme IV in that both the cell-centered, conserved quantities, and the edge-centered (non-conserved) quantities.  The form of the updates are distinctly different and derived accordingly.  The time-integrated edge value follows the polynomial form from Scheme IV,

\bar{u}_{j+1/2} = \int_{j+1/2}^{j+1/2-\nu} P(\xi) d\xi = P_0 + P_1(\frac{1}{2} - \frac{\nu}{2}) + P_2 (\frac{1}{2} - \frac{\nu}{2} + \frac{\nu^2}{3}) = u_{j+1/2}^n -\frac{\nu}{2} P_2 -(\frac{\nu}{2} - \frac{\nu^2}{3})P_3,

giving a cell centered update, u_j^{n+1} = u_j^n -\nu(\bar{u}_{j+1/2} - \bar{u}_{j-1/2}) .  The edge update is defined by a time-integrated, upstream-biased first derivative evaluated at the cell-edge position,

u_{j+1/2}^{n+1}= u_{j+1/2}^n - \nu P_2-\nu P_3+\nu^2 P_3 .

The amplification factor is A =1-\frac{1}{72} \left(\nu-2 \nu^2+2 \nu^3-\nu^4\right)\theta^4+O\left(\theta^6\right).

amp5

The phase error is \phi = 1+\frac{1}{540}(2-5\nu+5\nu^3-2\nu^4)\theta^4 + O(\theta^4)  now the amplification error is leading and the method is confirmed as being third-order. The scheme is very close in performance and formal error form to Scheme III, but is arguably simpler.

phase5

The semi-discrete form follows from the standard updates for both the variable and its moment without any problem.  The update has an accuracy of third order with small coefficients, u_{j+1/2} - u_{j-1/2} = \imath \theta + \frac{1}{72} \theta^4 + \frac{1}{270}\imath \theta^5+ O(\theta^6) .

imag-S5real-S5

Looking at the recent PPML (piecewise parabolic method – local) scheme of Popov and Norman produces an extremely interesting result.  The interpolation used by PPML is identical to Scheme V, but there is a key difference in the update of the edge values.  The differential form of evolution is not used and instead a semi-Lagrangian type of backwards characteristic interpolation is utilized,

u^{n+1}_{j+1/2} = P_2(1/2 -\nu) .

This produces exactly the same scheme as the original Scheme V for the scalar equation.  In many respects the semi-Lagrangian update is simpler and potentially more extensible to complex systems. Indeed this may be actualized by the PPML authors who have applied the method to integrating the equations of MHD.

Scheme VI. The Piecewise Parabolic Method with Evolving First and Second Moments.

This method uses a piecewise parabolic reconstruction

P(\xi) = P_0 + P_1 \xi + P_2(\xi^2 - \frac{1}{12}),

and uses the variables and its two edge values determining the fit.  The fit recovers the integral of the cell j and the value at the edges j\pm1/2.  The coefficients are then simply defined by P_0=u_j^n , P_1 =12 m_{1,j}^n , and P_2= 180 m_{2,j}^n - 15 u_j^n . m_1 is the first moment of the solution, and m_2 is the second moment.  The moments are computed by integrating the solution multiplied by the coordinate over the mesh cell.  This is a fifth-order method.

For all the power Mathematica provides, this scheme’s detailed analysis has remained elusive, so the results are only presented for the semi-discrete case where Mathematica remained up to the task.  Indeed I did get results for the fully discrete scheme, but the plot of the amplification error was suspicious and for that reason will not be shown, nor will the detailed truncation error associated with the Taylor series.

The semi-discrete form follows from the standard updates for both the variable and its moment without any problem.  The update has an accuracy of second order with small coefficients, u_{j+1/2} - u_{j-1/2} = \imath \theta + \frac{1}{720} \theta^6 + \frac{1}{4200}\imath \theta^7+ O(\theta^6), which the keen observer will notice being fifth-order accurate instead of the expected third order.  The semi-discrete update for the first moment is \frac{1}{2}(u_{j+1/2} + u_{j-1/2}) - u_j, and the second moment is

1/4(u_{j+1/2} - u_{j-1/2}) - 2 m_{1,j}.

imag-S6real-S6

Aside

Trying something new while writing a paper

04 Saturday Jan 2014

Posted by Bill Rider in Uncategorized

≈ Leave a comment

Tags

http://postto.wordpress.com/?abpost&bid=9619154

I’m writing a paper right now that reviews some suggestions for solving the ubiquitous scalar advection equation,

\partial_t u + \partial_x u = 0,

which is the basis for remap in arbitrary Lagrangian-Eulerian codes and more generally hyperbolic PDEs.  I will post different sections from the paper week to week as I work toward the submission of the manuscript in a month.  A lot of the paper is already written, so the posts will show the progression of polished text.

The standard method for solving scalar advection has been Van Leer’s piecewise linear method (with an appropriate limiter), and its generalizations.   The method is based on polynomial interpolation normalized within a cell

\xi =\frac{(x-x_j)}{\Delta x};P(\theta)=P_0 + P_1 \xi, where P_0=u_j; P_1 = \frac{\partial u}{\partial x}\Delta x.

Van Leer’s method is really only one of the many schemes he suggested in his 1977 paper with the other methods being largely ignored (with the exception of the PPM method).

My paper will review Van Leer’s schemes and the analysis from the 1977 paper and then suggest how these concepts can be extended.  Some of the methods have direct connections to discontinuous Galerkin, but others are unique.  Some of these methods are completely different than anything else out there and worth a lot of attention.  These methods have the character of both finite volume and finite difference method plus have compactness that might be quite advantageous on modern computers.  Their numerical error characteristics are superior to the standard methods.

Stay tuned.

Why do people prefer worse methods?

31 Tuesday Dec 2013

Posted by Bill Rider in Uncategorized

≈ Leave a comment

Here is the answer in a nutshell; people use an inappropriate approach and test problems to characterize their methods.

The reason I started doing work in verification was directly related to my work in developing better numerical methods.  In the early 90’s I was developing an improved implementation of a volume tracking method (or VOF where an interface is tracked through volumes, and the interface is reconstructed geometrically).   The standard implementation of the method resulted in horrific spaghetti code, and made subsequent development and debugging a nightmare.  I had developed a better, more object-oriented implementation and was eager to share it.  In the course of doing this implementation work, I engaged in extensive testing to put the method through its paces.  I felt that the standard tests were insufficiently taxing and provided poor code coverage, so I came up with several alternative test problems.  Existing tests involved translating objects on a grid, or engaging in solid body rotation.  Difficulty was defined by the complexity of the shape being moved.  My new problems used time dependent flowfields that had non-zero vorticity and could be reversed in time allowing exact error assessment.  These new problems were a hit and have largely replaced the earlier (too) simple tests.

Now fifteen years on the paper is very highly cited, but mostly because of the better tests, not because of the reason the paper was actually written.  I’d say about 75-90% of the citations are primarily for the tests, not the methodology.  There is a lesson in this that is positive; people want to test their methods rigorously.  For myself, I was a verification devotee thereafter largely because of the experience.

In terms of numerical methods and physics my greatest interest is the solution of hyperbolic PDEs.  I find the mathematics and physics of these equations both engaging on multiple levels, and the numerical methods to solve these equations spell-bindingly interesting.  It is also an area of numerical methods where verification is not the standard practice; at least for the application of the equations to full nonlinear PDEs that characterize the application of the method.  This isn’t entirely fair; for simple linear equations with smooth solutions or nonlinear equations with discontinuities verification is commonplace and expected.  Under these limited conditions authors regularly verify their methods and present results in virtually every paper.  When a discontinuity forms accuracy in the sense of high-order convergence is lost; solutions are limited to first-order convergence or lower.  Under these conditions verification is not done as a matter of course, in fact, it is exceedingly rare to the point of almost being unheard of.

This is extremely unfortunate.  Verification is not only about order of accuracy; it is also about estimating the magnitude of numerical error.

The standard approach when smoothness is not present is to plot results at some resolution against an exact solution and show these results graphically.  The main reason to show the results is to demonstrate a lack of oscillatory wiggles near the discontinuity.  The error itself is easily (almost trivial) computable, but almost never presented. The basis of this reasoning is that all results are basically first-order accurate, and the error isn’t important.  Implicit in this judgment is the belief that order of accuracy matter, and by virtue of this numerical error magnitude is unimportant.   Unfortunately, this point of view is not entirely correct, and misses a key aspect of method character.

When all solutions are first-order accurate there are actually significant differences in the level of numerical error.  Under mesh refinement these differences can become quite significant in terms of the quality of solution.   Consider the following example with regard to the quality of solution, for first-order a factor of two increases in mesh resolution results in a halving of numerical error.  If instead you changed the numerical method to halve the error on the original grid, the savings in compute time could be large.  A halving of the mesh results in the computational work increasing by a factor of four in one dimension (assuming the Courant number is held constant).  In two dimensions the increase in effort is a factor of eight and a factor of sixteen in three dimensions.  If the more accurate numerical method does not require four or eight or sixteen times the effort, it will be a win.    More accurate methods are most computationally intensive, but rarely so much that they aren’t more efficient than lower order methods. The cases where these dynamics play out are the closest to actual applications where the problems are discontinuous and never well behaved mathematically. Indeed it is this character that explains the rapid adoption of the second-order MUSCL (Van Leer differencing for you weapons’  lab folks) over the first-order methods.

The differences are not geometric as they would be with high-order accuracy and smooth solutions, but quite frankly no one really gives a damn about smooth solutions where high-order accuracy is achievable.  They are literally only of academic interest.   Seriously.  So why, isn’t more rigorous error estimation on more difficult problems the status quo?  Fundamentally people simply haven’t come to terms with the inevitable loss of accuracy for real problems, and the consequences this has on how methods should be designed and evaluated.

Despite this, the community solving hyperbolic PDEs believes that verification is being routinely done.  It is.  It is being done on cases that matter little, and only reflect on the order of accuracy of the method in cases no one cares about.  For people doing applications work, it makes verification seem like a pitiful and useless activity.  My suggestions that verification is done for error characterization and estimation is not considered useful.  I looked back through three and a half decades of the literature at results for the venerable Sod’s shock tube problem (en.wikipedia.org/wiki/Sod_shock_tube), and could find no examples of accuracy being quoted with computed results.  Sod showed runtime in his 1978 paper.  This implies that the accuracy of the results is comparable and the computational cost at fixed resolution only matters.  This seems to be an article of faith by the community, and it is wrong!

The fact is that both speed and accuracy matter, and results should be presented thusly.

There are large differences in the accuracy between methods.  If one takes the measure of accuracy of computational effort required to achieve constant accuracy one can find a factor of 30 difference in the efficiency of different approaches (http://www.sciencedirect.com/science/article/pii/S0021999107000897).  This difference in efficiency is in one dimension with the potential for dramatically larger differences in two or three dimensions.  Yet this community almost systematically ignores this issue, and verification of methods for practical problems.

This is a long-winded way of saying that there are a lot of cases where people prefer to use methods that are worse in performance.  Part of this preference is driven by the acceptance of the sort of verification practice discussed above.  An example would be WENO (http://www.sciencedirect.com/science/article/pii/S0021999196901308) where elegance and the promise of high-order would seem to drive people, despite its low resolution.  WENO stands for weighted essentially non-oscillatory, which blends order of accuracy preservation with lack of Gibbs oscillations in an elegant almost algebraic algorithm.  Most of the development of WENO since its inception has revolved around accuracy, which is compounded by computing results for these linear problems.   The focus of order-of-accuracy results in the method being quite poor at resolving discontinuities with performance at those points of the solution being on par with the minmod second-order method (the worst and most dissipative of the second-order TVD methods!).  Instead of improving WENO for practical problems and focusing on its efficiency, the mathematical community has focused on its ability to achieve formally high-order accuracy in situations no one really cares about.

This might explain the lack of penetration of this class of method into practical computations in spite of thousands of citations. It is an utterly absurd state of affairs.

MUSCL (http://en.wikipedia.org/wiki/MUSCL_scheme) is an example of much more successful method, and as I showed with Jeff Greenough a well-written MUSCL code kills a WENO code in actual efficiency (http://www.sciencedirect.com/science/article/pii/S0021999103005965).  This was just for one dimension and the gains for MUSCL over WENO in multiple dimensions are as great as expected.  It isn’t a surprise that WENO has not been as successful for applications as MUSCL.  A change in focus by the community doing WENO development might serve the method well.

PPM (http://crd.lbl.gov/assets/pubs_presos/AMCS/ANAG/A141984.pdf) is much better than either WENO or MUSCL.  The gains in efficiency are real (about a factor of two over MUSCL in 1-D).  Part of this resolution is due to the difference between a linear and parabolic local representation of the solution, and the parabolic profile’s capacity to represent local extrema.  PPM can also exhibit bona fide high-order accuracy through the selection of high-order values for the edge values used in determining the parabola.  The bottom line is that the PPM method has some intrinsic flexibility that can be ruthlessly exploited.  PPM is widely used by astrophysicists for many applications.  Beyond astrophysics it has been less successful and perhaps that should be studied.

WENO is elegant.  It is a beautiful method, and I loved coding it up.  I would have to say that my WENO code is extremely attractive, just as the mathematics underpinning these methods are appealing.  This appealing veneer can be superficial and hides the true performance of the method.  It is sort of like a beautiful person with a horrible personality or weak intellect; the attractive package hides the less than attractive interior.  This isn’t to say that the method can’t be improved in terms of efficiency. Indeed through taking various lessons learned from MUSCL and PPM and applying them systematically to WENO, the method could be dramatically improved.  In doing this some of the properties of WENO for smooth hyperbolic PDEs are undermined, while the properties for discontinuous problems are improved.  It could be done quite easily.

Ultimately the community has to decide where it wants impact.  Are hyperbolic PDEs important for applications?

Yes, and completely beyond a shadow of doubt.

Why does the published literature act the way it does? How has the community evolved to where the actual performance on more application-relevant testing matters less than beauty and elegance?

I think this whole issue says a great deal about how the applied mathematics community has drifted away from relevance.  That is a topic for another day, but keenly associated with everything discussed above.

Thanks for listening, happy New Year.

What makes a computational simulation good?

26 Thursday Dec 2013

Posted by Bill Rider in Uncategorized

≈ 2 Comments

Tags

algorithms, calibration, efficiency, models, Moore's law, supercomputer, validation, verification

A better question is how do we improve computational simulations most effectively?  Should we focus more on creating better computations instead of faster ones?

On Christmas morning I unwrapped a new iPhone 5S and got rid of my clunky ole iPhone 4S.  Amazingly enough, the Linpac benchmark runs on iPhones with a cool little app (https://itunes.apple.com/us/app/linpack/id380883195 or https://itunes.apple.com/us/app/linpack-benchmark/id390185550?mt=8 or https://play.google.com/store/apps/details?id=com.greenecomputing.linpack&hl=en for the Android).  More amazing than that the clunky iPhone 4 clocks in at around 130 Mflops, which coincidently enough is just about the same as the Cray XMP I used as a brand new professional in 1989.  Moreover, the XMP I had access to in 1989 was one of the most powerful computers in the World.   Yet today I happily chose to recycle my little “XMP” without a second thought.  In 1989 that would have been unthinkable, just as unthinkable as holding that computational horsepower in the palm of my hand!  The new iPhone 5S is just shy of a Gigaflop, and it mostly plays music, and surfs the Internet rather than computing turbulent fluid flows.  What a World we live in!

One of the things that the XMP had was an awesome operating system called CTSS.  In some ways it was a horror show with a flat file system, but in other ways it was a wonder.  It could create something called a “drop file” that saved a complete state of a code that could be picked up by a debugger.  You could change values of variables, or figure out exactly why your calculation had a problem.  Of course this power could be misused, but you had the power.  Soon Cray replaced CTSS with Unicos their version of Unix, and we had a modern hierarchical file system, but lost the power of the drop file.  A lot of computational scientists would enjoy the power of drop files much more than a brand new supercomputer!

What’s the point of this digression?  We have focused on supercomputing as the primary vehicle for progress in computational science for the past 25 years, while not putting nearly so much emphasis on how computations are done.  Computing power comes without asking for it, and yesterday’s supercomputing power provides gaming today, and today’s supercomputing will power the games of tomorrow.  None of this has really changed what we do on supercomputers, and changing what we do on supercomputers has real scientific value and importance.   

The truth of the matter is that the most difficult problems in simulation will not be solved through faster computers alone.  In areas I know a great deal about this is true; direct numerical simulation of turbulence has not yielded understanding, and the challenge of climate modeling is more dependent upon modeling.  Those who claim that a finer mesh will provide clarity have been shown to be overly optimistic.  Some characterized stockpile stewardship as being underground nuclear testing in a “box,” but like the other examples depends on greater acuity in modeling, numerical methods and physical theory. Computational simulation is a holistic undertaking dependent upon all the tools available, not simply the computer.  Likewise, improvement in this endeavor is dependent on all the constituent tools.

Most of the money flowing into scientific computing is focused on making computations faster through providing faster computers.  In my opinion we should be more focused upon improving the calculations themselves.  Improving them includes improving algorithms, methods, efficiency, and models not to mention improved practice in conducting and analyzing computations.  The standard approach to improving computational capability is the development of faster computers.  In fact, developing the fastest computer in the world is a measure of economic and military superiority.  The US government has made the development of the fastest computers a research priority with the exascale program gobbling up resources.  Is this the best way to improve?  I’m fairly sure it isn’t and our over emphasis on speed is extremely suboptimal.

Moore’s law has provided a fifty year glide path for supercomputing to ride.  Supercomputers weathered the storm of the initial generation of commodity-based computing development, and continued to provide the exponential growth in computing power.  The next ten years represents a significant challenge the nature of supercomputing.  Computers are changing dramatically with the fundamental physical limits of current technology hitting limits.  To achieve higher performances levels of parallelism need to grow to unpredicted levels.  Moreover, existing challenges with computer memory, disc access and communication all introduce additional challenges.  The power consumed by computers also poses a difficulty.  All of these factors are conspiring to make the development of supercomputing in the next decade an enormous challenge, and by no means a sure thing. 

I am going to question the default approach.

The signs pointing to the wastefulness of this approach have been with us for a while.  During the last twenty years the actual performance for the bulk of computational simulations has been far below the improvements that Moore’s law would have you believe.  Computational power is measured by the Linpac benchmark, which papers over many of the problems in making “real” applications work on computers.  It solves a seemingly important problem of inverting a matrix using dense linear algebra.  The problem in a nutshell is that dense linear algebra is not terribly important, and makes the computers look a lot better than they actually are.  The actual performance as a proportion of the peak Linpac measured performance has been dropping for decades.  Many practical applications run at much less than 1% of the quoted peak speed.  Everything I mentioned above makes this worse, much worse. 

Part of the problem is that many of methods, and algorithms used on computers are not changing or adapting to reflect the optimality of the new hardware.  In a lot of cases we simply move old codes onto new computers.  The codes run faster, but nowhere as fast as the Linpac benchmark would lead us to believe.  The investment in computer hardware isn’t paying off to the degree that people advertise. 

Computational modeling is extremely important to modern science.  It reflects substantial new capability to the scientific community.  Modeling is a reflection of our understanding of a scientific field.  If we can model something, we tend to understand that thing much better.  Lack of modeling capability usually reflects a gap in our understanding.  Better put, computational modeling is important to the progress of science, and its status reflects the degree of understanding that exists in a given field.  That said, faster computers do not provide any greater understanding in and of themselves.  Period.  Faster, more capable computers allow more complex models to be used, and those more complex models may yield better predictions.  These complex models can be contemplated with better computers, but their development is not spurred by the availability of supercomputing power.     Complex models are the product of physical understanding and algorithmic guile allowing for their solution. 

I am going to suggest that there be a greater focus on the development of better models, algorithms and practice instead of vast resources focused on supercomputers.  The lack of focus on models, algorithms and practice is limiting the effectiveness of computing far more greatly than the power of the computers.  A large part of the issue is the overblown degree of improvement that new supercomputers provide, only a fraction of the reported power.  There is a great deal of potential headroom for greater performance with computers already available and plugged in.  If we can achieve greater efficiency, we can compute much faster without any focus at all on hardware.  Restructuring existing methods or developing new methods with greater accuracy and/or greater data locality and parallelism can gain efficiency.  Compilers are another way to improve code and great strides could be made there to the good of any code using computers.

One of the key areas where supercomputing is designed to make a big impact is direct numerical simulation (DNS), or first principles physical simulation.  These calculations have endless appetites for computing power, but limited utility in solving real problems.  Turbulence, for example, has generally eluded understanding and our knowledge seems to be growing slowly.  DNS is often at the heart of the use case for cutting edge computing.  Given its ability to provide results, the case for supercomputing is weakened.  Perhaps now we ought to focus more greatly on modeling and physical understanding instead of brute force.

Advances in algorithms are another fruitful path for improving results.  Algorithmic advances are systematically under-estimated in terms of their impact.  Several studies have demonstrated that algorithmic improvements have added as much or more to computational power than Moore’s law.  Numerical linear algebra is one area where the case is clear; optimization methods are another.  Numerical discretization approaches may be yet another.  Taken together the gains from algorithms may dwarf those from pure computing power.  Despite this, algorithmic research is conducted as a mere after thought, and more often than not is cut first from a computational science program. 

One of the key issues with algorithmic research is the “quantum” nature of the improvements.  Rather coming is a steady, predictable stream, like Moore’s law, algorithm improvements are more like a phase transition where the performance jumps up changing my an order of magnitude when a break-through is made.  Such breakthroughs are rare and the consequence of many less fruitful research directions.  Once the breakthrough is made the efficiency of the method is improved in a small steady stream, but nothing like the original discovery.  Many examples of these quantum phase transition type of improvements exist: conjugate gradient, multigrid, flux-limited finite differences, artificial viscosity, Karmakar’s method, and others. 

The final area I will touch on is computational practice.  This where things like verification and validation come into the picture.  Modern computational science ought to be about being honest and straightforward about our capability, and V&V is one of the things at the heart of this.  Too often computations are steered into agreement with reality by the heavy hand of calibration.  In fact, calibration is almost always necessary in practice, but the magnitude of its impact is far too infrequently measured.  Even more importantly, the physical nature of the calibration is not identified.  In a crude sense calibration is a picture of our uncertainty.   Too often calibration uses one sort of physics to cover up our lack of knowledge of something else.  My experience has told me to look at turbulence and mixing physics as the first place for calibration to be identified.

If calibration is the public face of uncertainty, what is the truth?  In fact, the truth is hard to find.  Many investigations of uncertainty focus upon the lack of knowledge, which is distinctly different than physical uncertainty.  Lack of knowledge is often explored via parametric uncertainty of the models used to close the physics.  This lack of knowledge studied from parametric uncertainty often does not look like the physical sources of uncertainty, which arise from a lack of knowledge of precise initial conditions that blow up to large scale differences in physical states.  These distinctions loom large in many applications such as climate and weather modeling.  Unraveling the differences between the two types of uncertainty should be one of computational sciences greatest foci because of its distinct policy implications.  It also figures greatly in the determination of the proper placement of future scientific resources. 

Calibration is also used to paper over finite computational resolution.  Many models need to be retuned (i.e., recalibrated) when computational resolution changes.   This effect can easily be measured, but we stick our collective head in the sand.  All one has to do is take a calibrated solution and systematically change the resolution.  Repeatedly, people respond, “I can’t afford a refined calculation!”  Then coarsen the mesh and see how big the changes are.  If you can’t do this, you have big problems, and any predictive capability is highly suspect.  This sort of estimation should provide a very good idea of how much calibration is impacting your solution.   In most big computational studies calibration is important, and unmeasured.  It is time to stop this, and come clean.  Ending this sort of systematic delusion is far more important than buying bigger, faster computers.  In the long run “coming clean” will allow us to improve computational science’s positive impact on society far more than short-term focus on keeping Moore’s law alive.

Computational science isn’t just computers, it is modeling, it is physical theory, it is algorithmic innovation and efficiency, it is mathematics, it is programming languages, programming practice, it is validation against experiments and measurements, it is statistical science, and data analysis.  Computer hardware is only one of the things we should focus on, and that focus shouldn’t choke resources away from things that would actually make a bigger difference in the quality.  Today it does.  A balanced approach would recognize that greater opportunities exist in other aspects of computational science.

At least 13 Things a Systematic V&V Approach Will Do For You and Your Simulation

20 Friday Dec 2013

Posted by Bill Rider in Uncategorized

≈ Leave a comment

 One of the big problems that the entire V&V enterprise has is the sense of imposition on others.  Every simulation worth discussing does “V&V” at some level, and almost without exception they have weaknesses.  Doing V&V “right” or “well” is not easy or simple.  Usually, the proper conduct of V&V will expose numerous problems with a code, and/or simulation.  It’s kind of like exposing yourself to an annual physical; its good for you, but you might have to face some unpleasant realities.  In addition, the activity of V&V is quite broad and something almost always slips between the cracks (or chasms in many cases). 

To deal with this breadth, the V&V community has developed some frameworks to hold all the details together.  Sometimes these frameworks are approached as prescriptions for all the things you must do.  Instead I’ll suggest that these frameworks are not recipes, nor should they be thought of as prescriptions.  They are “thou should,” not “thou shalt,” or even “you might.“

 Several frameworks exist today and none of them is fit for all purposes, but all of them are instructive on the full range of activities that should be at least considered, if not engaged in.

 CSAU – Code Scaling Assessment and Uncertainty developed by the Nuclear Regulatory Committee to manage the quality of analyses done for power plant accidents.  It is principally applied to thermal-fluid (i.e. thermal-hydraulic) phenomena that could potentially threaten the ability of nuclear fuel to contain radioactive products.  This process led the way, but has failed in many respects to keep up to date.  Nonetheless it includes processes and perspectives that have not been fully replicated in subsequent work.  PCMM is attempting to utilize these lessons in improving its completeness.

 PCMM – Predictive Capability Maturity Model developed at Sandia National Laboratories for the stockpile stewardship program in the last 10 years.  As such it reflects the goals and objectives of this program and Sandia’s particular mission space.  It was inspired by the CMMI developed by Carnegie Mellon University to measure software process maturity.    PCMM was Sandia’s response to calls for greater attention to detail in defining the computational input into quantitative margins and uncertainty (QMU), the process for nuclear weapons’ certification completed annually.

 CAS – Credibility Assessment Scale developed by NASA.  They created a similar framework to PCMM for simulation quality in the wake of the shuttle accidents and specifically after Columbia where simulation quality played an unfortunate role.  In the process that unfolded with that accident, the practices and approach to modeling and simulation was found to be unsatisfactory.  The NASA approach has been adopted by the agency, but does not seem to be enforced.  This is a clear problem and potentially important lesson.  There is a difference between an enforced standard (i.e., CSAU) and one that comes across as well intentioned, but powerless directives.  Analysis should be done with substantial rigor when lives are on the line.  Ironically, formally demanding this rigor may not be the most productive way to achieve this end.

 PMI, Predictive Maturity Index developed at Los Alamos.  This framework is substantially more focused upon validation and uncertainty, and IMHO it gives a bit lax with respect to the code’s software and numerical issues.  In my view, these aspects are necessary to focus upon given advances in the past 25 years since CSAU came into us in the nuclear industry.

Computational simulations are increasingly used in our modern society to replace some degree of expensive or dangerous experiments and tests.  Computational fluid and solid mechanics are ever more commonplace in modern engineering practice.  The challenge of climate change may be another avenue where simulation quality is scrutinized and could benefit from a structured, disciplined approach to quality. Ultimately, these frameworks serve the role of providing greater confidence (faith) in the simulation results and their place in decision-making.   Climate modeling is a place where simulation and modeling plays a large role, and the decisions being made are huge.

The question lingers in the mind, “what can these frameworks do for me?”  My answer follows:

  1. V&V and UQ are both deep fields with numerous deep subfields.  Keeping all of this straight is a massive undertaking beyond the capacity of most professional scientists or engineers.
  2. Everyone will default to focusing on where they are strong and comfortable, or interested.  For some people it is mesh generation, for others it is modeling, and for yet others it is analysis of results.  Such deep focus may not lead (or is not likely to lead) to the right sort of quality.  Where quality is needed is dependent upon the problem itself and how the problem’s solution is used.
  3. These are useful outlines for all of the activities that a modeling and simulation project might consider.  Project planning can use the frameworks to develop objectives and subtasks, prioritize and review.
  4. These are menus of all the sort of things you might do, not all the things you must do. 
  5. They provide a sequenced set of activities, prepared in a sequenced rational manner with an eye toward what the modeling and simulation is used for.
  6. They help keep your activities in balance.  They will help keep you honest.
  7. You will understand what is fit for purpose, when you have put too much effort into a single aspect of quality.
  8. V&V and UQ are developing quickly and the frameworks provide a “cheat sheet” for all of the different aspects.
  9. The frameworks flexibility is key, not every application necessarily should focus on every quality aspect, or apply every quality approach in equal measure.

10. Validation itself is incredibly hard in both breadth and depth.  It should be engaged in a structured, thoughtful manner with a strong focus on the end application.  Validation is easy to do poorly.

11. The computational science community largely ignores verification of code and calculations.  Even when it is done, it is usually done poorly.

12. Error estimation and uncertainty too rarely include the impact of numerical error, and estimate uncertainty primarily through parametric changes in models.

13. Numerical error is usually much larger than acknowledged.  Lots of calibration is actually accounting for the numerical error, or providing numerical stability rather than physical modeling.

Referee or Coach? What is the right model for peer review?

13 Friday Dec 2013

Posted by Bill Rider in Uncategorized

≈ Leave a comment

Tags

anonymous, coach, open, peer review, Referee

The process of peer review is viewed as one of the cornerstones of our research system.  The approach to peer review is often cast as an adversarial interaction where the peer’s job is to critique the research, and in the process vet it for suitability for publication.  The most common form of this peer interaction is the refereeing of research papers for journals (and conferences). 

Very few scientists can’t share a few war stories about their experiences with peer review, and most of us have been on both the receiving and the giving end of a contentious exchange.  It is almost a form of professional hazing (or it IS a form of this).  Almost everyone can relate to the moment of dread upon opening the report from the referee.  What outlandish things will they demand?  Will they engage in ad hominem attacks?  What if it is an outright rejection?  I often left thinking, is this really the best system?  It is pretty clear that it’s not.

On the other hand, research without peer review is viewed dimly as it should be. Numerous times the referees report and my response to it made for a better paper.  A few times the referee has succeeded in making me publish a worse paper.  Seriously.  I’ve had referees defend backwards and ignorant practices and demand they continue. A better perspective on the topic is whether the current form is actually the most beneficial form for this practice?  I think we might do well to rethink our model.

I’ll make a modest suggestion for a huge problem; we should strive to engage more as “coaches” instead of as “referees”.

Sports analogies are often useful to may people, or at the very least they are useful to me.   Sports have four categories of people involved: players, coaches, referees and fans.  The players are the source of the action; coaches in training, planning and orchestrating the action and referees make sure the rules are followed.  Fans are primarily interested in the outcomes and only influence the outcome indirectly (or as much as they are allowed to).  Fans are basically like the (interested) public, and while this isn’t my focus today, although it might be an analogy worth developing.  I’m going to apply this analogy to work in science and how different people can or should adopt roles that parallel the actors in sports.

Referees are given authority; their authority is sanctioned.  This is true for a game and this is true for a peer review.  Despite this, they are generally reviled.  What happens when you try to referee without sanctioned authority?  Nothing good, that’s what. Even with the authority granted the act of refereeing is met with some degree of distain. 

At work, I spend a lot of time worrying about the quality of other people’s work. In fact I’m paid to do this, and sometimes have to do this professionally in reviewing papers or projects. This seems to be the lot of someone who does V&V.   What is the right way to do this?  It is easy for this sort of worrying-based activity to become horribly adversarial. 

After all, I know the right way to do things.  Of course I don’t, but the referee in me acts this way all the time.  This sort of interaction is bound to go south.

In this vein I’ve probably tended to act more like a referee in a game.  Everyone knows that players, coaches and fans generally despise referees.  Referees are not only hated, they are conned all the time.  Players and coaches love to get away with cheating the referee, and expect to get away with it.  This gets really ugly fast and I’ve come to realize this isn’t the best way to approach improving quality, which ought to be my main objective.  Even in the situation the sports referee is encouraged to engage in “social engineering” instead of hard-nosed policing.  Often you are negotiating and urging the players into behaving themselves.  This is much more effective than calling foul after foul because the game rapidly turns into a farce.  On the other hand there is fuzzy line that can’t be crossed where you must assert authority.  This line once crossed often results in a red card being shown.  Whenever I’ve had to do this it’s filled me with regret.  “What could I have done differently to prevent this?” 

Of course, referees also appear prominently in science as those anonymous people who review manuscripts.  Referees are viewed as being utterly essential to the entire system of the peer-reviewed literature.  They are also those who worry about quality, and are often despised.  Almost everyone who has published a paper has had to deal with an asshole referee.  I certainty have.  I might also honestly assess my own behavior as a referee; I’ve probably been an asshole too.  Hiding behind the cloak of anonymity has much to do with this; although the editors of journals know who the referees are and they have to tolerate the asshole behavior to some degree for it to persist.   It could be that some editors have a different threshold for what constitutes unreasonable or unprofessional conducts.  Of course some people are just assholes and enjoy being ones.

Of course, the game refereeing is a public, open activity and here the analogy gets a bit thin.  As poorly as refereeing goes in under the public eye, an anonymous referee would be worse.  The outright rejection of a paper is the proverbial red card, and its handed out anonymously.  I think this aspect of the system is ultimately harmful.  It isn’t that some work shouldn’t be rejected, but rather it shouldn’t be rejected behind a veil.  More than this, do we really give the rejected work a chance? Or the feedback it needs to succeed?  More often than not, the answer is no, and the system itself aids in the process.

Given this and lots of scandals appearing in the public eye one might reasonably ask whether the current system of peer review is working.  Most of the embarrassing cases are not in fields remotely related to computational science and physics, but it is probably only a matter of time till something bubbles up.  The flip side to the horrible, asshole referee is the crony referee.  The sports analogy is the corrupt referee who fixes the game.  In the academic literature this ought to be more clearly reviled. Again, the specter of anonymity raises its ugly head to assist the problem’s prominence.  Again, I might ask myself the question of whether I’m even slightly guilty of this sin?  The key point is that the nasty or the crony referees do not serve science well and their influence is hurting research.  This isn’t to say that most of the reviews aren’t well intentioned and professional.  Most are, but enough aren’t that it might be worth considering a different path.  This could be coupled to other aspects of the open-source, open-publishing movement, which is also coupled to reproducibility another big trend in publishing and peer review.

A more positive thing to think about is whether the current system is beneficial, or as beneficial as it could be.  The referees and editors are the gatekeepers, who can allow certain things to be published, and other things not to be published (in a given journal).  For the most part they do not work with the authors to improve the work.  A referee’s report usually only contains a small number of the items that might be improved.  The system limits the amount of feedback the referee might give.  Too much feedback is a sign that the paper should be rejected. 

I’ve seen cases where an earnest review on my part that was too detailed resulted in the editor rejecting the paper.  It was the exact opposite of what I wanted!  I ended regretting not writing a more effusive statement of the value and worth of the paper despite its flaws or necessary (in my opinion!) improvements.  The fact is that if I want to reject a paper I will not write a long review.  I will write why it isn’t “up to scratch” and what is valuable and worth focusing on when resubmitting.  A long review means, “I’m interested”.

The system is geared toward an adversarial relationship rather than a collaborative one.  The general adversarial approach permeates science and greatly undermines the collective ability of the community to solve deep problems. We should develop a peer review process that confers positive properties onto this essential interaction.  Negative behaviors are deeply empowered by the anonymity of the interaction.  If the referee was known to the authors and a collaborative model were established, the behavior of the referees would be better, and the spirit of collaboration could be established and used to improve the work.  Maybe an email exchange or phone call could be used to work out the details of how to improve a paper instead of the horribly inefficient system of review & revision.  Communication could take place in hours or days that now takes weeks or months. 

This is where the analogy of “coach” comes in (teacher or tutor works too).  The referee could be recast as a coach.  A coach for the different subfields the paper covers.  For example, I could be assigned as a “verification” or “numerical methods” coach for a paper.  As a coach I would suggest things to do in order for the work to improve.  Perhaps I might be given a minor, lesser stake in the paper.  In other words, the paper would still have authors, but the “coaches” would also be identified in a manner more significant than an acknowledgement, but less significant than authors.  This role would have to have value professionally for the idea to work. Maybe this could lead to a better, more collaborative community for publishing. A formal role would assist the reviewers in being more energetic and focused in their reviews.  They would become a stakeholder in the paper, not simply an anonymous servant.  Ultimately a positive interaction with a “coach” might lead to a more substantial interaction later.

Coaches operate in many ways, but their objectives are similar to win (i.e., succeed) and often winning is related to teaching.  Great coaches are very often great teachers.  The best coaches not only work toward the success of their team (or players), but also to develop their teams of players to their potential.  Coaches work in numerous ways to achieve these ends through drills, feedback, and challenges.  Each of these approaches develops something greater than what the team or players can do with themselves. As such, the coach is a model for developing an improved capability.  This is basically another way of stating the objectives of the model for replacing referees with something more positive and constructive.  I am saying that we should think more like a coach and less like a referee.

There is a need for some authority to be exercised.  There is a reason we have referees, and they are necessary.  Referees are adopted when the stakes are high and the outcome is meaningful.  Should we be viewing all of research as a high stakes “game”?  Or would approaching most research as a practice where we want to improve and saving the game approach to when things really matter better serve us.  This certainty should be applied to the educational setting where the current system does not maximize the training that young professionals receive.  Perhaps something more balanced can be used, where we have both coaches and referees being assigned to work with a given research.  The key for effective refereeing in fairness and impartiality, along with competence something we clearly don’t have enough of today. 

I think the same model should be adopted in looking at simulation quality in V&V and UQ work.  Those of us with such expertise should act more like coaches and less like referees.   This way we can work constructively to make computations better, we become part of the team instead of part of the “other” team, or the enemy in the middle.

Postscript on Trust (or Trust and Inequality)

06 Friday Dec 2013

Posted by Bill Rider in Uncategorized

≈ 1 Comment

The thoughts associuated with last week’s post continued to haunt me.  It hung over my thoughts as I grappled with the whole concept of thankfulness for last week’s holiday.  I know implicitly that I have a tremendous amont to be thankful for at both a personal and professional level.  By all but the most generous standards, I am very lucky, certainly far better off than the vast, vast majority of humanity.  Nonetheless, it is hard to shake the sense of missed opportunity and squandered potential.  The USA is not a healthy country right now, and that fills me with a sadness.

With a system that is so dysfunctional most of what we can be thankful for is close to home.  Things we used to take for granted, like freedom, and that your children had a better chance at a good life than you do are not secure any more.  We have transformed into a society where corporate rights exceed personal or human rights, and there is little trust in any institution public or private.   I wrote about the lack of trust in science last week and how it makes a scientists work both overly expensive and underly effective.  It is a tragic situation.

Of course the Internet has a lot to do with it.  Or at the very least the Internet has amplified many of the trends that were already apparent preceding its rise to societal prominence.  In a sense the Internet is a giant communication magnifier, but also allows anonymous communication at a scale unimaginable.  I am reminded of what Clay Shirky noted that we may be in the midst of huge reorganization of society similar to what happened after Gutenberg invented the printing press.  Its a big deal, but not much fun to be part of.  Anonymity can be a horrible thing if misused.  We are seeing this in spades.  You really can have too much information, or have it too unfiltered.  Worse yet than a filter is the propaganda that spins almost everything we read toward some world-view.  The money and talent is largely in the corporate camp, and we are being sold ideas they favor with a marketers talent for shaping our views.

Part of my inability to shake these ideas came in the form of a “longread” recommendation about the 40 year decline in the prospects of the American worker (http://prospect.org/article/40-year-slump) starting in 1974.  After massive growth in equality in the years following World War II, it ended in 1974, and began the steady march toward today’s perverse levels of inequality.  This was the same year as I had posited the beginning of the decline in trust for public institutions such as science.  While some institutions such as science, research and government have lost the public’s trust, the corporation has become the centerpiece for society. 

Stockholder return has become the benchmark for success.  This has become a stark contrast to the actual health or long-term prospects for a company.  Likewise, the government-funded research has become starkly risk adverse and commenserately short term focused.  Society as a whole will suffer the effects of these twin pathologies.  Neither public, nor private interests are investing in the future.  Our infrastructure crumbles and no one even thinks of creating a 21st Century infrastructure because we can’t keep our 20th Century infrastructure healthy.  R&D has become all development, and has to be attached to the bottom line preferably the next quarter.   This progress adverse corporate principle is adopted whole-cloth by the government because of a faith-driven belief that the “market knows best” and business practice is inherently superior to other approaches.  One might consider that such research would be application-mission oriented, but no it simply risk adverse and pursued because “success,” such as it is defined, is almost a sure thing.

In the process, the “1%” has risen to dizzying heights of wealth, largely through the decision making process where corporations invest far less capital in their futures and instead to pay massive dividends to their stockholders.  The implications for science, and society are far reaching.  Is there a connection?  I think it is all part of the same thing.  Any science that makes life difficult or complicated for a corporation is attacked because the profit motive and margin has become the sole measure of societal success.  The human toll is massive and our society shows no mercy or thought for the suffering all this wealth accumulation is unleashing.

I am reminded about the problems of correlation being linked to causation, but the seeming coincidence in timing might be more meaningful.  How are these two issues connected?  I’m not sure, but I do think there is one.  Perhaps science being an endeavour for the common good has lost favor because its benefits can be shared by all rather than a few.  Everything that produces a shared benefit seems to be in decline and the entire society is being structured to serve the very top of the food chain.  The wealthy have become apex predators in the human ecosystem, and seem have no qualm about “feeding” on the ever-expanding masses in poverty.  The loss of trust powers the problem because of the unique nature of the American phyche and its devotion to the individual above all.  Now we are transitioning to a view where the devotion is to the corporation (but really the rich who profit from the corporate “success”).

So these thems stretch across the expanse of years that define nearly my entire life, the mystery of how to leave something better for my children is unsolved.  This is what really haunts me, and why I’m not as thankful as I would like to be.

← Older posts
Newer posts →

Subscribe

  • Entries (RSS)
  • Comments (RSS)

Archives

  • February 2026
  • August 2025
  • July 2025
  • June 2025
  • May 2025
  • April 2025
  • March 2025
  • February 2025
  • January 2025
  • December 2024
  • November 2024
  • October 2024
  • September 2024
  • August 2024
  • July 2024
  • June 2024
  • May 2018
  • April 2018
  • March 2018
  • February 2018
  • January 2018
  • December 2017
  • November 2017
  • October 2017
  • September 2017
  • August 2017
  • July 2017
  • June 2017
  • May 2017
  • April 2017
  • March 2017
  • February 2017
  • January 2017
  • December 2016
  • November 2016
  • October 2016
  • September 2016
  • August 2016
  • July 2016
  • June 2016
  • May 2016
  • April 2016
  • March 2016
  • February 2016
  • January 2016
  • December 2015
  • November 2015
  • October 2015
  • September 2015
  • August 2015
  • July 2015
  • June 2015
  • May 2015
  • April 2015
  • March 2015
  • February 2015
  • January 2015
  • December 2014
  • November 2014
  • October 2014
  • September 2014
  • August 2014
  • July 2014
  • June 2014
  • May 2014
  • April 2014
  • March 2014
  • February 2014
  • January 2014
  • December 2013
  • November 2013
  • October 2013
  • September 2013

Categories

  • Uncategorized

Meta

  • Create account
  • Log in

Blog at WordPress.com.

  • Subscribe Subscribed
    • The Regularized Singularity
    • Join 56 other subscribers
    • Already have a WordPress.com account? Log in now.
    • The Regularized Singularity
    • Subscribe Subscribed
    • Sign up
    • Log in
    • Report this content
    • View site in Reader
    • Manage subscriptions
    • Collapse this bar
 

Loading Comments...