## Thursday, February 27, 2020

### ## USA Temperatures; averaging and graphics methods.

I have for many years been experimenting with methods for calculating average surface temperature anomalies from collections of station readings (including sea surface from grid, regarded as stations). I describe the history here. In the early days, I tried to calculate the averages for various countries and continents. With the coarse grids I was using at the time, and later meshes, I found the results unsatisfactory, and of limited scientific importance.

So I put more effort into developing global methods. There is a rather mathematical survey of these here. Briefly, there are six main types:

• The conventional grid analysis, averaging the measurements within each cell, and then getting an area-weighted average of the cells. I think this is very unsatisfactory since either the grid is very coarse, or there are cells with no data.
• Triangular mesh, with the sites as nodes. This has been my mainstay. But it is not very good at following national boundaries.
• A method where spherical harmonics are fitted and then integrated. This is now implemented as a general method for improving otherwise weak methods. The structure is instructive, but again, intrinsically global.
• A LOESS method, described here This has the characteristic of gathering information from as wide a net as needed; useful when stations get sparse, but not a respecter of boundaries.
• Most recently, a method using FEM shape functions. I think this may be best, in terms of getting the best representation on a relatively coarse grid. Again, not so good for boundaries, but it has as a special case, one of my earlier methods:
• Grid with infill (eg an early version here). The weakness of conventional gridding is that it does not use local information to estimate missing cells, and so the grid must be kept fairly coarse. But I have worked out a way of doing that systematically, which then allows much finer gridding. And that does have the incidental benefit of tracking national and land boundaries. It also allows a good graphics scheme. I'll say more about it in the next section. In this text, I'm using blue colors for the more technical text.

#### Infilling and diffusion - the Laplace equation.

The logical expression of using neighbour information is that empty cells should be assigned the average temperature of their neighbours. For an isolated cell, or maybe a few, you can do this directly. But with clumps of empty cells, some may have no known neighbours at all. But you can still maintain the requirement; a solution procedure is needed to make it happen.

The idealization of this is the Laplace differential equation, solved with Dirichlet boundary conditions expressing the known cells, and zero gradient (Neumann) conditions at the boundaries (not needed for the globe). That equation would describe a physical realisation in which a sheet of metal was kept insulated but held at the appropriate temperatures in specified locations. The temperatures in between would, at equilibrium, vary continuously according to the Laplace equation.

This is a very basic physical problem, and methods are well established for solving. You just write a huge set of linear equations linking the variables - basically, one for each cell saying that it should be the average of the neighbours. I used to solve that system just by letting it play out as a diffusion, but faster is to use conjugate gradients.

Once the data-free cells are filled, the whole grid can be integrated by taking the area-weighted average of the cell values.

In one dimension, requiring each unknown value to be the average of its neighbours would lead to linear interpolation. Solving the Laplace equation is thus the 2D analogue of linear interpolation.

#### Graphics and gridding the region

I mentioned the usefulness for graphics. I just draw a point of the appropriate color for each cell. For ConUS (contiguous) I typically use grid cells of about 20km edge, giving about 20000 for the country, and so quite good resolution. I will give examples shortly.

As far as getting a grid of the US, say, one might look on public data for a lat/lon land mask. There is a lot of choice for physical land, but I found it harder for ational boundaries. But anyway, there is another consideration; projection. A lat/lon grid is not isotropic. N-S neighbours are further away than E-W. Proper use of the Laplace equation would require allowance for this. And anyway, good graphics would require an equal area projection.

So I decided to make a map, after projection, using the R maps package. I can extract a specification of the US as a complex polygon. I remap the data, first onto a sphere, then project that on to a plane tangent to the sphere at the US center (I used 37N, 97W). Then I output that on a suitable small scale as a BMP file (uncompressed), and read off the pixels. That gives a regular grid, not lat/lon, which is very close to isotropic - cells are close to square (see notes below).

The task of mapping requires another solution of the Laplace equation, applied to the cell averages of the station anomalies.

#### Data sources

I was interested to see the result for ConUS because GHCN V4 monthly has a dense network of nearly 12000 US stations. I used both the unadjusted and adjusted versions. There is also the data from USCRN. Here I use just the 115 commissioned stations, for which some were reporting in 2002, but I use only since 2005, when there were enough to try for a regional average. These stations are evenly distributed, and of high quality, with almost no missing (monthly) data recently. But they are sparse.

NOAA posts monthly averages ClimDiv and USCRN which I can use for comparison. I'll do that in detail in my next post. GISS posts unfortunately only annual averages, but I can use that for a check. I'll talk more about NOAA's graphics below. They feature the anomaly graph in their national climate reports, although they describe them as "Average Temperature Departures" from an absolute spatial average, so it may not be the same thing.

#### An example - USCRN

For the graphics I solve the Laplace equation on a finer mesh of about 10x10 km. The plot of the solution for USCRN shows the good and bad features of the Laplace interpolation method:

The stations are shown with small dots. With this, and the next plot, you can remove the stations and the state borders with the checkboxes top right. Mostly the interpolation is smooth, with a few that stand out creating a small dimple. The color actually at the station is in this image exact. There are just 115 stations, so that is a fairly good result. The anomaly base is 2005-2019, which is what is available with USCRN, and so I use the same base when comparing with GHCN.

Now here is a plot of the same month, with the much more numerous GHCN V4 data (5263 stations reporting)

Obviously there is much more detail, not resolved by the CRN infill. The integral is almost identical (1.694 vs 1.689). I'm using unadjusted GHCN data here, although it makes very little difference. I think the results show very good spatial concordance between the GHCN stations, and that the patterns that they reveal are signiicant.

#### Comparison with NOAA

I mentioned that NOAA publishes anomaly maps in the monthly National Climate Reports (January 2020). Here is their plot:

Here is my plot, now using the colors (almost) and levels of the NOAA plot. I also use the same anomaly base (20th century) and now use adjusted GHCN, although again, it makes very little difference over recent times. Temperatures now in Fahrenheit.

It's pretty similar. Mine is a little cooler on the west coast. The agreement is not because of identical data, as the ClimDiv set used by NOAA is somewhat more extensive.

I've shown a smoothing factor; 0.25 is essentially no smoothing. Here is a plot which smooths out some of the speckles without changing the result substantially

#### Next steps

I'll post very soon a much more detailed comparison of time series. Generally, my GHCN based averages agree very well with the NOAA Climdiv series since 2005; the agreement with the CRN average, both my GHCN and Climdiv, is not as close. In earlier times, the effects of differing adjustment do start to show through. I think the inferior agreement of the CRN average emphasises that the main error in regional temperature averaging is not station imperfections, but coverage uncertainty. The finer grade GHCN shows how a different location of the CRN sites, no matter how good they are, would give a different result.

I started looking at the US results because they are often misused by sceptics, and I wanted a good independent method of checking on the sometimes opaque NOAA procedures.

I'll probably add monthly updates to the latest data page. I'll do something similar with Australia.

#### Some technical notes on implementation.

1. It took a while to get the Laplace operator right. I do the following steps:

1. Make an nx4 matrix of the neighboring values for the n stations. Calculate the row sums (not means).
2. Make a diagonal consisting of the number of items in each row. Add B ( a large number, say 100) to the cells with data.
3. Make another vector with B's on the cells with data, 0 elsewhere.
4. The symmetric part of the implementation is done by multiplying the coefficients from 2 by the values, and subtract the rowsums from row 1.
5. This is then scaled by the inverse of the diagonal, and right multiplied by the vector from 3. These operations can be deferred; the conjugate gradient iteration, which requires symmetry, should be done on the symmetric part and the scaling later.

I needed a scheme to deal with the imperfect coast alignment (and possibly inaccurate lat/lon in the GHCN inventory). Some stations are out to sea. I linked them with the nearest (usually adjacent) cell in the US grid. The Channel Islands of California resisted this process, and I had to omit them.

I use a TempLS-like convergence to get the station normals (offsets). They are not necessarily the same as those used in the global calculation. It is an independent analysis. It takes only a minute or so for the GHCN case (CG iterations are about 30-100 per month); CRN is longer, because there is more infilling to do. Those numbers are for the 20 km cells; 0 km are much slower.

Islands without data can be a problem, since they make the solution indeterminate, which slows the iterative solution. I added a very small number (1e-6) to the diagonal, which fixes the island solution at zero (anomaly). Fortunately this was a rare problem.

I described above the way I projected the US onto the plane. I rotated the spherical coords so that the centre (37N 97W), was at (0N, 0E), which is the (0,0,-1) 3D point. Then I projected onto the z=-1 plane from the point (0,0,2). This minimises the distortion near the centre, which is to say, for most of the USA.

Update: In comments, MMM suggested a difference plot between CRN and GHCN would be interesting. I agree. Although the differences are not large, they do not seem to be random. Here is the plot:

## Saturday, February 15, 2020

### ## GISS January global up by 0.08°C from December.

The GISS V4 land/ocean temperature anomaly was 1.18°C in January 2019, up from 1.10°C in December. That compares with a 0.047deg;C rise first reported in the TempLS V4 mesh index (now 0.079°C with later data). It was the warmest January in the GISS record, just 0.01°C warmer than the El Niño 2016.

As usual here, I will compare the GISS and earlier TempLS plots below the jump.

## Wednesday, February 5, 2020

### ## January global surface TempLS up 0.047°C from December.

The TempLS mesh anomaly (1961-90 base) was 0.951deg;C in January vs 0.904°C in December 2019. This was contrary to the 0.035°C fall in the NCEP/NCAR reanalysis base index. It makes it easily the second warmest January in the record, just behind the El Niño 2016 at 1.011°C.

The prominent feature was a huge band of warmth stretching from Europe through to E Siberia and China. N America was also warm, except for Alaska (cold). Greenland and the Arctic archipelago were also cool. Africa was rather cool, Antarctica mixed.

Here is the temperature map, using the LOESS-based map of anomalies.

As always, the 3D globe map gives better detail.

## Tuesday, February 4, 2020

### ## More trouble with Australian bushfire statistics.

I wrote last month about a fallacy that was entering the discussion of extent of bushfires in Australia. The very bad forest fires that we have been having have burned about 10 million hectares in SE Australia. That is a huge amount historically, and is an area larger than Hungary or Portugal (or Indiana). It is quite large proportion of our temperate forest, where many people live, and overlaps with a lot of agriculture.

Wikipedia keeps a list of past bushfires in which areas are quoted, and has often been used as a source. I noted that their list caused confusion, because along with forest fires they had included, for random reasons, just a few years with extraordinarily high totals, which are not otherwise known as bad bushfire years. One in particular was 1974/5, with a total of 117 million ha. Inferences have been drawn from this. However, the comparisons are spurious, because Wiki included in that year the huge areas that are regularly burnt in the northern savanna scrublands. Although that was an unusually large year, figures of 20-30 million hectares reported are common. I emphasised that it was important to separate these fires, which have little adverse impact, from the hugely destructive temperate forest fires in statistics for comparison.

Unfortunately the situation has got worse, and I think all Australian bushfire area statistics should now be regarded with scepticism. The Wikipedia list has got totally out of hand. I entered discussion there, pointing out that 1969 had an area burnt in the Northern Territory equal to that of 1974. This was intended as a reductio ad absurdum, but all it achieved was that that total was included too, making another year to be quoted for its extreme fires. And 1974/5 is a total mess, with not only the huge total but also the separate state totals, and then a collection of NSW fires.

But it is now worse again, in that the 2019/20 total has been boosted to 18.6 million ha, by inclusion of this year's NT fires (6.8 million ha). That totally corrupts the statistics. 6.8 was actually a low number for NT; next year it could easily be more than 20 million ha, and we will be told that 2019/20 was nothing in comparison.

And now I find that it is not just Wiki but the Australian Gov't Dept of Home Affairs which has put out numbers using similar arithmetic. I don't know why DHA is publishing bushfire stats at all, but I am very suspicious, because its minister is Peter Dutton, the conservative enforcer in the Cabinet, and I don't think he has the accurate appreciation of bushfire severity as a prime aim. I think the prospects of Wiki and similar sources getting it right are now bad, and the statistics are best ignored.

#### Reliability of savanna fire statistics

I think the savanna stats should be kept separate, because they represent very low impact fires, where there is no real effort at suppression. But I also think the numbers are very dubious. They are taken from a worldwide database compiled by Giglio et al (paper here) and based on satellite MODIS observations. As far as I can tell from the paper, they mostly rely on monitoring the actual flames, rather than deciding on the burnt appearance afterwards. So the translation of that into area burnt involves some guesswork, and then there is the question of what burnt even means where the vegetation is sparse. (See comment below by WT)

And sparse it often is. At WUWT a post displayed this map of the 2001 fire distribution in Australia:

By contrast here is a Wiki map of the named desert regions

As you can see, the area burnt includes a lot of desert, including the Great Sandy Desert, Tanami Desert, Gibson Desert. Whether these could really be said to have meaningfully burned, there is no useful comparison of a hectare of such desert with one of temperate forest in NSW.

Update: I found a survey paper on savanna fires in Australia here.

## Monday, February 3, 2020

### ## NCEP/NCAR reanalysis down 0.035°C in January 2020.

The Moyhu NCEP/NCAR index fell from 0.587°C in December to 0.552°C in January, on a 1994-2013 anomaly base. December was warm, so that still makes January the second-warmest month since last March.

Europe and all of Russia were warm. So was N America, except for Alaska, which was cold. N Africa was cool. Greenland and adjacent Arctic was cool; Antarctica was mixed. There was still warm blob SE of New Zealand.