This is meant mainly as a supplement to a Christmas Eve-Eve gathering/discussion I was at. I will make this post a little different than other posts, as, it will be “minimalist.” This is the second installment of the topics covered, which are polar bears, rising sea levels, CO2, Inconvenient Truth (the movie), nuclear power, warmest year, electric vehicles (EVs)/hybrid cars, and bullet trains.
Tarzan couldn’t take this kind of hot!
The question becomes this:
- what are the different ways to measure temperature,
- which one are many media sources using,
- and which are the most reliable.
These are the questions any serious person who tells people 2014 is the hottest year as if they are the final arbiter of truth. I have dealt with this in the past, but will again deal with it here.
The two main research groups tracking global lower-tropospheric temperatures (our UAH group, and the Remote Sensing Systems [RSS] group) show 2014 lagging significantly behind 2010 and especially 1998:
Viewed another way:
Another instance highlighting the NOAA’s overestimating temperature is when it said October was the hottest month ever, but RSS data showed October to be ninth warmest on record. And I wonder what family members on the East-Coast would say to there being more than 400 record lows and record cool highs set, via The Weather Channel (11/2014):
A third surge of cold air pushed into the Northern Plains, Upper Midwest and East Coast. This third surge reinforced the cold temperatures for millions of Americans who have already endured at least a week of January-like chill.
There have been more than 400 record lows and record cool highs set, covering 43 states, since Sunday. That leaves only five states in the contiguous U.S., all in New England, that have not experienced record cold temperatures this week.
On Wednesday morning record lows were broken or tied from New York to Houston. Thursday morning brought more record cold to parts of the Southeast.
– First arctic surge: Spread into the East last week (November 11-15).
– Second arctic surge: Blasted through the East, Midwest, and South through early Thursday (November 16-20). For parts of the mid-Mississippi Valley, Ohio Valley, Tennessee Valley and the Middle Atlantic States, this was the coldest of the surges, with numerous daily record lows broken.
– Third arctic surge: Reached the Northern Plains and Upper Midwest Thursday, then slid east across the Great Lakes and parts of the Northeast on Friday. It did not press nearly as far south as the first and second surges did.
– Cold relief: Relief began in the Rockies, then expanded into the southern Plains and Southeast Wednesday and Thursday. Midwest and Northeast relief arrives this weekend….
Okay, what we have already seen is that the satellite temperatures say 2014 will NOT BE the hottest year. One should ask what the hottest year was: 1934:
(L.A. Times) A slight adjustment to U.S. temperature records has bumped 1998 as the hottest year in the country’s history and made the Dust Bowl year of 1934 the new record holder, according to NASA.
That meant that 1998, which had been 0.02 degrees warmer than 1934, was now 0.04 degrees cooler.
This is where I transition to the NOAA temperature, but I wanted to take this transition with help from Dr. Willie Soon (NM), who is an Astrophysicist and Geoscientist at the Solar and Stellar Physics (SSP) Division, Harvard-Smithsonian Center for Astrophysics. (I will embolden the point made below.) More and more scientists have been coming out of the closet and Dr. Soon is one of them. He says in fact that if his community continues “to keep silent and do not express outrage like the one I now feel, the notion of science as a philosophy and way of life will soon be reduced to computer games and animation for the mind-controllers and beauty-contest institutions….” Dr. Soon continues:
2014 hottest year a manipulation
Is this a joke or simply my BAD dream? Prostituting science like this is now consider a virtue. It is no wonder that science writer Lord Ridley said that he has lost his faith on science as an institution.
Why would anyone even bother with claims and insistence of the globe in 2014 being the hottest to a relative colder years all within a few hundredths of a degree Celsius? Poor Anders Celsius should be dancing in his grave.
The claim is based on just one (from a half dozen or so) thermometer-based products whose measurement quality is fraught with uncertainty and with actual error bars at least ten times larger than those claimed “effects”. WMO and others simply pick and choose the “data” that produces the press news they want in time for the Lima, Peru political pow-wow.
In truth the datasets taken as a whole clearly show that the global temperature has been flat-trending for nearly two decades now and that the theory of rising CO2 leading to global warming is sorrowfully exaggerated.
This kind of manipulative science, exemplified by IPCC, WMO, NOAA and what have you, is serving its master in the realm of politics and policy, and is indeed very sickening.
All of them are essentially behaving in ways we would never want any of our school children to behave: cheating and manipulating that are accompanied by careful wording and clever rhetoric….
This sets us up for HOW the NOAA gets their temperatures, and why they are inaccurate. The below is posted elsewhere on my blog and is semi-technical for the layman. But the key is PLACEMENT, and you can see that in the photo’s below Dr. Mueller’s presentation on how the numbers are skewed/manipulated.
(Dr. Mueller is part of the Department of Physics at the University of California at Berkeley, and Faculty Senior Scientist at the Lawrence Berkeley Laboratory, where he is also associated with the Institute for Nuclear and Particle Astrophysics.)
There seems to be a misunderstanding by the general public of the NOAA and other organizations and how they misuse data points (or average them wrongly).
So, for instance, professor Mueller at Berkeley mentions how climate “scientists” were hiding the decline in the past:
They were skewing the numbers in other words. This is an example of fraud. But numbers can be skewed by faulty or outdated methods/equipment. For instance,
Here are some examples of fudged data because of sub-standard equipment:
You’d think the answer would be obvious, but here we have a NOAA operated USHCN climate station of record providing a live experiment. It always helps to illustrate with photos. Today I surveyed a sewage treatment plant, one of 4 stations surveyed today (though I tried for 5) and found that for convenience, they had made a nice concrete walkway to allow servicing the Fisher-Porter rain gauge, which needs a paper punch tape replaced one a month.
Here is what you see in visible light:
Here is what the infrared camera sees:
Note that the concrete surface is around 22-24°C, while the grassy areas are between 12-19°C
This station will be rated a CRN5 by this definition from the NOAA Climate Reference Network handbook, section 2.2.1:
Class 5 (error >~= 5C) – Temperature sensor located next to/above an artificial heating source, such a building, roof top, parking lot, or concrete surface.”
More than half of the stations the NOAA use are tainted or wrongly placed equipment.
Another example of changing averages was noted by Steve Goddard and others — even the NOAA have acknowledge it — have been discussing recently is exemplified in Dr. Judith Carry’s post on the matter (from a larger post of mine):
Even the Wall Street Journal chose the higher temperature reading to say that July of 2012 was July was the “hottest month in the contiguous U.S. since records began in 1895.” WUWT found this on accident and it has led to quite a few other revelations as we will see. Here is description in part of what we looking at:
Glaring inconsistencies found between State of the Climate (SOTC) reports sent to the press and public and the “official” climate database record for the United States. Using NCDC’s own data, July 2012 can no longer be claimed to be the “hottest month on record”.
I initially thought this was just some simple arithmetic error or reporting error, a one-off event, but then I began to find it in other months when I compared the output from the NCDC climate database plotter. Here is a table of the differences I found for the last two years between claims made in the SOTC report and the NCDC database output.
In almost every instance dating back to the inception of the CONUS Tavg value being reported in the SOTC report, there’s a difference. Some are quite significant. In most cases, the database value is cooler than the claim made in the SOTC report. Clearly, it is a systemic issue that spans over two years of reporting to the press and to the public.
It suggests that claims made by NCDC when they send out these SOTC reports aren’t credible because there are such differences between the data. Clearly, NCDC means for the plotter output they link to, to be an official representation to the public, so there cannot be a claim of me using some “not fit for purpose” method to get that data….
The Wall Street Journal made a graph showing this record setting month (below-left). The more accurate temperature for July likewise is shown in the same graph (below-right):
This looking at the data sets chosen and what is used and isn’t used to support an idea that fails in every way. Combine this obvious cherry-picking with the bias, collusion, and charges against the report that the President used to route Congress, all show we have a problem Houston! But this is only the tip of the proverbial iceberg. It seems the NOAA has been skewing these temps for some time. Why? Because the left uses this as a way to promote an ever growing government and the scientists get more-and-more funding. This data fudging story is newer, and it is evolving quickley, including this newest post via Real Science where Steve Goddard notes that More Than 40% Of USHCN Station Data Is Fabricated. Here is Dr. Judith carry’s synopsis (excerpted), in which she critiques a bit Goddard’s post… but then bows to the evidence:
OK, acknowledging that Goddard made some analysis errors, I am still left with some uneasiness about the actual data, and why it keeps changing. For example, Jennifer Marohasy has been writing about Corrupting Australian’s temperature record.
In the midst of preparing this blog post, I received an email from Anthony Watts, suggesting that I hold off on my post since there is some breaking news. Watts pointed me to a post by Paul Homewood entitled Massive Temperature Adjustments At Luling, Texas. Excerpt:
So, I thought it might be worth looking in more detail at a few stations, to see what is going on. In Steve’s post, mentioned above, he links to the USHCN Final dataset for monthly temperatures, making the point that approx 40% of these monthly readings are “estimated”, as there is no raw data.
From this dataset, I picked the one at the top of the list, (which appears to be totally random), Station number 415429, which is Luling, Texas.
Taking last year as an example, we can see that ten of the twelve months are tagged as “E”, i.e estimated. It is understandable that a station might be a month, or even two, late in reporting, but it is not conceivable that readings from last year are late. (The other two months, Jan/Feb are marked “a”, indicating missing days).
But, the mystery thickens. Each state produces a monthly and annual State Climatological Report, which among other things includes a list of monthly mean temperatures by station. If we look at the 2013 annual report for Texas, we can see these monthly temperatures for Luling.
Where an “M” appears after the temperature, this indicates some days are missing, i.e Jan, Feb, Oct and Nov. (Detailed daily data shows just one missing day’s minimum temperature for each of these months).
Yet, according to the USHCN dataset, all ten months from March to December are “Estimated”. Why, when there is full data available?
But it gets worse. The table below compares the actual station data with what USHCN describe as “the bias-adjusted temperature”. The results are shocking.
In other words, the adjustments have added an astonishing 1.35C to the annual temperature for 2013. Note also that I have included the same figures for 1934, which show that the adjustment has reduced temperatures that year by 0.91C. So, the net effect of the adjustments between 1934 and 2013 has been to add 2.26C of warming.
Note as well, that the largest adjustments are for the estimated months of March – December. This is something that Steve Goddard has been emphasising.
It is plain that these adjustments made are not justifiable in any way. It is also clear that the number of “Estimated” measurements made are not justified either, as the real data is there, present and correct.
Watts appears in the comments, stating that he has contacted John Nielsen-Gammon (Texas State Climatologist) about this issue. Nick Stokes also appears in the comments, and one commenter finds a similar problem for another Texas station.
Homewood’s post sheds light on Goddard’s original claim regarding the data drop out (not just stations that are no longer reporting, but reporting stations that are ‘estimated’). I infer from this that there seems to be a real problem with the USHCN data set, or at least with some of the stations. Maybe it is a tempest in a teacup, but it looks like something that requires NOAA’s attention. As far as I can tell, NOAA has not responded to Goddard’s allegations. Now, with Homewood’s explanation/clarification, NOAA really needs to respond….
So we see in the above, that temperatures can be changed years later as the totality of the data is included. What was considered the hottest falls to just an average month in the heat index.
And this has — within the past few months — turned into a very large debate.
EQUIPMENT FAIL II
Here is another example of older/faulty equipment:
A Quick Note about the Difference between RSS and UAH TLT data
There is a noticeable difference between the RSS and UAH lower troposphere temperature anomaly data. Dr. Roy Spencer discussed this in his July 2011 blog post On the Divergence Between the UAH and RSS Global Temperature Records. In summary, John Christy and Roy Spencer believe the divergence is caused by the use of data from different satellites. UAH has used the NASA Aqua AMSU satellite in recent years, while as Dr. Spencer writes:
…RSS is still using the old NOAA-15 satellite which has a decaying orbit, to which they are then applying a diurnal cycle drift correction based upon a climate model, which does not quite match reality.
I updated the graphs in Roy Spencer’s post in On the Differences and Similarities between Global Surface Temperature and Lower Troposphere Temperature Anomaly Datasets.
While the two lower troposphere temperature datasets are different in recent years, UAH believes their data are correct, and, likewise, RSS believes their TLT data are correct. Does the UAH data have a warming bias in recent years or does the RSS data have cooling bias? Until the two suppliers can account for and agree on the differences, both are available for presentation.
ROSE COLORED GLASSES
Another example of competing ideas is this example from two major UK papers, the first being from the Guardian:
This next one from the Daily Mail:
(Daily Mail) ….The most widely used measurements of Arctic ice extent are the daily satellite readings issued by the US National Snow and Ice Data Center, which is co-funded by Nasa. These reveal that – while the long-term trend still shows a decline – last Monday, August 25, the area of the Arctic Ocean with at least 15 per cent ice cover was 5.62 million square kilometres.
This was the highest level recorded on that date since 2006 (see graph, right), and represents an increase of 1.71 million square kilometres over the past two years – an impressive 43 per cent.
Other figures from the Danish Meteorological Institute suggest that the growth has been even more dramatic. Using a different measure, the area with at least 30 per cent ice cover, these reveal a 63 per cent rise – from 2.7 million to 4.4 million square kilometres.
Crucially, the ice is also thicker, and therefore more resilient to future melting. Professor Andrew Shepherd, of Leeds University, an expert in climate satellite monitoring, said yesterday: ‘It is clear from the measurements we have collected that the Arctic sea ice has experienced a significant recovery in thickness over the past year.
‘It seems that an unusually cool summer in 2013 allowed more ice to survive through to last winter. This means that the Arctic sea ice pack is thicker and stronger than usual, and this should be taken into account when making predictions of its future extent.’
Same data used, one says BEWARE, the END IS NEAR, the other says making gains, as the graph shows below:
(So are Polar Bears, BTW) And the Antarctic has made BIG GAINS. But if the left — yes, the left — says we should renter the little ice-age to be at a normal point of “climate disruption,” then they are living a pipe-dream. As the earth gets warmer life flourishes, as warmer periods in history have exemplified:
June was ranked one of the coldest months in a while. According to the NOAA, July 2014 ranked 29th coldest out of 120 using the Maximum temperature. And so we enter the discussion about if the pause is real… which is now being acknowledged by almost all (except the IPCC).
Here is Dr. Judith Curry posting ion the subject before getting into two papers that talk about it from two points of view:
With 39 explanations and counting, and some climate scientists now arguing that it might last yet another decade, the IPCC has sidelined itself in irrelevance until it has something serious to say about the pause and has reflected on whether its alarmism is justified, given its reliance on computer models that predicted temperature rises that have not occurred. – Rupert Darwall
The statement by Rupert Darwall concisely states what is at stake with regards to the ‘pause.’ This seriously needs to be sorted out….
For a running — updated — list of these excuses mentioned above, see here. What HAS been clearly shown is that while some wish to make CO2 illegal, CO2 is in fact not driving the climate:
What is being shown as of late is that more-and-more scientists are becoming concerned with the group-think in the climate-sciences. Again, Dr. Judith Curry was the biggest pro-AGW proponent, but has — like many others leaders in their respective field — changed or softened her/their positions on what the science is actually showing:
The implications of dogmatic groupthink and intimidation for the pursuit of sound science — and sound policy — are chilling. – Christopher Snowden
A collection of articles from the health science community on the fate of papers and scientists that challenge the consensus.
SUN & OCEAN
- Gross Scientific Negligence – IPCC Ignored Huge Body Of Peer-Reviewed Literature Showing Sun’s Clear Impact – See more at: http://tinyurl.com/kw47zcf (NASA is changing on this as well)
- NASA is saying the sun — thanks to the Goddard Institute studying this, is the driver of warming and cooling: http://tinyurl.com/m29yo99 (Via Hockey Shtick)
And yet another study is showing the sun as the major player… NOT greenhouse gases.
So what’s the bottom line? NOAA temps change over time. Many in this respective field are seeing group-think. Ideology is driving this group-think, not science.
Why should you be interested? Sea surface temperature records indicate El Niño and La Niña events are responsible for the warming of global sea surface temperature anomalies over the past 30 years, not man-made greenhouse gases. I’ve searched sea surface temperature records for more than 4 years and ocean heat content records for more than 3 years, and I can find no evidence of an anthropogenic greenhouse gas signal in either dataset. That is, the warming of the global oceans has been caused by naturally occurring, sunlight-fueled, coupled ocean-atmosphere processes, not anthropogenic greenhouse gases.
(From the newest global temp data-set at WUWT)
No matter what evidences one puts forward, until people remove their rose-colored glasses, they will continue to explain away the pause showing CO2 has nothing to do with global temperatures.