Archive for the ‘Philosophy’ Category
On simplicity
This is a long screed in response to a reading list posted by Massimo Pigliucci (so he bears no responsibility) where he nominated a post on simplicity in science by Elliot Sober on Aeon.
Why is simplicity better?
As it happens, I am in the midst of an argument with the climate science community over simplicity as applied to statistical inference. A couple of days ago I bought Probability, Confirmation and Simplicity: readings in the the philosophy of inductive logic Foster and Martin 1966, which contains six essays on simplicity. Not as simple as it’s cracked up to be – exactly the ammunition I require.
Accordingly, I disagree with Sober. He refers to the Akaike Information Criterion, which measures simplicity but says that it refers to the same underlying reality. But we see it being repeatedly used for different underlying realities by people who don’t read the small print. They are being simplistic (#OccamsRazor). By mixing probabilities with theory Sober is making a fundamental mistake. I can apply probabilities to an experiment or a test, but I cannot to a theory. At best I can severely test (Mayo) a hypothesis and by attaching it to probative criteria in such a way that the alternatives are as unlikely as the hypothesis is likely, then I have a chance of confirming that theory.
In climate science, simplicity is represented by trend-like change. Under increasing greenhouse gases, forcing leads to warming as the logarithm of the increasing forcing plus feedbacks. In the Earth system, this leads to monotonic warming, linear to forcing. Trouble is, most of this heat is absorbed by the ocean and it is the atmosphere that needs to respond. The atmosphere-ocean relationship is a dissipative system driven by thermodynamics and decidedly nonlinear. So, if I assume the atmosphere warms according to the linear radiative forcing concept, I have a simple model that is predictive over demi-century-long timescales. If I assume that warming obeys the dissipative pathway, then it proceeds via enhanced climate variability as a series of step-like regime changes. Over both pathways, warming reaches close to the same destination but its mode of getting there is very different. One contains more inherent risk than the other.
So, I can represent both pathways statistically. They get similar sum of squares residuals (trend-like change fails the heteroscedasticity test but almost no-one tests for this), but because the pathway of step-like change carries more adjustable parameters, it is penalised (actually that isn’t even true because the detection method is completely different). But they represent different realities – Sober does mention this but few have remembered this before, so why should they now?
Where simplicity works with this example, is that the natural greenhouse effect (average 155 Watts per square metre per year) is distributed through climate variability. The net anthropogenic greenhouse effect is 0.7 Watts per square metre per year and roughly 1% of that is assumed to be stored within the atmosphere (0.07 W m2/yr) producing a trend. So here we have created a very complex physical situation where most of the energy flux is controlled by climate variability and where perturbations in the climate of >1 W m2/y can be brought back to mean within months, but somehow a tiny amount of heat remains in the atmosphere in preference to an ocean with 24 times the heat conductivity and 3,200 the heat capacity.
Whereas we could accept that the simplest thermodynamic solution is for all heat to follow the same pathway, for climate change to behave like enhanced climate variability and for warming to follow a series of regime change producing a long-term, complex trend.
Theoretically and thermodynamically simple, statistically more complex. The problem with the simplicity argument is that it has to be very finely applied, and that confusing methodological simplicity with theoretical parsimony is an issue. In economic, climatology and a number of other disciplines, simplicity is being misapplied to methods rather than theory and this is a problem, because it means we apply simple solutions to complex, real-world problems.
Discussion paper for open review
After promising to have our flagship paper on reconciling the signal and noise of global warming on decadal timescales subject to open review, it is finally on. The paper has been submitted and accepted for open review at Earth System Dynamics.
Reconciling the signal and noise of atmospheric warming on decadal timescales
Roger N. Jones and James H. Ricketts
Victoria Institute of Strategic Economic Studies, Victoria University, Melbourne, Victoria 8001, Australia
Received: 13 Aug 2016 – Accepted: 22 Aug 2016 – Published: 23 Aug 2016
Abstract
Interactions between externally-forced and internally-generated climate variations on decadal timescales is a major determinant of changing climate risk. Severe testing is applied to observed global and regional surface and satellite temperatures and modelled surface temperatures to determine whether these interactions are independent, as in the traditional signal-to-noise model, or whether they interact, resulting in steplike warming. The multi-step bivariate test is used to detect step changes in temperature data. The resulting data are then subject to six tests designed to show strong differences between the two statistical hypotheses, hstep and htrend: (1) Since the mid-20th century, most of the observed warming has taken place in four events: in 1979/80 and 1997/98 at the global scale, 1988/89 in the northern hemisphere and 1968/70 in the southern hemisphere. Temperature is more steplike than trend-like on a regional basis. Satellite temperature is more steplike than surface temperature. Warming from internal trends is less than 40 % of the total for four of five global records tested (1880–2013/14). (2) Correlations between step-change frequency in models and observations (1880–2005), are 0.32 (CMIP3) and 0.34 (CMIP5). For the period 1950–2005, grouping selected events (1963/64, 1968–70, 1976/77, 1979/80, 1987/88 and 1996–98), correlation increases to 0.78. (3) Steps and shifts (steps minus internal trends) from a 107-member climate model ensemble 2006–2095 explain total warming and equilibrium climate sensitivity better than internal trends. (4) In three regions tested, the change between stationary and non-stationary temperatures is steplike and attributable to external forcing. (5) Steplike changes are also present in tide gauge observations, rainfall, ocean heat content, forest fire danger index and related variables. (6) Across a selection of tests, a simple stepladder model better represents the internal structures of warming than a simple trend – strong evidence that the climate system is exhibiting complex system behaviour on decadal timescales. This model indicates that in situ warming of the atmosphere does not occur; instead, a store-and-release mechanism from the ocean to the atmosphere is proposed. It is physically plausible and theoretically sound. The presence of steplike – rather than gradual – warming is important information for characterising and managing future climate risk.
Comments welcome: here or there. Deadline October 4.
Step change hypothesis and working paper
Imagine you didn’t know anything about climate change and the greenhouse effect but were interested and you know a bit about general science. Would you accept the following story?
“Earth’s climate is a large, complex system, affected by forces that produce both linear and nonlinear responses. Shortwave radiation – basically UV – from the sun comes in and heats up the planet, producing infrared radiation. Some UV gets reflected straight back out by clouds, snow and ice and stuff. The land can heat up quite a lot, but it cools back down again and doesn’t store much. If a forest is cleared and replaced by buildings, it will warm up a bit but the effect is only local.”
“But the ocean – that’s another story. It absorbs a lot of radiation, so is taking up heat all the time. Huge streams of energy are entering and leaving the ocean store each year. Some is ‘dry’ or sensible heat, which is ordinary warmth. Some is ‘wet heat’ or evaporated moisture. Energy gets taken up when the moisture is evaporated and it will be released again when the moisture cools, condenses and then gets rained out. In this way, the oceans provide a lot of heat to the land every year, largely as rainfall and a bit of snow.”
End of the hiatus
Understanding Climate Risk has been in something of a hiatus, or a pause for the last couple of years due your host being almost fully submerged, but maybe it’s time to rise to the surface and get things going again.
This is for a few reasons. One is that research, especially public good research and especially in CSIRO, is under serious threat in Australia. We have a government who tout innovation, but who wilfully ignore the role of the generation of underpinning knowledge in fuelling such innovation. They are interested only in commercial innovation – public-good innovation is not only being ignored, it is being excluded from processes such as the Cooperative Research Centre bids currently under way. Having sustainable cities, catchments and ecosystems is impossible without public good research and social innovation, with funding that extends across the sciences, the humanities and the arts. With an election going on, these harms need to be publicised. Read the rest of this entry »
The scientific origins of the gradualist adaptation narrative and how to move beyond it
The following statements are typical of the gradualist adaptation narrative:
- Within limits, the impacts of gradual climate change should be manageable.
- Therefore, climate change adaptation can be understood as: (a) adapting to gradual changes in average temperature, sea level and precipitation.
- Gradual climate change allows for a gradual shift in the mix of crops and to alternative farming systems.
So why are Gauss and Newton in the bath and Ed Lorenz in the hot tub?
Et tu, Chief Scientist?
Less than a fortnight ago, I wrote that those barracking for conventional scientific theories often maintain that science is not a matter of belief. Sorry guys, but assessing the probability of t (scientific truth) being T (absolute truth), is a matter of belief, as is anything that goes on in the mind regarding external evidence. But there is a difference between belief and true belief.
And then in a conversation with the Chief Scientist Ian Chubb on communicating science, with specific reference to climate science, The Conversation quotes him as saying:
“We scientists need to talk about evidence, and without being cornered into answering questions like ‘do you believe?’,” Professor Chubb said.
“I get asked that every day and every now and then I make a mistake and say yes or no…It’s not a belief, it’s an understanding and an encapsulation and interpretation of the evidence.”
Aaaaaaaaaaaaarrrrrggghhh!!!
Ian Chubb was speaking to the Royal Society of Victoria, which launched on August 30th a three-year program aimed at increasing the awareness of science among primary school children. And while I agree with most of what Ian Chubb says in his interview with The Conversation, the belief thing should be called for what it is – a full-blown fallacy.
Read the rest of this entry »
Grauniad on Kuhn
No, not an exotic Hungarian sandwich, but The Grauniad has an excellent 50th anniversary review of Kuhn’s The Structure of Scientific Revolutions written by John Naughton in their Sundy paper. In a total paradigm shift, the comments aren’t totally trolltown, either. See also, Howard Sankey on the The Conversation.
Before Kuhn, our view of science was dominated by philosophical ideas about how it ought to develop (“the scientific method”), together with a heroic narrative of scientific progress as “the addition of new truths to the stock of old truths, or the increasing approximation of theories to the truth, and in the odd case, the correction of past errors”
Manual on Scientific Evidence
My email had a notice for the US National Academies Press and an interesting link in the just-published section. Their blurb:
The Reference Manual on Scientific Evidence, Third Edition, assists judges in managing cases involving complex scientific and technical evidence by describing the basic tenets of key scientific fields from which legal evidence is typically derived and by providing examples of cases in which that evidence has been used.
The chapter on How Science Works is really useful. It’s by David Goodstein, a physicist, and gives a quick tour through the development of the scientific method. It is an entertaining read, perhaps surprising for a manual. It starts with Francis Bacon who said the scientist should be an impartial observer of nature. Goodstein quickly skewers this – no serious thinker accepts that observation is done without prejudice – she must have assumptions about how the world works. Yet look around today, and you’ll find the Baconian ideal still alive and kicking (Plimer has it, for example. Ok, go ahead and laugh).
Science in a democratic society
This post was precipitated by several stoushes held at Larvatus Prodeo over climate science but reflects more widely on the state of climate science and its public perception in the English-speaking democracies. It’s an issue I’ve been interested in for a number of years because of the attacks on climate science and the need to build better links between science and the risk management of global change. The post title is also the title of Phillip Kitcher’s new book released earlier this month. Kitcher is an English philosopher based at Colombia University and in 2006 won the Prometheus Prize of the US Philosopher’s Association – this book is the result. He has made his case really well.
The main points of this piece are that:
- Society needs to draw from a body of public knowledge in order to be successful. Psychological and cognitive limitations lead to the sum of individual decisions producing suboptimal outcomes.
- Attacks on public knowledge driven by self-interest and opaque values are being made under the cover of free speech and individual freedoms. The evidence used by these attacks is generally untrue, distorted or selective or fails basic tests for scientific proof.
- Science is a values-driven enterprise. Those values need to be made explicit in what Kitcher refers to as well-ordered science.
- Science is secular. Passing certain probative (proof) tests allows it to be shared as knowledge that has claims to objectivity.
- Belief is personal and can also be shared but does not require the same tests (Belief also expresses a set of human needs not necessarily addressed by science).
- Public knowledge in English-speaking democracies has become degraded. Science is vulnerable to vulgar democracy, where under the guise of free speech, any belief can masquerade as knowledge.
- Science also needs to become better ordered, through measures that cover:
- Education – for most students teaching what science does and what its impacts are, rather than how it works (technical), by separating pedagogy into liberal education and technical specialisation. This works on the presumption that most people need to understand the role science plays in society while fewer will become actual scientists.
- Bringing people into the scientific workplace to familiarise them with knowledge goals and probative values and methods of certifying science.
- Avoidance of universal punditry (experts speaking beyond their expertise) and overconfidence in findings in favour of communicating scientific evidence with the appropriate levels of confidence in theory and uncertainties in outcomes.
- A process that steps through claims of consensus, consequences of those claims, ethical exploration of a potential policy framework and an exploration of how current actions can be balanced against future harm. Read the rest of this entry »