Tuesday, December 22, 2009
Now, back to the polar bears. From the polarbearsinternational it is confirmed that the population of polar bears from the 1960's was very low, in the few thousands, and in the early 2000s was up to between 22000-25000. These are also reflected in the usgs site on polar bears where you can actually get some of the data.
If you go to the polar bear specialist group which advises for the IPCC, you can find a table of the status of the polar bear in various regions. The problem with this table is that there is a column titled "Observed or predicted trend". Hello? Why would you mix observed trends and predicted trends in a table? Just show me the data. Anyway, there is a document here which has an explanation of the projections, and possibly some data, although I haven't read the 200 pages of the document to see if it is buried in the text (there is no figure with the data...just model predictions). I'd love to find a straightforward presentation of the estimates of the current numbers of polar bears, complete with error-bars to denote uncertainty.
It seems reasonable that with the decline of the arctic ice that the polar bear populations can be affected, some more than others, but the role of hunting (the regulation of which caused the surge in the bear numbers from the 1970s) still plays a role can is difficult to disentangle...I've seen unsubstantiated claims that the areas with the decrease are primarily hunting areas. I haven't confirmed this, but this could also be the fact that ice retreat will be more substantial in the more habitable areas, where there would be more people. Correlation does not equal cause and effect.
So, it is a fact that there are many more polar bears now than, say 30-40 years ago. It can also be true, although I have difficulty tracking the data down in a readable form, that arctic ice retreat could impact polar bear numbers adversely.
One question that I have now is, if it was warmer 1000 years ago, is there evidence that there was a significant retreat of the ice back then? If that is the case, then the polar bear scare is just that...a scare. Again, many of global warming consequences that are being reported are tied to the question of the size, extent, and effects of the Medieval Warming Period. That's why, in my opinion, that is the most important question of all.
Tuesday, December 15, 2009
The scientific method's central motivation is the ubiquity of error--the awareness that mistakes and self-delusion can creep in absolutely anywhere and that the scientist's effort is primarily expended in recognizing and rooting out error.
In stark contrast to the sciences relying on deduction and empiricism, computational science is far less visibly concerned with the ubiquity of error. At conferences and in publications, it's now completely acceptable for a researcher to simply say "here is what I did, and here are my results." Presenters devote almost no time to explaining why the audience should believe that they found and corrected errors in their computations. The presenter's core isn't about the struggle to root out error--as it would be in mature fields--but is instead a sales pitch...
Many users of scientific computing aren't even trying to follow a systematic, rigorous discipline that would in principle allow others to verify the claims they make. How dare we imagine that computational science, as routinely practiced, is reliable!
On ClimateAudit, there is an older article (2005) about the Hockey Stick plot. Ross McKitrick makes the suggestion of an audit panel,
A group of experts fully independent of the IPCC should
be assembled immediately after the release of any future IPCC Reports to
prepare an audit report which will be released under the imprimatur of the IPCC itself. The audit will
identify the key studies on which the Report’s conclusions have been
based, and scrutinize those studies, with a view to verifying that, at a
- data are publicly available,
- The statistical methods were fully described, correctly
implemented and the computer code is published
- If the findings given maximum prominence are
at odds with other published evidence, good reason is provided in the
text as to why these findings have been given prominence.
Any competent scientist can assess these things. My strong
recommendation is that such a panel be
drawn from the ranks of competent mathematicians, statisticians,
physicists and computer scientists
outside the climatology profession, to prevent the conflict of
interest that arises because
climatologists face career repercussions from
publicly criticizing the IPCC. Also, participation should
exclude officials from environment ministries,
because of the conflict of interest entailed in the fact
that environment ministries are the main financial
beneficiaries of the promotion of global warming fears.
The second recommendation is for a "counter-weight panel", whose job would be to actively try to find holes in the analysis, assumptions, etc...
I'm not sure how I feel about the second one (I'll have to think about it), but the audit panel to me makes total sense. Why don't the scientific journals do this as a matter of policy?
Saturday, December 12, 2009
It's an entertaining introduction to risk analysis, and they use the word micromort, referring to a measurement of the quantity of an action which gives you a 1 in a million chance of dying. The website is www.understandinguncertainty.org.
Thursday, December 10, 2009
Why is the scientific consensus on Anthropomorphic Global Warming (AGW) different than the scientific consensus on Evolution? Let me list some of the ways:
- Evolution has many independent, very different, lines of evidence (fossils, embryology, immunology, molecular biology, paleontology, etc...). AGW has at best 50-100 different data sets, from the dozen or so tree rings, to the dozen or so ice cores, satellite and surface temperature records. Much of our inference comes from computer simulations, that a very few completely understand. Much of global warming consensus comes from a small minority that are directly involved with the data or the simulations.
- We can control aspects of evolution. With knowledge of DNA, we can make genetically modified foods, we can change the course of diseases, and breed bacteria to eat nylon. Our understanding of AGW is at such a low level that we can only possibly control the climate at the grossest level. Our lack of understanding of feedback loops prevents even the most basic possible control of the system.
- Although evolution occurs on long time scales, we can see its action on the small scale. AGW also occurs on longish time scales, but there is no short-term equivalent. This adds to our level of control (with evolution), or lack of it with AGW.
- Those that are denying evolution want to replace it with something that violates not just evolution, but all of physics, chemistry, astronomy...pretty much all of science. Although the extremists in the anti-global warming camp can seem pretty anti-science, they aren't trying to replace global warming with something that violates all of science (they still might be wrong!). There is also a much more nuanced camp that admits that the planet is warmer, but perhaps it is not as special as the AGW theory would suggest, and that draconian CO2 policies are unwarranted given the uncertainties. This puts it on a very different scale than the anti-evolution group.
It is dangerous to make the comparison. One is partly the demonizing of your opponent and, at the same time, angelizing (is that a word? :) ) ones own perspective: by saying that the AGW deniers are just like the evolution deniers, both makes the deniers seem unreasonable, but by association, implies that AGW is as solid as evolution. This latter claim, despite the claims of its proponents, is definitely hyperbole.
Wednesday, December 9, 2009
The evidence for anthropogenic (that is, human-caused) global warming is strong, comes from many sources, and has been subject to much scientific scrutiny. Plenty of data are freely available. The basic principles can be understood by just about anyone, and first- and second-order calculations can be perfomed by any physics grad student. Given these facts, questioning the occurrence of anthropogenic global warming seems crazy. (Predicting the details is much, much more complicated). And yet, I have seen discussions, articles, and blog posts from smart, educated people who seem to think that anthropogenic climate change is somehow called into question by the facts that (1) some scientists really, deeply believe that global warming skeptics are wrong in their analyses and should be shut out of the scientific discussion of global warming, and (2) one scientist may have fiddled with some of the numbers in making one of his plots. This is enough to make you skeptical of the whole scientific basis of global warming? Really?
I would love to go point by point in this quote and show the calculations, and I'd imagine that it would get stymied once I tried to put in the water vapor feedback. I need to read more about this, because from what I've read we don't understand the magnitude, or sign, of the cloud feedback and that it could easily wipe out any warming caused by CO2 increases.
Some of the comments are very good too, like:
A. zarkov: I'm really disappointed to see you engage in the usual group think about global warming. Have you read the Wegman report? How come you don't refer people to ClimateAudit for the other side of the debate? Did you know that Michael Mann had to be forced by Congress to provide the data and codes behind the hockey stick calculation? ClimateAudit give you everything, the data and the R code they use. The other side stonewalls, and no wonder-- their results are a fraud.
The blog he refers to, climateaudit.org, is very interesting and is exactly the way the commenter says: they are all for open information. They post the data, the code, everything right up front and simply ask everyone else to do the same. Why this isn't required for all scientific publications, I don't know. Why it is not required for all high-stakes publications (ones that could result in very high-stakes policy) I don't know either. It's a travesty.
If everyone were as open about the data and the code, Climategate couldn't have happened.
One final comment on this thread:
Radford, Neal: Few people ever disputed that the current temperatures are higher than those of earlier times back to four hundred years ago. The big issue has always been whether the Medieval Warm Period (usually seen as occuring around a thousand years ago) was warmer than at present, since if it was, that makes the present warming seem not so unusal and perhaps due to natural causes.
This is my point too: if it was warmer 1000 years ago, then the hysterical language of the global warming media is completely unjustified.
Phil Jones in discussing the presentation of temperature reconstructions stated that “I’ve just completed Mike’s Nature trick of adding in the real temps to each series for the last 20 years (ie from 1981 onwards) and from 1961 for Keith’s to hide the decline.” The paper in question is the Mann, Bradley and Hughes (1998) Nature paper on the original multiproxy temperature reconstruction, and the ‘trick’ is just to plot the instrumental records along with reconstruction so that the context of the recent warming is clear. Scientists often use the term “trick” to refer to a “a good way to deal with a problem”, rather than something that is “secret”, and so there is nothing problematic in this at all. As for the ‘decline’, it is well known that Keith Briffa’s maximum latewood tree ring density proxy diverges from the temperature records after 1960 (this is more commonly known as the “divergence problem”–see e.g. the recent discussion in this paper) and has been discussed in the literature since Briffa et al in Nature in 1998 (Nature, 391, 678-682). Those authors have always recommend not using the post 1960 part of their reconstruction, and so while ‘hiding’ is probably a poor choice of words (since it is ‘hidden’ in plain sight), not using the data in the plot is completely appropriate, as is further research to understand why this happens.
A must read is the article by David Holland,
which outlines the problems with the hockey-stick analysis. He explains the divergence problem, and the
history of all this far better than I can summarize here.
Sunday, December 6, 2009
I am struck by a few things.
First, in the blog post he mentions:
It assumes that all things we call religion or religious impulses are essentially the same or have some common core. This faces the philosophical problem of properties and propositions in general. For example, take the property of redness. Is there something that all objects we call red have in common? And if there is, is this the same kind of thing we call religious belief?
In fact there is something common to all things red: the wavelengths of light that are absorbed. I think what he is asking is whether we experience red in the same way as our friend. In fact, it is quite likely, and it is not a philosophical idea at all. It seems to me more and more that philosophy tries to handle questions that are out of reach for science (for the moment) but the solutions found in philosophy evaporate or are insubstantial once we really understand what is going on. Tom Mitchell has done some very interesting work with looking at fMRI data in his "Brains, Meaning, and Corpus Statistics" talk (talk slides on his home page).
In the work, he compares fMRI data from different individuals, and finds that he can correctly identify images and words from brain activity of one person, using the associations between the images and words derived from the brain activity of other people. This strongly suggests that the internal representations of words and concepts may be very similar between individuals. Not only that, but that we have the possibility of determining what those are and not just leave it up to philosophical ruminations.
Dawkins mentions belief in authorities as a psychological tendency that may lead to religious thinking under the right circumstances. I would further add the brain's tendency for seeing patterns where there are none as the other piece of the religious-thinking puzzle. It is evolutionary advantageous to see tigers where there are none as opposed to not seeing tigers where there are some. Not all errors are equally costly. Religious interpretation of experience seems to me to easily follow from these sorts of errors.
Monday, November 30, 2009
Here's a little pet peeve of mine: nothing rhymes with orange. You've heard that before, right? Orange is famous for its rhymelessness. There's even a comic strip called "Rhymes with Orange." Fine then, let me ask you something. What the heck rhymes with purple?
If you stop and think about it, you'll find that English is jam-packed with rhymeless common words. What rhymes with empty, or olive, or silver, or circle? You can even find plenty of one-syllable words like wolf, bulb, and beige. Yet orange somehow became notorious for its rhymelessness, with the curious result that people now assume its status is unique.
I was directed to the quote by Andrew Gelman's Statistical Modeling blog, and he has other posts about names and sounds.
We are in the information age, and I'd love the way the information in the first two links is portrayed, and I plan on playing in the infochimps site more.
Wednesday, November 18, 2009
Anyway, this particular project takes famous scientists, and puts their speech to music. The music is catchy, and really captures well the philosophies of Carl Sagan and others. It just makes me realize how much I miss Carl Sagan, which then makes me miss Stephen Jay Gould and E.T. Jaynes.
Saturday, November 14, 2009
I think the best quote which summarizes the danger and the bad thinking is:
Major General Jehad al-Jabiri is head of the Ministry of the Interior’s General Directorate for Combating Explosives. “I don’t care about Sandia or the Department of Justice or any of them,” he says. “Whether it’s magic or scientific, what I care about is it detects bombs.”
Of course, if it hasn't been shown to detect bombs by science, the other option is irrelevant.
Monday, October 12, 2009
If it is true what the article says, that the BellKor team’s algorithms used information such as "genre" and "actors", then they plainly cheated!
This information was not available in the contest. The contest data had only 4 pieces of information: people ID, movie, date, and the rating values.
If BellKor managed to get extra information about the people and movies, it is hard to understand how NetFlix overlooked it. Therefore I'm quite sure that the article got the story wrong – it is too extreme to be believable. But if the article is right, then the winning team used crooked means to get the prize and its 'solution' does not worth much. Hope it is the first option…
I remember looking at the data, and if I recall correctly, I remember a script to add to the database from IMDB. I assume this was allowed, because otherwise there would have been no way to get 10%.
Saturday, October 10, 2009
I wrote as a comment to the person posting it:
Actually, this is not a true story, nor does it capture even in the slightest way the views of Einstein on religion (see http://www.einsteinandreligion.com/).
Although it is a cute video, there are logical flaws such as the fact that neither "hot" nor "cold" exists except as labels on "temperature", neither label with any special role.
I'm mostly disturbed by the fact that this Christian organization is trying to legitimize a flawed argument, by falsely attributing it to a very famous non-Christian, and nearly non-religious, scientist.
Wednesday, September 16, 2009
S.K. could identify some shapes (triangles, squares, etc.) when they were side-by-side, but not when they overlapped. His brain was unable to distinguish the outlines of a whole shape; instead, he believed that each fragment of a shape was its own whole. For S.K. and other patients like him, "it seems like the world has been broken into many different pieces," says Sinha.
However, if a square or triangle was put into motion, S.K. (and the other two patients) could much more easily identify it. (With motion, their success rates improved from close to zero to around 75 percent.) Furthermore, motility of objects greatly influenced the patients' ability to recognize them in images.
This is very easily interpreted using the HTM framework. It's very interesting!
from random import randint
your_first_answer=input('Which door %s? ' % str(door_choices))
else: # automatic
if prize==your_first_answer: # happens 1/3 of the time
door_choices.remove(your_first_answer) # get the other two
your_second_answer=input('Which door %s? ' % str(door_choices))
else: # automatic
# always switch
print "You win!"
print "You Lose!"
print "Winning percentage: ",float(win)/turn*100
Wednesday, September 9, 2009
Wednesday, September 2, 2009
Wednesday, August 12, 2009
One advantage to replacing mobile power (e.g. gasoline) with stationary power (e.g. electric company) is that one can replace the stationary power with nuclear, which you just can't do with cars.
It is claimed that it will be cheaper for end users, "In Detroit, where off-peak electricity rates are 5 cents a kilowatt hours, it will cost about 40 cents to recharge batteries overnight." One does have to factor in the maintenance costs, battery replacement, and battery disposal costs into the cost of owning this new car.
Now, I am not against this development, as such, but it is important (as always) to be aware of both the benefits and the costs of the new technology. It is far too easy to read one side of the equation, pat oneself on the back, while ignoring the hidden costs.
Thursday, July 30, 2009
Friday, July 24, 2009
another post here, although written in spanish, has a number of interesting LOGO (or python turtle) example exercises.
"For each day that the high temperature in your hometown is at least 1 degree Fahrenheit above average, as listed by Weather Underground, you owe me $25. For each day that it is at least 1 degree Fahrenheit below average, I owe you $25."
He's trying to address recent statements by some conservatives, paraphrased as "It's cold this summer here in Minneapolis, so global warming must be wrong." That's a bit of a strawman, but from the Power Line blog post, there really is this sense of local vs global perspective.
Well, it's actually a pretty silly challenge to a pretty silly statement. No serious GW skeptic I've heard contests that their is warming on a global scale, but argues against the magnitude or, more commonly, the cause of the warming (human vs not). The statistical challenge here only addresses whether there is warming, and even there is rigged to win even if there were no real global warming, because of the urban heat island effect. Most of the thermometers started out in rural areas, or in fields outside of towns, and cities were built around them. Areas around pavement are warmer than the surrounding areas, so there would be a measured warming trend due to development, not due to atmospherics.
A better bet would involve predicting the global temperature for, say, 5 years from now (along with the uncertainty). Each side puts in their prediction, and pays $1 times the ratio of the posterior probabilities for the two models, P(M1)/P(M2). Would anyone take a bet like that?
Monday, July 20, 2009
Because supporters and opponents tend to break down along partisan lines. Democrats favor sampling because the people who are traditionally hardest to count are the urban poor, minorities, and immigrants, all of whom tend to live in Democratic strongholds and vote Democratic. These groups are often undercounted because they move so frequently and do not trust government employees asking questions. Republicans, by contrast, stress that the Constitution specifies an “actual enumeration” of the population, not an estimate. They also argue that statistical sampling is inferior to counting. “Anyone familiar with public opinion polling can tell you that statistical sampling carries a margin of error,” Republican Reps. Darrell Issa and Patrick McHenry recently wrote. “And error is the enemy of a full and accurate census.”
The notion that a national count is completely error free is ridiculous. I think everyone would agree that if you do a count, that you will not get everyone. It is known that mistakes are made, omissions occur, and that some people actively avoid the census. Because the census avoidance is not random, the omissions are biased in some way. One can argue in which ways the bias points, but the bias is there.
So what is the best plan of action in this case? You want to make an estimate of the number of people in the country. "Estimate" is the correct word, even for an enumeration, given the fact that the enumeration is known to be incomplete. The best thing to do, then, is to have a public and open statistical model of the process of sampling, with independent ways to confirm the validity of the model. If the model is simple, and open, it would be difficult to argue against. Without this approach, a "strict enumeration" is really an unstated statistical model where the assumptions are very difficult to see.
Sunday, July 19, 2009
In a strict way, this is an agnostic perspective. The description of the universe, as described by Laplace, does not need to use the concept of God in any way. This does not disprove the existence of God, or even deny God's existence. It merely states that the concept of God is not needed. This is the pure vision of science, and why science does not necessarily conflict with religion. However, there could be certain claims from specific religions that conflict with science. The 6000 year old Earth, part of some fundamentalist Christian beliefs, is one example. The God as the mystery in the Universe is not something that can conflict with science.
Saturday, July 18, 2009
In Chapter 10 he quotes Laplace:
"We may regard the present state of the universe as the effect of its past and the cause of its future. An intellect which at a certain moment would know all forces that set nature in motion, and all positions of all items of which nature is composed, if this intellect were also vast enough to submit these data to analysis, it would embrace in a single formula the movements of the greatest bodies of the universe and those of the tiniest atom; for such an intellect nothing would be uncertain and the future just like the past would be present before its eyes."
and Mlodinow states that this is an expression of determinism. He then further states
But for Laplace's dream to hold true, several conditions must be met. First, the laws of nature must dictate a definite future, and we must know those laws. Second, we must have access to data that completely describe the system of interest, allowing no unforeseen influences. Finally, we must have sufficient intelligence or computing power to be able to decide what, given the data about the present, the laws say the future will hold.
He then criticizes it with the following three problems:
- society is not governed (as far as we know) by definite and fundamental laws in the way physics is
- like Lorenz, we cannot obtain the precise data necessary for making predictions
- human affairs are so complex that it is doubtful we'd be able to make the calculations anyway
He concludes "as a result, determinism is a poor model for the human experience." His point seems to be, in some ways, obvious and in other ways irrelevant.
Laplace was simply saying that "God" would not find anything random, because of complete knowledge. The connection between knowledge and inference, which probability theory affords, was worked out by Laplace in great detail and it known to use today as Bayesian inference. The structure of Bayesian inference describes randomness simply as the product of our ignorance of the model, the parameters, the initial conditions, the measurement details, etc... Laplace was simply saying that with perfect knowledge, there is no randomness. E.T. Jaynes would describe the "random process" as a mind-projection fallacy: you have ignorance of the system, so you attribute its unpredictable behavior as a product of the system itself. A rolled die is following Newton's Laws, deterministically, and detailed knowledge of the die and the roll and the surface should allow you to predict 100% of the time what it will do. We lack that knowledge, thus the behavior becomes unpredictable. We often then attribute that unpredictable behavior as a "random die", as if it were the die that contains the randomness and not our own ignorance.
Bringing in Lorenz, and chaos theory, is irrelevant here. Lorenz's systems were completely deterministic, and it is theoretically possible for a being to know the state of the system out to a sufficient number of decimal places to provide any particularly set level of uncertainty in the system. With the quantization of states, it then becomes possible to know *exactly* what state something is in. Of course, quantum mechanics is a two-edged sword in this example: it solves the chaos problem, but adds an inherent, physical, randomness to the system which is very peculiar.
The problem with Mlodinow, it seems, is that he hold human activity to be a bit too special. We are, after all, made up of atoms and would thus be governed by the laws of physics. Certainly it would be too complex to handle, for us, but Laplace was not talking about us in his quote, or at least not us right now or in the near future.
Friday, July 10, 2009
Wednesday, July 8, 2009
Imagine a woman named Linda, thirty-two years old, single, out-spoken, and very bright. In college she majored in philosophy. While a student she was deeply concerned with discrimination and social justice and participated in antinuclear demonstrations.
Linda is active in the feminist movement: 2.1Linda is a bank teller and is active in the feminist movement: 4.1Linda is a bank teller: 6.2
They presented a group of internists with a serious medical problem: a pulmonary embolism (a blood clot in the lung). If you have that ailment, you might display one or more of a set of symptoms. Some of those symptoms, such as partial paralysis, are uncommon; others, such as shortness of breath, are probable. Which is more likely: that the victim of an embolism will experience partial paralysis or that the victim will experience both partial paralysis and shortness of breath? Kahneman and Tversky found that 91 percent of the doctors believed a clot was less likely to cause just a rare symptom than it was to cause a combination of the rare symptom and a common one. (In the doctor's defense, patients don't walk into their offices and say things like "I have a blood clot in my lungs. Guess my symptoms."
- steak with no potatoes
- steak with potatoes
- clot with paralysis and shortness of breath
- clot with paralysis and no shortness of breath
Linda owns an IHOP franchiseLinda had a sex-change and is now LarryLinda had a sex-change and is now Larry and owns an IHOP franchise
- someone is claiming that the patient has an embolism
- the patient is claiming, or someone has measured, that she has partial paralysis
- the patient is claiming, or someone has measured, that she has shortness of breath
- they had the means to measure shortness breath in the patient, but there was none
- they did not have the means to measure shortness of breath
Tuesday, July 7, 2009
I've just finished the book "Euclid's Window" by Leonard Mlodinow, and really enjoyed it. The book describe the history of geometry from Euclid, Descartes, Gauss, and Einsten. During his coverage of Euclid he presents a simple proof of the Pythagorean Theorem that really resonated with me. I don't recall ever seeing a proof of it, or at least no memorable proof. This one uses a minimum of jargon and formality...you just draw the picture, discuss it for a bit, and you see it!
You start with a right triangle, like:
which, by eye, you can see that the total area of the square is the area of 4 triangles (just like our original) plus the area of the inner square, which is c*c (which reminds me that I have to figure out how to do superscripts and subscripts in this blog. :) )
The second construction is nearly the same as the first, and looks like:
which, again by eye (with a little shading to make it a bit more obvious), the total area of the square is the area of 4 triangles (just like our original) plus the area of the two inner squares, which are a*a and b*b. Therefore:
for any triangle for which you can make this construction, which are right triangles.