Starting today August 7th, 2024, in order to post in the Married Couples, Courting Couples, or Singles forums, you will not be allowed to post if you have your Marital status designated as private. Announcements will be made in the respective forums as well but please note that if yours is currently listed as Private, you will need to submit a ticket in the Support Area to have yours changed.
Well, I think first you should remove the zero mode from the plot, because that mode means nothing. It's the ramp at the small-but-non-zero that is trend-like behavior. Here it looks so sharp that it's not clear by eye whether what we're seeing is mostly the zero mode or the small-but-nonzero ones.Just as a matter of interest here's some cyclical data with a clear secular trend:
and here's its spectral analysis:
So, impress me, MrMorton. Show me what that big peak on the left-hand side means?
Or more importantly show me how it cannot mean a secular trend.
Big mea culpa in order:
Now I noted that, indeed the Leroy (1998) 1-5 ranking scale includes only positive biases. However clearly negative biases are possible.
However:
Davey and Pielke suggested that poor sitinginduced bias could be positive or negative, (SOURCE)
The results presented here clearly support the theory that, if poor siting causes a bias, homogeneity adjustments account for the biases and contradict the hypothesis that poor current siting causes a warm bias or even any bias in the homogeneity-adjusted U.S. temperature change record. (ibid)
Emphasis added.
Actually I don't.
I mistakenly assumed the LeRoy metrics on siting were for both positive and negative biases. I am obviously mistaken since they only site location near heat sources. However I will also note that biases can be both positive and negative in sitings as pointed out in the Peterson paper.
Nice. Now's your chance to stop insulting me and "teach" a bit.
So your friend didn't think that a peak near zero on a periodogram could be an exceptionally long period trend? Since FT treats all trends as cyclical, what do you think the wavelength of something at that end of the spectrum is?
Now please explain to me how you would identify a secular trend in periodic data?
If you can, that is.
Did you show your "laughing friend" the data?
Just for an outsider's view, let's look at some temperature from Europe (SOURCE= ANALYSIS OF SECULAR TIME SERIES OF CLIMATOLOGICAL CHARACTERISTICS RNDr. Ladislav METELKA
Department of Meteorology and Environmental Protection
Faculty of Mathematics and Physics
Charles University
Prague
September 1997
(LINK)
Just as a matter of interest here's some cyclical data with a clear secular trend:
and here's its spectral analysis:
So, impress me, MrMorton. Show me what that big peak on the left-hand side means?
Or more importantly show me how it cannot mean a secular trend.
Or better yet have your "laughing friend" explain how he does time-series analyses such that secular trends cannot be modeled as just extremely long period cycles.
(Honestly, I'm curious because this all seems so common sense that when you find people who laugh at it, I have to learn the math behind it).[/qutoe]
It is a low frequency phenomenon as I said, but the phase plot is essential. And as I said, I have never met anyone who actually intuitively understands the phase spectrum. They always look like scatter-grams.
Oh and one other thing, in case you are interested in misrepresenting what I've asked or said here, I will remind you that some time series analyses treat secular trends as nothing more than just a very long period cycle, I suppose the key then is to limit drawing conclusions on a wavelength well outside of the sampling window and how you differentiate that from a pure secular trend.
But the big peak on the side is a DC shift above. That is just a constant number--a constant of integration added to the data.
So please, do teach. Don't just laugh and insult. Any fool can do that. It takes an real scientist to explain.
As long as you will actually stop acting like everything I say on topics I use professionally must automatically be wrong and stupid just because I reject GW. I find this attitude quite common among GW advocates. People usually have to be burned to start being more careful. I know I have been burned in GW debates when I wasn't careful, so, I have been exactly where you are. A few months ago I got my head handed to me in one of the debates. I went back and started examining everything I believed about GW was wrong. Those months greatly improved my arguments. Maybe this will do that for you.
(emphasis added)there are other potential sources of error, such as urban warming near meteorological stations, etc., many other methods have been used to verify the approximate magnitude of inferred global warming. These methods include inference of surface temperature change from vertical temperature profiles in the ground (bore holes) at many sites around the world, rate of glacier retreat at many locations, and studies by several groups of the effect of urban and other local human influences on the global temperature record. All of these yield consistent estimates of the approximate magnitude of global warming, which has now increased to about twice the magnitude that we reported in 1981. (SOURCE)
Based on data from Angell's global network of 63 radiosonde stations, over the period from 1958 through 2005, the global mean, near-surface air temperature warmed by approximately 0.17°C/decade,(SOURCE)
A warming signal has penetrated into the world's oceans over the past 40 years. The signal is complex, with a vertical structure that varies widely by ocean; it cannot be explained by natural internal climate variability or solar and volcanic forcing, but is well simulated by two anthropogenically forced climate models. We conclude that it is of human origin, a conclusion robust to observational sampling and model differences. (SOURCE)
(emphasis added)The data are displayed as a thick black line in the top left plot. The periodogram of the data is shown as dots in the top right
panel. Note the exceptionally high periodogram values at low frequencies. This comes from the trend in the data. Because
periodogram analysis explains everything in terms of waves, an upward trend shows up as a very long (low frequency) wave.(SOURCE)
Glenn,
I recommend your friend who "laughed out loud" at the interpretation of the time-series analysis learn some statistics.
He might wish to speak with SAS. Perhaps in your dabbling in statistics you are familiar with SAS? (In case he's not they are one of the leading software companies for statistical measurement and analysis in the world. They are one of the most widely used stats software providers and have published numerous books on statistical analysis. You can read about them here)
Here's some information from SAS
(emphasis added)
I would dearly love you to addres this point because my pride is being pricked quite heavily on this matter. You have now found people to "laugh out loud" at my statements as if I were some common crackpot or fool.
THE STATE OF THE DEBATE
1. Thaumaturgy has made a large error in that he failed to note that the Michel Leroy "Siting Index" only accounts for positive bias. Hence Glenn's criticism that there is a possibility of a significant positive bias to the stations so far sampled in the U.S. by surfacestations.org is correct.
2. Thaumaturgy is still correct that surfacestation.org has only sampled 43% of the U.S. national sitings. The maps from surfacestation.org and the methodology of casting a net for "volunteers" indicates that their sampling protocol is not random, therefore the fact that they have found about 69% of the currently non-random sampled sites have said potential for positive bias means little in the way of statistical value.
The map itself on surfacestations.org is clearly biased for those areas with higher populations and hence higher probability of finding an installation at or near a "bad siting" criterion per the Leroy (1998) standard.
3. Glenn has also introduced a "red herring" of sorts in trumpetting the fact that surfacestations.org has sampled "100% of the California Stations". Again, that does not mean anything statistically as the temperature doesn't care one way or another whether there is a "state line" arbitrarily drawn by the U.S. government (which itself may introduce a bias) or not.
That is 54 stations, of which 38 have Leroy rating of 4 or 5.
So, are we really to abandon the entire U.S. grid because in a single arbitrary box in one of the most populace states in the U.S. 38 stations were found to have a potential for positive bias?
Again, I cannot stress enough that I was in grave error in not noting that the Leroy system only accounts for the potential of positive bias.
However, this is hardly a deathblow for the entirety of the data that support global warming.
Now, the problem for Glenn's side of the debate comes forth that:
Global warming evidence is not solely based on U.S. or U.K. or Chinese surface temperature measurements. It is from a number of lines of evidence which correlate among each other to verify the general trend.
NASA:
(emphasis added)
Now, it is highly unlikely that a stray air-conditioner unit would affect borehole temperature measurements and it would surely take a very large parking lot to significantly melt a glacier.
But further we don't even have to stay on the ground. Weather balloon data supports global warming trends as well:
Let's not limit ourselves to land or sky, let's look at ocean data:
Now, I'll admit it's been about 17 years since I was on a research oceanographic cruise, but when I was out on the North Atlantic measuring gas exchange in sea water, I sure didn't see floating islands of air conditioner units or large parking structures. If I had you can be guaranteed I'd have been off that damn research ship in a second.
So, congrats, Glenn on pulling us off into anecdotal data by finding a likely non-random sampling of surface temperature sites with the potential for positive bias and dragging the conversation away from the mass of data that disconfirm your hypothesis.
It is still incumbent upon you to prove a statistically significant actual bias using proper statistics including true random (or near true random) sampling methodologies, and then explain how the various efforts by NASA, NOAA, and numerous international bodies to account for bias and error are somehow ineffective.
Then, please, feel free to get back to the little discussion of 95% Confidence Intervals vs standard deviation. (as you know sandard error of the mean is, by definition, smaller than the standard deviation unless you have only 1 data point. That's just the math:
(Now my math skills, having been maligned earlier may be in error on this but any time you divide a number by another number >1 means it gets smaller, so maybe you've found a way to get a standard error on the mean that is somehow larger than the standard deviation. I dunno.)
That is 54 stations, of which 38 have Leroy rating of 4 or 5.
No it is not random in the statistical sense, but Thaumaturgy continues to avoid the question I asked. 100% of the California stations have been surveyed, and 35% have a 2 deg bias and 35% have a 5 deg plus bias. Can't we agree that the California contribution to global warming is crap? Shouldn't it be excluded?
I have to say that, as a scientists, I am disgusted by this data. Especially since it is do to something as stupid as putting a thermometer in front of an air conditioning exhaust.
Glenn: I don't see the need to throw out all the CA data, however. What happens if the data with Leroy ratings of 4-5 are excluded?
Depends on what the original purpose of the thermometers was, does it not? "As a scientist" I often have to deal with data that is far from perfect, for example because it is collected by volunteers or because the data was never originally intended for the thing you want to use it for. This doesn't mean this data is useless, but it does mean you need to take into account it isn't perfect. From what I can gather, the data of the National Weather Services wasn't originally intended to specifically deal with long-term climate trends, nor do climate scientists have influence on how the data is gathered. If they don't own the stations, they can at best ask for suggestions to be taken into account. From what I can gather, the network is largely run by volunteers and the personnel involved is too low-staffed to check up on all sites. Next to that, a site may have been good to start with but deteriorated by local factors. Perhaps the thermometer was not placed next to the air vent but rather vice versa, for example?I have to say that, as a scientists, I am disgusted by this data. Especially since it is do to something as stupid as putting a thermometer in front of an air conditioning exhaust.
So, maybe it is time for you to stop being so [bless and do not curse][bless and do not curse][bless and do not curse][bless and do not curse] sure of yourself on issues you are new to and be willing to actually listen and learn rather than constantly assuming that anyone who doubts GW must be a nutter. How about that. Then I will try to teach a bit.
Because a peak at zero frequency is a constant shift. Don't confuse a ramp in frequency domain with a ramp in the time domain.
The data are displayed as a thick black line in the top left plot. The periodogram of the data is shown as dots in the top right
panel. Note the exceptionally high periodogram values at low frequencies. This comes from the trend in the data. Because
periodogram analysis explains everything in terms of waves, an upward trend shows up as a very long (low frequency) wave.(SOURCE)
I have clearly demonstrated that a high low frequency component doesn't have to mean a secular trend.
Somehow, I think you will continue along this thread on FFT insisting that the high amplitude in the power spectra means a secular trend.
While you may not respect my freind who laughed, he is one of the richest guys I know all based upon his ability to use mathematics.
Be mad at me if you will, but you have only yourself and your stubbornness on things about which you know little to blame.
Would you please answer my point where I calculated the standard deviations of the temperatures in small areas and all the standard deviations (the error) was bigger than the proclaimed amount of global warming?
Thaumaturgy said:In the earlier e-mail you stated that you parsed out the "linear" trend as a zero peak in the periodogram of the RESIDUALS of the linear fit to the data.
I'm a bit confused. To help me understand this a bit more I ginned up a fake data set that has one cyclical component (a sin wave) with and without an overlayed linear trend (Y trend) which was generated by taking the sin function and then adding on an (X-mean(X)) factor to give it a nice linear trend.
I ran a time series on both and saw that nice big spike at zero for the time series data on Y-Trend.
Am I correct in the statement:
Linear trends in time-series data are often represented by a peak in the frequency periodogram at zero
Or am I missing something altogether here?
(Also, the residuals of the linear fit of the Y-Trend data shown here plot with the same sine wave frequency as the original data set, which is what I'd expect).
PhD Statistician said:Yes, that is correct.
Depends on what the original purpose of the thermometers was, does it not? "As a scientist" I often have to deal with data that is far from perfect, for example because it is collected by volunteers or because the data was never originally intended for the thing you want to use it for. This doesn't mean this data is useless, but it does mean you need to take into account it isn't perfect. From what I can gather, the data of the National Weather Services wasn't originally intended to specifically deal with long-term climate trends, nor do climate scientists have influence on how the data is gathered. If they don't own the stations, they can at best ask for suggestions to be taken into account. From what I can gather, the network is largely run by volunteers and the personnel involved is too low-staffed to check up on all sites. Next to that, a site may have been good to start with but deteriorated by local factors. Perhaps the thermometer was not placed next to the air vent but rather vice versa, for example?
The US Climate Reference Network has been specifically designed to measure long-term climate trends, but this network has only be operational since 2001, I don't know whether these have been used for comparison purposes already.
This does not make the data à priori useless, it just means that you need to take into account the error they can give. For example by giving less weight to data or by correcting the data of the worse stations with the help of the better stations before aggregating the data. As far as I understand, the latter is being done with the national weather stations. If you can estimate the error and direction of a measurement, you can use the measurement if you take this into account. So rather than rejecting the stations out of hand, you need to study the whole path from observation, via correction to eventual conclusions. Whatever has been discussed so far, the method of correction for differences between stations has been virtually ignored so far. Which to me, "as an epidmiologist" is "disgusting".
"As a scientist", I'm wondering what kind of scientist you are. Not meant denigrating, but I've often noticed differences between different "kinds" of scientists, often stemming from a lack of insight in the different problems different fields have. Watch a discussion between an epidemiologist and a toxicologist for hilarious effect. The difference between an epidemiologist who often has to use an estimate of exposure and effect over large groups to make statements for whole populations to a toxicologist who knows exactly how much of a certain substance and which effect he measured on his very low number of (human or animal) guinea pigs is extremely interesting.
What is wrong with this data? The linear regression shows a 3 degree increase in mean annual temperature over the time period measured. The cyclic nature of the data is apparent, but no one is arguing with that. We are talking about a Global Warming trend, not a warming for one station, or one state, or even one country. I applaud your efforts at routing out bad data, Glenn, but perhaps you are not seeing the forest because you are looking too hard at the individual trees.I am downloading some of the stations I don't have and one of th e 'good' stations is Susanville CA. It is a class 1 station. Yep, a class 1 station. I don't know what the heck was going on in Susanville, but this is the temperature record that we are going to use to determine the global warming. I think it is utter incompetence. I know Thaumaturgy has doubted that, but if this is a good station..., lordy lordy!
Very good.After I posted the above, I downloaded Electra. It is even funnier, it is a class 2 station. I will post on it tonight. Split Rock, I think I can already answer your question, but I will go through the exercise to determine the proper answer to this question. I wouldn't want Thaumaturgy to accuse me of ignoring statistics.
IIRC, but this is from studying the temperature data long ago for a different project (correlating temperature changes with health outcomes) the data is normally used more for weather information. Specifically onset of seasons etc for farmers. Applications where a crude measure of temperature suffices.OK, Tom, if these stations were not setup for the studies they are being used for (or for accurate air temperature measurements in general), I can understand why the data is as error-prone as it seems to be.
Or statistician heavenThat makes it a statistical "dog's breakfast."
It is an option. But with removing data you substitute one problem with another, because you now have missing data. And missing data, if not randomly distributed, can lead to biased results. So if you are removing data from your dataset, you still need to make an estimate of the effects this creates. My experience is that it is often better to have a number where you know the error, than a missing value. At least in health research it is becoming more and more common to put an estimate of a value in the place of a missing value instead of leaving the value open. This can reduce bias in your data set.Nevertheless, I would be more inclined to remove such data from the analysis, rather than try to correct it after the fact. Maybe there are issues with doing this?
I see what you mean. I've worked with large datasets in the past, but never with trend analysis. And I really don't know enough about the data here to make accurate statements one way or the other, which is why I mostly read at this point.I am a plant physiologist by training, and an agricultural physiologist by experience. And no, I don't normally deal with large-scale, global projects like this one. That is why I am letting others here lead the discussion.
We use cookies and similar technologies for the following purposes:
Do you accept cookies and these technologies?
We use cookies and similar technologies for the following purposes:
Do you accept cookies and these technologies?