<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Archives statistics - Open Forecasting</title>
	<atom:link href="https://openforecast.org/tag/statistics/feed/" rel="self" type="application/rss+xml" />
	<link>https://openforecast.org/tag/statistics/</link>
	<description>How to look into the future</description>
	<lastBuildDate>Sun, 22 Mar 2026 15:30:09 +0000</lastBuildDate>
	<language>en-GB</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2015/08/cropped-usd-05-32x32.png&amp;nocache=1</url>
	<title>Archives statistics - Open Forecasting</title>
	<link>https://openforecast.org/tag/statistics/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>The real Dunning-Kruger effect</title>
		<link>https://openforecast.org/2026/03/23/the-real-dunning-kruger-effect/</link>
					<comments>https://openforecast.org/2026/03/23/the-real-dunning-kruger-effect/#respond</comments>
		
		<dc:creator><![CDATA[Ivan Svetunkov]]></dc:creator>
		<pubDate>Mon, 23 Mar 2026 09:03:35 +0000</pubDate>
				<category><![CDATA[Social media]]></category>
		<category><![CDATA[statistics]]></category>
		<category><![CDATA[theory]]></category>
		<guid isPermaLink="false">https://openforecast.org/?p=4096</guid>

					<description><![CDATA[<p>Many of you have seen this image on the Internet — I&#8217;ve seen it myself a few times on LinkedIn lately. People say it depicts the &#8220;Dunning-Kruger&#8221; effect&#8230; But did you know this is actually an internet meme with little to do with the original paper? Here is one of the recent examples, a screenshot [&#8230;]</p>
<p>Message <a href="https://openforecast.org/2026/03/23/the-real-dunning-kruger-effect/">The real Dunning-Kruger effect</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>Many of you have seen this image on the Internet — I&#8217;ve seen it myself a few times on LinkedIn lately. People say it depicts the &#8220;Dunning-Kruger&#8221; effect&#8230; But did you know this is actually an internet meme with little to do with the original paper?</p>
<p>Here is one of the recent examples, a screenshot of <a href="https://www.linkedin.com/posts/fotios-petropoulos-04536023_dear-mr-i-reduce-forecast-error-by-30-share-7437246645530140672-NXnT">the post of Fotios Petropoulos</a> about the effect.</p>
<div id="attachment_4098" style="width: 282px" class="wp-caption aligncenter"><a href="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2026/03/2026-03-22-Dunning-Kruger-Petropoulos.png&amp;nocache=1"><img fetchpriority="high" decoding="async" aria-describedby="caption-attachment-4098" src="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2026/03/2026-03-22-Dunning-Kruger-Petropoulos-272x300.png&amp;nocache=1" alt="A LinkedIn post by Fotios Petropoulos" width="272" height="300" class="size-medium wp-image-4098" srcset="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2026/03/2026-03-22-Dunning-Kruger-Petropoulos-272x300.png&amp;nocache=1 272w, https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2026/03/2026-03-22-Dunning-Kruger-Petropoulos.png&amp;nocache=1 556w" sizes="(max-width: 272px) 100vw, 272px" /></a><p id="caption-attachment-4098" class="wp-caption-text">A LinkedIn post by Fotios Petropoulos</p></div>
<p>In the original paper, <a href="https://psycnet.apa.org/doi/10.1037/0022-3514.77.6.1121">Kruger and Dunning (1999)</a> ran experiments with undergraduates on humour, logical reasoning, and grammar. Participants completed a test and estimated their percentile rank. The authors then sorted participants into four quartiles by actual performance and computed averages for actual and self-assessed performance for each quartile. The plots in their paper &#8211; the real Dunning–Kruger effect &#8211; are just four data points per line, not a smooth curve over a learning journey (second image).</p>
<p>What did they find? People in the bottom quartile substantially overestimated their performance, often believing they were average or above. Top performers slightly underestimated their standing. The key finding is an asymmetry in miscalibration: low performers overestimate, high performers slightly underestimate.</p>
<p>This has almost nothing to do with the popular &#8220;experience vs. confidence&#8221; image. The original X‑axis is performance quartile at a single point in time; the meme&#8217;s X‑axis is a vague notion of &#8220;experience&#8221; through time. The original Y‑axis is the assessed test percentile; the meme&#8217;s is a free‑floating &#8220;confidence&#8221; construct. In the actual data, perceived performance increases with actual performance &#8211; there is no early spike, no &#8220;valley of despair,&#8221; no &#8220;slope of enlightenment.&#8221; That swooping curve is an internet-era graphic never reported by Kruger and Dunning, and it misleadingly frames the effect as a personal development trajectory the paper never studied.</p>
<p>There is also a serious critique of the original paper from statistical point of view. For example, <a href="https://doi.org/10.1016/j.intell.2020.101449">Gignac and Zajenkowski (2020)</a> showed that sorting people into quartiles and plotting average self-assessment against average performance can, by itself, generate the characteristic pattern &#8211; purely as a statistical artefact. In their own empirical data, miscalibration was roughly constant across ability levels, consistent with measurement noise rather than a special cognitive deficit in low performers. You can actually reproduce the pattern using two random uncorrelated variables. Here is a simple example in R:</p>
<pre class="decode">set.seed(41)

x <- rnorm(10000, 100, 10)
y <- rnorm(10000, 100, 10)
plot(x,y)
xQ <- quantile(x)
yQ <- quantile(y)

yMeans <- xMeans <- vector("numeric",4)

for(i in 1:4){
    xMeans[i] <- mean(x[x<xQ[i+1] &#038; x>xQ[i]])
    yMeans[i] <- mean(y[x<xQ[i+1] &#038; x>xQ[i]])
}

plot(1:4, xMeans, type="b", ylim=range(xMeans,yMeans),
     xlab="Real performance", ylab="Assessed performance",
     lwd=2)
lines(yMeans, lwd=2, lty=2)
points(yMeans, lwd=2)
legend("topleft",
       legend=c("Actual performance", "Assessed performance"),
       lwd=2, lty=c(1,2), pch=1)</pre>
<p>Which produces the image like this:</p>
<div id="attachment_4100" style="width: 310px" class="wp-caption aligncenter"><a href="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2026/03/2026-03-22-Dunning-Kruger-R.png&amp;nocache=1"><img decoding="async" aria-describedby="caption-attachment-4100" src="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2026/03/2026-03-22-Dunning-Kruger-R-300x175.png&amp;nocache=1" alt="Dunning-Kruger plot reproduction" width="300" height="175" class="size-medium wp-image-4100" srcset="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2026/03/2026-03-22-Dunning-Kruger-R-300x175.png&amp;nocache=1 300w, https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2026/03/2026-03-22-Dunning-Kruger-R-1024x597.png&amp;nocache=1 1024w, https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2026/03/2026-03-22-Dunning-Kruger-R-768x448.png&amp;nocache=1 768w, https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2026/03/2026-03-22-Dunning-Kruger-R.png&amp;nocache=1 1200w" sizes="(max-width: 300px) 100vw, 300px" /></a><p id="caption-attachment-4100" class="wp-caption-text">Dunning-Kruger plot reproduction</p></div>
<p>If you introduce a correlation between the two variables, the images starts looking even more similar to the ones from the original paper.</p>
<p>So there might be a real effect &#8211; many follow-up studies have measured it with more rigorous tools &#8211; but Dunning and Kruger&#8217;s method was not the right one to establish it. And that image with experience vs confidence is just a meme and a serious misconception that should not be used.</p>
<p>P.S. If you wonder who the &#8220;leading expert&#8221; that Fotios Petropoulos refers to in his post is &#8211; it&#8217;s me. Not sure why he doesn&#8217;t tag me properly.</p>
<p>Message <a href="https://openforecast.org/2026/03/23/the-real-dunning-kruger-effect/">The real Dunning-Kruger effect</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://openforecast.org/2026/03/23/the-real-dunning-kruger-effect/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>There&#8217;s no such thing as &#8220;deterministic forecast&#8221;</title>
		<link>https://openforecast.org/2026/03/02/there-s-no-such-thing-as-deterministic-forecast/</link>
					<comments>https://openforecast.org/2026/03/02/there-s-no-such-thing-as-deterministic-forecast/#respond</comments>
		
		<dc:creator><![CDATA[Ivan Svetunkov]]></dc:creator>
		<pubDate>Mon, 02 Mar 2026 22:45:31 +0000</pubDate>
				<category><![CDATA[Social media]]></category>
		<category><![CDATA[Theory of forecasting]]></category>
		<category><![CDATA[extrapolation methods]]></category>
		<category><![CDATA[statistics]]></category>
		<category><![CDATA[theory]]></category>
		<guid isPermaLink="false">https://openforecast.org/?p=4081</guid>

					<description><![CDATA[<p>Sometimes I see people referring to a &#8220;deterministic&#8221; forecast, and I have some personal issues with this. Because if you apply a model to data then there is nothing deterministic about your forecasts! In many contexts, &#8220;deterministic&#8221; has a precise meaning: no randomness, no uncertainty. A deterministic solution to an optimisation problem (e.g. linear programming) [&#8230;]</p>
<p>Message <a href="https://openforecast.org/2026/03/02/there-s-no-such-thing-as-deterministic-forecast/">There&#8217;s no such thing as &#8220;deterministic forecast&#8221;</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>Sometimes I see people referring to a &#8220;deterministic&#8221; forecast, and I have some personal issues with this. Because if you apply a model to data then there is nothing deterministic about your forecasts!</p>
<p>In many contexts, &#8220;deterministic&#8221; has a precise meaning: no randomness, no uncertainty. A deterministic solution to an optimisation problem (e.g. linear programming) implies that there are no random inputs or outputs once the model and its parameters are fixed. Forecasting is different. As <a href="https://onlinelibrary.wiley.com/doi/10.1002/(SICI)1099-131X(199612)15:7%3C495::AID-FOR640%3E3.0.CO;2-O">Chatfield</a> and many others have pointed out, forecasting has multiple sources of uncertainty, and there is essentially zero chance that the future will unfold exactly as any single number suggests.</p>
<p>Yes, some people use &#8220;deterministic&#8221; as a synonym for &#8220;point forecast&#8221;. But that label is still misleading, because a point forecast is not uncertainty-free &#8211; it is just one summary of a predictive distribution (often the conditional mean, sometimes the median or another functional).</p>
<p>Here’s a quick reality check you can do yourself. Take a dataset, apply your model, and write down the point forecast for the next few observations. Now add one new observation, re-estimate, and forecast again (the image in this post depicts exactly that, but with 50 forecasts produced on different subsamples of data). The point forecast will change unless you are dealing with an exotic situation with non-random data (e.g. every day, you sell exactly 100 units). So, which of the two was the &#8220;deterministic&#8221; forecast? If forecasts were truly deterministic in the strict sense, you would not get multiple plausible values from small, reasonable changes in the sample.</p>
<p>This happens because any forecasting method (statistical or ML) depends on data and on modelling choices: parameter estimation, feature selection, splitting rules, tuning, even decisions like &#8220;use α=0.1&#8221;. Those choices can be fixed across samples of data, but fixing them does not remove uncertainty &#8211; it only hides it. The randomness is still there in the data and in the fact that we only observe a sample of it.</p>
<p>So when you see someone mentioning &#8220;deterministic forecast&#8221;, it&#8217;s worth translating it mentally to: &#8220;a point forecast, probably a conditional mean&#8221;. If you care about decisions and risk, you should know that there is an uncertainty associated with this so called &#8220;deterministic forecast&#8221;, and that it should not be ignored. But this is a topic for another discussion in another post.</p>
<p>Message <a href="https://openforecast.org/2026/03/02/there-s-no-such-thing-as-deterministic-forecast/">There&#8217;s no such thing as &#8220;deterministic forecast&#8221;</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://openforecast.org/2026/03/02/there-s-no-such-thing-as-deterministic-forecast/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Teaching Statistics and Descriptive Analytics in the world of AI</title>
		<link>https://openforecast.org/2026/01/07/teaching-statistics-and-descriptive-analytics-in-the-world-of-ai/</link>
					<comments>https://openforecast.org/2026/01/07/teaching-statistics-and-descriptive-analytics-in-the-world-of-ai/#respond</comments>
		
		<dc:creator><![CDATA[Ivan Svetunkov]]></dc:creator>
		<pubDate>Wed, 07 Jan 2026 17:32:28 +0000</pubDate>
				<category><![CDATA[Analytics]]></category>
		<category><![CDATA[Statistics]]></category>
		<category><![CDATA[statistics]]></category>
		<category><![CDATA[teaching]]></category>
		<guid isPermaLink="false">https://openforecast.org/?p=3946</guid>

					<description><![CDATA[<p>Teaching statistics as a flipped classroom with the help of AI? You heard that right! That’s exactly what I tried this year &#8211; and here are the results. Attached to this post is the student evaluation score for the module. Yes, the number of responses is quite low (only 50% of the cohort), but it [&#8230;]</p>
<p>Message <a href="https://openforecast.org/2026/01/07/teaching-statistics-and-descriptive-analytics-in-the-world-of-ai/">Teaching Statistics and Descriptive Analytics in the world of AI</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>Teaching statistics as a flipped classroom with the help of AI? You heard that right! That’s exactly what I tried this year &#8211; and here are the results.</p>
<p>Attached to this post is the student evaluation score for the module. Yes, the number of responses is quite low (only 50% of the cohort), but it should still give a sense of how students perceived Statistics and Descriptive Analytics. Of course, this reflects only their impression &#8211; coursework submissions are yet to come &#8211; but it’s still an encouraging sign that some things worked well.</p>
<p>I’ve taught this module since 2018, first with Dave Worthington and later with Alisa Yusupova. Normally, I focused on the second half, covering regression through lectures and workshops. But this year, I took on the full module and realised I didn’t want to teach probability theory and statistics in the traditional way &#8211; long monologues in lectures followed by awkward silence in workshops. That format, I believe, no longer works. After all, students can always ask their favourite LLM to explain concepts they don&#8217;t understand. And some don&#8217;t even do that &#8211; they just ask to solve problems without understanding them. So, what can be done in this brave new world?</p>
<p>I don’t yet have a definitive answer &#8211; only the results of an experiment.</p>
<p>Lectures. This year, I used Google Notebook ML to prepare lecture materials. I provided my existing notes, slides, and relevant texts, then asked it to produce podcasts on specific topics. This took more time than expected, as I had to review the generated content, adjust prompts, and refine focus areas, listening to the podcasts over and over again. Once ready, I uploaded the materials on Moodle and asked students to listen beforehand. In class, we skipped formal lectures and instead had whiteboard &#038; marker discussions. I asked questions, showed derivations, and encouraged debate. With a class of 26 students, it was possible to create much more interaction than in previous years.</p>
<p>Workshops. We still had problem-solving sessions, but I allowed (actually encouraged) students to use LLMs to solve tasks and explain why the solutions were correct. The aim was to emphasise reasoning and assumptions over simply obtaining the right number. This worked with mixed success, and I still need to think how it can be improved further.</p>
<p>Did it work overall?</p>
<p>I&#8217;m not entirely sure. Not all students engaged with the materials in advance, but those who did seemed to benefit and appreciated the approach. What I do know is that the &#8220;two-hour monologue while everyone tries not to fall asleep&#8221; format does not work any more. For universities, and for the (very!) expensive UK education, to remain relevant, we must innovate and rethink how we teach.</p>
<p>What would you change if you were teaching a technical subject at university in the era of AI? I’d love to hear your ideas.</p>
<p>Message <a href="https://openforecast.org/2026/01/07/teaching-statistics-and-descriptive-analytics-in-the-world-of-ai/">Teaching Statistics and Descriptive Analytics in the world of AI</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://openforecast.org/2026/01/07/teaching-statistics-and-descriptive-analytics-in-the-world-of-ai/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>On randomness and uncertainty</title>
		<link>https://openforecast.org/2025/04/28/on-randomness-and-uncertainty/</link>
					<comments>https://openforecast.org/2025/04/28/on-randomness-and-uncertainty/#respond</comments>
		
		<dc:creator><![CDATA[Ivan Svetunkov]]></dc:creator>
		<pubDate>Mon, 28 Apr 2025 11:05:29 +0000</pubDate>
				<category><![CDATA[Social media]]></category>
		<category><![CDATA[Statistics]]></category>
		<category><![CDATA[Theory of forecasting]]></category>
		<category><![CDATA[statistics]]></category>
		<category><![CDATA[theory]]></category>
		<category><![CDATA[uncertainty]]></category>
		<guid isPermaLink="false">https://openforecast.org/?p=3828</guid>

					<description><![CDATA[<p>Everything is random! Your data, your model, its parameter estimates, the forecasts it produces, and even the minimum of the loss function you used. There is no such thing as a &#8220;deterministic&#8221; forecast &#8211; everything is stochastic! Whenever you work with data, you are working with a sample from a population. In some cases, this [&#8230;]</p>
<p>Message <a href="https://openforecast.org/2025/04/28/on-randomness-and-uncertainty/">On randomness and uncertainty</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>Everything is random! Your data, your model, its parameter estimates, the forecasts it produces, and even the minimum of the loss function you used. There is no such thing as a &#8220;deterministic&#8221; forecast &#8211; everything is stochastic!</p>
<p>Whenever you work with data, you are working with a sample from a population. In some cases, this is more apparent than in others. In my statistics lectures, I typically give the following example. Consider that we are interested in the average height of students at the university. I could ask every student at the lecture to tell me their height, take the average, and get a number. Is this number random? Yes, indeed. Why? Because if a student who was late for the lecture comes in, I would need to recalculate the average, and the number would change. The average that I get depends on who specifically I have in the sample and how many observations I have. It will vary more in smaller samples and become more stable in larger ones. But this example gives you an idea about the inherent uncertainty of any estimates we deal with.</p>
<p>In time series, the situation is somewhat similar: you are dealing with a sample of values that you have observed up until a specific moment. If, for example, you want to forecast daily admissions in the emergency department of a hospital and apply a model, its forecast will change when a new day comes and a new cohort of patients arrives. This is because your sample changes, and you receive new information about the demand.</p>
<p>So, the parameter estimates of a model you use will change when you get a new observation (e.g., a new record of product sales). Yes, if you estimate the model properly (e.g., using Least Squares), the parameter estimates won’t change substantially, but they will change nonetheless. And this would affect point forecasts and any other statistics produced by your model. Your standard errors, p-values, conditional means, prediction intervals, error measures, model ranking &#8211; everything will change with a new observation. In fact, if you do model selection, the structure of the model might change as well. For example, in the case of ETS, you might switch from a model without a trend to one with a trend. So, every time you estimate anything on a sample of data, you should keep in mind that it is random and will change if your sample changes or gets updated.</p>
<p>Why is that important? Because we need to understand this inherent uncertainty, and ideally, we should somehow take it into account. In forecasting, this means you should not draw conclusions based on one application of a model to a dataset. At the very least, you should perform <a href="/adam/rollingOrigin.html">a rolling origin evaluation</a>. As Leonidas Tsaprounis says, &#8220;if you don&#8217;t roll the origin, you roll the dice&#8221;.</p>
<p>So, embrace the uncertainty and learn how to deal with it.</p>
<p>By the way, Kandrika Pritularga and I are holding a course on Demand Forecasting starting on 6th May. There is still time to <a href="https://online-payments.lancaster-university.co.uk/product-catalogue/courses/lancaster-university-management-school-lums/centre-for-marketing-analytics-forecasting-cmaf/demand-forecasting-principles-with-examples-in-r">sign up for it here</a>.</p>
<p>Message <a href="https://openforecast.org/2025/04/28/on-randomness-and-uncertainty/">On randomness and uncertainty</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://openforecast.org/2025/04/28/on-randomness-and-uncertainty/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Model vs Method &#8211; why should we care?</title>
		<link>https://openforecast.org/2025/02/04/model-vs-method-why-should-we-care/</link>
					<comments>https://openforecast.org/2025/02/04/model-vs-method-why-should-we-care/#respond</comments>
		
		<dc:creator><![CDATA[Ivan Svetunkov]]></dc:creator>
		<pubDate>Tue, 04 Feb 2025 12:14:44 +0000</pubDate>
				<category><![CDATA[Social media]]></category>
		<category><![CDATA[Statistics]]></category>
		<category><![CDATA[Theory of forecasting]]></category>
		<category><![CDATA[statistics]]></category>
		<category><![CDATA[theory]]></category>
		<guid isPermaLink="false">https://openforecast.org/?p=3771</guid>

					<description><![CDATA[<p>Image above depicts a fashion model making a presentation about a forecasting method. I like the forecast for the final period in that image&#8230; Over the last few years, I’ve seen phrases like &#8220;LightGBM model&#8221; or &#8220;Neural Network model&#8221; on LinkedIn many times, and the statistician in me shivers every time. So, I figured it’s [&#8230;]</p>
<p>Message <a href="https://openforecast.org/2025/02/04/model-vs-method-why-should-we-care/">Model vs Method &#8211; why should we care?</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p><em>Image above depicts a fashion model making a presentation about a forecasting method. I like the forecast for the final period in that image&#8230;</em></p>
<p>Over the last few years, I’ve seen phrases like &#8220;LightGBM model&#8221; or &#8220;Neural Network model&#8221; on LinkedIn many times, and the statistician in me shivers every time. So, I figured it’s time to discuss the difference between a model and a method.</p>
<p>Some of you might remember that I wrote <a href="/2020/06/08/forecasting-method-vs-forecasting-model-what-s-difference/">a post on this topic</a> a few years ago. But it seems it is worth revisiting.</p>
<p>John Boylan and I came up with the following definitions in <a href="/2023/09/08/iets-state-space-model-for-intermittent-demand-forecasting/">our paper</a>:</p>
<ul>
<li>A forecasting model is a mathematical representation of a real phenomenon with a complete specification of distribution and parameters;</li>
<li>A forecasting method is a mathematical procedure that generates point and/or interval forecasts, with or without a forecasting model.</li>
</ul>
<p>If these sound too technical, here’s a simpler explanation:</p>
<ul>
<li>A forecasting method is a way of generating forecasts;</li>
<li>A forecasting model is a way to describe the assumed structure of a real phenomenon.</li>
</ul>
<p>The key difference? A method focuses on producing something specific (e.g., point forecasts) with minimal assumptions, while a model relies on assumptions but can do much more:</p>
<ol>
<li>Rigorous estimation. Models can be constructed in ways that ensure their estimates of parameters are efficient and consistent.</li>
<li>Model selection using information criteria. A powerful approach that saves computational time and typically produces reasonable forecasts.</li>
<li>Predictive distribution. Models can generate moments (mean, variance, skewness) and quantiles, capturing uncertainty around future values.</li>
<li>Confidence intervals for parameters. While not crucial for forecasting, this is useful in other areas to quantify uncertainty.</li>
<li>Extendibility. Additional variables and components can be easily incorporated in a model.</li>
</ol>
<p>All of this comes at a price of <a href="/2025/01/07/there-is-no-such-thing-as-assumption-free-approach/">making assumptions about the reality</a>. If the assumptions don’t hold, the model won’t perform well. It might still be useful, but the risk of error increases. For example, you can apply a Random Walk model to purely random data, but you shouldn’t expect it to work well.</p>
<h3>Examples</h3>
<ol>
<li>A forecasting method: Naïve, defined by the simple equation:<br />
\( F_t = A _{t-1} \)<br />
This method is easy to explain, hard to break, and provides point forecasts, but nothing more.</li>
<li>A forecasting model: Random Walk, which underlies the Naïve method:<br />
\( A_t = A_{t-1} + \epsilon_t \)<br />
where \( \epsilon_t \) follows some distribution with zero mean and fixed variance. The Random Walk model has all the properties described above.</li>
</ol>
<p>In some cases, you can derive models underlying the methods. In my opinion, this typically enhances the latter, making them more powerful due to the reasons explained above. What is interesting about this general connection is that if we can identify a model underlying a method, we can do much more with it.</p>
<p>For example, when estimating a quantile regression, we typically minimize a pinball loss function, which gives us a method for generating quantiles. However, if we estimate the same linear regression model using likelihood, assuming that the error term follows the <a href="https://doi.org/10.1093/biostatistics/kxj039">Asymmetric Laplace distribution</a>, we arrive at exactly the same parameter estimates as in quantile regression. But now, we also gain additional benefits, such as model selection, predictive distribution, and confidence intervals for parameters &#8211; features outlined in the previous post. In a way, these benefits come &#8220;for free&#8221;, although at the cost of making explicit assumptions about the model. That said, I’d argue that assumptions exist in quantile regression anyway &#8211; they’re just not stated explicitly.</p>
<p>And here we finally come to the ML approaches. According to the definitions we discussed earlier, Decision Trees, k-Nearest Neighbors, Artificial Neural Networks (ANNs) and other ML approaches are not forecasting models. They do not attempt to capture the underlying structure of the data. Instead, they focus on identifying nonlinear patterns via engineered features to produce point forecasts. In other words, they are methods, not models.</p>
<p>This doesn’t make them inferior. Their strength lies in their flexibility, precisely because they don&#8217;t impose strong assumptions. However, treating them as forecasting models can lead to potential issues.</p>
<p>For example, plugging LightGBM’s point forecasts into a probability distribution doesn’t magically turn it into a model. It simply makes it a method that now generates quantiles, but without a solid theoretical foundation for why a specific distribution is chosen or used in a particular way.</p>
<p>Another example is model selection using information criteria, which is meaningless for ML approaches. Why? Because information criteria rely on the assumption that the model is estimated in a specific way (e.g., via maximum likelihood estimation), ensuring parameter consistency and model identifiability. However, some ML methods, such as ANNs, are fundamentally unidentifiable, as different architectures can produce the same output. So, the information criteria become meaningless in this setting.</p>
<p>So next time you see the term model, take a moment to consider whether it’s used correctly and whether it actually means what the author thinks.</p>
<p>Message <a href="https://openforecast.org/2025/02/04/model-vs-method-why-should-we-care/">Model vs Method &#8211; why should we care?</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://openforecast.org/2025/02/04/model-vs-method-why-should-we-care/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>There is no such thing as an &#8220;assumption-free approach&#8221;</title>
		<link>https://openforecast.org/2025/01/07/there-is-no-such-thing-as-assumption-free-approach/</link>
					<comments>https://openforecast.org/2025/01/07/there-is-no-such-thing-as-assumption-free-approach/#respond</comments>
		
		<dc:creator><![CDATA[Ivan Svetunkov]]></dc:creator>
		<pubDate>Tue, 07 Jan 2025 10:50:55 +0000</pubDate>
				<category><![CDATA[Social media]]></category>
		<category><![CDATA[Statistics]]></category>
		<category><![CDATA[Theory of forecasting]]></category>
		<category><![CDATA[statistics]]></category>
		<category><![CDATA[theory]]></category>
		<guid isPermaLink="false">https://openforecast.org/?p=3761</guid>

					<description><![CDATA[<p>One thing that bothers me when I read posts on social media or papers in peer-reviewed journals is the claim that a proposed approach is &#8220;assumption-free.&#8221; In forecasting, this is never true. Such an approach is like a spherical unicorn in a vacuum (see image above). Here&#8217;s why. Every model is a simplification of reality, [&#8230;]</p>
<p>Message <a href="https://openforecast.org/2025/01/07/there-is-no-such-thing-as-assumption-free-approach/">There is no such thing as an &#8220;assumption-free approach&#8221;</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>One thing that bothers me when I read posts on social media or papers in peer-reviewed journals is the claim that a proposed approach is &#8220;assumption-free.&#8221; In forecasting, this is never true. Such an approach is like a spherical unicorn in a vacuum (see image above). Here&#8217;s why.</p>
<p>Every model is a simplification of reality, meaning that it captures only a part of it. Simplifying implies that certain aspects of reality are irrelevant and can be ignored. For example, in forecasting, regardless of the approach used, we typically assume that the model captures the structure correctly, i.e. neither omitting important elements nor overfitting the data. Different approaches address this differently: statistical models do that explicitly, while a good ML approach seeks a balance between underfitting and overfitting, often in a non-linear way. When the structure is captured correctly, the forecast reflects the essential part of reality while ignoring small random fluctuations (see a post on <a href="/2024/08/13/structure-vs-noise-a-fundamental-concept-in-forecasting/">structure vs. noise</a>).</p>
<p>Depending on the assumptions we make, we can classify approaches as parametric, semiparametric, or nonparametric.</p>
<p>Parametric approaches assume that the model is correctly specified, its parameters are accurately estimated, and the chosen distribution is appropriate (often the normal one, though others can be used). In this case, we fully rely on the model. A classical example is the construction of conventional prediction intervals: the conditional expectation and variance are calculated and plugged into the normal distribution to derive the necessary quantiles for a specified confidence level. Specifically, in this case we assume that the model is correct, errors are uncorrelated and homoscedastic, and that they follow a normal distribution.</p>
<p>Semiparametric approaches relax some of these assumptions. For example, we might calculate statistics in a more robust manner or drop the assumption of a specific distribution. For example, instead of relying on textbook formulae, we could use in-sample multistep forecast errors to calculate conditional variances. This eliminates the need to assume uncorrelated and homoscedastic errors and allows for some flexibility in the model structure. However, in this example, we still rely on normality.</p>
<p>Nonparametric approaches avoid most of the above assumptions but come with their own hidden ones. For instance, the method proposed by <a href="https://doi.org/10.1287/mnsc.45.2.225">Taylor &#038; Bunn (1999)</a> for constructing prediction intervals fits quantile regressions to in-sample multistep forecast errors. This method does not assume a correct model, well-behaved residuals, or normality. However, it does assume the appropriateness of the chosen quantile regression function (Spoiler: they used polynomial regression, but my experiments suggest that a power function is a more robust alternative).</p>
<p>You might think that nonparametric approaches, with fewer assumptions, should always be preferred. But that&#8217;s not necessarily the case. It is &#8220;horses for courses&#8221;: you should select the approach that best fits your specific situation. For example, when working with small samples, introducing some assumptions might be necessary to get meaningful estimates. A nonparametric approach, while powerful, might require more data than you have available.</p>
<p>Finally, there is no such thing as a &#8220;best&#8221; method for every situation. As is often the case in forecasting, you need to try different approaches and choose the one that works best. Even then, remember that forecasting always rests on a fundamental assumption: the future will resemble the past. And no fancy method can guarantee that this assumption will hold.</p>
<p>Message <a href="https://openforecast.org/2025/01/07/there-is-no-such-thing-as-assumption-free-approach/">There is no such thing as an &#8220;assumption-free approach&#8221;</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://openforecast.org/2025/01/07/there-is-no-such-thing-as-assumption-free-approach/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Structure vs. Noise: A Fundamental Concept in Forecasting</title>
		<link>https://openforecast.org/2024/08/13/structure-vs-noise-a-fundamental-concept-in-forecasting/</link>
					<comments>https://openforecast.org/2024/08/13/structure-vs-noise-a-fundamental-concept-in-forecasting/#respond</comments>
		
		<dc:creator><![CDATA[Ivan Svetunkov]]></dc:creator>
		<pubDate>Tue, 13 Aug 2024 13:06:01 +0000</pubDate>
				<category><![CDATA[Social media]]></category>
		<category><![CDATA[Statistics]]></category>
		<category><![CDATA[Theory of forecasting]]></category>
		<category><![CDATA[statistics]]></category>
		<category><![CDATA[theory]]></category>
		<guid isPermaLink="false">https://openforecast.org/?p=3647</guid>

					<description><![CDATA[<p>One of the core ideas in statistics, which extends to many other fields including forecasting, is the concept of structure versus noise. You’ve probably heard of it, but it’s often overlooked by those without a strong quantitative background. So, let&#8217;s discuss. The core of the idea is that any data consists of two fundamental parts: [&#8230;]</p>
<p>Message <a href="https://openforecast.org/2024/08/13/structure-vs-noise-a-fundamental-concept-in-forecasting/">Structure vs. Noise: A Fundamental Concept in Forecasting</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>One of the core ideas in statistics, which extends to many other fields including forecasting, is the concept of structure versus noise. You’ve probably heard of it, but it’s often overlooked by those without a strong quantitative background. So, let&#8217;s discuss.</p>
<p>The core of the idea is that any data consists of two fundamental parts:</p>
<ol>
<li>Structure, which can take various forms, and might include trend, seasonality, calendar effects, and the influence of external factors on demand (e.g., price changes, promotions etc).</li>
<li>Noise, which is inherently unpredictable.</li>
</ol>
<p>Structure can be captured using models or methods, and this is what produces the fitted values or point forecasts. Noise, on the other hand, is unpredictable &#8211; like not knowing exactly who will visit a store and when they’ll make a purchase.</p>
<p>For example, consider a local Lancaster pub that has a nice selection of beers. Their sales likely follow a pattern, such as higher sales on weekends or during special events like football matches. These patterns are the structure we can capture and forecast. However, the pub can&#8217;t anticipate when my friend Yves will visit me, and when we&#8217;ll go out for drinks. This element of uncertainty forms the noise &#8211; while it’s explainable from my perspective, it’s a mystery to the pub owner.</p>
<p>But as I said, the idea of structure vs. noise isn’t just relevant in demand forecasting; it applies in many other areas too. Take classification, for instance. When identifying mushrooms, you might not be able to tell for sure whether you’re looking at a Rosy Brittlegill or The Sickener without a microscope. While certain characteristics (like stem shape or cap colour) make up the structure, there’s always some randomness that can make one mushroom look like another. So, in classification, you can only say that it’s more likely that we have one type of mushroom rather than the other, and you need to consider the uncertainty around this choice (the modern approach to this is to use conformal prediction).</p>
<p>Furthermore, we as humans are very good at finding patterns in the noise. If you look at clouds and see a mushroom, it’s not a real mushroom, just a random arrangement of vapour. So when you work with the data, remember this feature and don’t fall into the trap of finding patterns that don’t actually exist. Be critical and avoid overfitting the noise.</p>
<p>As you can see, the concept of structure versus noise is fundamental and shows up in many contexts. In forecasting, our job all is to capture the structure somehow, filter out the noise so that we can then produce point forecasts (future structure) and prediction intervals (representing the size of uncertainty) to be able to make adequate decisions.</p>
<p>Message <a href="https://openforecast.org/2024/08/13/structure-vs-noise-a-fundamental-concept-in-forecasting/">Structure vs. Noise: A Fundamental Concept in Forecasting</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://openforecast.org/2024/08/13/structure-vs-noise-a-fundamental-concept-in-forecasting/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Complex-Valued Econometrics with Examples in R</title>
		<link>https://openforecast.org/2024/08/04/complex-valued-econometrics-with-examples-in-r/</link>
					<comments>https://openforecast.org/2024/08/04/complex-valued-econometrics-with-examples-in-r/#respond</comments>
		
		<dc:creator><![CDATA[Ivan Svetunkov]]></dc:creator>
		<pubDate>Sun, 04 Aug 2024 14:33:40 +0000</pubDate>
				<category><![CDATA[Complex-valued models]]></category>
		<category><![CDATA[Statistics]]></category>
		<category><![CDATA[complex variables]]></category>
		<category><![CDATA[multivariate models]]></category>
		<category><![CDATA[R]]></category>
		<category><![CDATA[statistics]]></category>
		<guid isPermaLink="false">https://openforecast.org/?p=3632</guid>

					<description><![CDATA[<p>Back in 2022, my father asked me to help him in amending and editing a monograph he wrote on the topic of &#8220;Complex-Valued Econometrics&#8221;. The original book focused on dynamic models, but after looking through the material and a thorough discussion, we decided to write something more fundamental. The monograph is based on the research [&#8230;]</p>
<p>Message <a href="https://openforecast.org/2024/08/04/complex-valued-econometrics-with-examples-in-r/">Complex-Valued Econometrics with Examples in R</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>Back in 2022, my father asked me to help him in amending and editing a monograph he wrote on the topic of &#8220;Complex-Valued Econometrics&#8221;. The original book focused on dynamic models, but after looking through the material and a thorough discussion, we decided to write something more fundamental. The monograph is based on the research he has done over the years, working in Saint Petersburg. I developed an R package called &#8220;<a href="https://github.com/config-i1/complex">complex</a>&#8221; to support the book and then expanded the text with some derivations and examples of application. The result was then submitted to Springer and is <a href="https://doi.org/10.1007/978-3-031-62608-1">now finally published</a> in their &#8220;Contributions to Economics&#8221; series. Unfortunately, due to the agreement with the publisher, we cannot make the book freely available, but some of related materials can be found on a github repo, <a href="https://github.com/config-i1/complex-econometrics">here</a>.</p>
<p>We will receive royalties from selling this book, and we have decided to direct them to a charity to help Ukrainians (<a href="https://www.justgiving.com/campaign/ukraine-aid-help-now">this one</a>).</p>
<p>And here is how the cover of the book looks like:<br />
<div id="attachment_3638" style="width: 209px" class="wp-caption aligncenter"><a href="https://openforecast.org/wp-content/uploads/2024/08/title.webp"><img decoding="async" aria-describedby="caption-attachment-3638" src="https://openforecast.org/wp-content/uploads/2024/08/title-199x300.webp" alt="Complex-Valued Econometrics with Examples in R" width="199" height="300" class="size-medium wp-image-3638" srcset="https://openforecast.org/wp-content/uploads/2024/08/title-199x300.webp 199w, https://openforecast.org/wp-content/uploads/2024/08/title-680x1024.webp 680w, https://openforecast.org/wp-content/uploads/2024/08/title-768x1157.webp 768w, https://openforecast.org/wp-content/uploads/2024/08/title.webp 827w" sizes="(max-width: 199px) 100vw, 199px" /></a><p id="caption-attachment-3638" class="wp-caption-text">Complex-Valued Econometrics with Examples in R</p></div>
<p><a href="https://doi.org/10.1007/978-3-031-62608-1">Svetunkov S., Svetunkov I. (2024). Complex-Valued Econometrics with Examples in R: Modelling, Regression and Applications. Springer Cham. 154 pages. DOI: 10.1007/978-3-031-62608-1</a></p>
<p>Message <a href="https://openforecast.org/2024/08/04/complex-valued-econometrics-with-examples-in-r/">Complex-Valued Econometrics with Examples in R</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://openforecast.org/2024/08/04/complex-valued-econometrics-with-examples-in-r/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Multistep loss functions: Geometric Trace MSE</title>
		<link>https://openforecast.org/2024/06/04/multistep-loss-functions-geometric-trace-mse/</link>
					<comments>https://openforecast.org/2024/06/04/multistep-loss-functions-geometric-trace-mse/#respond</comments>
		
		<dc:creator><![CDATA[Ivan Svetunkov]]></dc:creator>
		<pubDate>Tue, 04 Jun 2024 09:05:56 +0000</pubDate>
				<category><![CDATA[Social media]]></category>
		<category><![CDATA[Theory of forecasting]]></category>
		<category><![CDATA[Univariate models]]></category>
		<category><![CDATA[ARIMA]]></category>
		<category><![CDATA[estimators]]></category>
		<category><![CDATA[ETS]]></category>
		<category><![CDATA[extrapolation methods]]></category>
		<category><![CDATA[statistics]]></category>
		<guid isPermaLink="false">https://openforecast.org/?p=3594</guid>

					<description><![CDATA[<p>While there is a lot to say about multistep losses, I&#8217;ve decided to write the final post on one of them and leave the topic alone for a while. Here it goes. Last time, we discussed MSEh and TMSE, and I mentioned that both of them impose shrinkage and have some advantages and disadvantages. One [&#8230;]</p>
<p>Message <a href="https://openforecast.org/2024/06/04/multistep-loss-functions-geometric-trace-mse/">Multistep loss functions: Geometric Trace MSE</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>While there is a lot to say about multistep losses, I&#8217;ve decided to write the final post on one of them and leave the topic alone for a while. Here it goes.</p>
<p>Last time, we discussed <a href="/2024/05/25/recursive-vs-direct-forecasting-strategy/">MSEh</a> and <a href="/2024/06/01/multistep-loss-functions-trace-mse/">TMSE</a>, and I mentioned that both of them impose shrinkage and have some advantages and disadvantages. One of the main advantages of TMSE was in reducing computational time in comparison with MSEh: you just fit one model with it instead of doing it h times. However, the downside of TMSE is that it averages things out, and we end up with model parameters that minimize the h-steps-ahead forecast error to a much larger extent than those that are close to the one-step-ahead. For example, if the one-step-ahead MSE was 500, while the six-steps-ahead MSE was 3000, the impact of the latter in TMSE would be six times higher than that of the former, and the estimator would prioritize the minimization of the longer horizon one.</p>
<p>A more balanced version of this was introduced in <a href="/2023/08/09/multi-step-estimators-and-shrinkage-effect-in-time-series-models/">our paper</a> and was called &#8220;Geometric Trace MSE&#8221; (GTMSE). The main idea of GTMSE is to take the geometric mean or, equivalently, the sum of logarithms of MSEh instead of taking the arithmetic mean. Because of that, the impact of MSEh on the loss becomes comparable with the effect of MSE1, and the model performs well throughout the whole horizon from 1 to h. For the same example of MSEs as above, the logarithm of 500 is approximately 2.7, while the logarithm of 3000 is 3.5. The difference between the two is much smaller, reducing the impact of the long-term forecast uncertainty. As a result, GTMSE has the following features:</p>
<ul>
<li>It imposes shrinkage on models parameters.</li>
<li>The strength of shrinkage is proportional to the forecast horizon.</li>
<li>But it is much milder than in case of MSEh or TMSE.</li>
<li>It leads to more balanced forecasts, performing well on average across the whole horizon.</li>
</ul>
<p>In that paper, we did extensive simulations to see how different estimators behave, and we found that:</p>
<ol>
<li>If an analyst is interested in parameters of models, they should stick with the conventional loss functions (based on one-step-ahead forecast error) because the multistep ones tend to produce biased estimates of parameters.</li>
<li>On the other hand, multistep losses kick off the redundant parameters faster than the conventional one, so there might be a benefit in the case of overparameterized models.</li>
<li>At the same time, if forecasting is of the main interest, then multistep losses might bring benefits, especially on larger samples.</li>
</ol>
<div id="attachment_3595" style="width: 310px" class="wp-caption aligncenter"><a href="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-04-Multistep-Example.png&amp;nocache=1"><img loading="lazy" decoding="async" aria-describedby="caption-attachment-3595" src="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-04-Multistep-Example-300x200.png&amp;nocache=1" alt="ETS(A,A,A) estimated using different loss functions applied to the data with multiplicative seasonality" width="300" height="200" class="size-medium wp-image-3595" srcset="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-04-Multistep-Example-300x200.png&amp;nocache=1 300w, https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-04-Multistep-Example-1024x681.png&amp;nocache=1 1024w, https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-04-Multistep-Example-768x511.png&amp;nocache=1 768w, https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-04-Multistep-Example.png&amp;nocache=1 1049w" sizes="auto, (max-width: 300px) 100vw, 300px" /></a><p id="caption-attachment-3595" class="wp-caption-text">ETS(A,A,A) estimated using different loss functions applied to the data with multiplicative seasonality</p></div>
<p>The image above shows an example from our paper, where we applied the additive model to the data, which exhibits apparent multiplicative seasonality. Despite that, we can see that multistep losses did a much better job than the conventional MSE, compensating for the misspecification.</p>
<p>Message <a href="https://openforecast.org/2024/06/04/multistep-loss-functions-geometric-trace-mse/">Multistep loss functions: Geometric Trace MSE</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://openforecast.org/2024/06/04/multistep-loss-functions-geometric-trace-mse/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Multistep loss functions: Trace MSE</title>
		<link>https://openforecast.org/2024/06/01/multistep-loss-functions-trace-mse/</link>
					<comments>https://openforecast.org/2024/06/01/multistep-loss-functions-trace-mse/#respond</comments>
		
		<dc:creator><![CDATA[Ivan Svetunkov]]></dc:creator>
		<pubDate>Sat, 01 Jun 2024 11:29:12 +0000</pubDate>
				<category><![CDATA[Social media]]></category>
		<category><![CDATA[Theory of forecasting]]></category>
		<category><![CDATA[Univariate models]]></category>
		<category><![CDATA[estimators]]></category>
		<category><![CDATA[extrapolation methods]]></category>
		<category><![CDATA[statistics]]></category>
		<category><![CDATA[theory]]></category>
		<guid isPermaLink="false">https://openforecast.org/?p=3586</guid>

					<description><![CDATA[<p>As we discussed last time, there are two possible strategies in forecasting: recursive and direct. The latter aligns with the estimation of a model using a so-called multistep loss function, such as Mean Squared Error for h-steps-ahead forecast (MSEh). But this is not the only loss function that can be efficiently used for model estimation. [&#8230;]</p>
<p>Message <a href="https://openforecast.org/2024/06/01/multistep-loss-functions-trace-mse/">Multistep loss functions: Trace MSE</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>As we discussed <a href="/2024/05/25/recursive-vs-direct-forecasting-strategy/">last time</a>, there are two possible strategies in forecasting: recursive and direct. The latter aligns with the estimation of a model using a so-called multistep loss function, such as Mean Squared Error for h-steps-ahead forecast (MSEh). But this is not the only loss function that can be efficiently used for model estimation. Let&#8217;s discuss another popular option.</p>
<p>But before that, let&#8217;s take a step back to recap what we are talking about. All the multistep losses imply that we fit the model to the data in the conventional way and then produce recursively 1 to h-steps-ahead point forecasts from each in-sample observation, from the very first to the very last one. We can then calculate the forecast errors and collect them in a matrix with observations in rows and horizon in columns, as shown in the image below, generated using the <code>rmultistep()</code> function from the <code>smooth</code> package in R:</p>
<div id="attachment_3587" style="width: 310px" class="wp-caption aligncenter"><a href="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-01-multistep-matrix.png&amp;nocache=1"><img loading="lazy" decoding="async" aria-describedby="caption-attachment-3587" src="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-01-multistep-matrix-300x185.png&amp;nocache=1" alt="An example of a matrix of multistep forecast errors" width="300" height="185" class="size-medium wp-image-3587" srcset="https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-01-multistep-matrix-300x185.png&amp;nocache=1 300w, https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-01-multistep-matrix-768x475.png&amp;nocache=1 768w, https://openforecast.org/wp-content/webpc-passthru.php?src=https://openforecast.org/wp-content/uploads/2024/06/2024-06-01-multistep-matrix.png&amp;nocache=1 791w" sizes="auto, (max-width: 300px) 100vw, 300px" /></a><p id="caption-attachment-3587" class="wp-caption-text">An example of a matrix of multistep forecast errors</p></div>
<p>After that, we can calculate any of the multistep loss functions. MSEh, for example, would simply be the mean of squared errors in the last column of that matrix.</p>
<p>One of the most straightforward modifications of MSEh is a loss function that can be called &#8220;Trace MSE&#8221;, which is the sum of MSEs of each of the columns in that matrix. It has some advantages and disadvantages in comparison with MSEh. Here are some:</p>
<ul>
<li>Because we sum up MSEs for different horizons, those closer to h will tend to be higher than those close to 1, simply because typically, with an increase of the horizon, uncertainty increases as well.</li>
<li>The previous point means that the model estimated via TMSE will care less about short-term forecasts and will focus more on longer ones.</li>
<li>But at least it will not be as myopic as a model estimated with a specific MSEh.</li>
<li>You do not need to estimate h models; you can estimate just one, and it will be optimized for the entire horizon from 1 to h.</li>
<li>This means that you save on computations, making the estimation and forecasting roughly h times faster than in the case of MSEh.</li>
<li><a href="https://doi.org/10.1016/j.ijpe.2019.107597">Kourentzes et al. (2019)</a> showed that TMSE slightly outperformed MSE1 and MSEh. In fact, in one of the early versions of that paper, <a href="https://kourentzes.com/forecasting/2015/08/10/true-models-trace-optimisation-and-parameter-shrinkage-2/">Kourentzes &#038; Trapero</a> showed how well TMSE performs in the example of solar irradiation forecasting with ETS.</li>
<li>TMSE imposes shrinkage on parameters of dynamic models, which makes them less reactive and avoids overfitting.</li>
<li>But the shrinkage is not as strong as in the case of MSEh.</li>
</ul>
<p>This is discussed in <a href="/2023/08/09/multi-step-estimators-and-shrinkage-effect-in-time-series-models/">the paper</a> I wrote together with Nikolaos Kourentzes and Rebecca Killick</p>
<p>Some examples of application of TMSE are provided in <a href="/adam/multistepLosses.html">Section 11.3 of ADAM</a>.</p>
<p>Also, Peter Laurinec did an independent exploration of multistep losses and wrote <a href="https://petolau.github.io/Multistep-loss-optimized-forecasting-with-ADAM/">this nice post</a>.</p>
<p>Message <a href="https://openforecast.org/2024/06/01/multistep-loss-functions-trace-mse/">Multistep loss functions: Trace MSE</a> first appeared on <a href="https://openforecast.org">Open Forecasting</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://openforecast.org/2024/06/01/multistep-loss-functions-trace-mse/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
