Hacker Newsnew | past | comments | ask | show | jobs | submit | teruakohatu's commentslogin

They will have a long time ago. By now Simon's meme will be well represented in training sets.

What is not generally understood is that these models don’t predict egg prices or inflation in Italy.

They decompose a time series into trends, seasonality and residuals. That’s what they are actually modelling.

They cannot predict wars in the Middle East influencing inflation unless there is a seasonal pattern(s).


> They cannot predict wars in the Middle East influencing inflation unless there is a seasonal pattern(s).

well...


Next you'll suggest something looney like a correlation with the 11-year solar cycle!

(for those who are lost: https://x.com/onionweigher/status/1936630237208469898)


The Middle East war season is upon us once again

Born too soon to deploy to the Middle East.

Born too late to deploy to the Middle East.

Born just in time to deploy to the Middle East.


That's what traditional time-series modelling does. This is a foundational model, which means it's just a neural network trained on lots of time series. (So maybe OP's question still stands? But it's the same question as "how can LLMs be good at so many different kinds of conversations?")

Because traditional time-series modelling (ARIMA, GARCH, ...) is too "simple" and "strict". Just like "simple" computer vision (OpenCV, edge-detection, ...) was crushed by neural networks when having to deal with real world images.

This seemed like a good answer at first. But on further thought, images on the whole really do seem to have quite a bit more standard structure / "grammar" to exploit compared to arbitrary time-series. Many images are of the world, where there is gravity so you might see preponderance of blobs at the bottom, or the repetitive types like people, animals, faces, eyes. Wildly abstract images still have some continuity, pixels in a neighborhood are likely to be similar.

Time series in general have none of this kind of structure that's strictly necessary. I'm sure that many real-world sensors typically have some gaussian distribution aspects + noise and/or smoothness and locality types of assumptions that are pretty safe, but presumably that simple stuff is exactly what traditional time-series modelling was exploiting.

Maybe the real question is just what kind of time-series are in the training data, and why do we think whatever implicit structure that is there actually generalizes? I mean, you can see how any training that mixes pictures of dogs and cats with picturing of people could maybe improve drawing hair, detecting hair, or let you draw people AND dogs. It's less clear to me how mixing sensor data / financial data / anything else together could be helpful.


> It's less clear to me how mixing sensor data / financial data / anything else together could be helpful.

Because many of these have the same underlying causal structures - humans doing things, weather correlations, holidays.

Well studied behavioral stuff like "the stock market takes the stairs up and the elevator down" which is not really captured by "traditional" modelling tools.

I'm sure people will be doing mechanical interpretation on these models to extract what they pattern match for prediction.


Personally, coming from an EE background and not finance or statistics, I would go about identifying these patterns with an Signals & Systems toolbox, like systems identification, various matched filters/classifiers.

This might be a totall wrong approach, but I think it might make sense to try to model a matched filter based on previous stock selloff/bullrun trigger events, and then see if the it has any predictive ability, likewise the market reaction seems to be usually some sort of delayed impulse-like activity, with the whales reacting quickly, and then a distribution of less savvy investors following up the signal with various delays.

I'm sure other smarter people have explored this approach much more in depth before me.


You're crafting features. The modern approach to ML (deep learning) is to use over-parameterized models and let them learn the features. Perhaps you remember this? https://www.nytimes.com/2012/06/26/technology/in-a-big-netwo...

Except that their success in the time series domain has been rather lackluster and elusive. It will s one of the few domains where old school models are not only less work to maintain but also more accurate. There are a few exceptions here and there. Every year there are a few neural nets based challengers. You can follow the M series of computations from its start to see this evolution.

Maybe because useful time-series modeling is usually really about causal modeling? My understanding is that mediated causality in particular is still very difficult, where adding extra hops in the middle takes CoT performance from like 90% to 10%.

Yes causal models are hard.

NNs do ok on those time series problems where it is really about learning a function directly off time. This is nonlinear regression where time is just another input variable.

Cases where one has to adjust for temporaly correlated errors, those seem to be harder for NNs. BTW I am talking about accuracies beyond what a typical RNN variants will achieve, which is pretty respectable. It's the case that more complicated DNNs don't seem to do much better inspite of their significant model complexity.


LightGBM won M5 and it wasn't even a competition.

The task was slightly different and favored GBMs. Note they aren't NNs whose underwhelming performance was what my comment was about.

The M series of competitions change the tasks every year to explore what models perform best under different scenarios. As I mentioned, neural network based models win here and there, but very spotty performance over all.


> Because many of these have the same underlying causal structures - humans doing things, weather correlations, holidays.

Or, you know, maybe they aren't. Thermometers and photon counts are related to weather sometimes, but not holidays. Holidays are related to traffic sensors and to markets, but not Geiger counters.

> Well studied behavioral stuff like "the stock market takes the stairs up and the elevator down" which is not really captured by "traditional" modelling tools.

Prices are the opposite, up like a shot during shocks, falling slowly like a feather. So that particular pattern seems like a great example of over-fitting danger and why you wouldn't expect mixing series of different types to be work very well.


Electricity demand is influenced very strongly by holidays, strongly by weather and from weak to strong by geopolitics (depending on location).

The model will have a library of patterns, and will be able to pattern match subtle ones to deduce "this time series has the kind of micro-patterns which appear in strongly weather influenced time-series", and use this to activate the weather pattern cluster.

To use your example, when served thermometer data, the model notices that the holiday pattern cluster doesn't activate/match at all, and will ignore it.

And then it makes sense to train it on the widest possible time series, so it can build a vast library of patterns and find correlations of activation between them.


Sometimes you want inductive bias. No universally true claim can be made like this.

Wars in the middle east seem to have increasingly regular patterns tied to stock market opening hours, unfortunately.

I totally agree with the sentiment but from what I can tell, I’d say they tend happen immediately before or after markets open and close. Essentially, and to their maximum, screwing absolutely everyone who isn’t in the clique from participating in the trade.

FWIW— the only sure fire way to win the trade is to buy time and assume both gross incompetence and negligence when it comes action. The only caveat is if the markets tank enough, this administration will signal capitulation before hand, e.g. Trump mildly capitulating on tariffs last April after the markets proceed to relentlessly defecate themselves.

0-DTE options are typically, and for good reason, stupid gambles. But, right now they can’t even be considered gambling, because there’s zero chance of winning. Not just bad odds, but no odds. Again just signaling how truly malicious this admin is and its disdain for anyone and everyone not close to them.


I mean it's super obvious, it's directly tied to scrubs popularity.

New season of scrubs = new war in the middle east.


Wow, I didn't know. Thank you! Such a great show.

It's suprisingly good, like it's it's 100% worth watching if you liked scrubs.

I am not familiar with time series models, but judging from your answer, it would be necessary to feed long time series into this model for it to detect trends. What is a token here? Can it, for the lack of a better example, take in all intraday movements of a stock for a day, a week, a month, etc?

I tend to avoid time series forecasting when I can help it because I find it hard to communicate to stakeholders that a neural network (or another method) is not an oracle.

If you are talking about granularity of observations, it would depend on what you are trying to predict (the price in an hour or the price in 12 months?) and how quickly you need the prediction (100ms? Tomorrow morning?). If I had infinite data I would use granularity as a hyper parameter and tune that to a level that produced the best test results.

I am for example currently using weekly averages for non-price data forecasting. I could use daily data but weekly is absolutely adequate for this purpose.


You can use lightgbm with appropriate feature engineering.

Using many different models, just not NN for this particular application.

What makes these models different from models used for e.g. audio?

Or other low-dimensional time domain signals?


You could abstract speech or other audio as a series of sounds, where time is indeed a factor. Speech, however, has patterns that are more similar to written language than to seasonal patterns that are typically assumed in time series. While trained on different data, the architecture of TimesFM is actually similar to LLMs. But not identical, as pointed out at https://research.google/blog/a-decoder-only-foundation-model...:

> Firstly, we need a multilayer perceptron block with residual connections to convert a patch of time-series into a token that can be input to the transformer layers along with positional encodings (PE).

> [...]

> Secondly, at the other end, an output token from the stacked transformer can be used to predict a longer length of subsequent time-points than the input patch length, i.e., the output patch length can be larger than the input patch length.


If "seasonal patterns" is the thing that differentiates between these two data sources, then perhaps time series models should be called seasonal models?

Do these models predict on just a single time series then?

it is far more useful for predictions to look for correlations between time series. This is far more complex than looking for correlations in general because most time series trend up or down and therefore correlate.


It is the Middle East. Wars are always in season. And supply is more than the demand.

The main issue is that people do use them to predict bitcoin prices intraday and that sort of things.

Is it an issue because it works, or because it doesn’t? Or because it’s bitcoin?

I genuinely want to know. Thank you


It is an issue because bitcoin is highly unpredictable.

These tools are good at predicting timeseries that are in fact quite predictable. Like insurances will use this to estimate the number of people who will die from cancer in the next year, the year after that, and so on up to 50 years in the future. The model will extrapolate the progresses made in cancer treatment from the current trend, etc. It is a prediction, cause it's still possible that a breakthrough comes in and suddenly people don't die from a certain form of cancer, but generally it should be roughly correct.

Bitcoin prices are a lot more chaotic, influenced by a ton of unrelated events that shape its path a certain way. There is absolutely no certainty that studying the shape of its past evolution will help in any way understand its future evolution.

Of course here I mean by studying its price alone. If you add more information, like who's behind each trend and why, you have a much better sense of what could happen next.


ARIMA and ARMA models

ar(k) stuff, sure. that's old news. i would expect the newfangled stuff to be good at 0-shot learning of pre-event signatures spread across multiple series, at a minimum.

> Brave Browser unable to block certain ads due to Rust language limitations

Deceptive HN title.

Someone whose GH account has since been deleted posted a comment on an issue a few months ago saying “the rust languages does not support lookaround regex”. Either a language barrier or a misunderstanding, since there is no fundamental issue with rust itself.

I flagged this submission.


Blocking Pinterest is not completely useless.


The top seven blocked domains on Kagi are all Pinterest: https://kagi.com/stats?stat=insights

Failing to block Pinterest and then also having no room for anything else is almost completely useless.


> “It was probably 40 degrees outside, but there’s a lot of heat going on in the back,”

Would this is safe to do on a sunny warm weather? Would body heat plus the sun ruin the cream?


> Would this is safe to do on a sunny warm weather? Would body heat plus the sun ruin the cream?

It's fairly safe. You can leave dairy products unrefrigerated for an uncomfortable amount of time :) Butter, in particular, can last for days outside a fridge.

The bacteria that tends to infest dairy products will usually (but not always) turn it into something tasty like yogurt.

Don't get me wrong, you can definitely get sick from spoiled dairy products, but it's not a 100% thing.


> Butter, in particular, can last for days outside a fridge.

I live in Ireland, and once we take butter out of the fridge (to replace the one that's now gone), it doesn't go back in, whatever the weather. All butter here is basically of Kerrygold quality (I'm talking real butter of course).


That's basically how we treated butter while I grew up. So long as it's salted, it rarely goes bad outside the fridge. We had a butter dish and that was about it. The cover keeps the butter from turning a darker yellow and drying out. But we'd still eat it even when that happened.

Gotta be honest, though, I'm not a fan of grassy dairy products :). I had dairy cows growing up and in the spring their milk definitely took on a distinct grassy flavor. I personally preferred it more when it was primarily hay flavored. Store milk tastes like basically nothing in particular.


Yes, also in Ireland and while I wouldn't leave homemade butter out for more than a day or two, Kerrygold salted will last two weeks at 19C without issue.


Yeah we keep butter in a butter dish in the cupboard, refill from the fridge as it is used up. I never knew this wasn’t what everyone did until my roommate in college was blown away about how good the butter was this way.


> This tracks with the science; according to Scientific American, room-temperature cream turns to butter much faster than cold cream because the molecules move more quickly at higher temperatures. Of course, if the temperature gets too high, everything will just melt, so their experiment probably wouldn’t have worked on a summer run.


Cold cream becomes whipped cream. colder yet icecream.


> Would this is safe to do on a sunny warm weather?

I’d be more concerned about having plastic bags against my skin when I’m sweating heavily than the effect the heat would have on the butter tbh. Hot weather is an excuse to wear less clothing, not wrap yourself in ziploc bags


On warmer days you could swap doing the washing for food preparation!


I assume they shamelessly were talking in Fahrenheit degrees.


> Yeah, but you're not normally going to random websites, downloading an exe and running it.

Most or many of secure mobile operating system zero days are caused by image parsing. There is a threat at least as big, if not bigger, in parsing complex file formats.

Sure gopher or gemini would be more secure, but even without JS the web ecosystem would be venerable.


> Do you think YC-backed companies would have respected this guideline if it was posted on some other website they wanted to operate in?

That is a false equivalence. What a YC-backed company does is not relevant to how a YC-owned web forum operates.


They're asking a question, not making an equivalence. And I'll add that YC founders/companies do have some specific advantages on this forum, so it's worth knowing if they are held to any standard.


Why?

If you have so little faith in them that they won’t honour the privacy controls you should also delete your non-consumer account too.


I was just wondering where our gypsum came from, I always assumed it was mined. Thank you for the information.


The best thing that can happen is a fork. There is no way this will make the app worse, at least if you use a phone to do the reviewing. Syncing is sure to become a subscription.


I still haven't quite understood how Anki can apparently be profitable when the web sync with Ankiweb apparently supports an unlimited amount of data (surely some people heavily abuse this - but one never hears about fair use being applied?). Even legit and honest use can mean massive collection sizes since you can have any rich media in an Anki collection.


A few things. Up until now Anki was not a business and was not meant to be profitable.

AnkiWeb also does not support an unlimited amount of data. It's evident that their storage requirements aren't infinite. They aggressively cull content. Not syncing for 6 months results in automatic deletion last I checked.


I would still expect ingress and egress costs to be significant - as well as storage for content that isn't culled. And while it was not 'meant to be profitable' my point is that it was profitable, apparently exclusively from revenue from the iOS app which I still find surprising that it would cover the Ankiweb costs.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: