Hacker News new | past | comments | ask | show | jobs | submit login

there's actually some work on training transformer models on time series data which is quite interesting (for prediction purposes)

see google TimesFM: https://github.com/google-research/timesfm

what i mean i guess is llms can -reason- linguistically about time manipulating language, but can't really experience it. a bit like physics. thats why they do bad on exercises/questions about physics/logic that their training corpus might not have seen.






Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: