A comprehensive survey on the time series papers from 2018-2022 (we will update it in time ASAP!) on the top conferences (NeurIPS, ICML, ICLR, SIGKDD, SIGIR, AAAI, IJCAI, WWW, CIKM, ICDM, WSDM, etc.)
We divided these papers into several fundamental tasks as follows.
- Up-to-date papers
- Summarize the contributions in papers
- Present the datasets used in papers
- [2022-05-31] Add papers published in ICML 2022
- [2022-05-31] Add papers published in NeurIPS, ICML, ICLR, SIGKDD, SIGIR, AAAI, IJCAI 2019!
- [2022-05-05] Add papers published in WWW 2022!
- [2022-04-25] TS-Paper v1.0 is released! We support the published time series papers from 2020 to 2022. Stay tuned!
- Add papers published in 2018. (v3.0)
Paper | Conference | Year | Code | Key Contribution |
---|---|---|---|---|
Transformers in Time Series: A Survey | - | 2022 | link | 1. This work summarizes the network structures from adaptations and modification. 2. This work categorizes these methods into three tasks, e.g., forecasting, anomaly detection, and classification. |
Time series data augmentation for deep learning: a survey | IJCAI | 2021 | - | 1. This work systematically reviews and empirically compares different data augmentation methods for time series. 2. They discuss and highlight five future directions to provide useful research guidance. |
Neural temporal point processes: a review | IJCAI | 2021 | - | 1. They focus on important design choices and general principles for defining neural TPP models. 2. They provide an overview of common application areas. 3. They conclude many open challenges and important directions. |
Time-series forecasting with deep learning: a survey | Philosophical Transactions of the Royal Society A | 2021 | - | 1. They survey common encoder and decoder designs used in both one-step-ahead and multi-horizon time series forecastingβ describing how temporal information is incorporated into predictions by each model. 2. They highlight recent developments in hybrid deep learning models, which combine well-studied statistical models with neural network components to improve pure methods in either category. 3. They outline some ways in which deep learning can also facilitate decision support with time series data. |
Deep learning for time series forecasting: a survey | Big Data | 2021 | - | 1. They formulate the time series forecasting problem along with its mathematical fundamentals. 2. They discuss the advantages and limitations in the feed forward networks, recurrent neural networks (including Elman, long-short term memory, gated recurrent units, and bidirectional networks), and convolutional neural networks. |
DL-Traff: Survey and Benchmark of Deep Learning Models for Urban Traffic Prediction | CIKM | 2021 | graph-data, grid-data | They synthetically review the deep traffic models and the widely used datasets, then build a standard benchmark to comprehensively evaluate their performances with the same settings and metrics. |
Graph Neural Network for Traffic Forecasting: A Survey | - | 2021 | - | The first comprehensive survey that explores the application of graph neural networks for traffic forecasting problems (e.g. road traffic flow and speed forecasting, passenger flow forecasting in urban rail transit systems, and demand forecasting in ride-hailing platforms). |
Deep learning for anomaly detection in time-series data: review, analysis, and guidelines | Access | 2021 | - | 1. This review provides a background on anomaly detection in time-series data and reviews the latest applications in the real world. 2. They comparatively analyze state-of-the-art deep-anomaly-detection models with several benchmark datasets. 3. They offer guidelines for appropriate model selection and training strategy for deep learning-based time series anomaly detection. |
A review on outlier/anomaly detection in time series data | ACM Computing Surveys | 2021 | - | 1. This review provides a structured and comprehensive state-of-the-art on outlier detection techniques in the context of time series. 2. a taxonomy is presented based on the main aspects that characterize an outlier detection technique. |
A unifying review of deep and shallow anomaly detection | Proceedings of the IEEE | 2021 | - | 1. This work draws connections between classic βshallowβ and novel deep approaches and show how this relation might cross-fertilize or extend both directions. 2. They outline some critical open challenges. |
Big Data for Traffic Estimation and Prediction: A Survey of Data and Tools | Applied System Innovation 5 | 2021 | - | This study presents an up-to-date survey of open data and big data tools used for traffic estimation and prediction. |
Fusion in stock market prediction: A decade survey on the necessity, recent developments, and potential future directions | Information Fusion | 2021 | - | 1. Survey information, feature, model fusion from 2011β2020. 2. Discuss their limitations and future directions are explored for various stock applications. |
Applications of deep learning in stock market prediction: Recent progress | ESA | 2021 | - | 1. Give a latest review of recent works on deep learning models for stock market prediction. 2. Discuss data sources, models, metrics and the implementation and reproducibility. |
Deep Learning for Spatio-Temporal Data Mining: A Survey | KDD | 2020 | - | 1. A review of recent progress in applying deep learning techniques for STDM. 2. They classify existing literature based on the types of spatio-temporal data, the data mining tasks, and the deep learning models. |
Urban flow prediction from spatiotemporal data using machine learning: A surveyΒ | Information Fusion | 2020 | - | 1. Urban flow prediction from spatiotemporal data. 2. methods based on machine learning. 3. The difficulties and some ideas. |
An empirical survey of data augmentation for time series classification with neural networks | - | 2020 | link | 1. a taxonomy and outline the four families in time series data augmentation, including transformation-based methods, pattern mixing, generative models, and decomposition methods. 2. empirically evaluate 12 time series data augmentation methods on 128 time series classification datasets with six different types of neural networks. 3. analyze the characteristics, advantages and disadvantages, and recommendations of each data augmentation method. |
Deep Learning on Traffic Prediction: Methods, Analysis and Future Directions | - | 2020 | - | 1. They summarize the existing traffic prediction methods, widely used public datasets, give an evaluation and analysis by conducting extensive experiments to compare the performance of different methods on a real-world public dataset. |
Neural forecasting: Introduction and literature overview | - | 2020 | - | An introduction and an overview of some of the advances of neural networks in machine learning. |
Financial time series forecasting with deep learning : A systematic literature review: 2005β2019 | ASC | 2019 | - | 1. categorized the studies according to the intended forecasting implementation areas, such as index, forex, commodity forecasting. 2. grouped them based on DL model choices, such as CNNs, Deep Belief Networks (DBNs), Long-Short Term Memory (LSTM). |
Deep learning for time series classification: a review | Data Mining and Knowledge Discovery | 2019 | link | They implemented existing approaches by training 8,730 deep learning models on 97 time series datasets. |
Financial time series forecasting with deep learning : A systematic literature review: 2005β2019 | ASC | 2019 | - | 1. They categorized the studies according to their intended forecasting implementation areas, such as index, forex, commodity forecasting. 2. They grouped DL model, such as Convolutional Neural Networks (CNNs), Deep Belief Networks (DBNs), Long-Short Term Memory (LSTM). |
Natural language based financial forecasting: a survey | Artificial Intelligence Review | 2018 | - | They show scopes, progress and hotspots in natural language based financial forecasting (NLFF). |
Paper | Conference | Year | Code | Used Datasets | Key Contribution |
---|---|---|---|---|---|
FEDformer: Frequency Enhanced Decomposed Transformer for Long-term Series Forecasting | ICML | 2022 | code | ETT, Electricity, Exchange, Weather, ILI | We propose to combine Transformer with the seasonal-trend decomposition method, in which the decomposition method captures the global profile of time series while Transformers capture more detailed structures. The proposed method, termed as Frequency Enhanced Decomposed Transformer (FEDformer), is more efficient than standard Transformer with a linear complexity to the sequence length. |
TACTiS: Transformer-Attentional Copulas for Time Series | ICML | 2022 | code | electricity, fred-md, kdd-cup, solar-10min, traffic | We propose a versatile method, based on the transformer architecture, that estimates joint distributions using an attentionbased decoder that provably learns to mimic the properties of non-parametric copulas. |
Domain Adaptation for Time Series Forecasting via Attention Sharing | ICML | 2022 | code | UCI, Wiki | we propose a novel domain adaptation framework, Domain Adaptation Forecaster (DAF). DAF leverages statistical strengths from a relevant domain with abundant data samples (source) to improve the performance on the domain of interest with limited data (target). |
Volatility Based Kernels and Moving Average Means for Accurate Forecasting with Gaussian Processes | ICML | 2022 | code | we take inspiration from well studied domains to introduce a new class of models, Volt and Magpie, that significantly outperform baselines in stock and wind speed forecasting, and naturally extend to the multitask setting. | |
DSTAGNN: Dynamic Spatial-Temporal Aware Graph Neural Network for Traffic Flow Forecasting | ICML | 2022 | code | PEMS | This paper proposes a novel Dynamic Spatial-Temporal Aware Graph Neural Network (DSTAGNN) to model the complex spatial-temporal interaction in road network. |
Multi-Granularity Residual Learning with Confidence Estimation for Time Series Prediction | WWW | 2022 | Code | Electricity, Stock | we design a novel residual learning net to model the prior knowledge of the fine-grained dataβs distribution through the coarse-grained one. Furthermore, to alleviate the side effect of validity dif- ferences, we introduce a self-supervised objective for confidence estimation, which delivers more effective optimization without the requirement of additional annotation efforts. |
CAMul: Calibrated and Accurate Multi-view Time-Series Forecasting | WWW | 2022 | Code | google-symptoms, covid19, power, tweet | We propose a general probabilistic multi-view forecasting framework CAMul, which can learn representations and uncertainty from diverse data sources. It integrates the information and uncertainty from each data view in a dynamic context-specific manner, assign- ing more importance to useful views to model a well-calibrated forecast distribution. |
EXIT: Extrapolation and Interpolation-based Neural Controlled Differential Equations for Time-series Classification and Forecasting | WWW | 2022 | Code | MuJoCo, Google Stock | we propose to i) generate another latent continuous path using an encoder-decoder archi- tecture, which corresponds to the interpolation process of NCDEs, i.e., our neural network-based interpolation vs. the existing explicit interpolation, and ii) exploit the generative characteristic of the de- coder, i.e., extrapolation beyond the time domain of original data if needed. |
RETE: Retrieval-Enhanced Temporal Event Forecasting on Unified Query Product Evolutionary Graph | WWW | 2022 | - | Yelp, E-commerce | RETE efficiently and dynamically retrieves relevant entities centrally on each user as high-quality subgraphs, preventing the noise propagation from the densely evolutionary graph structures that incorporate abun- dant search queries. |
CATN: Cross Attentive Tree-aware Network for Multivariate Time Series Forecasting | AAAI | 2022 | - | Traffic, Electricity, PeMSD7(M), METR-LA | studied the hierarchical and grouped correlation mining problem of multivariate time-series data and proposed CATN for multi-step forecasting. |
Reinforcement Learning based Dynamic Model Combination for Time Series Forecasting | AAAI | 2022 | - | DATA | a novel and practically effective online ensemble aggregation framework for time-series forecasting that employs a deep reinforcement learning approach as a meta-learning technique. |
Conditional Local Convolution for Spatio-temporal Meteorological Forecasting | AAAI | 2022 | Code link | WeatherBench (Rasp et al. 2020) | a local conditional convolution to capture and imitate the meteorological flows of local patterns on the whole sphere |
TLogic: Temporal Logical Rules for Explainable Link Forecasting on Temporal Knowledge Graphs | AAAI | 2022 | Code link | Integrated Cri- sis Early Warning System, Split method | the first symbolic framework that directly learns temporal logical rules from temporal knowl- edge graphs and applies these rules for link forecasting |
Spatio-Temporal Recurrent Networks for Event-Based Optical Flow Estimation | AAAI | 2022 | - | The MVSEC dataset (Zhu et al. 2018a) | novel input representation to effectively extract the spatio-temporal information from event input. |
A GNN-RNN Approach for Harnessing Geospatial and Temporal Information: Application to Crop Yield Prediction | AAAI | 2022 | - | Crop | a novel GNN-RNN framework to innovatively incorporate both geospatial and temporal knowledge into crop yield prediction. |
ST-GSP: Spatial-Temporal Global Semantic Representation Learning for Urban Flow Prediction | WSDM | 2022 | Code link | TaxiBJ, BikeNYC | our model explicitly models the correlation among temporal dependencies of different scales to extract global temporal dependencies + new simple fusion strategy + self-supervised learning |
PYRAFORMER: LOW-COMPLEXITY PYRAMIDAL ATTENTION FOR LONG-RANGE TIME SERIES MODELING AND FORECASTING | ICLR | 2022 | Code link | Electricity, Wind, ETT data and App Flow | a novel model based on pyramidal attention that can effectively describe both short and long temporal dependencies with low time and space complexity. |
DEPTS: DEEP EXPANSION LEARNING FOR PERIODIC TIME SERIES FORECASTING | ICLR | 2022 | Code link | ELECTRICITY, TRAFFIC2, and M4(HOURLY) | model complicated periodic dependencies and to capture sophisticated compositions of diversified periods simultaneously. |
TAMP-S2GCNETS: COUPLING TIME-AWARE MULTIPERSISTENCE KNOWLEDGE REPRESENTATION WITH SPATIO-SUPRA GRAPH CONVOLUTIONAL NETWORKS FOR TIME-SERIES FORECASTING | ICLR | 2022 | Code link | PeMSD3, PeMSD4, PeMSD8 and COVID-19 | The developed TAMP-S2GCNets model is shown to yield highly competitive forecasting performance on a wide range of datasets, with much lower computational costs. |
CoST: Contrastive Learning of Disentangled Seasonal-Trend Representations for Time Series Forecasting | ICLR | 2022 | Code link | ETT,Electricity,Weather | proposed CoST, a contrastive learning framework that learns disentangled seasonal-trend representations for time series forecasting tasks. |
REVERSIBLE INSTANCE NORMALIZATION FOR ACCURATE TIME-SERIES FORECASTING AGAINST DISTRIBUTION SHIFT | ICLR | 2022 | - | ETT, Electricity Consuming Load (ECL) | address the distribution shift problem in time series, proposing a simple yet ef- fective normalization-and-denormalization method, reversible instance normalization (RevIN) |
TEMPORAL ALIGNMENT PREDICTION FOR SUPERVISED REPRESENTATION LEARNING AND FEW-SHOT SEQUENCE CLASSIFICATION | ICLR | 2022 | Code link | MSR Action3D, MSR Daily Activity3D, βSpoken Arabic Digits (SAD)β dataset, ChaLearn | present TAP, which is a learnable distance for sequences. |
Deep Switching Auto-Regressive Factorization: Application to Time Series Forecasting | AAAI | 2021 | - | Pacific Ocean Temperature Dataset, Parking Birmingham Data Set, ........ | it parameterizes the weights in terms of a deep switching vector auto-regressive likelihood governed with a Markovian prior |
Dynamic Gaussian Mixture Based Deep Generative Model for Robust Forecasting on Sparse Multivariate Time Series | AAAI | 2021 | - | USHCN, KDD-CUP, MIMIC-III | provides a novel and general solution that explicitly defines temporal dependency between Gaussian mixture distributions at different time steps |
Temporal Latent Autoencoder: A Method for Probabilistic Multivariate Time Series Forecasting | AAAI | 2021 | - | Traffic, Electricity, Wiki | introduced a novel temporal latent auto-encoder method which enables nonlinear factorization of multivariate time series, learned end-to-end with a temporal deep learning latent space forecast model. By imposing a probabilistic latent space model, complex distributions of the input series are modeled via the decoder. |
Synergetic Learning of Heterogeneous Temporal Sequences for Multi-Horizon Probabilistic Forecasting | AAAI | 2021 | - | Electricity(UCI), Traffic, Environment(Li, L.;Yan,J.;Yang,X.;and Jin,Y.2019a.) | presented a novel approach based on the deep conditional generative model to jointly learn from heterogeneous temporal sequences. |
Time-Series Event Prediction with Evolutionary State Graph | WSDM | 2021 | Code link | DJIA30, WebTraffic, NetFlow, ClockErr, and AbServe | proposed a novel represen- tation, the evolutionary state graph, to present the time-varying re- lations among time-series states. |
Long Horizon Forecasting With Temporal Point Processes | WSDM | 2021 | Code link | Election, Taxi, Traffic-911, and EMS-911. | a novel MTPP model specif- ically designed for long-term forecasting of events. |
Modeling Inter-station Relationships with Attentive Temporal Graph Convolutional Network for Air Quality Prediction | WSDM | 2021 | - | Beijing, Tianjin and POIs data | encode multiple types of inter-station relationships into graphs and design parallel GCNbased encoding and decoding modules to aggregate features from related stations using different graphs. |
Predicting Crowd Flows via Pyramid Dilated Deeper Spatial-temporal Network | WSDM | 2021 | - | Wi-Fi connection log, bike in New York city and taxi ride in New York | ConvLSTM + pyramid dilated residual network + integrated attention |
Z-GCNETs: Time Zigzags at Graph Convolutional Networks for Time Series Forecasting | ICML | 2021 | Code link | Decentraland, Bytom, PeMSD4 and PeMSD8. | The new Z-GCNETs layer allows us to track the salient timeaware topological characterizations of the data persisting over time. |
Explaining Time Series Predictions with Dynamic Masks | ICML | 2021 | Code link | MIMIC-III | These masks are endowed with an insightful information theoretic interpretation and offer a neat improvement in terms of performance. |
End-to-End Learning of Coherent Probabilistic Forecasts for Hierarchical Time Series | ICML | 2021 | Code link | Labour, Traffic, Tourism, Tourism-L, and Wiki | a single, global model that does not require any adjustments to produce coherent, probabilistic forecasts, a first of its kind. |
Autoregressive Denoising Diffusion Models for Multivariate Probabilistic Time Series Forecasting | ICML | 2021 | Code link | Exchange, Solar and Electricity, Traffic, Taxi and Wikipedia | a combination of improved variance schedule and an L1 loss to allow sampling with fewer steps at the cost of a small reduction in quality if such a trade-off is required. |
Conformal prediction interval for dynamic time-series | ICML | 2021 | Code link | solar and wind energy data | present a predictive inference method for dynamic time-series. |
RNN with Particle Flow for Probabilistic Spatio-temporal Forecasting | ICML | 2021 | - | PeMSD3, PeMSD4, PeMSD7 and PeMSD8 | propose a state-space probabilistic model- ing framework for multivariate time-series prediction that can process information provided in the form of a graph that specifies (probable) predictive or causal relationships. |
ST-Norm: Spatial and Temporal Normalization for Multi-variate Time Series Forecasting | KDD | 2021 | Code link | BikeNYC, PeMSD7 and Electricity | propose two kinds of normalization modules -- temporal and spatial normalization -- which separately refine the high-frequency component and the local component underlying the raw data. |
MiniRocket: A Fast (Almost) Deterministic Transform for Time Series Classification | KDD | 2021 | Code link | UCR archive | reformulate Rocket into a new method, MiniRocket. MiniRocket is up to 75 times faster than Rocket on larger datasets, and almost deterministic. |
Dynamic and Multi-faceted Spatio-temporal Deep Learning for Traffic Speed Forecasting | KDD | 2021 | Code link | PeMSD4, PeMSD8 and England | design a dynamic graph construction method to learn the time-specific spatial dependencies of road segments. |
Forecasting Interaction Order on Temporal Graphs | KDD | 2021 | Code link | COLLEGEMSG, EMAIL-EU and FBWALL | devise an attention mechanism to aggregate neighborhoods' information based on their representations and time encodings attached to their specific edges. |
Quantifying Uncertainty in Deep Spatiotemporal Forecasting | KDD | 2021 | - | air quality PM2.5, road network traffic, and COVID-19 incident deaths | conduct benchmark studies on uncertainty quantification in deep spatiotemporal forecasting from both Bayesian and frequen- tist perspectives. |
Spatial-Temporal Graph ODE Networks for Traffic Flow Forecasting | KDD | 2021 | Code link | - | we capture spatial-temporal dynamics through a tensor-based ordinary differential equation (ODE), as a result, deeper networks can be constructed and spatial-temporal features are utilized synchronously. |
A PLAN for Tackling the Locust Crisis in East Africa: Harnessing Spatiotemporal Deep Models for Locust Movement Forecasting | KDD | 2021 | Code link | - | PLAN's novel spatio-temporal deep learning architecture enables representing PlantVillage's crowdsourced locust observation data using novel image-based feature representations, and its design is informed by several unique insights about this problem domain. |
Topological Attention for Time Series Forecasting | NeurIPS | 2021 | Code link | M4 competition dataset | propose topological attention, which allows attending to local topological features within a time horizon of historical data. |
MixSeq: Connecting Macroscopic Time Series Forecasting with Microscopic Time Series Data | NeurIPS | 2021 | - | Rossmann, Wiki and M5 | an end2end mixture model to cluster microscopic time series, where all the components come from a family of Seq2seq models parameterized by differ- ent parameters. |
Test-time Collective Prediction | NeurIPS | 2021 | - | Boston...... | Our approach takes inspiration from the literature in social science on human consensus-making. |
Bubblewrap: Online tiling and real-time flow prediction on neural manifolds | NeurIPS | 2021 | Code link | - | we propose a method that combines fast, stable dimensionality reduction with a soft tiling of the resulting neural manifold, allowing dynamics to be approximated as a probability flow between tiles. |
Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting | NeurIPS | 2021 | Code link | ETT, Electricity, Exchange, and Traffic | we propose Auto- former as a novel decomposition architecture with an Auto-Correlation mechanism. |
Learning to Learn the Future: Modeling Concept Drifts in Time Series Prediction | CIKM | 2021 | - | Climate Dataset, Stock Dataset and Synthetic Dataset | propose a novel framework called learning to learn the future. Specifically, we develop a learning method to model the concept drift during the inference stage, which can help the model generalize well in the future. |
AdaRNN: Adaptive Learning and Forecasting of Time Series | CIKM | 2021 | - | UCI activity, Air quality, Electric power and Stock price | AdaRNN is a general framework with flexible distribution distances integrated. |
Actionable Insights in Urban Multivariate Time-series | CIKM | 2021 | - | Gaussian, Insect, Wikipedia and so on ...... | introduce and formalize a novel problem RaTSS that aims to find such time-series (rationalizations), which are actionable for the segmentation. We also propose an algorithm Find-RaTSS to find them for any black-box segmentation. |
Historical Inertia: A Neglected but Powerful Baseline for Long Sequence Time-series Forecasting | CIKM | 2021 | - | ETT, Electricity | introduce a new baseline for LSTF, the historical inertia (HI), which refers to the most recent historical data-points in the input time series. |
AGCNT: Adaptive Graph Convolutional Network for Transformer-based Long Sequence Time-Series Forecasting | CIKM | 2021 | - | ETT | a probsparse adaptive graph self-attention+the stacked encoder with distilling probsparse graph self-attention integrates the graph attention mechanism+ the stacked decoder with generative inference generates all prediction values in one forward operation |
PIETS: Parallelised Irregularity Encoders for Forecasting with Heterogeneous Time-Series | ICDM | 2021 | - | Covid-19 | design a novel architecture, PIETS, to model heterogeneous time-series. |
Attentive Neural Controlled Differential Equations for Time-series Classification and Forecasting | ICDM | 2021 | Code link | Character Trajectories, PhysioNet Sepsis and Google Stock. | present Attentive Neural Controlled Differential Equations (ANCDEs) for time-series classification and forecasting, where dual NCDEs are used: one for generating attention values, and the other for evolving hidden vectors for a downstream machine learning task. |
SSDNet: State Space Decomposition Neural Network for Time Series Forecasting | ICDM | 2021 | - | Electricity,Exchange, Solar ...... | SSDNet combines the Transformer architecture with state space models to provide probabilistic and interpretable forecasts, including trend and seasonality components and previous time steps important for the prediction. |
Two Birds with One Stone: Series Saliency for Accurate and Interpretable Multivariate Time Series Forecasting | IJCAI | 2021 | - | electricity, Air-quality, Industry data | Series saliency is model agnostic and performs as an adap- tive data augmentation method for training deep models. Moreover, by slightly changing the objec- tive, we optimize series saliency to find a mask for interpretable forecasting in both feature and time dimensions. |
TE-ESN: Time Encoding Echo State Network for Prediction Based on Irregularly Sampled Time Series Data | IJCAI | 2021 | - | MG system, SILSO, USHCN, COVID-19 | propose a novel Time En- coding (TE) mechanism. TE can embed the time information as time vectors in the complex domain. |
DeepFEC: Energy Consumption Prediction under Real-World Driving Conditions for Smart Cities | WWW | 2021 | Code link | SPMD, VED | presents a novel framework that identifies vehicle/driving environment-dependent factors to predict energy consumption over a road network based on his- torical consumption data for different vehicle types. |
HINTS: Citation Time Series Prediction for New Publications viaDynamic Heterogeneous Information Network Embedding | WWW | 2021 | - | the AMiner Computer Science dataset and the American Physical Society (APS) Physics dataset | a novel end-to-end deep learning framework that converts citation signals from dynamic heterogeneous information networks (DHIN) into citation time series. |
Variable Interval Time Sequence Modeling for Career Trajectory Prediction: Deep Collaborative Perspective | WWW | 2021 | - | traffic data from 1988.1 to 2018.11 | propose a unified time-aware career trajectory prediction framework, namely TACTP, which is capable of jointly providing the above three abilities for better understanding the career trajectories of talents. |
REST: Reciprocal Framework for Spatiotemporal coupled predictions | WWW | 2021 | - | a traffic dataset released by Li et al. and a web dataset | come up with a novel Reciprocal SpatioTemporal (REST) framework, which introduces Edge Inference Networks (EINs) to couple with GCNs. |
AutoSTG: Neural Architecture Search for Predictions of Spatio-Temporal Graph | WWW | 2021 | Code link | PEMS-BAY and METR-LA | propose a novel framework, entitled AutoSTG, for automated spatio-temporal graph prediction. In our AutoSTG, spatial graph convolution and temporal convolution operations are adopted in our search space to capture complex spatio-temporal correlations. Besides, we employ the meta learning technique to learn the adjacency matrices of spatial graph convolution layers and kernels of temporal convolution layers from the meta knowledge of the attributed graph. |
Fine-grained Urban Flow Prediction | WWW | 2021 | - | TaxiBJ+, HappyValley | Spatio-Temporal Relation Net- work (STRN) to predict fine-grained urban flows. First, a backbone network is used to learn high-level representations for each cell. Second, we present a Global Relation Module (GloNet) that cap- tures global spatial dependencies much more efficiently compared to existing methods. Third, we design a Meta Learner that takes external factors and land functions (e.g., POI density) as inputs to produce meta knowledge and boost model performances. |
Probabilistic Time Series Forecasting with Shape and Temporal Diversity | NeurIPS | 2020 | Code link | - | Diversity is controlled via two proposed differentiable positive semi-definite kernels for shape and time and exploits a forecasting model with a disentangled latent space. |
Benchmarking Deep Learning Interpretability in Time Series Predictions | NeurIPS | 2020 | Code link | - | a comprehensive synthetic benchmark where positions of informative features are known. |
Adversarial Sparse Transformer for Time Series Forecasting | NeurIPS | 2020 | - | electricity,traffic, wind, solar, M4-Hourly | By adversarial learning, we improve the contiguous and fidelity at the sequence level. We further propose Sparse Transformer to improve the ability to pay more attention on relevant steps in time series. |
Deep Rao-Blackwellised Particle Filters for Time Series Forecasting | NeurIPS | 2020 | - | electricity, traffic, solar, exchange, wiki | proposed an extension of the classical SGLS that addresses two weaknesses |
Gamma-Models: Generative Temporal Difference Learning for Infinite-Horizon Prediction | NeurIPS | 2020 | - | - | introduced a new class of predictive model, a R-model, that is a hybrid between standard model-free and model-based mechanisms |
EvolveGraph: Multi-Agent Trajectory Prediction with Dynamic Relational Reasoning | NeurIPS | 2020 | - | Honda 3D Dataset (H3D), NBA SportVU Dataset (NBA), and Stanford Drone Dataset (SDD) | present a generic trajectory forecasting framework with explicit relational reasoning among multiple heterogeneous, interactive agents with a graph representation. |
Multi-agent Trajectory Prediction with Fuzzy Query Attention | NeurIPS | 2020 | Code link | ETH-UCY, Collisions, NGsim, Charges, NBA | a general architecture designed to predict trajectories in multi-agent systems while modeling the crucial inductive biases of motion, namely, inertia, relative motion, intents and interactions. |
Set Functions for Time Series | ICML | 2020 | Code link | MIMIC-III, Physionet 2012 Mortality Prediction Challenge | presented a novel approach for classifying time se- ries with irregularly-sampled and unaligned. |
Learning from Irregularly-Sampled Time Series: A Missing Data Perspective | ICML | 2020 | Code link | MNIST..... | introduced an encoder-decoder framework for modeling general missing data problems and introduced two model families leveraging this framework: P-VAE and P-BiGAN. |
Unsupervised Transfer Learning for Spatiotemporal Predictive Networks | ICML | 2020 | Code link | - | studied a new unsupervised transfer learn- ing problem of using multiple pretrained models to im- prove the performance of a new spatiotemporal predictive learning task. |
Connecting the Dots: Multivariate Time Series Forecasting with Graph Neural Networks | KDD | 2020 | Code link | Solar-Energy, Traffic, Electricity, Exchange-Rate and PeMS ...... | a novel framework for multivariate time series forecasting |
Deep State-Space Generative Model For Correlated Time-to-Event Predictions | KDD | 2020 | Code link | MIMIC-III | proposed a deep latent state-space generative model to capture the relations between patientsβ mortality risk and the associated organ failure risks. |
Attention based multi-modal new product sales time-series forecasting | KDD | 2020 | - | - | propose and empirically evaluate several novel attention-based multi-modal encoder-decoder models to forecast the sales for a new product purely based on product images, any available product attributes and also external factors like holidays, events, weather, and discount. |
BusTr: predicting bus travel times from real-time traffic | KDD | 2020 | - | - | demonstrates excellent generalization to test data that differs both spatially and temporally from the training examples we use, allowing our model to cope gracefully with the ever-changing world. |
CompactETA: A Fast Inference System for Travel Time Prediction | KDD | 2020 | - | - | encode high order spatial and temporal dependency into sophisticated representations by applying graph attention network on a spatiotemporal weighted road network graph. We further encode the sequential information of the travel route by positional encoding to avoid the recurrent network structure. |
DATSING: Data Augmented Time Series Forecasting with Adversarial Domain Adaptation | CIKM | 2020 | - | - | propose a two-phased frameworkwhich first clusters similar mixed domains time series data and thenperforms a fine-tuning procedure with domain adversarial regular-ization to achieve better out-of-sample generalization. |
Dual Sequential Network for Temporal Sets Prediction | SIGIR | 2020 | - | - | addressed the problem that most of the existing methods were designed for predicting time series or temporal events, which could not be directly used for temporal sets prediction due to the difficulties of multi-level representations of items and sets, complex temporal dependencies of sets, and evolving dynamics of sequential behaviors. |
Think Globally, Act Locally: A Deep Neural Network Approach to High-Dimensional Time Series Forecasting | NeurIPS | 2019 | Code | electricity, traffic, wik, PeMS07(M) | Our model can be trained effectively on high-dimensional but diverse time series, where different time series can have vastly different scales, without a priori normalization or rescaling. |
Shape and Time Distortion Loss for Training Deep Time Series Forecasting Models | NeurIPS | 2019 | Code | Synth, ECG, Traffic | We introduce a differentiable loss function suitable for training deep neural nets, and provide a custom back-prop implementation for speeding up optimization. We also introduce a variant of DILATE, which provides a smooth generalization of temporally-constrained Dynamic Time Warping (DTW). |
Enhancing the Locality and Breaking the Memory Bottleneck of Transformer on Time Series Forecasting | NeurIPS | 2019 | - | solar, wind | we first propose convolutional self-attention by producing queries and keys with causal convolution so that local context can be better incorporated into attention mechanism. |
Discovering Latent Covariance Structures for Multiple Time Series | ICML | 2019 | - | - | We present a pragmatic search algorithm which explores a larger structure space efficiently. |
BeatGAN: Anomalous Rhythm Detection using Adversarially Generated Time Series | IJCAI | 2019 | Code | - | BeatGAN outputs explainable results to pinpoint the anomalous time ticks of an input beat, by comparing them to ad- versarially generated beats. |
Learning Interpretable Deep State Space Model for Probabilistic Time Series Forecasting | IJCAI | 2019 | Code | MIT-BIH ECG dataset, CMU Motion Capture dataset | Our approach involves a deep network based embodiment of the state space model, to allow for non-linear emission and transition models design, which is flexible to deal with arbitrary data distribution. |
Explainable Deep Neural Networks for Multivariate Time Series Predictions | IJCAI | 2019 | - | - | We design a two stage convolutional neural network architec- ture which uses particular kernel sizes. This allows us to utilise gradient based techniques for generat- ing saliency maps for both the time dimension and the features. |
Paper | Conference | Year | Code | Used Datasets | Key Contribution |
---|---|---|---|---|---|
OMNI-SCALE CNNS: A SIMPLE AND EFFECTIVE KERNEL SIZE CONFIGURATION FOR TIME SERIES CLASSIFICATION | ICLR | 2022 | Code link | MEG-TLE, UEA 30 archive, UCR 85 archive, UCR 128 archive | presents a simple 1D-CNN block, namely OS-block. |
Correlative Channel-Aware Fusion for Multi-View Time Series Classification | AAAI | 2021 | - | EV-Action, NTU RGB+D, UCI Daily and Sports Activities | The global-local temporal encoders are developed to extract robust temporal representations for each view, and a learnable fusion mechanism is proposed to boost the multi-view label information. |
Learnable Dynamic Temporal Pooling for Time Series Classification | AAAI | 2021 | - | UCR/UEA | proposes a dynamic temporal pooling + a learning framework to simultaneously optimize the network parameters of a CNN classifier and the prototypical hidden series that encodes the latent semantic of the segments. |
ShapeNet: A Shapelet-Neural Network Approach for Multivariate Time Series Classification | AAAI | 2021 | Code link | UEA MTS datasets | We propose Mdc-CNN to learn time series subsequences of various lengths into unified space and propose a cluster-wise triplet loss to train the network in an unsupervised fashion. We adopt MST to obtain the MST representation of time series. |
Joint-Label Learning by Dual Augmentation for Time Series Classification | AAAI | 2021 | Code | UCR | a novel time-series data augmentation method. |
Explainable Multivariate Time Series Classification: A Deep Neural Network Which Learns To Attend To Important Variables As Well As Time Intervals | WSDM | 2021 | - | PM2.5w Seizure Movement | introduced LAXCAT, a novel, modular architecture for explainable multivariate time series classification. |
Voice2Series: Reprogramming Acoustic Models for Time Series Classification | ICML | 2021 | Code link | Coffee,...... | a novel approach to reprogram a pre-trained acoustic model for time series classification. |
Learning Saliency Maps to Explain Deep Time Series Classifiers | CIKM | 2021 | - | Wafer, GunPoint, Computers, Earthqakes, FordA, FordB, CricketX, PTB, ECG | a method that learns to highlight the timesteps that are most responsible and the degree to which they are important for the classifierβs prediction. |
Gaussian Process Model Learning for Time Series Classification | ICDM | 2021 | - | - | proposed a novel approach for time series classification called Local Gaussian Process Model Inference Classification (LOGIC). |
Contrast Profile: A Novel Time Series Primitive that Allows Classification in Real World Settings | ICDM | 2021 | Code link | UCR archive | We have shown that the MPdist is more robust to noise, irrelevant data, misalignment etc., than either Euclidian distance or DTW. |
Attentive Neural Controlled Differential Equations for Time-series Classification and Forecasting | ICDM | 2021 | Code link | - | a novel NCDE architecture that incorporates the concept of attention. |
Imbalanced Time Series Classification for Flight Data Analyzing with Nonlinear Granger Causality Learning | CIKM | 2020 | Code link | - | presented a neural network classification model for imbalanced multivariate time series by leveraging the information learned from normal class, which can also learn the nonlinear Granger causality for each class, so that we can pinpoint how time series classes differ from each other. |
Visualet: Visualizing Shapelets for Time Series Classification | CIKM | 2020 | Code link | UCR archive | Such efficiency has made it possible for demo attendees to interact with shapelet discovery and explore high-quality shapelets. In this demo, we present Visualet -- a tool for visualizing shapelets, and exploring effective and interpretable ones. |
Learning Discriminative Virtual Sequences for Time Series Classification | CIKM | 2020 | Code link | - | propose a novel time series classification method named Discriminative Virtual Sequence Learning (DVSL). |
Fast and Accurate Time Series Classification Through Supervised Interval Search | CIKM | 2020 | Code link | - | STSF improves the classification efficiency by examining only a (set of) sub-series of the original time series, and its tree-based structure allows for interpretable outcomes. |
Dataset | Conference | Year | Code | Used Datasets | Key Contribution |
---|---|---|---|---|---|
Unsupervised Model Selection For Time-series Anomaly Detection | ICLR | 2023 | UCR, SMD | In this paper, we explore how we can select accurate time-series anomaly detection models given an unlabeled dataset and a set of candidate models. Iβd like to point out that we use adjusted F1. The adjustment ensures that once an algorithm detects even a part of the anomaly, we consider that it has detected the entire anomaly. | |
Deep Variational Graph Convolutional Recurrent Network for Multivariate Time Series Anomaly Detection | ICML | 2022 | DND, SMD, MSL, SMAP | In this paper, we model channel dependency and stochasticity within MTS by developing an embedding-guided probabilistic generative network. We combine it with adaptive Variational Graph Convolutional Recurrent Network (VGCRN) to model both spatial and temporal fine-grained correlations in MTS. To explore hierarchical latent representations, we further extend VGCRN into a deep variational network, which captures multilevel information at different layers and is robust to noisy time series. | |
A Semi-Supervised VAE Based Active Anomaly Detection Framework in Multivariate Time Series for Online Systems | WWW | 2022 | - | online cloud server data from two different types of game business | SLA-VAE first defines anomalies based on feature extraction module, introduces semi-supervised VAE to identify anomalies in multivariate time series, and employs active learning to update the online model via a small number of uncertain samples. |
Towards a Rigorous Evaluation of Time-series Anomaly Detection | AAAI | 2022 | - | Secure water treatment (SWaT), ...... | applying PA can severely overestimate a TAD modelβs capability. |
DeepGPD: A Deep Learning Approach for Modeling Geospatio-Temporal Extreme Events | AAAI | 2022 | Code link | the Global Historical Climatology Network (GHCN) | proposed a novel deep learning architecture (DeepGPD) capable of learning the parameters of the generalized Pareto distribution while satisfying the conditions placed on those parameters. |
Graph-Augmented Normalizing Flows for Anomaly Detection of Multiple Time Series | ICLR | 2022 | - | PMU-B, PMU-C, SWaT, METR-LA | propose a novel flow model by imposing a Bayesian network among constituent series. |
Anomaly Transformer: Time Series Anomaly Detection with Association Discrepancy | ICLR | 2022 | - | SMD MSL SMAP SWaT PSM | propose the Anomaly Transformer with a new Anomaly-Attention mechanism to compute the association discrepancy. A minimax strategy is devised to amplify the normal-abnormal distinguishability of the association discrepancy. |
Graph Neural Network-Based Anomaly Detection in Multivariate Time Series | AAAI | 2021 | - | SWaT and WADI | proposed our Graph Deviation Network (GDN) approach, which learns a graph of relationships between sensors, and detects deviations from these patterns, while incorporating sensor embeddings |
Time Series Anomaly Detection with Multiresolution Ensemble Decoding | AAAI | 2021 | - | ECG, 2D-gesture and Power-demand, Yahooβs S5 | Its core is to use lower-resolution information to help long-range decoding at layers with higher resolutions. This is achieved by jointly learning multiple recurrent decoders where each decoder is with a different decoding length. |
Outlier Impact Characterization for Time Series Data | AAAI | 2021 | benchmark | Webscope, Physionet | study recurring outliers in time series data and aim to provide a systematic way of measuring the impact of such outliers on time series analysis |
F-FADE: Frequency Factorization for Anomaly Detection in Edge Streams | WSDM | 2021 | Code link | - | a new approach for detection of anomalies in edge streams, which uses a novel frequency-factorization technique to efficiently model the time-evolving distributions of frequencies of interactions between node-pairs. |
FluxEV: A Fast and Effective Unsupervised Framework for Time-Series Anomaly Detection | WSDM | 2021 | Code link | - | By converting the non-extreme anomalies to extreme values, our framework addresses the limitation of SPOT and achieves a huge improvement in the detection accuracy. Moreover, Method of Moments is adopted to speed up the parameter estimation in the automatic thresholding. |
Event Outlier Detection in Continuous Time | ICML | 2021 | Code link | MIMIC III | we develop outlier detection methods based on point processes thatcan take context information into account. Our methods are based on Bayesian decision theory and hypothesis testing with theoretical guarantees. |
Multivariate Time Series Anomaly Detection and Interpretation using Hierarchical Inter-Metric and Temporal Embedding | KDD | 2021 | Code link | - | Its core idea is to model the normal patterns inside MTS data through hierarchical Variational AutoEncoder with two stochastic latent variables, each of which learns low-dimensional inter-metric or temporal embeddings. Furthermore, we propose an MCMC-based method to obtain reasonable embeddings and reconstructions at anomalous parts for MTS anomaly interpretation. |
Practical Approach to Asynchronous Multi-variate Time Series Anomaly Detection and Localization | KDD | 2021 | Code link | - | Our solution is designed to leverage this behavior. The solution utilizes spectral analysis on the latent representation of a pre-trained autoencoder to extract dominant frequencies across the signals, which are then used in a subsequent network that learns the phase shifts across the signals and produces a synchronized representation of the raw multivariate. |
Time Series Anomaly Detection for Cyber-physical Systems via Neural System Identification and Bayesian Filtering | KDD | 2021 | Code link | - | a specially crafted neural network architecture is posed for system identification, i.e., capturing the dynamics of CPS in a dynamical state-space model; then a Bayesian filtering algorithm is naturally applied on top of the "identified" state-space model for robust anomaly detection by tracking the uncertainty of the hidden state of the system recursively over time. |
Multi-Scale One-Class Recurrent Neural Networks for Discrete Event Sequence Anomaly Detection | KDD | 2021 | Code link | - | a multi-scale one-class recurrent neural network for detecting anomalies in discrete event sequences. |
Online false discovery rate control for anomaly detection in time series | NeurIPS | 2021 | - | - | The methods proposed in this article overcome short-comings of previous FDRC rules in the context of anomaly detection, in particular ensuring that power remains high even when the alternative is exceedingly rare (typical in anomaly detection) and the test statistics are serially dependent (typical in time series). |
Detecting Anomalous Event Sequences with Temporal Point Processes | NeurIPS | 2021 | - | LOGS,STEAD | The proposed method can be combined with various TPP models, such as neural TPPs, and is easy to implement. |
You Only Look at One Sequence: Rethinking Transformer in Vision through Object Detection | NeurIPS | 2021 | Code link | - | have explored the transferability of the vanilla ViT pre-trained on mid-sized ImageNet-1k dataset to the more challenging COCO object detection benchmark |
Drop-DTW: Aligning Common Signal Between Sequences While Dropping Outliers | NeurIPS | 2021 | - | MNIST | introduced an extension to the classic DTW algorithm, which relaxes the constraints of matching endpoints of paired sequences and the continuity of the path cost. |
Conditional Score-based Diffusion Models for Probabilistic Time Series Imputation | NeurIPS | 2021 | Code link | healthcare, air quality | a novel approach to impute multivariate time series with conditional diffusion models. |
SDFVAE: Static and Dynamic Factorized VAE for Anomaly Detection of Multivariate CDN KPIs | WWW | 2021 | - | - | Our key insight is that different KPIs are constrained by certain time-invariant characteristics of the underlying system, and that explicitly modelling such invariance may help resist noise in the data. We thus propose a novel anomaly detection method called SDFVAE, short for Static and Dynamic Factorized VAE, that learns the representations of KPIs by explicitly factorizing the latent variables into dynamic and static parts. |
Time-series Change Point Detection with Self-Supervised Contrastive Predictive Coding | WWW | 2021 | - | Yahoo!Benchmark, HASC, USC-HAD | propose a novel self-supervised CPD method, ππ β πΆπ2 for time series. ππ β πΆπ2 learns an embedded representation predict a future interval of a times series from historical samples. |
NTAM: Neighborhood-Temporal Attention Model for Disk Failure Prediction in Cloud Platforms | WWW | 2021 | - | industrial datasets collected from millions of disks in Microsoft Azure,....... | NTAM is a novel approach that not only utilizes a diskβs own status data, but also considers its neighborsβ status data. Moreover, NTAM includes a novel attention-based temporal component to capture the temporal nature of the disk status data. Besides, we propose a data enhancement method, called Temporal Progressive Sampling (TPS), to handle the extreme data imbalance issue. |
Improving Irregularly Sampled Time Series Learning with Time-Aware Dual-Attention Memory-Augmented Networks | CIKM | 2021 | Code link | - | The proposed model can leverage both time irregularity, multi-sampling rates and global temporal patterns information inherent in IASS-MTS so as to learn more effective representations for improving prediction performance. |
BiCMTS: Bidirectional Coupled Multivariate Learning of Irregular Time Series with Missing Values | CIKM | 2021 | - | - | BiCMTS method to represent both forward and backward value couplings within a time series by RNNs and between MTS by self-attention networks; the learned bidirectional intra- and inter-time series coupling representations are fused to estimate missing values. |
Timeseries Anomaly Detection using Temporal Hierarchical One-Class Network | NeurIPS | 2020 | - | 2D-gesture, Power demand, KDD-Cup99 data, SWaT, MSL, SMAP | based on a set of hierarchical structured hyperspheres. The solution uses a probabilistic relevance on cluster centers to help the model access the whole temporal history. A center orthogonality loss and a temporal self-supervision loss are also introduced for improved feature representation. |
USAD : UnSupervised Anomaly Detection on multivariate time series | KDD | 2020 | - | SWaT, WADI, SMD, SMAP, MSL | adversely trained autoencoders |
Application Performance Anomaly Detection with LSTM on Temporal Irregularities in Logs | CIKM | 2020 | Code link | - | present a new method to perform anomaly detection, while maintaining the quantitative aspect of time, using a count of event types over time. |
Multivariate Time-series Anomaly Detection via Graph Attention Network | ICDM | 2020 | Code link | SMAP, MSL, TSA | propose a novel framework based on graph attention network for multivariate time-series anomaly detection. |
MERLIN: Parameter-Free Discovery of Arbitrary Length Anomalies in Massive Time Series Archives | ICDM | 2020 | - | - | an algorithm that can efficiently and exactly find discords of all lengths in massive time series archives. |
Outlier Detection for Time Series with Recurrent Autoencoder Ensembles | IJCAI | 2019 | Code | Traffic,...... | The two solutions are ensemble frameworks, specifically an indepen- dent framework and a shared framework, both of which combine multiple S-RNN based autoencoders to enable outlier detection. |
Regularization for Time Series Trend Filtering | IJCAI | 2019 | - | - | we adopt the Hu- ber loss to suppress outliers, and utilize a combina- tion of the first order and second order difference on the trend component as regularization to cap- ture both slow and abrupt trend changes. Further- more, an efficient method is designed to solve the proposed robust trend filtering based on majoriza- tion minimization (MM) and alternative direction method of multipliers (ADMM). |
Paper | Conference | Year | Code | Used Datasets | Key Contribution |
---|---|---|---|---|---|
Clustering Interval-Censored Time-Series for Disease Phenotyping | AAAI | 2022 | - | - | present our method, SubLign, to learn latent representations of disease progression that correct for temporal misalignment in real-world observations and consider conditions for identifiability of subtype and alignment values. |
Corsets for Time Series Clustering | NeurIPS | 2021 | - | synthetic data | address the problem of constructing coresets for time series data generated from Gaussian mixture models with auto-correlations across time. |
Temporal Phenotyping using Deep Predictive Clustering of Disease Progression | ICML | 2020 | Code link | UKCF, Alzheimerβs Disease Neuroimaging Initiative (ADNI) | defined novel loss functions to encourage each cluster to have homogeneous future outcomes and designed optimization procedures to avoid trivial solutions in identifying cluster as- signments and the centroids. |
Learning low-dimensional state embeddings and metastable clusters from time series data | NeurIPS | 2019 | - | simulated diffusion processes | his idea also leads to a kernel reshaping method for more accurate nonparametric estimation of the transition function. State embedding can be used to cluster states into metastable sets, thereby identifying the slow dynamics. Sharp statistical error bounds and misclassification rate are proved. |
Learning Representations for Time Series Clustering | NeurIPS | 2019 | - | - | ere we propose a novel unsupervised temporal representation learning model, named Deep Temporal Clustering Representation (DTCR), which integrates the temporal reconstruction and K-means objective into the seq2seq model. |
Paper | Conference | Year | Code | Used Datasets | Key Contribution |
---|---|---|---|---|---|
ClaSP-Time Series Segmentation | CIKM | 2021 | - | 98 datasets.... | a novel and highly accurate method for TSS. ClaSP hierarchically splits a TS into two parts, where each split point is determined by training a binary TS classifier for each possible split point and selecting the one with highest accuracy |
Multi-series Time-aware Sequence Partitioning for Disease Progression Modeling | IJCAI | 2021 | - | sEMG | improved the TICC by incorporating multi- series input (M-TICC) and time-awareness (MT-TICC). |
Linear Time Complexity Time Series Clustering with Symbolic Pattern Forest | IJCAI | 2019 | Code | - | This paper presents a novel time series clustering algorithm that has linear time complex- ity. The proposed algorithm partitions the data by checking some randomly selected symbolic pat- terns in the time series. |
Similarity Preserving Representation Learning for Time Series Clustering | IJCAI | 2019 | - | - | In this paper, we bridge this gap by proposing an efficient representation learning framework that is able to convert a set of time series with various lengths to an instance-feature matrix. |
Paper | Conference | Year | Code | Used Datasets | Key Contribution |
---|---|---|---|---|---|
Adaptive Conformal Predictions for Time Series | ICML | 2022 | code | Uncertainty quantification of predictive models is crucial in decision-making problems. Conformal prediction is a general and theoretically sound answer. However, it requires exchangeable data, excluding time series. While recent works tackled this issue, we argue that Adaptive Conformal Inference (ACI, Gibbs and Candes` , 2021), developed for distribution-shift time series, is a good procedure for time series with general dependency. We theoretically analyse the impact of the learning rate on its efficiency in the exchangeable and auto-regressive case. We propose a parameter-free method, AgACI, that adaptively builds upon ACI based on online expert aggregation. We lead extensive fair simulations against competing methods that advocate for ACIβs use in time series. We conduct a real case study: electricity price forecasting. The proposed aggregation algorithm provides efficient prediction intervals for day-ahead forecasting. All the code and data to reproduce the experiments is made available. | |
Modeling Irregular Time Series with Continuous Recurrent Units | ICML | 2022 | code | Pendulum Images, Climate Data (USHCN), Electronic Health Records (Physionet) | In many datasets (e.g. medical records) observation times are irregular and can carry important information. To address this challenge, we propose continuous recurrent units (CRUs) β a neural architecture that can naturally handle irregular intervals between observations. |
Unsupervised Time-Series Representation Learning with Iterative Bilinear Temporal-Spectral Fusion | ICML | 2022 | HAR, SleepEDF, ECG Waveform, ETT, Weather, SaaT, WADI, SMD, SMAP, MSL | We devise a novel iterative bilinear temporal-spectral fusion to explicitly encode the affinities of abundant time-frequency pairs, and iteratively refines representations in a fusion-and-squeeze manner with Spectrum-to-Time (S2T) and Time-to-Spectrum (T2S) Aggregation modules | |
Utilizing Expert Features for Contrastive Learning of Time-Series Representations | ICML | 2022 | code | HAR, SleepEDF, ECG Waveform | We present an approach that incorporates expert knowledge for time-series representation learning. Our method employs expert features to replace the commonly used data transformations in previous contrastive learning approaches. |
Neural Predicting Higher-order Patterns in Temporal Networks | WWW | 2022 | Code | Tags-math-sx, Tags-ask-ubuntu, Congress-bills, DAWN, Threads-ask-ubuntu | we proposed the first model HIT to predict higher- order patterns in temporal hypergraphs to answer what type of, when, and why interactions may expand in a node triplet. HIT can be further generalized to predict even higher-order patterns. |
ONBRA: Rigorous Estimation of the Temporal Betweenness Centrality in Temporal Networks | WWW | 2022 | Code | data, data2 | In this work we present ONBRA, the first sampling-based approximation algorithm for estimating the temporal betweenness centrality values of the nodes in a temporal network, providing rigorous probabilistic guar- antees on the quality of its output. |
Knowledge-based Temporal Fusion Network for Interpretable Online Video Popularity Prediction | WWW | 2022 | Code | medium-video dataset and a micro-video dataset from the server logs of Xigua and Douyin | In this paper, we propose a Knowledge-based Temporal Fusion Network (KTFN) that incorporates knowledge graph representation to address the aforementioned challenges in the task of online video popularity prediction. |
STAM: A Spatiotemporal Aggregation Method for Graph Neural Network-based Recommendation | WWW | 2022 | Code | MovieLens, Amazon, Taobao | In this work, we propose a spatiotemporal aggregation method STAM to efficiently incorporate temporal information into neighbor embedding learning. |
A Graph Temporal Information Learning Framework for Popularity Prediction | WWW | 2022 | Code | Sina Weibo | In this paper, we propose a graph temporal information learning framework based on an improved graph convolutional network (GTGCN), which can capture both the temporal information governing the spread of information in a snapshot, and the inherent temporal dependencies among different snapshots. |
PREP: Pre-training with Temporal Elapse Inference for Popularity Prediction | WWW | 2022 | Code | Sina Weibo, Twitter | We design a novel pretext task for pre-training, i.e., temporal elapse inference for two ran- domly sampled time slices of popularity dynamics, impelling the representation model to learn intrinsic knowledge about popularity dynamics. |
Conditional Loss and Deep Euler Scheme for Time Series Generation | AAAI | 2022 | - | - | - |
TS2Vec: Towards Universal Representation of Time Series | AAAI | 2022 | Code link | 128 UCR datasets,30 UEA datasets, 3 ETT datasets, Electricity, Yahoo dataset, KPI dataset | performs contrastive learning in a hierarchical way over augmented context views |
Time Masking for Temporal Language Models | WSDM | 2022 | Code link | - | - |
Long Short-Term Temporal Meta-learning in Online Recommendation | WSDM | 2022 | - | - | - |
Structure Meets Sequences: Predicting Network of Co-evolving Sequences | WSDM | 2022 | Code link | - | - |
EvoKG: Jointly Modeling Event Time and Network Structure for Reasoning over Temporal Knowledge Graphs | WSDM | 2022 | Code link | - | - |
FILLING THE GAPS: MULTIVARIATE TIME SERIES IMPUTATION BY GRAPH NEURAL NETWORKS | ICLR | 2022 | - | Air quality, Traffic, and Smart Grids | - |
PSA-GAN: PROGRESSIVE SELF ATTENTION GANS FOR SYNTHETIC TIME SERIES | ICLR | 2022 | Code link, Code on glueonts | Electricty, M4, Solar energy, Traffic | - |
Generative Semi-Supervised Learning for Multivariate Time Series Imputation | AAAI | 2021 | - | - | - |
Temporal Cross-Effects in Knowledge Tracing | WSDM | 2021 | - | - | - |
Learning Dynamic Embeddings for Temporal Knowledge Graphs | WSDM | 2021 | - | - | - |
Temporal Meta-path Guided Explainable Recommendation | WSDM | 2021 | - | - | - |
Generative Adversarial Networks for Markovian Temporal Dynamics: Stochastic Continuous Data Generation | ICML | 2021 | Code link | - | - |
Discrete-time Temporal Network Embedding via Implicit Hierarchical Learning | KDD | 2021 | Code link | - | - |
Time-series Generation by Contrastive Imitation | NeurIPS | 2021 | - | - | - |
Adjusting for Autocorrelated Errors in Neural Networks for Time Series | NeurIPS | 2021 | Code link | - | - |
Spikelet: An Adaptive Symbolic Approximation for Finding Higher-Level Structure in Time Series | ICDM | 2021 | - | - | - |
STING: Self-attention based Time-series Imputation Networks using GAN | ICDM | 2021 | Code link | - | - |
SMATE: Semi-Supervised Spatio-Temporal Representation Learning on Multivariate Time Series | ICDM | 2021 | Code link | - | - |
TCube: Domain-Agnostic Neural Time-series Narration | ICDM | 2021 | Code link | - | - |
Towards Interpretability and Personalization: A Predictive Framework for Clinical Time-series Analysis | ICDM | 2021 | Code link | - | - |
Continual Learning for Multivariate Time Series Tasks with Variable Input Dimensions | ICDM | 2021 | Code link | - | - |
CASPITA: Mining Statistically Significant Paths in Time Series Data from an Unknown Network | ICDM | 2021 | Code link | - | - |
Multi-way Time Series Join on Multi-length Patterns | ICDM | 2021 | Code link | - | - |
Temporal Event Profiling based on Multivariate Time Series Analysis over Long-term Document Archives | SIGIR | 2021 | Code link | - | - |
Time-Aware Multi-Scale RNNs for Time Series Modeling | ICDM | 2021 | Code link | - | - |
Deep reconstruction of strange attractors from time series | NeurIPS | 2020 | Code link | - | - |
One Detector to Rule Them All: Towards a General Deepfake Attack Detection Framework | WWW | 2021 | Code link | - | - |
High-recall causal discovery for autocorrelated time series with latent confounders | NeurIPS | 2020 | Code link | - | - |
Learning Long-Term Dependencies in Irregularly-Sampled Time Series | NeurIPS | 2020 | Code link | - | - |
ARMA Nets: Expanding Receptive Field for Dense Prediction | NeurIPS | 2020 | Code link | - | - |
Learnable Group Transform For Time-Series | ICML | 2020 | Code link | - | - |
Fast RobustSTL: Efficient and Robust Seasonal-Trend Decomposition for Time Series with Complex Patterns | KDD | 2020 | - | - | - |
Matrix Profile XXI: A Geometric Approach to Time Series Chains Improves Robustness | KDD | 2020 | Code link | - | - |
Multi-Source Deep Domain Adaptation with Weak Supervision for Time-Series Sensor Data | KDD | 2020 | Code link | - | - |
Personalized Imputation on Wearable-Sensory Time Series via Knowledge Transfer | CIKM | 2020 | Code link | - | - |
Hybrid Sequential Recommender via Time-aware Attentive Memory Network | CIKM | 2020 | Code link | - | - |
Order-Preserving Metric Learning for Mining Multivariate Time Series | ICDM | 2020 | Code link | - | - |
Fast Automatic Feature Selection for Multi-period Sliding Window Aggregate in Time Series | ICDM | 2020 | - | Tianchi, PLAsTiCC, NFL, MotionSense, Gas Sensors | a framework to fill the gap of the end-to-end automatic sliding window aggregate feature selection for time series |
Matrix Profile XXII: Exact Discovery of Time Series Motifs Under DTW | ICDM | 2020 | Code link | - | - |
Inductive Granger Causal Modeling for Multivariate Time Series | ICDM | 2020 | - | Finance, FMRI, Synthetic data | - |
Mining Recurring Patterns in Real-Valued Time Series using the Radius Profile | ICDM | 2020 | - | - | - |
Learning Periods from Incomplete Multivariate Time Series | ICDM | 2020 | - | - | - |
FilCorr: Filtered and Lagged Correlation on Streaming Time Series | ICDM | 2020 | - | - | - |
Unsupervised Scalable Representation Learning for Multivariate Time SeriesΒ | NeurIPS | 2019 | Code | Datasets | - |
Latent Ordinary Differential Equations for Irregularly-Sampled Time Series | NeurIPS | 2019 | - | Human Activity dataset | We generalize RNNs to have continuous-time hidden dynamics defined by ordinary differential equations (ODEs), a model we call ODE-RNNs. Furthermore, we use ODE-RNNs to replace the recognition network of the recently-proposed Latent ODE model. |
GRU-ODE-Bayes: Continuous Modeling of Sporadically-Observed Time Series | NeurIPS | 2019 | Code | Datasets | - |
Interpolation-Prediction Networks for Irregularly Sampled Time Series | ICLR | 2019 | Code | MIMIC-III, UWaveGestureLibraryAll | In this paper, we have presented a new framework for dealing with the problem of supervised learn- ing in the presence of sparse and irregularly sampled time series. The proposed framework is fully modular. |
SOM-VAE: Interpretable Discrete Representation Learning on Time Series | ICLR | 2019 | - | - | The SOM-VAE can recover topologically interpretable state representations on time series and static data. It provides an improvement to standard methods in terms of clustering performance and offers a way to learn discrete two-dimensional representations of the data manifold in concurrence with the reconstruction task. |
U-Time: A Fully Convolutional Network for Time Series Segmentation Applied to Sleep Staging | NeurIPS | 2019 | Code | Datasets | - |
E^2GAN: End-to-End Generative Adversarial Network for Multivariate Time Series Imputation | IJCAI | 2019 | - | - | - |
Name | Company | Stars | Explanation |
---|---|---|---|
π Darts | Unit8 | βοΈ 5.3K | Darts is a Python library for user-friendly forecasting and anomaly detection on time series. It contains a variety of models, from classics such as ARIMA to deep neural networks. The forecasting models can all be used in the same way, using fit() and predict() functions, similar to scikit-learn. The library also makes it easy to backtest models, combine the predictions of several models, and take external data into account. Darts supports both univariate and multivariate time series and models. The ML-based models can be trained on potentially large datasets containing multiple time series, and some of the models offer a rich support for probabilistic forecasting. |
π Prophet | Meta (Facebook) | βοΈ 15.5K | Prophet is a procedure for forecasting time series data based on an additive model where non-linear trends are fit with yearly, weekly, and daily seasonality, plus holiday effects. It works best with time series that have strong seasonal effects and several seasons of historical data. Prophet is robust to missing data and shifts in the trend, and typically handles outliers well. |
π Neural Prophet | - | βοΈ 2.8K | A Neural Network based Time-Series model, inspired by Facebook Prophet and AR-Net, built on PyTorch. |
π GluonTS | AWS | βοΈ 3.3K | GluonTS is a Python package for probabilistic time series modeling, focusing on deep learning based models, based on PyTorch and MXNet. |
π stumpy | TD Ameritrade | βοΈ 2.5K | STUMPY is a powerful and scalable Python library that efficiently computes something called the matrix profile, which is just an academic way of saying "for every subsequence within your time series, automatically identify its corresponding nearest-neighbor". What's important is that once you've computed your matrix profile (middle panel above) it can then be used for a variety of time series data mining tasks. |
π tsfresh | Blue Yonder GmbH | βοΈ 7K | The package provides systematic time-series feature extraction by combining established algorithms from statistics, time-series analysis, signal processing, and nonlinear dynamics with a robust feature selection algorithm. In this context, the term time-series is interpreted in the broadest possible sense, such that any types of sampled data or even event sequences can be characterised. |
π SKTIME | - | βοΈ 6.1K | sktime is a library for time series analysis in Python. It provides a unified interface for multiple time series learning tasks. Currently, this includes time series classification, regression, clustering, annotation and forecasting. It comes with time series algorithms and scikit-learn compatible tools to build, tune and validate time series models. |
π pmdarima | - | βοΈ 1.3K | Pmdarima (originally pyramid-arima, for the anagram of 'py' + 'arima') is a statistical library designed to fill the void in Python's time series analysis capabilities. |
π tslearn | - | βοΈ 2.4K | The machine learning toolkit for time series analysis in Python. |
π PyTorch Forecasting | - | βοΈ 2.6K | PyTorch Forecasting is a PyTorch-based package for forecasting time series with state-of-the-art network architectures. It provides a high-level API for training networks on pandas data frames and leverages PyTorch Lightning for scalable training on (multiple) GPUs, CPUs and for automatic logging. |
π StatsForecast | - | βοΈ 2.2K | StatsForecast offers a collection of widely used univariate time series forecasting models, including automatic ARIMA, ETS, CES, and Theta modeling optimized for high performance using numba. It also includes a large battery of benchmarking models. |
π Streamz | - | βοΈ 1.1K | Streamz helps you build pipelines to manage continuous streams of data. It is simple to use in simple cases, but also supports complex pipelines that involve branching, joining, flow control, feedback, back pressure, and so on. |
Paper | Conference | Year | Code | Key Contribution |
---|---|---|---|---|
Monash Time Series Forecasting Repository | NeurIPS | 2021 | paper link | There have been many deep time series evaluated on the same datasets in recent years. Even though this works for basic benchmarking, it may not hold up when applied to a variety of temporal tasks. Its goal is to create a "master list" of different time series datasets and serve as an authoritative benchmark. Over 20 different datasets are included in the repository, spanning industries as diverse as health, retail, ride-share, and demographics. |
Revisiting Time Series Outlier Detection: Definitions and Benchmarks | NeurIPS | 2021 | link | This paper critiques many existing time series anomaly/outlier detection datasets and proposes 35 brand-new synthetic datasets and 4 real-world datasets for benchmarking purposes. |
Subseasonal Forecasting Microsoft | Microsoft | 2021 | link | Microsoft has released a dataset to facilitate machine learning for improving subseasonal forecasting (e.g. two to six weeks in the future). Forecasting subseasonally helps government agencies and farmers prepare for weather events. In general, deep learning models performed quite poorly compared to other methods in Microsoft's benchmark. A simple feed-forward model proved to be the most accurate DL model, while the Informer performed poorly. |
We appreciate all contributions to improve this paper repo!
Please feel free to pull requests, open an issue or send me email ([email protected]) to add awesome papers.