Context Consistency Regularization for Label Sparsity in Time Series

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 45
  • Download : 0
Labels are typically sparse in real-world time series due to the high annotation cost. Recently, consistency regularization techniques have been used to generate artificial labels from unlabeled augmented instances. To fully exploit the sequential characteristic of time series in consistency regularization, we propose a novel method of data augmentation called context-attached augmentation, which adds preceding and succeeding instances to a target instance to form its augmented instance. Unlike the existing augmentation techniques that modify a target instance by directly perturbing its attributes, the context-attached augmentation generates instances augmented with varying contexts while maintaining the target instance. Based on our augmentation method, we propose a context consistency regularization framework, which first adds different contexts to a target instance sampled from a given time series and then shares unitary reliability-based cross-window labels across the augmented instances to maintain consistency. We demonstrate that the proposed framework outperforms the existing state-of-the-art consistency regularization frameworks through comprehensive experiments on real-world time-series datasets.
Publisher
ML Research Press
Issue Date
2023-07-28
Language
English
Citation

40th International Conference on Machine Learning, ICML 2023, pp.31579 - 31595

URI
http://hdl.handle.net/10203/314370
Appears in Collection
IE-Conference Papers(학술회의논문)CS-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0