transferlearning
TS-TCC
transferlearning | TS-TCC | |
---|---|---|
1 | 1 | |
12,867 | 311 | |
- | - | |
7.8 | 3.2 | |
9 days ago | about 1 month ago | |
Python | Python | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
transferlearning
-
[D] Medium Article: Adaptive Learning for Time Series Forecasting
The src is available in https://github.com/jindongwang/transferlearning I'll also publish about how to code the model for time series
TS-TCC
-
[R] Time-Series Representation Learning via Temporal and Contextual Contrasting
Abstract: Learning decent representations from unlabeled time-series data with temporal dynamics is a very challenging task. In this paper, we propose an unsupervised Time-Series representation learning framework via Temporal and Contextual Contrasting (TS-TCC), to learn time-series representation from unlabeled data. First, the raw time-series data are transformed into two different yet correlated views by using weak and strong augmentations. Second, we propose a novel temporal contrasting module to learn robust temporal representations by designing a tough cross-view prediction task. Last, to further learn discriminative representations, we propose a contextual contrasting module built upon the contexts from the temporal contrasting module. It attempts to maximize the similarity among different contexts of the same sample while minimizing similarity among contexts of different samples. Experiments have been carried out on three real-world time-series datasets. The results manifest that training a linear classifier on top of the features learned by our proposed TS-TCC performs comparably with the supervised training. Additionally, our proposed TS-TCC shows high efficiency in few- labeled data and transfer learning scenarios. The code is publicly available at this https URL.
What are some alternatives?
zshot - Zero and Few shot named entity & relationships recognition
thesis - MSc thesis on: Classifying brain activity using EEG and automated time tracking of computer use (using ActivityWatch)
stackoverflow-better-stats - Better statistics about Stack Overflow's 2023 Developer Survey
fooof - Parameterizing neural power spectra into periodic & aperiodic components.
PaddleHelix - Bio-Computing Platform Featuring Large-Scale Representation Learning and Multi-Task Deep Learning “螺旋桨”生物计算工具集
AdaTime - [TKDD 2023] AdaTime: A Benchmarking Suite for Domain Adaptation on Time Series Data
awesome-artificial-intelligence-research - A curated list of Artificial Intelligence (AI) Research, tracks the cutting edge trending of AI research, including recommender systems, computer vision, machine learning, etc.
eval_ssl_ssc - [TNSRE 2023] Self-supervised Learning for Label-Efficient Sleep Stage Classification: A Comprehensive Evaluation
Transfer-Learning-Library - Transfer Learning Library for Domain Adaptation, Task Adaptation, and Domain Generalization
Revisiting-Contrastive-SSL - Revisiting Contrastive Methods for Unsupervised Learning of Visual Representations. [NeurIPS 2021]
Efficient-VDVAE - Official Pytorch and JAX implementation of "Efficient-VDVAE: Less is more"
Awesome-SSL4TS - A professionally curated list of awesome resources (paper, code, data, etc.) on Self-Supervised Learning for Time Series (SSL4TS).