Spaces:
Running
Running
| title: VisionTSpp | |
| emoji: π | |
| colorFrom: purple | |
| colorTo: purple | |
| python_version: 3.10.14 | |
| sdk: gradio | |
| # sdk_version: 5.44.1 | |
| sdk_version: 5.34.0 | |
| app_file: app.py | |
| pinned: false | |
| license: mit | |
| short_description: space for VisionTSpp | |
| pipeline_tag: time-series-forecasting | |
| # VisionTS++: Cross-Modal Time Series Foundation Model with Continual Pre-trained Visual Backbones | |
| This repository hosts the **VisionTS++** model, a state-of-the-art time series foundation model based on continual pre-training of a visual Masked AutoEncoder (MAE) on large-scale time series data. It excels in multivariate and probabilistic time series forecasting by bridging modality gaps between vision and time series data. | |
| The model was introduced in the paper: | |
| [**VisionTS++: Cross-Modal Time Series Foundation Model with Continual Pre-trained Vision Backbones**](https://arxiv.org/abs/2508.04379) | |
| Official GitHub repository: [https://github.com/HALF111/VisionTSpp](https://github.com/HALF111/VisionTSpp) | |
| Experience **VisionTS++** directly in your browser on this [Hugging Face Space](https://huggingface.co/spaces/Lefei/VisionTSpp)! You can upload your own custom time series CSV file for zero-shot forecasting. | |
| ## About | |
| VisionTS++ is built upon continual pre-training of a vision model on large-scale time series, addressing key discrepancies in cross-modal transfer from vision to time series. It introduces three key innovations: | |
| 1. **Vision-model-based filtering**: Identifies high-quality sequences to stabilize pre-training and mitigate the data-modality gap. | |
| 2. **Colorized multivariate conversion**: Encodes multivariate series as multi-subfigure RGB images to enhance cross-variate modeling. | |
| 3. **Multi-quantile forecasting**: Uses parallel reconstruction heads to generate quantile forecasts for probabilistic predictions without parametric assumptions. | |
| These innovations allow VisionTS++ to achieve state-of-the-art performance in both in-distribution and out-of-distribution forecasting, demonstrating that vision models can effectively generalize to Time Series Forecasting with appropriate adaptation. | |
| ## Installation | |
| The VisionTS++ model is available through the `visionts` package on PyPI. | |
| First, install the package: | |
| ```shell | |
| pip install visionts | |
| ``` | |
| If you want to develop the inference code, you can also build from source: | |
| ```shell | |
| git clone https://github.com/HALF111/VisionTSpp.git | |
| cd VisionTSpp | |
| pip install -e . | |
| ``` | |
| For detailed inference examples and usage with clear visualizations of image reconstruction, please refer to the `demo.ipynb` notebook in the [official GitHub repository](https://github.com/HALF111/VisionTSpp/blob/main/demo.ipynb). | |
| ## Citation | |
| If you're using VisionTS++ or VisionTS in your research or applications, please cite them using this BibTeX: | |
| ```bibtex | |
| @misc{chen2024visionts, | |
| title={VisionTS: Visual Masked Autoencoders Are Free-Lunch Zero-Shot Time Series Forecasters}, | |
| author={Mouxiang Chen and Lefei Shen and Zhuo Li and Xiaoyun Joy Wang and Jianling Sun and Chenghao Liu}, | |
| year={2024}, | |
| eprint={2408.17253}, | |
| archivePrefix={arXiv}, | |
| url={https://arxiv.org/abs/2408.17253}, | |
| } | |
| @misc{shen2025visiontspp, | |
| title={VisionTS++: Cross-Modal Time Series Foundation Model with Continual Pre-trained Visual Backbones}, | |
| author={Lefei Shen and Mouxiang Chen and Xu Liu and Han Fu and Xiaoxue Ren and Jianling Sun and Zhuo Li and Chenghao Liu}, | |
| year={2025}, | |
| eprint={2508.04379}, | |
| archivePrefix={arXiv}, | |
| primaryClass={cs.CV}, | |
| url={https://arxiv.org/abs/2508.04379}, | |
| } | |
| ``` | |