Recognition: unknown
VTBench: A Multimodal Framework for Time-Series Classification with Chart-Based Representations
Pith reviewed 2026-05-07 10:30 UTC · model grok-4.3
The pith
Fusing simple chart images with raw time series data yields competitive classification accuracy when the visuals supply non-redundant cues.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
VTBench generates line, area, bar, and scatter plots from each time series and feeds them into modular fusion architectures alongside the raw sequence. On 31 UCR datasets, chart-only models prove competitive with raw-input baselines particularly on smaller sets; multi-chart combinations raise accuracy by supplying complementary visual patterns; and full multimodal fusion improves or preserves performance when the visual branch adds information absent from the numbers but lowers accuracy when the charts merely duplicate existing signals.
What carries the argument
The VTBench modular architecture that renders time series into line, area, bar, and scatter charts and supports single-chart, multi-chart, and full visual-numerical fusion strategies for classification.
If this is right
- Chart-only pipelines become a viable option for smaller datasets where raw-data deep models risk overfitting.
- Combining multiple chart types allows models to draw on different visual encodings of the same signal without additional preprocessing.
- Multimodal fusion should be applied selectively after checking that visual features do not duplicate information already present in the raw sequence.
- The distilled guidelines offer direct rules for choosing chart types, fusion method, and configuration on new time-series problems.
- VTBench supplies a reusable testbed for exploring other interpretable visual representations in time-series work.
Where Pith is reading between the lines
- The same chart-fusion idea could be tested on forecasting or anomaly detection tasks where visual trend or outlier cues in plots might add value.
- Adopting human-readable charts may lower reliance on specialized preprocessing pipelines that many current time-series models require.
- Pre-trained vision models could be swapped into the chart branch to see whether transfer learning amplifies the observed gains.
- Because the inputs remain directly viewable plots, the approach opens a route toward more inspectable time-series classifiers.
Load-bearing premise
The selected chart types deliver non-redundant visual features that standard fusion can exploit without injecting noise or needing extensive extra preprocessing.
What would settle it
A collection of UCR datasets on which every multimodal chart-plus-raw configuration produces lower accuracy than the raw-only baseline across all fusion variants and dataset sizes.
Figures
read the original abstract
Time-series classification (TSC) has advanced significantly with deep learning, yet most models rely solely on raw numerical inputs, overlooking alternative representations. While texture-based encodings such as Gramian Angular Fields (GAF) and Recurrence Plots (RP) convert time series into 2D images, they often require heavy preprocessing and yield less intuitive representations. In contrast, chart-based visualizations offer more interpretable alternatives and show promise in specific domains; however, their effectiveness remains underexplored, with limited systematic evaluation across chart types, visual encoding choices, and datasets. In this work, we introduce VTBench, a systematic and extensible framework that re-examines TSC through multimodal fusion of raw sequences and chart-based visualizations. VTBench generates lightweight, human-interpretable plots -- line, area, bar, and scatter, providing complementary views of the same signal. We develop a modular architecture supporting multiple fusion strategies, including single-chart visual-numerical fusion, multi-chart visual fusion, and full multimodal fusion with raw inputs. Through experiments across 31 UCR datasets, we show that: (1) chart-only models are competitive in selected settings, particularly on smaller datasets; (2) combining multiple chart types can improve accuracy by capturing complementary visual cues; and (3) multimodal models improve or maintain performance when visual features provide non-redundant information, but may degrade accuracy when they introduce redundancy. We further distill practical guidelines for selecting chart types, fusion strategies, and configurations. VTBench establishes a unified foundation for interpretable and effective multimodal time-series classification.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript introduces VTBench, a multimodal framework for time-series classification (TSC) that converts univariate series into lightweight chart visualizations (line, area, bar, scatter) and fuses them with raw numerical inputs via modular early/late/hybrid strategies. Experiments across 31 UCR datasets support three claims: (1) chart-only models are competitive, especially on smaller datasets; (2) multi-chart combinations improve accuracy by exploiting complementary visual cues; and (3) full multimodal fusion improves or maintains performance when visuals add non-redundant information but can degrade it when redundant. The work also distills practical guidelines for chart-type and fusion-strategy selection.
Significance. If the empirical patterns hold under rigorous controls, the contribution is moderately significant: it supplies a systematic, interpretable alternative to texture encodings such as GAF/RP and offers conditional evidence on when visual modalities help or hurt TSC. The practical guidelines and extensible modular design could influence follow-up work on multimodal time-series pipelines, though the absence of novel theory or large-scale gains limits broader impact.
major comments (2)
- [Experiments] Experimental section (results on 31 UCR datasets): the three numbered claims rest on comparisons of chart-only, multi-chart, and multimodal models, yet the manuscript supplies no architecture diagrams, layer counts, optimizer settings, learning-rate schedules, or number of runs with error bars. Without these, it is impossible to verify whether the reported accuracy patterns are robust or merely artifacts of under-specified training.
- [Fusion strategies] Fusion-strategy subsection: the claim that multimodal models 'improve or maintain performance when visual features provide non-redundant information' is load-bearing for claim (3), but the text does not quantify redundancy (e.g., via mutual information between chart embeddings and raw-series features) or show ablation results that isolate the redundancy effect from simple capacity differences.
minor comments (2)
- [Abstract] Abstract: the phrase 'lightweight, human-interpretable plots' is repeated without specifying chart resolution, axis scaling, or color-mapping choices; these parameters affect both reproducibility and the 'complementary visual cues' argument.
- [Introduction] Related-work paragraph: prior chart-based TSC papers (e.g., those using line plots with CNNs) are mentioned only in passing; a short table contrasting VTBench against them would clarify the incremental contribution.
Simulated Author's Rebuttal
Thank you for the constructive feedback on our manuscript. We address each major comment point by point below and describe the planned revisions to strengthen the work.
read point-by-point responses
-
Referee: [Experiments] Experimental section (results on 31 UCR datasets): the three numbered claims rest on comparisons of chart-only, multi-chart, and multimodal models, yet the manuscript supplies no architecture diagrams, layer counts, optimizer settings, learning-rate schedules, or number of runs with error bars. Without these, it is impossible to verify whether the reported accuracy patterns are robust or merely artifacts of under-specified training.
Authors: We agree that the experimental details provided are insufficient for full reproducibility and independent verification of robustness. The original manuscript emphasized the framework and aggregate results across the 31 UCR datasets but omitted granular implementation specifications. In the revised version we will add architecture diagrams for the visual encoders (ResNet-style for charts) and numerical encoder (1D CNN or Transformer), exact layer counts and dimensions, optimizer choice (Adam), learning-rate schedules, batch sizes, and training epochs. We will also report all accuracy figures as means and standard deviations over five independent runs with different random seeds. These changes will directly support verification of the three claims. revision: yes
-
Referee: [Fusion strategies] Fusion-strategy subsection: the claim that multimodal models 'improve or maintain performance when visual features provide non-redundant information' is load-bearing for claim (3), but the text does not quantify redundancy (e.g., via mutual information between chart embeddings and raw-series features) or show ablation results that isolate the redundancy effect from simple capacity differences.
Authors: We acknowledge that the current support for claim (3) relies primarily on empirical performance differences rather than explicit redundancy quantification. While the observed patterns across fusion strategies and chart types provide indicative evidence, we agree that stronger isolation is needed. In revision we will add (i) estimates of mutual information between the chart-derived embeddings and the raw time-series features and (ii) capacity-controlled ablations that match total parameter counts across fusion variants. These additions will better separate redundancy effects from mere increases in model capacity. revision: yes
Circularity Check
No significant circularity: purely empirical framework
full rationale
The paper introduces VTBench as an empirical multimodal TSC framework that generates standard chart visualizations (line/area/bar/scatter) from UCR time series and evaluates fusion strategies via controlled experiments on 31 public benchmarks. No derivation chain, equations, fitted parameters renamed as predictions, or load-bearing self-citations appear in the abstract or described methodology. All reported results (chart-only competitiveness on small data, multi-chart complementarity, conditional multimodal gains) are direct outcomes of the stated experimental design against external datasets, making the work self-contained without reduction to its own inputs.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
-
[1]
Alaa, Alex James Chan, and Mihaela van der Schaar
Ahmed M. Alaa, Alex James Chan, and Mihaela van der Schaar. 2021. Generative Time-series Modeling with Fourier Flows. In9th International Conference on Learning Representations, ICLR 2021. OpenReview.net. https://openreview.net/ forum?id=PpshD0AXfA
2021
-
[2]
Sarah Alnegheimish, Dongyu Liu, Carles Sala, Laure Berti-Équille, and Kalyan Veeramachaneni. 2022. Sintel: A Machine Learning Framework to Extract Insights from Signals. InSIGMOD 2022. ACM, 1855–1865. doi:10.1145/3514221.3517910
-
[3]
Anthony Bagnall, Jason Lines, Aaron Bostrom, James Large, and Eamonn Keogh
-
[4]
The great time series classification bake off: a review and experimental evaluation of recent algorithmic advances.Data mining and knowledge discovery 31 (2017), 606–660
2017
-
[5]
Yue Bai, Lichen Wang, Zhiqiang Tao, Sheng Li, and Yun Fu. 2021. Correlative Channel-Aware Fusion for Multi-View Time Series Classification. InThirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021. AAAI Press, 6714–6722. doi:10.1609/AAAI.V35I8.16830
-
[6]
Jun-Hao Chen and Yun-Cheng Tsai. 2020. Encoding candlesticks as images for pattern classification using convolutional neural networks.Financial Innovation 6, 1 (2020), 26
2020
-
[7]
Hoang Anh Dau, Anthony Bagnall, Kaveh Kamgar, Chin-Chia Michael Yeh, Yan Zhu, Shaghayegh Gharghabi, Chotirat Ann Ratanamahatana, and Eamonn Keogh
-
[8]
The UCR time series archive.IEEE/CAA Journal of Automatica Sinica6, 6 (2019), 1293–1305. doi:10.1109/JAS.2019.1911747
-
[9]
Angus Dempster, François Petitjean, and Geoffrey I Webb. 2020. ROCKET: excep- tionally fast and accurate time series classification using random convolutional kernels.Data Mining and Knowledge Discovery34, 5 (2020), 1454–1495
2020
-
[10]
Angus Dempster, Daniel F. Schmidt, and Geoffrey I. Webb. 2022. Hydra: compet- ing convolutional kernels for fast and accurate time series classification.Data Mining and Knowledge Discovery37 (2022), 1779–1805. doi:10.1007/S10618-023- 00939-3
-
[11]
Eskofier, and An Nguyen
Simon Dietz, Thomas Altstidl, Dario Zanca, Björn M. Eskofier, and An Nguyen
- [12]
-
[13]
Han Ding, Lei Guo, Cui Zhao, Fei Wang, Ge Wang, Zhiping Jiang, Wei Xi, and Jizhong Zhao. 2020. RFnet: Automatic gesture recognition and human identifi- cation using time series RFID signals.Mobile Networks and Applications25, 6 (2020), 2240–2253
2020
-
[14]
Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In9th International Conference on Learning Representations, ICLR 2...
2021
-
[15]
Schmidt, Jonathan Weber, Geoffrey I
Hassan Ismail Fawaz, Benjamin Lucas, Germain Forestier, Charlotte Pelletier, Daniel F. Schmidt, Jonathan Weber, Geoffrey I. Webb, Lhassane Idoumghar, Pierre- Alain Muller, and Franccois Petitjean. 2019. InceptionTime: Finding AlexNet for time series classification.Data Mining and Knowledge Discovery34 (2019), 1936 –
2019
-
[16]
doi:10.1007/S10618-020-00710-Y
-
[17]
Cristina Gómez, Joanne C White, and Michael A Wulder. 2016. Optical remotely sensed time series data for land cover classification: A review.ISPRS Journal of photogrammetry and Remote Sensing116 (2016), 55–72
2016
-
[18]
Nima Hatami, Yann Gavet, and Johan Debayle. 2017. Classification of time- series images using deep convolutional neural networks. InTenth Interna- tional Conference on Machine Vision, ICMV 2017, Antanas Verikas, Petia Radeva, Dmitry P. Nikolaev, and Jianhong Zhou (Eds.), Vol. 10696. SPIE, 106960Y. doi:10.1117/12.2309486
-
[19]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep Residual Learning for Image Recognition. In2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016. IEEE Computer Society, 770–778. doi:10.1109/ CVPR.2016.90
2016
-
[20]
Duy Tang Hoang, Xuan Toa Tran, Mien Van, and Hee Jun Kang. 2021. A deep neural network-based feature fusion for bearing fault diagnosis.Sensors21, 1 (2021), 244. doi:10.3390/S21010244
- [21]
-
[22]
Hassan Ismail Fawaz, Benjamin Lucas, Germain Forestier, Charlotte Pelletier, Daniel F Schmidt, Jonathan Weber, Geoffrey I Webb, Lhassane Idoumghar, Pierre- Alain Muller, and François Petitjean. 2020. Inceptiontime: Finding alexnet for time series classification.Data Mining and Knowledge Discovery34, 6 (2020), 1936–1962
2020
-
[23]
Xianya Jiang, Mo Hai, and Haifeng Li. 2019. Stock Classification Prediction Based on Spark. (2019), 243–250
2019
- [24]
-
[25]
Fazle Karim, Somshubra Majumdar, Houshang Darabi, and Shun Chen. 2018. LSTM Fully Convolutional Networks for Time Series Classification.IEEE Access 6 (2018), 1662–1669. doi:10.1109/ACCESS.2017.2779939
-
[26]
Jin-Hwa Kim, Kyoung Woon On, Woosang Lim, Jeonghee Kim, Jung-Woo Ha, and Byoung-Tak Zhang. 2017. Hadamard Product for Low-rank Bilinear Pooling. In5th International Conference on Learning Representations, ICLR 2017. https: //openreview.net/forum?id=r1rhWnZkg
2017
- [27]
-
[28]
Flynn, René Vidal, Austin Reiter, and Gregory D
Colin Lea, Michael D. Flynn, René Vidal, Austin Reiter, and Gregory D. Hager
-
[29]
Flynn, René Vidal, Austin Reiter, and Gregory D
Temporal Convolutional Networks for Action Segmentation and Detection. In2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017. 1003–1012. doi:10.1109/CVPR.2017.113
- [30]
-
[31]
Yingming Li, Ming Yang, and Zhongfei Zhang. 2018. A survey of multi-view representation learning.IEEE transactions on knowledge and data engineering31, 10 (2018), 1863–1883. doi:10.1109/TKDE.2018.2872063
-
[32]
Zekun Li, Shiyang Li, and Xifeng Yan. 2023. Time Series as Images: Vision Trans- former for Irregularly Sampled Time Series. InAdvances in Neural Information Processing Systems, NeurIPS 2023
2023
-
[33]
Yuxuan Liang, Haomin Wen, Yuqi Nie, Yushan Jiang, Ming Jin, Dongjin Song, Shirui Pan, and Qingsong Wen. 2024. Foundation models for time series analysis: A tutorial and survey. InProceedings of the 30th ACM SIGKDD conference on knowledge discovery and data mining. 6555–6565. doi:10.1145/3637528.3671451
-
[34]
Tsung-Yu Lin, Aruni RoyChowdhury, and Subhransu Maji. 2015. Bilinear CNN Models for Fine-Grained Visual Recognition. In2015 IEEE International Conference on Computer Vision, ICCV 2015. IEEE Computer Society, 1449–1457. doi:10.1109/ ICCV.2015.170
2015
-
[35]
Shiwei Liu, Liejun Wang, and Wenwen Yue. 2024. An efficient medical image classification network based on multi-branch CNN, token grouping Transformer and mixer MLP.Applied Soft Computing153 (2024), 111323. doi:10.1016/J.ASOC. 2024.111323
-
[36]
Xun Liu, Alex Hay-Man Ng, Linlin Ge, Fangyuan Lei, and Xuejiao Liao. 2024. Multi-branch Fusion: A Multi-branch Attention Framework by Combining Graph Convolutional Network and CNN for Hyperspectral Image Classification.IEEE Transactions on Geoscience and Remote Sensing(2024)
2024
-
[37]
Yong Liu, Tengge Hu, Haoran Zhang, Haixu Wu, Shiyu Wang, Lintao Ma, and Mingsheng Long. 2024. iTransformer: Inverted Transformers Are Effective for Time Series Forecasting. InThe Twelfth International Conference on Learning Representations, ICLR 2024. https://openreview.net/forum?id=JePfAI8fah
2024
-
[40]
Zhen Liu, Yicheng Luo, Boyuan Li, Emadeldeen Eldele, Min Wu, and Qianli Ma
- [41]
-
[42]
Zhengyi Liu, Song Shi, Quntao Duan, Wei Zhang, and Peng Zhao. 2019. Salient object detection for RGB-D image by single stream recurrent convolution neural network.Neurocomputing363 (2019), 46–57. doi:10.1016/J.NEUCOM.2019.07.012
-
[43]
Qianli Ma, Zhen Liu, Zhenjing Zheng, Ziyang Huang, Siying Zhu, Zhongzhong Yu, and James T. Kwok. 2024. A Survey on Time-Series Pre-Trained Models. IEEE Transactions on Knowledge and Data Engineering36, 12 (2024), 7536–7555. doi:10.1109/TKDE.2024.3475809
-
[44]
Maria Mariani, Prince Appiah, and Osei Tweneboah. 2025. Fusion of Recurrence Plots and Gramian Angular Fields with Bayesian Optimization for Enhanced Time-Series Classification.Axioms14, 7 (2025), 528
2025
-
[45]
Joty, and Enamul Hoque
Ahmed Masry, Do Xuan Long, Jia Qing Tan, Shafiq R. Joty, and Enamul Hoque
-
[46]
doi: 10.18653/v1/2022.findings-acl.177
ChartQA: A Benchmark for Question Answering about Charts with Vi- sual and Logical Reasoning. InFindings of the Association for Computational Linguistics: ACL 2022. Association for Computational Linguistics, 2263–2279. doi:10.18653/V1/2022.FINDINGS-ACL.177
-
[47]
Matthew Middlehurst, James Large, Michael Flynn, Jason Lines, Aaron Bostrom, and Anthony Bagnall. 2021. HIVE-COTE 2.0: a new meta ensemble for time series classification.Machine Learning110, 11 (2021), 3211–3243. doi:10.1007/S10994- 021-06057-9
-
[48]
Matthew Middlehurst, Patrick Schäfer, and Anthony Bagnall. 2024. Bake off redux: a review and experimental evaluation of recent time series classification algorithms.Data Mining and Knowledge Discovery38, 4 (2024), 1958–2031. doi:10.1007/S10618-024-01022-1
-
[49]
Arya Hadizadeh Moghaddam and Saeedeh Momtazi. 2021. Image processing meets time series analysis: Predicting Forex profitable technical pattern positions. Applied Soft Computing108 (2021), 107460. doi:10.1016/J.ASOC.2021.107460
-
[50]
Seo-Hyeong Park, Nur Suriza Syazwany, and Sang-Chul Lee. 2023. Meta-feature fusion for few-shot time series classification.IEEE Access11 (2023), 41400–41414. doi:10.1109/ACCESS.2023.3270493 Venkatesan et al
-
[51]
Boris Pyakillya, Natasha Kazachenko, and Nikolay Mikhailovsky. 2017. Deep learning for ECG classification. InJournal of physics: conference series, Vol. 913. IOP Publishing, 012004
2017
-
[52]
Wenqi Ren, Lin Ma, Jiawei Zhang, Jinshan Pan, Xiaochun Cao, Wei Liu, and Ming-Hsuan Yang. 2018. Gated Fusion Network for Single Image Dehazing. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018. Computer Vision Foundation / IEEE Computer Society, 3253–3261. doi:10.1109/ CVPR.2018.00343
- [53]
-
[54]
Alper Sarikaya, Michael Correll, Lyn Bartram, Melanie Tory, and Danyel Fisher. 2018. What Do We Talk About When We Talk About Dashboards? IEEE transactions on visualization and computer graphics25, 1 (2018), 682–692. doi:10.1109/TVCG.2018.2864903
-
[55]
Ahmed Shifaz, Charlotte Pelletier, François Petitjean, and Geoffrey I Webb. 2020. TS-CHIEF: a scalable and accurate forest algorithm for time series classification. Data Mining and Knowledge Discovery34, 3 (2020), 742–775
2020
-
[56]
Karen Simonyan and Andrew Zisserman. 2015. Very Deep Convolutional Net- works for Large-Scale Image Recognition. In3rd International Conference on Learning Representations, ICLR 2015. http://arxiv.org/abs/1409.1556
work page internal anchor Pith review arXiv 2015
-
[57]
Wensi Tang, Guodong Long, Lu Liu, Tianyi Zhou, Michael Blumenstein, and Jing Jiang. 2022. Omni-Scale CNNs: a simple and effective kernel size configuration for time series classification. InThe Tenth International Conference on Learning Representations, ICLR 2022. https://openreview.net/forum?id=PDYs7Z2XFGv
2022
-
[58]
Gomez, Lukasz Kaiser, and Illia Polosukhin
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. InAdvances in Neural Information Processing Systems, NeurIPS 2017. 5998–6008
2017
-
[59]
Yihe Wang, Nan Huang, Taida Li, Yujun Yan, and Xiang Zhang. 2024. Medformer: A Multi-Granularity Patching Transformer for Medical Time-Series Classification. InAdvances in Neural Information Processing Systems, NeurIPS 2024
2024
-
[60]
Zhiguang Wang and Tim Oates. 2015. Imaging Time-Series to Improve Clas- sification and Imputation. InProceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, IJCAI 2015. AAAI Press, 3939–3945. http://ijcai.org/Abstract/15/553
2015
-
[61]
Zhiguang Wang, Tim Oates, et al . 2015. Encoding time series as images for visual inspection and classification using tiled convolutional neural networks. In Workshops at the twenty-ninth AAAI conference on artificial intelligence, Vol. 1. Austin, TX, 1–7
2015
-
[62]
Xi Wei, Tianzhu Zhang, Yan Li, Yongdong Zhang, and Feng Wu. 2020. Multi- Modality Cross Attention Network for Image and Sentence Matching. In2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020. Computer Vision Foundation / IEEE, 10938–10947. doi:10.1109/CVPR42600.2020. 01095
-
[63]
Haixu Wu, Tengge Hu, Yong Liu, Hang Zhou, Jianmin Wang, and Mingsheng Long. 2023. TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis. InThe Eleventh International Conference on Learning Representations, ICLR 2023. https://openreview.net/forum?id=ju_Uqw384Oq
2023
-
[64]
Ke Xu, Jun Yuan, Yifang Wang, Cláudio T. Silva, and Enrico Bertini. 2021. mTSeer: Interactive Visual Exploration of Models on Multivariate Time-series Forecast. In CHI Conference on Human Factors in Computing Systems, CHI’21. ACM, 23:1–23:15. doi:10.1145/3411764.3445083
-
[65]
Yanbo Xu, Siddharth Biswal, Shriprasad R. Deshpande, Kevin O. Maher, and Jimeng Sun. 2018. RAIM: Recurrent Attentive and Intensive Model of Multimodal Patient Monitoring Data. InProceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2018. ACM, 2565–2573. doi:10.1145/3219819.3220051
-
[66]
Jin Zeng, Yanfeng Tong, Yunmu Huang, Qiong Yan, Wenxiu Sun, Jing Chen, and Yongtian Wang. 2019. Deep Surface Normal Estimation With Hierarchical RGB-D Fusion. InIEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019. Computer Vision Foundation / IEEE, 6153–6162. doi:10.1109/CVPR.2019.00631
- [67]
-
[68]
Ye Zhang, Yi Hou, Shilin Zhou, and Kewei Ouyang. 2020. Encoding Time Series as Multi-Scale Signed Recurrence Plots for Classification Using Fully Convolutional Networks.Sensors20, 14 (2020), 3818. doi:10.3390/S20143818
- [69]
-
[70]
Bolei Zhou, Yiyou Sun, David Bau, and Antonio Torralba. 2018. Interpretable Basis Decomposition for Visual Explanation. InComputer Vision - ECCV 2018 - 15th European Conference, Vol. 11212. Springer, 122–138. doi:10.1007/978-3-030- 01237-3_8
-
[71]
Tian Zhou, Peisong Niu, Liang Sun, Rong Jin, et al . 2023. One fits all: Power general time series analysis by pretrained lm.Advances in neural information processing systems36 (2023), 43322–43355. VTBench: A Multimodal Framework for Time-Series Classification with Chart-Based Representations A Experimental Setup A.1 Datasets To ensure a rigorous and fair...
2023
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.