Recognition: no theorem link
A GAN and LLM-Driven Data Augmentation Framework for Dynamic Linguistic Pattern Modeling in Chinese Sarcasm Detection
Pith reviewed 2026-05-10 18:10 UTC · model grok-4.3
The pith
A GAN and LLM framework augments Chinese sarcasm data with user history to reach F1 scores of 0.9138 and 0.9151.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The authors show that training a GAN on Sina Weibo data and applying GPT-3.5 augmentation produces the SinaSarc dataset containing target comments, contextual information, and user historical behavior; feeding this multi-dimensional input into an extended BERT architecture lets the model capture dynamic linguistic patterns and implicit sarcastic cues, delivering F1 scores of 0.9138 on non-sarcastic comments and 0.9151 on sarcastic ones that exceed all prior state-of-the-art results.
What carries the argument
The GAN plus GPT-3.5 pipeline that synthesizes the SinaSarc dataset of comments, context, and user historical behavior, together with the multi-dimensional BERT extension that ingests these elements to model long-term personal language patterns.
If this is right
- The augmented SinaSarc dataset supplies the missing user-history dimension that lets models track how the same person expresses sarcasm across topics and time.
- The extended BERT architecture demonstrates that incorporating historical behavior improves detection of implicit sarcasm beyond what text-only models achieve.
- The overall framework simultaneously enlarges the available training data and advances the modeling technique for Chinese sarcasm detection.
Where Pith is reading between the lines
- The same augmentation approach could be reused on other low-resource social-media tasks that benefit from user context, such as irony detection or personalized sentiment analysis.
- If the synthetic patterns generalize, the method might reduce reliance on expensive manual labeling for any subjective language task where individual style matters.
- Evaluating the model on sarcasm from entirely new platforms or recent events would test whether the learned user patterns transfer beyond the original Weibo collection.
Load-bearing premise
The comments generated by the GAN and GPT-3.5 faithfully reproduce real users' ongoing linguistic habits without introducing systematic biases or artifacts that artificially raise the reported F1 scores.
What would settle it
Retraining and testing the model exclusively on untouched real Weibo comments from held-out users, without any synthetic augmentation, would show whether the claimed F1 scores persist or fall when the data no longer contains GAN or GPT artifacts.
Figures
read the original abstract
Sarcasm is a rhetorical device that expresses criticism or emphasizes characteristics of certain individuals or situations through exaggeration, irony, or comparison. Existing methods for Chinese sarcasm detection are constrained by limited datasets and high construction costs, and they mainly focus on textual features, overlooking user-specific linguistic patterns that shape how opinions and emotions are expressed. This paper proposes a Generative Adversarial Network (GAN) and Large Language Model (LLM)-driven data augmentation framework to dynamically model users' linguistic patterns for enhanced Chinese sarcasm detection. First, we collect raw data from various topics on Sina Weibo. Then, we train a GAN on these data and apply a GPT-3.5 based data augmentation technique to synthesize an extended sarcastic comment dataset, named SinaSarc. This dataset contains target comments, contextual information, and user historical behavior. Finally, we extend the BERT architecture to incorporate multi-dimensional information, particularly user historical behavior, enabling the model to capture dynamic linguistic patterns and uncover implicit sarcastic cues in comments. Experimental results demonstrate the effectiveness of our proposed method. Specifically, our model achieves the highest F1-scores on both the non-sarcastic and sarcastic categories, with values of 0.9138 and 0.9151 respectively, which outperforms all existing state-of-the-art (SOTA) approaches. This study presents a novel framework for dynamically modeling users' long-term linguistic patterns in Chinese sarcasm detection, contributing to both dataset construction and methodological advancement in this field.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper proposes a GAN and GPT-3.5-driven data augmentation pipeline to synthesize the SinaSarc dataset (target comments, context, and user histories from Sina Weibo), then extends BERT to a multi-dimensional architecture that ingests these elements to capture dynamic user linguistic patterns for Chinese sarcasm detection. It reports F1 scores of 0.9138 (non-sarcastic) and 0.9151 (sarcastic) that exceed all listed SOTA baselines.
Significance. If the synthetic data faithfully reproduces real Weibo distributions without detectable artifacts, the framework would offer a practical route to scale sarcasm detection in data-scarce settings while explicitly modeling user-specific history, a dimension largely absent from prior Chinese sarcasm work.
major comments (3)
- [Abstract and §4] Abstract and §4 (Experiments): the headline F1 scores of 0.9138/0.9151 are stated without baseline implementation details, statistical significance tests, or confidence intervals, so it is impossible to verify that the gains over SOTA are robust rather than the result of implementation differences or random variation.
- [§3] §3 (Framework): the GAN + GPT-3.5 augmentation step that produces user histories and comments contains no quantitative fidelity checks (e.g., n-gram distribution overlap, perplexity against real data, or human realism ratings), which is load-bearing because the central claim that the extended BERT learns genuine linguistic patterns rather than generator artifacts rests on this untested assumption.
- [§3.2 and §4.3] §3.2 and §4.3: the multi-dimensional BERT extension is described at the architectural level but lacks ablation results isolating the contribution of user-history features versus context or target text alone, leaving open whether the reported gains derive from the claimed dynamic pattern modeling.
minor comments (2)
- [Abstract] The abstract refers to 'various topics' without stating the number of topics, collection period, or filtering criteria used for the raw Sina Weibo crawl.
- [Model section] Notation for the extended BERT input (e.g., how history sequences are tokenized and fused) is introduced without an accompanying equation or diagram in the model section.
Simulated Author's Rebuttal
We thank the referee for the constructive comments, which highlight important areas for strengthening the empirical rigor of our work. We address each major comment below and will incorporate the suggested improvements in the revised manuscript.
read point-by-point responses
-
Referee: [Abstract and §4] Abstract and §4 (Experiments): the headline F1 scores of 0.9138/0.9151 are stated without baseline implementation details, statistical significance tests, or confidence intervals, so it is impossible to verify that the gains over SOTA are robust rather than the result of implementation differences or random variation.
Authors: We agree that the reported F1 scores require additional supporting details to allow independent verification. In the revision, we will expand §4 to include complete implementation details for all baselines (hyperparameters, training procedures, and code references where possible), perform statistical significance tests (e.g., McNemar's test or bootstrap resampling), and report 95% confidence intervals for the F1 scores on both classes. revision: yes
-
Referee: [§3] §3 (Framework): the GAN + GPT-3.5 augmentation step that produces user histories and comments contains no quantitative fidelity checks (e.g., n-gram distribution overlap, perplexity against real data, or human realism ratings), which is load-bearing because the central claim that the extended BERT learns genuine linguistic patterns rather than generator artifacts rests on this untested assumption.
Authors: We acknowledge that quantitative validation of the synthetic SinaSarc data is essential to support the claim that the model captures genuine patterns. The original submission omitted these checks. We will add them in the revised §3, including n-gram distribution overlap statistics, perplexity measurements against held-out real Weibo data, and aggregated human realism ratings from annotators. revision: yes
-
Referee: [§3.2 and §4.3] §3.2 and §4.3: the multi-dimensional BERT extension is described at the architectural level but lacks ablation results isolating the contribution of user-history features versus context or target text alone, leaving open whether the reported gains derive from the claimed dynamic pattern modeling.
Authors: We agree that ablation studies are needed to isolate the contribution of user-history features. In the revised §4.3, we will present ablation results comparing the full multi-dimensional model against variants that remove user-history inputs, context, or target text alone, thereby quantifying the added value of dynamic linguistic pattern modeling. revision: yes
Circularity Check
No circularity: empirical pipeline with no self-referential derivations
full rationale
The paper presents a data-augmentation pipeline (GAN + GPT-3.5 to synthesize SinaSarc) followed by an extended BERT classifier whose performance is measured by held-out F1 scores. No equations, uniqueness theorems, or fitted parameters are defined in terms of the target predictions; the reported F1 values (0.9138 / 0.9151) are obtained from standard train/test splits on the augmented corpus rather than by algebraic identity or self-citation. The framework description contains no self-definitional loops, no renaming of known results, and no load-bearing citations to prior work by the same authors that would close a circular chain. The central claims therefore rest on external experimental outcomes and are self-contained against the listed circularity patterns.
Axiom & Free-Parameter Ledger
axioms (1)
- domain assumption BERT can be extended with additional user-historical-behavior features to capture dynamic linguistic patterns.
Reference graph
Works this paper leans on
-
[1]
Sarcasm as contrast between a positive sentiment and negative situation,
E. Riloff, A. Qadir, P. Surveet al., “Sarcasm as contrast between a positive sentiment and negative situation,” inProceedings of the 18th Conference on Empirical Methods in Natural Language Processing, Seattle, W A, USA, October 2013, pp. 704–714
work page 2013
-
[2]
From humor recognition to irony detection: The figurative language of social media,
A. Reyes, P. Rosso, and D. Buscaldi, “From humor recognition to irony detection: The figurative language of social media,”Data & Knowledge Engineering, vol. 74, pp. 1–12, 2012
work page 2012
-
[3]
Irony detection in twitter: The role of affective content,
D. I. H. Far ´ıas, V . Patti, and P. Rosso, “Irony detection in twitter: The role of affective content,”ACM Transactions on Internet Technology, vol. 16, no. 3, pp. 1–24, 2016
work page 2016
-
[4]
Detecting ironic intent in creative comparisons,
T. Veale and Y . Hao, “Detecting ironic intent in creative comparisons,” in Proceedings of the 19th European Conference on Artificial Intelligence, 2010, pp. 765–770
work page 2010
-
[5]
M. Nirmala, A. H. Gandomi, M. R. Babuet al., “An emoticon- based novel sarcasm pattern detection strategy to identify sarcasm in microblogging social networks,”IEEE Transactions on Computational Social Systems, vol. 11, no. 4, pp. 5319–5326, 2023
work page 2023
-
[6]
Cascade: Contextual sarcasm detection in online discussion forums,
D. Hazarika, S. Poria, S. Gorantlaet al., “Cascade: Contextual sarcasm detection in online discussion forums,” inProceedings of the 27th International Conference on Computational Linguistics, Santa Fe, NM, USA, August 2018, pp. 1837–1848
work page 2018
-
[7]
Contextualized sarcasm detection on twit- ter,
D. Bamman and N. Smith, “Contextualized sarcasm detection on twit- ter,” inProceedings of the 9th International Conference on Web and Social Media, Oxford, UK, May 2015, pp. 574–577
work page 2015
-
[8]
Harnessing context in- congruity for sarcasm detection,
A. Joshi, V . Sharma, and P. Bhattacharyya, “Harnessing context in- congruity for sarcasm detection,” inProceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, July 2015, pp. 757–762
work page 2015
-
[9]
The perfect solution for detecting sarcasm in tweets# not,
C. C. Liebrecht, F. A. Kunneman, and A. P. J. Van Den Bosch, “The perfect solution for detecting sarcasm in tweets# not,” inProceedings of the 4th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, Atlanta, GA, USA, June 2013, pp. 29–37
work page 2013
-
[10]
A. Rodriguez, Y . L. Chen, and C. Argueta, “Sincere: A hybrid framework with graph-based compact textual models using emotion classification and sentiment analysis for twitter sarcasm detection,”IEEE Transactions on Computational Social Systems, vol. 11, no. 5, pp. 5593–5606, 2024
work page 2024
-
[11]
I. Touahri and A. Mazroui, “Enhancement of a multi-dialectal sentiment analysis system by the detection of the implied sarcastic features,” Knowledge-Based Systems, vol. 227, p. 107232, 2021
work page 2021
-
[12]
Sarcasm analysis using conversation context,
D. Ghosh, A. R. Fabbri, and S. Muresan, “Sarcasm analysis using conversation context,”Computational Linguistics, vol. 44, no. 4, pp. 755–792, 2018
work page 2018
-
[13]
Deepmsd: Advancing multimodal sar- casm detection through knowledge-augmented graph reasoning,
Y . Wei, H. Zhou, S. Yuanet al., “Deepmsd: Advancing multimodal sar- casm detection through knowledge-augmented graph reasoning,”IEEE Transactions on Circuits and Systems for Video Technology, vol. 35, no. 7, pp. 6413–6423, 2025
work page 2025
-
[14]
Mimicking the brain’s cognition of sarcasm from multidisciplines for twitter sarcasm detection,
F. Yao, X. Sun, H. Yuet al., “Mimicking the brain’s cognition of sarcasm from multidisciplines for twitter sarcasm detection,”IEEE Transactions on Neural Networks and Learning Systems, vol. 34, no. 1, pp. 228–242, 2021
work page 2021
-
[15]
Multi-modal sarcasm detection on social media via multi-granularity information fusion,
L. Ou and Z. Li, “Multi-modal sarcasm detection on social media via multi-granularity information fusion,”ACM Transactions on Multimedia Computing, Communications and Applications, vol. 21, no. 3, pp. 1–23, 2025
work page 2025
-
[16]
Multi-modal sarcasm detection via knowledge-aware focused graph convolutional networks,
X. Zhuang, F. Zhou, and Z. Li, “Multi-modal sarcasm detection via knowledge-aware focused graph convolutional networks,”ACM Trans- actions on Multimedia Computing, Communications and Applications, vol. 21, no. 5, pp. 1–22, 2025
work page 2025
-
[17]
Fine-grained semantic disentanglement network for multimodal sarcasm analysis,
J. Tang, B. Ni, F. Zhouet al., “Fine-grained semantic disentanglement network for multimodal sarcasm analysis,”ACM Transactions on Mul- timedia Computing, Communications and Applications, vol. 21, no. 6, pp. 1–22, 2025
work page 2025
-
[18]
S3 agent: Unlocking the power of vllm for zero-shot multi-modal sarcasm detection,
P. Wang, Y . Zhang, H. Feiet al., “S3 agent: Unlocking the power of vllm for zero-shot multi-modal sarcasm detection,”ACM Transactions on Multimedia Computing, Communications and Applications, vol. 21, no. 11, pp. 1–16, 2025
work page 2025
-
[19]
Self-adaptive representation learning model for multi-modal sentiment and sarcasm joint analysis,
Y . Zhang, Y . Yu, M. Wanget al., “Self-adaptive representation learning model for multi-modal sentiment and sarcasm joint analysis,”ACM Transactions on Multimedia Computing, Communications and Applica- tions, vol. 20, no. 5, pp. 1–17, 2024
work page 2024
-
[20]
A novel retrospective-reading model for detecting chinese sarcasm comments of online social network,
L. Zhang, X. Zhao, Q. Maoet al., “A novel retrospective-reading model for detecting chinese sarcasm comments of online social network,”IEEE Transactions on Computational Social Systems, vol. 12, no. 2, pp. 792– 806, 2025. 13
work page 2025
-
[21]
Y . Liu, Y . Zhang, and D. Song, “A quantum probability driven frame- work for joint multi-modal sarcasm, sentiment and emotion analysis,” IEEE Transactions on Affective Computing, vol. 15, no. 1, pp. 326–341, 2023
work page 2023
-
[22]
Elevating knowledge-enhanced entity and relationship understanding for sarcasm detection,
X. Wang, Y . Wang, D. Heet al., “Elevating knowledge-enhanced entity and relationship understanding for sarcasm detection,”IEEE Transactions on Knowledge and Data Engineering, vol. 37, no. 6, pp. 3356–3371, 2025
work page 2025
-
[23]
Clues for detecting irony in user-generated contents: Oh...!! it’s so easy ;-),
P. Carvalho, L. Sarmento, M. J. Silvaet al., “Clues for detecting irony in user-generated contents: Oh...!! it’s so easy ;-),” inProceedings of the 1st International CIKM Workshop on Topic-sentiment Analysis for Mass Opinion, New York, NY , USA, November 2009, pp. 53–56
work page 2009
-
[24]
Who cares about sarcastic tweets? investigating the impact of sarcasm on sentiment analysis,
D. G. Maynard and M. A. Greenwood, “Who cares about sarcastic tweets? investigating the impact of sarcasm on sentiment analysis,” in Proceedings of the 9th International Conference on Language Resources and Evaluation, Reykjavik, Iceland, May 2014, pp. 4238–4243
work page 2014
-
[25]
Semi-supervised recognition of sarcastic sentences in twitter and amazon,
D. Davidov, O. Tsur, and A. Rappoport, “Semi-supervised recognition of sarcastic sentences in twitter and amazon,” inProceedings of the 14th International Conference on Computational Natural Language Learning, Uppsala, Sweden, July 2010, pp. 107–116
work page 2010
-
[26]
A deeper look into sarcastic tweets using deep convolutional neural networks,
S. Poria, E. Cambria, D. Hazarikaet al., “A deeper look into sarcastic tweets using deep convolutional neural networks,” inProceedings of the 26th International Conference on Computational Linguistics: Technical Papers, Osaka, Japan, December 2016, pp. 1601–1612
work page 2016
-
[27]
Fracking sarcasm using neural network,
A. Ghosh and T. Veale, “Fracking sarcasm using neural network,” in Proceedings of the 7th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, San Diego, CA, USA, June 2016, pp. 161–169
work page 2016
-
[28]
Reasoning with sarcasm by reading in-between,
Y . Tay, L. A. Tuan, S. C. Huiet al., “Reasoning with sarcasm by reading in-between,” inProceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia, July 2018, pp. 1010–1020
work page 2018
-
[29]
Sarcasm detection using deep learning and ensemble learning,
P. Goel, R. Jain, A. Nayyaret al., “Sarcasm detection using deep learning and ensemble learning,”Multimedia Tools and Applications, vol. 81, no. 30, pp. 43 229–43 252, 2022
work page 2022
-
[30]
Affective representations for sarcasm detection,
A. Agrawal and A. An, “Affective representations for sarcasm detection,” inThe 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, July 2018, pp. 1029–1032
work page 2018
-
[31]
Humans require context to infer ironic intent (so computers probably do, too),
B. C. Wallace, L. Kertz, and E. Charniak, “Humans require context to infer ironic intent (so computers probably do, too),” inProceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, Baltimore, MD, USA, June 2014, pp. 512–516
work page 2014
-
[32]
B. C. Wallace and E. Charniak, “Sparse, contextually informed models for irony detection: Exploiting user communities, entities and sentiment,” inProceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, July 2015, pp. 1035– 1044
work page 2015
-
[33]
Affective and contextual em- bedding for sarcasm detection,
N. Babanejad, H. Davoudi, A. Anet al., “Affective and contextual em- bedding for sarcasm detection,” inProceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain (Online), December 2020, pp. 225–243
work page 2020
-
[34]
A novel hierarchical bert architecture for sarcasm detection,
H. Srivastava, V . Varshney, S. Kumariet al., “A novel hierarchical bert architecture for sarcasm detection,” inProceedings of the Second Workshop on Figurative Language Processing, Online, July 2020, pp. 93–97
work page 2020
-
[35]
A transformer- based approach to irony and sarcasm detection,
R. A. Potamias, G. Siolas, and A. G. Stafylopatis, “A transformer- based approach to irony and sarcasm detection,”Neural Computing and Applications, vol. 32, no. 23, pp. 17 309–17 320, 2020
work page 2020
-
[36]
S. Yuan, Y . Wei, H. Zhouet al., “Enhancing semantic awareness by sentimental constraint with automatic outlier masking for multimodal sarcasm detection,”IEEE Transactions on Multimedia, vol. 27, pp. 5376–5386, 2025
work page 2025
-
[37]
A. Phukan, S. Pal, and A. Ekbal, “Hybrid quantum-classical neural network for multimodal multitask sarcasm, emotion, and sentiment analysis,”IEEE Transactions on Computational Social Systems, vol. 11, no. 5, pp. 5740–5750, 2024
work page 2024
-
[38]
B. Liang, L. Gui, Y . Heet al., “Fusion and discrimination: A multimodal graph contrastive learning framework for multimodal sarcasm detection,” IEEE Transactions on Affective Computing, vol. 15, no. 4, pp. 1874– 1888, 2024
work page 2024
-
[39]
Chinese irony corpus construction and ironic structure analysis,
Y . Tang and H. H. Chen, “Chinese irony corpus construction and ironic structure analysis,” inProceedings of the 25th International Conference on Computational Linguistics: Technical Papers, 2014, pp. 1269–1278
work page 2014
-
[40]
Sarcasm detection in chinese using a crowdsourced corpus,
S. K. Lin and S. K. Hsieh, “Sarcasm detection in chinese using a crowdsourced corpus,” inProceedings of the 28th Conference on Computational Linguistics and Speech Processing, 2016, pp. 299–310
work page 2016
-
[41]
Ciron: a new benchmark dataset for chinese irony detection,
R. Xiang, X. Gao, Y . Longet al., “Ciron: a new benchmark dataset for chinese irony detection,” inProceedings of the Twelfth Language Resources and Evaluation Conference, 2020, pp. 5714–5720
work page 2020
-
[42]
The design and construction of a chinese sarcasm dataset,
X. Gong, Q. Zhao, J. Zhanget al., “The design and construction of a chinese sarcasm dataset,” inProceedings of the Twelfth Language Resources and Evaluation Conference, 2020, pp. 5034–5039
work page 2020
-
[43]
A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, “Attention is all you need,” inProceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, California, USA, 2017, p. 6000–6010
work page 2017
-
[44]
Improved training of wasserstein gans,
I. Gulrajani, F. Ahmed, M. Arjovsky, V . Dumoulin, and A. Courville, “Improved training of wasserstein gans,” inProceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, California, USA, 2017, p. 5769–5779
work page 2017
-
[45]
Bert: Pre-training of deep bidirectional transformers for language understanding,
J. Devlin, M. W. Chang, K. Leeet al., “Bert: Pre-training of deep bidirectional transformers for language understanding,” inProceedings of the 17th Conference of the North American Chapter of the Asso- ciation for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, June 2019, pp. 4171–4186
work page 2019
-
[46]
Convolutional neural networks for sentence classification,
Y . Kim, “Convolutional neural networks for sentence classification,” in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, Oct. 2014, pp. 1746–1751
work page 2014
-
[47]
Generating behavior features for cold- start spam review detection with adversarial learning,
X. Tang, T. Qian, and Z. You, “Generating behavior features for cold- start spam review detection with adversarial learning,”Information Sciences, vol. 526, pp. 274–288, 2020
work page 2020
-
[48]
Sarcasm detection in social media based on imbalanced classification,
P. Liu, W. Chen, G. Ouet al., “Sarcasm detection in social media based on imbalanced classification,” inProceedings of the 15th International Conference on Web-age Information Management, Macau, China, June 2014, pp. 459–471
work page 2014
-
[49]
Sememe knowledge and auxiliary information enhanced approach for sarcasm detection,
Z. Wen, L. Gui, Q. Wanget al., “Sememe knowledge and auxiliary information enhanced approach for sarcasm detection,”Information Processing & Management, vol. 59, no. 3, p. 102883, 2022
work page 2022
-
[50]
Sarcasm detection on twitter: A behavioral modeling approach,
A. Rajadesingan, R. Zafarani, and H. Liu, “Sarcasm detection on twitter: A behavioral modeling approach,” inProceedings of the 8th International Conference on Web Search and Data Mining, New York, NY , USA, February 2015, pp. 97–106
work page 2015
-
[51]
An improved random forest classifier for multi-class classification,
A. Chaudhary, S. Kolhe, and R. Kamal, “An improved random forest classifier for multi-class classification,”Information Processing in Agri- culture, vol. 3, no. 4, pp. 215–222, 2016
work page 2016
-
[52]
Explaining the success of adaboost and random forests as interpolating classifiers,
A. J. Wyner, M. Olson, J. Bleichet al., “Explaining the success of adaboost and random forests as interpolating classifiers,”Journal of Machine Learning Research, vol. 18, no. 48, pp. 1–33, 2017
work page 2017
-
[53]
Atalaya at semeval 2019 task 5: Robust embeddings for tweet classification,
J. M. P ´erez and F. M. Luque, “Atalaya at semeval 2019 task 5: Robust embeddings for tweet classification,” inProceedings of the 13th International Workshop on Semantic Evaluation, Minneapolis, MN, USA, June 2019, pp. 64–69
work page 2019
-
[54]
A hybrid transformer based model for sarcasm detection from news headlines,
A. Khan, D. Majumdar, and B. Mondal, “A hybrid transformer based model for sarcasm detection from news headlines,”Journal of Intelligent Information Systems, vol. 63, no. 4, pp. 1339–1359, 2025
work page 2025
-
[55]
Addressing unintended bias in toxicity detection: An lstm and attention-based approach,
W. Dai, J. Tao, X. Yanet al., “Addressing unintended bias in toxicity detection: An lstm and attention-based approach,” inProceedings of the 5th International Conference on Artificial Intelligence and Computer Applications, Dalian, China, November 2023, pp. 375–379
work page 2023
-
[56]
M. Neog and N. Baruah, “A deep learning framework for assamese toxic comment detection: Leveraging lstm and bilstm models with attention mechanism,” inProceedings of the 2nd International Conference on Advances in Data-driven Computing and Intelligent Systems, BITS Pilani, India, September 2023, pp. 485–497
work page 2023
-
[57]
Revisiting pre-trained models for chinese natural language processing,
Y . Cui, W. Che, T. Liuet al., “Revisiting pre-trained models for chinese natural language processing,” inFindings of the Association for Computational Linguistics: EMNLP, Online, November 2020, pp. 657–668
work page 2020
-
[58]
H. Madhu, S. Satapara, S. Modhaet al., “Detecting offensive speech in conversational code-mixed dialogue on social media: A contextual dataset and benchmark experiments,”Expert Systems with Applications, vol. 215, p. 119342, 2023
work page 2023
-
[59]
A simple and interactive transformer for fine-grained emotion detection,
D. Sui, B. Li, H. Yanget al., “A simple and interactive transformer for fine-grained emotion detection,”IEEE Transactions on Audio, Speech, and Language Processing, vol. 33, pp. 347–358, 2025
work page 2025
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.