pith. machine review for the scientific record. sign in

arxiv: 2605.02712 · v1 · submitted 2026-05-04 · 💻 cs.CL · cs.AI

Recognition: 2 theorem links

· Lean Theorem

mdok-style at SemEval-2026 Task 10: Finetuning LLMs for Conspiracy Detection

Authors on Pith no claims yet

Pith reviewed 2026-05-08 18:45 UTC · model grok-4.3

classification 💻 cs.CL cs.AI
keywords conspiracy detectiondata augmentationself-traininglarge language modelsfinetuningtext classificationSemEvalReddit comments
0
0 comments X

The pith

Data augmentation and self-training let a large language model rank eighth for spotting conspiracy beliefs in Reddit comments

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper demonstrates that data augmentation and self-training, techniques originally used for machine-generated text detection, can be applied to finetune the Qwen3-32B model for binary classification of conspiracy beliefs expressed in Reddit comments. This approach tackles the issue of limited labeled training data in SemEval-2026 Task 10. The resulting system places eighth out of 52 submissions, showing solid competitive performance. A sympathetic reader would care because it offers a concrete way to adapt existing LLM adaptation methods to new problems in analyzing online beliefs without needing enormous amounts of new labeled examples.

Core claim

The submitted mdok-style system utilizes data augmentation and self-training to finetune the Qwen3-32B model for a binary text-classification task. The submitted system is very competitive, ranking in the 85th percentile (8th out of 52 submissions). The results show that the approach, which originated in machine-generated text detection, can be used for conspiracy detection as well.

What carries the argument

Data augmentation and self-training to finetune the Qwen3-32B large language model for binary classification of conspiracy beliefs in Reddit comments

Load-bearing premise

The data augmentation and self-training steps do not introduce systematic biases or label noise that would inflate performance on the specific test set used in the shared task.

What would settle it

Testing the finetuned model on a new collection of independently labeled Reddit comments about conspiracies and finding accuracy no higher than a basic classifier trained only on the original small set would show the augmentation and self-training steps are not adding real value.

read the original abstract

SemEval-2026 Task 10 is focused on conspiracy detection. Specifically, the goal is to detect whether a Reddit comment expresses a conspiracy belief. Our submitted mdok-style system utilizes data augmentation and self-training (to cope with a rather small amount of training data) to finetune the Qwen3-32B model for a binary text-classification task. The submitted system is very competitive, ranking in the 85th percentile (8th out of 52 submissions). The results shown that our approach, which originated in machine-generated text detection, can be used for conspiracy detection as well.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 1 minor

Summary. The manuscript describes a submission to SemEval-2026 Task 10 on binary conspiracy detection in Reddit comments. The authors apply an mdok-style pipeline of data augmentation and self-training to fine-tune Qwen3-32B, reporting an 8th-place ranking out of 52 submissions and claiming that the approach, originally developed for machine-generated text detection, transfers successfully to this task.

Significance. A substantiated demonstration that data augmentation and self-training from MGT detection generalize to conspiracy detection would be useful for low-resource binary text classification. The reported ranking provides a positive empirical signal, but the absence of any supporting analyses means the transfer claim cannot be evaluated on its merits.

major comments (2)
  1. The manuscript supplies no implementation details, ablation studies, or error analysis for the data augmentation and self-training steps. Without these, the 8/52 ranking cannot be attributed to the mdok-style pipeline rather than the base Qwen3-32B capacity or incidental test-set alignment, undermining the central transfer claim stated in the abstract.
  2. No discussion or validation is provided on how self-training pseudo-labels were generated or filtered (e.g., confidence thresholds or external validation set), leaving open the possibility that performance reflects reinforcement of label distributions in the small seed data rather than genuine cross-task transfer.
minor comments (1)
  1. Grammatical error: 'The results shown that our approach...' should read 'The results show that our approach...'.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive feedback on our SemEval-2026 Task 10 submission. We address each major comment below and will revise the manuscript to provide the requested clarifications and supporting details, thereby strengthening the presentation of the mdok-style pipeline's transfer to conspiracy detection.

read point-by-point responses
  1. Referee: The manuscript supplies no implementation details, ablation studies, or error analysis for the data augmentation and self-training steps. Without these, the 8/52 ranking cannot be attributed to the mdok-style pipeline rather than the base Qwen3-32B capacity or incidental test-set alignment, undermining the central transfer claim stated in the abstract.

    Authors: We agree that the current manuscript lacks these elements, which limits evaluation of the pipeline's specific contribution. As a concise SemEval system paper, we prioritized the end-to-end description and competition ranking over extensive analysis. The 8th-place result (out of 52) was achieved with limited training data, where the mdok-style combination of augmentation and self-training was intended to compensate for data scarcity, consistent with its prior use in MGT detection. In the revised manuscript we will add a dedicated methods subsection with implementation details for both steps, a brief error analysis of misclassified development-set examples, and a discussion of why the base model alone is unlikely to explain the ranking given the small seed set size. Full ablations remain infeasible within the revision timeline due to the scale of the 32B model, but the added material will allow readers to assess the transfer claim more directly. revision: yes

  2. Referee: No discussion or validation is provided on how self-training pseudo-labels were generated or filtered (e.g., confidence thresholds or external validation set), leaving open the possibility that performance reflects reinforcement of label distributions in the small seed data rather than genuine cross-task transfer.

    Authors: We acknowledge the omission of this procedural detail, which is important for ruling out simple label reinforcement. The self-training component follows the iterative pseudo-labeling approach from the original mdok work, applied here to unlabeled Reddit comments. In the revised version we will explicitly describe the pseudo-label generation process, the confidence-based selection criterion used to retain reliable examples, and any use of a held-out validation set to monitor distribution shift across iterations. This addition will directly address the concern and better demonstrate that the performance gain arises from the pipeline's ability to leverage the model's cross-task generalization rather than from the seed data alone. revision: yes

Circularity Check

0 steps flagged

No circularity: empirical shared-task ranking with no derivation or fitted-parameter predictions

full rationale

The paper describes an applied system for SemEval-2026 Task 10: finetuning Qwen3-32B via data augmentation and self-training on a small labeled seed, then reporting an 8/52 ranking on the official test set. No equations, first-principles derivations, or parameter-fitting steps are present. The central claim is an empirical performance result from a public competition, not a prediction derived from inputs by construction. Self-citation to prior MGT-detection work is mentioned but is not load-bearing for any mathematical result; the ranking itself is externally verifiable against the shared-task leaderboard. No self-definitional loops, fitted-input-as-prediction, or ansatz smuggling occur.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

No mathematical derivations, free parameters, axioms, or invented entities are present; the paper is a purely empirical system description for a shared task.

pith-pipeline@v0.9.0 · 5393 in / 1046 out tokens · 28631 ms · 2026-05-08T18:45:02.078076+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Reference graph

Works this paper leans on

300 extracted references · 149 canonical work pages

  1. [1]

    S em E val-2026 Task 10: P sy C o M ark -- Psycholinguistic Conspiracy Marker Extraction and Detection

    Samory, Mattia and Soldner, Felix and Batzdorfer, Veronika. S em E val-2026 Task 10: P sy C o M ark -- Psycholinguistic Conspiracy Marker Extraction and Detection. Proceedings of the 20th International Workshop on Semantic Evaluation (SemEval-2026). 2026

  2. [2]

    Machine learning , volume=

    Support-vector networks , author=. Machine learning , volume=. 1995 , publisher=

  3. [3]

    Journal of Information Science , volume=

    Detection of conspiracy propagators using psycho-linguistic characteristics , author=. Journal of Information Science , volume=. 2023 , publisher=

  4. [4]

    arXiv preprint arXiv:2403.06765 , year=

    Conspemollm: Conspiracy theory detection using an emotion-based large language model , author=. arXiv preprint arXiv:2403.06765 , year=

  5. [5]

    Self-training: A survey , journal =

    Massih-Reza Amini and Vasilii Feofanov and Loïc Pauletto and Liès Hadjadj and Émilie Devijver and Yury Maximov , keywords =. Self-training: A survey , journal =. 2025 , issn =. doi:https://doi.org/10.1016/j.neucom.2024.128904 , url =

  6. [6]

    2025 , eprint=

    Increasing the Robustness of the Fine-tuned Multilingual Machine-Generated Text Detectors , author=. 2025 , eprint=

  7. [7]

    Dominik Macko , year=. mdok of. 2506.01702 , archivePrefix=

  8. [8]

    Voight-Kampff

    Janek Bevendorff and Yuxia Wang and Jussi Karlgren and Matti Wiegmann and Maik Fr. Overview of the "Voight-Kampff" Generative. Working Notes of

  9. [9]

    Working Notes of

    Dominik Macko , title =. Working Notes of

  10. [10]

    2025 , eprint=

    Authorship Attribution in Multilingual Machine-Generated Texts , author=. 2025 , eprint=

  11. [11]

    Dettmers, Tim and Pagnoni, Artidoro and Holtzman, Ari and Zettlemoyer, Luke , booktitle =

  12. [12]

    2025 , eprint=

    Gemma 3 Technical Report , author=. 2025 , eprint=

  13. [13]

    2025 , eprint=

    Qwen3 Technical Report , author=. 2025 , eprint=

  14. [14]

    Journal of personality and social psychology , volume=

    Evidence for universality and cultural variation of differential emotion response patterning , author=. Journal of personality and social psychology , volume=. 1994 , publisher=

  15. [15]

    Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems. 2025

  16. [16]

    Research on LLM s-Empowered Conversational AI for Sustainable Behaviour Change

    Chen, Ben. Research on LLM s-Empowered Conversational AI for Sustainable Behaviour Change. Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems. 2025

  17. [17]

    Deep Reinforcement Learning of LLM s​ using RLHF

    Levandovsky, Enoch. Deep Reinforcement Learning of LLM s​ using RLHF. Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems. 2025

  18. [18]

    Conversational Collaborative Robots

    Kranti, Chalamalasetti. Conversational Collaborative Robots. Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems. 2025

  19. [19]

    Dialogue System using Large Language Model-based Dynamic Slot Generation

    Hashimoto, Ekai. Dialogue System using Large Language Model-based Dynamic Slot Generation. Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems. 2025

  20. [20]

    Towards Adaptive Human-Agent Collaboration in Real-Time Environments

    Nakae, Kaito. Towards Adaptive Human-Agent Collaboration in Real-Time Environments. Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems. 2025

  21. [21]

    Towards Human-Like Dialogue Systems: Integrating Multimodal Emotion Recognition and Non-Verbal Cue Generation

    Jiang, Jingjing. Towards Human-Like Dialogue Systems: Integrating Multimodal Emotion Recognition and Non-Verbal Cue Generation. Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems. 2025

  22. [22]

    Controlling Dialogue Systems with Graph-Based Structures

    Hilgendorf, Laetitia Mina. Controlling Dialogue Systems with Graph-Based Structures. Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems. 2025

  23. [23]

    Multimodal Agentic Dialogue Systems for Situated Human-Robot Interaction

    Sucal, Virgile. Multimodal Agentic Dialogue Systems for Situated Human-Robot Interaction. Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems. 2025

  24. [24]

    Knowledge Graphs and Representational Models for Dialogue Systems

    Walker, Nicholas Thomas. Knowledge Graphs and Representational Models for Dialogue Systems. Proceedings of the 21st Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems. 2025

  25. [25]

    Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.0

  26. [26]

    Fine-Tuning Large Language Models for Relation Extraction within a Retrieval-Augmented Generation Framework

    Efeoglu, Sefika and Paschke, Adrian. Fine-Tuning Large Language Models for Relation Extraction within a Retrieval-Augmented Generation Framework. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.1

  27. [27]

    Benchmarking Table Extraction: Multimodal LLM s vs Traditional OCR

    Nunes, Guilherme and Rolla, Vitor and Pereira, Duarte and Alves, Vasco and Carreiro, Andre and Baptista, M \'a rcia. Benchmarking Table Extraction: Multimodal LLM s vs Traditional OCR. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.2

  28. [28]

    Injecting Structured Knowledge into LLM s via Graph Neural Networks

    Li, Zichao and Ke, Zong and Zhao, Puning. Injecting Structured Knowledge into LLM s via Graph Neural Networks. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.3

  29. [29]

    Regular-pattern-sensitive CRF s for Distant Label Interactions

    Papay, Sean and Klinger, Roman and Pad \'o , Sebastian. Regular-pattern-sensitive CRF s for Distant Label Interactions. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.4

  30. [30]

    From Syntax to Semantics: Evaluating the Impact of Linguistic Structures on LLM -Based Information Extraction

    Swarup, Anushka and Bhandarkar, Avanti and Wilson, Ronald and Pan, Tianyu and Woodard, Damon. From Syntax to Semantics: Evaluating the Impact of Linguistic Structures on LLM -Based Information Extraction. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.5

  31. [31]

    Detecting Referring Expressions in Visually Grounded Dialogue with Autoregressive Language Models

    Willemsen, Bram and Skantze, Gabriel. Detecting Referring Expressions in Visually Grounded Dialogue with Autoregressive Language Models. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.6

  32. [32]

    Exploring Multilingual Probing in Large Language Models: A Cross-Language Analysis

    Li, Daoyang and Zhao, Haiyan and Zeng, Qingcheng and Du, Mengnan. Exploring Multilingual Probing in Large Language Models: A Cross-Language Analysis. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.7

  33. [33]

    Self-Contrastive Loop of Thought Method for Text-to- SQL Based on Large Language Model

    Kang, Fengrui and Tan, Mingxi and Huang, Xianying and Yang, Shiju. Self-Contrastive Loop of Thought Method for Text-to- SQL Based on Large Language Model. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.8

  34. [34]

    Combining Automated and Manual Data for Effective Downstream Fine-Tuning of Transformers for Low-Resource Language Applications

    Isaeva, Ulyana and Astafurov, Danil and Martynov, Nikita. Combining Automated and Manual Data for Effective Downstream Fine-Tuning of Transformers for Low-Resource Language Applications. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.9

  35. [35]

    Seamlessly Integrating Tree-Based Positional Embeddings into Transformer Models for Source Code Representation

    Bartkowiak, Patryk and Grali \'n ski, Filip. Seamlessly Integrating Tree-Based Positional Embeddings into Transformer Models for Source Code Representation. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.10

  36. [36]

    Enhancing AMR Parsing with Group Relative Policy Optimization

    Barta, Botond and Hamerlik, Endre and Nyist, Mil \'a n and Ito, Masato and Acs, Judit. Enhancing AMR Parsing with Group Relative Policy Optimization. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.11

  37. [37]

    Structure Modeling Approach for UD Parsing of Historical M odern J apanese

    Ozaki, Hiroaki and Omura, Mai and Komiya, Kanako and Asahara, Masayuki and Ogiso, Toshinobu. Structure Modeling Approach for UD Parsing of Historical M odern J apanese. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.12

  38. [38]

    BARTABSA ++: Revisiting BARTABSA with Decoder LLM s

    Pfister, Jan and V. BARTABSA ++: Revisiting BARTABSA with Decoder LLM s. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.13

  39. [39]

    Typed- RAG : Type-Aware Decomposition of Non-Factoid Questions for Retrieval-Augmented Generation

    Lee, DongGeon and Park, Ahjeong and Lee, Hyeri and Nam, Hyeonseo and Maeng, Yunho. Typed- RAG : Type-Aware Decomposition of Non-Factoid Questions for Retrieval-Augmented Generation. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.14

  40. [40]

    Do we still need Human Annotators? Prompting Large Language Models for Aspect Sentiment Quad Prediction

    Hellwig, Nils Constantin and Fehle, Jakob and Kruschwitz, Udo and Wolff, Christian. Do we still need Human Annotators? Prompting Large Language Models for Aspect Sentiment Quad Prediction. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.15

  41. [41]

    Can LLM s Interpret and Leverage Structured Linguistic Representations? A Case Study with AMR s

    Raut, Ankush and Zhu, Xiaofeng and Pacheco, Maria Leonor. Can LLM s Interpret and Leverage Structured Linguistic Representations? A Case Study with AMR s. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.16

  42. [42]

    LLM Dependency Parsing with In-Context Rules

    Ginn, Michael and Palmer, Alexis. LLM Dependency Parsing with In-Context Rules. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.17

  43. [43]

    Cognitive Mirroring for D oc RE : A Self-Supervised Iterative Reflection Framework with Triplet-Centric Explicit and Implicit Feedback

    Han, Xu and Wang, Bo and Sun, Yueheng and Zhao, Dongming and Qu, Zongfeng and He, Ruifang and Hou, Yuexian and Hu, Qinghua. Cognitive Mirroring for D oc RE : A Self-Supervised Iterative Reflection Framework with Triplet-Centric Explicit and Implicit Feedback. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025)...

  44. [44]

    Cross-Document Event-Keyed Summarization

    Walden, William and Kuchmiichuk, Pavlo and Martin, Alexander and Jin, Chihsheng and Cao, Angela and Sun, Claire and Allen, Curisia and White, Aaron. Cross-Document Event-Keyed Summarization. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.19

  45. [45]

    Transfer of Structural Knowledge from Synthetic Languages

    Budnikov, Mikhail and Yamshchikov, Ivan. Transfer of Structural Knowledge from Synthetic Languages. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.20

  46. [46]

    Language Models are Universal Embedders

    Zhang, Xin and Li, Zehan and Zhang, Yanzhao and Long, Dingkun and Xie, Pengjun and Zhang, Meishan and Zhang, Min. Language Models are Universal Embedders. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.21

  47. [47]

    D ia DP @ XLLM 25: Advancing C hinese Dialogue Parsing via Unified Pretrained Language Models and Biaffine Dependency Scoring

    Duan, Shuoqiu and Chen, Xiaoliang and Miao, Duoqian and Gu, Xu and Li, Xianyong and Du, Yajun. D ia DP @ XLLM 25: Advancing C hinese Dialogue Parsing via Unified Pretrained Language Models and Biaffine Dependency Scoring. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.22

  48. [48]

    LLMSR @ XLLM 25: Less is More: Enhancing Structured Multi-Agent Reasoning via Quality-Guided Distillation

    Yuan, Jiahao and Sun, Xingzhe and Yu, Xing and Wang, Jingwen and Du, Dehui and Cui, Zhiqing and Di, Zixiang. LLMSR @ XLLM 25: Less is More: Enhancing Structured Multi-Agent Reasoning via Quality-Guided Distillation. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.23

  49. [49]

    S peech EE @ XLLM 25: End-to-End Structured Event Extraction from Speech

    Chaudhuri, Soham and Biswas, Diganta and Saha, Dipanjan and Das, Dipankar and Bandyopadhyay, Sivaji. S peech EE @ XLLM 25: End-to-End Structured Event Extraction from Speech. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.24

  50. [50]

    Luu, Son and Van Nguyen, Kiet

    Pham Hoang Le, Nguyen and Dinh Thien, An and T. Luu, Son and Van Nguyen, Kiet. D oc IE @ XLLM 25: Z ero S emble - Robust and Efficient Zero-Shot Document Information Extraction with Heterogeneous Large Language Model Ensembles. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.25

  51. [51]

    D oc IE @ XLLM 25: In-Context Learning for Information Extraction using Fully Synthetic Demonstrations

    Popovic, Nicholas and Kangen, Ashish and Schopf, Tim and F. D oc IE @ XLLM 25: In-Context Learning for Information Extraction using Fully Synthetic Demonstrations. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.26

  52. [52]

    LLMSR @ XLLM 25: Integrating Reasoning Prompt Strategies with Structural Prompt Formats for Enhanced Logical Inference

    Tai, Le and Van, Thin. LLMSR @ XLLM 25: Integrating Reasoning Prompt Strategies with Structural Prompt Formats for Enhanced Logical Inference. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.27

  53. [53]

    D oc IE @ XLLM 25: UIEP rompter: A Unified Training-Free Framework for universal document-level information extraction via Structured Prompt

    Qiu, Chengfeng and Zhou, Lifeng and Wei, Kaifeng and Li, Yuke. D oc IE @ XLLM 25: UIEP rompter: A Unified Training-Free Framework for universal document-level information extraction via Structured Prompt. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.28

  54. [54]

    LLMSR @ XLLM 25: SWRV : Empowering Self-Verification of Small Language Models through Step-wise Reasoning and Verification

    Chen, Danchun. LLMSR @ XLLM 25: SWRV : Empowering Self-Verification of Small Language Models through Step-wise Reasoning and Verification. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.29

  55. [55]

    LLMSR @ XLLM 25: An Empirical Study of LLM for Structural Reasoning

    Li, Xinye and Wan, Mingqi and Sui, Dianbo. LLMSR @ XLLM 25: An Empirical Study of LLM for Structural Reasoning. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.30

  56. [56]

    LLMSR @ XLLM 25: A Language Model-Based Pipeline for Structured Reasoning Data Construction

    Xing, Hongrui and Liu, Xinzhang and Jiang, Zhuo and Yang, Zhihao and Yao, Yitong and Wang, Zihan and Deng, Wenmin and Wang, Chao and Song, Shuangyong and Yang, Wang and He, Zhongjiang and Li, Yongxiang. LLMSR @ XLLM 25: A Language Model-Based Pipeline for Structured Reasoning Data Construction. Proceedings of the 1st Joint Workshop on Large Language Model...

  57. [57]

    S peech EE @ XLLM 25: Retrieval-Enhanced Few-Shot Prompting for Speech Event Extraction

    Gedeon, M \'a t \'e. S peech EE @ XLLM 25: Retrieval-Enhanced Few-Shot Prompting for Speech Event Extraction. Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025). 2025. doi:10.18653/v1/2025.xllm-1.32

  58. [58]

    Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  59. [59]

    An introduction to computational identification and classification of Upam \= a alaṇk \= a ra

    Jadhav, Bhakti and Dutta, Himanshu and Kanitkar, Shruti and Kulkarni, Malhar and Bhattacharyya, Pushpak. An introduction to computational identification and classification of Upam \= a alaṇk \= a ra. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  60. [60]

    Aesthetics of S anskrit Poetry from the Perspective of Computational Linguistics: A Case Study Analysis on \'S ikṣ \= a ṣṭaka

    Sandhan, Jivnesh and Barbadikar, Amruta and Maity, Malay and Satuluri, Pavankumar and Sandhan, Tushar and Gupta, Ravi M and Goyal, Pawan and Behera, Laxmidhar. Aesthetics of S anskrit Poetry from the Perspective of Computational Linguistics: A Case Study Analysis on \'S ikṣ \= a ṣṭaka. Computational Sanskrit and Digital Humanities - World Sanskrit Confere...

  61. [61]

    Itaretara Dvandva: A challenge for Dependency Tree semantics

    Kulkarni, Amba and Neelamana, Vasudha. Itaretara Dvandva: A challenge for Dependency Tree semantics. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  62. [62]

    A Case Study of Handwritten Text Recognition from Pre-Colonial era S anskrit Manuscripts

    Chincholikar, Kartik and Dwivedi, Shagun and Gopalan, Kaushik and Awasthi, Tarinee. A Case Study of Handwritten Text Recognition from Pre-Colonial era S anskrit Manuscripts. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  63. [63]

    Towards Accent-Aware V edic S anskrit Optical Character Recognition Based on Transformer Models

    Tsukagoshi, Yuzuki and Kuroiwa, Ryo and Ohmukai, Ikki. Towards Accent-Aware V edic S anskrit Optical Character Recognition Based on Transformer Models. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  64. [64]

    Vedavani: A Benchmark Corpus for ASR on V edic S anskrit Poetry

    Kumar, Sujeet and Ray, Pretam and Beerukuri, Abhinay and Kamoji, Shrey and Jagadeeshan, Manoj Balaji and Goyal, Pawan. Vedavani: A Benchmark Corpus for ASR on V edic S anskrit Poetry. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  65. [65]

    Compound Type Identification in S anskrit

    Krishnan, Sriram and Satuluri, Pavankumar and Barbadikar, Amruta and Prasanna Venkatesh, T S and Kulkarni, Amba. Compound Type Identification in S anskrit. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  66. [66]

    IKML : A Markup Language for Collaborative Semantic Annotation of I ndic Texts

    Lakkundi, Chaitanya S and Rajaraman, Gopalakrishnan and Susarla, Sai Rama Krishna. IKML : A Markup Language for Collaborative Semantic Annotation of I ndic Texts. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  67. [67]

    Challenges in Processing V edic S anskrit: Towards creating a normalized dataset for the Ṛgveda-saṃhit \= a

    Krishnan, Sriram and Gayathri, Sepuri and Kulkarni, Amba. Challenges in Processing V edic S anskrit: Towards creating a normalized dataset for the Ṛgveda-saṃhit \= a. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  68. [68]

    P \= a ṇḍitya: Visualizing S anskrit Intellectual Networks

    Neill, Tyler. P \= a ṇḍitya: Visualizing S anskrit Intellectual Networks. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  69. [69]

    Anveshana: A New Benchmark Dataset for Cross-Lingual Information Retrieval on E nglish Queries and S anskrit Documents

    Jagadeeshan, Manoj Balaji and Raj, Prince and Goyal, Pawan. Anveshana: A New Benchmark Dataset for Cross-Lingual Information Retrieval on E nglish Queries and S anskrit Documents. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  70. [70]

    Concordance of S anskrit Synonyms

    Patel, Dhaval. Concordance of S anskrit Synonyms. Computational Sanskrit and Digital Humanities - World Sanskrit Conference 2025. 2025

  71. [71]

    Proceedings of the First Workshop on Writing Aids at the Crossroads of AI, Cognitive Science and NLP (WRAICOGS 2025). 2025

  72. [72]

    Chain-of- M eta W riting: Linguistic and Textual Analysis of How Small Language Models Write Young Students Texts

    Buhnila, Ioana and Cislaru, Georgeta and Todirascu, Amalia. Chain-of- M eta W riting: Linguistic and Textual Analysis of How Small Language Models Write Young Students Texts. Proceedings of the First Workshop on Writing Aids at the Crossroads of AI, Cognitive Science and NLP (WRAICOGS 2025). 2025

  73. [73]

    Semantic Masking in a Needle-in-a-haystack Test for Evaluating Large Language Model Long-Text Capabilities

    Shi, Ken and Penn, Gerald. Semantic Masking in a Needle-in-a-haystack Test for Evaluating Large Language Model Long-Text Capabilities. Proceedings of the First Workshop on Writing Aids at the Crossroads of AI, Cognitive Science and NLP (WRAICOGS 2025). 2025

  74. [74]

    Reading Between the Lines: A dataset and a study on why some texts are tougher than others

    Khallaf, Nouran and Eugeni, Carlo and Sharoff, Serge. Reading Between the Lines: A dataset and a study on why some texts are tougher than others. Proceedings of the First Workshop on Writing Aids at the Crossroads of AI, Cognitive Science and NLP (WRAICOGS 2025). 2025

  75. [75]

    P ara R ev : Building a dataset for Scientific Paragraph Revision annotated with revision instruction

    Jourdan, L \'e ane and Boudin, Florian and Dufour, Richard and Hernandez, Nicolas and Aizawa, Akiko. P ara R ev : Building a dataset for Scientific Paragraph Revision annotated with revision instruction. Proceedings of the First Workshop on Writing Aids at the Crossroads of AI, Cognitive Science and NLP (WRAICOGS 2025). 2025

  76. [76]

    Towards an operative definition of creative writing: a preliminary assessment of creativeness in AI and human texts

    Maggi, Chiara and Vitaletti, Andrea. Towards an operative definition of creative writing: a preliminary assessment of creativeness in AI and human texts. Proceedings of the First Workshop on Writing Aids at the Crossroads of AI, Cognitive Science and NLP (WRAICOGS 2025). 2025

  77. [77]

    Decoding Semantic Representations in the Brain Under Language Stimuli with Large Language Models

    Sato, Anna and Kobayashi, Ichiro. Decoding Semantic Representations in the Brain Under Language Stimuli with Large Language Models. Proceedings of the First Workshop on Writing Aids at the Crossroads of AI, Cognitive Science and NLP (WRAICOGS 2025). 2025

  78. [78]

    Proceedings of the The 9th Workshop on Online Abuse and Harms (WOAH). 2025

  79. [79]

    A Comprehensive Taxonomy of Bias Mitigation Methods for Hate Speech Detection

    Fillies, Jan and Wawerek, Marius and Paschke, Adrian. A Comprehensive Taxonomy of Bias Mitigation Methods for Hate Speech Detection. Proceedings of the The 9th Workshop on Online Abuse and Harms (WOAH). 2025

  80. [80]

    Sensitive Content Classification in Social Media: A Holistic Resource and Evaluation

    Antypas, Dimosthenis and Sen, Indira and Perez Almendros, Carla and Camacho-Collados, Jose and Barbieri, Francesco. Sensitive Content Classification in Social Media: A Holistic Resource and Evaluation. Proceedings of the The 9th Workshop on Online Abuse and Harms (WOAH). 2025

Showing first 80 references.