ISSN :2582-9793

Learning discriminative syntax-semantic patterns with transformer-based contrastive learning

Original Research (Published On: 29-Jan-2026 )
DOI : https://doi.org/10.54364/AAIML.2026.61273

Kiran Mayee Adavala and Om Adavala

Adv. Artif. Intell. Mach. Learn., 6 (1):4943-4964

1. Kiran Mayee Adavala: Kakatiya Institute of Technology & Science, Warangal

2. Om Adavala: NFSU, Gujarat

Download PDF Here

DOI: 10.54364/AAIML.2026.61273

Article History: Received on: 10-Nov-25, Accepted on: 22-Jan-26, Published on: 29-Jan-26

Corresponding Author: Kiran Mayee Adavala

Email: ak.csm@kitsw.ac.in

Citation: Kiran Mayee Adavala and Om Adavala. Learning Discriminative Syntax-Semantic Patterns With Transformer-Based Contrastive Learning. Advances in Artificial Intelligence and Machine Learning. 2026;6(1):273. https://dx.doi.org/10.54364/AAIML.2026.61273


Abstract

    

This paper proposes a dual-encoder framework that learns syntax- and semantics-aligned sentence embeddings using contrastive learning. The proposed Syntax-Semantic Contrastive Pretraining (SSCP) model employs two transformer encoders to separately model syntactic structure and semantic content, which are aligned in a shared embedding space via a symmetric contrastive objective. Across standard benchmarks, SSCP consistently outperforms strong baselines such as BERT, SimCSE, and SyntaxBERT. In particular, SSCP improves Spearman correlation on STS-B by +1.4 points over SimCSE, increases PAWS paraphrase accuracy by +2.5 points, and achieves 96.1% accuracy on TREC question classification, exceeding existing syntax-aware models. Probing experiments further show gains of up to +4–5 points on syntactic structure prediction tasks, confirming that SSCP preserves grammatical information while maintaining semantic robustness. These results demonstrate that explicitly aligning syntactic and semantic views yields representations that are more discriminative, interpretable, and robust than single-view or syntax￾augmented approaches, positioning SSCP as a principled multi-view pretraining strategy for structure-aware language understanding.

Statistics

   Article View: 981
   PDF Downloaded: 11