Kiran Mayee Adavala and Om Adavala
Adv. Artif. Intell. Mach. Learn., 6 (1):4943-4964
1. Kiran Mayee Adavala: Kakatiya Institute of Technology & Science, Warangal
2. Om Adavala: NFSU, Gujarat
DOI: 10.54364/AAIML.2026.61273
Article History: Received on: 10-Nov-25, Accepted on: 22-Jan-26, Published on: 29-Jan-26
Corresponding Author: Kiran Mayee Adavala
Email: ak.csm@kitsw.ac.in
Citation: Kiran Mayee Adavala and Om Adavala. Learning Discriminative Syntax-Semantic Patterns With Transformer-Based Contrastive Learning. Advances in Artificial Intelligence and Machine Learning. 2026;6(1):273. https://dx.doi.org/10.54364/AAIML.2026.61273
This paper proposes a dual-encoder framework that learns syntax- and semantics-aligned sentence embeddings using contrastive learning. The proposed Syntax-Semantic Contrastive Pretraining (SSCP) model employs two transformer encoders to separately model syntactic structure and semantic content, which are aligned in a shared embedding space via a symmetric contrastive objective. Across standard benchmarks, SSCP consistently outperforms strong baselines such as BERT, SimCSE, and SyntaxBERT. In particular, SSCP improves Spearman correlation on STS-B by +1.4 points over SimCSE, increases PAWS paraphrase accuracy by +2.5 points, and achieves 96.1% accuracy on TREC question classification, exceeding existing syntax-aware models. Probing experiments further show gains of up to +4–5 points on syntactic structure prediction tasks, confirming that SSCP preserves grammatical information while maintaining semantic robustness. These results demonstrate that explicitly aligning syntactic and semantic views yields representations that are more discriminative, interpretable, and robust than single-view or syntaxaugmented approaches, positioning SSCP as a principled multi-view pretraining strategy for structure-aware language understanding.