<h1>Transformer Encoder for Social Science (TESS)</h1>

TESS is a deep neural network model intended for social science related NLP tasks. The model is developed by Haosen Ge, In Young Park, Xuancheng Qian, and Grace Zeng.

We demonstrate in two validation tests that TESS outperforms BERT and RoBERTa by 16.7% on average, especially when the number of training samples is limited (<1,000 training instances). The results display the superiority of TESS on social science text processing tasks.

GitHub: TESS.

<h2>Training Corpus</h2>

TEXT SOURCE
Preferential Trade Agreements ToTA
Congressional Bills Kornilova and Eidelman (2019)
UNGA Resolutions UN
Firms' Annual Reports Loughran and McDonald (2016)
U.S. Court Opinions Caselaw Access Project

The model is trained on 4 NVIDIA A100 GPUs for 120K steps.