Skip to content

Notebooks for running and visualizing results using trained models for linguistic complexity.

Notifications You must be signed in to change notification settings

Farahn/Liguistic-Complexity

Repository files navigation

Liguistic-Complexity

Extension to hierarchical attention network for computing text difficulty from the paper Estimating Linguistic Complexity for Science Texts, Farah Nadeem and Mari Ostendorf, The Thirteenth Workshop on Innovative Use of NLP for Building Educational Applications, NAACL 2018.

The model uses attention across sentences in a text ("bi-directional context with attention"), and ordinal regression. The original implementation was from https://github.com/ilivans/tf-rnn-attention, modified to use word and sentence level structure, and ordinal instead of logistic regression.

trained_bca_ELA_grades_CCS.ipynb is the iPython notebook for running pre-trained models for computing text difficulty. The pretrained models can be obtained from https://1drv.ms/f/s!Ag4UUgKkf0ZPu55vBR6_-6WOuAO-Ug. Two test sets are also available at https://sites.google.com/site/nadeemf0755/research/linguistic-complexity. The CCS test set can be tested with both of the pretrained models.

Models have been updated to TF 1.8.

About

Notebooks for running and visualizing results using trained models for linguistic complexity.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published