Skip to content

epfl-ml4ed/ripple

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

63 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Ripple: Concept-Based Interpretation for Raw Time Series Models in Education

This is the official repository for the paper titled "Ripple: Concept-Based Interpretation for Raw Time Series Models in Education", written by Mohammad Asadi, Vinitra Swamy, Jibril Frej, Julien Vignoud, Mirko Marras, Tanja Kaser and featured at AAAI 23.

Jupyter notebooks corresponding to each of the research questions mentioned in the paper can be found in scripts/ folder.

Project overview

Time series is the most prevalent form of input data for educational prediction tasks. The vast majority of research using time series data focuses on hand-crafted features, designed by experts for predictive performance and interpretability. However, extracting these features is labor-intensive for humans and computers. In this paper, we propose an approach that utilizes irregular multivariate time series modeling with graph neural networks to achieve comparable or better accuracy with raw time series clickstreams in comparison to hand-crafted features.

Furthermore, we extend concept activation vectors for interpretability in raw time series models. We analyze these advances in the education domain, addressing the task of early student performance prediction for downstream targeted interventions and instructional support. Our experimental analysis on 23 MOOCs with millions of combined interactions over six behavioral dimensions show that models designed with our approach can (i) beat state-of-the-art educational time series baselines with no feature extraction and (ii) provide interpretable insights for personalized interventions.

Usage Guide

Jupyter notebooks and python scripts needed to reproduce the paper's results are provided in scripts\ folder. Short descriptions regarding the research question and methodology are also presented at the beginning of each of the notebooks.

  • Start by cloning this repository with git clone https://github.com/epfl-ml4ed/ripple.git.
  • Install the required packages to run these scripts with pip install -r requirements.txt.
  • Run the desired scripts below to reproduce experiments from the paper.

data_splits.ipynb
Splits the dataset for each of the courses found in data\prep_data folder into train, validation, and test with respect to the defined percentages (80-10-10 as default). It then stores the arguments of the students corresponding to each of the splits and stores them in data\split_args.

dataset_convert.ipynb
This script is responsible for transforming the course log datasets as provided by MOOC platforms into arrays of interaction×event_types, actual time_stamps, actual non-zero length of each time series and their corresponding pass/fail labels.

run_raindrop.ipynb
RQ1. Using the data splits and datasets resulted from the last two scripts, run_raindrop.ipynb alters Raindrop model w.r.t. each course dataset and trains the model on it. It then evaluates the the trained model's performance again on validation set and test set and saves them as .csv files for each course. This file uses two versions of Raindrop namely Raindrop_v2 and Raindrop_v3 where the last one is a slightly adjusted version of Raindrop_v2 from the original Raindrop repository to change the number of layers for the final MLP classification layer for grid search.

run_baselines.ipynb
Trains, evaluates, and stores validation and test results for all the handcrafted-feature models. Implementation adopted from the ML4ED lab's previous Meta-Transfer Learning for Early Success Prediction in MOOCs paper. The code for this model can be found at the epfl-ml4ed/meta-transfer-learning repository.

run_baselines_seft_trnsfrmr.ipynb
Trains, evaluates, and stores validation and test results for SEFT and Transformer as implemented by Raindrop's repository. It also adjusts SEFT and Transformer for altering the number of layers for TransformerEncoder and last MLP classification layer essential for performing the grid search.

tcav_rq2.ipynb
RQ2 & RQ3. Implements TCAV pipeline on Raindrop model trained on a specific course. Here all the concepts mentioned in the paper are defined and the groups of top and bottom student w.r.t. each concept are extracted. These students are then used to specify the concepts and feed them into TCAV pipeline for time series. This script also generates the results showen in paper given the same dataset. This script also generates TCAV scores and the corresponding plots for the 4 confusion matrix scenarios.

tcav_rq3.ipynb
RQ3. This script is responsible for normalizing the feature scores comprising the concepts and then finding the top and bottom students regarding the overall performance across all the concepts. It then generates the TCAV score plots for the model's prediction on these specific students.

Contributing

This code is provided for educational purposes and aims to facilitate reproduction of our results, and further research in this direction. We have done our best to document, refactor, and test the code before publication.

If you find any bugs or would like to contribute new models, training protocols, etc, please let us know. Feel free to file issues and pull requests on the repo and we will address them as we can.

Citations

If you find this code useful in your work, please cite our paper:

Asadi, M., Swamy, V., Frej, J., Vignoud, J., Marras, M., & Käser, T. (2023). 
Ripple: Concept-Based Interpretation for Raw Time Series Models in Education. 
In: Proceedings of the 37th AAAI Conference on Artificial Intelligence (EAAI Symposium).

License

This code is free software: you can redistribute it and/or modify it under the terms of the MIT License.

This software is distributed in the hope that it will be useful, but without any warranty; without even the implied warranty of merchantability or fitness for a particular purpose. See the MIT License for details.

About

Interpretability on raw time series with graph neural nets and concept activation vectors. Featured at AAAI 2023.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published