Skip to content

YiJingLin/ML2017Fall_Final_QA

Repository files navigation

Question Answering with Context in Traditional Chinese

Department of Electronic Enginering, National Taiwan University
Machine Learning (2017, Fall) - Final Project

  • kaggle competition (link)

Maintenance

Dependancy :

python version keras version numpy version gensim version jieba version

Introduction :

Abstract :

受到 SQuADBiDAF + Self Attention 的啟發,我們採用 self-attention 以及 Bidirectional Attention Flow 的技術實作 QA model。

Training :

採用既定json格式作為training data的input (請參照train-v1.1.json)。在過程中,我們將 1.文章 2.問題 及 3.答案 抓出,分別存成3個list。文章及問題的list會經過word2vector將字詞轉換成向量,最後將3個list都轉成numpy array。文章及問題array作為model input,答案array作為model output,並開始訓練model。完成訓練後,會將model存成model.h5,待predict時使用。

Testing :

一樣採用既定json格式作為testing data的input (請參照test-v1.1.json)。與traing data process相比,只抓出文章及對應的問題,並經過word2vector轉換,最後將文章及問題arrays作為model input,並得到predict結果。

bash test.sh <test.json> <output.csv>

Collaborators :

Team :

NTU_R04942032_伊恩好夥伴
(inspired from Ian Goodfellow)

member :

  • Huang.Ychen (黃彥澄) (github)
  • Lin.TzungYing (林宗穎) (github)
  • Lin.YiJing (林益璟) (github)

About

ML2017Fall Final Project : QA in chinese

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published