Skip to content

ljn3333/ML2017Fall_Final_QA

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Question Answering with Context in Traditional Chinese

Department of Electronic Enginering, National Taiwan University
Machine Learning (2017, Fall) - Final Project

  • kaggle competition (link)

Dependancy :

python version keras version numpy version gensim version jieba version

Introduction :

Abstract :

受到 SQuADBiDAF + Self Attention 的啟發,我們採用 self-attention 以及 Bidirectional Attention Flow 的技術實作 QA model。

Training :

採用既定json格式作為training data的input (請參照train-v1.1.json)。在過程中,我們將 1.文章 2.問題 及 3.答案 抓出,分別存成3個list。文章及問題的list會經過word2vector將字詞轉換成向量,最後將3個list都轉成numpy array。文章及問題array作為model input,答案array作為model output,並開始訓練model。完成訓練後,會將model存成model.h5,待predict時使用。

Testing :

一樣採用既定json格式作為testing data的input (請參照test-v1.1.json)。與traing data process相比,只抓出文章及對應的問題,並經過word2vector轉換,最後將文章及問題arrays作為model input,並得到predict結果。

python test.sh <test.json> <output.csv>

Collaborators :

Team :

NTU_R04942032_伊恩好夥伴
(inspired from Ian Goodfellow)

member :

  • Huang.Ychen (黃彥澄) (github)
  • Lin.ZhuangYing (林宗穎) (github)
  • Lin.YiJing (林益璟) (github)

About

ML2017Fall Final Project : QA in chinese

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 97.0%
  • Shell 3.0%