Daan de Jong Master Thesis Methodology and Statistics for the Biomedical, Behavioral and Social Sciences Negation Scope Detection with a BiLSTM NN
The Thesis-Report repository contains two folders: Text and Data and code.
- img folder:
Plaatjesmaker.ppt
andPlaatjesmaker2.ppt
were used to createmodel.png
,model2.png
andprediction types.png
model.png
is outdatedmodel2.png
was used in the final Report (bymain.tex
in the text folder)prediction types.png
was used in the final Report (bymain.tex
in the text folder)
main.tex
contains the text that generates the final reportrefs_report
is abib
file that contains the references used by main.textitle page.tex
is used bymain.tex
, to split up the first page from the actual content of the reportDraft 1 Report DDJ.pdf
is the main draft which was feedbacked by Ayoub Bagheri and Joost de Jong (feedback not included)Research Report Daan de Jong.pdf
is the final Report that was handed inmain.bbl
,main.blg
,main.log
,main.synctex.gz
,textut.log
,title page.log
are byproducts ofmain.tex
andtitle page.tex
__pycache__
is a folder that contains byproducts of the.py
files- Original is a folder that contains all data used in this study
- freely downloadable here
- the folder bioscope was not used
bioscope_abstract.csv
andbioscope_full.csv
are the data files that were used in this study (byPreprocessing.py
)
- All
.csv
files here are rendered byPreprocessing.py
, and are read inModel.py
, except forE.csv
(the word embeddings, fromword2vec.py
) CustomEmbedding.py
defines a keras layer class, to be used inModel.py
CustomLoss.py
defines the loss function used in the study, used byModel.py
CustomMetric.py
defines the metrics used in the study, used byModel.py
Data description table.py
renders the information used in Table 1 of the ReportModel.py
defines, compiles, trains, tests and evaluates the modelPreprocessing.py
bridges the gap between the raw data.csv
files and the train and test data suited for the modelword2vec.py
is an implementation of the word2vec algorithm and rendersE.csv
, the word embeddings