Skip to content

Latest commit

 

History

History
77 lines (59 loc) · 5.34 KB

README.md

File metadata and controls

77 lines (59 loc) · 5.34 KB

Experiments

Please see the corpus website for an introduction to this data set.

Requirements

The version numbers in parantheses indicate the version we used, other versions may or may not work.

We recommend a miniconda environment.

To obtain reproducible results, parallel execution is disabled at several points in the code. This means things could run quite a bit faster, but would not result in the exact same results, which is now the case, with the exception of the LSTM part.

Total duration: about 4.5 hours on a machine with the following specs:

  • Intel Core i7-6900K, 8x 3.20GHz
  • 64 GB RAM (4x16GB DDR4-2133)
  • 1TB SSD
  • NVIDIA Titan X (Pascal)

Running the Experiments

To run everything, simply execute

./run.sh

The script will ask you if you want to download the corpus (requiring wget or curl and bzip2).

If you are interested only in certain parts, uncomment what you don't need in run.sh and in src/run_evaluation.py (in particular, you can comment out entries from methodmodules if you want to run only certain methods).

The code produces some output in the directories logs, plots and tables. We have included our results here, so you can see what to expect.

The code also creates a directory models, which will be about 500 MB in size. The entire experiments folder, including the downloaded corpus, will be almost 1GB in size.

Results

CategoryMeasureBOWMNBNBSVMBOCIDD2VLSTM
Negative SentimentPrecision0.55210.56370.56600.53450.58420.5349
Recall0.51090.48670.45120.54520.56240.7197
F10.53070.52240.50210.53980.57310.6137
Positive SentimentPrecision0.10000.00000.23530.06620.03970.0000
Recall0.06980.00000.09300.20930.46510.0000
F10.08220.00000.13330.10060.07310.0000
Off-topicPrecision0.27540.61900.39690.22520.20650.2742
Recall0.23790.02240.13280.51210.62410.2638
F10.25530.04330.19900.31280.31030.2689
InappropriatePrecision0.16270.00000.17650.15160.13400.1964
Recall0.11220.00000.04950.39930.57760.1089
F10.13280.00000.07730.21980.21750.1401
DiscriminatingPrecision0.18470.00000.26830.13010.11110.1136
Recall0.10280.00000.07800.29430.39360.1418
F10.13210.00000.12090.18040.17330.1262
FeedbackPrecision0.65540.74650.73560.50940.52400.6307
Recall0.58030.40740.52190.68790.70560.6287
F10.61560.52710.61060.58530.60140.6297
Personal StoriesPrecision0.69810.54910.69160.57620.62470.6380
Recall0.59200.45780.47880.71200.81230.6658
F10.64070.49930.56580.63690.70630.6516
Arguments UsedPrecision0.61050.50860.60640.56420.56570.5685
Recall0.52150.31700.46280.61060.66140.6458
F10.56250.39060.52500.58650.60980.6047
WinsPrecision222011
Recall000071
F1001322