From 4dcb875bf67138fa9881f7ceb9d99ebe4867d4bf Mon Sep 17 00:00:00 2001 From: Martin Thoma Date: Sat, 11 Feb 2017 12:40:01 +0100 Subject: [PATCH] HASY: Add k-NN (k=3, k=5) --- publications/hasy/content.tex | 15 ++++++++++----- 1 file changed, 10 insertions(+), 5 deletions(-) diff --git a/publications/hasy/content.tex b/publications/hasy/content.tex index 388e4a0..b58c8a8 100644 --- a/publications/hasy/content.tex +++ b/publications/hasy/content.tex @@ -201,8 +201,8 @@ of any classifier being evaluated on \dbName{} as follows: \subsection{Model Baselines} Eight standard algorithms were evaluated by their accuracy on the raw image data. The neural networks were implemented with -Tensorflow~\cite{tensorflow2015-whitepaper}. All other algorithms are -implemented in sklearn~\cite{scikit-learn}. \Cref{table:classifier-results} +Tensorflow~0.12.1~\cite{tensorflow2015-whitepaper}. All other algorithms are +implemented in sklearn~0.18.1~\cite{scikit-learn}. \Cref{table:classifier-results} shows the results of the models being trained and tested on MNIST and also for \dbNameVersion{}: \begin{table}[h] @@ -215,6 +215,8 @@ shows the results of the models being trained and tested on MNIST and also for Random Forest & \SI{96.41}{\percent} & \SI{62.4}{\percent} & \SI{62.1}{\percent} -- \SI{62.8}{\percent}\\% & \SI{19.0}{\second}\\ MLP (1 Layer) & \SI{89.09}{\percent} & \SI{62.2}{\percent} & \SI{61.7}{\percent} -- \SI{62.9}{\percent}\\% & \SI{7.8}{\second}\\ LDA & \SI{86.42}{\percent} & \SI{46.8}{\percent} & \SI{46.3}{\percent} -- \SI{47.7}{\percent}\\% & \SI{0.2}{\second}\\ + $k$-NN ($k=3$)& \SI{92.84}{\percent} & \SI{28.4}{\percent} & \SI{27.4}{\percent} -- \SI{29.1}{\percent}\\% & \SI{196.2}{\second}\\ + $k$-NN ($k=5$)& \SI{92.88}{\percent} & \SI{27.4}{\percent} & \SI{26.9}{\percent} -- \SI{28.3}{\percent}\\% & \SI{196.2}{\second}\\ QDA & \SI{55.61}{\percent} & \SI{25.4}{\percent} & \SI{24.9}{\percent} -- \SI{26.2}{\percent}\\% & \SI{94.7}{\second}\\ Decision Tree & \SI{65.40}{\percent} & \SI{11.0}{\percent} & \SI{10.4}{\percent} -- \SI{11.6}{\percent}\\% & \SI{0.0}{\second}\\ Naive Bayes & \SI{56.15}{\percent} & \SI{8.3}{\percent} & \SI{7.9}{\percent} -- \hphantom{0}\SI{8.7}{\percent}\\% & \SI{24.7}{\second}\\ @@ -225,9 +227,12 @@ shows the results of the models being trained and tested on MNIST and also for % The test time is the time needed for all test samples in average. The number of test samples differs between the folds, but is $\num{16827} \pm - 166$. The decision tree - was trained with a maximum depth of 5. The exact structure - of the CNNs is explained in~\cref{subsec:CNNs-Classification}.} + 166$. The decision tree was trained with a maximum depth of~5. The + exact structure of the CNNs is explained + in~\cref{subsec:CNNs-Classification}. For $k$ nearest neighbor, + the amount of samples per class had to be reduced to 50 for HASY + due to the extraordinary amount of testing time this algorithm + needs.} \label{table:classifier-results} \end{table}