diff options
Diffstat (limited to 'report')
| -rw-r--r-- | report/pr-clustering/EMVSPR.pdf | bin | 0 -> 58054 bytes | |||
| -rw-r--r-- | report/pr-clustering/posterior.tex | 31 | 
2 files changed, 31 insertions, 0 deletions
| diff --git a/report/pr-clustering/EMVSPR.pdf b/report/pr-clustering/EMVSPR.pdfBinary files differ new file mode 100644 index 00000000..c03b41f2 --- /dev/null +++ b/report/pr-clustering/EMVSPR.pdf diff --git a/report/pr-clustering/posterior.tex b/report/pr-clustering/posterior.tex index 73c15dba..7597c8e1 100644 --- a/report/pr-clustering/posterior.tex +++ b/report/pr-clustering/posterior.tex @@ -191,3 +191,34 @@ where $\mathcal{L}_1$ and $\mathcal{L}_2$  are log-likelihood of  two models.  \section{Experiments} +As a sanity check, we looked at a few examples produced by +the basic model (EM)  +and the posterior regularization (PR) model +with sparsity constraints. Table \ref{tab:EMVSPR} +shows a few examples. + +\begin{table}[h] +  \centering +  \includegraphics[width=3.5in]{pr-clustering/EMVSPR} +  \caption[A few examples comparing EM and PR] +  {A few examples comparing EM and PR.  +    Count of most frequent category shows how  +    many instances of a phrase are concetrated on  +    the single most frequent tag.  +    Number of categories shows how many categories +    a phrase is labelled with. By experience as mentioned before,  +    we want a phrase to use fewer categories.  +	These numbers are fair indicators of sparsity. +    } +  \label{tab:EMVSPR} +\end{table} + +The models are formally evaluated with two kinds +of metrics. We feed the clustering output +through the whole translation pipeline  +to obtain a BLEU score. We also came up  +with an intrinsic evaluation of clustering quality +by comparing against a supervised CFG parser trained on the +tree bank. + + | 
