PMI Results

Our PMI results are very useful. We can clearly see the nondeterministic nature of some pairs, e.g., “should” occurs so many times that it becomes less correlated with any specific joining term. Whereas we see that high PMI values are associated with terms that do not occur on their own very frequently, such as “infantile […]

Pointwise Mutual Information

Pointwise Mutual Information (log base 2) was calculated based on our unigram and bigram tables. The results indicate that “should” is too ubiquitous to be meaningfully correlated. The PMI was determined from P(x,y) being equivalent to the frequency of the bigram appearing and P(x) and P(y) are frequencies taken from the unigram table. Our corpus […]

Probability Theory

One aspect of machine learning that is ubiquitous is probability theory. As we develop more complex programs, it seems natural to turn to a programming language that is built for this task. We are investigating Gen which is written in the Julia programming language. Julia is a very powerful and yet user-friendly language. It builds […]