Skip to content

Commit 473ef3d

Browse files
author
Vineet John
committed
Migrated to natbib reference format
1 parent 7e8d8de commit 473ef3d

File tree

2 files changed

+8
-8
lines changed

2 files changed

+8
-8
lines changed

project-report/cs698_project_report.bib

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -140,5 +140,5 @@ @article{SemEvalPaper
140140
title={UW-FinSent at SemEval-2017 Task 5: Fine-Grained Sentiment Analysis on Financial News Headlines},
141141
author={John, Vineet and Vechtomova, Olga},
142142
journal={Proceedings of the 11th international workshop on semantic evaluation},
143-
year={2017},
143+
year={2017}
144144
}

project-report/cs698_project_report.tex

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414

1515
\setlength\titlebox{5cm}
1616

17-
\title{A Survey of Neural Network Techniques\\ for Feature Extraction from Text\\ \large CS 698 (Winter 2017) - Project}
17+
\title{A Survey of Neural Network Techniques\\ for Feature Extraction from Text}
1818

1919
\author{
2020
Vineet John \\
@@ -319,7 +319,7 @@ \section{Linguistic Regularities in Continuous Space Word Representations} % (fo
319319
$f(z) = \frac{1}{1 + e^{-z}}$ and $g(z_m) = \frac{e^{z_m}}{\sum_k e^{z_k}} $
320320
\begin{figure}[ht]
321321
\centering
322-
\includegraphics[width=.4\textwidth]{rnn-lang-model}
322+
\includegraphics[width=.4\textwidth]{rnn-lang-model.png}
323323
\caption{RNN Language Model}
324324
\label{fig:rnn-lang-model}
325325
\end{figure}
@@ -350,7 +350,7 @@ \section{Better Word Representations with Recursive Neural Networks for Morpholo
350350
\texttt{morphoRNN} operates at the morpheme level rather than the word level. An example of the this is illustrated in Figure \ref{fig:rnn-morphology}.
351351
\begin{figure}[ht]
352352
\centering
353-
\includegraphics[width=.4\textwidth]{rnn-morphology}
353+
\includegraphics[width=.4\textwidth]{rnn-morphology.png}
354354
\caption{morphoRNN}
355355
\label{fig:rnn-morphology}
356356
\end{figure}
@@ -388,7 +388,7 @@ \section{Efficient Estimation of Word Representations in Vector Space} % (fold)
388388
\textbf{Approach:}
389389
\begin{itemize}
390390
\item
391-
The ideas presented in this paper build on the previous ideas presented by Bengio et.al.\cite{bengio2003neural}
391+
The ideas presented in this paper build on the previous ideas presented by \cite{bengio2003neural}.
392392
\item
393393
The objective was to obtain high-quality word embeddings that capture the syntactic and semantic characteristics of words in a manner that allows algebraic operations to proxy the distances in vector space.
394394
$$man - woman = king - queen$$ or $$tell - told = walk - walked$$
@@ -440,7 +440,7 @@ \section{Distributed Representations of Words and Phrases and their Compositiona
440440
\end{equation}
441441
where $c$ is the window or context surrounding the current word being trained on.
442442
\item
443-
As introduced by Morin, Bengio et.al.\cite{morin2005hierarchical}, a computationally efficient approximation of the full softmax is the hierarchical softmax. The hierarchical softmax uses a binary tree representation of the output layer with the W words as its leaves and, for each node, explicitly represents the relative probabilities of its child nodes. These define a random walk that assigns probabilities to words.
443+
As introduced by \cite{morin2005hierarchical}, a computationally efficient approximation of the full softmax is the hierarchical softmax. The hierarchical softmax uses a binary tree representation of the output layer with the W words as its leaves and, for each node, explicitly represents the relative probabilities of its child nodes. These define a random walk that assigns probabilities to words.
444444
\item
445445
The authors use a binary Huffman tree, as it assigns short codes to the frequent words which results in fast training. It has been observed before that grouping words together by their frequency works well as a very simple speedup technique for the neural network based language models.
446446
\item
@@ -483,7 +483,7 @@ \section{Glove: Global Vectors for Word Representation} % (fold)
483483
\item
484484
An additive shift is included in the logarithm, $$\log(X_{ik}) \Rightarrow log(1 + X_{ik})$$ which maintains the sparsity of X while avoiding the divergences while computing the co-occurrences matrix.
485485
\item
486-
The model obtained in the paper could be compared to a global skip-gram model as opposed to a fixed window-size skip-gram model as proposed by Mikolov et.al.\cite{mikolov2013efficient}.
486+
The model obtained in the paper could be compared to a global skip-gram model as opposed to a fixed window-size skip-gram model as proposed by \cite{mikolov2013efficient}.
487487
\item
488488
The performance seems to increase monotonically with an increase in training data.
489489
\end{itemize}
@@ -535,7 +535,7 @@ \section{Acknowledgements} % (fold)
535535
% section acknowledgements (end)
536536

537537

538-
\bibliographystyle{unsrt}
538+
\bibliographystyle{acl_natbib}
539539
\bibliography{cs698_project_report}
540540

541541
\end{document}

0 commit comments

Comments
 (0)