sentence1
stringlengths 16
446
| sentence2
stringlengths 14
436
|
---|---|
we perform the mert training to tune the optimal feature weights on the development set .
|
we tune model weights using minimum error rate training on the wmt 2008 test data .
|
bleu is smoothed , and it considers only matching up to bigrams because this has higher correlations with human judgments than when higher-ordered n-grams are included .
|
bleu is smoothed to be more appropriate for sentencelevel evaluation , and the bigram versions of bleu and hwcm are reported because they have higher correlations than when longer n-grams are included .
|
blacoe and lapata compare count and predict representations as input to composition functions .
|
blacoe and lapata compare different arithmetic functions across multiple representations on a range of compositionality benchmarks .
|
in section 3 , we describe the three resources we use in our experiments .
|
in section 3 , we describe the three resources we use in our experiments and how we model them .
|
word sense induction ( wsi ) is the task of automatically identifying the senses of words in texts , without the need for handcrafted resources or manually annotated data .
|
word sense induction ( wsi ) is the task of automatically inducing the different senses of a given word , generally in the form of an unsupervised learning task with senses represented as clusters of token instances .
|
inversion transduction grammar is a formalism for synchronous parsing of bilingual sentence pairs .
|
inversion transduction grammar is a synchronous grammar for synchronous parsing of source and target language sentences .
|
we then created trigram language models from a variety of sources using the srilm toolkit , and measured their perplexity on this data .
|
we trained a specific language model using srilm from each of these corpora in order to estimate n-gram log-probabilities .
|
we evaluate global translation quality with bleu and meteor .
|
we evaluate translations with bleu and meteor .
|
a 4-gram language model is trained on the xinhua portion of the gigaword corpus with the srilm toolkit .
|
gram language model with modified kneser-ney smoothing is trained with the srilm toolkit on the epps , ted , newscommentary , and the gigaword corpora .
|
in future work , we will try to collect and annotate data for microblogs in other languages .
|
in future work , we will try to collect and annotate data for microblogs in other languages to test the robustness of our method .
|
for all data sets , we trained a 5-gram language model using the sri language modeling toolkit .
|
we trained a 4-gram language model with kneser-ney smoothing and unigram caching using the sri-lm toolkit .
|
however , to the best of our knowledge , there is no attempt in the literature to build a resource .
|
however , to the best of our knowledge , there is no systematic attempt in the literature to build such a resource .
|
glove is an unsupervised learning algorithm for word embeddings .
|
glove is an unsupervised algorithm that constructs embeddings from large corpora .
|
evaluation on the ace data set shows that the ilp based entity-mention model is effective for the coreference resolution task .
|
and our experimental results on the ace data set shows the model is effective for coreference resolution .
|
in this case , it may be preferable to look for near-duplicate documents .
|
in this setting , reuse may be mixed with text derived from other sources .
|
sagae and lavie apply a notion of reparsing to a two stage parser combination chartbased approach .
|
sagae and lavie proposed a constituent reparsing method for multiple parsers combination .
|
in li and roth , they used wordnet for english and built a set of class-specific words as semantic features and achieved the high precision .
|
later li and roth used more semantic information sources including named entities , wordnet senses , class-specific related words , and distributional similarity based categories in question classification task .
|
given such parallel data , we can easily train an encoder-decoder model that takes a sentence and target syntactic template .
|
such data allows us to train a neural encoder-decoder model with extra inputs to specify the target syntax .
|
we trained a 5-gram language model on the xinhua portion of gigaword corpus using the srilm toolkit .
|
we then created trigram language models from a variety of sources using the srilm toolkit , and measured their perplexity on this data .
|
on the remaining tweets , we trained a 10-gram word length model , and a 5-gram language model , using srilm with kneyser-ney smoothing .
|
for the semantic language model , we used the srilm package and trained a tri-gram language model with the default goodturing smoothing .
|
evaluated on a news headline dataset , our model yielded higher accuracy .
|
this approach yielded a precision between 71 % and 82 % on the news headline dataset .
|
for our classifiers , we used the weka implementation of na茂ve bayes and the svmlight implementation of the svm .
|
here too , we used the weka implementation of the na茂ve bayes model and the svmlight implementation of the svm .
|
experiments show that the models can achieve 0 . 85 precision at a level of 0 . 89 recall , and even higher precision .
|
experimental results show that our system can achieve 0.85 precision at 0.89 recall , excluding exact matches .
|
we obtained a vocabulary of 320,935 unique words after eliminating words which occur only once , stemming by a part-ofspeech tagger , and stop word removal .
|
we obtained a vocabulary of 183,400 unique words after eliminating words which occur only once , stemming by a partof-speech tagger , and stop word removal .
|
we evaluated the system using bleu score on the test set .
|
we evaluated the translation quality using the bleu-4 metric .
|
in order to reduce the source vocabulary size translation , the german text was preprocessed by splitting german compound words with the frequencybased method described in .
|
for the translation from german into english , german compound words were split using the frequency-based method described in .
|
a sentiment lexicon is a list of words and phrases , such as “ excellent ” , “ awful ” and “ not bad ” , each of them is assigned with a positive or negative score reflecting its sentiment polarity and strength ( cite-p-18-3-8 ) .
|
a sentiment lexicon is a list of sentiment expressions , which are used to indicate sentiment polarity ( e.g. , positive or negative ) .
|
a 4-gram language model was trained on the target side of the parallel data using the srilm toolkit from stolcke .
|
a trigram language model with modified kneser-ney discounting and interpolation was used as produced by the srilm toolkit .
|
training is done using stochastic gradient descent over mini-batches with the adadelta update rule .
|
we train the model through stochastic gradient descent with the adadelta update rule .
|
in this task , we use the 300-dimensional 840b glove word embeddings .
|
for representing words , we used 100 dimensional pre-trained glove embeddings .
|
we used the google news pretrained word2vec word embeddings for our model .
|
we use the word2vec skip-gram model to train our word embeddings .
|
semantic parsing is the problem of deriving a structured meaning representation from a natural language utterance .
|
semantic parsing is the problem of translating human language into computer language , and therefore is at the heart of natural language understanding .
|
with regard to inputs , we use 50-d glove word embeddings pretrianed on wikipedia and gigaword and 5-d postion embedding .
|
we use glove vectors with 100 dimensions trained on wikipedia and gigaword as word embeddings , which we do not optimize during training .
|
this method does not use any parallel corpora for learning .
|
the method is a naive-bayes classifier which learns from noisy data .
|
we trained word vectors with the two architectures included in the word2vec software .
|
we also used word2vec to generate dense word vectors for all word types in our learning corpus .
|
the first is a reimplementation of the pronoun prediction neural network proposed by hardmeier et al .
|
the first component of our model is a modified reimplementation of the pronoun prediction network introduced by hardmeier et al .
|
we formulate the inference procedures in the training algorithm as integer linear programming ( ilp ) problems , ( ii ) we introduce a soft-constraint in the ilp objective to model noisy-or in training .
|
we formulate the inference procedures in training as integer linear programming ( ilp ) problems and implement the relaxation to the “ at least one ” heuristic via a soft constraint in this formulation .
|
system that participated in semeval-2013 task 2 : sentiment analysis in twitter .
|
our system participated in semeval-2013 task 2 : sentiment analysis in twitter ( cite-p-12-3-1 ) .
|
details about svm and krr can be found in .
|
more details about svm and krr can be found in .
|
to gain a more accurate basis for the pattern search , the oc uses the stanford parser to derive grammatical structures for each sentence .
|
the oc makes use of the stanford parser to derive grammatical structures for each sentence , which then form a more accurate basis for the later pattern search .
|
the mert was used to tune the feature weights on the development set and the translation performance was evaluated on the test set with the tuned weights .
|
the feature weights for each system were tuned on development sets using the moses implementation of minimum error rate training .
|
color ¨c name pairs obtained from an online color design forum , we evaluate our model on a ¡° color turing test ¡± and find that , given a name , the colors predicted by our model are preferred by annotators to color names created by humans .
|
using a large set of color¨cname pairs obtained from a color design forum , we evaluate our model on a ¡°color turing test¡± and find that , given a name , the colors predicted by our model are preferred by annotators to color names created by humans .
|
for this reason , we used glove vectors to extract the vector representation of words .
|
for input representation , we used glove word embeddings .
|
we presented a complete , correct , terminating extension of earley ' s algorithm that uses restriction .
|
in section 4 , we develop a correct , complete and terminating extension of earley 's algorithm for the patr-ii formalism using the restriction notion .
|
the semantic roles in the example are labeled in the style of propbank , a broad-coverage human-annotated corpus of semantic roles and their syntactic realizations .
|
the semantic roles in the examples are labeled in the style of propbank , a broad-coverage human-annotated corpus of semantic roles and their syntactic realizations .
|
we use the skll and scikit-learn toolkits .
|
for data preparation and processing we use scikit-learn .
|
our sd is the penn treebank of wall street journal text .
|
we used the wall street journal articles article boundary .
|
sentiment classification is the task of identifying the sentiment polarity of a given text .
|
sentiment classification is the task of labeling a review document according to the polarity of its prevailing opinion ( favorable or unfavorable ) .
|
also , li et al incorporate textual topic and userword factors with supervised topic modeling .
|
meanwhile , li et al present a topic model incorporating reviewer and item information for sentiment analysis .
|
for language modeling , we use kenlm to train 6-gram character-level language models on opensubs f iltered and huawei m onot r .
|
after standard preprocessing of the data , we train a 3-gram language model using kenlm .
|
as for ej translation , we use the stanford parser to obtain english abstraction trees .
|
we use the stanford dependency parser to extract nouns and their grammatical roles .
|
then we review the path ranking algorithm introduced by lao and cohen .
|
we briefly review the path ranking algorithm , described in more detail by lao and cohen .
|
event schema is a high-level representation of a bunch of similar events .
|
an event schema is a structured representation of an event , it defines a set of atomic predicates or facts and a set of role slots that correspond to the typical entities that participate in the event .
|
our baseline is a phrase-based mt system trained using the moses toolkit .
|
both systems are phrase-based smt models , trained using the moses toolkit .
|
which has the dual effect of factoring computationally costly null heads out from parsing ( but not from the resulting parse trees ) and rendering mgs fully compatible for the first time with existing supertagging techniques .
|
the second is a method for factoring computationally costly null heads out from bottom-up mg parsing ; this has the additional benefit of rendering the formalism fully compatible for the first time with highly efficient markovian supertaggers .
|
we assessed the statistical significance of f-measure improvements over baseline , using the approximate randomness test .
|
statistical significance of system differences in terms of f1 was assessed by an approximate randomization test .
|
fasttext is a library for efficient text classification and representation learning .
|
fasttext is a simple and effective method for classifying texts based on n-gram embeddings .
|
this is the first attempt at infusing general world knowledge for task specific training of deep learning .
|
to the best of our knowledge this is the first attempt to incorporate world knowledge from a knowledge base for learning models .
|
when used as the underlying input representation , word vectors have been shown to boost the performance in nlp tasks .
|
latent feature vectors have been recently successfully exploited for a wide range of nlp tasks .
|
later , their work was extended to take into account syntactic structure and grammars .
|
later , their work was extended to take into account the syntactic relation between words and grammars .
|
some researchers have applied the rule of transliteration to automatically translate proper names .
|
some researchers have found that transliteration is quite useful in proper name translation .
|
in addition , machine translation systems can be improved by training on sentences extracted from parallel or comparable documents mined from the web .
|
for instance , machine translation systems can benefit from training on sentences extracted from parallel or comparable documents retrieved from the web .
|
coreference resolution is the problem of identifying which noun phrases ( nps , or mentions ) refer to the same real-world entity in a text or dialogue .
|
coreference resolution is a well known clustering task in natural language processing .
|
we implemented the algorithms in python using the stochastic gradient descent method for nmf from the scikit-learn package .
|
within this subpart of our ensemble model , we used a svm model from the scikit-learn library .
|
a language model is a probability distribution that captures the statistical regularities of natural language use .
|
traditionally , a language model is a probabilistic model which assigns a probability value to a sentence or a sequence of words .
|
similarity is a kind of association implying the presence of characteristics in common .
|
similarity is a fundamental concept in theories of knowledge and behavior .
|
sentiment analysis is a much-researched area that deals with identification of positive , negative and neutral opinions in text .
|
sentiment analysis is a research area in the field of natural language processing .
|
in the translation tasks , we used the moses phrase-based smt systems .
|
we used moses with the default configuration for phrase-based translation .
|
for training the model , we use the linear kernel svm implemented in the scikit-learn toolkit .
|
we use a random forest classifier , as implemented in scikit-learn .
|
in this paper , we took a focused form of humorous tercets in hindi-dur se dekha , and performed an analysis of its structure and humour .
|
in this paper , we look at dur se dekha jokes , a restricted domain of humorous three liner poetry in hindi .
|
the annotation scheme is based on an evolution of stanford dependencies and google universal part-of-speech tags .
|
the ud scheme is built on the google universal part-of-speech tagset , the interset interlingua of morphosyntactic features , and stanford dependencies .
|
the present paper is a report of these investigations , their results and conclusions drawn therefrom .
|
the present paper is a contribution towards this goal : it presents the results of a large-scale evaluation of window-based dsms on a wide variety of semantic tasks .
|
the smt weighting parameters were tuned by mert using the development data .
|
the same data was used for tuning the systems with mert .
|
the automobile , kitchen and software reviews are taken from blitzer et al .
|
the automobile and software reviews 2 are taken from blitzer et al .
|
openccg uses a hybrid symbolic-statistical chart realizer which takes logical forms as input and produces sentences by using ccg com- binators to combine signs .
|
openccg uses a hybrid symbolic-statistical chart realizer which takes logical forms as input and produces sentences by using ccg combinators to combine signs .
|
despite its simplicity , our directional similarity approach provides a robust model for relational similarity .
|
empirically , mixing heterogeneous models tends to make the final relational similarity measure more robust .
|
the basic idea of this approach is to project the word indices onto a continuous space and to use a probability estimator operating on this space .
|
the basic idea of the neural network lm is to project the word indices onto a continuous space and to use a probability estimator operating on this space .
|
five-gram language models are trained using kenlm .
|
word-based lms were trained using the kenlm package .
|
we use svm-light-tk to train our reranking models , 9 which enables the use of tree kernels in svm-light .
|
we used svm-light-tk , which enables the use of the partial tree kernel .
|
one of the main stumbling blocks for spoken natural language understanding systems is the lack of reliability of automatic speech recognizers .
|
one of the main stumbling blocks for spoken dialogue systems is the lack of reliability of automatic speech recognizers .
|
sentiment analysis is the task of automatically identifying the valence or polarity of a piece of text .
|
sentiment analysis is a fundamental problem aiming to give a machine the ability to understand the emotions and opinions expressed in a written text .
|
the language model is a large interpolated 5-gram lm with modified kneser-ney smoothing .
|
we choose modified kneser ney as the smoothing algorithm when learning the ngram model .
|
all the feature weights and the weight for each probability factor are tuned on the development set with minimum-error-rate training .
|
the feature weights are tuned to optimize bleu using the minimum error rate training algorithm .
|
we use the stanford nlp pos tagger to generate the tagged text .
|
we use stanford log-linear partof-speech tagger to produce pos tags for the english side .
|
neural models can be categorized into two classes : recursive models and convolutional neural networks ( cnn ) models .
|
most methods fall into three types : unordered models , sequence models , and convolutional neural networks models .
|
thus , optimizing this objective remains straightforward with the expectation-maximization algorithm .
|
hence we use the expectation maximization algorithm for parameter learning .
|
as training examples , we formulate the learning problem as a structured prediction problem and derive a maximum-margin algorithm .
|
by taking a structured prediction approach , we provide a large-margin method that directly optimizes a convex relaxation of the desired performance measure .
|
we solve this problem by adding shortcut connections between different layers inspired by residual networks .
|
we further add skip connections between the lstm layers to the softmax layers , since they are proved effective for training neural networks .
|
throughout this work , we use mstperl , an unlabelled first-order non-projective single-best implementation of the mstparser of mcdonald et al , trained using 3 iterations of mira .
|
throughout this work , we use mstperl , an implementation of the mstparser of mcdonald et al , with first-order features and non-projective parsing .
|
we used the weka implementation of svm with 10-fold cross-validation to estimate the accuracy of the classifier .
|
we used the weka implementation of na茂ve bayes for this baseline nb system .
|
here , we propose s em a xis , a simple yet powerful framework to characterize word semantics .
|
in this work , we propose s em a xis , a lightweight framework to characterize domain-specific word semantics beyond sentiment .
|
abstract meaning representation is a sembanking language that captures whole sentence meanings in a rooted , directed , labeled , and acyclic graph structure .
|
abstract meaning representation is a semantic formalism which represents sentence meaning in a form of a rooted directed acyclic graph .
|
given in advance , we are interested in utilizing the emotion information in microblog messages for real-world event detection .
|
we evaluate our approach on large-scale microblog data sets by using real-world event list for each community .
|
a 5-gram lm was trained using the srilm toolkit 12 , exploiting improved modified kneser-ney smoothing , and quantizing both probabilities and back-off weights .
|
a 5-gram lm was trained using the srilm toolkit 5 , exploiting improved modified kneser-ney smoothing , and quantizing both probabilities and back-off weights .
|
topic models such as latent dirichlet allocation have emerged as a powerful tool to analyze document collections in an unsupervised fashion .
|
topic models , such as plsa and lda , have shown great success in discovering latent topics in text collections .
|
key ciphers also use a secret substitution function .
|
these ciphers use a substitution table as the secret key .
|
sentiment analysis is a research area where does a computational analysis of people ’ s feelings or beliefs expressed in texts such as emotions , opinions , attitudes , appraisals , etc . ( cite-p-12-1-3 ) .
|
sentiment analysis is the computational analysis of people ’ s feelings or beliefs expressed in texts such as emotions , opinions , attitudes , appraisals , etc . ( cite-p-11-3-3 ) .
|
coreference resolution is the problem of identifying which mentions ( i.e. , noun phrases ) refer to which real-world entities .
|
coreference resolution is the task of determining whether two or more noun phrases refer to the same entity in a text .
|
in the multi-agent decentralizedpomdp reach implicature-rich interpretations simply as a by-product of the way they reason about each other to maximize joint utility .
|
we show that agents in the dec-pomdp reach implicature-rich interpretations simply as a byproduct of the way they reason about each other to maximize joint utility .
|
sennrich et al also created synthetic parallel data by translating target-language monolingual text into the source language .
|
sennrich et al proposed a method using synthetic parallel texts , in which target monolingual corpora are translated back into the source language .
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.