Xem 1-20 trên 124 kết quả Probabilistic model
  • The PageRank algorithm, used in the Google search engine, greatly improves the results of Web search by applying probabilistic model on the link structure of Webs to evaluate the “importance” of Webs. In PageRank probabilistic model, the links and webs are uniform, so the rank score of webs are quite independent from their content. In practice, the researchers often hope that the web results can be ranked by their proposed topics. Moreover, when computer’s techniques solve given problems ineffectively, it’s necessary to do better research in theoretical problems. ...

    pdf12p tuanlocmuido 19-12-2012 15 2   Download

  • Recognizing entailment at the lexical level is an important and commonly-addressed component in textual inference. Yet, this task has been mostly approached by simplified heuristic methods. This paper proposes an initial probabilistic modeling framework for lexical entailment, with suitable EM-based parameter estimation. Our model considers prominent entailment factors, including differences in lexical-resources reliability and the impacts of transitivity and multiple evidence.

    pdf6p hongdo_1 12-04-2013 18 2   Download

  • Previous research applying kernel methods to natural language parsing have focussed on proposing kernels over parse trees, which are hand-crafted based on domain knowledge and computational considerations. In this paper we propose a method for defining kernels in terms of a probabilistic model of parsing. This model is then trained, so that the parameters of the probabilistic model reflect the generalizations in the training data. The method we propose then uses these trained parameters to define a kernel for reranking parse trees. ...

    pdf8p bunbo_1 17-04-2013 16 2   Download

  • This paper demonstrates that the use of ensemble methods and carefully calibrating the decision threshold can significantly improve the performance of machine learning methods for morphological word decomposition. We employ two algorithms which come from a family of generative probabilistic models. The models consider segment boundaries as hidden variables and include probabilities for letter transitions within segments. The advantage of this model family is that it can learn from small datasets and easily generalises to larger datasets. ...

    pdf9p hongdo_1 12-04-2013 15 1   Download

  • We describe two probabilistic models for unsupervised word-sense disambiguation using parallel corpora. The first model, which we call the Sense model, builds on the work of Diab and Resnik (2002) that uses both parallel text and a sense inventory for the target language, and recasts their approach in a probabilistic framework. The second model, which we call the Concept model, is a hierarchical model that uses a concept latent variable to relate different language specific sense labels.

    pdf8p bunbo_1 17-04-2013 11 1   Download

  • In this paper we deal with Named Entity Recognition (NER) on transcriptions of French broadcast data. Two aspects make the task more difficult with respect to previous NER tasks: i) named entities annotated used in this work have a tree structure, thus the task cannot be tackled as a sequence labelling task; ii) the data used are more noisy than data used for previous NER tasks. We approach the task in two steps, involving Conditional Random Fields and Probabilistic Context-Free Grammars, integrated in a single parsing algorithm.

    pdf11p bunthai_1 06-05-2013 22 3   Download

  • We present a statistical model for canonicalizing named entity mentions into a table whose rows represent entities and whose columns are attributes (or parts of attributes). The model is novel in that it incorporates entity context, surface features, firstorder dependencies among attribute-parts, and a notion of noise.

    pdf9p nghetay_1 07-04-2013 17 2   Download

  • This paper presents an incremental probabilistic learner that models the acquistion of syntax and semantics from a corpus of child-directed utterances paired with possible representations of their meanings. These meaning representations approximate the contextual input available to the child; they do not specify the meanings of individual words or syntactic derivations. The learner then has to infer the meanings and syntactic properties of the words in the input along with a parsing model.

    pdf11p bunthai_1 06-05-2013 18 2   Download

  • Expert search, in which given a query a ranked list of experts instead of documents is returned, has been intensively studied recently due to its importance in facilitating the needs of both information access and knowledge discovery. Many approaches have been proposed, including metadata extraction, expert profile building, and formal model generation. However, all of them conduct expert search with a coarse-grained approach. With these, further improvements on expert search are hard to achieve. ...

    pdf9p hongphan_1 15-04-2013 10 1   Download

  • In this paper we investigate the benefit of stochastic predictor components for the parsing quality which can be obtained with a rule-based dependency grammar. By including a chunker, a supertagger, a PP attacher, and a fast probabilistic parser we were able to improve upon the baseline by 3.2%, bringing the overall labelled accuracy to 91.1% on the German NEGRA corpus. We attribute the successful integration to the ability of the underlying grammar model to combine uncertain evidence in a soft manner, thus avoiding the problem of error propagation. ...

    pdf8p hongvang_1 16-04-2013 16 1   Download

  • Tuyển tập báo cáo các nghiên cứu khoa học quốc tế ngành hóa học dành cho các bạn yêu hóa học tham khảo đề tài: Research Article Matrix-Variate Probabilistic Model for Canonical Correlation Analysis

    pdf7p sting05 09-02-2012 17 4   Download

  • Tuyển tập báo cáo các nghiên cứu khoa học quốc tế ngành hóa học dành cho các bạn yêu hóa học tham khảo đề tài: A Probabilistic Model for Face Transformation with Application to Person Identification

    pdf12p sting12 11-03-2012 19 4   Download

  • This paper establishes a connection between two apparently very different kinds of probabilistic models. Latent Dirichlet Allocation (LDA) models are used as “topic models” to produce a lowdimensional representation of documents, while Probabilistic Context-Free Grammars (PCFGs) define distributions over trees. The paper begins by showing that LDA topic models can be viewed as a special kind of PCFG, so Bayesian inference for PCFGs can be used to infer Topic Models as well. Adaptor Grammars (AGs) are a hierarchical, non-parameteric Bayesian extension of PCFGs. ...

    pdf10p hongdo_1 12-04-2013 22 2   Download

  • Probabilistic models of sentence comprehension are increasingly relevant to questions concerning human language processing. However, such models are often limited to syntactic factors. This paper introduces a novel sentence processing model that consists of a parser augmented with a probabilistic logic-based model of coreference resolution, which allows us to simulate how context interacts with syntax in a reading task. Our simulations show that a Weakly Interactive cognitive architecture can explain data which had been provided as evidence for the Strongly Interactive hypothesis. ...

    pdf10p hongdo_1 12-04-2013 25 2   Download

  • We present a probabilistic model extension to the Tesni` re Dependency Structure e (TDS) framework formulated in (Sangati and Mazza, 2009). This representation incorporates aspects from both constituency and dependency theory. In addition, it makes use of junction structures to handle coordination constructions. We test our model on parsing the English Penn WSJ treebank using a re-ranking framework.

    pdf6p hongdo_1 12-04-2013 10 2   Download

  • We present an unsupervised model for joint phrase alignment and extraction using nonparametric Bayesian methods and inversion transduction grammars (ITGs). The key contribution is that phrases of many granularities are included directly in the model through the use of a novel formulation that memorizes phrases generated not only by terminal, but also non-terminal symbols. This allows for a completely probabilistic model that is able to create a phrase table that achieves competitive accuracy on phrase-based machine translation tasks directly from unaligned sentence pairs. ...

    pdf10p hongdo_1 12-04-2013 17 2   Download

  • Linking entities with knowledge base (entity linking) is a key issue in bridging the textual data with the structural knowledge base. Due to the name variation problem and the name ambiguity problem, the entity linking decisions are critically depending on the heterogenous knowledge of entities. In this paper, we propose a generative probabilistic model, called entitymention model, which can leverage heterogenous entity knowledge (including popularity knowledge, name knowledge and context knowledge) for the entity linking task. ...

    pdf10p hongdo_1 12-04-2013 15 2   Download

  • This paper defines a generative probabilistic model of parse trees, which we call PCFG-LA. This model is an extension of PCFG in which non-terminal symbols are augmented with latent variables. Finegrained CFG rules are automatically induced from a parsed corpus by training a PCFG-LA model using an EM-algorithm. Because exact parsing with a PCFG-LA is NP-hard, several approximations are described and empirically compared. In experiments using the Penn WSJ corpus, our automatically trained model gave a per40 formance of 86.

    pdf8p bunbo_1 17-04-2013 11 2   Download

  • Tuyển tập các báo cáo nghiên cứu về sinh học được đăng trên tạp chí sinh học quốc tế đề tài: Does probabilistic modelling of linkage disequilibrium evolution improve the accuracy of QTL location in animal pedigree?

    pdf10p toshiba18 08-11-2011 25 1   Download

  • Tham khảo sách 'probabilistic modeling in bioinformatics and medical informatics', y tế - sức khoẻ, y học thường thức phục vụ nhu cầu học tập, nghiên cứu và làm việc hiệu quả

    pdf510p hyperion75 22-01-2013 20 1   Download

Đồng bộ tài khoản