This paper reports the development of loglinear models for the disambiguation in wide-coverage HPSG parsing. The estimation of log-linear models requires high computational cost, especially with widecoverage grammars. Using techniques to reduce the estimation cost, we trained the models using 20 sections of Penn Treebank. A series of experiments empirically evaluated the estimation techniques, and also examined the performance of the disambiguation models on the parsing of real-world sentences. ...
This paper describes a wide-coverage Japanese grammar based on HPSG. The aim of this work is to see the coverage and accuracy attainable using an underspecified grammar. Underspecification, allowed in a typed feature structure formalism, enables us to write down a wide-coverage grammar concisely. The grammar we have implemented consists of only 6 ID schemata, 68 lexical entries (assigned to functional words), and 63 lexical entry templates (assigned to parts of speech ( B O S s ) ) .
Broad coverage lexicons for the English language have traditionally been handmade. This approach, while accurate, requires too much human labor. Furthermore, resources contain gaps in coverage, contain specific types of information, or are incompatible with other resources. We believe that the state of open-license technology is such that a comprehensive syntactic lexicon can be automatically compiled.
Common approaches to assessing document quality look at shallow aspects, such as grammar and vocabulary. For many real-world applications, deeper notions of quality are needed. This work represents a ﬁrst step in a project aimed at developing computational methods for deep assessment of quality in the domain of intelligence reports. We present an automated system for ranking intelligence reports with regard to coverage of relevant material.
This paper shows how ﬁnite approximations of long distance dependency (LDD) resolution can be obtained automatically for wide-coverage, robust, probabilistic Lexical-Functional Grammar (LFG) resources acquired from treebanks. We extract LFG subcategorisation frames and paths linking LDD reentrancies from f-structures generated automatically for the Penn-II treebank trees and use them in an LDD resolution algorithm to parse new text.
This paper describes a wide-coverage statistical parser that uses Combinatory Categorial Grammar (CCG) to derive dependency structures. The parser differs from most existing wide-coverage treebank parsers in capturing the long-range dependencies inherent in constructions such as coordination, extraction, raising and control, as well as the standard local predicate-argument dependencies. A set of dependency structures used for training and testing the parser is obtained from a treebank of CCG normal-form derivations, which have been derived (semi-) automatically from the Penn Treebank. ...
PCFGs can be accurate, they suffer from vocabulary coverage problems: treebanks are small and lexicons induced from them are limited. The reason for this treebank-centric view in PCFG learning is 3-fold: the English treebank is fairly large and English morphology is fairly simple, so that in English, the treebank does provide mostly adequate lexical coverage1 ; Lexicons enumerate analyses, but don’t provide probabilities for them; and, most importantly, the treebank and the external lexicon are likely to follow different annotation schemas, reﬂecting different linguistic perspectives.
Three key messages from The world health report: Universal health coverage, with full access to high-quality services for health promotion, prevention, treatment, rehabilitation, palliation and financial risk protection, cannot be achieved without evidence from research. Research has the power to address a wide range of questions about how we can reach universal coverage, providing answers to improve human health, well-being and development...
We describe an extension of the Wysiwym technology for knowledge editing through natural language feedback. Previous applications have addressed relatively simple tasks requiring a very limited range of nominal and clause patterns. We show that by adding a further editing operation called reconﬁguration, the technology can achieve a far wider coverage more in line with other general-purpose generators. The extension will be included in a Java-based library package for producing Wysiwym applications. ...
We present an approach to grammar development where the task is decomposed into two separate subtasks. The first task is hnguistic, with the goal of producing a set of rules that have a large coverage (in the sense that the correct parse is among the proposed parses) on a bhnd test set of sentences. The second task is statistical, with the goal of developing a model of the grammar which assigns maximum probability for the correct parse.
The paper demonstrates that exponential complexities with respect to grammar size and input length have little impact on the performance of three unification-based parsing algorithms, using a wide-coverage grammar. The results imply that the study and optimisation of unification-based parsing must rely on empirical data until complexity theory can more accurately predict the practical behaviour of such parserQ. 1.
We describe how unknown lexical entries are processed in a unification-based framework with large-coverage grammars and how from their usage lexical entries are extracted. To keep the time and space usage during parsing within bounds, information from external sources like Part of Speech (PoS) taggers and morphological analysers is taken into account when information is constructed for unknown words.
This paper describes a method using morphological rules and heuristics, for the automatic extraction of large-coverage lexicons of stems and root word-forms from a raw text corpus. We cast the problem of high-coverage lexicon extraction as one of stemming followed by root word-form selection. We examine the use of POS tagging to improve precision and recall of stemming and thereby the coverage of the lexicon. We present accuracy, precision and recall scores for the system on a Hindi corpus.
We discuss text summarization in terms of maximum coverage problem and its variant. We explore some decoding algorithms including the ones never used in this summarization formulation, such as a greedy algorithm with performance guarantee, a randomized algorithm, and a branch-andbound method. On the basis of the results of comparative experiments, we also augment the summarization model so that it takes into account the relevance to the document cluster.
The performance of a CDMA system can be difficult to assess. Link budget
calculations can be used to verify performance margins for a particular set of
parameters, but fail to tell the “whole” story in an easy to understand, visual
way. In this paper, a graphical depiction of CDMA coverage vs. capacity is
shown to be a useful tool in evaluating the performance of a CDMA system.
The graphical results are used to show scenarios in which CDMA can be both
uplink and downlink limited.
Maximum Security, provides updated, comprehensive, platform-by-platform coverage of security issues, and includes clear, to the point descriptions of the most common techniques hackers use to penetrate systems. This book provides information for security administrators and others interested in computer and network security and provides them with techniques to take steps to protect their systems.
The Digivance Indoor Coverage Solution (ICS) is used to extend coverage to specific areas
within a building, multiple buildings, or throughout a campus environment. This digital
distributed antenna system (D-DAS) solution features unique, patented technology that
distributes wireless coverage digitally over optical fiber, making Digivance ICS the best
choice for signal quality, flexibility, and overall performance. Its digitized optical RF transport
also future-proofs in-building wireless infrastructure for higher data rate services.
help for using digital recording devices. A full model student essay in
Chapter 3, entitled “Out Patients,” demonstrates effective research and
documentation of an online community.
More writing coverage throughout the book includes expanded fi eld-writing sections in each chapter, focusing on important topics such as
using language effectively, considering an audience, and working with
rhetoric. New objectives at the start of each chapter indicate the writing
skills covered in the chapter, guiding students to develop essential critical-thinking and rhetorical skills....
A Circuits and Systems perspective presents broad and in-depth coverage of the entire field of modern CMOS VLSI Design. The authors draw upon extensive industry and classroom experience to introduce today’s most advanced and effective chip design practices. They present extensively updated coverage of every key element of VLSI design, and illuminate the latest design challenges with 65 nm process examples. This book contains unsurpassed circuit-level coverage, as well as a rich set of problems and worked examples that provide deep practical insight to readers at all levels....
The increase in popularity of wireless devices has caused mobile
subscribers to demand coverage anytime and anywhere. Wireless
service providers are challenged with cost-effectively providing adequate
coverage and capacity precisely where it is needed. This is particularly
difficult in urban areas due to zoning restrictions and congestion from
tall buildings that limit the placement of cell sites for additional capacity.