Ontology Learning for the Semantic Web explores techniques for applying knowledge discovery techniques to different web data sources (such as HTML documents, dictionaries, etc.), in order to support the task of engineering and maintaining ontologies. The approach of ontology learning proposed in Ontology Learning for the Semantic Web includes a number of complementary disciplines that feed in different types of unstructured and semi-structured data. This data is necessary in order to support a semi-automatic ontology engineering process. Ontology Learning for the Semantic Web is designed for researchers and developers of semantic web applications. It also serves as an excellent supplemental reference to advanced level courses in ontologies and the semantic web.
The promise of the Semantic Web is that future web pages will be annotated not only with bright colors and fancy fonts as they are now, but with annotation extracted from large domain ontologies that specify, to a computer in a way that it can exploit, what information is contained on the given web page. The presence of this information will allow software agents to examine pages and to make decisions about content as humans are able to do now. The classic method of building an ontology is to gather a committee of experts in the domain to be modeled by the ontology, and to have this committee.
This publication advances the state-of-the-art in ontology learning by presenting a set of novel approaches to the semi-automatic acquisition, refinement and evaluation of logically complex axiomatizations. It has been motivated by the fact that the realization of the semantic web envisioned by Tim Berners-Lee is still hampered by the lack of ontological resources, while at the same time more and more applications of semantic technologies emerge from fast-growing areas such as e-business or life sciences. Such knowledge-intensive applications, requiring large scale reasoning over complex domains of interest, even more than the semantic web depend on the availability of expressive, high-quality axiomatizations. This knowledge acquisition bottleneck could be overcome by approaches to the automatic or semi-automatic construction of ontologies. Hence a huge number of ontology learning tools and frameworks have been developed in recent years, all of them aiming for the automatic or semi-automatic generation of ontologies from various kinds of data. However, both the quality and the expressivity of ontologies that can be acquired by the current state-of-the-art in ontology learning so far have failed to meet the expectations of people who argue in favor of powerful, knowledge-intensive applications based on logical inference. This work therefore takes a first, yet important, step towards the semi-automatic generation and maintenance of expressive ontologies.
This book constitutes carefully reviewed and revised selected papers from the 13th Chinese Lexical Semantics Workshop, CLSW 2012, held in Wuhan, China, in July 2012. The 67 full papers and 17 short papers presented in this volume were carefully reviewed and selected from 169 submissions. They are organized in topical sections named: applications on natural language processing; corpus linguistics; lexical computation; lexical resources; lexical semantics; new methods for lexical semantics; and other topics.
Pharmacogenomics is the study of how variation in the human genome impacts drug response in patients. It is a major driving force of "personalized medicine" in which drug choice and dosing decisions are informed by individual information such as DNA genotype. The field of pharmacogenomics is in an era of explosive growth; massive amounts of data are being collected and knowledge discovered, which promises to push forward the reality of individualized clinical care. However, this large amount of data is dispersed in many journals in the scientific literature and pharmacogenomic findings are discussed in a variety of non-standardized ways. It is thus challenging to identify important associations between drugs and molecular entities, particularly genes and gene variants. Thus, these critical connections are not easily available to investigators or clinicians who wish to survey the state of knowledge for any particular gene, drug, disease or variant. Manual efforts have attempted to catalog this information, however the rapid expansion of pharmacogenomic literature has made this approach infeasible. Natural Language Processing and text mining techniques allow us to convert free-style text to a computable, searchable format in which pharmacogenomic concepts such as genes, drugs, polymorphisms, and diseases are identified, and important links between these concepts are recorded. My dissertation describes novel computational methods to extract and predict pharmacogenomic relationships from text. In one project, we extract pharmacogenomic relationships from the primary literature using text-mining. We process information at the fine-grained sentence level using full text when available. In a second project, we investigate the use of these extracted relationships in place of manually curated relationships as input into an algorithm that predicts pharmacogenes for a drug of interest. We show that for this application we can perform as well with text-mined relationships as with manually curated information. This approach holds great promise as it is cheaper, faster, and more scalable than manual curation. Our method provides us with interesting drug-gene relationship predictions that warrant further experimental investigation. In the third project, we describe knowledge inference in the context of pharmacogenomic relationships. Using cutting-edge natural language processing tools and automated reasoning, we create a rich semantic network of 40,000 pharmacogenomic relationships distilled from 17 million Medline abstracts. This network connects over 200 entity types with clear semantics using more than 70 unique types of relationships. We use this network to create collections of precise and specific types of knowledge, and infer relationships not stated explicitly in the text but rather inferred from the large number of related sentences found in the literature. This is exciting because it demonstrates that we are able to overcome the heterogeneity of written language and infer the correct semantics of the relationship described by authors. Finally, we can use this network to identify conflicting facts described in the literature, to study change in language use over time, and to predict drug-drug interactions. These achievements provide us with new ways of interacting with the literature and the knowledge embedded within it, and help ensure that we do not bury the knowledge embodied in the publications, but rather connect the often fragmented and disconnected pieces of knowledge spread across millions of articles in hundreds of journals. We are thereby brought one step closer to the realization of personalized medicine and ensure that as scientists, we continue to build on the knowledge discovered by past generations and truly to stand on the shoulders of giants.
The latest title in Black Library's premium line. Perturabo - master of siegecraft, and executioner of Olympia. Long has he lived in the shadow of his more favoured primarch brothers, frustrated by the mundane and ignominious duties which regularly fall to his Legion. When Fulgrim offers him the chance to lead an expedition in search of an ancient and destructive xenos weapon, the Iron Warriors and the Emperor's Children unite and venture deep into the heart of the great warp-rift known only as 'the Eye'. Pursued by a ragged band of survivors from Isstvan V and the revenants of a dead eldar world, they must work quickly if they are to unleash the devastating power of the Angel Exterminatus