Historical research on World War II and the impact of large-scale violence largely depends on the availability of source materials: diaries, newspapers, eyewitness accounts, archival documents, photographs and videos, etc. Currently, these resources are held by a large number of memory institutions, often in analogue formats. For scholars, it can be challenging to find out which collections are relevant for their research and also what information can be found in these collections. In this...
Since the Simple Knowledge Organization System (SKOS) specification and its SKOS eXtension for Labels (SKOS-XL) became formal W3C recommendations in 2009, a significant number of conventional Knowledge Organization Systems (KOS) (including thesauri, classification schemes, name authorities, and lists of codes and terms, produced before the arrival of the ontology-wave) have made their journeys to join the Semantic Web mainstream. This paper uses "LOD KOS" as an umbrella term to refer to all...
At the present time, the Web is primarily designed for human consumption and not for computer consumption. This may seem like an unusual state of affairs, given that the web is vast and mature computerized information resource. However, we must recognize that the computer is presently used as the carrier of this information, and not as the consumer of the information. As a result, a great deal of the potential of the Web has yet to be realized. This book explores the challenges of automatic...
In this book we have been consistently directed by the vision of the Semantic Web. This vision can be summarized as the ability for computers to automatically use information on the Web in a similar way to humans. In particular, we want to be able to retrieve, comprehend, and exchange knowledge using automated techniques. At this point we have defined all of the main techniques that can be used to realize these goals. A summary of the four key techniques that we now have at our disposal is...
Dataset of terms from the Semantic Web Domain. The terms have been extracted automatically from a corpus of the Semantic Web Journal and the Journal of Semantic Web
A dataset of Austrian court decisions in German language prepared by Christian Sageder from Cybly in JSON-LD format compliant with LynxDocument schema (https://lynx-project.eu/doc/lkg/) - folder "original_json". Additionally, named entities annotations produced by a Bert-based transformer trained on WikiNer corpus - Per, Loc, Org, Misc - by DFKI team in N3 RDF notations, compliant with NIF2.1...
Stackexchange is also a collection of questions, some of which have been marked as duplicates by site-adminsitrators. I downloaded a data dump that contains the questions sorted by subject. I picked the chemistry subject because MEsH includes many chemistry-related concepts. There are 660 questions that have been marked as duplicates (class 1). I generated a further 861 pairs of non-duplicates (class 2) by looking for pairs of questions that aren't marked as either duplicates or related.
Formally, WiC is framed as a binary classification task. Each instance in WiC-TSV consists of a target word w with a corresponding target sense s represented by either its definition (subtask 1) or its hypernym/s (subtask 2), and a context c containing the target word w. The task aims to determine whether the meaning of the word w used in the context c matches the target sense s. In the following table there are some examples from the dataset. Subtasks WiC-TSV has three subtasks -...
PerfectO references, classifies and provides tools to encourage Semantic Web Best Practices to achieve Semantic Interoperability by focusing on ontology improvement. PerfectO is an ontology improvement methodology, implemented as a tool implemented as a tool PerfectO references, classifies and provides tools to encourage Semantic Web Best Practices to achieve Semantic Interoperability by focusing on ontology improvement. Permanent URL: http://purl.org/perfecto This URL redirects to the...
v0.0.1 August 2015 version of All UF triples from vivo.ufl.edu 2015-08-06 release 0.0.1 File: all-uf-triples-august-2015.tar.gz (178M) MD5: (all-uf-triples-august-2015.tar.gz) = 262c4e82be87cc24b04ccad188e75487 Download: http://vivo.ufl.edu/all-uf-triples-august-2015.tar.gz Unzips to all-uf-triples-august-2015.nt (~2.5GB) File Contains about 18.7 million lines $ wc -l all-uf-triples-august-2015.nt 18785880 all-uf-triples-august-2015.nt
We revise two relatively rare ensign wasp genera, whose species are restricted to Sub-Saharan Africa: Afrevania and Trissevania. Afrevania longipetiolata sp. nov., Trissevania heatherae sp. nov., T. hugoi sp. nov., T. mrimaensis sp. nov. and T. slideri sp. nov. are described, males and females of T. anemotis and Afrevania leroyi are redescribed, and an identification key for Trissevaniini is provided. We argue that Trissevania mrimaensis sp. nov. and T. heatherae sp. nov. populations are...
Collection movie, actor, director, genre, producer and gorss from DBpedia and Wikidata using Federated SPARQL Query
The Physical Semantic Web is a novel paradigm built upon the Google Physical Web approach and devoted to improve the quality of interactions in the Internet of Things. Semantic annotations accompany beacons instead of simple identifiers, i.e., machine-understandable descriptions of physical resources lay the foundations for novel ontology-based object advertisement and discovery. They in turn enable advanced user-to-thing and autonomous thing-to-thing interactions. An overall toolkit for...
The goal of `wbdataset` is to create tidy datasets from Wikidata or a Wikibase instance. The `wbdataset` package is an extension of the `dataset`, which in turn is an R package that helps to exchange, publish and combine datasets more easily by improving their semantics. The `wbdataset` extends the usability of dataset by connecting the Wikibase API with the R statistical environment.
Introduction This repository contains the code (and data) corresponding to the paper we presented at the 2020 workshop on Large Scale RDF Analytics (LASCAR), a workshop co-located with the Extended Semantic Web Conference (ESWC). The original paper is attached but the website (https://submassive.cc) where the original dataset and documentation are no longer maintained. Data preprocessing generate_cycles.py : generate the subgraph of the entire LOD-a-lot. This step is for the sake of memory...
The file contains the graph as json. Where do I add a repository? ;) https://github.com/AntonioNoack/WebPageRank/commit/0be63ef218f32676ef74b1077e375470233f2b9b was my last commit, when I created this file. The project there was used to calculate the PageRank and HITS values. Normalization: None PageRank random jump probability: 15% PageRank preference vector: None
v0.9 without all references properly written.
The EventKG is a multilingual resource incorporating event-centric information extracted from several large-scale knowledge graphs such as Wikidata, DBpedia and YAGO, as well as less structured sources such as the Wikipedia Current Events Portal and Wikipedia event lists in 15 languages. The EventKG is an extensible event-centric resource modelled in RDF. It relies on Open Data and best practices to make event data spread across different sources available through a common representation and...
Combining Knowledge Graphs and Deep Learning for categorizing Tweets
Linked lists represent a countable number of ordered values, and are among the most important abstract data types in computer science. With the advent of RDF as a highly expressive knowledge representation language for the Web, various implementations for RDF lists have been proposed. Yet, there is no benchmark so far dedicated to evaluate the performance of triple stores and SPARQL query engines on dealing with ordered linked data. Moreover, essential tasks for evaluating RDF lists, like...
Indigenising the Semantic Web: Ontologies for Indigenous knowledge and heritage resources on a machine-readable Web. This project will put Australia at the forefront of international efforts to realise a functioning Semantic Web in which all data transactions are handled by machines talking to machines. It addresses the government's call for the creation of infrastructure and e-research tools that enable high-speed distributed access to Indigenous knowledge and culture resources, and its...
Intelligent Applications Through The Semantic Web. The primary aim of the proposed bid is to build a network of Australian researchers and their international peers for condresearch into the fundamental as well as applied aspects of the Semantic Web. By incorporating meaning of web-content in a form that can be accessed and processed by intelligent software agents, the Semantic Web will allow computers and humans to work in cooperation. This research will address the needs of both the...
The World Wide Web has changed ways people find and sue information. Since its beginnings as a collaboration tool for physicists, it has become a truly global channel for distributed publication of information. More recently, where once content was passive text on a page, it is fast becoming dynamic, flexible and responsive to the needs of its users. The result is a fundamental change in what it means to be a knowledge-based industry. As key sectors within the knowledge economy, the future...
Personalized, Adaptive, and Semantic-driven Selection and Composition of Web Services. Web services are the pillar of the new generation of Internet technologies. They provide standardized access to functionality that would otherwise be hidden inside enterprise information systems. As the existing base of web services expands, there is a need for techniques to select, configure, assemble, and coordinate web services to perform complex user tasks. This project will advance the fundamental...
During the last decade the Semantic Web community has established basic standards for representing data and the conceptual systems (ontologies) through which they are defined. However, encoding information in these formalisms (OWL, RDF) remains a technically difficult task. Widespread adoption of these technologies (with their important potential benefits) would be facilitated if transparent interfaces to the technical formalisms were available.The project aims to show that metadata in OWL...
My work has largely been concerned with how to improve access to the right information at the right time on the Web via effective presentation both of how to explore information as well as how to represent the confluence of sources that can be explored. My goal is to find ways to make it ever easier for anyone to start with what they know and be able to use that to help them explore information to build the knowledge they want. For instance, one challenge was to help someone who knew nothing...
"The main purpose of this book is to sum up the vital and highly topical research issue of knowledge representation on the Web and to discuss novel solutions by combining benefits of folksonomies and Web 2.0 approaches with ontologies and semantic technologies. The book contains an overview of knowledge representation approaches in past, present and future, introduction to ontologies, Web indexing and in first case the novel approaches of developing ontologies. It ombines aspects of...
The social web has become a major repository of social and behavioral data that is of exceptional interest to the social science and humanities research community. Computer science has only recently developed various technologies and techniques that allow for harvesting, organizing and analyzing such data and provide knowledge and insights into the structure and behavior or people on-line. Some of these techniques include social web mining, conceptual and social network analysis and...
Imagine if you could teach your computer to make decisions that lead to faster B2B trading implementations, more effective customer management, and seamless data warehousing. Now, your imagination can be turned into reality. This authoritative book explores the Semantic Web, a revolutionary new framework for creating intelligent software applications that automate the processes of reasoning and decision-making. The Semantic Web first shows you how to vastly improve knowledge management in...
The Social Web (including services such as MySpace, Flickr, last.fm, and WordPress) has captured the attention of millions of users as well as billions of dollars in investment and acquisition. Social websites, evolving around the connections between people and their objects of interest, are encountering boundaries in the areas of information integration, dissemination, reuse, portability, searchability, automation and demanding tasks like querying. The Semantic Web is an ideal platform for...
"Semantic Web Programming" takes the semantic web directly and boldly into solving practical, real-world problems. Hands-on code examples drive the understanding of this new powerful technology that can unify and fully leverage the growing sea of data, information, and services available on the Internet. As the code examples build, the reader explores the many technologies that form the semantic web including the knowledge representations such as Resource Description Framework (RDF), Web...
Semantic Web Tutorial about turtle format using vocabulary.
The application showcases the ideas of a distributed, Semantic-Web enabled micro blogging architecture, providing a way to leverage this new Web 2.0 practice to the Semantic Web.
Organizations can benefit from integrating multilingual information from both textual and structured sources, and from its retrieval by means of Question Answering (QA) systems. Hybrid QA approaches, capable of finding answers in both documents and KGs, usually rely on translating textual sources into KG statements or vice-versa, and are often not leveraging the whole extent of a graph or the richness of the natural language text. Here we propose PORQUE, a hybrid QA system that utilizes...
.
This presentation provides deep insights into the methodology used to develop the Semantic Web Company’s Knowledge Hub as a core element of a broader knowledge management strategy. It will cover methodologies, architecture, implemented governance models, and the added value from a business perspective.
We outline some of the core features and capabilities of PoolParty Semantic Suite along the enterprise knowledge graph life cycle (EKG-LC), which provides also a comprehensive overview of the actors and agents involved during the most important operational steps for the (ongoing) development of the graph. This ranges from data inventory, extraction and curation, modeling (authoring), various transformation steps, to linking and enrichment (e.g., inferred data), and analysis or feedback of...
Taxonomies and ontologies are often merged. Ontologists might look upon taxonomies as the component or extension of ontologies to include hierarchies of subclasses. But it also goes the other way. Taxonomies can be semantically enriched to become ontologies with the addition of a high-level ontology model. This bottom-up approach to ontology creation leverages existing taxonomies. While stakeholder input and analysis of data are standard inputs into ontology design, taxonomies can serve as...
In this presentation, we introduce the ESG topic complex and we discuss how the use of a comprehensive ESG knowledge model can support the implementation of ESG strategies in a multitude of ways. For example, to optimize ESG risk management, enhance sustainability reporting, and even provide rich information around ESG topics to all types of stakeholders, including supply chain managers, risk managers, and ESG strategists. Based on Semantic Web Company’s ESG knowledge model, we mention...
This presentation shows the approach of making use of Knowledge Graphs in Data Spaces and Data Markets to foster data- and semantic interoperability. Interoperability is the enabler of efficient and sustainable data sharing between organisations, either in a certain industry or across industries, either in the form of data trading or as data collaborations. This talk will explain the basic principles of Data Spaces, draw the problem statement of interoperability for value-added data sharing,...
In a multi-stakeholder ecosystem, data exchange is not sufficient and semantic interoperability should be achieved to ensure different information systems can communicate without loss of meaning. A semantic conversion procedure relying on a reference ontology can enable an efficient conversion between data formats sharing common semantics, whilst generating an integrated and interoperable knowledge graph. The Chimera framework proposes a flexible and configurable solution to address...
Presentation by Dr Samantha Pearman-Kanza on A Pragmatic view of the Semantic Web for the Physical Sciences for The joint meeting on Units, Symbols, and Terminology in the Physical Sciences in and for the Digital Era that had representatives from: PSDI – Physical Sciences Data Infrastructure, CODATA DRUM Task Group – Digital Representation of Units of Measure, IUPAC Green Book – Quantities, Units, and Symbols in Physical Chemistry and IUPAC Gold Book – Compendium of Chemical Terminology....
