Project RSS Feeds
New York, US
More information to appear at http://www.kdd.org/kdd2014/
What are best examples of data-driven Web applications you've ever seen? The updates to Open Street Map after the Haiti earthquake? The mapping of all 9,966,539 buildings in the Netherlands? The NHS Prescription data? Things like SF Park that help you 'park your car smarter' in San Francisco using real time data? Bing maps and Google Earth?
All these and many, many more data-driven applications have geospatial information at their core. Very often the common factor across multiple data sets is the location data, and maps are crucial in visualizing correlations between data sets that may otherwise be hidden.
It's this desire to work with multiple data sets in different formats about different topics and link those with the powerful technologies used in geospatial information systems that is behind the linking geospatial data workshop.
How can geographic information best be integrated with other data on the Web? How can we discover that different facts in different data sets relate to the same place, especially when 'place' can be expressed in different ways and at different levels of granularity?
On behalf of the Smart Open Data project, the World Wide Web Consortium (W3C), in partnership with the Open Geospatial Consortium (OGC) and the OGC GeoSPARQL Standards Working Group, the UK Government Linked Data Working Group, Google and Ordnance Survey, invite you to share your experiences, successes and frustrations in using GI.
The workshop is open to all and will take place at Campus London on Wednesday 5th - Thursday 6th March, 2014.
The tutorial provides a comprehensive view of the RDF-Stream Processing (RSP) research area. It consists of four parts. The first one introduces the RSP basic concepts: RDF streams to represent temporally-ordered sequence of data items; continuous SPARQL extensions to query RDF streams, and RSP engines to execute continuous query answering over RDF streams. The second part presents the available RSP engine implementations. It starts with an overview on the existing RSP engines, highlighting similarities and differences among them. Next, two existing implementations are analysed in depth: C-SPARQL and SPARQLstream. The third part is a hands-on session where the attendees learn how to (1) use the three presented RSP engines presented above and (2) let the systems interact among them. Finally, the fourth part of the tutorial provides an overview on RSP-related topics: RSP engine benchmarking, stream reasoning and real-world deployments. The tutorial closes with a discussion on the open challenges and the research problems of this research field.
The Centre for Applied Linguistics of the Santiago de Cuba’s branch of the Ministry of Science, Technology and the Environment, is pleased to announce the Fourteenth International Symposium on Social Communication. The event will be held in Santiago de Cuba, January 19 through the 23, 2015 and in this occasion will be dedicated to the 500 years of the foundation of the Santiago de Cuba's city. This interdisciplinary event will focus on social communication processes from the points of view of Linguistics, Computational Linguistics, Medicine, Mass Media, and Art, Ethnology and Folklore.
In the context of the XIV Symposium, will be held also the Workshop "Resources and tools of the Spanish and Portuguese languages and his variants in Latin America" sponsored by the Centre for Applied Linguistics and the Spanish Association on Natural Language Processing (SEPLN). The aims of the workshop are to know the new tools on NLP developed in the Spanish-speaking countries and Portuguese of Latin America and to know about Linguistic studies on Latin-America where NLP's instruments are applied.
More information to appear on the website: http://www.santiago.cu/hosting/linguistica/index.php?id=en
New York, US
More information to appear at the conference website
Download slides here
Download the slides here
The evolution of ontologies is an undisputed necessity in current research community. The problem of understanding this evolution is a fundamental problem as, based on this understanding, maintainers of depending artifacts need to take a decision about possible changes. Moreover, as ontologies are often developed by several ontology engineers, it is also important for them to understand what changes have been made by each other. Recent research focuses on just identifying and presenting the changes from one ontology version to another. In this paper, we argue that this is not enough and that we need more fine-grained methods for understanding how the ontology evolved. To this direction, we present a module, named ProvenanceTracker, which gets as input the list of changes between two or more RDF/S ontology versions and can answer fine-grained provenance queries about ontology resources. Our module can identify when a resource was created and how. The sequence of changes that led to the creation of that specific resource can be identified and presented to the user. We evaluate the time complexity of our approach and show that it can possibly reduce the human effort spent on understanding ontology evolution.AttachmentSize paper-30.pdf942.37 KB
In this work we describe and evaluate Hippalus, a system that offers exploratory search enriched with preferences. Hippalus supports the very popular interaction model of Faceted and Dynamic Taxonomies (FDT), enriched with user actions which allow the users to express their preferences. The underlying preference framework allows expressing preferences over attributes (facets), whose values can be hierarchically valued and/or multi-valued, and offers automatic conflict resolution. To evaluate the system we conducted a user study with a number of tasks related to a "car selection" scenario. The results of the comparative evaluation, with and without the preference actions, were impressive: with the preference-enriched FDT, all users completed all the tasks successfully in 1/3 of the time, performing 1/3 of the actions compared to the plain FDT. Moreover all users (either plain or expert) preferred the preference enriched interface. The benefits are also evident through various other metrics.AttachmentSize Papadakos_2014_ExploreDB.pdf1.76 MB
We report on our experiences with integrating geospatial datasets using Linked Data technologies. We describe NeoGeo, an integration vocabulary, and an integration scenario involving two geospatial
datasets: the GADM database of Global Administrative Areas and NUTS, the Nomenclature of Territorial Units for Statistics. We identify the need for provenance to be able to correctly interpret query results over the integrated dataset.AttachmentSize lgd14_submission_54.pdf272.52 KB
In this paper we describe Map4RDFiOS, a tool that allows visualizing and navigating through RDFbased geographic datasets available via a SPARQL endpoint, as well as connecting that data with statistical data represented with the W3C DataCube vocabulary or sensor data represented with the W3C Semantic Sensor Network ontology.AttachmentSize lgd14_submission_46.pdf5.56 MB
As the amount of available linked data expand and the number of related applications increases, the management of aspects such as provenance and access control of such data begin to become an issue. Current approaches do not provide sufficient support for automatic reasoning over different metadata and their possible interdependencies. MetaReasons is a framework that supports the representation of metadata in a logical formalism and consequently to support automated reasoning on metadata. Different types of metadata, such as data-provenance and accessibility-restrictions are represented as distinct meta-theories, and dependencies between types of metadata are represented by rules between different meta-theories. In this paper we present the logic based definition of the MetaReasons framework and two examples of meta-theories for provenance and access control. Moreover, we propose a materialization calculus for concrete forward reasoning on the two aspects.AttachmentSize TR-FBK-DKM-2014-01.pdf689.22 KB
BeRSys 2014 is the second edition of the series of BeRSys workshops; it provides a forum where topics related to the evaluation (included, but not limited to, expressive power, usability and performance) of RDF data management platforms can be discussed and elaborated.
The objectives of this workshop are to:
- create a discussion forum where researchers and industrials can meet and discuss topics related to the performance of RDF systems
- expose and initiate discussions on best practices, different application needs and scenarios related to RDF data management
Topics Of Interest
We welcome contributions presenting experiences with benchmarking RDF systems as well as technical contributions regarding the development of benchmarks for different aspects of RDF data management ranging from query processing and reasoning to data integration and ETL techniques. We welcome contributions from a diverse set of domain areas such as life science (bio-informatics, pharmaceutical domain), social networks, cultural informatics, news, digital forensics among others.
More specifically, the topics of interest include but are not limited
- descriptions of RDF data management use cases and query workloads
- benchmarks for SPARQL query and reasoning workloads
- benchmarks for RDF data integration tasks including but not limited to
ontology alignment, instance matching and ETL techniques
- benchmark metrics
- temporal and geospatial benchmarks
- evaluation of benchmark performance results on RDF engines
- benchmark principles
- query processing and optimisation algorithms for RDF systems
* Submission Deadline: June 15, 2014
* Notification of Acceptance: July 15, 2014
* Camera Ready Copy: August 1, 2014
* Workshop Day: September 5, 2014
More information is available at: http://events.sti2.at/bersys2014/Semantic WebComputingData managementInformationQuery languagesKnowledge representationRDFKnowledge engineeringRDF query languageSPARQLBenchmarkOntologyTechnologyquery processingsocial networkspharmaceutical domaindata managementBenchmarking RDF Systems
Submissions are open for the 3rd issue of the journal, which will be based on research work accepted for presentation at the 11th European Semantic Web Conference ESWC 2014 to be held in in Greece at Limenas Hersonissou, Crete from November 11 to 15.
More information is available at: http://planet-data.eu/semantic-data-management-video-journal/vol3Semantic WebElectronic submissionWorld Wide WebMedia technologyTechnologyGreeceWeb Conference ESWC
Submissions are open for the 3rd issue of the journal, which will be based on research work accepted for presentation at the 11th European Semantic Web Conference ESWC 2014 to be held in Greece at Limenas Hersonissou, Crete from May 25 to 29.Format of the Submissions
- Non-scripted, interview style video recordings or abstracts of research describing work from the area of semantic data management.
- Length of videos of 5 to 6 minutes.
- Filmed at a separate filming session at ESWC 2014
The videos will be peer reviewed by the Editors and members of the Editorial Board according to the following criteria: relevance to the area of semantic data management; technical quality of the content; educational quality of the content; accessibility to novice audiences; entertainment value.Benefits and Results
Higher visibility - we will provide links to the videos from the ESWC 2014 conference website and Springer Verlag will also provide links from their dedicated conference page. Each recording will have a link to the paper.Requirements
Authors should submit an abstract of their paper, slides and sign a Videolectures.Net release form. Click here to downloadExtra
Work package 4 is happy to announce the release of a corpus containing 147 million quasi-relational Web tables.
The Web contains vast amounts of HTML tables. Most of these tables are used for layout purposes, but a fraction of the tables is also quasi-relational, meaning that they contain structured data describing a set of entities. A corpus of Web tables can be useful for research and applications in areas such as data search, table augmentation, knowledge base construction, and for various NLP tasks.
The WDC Web Tables corpus has been extracted from the 2012 version of the Common Crawl, the largest Web crawl that is available to the public. The corpus contains the subset of the 11 billion HTML tables found in the Common Crawl that are likely quasi-relational. More information about the corpus, its application domains as well as information about how to download the corpus is found at: http://webdatacommons.org/webtables/
We want to thanks the Common Crawl Foundation for providing their great web crawl and thus enabling the creation of the WDC Web Tables corpus.
The creation of the WDC Web Tables corpus was supported by the German Research Foundation (DFG), the EU FP7 project PlanetData and by Amazon Web Services. We thank our sponsors a lot.
Applied linguisticsCorporaInternational Corpus of EnglishTechnologyHTMLWeb Tables corpusdata searchquasi-relational Web tablesWeb CrawlWeb tables
HealthCare Use Case
The Health Use Case Demo showcases the use of the access control enforcement techniques developed in PlanetData in order to provide selective exposure of patients' Personal Health Record information to various users/roles (doctors, medical staff, public services, organizations, hospitals etc) according to the access rights that the patient himself/herself has provided. More details on the demo can be found bellow:
Event Registry Use Case
The “Event Registry” system is developed as a prototype to support a standardization working group at the IPTC level (publishers’ standardization organization – IPTC.org). The aim is to release recommendations to collect, annotate and interoperate information on global events and storylines across languages, domains and granularities.
Smart Cities Use Case
The usecase describes some of the challenges and opportunities that arise from the existence of diverse sets of open (and closed) public and private data related to city infrastructures and territory, demography, public transport facilities and commercial activities across the city, specifically, focusing on the area of geomarketing.
A System for Aligning Taxonomies and Debugging Taxonomies and Their Alignments
Valentina Ivanova, Patrick Lambrix
PCKS: Commandments for Benchmarking Semantic Flow Processing Systems
Thomas Scharrenbach, Emanuele Della Valle
Applying SemanticWeb Technologies for Diagnosing Road Traffic Congestions
ourSpaces - Design and Deployment of a Semantic Virtual Research Environment
Google: “Get the best summary”
Andreas Thalhammer, Magnus Knuth
Market Sensing Using a Graph Database
Formal Verification of Data Provenance Records
A Formal Semantics for Weighted Ontology Mappings
Feature LDA: a Supervised Topic Model for Automatic Detection of Web API Documentations from the Web
Link Discovery with Guaranteed Reduction Ratio in Ane Spaces with Minkowski Measures
Axel-Cyrille Ngonga Ngomo
MORE: Modular Classification of OWL 2 Ontologies
Rapidly Integrating Services into the Linked Data Cloud
On the Diversity and Availability of Temporal Information in Linked Open Data
Discovering Concept Coverings in Aligning Ontologies of Linked Data