exploring lod through metadata extraction and data-driven...

21
Exploring LOD through metadata extraction and data-driven visualizations Abstract Purpose: This article presents a new approach towards automatically visualising Linked Open Data through metadata analysis. Approach: By focusing on the data within a LOD dataset, we can infer its structure in a much better way than current approaches, generating more intuitive models to progress towards visual representations. Findings: With no technical knowledge required, focusing on metadata properties from a semantically annotated dataset could lead to automat- ically generated charts that allow to understand the dataset in an ex- ploratory manner. Through interactive visualizations, users can navigate Linked Open Data sources using a natural approach, in order to save time and resources when dealing with an unknown resource for the first time. Research limitations*: This approach is suitable for available SPARQL endpoints and could be extended for RDF dumps loaded locally. Originality: Most works dealing with LOD visualization are customized for a specific domain or dataset. This article proposes a generic approach based on traditional data visualization and Exploratory Data Analysis literature. Keywords: Linked Open Data, LOD Visualization, Semantic Web, Ex- ploratory Data Analysis, Metadata Extraction, Datatype Inference Article classification: Research paper 1 Introduction We can undoubtedly declare that we live in the age of data. Facts are present in every and each single event within our environment. Such is the significance of measuring and assigning a value to everything, that we have developed a wide variety of expressions to name all the interactions we perform with data: we talk about big data when working with colossal amounts of it; we mine data in order to detect patterns (in a clear reference to the extraction of gold and other precious metals/stones from dirt), we structure; farm; curate; manage; analyse; model; process and link it... but all this is senseless if it does not ultimately lead to a greater understanding of our surroundings. With an increased ease to publish facts on the Internet, more and more data is being made accessible from anywhere in the world, by anybody. This 1

Upload: others

Post on 06-Jul-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

Exploring LOD through metadata extraction anddata-driven visualizations

Abstract

Purpose: This article presents a new approach towards automaticallyvisualising Linked Open Data through metadata analysis.Approach: By focusing on the data within a LOD dataset, we can inferits structure in a much better way than current approaches, generatingmore intuitive models to progress towards visual representations.Findings: With no technical knowledge required, focusing on metadataproperties from a semantically annotated dataset could lead to automat-ically generated charts that allow to understand the dataset in an ex-ploratory manner. Through interactive visualizations, users can navigateLinked Open Data sources using a natural approach, in order to save timeand resources when dealing with an unknown resource for the first time.Research limitations*: This approach is suitable for available SPARQLendpoints and could be extended for RDF dumps loaded locally.Originality: Most works dealing with LOD visualization are customizedfor a specific domain or dataset. This article proposes a generic approachbased on traditional data visualization and Exploratory Data Analysisliterature.

Keywords: Linked Open Data, LOD Visualization, Semantic Web, Ex-ploratory Data Analysis, Metadata Extraction, Datatype InferenceArticle classification: Research paper

1 IntroductionWe can undoubtedly declare that we live in the age of data. Facts are present inevery and each single event within our environment. Such is the significance ofmeasuring and assigning a value to everything, that we have developed a widevariety of expressions to name all the interactions we perform with data: wetalk about big data when working with colossal amounts of it; we mine data inorder to detect patterns (in a clear reference to the extraction of gold and otherprecious metals/stones from dirt), we structure; farm; curate; manage; analyse;model; process and link it... but all this is senseless if it does not ultimatelylead to a greater understanding of our surroundings.

With an increased ease to publish facts on the Internet, more and moredata is being made accessible from anywhere in the world, by anybody. This

1

Page 2: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

concept was further enriched by Sir Tim Berners-Lee’s proposal to evolve froma document-based web (in which text documents were interlinked) to a data-driven space (Berners-Lee et al. 2001), which was later described in detail inwhat we actually know as the Linked Data (LD) principles (Berners-Lee 2006):a set of premises to publish data on the web in order to make it understandableand processable by machines, with the potential to connect it to external datasources to avoid redundancy and foster knowledge discovery.

Some authors have already discussed the milestones reached by the LD com-munity since its conception (Bizer et al. 2009, Shadbolt et al. 2006), draftingthe path towards its full realization. Actually, the two main pillars of LD are itspublication and its consumption. Most developed technologies and frameworkshave focused on the former, making applications and tools available to allowlinkage, provide provenance and fact validation, making queries easier to en-code, etc. However, LD consumption has been left to automatic algorithms andSemantic Web experts. Despite LD was originally designed for machine con-sumption, the truth is that at some point users become involved in the process.The extra efforts required to understand some of the concepts (for example,SPARQL querying when most developers are familiar with SQL), the require-ments to deploy and maintain SPARQL endpoints (versus the most extendedapproach of REST APIs) and other related issues diminish the usefulness per-ception from humans, which may explain why LD and semantic technologies arestill waiting to get mass traction.

Our approach tries to exhibit the main benefits of Linked Open Data (LOD,LD with Open Licenses) through visual representations, with no prior knowledgerequired of Semantic Web nor Visualization-related technologies. To that end,we have implemented a LOD visualization tool which analyses the structure ofthe dataset, extracts different metadata values and generates visual representa-tions which are suitable to render the data. On section 2 we review previousapproaches which have dealed with LOD visualization. Section 3 states theproblem and sets the goals to be achieved by the envisaged visualization pro-totype. Next, on section 4 our approach is presented, describing the differentmodules which constitute John Snow, our designed and implemented tool. Sec-tion 5 presents the evaluation stage, during which real users where used tovalidate the proposal. We finish our article by summarising the conclusions andlisting some future lines of work.

2 Literature reviewFor the last decade, the Semantic Web and Linked Data communities haveworked unrelentingly towards Berners-Lee’s vision for the next Internet stage.Regarding LOD communication and exploitation, many works have relied on vi-sualization as a mean to reach these goals, entrusting humans visual abstractionabilities.

After the first years of work from the community, when many tools weredesigned and developed to realise the Semantic Web, a seminal survey was pub-

2

Page 3: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

lished (Dadzie & Rowe 2011) analysing the State of the Art on LD exploration.The authors evaluated the most relevant browsers designed to work with LD,splitting them in those only capable of answering in plain text format, and thosewith visualization features. They concluded that most tools were still in theirinfancy, with a clear target on technical users (those already contributing withinthe LD community), so more elaborated visualization and browsing tools wouldplay a crucial role in the future of LD consumption.

We provide a brief summary of some tools developed since their survey,focusing on those works which produce visualizations as output. We couldclassify the existing visualization tools according to their approach.

2.1 Domain or ontology specificThis group is formed by all those tools and applications which are tailored fora specific knowledge area or vocabulary. Most existing works fall within thiscategory, as knowing what the user expects to have visualized beforehand is agreat advantage. Also, certain domains have standardised ways to present dataand results through visualizations, so the design will have those set of chartsand graphs as their goal.

These features can be observed in Map4RDF (Leon et al. 2012), a tooldeveloped to map and browse geospatial datasets, offering the possibility tofilter the presented data through facets. LODVizSuite (Mechant et al. 2014) isa prototype customised to represent the dataset containing information aboutFlemish researchers, universities and projects, rendering graphs and histogramswhich summarise the research landscape in Belgium. A similar application wasdesigned for the Semantic Web Journal, by IOS Press1, named as the SWGETPortal (Fionda et al. 2014).

Regarding ontology-driven visualizers, FOAF.Vix2 allows to explore the rela-tions defined by the FOAF (Friend-Of-A-Friend) vocabulary (Brickley & Miller2012) using a tabular-based template. FOAF is widely used across differentdomains to portray person instances.

The main drawbacks of these approaches are that they either need to beredesigned in order to work with new data sources or vocabularies, which werenot originally considered when devoloping the tool, or are completely unable torepresent data from other fields (Map4RDF, for example, will never be able torender data which has no geographical features). Our proposal is independentfrom both vocabularies and domains, avoiding the need for further adjustmentsin order to make the tool work in unforeseen scenarios.

2.2 Domain agnostic or generic approachesThe previously listed applications are a great starting point to move away fromthe classical LOD presentation in tabular format (Cyganiak & Bizer 2008) ornode-link graphs (Deligiannidis et al. 2007, Hoefler et al. 2013), which seems

1http://semantic-web-journal.net/2http://foaf-visualizer.gnu.org.ua/

3

Page 4: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

to be the de-facto approach to visualize LOD datasets, probably due to theirencoding as Resource Description Framework (RDF)3 triples. However, to pro-mote tool reusability and adopt the benefits from visualising data, more genericapproaches are required, i.e., visualization tools which are able to generate anoutput for a wide variety of knowledge areas.

The Linked Data Visualization Model (LDVM) (Brunetti et al. 2013) pro-poses an interesting adaptation of the Data State Reference Model (Chi 2000) tothe LOD’s scene, establishing a visualization pipeline to transform raw data inRDF into visualizations, thanks to a set of transformations and abstractions ap-plied in each pipeline’s step. A LDVM’s implementation; Payola (Klımek et al.2014) demonstrates how the conceptual pipeline can be used with real data.Still, LDVM’s implementations are aimed to technical users with previous ex-perience in the SW field, and the chart rendering process might be challengingfor more standard user profiles.

A tool conceived to be truly generic should lower the entry barrier anddemocratise access to LD resources. Casual and non-technical users are notable to download and deploy a complex tool in order to perform a basic dataexploratory analysis, thus they usually use the built-in features of existing soft-ware solutions for data management, both online an desktop-based (any spread-sheets software, Tableau4, etc.). The main drawback of these tools are that theyare not designed to work with some LOD formats (especially RDF), and the re-quirement to have some basic analitical skills, as the users need to know thestatistical functions beforehand in order to start their analyses.

Our goal is to provide an online LOD visualization tool (accessible throughany standard web browser), agnostic from the domains and vocabularies covered,which provides insights and suitable visualizations for any user who wants toexplore a dataset.

3 Problem statementInformation visualization was defined as the use of computers to amplify cog-nition interactively, using visual representations of abstract data (Card et al.1999). Due to the variety of topics and ontologies present on those LD reposi-tories with an Open License, visualization could encourage end users to adoptLOD in their exploration tasks, which would eventually lead to a better supportfrom data publishers to make more facts available as LOD. Unless we providesolutions to involve casual users, i.e., those without a strong technical back-ground, but with a curious and analytical attitude (Elmqvist 2014), the usageof LOD will continue to diminish, failing with the best opportunity we have topopularise Berners-Lee’s vision.

3http://www.w3.org/RDF/4http://tableau.com

4

Page 5: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

In an effort to design a generic approach to deal with the visualization ofLOD, we tried to answer the following research question: How can we generatesmart, coherent and automatic visualizations for LOD?.

First, we must describe what we mean for the highlighted words above:

• Smart: The main contribution of LD is the annotation of concepts usingshared schemas, allowing two different actors to agree on the definitionsof a set of resources. This annotations are understandable by machines,so algorithms with a semantic input should provide richer outputs (visu-alizations) than those created from raw data.

• Coherent: The generated visual representations should be suitable for thedescribed datum. If geographical features are identified within the datasetin the format of latitude—longitude pairs of values, a map should be thefirst option to plot the resources. Other tools that visualise structureddata (e.g., Google Spreadsheet’s Explore feature, the chart generation toolof most spreadsheet software, etc) detect that these properties are typedas numbers, and render them on either a scatterplot or a column chart,missing the opportunity to provide real value to end users.

• Automatic: Our intention is to lower the knowledge barrier requiredto explore semantic datasets, independently of the users’ background.Whereas less interaction means less customization of the resulting visual-izations, it also reduces the resources and skills needed to play with thedata, thus easing the path towards knowledge discovery using LOD.

Consequently, we envisaged a prototype tool to explore LOD repositories:John Snow. The tool has been named after the famous English physician whotraced the origin of the 1854’s cholera breakout in London to a particular waterpipe by plotting the deaths on a map. John Snow’s design was conceived tofavour class and property exploration, generating different visualizations whenpossible with the following requirements in mind:

• No field expertise or data analysis skills are needed in order toconduct an exploratory study.

• Regardless of the topics covered by the datasets, or the vocabularies usedto describe them, the prototype tool should provide visual feedbacks,in the form of summaries and insights for the data.

• Take a data-driven approach. The automatic generation of many LODrepositories produces lots of errors (missing attributes, blank axioms, in-correct typing, etc.), thus we can not assume data to be as correct asintended by their publishers.

• Generate meaningful visualizations, i.e., those which could really con-tribute to a greater understanding of the data, instead of trying to renderas many charts as possible.

5

Page 6: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

4 A generic approach towards LOD visualiza-tion

In order to answer the previous issues, we took Information Visualization’s rootsand adapted them to the LOD field. When a user is presented a dataset shedoes not have any prior experience with, the same naıve question is alwaysformulated: What is this data about?

In accordance with Information Theory, vision is the sense with the largestbandwith for sending information to the brain (Ware 2012), so it seems logical topromote data visualization as a valid approach towards knowledge acquisition.This idea is also supported by the a picture is worth a thousand words adage,which summarises humans ability to quickly understand complex data in anabstract way. Mathematician John W. Tukey explored these concepts more thanfour decades ago, introducing Exploratory Data Analysis (EDA) (Tukey 1977)as a set of basic statistical techniques supported by graphical representationsto deal with new data sources. EDA does not require a model to fit the datain order to start the analysis stage. Instead, it takes data as-is (allowing somepre-processing to clean the data), and creates a data-driven model, lacking therigidity of more formal approaches.

By looking at the data itself, the exploratory analisis can take a more nat-ural and suggestive approach, avoiding the biases produced by pre-establishedassumptions of more traditional methodologies. This path was also pursued bythe authors of the follow your nose principle (Yu 2011), a process to dynamicallydiscover new data by playing with the internal connections between resourcesand their attributes.

4.1 Understanding the essence of the dataThe first task consists on figuring out what is being portrayed. LOD is pub-lished following publicly available ontologies, describing which conceptual classesthe resources belong to and the properties that characterise each data entity.OWL5 ontologies allow to declare ranges and domains for each property. Do-mains list the classes those properties are intended for, whereas ranges restrictthe datatypes values can adopt. These values fall in one of the following groups:

• An object-property, which relates two resource individuals, that is, theaxiom’s value will be an instance of an ontological class, thus allowing theconnection betweem them and fostering data discovery (e.g., a Documentinstance is related to a Person instance through the author property link).

• A datatype-property, which provides a value defined by a certain XMLSchema Datatype (XSD)6. For example, the same Document instance asbefore, will have a “14 ” literal value for the numberOfPages property.

5http://www.w3.org/TR/owl2-primer/6http://www.w3.org/TR/xmlschema11-2/

6

Page 7: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

Due to the unsupervised, automatic generation of most LOD datasets, manyerrors are made due to datatype assignment. Querying the english chapter ofDBpedia’s SPARQL endpoint (a principal actor in the LOD community) as ofnovember 2015, just 2546 out of the 60247 properties used (4.23%) have a valuefor rdfs:range defined, resulting in only 15.25% of the total amount of valueaxioms declaring its datatype explicitly, leaving the remaining values typed asplain, literal strings (Zembowicz et al. 2010). OOPS! OntOlogy Pitfall Scanner!(Poveda-Villalon et al. 2014) lets any ontology developer validate her ontologydescriptions, avoiding common pitfalls made when designing OWL vocabularies.To the best of our knowledge, no application exists with a similar approach tovalidate datatype values themselves.

Due to these innacuracies, automatic visualization based only on propertyranges is not feasible for producing satisfactory results, as not knowing thenature of the features reduces the understanding of the instances we focus ourinterests on and, more importantly, John Snow would not realise how thosevalues could be visualised. Our proposal is to use a mixed solution, usingboth the defined range (if any) and the result from a custom datatype inferencealgorithm for each property. The inference stage will try to map each value to thedatatype categories where they best fit, profiling the manner each value shouldbe interpreted, how is implemented, encoded and stored in the repository, theset of available operations, its meaning and the value ranges for each observation(Parnas et al. 1976).

For our classification purposes, the following primitive datatypes were se-lected. Please note that these categories are intended to be conceptual andprogramming-language agnostic:

• Integer: Composed by the finite computer representable subset of wholenumbers, such as the height of a person in centimetres, or the number ofwheels in a given vehicle. Negative values are permitted.

• Float: The representation of any real number, as the height of a personin metres. Negative values are permitted.

• Boolean: A value meaning a logical truth, such as “true/false”, “0/1 ”,“yes/no” pairs of values.

• IRI: Internationalised Resource Identifiers are a standard defined uponthe URI scheme, formed by any Unicode character sequence which uniquelyidentifies any resource over the Internet. IRIs are especially relevant inthe SW, as they constitute one of its core components, making resourcesto be linkable and discoverable between them. IRI values are the onlyrepresentatives of object-datatypes.

• String: Defined as any sequence of characters, they are understood asa superset covering the rest. In fact, many values in LD are typed asplain strings (xsd:string, rdfs:Literal, etc.), without any more concrete“ˆˆxsd:datatype” defined for the object’s values or the property’s rdfs:range.

7

Page 8: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

• Datetime component: A part of either a date, a time or both, expressedin any standardised format (preferably following ISO 8601’s directives).

• Geographical component: Any geographical dimension which couldhelp locating a resource in space, e.g., a pair of latitude-longitude coordi-nates and its projection system, a geographical feature or point, etc.

• Categorical data: Marking a property as categorical means that therange of used values is limited or within a certain range, which enablesnew visualizations to represent the property, e.g., histograms which displaythe instance count per occurrence, value distribution or usage.

Figure 1: The inferred primitive datatypes are highlighted for each property.From left to right, they are: Integer, Float, Boolean, IRI, String, Datetimecomponent, Geographical component and Categorical.

Figure 1 depicts how John’s Snow interface displays the inferred primitivedatatypes for each ontological property. Datetime and geographical componentsare non-exclusive, e.g., by analysing geo:lat values, the property can be classifiedboth as Float and Geographical component. Moreover, any property can beclassified as Categorical, should their distinct values be comprised between asmall range.

The primitive datatype inference algorithm was presented in (Pena et al.2015). Although there is still room for improvement, the good results obtainedexhibit a promising approach to classify resources according to their types. Wehave verified that most of our misses in datatype classification were caused bythe algorithm being stricter than the agreement from the expert panel. Forexample, given a list of years, researchers typed the property’s range to be In-teger, whereas the algorithm also assigned the Datetime component datatype.Another interesting example were partial URIs, which uniquely identified a re-source starting with the “/” character. Whilst some experts classified the valuesas IRI instances, the inference resulted in just assigning the String datatype.

8

Page 9: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

4.2 Focusing on relevant conceptsWith dense and high-detail datasets users could suffer from Information Over-load (Eppler & Mengis 2004), losing the ability to center their attention onthe most important aspects of the data. As stated by (Carvalho et al. 2014),“Metadata provides the means to discover datasets, access them and understandthem”. By definition, metadata exposes an abstraction on the dataset itself,thus different tools have relied on metadata extraction in order to comply withInfoVis Mantra’s first visualization task: Overview (Shneiderman 1996). Bysummarising the dataset’s structure, an analyst can get a broad picture of itscontents with little effort.

Different initiatives such as RDFStats (Langegger & Woss 2009), LOD-Stats (Demter et al. 2012), VoID (Hausenblas & Cyganiak 2009, Makela 2014),rdf:SynopsViz (Bikakis et al. 2014) and so on propose different metrics in orderto describe dataset dimensions. Although each project has its own set of prop-erties and focuses on particular statistics, they agree on some common metrics,such as: total number of classes and properties, number of axioms, count ofunique instances, licensing, in\out-degrees (number of incoming and outgoinglinks), and so forth. These indicators are usually listed in tabular format, leav-ing their interpretation to analysts. For example, the LDVM includes thesemetrics in a visual representation, aggregating class and property hierarchiesand plotting them as a treemap, allowing users to explore the components ofeach parent entity in a naıve fashion.

In order to complement existing works, we propose to extract further met-rics which will be used by our visualization prototype to provide insights on thedata’s structure and knowledge value. With greater number of axioms, unexpe-rienced users might get lost in the first stages of the data exploration process.By extracting parameters about the dataset’s conformation, John Snow can re-arrange the rendering of properties and classes to prioritise the most used ones,or enrich some charts with pre-computed basic statistical indicators.

4.2.1 Property usage

Properties are used to define which features could be used within a vocabulary,allowing to relate individuals and values through ?subject→ ?property→ ?objectsemantic axioms. For any selected ontological class, its instances can be definedby a long list of features, not all of which are as suitable as the others tosummarise the instances. The usage of each property is therefore assessed as:

pu = cs

ci

where:

pu: property usage ratio

cs: number of unique class individuals which make use of the property.

9

Page 10: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

ci: number of instance objects for that class

For example, in an example dataset with descriptions for 300 foaf:Personindividuals, if 225 of them provide a value for the foaf:depiction property, itscomputed property usage will have a value of .75 ( 225

300 ), in other words, 3 out of4 foaf:Person instances offer a related photograph. Property usage values arespanned in the (0, 1] range.

• If pu = 0, the property has never been used to describe any resource of theselected class, although it is defined within the ontology. In contrast withother tools, John Snow does not list those properties to avoid visual clut-tering. This is specially useful for big datasets (e.g., DBpedia), where dueto inheritance issues some resources have hundreds of properties related,but only a few dozens are assigned a value.

• The closer the pu value is to 1, the more the property is used. Whenpu = 1, all the instances have provided at least one value for the property.Ordering properties according to pu value allows to distinguish the mostreliable properties to characterise a resource from those which are not sorepresentative.

4.2.2 Completeness ratio

The property usage just exhibits the preferred properties to depict resourcesbelonging to a certain class, but does not reflect the intended purpose of datapublishers when assigning values to the property. The completeness ratio, for-mulated as:

cr = pi

ci

where:

cr: completeness ratio

pi: total amount of values related through the selected property

tries to reflect this dimension. cr calculates how many values are assigned toeach unique subject, and depending on its value, three scenarios can be observed:

• 0 < cr < 1: not all class individuals have this property defined, concludingthat it is an optional argument when defining the resource.

• cr = 1: each individual from the class defines a single value for the prop-erty. An example is the foaf:name attribute for each foaf:Person instance.Each instance must provide one (and only one) valid value for this prop-erty.

10

Page 11: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

• cr > 1: some class individuals are given more than one value for the se-lected property. For example, the same foaf:Person could take differentfoaf:nick values if any individual takes different aliases through the infor-mation management system. A histogram renders the number of instancesper unique subject on a modal panel, as shown in Figure 2.

Figure 2: The completeness ratio histogram exhibits the instance count groupedby the number of values adopted for each property. In the image, 17 foaf:Personinstances provide exactly 2 values for the foaf:nick property.

4.2.3 Extra metrics for certain primitive datatypes

Finally, once primitive datatypes (see section 4.1) are inferred for each property,extra metrics are computed based on their classification using built-in functionsof SPARQL v1.1. Some examples are listed below:

• For numerical datatypes, statistical indicators such as maximum, mini-mum, mean and average values are extracted, as well as quartile values ifpossible. Showing these values together with the generated visualizationscan minimise the time required to find an answer when exploring the data(Figure 3).

• Regarding properties containing geographical features, their bounding boxesare calculated. A bounding box is the smallest area within which all thepoints lie, attracting the user’s focus. The most basic rectangular bound-ing box is formed by the minimum and maximum values for both latitudeand longitude properties.

• When a temporal dimension is encoded, the start/end marks for eachevent could be plotted in a timeline, allowing to see instance overlappings

11

Page 12: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

and gaps. If some events coincide in the same date, a calendar heatmappermits pattern detection.

Figure 3: The depicted property in John Snow featured the maximum allowedguests for a list of restaurants. The table summarised the values adopted by allthe instances.

These extended features guide users during their analyses, as they hint thereal status of the dataset’s contents. Without the need to start digging in thedata itself (yet), users can detect inconsistencies (e.g., off-limits data entrieswhen focusing on a specific geographical region, or a highly-dispersed set ofmeasurements for a numeric attribute) or missing-value issues (i.e., low com-pleteness scores for most of the defined properties) early on the analysis stage.Presenting summarised values in an understandable, non-invasive manner couldlead to improve the dataset exploration experience.

4.3 Exploratory analysis goalTogether with the previous approaches to infer the dataset’s structure and com-position, John Snow should recommend the most suitable visual representa-tions taking into account the expected outcomes from the end users perspective.When dealing with a new set of data, despite the analytic skills and previous ex-periences in similar scenarios, users intrinsically have an analytic goal in mind:whether they want to study the trends of a particular set of values throughtime, study the distribution of values for a particular property, the correlationsbetween a set of data dimensions, and so on. Based on the message the dataanalysts want to communicate, we can classify the analysis intention under thefollowing categories (Borner 2015):

• Categorization: Cluster data in groups with similar meaning or features.

• Comparison: Contrast two or more equivalent items to observe theirdifferences.

12

Page 13: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

• Composition: Aggregation of data which are part of a whole, in orderto represent the full description of the item.

• Distribution: Reflection about data dispersion, in order to analyse po-tential patterns.

• Geospatial location: Rendering of resources with a geographical dimen-sion on a map.

• Ordering: Also referred as sorting, consists on arranging the items in aparticular manner, attending to a certain feature of interest to the user.

• Relationship: Sometimes individual items are connected between themby common characteristics.

• Temporal trends: Finally, resources with a temporal dimension canform patterns by focusing on the evolution of their value series throughtime.

For the first version of our prototype (see Figure 4), only the most knowncharts were encoded, e.g., pie charts, column charts (to render histograms),maps, force directed graphs and boxplots. These charts are easily understand-able by most users (despite the latter maybe, which is why the prototype offersalternative indicators), as they allow to encode all the primitive datatypes in-ferred from Section 4.1, and cover all the analysis messages exposed before (seeTable 1).

Figure 4: Through all the metadata extracted in previous steps (A), John Snowpresents different visualizations suitable to represent the data (B). Selecting anyavailable chart, will render the corresponding values using the most appropriatevisual abstractions (C).

13

Page 14: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

Table 1: Common charts suitability based on the message intention: Catego-rization, Comparison, Composition, Distribution, Geospatial location, Order-ing, Relationship and Temporal trends.

Chart Cat. Compa. Compo. Dist. Geo. Ord. Rel. Temp.Line * * * *Bar/Column * * * * *Bullet bar * *Stacked bar * * * *Area * * *Pie * *Scatter plot * * * *Bubble * * * *Waterfall * *Map *Table * * *

5 Evaluation and results discussionIn order to check the validity of our visualization prototype, we gathered a groupof 14 volunteers and conducted individual goal-directed evaluations with them,using John Snow as a visual exploratory analysis tool. Most participants had astrong technical background (different engineering degrees), but just a few knewhad heard about the Semantic Web. Each session lasted around one hour, beingrecorded for further study.

To set the experiment, open-license datasets were selected from a variety ofOpen Data portals. To be eligible for selection, data must be provided boththrough an available SPARQL endpoint and in traditional structured formats.Eash selected dataset covers a different topic, out of the specialization area of theparticipants, thus avoiding preconceptions about the data. As confirmed dur-ing interviews, all participants declared to be unaware of the selected datasetsbeforehand. The datasets presented to participants were:

• Restaurants7

The first dataset held information about the restaurants within the cityof Zaragoza, which maintains on of the most enowned Open Data Portalsin Spain.

• Animal intake service8

The second dataset dealed with the Animal Intake Service deparment fromthe City of Los Angeles, USA. Data.gov is a world reference in Open Data,and provides nearly 200K highly curated datasets.

7https://www.zaragoza.es/ciudad/risp/detalle_Risp?id=2858http://catalog.data.gov/dataset/animal-services-intake-data-57f73

14

Page 15: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

• Student enrollments9

The third dataset included information about new enrollments for thePablo de Olavide University in Seville, south of Spain, during the 2012-2013 academic year.

• Street crime10

Finally, the last dataset contained information about street crimes com-mitted in Camden’s area, London (this information was hidden from par-ticipants, as it was one of the questions that needed answering).

After the setup, a few questions about the expertise level of each partici-pant in different areas was collected, to evaluate their skills using a Likert scale.Results are displayed in Figure 5. Readers can realise the participants’ lit-tle experience with Semantic Web technologies, which contrasts with the solidbackground with both spreadsheets software and visualization methodologies.Accordingly, we favoured the usage of CSV as the traditional approach towardsexploratory analysis, which would be compared to John Snow. The results alsoexhibited the familiarity with structured data, so they quickly understood theexpression of LOD as RDF triples. Regarding data visualization, even thoughparticipants expressed a better than average knowledge, they were surprised bysome chart rendering decisions on John Snow. The most repeated one was thepredominancy of bar/column charts over pie charts. After explaining that thepie chart generation heuristic took into account if the expressed values added-up to a whole, the categorical classification of values, etc. most recognised thesuitability of the resulting solution to understand the data.

● ●Semantic Web

Visualization

Spreadsheets

2 4 6 8 101 2 3 4 5 6 7 8 9 10

Figure 5: Skills level for the people involved in the evaluation, from none /very low (1) to very high / expert (10). The questions asked for the knowledgeof spreadsheets software, the expertise with data visualization tools, and theexperience with Semantic Web / Linked Data technologies.

Later, the John Snow tool was presented by exploring a semantically an-notated data dump from our research group’s internal database. By taking a

9https://datos.upo.gob.es/dataset/?id=estudiantes-nuevo-grado10https://data.gov.uk/dataset/on-street-crime-in-camden

15

Page 16: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

guided exploratory analysis approach, we could present the interface layout andhow could users explore the data navigating through their classes and proper-ties. After introducing the visualization tool, we trained participants by asking afew questions which highlighted the benefits of LOD to answer complex queries,and showed its limitations. During this stage, each participant could ask theinterviewer for hints to reach the correct answers.

Finally, participants were asked to answer some questions about the selecteddatasets, both using traditional software tools and John Snow. The set ofquestions was designed based on previous experiences in hackathon events andthe like, when we needed to deal with new data sources for the first time. Oneof the first tasks when starting a new data analysis is to figure out the structureof the dataset, how resources are organised and described, the relations betweeninstances (if any), and what values are expected for each feature. The order ofwhich tool was used first was determined by each participant, to remove anypossible biases derived from the second approach being easier than the initialone. The goal of the exploratory task was not to find the exact answers to eachquestion, but to understand what paths did a real user follow towards them,and which were the most common methodologies. We must declare that allparticipants reached the solutions both by means of John Snow and downloadedCSVs.

Once the goal-directed evaluation was finished, each participant filled in abrief survey about John Snow. These questions were designed to validate theproposed approaches, and identify future lines of work. Each item was scoredfrom 1 (very low) to 10 (very much), and the results are depicted in Figure 6.

a) How suitable did you find the datatype categories (i.e., integer, float,boolean, IRI, string, datetime component, geographical component andcategorical) in order to classify property values?

b) How would you assess chart availability in order to answer the ques-tions?

c) How useful did you find the primitive datatype inference section for eachproperty, which highlighted the identified datatypes for each of them?

d) Some metadata values were rendered visually (see Section 4.2) in JohnSnow’s interface. Did you rely on them during your exploratory analisis?

e) For some primitive datatypes, specific extended metrics were displayed(consult Section 4.2.3). Did these indicators help you reach some answersmore quickly, or in a more intuitive way?

The overall impression is that the prototype improved how users faced newdata, as the answers obtained using John Snow were reached quicker and moreintuitively than using spreadsheets software, with no training nor previous ex-perience with the tool. By inferencing the primitive datatype, the users got aquick summary of how data values could be encoded. Even though this featurereceived some of the lowest scores (users did not focus much on the inference

16

Page 17: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

e)

d)

c)

b)

a)

2 4 6 8 101 2 3 4 5 6 7 8 9 10

Figure 6: Users evaluation of John Snow’s features: a) datatype suitability,b) specific metrics, c) metadata charts, d) primitive datatype inference and e)chart availability.

rendering during data exploration, as stated when asked about this particular-ity), the datatype inference stage is a fundamental part of our LOD visualizationpipeline, as this information is required for chart generation heuristics. If fur-ther information was required, they clicked on the “sample values” button, whichlisted random values found in the dataset for the property, so they could confirmtheir expectations. It is also noteworthy that without the datatype inferencestep, no visualizations could be generated, thus missing the opportunities toachieve the goals determined for this research.

By automatically mapping data values to graphical representations, userscould explore the datasets in an interactive fashion, which was addressed bymany participants non familiarised with LOD to be a really intuitive and efficientapproach. With a very low technical profile required (just to know how to use aweb browser and a little explanation about how John Snow’s interface worked),users could start navigating through classes and properties from the very firstmoment. When they reached an exploratory dead-end, they could just go backto the previous view and try a different path.

6 Conclusions and future workBy focusing our attention on metadata and the data contained within anydataset, we have designed a generic LOD visualization pipeline which allows anyuser to explore the Web of Data. Through an initial analysis on the structureand relationships between the resources of any data source, and visualization

17

Page 18: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

generator heuristics which implement some ideas from the Exploratory DataAnalysis Field, the presented visualization tool John Snow is able to presentdifferent charts and graphs adapted to the semantically annotated data takenas input.

One of the main advantages of LOD compared to traditional structured for-mats (CSV, JSON, relational database dumps, etc.), is that the schemas usedfor description are publicly available over the Internet, allowing different actorsto share standardised vocabularies to describe their resources. The techniquespresented in this work are not valid only for each exploratory analisis session.All the extracted metrics are stored, and when the same vocabularies or simi-lar data structures are found in later analyses, previously obtained knowledgeis consulted in order to ease new data discovery. Thanks to this approach,our prototype is able to represent data from different domains, with no extrainformation required but that already encoded within the dataset.

The participants which took part in the evaluation stage praised the inter-action with the tool, which allowed an intuitive data exploratory analysis withno previous background required. Their comments during the experimentationsresulted in the improvement on some features of the tool, for which we arethankful. When participants became aware of the lack of customization of theresulting visualizations, some declared they preferred a more autonomous visu-alization tool at the expense of chart tailoring, as those charts are not intendedfor public exhibition (in a report, a research article, etc.), but to understandthe inner structure and data-relationships in a quick fashion.

In order to support new graphs, the implementers must define the heuristicto draw the new chart, as well as the conditions the data needs to satisfy tothat end and how are the data dimensions encoded on the depiction. Besidesimplementing new charts, our intention is to improve the datatype inferencealgorithm, to cover most of the valid values present in the LOD Cloud. We willalso publish all the extracted knowledge for the analysed datasets, so that futureresearchers can reuse the identified patterns in their exploratory analyses andavoid common pitfalls and waste of resources. This way, we hope to contributeto LOD discovery by easing the initial steps.

After nearly fifteen years since the principles of the Semantic Web, we needto make lay users aware of the advantages of the future’s Web. We can not holdusers accountable for the direct manipulation of semantic technologies. Boththe SW and LD communities should provide “smarter” tools which can displayinformation in the most convenient formats, using approaches which do not needconstant revision in order to work with new vocabularies or domains.

18

Page 19: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

ReferencesBerners-Lee, T. (2006), ‘Linked Data - Design Issues’.

URL: http://www.w3.org/DesignIssues/LinkedData.html

Berners-Lee, T., Hendler, J. & Lassila, O. (2001), ‘The semantic web’, Scientificamerican 284(5), 28–37.

Bikakis, N., Skourla, M. & Papastefanatos, G. (2014), rdf:SynopsViz – A Frame-work for Hierarchical Linked Data Visual Exploration and Analysis, in ‘TheSemantic Web: ESWC 2014 Satellite Events’, number 8798 in ‘Lecture Notesin Computer Science’, Springer International Publishing, pp. 292–297. DOI:10.1007/978-3-319-11955-7 37.

Bizer, C., Heath, T. & Berners-Lee, T. (2009), ‘Linked Data - The Story So Far:’,International Journal on Semantic Web and Information Systems 5(3), 1–22.

Brickley, D. & Miller, L. (2012), ‘FOAF vocabulary specification 0.98’, Names-pace document 9.

Brunetti, J. M., Auer, S., Garcıa, R., Klımek, J. & Necasky, M. (2013), FormalLinked Data Visualization Model, in ‘Proceedings of International Conferenceon Information Integration and Web-based Applications & Services’, IIWAS’13, ACM, New York, NY, USA, pp. 309:309–309:318.

Borner, K. (2015), Atlas of Knowledge: Anyone Can Map, The MIT Press,Cambridge, Massachusetts.

Card, S. K., Mackinlay, J. D. & Shneiderman, B. (1999), Readings in informa-tion visualization: using vision to think, Morgan Kaufmann.

Carvalho, P., Hitzelberger, P., Otjacques, B., Bouali, F. & Gilles, V. (2014),Open Data Integration - Visualization as an Asset, pp. 41–47.

Chi, E. H. (2000), A taxonomy of visualization techniques using the data statereference model, in ‘IEEE Symposium on Information Visualization, 2000.InfoVis 2000’, pp. 69–75.

Cyganiak, R. & Bizer, C. (2008), ‘Pubby-a linked data frontend for sparql end-points’, Retrieved from http://www4.wiwiss.fu-berlin.de/pubby 28, 2011.

Dadzie, A.-S. & Rowe, M. (2011), ‘Approaches to Visualising Linked Data: ASurvey’, Semant. web 2(2), 89–124.

Deligiannidis, L., Kochut, K. J. & Sheth, A. P. (2007), RDF Data Explorationand Visualization, in ‘Proceedings of the ACM First Workshop on CyberIn-frastructure: Information Management in eScience’, CIMS ’07, ACM, NewYork, NY, USA, pp. 39–46.

19

Page 20: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

Demter, J., Auer, S., Martin, M. & Lehmann, J. (2012), LODStats—An Exten-sible Framework for High-performance Dataset Analytics, in ‘Proceedings ofthe EKAW 2012’, Lecture Notes in Computer Science (LNCS) 7603, Springer.

Elmqvist, N. (2014), Visualization reloaded: redefining the scientific agenda forvisualization research, in ‘Proceedings of HCI Korea’, Hanbit Media, Inc.,pp. 132–137.

Eppler, M. J. & Mengis, J. (2004), ‘The Concept of Information Overload: AReview of Literature from Organization Science, Accounting, Marketing, MIS,and Related Disciplines’, The Information Society 20(5), 325–344.

Fionda, V., Gutierrez, C. & Pirro, G. (2014), ‘The swget portal: Navigatingand acting on the web of linked data’, Web Semantics: Science, Services andAgents on the World Wide Web 26, 29–35.

Hausenblas, M. & Cyganiak, R. (2009), ‘Describing Linked Datasets - On theDesign and Usage of voiD, the ’Vocabulary of Interlinked Datasets”.

Hoefler, P., Granitzer, M., Sabol, V. & Lindstaedt, S. (2013), Linked Data QueryWizard: A Tabular Interface for the Semantic Web, in ‘The Semantic Web:ESWC 2013 Satellite Events’, number 7955 in ‘Lecture Notes in ComputerScience’, Springer Berlin Heidelberg, pp. 173–177. DOI: 10.1007/978-3-642-41242-4 19.

Klımek, J., Helmich, J. & Necasky, M. (2014), Application of the Linked DataVisualization Model on Real World Data from the Czech LOD Cloud, in ‘6thInternational Workshop on the Linked Data on the Web (LDOW’14)’.

Langegger, A. & Woss, W. (2009), RDFStats - An Extensible RDF StatisticsGenerator and Library, in ‘20th International Workshop on Database andExpert Systems Application, 2009. DEXA ’09’, pp. 79–83.

Leon, A. d., Wisniewki, F., Villazon-Terrazas, B. & Corcho, O. (2012), Map4rdf- Faceted Browser for Geospatial Datasets, in ‘PMOD workshop’, Facultadde Informatica (UPM).

Mechant, P., Van Compernolle, M., Dimou, A., De Vocht, L. & De Marez,L. (2014), Linked open data and research knowledge management: an ex-ploratory search and visualization framework, in ‘6th International Confer-ence on Education and New Learning Technologies, Proceedings’, IATEDAcademy.

Makela, E. (2014), Aether – Generating and Viewing Extended VoID StatisticalDescriptions of RDF Datasets, in ‘The Semantic Web: ESWC 2014 Satel-lite Events’, number 8798 in ‘Lecture Notes in Computer Science’, SpringerInternational Publishing, pp. 429–433. DOI: 10.1007/978-3-319-11955-7 61.

20

Page 21: Exploring LOD through metadata extraction and data-driven ...pdfs.semanticscholar.org/fc9b/4a999a773b5e1b1b0b2... · For the last decade, the Semantic Web and Linked Data communities

Parnas, D. L., Shore, J. E. & Weiss, D. (1976), Abstract Types Defined AsClasses of Variables, in ‘Proceedings of the 1976 Conference on Data : Ab-straction, Definition and Structure’, ACM, New York, NY, USA, pp. 149–154.

Pena, O., Aguilera, U. & Lopez-de Ipina, D. (2015), Resource Classificationas the Basis for a Visualization Pipeline in LOD Scenarios, in ‘Metadataand Semantics Research’, number 544 in ‘Communications in Computer andInformation Science’, Springer International Publishing, pp. 457–460. DOI:10.1007/978-3-319-24129-6 40.

Poveda-Villalon, M., Gomez-Perez, A. & Suarez-Figueroa, M. C. (2014), ‘OOPS!(OntOlogy Pitfall Scanner!):: An On-line Tool for Ontology Evaluation’, In-ternational Journal on Semantic Web and Information Systems 10(2), 7–34.

Shadbolt, N., Hall, W. & Berners-Lee, T. (2006), ‘The Semantic Web Revisited’,IEEE Intelligent Systems 21(3), 96–101.

Shneiderman, B. (1996), The eyes have it: a task by data type taxonomy forinformation visualizations, in ‘, IEEE Symposium on Visual Languages, 1996.Proceedings’, pp. 336–343.

Tukey, J. W. (1977), Exploratory Data Analysis, 1 edition edn, Pearson, Read-ing, Mass.

Ware, C. (2012), Information Visualization, Third Edition: Perception for De-sign, 3 edition edn, Morgan Kaufmann, Waltham, MA.

Yu, L. (2011), Follow Your Nose: A Basic Semantic Web Agent, in ‘A Devel-oper’s Guide to the Semantic Web’, Springer Berlin Heidelberg, pp. 533–557.DOI: 10.1007/978-3-642-15970-1 14.

Zembowicz, F., Opolon, D. & Miles, S. (2010), OpenChart: Charting Quantita-tive Properties in LOD., in ‘LDOW’, Vol. 628 of CEUR Workshop Proceedings,CEUR-WS.org.

21