interoperability - is it feasible -
DESCRIPTION
Interoperability - is it feasible -. Peter Wittenburg. Why care about interoperability?. e-Science & e-Humanities “data is the currency of modern research” thus need to get integrated access to many data sets data sets are scattered across many repositories => (virtual) integration - PowerPoint PPT PresentationTRANSCRIPT
The Language Archive – Max Planck Institute for Psycholinguistics
Nijmegen, The Netherlands
Interoperability- is it feasible -
Peter Wittenburg
Why care about interoperability?
• e-Science & e-Humanities• “data is the currency of modern research”• thus need to get integrated access to many data sets • data sets are
• scattered across many repositories => (virtual) integration
• created by different research teams using different conventions (formats, semantics)• often in bad states and quality => curation
• thus interoperability most used word at ICRI conference
• Big Questions: • What is meant with interoperability?• How to remove interoperability barriers to analyze
large heterogeneous and probably distributed data sets?
• Is interoperability something we need/want to achieve?
What is interoperability?
1. Wikipedia: Interoperability is a property of a system, whose interfaces are completely understood, to work with other systems, present or future, without any restricted access or implementation.
2. IEEE: Interoperability is the ability of two or more systems or components to exchange information and to use the information that has been exchanged.
3. O’Brian/Marakas: Being able to accomplish end-user applications using different types of computer system, operating systems, and application software, interconnected by different types of local and wide-area networks.
4. OSLC: To be interoperable one should actively be engaged in the ongoing process of ensuring that the systems, procedures and culture of an organization are managed in such a way as to maximise opportunities for exchange and re-use of information.
What is interoperability?
• Technical Interoperability (techn. encoding, format, structure, API, protocol)
• Semantic Interoperability • is it also about bridging
understanding between two or more humans?
<hund>
<köter>
<dog>
humans – humans we better speak about understandinghumans – machine same or?machine – machine well here interoperability makes sense
What is interoperability?
• seems that every one speaks about technical systems when talking about interoperability
• do we include feeding machines with some mapping rules specified by human users and then carrying out some automatic functions? • when linguists hear about mapping tag sets some
immediately say that it is impossible and does not make sense
• why: tags are part of a whole theory behind it • well if you look to other disciplines (life sciences, earth
observation sciences etc.) that’s exactly what they do • why
• people want to work across collections and ignore theories
• some see tag sets just as first help but want to work on raw data
• some see the demand of politicians and society to come up with answers and not with statements about problems
• AND there is much money (is it useless?)
Big Data in Natural Science
• numbers in regular structures • how to find relevant data sets
• volcanology/earthquakes/Tsunamies/etc. • X sensor datastreams (seismology)
(time, location, parameters)
• X human observations (biodiversity)(time, location, nr.
frogs (etc))
• window extraction to transfer and manage data
• interpret regular structures (even frogs)
• time normalization, take care of dynamics etc.
• visualize things coherently
Big Data in Natural Science
• numbers in regular structures • how to find relevant data sets
• volcanology/earthquakes/Tsunamies/etc. • X sensor datastreams (seismology)
(time, location, parameters)
• X human observations (biodiversity)(time, location, nr.
frogs (etc))
• window extraction to transfer and manage data
• interpret regular structures (even frogs)
• time normalization, take care of dynamics etc.
• visualize things coherently
interoperability looks simple enough
just find patterns in sequences of numbers the
format you need to know
(well – not quite as simple, but ...)
Big Data in Environmental Sciences
• many different types of observations• climate, weather, etc.• species and populations according to multitude of
classification systems and schools
• grand challenge• how can all these observations be used to stabilize
our environment• how can it all be used to maintain diversity• etc.
Big Data in Environmental Sciences
• many different types of observations• climate, weather, etc.• species and populations according to multitude of
classification systems and schools
• grand challenge• how can all these observations be used to stabilize
our environment• how can it all be used to maintain diversity• etc.
sounds similar to our field
interoperability is tough
but there are expected gains
and there is more money
intensive work also in social science
many layers of interop: access
Scientists, Data Curators,End Users, Applications
EnablingTechnologies
Discovery
Access(ref. resolution, protocols, AAI)
Interpretation
ReuseAccessed via Repositories
01000101..
ID
ID
ID
ID
ID
ID
ID
ID
ID
Datasets
01000101..
ID
ID
ID
ID
01000101..
ID
01000101..
ID
01000101..
IDID
ID
metadata search resulting in Handles (PID)
and some properties
Handle (PID) resolution
and you get the data
here linguistics is playing a role
(get schemas and semantics)
here linguistics is playing an even bigger role
(get context information)
need a high degree of automation
what can be automated
many layers of interop: management(mostly underestimated!!!)
Data ManagersData Scientists
EnablingTechnologies
Collections +Properties
Access(ref. resolution, protocols, AAI)
formalized policiesworkflow engine
AssessmentAccessed via Repositories
01000101..
ID
ID
ID
ID
ID
ID
ID
ID
ID
Datasets
01000101..
ID
ID
ID
ID
01000101..
ID
01000101..
ID
01000101..
IDID
ID
metadata gathering resulting in Handles (PID)
and some properties
Handle (PID) resolution
and you get the data
check of rules and engine
it’s all about establishing trust
formal rules manipulate properties of
Handles and metadata and may generate new DOs
need a high degree of automation
need a high degreeof automation
simple but essential example: PIDs
• it’s similar to TCP/IP with all its core machinery that brought us the Internet and thus interoperability with respect to communication
• email system works when we abstract from content and thus the semantics of our human messages and focus on the semantics of attributes, parameters etc.
• let’s assume that you want to use a certain file and first want to be sure that the file has not been modified
• you look up in metadata• that automatically looks for the PID• the PID is resolved automatically and a checksum is
retrieved• the checksum is automatically compared with the
checksum of the file accessed• a warning is given automatically if the two don’t match
• this would be a great service (and will come)
email Value AddedServices
NetworkTechnology
InternetProtocol Suite
WWW phone
SMTP HTTP RTP…TCP UDP…
IP
copper fiber radio
CSMA async sonet…
ethernet PPP…
Internet machinery (collaboration CNRI and MPI)
DNS
all applications making use of the same basic protocol where the “packet” is the basic object and where endpoints have addresses and names
Value AddedServices
DataSources
PersistentIdentifiers
PersistentReferenceAnalysis Citation
AppsCustomClients Plug-Ins
Resolution System Typing
PID
Local Storage Cloud Computed
Data Sets RDBMS Files
Digital Objects
Data machinery (collaboration CNRI and MPI)
all applications making use of the same basic protocol where “data” is the basic object and where PID and metadata attributes describe object properties
PID recordattributes
bit sequence(instance)
metadataattributes
points to instances describes properties
describes properties& context
point toeach other
Layers of interoperability
• Protocols/APIs: defined formats, semantics, processes• SCSI: how to read/write/etc. blocks to/from SCSI disc • File System: how to read/write/etc. to/from logical entities
how to organize files on a machine (virtualization)
how to organize files across machines
• OAI/PMH: how to serve metadata descriptions • SRU/CQL: how to do distributed content search • etc. etc.
• all based on standards or widely accepted best practices • advantage: standards establish a 1:N relation constant
over time• large number of standards/BP for metadata (structure,
semantics)
back to linguistics
• where are we in the linguistics domain?• what happened in some well-known projects • do we miss the big challenges which other disciplines
have and that would force us to ignore schools, vainness, etc.
• 4 examples• metadata• DOBES• TDS• CLARIN
metadata is kind of easy
• DC/OLAC – CMDI mapping examples: • DC:language CMDI:languageIn• DC:language CMDI:dominantLanguage• DC:language CMDI:sourceLanguage• DC:language CMDI: targetLanguage• DC:date CMDI:creationDate• DC:date CMDI:publicationDate• DC:date CMDI:startYear• DC:date CMDI:derivationDate • DC:format CMDI:mediaType• DC:format CMDI:mimeType• DC:format CMDI:annotationFormat• DC:format CMDI:characterEncoding
• everyone accepts now: metadata is for pragmatic purposes and not replacing the one and only one true categorization
• mapping errors may influence recall and precision – but who cares really
crucial for machine
processing
semantic mapping doable due to limited element
sets and to now well-described semantics
(except for recursive machines such as TEI)
if mapping is used for discovery – no problem
if mapping is used for statistics – well ...
truth in metadata usage – still !!
Rebecca Koskela: DataONE
DOBES – some facts
DOBES = Documentation of Endangered Languages
some facts•started 2000 with 7 international teams and 1 archive team•2012: now 68 documentation teams working almost every where
• cross-disciplinary approach: linguists, ethnologists, musicologists, biologists, ship builders, etc.
• every year one workshop and two training courses
DOBES Agreements
• in first 2-3 years quite some joint agreements• formats to be stored in the archive – interoperability• principles of archiving such as PIDs • workflows determining the archive-team interaction • organizational principles to manage and manipulate data • metadata to be used to manage and find data (pragmatics vs. theory) • joint agreement on Code of Conduct
• short discussions on more linguistic aspects failed• agreement on joint tag set - NO• agreement on joint lexical structures - NO• etc. • good reason: the languages are so different • “bad” reason: agreements require effort
recent DOBES Questions
• now after >10 years we have so much good data in the archive• what can we do with it ????• traditional: every researcher looks at his/her data and
publishes of course taking into account what has been published by others
• new: can researcher teams come to new results while working on the raw and annotated data?
• what does this require in case of “automatic” or “blind” procedures?
(remember that the researchers do not understand the language)• you need to know the tier labels to understand the type of
annotation• you need to know the tags used to understand the results of
the linguistic analysis work (morphological, syntactical, etc.)
an text example
Example from Kilivila (Trobriand Islands – New Guinea)
p1tr Ambeyap1en Where do you go?
p2tr Bala bakakayap2w-en I will go I will take a bathp2en I will go to have a bath
p2tr Bila bikakaya bike’ita bisisu bipaisewap3gl 3.Fut-go 3.Fut-bath 3.Fut-come back 3.Fut-be 3.Fut-workp2w-en He will go - he will have a bath - he will come back – he will stay -
he will work.p2en He will take a bath and afterwards work with us.
what’s this?
what’s this?
big question: what can we do with searches, statistics – thus semi-automatic procedures across different corpora
Hum. Example: Multi-verb Expressions
mixed glossing
POS tagging
a multimodal example
tier names from Toolbox
Interaction Study: 12 participants + exper; per part. 7 tiers
tier names – an area of creativity
5 cross-corpora projects
• demonstratives with exophoric reference (morpho-syntactic and discourse pragmatic analysis incl. gestures)
• discourse and prosody – convergence in information structure
• relative frequencies of nouns, pronouns and verbs
• cross-linguistic patterns in 3-participant events
• one rather large program with 13 teams covering different languages
• primary topic is “referentiality”• bigger question: how to do this kind of cross-corpus work • strategy: define new tag set and add a manually created
tier • yet no agreed tags – committee has been formed• now in a process to determine selection of corpora • question: will existing tags help to find spots of relevance
in general:
additional tagging based on specific agreements
are existing annotations of any help?
finally everyone works in his/her data
TDS (LOT etc.)
Typology Database System - offering one semantic domain to look for phenomena in 11 different typological databases created independently and covering many languages.
Domain expertDatabase developer TDS Knowledge engineer
Topic taxonomies(SKOS)
Global linguistic ontology(OWL)
Local database ontologies(DTL)
Database schemata(any DDL)
straight mapping
complexmapping
many descriptive parameters
&differences in
structure, terminology
and theoretical
assumptions
TDS (LOT etc.)
Typology Database System - offering one semantic domain to look for phenomena in 11 different typological databases created independently and covering many languages.
Domain expertDatabase developer TDS Knowledge engineer
Topic taxonomies(SKOS)
Global linguistic ontology(OWL)
Local database ontologies(DTL)
Database schemata(any DDL)
straight mapping
complexmapping
many descriptive parameters
&differences in
structure, terminology
and theoretical
assumptions
thus an ontology based approach to interoperability
instead of an attempt to redo type specification
(WALS)
good: get typology specs out of individual boxes
thus: TDS was also curation work
subject-verb agreement
• Q1: which languages have subject-verb agreement?• db A: exactly this question with Boolean answer
• no distinction thus simple • db B: bundle of information
• sole argument of an intransitive verb• agent/patient/recipient-like arguments of transitive
verb• in general “yes” for s and a cases (but not always
clear) • Q2: which languages are of type a for transitive verbs
• db A: ambiguous – so give all languages or none • db B: simple answer
• a pre-query stage allows user to decide about options
• what when several parameters are used to describe a phenomenon
Did TDS work?
• let’s assume that• the local ontologies represent the conceptualization
correctly• the global ontology forms a useful unifying
conceptualization (is there such an accepted unifying ontology?)• the 2-stage query interface offers proper help• THEN TDS sounds like an excellent, scalable approach
• why did TDS not yet take up?• TRs rely on papers and are not interested in databases ?• TRs don’t understand and rely on the formal semantics
blurb ?• TRs would need to invest time – do they take it ? (occasional usage, small community of experts)
• what is WALS then – just a glossary for non-experts ?
What happens in CLARIN?
• well Metadata is obvious –> Virtual Language Observatory • harvesting and mapping is not the problem• bad quality is the problem (as for Europeana
etc.)
• planned is f.e. distributed content search
SRU/CQL
what is comparatively easy?
• what if we only look at Dutch or German texts? • searching just for textual patterns (collocations)• could make use of SUMO, Wordnets to extend
query etc• but can/should we compete with Google?
• what if we search across languages?• well – need some translation mechanism for
textual patterns – could be trivial translation• does it make sense – will people use it?• AND: it is mainstream – so Google will do it
this is what currently is being worked on
not so inspiring
seems that researchers are not really interested in this
at this moment
what is more difficult and special?
• assume some annotated texts, audios, videos • assume some standard type of linguistic annotations such as
morphosyntax, POS, etc.
1. Select corpora 2. Select Tag sets 3. Formulate query 4. Expand by rules (relations between tags)
2
43
1
this was rejected across country borders
but is it this what we need?
is there a potential or just a myth?
semantic bridges: how?
• assume that we have two corpora: one encoded by STTS and the other one by CGN and assume that they have some linguistic annotation (morphosyntax, POS, etc.) to be used in a distributed search or statistics
(take care: searching != statistics)• what to do now to exploit both collections?
1. do separate searches – well ...2. create rich umbrella ontology and complex refs (comparable to TDS)
• well - could become a never ending story ...• people disagree on relations etc.• relations partly depend on pragmatic considerations • expensive, static, require experts, not
understandable, etc.
Are flat category registries ok?
3. flat registries of linguistic categories such as ISOcat (12620) sound like a solution for some tasks
• easy mapping between two (or more) categories • users can easily create their own mappings or re-use some • maintenance is more easy and thus allows dynamics • etc.• so it seems that we could overcome the TDS barrier
•but we are reducing accuracy and losing much information•too simple for statistics ??•sufficient for searches ??
What about Jan’s examples?
• e0: annotations are structured: “np\s/np”• e1: “JJR” -> “POS=adjective & degree=comparative”• e2: “Transitive” -> “thetavp=vp120 &
synvps=[synNP] & caseAssigner=True”• e3: “VVIMP” -> “POS= verb & main verb &
mood=imperative”
• where to put annotation complexity if “ontology” is simple
• complexity needs to be put into schemas • who can do it – is it feasible?
• mapping must be between combinations of cats or graphs• who can do it – is it feasible?
are there conclusions?
1. do we want/need cross-corpora operations?• for many other communities this is a MUST • don’t we have “society relevant” challenges?• do they just get more money?• given all regularity finding machines – is linguistic
annotation relevant at all?
2. is it for us more difficult to do?• well - that’s what all claim – don’t believe that anymore
are there conclusions?
3. are we interested to try it out?• well – yet there are not so many people committed
• is it not of relevance?• is it lack of money?
• some are opposing strictly • is it a sense of reality?• is it lack of vision?• is it vainness?
4. if interested, how do we want to tackle things?• pragmatic – stepwise – simple first • will people use it then? • do we have evangelists?
Bedankt voor het luisteren.
useless Cloud debate
some just call for Cloud – what does it solve?
just collect also all content into
one big pot
all the issues about interoperability remain the samesearching will be more efficient – no transport etc.
What about metadata?
• TEI example 1resp annotation supervisor and developer
date from="1997" to="2004" name Claudia Kunze
• which date is it? need to interpret context• which role is it? need to interpret context
• TEI example 2name Dan Tufiş
resp Overal editorshipname Ştefan Bruda
resp Error correction and CES1 conformance• which role is it? need to interpret context
• very simple examples show• meant to be read by humans • (too) much degree of freedom• no CV for responsibility role
just a bit of school
C.K. Ogden/I.A. Richards, The Meaning of MeaningA Study in the Influence of Language upon Thought and The Science of SymbolismLondon 1923, 10th edition 1969
Concept
Referent
Refers To Symbolizes
Stands For“Orange”
from the slide of [Bargmeyer, Bruce, Open Metadata Forum, Berlin, 2005]
Slide adapted from (c) Key-Sun Choi for Pan Localization 2005
Term