Procurement and interoperability
Welcome & instructions
Achieving simplicity in transacting:
The role of contracts in our society and how we can model the EU economy via contracts
Keynote – Sally Guyer & Luigi Telesca
Global Chief Executive Officer for the World Commerce & Contracting WordCC &
Co-Founder and CEO of TRAKTI and CEO of EXRADE Srl
Track 1 – TOPIC OF THE DAY
Track 2 – SOLUTIONS
Track 3 – INNOVATION
Creation of a semantic base for European Public Procurement
Natalie Muric (Publications Office of the European Union)
A lot of work has been done in the area of public procurement interoperability, however, until now the semantic interoperability ensuring reuse of concepts across systems and the procurement lifecycle has been largely ignored. The Publications Office funded by the ISA2 programme has been tackling this problem by providing reference data in the field of public procurement and developing the eProcurement Ontology in close coordination with the implementation of eForms for the Supplement to the Official Journal (OJ/S). The availability of these artefacts promotes a common understanding of concepts used in procurement, reducing the need for extensive mapping, whilst facilitating the move to linked open data. The Ontology is created by stakeholder participation, which is open and free. Participation in the eProcurement Ontology Working Group and input on authority table requirements are encouraged to provide maximum experience in all areas of public procurement.
AGROVOC: managing a multilingual thesaurus through networked collaboration
Imma Subirats Coll & Kristin Kolshus (Food and Agriculture Organization of the United Nations)
AGROVOC is a multilingual thesaurus, coordinated by FAO, and maintained by an international community of experts and institutions active in the area of agriculture, fisheries, forestry and related domains. AGROVOC is available as an SKOS concept scheme, also published as a Linked Data set composed of +38,000 concepts available in up to 40 languages. By means of Linked Data, AGROVOC is aligned to other open datasets related to agriculture and the environment. This presentation will address multilingualism in thesauri and controlled vocabularies, especially how AGROVOC manages this. Internationalization is not only about translations, it is also about context and knowledge. The AGROVOC Editorial network comprises national and international institutions and expert networks, who curate a language or a technical topic (for example land governance, or aquatic sciences and fisheries). Working with expert communities is essential to keep AGROVOC as up to date as possible with current scientific terminology. In this regard, AGROVOC is now including specialized schemes, which is proving to be very interesting. Having the new AGROVOC Editorial guidelines for shared understanding and maintaining a strong community focus is also essential, as this work is done on a volunteer basis by participating organizations.
Kohesio: a knowledge platform on projects funded by Regional Policy, using a wikibase (semantic web and linked data open source tool)
Anne Thollard & Max De Wilde (European Commission)
Kohesio project aims at building a knowledge base on projects and beneficiaries funded by Regional policy (European Regional Development Fund - ERDF, Cohesion Fund - CF) and European Social Fund (ESF). Objectives are threefold: · Increase visibility and transparency on projects and beneficiaries mainly for the general public, · Re-use and link existing data thanks to the use of a Wikibase, open source and free tool developed by Wikidata, a German branch of the Wikimedia Foundation. Providing extensive open and linked datasets. · Create a community of practice with internal and external stakeholders, including European Commission communication partners in the EU member states.
The European vision for Public Procurement Data
Isabel da Rosa (European Commission)
Building a ‘common language’ for semantic interoperability
Elena Shulman (D.E.Solution)
This talk will describe the steps, collaborations and challenges involved in building a shared language (semantic interoperability) across public facing websites of the European Commission. First part of this undertaking involved the design of an information model and controlled vocabularies for describing and restructuring content from approximately 60 websites of various EC departments. The second part involved mapping heterogeneous legacy vocabularies to a common and centrally managed set of controlled vocabularies with the support of the Publications Office of the European Union. Lessons learned confirmed the necessity for establishing, early on, a central point of coordination and communication dedicated to promoting and enabling semantic interoperability, strong management buy-in, as well as vocabulary management tools and an institutional commitment to provide long-term vocabulary stewardship.
Crowdsourcing and citizen engagement for people-centric smart cities
Elena Simperl (King’s College London)
Smart cities are as much about the needs, expectations and values of the people they serve as they are about the underlying technology. In this talk, I am going to present several areas of system design where human and social factors play a critical role, from fostering participation to augmented intelligence and responsible innovation. I will present two Horizon 2020 programmes: Qrowd, which delivered a humane AI approach for transport and mobility using digital tools, crowdsourcing and citizen-led innovation; and ACTION, which runs an accelerator for participatory initiatives against pollution with 15 pilots in several European countries. In both contexts, I will present approaches to collect and curate open and reference datasets and comment on ongoing challenges, solutions and opportunities.
Towards a comprehensive EU data strategy for eProcurement
Roberto Reale (Eutopian)
The eProcurement Ontology, as developed by the Publications Office of the EU, is a foundational work for shaping the eProcurement landscape across the EU, especially so as key components (eForms, ESPD, eCertis, ...) are continuously improved and harmonised. Whilst the Ontology per se is a very comprehensive effort, there appears to be a missing link as far as new data architecture initiatives (notably the Data Strategy, the European Blockchain Services Infrastructure and GAIA-X) are concerned. Liasing with such initiatives, at the governance, semantic and data architecture levels, is of the utmost importance to create a rich and effective data ecosystem. We will examine governance and technical aspects of such liasing, with a special focus on the European Blockchain Services Infrastructure.
Using a Covid taxonomy to improve the findability of e-learning sessions for health professionals
Maura Moran (Mekon Ltd.)
Health Education England is a body in the Department of Health with a responsibility to provide online education and training for the health and public health workforce. It currently provides over 400 e-learning programmes to 1.5 million registered users.
Feedback indicated that it was difficult to find useful search results on HEE’s e-learning website. This is partly because of the complexity of the health domain itself, where content could be described with many synonyms and terms from many facets. In addition, the content is sourced from a large number of providers in the government and the healthcare community, and this has presented a challenge in describing the content in a useful and consistent manner.
The Covid-19 pandemic created an urgent need to improve the site so that new information could be disseminated, and refresher training could be offered to staff being re-deployed.
This presentation will explain how HEE improved the search accuracy and navigation features by the use of taxonomy. The main focus is the development of a Coronavirus taxonomy to describe the e-learning content. This allowed us to make use of synonyms and hierarchical relationships to improve search accuracy. It also includes facets that healthcare workers need for search. In addition, we licensed additional taxonomies from other government bodies. Together with the new search tools, the taxonomies have made the search and navigation more accurate and useful.
You said interoperability? Lessons Lessons learned from two #AI #ML projects
Valérie Saintot (European Central Bank)
A cultural shift across the board (make data sexy for non-data scientists and engage the executives and leaders to practice evidence-based decision-making and communication) - Show case with the Problem-Data-Facilitation dialogic framework how to create the conditions to achieve the above. PURPOSE: realise that data science is about data science and much more. VALUE PROPOSITION: to work with data requires a lot of thinking, or at least it should. Three different forms of thinking are clearly needed: analytical, creative and critical thinking. Each of these types of thinking corresponds to a different part of the encoding and decoding phases of the data analysis process. When we define the PROBLEM we are trying to solve, we need to mobilise analytical thinking. When searching the DATA to inform the decision-makers about the problem, we need to be thinking creatively to gain insights from the data. When we have harvested a useful dataset and insight, we need critical thinking for an evidenced-based FACILITATION of the necessary discussions. With the PDF framework (Problem-Data-Facilitation) we enable rich discussions to support decision-makers in their work and outreach communication.
Putting in place an eProcurement BDTI pilot
Cécile Guasch & Sander van Dooren (European Commission)
The FAO Caliper Platform
Caterina Caracciolo & Carola Fabi (Food and Agriculture Organization of the United Nations)
In this talk I will present Caliper, a platform developed at FAO dedicated to improving the management and dissemination of statistical classifications. Caliper is inspired by the principles of open and linked data. While all the classifications and correspondences it disseminates are already in the public domain, it adds new possibilities to the ways they can be accessed and consumed. In Caliper, the layman can use its user-friendly, multilingual look-up services, the IT-savvy can access and reuse data in a variety of ways, while classifications’ custodians can have an integrated platform supporting both curation and publication of classifications. Since it relies on open source and free of charge tools, it talks to people and organizations interested in transparency and, possibly, with low budget.
Information transferring in the benefit of policy building
Sabrina Medeiros (InterAgency Institute)
The adaptive approach of public policies suggests that policy building is dependent on the various perceptions of efficiency actors (Walker, Rahman, and Cave 2001). Efficiency is mostly qualified by the original expectations that public policies will come from the arising demands, naturally and systematically organized by institutional processes. Both the expectations of the society and the electoral arena are conditioners of the demands that arise in terms of public policies. A great variety of theories was dedicated to explaining how public policies emerge and why institutional development is done (Wallner 2008, Walker, Rahman, and Cave 2001, Shipan and Volden 2008). Rational-choice theories point a great amount of rationality inside the process of institutional development (Simon 1957, Simon 1991, Simon et al. 1987), while institutionalists put their eyes on the various frameworks that eventually sub-judge interferences. Still within an institutional framework of theory, there is the possibility of observing and understanding institutional process and policy building taking into account the implications of the actors’ trajectories in their natural institutional developments. As Walker have stated, “most policies must be devised in spite of profound uncertainties about the future” (Walker, Rahman, and Cave 2001). At the same time, the governance schemes are built to observe the necessity of changes and institutional evolution, but less effort is dedicated to appropriate practices and experience than to the general view from the top side of institutions. Whereas developments are integrated naturally and actors along those institutional lines of communication may incorporate them, this appropriation is not visible. From the bottom side of the institutional processes, practices are mostly considered natural responses of the institutional development level and strides (Plessner, Betsch, and Betsch 2011, Chhotray and Stoker 2008, Bardach 1998). So, there are two main problems connected: guarantee more visibility of how policy innovation is manifested through institutional developments, and how to better achieve this appropriation from actors’ behaviour in a systematic manner, including semantics. Walker’s adaptive approach diagnostics that policymaking is different in three different aspects, once it is the option: the analytical approach; the types of policies considered, and the decision-making processes (Walker, Rahman, and Cave 2001). So, I propose simulations as methods of observing and analysing institutional change. In sequence, the proposal is to add to the simulation presentation, the schemes for observing and treating information from the various participants as to build policies oriented by fundamental concepts derived from the debates induced. That is why my proposal is both centred on a regular presentation and/or a dynamic eighter to be presented as innovative method and software (online) or to use it inside a workshop as to achieve conclusions for the events main themes.
Break & demonstration
Juremy Legal Term Search
Robin & Timea Palotai (Juremy.com)
Juremy is an EU Corpus-based legal and technical terminology search tool. It provides fast and authentic concordance search within the Eur-Lex database in all 24 official languages. Juremy is streamlined for exact phrase lookup, so matches are presented in the narrow bilingual paragraph context. First, we present how Juremy makes terminology search on the Eur-Lex database fast and customizable, with special regard to the requirements of the EU linguists' translation workflow. Then we present the advantages of our search tool by looking at some typical terminology examples. Finally, the special functionalities are presented. Since Juremy focuses on legal terminology research, it can provide specific, EU-translation focused functionality such as sorting, filters and additional metadata, which will be shown in detail during the demonstration.
Break & demonstration
Caterina Caracciolo (Food and Agriculture Organization of the United Nations)
Towards semantic interdisciplinarity by linking diverse scientific knowledge graphs
Panagiotis-Marios Filippidis, Charalampos Bratsas & Evangelos Chondrokostas (Aristotle University of Thessaloniki)
In this work we build a scientific Knowledge Graph (KG) consisting of generic and specific scientific taxonomies, by joining them through their related concepts. The purpose is to combine the scope of the generic taxonomies with the level of detail of the specific taxonomies, in order to have a KG that can be used in all scientific fields properly. The unified KG is used for semantic enhancement and search of science related content, enriching its annotation with multiple concepts of heterogeneous classifications. The join of the various taxonomies is accomplished by using ontologies matching applications such as the Silk Framework and the Alignment Tool and it is based on similarity metrics software suggestions and domain experts’ contribution and verification. A total of 3 500 links between generic and specific classifications have been created and imported into the KG, benefiting semantic annotation of science related content. To this end, we built a Semantic Engine (SE) to enhance numerous PhD offers with concepts of the semantic KG, analysing the offer text. These concepts are stored as semantic annotations for each offer, including their KG links and the offer is indexed with concepts of various classifications. Another implication is the tagging system of journals and the indexing and search of publications with related content, providing more concepts that correspond to the journal’s content facilitating its search through scientific publications repositories. Finally, graph analysis features provide valuable insights about the similarity and the relevance of diverse scientific fields allowing better results through statistical concepts modelling.
Diverse Uses of a Semantic Graph Database for Knowledge Organization and Research
Vladimir Alexiev (Ontotext Corp.)
Semantic Graph Databases are the foundation of Enterprise Knowledge Graphs. They are used in numerous industrial applications, but also Knowledge Organization Management systems ( thesaurus and ontology management systems), such as VocBench, SWC PoolParty, Synaptica Semaphore. Through VocBench, semantic databases manage or publish some of the most important thesauri: EuroVoc, AgroVoc, the Getty Vocabularies, etc. Semantic databases are also used in a wide variety of research domains and projects. Some have open source or free editions that make them an easy choice for academic research. We searched on Google Scholar and found 1000-1200 academic papers and theses mentioning one of the popular databases. We also found at least 50 books on Google Books that mention it. We started a Zotero bibliography on the topic (currently about 150 papers), and captured about 220 research topics, based on the titles of about 250 papers. We will present an analysis of reference data and research domains using a semantic database. Some of the traditional topics include: social media analytics, data marketplaces, business process management, enterprise data integration, statistical data, engineering, smart cities, sensor networks, life sciences, biomedical ontologies, medicines, chemistry, linguistic data, semantic publishing, semantic text analysis, geographic information, master data management. Newer or more exotic topics include academic/research data, COVID and Zika viruses, Quran and bilingual Arabic-English data, art history, Holocaust research, musical events and adaptations, iconography, food and drink, tourism, investment decision support, economic research, offshore leaks, maritime data, construction projects, building information management, crisis management, critical incidents and infrastructures, data journalism, clinical trials and specific medical topics (e.g. intestinal cells, intracoronal tooth restorations, vaccines, toxicology), investment recommendations, data journalism, etc.
Break & solution room
Common Mapping of Innovation supporting actors – How the interoperable mapping helps SMEs innovate?
Szabolcs Szekacs (European Commission)
There are several websites that support businesses in innovating by providing information on advanced technologies, testing facilities, financing, etc. in the area of innovative solutions. However, in practice, many of those websites provide only fragmented information rather than a comprehensive overview, implying authorities and businesses to visit many websites to obtain all needed data about innovation aspects. This fragmentation and lack of data interoperability represents burden and additional costs for the different interested stakeholders. The European Commission's CMISA project aims to address the aforementioned challenges and needs of the stakeholders and users by allowing the websites that collect information on innovation supporting actors to easily share and reuse each other’s datasets using common, interoperable tools.
Creating Semantic Reference Models for Spaceflight - An Ontology and Knowledge Graph Suite to support Astronautics
Robert Rovetto (Independent consultant)
This work is about creating semantic models of the discipline of astronautics. These domain models will capture the data, knowledge and semantics of the discipline in order to support safe spaceflight, improve space situational awareness, and help resolve the hazard of orbital debris for future generations. More specifically, I am developing a suite of ontologies to encode domain knowledge, and annotate data, for such things as artificial intelligence applications, knowledge representation and reasoning, data fusion, and search and retrieval. The project is open to sponsor, formal collaborations, and is currently described at http://ontospace.wordpress.com and https://purl.org/space-ontology.
Data strategy for eProcurement & interoperability: future actions
Giorgia Lodi (Institute of Cognitive Sciences and Technologies of the Italian National Council of Research)
Patrizia Cannuli (Consip S.p.A)
Carmen Ciciriello (European Commission)
Cécile Guasch (European Commission)
Isabel da Rosa (European Commission)
Steve Graham (Open Peppol)
Cross-organisational collaborative activities around open-source semantic applications to better maintain, link, visualise and disseminate semantic assets
Caterina Caracciolo (Food and Agriculture Organization of the United Nations)
Luca Gramaglia (Eurostat)
Anthony Camilleri (European Commission)
Anikó Gerencsér (Publications Office of the European Union)
Denis Dechandon (Publications Office of the European Union)
This workshop will be run by representatives of the Food and Agriculture Organization of the United Nations, the statistical office of the European Union (Eurostat) and the Publications Office of the European Union. It will mainly focus on the maintenance, alignment and dissemination of 3 major assets: the Statistical Classification of Products by Activity (CPA), the Central Product Classification (CPC) and EuroVoc, the multilingual, multidisciplinary thesaurus covering the activities of the EU. Furthermore, it will demonstrate that open-source semantic technologies and online accessible collaborative platforms play a crucial role in and for developing the semantic interoperability and supporting the sharing and implementation of international standards, as well as further strengthening the dissemination of semantic assets and supporting the interlinking of data.
Prêt-à-LLOD: Making Linguistic Data Ready-to-use with Linked Data
John McCrae (National University of Ireland Galway)
The Prêt-à-LLOD project is building a data value chain that makes linguistic data ready-to-use in a variety of applications, and in this talk we will provide an overview of the toolkit that the project is developing for this data value chain. There are six main steps to this chain: data discovery, which is being supported by a new data portal for linguistic data, Linghub, that combines different sources using state-of-the-art linked data technologies. Secondly, we are concerned with the preparation of data, including converting it into a format that enables further analysis, for which we are developing an innovative tool called FINTAN. The next step consists of organization and our focus here is on the legal aspects of data reuse especially around licensing. For this we are looking at the use of blockchain technologies such as IPFS to ensure that information about resources can be shared and combined through open standards such as ODRL. The next step is the integration of data where we are developing a new tool called Naisc to semi-automatically integrate datasets. Then, we must actually develop our analysis using Docker and linked data combined in a new workflow management platform called Teanga. Finally, these tools will lead to real-world actions especially relative to under-resourced languages through Europe and the World.
An Opportunity to Unite us in Terminology
Stine Jensen (DALeXI)
Knowledge management and terminology goes hand in hand as terminology is the vocabulary of professional knowledge covering a given knowledge field; both are ordered and categorised through ontologies and meta data, respectively. To give an example, EuroVoc is used as a terminology classification system for IATE and the Eurotermbank. Knowledge sharing of the scale that ISA2 and the Publication Office offer – in terms of support and guidance on best practices drawn from experience with ISA2 – is an extraordinary opportunity to perform large scale terminology management which would involve not only sharing but revising the terminology as all the field work and thus first step of a collaborative terminology management (involving users, experts, creators, legislators, aligning material, etc.) has already been taken. Such terminology management would create one source of reference that could serve member states with a weak or no public terminology management and most likely also attract private sector engagement. This would only enrich a resource of this kind due to the interconnection of the private and public sector, encouraging that way also cross-border and cross-sector interoperability. Furthermore, adding a promotional aspect and contribution motivation that might end up easing silos. One of which could be that updated terminology in an AI perspective plays an essential role in areas of interest, such as machine learning, digitalisation, medical report writing, chatbot technology, machine translation, etc. ISA2 presents a chance of preparing terminology for when the (language) technology is ready to integrate it!
Endorse 5 star deployment scheme for Linked Data
Ivo Velitchkov (KVISTGAARD)
The promise of Linked Data was to complement the web of documents with the web of data so that humans and machines can use the Internet as if it is a single database, while enjoying the benefits of decentralisation. Today, 15 years later, the web of data grew to 1 470 Linked Open Data publications. Yet, that is just 0.005% of all publicly known datasets. Apart from LOD, there is Linked Enterprise Data that can unite the heterogeneous corporate data sources into a single enterprise knowledge graph and in doing so achieve data integration at a small fraction of the cost of current data integration projects and bring additional value due to its low cost of change. Yet, corporations keep wasting IT investments in creating data silos and applying application-centric approaches to data integration. So why is it that Linked Data, both open and restricted, is still marginal? The main reasons are that Linked Data is still not attractive for big software vendors and for developers; and it is still perceived as too academic. But there are a few other, less obvious reasons, hiding where nobody would look for them: the successful LOD practices. This talk will examine this finding, recommend some remedies and suggest to complement the 5-star open data deployment scheme with 5-star Linked Data deployment scheme to distinguish the different impact of the way Linked Data is implemented and used.
Track 1 – EVOLUTION
Track 2 – SOLUTIONS
Track 3 – INNOVATION
Search of Knowledge (Tables of Contents and Indexes)
Meral Alakuş (Dahousie University)
Looking into the future and realizing the fast accumulation of knowledge specially during the last century and to these days, it is obvious that controlling knowledge will be the most crucial problem people will have to deal with in the coming years. Coping with the information explosion will be the responsibility of “Knowledge and/or Information Managers,” e.g. librarians and indexers, subject specialists and researchers as well as computer scientists. Organizing knowledge will be carried through analytic study of content matter, topics and subtopics will be identified with keywords; personal and other names will also be given just as they are recorded in the text, indicating relationships between them and showing page / location numbers. All this is done with the help of human intellect formulating and rearranging content matter under “Tables of Contents and Index”. Same rules are applied to books, journals, reports and other recorded materials, print or virtual. It is rightly claimed by the British Council Information Centre that “for the organization of information and of documents, the contribution of technology in comparison to human intellect, is 85% to 15% in favour of intellect.” In this paper, in search of knowledge, I endeavour to explore the uses of two traditional tools made available in the publishing world: namely the “Book Index” and “Table of Contents. It certainly seems that indexing will never disappear as long as knowledge in various formats grows and expands; in application, it might require different ways to approach the full content, which is in fact content analysis and defining topics with keywords. With the continuing increases in computer processing and storage capabilities, the barriers to and benefits of electronic access to more information content are becoming serious issues in information science research.
Interoperability of Semantically-Enabled Web Services on the WoT: Challenges and Prospects
Mahdi Bennara (École des Mines de Saint-Étienne)
The advent of the Web of Things as an application layer for the Internet of Things has led to the proliferation of Web services exposing data and functionality of the networked objects. Since the resource-oriented paradigm aligns well with the WoT architectures, RESTful services have been the go-to interface to expose the connected devices on the Web in a lightweight resource-oriented manner. However, the heterogeneity of descriptions of devices and services as well as the data-formats they exchange has led to a number of interoperability issues. Recently, the growing popularity of semantically-enabled services has led to the emergence of services described with and exchanging RDF. RDF can be seen as a universal abstract data model, or as a lingua franca for the data formats of the Web. By leveraging the power of REST and RDF and combining them with the Thing-oriented nature of the WoT, we believe that true semantic interoperability on the Web of Things can be achieved. In this work, we attempt to frame the challenges encountered in enabling semantic interoperability of heterogeneous WoT services, with the help of a real world production line scenario. We also propose preliminary solutions to the main issues hampering the establishment of true semantic interoperability on the WoT.
A Web-Based Operating System for Integrating Heterogeneous Data and Code Using Semantic Web Technologies as Interlingua
Sergejs Kozlovičs (Institute of Mathematics and Computer Science, University of Latvia)
The purpose of the lightning talk is to discuss the idea of the web operating system which would unify the web at both the data and computational levels. Instead of imposing a particular standard or specific requirements, our idea is to allow individuals, government, and private companies to publish their data using formats of their choice. Similarly, we allow them to provide access to their code using technologies of their choice. The proposed web operating system will provide a means to integrate these data and code by utilizing the OS driver metaphor. Besides, it will raise the level of abstraction by factoring out network-specific aspects, including security and access management. During the talk, we will demonstrate an initial working version of the proposed OS prototype, currently known as webAppOS (an open-source project, licensed under EUPL). We propose to use semantic web technologies (RDF, OWL, and SPARQL) as an interlingua for integrating heterogeneous data and code. Drivers will provide on-the-fly access to concrete data formats and storages (including cloud drives) and implement different code invocation methods. Besides, we will join heterogeneous data and distributed code by introducing the innovative web memory and web calls concepts. The proposed OS will draw us closer to the idea of a web-based computational infrastructure of migrating objects, offered by Alan Kay. That would facilitate cooperation between individuals, government, and private companies and boost interoperability between their services.
Challenges to promote Interoperability in a bicameral Parliament: the example of Cortes Generales
Marina Cueto (Senate of Spain)
I would like to speak about of the first steps reached by the Spanish Senate towards interoperability through using semantic assets and drafting work. The first conclusion of my presentation is that a collaboration with the other chamber of the Spanish Parliament, the Congress of Deputies, is necessary: both Chambers began together in the 80’s in the beginning of the Eurovoc thesaurus; it is important to promote a better understanding of parliamentary work, in the digital world, in which both chambers must speak thorough unique semantic; and because the necessary collaboration with Government (Ministry of the Presidency, Relations with the Parliament and Democratic Memory) in the making of legal documents, traditionally on paper and sendings, today must be reached with draft techniques where LEOS seems to be the tool to share and to amend documents.
Semantic data for better operations
Lambert Hogenhout (United Nations Secretariat)
The Emerging Tech Lab at the UN Secretariat is building a semantic layer to connect the overlapping areas of our different departments and SDG-related programmes with the aim to facilitate research, strengthen IT systems and foster collaboration among staff. By using existing taxonomies and linked data we established a basic knowledge graph (for various SDG-related topics and some specific topics such as humanitarian affairs). We have enriched this knowledge graph with automated methods (NLP and Machine Learning). We have developed various algorithms to leverage this graph for particular purposes, for instance to determine document similarity or assess the alignment of documents (such as project proposals) to strategic frameworks. The resulting graph will be used (and is already being used) in various ways: (a) to give human users a better insight in the relation between topics, including though immersive experiences such as augmented reality (we have implemented this on smartphones and are working on a version for headsets), (b) to underpin IT applications with a layer of knowledge, e.g. our conversational AI platform (chatbot) may use this layer in the future to improve its natural language understanding (NLU) and our Search Engine could use this to allow interactive discovery as an alternative to search results (c) allow UN staff to more easily find peers in the organization working on issues or possessing expertise that is relevant to their work.
Using Wikibase to connect & collaborate around structured data
Sam Alípio (Wikimedia)
Wikibase, the free and open source software behind Wikidata, is being increasingly adopted by organizations and researchers as a platform for creating knowledge bases of linked (open) data. Through a selection of case studies, learn how projects around the world are using Wikibase to power structured data repositories that break down knowledge silos and enable cross-disciplinary collaboration.
Europeana, the reference of European digital cultural heritage
Fulgencio Sanmartín & Katerina Moutogianni (European Commission)
Europeana is the Commission’s flagship in the wider policy area of digitisation, online accessibility and digital preservation of European cultural heritage. It is currently funded as a Digital Service Infrastructure under the Connecting Europe Facility programme. Europeana incorporates and works with thousands of European archives, libraries and museums to share cultural heritage for enjoyment, education and research. There are currently over 50 million records in its repository (images, publications, music, artworks and more), and 2.8 billion linked open data about them. The Europeana Data Model, the Europeana Publishing Framework (content and metadata components) and the International Rights Statements are key components for interoperability, sharing and re-use of cultural heritage data in Europeana. A future Common Data Space for Cultural Heritage will build on this initiative.
Lucy Walhain & Anikó Gerencsér (Publications Office of the European Union)
Najeh Hajlaoui & Denis Dechandon (Publications Office of the European Union)
Bernhard Krabina (KDZ - Zentrum für Verwaltungsforschung)
Wikidata as a central hub in the linked open data web
Lydia PINTSCHER (Wikimedia)
Wikidata, Wikipedia’s sister project, is a large knowledge graph containing general purpose data about the world, collected and maintained by thousands of people. Wikidata’s data is used in a wide variety of applications from digital personal assistants to data visualisations to making Wikipedia better. But it has another super power. It connects more than 5 000 catalogues, databases, social networking sites and more through identifiers, and has thereby become the biggest identifier hub on the web today. This talk will dive into what this means and why it matters for you.
Closure & SliDo survey (input from participants in the form of recommendations for the future)
Closure & SliDo survey (input from participants in the form of recommendations for the future)
Closure & SliDo survey (input from participants in the form of recommendations for the future)