Tag Archives: DBpedia

Call for Participants: DBpedia Autumn Hackathon

Dear DBpedians, Linked Data savvies and Ontologists,

We would like to invite you to join the DBpedia Autumn Hackathon 2020 as a new format to contribute to DBpedia, gain fame, win small prizes and experience the latest technology provided by DBpedia Association members. 
The hackathon is part of the Knowledge Graphs in Action conference on October 6, 2020. 

Timeline 

  • Registration of participants – main communication channel will be the #hackathon channel in DBpedia Slack (sign up, then add yourself to the channel). If you wish to receive a reminder email on Sep 21st, 2020 you can leave your email address in this form.
  • Until September 14th – preparation phase, participating organizations prepare details and track formation. Additional tracks can be proposed, please contact dbpedia-events@infai.org.
  • Announcement of details for each track, including prizes, participating data, demos as well as tools and tasks. Please check updates on the Hackathon website. – September 21st, 2020
  • Hacking period, coordinated via DBpedia slack September 21st to October 1st, 2020
  • Submission of hacking result (3 min video and 2-3 paragraph summary with links, if not stated otherwise in the track) – October 1st, 2020 at 23:59 Hawaii Time
  • Final Event, Each track chair will present a short recap of the track and announces prizes or summarizes the result of hacking. – October 5th, 2020 at 16:00 CEST
  • Knowledge Graphs in Action Event (see program) – October 6th, 2020 at 9:50 – 15:30 CEST
  • Results and videos are documented on the DBpedia Website and the DBpedia Youtube channel.

Member Tracks 

The member tracks are hosted by DBpedia Association members, who are technology leaders in the area of Knowledge Engineering. Additional tracks can be proposed until Sep 14th, please contact dbpedia-events@infai.org.

  • timbr SQL Knowledge Graph: Learn how to model, map and query ontologies in timbr and then model an ontology of GDELT, map it to the GDELT database, and answer a number of questions that currently are quite impossible to get from the BigQuery GDELT database. Cash prizes planned. 
  • GNOSS Knowledge Graph Builder: Give meaning to your organisation’s documents and data with a Knowledge Graph. 
  • ImageSnippets: Labeling images with semantic descriptions. Use DBpedia spotlight and an entity matching lookup to select DBpedia terms to describe images. Then explore the resulting dataset through searches over inference graphs and explore the ImageSnippets dataset through our SPARQL endpoint. Prizes planned. 
  • Diffbot: Build Your Own Knowledge Graph! Use the Natural Language API to extract triples from natural language text and expand these triples with data from the Diffbot Knowledge Graph (10+ billion entities, 1+ trillion facts). Check out the demo. All participants will receive access to the Diffbot KG and tools for (non-commercial) research for one year ($10,000 value).

Dutch National Knowledge Graph Track

Following the DBpedia FlexiFusion approach, we are currently flexi-fusing a huge, dbpedia-style knowledge graph that will connect many Linked Data sources and data silos relevant to the country of the Netherlands. We hope that this will eventually crystallize a well-connected sub-community linked open data (LOD) cloud in the same manner as DBpedia crystallized the original LOD cloud with some improvements (you could call it LOD Mark II). Data and hackathon details will be announced on 21st of September.

Organising committee:

Improve DBpedia Track

A community track, where everybody can participate and contribute in improving existing DBpedia components, in particular the extraction framework, the mappings, the ontology, data quality test cases, new extractors, links and other extensions. Best individual contributions will be acknowledged on the DBpedia website by anointing the WebID/Foaf profile.

(chaired by Milan Dojchinovski and Marvin Hofer from the DBpedia Association & InfAI and the DBpedia Hacking Committee, please message @m1ci to volunteer to the hacking committee)

DBpedia Open Innovation Track 

(not part of the hackathon, pre-announcement)

For the DBpedia Spring Event 2021, we are planning an Open Innovation Track, where DBpedians can showcase their applications. This endeavour will not be part of the hackathon as we are looking for significant showcases with development effort of months & years built on the core infrastructure of DBpedia such as the SPARQL endpoint, the data, lookup, spotlight, DBpedia Live, etc. Details will be announced during the Hackathon Final Event on October 5.  

(chaired by Heiko Paulheim et al.)

Stay tuned and check Twitter, Facebook and our Website or subscribe to our Newsletter for latest news and information.

The DBpedia Organizing Team


‘Knowledge Graphs in Action’ online event on Oct 6, 2020

Due to current circumstances, the SEMANTiCS Onsite Conference 2020 had, unfortunately, to be postponed till September 2021. To bridge the gap until 2021, DBpedia, PLDN and EuroSDR will organize a SEMANTiCS satellite event online, on October 6, 2020. We set up an exciting themed program around ‘Knowledge Graphs in Action: DBpedia, Linked Geodata and Geo-information Integration’.

This new event is a combination of two already existing ones: the DBpedia Community Meeting, which is regularly held as part of the SEMANTiCS, and the annual Spatial Linked Data conference organised by EuroSDR and the Platform Linked Data Nederland. We fused both together and as a bonus, we added a track about Geo-information Integration hosted by EuroSDR. For the joint opening session, we recruited four amazing keynote speakers to kick the event off.    

Highlights of the Knowledge Graph in Action event

– Hackathon (starts 2 weeks earlier)

– Keynote by Carsten Hoyer-Click, German Aerospace Center

– Keynote by Marinos Kavouras, National Technical University of Athens

– Keynote by Peter Mooney, Maynooth University

– Spatial Linked Data Country Session

– DBpedia Chapter Session

– Self Service GIS Session

– DBpedia Showcase Session

Quick Facts

– Web URL: https://wiki.dbpedia.org/meetings/KnowledgeGraphsInAction

– When: October 6, 2020

– Where: The conference will take place fully online.

Schedule

– Please check the schedule for the upcoming Knowledge Graphs in Action event here: https://wiki.dbpedia.org/meetings/KnowledgeGraphsInAction  

Registration 

– Attending the conference is free. Registration is required though. Please get in touch with us if you have any problems during the registration stage. Register here to be part of the meeting: https://wiki.dbpedia.org/meetings/KnowledgeGraphsInAction 

Organisation

– Benedicte Bucher, University Gustave Eiffel, IGN, EuroSDR

– Erwin Folmer, Kadaster, University of Twente, Platform Linked Data Netherlands

– Rob Lemmens, University of Twente

– Sebastian Hellmann, AKSW/KILT, DBpedia Association

– Julia Holze, DBpedia Association

Don’t think twice and register now! Join the Knowledge Graph in Action event on October 6, 2020 to catch up with the latest research results and developments in the Semantic Web Community. Register here and meet us and other SEMANTiCS enthusiasts.

For latest news and updates check Twitter, LinkedIn, the DBpedia blog and our Website or subscribe to our newsletter.

We are looking forward to meeting you online!

Julia

on behalf of the DBpedia Association

DBpedia Workshop at LDAC

More than 90 DBpedia enthusiasts joined the DBpedia Workshop colocated with LDAC2020

On June 19, 2020 we organized a DBpedia workshop co-located with the LDAC workshop series to exchange knowledge regarding new technologies and innovations in the fields of Linked Data and Semantic Web. This workshop series provides a focused overview on technical and applied research on the usage of Semantic Web, Linked Data and Web of Data technologies for the architecture and construction domains (design, engineering, construction, operation, etc.). The workshop aims at gathering researchers, industry stakeholders, and standardization bodies of the broader Linked Building Data (LBD) community.

First and foremost, we would like to thank the LDAC committee for hosting our virtual meeting and many thanks to Beyza Yaman, Milan Dojchinovski, Johannes Frey and Kris McGlinn for organizing and chairing the DBpedia workshop. 

Following, we will give you a brief retrospective about the presentations.

Opening & Keynote 

The first virtual DBpedia meeting was opened with a keynote presentation ‘{RDF} Data quality assessment – connecting the pieces’ by Dimitris Kontokostas (diffbot, US). He gave an overview on the latest developments and achievements around Data Quality. His presentation was focused on defining data quality and identification of data quality issues.  

Sebastian Hellmann gave a brief overview of DBpedia’s history. Furthermore, he presented the updated DBpedia Organisational architecture, including the vision of the new DBpedia chapters and benefits of the DBpedia membership.

Shortly after,  Milan Dojchinovski (InfAI/CTU in Prague) gave a presentation on  ‘Querying and Integrating (Architecture and Construction) Data with DBpedia’. ‘The New DBpedia Release Cycle’ was introduced by Marvin Hofer (InfAI). Closing the Showcase Session, Johannes Frey, InfAI, presented the Databus Archivo and demonstrated the downloading process with the DBpedia Databus

For further details of the presentations follow the links to the slides.

  • Keynote: {RDF} Data quality assessment – connecting the pieces, by Dimitris Kontokostas, diffbot, US (slides)
  • Overview of DBpedia Organisational Architecture, by Sebastian Hellmann, Julia Holze, Bettina Klimek, Milan Dojchinovski, INFAI / DBpedia Association (slides)
  • Querying and Integrating (Architecture and Construction) Data with DBpedia by Milan Dojchinovski, INFAI/CTU in Prague (slides)
  • The New DBpedia Release Cycle by Marvin Hofer and Milan Dojchinovski, INFAI (slides)
  • Databus Archivo and Downloading with the Databus by Johannes Frey, Fabian Goetz and Milan Dojchinovski, INFAI (slides)

Geospatial Data & DBpedia Session

After the opening session we had the Geospatial Data & DBpedia Session. Milan Dojchinovski (InfAI/CTU in Prague) chaired this session with three very stimulating talks. Hereafter you will find all presentations given during this session:

  • Linked Geospatial Data & Data Quality by Wouter Beek, Triply Ltd. (slides)
  • Contextualizing OSi’s Geospatial Data with DBpedia by Christophe Debruyne, Vrije Universiteit Brussel and ADAPT at Trinity College Dublin
  • Linked Spatial Data: Beyond The Linked Open Data Cloud by Chaidir A. Adlan, The Deutsche Gesellschaft für Internationale Zusammenarbeit GmbH (slides)

Data Quality & DBpedia Session

The first online DBpedia workshop also covered a special data quality session. Johannes Frey (InfAI) chaired this session with three very stimulating talks. Hereafter you will find all presentations given during this session:

  • SeMantic AnsweR Type prediction with DBpedia – ISWC 2020 Challenge by Nandana Mihindukulasooriya, MIT-IBM Watson AI Lab (slides)
  • RDF Doctor: A Holistic Approach for Syntax Error Detection and Correction of RDF Data by Ahmad Hemid, Fraunhofer IAIS (slides)
  • The Best of Both Worlds: Unlocking the Power of (big) Knowledge Graphs with SANSA by Gezim Sejdiu,  Deutsche Post DHL Group and University of Bonn (slides)
  • Closing words by the workshop organizers

In case you missed the event, all slides and presentations are also available on the DBpeda workshop website. Further insights, feedback and photos about the event are available on Twitter (#DBpediaDay hashtag).

We are now looking forward to our first DBpedia Stack tutorial, which will be held online on July 1st, 2020. Over the last year, the DBpedia core team has consolidated a great amount of technology around DBpedia. The tutorial primarily targets developers (in particular of DBpedia Chapters) that wish to learn how to replicate local infrastructure such as loading and hosting an own SPARQL endpoint. A core focus will also be the new DBpedia Stack, which contains several dockerized applications that are automatically loading data from the Databus. Attending the DBpedia Stack tutorial is free and will be organized online. Please register to be part of the meeting.

Stay tuned and check Twitter, Facebook and our Website or subscribe to our Newsletter for latest news and information.

Julia and Milan 

on behalf of the DBpedia Association

GSoC2020 – Call for Contribution

James: Sherry with the soup, yes… Oh, by the way, the same procedure as last year, Miss Sophie?

Miss Sophie: Same procedure as every year, James.

…and we are proud of it. We are very grateful to be accepted as an open-source organization in this years’  Google Summer of Code (GSoC2020) edition, again. The upcoming GSoC2020 marks the 16th consecutive year of the program and is the 9th year in a row for DBpedia. 

We did it again – We are mentoring organization!

What is GSoC2020? 

Google Summer of Code is a global program focused on bringing student developers into open source software development. Funds will be given to students (BSc, MSc, PhD.) to work for three months on a specific task. For GSoC-Newbies, this short video and the information provided on their website will explain all there is to know about GSoC2020.

This year’s Narrative

Last year we tried to increase female participation in the program and we will continue to do so this year. We want to encourage explicitly female students to apply for our projects. That being said, we already engaged excellent female mentors to also raise the female percentage in our mentor team. 

In the following weeks, we invite all students, female and male alike, who are interested in Semantic Web and Open Source development to apply for our projects. You can also contribute your own ideas to work on during the summer. 

And this is how it works: 4 steps to GSoC2020 stardom

  1. Open source organizations such as DBpedia announce their projects ideas. You can find our project here
  2. Students contact the mentor organizations they want to work with and write up a project proposal. Please get in touch with us via the DBpedia Forum or dbpedia@infai.org as soon as possible.
  3. The official application period at GSoC starts March, 16th. Please note, you have to submit your final application not through our Forum, but the GSoC Website
  4. After a selection phase, students are matched with a specific project and a set of mentors to work on the project during the summer.

To all the smart brains out there, if you are a student who wants to work with us during summer 2020, check our list of project ideas, warm-up tasks or come up with your own idea and get in touch with us.

Application Procedure

Further information on the application procedure is available in our DBpedia Guidelines. There you will find information on how to contact us and how to appropriately apply for GSoC2020. Please also note the official GSoC 2020 timeline for your proposal submission and make sure to submit on time.  Unfortunately, extensions cannot be granted. Final submission deadline is March 31st, 2020, 8 pm, CEST.

Finally, check our website for information on DBpedia, follow us on Twitter or subscribe to our newsletter.

And in case you still have questions, please do not hesitate to contact us via praetor@infai.org.

We are thrilled to meet you and your ideas.

Your DBpedia-GSoC-Team


GlobalFactSync and WikiDataCon2019

We will be spending the next three days in Berlin at WikidataCon 2019, the conference for open data enthusiasts. From October 24th till 26th we will be presenting the latest developments and first results of our work in the GlobalFactSyncRE-Project. 

Short Project Intro

Funded by the Wikimedia Foundation, the project started in June 2019 and has two goals:

  • Answer the following questions:
    • How is data edited in Wikipedia and Wikidata?
    • Where does it come from?
    • How can we synchronize it globally?
  • Build an information system to synchronize facts between all Wikipedia language-editions, Wikidata, DBpedia and eventually multiple external sources, while also providing respective references. 

In order to help Wikipedians to maintain their infoboxes, check for factual correctness, and also improve data in Wikidata, we use data from Wikipedia infoboxes of different languages, Wikidata, and DBpedia and fuse them into our PreFusion dataset (in JSON-LD). More information on the fusion process, which is the engine behind GFS, can be found in the FlexiFusion paper.

Can’t join the conference or want to find out more about GlobalFactSync?

No problem, the poster we are presenting at the conference is currently available here and will soon be available here. Additionally, why not go through our project timeline, follow up on our progress so far and find out what’s coming up next.

In case you have specific questions regarding GlobalfactSync or even some helpful feedback just ping us via dbpedia@infai.org. We also have our new DBpedia Forum, home to the DBpedia Comunity, which just waits for you to initialize a discussion around GlobalFactSync. Why not start it now?

For general DBpedia news and updates follow us on Twitter.

…And if you are in Berlin at WikiDataCon2019 stop by our poster and talk to our developers. They are looking forward to vital exchanges with you.

All the best

yours,


DBpedia Association


SEMANTiCS 2019 Interview: Katja Hose

Today’s post features an interview with our DBpedia Day keynote speaker Katja Hose, a Professor of Computer Science at Aalborg University, Denmark. In this Interview, Katja talks about increasing the reliability of Knowledge Graph Access as well as her expectations for SEMANTiCS 2019

Prior to joining Aalborg University, Katja was a postdoc at the Max Planck Institute for Informatics in Saarbrücken. She received her doctoral degree in Computer Science from Ilmenau University of Technology in Germany.

Can you tell us something about your research focus?

The most important focus of my research has been querying the Web of Data, in particular, efficient query processing over distributed knowledge graphs and Linked Data. This includes indexing, source selection, and efficient query execution. Unfortunately, it happens all too often that the services needed to access remote knowledge graphs are temporarily not available, for instance, because a software component crashed. Hence, we are currently developing a decentralized architecture for knowledge sharing that will make access to knowledge graphs a reliable service, which I believe is the key to a wider acceptance and usage of this technology.

How do you personally contribute to the advancement of semantic technologies?

I contribute by doing research, advancing the state of the art, and applying semantic technologies to practical use cases.  The most important achievements so far have been our works on indexing and federated query processing, and we have only recently published our first work on a decentralized architecture for sharing and querying semantic data. I have also been using semantic technologies in other contexts, such as data warehousing, fact-checking, sustainability assessment, and rule mining over knowledge bases.

Overall, I believe the greatest ideas and advancements come when trying to apply semantic technologies to real-world use cases and problems, and that is what I will keep on doing.

Which trends and challenges do you see for linked data and the semantic web?

The goal and the idea behind Linked Data and the Semantic Web is the second-best invention after the Internet. But unlike the Internet, Linked Data and the Semantic Web are only slowly being adopted by a broader community and by industry.

I think part of the reason is that from a company’s point of view, there are not many incentives and added benefit of broadly sharing the achievements. Some companies are simply reluctant to openly share their results and experiences in the hope of retaining an advantage over their competitors. I believe that if these success stories were shared more openly, and this is the trend we are witnessing right now, more companies will see the potential for their own problems and find new exciting use cases.

Another particular challenge, which we will have to overcome, is that it is currently still far too difficult to obtain and maintain an overview of what data is available and formulate a query as a non-expert in SPARQL and the particular domain… and of course, there is the challenge that accessing these datasets is not always reliable.

As artificial intelligence becomes more and more important, what is your vision of AI?

AI and machine learning are indeed becoming more and more important. I do believe that these technologies will bring us a huge step ahead. The process has already begun. But we also need to be aware that we are currently in the middle of a big hype where everybody wants to use AI and machine learning – although many people actually do not truly understand what it is and if it is actually the best solution to their problems. It reminds me a bit of the old saying “if the only tool you have is a hammer, then every problem looks like a nail”. Only time will tell us which problems truly require machine learning, and I am very curious to find out which solutions will prevail.

However, the current state of the art is still very far away from the AI systems that we all know from Science Fiction. Existing systems operate like black boxes on well-defined problems and lack true intelligence and understanding of the meaning of the data. I believe that the key to making these systems trustworthy and truly intelligent will be their ability to explain their decisions and their interpretation of the data in a transparent way.

What are your expectations about Semantics 2019 in Karlsruhe?

First and foremost, I am looking forward to meeting a broad range of people interested in semantic technologies. In particular, I would like to get in touch with industry-based research and to be exposed 

The End

We like to thank Katje Hose for her insights and are happy to have her as one of our keynote speakers.

Visit SEMANTiCS 2019 in Karlsruhe, Sep 9-12 and get your tickets for our community meeting here. We are looking forward to meeting you during DBpedia Day.

Yours DBpedia Association

Global Fact Sync – Synchronizing Wikidata & Wikipedia’s infoboxes

How is data edited in Wikipedia/Wikidata? Where does it come from? And how can we synchronize it globally?  

The GlobalFactSync (GFS) Project — funded by the Wikimedia Foundation — started in June 2019 and has two goals:

  • Answer the above-mentioned three questions.
  • Build an information system to synchronize facts between all Wikipedia language-editions and Wikidata. 

Now we are seven weeks into the project (10+ more months to go) and we are releasing our first prototypes to gather feedback. 

How – Synchronization vs Consensus

We follow an absolute “Human(s)-in-the-loop” approach when we talk about synchronization. The final decision whether to synchronize a value or not should rest with a human editor who understands consensus and the implications. There will be no automatic imports. Our focus is to drastically reduce the time to research all references for individual facts.

A trivial example to illustrate our reasoning is the release date of the single “Boys Don’t Cry” (March 16th, 1989) in the English, Japanese, and French Wikipedia, Wikidata and finally in the external open database MusicBrainz.  A human editor might need 15-30 minutes finding and opening all different sources, while our current prototype can spot differences and display them in 5 seconds.

We already had our first successful edit where a Wikipedia editor fixed the discrepancy with our prototype: “I’ve updated Wikidata so that all five sources are in agreement.” We are now working on the following tasks:

  • Scaling the system to all infoboxes, Wikidata and selected external databases (see below on the difficulties there)
  • Making the system:
    •  “live” without stale information
    • “reliable” with less technical errors when extracting and indexing data
    • “better referenced” by not only synchronizing facts but also references 

Contributions and Feedback

To ensure that GlobalFactSync will serve and help the Wikiverse we encourage everyone to try our data and micro-services and leave us some feedback, either on our Meta-Wiki page or via email. In the following 10+ months, we intend to improve and build upon these initial results. At the same time, these microservices are available to every developer to exploit it and hack useful applications. The most promising contributions will be rewarded and receive the book “Engineering Agile Big-Data Systems”. Please post feedback or any tool or GUI here. In case you need changes to be made to the API, please let us know, too.
For the ambitious future developers among you, we have some budget left that we will dedicate to an internship. In order to apply, just mention it in your feedback post. 

Finally, to talk to us and other GlobalfactSync-Users you may want to visit WikidataCon and Wikimania, where we will present the latest developments and the progress of our project. 

Data, APIs & Microservices (Technical prototypes) 

Data Processing and Infobox Extraction

For GlobalFactSync we use data from Wikipedia infoboxes of different languages, as well as Wikidata, and DBpedia and fuse them to receive one big, consolidated dataset – a PreFusion dataset (in JSON-LD). More information on the fusion process, which is the engine behind GFS, can be found in the FlexiFusion paper. One of our next steps is to integrate MusicBrainz into this process as an external dataset. We hope to implement even more such external datasets to increase the amount of available information and references. 

First microservices 

We deployed a set of microservices to show the current state of our toolchain.

  • [Initial User Interface] The GlobalFactSync UI prototype (available at http://global.dbpedia.org) shows all extracted information available for one entity for different sources. It can be used to analyze the factual consensus between different Wikipedia articles for the same thing. Example: Look at the variety of population counts for Grimma.
  • [Reference Data Download] We ran the Reference Extraction Service over 10 Wikipedia languages. Download dumps here.
  • [ID service] Last but not least, we offer the Global ID Resolution Service. It ties together all available identifiers for one thing (i.e. at the moment all DBpedia/Wikipedia and Wikidata identifiers – MusicBrainz coming soon…) and shows their stable DBpedia Global ID. 

Finding sync targets

In order to test out our algorithms, we started by looking at various groups of subjects, our so-called sync targets. Based on the different subjects a set of problems were identified with varying layers of complexity:

  • identity check/check for ambiguity — Are we talking about the same entity? 
  • fixed vs. varying property — Some properties vary depending on nationality (e.g., release dates), or point in time (e.g., population count).
  • reference — Depending on the entity’s identity check and the property’s fixed or varying state the reference might vary. Also, for some targets, no query-able online reference might be available.
  • normalization/conversion of values — Depending on language/nationality of the article properties can have varying units (e.g., currency, metric vs imperial system).

The check for ambiguity is the most crucial step to ensure that the infoboxes that are being compared do refer to the same entity. We found, instances where the Wikipedia page and the infobox shown on that page were presenting information about different subjects (e.g., see here).

Examples

As a good sync target to start with the group ‘NBA players’ was identified. There are no ambiguity issues, it is a clearly defined group of persons, and the amount of varying properties is very limited. Information seems to be derived from mainly two web sites (nba.com and basketball-reference.com) and normalization is only a minor issue. ‘Video games’ also proved to be an easy sync target, with the main problem being varying properties such as different release dates for different platforms (Microsoft Windows, Linux, MacOS X, XBox) and different regions (NA vs EU).

More difficult topics, such as ‘cars’, ’music albums’, and ‘music singles’ showed more potential for ambiguity as well as property variability. A major concern we found was Wikipedia pages that contain multiple infoboxes (often seen for pages referring to a certain type of car, such as this one). Reference and fact extraction can be done for each infobox, but currently, we run into trouble once we fuse this data. 

Further information about sync targets and their challenges can be found on our Meta-Wiki discussion page, where Wikipedians that deal with infoboxes on a regular basis can also share their insights on the matter. Some issues were also found regarding the mapping of properties. In order to make GlobalFactSync as applicable as possible, we rely on the DBpedia community to help us improve the mappings. If you are interested in participating, we will connect with you at http://mappings.dbpedia.org and in the DBpedia forum.  

BottomlineWe value your feedback

Your DBpedia Association

Vítejte v Praze!

After our meetups in Poland and France last year, we delighted the Czech DBpedia community with a DBpedia meetup. It was co-located with the XML Prague conference on February 7th, 2019.

First and foremost, we would like to thank Jirka Kosek (University of Economics, Prague), Milan Dojchinovski (AKSW/KILT, Czech Technical University in Prague), Tomáš Kliegr (KIZI/University of Economics, Prague) and, the XML Prague conference for co-hosting and support the event.

Opening the DBpedia community meetup

The Czech DBpedia community and the DBpedia Databus were in the focus of this meetup. Therefore, we invited local data scientists as well as DBpedia enthusiasts to discuss the state-of-the-art of the DBpedia databus. Sebastian Hellmann (AKSW/KILT) opened the meeting with an introduction to DBpedia and the DBpedia Databus. Following, Marvin Hofer explained how to use the DBpedia databus in combination with the Docker technology and, Johannes Frey (AKSW/KILT) presented the methods behind the DBpedia’s Data Fusion and Global ID Management.

Showcase Session

Marek Dudáš (KIZI/UEP) started the DBpedia Showcase Session with a presentation on “Concept Maps with the help of DBpedia”, where he showed the audience how to create a “concept map” with the ContextMinds application. Furthermore, Tomáš Kliegr (KIZI/UEP) presented “Explainable Machine Learning and Knowledge Graphs”. He explained his contribution to a rule-based classifier for business use cases. Two other showcases followed: Václav Zeman (KIZI/UEP), who presented “RdfRules: Rule Mining from DBpedia” and Denis Streitmatter (AKSW/KILT), who demonstrated the “DBpedia API”.

Miroslav Blasko presents “Ontology-based Dataset Exploration”

Closing this Session, Miroslav Blasko (CTU, Prague) gave a presentation on “Ontology-based Dataset Exploration”. He explained a taxonomy developed for dataset description. Additionally, he presented several use cases that have the main goal of improving content-based descriptors.

Summing up, the DBpedia meetup in Prague brought together more than 50 DBpedia enthusiasts from all over Europe. They engaged in vital discussions about Linked Data, the DBpedia databus, as well as DBpedia use cases and services.

 

 

 

In case you missed the event, all slides and presentations are available on our website. Further insights  feedback, and photos about the event can be found on Twitter via #DBpediaPrague.

We are currently looking forward to the next DBpedia Community Meeting, on May 23rd, 2019 in Leipzig, Germany. This meeting is co-located with the Language, Data and Knowledge (LDK) conference. Stay tuned and check Twitter, Facebook and the website or subscribe to our newsletter for the latest news and updates.

Your DBpedia Association

A year with DBpedia – Retrospective Part 3

This is the final part of our journey around the world with DBpedia. This time we will take you from Austria, to Mountain View, California and to London, UK.

Come on, let’s do this.

Welcome to Vienna, Austria  – Semantics

More than 110 DBpedia enthusiasts joined our Community Meeting in Vienna, on September 10th, 2018. The event was again co-located with SEMANTiCS, a very successful collaboration. Lucky us, we got hold of two brilliant Keynote speakers, to open our meeting. Javier David Fernández García, Vienna University of Economics, opened the meeting with his keynote Linked Open Data cloud – act now before it’s too late. He reflected on challenges towards arriving at a truly machine-readable and decentralized Web of Data. Javier reviewed the current state of affairs, highlighted key technical and non-technical challenges, and outlined potential solution strategies. The second keynote speaker was Mathieu d’Aquin, Professor of Informatics at the Insight Centre for Data Analytics at NUI Galway. Mathieu, who is specialized in data analytics, completed the meeting with his keynote Dealing with Open-Domain Data.

The 12th edition of the DBpedia Community Meeting also covered a special chapter session, chaired by Enno Meijers, from the Dutch DBpedia Language Chapter. The speakers presented the latest technical or organizational developments of their respective chapter. This session has mainly created an exchange platform for the different DBpedia chapters. For the first time, representatives of the European chapters discussed problems and challenges of DBpedia from their point of view. Furthermore, tools, applications, and projects were presented by each chapter’s representative.

In case you missed the event, a more detailed article can be found here. All slides and presentations are also available on our Website. Further insights, feedback, and photos about the event are available on Twitter via #DBpediaDay.

Welcome to Mountain View  – GSoC mentor summit

GSoC was a vital part of DBpedia’s endeavors in 2018. We had three very talented students that with the help of our great mentors made it to the finish line of the program. You can read about their projects and success story in a dedicated post here.

After a successful 3-month mentoring, two of our mentors had the opportunity to attend the annual Google Summer of Code mentor summit. Mariano Rico and Thiago Galery represented DBpedia at the event this year. They engaged in a vital discussion about this years program, about lessons learned, highlights and drawbacks they experienced during the summer. A special focus was put on how to engage potential GSoC students as early as possible to get as much commitment as possible. The ideas the two mentors brought back in their suitcases will help to improve DBpedia’s part of the program for 2019. And apparently, chocolate was a very big thing there ;).

In case you have a project idea for GSoC2019 or want to mentor a DBpedia project next year, just drop us a line via dbpedia@infai.org. Also, as we intend to participate in the upcoming edition, please spread the word amongst students, and especially female students,  that fancy spending their summer coding on a DBpedia project. Thank you.

 

Welcome to London, England – Connected Data London 2018

In early November, we were invited to Connected Data London again. After 2017 this great event seems to become a regular in our DBpedia schedule.

Executive Director of the DBpedia Association, Sebastian Hellmannparticipated as panel candidate in the discussion around “Building Knowledge Graphs in the Real World”. Together with speakers from Thomson Reuters, Zalando, and Textkernel, he discussed definitions of KG, best practices of how to build and use knowledge graphs as well as the recent hype about it.

Visitors of CNDL2018 had the chance to grab a copy of our brand new flyer and exchange with us about the DBpedia Databus. This event gave us the opportunity to already met early adopters of our databus  – a decentralized data publication, integration, and subscription platform. Thank you very much for that opportunity.

A year went by

2018 has gone by so fast and brought so much for DBpedia. The DBpedia Association got the chance to meet more of DBpedia’s language chapters, we developed the DBpedia Databus to an extent that it can finally be launched in spring 2019. DBpedia is a community project relying on people and with the DBpedia Databus, we create a platform that allows publishing and provides a networked data economy around it. So stay tuned for exciting news coming up next year. Until then we like to thank all DBpedia enthusiasts around the world for their research with DBpedia, and support and contributions to DBpedia. Kudos to you.

 

All that remains to say is have yourself a very merry Christmas and a dazzling New Year. May 2019 be peaceful, exciting and prosperous.

 

Yours – being in a cheerful and festive mood –

 

DBpedia Association

 

A year with DBpedia – Retrospective Part Two

Retrospective Part II. Welcome to the second part of our journey around the world with DBpedia. This time we are taking you to Greece, Germany, to Australia and finally France.

Let the travels begin.

Welcome to Thessaloniki, Greece & ESWC

DBpedians from the Portuguese Chapter presented their research results during ESWC 2018 in Thessaloniki, Greece.  the team around Diego Moussalem developed a demo to extend MAG  to support Entity Linking in 40 different languages. A special focus was put on low-resources languages such as Ukrainian, Greek, Hungarian, Croatian, Portuguese, Japanese and Korean. The demo relies on online web services which allow for an easy access to (their) entity linking approaches. Furthermore, it can disambiguate against DBpedia and Wikidata. Currently, MAG is used in diverse projects and has been used largely by the Semantic Web community. Check the demo via http://bit.ly/2RWgQ2M. Further information about the development can be found in a research paper, available here

 

Welcome back to Leipzig Germany

With our new credo “connecting data is about linking people and organizations”, halfway through 2018, we finalized our concept of the DBpedia Databus. This global DBpedia platform aims at sharing the efforts of OKG governance, collaboration, and curation to maximize societal value and develop a linked data economy.

With this new strategy, we wanted to meet some DBpedia enthusiasts of the German DBpedia Community. Fortunately, the LSWT (Leipzig Semantic Web Tag) 2018 hosted in Leipzig, home to the DBpedia Association proofed to be the right opportunity.  It was the perfect platform to exchange with researchers, industry and other organizations about current developments and future application of the DBpedia Databus. Apart from hosting a hands-on DBpedia workshop for newbies we also organized a well-received WebID -Tutorial. Finally,  the event gave us the opportunity to position the new DBpedia Databus as a global open knowledge network that aims at providing unified and global access to knowledge (graphs).

Welcome down under – Melbourne Australia

Further research results that rely on DBpedia were presented during ACL2018, in Melbourne, Australia, July 15th to 20th, 2018. The core of the research was DBpedia data, based on the WebNLG corpus, a challenge where participants automatically converted non-linguistic data from the Semantic Web into a textual format. Later on, the data was used to train a neural network model for generating referring expressions of a given entity. For example, if Jane Doe is a person’s official name, the referring expression of that person would be “Jane”, “Ms Doe”, “J. Doe”, or  “the blonde woman from the USA” etc.

If you want to dig deeper but missed ACL this year, the paper is available here.

 

Welcome to Lyon, France

In July the DBpedia Association travelled to France. With the organizational support of Thomas Riechert (HTWK, InfAI) and Inria, we finally met the French DBpedia Community in person and presented the DBpedia Databus. Additionally, we got to meet the French DBpedia Chapter, researchers and developers around Oscar Rodríguez Rocha and Catherine Faron Zucker.  They presented current research results revolving around an approach to automate the generation of educational quizzes from DBpedia. They wanted to provide a useful tool to be applied in the French educational system, that:

  • helps to test and evaluate the knowledge acquired by learners and…
  • supports lifelong learning on various topics or subjects. 

The French DBpedia team followed a 4-step approach:

  1. Quizzes are first formalized with Semantic Web standards: questions are represented as SPARQL queries and answers as RDF graphs.
  2. Natural language questions, answers and distractors are generated from this formalization.
  3. We defined different strategies to extract multiple choice questions, correct answers and distractors from DBpedia.
  4. We defined a measure of the information content of the elements of an ontology, and of the set of questions contained in a quiz.

Oscar R. Rocha and Catherine F. Zucker also published a paper explaining the detailed approach to automatically generate quizzes from DBpedia according to official French educational standards. 

 

 

Thank you to all DBpedia enthusiasts that we met during our journey. A big thanks to

With this journey from Europe to Australia and back we provided you with insights into research based on DBpedia as well as a glimpse into the French DBpedia Chapter. In our final part of the journey coming up next week, we will take you to Vienna,  San Francisco and London.  In the meantime, stay tuned and visit our Twitter channel or subscribe to our DBpedia Newsletter.

 

Have a great week.

Yours DBpedia Association