Discussion:
[Wikidata] Fwd: [Wikimedia-l] Wikipedia in an abstract language
Pine W
2018-09-29 18:42:57 UTC
Permalink
Forwarding because this (ambitious!) proposal may be of interest to people
on other lists. I'm not endorsing the proposal at this time, but I'm
curious about it.

Pine
( https://meta.wikimedia.org/wiki/User:Pine )


---------- Forwarded message ---------
From: Denny Vrandečić <***@gmail.com>
Date: Sat, Sep 29, 2018 at 6:32 PM
Subject: [Wikimedia-l] Wikipedia in an abstract language
To: Wikimedia Mailing List <wikimedia-***@lists.wikimedia.org>


Semantic Web languages allow to express ontologies and knowledge bases in a
way meant to be particularly amenable to the Web. Ontologies formalize the
shared understanding of a domain. But the most expressive and widespread
languages that we know of are human natural languages, and the largest
knowledge base we have is the wealth of text written in human languages.

We looks for a path to bridge the gap between knowledge representation
languages such as OWL and human natural languages such as English. We
propose a project to simultaneously expose that gap, allow to collaborate
on closing it, make progress widely visible, and is highly attractive and
valuable in its own right: a Wikipedia written in an abstract language to
be rendered into any natural language on request. This would make current
Wikipedia editors about 100x more productive, and increase the content of
Wikipedia by 10x. For billions of users this will unlock knowledge they
currently do not have access to.

My first talk on this topic will be on October 10, 2018, 16:45-17:00, at
the Asilomar in Monterey, CA during the Blue Sky track of ISWC. My second,
longer talk on the topic will be at the DL workshop in Tempe, AZ, October
27-29. Comments are very welcome as I prepare the slides and the talk.

Link to the paper: http://simia.net/download/abstractwikipedia.pdf

Cheers,
Denny
_______________________________________________
Wikimedia-l mailing list, guidelines at:
https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and
https://meta.wikimedia.org/wiki/Wikimedia-l
New messages to: Wikimedia-***@lists.wikimedia.org
Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/wikimedia-l,
<mailto:wikimedia-l-***@lists.wikimedia.org?subject=unsubscribe>
Luca Martinelli
2018-09-29 20:17:55 UTC
Permalink
Wow! Please share the slides or the video! I'm interested too.

L.
Post by Pine W
Forwarding because this (ambitious!) proposal may be of interest to people
on other lists. I'm not endorsing the proposal at this time, but I'm
curious about it.
Pine
( https://meta.wikimedia.org/wiki/User:Pine )
---------- Forwarded message ---------
Date: Sat, Sep 29, 2018 at 6:32 PM
Subject: [Wikimedia-l] Wikipedia in an abstract language
Semantic Web languages allow to express ontologies and knowledge bases in a
way meant to be particularly amenable to the Web. Ontologies formalize the
shared understanding of a domain. But the most expressive and widespread
languages that we know of are human natural languages, and the largest
knowledge base we have is the wealth of text written in human languages.
We looks for a path to bridge the gap between knowledge representation
languages such as OWL and human natural languages such as English. We
propose a project to simultaneously expose that gap, allow to collaborate
on closing it, make progress widely visible, and is highly attractive and
valuable in its own right: a Wikipedia written in an abstract language to
be rendered into any natural language on request. This would make current
Wikipedia editors about 100x more productive, and increase the content of
Wikipedia by 10x. For billions of users this will unlock knowledge they
currently do not have access to.
My first talk on this topic will be on October 10, 2018, 16:45-17:00, at
the Asilomar in Monterey, CA during the Blue Sky track of ISWC. My second,
longer talk on the topic will be at the DL workshop in Tempe, AZ, October
27-29. Comments are very welcome as I prepare the slides and the talk.
Link to the paper: http://simia.net/download/abstractwikipedia.pdf
Cheers,
Denny
_______________________________________________
https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and
https://meta.wikimedia.org/wiki/Wikimedia-l
Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/wikimedia-l,
_______________________________________________
Wikidata mailing list
https://lists.wikimedia.org/mailman/listinfo/wikidata
f***@imm.dtu.dk
2018-10-04 15:50:51 UTC
Permalink
Denny's project is a very interesting.

We already have Wikidata and Magnus Manske's autodesc which can create
paragraph-length natural language for some types of items.

Example:

https://tools.wmflabs.org/autodesc/?q=Q18618629&lang=&mode=long&links=text&redlinks=&format=html&get_infobox=yes&infobox_template=

"""Denny Vrandečić is a Croatia researcher, programmer, and computer
scientist.
He was born on February 27, 1978 in Stuttgart.
He studied at Karlsruhe Institute of Technology from October 2004 until
June 2010, University of Stuttgart from September 1998 until February
2004, University of Stuttgart from September 1997 until February 2004,
and Geschwister-Scholl-Gymnasium. He worked for Google from October
2013, for Wikimedia Deutschland from March 2012 until September 2013,
and for Karlsruhe Institute of Technology from 2004 until 2012."""

Currently he seems to support English, French and Dutch.

I think Magnus Manske would accept pull request to other languages at
https://bitbucket.org/magnusmanske/reasonator/src/9c58fadb7b72a791142fc158aebe38d9a4b98d92/public_html/auto_long_desc.js?at=master&fileviewer=file-view-default

So how would we go beyond Magnus? Would the Wikidata representation
suffice? I have seen Q50827579 and Q28819478 for Wikidata to language
generation, but I am not aware of running applications and are they
better than Magnus' hard-coded approach?

I have been experimenting a bit the other way. Ordia can go from natural
Post by Pine W
from ordia.base import Base
base = Base()
base.words_to_form_ids('der kom en soldat marcherende henad
landevejen'.split(), language='da')
[['L3064-F1'], ['L3065-F3', 'L3065-F6'], ['L2022-F1', 'L3073-F3'],
['L3074-F1'], ['L3075-F5'], ['L3215-F1'], ['L3216-F2']]

Writing the encyclopedic text in "Wikidata-lexemesh" could perhaps ease
translation, particularly after 18 October when senses are planned to be
enabled.

/Finn
Post by Pine W
Forwarding because this (ambitious!) proposal may be of interest to
people on other lists. I'm not endorsing the proposal at this time, but
I'm curious about it.
Pine
( https://meta.wikimedia.org/wiki/User:Pine )
---------- Forwarded message ---------
Date: Sat, Sep 29, 2018 at 6:32 PM
Subject: [Wikimedia-l] Wikipedia in an abstract language
Semantic Web languages allow to express ontologies and knowledge bases in a
way meant to be particularly amenable to the Web. Ontologies formalize the
shared understanding of a domain. But the most expressive and widespread
languages that we know of are human natural languages, and the largest
knowledge base we have is the wealth of text written in human languages.
We looks for a path to bridge the gap between knowledge representation
languages such as OWL and human natural languages such as English. We
propose a project to simultaneously expose that gap, allow to collaborate
on closing it, make progress widely visible, and is highly attractive and
valuable in its own right: a Wikipedia written in an abstract language to
be rendered into any natural language on request. This would make current
Wikipedia editors about 100x more productive, and increase the content of
Wikipedia by 10x. For billions of users this will unlock knowledge they
currently do not have access to.
My first talk on this topic will be on October 10, 2018, 16:45-17:00, at
the Asilomar in Monterey, CA during the Blue Sky track of ISWC. My second,
longer talk on the topic will be at the DL workshop in Tempe, AZ, October
27-29. Comments are very welcome as I prepare the slides and the talk.
Link to the paper: http://simia.net/download/abstractwikipedia.pdf
Cheers,
Denny
_______________________________________________
https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and
https://meta.wikimedia.org/wiki/Wikimedia-l
Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/wikimedia-l,
_______________________________________________
Wikidata mailing list
https://lists.wikimedia.org/mailman/listinfo/wikidata
Loading...