New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Schema.org should have mappings to Wikidata terms where possible #280

Open
danbri opened this Issue Jan 23, 2015 · 186 comments

Comments

@danbri
Contributor

danbri commented Jan 23, 2015

From Lydia Pintscher in https://twitter.com/nightrose/status/558549091844886528

@danbri any issue to track progress on http://schema.org  mapping to Wikidata? 
Maybe even get people to help out?

Update 2016-01-26 - since the original post there have been some improvements at both Wikidata and Schema.org:

  • Wikidata: mappings (exact, super/sub) from properties and (perhaps to a lesser extent in that the notion isn't so built-in) types to schema.org can be expressed within Wikidata.
  • Wikidata now has a SPARQL endpoint at https://query.wikidata.org which is the most natural way of retrieving data; other explorations such as JSON dumps below are less important now.
  • Schema.org has updated its extension mechanism and is encouraging both hosted and external extensions.
  • D3-compatible RDFS JSON-LD is published from schema.org and can be used for visualization; this would also be a good model for getting an overview of Wikidata. See http://bl.ocks.org/danbri/1c121ea8bd2189cf411c for example visualization.
  • Various notes towards using Wikidata as an extension language for Schema.org are explored towards the end of this issue, as are SPARQL queries for extracting Wikidata's structure and property metadata for use in mappings.

 Nearby

@danbri

This comment has been minimized.

Show comment
Hide comment
@danbri

danbri Jan 23, 2015

Contributor

Notes from IRC,

Contributor

danbri commented Jan 23, 2015

Notes from IRC,

@lydiapintscher

This comment has been minimized.

Show comment
Hide comment
@lydiapintscher

lydiapintscher Jan 23, 2015

Here is how mapping can be done on the Wikidata side for example: https://www.wikidata.org/wiki/Property:P31

The JSON dumps are the best dumps.

lydiapintscher commented Jan 23, 2015

Here is how mapping can be done on the Wikidata side for example: https://www.wikidata.org/wiki/Property:P31

The JSON dumps are the best dumps.

@innovimax

This comment has been minimized.

Show comment
Hide comment
@innovimax

innovimax commented Jan 24, 2015

+1

@elf-pavlik

This comment has been minimized.

Show comment
Hide comment
@elf-pavlik

elf-pavlik Jan 25, 2015

Contributor

happy to help here a little! I had chance to meet few people from Wikidata crew during 31C3 and remember that serving turtle also needs some fixing... but it already uses schema.org quite a lot!

$ curl http://www.wikidata.org/entity/Q80 -iL -H "Accept: text/turtle"
Contributor

elf-pavlik commented Jan 25, 2015

happy to help here a little! I had chance to meet few people from Wikidata crew during 31C3 and remember that serving turtle also needs some fixing... but it already uses schema.org quite a lot!

$ curl http://www.wikidata.org/entity/Q80 -iL -H "Accept: text/turtle"
@danbri

This comment has been minimized.

Show comment
Hide comment
@danbri

danbri Jan 25, 2015

Contributor

I went looking for the code that generates this. For those without turtle, an excerpt from running

curl http://www.wikidata.org/entity/Q42 -iL -H "Accept: text/turtle"

(full response is at https://gist.github.com/danbri/66616096d42e595376f6 )

[update]Hmm actually you can get it all in the browser without using content negotiation, just via suffixes:

( edit! I have moved a big chunk of text to https://gist.github.com/danbri/181ff7763f479c397e10 - apologies to those who got accidental notifications due to the '@' symbol.)

This is great but also unfortunately "the easy part" in that these are fixed built-in properties that each Wikidata entry will always carry.

Looking around for relevant source code,

It would be interesting to see how addEntityMetaData might be amended to exploit equivalentProperty information in Wikidata, at @lydiapintscher mentioned re https://www.wikidata.org/wiki/Property:P31

Contributor

danbri commented Jan 25, 2015

I went looking for the code that generates this. For those without turtle, an excerpt from running

curl http://www.wikidata.org/entity/Q42 -iL -H "Accept: text/turtle"

(full response is at https://gist.github.com/danbri/66616096d42e595376f6 )

[update]Hmm actually you can get it all in the browser without using content negotiation, just via suffixes:

( edit! I have moved a big chunk of text to https://gist.github.com/danbri/181ff7763f479c397e10 - apologies to those who got accidental notifications due to the '@' symbol.)

This is great but also unfortunately "the easy part" in that these are fixed built-in properties that each Wikidata entry will always carry.

Looking around for relevant source code,

It would be interesting to see how addEntityMetaData might be amended to exploit equivalentProperty information in Wikidata, at @lydiapintscher mentioned re https://www.wikidata.org/wiki/Property:P31

@ppKrauss

This comment has been minimized.

Show comment
Hide comment
@ppKrauss

ppKrauss Feb 19, 2015

I agree, "Schema.org should have mappings to Wikidata terms where possible". How to vote? or how to colaborate and/or check work in progress? There are a link about work in this issue?

ppKrauss commented Feb 19, 2015

I agree, "Schema.org should have mappings to Wikidata terms where possible". How to vote? or how to colaborate and/or check work in progress? There are a link about work in this issue?

@elf-pavlik

This comment has been minimized.

Show comment
Hide comment
@elf-pavlik

elf-pavlik Feb 22, 2015

Contributor

@danbri please remember to fence code snippets with three backticks which can also include clue for syntax highlighting

```ttl
  code goes here @bg @dr @mr
  @prefix data: http://www.wikidata.org/wiki/Special:EntityData/ .
  @prefix schema: http://schema.org/ .
  no mentions using @foo

```

also see code tab in Examples of github markdown https://guides.github.com/features/mastering-markdown/#examples

Contributor

elf-pavlik commented Feb 22, 2015

@danbri please remember to fence code snippets with three backticks which can also include clue for syntax highlighting

```ttl
  code goes here @bg @dr @mr
  @prefix data: http://www.wikidata.org/wiki/Special:EntityData/ .
  @prefix schema: http://schema.org/ .
  no mentions using @foo

```

also see code tab in Examples of github markdown https://guides.github.com/features/mastering-markdown/#examples

@elf-pavlik

This comment has been minimized.

Show comment
Hide comment
@elf-pavlik

elf-pavlik Feb 22, 2015

Contributor

@ppKrauss I think people would appreciate more machine readable mappings using owl:equivalentProperty etc.
e.g.

<link property="owl:equivalentProperty" href="http://purl.org/dc/terms/description"/>

IMO we could consider everything from subset of OWL used by RDFa Vocabulary Entailment
http://www.w3.org/TR/rdfa-syntax/#s_vocab_expansion

Contributor

elf-pavlik commented Feb 22, 2015

@ppKrauss I think people would appreciate more machine readable mappings using owl:equivalentProperty etc.
e.g.

<link property="owl:equivalentProperty" href="http://purl.org/dc/terms/description"/>

IMO we could consider everything from subset of OWL used by RDFa Vocabulary Entailment
http://www.w3.org/TR/rdfa-syntax/#s_vocab_expansion

@ppKrauss

This comment has been minimized.

Show comment
Hide comment
@ppKrauss

ppKrauss Feb 22, 2015

@elf-pavlik thanks (!), so the issue now is only to add something as
<link property="owl:equivalentProperty" href="http://WikiDataURL"/>
in each rdf:Property and each rdfs:Class ... is it?

New suggestion: we may colaborate with an online interface or (initially) by a spreadsheet (ex. Excel) at github, with the columns wikidataID and Property or wikidataID and Class.

ppKrauss commented Feb 22, 2015

@elf-pavlik thanks (!), so the issue now is only to add something as
<link property="owl:equivalentProperty" href="http://WikiDataURL"/>
in each rdf:Property and each rdfs:Class ... is it?

New suggestion: we may colaborate with an online interface or (initially) by a spreadsheet (ex. Excel) at github, with the columns wikidataID and Property or wikidataID and Class.

@lydiapintscher

This comment has been minimized.

Show comment
Hide comment
@lydiapintscher

lydiapintscher Feb 22, 2015

Why not add it directly in Wikidata?

lydiapintscher commented Feb 22, 2015

Why not add it directly in Wikidata?

@ppKrauss

This comment has been minimized.

Show comment
Hide comment
@ppKrauss

ppKrauss Feb 22, 2015

@lydiapintscher , perhaps I am not understanding your point, sorry... The objetive in this issue is to map the Schema.org's definitions into the Wikidata.org's concept-definitions, not the inverse.

ppKrauss commented Feb 22, 2015

@lydiapintscher , perhaps I am not understanding your point, sorry... The objetive in this issue is to map the Schema.org's definitions into the Wikidata.org's concept-definitions, not the inverse.

@lydiapintscher

This comment has been minimized.

Show comment
Hide comment
@lydiapintscher

lydiapintscher Feb 22, 2015

Both should happen, no? ;-)

lydiapintscher commented Feb 22, 2015

Both should happen, no? ;-)

@ppKrauss

This comment has been minimized.

Show comment
Hide comment
@ppKrauss

ppKrauss Feb 22, 2015

@lydiapintscher , I think it is a matter of scope. You can imagine Wikidata as an (external and closed) didictionary, like Webster, not like an open project like Wiipedia.

ppKrauss commented Feb 22, 2015

@lydiapintscher , I think it is a matter of scope. You can imagine Wikidata as an (external and closed) didictionary, like Webster, not like an open project like Wiipedia.

@lydiapintscher

This comment has been minimized.

Show comment
Hide comment
@lydiapintscher

lydiapintscher Feb 22, 2015

Wikidata is just as open as Wikipedia.

lydiapintscher commented Feb 22, 2015

Wikidata is just as open as Wikipedia.

@nemobis

This comment has been minimized.

Show comment
Hide comment
@nemobis

nemobis Feb 22, 2015

Peter, 22/02/2015 18:39:

wikipedia.org concept definitions

Does such a thing exist?

nemobis commented Feb 22, 2015

Peter, 22/02/2015 18:39:

wikipedia.org concept definitions

Does such a thing exist?

@elf-pavlik

This comment has been minimized.

Show comment
Hide comment
@elf-pavlik

elf-pavlik Feb 23, 2015

Contributor

@lydiapintscher once schema.org URIs have mappings to wikidata URIs added, do you see a way to add them to wikidata in programmable way? IMO it doesn't make sense to do it manualy via web UI... maybe wikidata team could just import them from schema.rdfa?

BTW I'll stay most of march ~Berlin and could meet IRL with you and anyone else from wikidata interested in this issue... Whenever in Berlin I go anyways to #OKLab / CodeForBerlin on every monday evening at Wikimedia HQ 😄 (we can discuss details over pm - just see my gh profile)

Contributor

elf-pavlik commented Feb 23, 2015

@lydiapintscher once schema.org URIs have mappings to wikidata URIs added, do you see a way to add them to wikidata in programmable way? IMO it doesn't make sense to do it manualy via web UI... maybe wikidata team could just import them from schema.rdfa?

BTW I'll stay most of march ~Berlin and could meet IRL with you and anyone else from wikidata interested in this issue... Whenever in Berlin I go anyways to #OKLab / CodeForBerlin on every monday evening at Wikimedia HQ 😄 (we can discuss details over pm - just see my gh profile)

@ppKrauss

This comment has been minimized.

Show comment
Hide comment
@ppKrauss

ppKrauss Feb 23, 2015

I am trying (with bad English) to consolidate this issue in a draft of the proposal, can you help?

A next step will be to create a Readme.md for everybody edit this text, perhaps with the #352 mechanism, and (phase1) implement "by hand" some examples in schema.rdfa.


Foundations collected from comments posted in this discussion:

  1. @danbri and Lydia Pintscher summary, "schema.org mapping to Wikidata".
  2. Techinal suggestion to "schema.org property marked as equiv to another: schema:description ", @danbri.
  3. @danbri and @elf-pavlik looking for some automation ... or "how addEntityMetaData might be amended to exploit equivalentProperty information in Wikidata".
  4. ...
  5. @elf-pavlik suggestion to add the tag <link property="owl:equivalentProperty" href="http://WikiDataURL"/>, into each rdfs:Class and each rdf:Property resource definitions.
    The equivalentProperty is the same as showed in the Property:P31 example) of @lydiapintscher.
  6. Proposal of @ppKrauss to start at Schema.org and with human work, with no automation (for test and start).
  7. Suggestion of @lydiapintscher for think also about Wikidata mapping to Schema.org...

PROPOSAL OF THE ISSUE #280

Proposal for enhance schema.rdfa definition descriptors (rdfs:comment) and semantics, mapping each vocabulary item to a Wikidata item.

A sibling project at Wikidata will be the Wikidata.org-to-Schema.org mapping.

PART 1 - SchemaOrg mapping to Wikidata

Actions: add <link property="{$OWL}" href="{$WikiDataURL}"/> with the correct $WikiDataURL.

  • At each rdfs:Class add the <link> tag with $OWL="owl:equivalentClass" or, when not possible, use$OWL="rdfs:subClassOf".
  • At each rdf:Property add the <link> tag with $OWL="owl:equivalentProperty" or, when not possible, use$OWL="rdfs:subPropertyOf".

Actions on testing phase: do some with no automation. Example: start with classes Person and Organization, and its properties.

Examples


PART 2 - Wikidata mapping to SchemaOrg

... under construction... see similar mappings at schema.rdfs.org/mappings.html... Wikidata also have a lot of iniciatives maping Wikidata to external vocabularies (ex. there are a map from Wikidata to BNCF Thesaurus)...

ppKrauss commented Feb 23, 2015

I am trying (with bad English) to consolidate this issue in a draft of the proposal, can you help?

A next step will be to create a Readme.md for everybody edit this text, perhaps with the #352 mechanism, and (phase1) implement "by hand" some examples in schema.rdfa.


Foundations collected from comments posted in this discussion:

  1. @danbri and Lydia Pintscher summary, "schema.org mapping to Wikidata".
  2. Techinal suggestion to "schema.org property marked as equiv to another: schema:description ", @danbri.
  3. @danbri and @elf-pavlik looking for some automation ... or "how addEntityMetaData might be amended to exploit equivalentProperty information in Wikidata".
  4. ...
  5. @elf-pavlik suggestion to add the tag <link property="owl:equivalentProperty" href="http://WikiDataURL"/>, into each rdfs:Class and each rdf:Property resource definitions.
    The equivalentProperty is the same as showed in the Property:P31 example) of @lydiapintscher.
  6. Proposal of @ppKrauss to start at Schema.org and with human work, with no automation (for test and start).
  7. Suggestion of @lydiapintscher for think also about Wikidata mapping to Schema.org...

PROPOSAL OF THE ISSUE #280

Proposal for enhance schema.rdfa definition descriptors (rdfs:comment) and semantics, mapping each vocabulary item to a Wikidata item.

A sibling project at Wikidata will be the Wikidata.org-to-Schema.org mapping.

PART 1 - SchemaOrg mapping to Wikidata

Actions: add <link property="{$OWL}" href="{$WikiDataURL}"/> with the correct $WikiDataURL.

  • At each rdfs:Class add the <link> tag with $OWL="owl:equivalentClass" or, when not possible, use$OWL="rdfs:subClassOf".
  • At each rdf:Property add the <link> tag with $OWL="owl:equivalentProperty" or, when not possible, use$OWL="rdfs:subPropertyOf".

Actions on testing phase: do some with no automation. Example: start with classes Person and Organization, and its properties.

Examples


PART 2 - Wikidata mapping to SchemaOrg

... under construction... see similar mappings at schema.rdfs.org/mappings.html... Wikidata also have a lot of iniciatives maping Wikidata to external vocabularies (ex. there are a map from Wikidata to BNCF Thesaurus)...

@ppKrauss

This comment has been minimized.

Show comment
Hide comment
@ppKrauss

ppKrauss Feb 23, 2015

@lydiapintscher , Sorry again... I not saw that there are also a proposal of "sibling project at Wikidata" (!)... Can you please check if my "draft of this proposal" text is now on the rails? I am trying to "translate" and consolidate all comments in one document... To start all with the same scope, objective, etc.

ppKrauss commented Feb 23, 2015

@lydiapintscher , Sorry again... I not saw that there are also a proposal of "sibling project at Wikidata" (!)... Can you please check if my "draft of this proposal" text is now on the rails? I am trying to "translate" and consolidate all comments in one document... To start all with the same scope, objective, etc.

@ppKrauss

This comment has been minimized.

Show comment
Hide comment
@ppKrauss

ppKrauss Feb 23, 2015

@danbri , @elf-pavlik , and others, I not understand if there are a "formal procedure for create proposals" here...

Can you please check if my "draft of this proposal" text is now on the rails? I need your help to "translate" and consolidate it.


About automation, I still do not understand well, you want to automate?
My opinion. I think we can start with non-automated procedures, that will be util to check automated ones, which happen to be introduced later... Or to check the "size" of the non-automated task (~1000 items!). I think that a reliable mapping needs human control.

ppKrauss commented Feb 23, 2015

@danbri , @elf-pavlik , and others, I not understand if there are a "formal procedure for create proposals" here...

Can you please check if my "draft of this proposal" text is now on the rails? I need your help to "translate" and consolidate it.


About automation, I still do not understand well, you want to automate?
My opinion. I think we can start with non-automated procedures, that will be util to check automated ones, which happen to be introduced later... Or to check the "size" of the non-automated task (~1000 items!). I think that a reliable mapping needs human control.

@elf-pavlik

This comment has been minimized.

Show comment
Hide comment
@elf-pavlik

elf-pavlik Feb 23, 2015

Contributor

@ppKrauss thanks for trying to summarize this thread into a proposal!

http://schema.org/Organization is owl:equivalentProperty to Q43229

please don't confuse owl:equivalentClass with owl:equivalentProperty

if you look at schema.rdf we need accordingly

  • typeof="rdfs:Class" needs owl:equivalentClass or rdfs:subClassOf
  • typeof="rdf:Property" needs owl:equivalentProperty or rdfs:subPropertyOf

for the automation, once we map one way schema.org -> wikidata (however we manage to do it) then we can automate importing most of that mapping into wikidata so no one needs to click and copy&paste...

Last but not least, schema.org just starts using github recently and also seems to go through various other processes, I would encourage you to stay patient and give people time to reply 😄

Contributor

elf-pavlik commented Feb 23, 2015

@ppKrauss thanks for trying to summarize this thread into a proposal!

http://schema.org/Organization is owl:equivalentProperty to Q43229

please don't confuse owl:equivalentClass with owl:equivalentProperty

if you look at schema.rdf we need accordingly

  • typeof="rdfs:Class" needs owl:equivalentClass or rdfs:subClassOf
  • typeof="rdf:Property" needs owl:equivalentProperty or rdfs:subPropertyOf

for the automation, once we map one way schema.org -> wikidata (however we manage to do it) then we can automate importing most of that mapping into wikidata so no one needs to click and copy&paste...

Last but not least, schema.org just starts using github recently and also seems to go through various other processes, I would encourage you to stay patient and give people time to reply 😄

@danbri

This comment has been minimized.

Show comment
Hide comment
@danbri

danbri Feb 23, 2015

Contributor

Thanks all. Indeed I'm on a trip and can't currently give this the
attention it deserves, but I would try to nudge the focus towards actual
mappings and away from the specific implementation details at schema.org.
We will be making some changes in the site tooling to support mechanisms
for extension that may be relevant here.

How about we just jump into the details and start a spreadsheet with a
table of schema.org types and properties? Eg on google docs...?

On Mon, 23 Feb 2015 09:06 ☮ elf Pavlik ☮ notifications@github.com wrote:

@ppKrauss https://github.com/ppKrauss thanks for trying to summarize
this thread into a proposal!

http://schema.org/Organization is owl:equivalentProperty to Q43229

please don't confuse owl:equivalentClass with owl:equivalentProperty

if you look at schema.rdf
https://github.com/schemaorg/schemaorg/blob/sdo-gozer/data/schema.rdfa
we need accordingly

  • typeof="rdfs:Class" needs owl:equivalentClass or
    rdfs:subClassOf
  • typeof="rdf:Property" needs owl:equivalentProperty or
    rdfs:subPropertyOf

for the automation, once we map one way schema.org -> wikidata (however
we manage to do it) then we can automate importing most of that mapping
into wikidata so no one needs to click and copy&paste...

Last but not least, schema.org just starts using github recently and also
seems to go through various other processes, I would encourage you to stay
patient and give people time to reply [image: 😄]


Reply to this email directly or view it on GitHub
#280 (comment).

Contributor

danbri commented Feb 23, 2015

Thanks all. Indeed I'm on a trip and can't currently give this the
attention it deserves, but I would try to nudge the focus towards actual
mappings and away from the specific implementation details at schema.org.
We will be making some changes in the site tooling to support mechanisms
for extension that may be relevant here.

How about we just jump into the details and start a spreadsheet with a
table of schema.org types and properties? Eg on google docs...?

On Mon, 23 Feb 2015 09:06 ☮ elf Pavlik ☮ notifications@github.com wrote:

@ppKrauss https://github.com/ppKrauss thanks for trying to summarize
this thread into a proposal!

http://schema.org/Organization is owl:equivalentProperty to Q43229

please don't confuse owl:equivalentClass with owl:equivalentProperty

if you look at schema.rdf
https://github.com/schemaorg/schemaorg/blob/sdo-gozer/data/schema.rdfa
we need accordingly

  • typeof="rdfs:Class" needs owl:equivalentClass or
    rdfs:subClassOf
  • typeof="rdf:Property" needs owl:equivalentProperty or
    rdfs:subPropertyOf

for the automation, once we map one way schema.org -> wikidata (however
we manage to do it) then we can automate importing most of that mapping
into wikidata so no one needs to click and copy&paste...

Last but not least, schema.org just starts using github recently and also
seems to go through various other processes, I would encourage you to stay
patient and give people time to reply [image: 😄]


Reply to this email directly or view it on GitHub
#280 (comment).

@ppKrauss

This comment has been minimized.

Show comment
Hide comment
@ppKrauss

ppKrauss Feb 24, 2015

@elf-pavlik thanks (!), I edited with your correction (and now coping also to my issue280 "ahead of work" :-)


@danbri Ok I send to to this googleDoc and updated my #352 with the tool that generates the spreadsheet.


@elf-pavlik and @danbri , no urgence (!). As a novice here, I am experimenting/testing the collaboration possibilities, and studing schemaOrg as a project ... Now I have a better "schema.org big picture", I see a good work(!), by moderators and vibrant community. My only help/clue about "better Github use" is at #352, and perhaps still a little messy.

Returning to talk about the spreadsheet, there are ~1500 items (!)... A good starting point is the classes Person and Organization, the "vCard semantic" is the more used in the Web,

http://webdatacommons.org/structureddata/index.html#toc2

so, I am starting to work with them (Person and Organization)... It is ok, good starting point?

ppKrauss commented Feb 24, 2015

@elf-pavlik thanks (!), I edited with your correction (and now coping also to my issue280 "ahead of work" :-)


@danbri Ok I send to to this googleDoc and updated my #352 with the tool that generates the spreadsheet.


@elf-pavlik and @danbri , no urgence (!). As a novice here, I am experimenting/testing the collaboration possibilities, and studing schemaOrg as a project ... Now I have a better "schema.org big picture", I see a good work(!), by moderators and vibrant community. My only help/clue about "better Github use" is at #352, and perhaps still a little messy.

Returning to talk about the spreadsheet, there are ~1500 items (!)... A good starting point is the classes Person and Organization, the "vCard semantic" is the more used in the Web,

http://webdatacommons.org/structureddata/index.html#toc2

so, I am starting to work with them (Person and Organization)... It is ok, good starting point?

@danbri

This comment has been minimized.

Show comment
Hide comment
@danbri

danbri Feb 24, 2015

Contributor

Thanks. Yes starting with the more most general / common types makes sense.

Where I got stuck: I could not figure out a good programmatic way to access
Wikidata's schema information in all its richness.

Maybe there is a way to take the JSON dumps, load them into some
fast-access NoSQL-ish database, so that things can be
searched/matched/retrieved easily?

nearby: https://gist.github.com/chrpr/23926c4650ce4363c51b dumps DBpedia's vocab (not Wikidata, but worth a look for comparison)

Contributor

danbri commented Feb 24, 2015

Thanks. Yes starting with the more most general / common types makes sense.

Where I got stuck: I could not figure out a good programmatic way to access
Wikidata's schema information in all its richness.

Maybe there is a way to take the JSON dumps, load them into some
fast-access NoSQL-ish database, so that things can be
searched/matched/retrieved easily?

nearby: https://gist.github.com/chrpr/23926c4650ce4363c51b dumps DBpedia's vocab (not Wikidata, but worth a look for comparison)

@jimkont

This comment has been minimized.

Show comment
Hide comment
@jimkont

jimkont Feb 24, 2015

Wikidata provides RDF dumps here: http://tools.wmflabs.org/wikidata-exports/rdf/exports/20150126/

It is easy to get the classes from the wikidata-taxonomy dump but needs to be joined with the wikidata-terms dump to get the labels. For properties you can use the wikidata-properties dump

If you want something more fine-grained you can try the WKDT toolkit
https://github.com/Wikidata/Wikidata-Toolkit

Or create a DBpedia extractor, we have experimental support for wikidata in this branch:
https://github.com/alismayilov/extraction-framework/tree/wikidataAllCommits

RDF dumps can be directly loaded in a SPARQL endpoint or easily manipulated in CLI/code and load in any store.

jimkont commented Feb 24, 2015

Wikidata provides RDF dumps here: http://tools.wmflabs.org/wikidata-exports/rdf/exports/20150126/

It is easy to get the classes from the wikidata-taxonomy dump but needs to be joined with the wikidata-terms dump to get the labels. For properties you can use the wikidata-properties dump

If you want something more fine-grained you can try the WKDT toolkit
https://github.com/Wikidata/Wikidata-Toolkit

Or create a DBpedia extractor, we have experimental support for wikidata in this branch:
https://github.com/alismayilov/extraction-framework/tree/wikidataAllCommits

RDF dumps can be directly loaded in a SPARQL endpoint or easily manipulated in CLI/code and load in any store.

@nitmws

This comment has been minimized.

Show comment
Hide comment
@nitmws

nitmws Apr 10, 2018

Re @VladimirAlexiev

  • IPTC manages vocabularies since the mid-1990ies and this results in a tradition of some terms ...
  • ... therefore an IPTC KOS was created to cover all the terms used since two decades - http://dev.iptc.org/IKOS- we know a few overlap with Dublin Core.
  • datatype of timestamps: ok, will be added
  • WD URLs: you are right. But dealing with WD URLs is not easy: the data URL uses https, the concept URI uses http. Ok, needs to be corrected.
  • A generic rule applies to all 200 IPTC vocabularies: free text in English uses British English - en-GB.

Cheers!

nitmws commented Apr 10, 2018

Re @VladimirAlexiev

  • IPTC manages vocabularies since the mid-1990ies and this results in a tradition of some terms ...
  • ... therefore an IPTC KOS was created to cover all the terms used since two decades - http://dev.iptc.org/IKOS- we know a few overlap with Dublin Core.
  • datatype of timestamps: ok, will be added
  • WD URLs: you are right. But dealing with WD URLs is not easy: the data URL uses https, the concept URI uses http. Ok, needs to be corrected.
  • A generic rule applies to all 200 IPTC vocabularies: free text in English uses British English - en-GB.

Cheers!

@nitmws

This comment has been minimized.

Show comment
Hide comment
@nitmws

nitmws Apr 10, 2018

As changes of definitions and translations were pending the Media Topics scheme was updated today: http://cv.iptc.org/newscodes/mediatopic/?format=rdfttl

  • Wikidata URIs corrected
  • ^^xsd:dateTime added
  • for created and modified additional dct properties exist

nitmws commented Apr 10, 2018

As changes of definitions and translations were pending the Media Topics scheme was updated today: http://cv.iptc.org/newscodes/mediatopic/?format=rdfttl

  • Wikidata URIs corrected
  • ^^xsd:dateTime added
  • for created and modified additional dct properties exist
@danbri

This comment has been minimized.

Show comment
Hide comment
@danbri

danbri Apr 10, 2018

Contributor

@thadguidry - maybe IPTC could publish the mappings you're working on?

Contributor

danbri commented Apr 10, 2018

@thadguidry - maybe IPTC could publish the mappings you're working on?

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Apr 10, 2018

@danbri sure IPTC could. but my need and others, is to have the mappings in Wikidata. There's a growing need to store the actual links for LinkedData in Wikidata.

and I'm testing OpenRefine's new Reconciling against Wikidata at the same time :)

thadguidry commented Apr 10, 2018

@danbri sure IPTC could. but my need and others, is to have the mappings in Wikidata. There's a growing need to store the actual links for LinkedData in Wikidata.

and I'm testing OpenRefine's new Reconciling against Wikidata at the same time :)

@danbri

This comment has been minimized.

Show comment
Hide comment
@danbri

danbri Jun 14, 2018

Contributor

@thadguidry et al., what's the status of these mappings? what do we have where? is there a handy query.wikidata.org sparql query to pull them out?

Contributor

danbri commented Jun 14, 2018

@thadguidry et al., what's the status of these mappings? what do we have where? is there a handy query.wikidata.org sparql query to pull them out?

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Jun 15, 2018

@danbri I'll give you a choice... do you want 5G this year ? or do you want me to complete the mappings up to the latest release forthcoming ? :) Truthfully, Ericsson (day job) is putting a bit of load on me at the moment, but I'm waiting on getting the bugs worked out in OpenRefine's Wikidata reconciling with Antonin, where we just released 3.0 Beta .. .so that we can get the wider community more involved here with mappings as well using OpenRefine 3.0. Probably, this month-ish ?

Here's our quick mapping query http://tinyurl.com/y8cqh2vj

As you will see in the query output, I began to map a few of the meta classes as well, but don't have any snazzy viz to showcase for all this yet. But here's an adhoc viz hack you can toy with: http://wikidata.rawgraphs.io/?url=https://query.wikidata.org/sparql?query=CONSTRUCT%20%7B%20%3Fs%20%3Chttp%3A%2F%2Fwww.w3.org%2F2002%2F07%2Fowl%23equivalentProperty%3E%20%3Fo.%20%7D%0AWHERE%20%7B%0A%20%20SELECT%20%3Fs%20%3FequivalentProperty%20%3Fo%20WHERE%20%7B%0A%20%20%20%20%3Fs%20wdt%3AP1628%20%3Fo.%0A%20%20%20%20BIND(wdt%3AP1628%20AS%20%3FequivalentProperty)%0A%20%20%20%20FILTER(REGEX(STR(%3Fo)%2C%20%22schema.org%22))%0A%20%20%7D%0A%7D%0A

and that adhoc viz just shows we have multiple mappings for "start time", the larger box for https://www.wikidata.org/wiki/Property:P580

thadguidry commented Jun 15, 2018

@danbri I'll give you a choice... do you want 5G this year ? or do you want me to complete the mappings up to the latest release forthcoming ? :) Truthfully, Ericsson (day job) is putting a bit of load on me at the moment, but I'm waiting on getting the bugs worked out in OpenRefine's Wikidata reconciling with Antonin, where we just released 3.0 Beta .. .so that we can get the wider community more involved here with mappings as well using OpenRefine 3.0. Probably, this month-ish ?

Here's our quick mapping query http://tinyurl.com/y8cqh2vj

As you will see in the query output, I began to map a few of the meta classes as well, but don't have any snazzy viz to showcase for all this yet. But here's an adhoc viz hack you can toy with: http://wikidata.rawgraphs.io/?url=https://query.wikidata.org/sparql?query=CONSTRUCT%20%7B%20%3Fs%20%3Chttp%3A%2F%2Fwww.w3.org%2F2002%2F07%2Fowl%23equivalentProperty%3E%20%3Fo.%20%7D%0AWHERE%20%7B%0A%20%20SELECT%20%3Fs%20%3FequivalentProperty%20%3Fo%20WHERE%20%7B%0A%20%20%20%20%3Fs%20wdt%3AP1628%20%3Fo.%0A%20%20%20%20BIND(wdt%3AP1628%20AS%20%3FequivalentProperty)%0A%20%20%20%20FILTER(REGEX(STR(%3Fo)%2C%20%22schema.org%22))%0A%20%20%7D%0A%7D%0A

and that adhoc viz just shows we have multiple mappings for "start time", the larger box for https://www.wikidata.org/wiki/Property:P580

@ppKrauss

This comment has been minimized.

Show comment
Hide comment
@ppKrauss

ppKrauss Jun 15, 2018

Hi @danbri, @thadguidry et al. there are some consensus about the "mapping task" at Wikidata, we could join our efforts... But the production of results was growing slowly,

There are people interested in go faster?

PS: my suggestion is to restart it as a formal project, with a little team, with ambitious goals ...

ppKrauss commented Jun 15, 2018

Hi @danbri, @thadguidry et al. there are some consensus about the "mapping task" at Wikidata, we could join our efforts... But the production of results was growing slowly,

There are people interested in go faster?

PS: my suggestion is to restart it as a formal project, with a little team, with ambitious goals ...

@jaygray0919

This comment has been minimized.

Show comment
Hide comment
@jaygray0919

jaygray0919 Jun 15, 2018

Thank you Thad @thadguidry One could make a case for @ppKrauss Peter's advice - the sprint concept. An alternative might be a more formal divide-and-conquer approach. Thad has developed a process and is using tools. Perhaps Thad could document that process and how to use tools (OpenRefine is well documented so a light overview may be all that is needed). If we divided the work into buckets (e.g. A-C, D-G, H- ...) and applied Thad's process/tools we might get a consistent, repeatable result from a group with the same goal - map the two vocabularies. I'm not sure how to organize the buckets, but I bet Thad has an idea. I will sign up for a bucket with a reasonable 'volume'. I could dedicate 10-20 hours to this project if others made a comparable commitment.

jaygray0919 commented Jun 15, 2018

Thank you Thad @thadguidry One could make a case for @ppKrauss Peter's advice - the sprint concept. An alternative might be a more formal divide-and-conquer approach. Thad has developed a process and is using tools. Perhaps Thad could document that process and how to use tools (OpenRefine is well documented so a light overview may be all that is needed). If we divided the work into buckets (e.g. A-C, D-G, H- ...) and applied Thad's process/tools we might get a consistent, repeatable result from a group with the same goal - map the two vocabularies. I'm not sure how to organize the buckets, but I bet Thad has an idea. I will sign up for a bucket with a reasonable 'volume'. I could dedicate 10-20 hours to this project if others made a comparable commitment.

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Jun 15, 2018

@jaygray0919 That's the idea. I'm waiting on my team in OpenRefine to get a good tool out for the community to handle this long term and much more easily. And yes we would be providing documentation and tutorials for this process. Here's the full docs for Reconciling https://github.com/OpenRefine/OpenRefine/wiki/Reconciliation but we're going to make a simpler, separate tutorial just for mapping Schema.org -> Wikidata.

thadguidry commented Jun 15, 2018

@jaygray0919 That's the idea. I'm waiting on my team in OpenRefine to get a good tool out for the community to handle this long term and much more easily. And yes we would be providing documentation and tutorials for this process. Here's the full docs for Reconciling https://github.com/OpenRefine/OpenRefine/wiki/Reconciliation but we're going to make a simpler, separate tutorial just for mapping Schema.org -> Wikidata.

@jaygray0919

This comment has been minimized.

Show comment
Hide comment
@jaygray0919

jaygray0919 Jun 15, 2018

@thadguidry good. we'll invest 10-20 hours in an organized 'stone soup' project (https://en.wikipedia.org/wiki/Stone_Soup). send up a flare when we should begin work on our 'bucket.'

jaygray0919 commented Jun 15, 2018

@thadguidry good. we'll invest 10-20 hours in an organized 'stone soup' project (https://en.wikipedia.org/wiki/Stone_Soup). send up a flare when we should begin work on our 'bucket.'

@pigsonthewing

This comment has been minimized.

Show comment
Hide comment
@pigsonthewing

pigsonthewing Jun 16, 2018

My proposal for a Wikidata property for IPTC subject codes has stalled, over the question of whether to have a property for each type of IPTC code, or just one over-arching property for all of them:

https://www.wikidata.org/wiki/Wikidata:Property_proposal/IPTC_subject_code

More views/ arguments there would help.

pigsonthewing commented Jun 16, 2018

My proposal for a Wikidata property for IPTC subject codes has stalled, over the question of whether to have a property for each type of IPTC code, or just one over-arching property for all of them:

https://www.wikidata.org/wiki/Wikidata:Property_proposal/IPTC_subject_code

More views/ arguments there would help.

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Jun 16, 2018

@pigsonthewing Added my Support comment to the Wikidata Property proposal.

thadguidry commented Jun 16, 2018

@pigsonthewing Added my Support comment to the Wikidata Property proposal.

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Jul 4, 2018

UPDATE: IPTC Newscode property is now live on Wikidata. Gotta get mapping now ! ;-) Thanks so much to Andy @pigsonthewing for making that happen !

thadguidry commented Jul 4, 2018

UPDATE: IPTC Newscode property is now live on Wikidata. Gotta get mapping now ! ;-) Thanks so much to Andy @pigsonthewing for making that happen !

@rtroncy

This comment has been minimized.

Show comment
Hide comment
@rtroncy

rtroncy Jul 17, 2018

@thadguidry I'm interested in getting an exhaustive list of mappings between the IPTC media topics codes and Wikidata. Running a simple query using the new P5429 property yields 74 results that actually mix up different code taxonomy from IPTC (which is fine).

Are you aware of any Phabricator task that aims to collect mappings between Wikidata and IPTC (media topics) codes? Are you working yourself on this?

rtroncy commented Jul 17, 2018

@thadguidry I'm interested in getting an exhaustive list of mappings between the IPTC media topics codes and Wikidata. Running a simple query using the new P5429 property yields 74 results that actually mix up different code taxonomy from IPTC (which is fine).

Are you aware of any Phabricator task that aims to collect mappings between Wikidata and IPTC (media topics) codes? Are you working yourself on this?

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Jul 17, 2018

@rtroncy what do you mean by mixing ? Can you give me 1 example here so I can see ?

thadguidry commented Jul 17, 2018

@rtroncy what do you mean by mixing ? Can you give me 1 example here so I can see ?

@rtroncy

This comment has been minimized.

Show comment
Hide comment
@rtroncy

rtroncy Jul 17, 2018

@thadguidry Click here, some Wikidata entities are mapped with "subject codes" (now deprecated), some with "audio codec", some with "media topics", some with "product genre", etc. ... all those are different codes list maintained (and sometimes deprecated) by IPTC and the mapping is done with the sole P5429 property. Fair enough, we can then filter if we want just the mappings to a specific code list, but this is what I meant with "mixing".

rtroncy commented Jul 17, 2018

@thadguidry Click here, some Wikidata entities are mapped with "subject codes" (now deprecated), some with "audio codec", some with "media topics", some with "product genre", etc. ... all those are different codes list maintained (and sometimes deprecated) by IPTC and the mapping is done with the sole P5429 property. Fair enough, we can then filter if we want just the mappings to a specific code list, but this is what I meant with "mixing".

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Jul 17, 2018

@rtroncy Ah, well, that's not mixing, that is called N:1 mapping. That is on purpose. "architecture" as an separate IPTC subject and media code, is still the concept of "architecture" in Wikidata. Having IPTC codes mapped to the correct Wikidata entity is a "good thing", even still when those IPTC codes have deprecated, then some old dataset can still be made useful against Wikidata because we took the time to map them for posterity sake.

thadguidry commented Jul 17, 2018

@rtroncy Ah, well, that's not mixing, that is called N:1 mapping. That is on purpose. "architecture" as an separate IPTC subject and media code, is still the concept of "architecture" in Wikidata. Having IPTC codes mapped to the correct Wikidata entity is a "good thing", even still when those IPTC codes have deprecated, then some old dataset can still be made useful against Wikidata because we took the time to map them for posterity sake.

@rtroncy

This comment has been minimized.

Show comment
Hide comment
@rtroncy

rtroncy Jul 17, 2018

Note that I'm not criticizing! I'm also all for doing mappings even for deprecated terms for the same exact reasons you provide. My comment was just that a single property is used (P5429) for doing N:1 mapping between Wikidata and different code lists which have just in common to be published by IPTC. One could have imagined minting different properties for mappings about subject codes, media topics, etc. But perhaps this would have been an overkill.

This brings me to my original question: who is working on this at the moment? If none, I may propose a mapping between Wikidata and the full media topics thesaurus.

rtroncy commented Jul 17, 2018

Note that I'm not criticizing! I'm also all for doing mappings even for deprecated terms for the same exact reasons you provide. My comment was just that a single property is used (P5429) for doing N:1 mapping between Wikidata and different code lists which have just in common to be published by IPTC. One could have imagined minting different properties for mappings about subject codes, media topics, etc. But perhaps this would have been an overkill.

This brings me to my original question: who is working on this at the moment? If none, I may propose a mapping between Wikidata and the full media topics thesaurus.

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Jul 17, 2018

@pigsonthewing and I decided that using many Wikidata properties was not needed, to have multiple properties, so we instead chose to go with just 1, and use it wisely. This reasoning is discussed in the initial Wikidata property proposal for the "IPTC Newscode" here: https://www.wikidata.org/wiki/Wikidata:Property_proposal/IPTC_subject_code

I'm working on it through OpenRefine reconciling currently. There's 1186 concepts that need to be mapped in total, and its not a completely automatic thing...human judgement has to be applied.
For instance, there's 20+ concepts of a "series" in Wikidata... and things like that, so its a bit slow process and I put about 4 hours a week into it, between my day job and volunteerism on open source projects.
capture

And crazy definitions used by IPTC sometimes like "food" with http://cv.iptc.org/newscodes/mediatopic/20000248 "Selling goods for human consumption to the human end user" which makes you immediately think they mean "food industry" but looking at the broader concept they have mapped, "Consumer Goods - Items produced for and sold to individuals" , then its clear they really did mean "food", but it has a food industry definition! Don't you just love vocabulary "opinions" and trying to map someone's state of mind at the time ? lololol

thadguidry commented Jul 17, 2018

@pigsonthewing and I decided that using many Wikidata properties was not needed, to have multiple properties, so we instead chose to go with just 1, and use it wisely. This reasoning is discussed in the initial Wikidata property proposal for the "IPTC Newscode" here: https://www.wikidata.org/wiki/Wikidata:Property_proposal/IPTC_subject_code

I'm working on it through OpenRefine reconciling currently. There's 1186 concepts that need to be mapped in total, and its not a completely automatic thing...human judgement has to be applied.
For instance, there's 20+ concepts of a "series" in Wikidata... and things like that, so its a bit slow process and I put about 4 hours a week into it, between my day job and volunteerism on open source projects.
capture

And crazy definitions used by IPTC sometimes like "food" with http://cv.iptc.org/newscodes/mediatopic/20000248 "Selling goods for human consumption to the human end user" which makes you immediately think they mean "food industry" but looking at the broader concept they have mapped, "Consumer Goods - Items produced for and sold to individuals" , then its clear they really did mean "food", but it has a food industry definition! Don't you just love vocabulary "opinions" and trying to map someone's state of mind at the time ? lololol

@rtroncy

This comment has been minimized.

Show comment
Hide comment
@rtroncy

rtroncy Jul 18, 2018

Indeed! We need to rely on human judgments and the short/vague definition of some terms does not make the life easy. Do you need help with the task? Is your openrefine project local or can we collaboratively work on it? By when do you expect to complete the mappings if you're work on it solely (and thanks a lot for doing this as an aside of a real job :-)

rtroncy commented Jul 18, 2018

Indeed! We need to rely on human judgments and the short/vague definition of some terms does not make the life easy. Do you need help with the task? Is your openrefine project local or can we collaboratively work on it? By when do you expect to complete the mappings if you're work on it solely (and thanks a lot for doing this as an aside of a real job :-)

@nitmws

This comment has been minimized.

Show comment
Hide comment
@nitmws

nitmws Jul 18, 2018

May I ask where these Wikidata / IPTC NewsCodes mappings come from? The IPTC Media Topics are the still maintained taxonomy of IPTC and it has a top level sports concept with a mapping to Wikidata: http://cv.iptc.org/newscodes/mediatopic/15000000 - but the Wikidata query refered to by @rtroncy ("Click here") shows a spct/sport for Q349 (Sport). From IPTC's point of view it would be great to use these about 100 Media Topic mappings as they have been done with a lot of human judgement by our team of taxonomists. (Note: the IPTC Subject Codes still exist but are not maintained anymore, the Media Topics are the successor)

nitmws commented Jul 18, 2018

May I ask where these Wikidata / IPTC NewsCodes mappings come from? The IPTC Media Topics are the still maintained taxonomy of IPTC and it has a top level sports concept with a mapping to Wikidata: http://cv.iptc.org/newscodes/mediatopic/15000000 - but the Wikidata query refered to by @rtroncy ("Click here") shows a spct/sport for Q349 (Sport). From IPTC's point of view it would be great to use these about 100 Media Topic mappings as they have been done with a lot of human judgement by our team of taxonomists. (Note: the IPTC Subject Codes still exist but are not maintained anymore, the Media Topics are the successor)

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Jul 18, 2018

@rtroncy I am able to do about 10% a day and just started this week (was waiting on OpenRefine bug to get fixed from my team), so should be done in about another week.

@nitmws The approx. 75 mappings that you see currently in Wikidata come from me. Once I am done hopefully next week, then I'll bulk upload into Wikidata from OpenRefine. And yes I am aware of the mappings done by the IPTC team themselves in IPTC Newscodes and am looking at them also when they show up. The task however is to get more Linked Data into Wikidata from lots of vocabularies. After I'm done, your IPTC team is welcome to query them and then notify on Wikidata talk pages for discussion of any quality issues...but let's wait until I'm done next week.

thadguidry commented Jul 18, 2018

@rtroncy I am able to do about 10% a day and just started this week (was waiting on OpenRefine bug to get fixed from my team), so should be done in about another week.

@nitmws The approx. 75 mappings that you see currently in Wikidata come from me. Once I am done hopefully next week, then I'll bulk upload into Wikidata from OpenRefine. And yes I am aware of the mappings done by the IPTC team themselves in IPTC Newscodes and am looking at them also when they show up. The task however is to get more Linked Data into Wikidata from lots of vocabularies. After I'm done, your IPTC team is welcome to query them and then notify on Wikidata talk pages for discussion of any quality issues...but let's wait until I'm done next week.

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Jul 21, 2018

UPDATE: 70% done with IPTC mapping.

thadguidry commented Jul 21, 2018

UPDATE: 70% done with IPTC mapping.

@bquinn

This comment has been minimized.

Show comment
Hide comment
@bquinn

bquinn Jul 24, 2018

Hi @thadguidry and all, I'm the new MD of IPTC, taking over from Michael who retired a few weeks ago (but will no doubt still stay involved in some projects).

It's great to see you working on this, and we'll be happy to take a look at your mappings when they're done (or before that if you want to share what you've done so far).

Also if you could flag up with us when you see a label or description that looks a bit strange, that would be great - we're in the process of reviewing the Media Topics vocabulary and would welcome any pointers to entries that look wrong.

Thanks again for your work!

bquinn commented Jul 24, 2018

Hi @thadguidry and all, I'm the new MD of IPTC, taking over from Michael who retired a few weeks ago (but will no doubt still stay involved in some projects).

It's great to see you working on this, and we'll be happy to take a look at your mappings when they're done (or before that if you want to share what you've done so far).

Also if you could flag up with us when you see a label or description that looks a bit strange, that would be great - we're in the process of reviewing the Media Topics vocabulary and would welcome any pointers to entries that look wrong.

Thanks again for your work!

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Jul 24, 2018

@bquinn Sure thing Brendan. I've been keeping a few notes, suggestions on simple tweaks, and deeper discussion issues. For starters to get you started on the simple stuff, there's a few topic names that ideally should have "and" replaced with "or" like this these : mediatopic/20000044 mediatopic/20000043 mediatopic/20000147

thadguidry commented Jul 24, 2018

@bquinn Sure thing Brendan. I've been keeping a few notes, suggestions on simple tweaks, and deeper discussion issues. For starters to get you started on the simple stuff, there's a few topic names that ideally should have "and" replaced with "or" like this these : mediatopic/20000044 mediatopic/20000043 mediatopic/20000147

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Aug 2, 2018

UPDATE: 90% done with IPTC mapping.

thadguidry commented Aug 2, 2018

UPDATE: 90% done with IPTC mapping.

@thadguidry

This comment has been minimized.

Show comment
Hide comment
@thadguidry

thadguidry Aug 13, 2018

UPDATE: 100% done with IPTC mapping. Edits being uploaded now from my OpenRefine instance, should be done in another 15 minutes. ;-)

753 IPTC Newscodes matched to Wikipedia topics.

thadguidry commented Aug 13, 2018

UPDATE: 100% done with IPTC mapping. Edits being uploaded now from my OpenRefine instance, should be done in another 15 minutes. ;-)

753 IPTC Newscodes matched to Wikipedia topics.

@bquinn

This comment has been minimized.

Show comment
Hide comment
@bquinn

bquinn Aug 13, 2018

Great, congratulations! I look forward to going through your mappings with the NewsCodes WG, we'll let you know if we spot any issues.

bquinn commented Aug 13, 2018

Great, congratulations! I look forward to going through your mappings with the NewsCodes WG, we'll let you know if we spot any issues.

@ettorerizza

This comment has been minimized.

Show comment
Hide comment
@ettorerizza

ettorerizza Sep 22, 2018

Dear all. I would be interested to take a look at the possibilities of mapping between the properties of schema.org and those of Wikidata. Can I use this file to do some testing, or is there a more complete and up-to-date list?

ettorerizza commented Sep 22, 2018

Dear all. I would be interested to take a look at the possibilities of mapping between the properties of schema.org and those of Wikidata. Can I use this file to do some testing, or is there a more complete and up-to-date list?

@Dataliberate

This comment has been minimized.

Show comment
Hide comment
@Dataliberate

Dataliberate Sep 22, 2018

Contributor
Contributor

Dataliberate commented Sep 22, 2018

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment