Christopher added a comment.
Do you foresee any changes to the context/vocabulary/ontology in the future
(e.g. implementing processing features of JSON-LD 1.1)? How will context
changes be versioned / published?
Could not also the ontology <http://wikiba.se/ontology-1.0.owl#&
Christopher added a comment.
@JanZerebecki It would nice to consolidate the post-processed "ready to use"
export data sets in a single directory that can be accessed by different
presentation clients simply over http. Like here
http://datasets.wikimedia.org/aggregate-datasets/{wik
Christopher added a comment.
FYI: I am working on the dashboards and have made some progress using the
shiny-server.
Check out the very preliminary prototype at http://wdm.wmflabs.org/wdm
The repo is here
https://git.wikimedia.org/tree/wikidata%2Fanalytics%2Fdashboard/HEAD
@Addshore if you
Christopher added a comment.
@Addshore yes, I will just create a separate remote download set function and
point it at your sources so that we can use both local and remote data.
One thing to note (particular to Dygraphs) is that the order of the columns in
the datasource matters. The first
Christopher created this task.
Christopher claimed this task.
Christopher added subscribers: EBernhardson, Ricordisamoa, Deskana,
JanZerebecki, Aklapper, Christopher.
Christopher added projects: Wikidata-Sprint-2015-08-11,
Wikidata-Sprint-2015-06-30, Wikidata.
TASK DESCRIPTION
This will
Christopher added subscribers: Lydia_Pintscher, Addshore.
Christopher added a comment.
This has progressed as follows:
1. Dashboard Repository created at wikidata/analytics/dashboard
2. Datasource Repository created at limn-wikidata-data
(https://phabricator.wikimedia.org/T112506)
3. Prototype
Christopher added a comment.
Using the analytics supported limn infrastructure is part of the dynamic metric
data set creation. (This should be created as a new and independent task).
From what I have gathered, the limn front end graphs have been deprecated.
Whether or not shiny is
Christopher added a comment.
The objective of using RDF/OWL as a format for the publication of the wikidata
metric definitions is to create a canonical data model. This is important to
establish the globally defined bases for interpretation of the values. The
consumers of the definitions are
Christopher created this task.
Christopher added subscribers: Luke081515, Revi, RP88, Aklapper, TempleM,
Matthewrbowker, scfc, coren, yuvipanda, Andrew, Christopher.
Christopher added projects: Labs, Wikidata, Wikidata-Query-Service.
Herald added a project: Discovery.
TASK DESCRIPTION
Project
Christopher added a subscriber: Smalyshev.
Christopher added a comment.
After researching this, I have discovered that the Munger that processes the
RDF dump removes several ontology types (wikibase:Item, wikibase:Statement,
wikibase:Reference, and wikibase:Value) that are needed for object
Christopher created this task.
Christopher assigned this task to Smalyshev.
Christopher added a subscriber: Christopher.
Christopher added projects: Wikidata-Query-Service, Wikidata.
Christopher moved this task to All WDQS-related tasks on the
Wikidata-Query-Service workboard.
Herald added a
Christopher added a comment.
@Andrew Is there something else that needs to be said/done in order to make
this happen?
Currently, the development dashboard is running on the scrumbugz project
(http://wdm.wmflabs.org/wdm/), so this will just be reallocated. Additional
note: If the RDF dumps
Christopher added a project: WMDE-Analytics-Engineering.
TASK DETAIL
https://phabricator.wikimedia.org/T108404
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: gerritbot, Addshore, Lydia_Pintscher, EBernhardson, Ricordisamoa, Deskana
Christopher added a project: WMDE-Analytics-Engineering.
Christopher set Security to None.
TASK DETAIL
https://phabricator.wikimedia.org/T113180
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: gerritbot, Christopher, Aklapper
Christopher added a project: WMDE-Analytics-Engineering.
Christopher set Security to None.
TASK DETAIL
https://phabricator.wikimedia.org/T115120
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Smalyshev, Christopher, Andrew, yuvipanda
Christopher added a subscriber: Christopher.
Christopher added a comment.
I cannot reproduce this now. I assume that this is fixed. The file is local
and in the repo now.
https://github.com/wikimedia/wikidata-analytics-dashboard/blob/master/data/wikidata_eng_social_media.tsv
TASK DETAIL
Christopher closed this task as "Resolved".
Christopher set Security to None.
TASK DETAIL
https://phabricator.wikimedia.org/T116150
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Christopher, Addshore, Aklapper, Wikidata-
Christopher added a comment.
This is why there is the config.R file. The only path variable that needs to
be changed is there.
See base_uri <- "/srv/dashboards/shiny-server/wdm/". In windows this would be
C:\whatever\whatever I guess.
TASK DETAIL
https://phabricator.wikimedi
Christopher added a subscriber: Christopher.
Christopher added a comment.
What is the benefit of having all properties on one graph? To me, the simplest
approach is to pass a parameter with a single property id from ordered table
link to a chart. To analyse the trend of a single property over
Christopher closed this task as "Resolved".
TASK DETAIL
https://phabricator.wikimedia.org/T116150
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: gerritbot, Christopher, Addshore, Aklapper, Wikidata-
Christopher added a subscriber: Christopher.
Christopher added a comment.
Does this mean that you would prefer the KPI tab on the dashboard sidebar to be
first in the list?
TASK DETAIL
https://phabricator.wikimedia.org/T117206
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings
Christopher added a blocked task: T117234: Reproduce wikidata-todo data using
analytics infrastructure .
TASK DETAIL
https://phabricator.wikimedia.org/T115242
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Smalyshev, Christopher
Cc: JanZerebecki
Christopher added subscribers: Addshore, Christopher.
Christopher added a comment.
Do you mean this https://searchdata.wmflabs.org/external/ ?
This should be able to be retrieved on short interval from Graphite?
@Addshore? The KPI is defined with a "rolling 30 day window".
Christopher moved this task to Doing on the WMDE-Analytics-Engineering
workboard.
TASK DETAIL
https://phabricator.wikimedia.org/T113180
WORKBOARD
https://phabricator.wikimedia.org/project/board/1585/
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To
Christopher moved this task to Doing on the WMDE-Analytics-Engineering
workboard.
TASK DETAIL
https://phabricator.wikimedia.org/T116009
WORKBOARD
https://phabricator.wikimedia.org/project/board/1585/
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To
Christopher added blocking tasks: T117194: Evaluate Spark on YARN, T117195:
Develop Wikidata (JSON or RDF) Dump Processing API for use with Spark.
TASK DETAIL
https://phabricator.wikimedia.org/T116547
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To
Christopher added a project: WMDE-Analytics-Engineering.
Christopher set Security to None.
TASK DETAIL
https://phabricator.wikimedia.org/T116547
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Wikidata-bugs, Addshore, Christopher
Christopher added a comment.
I have observed that the property data does not have a persistent frequency.
(i.e some days there are no values reported). It may be better to generate
null values for properties regularly if they do not report usage.
There are two options with the aggregate table
Christopher added a comment.
See the change here:
http://wdm.wmflabs.org/?t=wikidata_daily_getclaims_property_use
TASK DETAIL
https://phabricator.wikimedia.org/T116009
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Christopher
Christopher closed blocking task T117194: Evaluate Spark on YARN as "Declined".
TASK DETAIL
https://phabricator.wikimedia.org/T116547
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Wikidata-bugs, Addshore, Christopher, Ja
Christopher added a comment.
Can we agree that Graphite is the way forward for the backend and close this
task?
TASK DETAIL
https://phabricator.wikimedia.org/T116547
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Wikidata-bugs
Christopher reopened blocking task T117194: Evaluate Spark on YARN as "Open".
TASK DETAIL
https://phabricator.wikimedia.org/T116547
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Wikidata-bugs, Addshore, Christopher, Ja
Christopher closed blocking task T117195: Develop Wikidata (JSON or RDF) Dump
Processing API for use with Spark as "Resolved".
TASK DETAIL
https://phabricator.wikimedia.org/T116547
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To
Christopher closed blocking task T117194: Evaluate Spark on YARN as "Resolved".
TASK DETAIL
https://phabricator.wikimedia.org/T116547
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Addshore, Christopher
Cc: Wikidata-bugs, Addshore, C
Christopher added a comment.
Note: A new task will be created for measuring SPARQL performance for counting
tasks in different environments. This has some relationship to Hadoop and
Spark potentially, but the first step is profile Blazegraph with complex
counting queries and use this as a
Christopher added a comment.
Update: All data loaded into Blazegraph (it took over 24 hours). Sync now
running and up to 27 October.
Using Fast Range Counts returns counts of content objects instantly.
Examples:
curl -G http://wdm-rdf.wmflabs.org/bigdata/namespace/wdq/sparql
--data
Christopher added a comment.
No. the blocking task code enables an option to not filter item, statement,
value and reference rdf:types in the munger. I decided not to wait for this,
so that I could get started, but having it in master is very helpful going
forward.
In order to have these
Christopher added a subscriber: Christopher.
Christopher added a comment.
To expand on the use cases for a metrics storage backend here is appropriate.
I think that Wikidata content metrics favor long term retention (i.e. forever)
because their purpose is to evaluate dynamics over both short
Christopher added a comment.
If you are going to use HDFS, why not just use HBase instead of Graphite?
TASK DETAIL
https://phabricator.wikimedia.org/T117732
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Joe, Lydia_Pintscher
Christopher added a comment.
If not HBase, what about Cassandra? This is already puppetized. At least you
will be using a storage solution that is designed for HDFS.
TASK DETAIL
https://phabricator.wikimedia.org/T117732
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel
Christopher added a comment.
I am not sure why this is considered to be "a simple use case" since as
mentioned in https://phabricator.wikimedia.org/T117735 there are at least two
different requirements. Content metrics require long term (non-decaying)
storage, operational metr
Christopher added a comment.
Yes. It seems I need to disable the 10 minute query timeout set here first:
https://github.com/wikimedia/wikidata-query-rdf/blob/b3e646284f0b74131bce99a1b7d5fc6bfe675ec1/war/src/config/web.xml#L55
A fat query like this:
PREFIX wikibase: <http://wikiba
Christopher added a comment.
True, a statement is either referenced or "unreferenced". Getting the number
of referenced statements (currently 41,735,203) is easy and fast with:
curl -G https://query.wikidata.org/bigdata/namespace/wdq/sparql
--data-urlencode ESTCARD --data-url
Christopher added a comment.
OK. So the title "Referenced Statements by Statement Type" is just wrong then.
Rather, it shows **All Statements ** by Type"
| Date | itemlink | string | globecoordinate | time | quantity
| somevalue | novalue | Total
Christopher added a comment.
OK. I may have found an answer to the question of wildcard "Prefix Matching"
that is necessary in order to query for number of statements in an item.
PREFIX bds: <http://www.bigdata.com/rdf/search#>
prefix wikibase: <http://wikiba.se/ont
Christopher added a comment.
Truthy statement counts per Item can be done like this:
PREFIX wd: <http://www.wikidata.org/entity/>
SELECT (count(distinct(?o)) AS ?ocount) WHERE {
wd:Q7239 ?p ?o
FILTER(STRSTARTS(STR(?p), "http://www.wikidata.org/prop/direct";))
Christopher added a comment.
I am blocked on this by several problems with the data model/ontology. The
question of the relationship of the data model and the RDF node definitions is
a bit complicated, perhaps more so than it should be. A reference is a special
type of statement defined by
Christopher added a subscriber: Christopher.
TASK DETAIL
https://phabricator.wikimedia.org/T119775
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Christopher, Smalyshev, Aklapper, Addshore, StudiesWorld, jkroll,
Wikidata-bugs
Christopher added a comment.
You can get reference hashes for objects using the
http://www.wikidata.org/prop/reference/ predicate.
For example,
PREFIX wd: <http://www.wikidata.org/entity/>
PREFIX wdt: <http://www.wikidata.org/prop/direct/>
PREFIX prov: <http://www
Christopher added a member: Christopher.
PROJECT DETAIL
https://phabricator.wikimedia.org/project/profile/891/
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher, Gage, ksmith, Jdouglas, DanielFriesen, hoo, Addshore, Tpt,
JeroenDeDauw, Joe
Christopher added a comment.
The only way to get a count of statements with references in the current
model/format is like this:
PREFIX wd: <http://www.wikidata.org/entity/>
PREFIX wdt: <http://www.wikidata.org/prop/direct/>
PREFIX prov: <http://www.w3.org/ns/prov#>
Christopher created this task.
Christopher added a subscriber: Christopher.
Christopher added projects: Wikidata, Wikidata-Query-Service,
Wikibase-DataModel.
Herald added subscribers: StudiesWorld, Aklapper.
Herald added a project: Discovery.
TASK DESCRIPTION
This is data model and RDF
Christopher added a blocking task: T120166: Semantically define arity of
statement -> reference relations.
TASK DETAIL
https://phabricator.wikimedia.org/T117234
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Lydia_Pintsc
Christopher added a blocked task: T117234: Reproduce wikidata-todo/stats data
using analytics infrastructure .
TASK DETAIL
https://phabricator.wikimedia.org/T120166
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Aklapper
Christopher added a comment.
So basically a clever adaptation as to what I suggested in
https://phabricator.wikimedia.org/T119775 to get statements referenced to the
Wikipedias. It works, but seems a very hacky approach around the core problem
of not having a way to ask how many references a
Christopher added a comment.
@Jheald Thank you for your suggestions. What is fairly clear in my research is
that counting type queries on large (or undefined ranges) with an unbound
domain are just not possible (without huge resource consumption) when the
namespace contains millions and
Christopher added a comment.
Quick edit: I ran this query successfully in 13min, 11sec, 476m returning
312,068 results returning the arity of GND
(https://phabricator.wikimedia.org/P227) property statements. So it is
possible, but really, really slow.
prefix wikibase: <http://wikiba
Christopher added a comment.
@Jheald Perfect. This works, even with adding optional it runs in 10 seconds.
Yea, definitely outputting the statements is unnecessary and adds a lot of time.
Total results: 5, duration: 10445 ms
nrefs count
0 39775
1 339700
2 10050
3
Christopher added a comment.
@Addshore Some progress was made on this in
https://phabricator.wikimedia.org/T120166. The only "practical" way to get the
statement and reference metrics is to facet the data by property. It is just
not possible to run counting queries against the whol
Christopher added a comment.
I think that you may have missed the point. I added the $property variable in
the above query to indicate that this has to be run for **every** property.
p:https://phabricator.wikimedia.org/P227 is a random example.
TASK DETAIL
https
Christopher added a subscriber: Christopher.
Christopher added a comment.
Question: Why is not the GUI a completely independent project / repo / build
/deployment from WDQS?
One reason to not have to do a full maven build for every GUI patch can be seen
here: https://integration.wikimedia.org
Christopher added a comment.
I have actively started working on this. You can see the work here:
https://github.com/christopher-johnson/wdqs-gui
Since using node requires a lot of refactoring and code style changes, I am
interested in the developing the GUI as a separate dev branch or
Christopher closed this task as "Resolved".
TASK DETAIL
https://phabricator.wikimedia.org/T122848
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Christopher, Aklapper, Abraham, Wikidata-bugs, aud
Christopher added a comment.
done.
TASK DETAIL
https://phabricator.wikimedia.org/T122848
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Christopher, Aklapper, Abraham, Wikidata-bugs, aude, Mbch331
Christopher added a comment.
@smalyshev no, I think that this specific issue has been practically resolved.
TASK DETAIL
https://phabricator.wikimedia.org/T120166
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Smalyshev, Jheald
Christopher added a subscriber: Christopher.
Christopher added a comment.
question: why is this task limited in scope to the Graph extension?
TASK DETAIL
https://phabricator.wikimedia.org/T126730
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To
Christopher added a comment.
@smalyshev I completely agree with the concept of an intermediate service
between the nanosparqlserver and the client. I think that this service should
"broker" requests (based on an options configuration object), and eval whether
a query is re-execut
Christopher added a comment.
I perceive the use of Varnish as not directly related to how an object broker
could manage this use case (expensive querying of the wdqs nano sparql api),
though it is probably related to any UI elements (i.e. the query editor or
results renderer) that may
Christopher added a comment.
I may be wrong, but the headers that are returned from a request to the nginx
server wdqs1002 say that varnish 1.1 is already being used there. And, for
whatever reason,** it misses**, because repeating the same query gives the same
response time. For example
Christopher created this task.
Christopher moved this task to All WDQS-related tasks on the
Wikidata-Query-Service workboard.
Herald added a subscriber: Aklapper.
Herald added a project: Discovery.
TASK DESCRIPTION
See http://tinyurl.com/grkd7qw for an example query that returns the
Christopher added a comment.
Intentional or not., It is wrong. Why is it necessary? The problem is that
it breaks parsing of geosparql literals. For example, if I ask for instance of
volcanoes, I have to make exceptions for weird non-Earth coordinates.
TASK DETAIL
https
Christopher added a comment.
Thanks for the clarification. However, the Req 10 of the geoSPARQL
specification seems to be at odds with the definition of a "literal value".
(According to https://www.w3.org/TR/rdf11-concepts/#section-Graph-Literal).
The way that I read this spe
Christopher added a comment.
Eh,
http://schemas.opengis.net/geosparql/1.0/geosparql_vocab_all.rdf#wktLiteral is
an RDFS Datatype so the semantics are defined by the RDF schema, right? But, I
found this http://docs.opengeospatial.org/is/12-063r5/12-063r5.html that
demonstrates that the WKS
Christopher added a comment.
@Deskana - there are two relevant Christopher Johnson's, one that works for
Operations in VA, (who I believe goes by "Chris") and myself, (I go by
"Christopher" and work for WMDE).
While I have no prior knowledge of this task, it seems i
Christopher added a comment.
@JanZerebecki I built a test instance here: http://wdm.wmflabs.org/ The
vagrant script worked fine. I think I get the basics of the shiny-server (from
reading this http://rstudio.github.io/shiny-server/latest/ guide).
I guess I am ready to dig into the dashboard
Christopher added a comment.
@Smalyshev have you tried to read the updated WKT CRS specification
http://docs.opengeospatial.org/is/12-063r5/12-063r5.html yet? From what I can
interpret, they have now deprecated the 2012 "non-ISO compliant" concatenation
of a URI form of CRS an
Christopher added a comment.
Please see geoSPARQL CRS design is debatable
<https://www.w3.org/2015/spatial/wiki/Coordinate_Reference_Systems#GeoSPARQL>
from the W3C Coordinate Reference System website.
Also, #7 here: the conflation of CRS with with the WKT in a literal ha
Christopher added a comment.
@Smalyshev so, by stating that geometry and CRS are different, you then
concur with the main arguments referenced above that they should not be
conflated in a simple literal. @Daniel I agree with the idea of specifying the
CRS as an additional component of the
Christopher added a comment.
Coincidentally, it seems that there are people who know a lot more about this
than I do that have debated this issue at length in a long and very informative
thread:
CRS specification (was: Re: ISA Core Location Vocabulary)
<https://lists.w3.org/Archi
Christopher created this task.
Christopher moved this task to Blazegraph on the Wikidata-Query-Service
workboard.
Herald added a subscriber: Aklapper.
Herald added projects: Wikidata, Discovery.
TASK DESCRIPTION
openvirtuoso (dbpedia) can do this.
there is not a maven artifact similar to
Christopher closed this task as "Invalid".
TASK DETAIL
https://phabricator.wikimedia.org/T130799
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Aklapper, Christopher, debt, Gehel, D3r1ck01, FloNight, Izno, jkroll,
Christopher reopened this task as "Open".
TASK DETAIL
https://phabricator.wikimedia.org/T130799
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Aklapper, Christopher, debt, Gehel, D3r1ck01, FloNight, Izno, jkroll,
Smalyshev
Christopher added a comment.
the node.js version of the TPF server is actually why I created this issue.
My concept of the fragment server was that it could decentralize a big
dataset by distributing data fragments to it with selectors,
<http://www.hydra-cg.com/spec/latest/linked-d
Christopher added a comment.
it seems that with a CONSTRUCT query, sending an Accept: text/turtle works.
http://wdm-rdf.wmflabs.org/short/NyJpTCnpl
this is actually all that is required to get a linked data fragment from the
SPARQL interface.
The problem with TPF access to the
Christopher added a comment.
I have worked around the counting problem. The experimental TPF Server is
here:
http://orbeon-bb.wmflabs.org/
This wikidata datasource uses SPARQL interface at
http://query.wikidata.org/sparql
I think that this issue can be closed.
TASK DETAIL
Christopher created this task.
Herald added a subscriber: Aklapper.
Herald added projects: Wikidata, Discovery.
TASK DESCRIPTION
It seems that using shorthand rather than a lexical form for decimal
coordinates breaks (xsd schema) validation of the munged/split wikibase turtle
dumps. Example
Christopher added a comment.
The PRETTY_PRINT setting of the TurtleWriter is set to "true" by default.
This causes the writer to only write the literal "label" without the datatype.
This affects boolean, decimal, integer and double literals.
To fix make the follow
Christopher created this task.
Christopher moved this task to Need investigation on the Wikidata-Query-Service
workboard.
Herald added a subscriber: Aklapper.
Herald added projects: Wikidata, Discovery.
TASK DESCRIPTION
Wikipedia and Commons URIs do not match their RDF representation in
Christopher created blocking task T132319: Sitelink URIs should be IRIs.
TASK DETAIL
https://phabricator.wikimedia.org/T131960
EMAIL PREFERENCES
https://phabricator.wikimedia.org/settings/panel/emailpreferences/
To: Christopher
Cc: Christopher, Aklapper, Avner, debt, Gehel, D3r1ck01
Christopher created this task.
TASK DESCRIPTION
The RDF representation of **all** Wikidata sitelinks as URIs rather than IRIs
is problematic.
These sitelinks (as primary source identifiers) should be represented as
**unencoded** IRIs in the RDF because the sitelinks **refer directly** to
Christopher added a comment.
The main issue is with string comparison of the percent encoded and unencoded
forms of Unicode IRIs as **resources**.
Per https://tools.ietf.org/html/rfc3987#section-5.3.1
> When comparing character by character, the comparison function MUST NOT
Christopher added a comment.
The RDF standard that you reference explicitly supports my point.
> IRI normalization: Interoperability problems can be avoided by minting only
IRIs that are normalized according to Section 5 of [RFC3987].
> Non-normalized forms that are best a
Christopher added a comment.
according to mailing list (Wikidata Digest, Vol 83, Issue 18), this now enabled on beta. Yet when one requests the link: https://wikidata.beta.wmflabs.org/wiki/Special:EntityData/Q64.jsonld, it does not work?TASK DETAILhttps://phabricator.wikimedia.org/T207168EMAIL
Christopher added a comment.
thanks, I look forward to this being deployed. json-ld will be very useful for wikidata, particularly framing. You might want to consider providing the context as a remote link to reduce the payloads (and "noise" in the data). Here is that test entity,
Christopher created this task.Christopher added projects: Wikidata-Query-Service, Wikibase-DataModel-Serialization.Herald added a subscriber: Aklapper.Herald added projects: Wikidata, Discovery.
TASK DESCRIPTIONNote: this relates more to my localized use of Wikibase RDF serialization than to the
Christopher created this task.Christopher added projects: Wikidata-Query-Service, Wikibase-DataModel-Serialization.Herald added a subscriber: Aklapper.Herald added projects: Wikidata, Discovery.
TASK DESCRIPTIONNote: this relates more to my localized use of Wikibase RDF serialization than to the
Christopher edited the task description. (Show Details)
EDIT DETAILS...to produce the intended output attached.
{F5323223}
[[[ https://www.w3.org/2011/rdf-wg/wiki/Skolemisation | Skolemization ]]...TASK DETAILhttps://phabricator.wikimedia.org/T155891EMAIL PREFERENCEShttps
Christopher edited the task description. (Show Details)
EDIT DETAILS...to produce the intended output attached.
{F5323223} 350}
[[ https://www.w3.org/2011/rdf-wg/wiki/Skolemisation | Skolemization ]]
[[ https://tools.ietf.org/html/rfc5785 | (RFC5785]]TASK DETAILhttps://phabricator.wikimedia.org
Christopher edited the task description. (Show Details)
EDIT DETAILS...{F5323350}64}
[[ https://www.w3.org/2011/rdf-wg/wiki/Skolemisation | Skolemization ]]...TASK DETAILhttps://phabricator.wikimedia.org/T155891EMAIL PREFERENCEShttps://phabricator.wikimedia.org/settings/panel/emailpreferences/To
Christopher edited the task description. (Show Details)
EDIT DETAILS...{F5323364}75}
[[ https://www.w3.org/2011/rdf-wg/wiki/Skolemisation | Skolemization ]]...TASK DETAILhttps://phabricator.wikimedia.org/T155891EMAIL PREFERENCEShttps://phabricator.wikimedia.org/settings/panel/emailpreferences/To
1 - 100 of 108 matches
Mail list logo