{"id":221,"date":"2011-08-23T18:16:35","date_gmt":"2011-08-24T00:16:35","guid":{"rendered":"http:\/\/notebooks.dataone.org\/lod4dataone\/?p=221"},"modified":"2013-05-15T15:31:53","modified_gmt":"2013-05-15T15:31:53","slug":"week-ten-update-final-update","status":"publish","type":"post","link":"https:\/\/notebooks.dataone.org\/linked-data\/week-ten-update-final-update\/","title":{"rendered":"Week Ten Update – Final Update"},"content":{"rendered":"

Week Ten Goals<\/h3>\n

The goals for week 10, being the final week of the LOD4DataONE internship, were to finish off the details for a SPARQL query demonstration using the RDF data. In addition, all code, documents and RDF was to be collected on the github project page. Links to the query page and github are found below.<\/p>\n

Results<\/h3>\n

The demonstration is via a Webpage, found at http:\/\/manaus.cs.utep.edu\/lod4d1<\/a>, running on a Drupal server. The Webpage, implemented through a Drupal module, describes the project and queries that relate to the two use cases for the project. The queries search over the DataONE RDF that was created for this project as well as two remote stores: Data.gov and DBpedia. Answering the queries required a synchronization of data, requiring that I modify the RDF that I had for the three repositories so that RDF could be found when expected, with a single query. In some cases, the results don’t get everything I want but this is due to the different ways that data is expressed through the three repositories. I would expect this to be the case with any effort to combine data, I also find this to be an inherent issue with multiple sources of data.
\nIn order to support Use Case 1, I created a series of queries that show how to find data across the three DataONE member repositories. The result data can then be browsed via an RDF browser, as was done with the examples throughout this research.
\nIn order to support Use Case 2, I created queries that extract data from Dbpedia, data.gov and the DataONE repositories, creating a mashup of RDF that could then be browsed. One thing I had to consider when accessing remote stores was whether I wanted to duplicate the RDF data in the LOD4DataONE triple store for querying or if it made sense to limit the data to only what was needed for the particular mashup. Either way, using generic queries for accessing data from the remote stores generated too many triples for my process thus I had to limit them to some random number, e.g., 1000. As a result, I provided sample queries that extracted more focused data, i.e., pulling in less RDF from the remote stores by limiting the data with a SPARQL filter vs a random number. It seems that as the data over the Semantic Web increases, configuring stores and queries will be a norm for accessing the immense amounts of content that will be available.
\nAt the bottom of the demonstration Webpage are several queries that I tested over the last two weeks.
\nFinally, all php code for the Drupal module, java code for the RDF generation, Powerpoint presentations, javadoc and RDF are available on the github wiki, along with a Readme file.<\/p>\n

Observations<\/h3>\n