{"id":1179,"date":"2013-05-31T16:22:23","date_gmt":"2013-05-31T16:22:23","guid":{"rendered":"https:\/\/notebooks.dataone.org\/?p=1179"},"modified":"2013-05-31T16:22:23","modified_gmt":"2013-05-31T16:22:23","slug":"gathering-dataset","status":"publish","type":"post","link":"https:\/\/notebooks.dataone.org\/ontology-coverage\/gathering-dataset\/","title":{"rendered":"Gathering dataset"},"content":{"rendered":"

This week my focus was on meeting with my mentors, understanding my specific project requirements and gathering my datasets.\u00a0 My initial work was about creating a week-by-week plan to create a meaningful, generalizable ontology coverage tool for OWL ontologies.<\/p>\n

I spent some time looking over existing scripts and testing them to ensure the results were what I expected.\u00a0 Using these scripts I acquired the corpus I will be using in my research this summer.\u00a0 However, I also wrote some small scripts that will parse existing documents so that it will be easier to create ontologies (to allow for further testing).<\/p>\n

My second major step was gathering the ontologies.\u00a0 My mentors provided a link to some likely candidates, and we agreed to use the OWL API.\u00a0 However, as I was completely unfamiliar with the this API, I spent some time using tutorials and walking through example code to understand it.\u00a0 At this point, I have written some code that can read in existing ontologies, add various features to those ontologies, and check for whether features exist within an ontology.\u00a0 I have also written scripts that download a series of ontologies (to create a sufficient dataset for my research for the rest of the summer).<\/p>\n

I currently have both a corpus and all the SWEET ontologies.\u00a0 However, with my current scripts, it would be somewhat trivial to acquire a larger dataset (e.g., more corpses and ontologies) if that proves necessary.<\/p>\n","protected":false},"excerpt":{"rendered":"

This week my focus was on meeting with my mentors, understanding my specific project requirements and gathering my datasets.\u00a0 My initial work was about creating a week-by-week plan to create a meaningful, generalizable ontology coverage tool for OWL ontologies. I spent some time looking over existing scripts and testing them Continue reading Gathering dataset<\/span>→<\/span><\/a><\/p>\n","protected":false},"author":42,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[15],"tags":[143,140,141,56,142],"_links":{"self":[{"href":"https:\/\/notebooks.dataone.org\/wp-json\/wp\/v2\/posts\/1179"}],"collection":[{"href":"https:\/\/notebooks.dataone.org\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/notebooks.dataone.org\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/notebooks.dataone.org\/wp-json\/wp\/v2\/users\/42"}],"replies":[{"embeddable":true,"href":"https:\/\/notebooks.dataone.org\/wp-json\/wp\/v2\/comments?post=1179"}],"version-history":[{"count":1,"href":"https:\/\/notebooks.dataone.org\/wp-json\/wp\/v2\/posts\/1179\/revisions"}],"predecessor-version":[{"id":1181,"href":"https:\/\/notebooks.dataone.org\/wp-json\/wp\/v2\/posts\/1179\/revisions\/1181"}],"wp:attachment":[{"href":"https:\/\/notebooks.dataone.org\/wp-json\/wp\/v2\/media?parent=1179"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/notebooks.dataone.org\/wp-json\/wp\/v2\/categories?post=1179"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/notebooks.dataone.org\/wp-json\/wp\/v2\/tags?post=1179"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}