| 1 | = Next Steps , notes & links on OpenData, nano publications & Lucene Build Index on cvs file extracted by molgenis csv extract. = |
| 2 | == Next Steps : == |
| 3 | 1. Build an index (using Lucene) in Molgenis on a single database table. (Gene table from SyndromeBook). |
| 4 | * we have two options: |
| 5 | 1. use command line lucene . Need files for that. So we can export csv (molgenis csv export - Joeri) . The problem with that is that we wil have to export every single column or row in a file (see how that can be done) so the search is efficient. We need separate files for every single information (valid form) from the table . So an option is per column? or per row?. Explore that . |
| 6 | 1. There is an experimental database (Joeri ) we could also try that . The main idea is to generalize that and later make an index on more db tables, and potentially build a generator . The goal later is to place a google search on patient data on top of Molgenis. |
| 7 | 1. The fact is that there are several sets (molgenis) : database- model - system and we can use this lucene based indexing engine to create and index and be used by the search box that could be able to look in these dbs .''' If we generalize that to ALL DBs in molgenis we have the search inside patient data.''' |
| 8 | 1. use lucene java call inside a plugin in molgenis for the table in (1). |
| 9 | 1. Search not though ontocat but lucene in specific DBs . A start point is SyndromeBook's DB table : gene . |
| 10 | |
| 11 | http://www.ebi.ac.uk/ebisearch/search.ebi?db=allebi&requestFrom=searchBox&query=brca&FormsButton3=Go |
| 12 | |
| 13 | == Other notes : == |
| 14 | * Peregrine is running in Concept wiki so that data production in TRIPLEs is feasible. --> nano publications (some steps are missing , but we get the idea) --> rdf |
| 15 | * Hypothesis data / Real data /evidence --> experiments ---> STATEMENTS --> triples --> semantic web |
| 16 | * molgnenis producing triples ?? (experimental DB - Joeri) future plans. |
| 17 | * About the servlet version of search on top of ontocat , if you use a servlet , you are not REST (architecture) . |
| 18 | |
| 19 | Ontocat is retuning keywords , how about links? or more specific studies about the specific term. |
| 20 | |
| 21 | * http://www.nbic.nl/about-nbic/affiliated-organisations/cwa/introduction/ |
| 22 | * http://esw.w3.org/images/c/c0/HCLSIG$$SWANSIOC$$Actions$$RhetoricalStructure$$meetings$$20100215$cwa-anatomy-nanopub-v3.pdf |
| 23 | * [http://www.w3.org/TR/xhtml-rdfa-primer/ Rdfa] : "I''n this paper we explore the extra components that would need to be available to reinforce the value of a statement to the point where it could in itself be considered a publication. " - nanoPublication..''" |
| 24 | * "''[http://www.nbic.nl/about-nbic/affiliated-organisations/cwa/introduction/ The Concept Web Alliance (CWA) is a non- profit organization whose mission is “to enable an open collaborative environment to jointly address the challenges associated with high volume scholarly and professional data production, storage, interoperability and analyses for knowledge discovery]''.” "==== '''...and more analyric perpective on each ''' |
| 25 | * [http://www.nbic.nl/about-nbic/affiliated-organisations/cwa/declaration/ A declaration'] |
| 26 | |
| 27 | ==== aspect of the as called "core model":''' ''' ==== |
| 28 | * "''Our core model addresses some key requirements that stem from existing publication practices and the need to aggregate information from distributed sources. Similar to standard scientific publications, nano-publications need to be citable, attributable, and reviewable. Furthermore, they need to be easily curated. Nano-publications must be easily aggregated and identified across the Web. Finally, they need to be extensible to cater for new forms of both metadata and descriptio''n." |
| 29 | * '''''"aggregate Information from distributed sources".''.this is really important . The key is not to create more and more resources out of the existing ones, but try to provide and efficient and accurate serving/presentation of the existing valid ones. use standards that could actually point/refer to the core of your actual search, in a way that the information is distributed in an organized and consistent manner. ''' |
| 30 | |
| 31 | ==== links ==== |
| 32 | http://www2005.org/cdrom/docs/p613.pdf |
| 33 | |
| 34 | http://4store.org/ |
| 35 | |
| 36 | http://tagora.ecs.soton.ac.uk/eswc2009/ |
| 37 | |
| 38 | http://wiki.dbpedia.org/Downloads351 |
| 39 | |
| 40 | * ''"Numerous scientists have pointed out the irony that right at the historical moment when we have the technologies to permit worldwide availability and distributed process of scientific data, broadening collaboration and accelerating the pace and depth of discovery…..we are busy locking up that data and preventing the use of correspondingly advanced technologies on knowledge"'' |
| 41 | * ''http://sciencecommons.org/'' |
| 42 | |
| 43 | === STEP 1 : Create a Lucene Index (command line) using Molgenis svn extract for database hvp_pilot . === |
| 44 | 1. New test csv class : call csv export (/Users/despoina/Documents/workspace/hvp_pilot/handwritten/java/plugins/test_csv.java) |
| 45 | 1. . Done . File in : CVS molgenis export directory : /private/var/folders/to/toww8wCyG3a88-qsfLyIV++++TI/-Tmp-/ |
| 46 | 1. Cvs export in molgenis does not export every single valid quantity of information , like columns. just the database . |
| 47 | 1. In command prompt cd to cdLucene lucene directory , and try : (after you have copied your csv file in a directory here _syndrome_book_data_ |
| 48 | 1. $$$$ java org.apache.lucene.demo.!IndexFiles _syndrome_book_data_/ |
| 49 | |
| 50 | * Now you can search your index by typing : |
| 51 | |
| 52 | 1. $$$$$$ java org.apache.lucene.demo.!SearchFiles |
| 53 | 1. example search :glycoprotein |
| 54 | 1. ok ther it is in a single file ... |
| 55 | 1. CUSTOMIZING lucene : ..more output ...[http://lucene.apache.org/java/3_0_1/queryparsersyntax.html#Fuzzy Searches http://lucene.apache.org/java/3_0_1/queryparsersyntax.html#Fuzzy Searches] |
| 56 | * http://lucene.apache.org/java/3_0_1/gettingstarted.html |
| 57 | |
| 58 | == == |