You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+7-12Lines changed: 7 additions & 12 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,13 +4,17 @@ This repository contains the resources necessary to populate and curate the mode
4
4
* AutomaticLayoutDetection: Script that given a layout file extracted automatically from a dataset, it transforms it into the DataCubes representation from W3C.
5
5
* CSVToRDF: Scripts to transform all model contents in /Data to RDF.
6
6
* Data: All CSV files containing information of the models, their metadata and variables. The data is organized so the columns represent properties and data properties of the ontology, while the rows are instances described and linked in the models (e.g., input types, variables, etc.)
7
+
* ExportCatalog: Scripts to extract the contents of the model catalogs as a series of CSVs.
7
8
* GSNVariableImport: Scripts that query the current contents of the model catalog to extract GSNs (now SVOs) and bring the appropriate context into the model catalog. Additional links to Wikidata are created in the process.
8
9
* OtherTransformations: Legacy scripts to organize information about units.
9
10
* UnitToRDF: Scripts designed to align unit labels provided by modelers (e.g., "m/day") to a semantic representation.
10
-
11
+
12
+
## Process to export the model catalog.
13
+
Just execute `exportModelCatalog.py`. There is a config.yaml script to indicate the graphs desired to export (each user has a graph, right now it is configured to extract the mint and texas graphs). As a result, the script will write a series of CSVs, where the first column represents an instance, the header of each column represents the property and the cell rows represent the different values.
14
+
11
15
## Process to populate the model catalog.
12
16
13
-
1) Execute CSVToRDF. Compile and run the Java project, which will create an initial version of the turtle file with all contents from the Data folder integrated and linked.
17
+
1) Execute CSVToRDF. Compile and run the Java project, which will create an initial version of the turtle file with all contents from the Data folder integrated and linked. You should point the folder produced by the export Python script.
14
18
15
19
2) Extract units from labels and connect to WikiData:
16
20
@@ -40,15 +44,6 @@ This repository contains the resources necessary to populate and curate the mode
40
44
1. cd GSNVariableImport
41
45
2. python gsnvariableimport.py i
42
46
The "i" option makes the process interactive. If "a" is entered instead, the system will always pick the first definition found.
43
-
44
-
## Ongoing work
45
-
46
-
* Container descriptions: We will be adding container description files in RDF so one can explore the insights of a container (e.g., software dependencies, vulnerabilities, etc.)
47
-
48
-
* Layout description files: We are in the process of generating layout description files which contain a data cube of how we expect an input/output dataset to be. For example, in case of CSVs, the position of the columns in the described file.
49
-
50
-
## Future plans
51
-
52
-
* Automated CAG extraction: We plan to align our causal diagrams by extracting dependencies directly from the code.
0 commit comments