if you are browsing the repository you can realise that now there are several data models ended with the term ‘_incubated’.
This suffix denotes that there will be ‘soon’ a new data model.
The readme in the folder points to the repository where this data model is being created.
Once finished (all the docs, examples, etc) the _incubated suffix will be removed and the data model will be fully included into the data models.
This is the first survey to the USERS of the Smart Data Models. We want to know what is more important to you and try to prioritize our efforts. This is an agile initiative, therefore we will take into very consideration your comments. Your survey is anonymous although if you wish you can include your email in comments and we will contact you regarding your comments.
You can answer it here.
Soon there will be another survey for the CONTRIBUTORS.
In order to provide a more transparent and participative approach to the creation and maintenance of data models, here you can see the three stages for a usual data model to be part of the official list of data models.
1.- Pending is the ‘wild’ zone where developers share their thoughts and create their data models. When finished (ready) they can create a PR on the right Subject repository.
2.- Harmonization is the repository for those new accepted models while they are fine-tuned or completed
3.- Subject means the repository in which the data model is officially published. Once there they are versioned when necessary.
|Control of contributions
|Scripts for checking
||Yes (only acceptance and on update)
|Code inside json schema
||Contributor + SDM control
||Yes but not managed
||Yes (Here the open ones)
- Anyone previous pending acceptance
- Members direct
SDM: Smart Data Models initiative
The European open data portal is possibly the biggest open data portal in the world with more than one million free datasets coming from 36 countries.
One of the mechanisms to access its data is the use of the standard of DCAT-AP standard 2.0 which defines its catalogue of resources.
In the repository pending there are two data models, in progress to connect a dataset and a distribution as an additional resource by using NGSI. The idea is to create a script to map this resource into NGSI.
Every data model has a directory with examples of the payloads (json and jsonld) in key-values and normalized formats.
But it also includes examples of csv that it could help to connect the data models for other platforms.
Now, these examples are re-created automatically once a day.
The naming convention for these examples are the same name of the original payload plus a .csv suffix.
An example for the Battery data model examples.
In the upper menu (option Search descriptions) there is a searchable database of the descriptions for all the properties across the different data models.
Try it out here.
This DDBB is updated daily.
1.- It is just a technical improvement but now you can access everything in this site with https://data-models.fiware.org.
2.- The issue form has also been improved with new options but simplified structure.
Check it out and please report any issue.
Every subject had a file named CONTRIBUTORS.md which compiles all contributors to the different data models in this subject. However, from the point of view of the management, it was not a structured format. Due to this, it is going to be replaced with a new file named CONTRIBUTORS.yaml with the same info but structured so we could manage properly and answer questions like:
- How many contributors collaborate in this subject?
- How many subjects are being contributed by this person?
- What organisations are contributing to the initiative?
Now there is also a field for allocating comments.
The format is friendly enough to be simply edited but at the same time, it can be automatically processed.
Next week CONTRIBUTORS.md will be removed from the different repos.
Now it is based on a responsive interface with these options:
- Sort any column
- Search on any value across the different columns
- 50-pages pagination
check it out
In order to create a new data model is always interesting what others have done in order copy and to maintain interoperability.
That’s why we have available a database with all attributes and enumerations across all data models.
Now it’s updated daily and the number of occurrences is also available.