-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Forking the vocab graph #67
Comments
@nissimsan This is a very deep question because it's core to the governance and even "value" of this project. I vote for modernization (thus "breaking with tradition" where warranted), but what do the others think? |
@VladimirAlexiev, noting that we are in agreement about the direction. I love your analogy! After having thought about this exact problem for several years I have found it increasingly less controversial:
I suggest that we start today's meeting to discuss this topic, flesh out the pros and cons, and make a decision so we can push forward. It seems pretty clear that we are in a holding pattern until this gets sorted. |
Discussed on today's call: #69 |
I love the biyearly train model approach of UN/CEFACT (as an interim solution until our sci-fi dream comes true: all dependent software can be generated instantly by some CI/CD process overnight from the blueprint/specification). Only this centralized semantic model that everyone worldwide agreed upon gives us the ability to throw at any given time data over the digital fence, so the data can be processed from the receiver by automation (without human interaction nor bilateral agreements. This is the core of digitalization. I am new to this project and surely was not able to understand all implications that come along with this issue, yet I still tend to generate the "syntax" of JSON-LD from a shared centralistic model. If this is a utopian dream, I would like to understand the problems in more detail. Thanks for the interesting web conference today. I (will) be back! :-) |
Good push back on today's meeting - thank you (especially @cmsdroff and @svanteschubert). We are debating the nature of this project, and I fully accept your viewpoints. The realistic alternative to what I suggest on this issue is to distinguish the feedback we get on tickets between "Content" and "NDR". Everything content (for example #42) will be fed back to the relevant CEFACT modelling teams; we wouldn't even try to fix for example Only issues dealing with transformation/syntax/and other NDR-aspects (for example #25) will be actually dealt with in the scope of the vocab project. |
Based on the above, I just took a quick pass through the current issues, tagging the relevant ones with either "NDR" or "semantics": |
After further thought, digesting last meeting's input, I'm fine to abandon this idea. I suggest we aim to make a quick decision on Friday and get this closed. |
@nissimsan For info some of the same questions have been asked by the API JSON schema export tech spec team. Not sure who is on that team or if how we should align this, but it feels like passing the semantic issues back is the right approach. Im in the Library maintenance team call but it will likely be rolled to next week due to the volume of submissions. I'll follow this along with the library submission team |
Decision made - we will not fork, but rather pass back semantic issues to modelling teams. Closing ticket. |
Recent issues (such as #42, #48 and #41) raises bugs which will be easy and trivial to fix directly in the vocab graph, but demand complex special case treatment in the current NDR/programmatic approach. This raises the natural question: what is the value of sticking with the constraint of manipulating the vocab only programmatically? It was a natural place to start, but now that we have a more mature vocab to work from, allowing ourselves to make updates directly in the praph will significantly simplify the path forward.
One aspect to consider is future updates, which will have to be switched from a "transform and replace all" to a "transform increments only, merge into the graph". In theory at least this could be seen as more difficult; however, with the recent influx of issues the idea of doing everything with a "do everything automatically" script no longer seems practically possible.
I suggest we pivot away from the constraint we have unnecessarily forced upon ourselves and allow ourselves the freedom to continue manipulating the vocab graph directly too.
The text was updated successfully, but these errors were encountered: