-
Notifications
You must be signed in to change notification settings - Fork 35
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How can we best make HTS codes referenceable? #178
Comments
The best way would be to get the ITC, and more specifically, the Office of Tariff Affairs and Trade Agreements, to publish that data as Linked Data, with URIs properly minted for each relevant entity. This would not be very difficult, given the starting point of those documents you linked. (Most of the heavy lifting was already done, to produce those docs.) They -- or even you, as the licensing is very permissible -- could do it in hours if not minutes with Virtuoso; other Linked Data servers could also be used, but I cannot speak to how easily or quickly the task could be done with such others. If the ITC does it, they can mint permanent URIs under a domain/namespace they control. This would be optimal, all around. |
Thanks @TallTed! I completely agree, ITC would (should) ideally expose the HTS codes; and WCO expose HS codes. I'll see what I can do to convince them to do so. One of my peers at UN/CEFACT (thanks @colugo) has produced the attached JSON-LD representations. As a practical way to fill the void, might it make sense to host these lists here, though? Similar to how we also have the periodic table - could that be a model? That's also work great as an example of what we are requesting from ITC and WCO. |
I would not recommend storing monolithic dump files in a change-management-system such as git nor a site based on the same such as github. I would strongly recommend against storing such monolithic files in some other organization's space within such system -- including here. I've not looked into the JSON-LD ZIPs you posted here; it's all-but-certain that they are not based on official URIs/domains, without which they are of limited if any use. (If they are based in an official namespace, some of what follows may be implementable now, without further official assistance or involvement.) Significantly, the periodic table doesn't change (except when the occasional new element is added). These HS/HTS data sets appear to change annually, if not faster, so whatever exercise is performed now will need to be repeated at least every year going forward. During the Obama administration, data.gov made a great many Federal data sets public as live, linked data, accessible through SPARQL among other methods, hosted by a variety of back-ends initially stood up as proofs-of-concept and for comparative testing. Those efforts were stopped early in 2017, for reasons which are probably obvious. I think the current administration may recognize the value in reviving and advancing what was in place circa December 2016. The more voices pressing this idea, the more likely it will prevail -- so make your voice heard there, as well! I believe the UN and/or some sub-agency/ies thereof has or had been working on similar initiatives with their data sets. As to a practical way to fill the void, it would be best if the authoritative organizations could at least make known the (previously or newly) intended permanent URIs, whether or not they actually publish the relevant datasets there. Then someone willing and able to set up a long-lived deployment, with URIs within their own domain, could easily add (Similarly, though it would add some work to the wished-for agency publication, some third party could publish the same datasets as SPARQL-accessible Linked Data, at long-lived URIs; the agency could then add the These publications can be locally mirrored by entities who need to make queries against them, through creative DNS and other techniques, and such mirrors can persist and be replicated into the future, whether the original publications remain up or not (though, of course, it is preferable that they remain up, such that the URIs of these entities can be dereferenced by any querent, not only those who have local mirrors). Once official URIs for the HS/HTS entities are minted, those URIs may be used in perpetuity to refer to — to denote — those entities, whether or not those same URIs are ever dereferenceable. This would limit the utility of the HTTP-based "superkeys" — the URIs — of Linked Data in these data sets, but it would still mean that every other entity who needed to refer to the HS/HTS entities could use the same identifiers to do so, and thereby be unambiguous in their references. There's much more to this subject, but neither this issue nor this repo is the best space for such discussion. The W3C's Semantic Web mailing list is likely to reach a more relevant and appropriate audience. |
@TallTed, many thanks. It's hard to disagree with you on this. Better to focus our efforts on the proper solution than an interim one. I actually did file a request via https://www.data.gov/contact back when I raised this. I haven't heard back from them. But now noticed their github and raised this: GSA/data.gov#923 I've been deeply involved in the UN work you are referring to: https://service.unece.org/trade/uncefact/vocabulary/uncefact/ This includes many code lists, but not HS nor HTS (which as WCO and US Gov respectively). Thanks again! |
@TallTed et al., I've sent [email protected] three emails now (ref GSA/data.gov#923) but have not heard anything back. I'll keep trying, but don't see a reason to keep this ticket alive for something out of our control and may never happen. So I propose we close this issue and stick with the model we have: Any objections to closing? |
Below is email from Jeremy Wise, a super proactive and helpful contact at USITC. I wasn't aware of the first option he is highlighting: From: Wise, Jeremy [email protected] |
That could be the best we're going to get. I am somewhat disappointed in the of dereferencing that new URI --
Note that it is plain JSON, not JSON-LD. Conneg for So it looks like the next step is to introduce Mr. Wise to some of the wonders of Linked Data, if not RDF, such that they at least add an |
Thanks @TallTed, |
@nissimsan can you give us an update on this issue? |
I have nothing to report on this except passing of time. I did pass along the idea of adding |
Update from USITC: Updating the HTS DMS is underway. Schedule release of next version Fall 2022. |
I will reach out when it's fall. |
It is now fall. I will reach out. |
Just followed up with Jeremy at USITC. Expecting that he will follow the link to here too (but be precluded from responding) - hi Jeremy! 👋 |
@nissimsan any updates on this issue? |
Suggest closing this issue, it's not actionable. |
[email protected] is the contact... He doesn't appear to be a GitHub user (or at least, I can't find his handle) |
Closing as non-actionable |
"HTS" codes is an American extension of the World Customs Organization's Harmonized System codes. Expressing commodity type is used for determining tariffs and associated documentation.
Here is a source of these HTS lists:
https://catalog.data.gov/dataset/harmonized-tariff-schedule-of-the-united-states-2019
This includes a json formatted file - I don't suppose that's sufficient for proper referencing? I am looking to indicate something like "htsno":"6204.11.00.00" on an import declaration.
My question applies to the "normal" HS codes too:
https://github.com/datasets/harmonized-system/blob/master/data/harmonized-system.csv
The text was updated successfully, but these errors were encountered: