-
Notifications
You must be signed in to change notification settings - Fork 13
Process to handle scan programs #362
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Craig compiled a list of unique scripts that @smarshall-bmr can annotate: In addition, the metadata field (appears to) provide a link to a copy of the script version run on a particular on an ftp server ftp://10.160.21.2//gantry_data/LemnaTec/ScriptBackup/, but this is not resolving. the metadata looks like "gantry_system_variable_metadata": {
"...":,
"Script path on local disk": "C:\\LemnaTec\\StoredScripts\\SWIR_VNIR_Day1.cs",
"Script copy path on FTP server": "ftp://10.160.21.2//gantry_data/LemnaTec/ScriptBackup/SWIR_VNIR_Day1_7e5b6f22-f990-43fc-92d4-fee2db330d4a.cs",
"..."
}
|
we have the /gantry_data directory accessible on the cache server, but the ScriptBackup is not currently whitelisted for the transfer pipeline. there are 1,687 scripts listed there, e.g.:
@craig-willis this probably obviates your scan unless we have scripts in your results that dont appear here. Here is a list I generated. |
The list I generated has 80 entries. https://docs.google.com/spreadsheets/d/1tMkPT2jtMgTficfSDx80-RpB6PXuzry2gLEZG_ajvl8/edit I gather what you've got is specific versions on specific dates? That would be better for true traceability, if we can reference the specific script from the dataset metadata. |
@dlebauer Along with the program descriptions, should we consider linking to the fieldbook spreasheet (or something that references the fieldbook) from the dataset metadata? |
@craig-willis that would seem reasonable. Not sure the best way to do this. Much of this could be inserted into the BETYdb managements table (and some of it is there). Only issue is that the record keeping hasn't been consistent over the years. @smarshall-bmr what are your thoughts? |
@craig-willis @robkooper here's an exam question for ya. right now we trigger the full field extractor by:
I have added code to account for multiple scans per day in our full field unique key (day+sensor+scan). BUT that breaks our count check. I need a new method to know when to trigger full field without triggering when only half the geotiffs are generated. Ideas....
|
@max-zilla As discussed, your first option seems practical for November. |
Description of the scans is available here: |
we are accounting for this now. |
Per discussion with @smarshall-bmr, there are potentially multiple scans and scan programs on a given day. We need a process to capture the scan program information, update the cleaned metadata with the final program name/tag and update downstream extractors to respect this information (e.g., fullfield needs to run on a day for files associated with a program).
We need a way to determine if we have all files for that program in the raw data for the rulechecker.
@smarshall-bmr mentioned that program names may change over time, so we need to map to a standard name.
There is another issue with multiple scans on the same day with the same program. It's unclear whether this is something we need to address.
Completion criteria:
The text was updated successfully, but these errors were encountered: