-
Notifications
You must be signed in to change notification settings - Fork 22
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CWB data ingestion into seiscomp3 #8
Comments
You might also have to use antelope sql to filer events by date for this POC. |
This is the POC we want....events captured in antelope via primary stations data, which also overlaps some temporary stations data. Therefore, for events from antelope, and time series data from CWB, make sure that we have at least one temporary survey which overlaps the events/primary stations waveforms. |
We are ingesting the waveform and events.xml data for the month of March 2017 into sc3 for this POC. So it would be good to have temporary waveform data for the same period as per Sudipta's comment above. |
We cannot specify when temporary stations data are available, and because we always have data from primary statons, the strategy should be to select events/primary stations data that overlap with some available temporary stations data. |
The antelope events for the month of March 2015 (#=695) have been ingested into the seiscomp3 instance "niket_pst_poc_latest". FYI: Out of the total 780 events that are registered in antelope DB, some of the events were found to be having fatal errors while being parsed by obspy library. So around 85 such events were filtered out during the event extraction process. |
This are the tests we need to ensure/establish the procedure and gain confidence on the
Step 4 is critical and should be done with wide range of time ranges and many stations. We probably need to make this comparison for every miniseed we ingest into the production seiscomp3. These steps will ensure a robust data migration from CWB to seiscomp3. |
The test of using scevtstreams in conjunction with scart to export the waveform timeseries data for a historical event from March 2015 to a miniseed file is successful. Attaching the waveform of such miniseed file for a 7.4 magnitude event on 2nd March 2015 as seen on scrttv. There is a small issue with the event data being imported from antelope though. scevtstreams doesn't pre-pend the network code to the output for the antelope events, while does it for the native seiscomp3 events. For e.g. the output for native seiscomp3 events looks like: centos@ip-172-31-26-134:/opt/seiscomp3/var/lib/archive/2015/IU/GNI/BHZ.D$ scevtstreams -E ga2017skmtsa -d mysql://sysop:sysop@localhost/seiscomp3 -L 0 -m 300 but the output for imported events from antelope looks like: centos@ip-172-31-26-134:~$ scevtstreams -E "quakeml:ga.ga.gov.au/event/00967663" -d mysql://sysop:sysop@localhost/seiscomp3 -L 0 -m 300 with the network code "IU" missing in this case. This is just to make a note. |
Our POC seiscomp3 machine should have all other utilities installed, specifically needs to run following bash scripts:
Make sure you use the |
This is how this can proceed:
slarchive
(can usescart
).scevtstreams
andscart
combination)POC target: 1 month worth of historical data for all primary stations in aws.
Example of cwb query:
query -h localhost -t ms -s ".*" -b "2005-02-08 05:37:52.98" -d 1000
Clarification on 3:
antelope
virtualenv inside our sc3 image in aws sinceantelope
uses propriatary python libraries. Instead you will have to export the events in seiscomp3 xml format and copy them across to the sc3 image. This part will require some ingenuity to automate.The text was updated successfully, but these errors were encountered: