PhaseNet / phasenet /predict.py
zhuwq0's picture
upload phasenet
7b07ad9
import argparse
import logging
import multiprocessing
import os
import pickle
import time
from functools import partial
import h5py
import numpy as np
import pandas as pd
import tensorflow as tf
from data_reader import DataReader_mseed_array, DataReader_pred
from model import ModelConfig, UNet
from postprocess import (
extract_amplitude,
extract_picks,
save_picks,
save_picks_json,
save_prob_h5,
)
from pymongo import MongoClient
from tqdm import tqdm
from visulization import plot_waveform
tf.compat.v1.disable_eager_execution()
tf.compat.v1.logging.set_verbosity(tf.compat.v1.logging.ERROR)
username = "root"
password = "quakeflow123"
# client = MongoClient(f"mongodb://{username}:{password}@127.0.0.1:27017")
client = MongoClient(f"mongodb://{username}:{password}@quakeflow-mongodb-headless.default.svc.cluster.local:27017")
# db = client["quakeflow"]
# collection = db["waveform"]
def upload_mongodb(picks):
db = client["quakeflow"]
collection = db["waveform"]
try:
collection.insert_many(picks)
except Exception as e:
print("Warning:", e)
collection.delete_many({"_id": {"$in": [p["_id"] for p in picks]}})
collection.insert_many(picks)
def read_args():
parser = argparse.ArgumentParser()
parser.add_argument("--batch_size", default=20, type=int, help="batch size")
parser.add_argument("--model_dir", help="Checkpoint directory (default: None)")
parser.add_argument("--data_dir", default="", help="Input file directory")
parser.add_argument("--data_list", default="", help="Input csv file")
parser.add_argument("--hdf5_file", default="", help="Input hdf5 file")
parser.add_argument("--hdf5_group", default="data", help="data group name in hdf5 file")
parser.add_argument("--result_dir", default="results", help="Output directory")
parser.add_argument("--result_fname", default="picks", help="Output file")
parser.add_argument("--highpass_filter", default=0.0, type=float, help="Highpass filter")
parser.add_argument("--min_p_prob", default=0.3, type=float, help="Probability threshold for P pick")
parser.add_argument("--min_s_prob", default=0.3, type=float, help="Probability threshold for S pick")
parser.add_argument("--mpd", default=50, type=float, help="Minimum peak distance")
parser.add_argument("--amplitude", action="store_true", help="if return amplitude value")
parser.add_argument("--format", default="numpy", help="input format")
parser.add_argument("--s3_url", default="localhost:9000", help="s3 url")
parser.add_argument("--stations", default="", help="seismic station info")
parser.add_argument("--plot_figure", action="store_true", help="If plot figure for test")
parser.add_argument("--save_prob", action="store_true", help="If save result for test")
parser.add_argument("--upload_waveform", action="store_true", help="If upload waveform to mongodb")
parser.add_argument("--pre_sec", default=1, type=float, help="Window length before pick")
parser.add_argument("--post_sec", default=4, type=float, help="Window length after pick")
args = parser.parse_args()
return args
def pred_fn(args, data_reader, figure_dir=None, prob_dir=None, log_dir=None):
current_time = time.strftime("%y%m%d-%H%M%S")
if log_dir is None:
log_dir = os.path.join(args.log_dir, "pred", current_time)
if not os.path.exists(log_dir):
os.makedirs(log_dir)
if (args.plot_figure == True) and (figure_dir is None):
figure_dir = os.path.join(log_dir, "figures")
if not os.path.exists(figure_dir):
os.makedirs(figure_dir)
if (args.save_prob == True) and (prob_dir is None):
prob_dir = os.path.join(log_dir, "probs")
if not os.path.exists(prob_dir):
os.makedirs(prob_dir)
if args.save_prob:
h5 = h5py.File(os.path.join(args.result_dir, "result.h5"), "w", libver="latest")
prob_h5 = h5.create_group("/prob")
logging.info("Pred log: %s" % log_dir)
logging.info("Dataset size: {}".format(data_reader.num_data))
with tf.compat.v1.name_scope("Input_Batch"):
if args.format == "mseed_array":
batch_size = 1
else:
batch_size = args.batch_size
dataset = data_reader.dataset(batch_size)
batch = tf.compat.v1.data.make_one_shot_iterator(dataset).get_next()
config = ModelConfig(X_shape=data_reader.X_shape)
with open(os.path.join(log_dir, "config.log"), "w") as fp:
fp.write("\n".join("%s: %s" % item for item in vars(config).items()))
model = UNet(config=config, input_batch=batch, mode="pred")
# model = UNet(config=config, mode="pred")
sess_config = tf.compat.v1.ConfigProto()
sess_config.gpu_options.allow_growth = True
# sess_config.log_device_placement = False
with tf.compat.v1.Session(config=sess_config) as sess:
saver = tf.compat.v1.train.Saver(tf.compat.v1.global_variables(), max_to_keep=5)
init = tf.compat.v1.global_variables_initializer()
sess.run(init)
latest_check_point = tf.train.latest_checkpoint(args.model_dir)
logging.info(f"restoring model {latest_check_point}")
saver.restore(sess, latest_check_point)
picks = []
amps = [] if args.amplitude else None
if args.plot_figure:
multiprocessing.set_start_method("spawn")
pool = multiprocessing.Pool(multiprocessing.cpu_count())
for _ in tqdm(range(0, data_reader.num_data, batch_size), desc="Pred"):
if args.amplitude:
pred_batch, X_batch, amp_batch, fname_batch, t0_batch, station_batch = sess.run(
[model.preds, batch[0], batch[1], batch[2], batch[3], batch[4]],
feed_dict={model.drop_rate: 0, model.is_training: False},
)
# X_batch, amp_batch, fname_batch, t0_batch = sess.run([batch[0], batch[1], batch[2], batch[3]])
else:
pred_batch, X_batch, fname_batch, t0_batch, station_batch = sess.run(
[model.preds, batch[0], batch[1], batch[2], batch[3]],
feed_dict={model.drop_rate: 0, model.is_training: False},
)
# X_batch, fname_batch, t0_batch = sess.run([model.preds, batch[0], batch[1], batch[2]])
# pred_batch = []
# for i in range(0, len(X_batch), 1):
# pred_batch.append(sess.run(model.preds, feed_dict={model.X: X_batch[i:i+1], model.drop_rate: 0, model.is_training: False}))
# pred_batch = np.vstack(pred_batch)
waveforms = None
if args.upload_waveform:
waveforms = X_batch
if args.amplitude:
waveforms = amp_batch
picks_ = extract_picks(
preds=pred_batch,
file_names=fname_batch,
station_ids=station_batch,
begin_times=t0_batch,
config=args,
waveforms=waveforms,
use_amplitude=args.amplitude,
upload_waveform=args.upload_waveform,
)
if args.upload_waveform:
upload_mongodb(picks_)
picks.extend(picks_)
if args.plot_figure:
if not (isinstance(fname_batch, np.ndarray) or isinstance(fname_batch, list)):
fname_batch = [fname_batch.decode().rstrip(".mseed") + "_" + x.decode() for x in station_batch]
else:
fname_batch = [x.decode() for x in fname_batch]
pool.starmap(
partial(
plot_waveform,
figure_dir=figure_dir,
),
# zip(X_batch, pred_batch, [x.decode() for x in fname_batch]),
zip(X_batch, pred_batch, fname_batch),
)
if args.save_prob:
# save_prob(pred_batch, fname_batch, prob_dir=prob_dir)
if not (isinstance(fname_batch, np.ndarray) or isinstance(fname_batch, list)):
fname_batch = [fname_batch.decode().rstrip(".mseed") + "_" + x.decode() for x in station_batch]
else:
fname_batch = [x.decode() for x in fname_batch]
save_prob_h5(pred_batch, fname_batch, prob_h5)
if len(picks) > 0:
# save_picks(picks, args.result_dir, amps=amps, fname=args.result_fname+".csv")
# save_picks_json(picks, args.result_dir, dt=data_reader.dt, amps=amps, fname=args.result_fname+".json")
df = pd.DataFrame(picks)
# df["fname"] = df["file_name"]
# df["id"] = df["station_id"]
# df["timestamp"] = df["phase_time"]
# df["prob"] = df["phase_prob"]
# df["type"] = df["phase_type"]
if args.amplitude:
# df["amp"] = df["phase_amp"]
df = df[
[
"file_name",
"begin_time",
"station_id",
"phase_index",
"phase_time",
"phase_score",
"phase_amp",
"phase_type",
]
]
else:
df = df[
["file_name", "begin_time", "station_id", "phase_index", "phase_time", "phase_score", "phase_type"]
]
# if args.amplitude:
# df = df[["file_name","station_id","phase_index","phase_time","phase_prob","phase_amplitude", "phase_type","dt",]]
# else:
# df = df[["file_name","station_id","phase_index","phase_time","phase_prob","phase_type","dt"]]
df.to_csv(os.path.join(args.result_dir, args.result_fname + ".csv"), index=False)
print(
f"Done with {len(df[df['phase_type'] == 'P'])} P-picks and {len(df[df['phase_type'] == 'S'])} S-picks"
)
else:
print(f"Done with 0 P-picks and 0 S-picks")
return 0
def main(args):
logging.basicConfig(format="%(asctime)s %(message)s", level=logging.INFO)
with tf.compat.v1.name_scope("create_inputs"):
if args.format == "mseed_array":
data_reader = DataReader_mseed_array(
data_dir=args.data_dir,
data_list=args.data_list,
stations=args.stations,
amplitude=args.amplitude,
highpass_filter=args.highpass_filter,
)
else:
data_reader = DataReader_pred(
format=args.format,
data_dir=args.data_dir,
data_list=args.data_list,
hdf5_file=args.hdf5_file,
hdf5_group=args.hdf5_group,
amplitude=args.amplitude,
highpass_filter=args.highpass_filter,
)
pred_fn(args, data_reader, log_dir=args.result_dir)
return
if __name__ == "__main__":
args = read_args()
main(args)