SCEC_2019

所属分类:地理学
开发工具:Jupyter Notebook
文件大小:138423KB
下载次数:0
上传日期:2019-07-30 22:54:21
上 传 者sh-1993
说明:  南加州地震中心。模拟EQ数据集(RSQsim)、ML模型、预测概率和其他。
(Southern California Earthquake Center. Simulated EQ datasets (RSQsim), ML models, forecasted probabilities, and misc.)

文件列表:
50-50-30 models (0, 2019-07-31)
50-50-30 models\LR-50-50-30 (51619, 2019-07-31)
50-50-30 models\NN_50_50_30 (85778, 2019-07-31)
50-50-30 models\RF_model_50_50_30.zip (25963224, 2019-07-31)
code (0, 2019-07-31)
code\dimension_reduction_via_scikit_learn.ipynb (8408, 2019-07-31)
code\pipelines_and_scoring_jareds_ds.ipynb (131496, 2019-07-31)
code\random_sampling_LR_model.ipynb (62007, 2019-07-31)
code\useless_exercises (0, 2019-07-31)
code\useless_exercises\deep_neural_network_example.ipynb (70267, 2019-07-31)
code\useless_exercises\t10k-images-idx3-ubyte (7840016, 2019-07-31)
code\useless_exercises\t10k-labels-idx1-ubyte (10008, 2019-07-31)
code\useless_exercises\train-images-idx3-ubyte (47040016, 2019-07-31)
code\useless_exercises\train-labels-idx1-ubyte (60008, 2019-07-31)
data (0, 2019-07-31)
data\10-1-count - 10-1-count.csv (2978367, 2019-07-31)
data\100-10-count - 100-10-count.csv (2977594, 2019-07-31)
data\3-1-count - 3-1-count.csv (2978576, 2019-07-31)
data\30-10-count - 30-10-count.csv (2977497, 2019-07-31)
data\RFC-rupture-probs-100-30-special - RFC-rupture-probs-100-30-special.csv (983, 2019-07-31)
data\RFC-rupture-probs-100-30-special-format - RFC-rupture-probs-100-30-special-format.csv (1621, 2019-07-31)
data\RFC-rupture-probs-50-50-30.csv (410833, 2019-07-31)
data\RFC-rupture-probs.csv (1368, 2019-07-31)
data\dataset_4d.csv (2282201, 2019-07-31)
data\dataset_4d_in_0s_and_1s.csv (1133497, 2019-07-31)
data\dependent-probabilities-CCMC.csv (2163456, 2019-07-31)
data\eqdata.csv (2141738, 2019-07-31)
data\eqdata_30_year_intervals.csv (615618, 2019-07-31)
data\eqdata_33_sections.csv (9741537, 2019-07-31)
data\eqdata_33_sections_30_year_intervals.csv (2768920, 2019-07-31)
data\eqdata_disjoint_1000in_100out.csv (25576, 2019-07-31)
data\eqdata_disjoint_100in_30out.csv (221529, 2019-07-31)
data\eqdata_disjoint_100in_50out.csv (191862, 2019-07-31)
data\eqdata_disjoint_10in_10out.csv (1484085, 2019-07-31)
data\eqdata_disjoint_150in_70out.csv (130556, 2019-07-31)
data\eqdata_disjoint_200in_50out.csv (114809, 2019-07-31)
data\eqdata_disjoint_30in_100out.csv (221529, 2019-07-31)
data\eqdata_disjoint_30in_10out.csv (736635, 2019-07-31)
... ...

# SCEC_2019 Repository for the 2019 Machine Learning team. A place to host datasets, probabilities, and code. ### About this repo, and which folder to use in 2020 and beyond: The final files folder is a distilled compilation of everything needed to use and run the 2019 team's code / results. Everything else was used prior to the creation of the final files folder. We kept the other folders the same for our convenience with not having to change url's in some of our code. Moving forward we suggest working strictly out of the final files folder. The models were built based on two separate interval/binning methods on 'eqdata_33_sections'. The first one was open interval (ask Bill Savran), the second was a 100 year input bin, 30 year output bin. #### About the final files folder: ###### Inside the pickled files folder are pickled models: 1. Logistic Regression models (33 total) 2. Random Forest 3. Neural Network 4. There's a CSV file for the baseline probabilities. ###### About the datasets folder: 1. 'eqdata_33_sections' is the base final, from which 'eqdata_33_sects_testing' and 'eqdata_33_sects_training' are a straight split of eqdata_33_sections. From there, the other training/testing files are made based on the two separate binning/interval interpretations mentioned above. Namely the open interval, and the 100 year input / 30 year output sum interval method. As an example, 'eqdata_33_sections_30_year_intervals_testing' is the open interval testing set. 2. Why some files are binary files: if you don't use the binary files, colab notebooks had some issues, so we had to make some binary files. #### Not found in this repo, but found in the google drive: * Code to build the ML models * Code for generating probabilities * Visualization tools (ROC curves, scoring plots, etc) * Results files

近期下载者

相关文件


收藏者