Loading etl/prs/prs_inspect_results.py +4 −2 Original line number Diff line number Diff line Loading @@ -236,14 +236,16 @@ def main(run_id: str, ) avg_density_map = np.nansum(np.where(density_grid == 0, np.nan, density_grid ** 2), axis=2) / \ np.nansum(np.where(density_grid == 0, np.nan, density_grid), axis=2) std_density_map = np.nanstd(np.where(density_grid == 0, np.nan, density_grid), axis=2) avg_temperature_map = np.nansum(np.where(density_grid == 0, np.nan, temperature_grid * density_grid), axis=2) / \ np.nansum(np.where(density_grid == 0, np.nan, density_grid), axis=2) correlation_data = np.array([avg_density_map.flatten(), avg_temperature_map.flatten(), std_density_map.flatten(), ratio_values[::zoom_ratios[0], ::zoom_ratios[1]].flatten()]) results_dict[f'{"-".join(lines)}'][f'{"_".join([str(tdust), str(nh2)])}'] = correlation_data plt.scatter(correlation_data[0], correlation_data[2]) plt.scatter(correlation_data[0], correlation_data[3]) df_coldens_mom0_list.append(get_column_density_vs_mom0(volume_density_grid=density_grid, mom_zero_name_1=mom0_name_1, mom_zero_name_2=mom0_name_2, Loading @@ -265,7 +267,7 @@ def main(run_id: str, for lines in line_pairs: df = pd.DataFrame( data=results_dict[f'{"-".join(lines)}'][f'{"_".join([str(tdust), str(nh2)])}'].transpose(), columns=['avg_nh2', 'avg_tdust', f'ratio_{"-".join(lines)}']) columns=['avg_nh2', 'avg_tdust', 'std_nh2', f'ratio_{"-".join(lines)}']) df_list.append(df.dropna()) df_tmp = pd.concat(df_list, axis=1) df_tmp = df_tmp.loc[:, ~df_tmp.columns.duplicated()].copy() Loading Loading
etl/prs/prs_inspect_results.py +4 −2 Original line number Diff line number Diff line Loading @@ -236,14 +236,16 @@ def main(run_id: str, ) avg_density_map = np.nansum(np.where(density_grid == 0, np.nan, density_grid ** 2), axis=2) / \ np.nansum(np.where(density_grid == 0, np.nan, density_grid), axis=2) std_density_map = np.nanstd(np.where(density_grid == 0, np.nan, density_grid), axis=2) avg_temperature_map = np.nansum(np.where(density_grid == 0, np.nan, temperature_grid * density_grid), axis=2) / \ np.nansum(np.where(density_grid == 0, np.nan, density_grid), axis=2) correlation_data = np.array([avg_density_map.flatten(), avg_temperature_map.flatten(), std_density_map.flatten(), ratio_values[::zoom_ratios[0], ::zoom_ratios[1]].flatten()]) results_dict[f'{"-".join(lines)}'][f'{"_".join([str(tdust), str(nh2)])}'] = correlation_data plt.scatter(correlation_data[0], correlation_data[2]) plt.scatter(correlation_data[0], correlation_data[3]) df_coldens_mom0_list.append(get_column_density_vs_mom0(volume_density_grid=density_grid, mom_zero_name_1=mom0_name_1, mom_zero_name_2=mom0_name_2, Loading @@ -265,7 +267,7 @@ def main(run_id: str, for lines in line_pairs: df = pd.DataFrame( data=results_dict[f'{"-".join(lines)}'][f'{"_".join([str(tdust), str(nh2)])}'].transpose(), columns=['avg_nh2', 'avg_tdust', f'ratio_{"-".join(lines)}']) columns=['avg_nh2', 'avg_tdust', 'std_nh2', f'ratio_{"-".join(lines)}']) df_list.append(df.dropna()) df_tmp = pd.concat(df_list, axis=1) df_tmp = df_tmp.loc[:, ~df_tmp.columns.duplicated()].copy() Loading