csiborgtools/notebooks/MAH/mah.ipynb
Richard Stiskalek ee222cd010
Fix overlap runs (#125)
* Update nb

* Update script

* Update script

* Rename

* Update script

* Update script

* Remove warning

* Ignore minors when extracting MAH

* Fix paths bug

* Move notebooks

* Move files

* Rename and delete things

* Rename file

* Move file

* Rename things

* Remove old print statement

* Add basic MAH plot

* Add random MAH path

* Output snapshot numbers

* Add MAH random extraction

* Fix redshift bug

* Edit script

* Add extracting random MAH

* Little updates

* Add CB2 redshift

* Add some caching

* Add diagnostic plots

* Add caching

* Minor updates

* Update nb

* Update notebook

* Update script

* Add Sorce randoms

* Add CB2 varysmall

* Update nb

* Update nb

* Update nb

* Use catalogue HMF

* Move definition of radec2galactic

* Update nb

* Update import

* Update import

* Add galatic coords to catalogues

* Update nb
2024-04-08 11:23:21 +02:00

390 KiB

Process PV catalogues

In [2]:
from os.path import join
import numpy as np
import matplotlib.pyplot as plt
from h5py import File
from astropy.coordinates import match_coordinates_sky, SkyCoord
from astropy import units as u

%matplotlib inline
%load_ext autoreload
%autoreload 2

SPEED_OF_LIGHT = 299_792.458

Supernovae data

LOSS

In [3]:
a2dir = "/mnt/extraspace/rstiskalek/catalogs/PV/PV_Supranta/A2"

names = ["z_CMB", "mB", "x1", "c", "e_mB", "e_x1", "e_c", "RA", "DEC"]
dtype = [(n, np.float32) for n in names]
data = np.genfromtxt(join(a2dir, "loss.csv"), delimiter=",", skip_header=1,
                     usecols=[5 + n for n in range(len(names))])

loss_data = np.empty(len(data), dtype=dtype)
for i, n in enumerate(names):
    loss_data[n] = data[:, i]

Foundation

In [4]:
names = ["z_CMB", "RA", "DEC", "x1", "mB", "c", "peak", "e_peak", "e_x1", "e_mB", "e_c"]
dtype = [(n, np.float32) for n in names]
data = np.genfromtxt(join(a2dir, "foundation.csv"), delimiter=",", skip_header=1,
                     usecols=[3 + n for n in range(len(names))])

foundation_data = np.empty(len(data), dtype=dtype)
for i, n in enumerate(names):
    foundation_data[n] = data[:, i]

Pantheon+, all

In [5]:
fpath_full = "/mnt/extraspace/rstiskalek/catalogs/PV/Pantheon+SH0ES.dat"
fpath_group = "/mnt/extraspace/rstiskalek/catalogs/PV/pantheon+_groups.hdf5"

data_full = np.genfromtxt(fpath_full, names=True, dtype=None, encoding=None)
data_group = {}

# Read in the groups
with File(fpath_group, "r") as f:
    print(f.keys())
    for key in f.keys():
        try:
            data_group[key] = f[key][...]
        except IndexError:
            print(f"Failed to read {key}")
        # data[key] = f[key][...]

# data = data[data["zCMB"] < 0.1]

keys = ["zCMB", "zCMBERR", "mB", "mBERR", "x1", "x1ERR", "c", "cERR", "RA",
        "DEC", "VPEC", "VPECERR", "biasCor_m_b", "biasCorErr_m_b"]
pantheon_data = {}
for key in keys:
    pantheon_data[key] = data_full[key]
<KeysViewHDF5 ['DEGdeg', 'DEHdeg', 'DEJ2000', 'Host', 'IAUC', 'RAGdeg', 'RAHdeg', 'RAJ2000', 'SNID', 'SimbadName', 'e_vpec', 'e_zHD', 'e_zhel', 'recno', 'vpec', 'vpecG', 'zGHD', 'zGcmb', 'zGhel', 'zHD', 'zcmb', 'zhel']>
In [25]:
match_coord = SkyCoord(data_full["RA"], data_full["DEC"], unit="deg")
catalogue_coord = SkyCoord(data_group["RAJ2000"], data_group["DEJ2000"],
                           unit="deg")

idxs, sep2d, sep3d = match_coordinates_sky(match_coord, catalogue_coord, nthneighbor=1)
print(f"Maximum 2D separation is: {sep2d.max()}")
Maximum 2D separation is: 0.0005450609539575544 deg
In [26]:
zcmb_supernovae = data_group["zcmb"][idxs]
zcmb_group = data_group["zGcmb"][idxs]

pantheon_data["zCMB_SN"] = zcmb_supernovae
pantheon_data["zCMB_Group"] = zcmb_group

mask = pantheon_data["zCMB"] < 0.1
for key in pantheon_data.keys():
    pantheon_data[key] = pantheon_data[key][mask]
In [27]:
plt.figure()
plt.scatter(pantheon_data["zCMB_SN"], pantheon_data["zCMB_Group"], s=1)
plt.show()
No description has been provided for this image

Tully-Fisher galaxies

SFI++ galaxies

In [28]:
tf_folder = "/mnt/extraspace/rstiskalek/catalogs/PV/PV_Supranta/tf"

names = ["RA", "DEC", "z_CMB", "mag", "eta", "e_mag", "e_eta"]

dtype = [(n, np.float32) for n in names]
data = np.genfromtxt(join(tf_folder, "sfi_gals_tf.csv"), delimiter=",", skip_header=1,
                     usecols=[2 + n for n in range(len(names))])

sfi_gals = np.empty(len(data), dtype=dtype)
for i, n in enumerate(names):
    sfi_gals[n] = data[:, i]

SFI++ galaxies masked

In [29]:
names = ["RA", "DEC", "z_CMB", "mag", "eta", "e_mag", "e_eta"]

dtype = [(n, np.float32) for n in names]
data = np.genfromtxt(join(tf_folder, "sfi_gals_tf_masked.csv"), delimiter=",", skip_header=1,
                     usecols=[2 + n for n in range(len(names))])

sfi_gals_masked = np.empty(len(data), dtype=dtype)
for i, n in enumerate(names):
    sfi_gals_masked[n] = data[:, i]

SFI++ groups

In [30]:
names = ["RA", "DEC", "z_CMB", "r_hMpc", "e_r_hMpc"]

dtype = [(n, np.float32) for n in names]
data = np.genfromtxt(join(tf_folder, "sfi_grps.csv"), delimiter=",", skip_header=1,
                     usecols=[1 + n for n in range(len(names))])

sfi_groups = np.empty(len(data), dtype=dtype)
for i, n in enumerate(names):
    sfi_groups[n] = data[:, i]

Cross $\texttt{SFI++ galaxies}$ and $\texttt{SFI++ groups}$

In [31]:
match_coord = SkyCoord(sfi_gals["RA"], sfi_gals["DEC"], unit="deg")
catalogue_coord = SkyCoord(sfi_groups["RA"], sfi_groups["DEC"],
                           unit="deg")

idxs, sep2d, sep3d = match_coordinates_sky(match_coord, catalogue_coord, nthneighbor=1)
sep2d.to(u.degree)
Out[31]:
[$3^\circ40{}^\prime57.6210022{}^{\prime\prime}$ $0^\circ45{}^\prime09.98210907{}^{\prime\prime}$ $1^\circ47{}^\prime48.02945137{}^{\prime\prime}$ ... $0^\circ59{}^\prime48.39719296{}^{\prime\prime}$ $1^\circ27{}^\prime20.76476097{}^{\prime\prime}$ $2^\circ01{}^\prime03.90094757{}^{\prime\prime}$]
In [32]:
m = sep2d.value < 0.5
plt.figure()
plt.hist(sep2d, bins="auto")
plt.xlabel("Separation [deg]")
plt.ylabel("Counts")
plt.tight_layout()
plt.savefig("../../plots/sfi_gals_to_sfi_groups.png", bbox_inches="tight",
            dpi=450)
plt.show()
No description has been provided for this image

2MTF

In [33]:
names = ["RA", "DEC", "mag", "e_mag", "z_CMB", "r_hMpc", "e_rhMpc", "M", "eta", "e_eta"]

dtype = [(n, np.float32) for n in names]
data = np.genfromtxt(join(tf_folder, "twomtf_k.csv"), delimiter=",", skip_header=1,
                     usecols=[2 + n for n in range(len(names))])

twomtf_gals = np.empty(len(data), dtype=dtype)
for i, n in enumerate(names):
    twomtf_gals[n] = data[:, i]

Write to HDF5

In [34]:
outdir = "/mnt/extraspace/rstiskalek/catalogs"
fname = "PV_compilation.hdf5"

with File(join(outdir, fname), 'w') as f:
    # Write LOSS
    grp = f.create_group("LOSS")
    for name in loss_data.dtype.names:
        grp.create_dataset(name, data=loss_data[name])

    # Write Foundation
    grp = f.create_group("Foundation")
    for name in foundation_data.dtype.names:
        grp.create_dataset(name, data=foundation_data[name])

    # Write SFI gals
    grp = f.create_group("SFI_gals")
    for name in sfi_gals.dtype.names:
        grp.create_dataset(name, data=sfi_gals[name])

    # Write SFI gals masked
    grp = f.create_group("SFI_gals_masked")
    for name in sfi_gals_masked.dtype.names:
        grp.create_dataset(name, data=sfi_gals_masked[name])

    # Write SFI groups
    grp = f.create_group("SFI_groups")
    for name in sfi_groups.dtype.names:
        grp.create_dataset(name, data=sfi_groups[name])

    # Write 2MTF gals
    grp = f.create_group("2MTF")
    for name in twomtf_gals.dtype.names:
        grp.create_dataset(name, data=twomtf_gals[name])

    # Write Pantheon
    grp = f.create_group("Pantheon+")
    for name in pantheon_data.keys():
        grp.create_dataset(name, data=pantheon_data[name])
In [ ]: