diff --git a/.github/workflows/test.yml b/.github/workflows/test.yml index c60449c63..2d972d96a 100644 --- a/.github/workflows/test.yml +++ b/.github/workflows/test.yml @@ -71,8 +71,8 @@ jobs: fail-fast: false matrix: include: - - python: "3.7" - pyshort: "37" + # - python: "3.8" + # pyshort: "38" - python: "3.9" pyshort: "39" # - python: "3.10" diff --git a/setup.py b/setup.py index d6235a5db..879d57b95 100644 --- a/setup.py +++ b/setup.py @@ -275,6 +275,7 @@ def readme(): "pyyaml", "astropy", "healpy", + "pixell", "ephem", ] conf["extras_require"] = { diff --git a/src/toast/CMakeLists.txt b/src/toast/CMakeLists.txt index 068912526..e021eaa5e 100644 --- a/src/toast/CMakeLists.txt +++ b/src/toast/CMakeLists.txt @@ -110,7 +110,8 @@ install(FILES config.py job.py pixels.py - pixels_io.py + pixels_io_healpix.py + pixels_io_wcs.py covariance.py dist.py data.py diff --git a/src/toast/_libtoast/ops_mapmaker_utils.cpp b/src/toast/_libtoast/ops_mapmaker_utils.cpp index ddc98c493..bd47a7805 100644 --- a/src/toast/_libtoast/ops_mapmaker_utils.cpp +++ b/src/toast/_libtoast/ops_mapmaker_utils.cpp @@ -265,7 +265,7 @@ void init_ops_mapmaker_utils(py::module & m) { use_det_flags ); for (int64_t iw = 0; iw < nnz; iw++) { - #pragma omp atomic + # pragma omp atomic raw_zmap[zoff + iw] += zmap_val[iw]; } } diff --git a/src/toast/_libtoast/ops_pixels_healpix.cpp b/src/toast/_libtoast/ops_pixels_healpix.cpp index 639a1f9ce..311a70fe5 100644 --- a/src/toast/_libtoast/ops_pixels_healpix.cpp +++ b/src/toast/_libtoast/ops_pixels_healpix.cpp @@ -27,7 +27,7 @@ // So we must duplicate this across compilation units. void pixels_healpix_qa_rotate(double const * q_in, double const * v_in, - double * v_out) { + double * v_out) { // The input quaternion has already been normalized on the host. double xw = q_in[3] * q_in[0]; diff --git a/src/toast/_libtoast/ops_stokes_weights.cpp b/src/toast/_libtoast/ops_stokes_weights.cpp index ef2f7dbae..b69195948 100644 --- a/src/toast/_libtoast/ops_stokes_weights.cpp +++ b/src/toast/_libtoast/ops_stokes_weights.cpp @@ -19,7 +19,7 @@ // So we must duplicate this across compilation units. void stokes_weights_qa_rotate(double const * q_in, double const * v_in, - double * v_out) { + double * v_out) { // The input quaternion has already been normalized on the host. double xw = q_in[3] * q_in[0]; diff --git a/src/toast/_libtoast/template_offset.cpp b/src/toast/_libtoast/template_offset.cpp index 39b7d000b..b8f7758ff 100644 --- a/src/toast/_libtoast/template_offset.cpp +++ b/src/toast/_libtoast/template_offset.cpp @@ -201,7 +201,7 @@ void init_template_offset(py::module & m) { } else { contrib = raw_det_data[d]; } - #pragma omp atomic + # pragma omp atomic raw_amplitudes[amp] += contrib; } offset += raw_n_amp_views[iview]; diff --git a/src/toast/dipole.py b/src/toast/dipole.py index 7a8a9b737..6bdd338f9 100644 --- a/src/toast/dipole.py +++ b/src/toast/dipole.py @@ -2,7 +2,6 @@ # All rights reserved. Use of this source code is governed by # a BSD-style license that can be found in the LICENSE file. -import healpy as hp import numpy as np import scipy.constants as constants from astropy import units as u diff --git a/src/toast/observation.py b/src/toast/observation.py index 098876a6d..e23483709 100644 --- a/src/toast/observation.py +++ b/src/toast/observation.py @@ -185,12 +185,13 @@ def __init__( self._uid = name_UID(self._name) if self._session is None: - self._session = Session( - name=self._name, - uid=self._uid, - start=None, - end=None, - ) + if self._name is not None: + self._session = Session( + name=self._name, + uid=self._uid, + start=None, + end=None, + ) elif not isinstance(self._session, Session): raise RuntimeError("session should be a Session instance or None") diff --git a/src/toast/ops/CMakeLists.txt b/src/toast/ops/CMakeLists.txt index a8feed70b..75ef90c7f 100644 --- a/src/toast/ops/CMakeLists.txt +++ b/src/toast/ops/CMakeLists.txt @@ -27,6 +27,7 @@ install(FILES pointing_detector.py scan_map.py scan_healpix.py + scan_wcs.py mapmaker_utils.py mapmaker_binning.py mapmaker_solve.py @@ -53,6 +54,7 @@ install(FILES demodulation.py stokes_weights.py pixels_healpix.py + pixels_wcs.py filterbin.py save_hdf5.py load_hdf5.py diff --git a/src/toast/ops/__init__.py b/src/toast/ops/__init__.py index 88d8747ab..ee52917ba 100644 --- a/src/toast/ops/__init__.py +++ b/src/toast/ops/__init__.py @@ -36,6 +36,7 @@ from .operator import Operator from .pipeline import Pipeline from .pixels_healpix import PixelsHealpix +from .pixels_wcs import PixelsWCS from .pointing import BuildPixelDistribution from .pointing_detector import PointingDetectorSimple from .polyfilter import CommonModeFilter, PolyFilter, PolyFilter2D @@ -45,6 +46,7 @@ from .save_spt3g import SaveSpt3g from .scan_healpix import ScanHealpixMap, ScanHealpixMask from .scan_map import ScanMap, ScanMask, ScanScale +from .scan_wcs import ScanWCSMap, ScanWCSMask from .sim_cosmic_rays import InjectCosmicRays from .sim_crosstalk import CrossTalk, MitigateCrossTalk from .sim_gaindrifts import GainDrifter diff --git a/src/toast/ops/crosslinking.py b/src/toast/ops/crosslinking.py index e6ca3163f..4aeb90371 100644 --- a/src/toast/ops/crosslinking.py +++ b/src/toast/ops/crosslinking.py @@ -12,7 +12,7 @@ from ..mpi import MPI from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..timing import Timer, function_timer from ..traits import Bool, Float, Instance, Int, Unicode, trait_docs from ..utils import Logger diff --git a/src/toast/ops/filterbin.py b/src/toast/ops/filterbin.py index 9b7f250eb..0eb6bd122 100644 --- a/src/toast/ops/filterbin.py +++ b/src/toast/ops/filterbin.py @@ -22,7 +22,7 @@ from ..mpi import MPI from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import ( +from ..pixels_io_healpix import ( filename_is_fits, filename_is_hdf5, read_healpix_fits, diff --git a/src/toast/ops/groundfilter.py b/src/toast/ops/groundfilter.py index 5dbd85815..26bba3cfb 100644 --- a/src/toast/ops/groundfilter.py +++ b/src/toast/ops/groundfilter.py @@ -4,7 +4,6 @@ from time import time -import healpy as hp import numpy as np import traitlets from astropy import units as u diff --git a/src/toast/ops/madam_utils.py b/src/toast/ops/madam_utils.py index 84c0970c1..dacf7b5a9 100644 --- a/src/toast/ops/madam_utils.py +++ b/src/toast/ops/madam_utils.py @@ -4,7 +4,6 @@ import os -import healpy as hp import numpy as np from ..timing import function_timer diff --git a/src/toast/ops/mapmaker.py b/src/toast/ops/mapmaker.py index 435e548c3..e2676afc2 100644 --- a/src/toast/ops/mapmaker.py +++ b/src/toast/ops/mapmaker.py @@ -10,7 +10,8 @@ from ..mpi import MPI from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import write_healpix_fits, write_healpix_hdf5 +from ..pixels_io_healpix import write_healpix_fits, write_healpix_hdf5 +from ..pixels_io_wcs import write_wcs_fits from ..timing import Timer, function_timer from ..traits import Bool, Float, Instance, Int, Unicode, trait_docs from ..utils import Logger @@ -145,7 +146,7 @@ class MapMaker(Operator): write_rcond = Bool(True, help="If True, write the reciprocal condition numbers.") write_solver_products = Bool( - True, help="If True, write out equivalent solver products." + False, help="If True, write out equivalent solver products." ) keep_solver_products = Bool( @@ -423,9 +424,14 @@ def _exec(self, data, detectors=None, **kwargs): ] ).apply(data) - # FIXME: This all assumes the pointing operator is an instance of the - # PointingHealpix class. We need to generalize distributed pixel data - # formats and associate them with the pointing operator. + # FIXME: This I/O technique assumes "known" types of pixel representations. + # Instead, we should associate read / write functions to a particular pixel + # class. + + is_pix_wcs = hasattr(map_binning.pixel_pointing, "wcs") + is_hpix_nest = None + if not is_pix_wcs: + is_hpix_nest = map_binning.pixel_pointing.nest write_del = list() write_del.append((self.hits_name, self.write_hits)) @@ -439,25 +445,31 @@ def _exec(self, data, detectors=None, **kwargs): wtimer.start() for prod_key, prod_write in write_del: if prod_write: - if self.write_hdf5: - # Non-standard HDF5 output - fname = os.path.join(self.output_dir, "{}.h5".format(prod_key)) - write_healpix_hdf5( - data[prod_key], - fname, - nest=map_binning.pixel_pointing.nest, - single_precision=True, - force_serial=self.write_hdf5_serial, - ) - else: - # Standard FITS output + if is_pix_wcs: fname = os.path.join(self.output_dir, "{}.fits".format(prod_key)) - write_healpix_fits( - data[prod_key], - fname, - nest=map_binning.pixel_pointing.nest, - report_memory=self.report_memory, - ) + write_wcs_fits(data[prod_key], fname) + else: + if self.write_hdf5: + # Non-standard HDF5 output + fname = os.path.join(self.output_dir, "{}.h5".format(prod_key)) + write_healpix_hdf5( + data[prod_key], + fname, + nest=is_hpix_nest, + single_precision=True, + force_serial=self.write_hdf5_serial, + ) + else: + # Standard FITS output + fname = os.path.join( + self.output_dir, "{}.fits".format(prod_key) + ) + write_healpix_fits( + data[prod_key], + fname, + nest=is_hpix_nest, + report_memory=self.report_memory, + ) log.info_rank(f"Wrote {fname} in", comm=comm, timer=wtimer) if not self.keep_final_products: if prod_key in data: diff --git a/src/toast/ops/mapmaker_templates.py b/src/toast/ops/mapmaker_templates.py index ef0bb3469..b2f16d22a 100644 --- a/src/toast/ops/mapmaker_templates.py +++ b/src/toast/ops/mapmaker_templates.py @@ -11,7 +11,8 @@ from ..mpi import MPI from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import write_healpix_fits, write_healpix_hdf5 +from ..pixels_io_healpix import write_healpix_fits, write_healpix_hdf5 +from ..pixels_io_wcs import write_wcs_fits from ..templates import AmplitudesMap, Template from ..timing import Timer, function_timer from ..traits import Bool, Float, Instance, Int, List, Unicode, trait_docs @@ -785,6 +786,15 @@ def _exec(self, data, detectors=None, **kwargs): memreport.prefix = "After solving for amplitudes" memreport.apply(data) + # FIXME: This I/O technique assumes "known" types of pixel representations. + # Instead, we should associate read / write functions to a particular pixel + # class. + + is_pix_wcs = hasattr(self.binning.pixel_pointing, "wcs") + is_hpix_nest = None + if not is_pix_wcs: + is_hpix_nest = self.binning.pixel_pointing.nest + write_del = [ self.solver_hits_name, self.solver_cov_name, @@ -794,25 +804,31 @@ def _exec(self, data, detectors=None, **kwargs): ] for prod_key in write_del: if self.write_solver_products: - if self.write_hdf5: - # Non-standard HDF5 output - fname = os.path.join(self.output_dir, "{}.h5".format(prod_key)) - write_healpix_hdf5( - data[prod_key], - fname, - nest=self.binning.pixel_pointing.nest, - single_precision=True, - force_serial=self.write_hdf5_serial, - ) - else: - # Standard FITS output + if is_pix_wcs: fname = os.path.join(self.output_dir, "{}.fits".format(prod_key)) - write_healpix_fits( - data[prod_key], - fname, - nest=self.binning.pixel_pointing.nest, - report_memory=self.report_memory, - ) + write_wcs_fits(data[prod_key], fname) + else: + if self.write_hdf5: + # Non-standard HDF5 output + fname = os.path.join(self.output_dir, "{}.h5".format(prod_key)) + write_healpix_hdf5( + data[prod_key], + fname, + nest=is_hpix_nest, + single_precision=True, + force_serial=self.write_hdf5_serial, + ) + else: + # Standard FITS output + fname = os.path.join( + self.output_dir, "{}.fits".format(prod_key) + ) + write_healpix_fits( + data[prod_key], + fname, + nest=is_hpix_nest, + report_memory=self.report_memory, + ) if not self.mc_mode and not self.keep_solver_products: if prod_key in data: data[prod_key].clear() diff --git a/src/toast/ops/noise_estimation.py b/src/toast/ops/noise_estimation.py index b9f797231..c8cfba045 100644 --- a/src/toast/ops/noise_estimation.py +++ b/src/toast/ops/noise_estimation.py @@ -128,7 +128,8 @@ class NoiseEstim(Operator): ) output_dir = Unicode( - ".", + None, + allow_none=True, help="If specified, write output data products to this directory", ) @@ -1075,21 +1076,21 @@ def process_noise_estimate( ) log.debug_rank("Discard outliers", timer=timer) - self.save_psds( - binfreq, all_psds, all_times, det1, det2, fsample, fileroot, all_cov - ) - - if nbad > 0: + if self.output_dir is not None: self.save_psds( - binfreq, - good_psds, - good_times, - det1, - det2, - fsample, - fileroot + "_good", - good_cov, + binfreq, all_psds, all_times, det1, det2, fsample, fileroot, all_cov ) + if nbad > 0: + self.save_psds( + binfreq, + good_psds, + good_times, + det1, + det2, + fsample, + fileroot + "_good", + good_cov, + ) final_freqs = binfreq final_psd = np.mean(np.array(good_psds), axis=0) diff --git a/src/toast/ops/noise_model.py b/src/toast/ops/noise_model.py index 4686482ca..7589acaab 100644 --- a/src/toast/ops/noise_model.py +++ b/src/toast/ops/noise_model.py @@ -5,7 +5,7 @@ import numpy as np import traitlets from astropy import units as u -from scipy.optimize import Bounds, least_squares +from scipy.optimize import least_squares from ..noise import Noise from ..noise_sim import AnalyticNoise @@ -105,8 +105,12 @@ def _exec(self, data, detectors=None, **kwargs): freqs = in_model.freq(det) in_psd = in_model.psd(det) fitted, result = self._fit_psd(freqs, in_psd, params) - if result.success: + if result.success and len(result.x) == 3: + # This was a good fit params = result.x + else: + msg = f"FitNoiseModel observation {ob.name}, det {det} failed. Params = {result.x}" + log.warning(msg) nse_freqs[det] = freqs nse_psds[det] = fitted if ob.comm.comm_group is not None: @@ -156,12 +160,18 @@ def _evaluate_model(self, freqs, fmin, net, fknee, alpha): return psd def _fit_fun(self, x, *args, **kwargs): - net = x[0] - fknee = x[1] - alpha = x[2] freqs = kwargs["freqs"] data = kwargs["data"] fmin = kwargs["fmin"] + if "net" in kwargs: + # We are fixing the NET value + net = kwargs["net"] + fknee = x[0] + alpha = x[1] + else: + net = x[0] + fknee = x[1] + alpha = x[2] current = self._evaluate_model(freqs, fmin, net, fknee, alpha) # We weight the residual so that the high-frequency values specifying # the white noise plateau / NET are more important. Also the lowest @@ -186,6 +196,9 @@ def _fit_psd(self, freqs, data, guess=None): result = least_squares( self._fit_fun, x_0, + xtol=1.0e-8, + gtol=1.0e-8, + ftol=1.0e-8, kwargs={"freqs": raw_freqs, "data": raw_data, "fmin": raw_fmin}, ) fit_data = data @@ -196,6 +209,31 @@ def _fit_psd(self, freqs, data, guess=None): ) * psd_unit ) + # else: + # # Try fixing the NET based on the last few elements + # fixed_net = np.sqrt(np.mean(raw_data[-5:])) + # x_0 = np.array([0.1, 1.0]) + # result = least_squares( + # self._fit_fun, + # x_0, + # xtol=1.0e-8, + # gtol=1.0e-8, + # ftol=1.0e-8, + # kwargs={ + # "freqs": raw_freqs, + # "data": raw_data, + # "fmin": raw_fmin, + # "net": fixed_net, + # }, + # ) + # if result.success: + # fit_data = ( + # self._evaluate_model( + # raw_freqs, raw_fmin, fixed_net, result.x[0], result.x[1] + # ) + # * psd_unit + # ) + return fit_data, result def _finalize(self, data, **kwargs): diff --git a/src/toast/ops/pixels_wcs.py b/src/toast/ops/pixels_wcs.py new file mode 100644 index 000000000..034eda6d4 --- /dev/null +++ b/src/toast/ops/pixels_wcs.py @@ -0,0 +1,577 @@ +# Copyright (c) 2015-2020 by the parties listed in the AUTHORS file. +# All rights reserved. Use of this source code is governed by +# a BSD-style license that can be found in the LICENSE file. + +import warnings + +import numpy as np +import pixell +import pixell.enmap +import traitlets +from astropy import units as u + +from .. import qarray as qa +from ..mpi import MPI +from ..observation import default_values as defaults +from ..pixels import PixelDistribution +from ..timing import function_timer +from ..traits import Bool, Instance, Int, Tuple, Unicode, trait_docs +from ..utils import Environment, Logger +from .delete import Delete +from .operator import Operator + + +@trait_docs +class PixelsWCS(Operator): + """Operator which generates detector pixel indices defined on a flat projection. + + When placing the projection on the sky, either the `center` or `bounds` + traits must be specified, but not both. + + When determining the pixel density in the projection, exactly two traits from the + set of `bounds`, `resolution` and `dimensions` must be specified. + + If the view trait is not specified, then this operator will use the same data + view as the detector pointing operator when computing the pointing matrix pixels. + + This uses the pixell package to construct the WCS projection parameters. By + default, the world to pixel conversion is performed with internal, optimized code + unless use_astropy is set to True. + + """ + + # Class traits + + API = Int(0, help="Internal interface version for this operator") + + detector_pointing = Instance( + klass=Operator, + allow_none=True, + help="Operator that translates boresight pointing into detector frame", + ) + + projection = Unicode("CAR", help="Supported values are CAR, CEA, MER, ZEA, TAN") + + center = Tuple( + (180 * u.degree, 0 * u.degree), + allow_none=True, + help="The center Lon/Lat coordinates (Quantities) of the projection", + ) + + center_offset = Unicode( + None, + allow_none=True, + help="Optional name of shared field with lon, lat offset in degrees", + ) + + bounds = Tuple( + None, + allow_none=True, + help="The lower left and upper right Lon/Lat corners (Quantities)", + ) + + auto_bounds = Bool( + False, + help="If True, set the bounding box based on boresight and field of view", + ) + + dimensions = Tuple( + (710, 350), + allow_none=True, + help="The Lon/Lat pixel dimensions of the projection", + ) + + resolution = Tuple( + (0.5 * u.degree, 0.5 * u.degree), + allow_none=True, + help="The Lon/Lat projection resolution (Quantities) along the 2 axes", + ) + + view = Unicode( + None, allow_none=True, help="Use this view of the data in all observations" + ) + + pixels = Unicode("pixels", help="Observation detdata key for output pixel indices") + + quats = Unicode( + None, + allow_none=True, + help="Observation detdata key for output quaternions", + ) + + submaps = Int(10, help="Number of submaps to use") + + create_dist = Unicode( + None, + allow_none=True, + help="Create the submap distribution for all detectors and store in the Data key specified", + ) + + single_precision = Bool(False, help="If True, use 32bit int in output") + + use_astropy = Bool(False, help="If True, use astropy for world to pix conversion") + + @traitlets.validate("detector_pointing") + def _check_detector_pointing(self, proposal): + detpointing = proposal["value"] + if detpointing is not None: + if not isinstance(detpointing, Operator): + raise traitlets.TraitError( + "detector_pointing should be an Operator instance" + ) + # Check that this operator has the traits we expect + for trt in [ + "view", + "boresight", + "shared_flags", + "shared_flag_mask", + "quats", + "coord_in", + "coord_out", + ]: + if not detpointing.has_trait(trt): + msg = f"detector_pointing operator should have a '{trt}' trait" + raise traitlets.TraitError(msg) + return detpointing + + @traitlets.validate("wcs_projection") + def _check_wcs_projection(self, proposal): + check = proposal["value"] + if check not in ["CAR", "CEA", "MER", "ZEA", "TAN"]: + raise traitlets.TraitError("Invalid WCS projection name") + return check + + def __init__(self, **kwargs): + super().__init__(**kwargs) + # If running with all default values, the 'observe' function will not + # have been called yet. + if not hasattr(self, "_local_submaps"): + self._set_wcs( + self.projection, + self.center, + self.bounds, + self.dimensions, + self.resolution, + ) + self._done_auto = False + + @traitlets.observe("auto_bounds") + def _reset_auto_bounds(self, change): + # Track whether we need to recompute the bounds. + if change["new"]: + # enabling + self._done_auto = False + else: + self._done_auto = True + + @traitlets.observe("center_offset") + def _reset_auto_bounds(self, change): + # Track whether we need to recompute the bounds. + if change["new"] is not None: + if self.auto_bounds: + self._done_auto = False + + @traitlets.observe("projection", "center", "bounds", "dimensions", "resolution") + def _reset_wcs(self, change): + # (Re-)initialize the WCS projection when one of these traits change. + # Current values: + proj = str(self.projection) + center = self.center + if center is not None: + center = tuple(self.center) + bounds = self.bounds + if bounds is not None: + bounds = tuple(self.bounds) + dims = self.dimensions + if dims is not None: + dims = tuple(self.dimensions) + res = self.resolution + if res is not None: + res = tuple(self.resolution) + + # Update to the trait that changed + if change["name"] == "projection": + proj = change["new"] + if change["name"] == "center": + center = change["new"] + if center is not None: + bounds = None + if change["name"] == "bounds": + bounds = change["new"] + if bounds is not None: + center = None + if dims is not None and res is not None: + # Most likely the user cares about the resolution more... + dims = None + if change["name"] == "dimensions": + dims = change["new"] + if dims is not None and bounds is not None: + res = None + if change["name"] == "resolution": + res = change["new"] + if res is not None and bounds is not None: + dims = None + self._set_wcs(proj, center, bounds, dims, res) + self.projection = proj + self.center = center + self.bounds = bounds + self.dimensions = dims + self.resolution = res + + def _set_wcs(self, proj, center, bounds, dims, res): + log = Logger.get() + log.verbose(f"PixelsWCS: set_wcs {proj}, {center}, {bounds}, {dims}, {res}") + if res is not None: + res = np.array( + [ + res[0].to_value(u.degree), + res[1].to_value(u.degree), + ] + ) + if dims is not None: + dims = np.array([self.dimensions[0], self.dimensions[1]]) + + if bounds is None: + # Using center, need both resolution and dimensions + if center is None: + # Cannot calculate yet + return + if res is None or dims is None: + # Cannot calculate yet + return + pos = np.array( + [ + center[0].to_value(u.degree), + center[1].to_value(u.degree), + ] + ) + else: + # Using bounds, exactly one of resolution or dimensions specified + if res is not None and dims is not None: + # Cannot calculate yet + return + pos = np.array( + [ + [ + bounds[0][0].to_value(u.degree), + bounds[0][1].to_value(u.degree), + ], + [ + bounds[1][0].to_value(u.degree), + bounds[1][1].to_value(u.degree), + ], + ] + ) + + self.wcs = pixell.wcsutils.build(pos, res, dims, rowmajor=False, system=proj) + if dims is None: + # Compute from the bounding box corners + lower_left = self.wcs.wcs_world2pix(np.array([[pos[0, 0], pos[0, 1]]]), 0)[ + 0 + ] + upper_right = self.wcs.wcs_world2pix(np.array([[pos[1, 0], pos[1, 1]]]), 0)[ + 0 + ] + self.wcs_shape = tuple( + np.round(np.abs(upper_right - lower_left)).astype(int) + ) + else: + self.wcs_shape = tuple(dims) + log.verbose(f"PixelsWCS: wcs_shape = {self.wcs_shape}") + + self.pix_ra = self.wcs_shape[0] + self.pix_dec = self.wcs_shape[1] + self._n_pix = self.pix_ra * self.pix_dec + self._n_pix_submap = self._n_pix // self.submaps + if self._n_pix_submap * self.submaps < self._n_pix: + self._n_pix_submap += 1 + self._local_submaps = None + return + + @function_timer + def _exec(self, data, detectors=None, **kwargs): + env = Environment.get() + log = Logger.get() + + if self.detector_pointing is None: + raise RuntimeError("The detector_pointing trait must be set") + + if not self.use_astropy: + raise NotImplementedError("Only astropy conversion is currently supported") + + if self.auto_bounds and not self._done_auto: + # Pass through the boresight pointing for every observation and build + # the maximum extent of the detector field of view. + lonmax = 0 * u.radian + lonmin = 2 * np.pi * u.radian + latmax = (-np.pi / 2) * u.radian + latmin = (np.pi / 2) * u.radian + for ob in data.obs: + # Get the detectors we are using for this observation + dets = ob.select_local_detectors(detectors) + if len(dets) == 0: + # Nothing to do for this observation + continue + lnmin, lnmax, ltmin, ltmax = self._get_scan_range(ob) + lonmin = min(lonmin, lnmin) + lonmax = max(lonmax, lnmax) + latmin = min(latmin, ltmin) + latmax = max(latmax, ltmax) + if ob.comm.comm_group_rank is not None: + # Synchronize between groups + if ob.comm.group_rank == 0: + lonmin = ob.comm.comm_group_rank.allreduce(lonmin, op=MPI.MIN) + latmin = ob.comm.comm_group_rank.allreduce(latmin, op=MPI.MIN) + lonmax = ob.comm.comm_group_rank.allreduce(lonmax, op=MPI.MAX) + latmax = ob.comm.comm_group_rank.allreduce(latmax, op=MPI.MAX) + # Broadcast result within the group + if ob.comm.comm_group is not None: + lonmin = ob.comm.comm_group.bcast(lonmin, root=0) + lonmax = ob.comm.comm_group.bcast(lonmax, root=0) + latmin = ob.comm.comm_group.bcast(latmin, root=0) + latmax = ob.comm.comm_group.bcast(latmax, root=0) + new_bounds = ( + (lonmax.to(u.degree), latmin.to(u.degree)), + (lonmin.to(u.degree), latmax.to(u.degree)), + ) + log.verbose(f"PixelsWCS auto_bounds set to {new_bounds}") + self.bounds = new_bounds + self._done_auto = True + + if self._local_submaps is None and self.create_dist is not None: + self._local_submaps = np.zeros(self.submaps, dtype=np.bool) + + # Expand detector pointing + if self.quats is not None: + quats_name = self.quats + else: + if self.detector_pointing.quats is not None: + quats_name = self.detector_pointing.quats + else: + quats_name = defaults.quats + + view = self.view + if view is None: + # Use the same data view as detector pointing + view = self.detector_pointing.view + + self.detector_pointing.quats = quats_name + self.detector_pointing.apply(data, detectors=detectors) + + for ob in data.obs: + # Get the detectors we are using for this observation + dets = ob.select_local_detectors(detectors) + if len(dets) == 0: + # Nothing to do for this observation + continue + + # FIXME: remove this workaround after #557 is merged + if view is None: + view_slices = [slice(None)] + else: + view_slices = [ + slice(x.first, x.last + 1, 1) for x in ob.intervals[view] + ] + + # Create (or re-use) output data for the pixels, weights and optionally the + # detector quaternions. + + if self.single_precision: + exists = ob.detdata.ensure( + self.pixels, sample_shape=(), dtype=np.int32, detectors=dets + ) + else: + exists = ob.detdata.ensure( + self.pixels, sample_shape=(), dtype=np.int64, detectors=dets + ) + + # Do we already have pointing for all requested detectors? + if exists: + # Yes... + if self.create_dist is not None: + # but the caller wants the pixel distribution + for det in ob.select_local_detectors(detectors): + for vslice in view_slices: + good = ob.detdata[self.pixels][det][vslice] >= 0 + self._local_submaps[ + ob.detdata[self.pixels][det][vslice][good] + // self._n_pix_submap + ] = True + + if data.comm.group_rank == 0: + msg = ( + f"Group {data.comm.group}, ob {ob.name}, WCS pixels " + f"already computed for {dets}" + ) + log.verbose(msg) + continue + + # Focalplane for this observation + focalplane = ob.telescope.focalplane + + # Get the flags if needed. Use the same flags as + # detector pointing. + flags = None + if self.detector_pointing.shared_flags is not None: + flags = np.array(ob.shared[self.detector_pointing.shared_flags]) + flags &= self.detector_pointing.shared_flag_mask + n_good = np.sum(flags == 0) + n_bad = np.sum(flags != 0) + + center_lonlat = None + if self.center_offset is not None: + center_lonlat = ob.shared[self.center_offset].data + + # Process all detectors + for det in ob.select_local_detectors(detectors): + for vslice in view_slices: + # Timestream of detector quaternions + quats = ob.detdata[quats_name][det][vslice] + view_samples = len(quats) + + # print(f"det {det} quats = {quats}") + theta, phi = qa.to_position(quats) + # print(f"det {det} rad theta, phi = {theta}, {phi}") + + to_deg = 180.0 / np.pi + theta *= to_deg + phi *= to_deg + # print(f"det {det} deg theta, phi = {theta}, {phi}") + + world_in = np.column_stack([phi, 90.0 - theta]) + + if center_lonlat is not None: + # print(f"orig = {world_in}") + # print(f"center_lonlat = {center_lonlat}") + world_in[:, 0] -= center_lonlat[vslice, 0] + world_in[:, 1] -= center_lonlat[vslice, 1] + # print(f"final = {world_in}") + + # print(f"det {det} world_in = {world_in}") + rdpix = self.wcs.wcs_world2pix(world_in, 0) + if flags is not None: + # Set bad pointing to pixel -1 + bad_pointing = flags[vslice] != 0 + rdpix[bad_pointing] = -1 + rdpix = np.array(np.around(rdpix), dtype=np.int64) + # print(f"det {det} rdpix = {rdpix}") + + ob.detdata[self.pixels][det][vslice] = ( + rdpix[:, 0] * self.pix_dec + rdpix[:, 1] + ) + + if self.create_dist is not None: + good = ob.detdata[self.pixels][det][vslice] >= 0 + # print(f"det {det} has {np.sum(good)} good pixels") + # print((ob.detdata[self.pixels][det][vslice])[good]) + self._local_submaps[ + (ob.detdata[self.pixels][det][vslice])[good] + // self._n_pix_submap + ] = 1 + + @function_timer + def _get_scan_range(self, obs): + # FIXME: mostly copied from the atmosphere simulation code- we should + # extract this into a more general helper routine somewhere. + fov = obs.telescope.focalplane.field_of_view + fp_radius = 0.5 * fov.to_value(u.radian) + + # work in parallel + rank = obs.comm.group_rank + ntask = obs.comm.group_size + + # Create a fake focalplane of detectors in a circle around the boresight + xaxis, yaxis, zaxis = np.eye(3) + ndet = 64 + phidet = np.linspace(0, 2 * np.pi, ndet, endpoint=False) + detquats = [] + thetarot = qa.rotation(yaxis, fp_radius) + for phi in phidet: + phirot = qa.rotation(zaxis, phi) + detquat = qa.mult(phirot, thetarot) + detquats.append(detquat) + + # Get fake detector pointing + + center_lonlat = None + if self.center_offset is not None: + center_lonlat = np.array(obs.shared[self.center_offset].data) + center_lonlat[:, :] *= np.pi / 180.0 + + lon = [] + lat = [] + quats = obs.shared[self.detector_pointing.boresight][rank::ntask].copy() + for idet, detquat in enumerate(detquats): + theta, phi = qa.to_position(qa.mult(quats, detquat)) + if center_lonlat is None: + lon.append(phi) + lat.append(np.pi / 2 - theta) + else: + lon.append(phi - center_lonlat[rank::ntask, 0]) + lat.append((np.pi / 2 - theta) - center_lonlat[rank::ntask, 1]) + lon = np.unwrap(np.hstack(lon)) + lat = np.hstack(lat) + + # find the extremes + lonmin = np.amin(lon) + lonmax = np.amax(lon) + latmin = np.amin(lat) + latmax = np.amax(lat) + + if lonmin < -2 * np.pi: + lonmin += 2 * np.pi + lonmax += 2 * np.pi + elif lonmax > 2 * np.pi: + lonmin -= 2 * np.pi + lonmax -= 2 * np.pi + + # Combine results + if obs.comm.comm_group is not None: + lonmin = obs.comm.comm_group.allreduce(lonmin, op=MPI.MIN) + lonmax = obs.comm.comm_group.allreduce(lonmax, op=MPI.MAX) + latmin = obs.comm.comm_group.allreduce(latmin, op=MPI.MIN) + latmax = obs.comm.comm_group.allreduce(latmax, op=MPI.MAX) + + return ( + lonmin * u.radian, + lonmax * u.radian, + latmin * u.radian, + latmax * u.radian, + ) + + def _finalize(self, data, **kwargs): + if self.create_dist is not None: + submaps = None + if self.single_precision: + submaps = np.arange(self.submaps, dtype=np.int32)[ + self._local_submaps == 1 + ] + else: + submaps = np.arange(self.submaps, dtype=np.int64)[ + self._local_submaps == 1 + ] + + # print(f"create WCS pixdist {self._n_pix}, {self.submaps}, {submaps}") + data[self.create_dist] = PixelDistribution( + n_pix=self._n_pix, + n_submap=self.submaps, + local_submaps=submaps, + comm=data.comm.comm_world, + ) + # Store a copy of the WCS information in the distribution object + data[self.create_dist].wcs = self.wcs.deepcopy() + data[self.create_dist].wcs_shape = tuple(self.wcs_shape) + return + + def _requires(self): + req = self.detector_pointing.requires() + if self.view is not None: + req["intervals"].append(self.view) + return req + + def _provides(self): + prov = self.detector_pointing.provides() + prov["detdata"].append(self.pixels) + if self.create_dist is not None: + prov["global"].append(self.create_dist) + return prov diff --git a/src/toast/ops/reset.py b/src/toast/ops/reset.py index 380cfb6c3..457540875 100644 --- a/src/toast/ops/reset.py +++ b/src/toast/ops/reset.py @@ -62,11 +62,6 @@ def _exec(self, data, detectors=None, **kwargs): for key in self.detdata: for d in dets: ob.detdata[key][d, :] = 0 - print( - "Reset detdata {}, dets {}, result = ".format(key, dets), - ob.detdata[key], - flush=True, - ) if self.shared is not None: for key in self.shared: scomm = ob.shared[key].nodecomm diff --git a/src/toast/ops/scan_healpix.py b/src/toast/ops/scan_healpix.py index 7f86b4493..d6168d02d 100644 --- a/src/toast/ops/scan_healpix.py +++ b/src/toast/ops/scan_healpix.py @@ -7,7 +7,7 @@ from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import ( +from ..pixels_io_healpix import ( filename_is_fits, filename_is_hdf5, read_healpix_fits, diff --git a/src/toast/ops/scan_wcs.py b/src/toast/ops/scan_wcs.py new file mode 100644 index 000000000..4f84712a3 --- /dev/null +++ b/src/toast/ops/scan_wcs.py @@ -0,0 +1,344 @@ +# Copyright (c) 2015-2020 by the parties listed in the AUTHORS file. +# All rights reserved. Use of this source code is governed by +# a BSD-style license that can be found in the LICENSE file. + +import numpy as np +import traitlets + +from ..observation import default_values as defaults +from ..pixels import PixelData, PixelDistribution +from ..pixels_io_wcs import read_wcs_fits +from ..timing import function_timer +from ..traits import Bool, Instance, Int, Unicode, trait_docs +from ..utils import Logger +from .operator import Operator +from .pipeline import Pipeline +from .pointing import BuildPixelDistribution +from .scan_map import ScanMap, ScanMask + + +@trait_docs +class ScanWCSMap(Operator): + """Operator which reads a WCS format map from disk and scans it to a timestream. + + The map file is loaded and distributed among the processes. For each observation, + the pointing model is used to expand the pointing and scan the map values into + detector data. + + """ + + # Class traits + + API = Int(0, help="Internal interface version for this operator") + + file = Unicode(None, allow_none=True, help="Path to FITS file") + + det_data = Unicode( + defaults.det_data, help="Observation detdata key for accumulating output" + ) + + subtract = Bool( + False, help="If True, subtract the map timestream instead of accumulating" + ) + + zero = Bool(False, help="If True, zero the data before accumulating / subtracting") + + pixel_dist = Unicode( + "pixel_dist", + help="The Data key where the PixelDistribution object is located", + ) + + pixel_pointing = Instance( + klass=Operator, + allow_none=True, + help="This must be an instance of a pixel pointing operator", + ) + + stokes_weights = Instance( + klass=Operator, + allow_none=True, + help="This must be an instance of a Stokes weights operator", + ) + + save_map = Bool(False, help="If True, do not delete map during finalize") + + save_pointing = Bool( + False, + help="If True, do not clear detector pointing matrices if we " + "generate the pixel distribution", + ) + + @traitlets.validate("pixel_pointing") + def _check_pixel_pointing(self, proposal): + pixels = proposal["value"] + if pixels is not None: + if not isinstance(pixels, Operator): + raise traitlets.TraitError( + "pixel_pointing should be an Operator instance" + ) + # Check that this operator has the traits we expect + for trt in ["pixels", "create_dist", "view"]: + if not pixels.has_trait(trt): + msg = f"pixel_pointing operator should have a '{trt}' trait" + raise traitlets.TraitError(msg) + return pixels + + @traitlets.validate("stokes_weights") + def _check_stokes_weights(self, proposal): + weights = proposal["value"] + if weights is not None: + if not isinstance(weights, Operator): + raise traitlets.TraitError( + "stokes_weights should be an Operator instance" + ) + # Check that this operator has the traits we expect + for trt in ["weights", "view"]: + if not weights.has_trait(trt): + msg = f"stokes_weights operator should have a '{trt}' trait" + raise traitlets.TraitError(msg) + return weights + + def __init__(self, **kwargs): + super().__init__(**kwargs) + self.map_name = "{}_map".format(self.name) + + @function_timer + def _exec(self, data, detectors=None, **kwargs): + log = Logger.get() + + # Check that the file is set + if self.file is None: + raise RuntimeError("You must set the file trait before calling exec()") + + # Construct the pointing distribution if it does not already exist + + if self.pixel_dist not in data: + pix_dist = BuildPixelDistribution( + pixel_dist=self.pixel_dist, + pixel_pointing=self.pixel_pointing, + save_pointing=self.save_pointing, + ) + pix_dist.apply(data) + + dist = data[self.pixel_dist] + if not isinstance(dist, PixelDistribution): + raise RuntimeError("The pixel_dist must be a PixelDistribution instance") + + # Use the pixel odistribution and pointing configuration to allocate our + # map data and read it in. + nnz = None + if self.stokes_weights is None or self.stokes_weights.mode == "I": + nnz = 1 + elif self.stokes_weights.mode == "IQU": + nnz = 3 + else: + msg = "Unknown Stokes weights mode '{}'".format(self.stokes_weights.mode) + raise RuntimeError(msg) + + # Create our map to scan named after our own operator name. Generally the + # files on disk are stored as float32, but even if not there is no real benefit + # to having higher precision to simulated map signal that is projected into + # timestreams. + if self.map_name not in data: + data[self.map_name] = PixelData(dist, dtype=np.float32, n_value=nnz) + read_wcs_fits(data[self.map_name], self.file) + + # The pipeline below will run one detector at a time in case we are computing + # pointing. Make sure that our full set of requested detector output exists. + # FIXME: This seems like a common pattern, maybe move to a "Create" operator? + for ob in data.obs: + # Get the detectors we are using for this observation + dets = ob.select_local_detectors(detectors) + if len(dets) == 0: + # Nothing to do for this observation + continue + # If our output detector data does not yet exist, create it + exists_data = ob.detdata.ensure(self.det_data, detectors=dets) + + # Configure the low-level map scanning operator + + scanner = ScanMap( + det_data=self.det_data, + pixels=self.pixel_pointing.pixels, + weights=self.stokes_weights.weights, + map_key=self.map_name, + subtract=self.subtract, + zero=self.zero, + ) + + # Build and run a pipeline that scans from our map + scan_pipe = Pipeline( + detector_sets=["SINGLE"], + operators=[self.pixel_pointing, self.stokes_weights, scanner], + ) + scan_pipe.apply(data, detectors=detectors) + + return + + def _finalize(self, data, **kwargs): + # Clean up our map, if needed + if not self.save_map: + data[self.map_name].clear() + del data[self.map_name] + return + + def _requires(self): + req = self.pixel_pointing.requires() + req.update(self.stokes_weights.requires()) + return req + + def _provides(self): + prov = {"global": list(), "detdata": [self.det_data]} + if self.save_map: + prov["global"] = [self.map_name] + return prov + + +@trait_docs +class ScanWCSMask(Operator): + """Operator which reads a WCS mask from disk and scans it to a timestream. + + The mask file is loaded and distributed among the processes. For each observation, + the pointing model is used to expand the pointing and scan the mask values into + detector data. + + """ + + # Class traits + + API = Int(0, help="Internal interface version for this operator") + + file = Unicode(None, allow_none=True, help="Path to FITS file") + + det_flags = Unicode( + defaults.det_flags, + allow_none=True, + help="Observation detdata key for flags to use", + ) + + det_flags_value = Int( + 1, help="The detector flag value to set where the mask result is non-zero" + ) + + mask_bits = Int( + 255, help="The number to bitwise-and with each mask value to form the result" + ) + + pixel_dist = Unicode( + "pixel_dist", + help="The Data key where the PixelDistribution object is located", + ) + + pixel_pointing = Instance( + klass=Operator, + allow_none=True, + help="This must be an instance of a pixel pointing operator", + ) + + save_mask = Bool(False, help="If True, do not delete mask during finalize") + + save_pointing = Bool( + False, + help="If True, do not clear detector pointing matrices if we " + "generate the pixel distribution", + ) + + @traitlets.validate("pixel_pointing") + def _check_pixel_pointing(self, proposal): + pixels = proposal["value"] + if pixels is not None: + if not isinstance(pixels, Operator): + raise traitlets.TraitError( + "pixel_pointing should be an Operator instance" + ) + # Check that this operator has the traits we expect + for trt in ["pixels", "create_dist", "view"]: + if not pixels.has_trait(trt): + msg = f"pixel_pointing operator should have a '{trt}' trait" + raise traitlets.TraitError(msg) + return pixels + + def __init__(self, **kwargs): + super().__init__(**kwargs) + self.mask_name = "{}_mask".format(self.name) + + @function_timer + def _exec(self, data, detectors=None, **kwargs): + log = Logger.get() + + # Check that the file is set + if self.file is None: + raise RuntimeError("You must set the file trait before calling exec()") + + # Construct the pointing distribution if it does not already exist + + if self.pixel_dist not in data: + pix_dist = BuildPixelDistribution( + pixel_dist=self.pixel_dist, + pixel_pointing=self.pixel_pointing, + save_pointing=self.save_pointing, + ) + pix_dist.apply(data) + + dist = data[self.pixel_dist] + if not isinstance(dist, PixelDistribution): + raise RuntimeError("The pixel_dist must be a PixelDistribution instance") + + # Create our map to scan named after our own operator name. Generally the + # files on disk are stored as float32, but even if not there is no real benefit + # to having higher precision to simulated map signal that is projected into + # timestreams. + if self.mask_name not in data: + data[self.mask_name] = PixelData(dist, dtype=np.uint8, n_value=1) + read_wcs_fits(data[self.mask_name], self.file) + + # The pipeline below will run one detector at a time in case we are computing + # pointing. Make sure that our full set of requested detector output exists. + # FIXME: This seems like a common pattern, maybe move to a "Create" operator? + for ob in data.obs: + # Get the detectors we are using for this observation + dets = ob.select_local_detectors(detectors) + if len(dets) == 0: + # Nothing to do for this observation + continue + # If our output detector data does not yet exist, create it + exists_flags = ob.detdata.ensure( + self.det_flags, dtype=np.uint8, detectors=dets + ) + + # Configure the low-level map scanning operator + + scanner = ScanMask( + det_flags=self.det_flags, + det_flags_value=self.det_flags_value, + pixels=self.pixel_pointing.pixels, + mask_key=self.mask_name, + mask_bits=self.mask_bits, + ) + + # Build and run a pipeline that scans from our map + scan_pipe = Pipeline( + detector_sets=["SINGLE"], + operators=[self.pixel_pointing, scanner], + ) + scan_pipe.apply(data, detectors=detectors) + + return + + def _finalize(self, data, **kwargs): + # Clean up our map, if needed + if not self.save_mask: + data[self.mask_name].clear() + del data[self.mask_name] + return + + def _requires(self): + req = self.pixel_pointing.requires() + req.update(self.stokes_weights.requires()) + return req + + def _provides(self): + prov = {"global": list(), "detdata": [self.det_data]} + if self.save_map: + prov["global"] = [self.map_name] + return prov diff --git a/src/toast/ops/sim_ground.py b/src/toast/ops/sim_ground.py index 1f8fcb825..81b0b7402 100644 --- a/src/toast/ops/sim_ground.py +++ b/src/toast/ops/sim_ground.py @@ -5,7 +5,6 @@ import copy import datetime -import healpy as hp import numpy as np import traitlets from astropy import units as u diff --git a/src/toast/ops/sim_satellite.py b/src/toast/ops/sim_satellite.py index 900a271e3..1c88566ec 100644 --- a/src/toast/ops/sim_satellite.py +++ b/src/toast/ops/sim_satellite.py @@ -2,7 +2,6 @@ # All rights reserved. Use of this source code is governed by # a BSD-style license that can be found in the LICENSE file. -import healpy as hp import numpy as np import traitlets from astropy import units as u diff --git a/src/toast/ops/yield_cut.py b/src/toast/ops/yield_cut.py index 86deb8569..13c1595db 100644 --- a/src/toast/ops/yield_cut.py +++ b/src/toast/ops/yield_cut.py @@ -4,7 +4,6 @@ from time import time -import healpy as hp import numpy as np import traitlets from astropy import units as u diff --git a/src/toast/pixels.py b/src/toast/pixels.py index a099fd049..1b326e623 100644 --- a/src/toast/pixels.py +++ b/src/toast/pixels.py @@ -316,6 +316,7 @@ def alltoallv_info(self): - The locations in the receive buffer of each submap. """ + log = Logger.get() if self._alltoallv_info is not None: # Already computed return self._alltoallv_info @@ -390,6 +391,17 @@ def alltoallv_info(self): recv_locations, ) + msg_rank = 0 + if self._comm is not None: + msg_rank = self._comm.rank + msg = f"alltoallv_info[{msg_rank}]:\n" + msg += f" send_counts={send_counts} " + msg += f"send_displ={send_displ}\n" + msg += f" recv_counts={recv_counts} " + msg += f"recv_displ={recv_displ} " + msg += f"recv_locations={recv_locations}" + log.verbose(msg) + return self._alltoallv_info @@ -744,6 +756,17 @@ def setup_alltoallv(self): for sm, locs in recv_locations.items(): self._recv_locations[sm] = scale * np.array(locs, dtype=np.int32) + msg_rank = 0 + if self._dist.comm is not None: + msg_rank = self._dist.comm.rank + msg = f"setup_alltoallv[{msg_rank}]:\n" + msg += f" send_counts={self._send_counts} " + msg += f"send_displ={self._send_displ}\n" + msg += f" recv_counts={self._recv_counts} " + msg += f"recv_displ={self._recv_displ} " + msg += f"recv_locations={self._recv_locations}" + log.verbose(msg) + # Allocate a persistent single-submap buffer self._reduce_buf_raw = self.storage_class.zeros(self._n_submap_value) self.reduce_buf = self._reduce_buf_raw.array() @@ -774,6 +797,9 @@ def setup_alltoallv(self): buf_check_fail = True # Allocate a persistent receive buffer + msg = f"{msg_rank}: allocate receive buffer of " + msg += f"{recv_buf_size} elements" + log.verbose(msg) self._receive_raw = self.storage_class.zeros(recv_buf_size) self.receive = self._receive_raw.array() except: @@ -796,6 +822,7 @@ def forward_alltoallv(self): None. """ + log = Logger.get() gt = GlobalTimers.get() self.setup_alltoallv() @@ -887,11 +914,11 @@ def broadcast_map(self, fdata, comm_bytes=10000000): rank = 0 if self._dist.comm is not None: rank = self._dist.comm.rank - comm_submap = self._dist.comm_nsubmap(comm_bytes) + comm_submap = self.comm_nsubmap(comm_bytes) # we make the assumption that FITS binary tables are still stored in # blocks of 2880 bytes just like always... - dbytes = self._dtype(1).itemsize + dbytes = self._dtype.itemsize rowbytes = self._n_value * dbytes optrows = int(2880 / rowbytes) @@ -918,8 +945,8 @@ def broadcast_map(self, fdata, comm_bytes=10000000): # is this the last block for this communication? islast = False copyrows = rows - if out_off + rows > (comm_submap * self._dist.npix_submap): - copyrows = (comm_submap * self._dist.npix_submap) - out_off + if out_off + rows > (comm_submap * self._dist.n_pix_submap): + copyrows = (comm_submap * self._dist.n_pix_submap) - out_off islast = True if rank == 0: diff --git a/src/toast/pixels_io.py b/src/toast/pixels_io_healpix.py similarity index 99% rename from src/toast/pixels_io.py rename to src/toast/pixels_io_healpix.py index a0a80cd7c..8552cb6b6 100644 --- a/src/toast/pixels_io.py +++ b/src/toast/pixels_io_healpix.py @@ -134,7 +134,7 @@ def read_healpix_fits(pix, path, nest=True, comm_bytes=10000000): @function_timer -def collect_submaps(pix, comm_bytes=10000000): +def collect_healpix_submaps(pix, comm_bytes=10000000): # The distribution dist = pix.distribution @@ -271,7 +271,7 @@ def write_healpix_fits(pix, path, nest=True, comm_bytes=10000000, report_memory= if dist.comm is not None: rank = dist.comm.rank - fdata, fview = collect_submaps(pix, comm_bytes=comm_bytes) + fdata, fview = collect_healpix_submaps(pix, comm_bytes=comm_bytes) if rank == 0: if os.path.isfile(path): diff --git a/src/toast/pixels_io_wcs.py b/src/toast/pixels_io_wcs.py new file mode 100644 index 000000000..fe74c02c8 --- /dev/null +++ b/src/toast/pixels_io_wcs.py @@ -0,0 +1,332 @@ +# Copyright (c) 2015-2021 by the parties listed in the AUTHORS file. +# All rights reserved. Use of this source code is governed by +# a BSD-style license that can be found in the LICENSE file. + +import os + +import astropy.io.fits as af +import numpy as np +import pixell +import pixell.enmap + +from .timing import function_timer, Timer +from .mpi import MPI, use_mpi +from .utils import Logger, memreport + + +def submap_to_enmap(dist, submap, sdata, endata): + """Helper function to unpack our data into pixell format + + This takes a single submap of data with flat pixels x n_values and unpacks + that into an ndmap with n_values x 2D pixels. The ndmaps are column + major, like a FITS image: + + | OOXXXOO + | OOXXXOO + | OXXXOOO + V OXXXOOO + + So a submap covers a range of columns, and can start and end in the middle + of a column. + + Args: + dist (PixelDistribution): The pixel dist + submap (int): The global submap index + sdata (array): The data for a single submap + endata (ndmap): The pixell array + + Returns: + None + + """ + enshape = endata.shape + n_value = enshape[0] + n_cols = enshape[1] + n_rows = enshape[2] + + # Global pixel range of this submap + s_offset = submap * dist.n_pix_submap + s_end = s_offset + dist.n_pix_submap + + # Find which ndmap cols are covered by this submap + first_col = s_offset // n_rows + last_col = s_end // n_rows + if last_col >= n_cols: + last_col = n_cols - 1 + + # Loop over output rows and assign data + # print(f"endata shape = {endata.shape}") + for ival in range(n_value): + for col in range(first_col, last_col + 1): + pix_offset = col * n_rows + row_offset = 0 + if s_offset > pix_offset: + row_offset = s_offset - pix_offset + n_copy = n_rows - row_offset + if pix_offset + n_copy > s_end: + n_copy = s_end - pix_offset + sbuf_offset = pix_offset + row_offset - s_offset + # print( + # f"endata[{ival}, {col}, {row_offset}:{row_offset+n_copy}] = sdata[{sbuf_offset}:{sbuf_offset+n_copy}, {ival}]", + # flush=True, + # ) + endata[ival, col, row_offset : row_offset + n_copy] = sdata[ + sbuf_offset : sbuf_offset + n_copy, ival + ] + + +def enmap_to_submap(dist, endata, submap, sdata): + """Helper function to fill our data from pixell format + + This takes a single submap of data with flat pixels x n_values and fills + it from an ndmap with n_values x 2D pixels. + + Args: + dist (PixelDistribution): The pixel dist + endata (ndmap): The pixell array + submap (int): The global submap index + sdata (array): The data for a single submap + + Returns: + None + + """ + enshape = endata.shape + n_value = enshape[0] + n_cols = enshape[1] + n_rows = enshape[2] + + # Global pixel range of this submap + s_offset = submap * dist.n_pix_submap + s_end = s_offset + dist.n_pix_submap + + # Find which ndmap rows are covered by this submap + first_col = s_offset // n_rows + last_col = s_end // n_rows + if last_col >= n_cols: + last_col = n_cols - 1 + + # Loop over output rows and assign data + for ival in range(n_value): + for col in range(first_col, last_col + 1): + pix_offset = col * n_rows + row_offset = 0 + if s_offset > pix_offset: + row_offset = s_offset - pix_offset + n_copy = n_rows - row_offset + if pix_offset + n_copy > s_end: + n_copy = s_end - pix_offset + sbuf_offset = pix_offset + row_offset - s_offset + # print( + # f"sdata[{sbuf_offset}:{sbuf_offset+n_copy}, {ival}] = endata[{ival}, {col}, {row_offset}:{row_offset+n_copy}]", + # flush=True, + # ) + sdata[sbuf_offset : sbuf_offset + n_copy, ival] = endata[ + ival, col, row_offset : row_offset + n_copy + ] + + +@function_timer +def collect_wcs_submaps(pix, comm_bytes=10000000): + # The distribution + dist = pix.distribution + + rank = 0 + if dist.comm is not None: + rank = dist.comm.rank + + # We will reduce some number of whole submaps at a time. + # Find the number of submaps that fit into the requested + # communication size. + comm_submap = pix.comm_nsubmap(comm_bytes) + + # Determine which processes should send submap data. We do not use the + # PixelDistribution.submap_owners here, since that is intended for operations + # parallelized over submaps, and the submap owners do not necessarily have the + # owned submaps in local memory. Instead, we do a buffered allreduce. + + not_owned = None + allowners = None + if dist.comm is None: + not_owned = 1 + allowners = np.zeros(dist.n_submap, dtype=np.int32) + allowners.fill(not_owned) + for m in dist.local_submaps: + allowners[m] = rank + else: + not_owned = dist.comm.size + owners = np.zeros(dist.n_submap, dtype=np.int32) + owners.fill(not_owned) + for m in dist.local_submaps: + owners[m] = dist.comm.rank + allowners = np.zeros_like(owners) + dist.comm.Allreduce(owners, allowners, op=MPI.MIN) + + # Create a pixell map structure for the output + endata = None + if rank == 0: + endata = pixell.enmap.zeros((pix.n_value,) + dist.wcs_shape, wcs=dist.wcs) + + n_val_submap = dist.n_pix_submap * pix.n_value + + if dist.comm is None: + # Just copy our local submaps into the pixell buffer + for lc, sm in enumerate(dist.local_submaps): + submap_to_enmap(dist, sm, pix.data[lc], endata) + else: + sendbuf = np.zeros(comm_submap * n_val_submap, dtype=pix.dtype) + sendview = sendbuf.reshape(comm_submap, dist.n_pix_submap, pix.n_value) + + recvbuf = None + recvview = None + if rank == 0: + recvbuf = np.zeros(comm_submap * n_val_submap, dtype=pix.dtype) + recvview = recvbuf.reshape(comm_submap, dist.n_pix_submap, pix.n_value) + + submap_off = 0 + ncomm = comm_submap + while submap_off < dist.n_submap: + if submap_off + ncomm > dist.n_submap: + ncomm = dist.n_submap - submap_off + if np.any(allowners[submap_off : submap_off + ncomm] != not_owned): + # at least one submap has some hits. reduce. + for c in range(ncomm): + if allowners[submap_off + c] == dist.comm.rank: + # print(f"rank {dist.comm.rank}: set sendview[{c}, :, :]") + sendview[c, :, :] = pix.data[ + dist.global_submap_to_local[submap_off + c], :, : + ] + dist.comm.Reduce(sendbuf, recvbuf, op=MPI.SUM, root=0) + if rank == 0: + # copy into pixell buffer + for c in range(ncomm): + submap_to_enmap(dist, (submap_off + c), recvview[c], endata) + sendbuf.fill(0) + if rank == 0: + recvbuf.fill(0) + submap_off += ncomm + return endata + + +@function_timer +def write_wcs_fits(pix, path, comm_bytes=10000000, report_memory=False): + """Write pixel data to a FITS image + + The data across all processes is assumed to be synchronized (the data for a given + submap shared between processes is identical). The submap data is sent to the root + process which writes it out. + + Args: + pix (PixelData): The distributed pixel object. + path (str): The path to the output FITS file. + comm_bytes (int): The approximate message size to use. + report_memory (bool): Report the amount of available memory on the root + node just before writing out the map. + + Returns: + None + + """ + log = Logger.get() + timer = Timer() + timer.start() + + # The distribution + dist = pix.distribution + + # Check that we have WCS information + if not hasattr(dist, "wcs"): + raise RuntimeError("Pixel distribution does not have WCS information") + + rank = 0 + if dist.comm is not None: + rank = dist.comm.rank + + endata = collect_wcs_submaps(pix, comm_bytes=comm_bytes) + + if rank == 0: + if os.path.isfile(path): + os.remove(path) + # Basic wcs header + header = endata.wcs.to_header(relax=True) + # Add map dimensions + header["NAXIS"] = endata.ndim + for i, n in enumerate(endata.shape[::-1]): + header[f"NAXIS{i+1}"] = n + hdus = af.HDUList([af.PrimaryHDU(endata, header)]) + hdus.writeto(path) + + del endata + return + + +@function_timer +def read_wcs_fits(pix, path, ext=0, comm_bytes=10000000): + """Read and broadcast pixel data stored in a FITS image. + + The root process opens the FITS file and broadcasts the data in units + of the submap size. + + Args: + pix (PixelData): The distributed PixelData object. + path (str): The path to the FITS file. + ext (int, str): Then index or name of the FITS image extension to load. + comm_bytes (int): The approximate message size to use in bytes. + + Returns: + None + + """ + dist = pix.distribution + rank = 0 + if dist.comm is not None: + rank = dist.comm.rank + + endata = None + if rank == 0: + # Load from disk + endata = pixell.enmap.read_map(path, fmt="fits") + # Check dimensions + enpix = 1 + for s in endata.shape: + enpix *= s + tot_pix = dist.n_pix * pix.n_value + if tot_pix != enpix: + raise RuntimeError( + f"Input file has {enpix} pixel values instead of {tot_pix}" + ) + + n_val_submap = dist.n_pix_submap * pix.n_value + + if dist.comm is None: + # Single process, just copy into place + for sm in range(dist.n_submap): + if sm in dist.local_submaps: + loc = dist.global_submap_to_local[sm] + enmap_to_submap(dist, endata, sm, pix.data[loc]) + else: + # One reader broadcasts + comm_submap = pix.comm_nsubmap(comm_bytes) + + buf = np.zeros(comm_submap * n_val_submap, dtype=pix.dtype) + view = buf.reshape(comm_submap, dist.n_pix_submap, pix.n_value) + submap_off = 0 + ncomm = comm_submap + while submap_off < dist.n_submap: + if submap_off + ncomm > dist.n_submap: + ncomm = dist.n_submap - submap_off + if rank == 0: + # Fill the bcast buffer + for c in range(ncomm): + enmap_to_submap(dist, endata, (submap_off + c), view[c]) + # Broadcast + dist.comm.Bcast(buf, root=0) + # Copy these submaps into local data + for sm in range(submap_off, submap_off + ncomm): + if sm in dist.local_submaps: + loc = dist.global_submap_to_local[sm] + pix.data[loc, :, :] = view[sm - submap_off, :, :] + submap_off += comm_submap + buf.fill(0) + + return diff --git a/src/toast/scripts/toast_healpix_coadd.py b/src/toast/scripts/toast_healpix_coadd.py index d262f713e..e8e4a19ef 100644 --- a/src/toast/scripts/toast_healpix_coadd.py +++ b/src/toast/scripts/toast_healpix_coadd.py @@ -22,7 +22,7 @@ from toast._libtoast import cov_apply_diag, cov_eigendecompose_diag from toast.covariance import covariance_apply, covariance_invert from toast.mpi import MPI, Comm, get_world -from toast.pixels_io import ( +from toast.pixels_io_healpix import ( filename_is_fits, filename_is_hdf5, read_healpix, diff --git a/src/toast/scripts/toast_healpix_convert.py b/src/toast/scripts/toast_healpix_convert.py index 1944e88c1..dc388261f 100644 --- a/src/toast/scripts/toast_healpix_convert.py +++ b/src/toast/scripts/toast_healpix_convert.py @@ -18,7 +18,7 @@ import toast from toast.mpi import Comm, get_world -from toast.pixels_io import ( +from toast.pixels_io_healpix import ( filename_is_fits, filename_is_hdf5, read_healpix, diff --git a/src/toast/tests/CMakeLists.txt b/src/toast/tests/CMakeLists.txt index 7cc37a655..8df439a21 100644 --- a/src/toast/tests/CMakeLists.txt +++ b/src/toast/tests/CMakeLists.txt @@ -36,9 +36,11 @@ install(FILES ops_mapmaker.py covariance.py ops_pointing_healpix.py + ops_pointing_wcs.py ops_memory_counter.py ops_scan_map.py ops_scan_healpix.py + ops_scan_wcs.py ops_madam.py ops_gainscrambler.py template_amplitudes.py diff --git a/src/toast/tests/_helpers.py b/src/toast/tests/_helpers.py index 27947b336..4acac86d5 100644 --- a/src/toast/tests/_helpers.py +++ b/src/toast/tests/_helpers.py @@ -101,6 +101,52 @@ def create_space_telescope(group_size, sample_rate=10.0 * u.Hz, pixel_per_proces return Telescope("test", focalplane=fp, site=site) +def create_boresight_telescope(group_size, sample_rate=10.0 * u.Hz): + """Create a fake telescope with one boresight detector per process.""" + nullquat = np.array([0, 0, 0, 1], dtype=np.float64) + n_det = group_size + det_names = [f"d{x:03d}" for x in range(n_det)] + pol_ang = np.array([(2 * np.pi * x / n_det) for x in range(n_det)]) + + det_table = QTable( + [ + Column(name="name", data=det_names), + Column(name="quat", data=[nullquat for x in range(n_det)]), + Column(name="pol_leakage", length=n_det, unit=None), + Column(name="psi_pol", data=pol_ang, unit=u.rad), + Column(name="fwhm", length=n_det, unit=u.arcmin), + Column(name="psd_fmin", length=n_det, unit=u.Hz), + Column(name="psd_fknee", length=n_det, unit=u.Hz), + Column(name="psd_alpha", length=n_det, unit=None), + Column(name="psd_net", length=n_det, unit=(u.K * np.sqrt(1.0 * u.second))), + Column(name="bandcenter", length=n_det, unit=u.GHz), + Column(name="bandwidth", length=n_det, unit=u.GHz), + Column(name="pixel", data=[0 for x in range(n_det)]), + ] + ) + + fwhm = 5.0 * u.arcmin + + for idet in range(len(det_table)): + det_table[idet]["pol_leakage"] = 0.0 + det_table[idet]["fwhm"] = fwhm + det_table[idet]["bandcenter"] = 150.0 * u.GHz + det_table[idet]["bandwidth"] = 20.0 * u.GHz + det_table[idet]["psd_fmin"] = 1.0e-5 * u.Hz + det_table[idet]["psd_fknee"] = 0.05 * u.Hz + det_table[idet]["psd_alpha"] = 1.0 + det_table[idet]["psd_net"] = 100 * (u.K * np.sqrt(1.0 * u.second)) + + fp = Focalplane( + detector_data=det_table, + sample_rate=sample_rate, + field_of_view=1.1 * (2 * fwhm), + ) + + site = SpaceSite("L2") + return Telescope("test", focalplane=fp, site=site) + + def create_ground_telescope( group_size, sample_rate=10.0 * u.Hz, pixel_per_process=1, fknee=None ): diff --git a/src/toast/tests/noise.py b/src/toast/tests/noise.py index 9903c44e2..7767adc98 100644 --- a/src/toast/tests/noise.py +++ b/src/toast/tests/noise.py @@ -10,11 +10,11 @@ import numpy.testing as nt from astropy import units as u +from .. import ops as ops from ..instrument_sim import fake_hexagon_focalplane from ..io import H5File from ..noise import Noise from ..noise_sim import AnalyticNoise -from .. import ops as ops from ._helpers import create_outdir, create_satellite_data from .mpi import MPITestCase diff --git a/src/toast/tests/ops_cadence_map.py b/src/toast/tests/ops_cadence_map.py index f6a9611a5..dd2abf2e0 100644 --- a/src/toast/tests/ops_cadence_map.py +++ b/src/toast/tests/ops_cadence_map.py @@ -4,7 +4,6 @@ import os -import healpy as hp import numpy as np import numpy.testing as nt from astropy import units as u @@ -15,7 +14,7 @@ from ..noise import Noise from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import create_outdir, create_satellite_data from .mpi import MPITestCase diff --git a/src/toast/tests/ops_demodulate.py b/src/toast/tests/ops_demodulate.py index 49283c15e..d3138ecf3 100644 --- a/src/toast/tests/ops_demodulate.py +++ b/src/toast/tests/ops_demodulate.py @@ -12,7 +12,7 @@ from .. import qarray as qa from ..observation import default_values as defaults from ..pixels import PixelData -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import create_ground_data, create_outdir from .mpi import MPITestCase diff --git a/src/toast/tests/ops_filterbin.py b/src/toast/tests/ops_filterbin.py index 0d3e7619c..3219ea0b1 100644 --- a/src/toast/tests/ops_filterbin.py +++ b/src/toast/tests/ops_filterbin.py @@ -16,7 +16,7 @@ from ..noise import Noise from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import read_healpix +from ..pixels_io_healpix import read_healpix from ..vis import set_matplotlib_backend from ._helpers import create_fake_sky, create_ground_data, create_outdir, fake_flags from .mpi import MPITestCase diff --git a/src/toast/tests/ops_flag_sso.py b/src/toast/tests/ops_flag_sso.py index f950b89df..6b797463b 100644 --- a/src/toast/tests/ops_flag_sso.py +++ b/src/toast/tests/ops_flag_sso.py @@ -4,14 +4,13 @@ import os -import healpy as hp import numpy as np from astropy import units as u from .. import ops as ops from .. import qarray as qa from ..observation import default_values as defaults -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import create_ground_data, create_outdir from .mpi import MPITestCase diff --git a/src/toast/tests/ops_mapmaker.py b/src/toast/tests/ops_mapmaker.py index 06627d7b4..1284d18ea 100644 --- a/src/toast/tests/ops_mapmaker.py +++ b/src/toast/tests/ops_mapmaker.py @@ -15,7 +15,7 @@ from ..noise import Noise from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import create_fake_sky, create_outdir, create_satellite_data from .mpi import MPITestCase @@ -120,6 +120,8 @@ def test_offset(self): det_data=defaults.det_data, binning=binner, template_matrix=tmatrix, + solve_rcond_threshold=1.0e-1, + map_rcond_threshold=1.0e-1, write_hits=False, write_map=False, write_cov=False, diff --git a/src/toast/tests/ops_mapmaker_binning.py b/src/toast/tests/ops_mapmaker_binning.py index cf884d441..3b11c7dfb 100644 --- a/src/toast/tests/ops_mapmaker_binning.py +++ b/src/toast/tests/ops_mapmaker_binning.py @@ -15,7 +15,7 @@ from ..noise import Noise from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import create_outdir, create_satellite_data from .mpi import MPITestCase diff --git a/src/toast/tests/ops_mapmaker_solve.py b/src/toast/tests/ops_mapmaker_solve.py index 8a0bc0cff..78bf032fe 100644 --- a/src/toast/tests/ops_mapmaker_solve.py +++ b/src/toast/tests/ops_mapmaker_solve.py @@ -4,7 +4,6 @@ import os -import healpy as hp import numpy as np import numpy.testing as nt from astropy import units as u diff --git a/src/toast/tests/ops_noise_estim.py b/src/toast/tests/ops_noise_estim.py index 03b14af6d..93837d206 100644 --- a/src/toast/tests/ops_noise_estim.py +++ b/src/toast/tests/ops_noise_estim.py @@ -15,7 +15,7 @@ from ..noise import Noise from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import ( create_fake_sky, @@ -112,8 +112,8 @@ def test_noise_estim(self): if data.comm.world_rank == 0: set_matplotlib_backend() - import matplotlib.pyplot as plt import astropy.io.fits as pf + import matplotlib.pyplot as plt obs = data.obs[0] det = obs.local_detectors[0] diff --git a/src/toast/tests/ops_pointing_wcs.py b/src/toast/tests/ops_pointing_wcs.py new file mode 100644 index 000000000..2e501a991 --- /dev/null +++ b/src/toast/tests/ops_pointing_wcs.py @@ -0,0 +1,619 @@ +# Copyright (c) 2021-2022 by the parties listed in the AUTHORS file. +# All rights reserved. Use of this source code is governed by +# a BSD-style license that can be found in the LICENSE file. + +import os + +import astropy.io.fits as af +import numpy as np +from astropy import units as u +from astropy.wcs import WCS + +from .. import ops as ops +from .. import qarray as qa +from .. import templates +from ..data import Data +from ..observation import Observation +from ..observation import default_values as defaults +from ..pixels_io_wcs import write_wcs_fits +from ..vis import set_matplotlib_backend +from ._helpers import ( + create_boresight_telescope, + create_comm, + create_fake_sky, + create_ground_data, + create_outdir, + create_space_telescope, +) +from .mpi import MPITestCase + + +class PointingWCSTest(MPITestCase): + def setUp(self): + fixture_name = os.path.splitext(os.path.basename(__file__))[0] + self.outdir = create_outdir(self.comm, fixture_name) + # For debugging, change this to True + self.write_extra = False + + def check_hits(self, prefix, pixels): + wcs = pixels.wcs + + toastcomm = create_comm(self.comm) + data = Data(toastcomm) + tele = create_boresight_telescope( + toastcomm.group_size, + sample_rate=1.0 * u.Hz, + ) + + # Make some fake boresight pointing + npix_ra = pixels.pix_ra + npix_dec = pixels.pix_dec + px = list() + for ra in range(npix_ra): + px.extend( + np.column_stack( + [ + ra * np.ones(npix_dec), + np.arange(npix_dec), + ] + ).tolist() + ) + px = np.array(px, dtype=np.float64) + coord = wcs.wcs_pix2world(px, 0) + coord *= np.pi / 180.0 + phi = np.array(coord[:, 0], dtype=np.float64) + half_pi = np.pi / 2 + theta = np.array(half_pi - coord[:, 1], dtype=np.float64) + bore = qa.from_position(theta, phi) + + nsamp = npix_ra * npix_dec + data.obs.append(Observation(toastcomm, tele, n_samples=nsamp)) + data.obs[0].shared.create_column( + defaults.boresight_radec, (nsamp, 4), dtype=np.float64 + ) + data.obs[0].shared.create_column( + defaults.shared_flags, (nsamp,), dtype=np.uint8 + ) + if toastcomm.group_rank == 0: + data.obs[0].shared[defaults.boresight_radec].set(bore) + else: + data.obs[0].shared[defaults.boresight_radec].set(None) + + pixels.apply(data) + + # Hitmap + + build_hits = ops.BuildHitMap( + pixel_dist="dist", + pixels=pixels.pixels, + det_flags=None, + ) + build_hits.apply(data) + + if self.write_extra: + outfile = os.path.join(self.outdir, f"{prefix}.fits") + write_wcs_fits(data[build_hits.hits], outfile) + + if toastcomm.world_rank == 0: + set_matplotlib_backend() + + import matplotlib.pyplot as plt + + hdu = af.open(outfile)[0] + wcs = WCS(hdu.header) + + fig = plt.figure(figsize=(8, 8), dpi=100) + ax = fig.add_subplot(projection=wcs, slices=("x", "y", 0)) + # plt.imshow(hdu.data, vmin=-2.e-5, vmax=2.e-4, origin='lower') + im = ax.imshow( + np.transpose(hdu.data[0, :, :]), vmin=0, vmax=4, cmap="jet" + ) + ax.grid(color="white", ls="solid") + ax.set_xlabel("Longitude") + ax.set_ylabel("Latitude") + plt.colorbar(im, orientation="vertical") + fig.savefig(os.path.join(self.outdir, f"{prefix}.pdf"), format="pdf") + + np.testing.assert_array_equal( + data[build_hits.hits].data, + data.comm.ngroups + * len(data.obs[0].all_detectors) + * np.ones_like(data[build_hits.hits].data), + ) + del data + + def test_projections(self): + centers = list() + for lon in [130.0, 180.0, 230.0]: + for lat in [-40.0, 0.0, 40.0]: + centers.append((lon * u.degree, lat * u.degree)) + + detpointing_radec = ops.PointingDetectorSimple( + boresight=defaults.boresight_radec + ) + + for proj in ["CAR", "TAN", "CEA", "MER", "ZEA"]: + for center in centers: + pixels = ops.PixelsWCS( + projection=proj, + detector_pointing=detpointing_radec, + create_dist="dist", + use_astropy=True, + center=center, + dimensions=(710, 350), + resolution=(0.1 * u.degree, 0.1 * u.degree), + ) + self.check_hits( + f"hits_{proj}_{center[0].value}_{center[1].value}", pixels + ) + if self.comm is not None: + self.comm.barrier() + + def plot_maps(self, hitfile=None, mapfile=None): + figsize = (8, 8) + figdpi = 100 + + def plot_single(wcs, hdata, hindx, vmin, vmax, out): + set_matplotlib_backend() + + import matplotlib.pyplot as plt + + fig = plt.figure(figsize=figsize, dpi=figdpi) + ax = fig.add_subplot(projection=wcs, slices=("x", "y", hindx)) + im = ax.imshow(np.transpose(hdu.data[hindx, :, :]), cmap="jet") + ax.grid(color="white", ls="solid") + ax.set_xlabel("Longitude") + ax.set_ylabel("Latitude") + plt.colorbar(im, orientation="vertical") + plt.savefig(out, format="pdf") + plt.close() + + def map_range(hdata): + minval = np.amin(hdata) + maxval = np.amax(hdata) + margin = 0.1 * (maxval - minval) + if margin == 0: + margin = -1 + minval += margin + maxval -= margin + return minval, maxval + + if hitfile is not None: + hdu = af.open(hitfile)[0] + wcs = WCS(hdu.header) + maxhits = np.amax(hdu.data[0, :, :]) + plot_single(wcs, hdu, 0, 0, maxhits, f"{hitfile}.pdf") + + if mapfile is not None: + hdu = af.open(mapfile)[0] + wcs = WCS(hdu.header) + + mmin, mmax = map_range(hdu.data[0, :, :]) + plot_single(wcs, hdu, 0, mmin, mmax, f"{mapfile}_I.pdf") + try: + mmin, mmax = map_range(hdu.data[1, :, :]) + plot_single(wcs, hdu, 1, mmin, mmax, f"{mapfile}_Q.pdf") + except: + pass + + try: + mmin, mmax = map_range(hdu.data[2, :, :]) + plot_single(wcs, hdu, 2, mmin, mmax, f"{mapfile}_U.pdf") + except: + pass + + def test_mapmaking(self): + rank = 0 + if self.comm is not None: + rank = self.comm.rank + + # Test several projections + resolution = 0.1 * u.degree + + for proj in ["CAR", "TAN", "CEA", "MER", "ZEA"]: + # Create fake observing of a small patch + data = create_ground_data(self.comm) + + # Simple detector pointing + detpointing_radec = ops.PointingDetectorSimple( + boresight=defaults.boresight_radec, + ) + + # Stokes weights + weights = ops.StokesWeights( + mode="IQU", + hwp_angle=defaults.hwp_angle, + detector_pointing=detpointing_radec, + ) + + # Pixelization + pixels = ops.PixelsWCS( + detector_pointing=detpointing_radec, + projection=proj, + resolution=(0.5 * u.degree, 0.5 * u.degree), + auto_bounds=True, + use_astropy=True, + ) + + pix_dist = ops.BuildPixelDistribution( + pixel_dist="pixel_dist", + pixel_pointing=pixels, + ) + pix_dist.apply(data) + + # Create fake polarized sky pixel values locally + create_fake_sky(data, "pixel_dist", "fake_map") + + if self.write_extra: + # Write it out + outfile = os.path.join(self.outdir, f"mapmaking_{proj}_input.fits") + write_wcs_fits(data["fake_map"], outfile) + if rank == 0: + self.plot_maps(mapfile=outfile) + + # Scan map into timestreams + scanner = ops.Pipeline( + operators=[ + pixels, + weights, + ops.ScanMap( + det_data=defaults.det_data, + pixels=pixels.pixels, + weights=weights.weights, + map_key="fake_map", + ), + ], + detsets=["SINGLE"], + ) + scanner.apply(data) + + # Create an uncorrelated noise model from focalplane detector properties + default_model = ops.DefaultNoiseModel(noise_model="noise_model") + default_model.apply(data) + + # Simulate noise and accumulate to signal + sim_noise = ops.SimNoise( + noise_model=default_model.noise_model, det_data=defaults.det_data + ) + sim_noise.apply(data) + + # Set up binning operator for solving + binner = ops.BinMap( + pixel_dist="pixel_dist", + pixel_pointing=pixels, + stokes_weights=weights, + noise_model=default_model.noise_model, + ) + + # Set up template matrix with just an offset template. + + # Use 1/10 of an observation as the baseline length. Make it not evenly + # divisible in order to test handling of the final amplitude. + ob_time = ( + data.obs[0].shared[defaults.times][-1] + - data.obs[0].shared[defaults.times][0] + ) + step_seconds = float(int(ob_time / 10.0)) + tmpl = templates.Offset( + times=defaults.times, + noise_model=default_model.noise_model, + step_time=step_seconds * u.second, + ) + tmatrix = ops.TemplateMatrix(templates=[tmpl]) + + # Map maker + mapper = ops.MapMaker( + name=f"test_{proj}", + det_data=defaults.det_data, + binning=binner, + template_matrix=tmatrix, + solve_rcond_threshold=1.0e-2, + map_rcond_threshold=1.0e-2, + write_hits=False, + write_map=False, + write_cov=False, + write_rcond=False, + output_dir=self.outdir, + keep_solver_products=True, + keep_final_products=True, + ) + mapper.apply(data) + + if self.write_extra: + # Write outputs manually + for prod in ["hits", "map"]: + outfile = os.path.join(self.outdir, f"mapmaking_{proj}_{prod}.fits") + write_wcs_fits(data[f"{mapper.name}_{prod}"], outfile) + + if rank == 0: + outfile = os.path.join(self.outdir, f"mapmaking_{proj}_hits.fits") + self.plot_maps(hitfile=outfile) + + outfile = os.path.join(self.outdir, f"mapmaking_{proj}_map.fits") + self.plot_maps(mapfile=outfile) + + def fake_source(self, mission_start, ra_start, dec_start, times, deg_per_hour=1.0): + deg_sec = deg_per_hour / 3600.0 + t_start = float(times[0]) + first_ra = ra_start + (t_start - mission_start) * deg_sec + first_dec = dec_start + (t_start - mission_start) * deg_sec + incr = (times - t_start) * deg_sec + return first_ra + incr, first_dec + incr + + def create_source_data( + self, data, proj, res, signal_name, deg_per_hour=1.0, dbg_dir=None + ): + detpointing = ops.PointingDetectorSimple( + boresight=defaults.boresight_radec, + quats="temp_quats", + ) + detpointing.apply(data) + + # Normal autoscaled projection + pixels = ops.PixelsWCS( + projection=proj, + resolution=(res, res), + detector_pointing=detpointing, + pixels="temp_pix", + use_astropy=True, + auto_bounds=True, + ) + + pix_dist = ops.BuildPixelDistribution( + pixel_dist="source_pixel_dist", + pixel_pointing=pixels, + ) + pix_dist.apply(data) + + # Create fake polarized sky pixel values locally + create_fake_sky(data, "source_pixel_dist", "fake_map") + + if self.write_extra: + # Write it out + outfile = os.path.join(self.outdir, f"source_{proj}_input.fits") + write_wcs_fits(data["fake_map"], outfile) + if data.comm.world_rank == 0: + self.plot_maps(mapfile=outfile) + + weights = ops.StokesWeights( + mode="IQU", + hwp_angle=defaults.hwp_angle, + weights="temp_weights", + detector_pointing=detpointing, + ) + weights.apply(data) + + # Scan map into timestreams + scanner = ops.Pipeline( + operators=[ + pixels, + weights, + ops.ScanMap( + det_data=signal_name, + pixels=pixels.pixels, + weights=weights.weights, + map_key="fake_map", + ), + ], + detsets=["SINGLE"], + ) + scanner.apply(data) + + # Use this overall projection window to determine our source + # movement. The source starts at the center of the projection. + px = np.array( + [ + [ + int(0.6 * pixels.pix_ra), + int(0.2 * pixels.pix_dec), + ], + ], + dtype=np.float64, + ) + source_start = pixels.wcs.wcs_pix2world(px, 0) + + # Create the fake ephemeris data and accumulate to signal. + for ob in data.obs: + n_samp = ob.n_local_samples + times = np.array(ob.shared[defaults.times].data) + + source_ra, source_dec = self.fake_source( + data.obs[0].shared["times"][0], + source_start[0][0], + source_start[0][1], + times, + deg_per_hour=deg_per_hour, + ) + source_coord = np.column_stack([source_ra, source_dec]) + + # Create a shared data object with the fake source location + ob.shared.create_column("source", (n_samp, 2), dtype=np.float64) + if ob.comm.group_rank == 0: + ob.shared["source"].set(source_coord) + else: + ob.shared["source"].set(None) + + zaxis = np.array([0.0, 0.0, 1.0], dtype=np.float64) + sphi = ob.shared["source"].data[:, 0] * np.pi / 180.0 + stheta = (90.0 - ob.shared["source"].data[:, 1]) * np.pi / 180.0 + spos = qa.from_position(stheta, sphi) + sdir = qa.rotate(spos, zaxis) + for det in ob.local_detectors: + fwhm = 2 * ob.telescope.focalplane[det]["fwhm"].to_value(u.arcmin) + coeff = 1.0 / (fwhm * np.sqrt(2 * np.pi)) + pre = -0.5 * (1.0 / fwhm) ** 2 + quats = ob.detdata[detpointing.quats][det] + dir = qa.rotate(quats, zaxis) + sdist = np.arccos([np.dot(x, y) for x, y in zip(sdir, dir)]) + sdist_arc = sdist * 180.0 * 60.0 / np.pi + seen = sdist_arc < 10 + seen_samp = np.arange(len(sdist), dtype=np.int32) + amp = 50.0 * coeff * np.exp(pre * np.square(sdist_arc)) + ob.detdata[signal_name][det, :] += amp[:] + + if dbg_dir is not None and ob.comm.group_rank == 0: + set_matplotlib_backend() + + import matplotlib.pyplot as plt + + fig = plt.figure(figsize=(8, 8), dpi=100) + ax = fig.add_subplot() + ax.scatter( + ob.shared["source"].data[:, 0], + ob.shared["source"].data[:, 1], + ) + ax.set_xlabel("Longitude") + ax.set_ylabel("Latitude") + plt.savefig( + os.path.join(dbg_dir, f"source_coord_{proj}_{ob.name}.pdf"), + format="pdf", + ) + plt.close() + + default_model = ops.DefaultNoiseModel(noise_model="source_noise_model") + default_model.apply(data) + + # Simulate noise and accumulate to signal + sim_noise = ops.SimNoise( + noise_model=default_model.noise_model, det_data=signal_name + ) + sim_noise.apply(data) + + if dbg_dir is not None: + # Make a binned map of the signal in the non-source-centered projection + binner = ops.BinMap( + pixel_dist="source_pixel_dist", + pixel_pointing=pixels, + stokes_weights=weights, + noise_model=default_model.noise_model, + ) + mapper = ops.MapMaker( + name=f"source_{proj}_notrack", + det_data=signal_name, + solve_rcond_threshold=1.0e-2, + map_rcond_threshold=1.0e-2, + iter_max=10, + binning=binner, + template_matrix=None, + output_dir=dbg_dir, + write_hits=True, + write_map=True, + write_binmap=False, + ) + mapper.apply(data) + if data.comm.world_rank == 0: + outfile = os.path.join(self.outdir, f"source_{proj}_notrack_hits.fits") + self.plot_maps(hitfile=outfile) + outfile = os.path.join(self.outdir, f"source_{proj}_notrack_map.fits") + self.plot_maps(mapfile=outfile) + + # Cleanup our temp objects + ops.Delete( + detdata=[detpointing.quats, pixels.pixels, weights.weights], + meta=["source_pixel_dist", "source_noise_model"], + ).apply(data) + + def test_source_map(self): + rank = 0 + if self.comm is not None: + rank = self.comm.rank + + # Test several projections + resolution = 0.5 * u.degree + + for proj in ["CAR", "TAN"]: + # Create fake observing of a small patch + data = create_ground_data(self.comm, pixel_per_process=10) + + # Create source motion and simulated detector data. + dbgdir = None + if proj == "CAR" and self.write_extra: + dbgdir = self.outdir + self.create_source_data( + data, proj, resolution, defaults.det_data, dbg_dir=dbgdir + ) + + # Simple detector pointing + detpointing_radec = ops.PointingDetectorSimple( + boresight=defaults.boresight_radec, + ) + + # Stokes weights + weights = ops.StokesWeights( + mode="IQU", + hwp_angle=defaults.hwp_angle, + detector_pointing=detpointing_radec, + ) + + # Source-centered pointing + pixels = ops.PixelsWCS( + projection=proj, + resolution=(resolution, resolution), + center_offset="source", + detector_pointing=detpointing_radec, + use_astropy=True, + auto_bounds=True, + ) + + pix_dist = ops.BuildPixelDistribution( + pixel_dist="pixel_dist", + pixel_pointing=pixels, + ) + pix_dist.apply(data) + + default_model = ops.DefaultNoiseModel(noise_model="noise_model") + default_model.apply(data) + + # Set up binning operator for solving + binner = ops.BinMap( + pixel_dist="pixel_dist", + pixel_pointing=pixels, + stokes_weights=weights, + noise_model=default_model.noise_model, + ) + + # Set up template matrix with just an offset template. + + # Use 1/10 of an observation as the baseline length. Make it not evenly + # divisible in order to test handling of the final amplitude. + ob_time = ( + data.obs[0].shared[defaults.times][-1] + - data.obs[0].shared[defaults.times][0] + ) + step_seconds = float(int(ob_time / 10.0)) + tmpl = templates.Offset( + times=defaults.times, + det_flags=None, + noise_model=default_model.noise_model, + step_time=step_seconds * u.second, + ) + tmatrix = ops.TemplateMatrix(templates=[tmpl]) + + # Map maker + mapper = ops.MapMaker( + name=f"source_{proj}", + det_data=defaults.det_data, + solve_rcond_threshold=1.0e-2, + map_rcond_threshold=1.0e-2, + iter_max=10, + binning=binner, + template_matrix=tmatrix, + output_dir=self.outdir, + write_hits=True, + write_map=True, + write_binmap=True, + ) + mapper.apply(data) + + if rank == 0: + outfile = os.path.join(self.outdir, f"source_{proj}_hits.fits") + self.plot_maps(hitfile=outfile) + outfile = os.path.join(self.outdir, f"source_{proj}_map.fits") + self.plot_maps(mapfile=outfile) + outfile = os.path.join(self.outdir, f"source_{proj}_binmap.fits") + self.plot_maps(mapfile=outfile) + + if data.comm.comm_world is not None: + data.comm.comm_world.barrier() + + del data diff --git a/src/toast/tests/ops_scan_healpix.py b/src/toast/tests/ops_scan_healpix.py index 6aeb5de6e..eac556673 100644 --- a/src/toast/tests/ops_scan_healpix.py +++ b/src/toast/tests/ops_scan_healpix.py @@ -11,7 +11,7 @@ from .. import ops as ops from ..observation import default_values as defaults from ..pixels import PixelData -from ..pixels_io import write_healpix_fits, write_healpix_hdf5 +from ..pixels_io_healpix import write_healpix_fits, write_healpix_hdf5 from ._helpers import ( create_fake_mask, create_fake_sky, @@ -77,7 +77,9 @@ def test_healpix_fits(self): for ob in data.obs: for det in ob.local_detectors: np.testing.assert_almost_equal( - ob.detdata["test"][det], ob.detdata[defaults.det_data][det] + ob.detdata["test"][det], + ob.detdata[defaults.det_data][det], + decimal=5, ) del data @@ -132,7 +134,7 @@ def test_healpix_mask(self): for ob in data.obs: for det in ob.local_detectors: - np.testing.assert_almost_equal( + np.testing.assert_equal( ob.detdata["test_flags"][det], ob.detdata[defaults.det_flags][det] ) @@ -189,7 +191,9 @@ def test_healpix_hdf5(self): for ob in data.obs: for det in ob.local_detectors: np.testing.assert_almost_equal( - ob.detdata["test"][det], ob.detdata[defaults.det_data][det] + ob.detdata["test"][det], + ob.detdata[defaults.det_data][det], + decimal=5, ) del data diff --git a/src/toast/tests/ops_scan_wcs.py b/src/toast/tests/ops_scan_wcs.py new file mode 100644 index 000000000..2d5b1159a --- /dev/null +++ b/src/toast/tests/ops_scan_wcs.py @@ -0,0 +1,158 @@ +# Copyright (c) 2015-2020 by the parties listed in the AUTHORS file. +# All rights reserved. Use of this source code is governed by +# a BSD-style license that can be found in the LICENSE file. + +import os + +import numpy as np +import numpy.testing as nt +from astropy import units as u + +from .. import ops as ops +from ..observation import default_values as defaults +from ..pixels import PixelData +from ..pixels_io_wcs import write_wcs_fits +from ._helpers import ( + create_fake_mask, + create_fake_sky, + create_ground_data, + create_outdir, +) +from .mpi import MPITestCase + + +class ScanWCSTest(MPITestCase): + def setUp(self): + fixture_name = os.path.splitext(os.path.basename(__file__))[0] + self.outdir = create_outdir(self.comm, fixture_name) + np.random.seed(123456) + + def test_wcs_fits(self): + # Create fake observing of a small patch + data = create_ground_data(self.comm) + + # Simple detector pointing + detpointing_radec = ops.PointingDetectorSimple( + boresight=defaults.boresight_radec, + ) + + # Pixelization + pixels = ops.PixelsWCS( + projection="CAR", + resolution=(0.05 * u.degree, 0.05 * u.degree), + auto_bounds=True, + detector_pointing=detpointing_radec, + create_dist="pixel_dist", + use_astropy=True, + ) + pixels.apply(data) + pixels.create_dist = None + + # Stokes weights + weights = ops.StokesWeights( + mode="IQU", + hwp_angle=defaults.hwp_angle, + detector_pointing=detpointing_radec, + ) + weights.apply(data) + + # Create fake polarized sky pixel values locally + create_fake_sky(data, "pixel_dist", "fake_map") + + # Write this to a file + input_file = os.path.join(self.outdir, "fake.fits") + write_wcs_fits(data["fake_map"], input_file) + + # Scan map into timestreams + scanner = ops.ScanMap( + det_data=defaults.det_data, + pixels=pixels.pixels, + weights=weights.weights, + map_key="fake_map", + ) + scanner.apply(data) + + # Run the scanning from the file + scan_wcs = ops.ScanWCSMap( + file=input_file, + det_data="test", + pixel_pointing=pixels, + stokes_weights=weights, + ) + scan_wcs.apply(data) + + # Check that the sets of timestreams match. + + for ob in data.obs: + for det in ob.local_detectors: + np.testing.assert_almost_equal( + ob.detdata["test"][det], ob.detdata[defaults.det_data][det] + ) + + del data + return + + def test_wcs_mask(self): + # Create fake observing of a small patch + data = create_ground_data(self.comm) + + # Simple detector pointing + detpointing_radec = ops.PointingDetectorSimple( + boresight=defaults.boresight_radec, + ) + + # Pixelization + pixels = ops.PixelsWCS( + projection="CAR", + resolution=(0.05 * u.degree, 0.05 * u.degree), + auto_bounds=True, + detector_pointing=detpointing_radec, + create_dist="pixel_dist", + use_astropy=True, + ) + pixels.apply(data) + pixels.create_dist = None + + # Stokes weights + weights = ops.StokesWeights( + mode="IQU", + hwp_angle=defaults.hwp_angle, + detector_pointing=detpointing_radec, + ) + weights.apply(data) + + # Create fake mask pixel values locally + create_fake_mask(data, "pixel_dist", "fake_mask") + + # Write this to a file + input_file = os.path.join(self.outdir, "fake_mask.fits") + write_wcs_fits(data["fake_mask"], input_file) + + # Scan map into timestreams + scanner = ops.ScanMask( + det_flags=defaults.det_flags, + det_flags_mask=defaults.det_mask_invalid, + pixels=pixels.pixels, + mask_key="fake_mask", + ) + scanner.apply(data) + + # Run the scanning from the file + scan_wcs = ops.ScanWCSMask( + file=input_file, + det_flags="test_flags", + det_flags_mask=defaults.det_mask_invalid, + pixel_pointing=pixels, + ) + scan_wcs.apply(data) + + # Check that the sets of timestreams match. + + for ob in data.obs: + for det in ob.local_detectors: + np.testing.assert_equal( + ob.detdata["test_flags"][det], ob.detdata[defaults.det_flags][det] + ) + + del data + return diff --git a/src/toast/tests/ops_sim_cosmic_rays.py b/src/toast/tests/ops_sim_cosmic_rays.py index 24c81a65e..5f37062ec 100644 --- a/src/toast/tests/ops_sim_cosmic_rays.py +++ b/src/toast/tests/ops_sim_cosmic_rays.py @@ -4,14 +4,13 @@ import os -import healpy as hp import numpy as np from .. import ops as ops from .. import rng from ..covariance import covariance_apply from ..pixels import PixelData, PixelDistribution -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import ( create_fake_sky, diff --git a/src/toast/tests/ops_sim_crosstalk.py b/src/toast/tests/ops_sim_crosstalk.py index bc1fd78d8..b29b58222 100644 --- a/src/toast/tests/ops_sim_crosstalk.py +++ b/src/toast/tests/ops_sim_crosstalk.py @@ -4,14 +4,13 @@ import os -import healpy as hp import numpy as np from .. import ops as ops from .. import rng from ..covariance import covariance_apply from ..pixels import PixelData, PixelDistribution -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ._helpers import ( create_fake_sky, create_outdir, diff --git a/src/toast/tests/ops_sim_gaindrifts.py b/src/toast/tests/ops_sim_gaindrifts.py index 258f4d185..11a9726e1 100644 --- a/src/toast/tests/ops_sim_gaindrifts.py +++ b/src/toast/tests/ops_sim_gaindrifts.py @@ -12,7 +12,7 @@ from ..covariance import covariance_apply from ..observation import default_values as defaults from ..pixels import PixelData, PixelDistribution -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import ( create_fake_sky, diff --git a/src/toast/tests/ops_sim_ground.py b/src/toast/tests/ops_sim_ground.py index ccb85df85..dc8f77f3b 100644 --- a/src/toast/tests/ops_sim_ground.py +++ b/src/toast/tests/ops_sim_ground.py @@ -17,7 +17,7 @@ from ..instrument_sim import fake_hexagon_focalplane from ..mpi import MPI, Comm from ..observation import default_values as defaults -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..schedule import GroundSchedule from ..schedule_sim_ground import run_scheduler from ..vis import set_matplotlib_backend diff --git a/src/toast/tests/ops_sim_satellite.py b/src/toast/tests/ops_sim_satellite.py index ac273ecf7..f436e3e62 100644 --- a/src/toast/tests/ops_sim_satellite.py +++ b/src/toast/tests/ops_sim_satellite.py @@ -17,7 +17,7 @@ from ..instrument_sim import fake_hexagon_focalplane from ..mpi import MPI, Comm from ..observation import default_values as defaults -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..schedule_sim_satellite import create_satellite_schedule from ..vis import set_matplotlib_backend from ._helpers import create_comm, create_outdir diff --git a/src/toast/tests/ops_sim_tod_atm.py b/src/toast/tests/ops_sim_tod_atm.py index 64a3721be..b80afdcd1 100644 --- a/src/toast/tests/ops_sim_tod_atm.py +++ b/src/toast/tests/ops_sim_tod_atm.py @@ -13,7 +13,7 @@ from ..atm import available_atm, available_utils from ..dipole import dipole from ..observation import default_values as defaults -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import create_ground_data, create_outdir from .mpi import MPITestCase diff --git a/src/toast/tests/ops_sim_tod_conviqt.py b/src/toast/tests/ops_sim_tod_conviqt.py index 148111b85..f77367136 100644 --- a/src/toast/tests/ops_sim_tod_conviqt.py +++ b/src/toast/tests/ops_sim_tod_conviqt.py @@ -11,7 +11,7 @@ from .. import ops as ops from .. import qarray as qa from ..observation import default_values as defaults -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import ( create_fake_beam_alm, diff --git a/src/toast/tests/ops_sim_tod_dipole.py b/src/toast/tests/ops_sim_tod_dipole.py index b821e7b12..6ec89b11d 100644 --- a/src/toast/tests/ops_sim_tod_dipole.py +++ b/src/toast/tests/ops_sim_tod_dipole.py @@ -11,7 +11,7 @@ from .. import ops as ops from .. import qarray as qa from ..dipole import dipole -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import create_healpix_ring_satellite, create_outdir from .mpi import MPITestCase diff --git a/src/toast/tests/ops_sim_tod_totalconvolve.py b/src/toast/tests/ops_sim_tod_totalconvolve.py index bdbdeefd1..eed3c0e7e 100644 --- a/src/toast/tests/ops_sim_tod_totalconvolve.py +++ b/src/toast/tests/ops_sim_tod_totalconvolve.py @@ -10,7 +10,7 @@ from .. import ops as ops from .. import qarray as qa -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import ( create_fake_beam_alm, diff --git a/src/toast/tests/ops_sss.py b/src/toast/tests/ops_sss.py index 6043fd85d..c8a4393fe 100644 --- a/src/toast/tests/ops_sss.py +++ b/src/toast/tests/ops_sss.py @@ -4,14 +4,13 @@ import os -import healpy as hp import numpy as np from astropy import units as u from .. import ops as ops from .. import qarray as qa from ..observation import default_values as defaults -from ..pixels_io import write_healpix_fits +from ..pixels_io_healpix import write_healpix_fits from ..vis import set_matplotlib_backend from ._helpers import create_ground_data, create_outdir from .mpi import MPITestCase diff --git a/src/toast/tests/pixels.py b/src/toast/tests/pixels.py index 936d36075..3de73a468 100644 --- a/src/toast/tests/pixels.py +++ b/src/toast/tests/pixels.py @@ -8,7 +8,7 @@ import numpy as np import numpy.testing as nt -from .. import pixels_io as io +from .. import pixels_io_healpix as io from ..pixels import PixelData, PixelDistribution from ._helpers import create_outdir from .mpi import MPITestCase diff --git a/src/toast/tests/runner.py b/src/toast/tests/runner.py index 75b19c2b1..e802d9425 100644 --- a/src/toast/tests/runner.py +++ b/src/toast/tests/runner.py @@ -41,9 +41,11 @@ from . import ops_memory_counter as test_ops_memory_counter from . import ops_noise_estim as test_ops_noise_estim from . import ops_pointing_healpix as test_ops_pointing_healpix +from . import ops_pointing_wcs as test_ops_pointing_wcs from . import ops_polyfilter as test_ops_polyfilter from . import ops_scan_healpix as test_ops_scan_healpix from . import ops_scan_map as test_ops_scan_map +from . import ops_scan_wcs as test_ops_scan_wcs from . import ops_sim_cosmic_rays as test_ops_sim_cosmic_rays from . import ops_sim_crosstalk as test_ops_sim_crosstalk from . import ops_sim_gaindrifts as test_ops_sim_gaindrifts @@ -167,6 +169,7 @@ def test(name=None, verbosity=2): suite.addTest(loader.loadTestsFromModule(test_ops_sim_ground)) suite.addTest(loader.loadTestsFromModule(test_ops_memory_counter)) suite.addTest(loader.loadTestsFromModule(test_ops_pointing_healpix)) + suite.addTest(loader.loadTestsFromModule(test_ops_pointing_wcs)) suite.addTest(loader.loadTestsFromModule(test_ops_sim_tod_noise)) suite.addTest(loader.loadTestsFromModule(test_ops_sim_tod_dipole)) suite.addTest(loader.loadTestsFromModule(test_ops_sim_tod_atm)) @@ -182,6 +185,7 @@ def test(name=None, verbosity=2): suite.addTest(loader.loadTestsFromModule(test_ops_mapmaker)) suite.addTest(loader.loadTestsFromModule(test_ops_scan_map)) suite.addTest(loader.loadTestsFromModule(test_ops_scan_healpix)) + suite.addTest(loader.loadTestsFromModule(test_ops_scan_wcs)) suite.addTest(loader.loadTestsFromModule(test_ops_madam)) suite.addTest(loader.loadTestsFromModule(test_ops_gainscrambler)) suite.addTest(loader.loadTestsFromModule(test_ops_sim_gaindrifts)) diff --git a/wheels/build_requirements.txt b/wheels/build_requirements.txt index 0461c8014..e865d620a 100644 --- a/wheels/build_requirements.txt +++ b/wheels/build_requirements.txt @@ -10,4 +10,5 @@ pshmem>=0.2.10 pyyaml astropy healpy +pixell ephem