kima fits Keplerian curves to a set of RV measurements, using the Diffusive Nested Sampling (ascl:1010.029) algorithm to sample the posterior distribution for the model parameters. Additionally, the code can calculate the fully marginalized likelihood of a model with a given number of Keplerians and also infer the number of Keplerian signals detected in a given dataset. kima implements dedicated models for different analyses of a given dataset. The models share a common organization, but each has its own parameters (and thus priors) and settings.
SASHIMI-C calculates various subhalo properties efficiently using semi-analytical models for cold dark matter (CDM), providing a full catalog of dark matter subhalos in a host halo with arbitrary mass and redshift. Each subhalo is characterized by its mass and density profile both at accretion and at the redshift of interest, accretion redshift, and effective number (or weight) corresponding to that particular subhalo. SASHIMI-C computes the subhalo mass function without making any assumptions such as power-law functional forms; the only assumed power law is that for the primordial power spectrum predicted by inflation. The code is not limited to numerical resolution nor to Poisson shot noise, and its results are well in agreement with those from numerical N-body simulations.
Diffstar fits the star formation history (SFH) of galaxies to a smooth parametric model. Diffstar differs from existing SFH models because the parameterization of the model is directly based on basic features of galaxy formation physics, including halo mass assembly history, accretion of gas into the dark matter halo, the fraction of gas that is converted into stars, the time scale over which star formation occurs, and the possibility of rejuvenated star formation. The SFHs of a large number of simulated galaxies can be fit in parallel using mpi4py.
The UniverseMachine applies simple empirical models of galaxy formation to dark matter halo merger trees. For each model, it generates an entire mock universe, which it then observes in the same way as the real Universe to calculate a likelihood function. It includes an advanced MCMC algorithm to explore the allowed parameter space of empirical models that are consistent with observations.
SASHIMI-W calculates various subhalo properties efficiently using semi-analytical models for warm dark matter (WDM); the code is based on the extended Press-Schechter formalism and subhalos' tidal evolution prescription. The calculated constraints are independent of physics of galaxy formation and free from numerical resolution and the Poisson noise, and its results are well in agreement with those from numerical N-body simulations.
EXOTIC (EXOplanet Transit Interpretation Code) analyzes photometric data of transiting exoplanets into lightcurves and retrieves transit epochs and planetary radii. The software reduces images of a transiting exoplanet into a lightcurve, and fits a model to the data to extract planetary information crucial to increasing the efficiency of larger observational platforms. EXOTIC is written in Python and supports the citizen science project Exoplanet Watch. The software runs on Windows, Macintosh, and Linux/Unix computer, and can also be used via Google Colab.
HawkingNet searches for Hawking points in large Cosmic Microwave Background (CMB) data sets. It is based on the deep residual network ResNet18 and consists of eighteen neural layers. Written in Paython, HawkingNet inputs the CMB data, processes the data through its internal network trained for data classification, and outputs the result in a form of a classification score that indicates how confident it is that a Hawking point is contained in the image patch.
AnalyticLC generates an analytic light-curve, and optionally RV and astrometry data, from a set of initial (free) orbital elements and simultaneously fits these data. Written in MATLAB, the code is fast and efficient, and provides insight into the motion of the orbital elements, which is difficult to obtain from numerical integration. A Python wrapper for AnalyticLC is available separately.
RCR provides advanced outlier rejection that is easy to use. Both sigma clipping, the simplest form of outlier rejection, and traditional Chauvenet rejection make use of non-robust quantities, the mean and standard deviation, which are sensitive to the outliers that they are being used to reject. This limits such techniques to samples with small contaminants or small contamination fractions. RCR instead first makes use of robust replacements for the mean, such as the median and the half-sample mode, and similar robust replacements for the standard deviation. RCR has been carefully calibrated and can be applied to samples with both large contaminants and large contaminant fractions (sometimes in excess of 90% contaminated).
celmech provides a variety of analytical and semianalytical tools for celestial mechanics and dynamical astronomy. The package interfaces closely with the REBOUND N-body integrator (ascl:1110.016), thus facilitating comparisons between calculation results and direct N-body integrations. celmech can isolate the contribution of particular resonances to a system's dynamical evolution, and can develop simple analytical models with the minimum number of terms required to capture a particular dynamical phenomenon.
SFQEDtoolkit implements strong-field QED (SFQED) processes in existing particle-in-cell (PIC) and Monte Carlo codes to determine the dynamics of particles and plasmas in extreme electromagnetic fields, such as those present in the vicinity of compact astrophysical objects. The code uses advanced function approximation techniques to calculate high-energy photon emission and electron-positron pair creation probability rates and energy distributions within the locally-constant-field approximation (LCFA) as well as with more advanced models.
PHOTOe simulates the slowing down of photoelectrons in a gas with arbitrary amounts of H, He and O atoms, and thermal electrons, making PHOTOe useful for investigating the atmospheres of exoplanets. The multi-score scheme used in this code differs from other Monte Carlo approaches in that it efficiently handles rare collisional channels, as in the case of low-abundance excited atoms that undergo superelastic and inelastic collisions. PHOTOe outputs include production and energy yields, steady-state photoelectron flux, and estimates of the 'relaxation' time required by the photoelectrons to slow down from the injection energy to the cutoff energy. The model can also estimate the pathlength travelled by the photoelectrons while relaxing.
Deconfuser performs fast orbit fitting to directly imaged multi-planetary systems. It quickly fits orbits to planet detections in 2D images and ensures that all orbits within a certain tolerance are found. The code also tests all groupings of detections by planets (which detection belongs to which planet), and ranks partitions of detections by planets by deciding which assignment of detection-to-planet best fits the data.
nicaea calculates cosmology and weak-lensing quantities and functions from theoretical models of the large-scale structure. Written in C, it can compute the Hubble parameter, distances, and geometry for background cosmology, and linear perturbations, including growth factor, transfer function, cluster mass function, and linear 3D power spectra. It also calculates fitting formulae for non-linear power spectra, emulators, and halo model for Non-linear evolution, and the HOD model for galaxy clustering. In addition, nicaea can compute quantities for cosmic shear such as the convergence power spectrum, second-order correlation functions and derived second-order quantities, and third-order aperture mass moment; it can also calculate CMB anisotropies via CAMB (ascl:1102.026).
PREVIS is a Python module that provides functions to help determine the observability of astronomical sources from long-baseline interferometers worldwide: VLTI (ESO, Chile) and CHARA (USA). PREVIS uses data from the Virtual Observatory (OV), such as magnitudes, Spectral Energy Distribution (SED), celestial coordinates or Gaia distances. Then, it compares the target brightness to the limiting magnitudes of each instrument to determine whether the target is observable with present performances. PREVIS includes main facilities at the VLTI with PIONIER (H band), GRAVITY (K band) and MATISSE (L, M, N bands), and at CHARA array with VEGA (V band), PAVO (R bands), MIRC (H band), CLIMB (K band) and CLASSIC (H, K bands). PREVIS also uses the V or G magnitudes to check the guiding restriction or the tip/tilt correction limit. For the VLTI: if the star is too faint in G mag, PREVIS will look for the list of stars around the target (57 arcsec) with the appropriate magnitude and give the list of celestial coordinates usable as the guiding star.
HIPP (HIgh-Performance Package for scientific computation) provides elegant interfaces for some well-known HPC libraries. Some libraries are wrapped with full-OOP interfaces, and many new extensions based on those raw-interfaces are also provided. This C++ toolkit for HPC can significantly reduce the length of your code, making programming more productive.
ALMA3 computes loading and tidal Love numbers for a spherically symmetric, radially stratified planet. Both real (time-domain) and complex (frequency-domain) Love numbers can be computed. The planetary structure can include an arbitrary number of layers, and each layer can have a different rheological law. ALMA3 can model numerous linear rheologies, including Elastic, Maxwell visco-elastic, Newtonian viscous fluid, Kelvin-Voigt solid, Burgers and Andrade transient rheologies.
special (SPEctral Characterization of directly ImAged Low-mass companions) characterizes low-mass (M, L, T) dwarfs down to giant planets at optical/IR wavelengths. It can also be used more generally to characterize any type of object with a measured spectrum, provided a relevant input model grid, regardless of the observational method used to obtain the spectrum (direct imaging or not) and regardless of the format of the spectra (multi-band photometry, low-resolution or medium-resolution spectrum, or a combination thereof). It analyzes measured spectra, calculating the spectral correlation between channels of an IFS datacube and empirical spectral indices for MLT-dwarfs. It fits input spectra to either photo-/atmospheric model grids or a blackbody model, including additional parameters such as (extra) black body component(s), extinction and total-to-selective extinction ratio, and can use emcee (ascl:1303.002), nestle (ascl:2103.022), or UltraNest (ascl:1611.001) samplers infer posterior distributions on spectral model parameters in a Bayesian framework, among other tasks.
Puri-Psi addresses radio interferometric imaging problems using state-of-the-art optimization algorithms and deep learning. It performs scalable monochromatic, wide-band, and polarized imaging. It also provide joint calibration and imaging, and scalable uncertainty quantification. A scalable framework for wide-field monochromatic intensity imaging is also available, which encompasses a pure optimization algorithm, as well as an AI-based method in the form of a plug-and-play algorithm propelled by Deep Neural Network denoisers.
The 2-D wavelet transformation code MGwave detects kinematic moving groups in astronomical data; it can also investigate underdensities which can eventually provide further information about the MW's non-axisymmetric features. The code creates a histogram of the input data, then performs the wavelet transformation at the specified scales, returning the wavelet coefficients across the entire histogram in addition to information about the detected extrema. MGwave can also run Monte Carlo simulations to propagate uncertainties. It runs the wavelet transformation on simulated data (pulled from Gaussian distributions) many times and tracks the percentage of the simulations in which a given extrema is detected. This quantifies whether a detected overdensity or underdensity is robust to variations of the data within the provided errors.
desitarget selects targets for spectroscopic follow-up by Dark Energy Spectroscopic Instrument (DESI). The pipeline uses bitmasks to record that a specific source has been selected by a particular targeting algorithm, setting bit-values in output data files in a number of different columns that indicate whether a particular target meets specific selection criteria. desitarget also outputs a unique TARGETID that allows each target to be tracked throughout the DESI survey. This TARGETID encodes information about each DESI target, such as the catalog the target was selected from, whether a target is a sky location or part of a random catalog, and whether a target is part of a secondary program.
nFITSview is a simple, user-friendly and open-source FITS image viewer available for Linux and Windows. One of the main concepts of nFITSview is to provide an intuitive user interface which may be helpful both for scientists and for amateur astronomers. nFITSview has different color mapping and manipulation schemes, supports different formats of FITS data files as well as exporting them to different popular image formats. It also supports command-line exporting (with some restrictions) of FITS files to other image formats.
The application is written in C++/Qt for achieving better performance, and with every next version the performance aspect is taken into account.
nFITSview uses its own libnfits library (can be used separately as well) for parsing the FITS files.
SOXS creates simulated X-ray observations of astrophysical sources. The package provides a comprehensive set of tools to design source models and convolve them with simulated models of X-ray observatories. In particular, SOXS is the primary simulation tool for simulations of Lynx and Line Emission Mapper observations. SOXS provides facilities for creating spectral models, simple spatial models for sources, astrophysical background and foreground models, as well as a Python implementation of the SIMPUT file format.
PoWR (Potsdam Wolf-Rayet Models) calculates synthetic spectra for Wolf-Rayet and OB stars from model atmospheres which account for Non-LTE, spherical expansion and metal line blanketing. The model data is provided through a web interface and includes Spectral Energy Distribution, line spectrum in high resolution for different wavelength bands, and atmosphere stratification. For Wolf-Rayet stars of the nitrogen subclass, there are grids of hydrogen-free models and of models with a specified mass fraction of hydrogen. The iron-group and total CNO mass fractions correspond to the metallicity of the Galaxy, the Large Magellanic Cloud, or the Small Magellanic Cloud, respectively. The source code is available as a tarball on the same web interface.
GalCEM (GALactic Chemical Evolution Model) tracks isotope masses as a function of time in a given galaxy. The list of tracked isotopes automatically adapts to the complete set provided by the input yields. The prescription includes massive stars, low-to-intermediate mass stars, and Type Ia supernovae as enrichment channels. Multi-dimensional interpolation curves are extracted from the input yield tables with a preprocessing tool; these interpolation curves improve the computation speeds of the full convolution integrals, which are computed for each isotope and for each enrichment channel. GalCEM also provides tools to track the mass rate change of individual isotopes on a typical spiral galaxy with a final baryonic mass of 5×1010M⊙.
WALDO (Waveform AnomaLy DetectOr) flags possible anomalous Gravitational Waves from Numerical Relativity catalogs using deep learning. It uses a U-Net architecture to learn the waveform features of a dataset. After computing the mismatch between those waveforms and the neural predictions, WALDO isolates high mismatch evaluations for anomaly search.
void-dwarf-analysis analyzes Keck Cosmic Web Imager datacubes to produce maps of kinematic properties (velocity and velocity dispersion), emission line fluxes, and gas-phase metallicities of void dwarf galaxies.
KCWI_DRP, written in Python and based on kderp (ascl:2301.018), is the official DRP for the Keck Cosmic Web Imager at the W. M. Keck Observatory. It provides all of the functionality of the older pipeline and has three execution modes: multi-threading for CPU intensive tasks such as wavelength calibration, and multi-processing for large datasets. It offers vacuum to air and heliocentric or barycentric correction and the ability to use KOA file names or original file names. KCWI_DRP also improves the provenance and traceability of DRP versions and execution steps in the headers over kderp, and has versatile sky subtraction modes including using external sky frames and ability of masking regions.
kderp (KCWI Data Extraction and Reduction Pipeline) reduces data for the Keck Cosmic Web Imager. Written in IDL, it performs basic CCD reduction on raw images to produce bias and overscan subtracted, gain-corrected, trimmed and cosmic ray removed images; it can also subtract the sky. It defines the geometric transformations required to map each pixel in the 2d image into slice, postion, and wavelength, and performs flat field and illumination corrections. It generates cubes, applying the transformations previously solved to the object intensity, variance and mask images output from any of the previous stages, and uses a standard star observation to generate an inverse sensitivity curve which is applied to the corresponding observations to flux calibrate them.
This pipeline has been superseded by KCWI_DRP (ascl:2301.019).
ReACT extends the Copter (ascl:1304.022) and MG-Copter packages, which calculate redshift and real space large scale structure observables for a wide class of gravity and dark energy models. Additions to Copter include spherical collapse in modified gravity, halo model power spectrum for general theories, and real and redshift space LSS 2 point statistics for modified gravity and dark energy. ReACT also includes numerical perturbation theory kernel solvers, real space bispectra in modified gravity, and a numerical perturbation theory kernel solver up to 4th order for 1-loop bispectrum.
FERRE matches physical models to observed data, taking a set of observations and identifying the model parameters that best reproduce the data, in a chi-squared sense. It solves the common problem of having numerical parametric models that are costly to evaluate and need to be used to interpret large data sets. FERRE provides flexibility to search for all model parameters, or hold constant some of them while searching for others. The code is written to be truly N-dimensional and fast. Model predictions are to be given as an array whose values are a function of the model parameters, i.e., numerically. FERRE holds this array in memory, or in a direct-access binary file, and interpolates in it. The code returns, in addition to the optimal set of parameters, their error covariance, and the corresponding model prediction. The code is written in FORTRAN90.
SOAP-GPU is a revision of SOAP 2 (ascl:1504.021), which simulates spectral time series with the effect of active regions (spot, faculae or both). In addition to the traditional outputs of SOAP 2.0 (the cross-correlation function and extracted parameters: radial velocity, bisector span, full width at half maximum), SOAP-GPU generates the integrated spectra at each phase for given input spectra and spectral resolution. Additional capabilities include fast spectral simulation of stellar activity due to GPU acceleration, simulation of more complicated active region structures with superposition between active regions, and more realistic line bisectors, based on solar observations, that varies as function of mu angle for both quiet and active regions. In addition, SOAP-GPU accepts any input high resolution observed spectra. The PHOENIX synthetic spectral library are already implemented at the code level which allows users to simulate stellar activity for stars other than the Sun. Furthermore, SOAP-GPU simulates realistic spectral time series with either spot number/SDO image as additional inputs. The code is written in C and provides python scripts for input pre-processing and output post-processing.
LBL derives velocity measurements from high-resolution (R>50 000) datasets by accounting for outliers in the spectra data. It is tailored for fiber-fed multi-order spectrographs, both in optical and near-infrared (up to 2.5µm) domains. The domain is split into individual units (lines) and the velocity and its associated uncertainty are measured within each line and combined through a mixture model to allow for the presence of spurious values. In addition to the velocity, other quantities are also derived, the most important being a value (dW) that can be understood (for a Gaussian line) as a change in the line FWHM. These values provide useful stellar activity indicators. LBL works on data from a variety of instruments, including SPIRou, NIRPS, HARPS, and ESPRESSO. The code's output is an rdb table that can be uploaded to the online DACE pRV analysis tool.
pyExoRaMa visualizes and manipulates data related to exoplanets and their host stars in a multi-dimensional parameter space. It enables statistical studies based on the large and constantly increasing number of detected exoplanets, identifies possible interdependence among several physical parameters, and compares observables with theoretical models describing the exoplanet composition and structure.
XGA (X-ray: Generate and Analyse) analyzes X-ray sources observed by the XMM-Newton Space telescope. It is based around declaring different types of source and sample objects which correspond to real X-ray sources, finding all available data, and then insulating the user from the tedious generation and basic analysis of X-ray data products. XGA generates photometric products and spectra for individual sources, or whole samples, with just a few lines of code. Though not a pipeline, pipelines for complex analysis can be built on top of it. XGA provides an easy to use (and parallelized) Python interface with XMM's Science Analysis System (ascl:1404.004), as well as with XSPEC (ascl:9910.005). All XMM products and fit results are read into an XGA source storage structure, thus avoiding the need to leave a Python environment at any point during the analysis. This module also supports more complex analyses for specific object types such as the easy generation of scaling relations, the measurement of gas masses for galaxy clusters, and the PSF correction of images.
Rosetta runs tasks for resource-intensive, interactive data analysis as software containers. The code's architecture frames user tasks as microservices – independent and self-contained units – which fully support custom and user-defined software packages, libraries and environments. These include complete remote desktop and GUI applications, common analysis environments such as the Jupyter Notebooks. Rosetta relies on Open Container Initiative containers, allowing for safe, effective and reproducible code execution. It can use a number of container engines and runtimes and seamlessly supports several workload management systems, thus enabling containerized workloads on a wide range of computing resources.
Fastcc returns color corrections for different spectra for various Cosmic Microwave Background experiments. Available in both Python and IDL, the script is easy to use when analyzing radio spectra of sources with data from multiple wide-survey CMB experiments in a consistent way across multiple experiments.
Xpol computes angular power spectra based on cross-correlation between maps and covariance matrices. The code is written in C and is fully MPI parallelized in CPU and memory using spherical transform by s2hat (ascl:1110.013). It has been used to derive CMB and dust power spectra for Archeops and CMB, dust, CIB, SZ, SZ-CIB for Planck, among others.
HiLLiPoP is a multifrequency CMB likelihood for Planck data. The likelihood is a spectrum-based Gaussian approximation for cross-correlation spectra from Planck 100, 143 and 217GHz split-frequency maps, with semi-analytic estimates of the Cl covariance matrix based on the data. The cross-spectra are debiased from the effects of the mask and the beam leakage using Xpol (ascl:2301.009) before being compared to the model, which includes CMB and foreground residuals. They cover the multipoles from ℓ=30 to ℓ=2500. HiLLiPoP is interfaced with the Cobaya (ascl:1910.019) MCMC sampler.
LoLLiPoP is a Planck low-l polarization likelihood based on cross-power-spectra for which the bias is zero when the noise is uncorrelated between maps. It uses a modified approximation to apply to cross-power spectra and is interfaced with the Cobaya (ascl:1910.019) MCMC sampler. Cross-spectra are computed on the CMB maps from Commander component separation applied on each detset-split Planck frequency maps.
Self-cal produces radio-interferometric images of an astrophysical object. The code is an adaptation of the self-calibration algorithm to optical/infrared long-baseline interferometry, especially to make use of differential phases and differential visibilities. It works together with the Mira image reconstruction software and has been used mainly on VLTI data. Self-cal, written in Yorick, is also available as part of fitsOmatic (ascl:2301.005).
The fitOmatic model-fitting prototyping tool tests multi-wavelength model-fitting and exploits VLTI data. It provides tools to define simple geometrical models and conveniently adjust the model's parameters. Written in Yorick, it takes optical interferometry FITS (oifits) files as input and allows the user to define a model of the source from a set of pre-defined models, which can be combined to make more complicated models. fitOmatic then computes the Fourier Transform of the modeled brightness distribution and synthetic observables are computed at the wavelengths and projected baselines of the observations. fitomatic's strength is its ability to define vector-parameters, i.e., parameters that may depend on wavelength and/or time. The self-cal (ascl:2301.006) component of fitOmatic is also available as a separate code.
HEADSS (HiErArchical Data Splitting and Stitching) facilitates clustering at scale, unlike clustering algorithms that scale poorly with increased data volume or that are intrinsically non-distributed. HEADSS automates data splitting and stitching, allowing repeatable handling, and removal, of edge effects. Implemented in conjunction with scikit's HDBSCAN, the code achieves orders of magnitude reduction in single node memory requirements for both non-distributed and distributed implementations, with the latter offering similar order of magnitude reductions in total run times while recovering analogous accuracy. HEADSS also establishes a hierarchy of features by using a subset of clustering features to split the data.
WF4Py implements frequency-domain gravitational wave waveform models in pure Python, thus enabling parallelization over multiple events at a time. Waveforms in WF4Py are built as classes; the functions take dictionaries containing the parameters of the events to analyze as input and provide Fourier domain waveform models. All the waveforms are accurately checked with their implementation in LALSuite (ascl:2012.021) and are a core element of GWFAST (ascl:2212.001).
Pyxel hosts and pipelines models (analytical, numerical, statistical) simulating different types of detector effects on images produced by Charge-Coupled Devices (CCD), Monolithic, and Hybrid CMOS imaging sensors. Users can provide one or more input images to Pyxel, set the detector and model parameters, and select which effects to simulate, such as cosmic rays, detector Point Spread Function (PSF), electronic noises, Charge Transfer Inefficiency (CTI), persistence, dark current, and charge diffusion, among others. The output is one or more images including the simulated detector effects combined. The Pyxel framework, written in Python, provides basic image analysis tools, an input image generator, and a parametric mode to perform parametric and sensitivity analysis. It also offers a model calibration mode to find optimal values of its parameters based on a target dataset the model should reproduce.
CALSAGOS (Clustering ALgorithmS Applied to Galaxies in Overdense Systems) selects cluster members and searches, finds, and identifies substructures and galaxy groups in and around galaxy clusters using the redshift and position in the sky of the galaxies. The package offers two ways to determine cluster members, ISOMER and CLUMBERI. The ISOMER (Identifier of SpectrOscopic MembERs) function selects the spectroscopic cluster members by defining cluster members as those galaxies with a peculiar velocity lower than the escape velocity of the cluster. The CLUMBERI (CLUster MemBER Identifier) function select the cluster members using a 3D-Gaussian Mixture Modules (GMM). Both functions remove the field interlopers by using a 3-sigma clipping algorithm. CALSAGOS uses the function LAGASU (LAbeller of GAlaxies within SUbstructures) to search, find, and identify substructures and groups in and around a galaxy cluster; this function is based on clustering algorithms (GMM and DBSCAN), which search areas with high density to define a substructure or groups.
unWISE-verse is an integrated Python pipeline for downloading sets of unWISE time-resolved coadd cutouts from the WiseView image service and uploading subjects to Zooniverse.org for use in astronomical citizen science research. This software was initially designed for the Backyard Worlds: Cool Neighbors research project and is optimized for target sets containing low luminosity brown dwarf candidates. However, unWISE-verse can be applied to other future astronomical research projects that seek to make use of unWISE infrared sky maps, such as studies of infrared variable/transient sources.
Spender establishes a restframe for galaxy spectra that has higher resolution and larger wavelength range than the spectra from which it is trained. The model can be trained from spectra at different redshifts or even from different instruments without the need to standardize the observations. Spender also has an explicit, differentiable redshift dependence, which can be coupled with a redshift estimator for a fully data-driven spectrum analysis pipeline. The code describes the restframe spectrum by an autoencoder and transforms the restframe model to the observed redshift; it also matches the spectral resolution and line spread function of the instrument.
CONTROL (CUTE autONomous daTa ReductiOn pipeLine) produces science-quality output with a single command line with zero user interference for CUTE (Colorado Ultraviolet Transit Experiment) data. It can be used for any single order spectral data in any wavelength without any modification. The pipeline is governed by a parameter file, which is available with this distribution. CONTROL is fully automated and works in a series of steps following standard CCD reduction techniques. It creates a reduction log to track processes carried out and any parameters used.
Burning Arrow determines the destabilization of massive particle circular orbits due to thermal radiation, emitted in X-ray, from the hot accretion disk material. This code requires the radiation forces exerted on the material at the point of interest found by running the code Infinity (ascl:2212.021). Burning Arrow begins by assuming a target particle in the disk that moves in a circular orbit. It then introduces the recorded radiation forces from Infinity code for the target region. The forces are subsequently introduced into the target particle equations of motion and the trajectory is recalculated. Burning Arrow then produces images of the black hole - accretion disk system that includes the degenerated particle trajectories that obey the assorted velocity profiles.
Tranquillity creates an observing screen looking toward a black hole - accretion disk system, seeks the object, then searches and locates its contour. Subsequently, it attempts to locate the first Einstein "echo" ring and its location. Finally, it collates the retrieved information and draws conclusions; these include the accretion disk level inclination compared to the line of sight and the main disk and the first echo median. The displacement, and thus the divergence of the latter two, is the required information in order to construct the divergence plots. Other programs can later on automatically read these plots and provide estimations of the central black hole spin.
Elysium creates an observing screen at the desirable distance away from a black hole system. Observers set on every pixel of this screen then photograph the area toward the black hole - accretion disk system and report back what they record. This can be the accretion disk (incoming photons bring in radiation and thus energy), the black hole event horizon, or the empty space outside and beyond the system (there are no incoming photons or energy). The central black hole can be either Schwarzschild (nonrotating) or Kerr (rotating) by choice of the user.
Infinity sets an observer in a black hole - accretion disk system. The black hole can be either Schwarzschild (nonrotating) or Kerr (rotating) by choice of the user. This observer can be on the surface of the disk, in its exterior or its interior (if the disk is not opaque). Infinity then scans the entire sky around the observer and investigates whether photons emitted by the hot accretion disk material can reach them. After recording the incoming radiation, the program calculates the stress-energy tensor of the radiation. Afterwards, the program calculates the radiation flux and hence, the radiation force exerted on target particles of various velocity profiles.
Omega solves the photon equations of motion in the environment surrounding a black hole. This black hole can be either Schwarzschild (nonrotating) or Kerr (rotating) by choice of the user. The software offers numerous options, such as the geometrical setup of the accretion disk around the black hole (including no disk, band, slab, wedge, among others, the spin parameter of the central black hole, and the thickness of the accretion disk. Other options that can be set includ the azimuthal angle of the photon emission/reception, the poloidal angle of the photon emission/reception, and how far away or close to the system to look.
m2mcluster performs made-to-measure modeling of star clusters, and can fit target observations of a Galactic globular cluster's 3D density profile and individual kinematic properties, including proper motion velocity dispersion, and line of sight velocity dispersion. The code uses AMUSE (ascl:1107.007) to model the gravitational N-body evolution of the system between time steps; GalPy (ascl:1411.008) is also required.
SourceXtractor++ extracts a catalog of sources from astronomical images; it is the successor to SExtractor (ascl:1010.064). SourceXtractor++ has been completely rewritten in C++ and improves over its predecessor in many ways. It provides support for multiple “measurement” images, has an optimized multi-object, multi-frame model-fitting engine, and can define complex priors and dependencies for model parameters. It also offers efficient image data caching and multi-threaded processing, and has a modular design with support for third-party plug-ins.
powspec provides functions to compute power and cross spectral density of 2D arrays. Units are properly taken into account. It can, for example, create fake Gaussian field images, compute power spectra P(k) of each image, shrink a mask with regard to a kernel, generate a Gaussian field, and plot various results.
The AbundanceMatching Python module creates (interpolates and extrapolates) abundance functions and also provides fiducial deconvolution and abundance matching.
SImMER (Stellar Image Maturation via Efficient Reduction) reduces astronomical imaging data. It performs standard dark-subtraction and flat-fielding operations on data from, for example, the ShARCS camera on the Shane 3-m telescope at Lick Observatory and the PHARO camera on the Hale 5.1-m telescope at Palomar Observatory; its object-oriented design allows the software to be extended to other instruments. SImMER can also perform sky-subtraction, image registration, FWHM measurement, and contrast curve calculation, and can generate tables and plots. For widely separated stars which are of somewhat equal brightness, a “wide binary” mode allows the user to selects which star is the primary around which each image should be centered.
pyTANSPEC extracts XD-mode spectra automatically from data collected by the TIFR-ARIES Near Infrared Spectrometer (TANSPEC) on India's ground-based 3.6-m Devasthal Optical Telescope at Nainital, India. The TANSPEC offers three modes of observations, imaging with various filters, spectroscopy in the low-resolution prism mode with derived R~ 100-400 and the high-resolution cross-dispersed mode (XD-mode) with derived median R~ 2750 for a slit of width 0.5 arcsec. In the XD-mode, ten cross-dispersed orders are packed in the 2048 x 2048 pixels detector to cover the full wavelength regime. The XD-mode is most utilized; pyTANSPEC provides a dedicated pipeline for consistent data reduction for all orders and to reduces data reduction time. The code requires nominal human intervention only for the quality assurance of the reduced data. Two customized configuration files are used to guide the data reduction. The pipeline creates a log file for all the fits files in a given data directory from its header, identifies correct frames (science, continuum and calibration lamps) based on the user input, and offers an option to the user for eyeballing and accepting/removing of the frames, does the cleaning of raw science frames and yields final wavelength calibrated spectra of all orders simultaneously.
PACMAN (Planetary Atmosphere, Crust, and MANtle geochemical evolution) runs a coupled redox-geochemical-climate evolution model. It runs Monte Carlo calculations over nominal parameter ranges, including number of iterations and number of cores for parallelization, which can be altered to reproduce different scenarios and sensitivity tests. Model outputs and corresponding input parameters are saved in separate files which are used to plot results; the the user can choose which outputs to plot, including all successful outputs, nominal Earth outputs, waterworld false positives, desertworld false positives, and high CO2:H2O false positives. Among other functions, PACMAN contains functions for interpolating the pre-computed Outgoing Longwave Radiation (OLR) grid, the atmosphere-ocean partitioning grid, and the stratospheric water vapor grid, calculating bond albedo and outgassing fluxes.
The BANZAI-NRES pipeline processes data from the Network of Robotic Echelle Spectrographs (NRES) on the Las Cumbres Observatory network and provides extracted, wavelength calibrated spectra. If the target is a star, it provides stellar classification parameters (e.g., effective temperature and surface gravity) and a radial velocity measurement. The automated radial velocity measurements from this pipeline have a precision of ~ 10 m/s for high signal-to-noise observations. The data flow and infrastructure of this code relies heavily on BANZAI (ascl:2207.031), enabling BANZAI-NRES to focus on analysis that is specific to spectrographs. The wavelength calibration is primarily done using xwavecal (ascl:2212.011). The pipeline propagates an estimate of the formal uncertainties from all of the data processing stages and includes these in the output data products. These are used as weights in the cross correlation function to measure the radial velocity.
The xwavecal library automatically wavelength calibrates echelle spectrographs for high precision radial velocity work. The routines are designed to operate on data with extracted 1D spectra. The library provides a convienience function which returns a list of wavelengths from just a list of spectral feature coordinates (pixel and order) and a reference line list. The returned wavelengths are the wavelengths of the measured spectral features under the best fit wavelength model. xwavecal also provides line identification and spectral reduction utilities. The library is modular; each step of the wavelength calibration is a stage which can be disabled by removing the associated line in the config.ini file. Wavelength calibrating data which already have spectra means only using the wavelength calibration stages. Using the full experimental pipeline means enabling the other data reduction stages, such as overscan subtraction.
sf_deconvolve performs PSF deconvolution using a low-rank approximation and sparsity. It can handle a fixed PSF for the entire field or a stack of PSFs for each galaxy position. The code accepts Numpy binary files or FITS as input, takes the observed (i.e. with PSF effects and noise) stack of galaxy images and a known PSF, and attempts to reconstruct the original images. sf_deconvolve can be run in a terminal or in an active Python session, and includes options for initialization, optimization, low-Rank approximation, sparsity, PSF estimation, and other attributes.
Hazma enables indirect detection of sub-GeV dark matter. It computes gamma-ray and electron/positron spectra from dark matter annihilations, sets limits on sub-GeV dark matter using existing gamma-ray data, and determines the discovery reach of future gamma-ray detectors. The code also derives accurate CMB constraints. Hazma comes with several sub-GeV dark matter models, for which it provides functions to compute dark matter annihilation cross sections and mediator decay widths. A variety of low-level tools are provided to make it straightforward to define new models.
panco2 extracts measurements of the pressure profile of the hot gas inside galaxy clusters from millimeter-wave observations. The extraction is performed using forward modeling the millimeter-wave signal of clusters and MCMC sampling of a posterior distribution for the parameters given the input data. Many characteristic features of millimeter-wave observations can be taken into account, such as filtering (both through PSF smearing and transfer functions), point source contamination, and correlated noise.
PyMCCF (Python Modernized Cross Correlation Function), also known as MCCF, cross correlates two light curves that are unevenly sampled using linear interpolation and measures the peak and centroid of the cross-correlation function. Based on PyCCF (ascl:1805.032) and ICCF, it introduces a new parameter, MAX, to reduce the number of interpolated points used to just those which are not farther from the nearest real one than the MAX. This significantly reduces noise from interpolation errors. The estimation of the errors in PyMCCF is exactly the same as in PyCCF.
GPry efficiently obtains marginal quantities from computationally expensive likelihoods. It works best with smooth (continuous) likelihoods and posteriors that are slow to converge by other methods, which is dependent on the number of dimensions and expected shape of the posterior distribution. The likelihood should be low-dimensional (d<20 as a rule of thumb), though the code may still provide considerable improvements in speed in higher dimensions, despite an increase in the computational overhead of the algorithm. GPry is an alternative to samplers such as MCMC and Nested Sampling with a goal of speeding up inference in cosmology, though the software will work with any likelihood that can be called as a python function. It uses Cobaya's (ascl:1910.019) model framework so all of Cobaya's inbuilt likelihoods work, too.
MTNeedlet uses needlets to filter spherical (Healpix) maps and detect and analyze the maxima population using a multiple testing approach. It has been developed with the CMB in mind, but it can be applied to other spherical maps. It pivots around three basic steps: 1.) The calculation of several types of needlets and their possible use to filter maps; 2.) The detection of maxima (or minima) on spherical maps, their visualization and basic analysis; and 3.) The multiple testing approach in order to detect anomalies in the maxima population of the maps with respect to the expected behavior for a random Gaussian map. MTNeedlet relies on Healpy (ascl:2008.022) to efficiently deal with spherical maps.
FastDF (Fast Distribution Function) integrates relativistic particles along geodesics in a comoving periodic volume with forces determined by cosmological linear perturbation theory. Its main application is to set up accurate particle realizations of the linear phase-space distribution of massive relic neutrinos by starting with an analytical solution deep in radiation domination. Such particle realizations are useful for Monte Carlo experiments and provide consistent initial conditions for cosmological N-body simulations. Gravitational forces are calculated from three-dimensional potential grids, which are obtained by convolving random phases with linear transfer functions using Fast Fourier Transforms. The equations of motion are solved using a symplectic leapfrog integration scheme to conserve phase-space density and prevent the build-up of errors. Particles can be exported in different gauges and snapshots are provided in the HDF5 format, compatible with N-body codes like SWIFT (ascl:1805.020) and Gadget-4 (ascl:2204.014). The code has an interface with CLASS (ascl:1106.020) for calculating transfer functions and with monofonIC (ascl:2008.024) for setting up initial conditions with dark matter, baryons, and neutrinos.
MGCosmoPop implements a hierarchical Bayesian inference method for constraining the background cosmological history, in particular the Hubble constant, together with modified gravitational-wave propagation and binary black holes population models (mass, redshift and spin distributions) with gravitational-wave data. It includes support for loading and analyzing data from the GWTC-3 catalog as well as for generating injections to evaluate selection effects, and features a module to run in parallel on clusters.
Eventdisplay reconstructs and analyzes data from the Imaging Atmospheric Cherenkov Telescopes (IACT). It has been primarily developed for VERITAS and CTA analysis. The package calibrates and parametrizes images, event reconstruction, and stereo analysis, and provides train boosted decision trees for direction and energy reconstruction. It fills and uses lookup tables for mean scaled width and length calculation, energy reconstruction, and stereo reconstruction, and calculates radial camera acceptance from data files and instrument response functions such as effective areas, angular point-spread function, and energy resolution. Eventdisplay offers additional tools as well, including tools for calculating sky maps and spectral energy distribution, and to plot instrument response function, spectral energy distributions, light curves, and sky maps, among others.
GWFAST forecasts the signal-to-noise ratios and parameter estimation capabilities of networks of gravitational-wave detectors, based on the Fisher information matrix approximation. It is designed for applications to third-generation gravitational-wave detectors. It is based on Automatic Differentiation, which makes use of the library JAX (ascl:2111.002). This allows efficient parallelization and numerical accuracy. The code includes a module for parallel computation on clusters.
Many fields in science and engineering measure data that inherently live on non-Euclidean geometries, such as the sphere. Techniques developed in the Euclidean setting must be extended to other geometries. Due to recent interest in geometric deep learning, analogues of Euclidean techniques must also handle general manifolds or graphs. Often, data are only observed over partial regions of manifolds, and thus standard whole-manifold techniques may not yield accurate predictions. In this thesis, a new wavelet basis is designed for datasets like these.
Although many definitions of spherical convolutions exist, none fully emulate the Euclidean definition. A novel spherical convolution is developed, designed to tackle the shortcomings of existing methods. The so-called sifting convolution exploits the sifting property of the Dirac delta and follows by the inner product of a function with the translated version of another. This translation operator is analogous to the Euclidean translation in harmonic space and exhibits some useful properties. In particular, the sifting convolution supports directional kernels; has an output that remains on the sphere; and is efficient to compute. The convolution is entirely generic and thus may be used with any set of basis functions. An application of the sifting convolution with a topographic map of the Earth demonstrates that it supports directional kernels to perform anisotropic filtering.
Slepian wavelets are built upon the eigenfunctions of the Slepian concentration problem of the manifold - a set of bandlimited functions which are maximally concentrated within a given region. Wavelets are constructed through a tiling of the Slepian harmonic line by leveraging the existing scale-discretised framework. A straightforward denoising formalism demonstrates a boost in signal-to-noise for both a spherical and general manifold example. Whilst these wavelets were inspired by spherical datasets, like in cosmology, the wavelet construction may be utilised for manifold or graph data.
EXCEED-DM (EXtended Calculation of Electronic Excitations for Direct detection of Dark Matter) provides a complete framework for computing DM-electron interaction rates. Given an electronic configuration, EXCEED-DM computes the relevant electronic matrix elements, then particle physics specific rates from these matrix elements. This allows for separation between approximations regarding the electronic state configuration, and the specific calculation being performed.
APERO (A PipelinE to Reduce Observations) performs data reduction for the Canada-France-Hawaii Telescope's near-infrared spectropolarimeter SPIRou and offers different recipes or modules for performing specific tasks. APERO can individually run recipes or process a set of files, such as cleaning a data file of detector effects, collecting all dark files and creating a master dark image to use for correction, and creating a bad pixel mask for identifying and dealing with bad pixels. It can extract out flat images to measure the blaze and produced blaze correction and flat correction images, extract dark frames to provide correction for the thermal background after extraction of science or calibration frames, and correct extracted files for leakage coming from a FP (for OBJ_FP files only). It can also take a hot star and calculate telluric transmission, and then use the telluric transmission to calculate principle components (PCA) for correcting input images of atmospheric absorption, among many other tasks.
ODNet uses a convolutional neural network to examine frames of a given observation, using the flux of a targeted star along time, to detect occultations. This is particularly useful to reliably detect asteroid occultations for the Unistellar Network, which consists of 10,000 digital telescopes owned by citizen scientists that is regularly used to record asteroid occultations. ODNet is not costly in term of computing power, opening the possibility for embedding the code on the telescope directly. ODNet's models were developed and trained using TensorFlow version 2.4.
BiGONLight (Bi-local geodesic operators framework for numerical light propagation) encodes the Bi-local Geodesic Operators formalism (BGO) to study light propagation in the geometric optics regime in General Relativity. The parallel transport equations, the optical tidal matrix, and the geodesic deviation equations for the bilocal operators are expressed in 3+1 form and encoded in BiGONLight as Mathematica functions. The bilocal operators are used to obtain all possible optical observables by combining them with the observer and emitter four-velocities and four-accelerations. The user can choose the position of the source and the observer anywhere along the null geodesic with any four-velocities and four-accelerations.
Korg computes stellar spectra from 1D model atmospheres and linelists assuming local thermodynamic equilibrium and implements both plane-parallel and spherical radiative transfer. The code is generally faster than other codes, and is compatible with automatic differentiation libraries and easily extensible, making it ideal for statistical inference and parameter estimation applied to large data sets.
H-FISTA (Hierarchical Fast Iterative Shrinkage Thresholding Algorithm) retrieves the phases of the wavefield from intensity measurements for pulsar spectroscopy. The code accepts input data in ASCII format as produced by PSRchive's (ascl:1105.014) psrflux function, a FITS file, or a pickle. If using a notebook, any custom reader can be used as long as the data ends up in a NumPy array. H-FISTA obtains sparse models of the wavefield in a hierarchical approach with progressively increasing depth. Once the tail of the noise distribution is reached, the hierarchy terminates with a final unregularized optimization, resulting in a fully dense model of the complex wavefield that permits the discovery of faint signals by appropriate averaging.
PDFchem models the cold ISM at moderate and large scales using functions connecting the quantities of the local and the observed visual extinctions and the local number density with probability density functions. For any given observed visual extinction sampled with thousands of clouds, the algorithm instantly computes the average abundances of the most important species and performs radiative transfer calculations to estimate the average emission of the most commonly observed lines.
The Python module 2DFFTUtils implements tasks associated with measuring spiral galaxy pitch angle with 2DFFT (ascl:1608.015). Since most of the 2DFFT utilities are implemented in one place, it makes preparing images for 2DFFT and dealing with 2DFFT data interactively or in scripts event easier.
gsf fits photometric data points, simultaneously with grism spectra if provided, to get posterior probability of galaxy physical properties, such as stellar mass, dust attenuation, metallicity, as well as star formation and metallicity enrichment histories. Designed for extra-galactic science, this flexible, python-based SED fitting code involves a Markov-Chain Monte-Carlo (MCMC) process, and may take more time (depending on the number of parameters and length of MCMC chains) than other SED fitting codes based on chi-square minimization.
fastSHT performs spherical harmonic transforms on a large number of spherical maps. It converts massive SHT operations to a BLAS level 3 problem and uses the highly optimized matrix multiplication toolkit to accelerate the computation. GPU acceleration is supported and can be very effective. The core code is written in Fortran, but a Python wrapper is provided and recommended.
BlackJAX is a sampling library designed for ease of use, speed, and modularity and works on CPU as well as GPU. It is not a probabilistic programming library (PLL), though it integrates well with PPLs as long as they can provide a (potentially unnormalized) log-probability density function compatible with JAX. BlackJAX is written in pure Python and depends on XLA via JAX (ascl:2111.002). It can be used by those who have a logpdf and need a sampler or need more than a general-purpose sampler. It is also useful for building a sample on GPU and for users who want to learn how sampling algorithms work.
ovejero conducts hierarchical inference of strongly-lensed systems with Bayesian neural networks. It requires lenstronomy (ascl:1804.012) and fastell (ascl:9910.003) to run lens models with elliptical mass distributions. The code trains Bayesian Neural Networks (BNNs) to predict posteriors on strong gravitational lensing images and can integrate with forward modeling tools in lenstronomy to allow comparison between BNN outputs and more traditional methods. ovejero also provides hierarchical inference tools to generate population parameter estimates and unbiased posteriors on independent test sets.
The Population Monte-Carlo (PMC) sampling code pmclib performs fast end efficient parallel iterative importance sampling to compute integrals over the posterior including the Bayesian evidence.
mgcnn is a Convolutional Neural Network (CNN) architecture for classifying standard and modified gravity (MG) cosmological models based on the weak-lensing convergence maps they produce. It is implemented in Keras using TensorFlow as the backend. The code offers three options for the noise flag, which correspond to noise standard deviations, and additional options for the number of training iterations and epochs. Confusion matrices and evaluation metrics (loss function and validation accuracy) are saved as numpy arrays in the generated output/ directory after each iteration.
baobab generates images of strongly-lensed systems, given some configurable prior distributions over the parameters of the lens and light profiles as well as configurable assumptions about the instrument and observation conditions. Wrapped around lenstronomy (ascl:1804.012), baobab supports prior distributions ranging from artificially simple to empirical. A major use case for baobab is the generation of training and test sets for hierarchical inference using Bayesian neural networks (BNNs); the code can generate the training and test sets using different priors.
unTimely Catalog Explorer searches for and visualizes detections in the unTimely Catalog, a full-sky, time-domain catalog of detections based on WISE and NEOWISE image data acquired between 2010 and 2020. The tool searches the catalog by coordinates to create finder charts for each epoch with overplotted catalog positions and light curves using the unTimely photometry, to overplot these light curves with AllWISE multi-epoch and NEOWISE-R single exposure (L1b) photometry, and to create image blinks with overlaid catalog positions in GIF format.
PAHDecomp models mid-infrared spectra of galaxies; it is based on the popular PAHFIT code (ascl:1210.009). In contrast to PAHFIT, this model decomposes the continuum into a star-forming component and an obscured nuclear component based on Bayesian priors on the shape of the star-forming component (using templates + prior on extinction), making this tool ideally suited for modeling the spectra of heavily obscured galaxies. PAHDecomp successfully recovers properties of Compact Obscured Nuclei (CONs) where the inferred nuclear optical depth strongly correlates with the surface brightness of HCN-vib emission in the millimeter. This is currently set up to run on the short low modules of Spitzer IRS data (5.2 - 14.2 microns) but will be ideal for JWST/MIRI MRS data in the future.
AMBER (Abundance Matching Box for the Epoch of Reionization) models the cosmic dawn. The semi-numerical code allows users to directly specify the reionization history through the redshift midpoint, duration, and asymmetry input parameters. The reionization process is further controlled through the minimum halo mass for galaxy formation and the radiation mean free path for radiative transfer. The parallelized code is over four orders of magnitude faster than radiative transfer simulations and will efficiently enable large-volume models, full-sky mock observations, and parameter-space studies.
The analytic propagator Kepler-Collisions calculates collisions for Keplerian systems. The algorithm maintains a list of collision possibilities and jumps from one collision to the next; since collisions are rare in astronomical scales, jumping from collision to collision and calculating each one is more efficient than calculating all the time steps that are between collisions.
PTAfast calculates the overlap reduction function in Pulsar Timing Array produced by the stochastic gravitational wave background for arbitrary polarizations, propagation velocities, and pulsar distances.
cuvarbase provides a Python library for performing period finding (Lomb-Scargle, Phase Dispersion Minimization, Conditional Entropy, Box-least squares) on astronomical time-series datasets. Speedups over CPU implementations depend on the algorithm, dataset, and GPU capabilities but are typically ~1-2 orders of magnitude and are especially high for BLS and Lomb-Scargle.
paltas conducts simulation-based inference on strong gravitational lensing images. It builds on lenstronomy (ascl:1804.012) to create large datasets of strong lensing images with realistic low-mass halos, Hubble Space Telescope (HST) observational effects, and galaxy light from HST's COSMOS field. paltas also includes the capability to easily train neural posterior estimators of the parameters of the lensing system and to run hierarchical inference on test populations.
Cloud Killer recovers surface albedo maps by using reflected light photometry to map the clouds and surface of unresolved exoplanets. For light curves with negligible photometric uncertainties, the minimal top-of-atmosphere albedo at a location is a good estimate of its surface albedo. On synthetic data, it shows little bias, good precision, and accuracy, but slightly underestimated uncertainties; exoplanets with large, changing cloud structures observed near quadrature phases are good candidates for Cloud Killer cloud removal.
LensingETC optimizes observing strategies for multi-filter imaging campaigns of galaxy-scale strong lensing systems. It uses the lens modelling software lenstronomy (ascl:1804.012) to simulate and model mock imaging data, forecasts the lens model parameter uncertainties, and optimizes observing strategies.
PGOPHER simulates and fits rotational, vibrational, and electronic spectra. It handles linear molecules and symmetric and asymmetric tops, including effects due to unpaired electrons and nuclear spin, with a separate mode for vibrational structure. The code performs many sorts of transitions, including Raman, multiphoton, and forbidden transitions. It can simulate multiple species and states simultaneously, including special effects such as perturbations and state dependent predissociation. Fitting can be to line positions, intensities, or band contours. PGOPHER uses a standard graphical user interface and makes comparison with, and fitting to, spectra from various sources easy. In addition to overlaying numerical spectra, it is also possible to overlay pictures from pdf files and even plate spectra to assist in checking that published constants are being used correctly.
tvguide determines whether stars and galaxies are observable by TESS. It uses an object's right ascension and declination and estimates the pointing of TESS's cameras using predicted spacecraft ephemerides to determine whether and for how long the object is observable with TESS. tvguide returns a file with two columns, the first the minimum number of sectors the target is observable for and the second the maximum.
The Faiss library performs efficient similarity search and clustering of dense vectors. It contains algorithms that search in sets of vectors of any size, up to ones that possibly do not fit in RAM. It also contains supporting code for evaluation and parameter tuning. Faiss is written in C++ with complete wrappers for Python/numpy. Some of the most useful algorithms are implemented on the GPU.
BornRaytrace uses neural data compression of weak lensing map summary statistics to simulate weak gravitational lensing effects. It can raytrace through overdensity Healpix maps to return a convergence map, include shear-kappa transformation on the full sphere, and also include intrinsic alignments (NLA model).
MCCD (Multi-CCD) generates a Point Spread Function (PSF) model based on stars observations in the field of view. After defining the MCCD model parameters and running and validating the training, the model can recover the PSF at any position in the field of view. Written in Python, MCCD also calculates various statistics and can plot a random test star and its model reconstruction.
The photometric redshift-aided classification pipeline SHEEP uses ensemble learning to classify astronomical sources into galaxies, quasars and stars. It uses tabular data and also allows the use of sparse data. The approach uses SDSS and WISE photometry, but SHEEP can also be used with other types of tabular data, such as radio fluxes or magnitudes.
The simulation and analysis framework ixpeobssim was specifically developed for the Imaging X-ray Polarimetry Explorer (IXPE). It produces realistic simulated observations, in the form of event lists in FITS format, that also contain a strict superset of the information included in the publicly released IXPE data products. The framework's core simulation capabilities are complemented by post-processing applications that support the spatial, spectral, and temporal models needed for analysis of typical polarized X-ray sources, allowing implementation of complex, polarization-aware analysis pipelines. Where applicable, the data formats are consistent with the common display and analysis tools used by the community, e.g., the binned count spectra can be fed into XSPEC (ascl:9910.005), along with the corresponding response functions, for doing standard spectral analysis. All ixpeobssim simulation and analysis tools are fully configurable via the command line.
POSYDON (POpulation SYnthesis with Detailed binary-evolution simulatiONs) incorporates full stellar structure and evolution modeling for single and binary-star population synthesis. The code is modular and allows the user to specify initial population properties and adopt choices that determine how stellar evolution proceeds. Populations are simulated with the use of MESA (ascl:1010.083) evolutionary tracks for single, non-interacting, and interacting binaries organized in grids. Machine-learning methods are incorporated and applied on the grids for classification and various interpolation calculations, and the development of irregular grids guided by active learning, for computational efficiency.
LavAtmos performs gas-melt equilibrium calculations for a given temperature and melt composition. The thermodynamics of the melt are modeled by the MELTS code as presented in the Thermoengine package (ascl:2208.006). In combination with atmospheric chemistry codes, LavAtmos enables the characterization of interior compositions through atmospheric signatures.
PySME is a partial reimplementation of Spectroscopy Made Easy (SME, ascl:1202.013), which fits an observed spectrum of a star with a model spectrum. The IDL routines of SME used to call a dynamically linked library of compiled C++ and Fortran programs have been rewritten in Python. In addition, an object oriented paradigm and continuous integration practices, including build automation, self-testing, and frequent builds, have been added.
PETSc (Portable, Extensible Toolkit for Scientific Computation) provides a suite of data structures and routines for the scalable (parallel) solution of scientific applications modeled by partial differential equations, and is intended for use in large-scale application projects. The toolkit includes a large suite of parallel linear, nonlinear equation solvers and ODE integrators that are easily used in application codes written in C, C++, Fortran and Python. PETSc provides many of the mechanisms needed within parallel application codes, such as simple parallel matrix and vector assembly routines that allow the overlap of communication and computation. In addition, PETSc (pronounced PET-see) includes support for managing parallel PDE discretizations.
Solar-MACH (Solar MAgnetic Connection HAUS) derives and visualizes the spatial configuration and solar magnetic connection of different observers (i.e., spacecraft or planets) in the heliosphere at different times. It provides publication-ready figures for analyzing Solar Energetic Particle events (SEPs) or solar transients such as Coronal Mass Ejections (CMEs). Solar-MACH is available as a Python package; a Streamlit-enabled tool that runs in a browser is also available (solar-mach.github.io)
Blacklight postprocesses general-relativistic magnetohydrodynamic simulation data and produces outputs for analyzing data sets, including maps of auxiliary quantities and false-color renderings. The code can use Athena++ (ascl:1912.005) outputs directly, and also supports files in HARM (ascl:1209.005) and iHARM3d (ascl:2210.013) format. Written in C++, Blacklight offers support for adaptive mesh refinement input, slow-light calculations, and adaptive ray tracing.
iharm3D implements the HARM algorithm (ascl:1209.005) with modifications and enables a second-order, conservative, shock-capturing scheme for general-relativistic magnetohydrodynamics (GRMHD). Written in C, it simulates black hole accretion systems in arbitrary stationary spacetimes.
pixmappy provides a Python interface to gbdes pixel map (astrometry) solutions. It reads the YAML format astrometry solutions produced by gbdes (ascl:2210.011) and issues a PixelMap instance, which is a map from one 2d coordinate system ("pixel") to another ("world") 2d system. A PixelMap instance can be used as a function mapping one (or many) coordinate pairs. An inverse method does reverse mapping, and the local jacobian of the map is available also. The type of mapping that can be expressed is very flexible, and PixelMaps can be compounded into chains of tranformations.
gbdes derives photometric and astrometric calibration solutions for complex multi-detector astronomical imagers. The package includes routines to filter catalogs down to useful stellar objects, collect metadata from the catalogs and create a config file holding FITS binary tables describing exposures, instruments, fields, and other available information in the data, and uses a friends-of-friends matching algorithm to link together all detections of common objects found in distinct exposures. gbdes also calculates airmasses and parallactic angles for each exposure, calculates and saves the expected differential chromatic refraction (DCR) needed for precision astrometry, optimizes the parameters of a photometric model to maximize agreement between magnitudes measured in different exposures of the same source, and optimizing the parameters of an astrometric model to maximize agreement among the exposures and any reference catalogs, and performs other tasks. The solutions derived and used by gbdes are stored in YAML format; gbdes uses the Python code pixmappy (ascl:2210.012) to read the astrometric solution files and execute specified transformations.
The time series reconstruction method of massive astronomical catalogs reconstructs all celestial objects' time series data for astronomical catalogs with great accuracy. In addition, the program, which requires a Spark cluster, solves the boundary source leakage problem on the premise of ensuring accuracy, and the user can set different parameters for different data sets to filter the error records in the catalogs.
NEMESIS (Non-linear optimal Estimator for MultivariatE spectral analySIS) is the general purpose correlated-k/LBL retrieval code developed from the RADTRAN project (ascl:2210.008). Originally based on the correlated-k approximation, NEMESIS also works in line-by-line (LBL) mode. It has been designed to be generally applicable to any planet and with any observing mode and so is suitable for both solar-system studies and also exoplanetary studies.
RADTRAN calculates the transmission, absorption or emission spectra emitted by planetary atmospheres using either line-by-line integration, spectral band models, or 'correlated-K' approaches. Part of the NEMESIS project (ascl:2210.009), the code also incorporates both multiple scattering and single scattering calculations. RADTRAN is general purpose and not hard-wired to any specific planet.
COMET (Clustering Observables Modelled by Emulated perturbation Theory) provides emulated predictions of large-scale structure observables from models that are based on perturbation theory. It substantially speeds up these analytic computations without any relevant sacrifice in accuracy, enabling an extremely efficient exploration of large-scale structure likelihoods. At its core, COMET exploits an evolution mapping approach which gives it a high degree of flexibility and allows it to cover a wide cosmology parameter space at continuous redshifts up to z∼3z \sim 3z∼3. Among others, COMET supports parameters for cold dark matter density (ωc\omega_cωc), baryon density (ωb\omega_bωb), Scalar spectral index (nsn_sns), Hubble expansion rate (hhh) and Curvature density (ΩK\Omega_KΩK). The code can obtain the real-space galaxy power spectrum at one-loop order multipoles (monopole, quadrupole, hexadecapole) of the redshift-space, power spectrum at one-loop order, the linear matter power spectrum (with and without infrared resummation), Gaussian covariance matrices for the real-space power spectrum, and redshift-space multipoles and χ2\chi^2χ2's for arbitrary combinations of multipoles. COMET provides an easy-to-use interface for all of these computations.
ExoRad 2.0, a generic point source radiometric model, interfaces with any instrument to provide an estimate of several Payload performance metrics. For each target and for each photometric and spectroscopic channel, the code provides estimates of signals in pixels, saturation times, and read, photon, and dark current noise. ExoRad also provides estimates for the zodiacal background, inner sanctum, and sky foreground.
PSFr empirically reconstructs an oversampled version of the point spread function (PSF) from astronomical imaging observations. The code provides a light-weighted API of a refined version of an algorithm originally implemented in lenstronomy (ascl:1804.012). It provides user support with different artifacts in the data and supports the masking of pixels, or the treatment of saturation levels. PSFr has been used to reconstruct the PSF from multiply imaged lensed quasar images observed by the Hubble Space Telescope in a crowded lensing environment and more recently with James Webb Space Telescope (JWST) imaging data for a wide dynamical flux range.
Finder_charts creates multi-band finder charts from image data of various partial- and all-sky surveys such as DSS, 2MASS, WISE, UKIDSS, VHS, Pan-STARRS, and DES. It also creates a WISE time series of image data acquired between 2010 and 2021. All images are reprojected so that north is up and east is to the left. The resulting finder charts can be overplotted with corresponding catalog positions. All catalog entries within the specified field of view can be saved in a variety of formats, including ipac, csv, and tex, as can the finder charts in png, pdf, eps, and other common graphics formats. Finder_charts consists of a single Python module, which depends only on well-known packages, making it easy to install.
NIRDust uses K-band (2.2 micrometers) spectra to measure the temperature of the dust heated by an Active Galactic Nuclei (AGN) accretion disk. The package provides several functionalities to pre-process spectra and fit the hot dust component of a AGN K-band spectrum with a blackbody function. NIRDust needs a minimum of two spectra to run: a target spectrum, where the dust temperature will be estimated, and a reference spectrum, where the emission is considered to be purely stellar. The reference spectrum will be used by NIRDust to model the stellar emission from the target spectrum.
SPINspiral analyzes gravitational-wave signals from stellar-mass binary inspirals detected by ground-based interferometers such as LIGO and Virgo. It performs parameter estimation on these signals using Markov-chain Monte-Carlo (MCMC) techniques. This analysis includes the spins of the binary components. Written in C, the package is modular; its main routine is as small as possible and calls other routines, which perform tasks such as reading input, choosing and setting (starting or injection) parameters, and handling noise. Other routines compute overlaps and likelihoods, contain the MCMC core, and manage more general support functions and third-party routines.
Pulsar Survey Scraper aggregates pulsar discoveries before they are included in the ATNF pulsar catalog and enables searching and filtering based on position and dispersion measure. This facilitates identifying new pulsar discoveries. Pulsar Survey Scraper can be downloaded or run online using the Pulsar Survey Scraper webform.
EleFits is a modern C++ package to read and write FITS files which focuses on safety, user-friendliness, and performance.
FastQSL calculate the squashing factor Q at the photosphere, a cross section, or a box volume, given a 3D magnetic field with Cartesian, uniform or stretched grids. It is available in IDL and in an optimized version using Fortran for calculations and field line tracing. Use of a GPU accelerates a step-size adaptive scheme for the most computationally intensive part, the field line tracing, making the code fast and efficient.
SolTrack computes the position of the Sun, the rise and set times and azimuths, and transit times and altitudes. It includes corrections for aberration and parallax, and has a simple routine to correct for atmospheric refraction, taking into account local atmospheric conditions. SolTrack is derived from the Fortran library libTheSky (ascl:2209.018). The package can be used to track the Sun on a low-specs machine, such as a microcontroller or PLC, and can be used for (highly) concentrated (photovoltaic) solar power or accurate solar-energy modeling.
libTheSky compute the positions of celestial bodies, such as the Moon, planets, and stars, and events, including conjunctions and eclipses, with great accuracy. Written in Fortran, libTheSky can use different reference frames (heliocentric, geocentric, topocentric) and coordinate systems (ecliptic, equatorial, galactic; spherical, rectangular), and the user can choose low- or high-accuracy calculations, depending on need.
SpectraPy collects algorithms and methods for data reduction of astronomical spectra obtained by a through slits spectrograph. It produces two-dimensional wavelength calibrated spectra corrected by instrument distortions. The library is designed to be spectrograph independent and can be used on both longslit (LS) and multi object spectrograph (MOS) data. SpectraPy comes with a set of already configured spectrographs, but it can be easily configured to reduce data of other instruments.
RAPOC (Rosseland and Planck Opacity Converter) uses molecular absorption measurements (i.e., wavelength-dependent opacities) for a given temperature, pressure, and wavelength range to calculate Rosseland and Planck mean opacities for use in atmospheric modeling. The code interpolates between discrete data points and can use ExoMol and DACE data, or any user-defined data provided in a readable format. RAPOC is simple, straightforward, and easily incorporated into other codes.
TauREx 3 (Tau Retrieval for Exoplanets) provides a fully Bayesian inverse atmospheric retrieval framework for exoplanetary atmosphere modeling and retrievals. It is fully customizable, allowing the user to mix and match atmospheric parameters and add additional ones. The framework builds forward models, simulates instruments, and performs retrievals, and provides a rich library of classes for building additional programs and using new atmospheric parameters.
Synthetic Population of Interstellar Objects generates a synthetic population of interstellar objects (orbits and sizes) in arbitrary volume of space around the Sun. The only necessary assumption is that the population of ISOs in the interstellar space (far from any massive body) is homogeneous and isotropic. The assumed distribution of interstellar velocities of ISOs has to be provided as an input. This distribution can be defined analytically, but also in a discrete form. The algorithm, based on the multivariate inverse transform sampling method, is implemented in Python.
wsynphot provides a broad set of filters, including observation facility, instrument, and wavelength range, and functions for imaging stars to produce a filter curve showing the transmission of light for each wavelength value. It can create a filter curve object, plot the curve, and allows the user to do calculations on the filter curve object.
The time dependent Monte-Carlo code URILIGHT, written in Fortran 90, assumes homologous expansion. Energy deposition resulting from the decay of radioactive isotopes is calculated by a Monte-Carlo solution of the γ-ray transport, for which interaction with matter is included through Compton scattering and photoelectric absorption. The temperature is iteratively solved for in each cell by requiring that the total emissivity equals the total absorbed energy.
GaLight (Galaxy shapes of Light) performs two-dimensional model fitting of optical and near-infrared images to characterize the light distribution of galaxies with components including a disk, bulge, bar and quasar. Light is decomposes into PSF and Sersic, and the fitting is based on lenstronomy (ascl:1804.01). GaLight's automated features including searching PSF stars in the FOV, automatically estimating the background noise level, and cutting out the target object galaxies (QSOs) and preparing the materials to model the data. It can also detect objects in the cutout stamp and quickly create Sersic keywords to model them, and model QSOs and galaxies using 2D Sersic profile and scaled point source.
HyPhy maps from dark matter only simulations to full hydrodynamical physics models. It uses a fully convolutional variational auto-encoder (VAE) to synthesize hydrodynamic fields conditioned on dark matter fields from N-body simulations. After training, HyPhy can probabilistically map new dark matter only simulations to corresponding full hydrodynamical outputs and generate posterior samples for studying the variance of the mapping. This conditional deep generative model is implemented in TensorFlow.
GRUMPY (Galaxy formation with RegUlator Model in PYthon) models the formation of dwarf galaxies. When coupled with realistic mass accretion histories of halos from simulations and reasonable choices for model parameter values, this simple regulator-type framework reproduces a broad range of observed properties of dwarf galaxies over seven orders of magnitude in stellar mass. GRUMPY matches observational constraints on the stellar mass--halo mass relation and observed relations between stellar mass and gas phase and stellar metallicities, gas mass, size, and star formation rate. It also models the general form and diversity of star formation histories (SFHs) of observed dwarf galaxies. The software can be used to predict photometric properties of dwarf galaxies hosted by dark matter haloes in N-body simulations, such as colors, surface brightnesses, and mass-to-light ratios and to forward model observations of dwarf galaxies.
PINION (Physics-Informed neural Network for reIONization) predicts the complete 4-D hydrogen fraction evolution from the smoothed gas and mass density fields from pre-computed N-body simulations. Trained on C2-Ray simulation outputs with a physics constraint on the reionization chemistry equation, PINION accurately predicts the entire reionization history between z = 6 and 12 with only five redshift snapshots and a propagation mask as a simplistic approximation of the ionizing photon mean free path. The network's predictions are in good agreement with simulation to redshift z > 7, though the oversimplified propagation mask degrades the network's accuracy for z < 7.
AMBER (Apertif Monitor for Bursts Encountered in Real-time) detects single-pulse radio phenomena, such as pulsars and fast radio bursts, in real time. It is a fully auto-tuned pipeline that offloads compute-intensive kernels to many-core accelerators; the software automatically tunes these kernels to achieve high performance on different platforms.
KaRMMa (Kappa Reconstruction for Mass MApping) performs curved-sky mass map reconstruction using a lognormal prior from weak-lensing surveys. It uses a fully Bayesian approach with a physically motivated lognormal prior to sample from the posterior distribution of convergence maps. The posterior distribution of KaRMMa maps are nearly unbiased in one-point and two-point functions and peak/void counts. KaRMMa successfully captures the non-Gaussian nature of the distribution of κ values in the simulated maps, and KaRMMa posteriors correctly characterize the uncertainty in summary statistics.
The Shape COnstraint REstoration algorithm (SCORE) is a proximal algorithm based on sparsity and shape constraints to restore images. Its main purpose is to restore images while preserving their shape information. It can, for example, denoise a galaxy image by instanciating SCORE and using its denoise method and then visualize the results, and can deconvolve multiple images with different parameter values.
Cluster Toolkit calculates weak lensing signals from galaxy clusters and cluster cosmology. It offers 3D density and correlation functions, halo bias models, projected density and differential profiles, and radially averaged profiles. It also calculates halo mass functions, mass-concentration relations, Sunyaev-Zel’dovich (SZ) cluster signals, and cluster magnification. Cluster Toolkit consists of a Python front end wrapped around a well optimized back end in C.
DeepMass infers dark matter maps from weak gravitational lensing measurements and uses deep learning to reconstruct cosmological maps. The code can also be incorporated into a Moment Network to enable high-dimensional likelihood-free inference.
Herculens models imaging data of strong gravitational lenses. The package supports various degrees of model complexity, ranging from standard smooth analytical profiles to pixelated models and machine learning approaches. In particular, it implements multiscale pixelated models regularized with sparsity constraints and wavelet decomposition, for modeling both the source light distribution and the lens potential. The code is fully differentiable - based on JAX (ascl:2111.002) - which enables fast convergence to the solution, access to the parameters covariance matrix, efficient exploration of the parameter space including the sampling of posterior distributions using variational inference or Hamiltonian Monte-Carlo methods.
A-SLOTH (Ancient Stars and Local Observables by Tracing Halos) connects the formation of the first stars and galaxies to observables. The model is based on dark matter merger trees, on which A-SLOTH applies analytical recipes for baryonic physics to model the formation of both metal-free and metal-poor stars and the transition between them. The software samples individual stars and includes radiative, chemical, and mechanical feedback. A-SLOTH has versatile applications with moderate computational requirements. It can be used to constrain the properties of the first stars and high-z galaxies based on local observables, predicts properties of the oldest and most metal-poor stars in the Milky Way, can serve as a subgrid model for larger cosmological simulations, and predicts next-generation observables of the early Universe, such as supernova rates or gravitational wave events.
YONDER uses singular value decomposition to perform low-rank data denoising and reconstruction. It takes a tabular data matrix and an error matrix as input and returns a denoised version of the original dataset as output. The approach enables a more accurate data analysis in the presence of uncertainties. Consequently, this package can be used as a simple toolbox to perform astronomical data cleaning.
The toise framework estimates the sensitivity of natural-medium neutrino detectors such as IceCube-Gen2 to sources of high-energy astrophysical neutrinos. It uses parameterizations of a detector's fiducial area or volume, selection efficiency, energy resolution, angular resolution, and event classification efficiency to convert (surface) neutrino fluxes into mean event rates in bins of observable space. These are then used to estimate statistical quantities of interest, e.g., the median sensitivity to some flux (i.e., 90% upper limit assuming the true flux is zero) or the median discovery potential (i.e., the flux level at which the null hypothesis would be rejected at 5 sigma in 50% of realizations).
Cubefit is an OXY class that performs spectral fitting with spatial regularization in a spectro-imaging context. The 3D model is based on a 1D model and 2D parameter maps; the 2D maps are regularized using an L1L2 regularization by default. The estimator is a compound of a chi^2 based on the 1D model, a regularization term based of the 2D regularization of the various 2D parameter maps, and an optional decorrelation term based on the cross-correlation of specific pairs of parameter maps.
PyNAPLE (PYthon Nac Automated Pair Lunar Evaluator) detects changes and new impact craters on the lunar surface using Lunar Reconnaissance Orbiter Narrow Angle Camera (LRO NAC) images. The code enables large scale analyses of sub-kilometer scale cratering rates and refinement of both scaling laws and the luminous efficiency.
GSSP (Grid Search in Stellar Parameters) is based on a grid search in the fundamental atmospheric parameters and (optionally) individual chemical abundances of the star (or binary stellar components) in question. It uses atmosphere models and spectrum synthesis, which assumes a comparison of the observations with each theoretical spectrum from the grid. The code can optimize five stellar parameters at a time (effective temperature, surface gravity, metallicity, microturbulent velocity, and projected rotational velocity of the star) and synthetic spectra can be computed in any number of wavelength ranges. GSSP builds the grid of theoretical spectra from all possible combinations of the above mentioned parameters, and delivers the set of best fit parameters, the corresponding synthetic spectrum, and the ASCII file containing the individual parameter values for all grid points and the corresponding chi-square values.
GStokes performs simple multipolar fits to circular polarization data to provide information about the field strength and geometry. It provides forward calculation of the disc-integrated Stokes parameter profiles as well as magnetic inversions under several widely used simplifying approximations of the polarized line formation. GStokes implements the Unno–Rachkovsky analytical solution of the polarized radiative transfer equation and the weak-field approximation with the Gaussian local profiles. The magnetic field geometry is described with one of the common low-order multipolar field parametrizations. Written in IDL, GStokes provides a user-friendly graphical front-end.
RadioLensfit measures star-forming galaxy ellipticities using a Bayesian model fitting approach. The software uses an analytical exponential Sersic model and works in the visibility domain avoiding Fourier Transform. It also simulates visibilities of observed SF galaxies given a source catalog and Measurement Sets containing the description of the radio interferometer and of the observation. It provides both serial and MPI versions.
EstrellaNueva calculates expected rates of supernova neutrinos in detectors. It provides a link between supernova simulations and the expected events in detectors by calculating fluences and event rates in order to ease any comparison between theory and observation. The software is a standalone tool for exploring many physics scenarios, and offers an option to add analytical cross sections and define any target material.
HOCHUNK3D is an updated version of the HOCHUNK radiative equilibrium code (ascl:1711.013); the code has been converted to Fortran 95, which allows a specification of one-dimensional (1D), 2D, or 3D grids at runtime. The code is parallelized so it can be run on multiple processors on one machine, or on multiple machines in a network. It includes 3-D functionality and several other additional geometries and features. The code calculates radiative equilibrium temperature solution, thermal and PAH/vsg emission, scattering and polarization in protostellar geometries. HOCHUNK3D also computes spectral energy distributions (SEDs), polarization spectra, and images.
CRPropa3, an improved version of CRPropa2 (ascl:1412.013), provides a simulation framework to study the propagation of ultra-high-energy nuclei up to iron on their voyage through an (extra)galactic environment. It takes into account pion production, photodisintegration, and energy losses by pair production of all relevant isotopes in the ambient low-energy photon fields, as well as nuclear decay. CRPropa3 can model the deflection in (inter)galactic magnetic fields, the propagation of secondary electromagnetic cascades, and neutrinos for a multitude of scenarios for different source distributions and magnetic environments. It enables the user to predict the spectra of UHECR (and of their secondaries), their composition and arrival direction distribution. Additionally, the low-energy Galactic propagation can be simulated by solving the transport equation using stochastic differential equations. CRPropa3 features a very flexible simulation setup with python steering and shared-memory parallelization.
J-comb combines high-resolution data with large-scale missing information with low-resolution data containing the short spacing. Based on uvcombine (ascl:2208.014), it takes as input FITS files of low- and high-resolution images, the angular resolution of the input images, and the pixel size of the input images, and outputs a FITS file of the combined image.
uvcombine combines single-dish and interferometric data. It can combine high-resolution images that are missing large angular scales (Fourier-domain short-spacings) with low-resolution images containing the short/zero spacing. uvcombine includes the "feathering" technique for interferometry data, implementing a similar approach to CASA’s (ascl:1107.013) feather task but with additional options. Also included are consistency tests for the flux calibration and single-dish scale by comparing the data in the uv-overlap range.
SPAMMS (Spectroscopic PAtch Model for Massive Stars), designed with geometrically deformed systems in mind, combines the eclipsing binary modelling code PHOEBE 2 (ascl:1106.002) and the NLTE radiative transfer code FASTWIND to produce synthetic spectra for systems at given phases, orientations and geometries. SPAMMS reproduces the morphology of observed spectral line profiles for overcontact systems and the Rossiter-Mclaughlin and Struve-Sahade effects.
DELIGHT (Deep Learning Identification of Galaxy Hosts of Transients) automatically identifies host galaxies of transient candidates using multi-resolution images and a convolutional neural network. This library has a class with several methods to get the most likely host coordinates starting from given transient coordinates. In order to do this, the DELIGHT object needs a list of object identifiers and coordinates (oid, ra, dec). With this information, it downloads PanSTARRS images centered around the position of the transients (2 arcmin x 2 arcmin), gets their WCS solutions, creates the multi-resolution images, does some extra preprocessing of the data, and finally predicts the position of the hosts using a multi-resolution image and a convolutional neural network. DELIGHT can also estimate the host's semi-major axis if requested, taking advantage of the multi-resolution images.
POIS (Python Optical Interferometry Simulation) provides the building blocks to simulate the operation of a ground-based optical interferometer perturbed by atmospheric seeing perturbations. The package includes functions to generate simulated atmospheric turbulent wavefront perturbations, correct these perturbations using adaptive optics, and combine beams from an arbitrary number of telescopes, with or without spatial filtering, to provide complex fringe visibility measurements.
FFD (Flare Frequency Distribution) fits power-laws to FFDs. FFDs relate the frequency (i.e., occurrence rate) of flares to their energy, peak flux, photometric equivalent width, or other parameters. This module was created to handle disparate datasets between which the flare detection limit varies; in essence, the number of flares detected is treated as following a Poisson distribution while the flare energies are treated as following a power law.
LeXInt (Leja interpolation for eXponential Integrators) is a temporal exponential integration package using the method of polynomial interpolation at Leja points. Exponential Rosenbrock (EXPRB) and Exponential Propagation Iterative Runge-Kutta (EPIRK) methods use the Leja interpolation method to compute the functions. For linear PDEs, one can get the exact solution (in time) by directly computing the matrix exponential.
RJ-plots uses a moments of inertia method to disentangle a 2D structure's elongation from its centrally over/under-density, thus providing a means for the automated and objective classification of such structures. It may be applied to any 2D pixelated image such as column density maps or moment zero maps of molecular lines. This method is a further development of J-plots (ascl:2009.007).
VapoRock calculates the equilibrium partial pressures of metal-bearing gas species of specific elements above the magma ocean surface to determine the metal-bearing composition of the atmosphere as a function of temperature and the bulk composition of the magma ocean. It utilizes ENKI's ThermoEngine (ascl:2208.006) and combines estimates for element activities in silicate melts with thermodynamic data for metal and metal oxide vapor species.
ThermoEngine estimates the thermodynamic properties of minerals, fluids, and melts, and calculates phase equilibriums. The Equilibrate module of ThermoEngine provides Python functions and classes for computing equilibrium phase assemblages with focus on MELTS calculations. The Phases module includes Python functions and classes for computing standard thermodynamic calculations utilizing the Berman, Holland and Powell, or Stixrude-Lithgow-Bertelloni endmember databases, and calculations based on solution properties utilized by MELTS. There are many helper functions available in this module that assist in the calculation of pseudosections, univariant equilibria and the construction of phase diagrams.
Asymmetric Uncertainty implements and provides an object class for dealing with uncertainties for physical quantities that are not symmetric. Instances of the class behave appropriately with other numeric objects under most mathematical operations, and the associated errors propagate accordingly. The class also provides utilities such as methods for evaluating and plotting probability density functions, as well as capabilities for handling arrays of such objects. Standard and symmetric uncertainties are also supported.
The TOM Toolkit combines a flexible, searchable database of all information related to a scientific research project, with an observation and data analysis control system, and communication and data visualization tools. This Toolkit includes a fully operational TOM (Target and Observation Manager) system in addition to a range of optional tools for specific tasks, including interfaces to widely-used observing facilities and data archives and data visualization tools. With TOM Toolkit, project teams can develop and customize a system for their own science goals, without needing specialist expertise in databasing.
Scatfit models observed burst signals of impulsive time domain radio signals ( e.g., Fast Radio Bursts (FRBs) or pulsars pulses), which usually are convolution products of various effects, and fits them to the experimental data. It includes several models for scattering and instrumental effects. The code loads the experimental time domain radio data, cleans them, fits an aggregate scattering model to the data, and robustly estimates the model parameters and their uncertainties. Additionally, scatfit determines the scaling of the scattering time with frequency, i.e. the scattering index, and the scattering-corrected dispersion measure of the burst or pulse.
qrpca uses QR-decomposition for fast principal component analysis. The software is particularly suited for large dimensional matrices. It makes use of torch for internal matrix computations and enables GPU acceleration, when available. Written in both R and python languages, qrpca provides functionalities similar to the prcomp (R) and sklearn (python) packages.
BlaST (Blazar Synchrotron Tool) estimates the synchrotron peak of blazars given their spectral energy distribution. It uses a machine-learning algorithm that simplifies the estimation and also provides a reliable uncertainty estimation. The package naturally accounts for additional SED components from the host galaxy and the disk emission. BlaST also supports bulk estimation, e.g. estimating a whole catalog, by providing a directory or zip file containing the seds as well as an output file in which to write the results.
Eidein interactively visualizes a data sample for the selection of an informative (contains data with high predictive uncertainty, is diverse, but not redundant) data subsample for deep active learning. The data sample is projected to 2-D with a dimensionality reduction technique. It is visualized in an interactive scatter plot that allows a human expert to select and annotate the data subsample.
BMarXiv scans new (i.e., since the last time checked) submissions from arXiv, ranks submissions based on keyword matches, and produces an HTML page as an output.
The keywords are looked for (with regex capabilities) in the title, abstract, but also the author list, so it is possible to look for people too. The score is calculated for each specific entry but additional (and optional) scoring is performed using the first author recent submissions and/or the other authors' recent submissions.
It is possible to include/exclude any arXiv categories (within astro-ph or not). New astronomical conferences (from CADC by default) and new codes (from ASCL.net) are also checked and can also be scanned for keywords.
A local bibliography file can be scanned to find frequent words/groups of words that could become scanned keywords.
massmappy recovers convergence mass maps on the celestial sphere from weak lensing cosmic shear observations. It relies on SSHT (ascl:2207.034) and HEALPix (ascl:1107.018) to handle sampled data on the sphere. The spherical Kaiser-Squires estimator is implemented.
SSHT performs fast and exact spin spherical harmonic transforms; functionality is also provided to perform fast and exact adjoint transforms, forward and inverse transforms, and spherical harmonic transforms for a number of alternative sampling schemes. The code can interface with DUCC (ascl:2008.023) and use it as a backend for spherical harmonic transforms and rotations.
piXedfit provides a self-contained set of tools for analyzing spatially resolved properties of galaxies using imaging data or a combination of imaging data and the integral field spectroscopy (IFS) data. piXedfit has six modules that can handle all tasks in the analysis of the spatially resolved SEDs of galaxies, including images processing, a spatial-matching between reduced broad-band images with an IFS data cube, pixel binning, performing SED fitting, and making visualization plots for the SED fitting results.
gwdet computes the probability of detecting a gravitational-wave signal from compact binaries averaging over sky-location and source inclination. The code has two classes, averageangles and detectability. averageangles computes the detection probability, averaged over all angles (such as sky location, polarization, and inclination), as a function of the projection parameter. detectability computes the detection probability of a non-spinning compact binary.
BANZAI (Beautiful Algorithms to Normalize Zillions of Astronomical Images) processes raw data taken from Las Cumbres Observatory and produces science quality data products. It is capable of reducing single or multi-extension fits files. For historical data, BANZAI can also reduce the data cubes that were produced by the Sinistro cameras.
This code analyzes a dipole axis in the distribution of galaxy spin directions. The code takes as input a list of galaxies, their equatorial coordinates, and their spin directions. It then determines the statistical significance of possible dipole axis at any point in the sky by comparing the cosine dependence of the spin directions to the mean and standard deviation of the cosine dependence after 2000 runs with random spin directions. A code to analyze the binomial distribution of the spin directions using Monte Carlo simulation is also available.
ParticleGridMapper.jl interpolates particle data onto either a Cartesian (uniform) grid or an adaptive mesh refinement (AMR) grid where each cell contains no more than one particle. The AMR grid can be trimmed with a user-defined maximum level of refinement. Three different interpolation schemes are supported: nearest grid point (NGP), smoothed-particle hydrodynamics (SPH), and Meshless finite mass (MFM). It is multi-threading parallel.
disksurf measures the height of optically thick emission or photosphere in moderately inclined protoplanetary disks. The package is dependent on AstroPy (ascl:1304.002) and uses GoFish (ascl:2011.016) to retrieve data from FITS data cubes and user-specified parameters to return a surface object containing the disk-centric coordinates of the surface and the gas temperature and rotation velocity at those locations. disksurf provides clipping, smoothing, and diagnostic functions as well.
ConeRot extracts velocity perturbations in protoplanetary disks from observed line centroids maps ν∘, by creating axially-symmetric centroid maps. It also derives 3D rotation curves in disk-centered cylindrical coordinates, and can estimate the disk orientation based on line data alone. It approximates the unit opacity surface of an axially symmetric disc by a series of cones whose orientations are fit to the observed velocity centroid in concentric radial domains, or regions, with the disc orientation and the rotation curve both optimized to fit ν∘ in each region. ConeRot extracts the perturbations directly from observations without strong assumptions about the underlying disk model and employs a reduced number of free parameters.
pdspy fits Monte Carlo radiative transfer models for protostellar/protoplanetary disks to ALMA continuum and spectral line datasets using Markov Chain Monte Carlo fitting. It contains two tools, one to fit ALMA continuum visibilities and broadband spectral energy distributions (SEDs) with full radiative transfer models, and another to fit ALMA spectral line visibilities with protoplanetary disk models that include a vertically isothermal, power law temperature distribution. No radiative equilibrium calculation is done.
casa_cube provides an interface to data cubes generated by CASA (ascl:1107.013) or Gildas (ascl:1305.010). It performs simple tasks such as plotting given channel maps, moment maps, and line profile in various units, and also corrects for cloud extinction, reconvolves with a beam taper, and permits quick and easy comparisons with models.
pymcfost provides an interface to and can be used to visualize results from the 3D radiative transfer code MCFOST (ascl:2207.023). pymcfost can set up continuum and line models, read a single model or library of models, plot basic quantities such as density structures and temperature maps, and plot observables, including SEDs, polarization maps, visibilities, and channels maps (with spatial and spectral convolution). It can also convert units (e.g. W.m-2 to Jy or brightness temperature), and it provides an interface to the ALMA CASA simulator (ascl:1107.013).
MCFOST is a 3D continuum and line radiative transfer code based on an hybrid Monte Carlo and ray-tracing method. It is mainly designed to study the circumstellar environment of young stellar objects, but has been used for a wide range of astrophysical problems. The calculations are done exactly within the limitations of the Monte Carlo noise and machine precision, i.e., no approximation are used in the calculations. The code has been strongly optimized for speed.
MCFOST is primarily designed to study protoplanetary disks. The code can reproduce most of the observations of disks, including SEDs, scattered light images, IR and mm visibilities, and atomic and molecular line maps. As the Monte Carlo method is generic, any complex structure can be handled by MCFOST and its use can be extended to other astrophysical objects. For instance, calculations have succesfully been performed on infalling envelopes and AGB stars. MCFOST also includes a non-LTE line transfer module, and NLTE level population are obtained via iterations between Monte Carlo radiative transfer calculations and statistical equilibrium.
triple-stability uses a simple form of an artificial neural network, a multi-layer perceptron, to check whether a given configuration of a triple-star system is dynamically stable. The code is written in Python and the MLP classifier can be imported to other custom Python3 scripts.
BAYGAUD (BAYesian GAUssian Decomposer) implements the decomposition of velocity profiles in a data cube and subsequent classification. It uses MultiNest (ascl:1109.006) for calculating the posterior distribution and the evidence for a given likelihood function. The code models a given line profile with an optimal number of Gaussians based on the Bayesian Markov Chain Monte Carlo (MCMC) techniques. BAYGAUD is parallelized using the Message-Passing Interface (MPI) standard, which reduces the time needed to calculate the evidence using MCMC techniques.
vKompth fits the energy-dependent rms-amplitude and phase-lag spectra of low-frequency quasi-periodic oscillations in low mass black-hole X-ray binaries using a variable Comptonization model. The accretion disc is modeled as a multi-temperature blackbody source emitting soft photons which are then Compton up-scattered in a spherical corona, including feedback of Comptonized photons that return to the disc.
walter calculates the number density of stars detected in a given observation aiming to resolve a stellar population. The code also calculates the exposure time needed to reach certain population features, such as the horizontal branch, and provides an estimate of the crowding limit. walter was written with the expectation that such calculations will be very useful for planning surveys with the Nancy Grace Roman Space Telescope (RST, formerly WFIRST).
pocoMC performs Bayesian inference, including model comparison, for challenging scientific problems. The code utilizes a normalizing flow to precondition the target distribution by removing any correlations between its parameters. pocoMC then generates posterior samples, used for parameter estimation, with a powerful adaptive Sequential Monte Carlo algorithm manifesting a sampling efficiency that can be orders of magnitude higher than without precondition. Furthermore, pocoMC also provides an unbiased estimate of the model evidence that can be used for the task of Bayesian model comparison. The code is designed to excel in demanding parameter estimation problems that include multimodal and highly non–Gaussian target distributions.
LOTUS (non-LTE Optimization Tool Utilized for the derivation of atmospheric Stellar parameters) derives stellar parameters via Equivalent Width (EW) method with the assumption of 1D non-local thermodynamic equilibrium. It mainly applies on the spectroscopic data from high resolution spectral survey. It can provide extremely accurate measurement of stellar parameters compared with non-spectroscopic analysis from benchmark stars. LOTUS provides a fast optimizer for obtaining stellar parameters based on Differential Evolution algorithm, well constrained uncertainty of derived stellar parameters from slice-sampling MCMC from PyMC3 (ascl:1610.016), and can interpolate the Curve of Growth from theoretical EW grid under the assumptions of LTE and Non-LTE. It also visualizes excitation and ionization balance when at the optimal combination of stellar parameters.
DustPy simulates the radial evolution of gas and dust in protoplanetary disks, involving viscous evolution of the gas disk and advection and diffusion of the dust disk, as well as dust growth by solving the Smoluchowski equation. The package provides a standard simulation and the ability to plot results, and also allows modification of the initial conditions for dust, gas, the grid, and the central star.
calviacat calibrates star photometry by comparison to a catalog, including PanSTARRS 1, ATLAS-RefCat2, and SkyMapper catalogs. Catalog queries are cached so that subsequent calibrations of the same or similar fields can be more quickly executed.
petitRADTRANS (pRT) calculates transmission and emission spectra of exoplanets for clear and cloudy planets. It also incorporates an easy subpackage for running retrievals with nested sampling. It allows the calculation of emission or transmission spectra, at low or high resolution, clear or cloudy, and includes a retrieval module to fit a petitRADTRANS model to spectral data. pRT has two different opacity treatment modes. The low resolution mode runs calculations at λ/Δλ ≤ 1000 using the so-called correlated-k treatment for opacities. The high resolution mode runs calculations at λ/Δλ ≤ 106, using a line-by-line opacity treatment.
MuSCAT2_transit_pipeline provides photometry and transit analysis pipelines for MuSCAT2. It consists of a set of executable scripts and two Python packages: muscat2ph for photometry, and muscat2ta for transit analysis. The MuSCAT2 photometry can be carried out using the scripts only. The transit analysis can also in most cases be done using the main transit analysis script m2fit, but the muscat2ta package also offers high-level classes that can be used to carry out more customized transit analysis as a Python script (or Jupyter notebook).
The Exoplanet Characterization ToolKit (ExoCTK) focuses primarily on the atmospheric characterization of exoplanets and provides tools for time-series observation planning, forward modeling, data reduction, limb darkening, light curve fitting, and retrievals. It contains calculators for contamination, visibility, integrations and groups, and includes several Jupyter Notebooks to aid in learning how to use the various tools included in the ExoCTK package.
The samsam package provides two samplers, a scaled adaptive metropolis algorithm to robustly obtain samples from a target distribution, and a covariance importance sampling algorithm to efficiently compute the model evidence (or other integrals). It also includes tools to assess the convergence of the sam sampler and a few commonly used prior distributions.
Helios-r2 performs atmospheric retrieval of brown dwarf and exoplanet spectra. It uses a Bayesian statistics approach by employing a nested sampling method to generate posterior distributions and calculate the Bayesian evidence. The nested sampling itself is done by Multinest (ascl:1109.006). The computationally most demanding parts of the model have been written in NVIDIA's CUDA language for an increase in computational speed. Successful applications include retrieval of brown dwarf emission spectra and secondary eclipse measurements of exoplanets.
SolAster provides querying, analysis, and calculation methods to independently derive 'sun-as-a-star' RV variations using SDO/HMI data for any time span since SDO has begun observing. Scaling factors are provided in order to calculate RVs comparable to magnitudes measured by ground-based spectrographs (HARPS-N and NEID). In addition, there are routines to calculate magnetic observables to compare with RV variations and determine what is driving Solar activity.
TESS_PRF displays the TESS pixel response function (PRF) at any location on the detector. The package is primarily for estimating how the light from a point source is distributed given its position in a TESS Target Pixel File (TPF) or TESScut postage stamp. By default, it accesses the relevant PRF files on MAST, but can also reference files on a local directory. TESS_PRF assumes the PRF doesn't change considerably within a small TPF. The PRF model can be positioned by passing the relative row and column location within the TPF to the "resample" method. The pixel locations follow WCS convention, that an integer value corresponds to the center of a pixel.
Pyriod provides basic period detection and fitting routines for astronomical time series. Written in Python and designed to be run interactively in a Jupyter notebook, it displays and allows the user to interact with time series data, fit frequency solutions, and save figures from the toolbar. It can display original or residuals time series, fold the time series on some frequency, add selected peaks from the periodogram to the model, and refine the fit by computing a least-squared fit of the model using Lmfit (ascl:1606.014).
MultiModes extracts the most significant frequencies of a sample of classical pulsating stars. The code takes a directory with light curves and initial parameters as input. For every light curve, the code calculates the frequencies spectrum, or periodogram, with the Fast Lomb Scargle algorithm, extracts the higher amplitude peak, and evaluates whether it is a real signal or noise. It fits frequency, amplitude, and phase through non-linear optimization, using a multisine function. This function is redefined with the new calculated parameters. MultiModes then does a simultaneous fit of a number of peaks (20 by default), subtracts them from the original signal, and goes back to the beginning of the loop with the residual, repeating the same process until the stop criterion is reached. After that, the code can filter suspicious spurious frequencies, those of low amplitude below the Rayleigh resolution, and possible combined frequencies.
Echelle diagrams are used mainly in asteroseismology, where they function as a diagnostic tool for estimating Δν, the separation between modes of the same degree ℓ; the amplitude spectrum of a star is stacked in equal slices of Δν, the large separation. The echelle Python code creates and manipulates echelle diagrams. The code provides the ability to dynamically change Δν for rapid identification of the correct value. echelle features performance optimized dynamic echelle diagrams and multiple backends for supporting Jupyter or terminal usage.
Cosmic-kite performs a fast estimation of the TT Cosmic Microwave Background (CMB) power spectra corresponding to a set of cosmological parameters; it can also estimate the maximum-likelihood cosmological parameters from a power spectra. This software is an auto-encoder that was trained and calibrated using power spectra from random cosmologies computed with the CAMB code (ascl:1102.026).
MeSsI performs an automatic classification between merging and relaxed clusters. This method was calibrated using mock catalogues constructed from the millennium simulation, and performs the classification using some machine learning techniques, namely random forest for classification and mixture of gaussians for the substructure identification.
pynucastro interfaces to the nuclear reaction rate databases, including the JINA Reaclib nuclear reactions database. This set of Python interfaces enables interactive exploration of rates and collection of rates (networks) in Jupyter notebooks and easy creation of the righthand side routines for reaction network integration (the ODEs) for use in simulation codes.
MULTIGRIS (also called mgris) uses the sequential Monte Carlo method in PyMC (ascl:1506.005) to extract the posterior distributions of primary grid parameters and predict unobserved parameters/observables. The code accepts either a discrete number of components and/or continuous (e.g., power-law, normal) distributions for any given parameter. MULTIGRIS, written in Python, infers the posterior probability functions of parameters in a multidimensional potentially incomplete grid with some observational tracers defined for each parameter set. Observed values and their potentially asymmetric uncertainties are used to calculate a likelihood which, together with predefined or custom priors, produces the posterior distributions. Linear combinations of parameter sets may be used with inferred mixing weights and nearest neighbor or linear interpolation may be used to sample the parameter space.
Modern cosmological surveys are delivering datasets characterized by unprecedented quality and statistical completeness. In order to maximally extract cosmological information from these observations, matching theoretical predictions are needed. In the nonlinear regime of structure formation, cosmological simulations are the primary means of obtaining the required information but the computational cost of sufficiently resolved large-volume simulations makes it prohibitive to run very large ensembles. Nevertheless, precision emulators built on a tractable number of high-quality simulations can be used to build very fast prediction schemes to enable a variety of cosmological inference studies. The "Mira-Titan Universe" simulation suite covers the standard six cosmological parameters and, in addition, includes massive neutrinos and a dynamical dark energy equation of state. It is based on 111 cosmological simulations, each covering a (2.1Gpc)^3 volume and evolving 3200^3 particles, and augments these higher-resolution simulations with an additional set of 1776 lower-resolution simulations and TimeRG perturbation theory results to cover scales straddling the linear to mildly nonlinear regimes. The emulator built on this suite, the CosmicEmu, provides predictions at the two to three percent level of accuracy over a wide range of cosmological parameters. Presented in: https://arxiv.org/abs/2207.12345.
The software used to transform the tabular USNO/AE98 asteroid ephemerides into a Chebyshev polynomial representations, and evaluate them at an arbitrary time is available. The USNO/AE98 consisted of the ephemerides of fifteen of the largest asteroids, and were used in The Astronomical Almanac from 2000 through 2015. These ephemerides are outdated and no longer available, but the software used to store and evaluate them is still available and provides a robust method for storing compact ephemerides of solar system bodies.
The object of the software is to provide a compact binary representation of solar system bodies with eccentric orbits, which can produce the body's position and velocity at an arbitrary instant within the ephemeris' time span. It uses a modification of the Newhall (1989) algorithm to achieve this objective. The Newhall algorithm is used to store both the Jet Propulsion Laboratory DE and the Institut de mécanique céleste et de calcul des éphémérides INPOP high accuracy planetary ephemerides. The Newhall algorithm breaks an ephemeris into a number time contiguous segments, and each segment is stored as a set of Chebyshev polynomial coefficients. The length of the time segments and the maximum degree Chebyshev polynomial coefficient is fixed for each body. This works well for bodies with small eccentricities, but it becomes inefficient for a body in a highly eccentric orbit. The time segment length and maximum order Chebyshev polynomial coefficient must be chosen to accommodate the strong curvature and fast motion near pericenter, while the body spends most of its time either moving slowly near apocenter or in the lower curvature mid-anomaly portions of its orbit. The solution is to vary the time segment length and maximum degree Chebyshev polynomial coefficient with the body's position. The portion of the software that converts tabular ephemerides into a Chebyshev polynomial representation (CPR) performs this compaction automatically, and the portion that evaluates that representation requires only a modest increase in the evaluation time.
The software also allows the user to choose the required tolerance of the CPR. Thus, if less accuracy is required a more compact, somewhat quicker to evaluate CPR can be manufactured and evaluated. Numerical tests show that a fractional precision of 4e-16 may be achieved, only a factor of 4 greater than the 1e-16 precision of a 64-bit IEEE (2019) compliant floating point number.
The software is written in C and designed to work with the C edition of the Naval Observatory Vector Astrometry Software (NOVAS). The programs may be used to convert tabular ephemerides of other solar system bodies as well. The included READ.ME file provides the details of the software and how to use it.
REFERENCES
IEEE Computer Society 2019, IEEE Standard for Floating-Point Arithmetic. IEEE STD 754-2019, IEEE, pp. 1–84
Newhall, X X 1989, 'Numerical Representation of Planetary Ephemerides,' Celest. Mech., 45, 305 - 310
The Spritz code is a fully general relativistic magnetohydrodynamic code based on the Einstein Toolkit (ascl:1102.014). The code solves the GRMHD equations in 3D Cartesian coordinates and on a dynamical spacetime. Spritz supports tabulated equations of state, takes finite temperature effects into account and allows for the inclusion of neutrino radiation.
DustFilaments paints filaments in the Celestial Sphere to generate a full sky map of the Thermal Dust emission at millimeter frequencies by integrating a population of 3D filaments. The code requires a magnetic field cube, which can be calculated separately or by DustFilaments. With the magnetic field cube as input, the package creates a random filament population with a given seed, and then paints a filament into a healpix map provided as input; the healpix map is updated in place.
ShapePipe processes single-exposure images and stacked images. Input images have to be calibrated beforehand for astrometry and photometry. The code can handle different image and file types, such as single-exposure mosaic, single-exposure single-CCD, stacked images, database catalog files, and PSF files, some of which are created by the pipeline during the analysis, among others. The end product of ShapePipe is a final catalog containing information for each galaxy, including its shape parameters and the ellipticity components :math:e_1 and :math:e_2. This catalog also contains shapes of artificially sheared images. This information is used in post-processing to compute calibrated shear estimates via metacalibration.
CuspCore describes the formation of flat cores in dark matter haloes and ultra-diffuse galaxies from feedback-driven outflow episodes. The halo response is divided into an instantaneous change of potential at constant velocities followed by an energy-conserving relaxation. The core assumption of the model is that the total energy E=U+K is conserved for each shell enclosing a given dark matter mass, which is treated in the code as a least-square minimization of the difference between the final and the initial energy of each shell.
Wavetrack recognizes and tracks CME shock waves, filaments, and other solar objects. The code creates base images by averaging а series of images a few minutes prior to the start of the eruption and constructs base difference images by subtracting base images from the current raw image of the sequence. This enhances the change in intensity caused by coronal bright fronts, omits static details, and reduces noise. Wavetrack then chooses an appropriate intensity interval and decomposes the base difference or running difference image with an A-Trous wavelet transform, where each wavelet coefficient is obtained by convolving the image array with a corresponding iteration of the wavelet kernel. When the maximum value of the wavelet coefficients for a connected set of pixels satisfies certain conditions, this region is considered as a structure on the respective wavelet coefficient. Separate stand-alone object masks are obtained with a clustering algorithm and objects are renumbered according to the number of the quadrant they belong at each iteration.
The spectroscopy analysis pipeline pyPipe3D produces coherent and easy to distribute and compare parameters of stellar populations and ionized gas; it is suited in particular for data from the most recent optical IFS surveys. The pipeline is build using pyFIT3D, which is the main spectral fitting module included in this package.
RealSim-IFS generates survey-realistic integral-field spectroscopy (IFS) observations of galaxies from numerical simulations of galaxy formation. The tool is designed primarily to emulate current and experimental observing strategies for IFS galaxy surveys in astronomy, and can reproduce both the flux and variance propagation of real galaxy spectra to cubes. RealSim-IFS has built-in functions supporting SAMI and MaNGA IFU footprints, but supports any fiber-based IFU design, in general.
PyCASSO runs the STARLIGHT code (ascl:1108.006) in integral field spectra (IFS). Cubes from various instruments are supported, including PMAS/PPAK (CALIFA), MaNGA, GMOS and MUSE. Emission lines can be measured using DOBBY, which is included in the package. The package also includes tools for IFS cubes analysis and plotting.
CCDLAB provides graphical user interface functionality for FITS image viewing and data reduction based on the JPFITS FITS-file interface. It can view, manipulate, and save FITS primary image data and image extensions, view and manipulate FITS image headers, and view FITS Bintable extensions. The code enables batch processing, viewing, and saving of FITS images and searching FITS files on disk. CCDLAB also provides general image reduction techniques, source detection and characterization, and can create World Coordinate Solutions automatically or manually for FITS images.
The population synthesis code SEVN (Stellar EVolution for N-body) includes up-to-date stellar evolution (through look-up tables), binary evolution, and different recipes for core-collapse supernovae. SEVN also provides an up-to-date formalism for pair-instability and pulsational pair-instability supernovae, and is designed to interface with direct-summation N-body codes such as STARLAB (ascl:1010.076) and HiGPUs (ascl:1207.002).
MADYS (Manifold Age Determination for Young Stars) determines the age and mass of young stellar and substellar objects. The code automatically retrieves and cross-matches photometry from several catalogs, estimates interstellar extinction, and derives age and mass estimates for individual objects through isochronal fitting. MADYS harmonizes the heterogeneity of publicly-available isochrone grids and the user can choose amongst several models, some of which have customizable astrophysical parameters. Particular attention has been dedicated to the categorization of these models, labeled through a four-level taxonomical classification.
atoMEC simulates high energy density phenomena such as in warm dense matter. It uses Kohn-Sham density functional theory, in combination with an average-atom approximation, to solve the electronic structure problem for single-element materials at finite temperature.
wdwarfdate derives the Bayesian total age of a white dwarf from an effective temperature and a surface gravity. It runs a chain of models assuming single star evolution and estimates the following parameters and their uncertainties: total age of the object, mass and cooling age of the white dwarf, and mass and lifetime of the progenitor star.
Smart provides pre-processing for LP-VIcode (ascl:1501.007). It computes the accelerations and variational equations given a generic user-defined potential function, eliminating the need to calculate manually the accelerations and variational equations.
SpinSpotter calculates stellar rotation periods from high-cadence photometry. The code uses the autocorrelation function (ACF) to identify stellar rotation periods up to one-third the observational baseline of the data. SpinSpotter includes diagnostic tools that describe features in the ACF and allows tuning of the tolerance with which to accept a period detection.
Smooth calculates several mean quantities for all particles in an N-Body simulation output file. The program produces a file for each type of output specified on the command line. This output file is in ASCII format with one smoothed quantity for each particle. The program uses a symmetric SPH (Smoothed Particle Hydrodynamics) smoothing kernel to find the mean quantities.
WDPhotTools generates color-color diagrams and color-magnitude diagrams in various photometric systems, plots cooling profiles from different models, and computes theoretical white dwarf luminosity functions based on the built-in or supplied models of the (1) initial mass function, (2) total stellar evolution lifetime, (3) initial-final mass relation, and (4) white dwarf cooling time. The software has three main parts: the formatters that handle the output models from various works in the format as they are downloaded; the photometric fitter that solves for the WD parameters based on the photometry, with or without distance and reddening; and the generator of the white dwarf luminosity function in bolometric magnitudes or in any of the photometric systems available from the atmosphere model.
IFSCube performs analysis tasks in data cubes of integral field spectroscopy. It contains routines for fitting spectral features in 1D spectra and data cubes and rotation models to velocity fields; it also contains a routine that inspects the fit results. Though originally intended to make user scripts more concise, analysis can also be performed on the fly by using an interactive interpreter such as ipython. By default, IFSCube assumes data are in the Flexible Image Transport System (FITS) standard, but the package can be modified easily to allow use of other data formats.
FITS File interaction written in Visual Studio C# .Net.
JPFITS is not based upon any other implementation and is written from the ground-up, consistent with the FITS standard, designed to interact with FITS files as object-oriented structures.
JPFITS provides functionality to interact with FITS images and binary table extensions, as well as providing common mathematical methods for the manipulation of data, data reductions, profile fitting, photometry, etc.
JPFITS also implements object-oriented classes for Point Source Extraction, World Coordinate Solutions (WCS), WCS automated field solving, FITS Headers and Header Keys, etc.
The automatic world coordinate solver is based on the trigonometric algorithm as described here:
https://iopscience.iop.org/article/10.1088/1538-3873/ab7ee8
All function parameters, methods, properties, etc., are coded with XML descriptions which will function with Visual Studio. Other code editors may or may not read the XML files.
Everything which is reasonable to parallelize in order to benefit from the computation speed increase for multi-threaded systems has been done so. In all such cases function options are given in order to specify the use of parallelism or not. Generally, most image manipulation functions are highly amenable to parallelism. No parallelism is forced, i.e., any code which may execute parallelized is given a user option to do so or not.
Fastrometry is a Python implementation of the fast world coordinate solution solver for the FITS standard astronomical image. When supplied with the approximate field center (+-25%) and the approximate field scale (+-10%) of the telescope and detector system the astronomical image is from, fastrometry provides WCS solutions almost instantaneously. The algorithm is also originally implemented with parallelism enabled in the Windows FITS image processor and viewer CCDLAB (ascl:2206.021).
pyHIIexplorerV2 extracts the integrated spectra of HII regions from integral field spectroscopy (IFS) datacubes. The detection of HII regions performed by pyHIIexplorer is based on two assumptions: 1) HII regions have strong emission lines that are clearly above the continuum emission and the average ionized gas emission across each galaxy, and 2) the typical size of HII regions is about a few hundreds of parsecs, which corresponds to a usual projected size of a few arcsec at the distance of our galaxies. These assumptions will define clumpy structures with a high Ha emission line contrast in comparison to the continuum. pyHIIexplorerV2 is written in Python; it is based on and is a successor to HIIexplorer (ascl:1603.017).
Craterstats3 analyzes and plots crater count data for planetary surface dating. It is a Python implementation of Craterstats2 (ascl:2206.008) and is designed to replicate the output of the previous version as closely as possible. As before, it produces plots in cumulative, differential, Hartmann, and R-plot styles with possible overlays of crater counts, isochrons, equilibrium functions and epoch boundaries, as well aschronology and impact rate functions. Data can be shown with various binnings or unbinned, and age estimates made by either cumulative fitting, differential fitting, or Poisson timing evaluation. Numerical results can be output as text for further processing elsewhere. A number of published chronology systems are already set up for use, but new ones may be added by the user. The software is designed to be easily integrated into other software, which could allow the addition of a graphical interface or the inclusion of some Craterstats functions into a GIS.
Craterstats2 plots crater counts and determining surface ages. The software plots isochrons in cumulative, differential, R-plot and Hartmann presentations, and makes isochron fits to both cumulative and differential data. Hartmann-style piecewise production functions may also be used. A Python implementation of the software, Craterstats3, is also available.
CircleCraters is a projection independent crater counting plugin for QGIS. It has the flexibility to crater count in a GIS environment on Windows, OS X, or Linux, and uses three-click input to define crater rims as a circle.
MYRaf is a practicable astronomical image reduction and photometry software and interface for IRAF (ascl:9911.002). The library uses IRAF, PyRAF (ascl:1207.011), Ginga (ascl:1303.020), and other python packages with a Qt framework for automated software processing of data from robotic telescopes.
NonnegMFPy solves nonnegative matrix factorization (NMF) given a dataset with heteroscedastic uncertainties and missing data with a vectorized multiplicative update rule; this can be used create a mask and iterate the process to exclude certain new data by updating the mask. The code can work on multi-dimensional data, such as images, if the data are first flattened to 1D.
pystortion provides support for distortion measurements in astronomical imagers. It includes classes to support fitting of bivariate polynomials of arbitrary degree and helper functions for crossmatching catalogs. The crossmatching uses an iterative approach in which a two-dimensional distortion model is fit at every iteration and used to continuously refine the position of extracted sources.
ExoJAX provides auto-differentiable line-by-line spectral modeling of exoplanets/brown dwarfs/M dwarfs using JAX (ascl:2111.002). In a nutshell, ExoJAX allows the user to do a HMC-NUTS fitting using the latest molecular/atomic data in ExoMol, HITRAN/HITEMP, and VALD3. The code enables a fully Bayesian inference of the high-dispersion data to fit the line-by-line spectral computation to the observed spectrum, from end-to-end (i.e. from molecular/atomic databases to real spectra), by combining it with the Hamiltonian Monte Carlo in recent probabilistic programming languages such as NumPyro (ascl:2505.005).
A stand-alone spectral gridder and imager for the Green Bank Telescope, as well as functionality for any diameter telescope. Based around the cygrid package from Benjamin Winkel and Daniel Lenz
TCF calculates a periodogram designed to detect exoplanet transits after the light curve has been differenced. It is a matched filter for a periodic double-spike pattern. The difference operator that can be used independently for detrending a light curve; it is also embedded in ARIMA (autoregressive integrated moving average) Box-Jenkins modeling.
vortex performs a Helmholtz-Hodge decomposition on vector fields defined on AMR grids, decomposing a vector field in its solenoidal (divergence-less) and compressive (curl-less) parts. It works natively on vector fields defined on Adaptive Mesh Refinement (AMR) grids, so that it can perform the decomposition over large dynamical ranges; it is also applicable to particle-based simulations. As vortex is devised primarily to investigate the properties of the turbulent velocity field in the Intracluster Medium (ICM), it also includes routines for multi-scale filtering the velocity field.
simulateSearch simulates high time-resolution data in radio astronomy. The code is built around producing multiple binary data files that contain information on the radiometer noise and sources that are being simulated. These binary data files subsequently get combined and output PSRFITS
search mode files produced. The PSRFITS files can be processed using standard pulsar software packages such as PRESTO (ascl:1107.017).
MM-LSD (Multi-Mask Least-Squares Deconvolution) performs continuum normalization of 2D spectra (echelle order spectra). It also masks and partially corrects telluric lines and extracts RVs from spectra. The code requires RASSINE (ascl:2102.022) and uses spectral line data from VALD3.
PyWPF (Waterfall Principal Component Analysis Folding) finds periodicity in one-dimensional timestream data sets; it is particularly designed for very high noise situations where traditional methods may fail. Given a timestream, with each point being the arrival times of a source, the software computes the estimated period. The core function of the package requires several initial parameters to run, and using the best known period of the source (T_init) is recommended.
BANG (BAyesian decomposiotioN of Galaxies) models both the photometry and kinematics of galaxies. The underlying model is the superposition of different components with three possible combinations: 1.) Bulge + inner disc + outer disc + Halo; 2.) Bulge + disc + Halo; and 3.) inner disc + outer disc + Halo. As CPU parameter estimation can take days, running BANG on GPU is recommended.
CPNest performs Bayesian inference using the nested sampling algorithm. It is designed to be simple for the user to provide a model via a set of parameters, their bounds and a log-likelihood function. An optional log-prior function can be given for non-uniform prior distributions. The nested sampling algorithm is then used to compute the marginal likelihood or evidence. As a by-product the algorithm produces samples from the posterior probability distribution. The implementation is based on an ensemble MCMC sampler which can use multiple cores to parallelize computation.
ASTROMER is a Transformer-based model trained on millions of stars for the representation of light curves. Pretrained models can be directly used or finetuned on specific datasets. ASTROMER is useful in downstream tasks in which data are limited to train deep learning models.
HOPS (Haystack Observatory Postprocessing System) analyzes the data generated by DiFX VLBI correlators. It is written in C for Linux computers, and emphasizes quality-control aspects of data processing. It sits between the correlator and an image-processing and/or geodetic-processing package, and performs basic fringe-fitting, data editing, problem diagnosis, and correlator support functions.
ASOHF (Adaptive Spherical Overdensity Halo Finder) identifies bound dark matter structures (dark matter haloes) in the outputs of cosmological simulations, and works directly on an input particle list. The computational cost of running ASOHF in simulations with a large number of particles can be reduced by using a domain decomposition to split the simulation box into smaller boxes, or subdomains, which are then processed independently. The basic output of ASOHF is a halo catalog. The package includes a python code to build a merger tree from ASOHF outputs.
The LIghtweight Source finding Algorithms (LiSA) library finds HI sources in next generation radio surveys. LiSA can analyze input data cubes of any size with pipelines that automatically decompose data into different domains for parallel distributed analysis. For source finding, the library contains python modules for wavelet denoising of 3D spatial and spectral data, and robust automatic source finding using null-hypothesis testing. The source-finding algorithms all have options to automatically choose parameters, minimizing the need for manual fine tuning. Finally, LiSA also contains neural network architectures for classification and characterization of 3D spectral data.
Pryngles produces visualizations of the geometric configuration of a ringed exoplanet (an exoplanet with a ring or exoring for short) and calculates the light-curve signatures produced by these kind of planets. The model behind the package has been developed in an effort to predict the signatures that exorings may produce not only in the light-curve of transiting exoplanets (a problem that has been extensively studied) but also in the light of stars having non-transiting exoplanets.
Would you like to view a random code?