Results 1901-2000 of 2545 (2498 ASCL, 47 submitted)

[ascl:1710.013]
Ramses-GPU: Second order MUSCL-Handcock finite volume fluid solver

RamsesGPU is a reimplementation of RAMSES (ascl:1011.007) which drops the adaptive mesh refinement (AMR) features to optimize 3D uniform grid algorithms for modern graphics processor units (GPU) to provide an efficient software package for astrophysics applications that do not need AMR features but do require a very large number of integration time steps. RamsesGPU provides an very efficient C++/CUDA/MPI software implementation of a second order MUSCL-Handcock finite volume fluid solver for compressible hydrodynamics as a magnetohydrodynamics solver based on the constraint transport technique. Other useful modules includes static gravity, dissipative terms (viscosity, resistivity), and forcing source term for turbulence studies, and special care was taken to enhance parallel input/output performance by using state-of-the-art libraries such as HDF5 and parallel-netcdf.

[ascl:1011.007]
RAMSES: A new N-body and hydrodynamical code

A new N-body and hydrodynamical code, called RAMSES, is presented. It has been designed to study structure formation in the universe with high spatial resolution. The code is based on Adaptive Mesh Refinement (AMR) technique, with a tree based data structure allowing recursive grid refinements on a cell-by-cell basis. The N-body solver is very similar to the one developed for the ART code (Kravtsov et al. 97), with minor differences in the exact implementation. The hydrodynamical solver is based on a second-order Godunov method, a modern shock-capturing scheme known to compute accurately the thermal history of the fluid component. The accuracy of the code is carefully estimated using various test cases, from pure gas dynamical tests to cosmological ones. The specific refinement strategy used in cosmological simulations is described, and potential spurious effects associated to shock waves propagation in the resulting AMR grid are discussed and found to be negligible. Results obtained in a large N-body and hydrodynamical simulation of structure formation in a low density LCDM universe are finally reported, with 256^3 particles and 4.1 10^7 cells in the AMR grid, reaching a formal resolution of 8192^3. A convergence analysis of different quantities, such as dark matter density power spectrum, gas pressure power spectrum and individual haloes temperature profiles, shows that numerical results are converging down to the actual resolution limit of the code, and are well reproduced by recent analytical predictions in the framework of the halo model.

[ascl:2008.021]
ramses2hsim: RAMSES output to 3D data cube for HSIM

Richardson, Mark L. A.; Routledge, L.; Thatte, N.; Tecza, M.; Houghton, Ryan C. W.; Pereira-Santaella, M.; Rigopoulou, D

The ramses2hsim pipeline converts a simulated galaxy in a RAMSES (ascl:1011.007) output into an 3D input data cube for HSIM (ascl:1912.006). The code incorporates gas kinematics (both bulk and turbulence), line emission and line width for Hα, and accounts for dust extinction.

[ascl:2105.019]
RandomQuintessence: Integrate the Klein-Gordon and Friedmann equations with random initial conditions

RandomQuintessence integrates the Klein-Gordon and Friedmann equations for quintessence models with random initial conditions and functional forms for the potential. Quintessence models generically impose non-trivial structure on observables like the equation of state of dark energy. There are three main modules; montecarlo_nompi.py sets initial conditions, loops over a bunch of randomly-initialised models, integrates the equations, and then analyses and saves the resulting solutions for each model. Models are defined in potentials.py; each model corresponds to an object that defines the functional form of the potential, various model parameters, and functions to randomly draw those parameters. All of the equation-solving code and methods to analyze the solution are kept in solve.py under the base class DEModel(). Other files available analyze and plot the data in a variety of ways.

[ascl:2003.007]
RAPID: Real-time Automated Photometric IDentification

RAPID (Real-time Automated Photometric IDentification) classifies multiband photometric light curves into several different transient classes. It uses a deep recurrent neural network to produce time-varying classifications, and because it does not rely on deriving computationally expensive features from the data, it is well suited for processing alerts that wide-field surveys such as the Zwicky Transient Facility (ZTF) and the Large Synoptic Survey Telescope (LSST) will produce.

[ascl:2005.016]
RAPP: Robust Automated Photometry Pipeline

RAPP is a robust automated photometry pipeline for blurred images. RAPP requires that the observed images contain at least three stars and applies bias, dark, and flat field correction on blurred observational raw data; it also uses the median of adjacent pixels to eliminate outliers. It also uses star enhancement and robust image matching, extracts stars, and performs aperture photometry to extract information from blurred images.

[ascl:1803.015]
RAPTOR: Imaging code for relativistic plasmas in strong gravity

Bronzwaer, Thomas; Davelaar, Jordy; Younsi, Ziri; Mościbrodzka, Monika; Falcke, Heino; Kramer, Michael; Rezzolla, Luciano

RAPTOR produces accurate images, animations, and spectra of relativistic plasmas in strong gravity by numerically integrating the equations of motion of light rays and performing time-dependent radiative transfer calculations along the rays. The code is compatible with any analytical or numerical spacetime, is hardware-agnostic and may be compiled and run on both GPUs and CPUs. RAPTOR is useful for studying accretion models of supermassive black holes, performing time-dependent radiative transfer through general relativistic magneto-hydrodynamical (GRMHD) simulations and investigating the expected observational differences between the so-called fastlight and slow-light paradigms.

[ascl:1909.008]
RascalC: Fast code for galaxy covariance matrix estimation

RascalC quickly estimates covariance matrices from two- or three-point galaxy correlation functions. Given an input set of random particle locations and a two-point correlation function (or input set of galaxy positions), RascalC produces an estimate of the associated covariance for a given binning strategy, with non-Gaussianities approximated by a ‘shot-noise-rescaling’ parameter. For the 2PCF, the rescaling parameter can be calibrated by dividing the particles into jackknife regions and comparing sample to theoretical jackknife covariance. RascalC can also be used to compute Legendre-binned covariances and cross-covariances between different two-point correlation functions.

[ascl:2002.002]
RASCAS: Resonant line transfer in AMR simulations

Michel-Dansac, Leo; Blaizot, Jeremy; Garel, Thibault; Verhamme, Anne; Kimm, Taysun; Trebitsch, Maxime

The massively parallel code RASCAS (RAdiative SCattering in Astrophysical Simulations) performs radiative transfer on an adaptive mesh with an octree structure using the Monte Carlo technique. The code features full MPI parallelization, domain decomposition, adaptive load-balancing, and a standard peeling algorithm to construct mock observations. The radiative transport of resonant line photons through different mixes of species (e.g. HI, SiII, MgII, FeII), including their interaction with dust, is implemented in a modular fashion to allow new transitions to be easily added to the code. RASCAS may also be used to propagate photons at any wavelength (e.g. stellar continuum or fluorescent lines), and has been designed to be easily customizable and to process simulations of arbitrarily large sizes on large supercomputers.

[ascl:2102.022]
RASSINE: Normalizing 1D stellar spectra

RASSINE normalizes merged 1D spectra using the concept of convex hulls. The code uses six parameters that can be fine-tuned, and provides an interactive interface, including graphical feedback, for easily choosing the parameters. RASSINE can also provide a first guess for the parameters that are derived directly from the merged 1D spectrum based on previously performed calibrations.

[ascl:1904.014]
rate: Reliable Analytic Thermochemical Equilibrium

rate computes thermochemical-equilibrium abundances for a H-C-N-O system with known pressure, temperature, and elemental abundances. The output abundances are H2O, CH4, CO, CO2, NH3, C2H2, C2H4, HCN, and N2, H2, H, and He.

[ascl:0008.002]
RATRAN: Radiative Transfer and Molecular Excitation in One and Two Dimensions

RATRAN is a numerical method and computer code to calculate the radiative transfer and excitation of molecular lines. The approach is based on the Monte Carlo method, and incorporates elements from Accelerated Lambda Iteration. It combines the flexibility of the former with the speed and accuracy of the latter. Convergence problems known to plague Monte Carlo methods at large optical depth (>100) are avoided by separating local contributions to the radiation field from the overall transfer problem. The random nature of the Monte Carlo method serves to verify the independence of the solution to the angular, spatial, and frequency sampling of the radiation field. This allows the method to be used in a wide variety of astrophysical problems without specific adaptations. Moreover, the code can be applied to all atoms or molecules for which collisional rate coefficients are available and any axially symmetric source model. Continuum emission and absorption by dust is explicitly taken into account but scattering is neglected. We expect this program to be an important tool in analyzing data from present and future infrared and (sub-)millimeter telescopes.

[ascl:1105.009]
Ray Tracing Codes: run_tau, run_raypath, and ray_kernel

Time-distance helioseismology aims to measure and interpret the travel times of waves propagating between two points located on the solar surface. The travel times are then inverted to infer sub-surface properties that are encoded in the measurements. The trajectory of the waves generally follows that of the infinite-frequency ray path, although they are sensitive to perturbations off of this path. Finite-frequency sensitivity kernels are thus needed to give more accurate inversion results.

Ray tracing codes calculate travel time kernels for a ray. There are three main codes which calculate the group time as a function of distance, the ray paths as well as the phase and group times along the path, and the ray kernels for the sound speed squared.

[ascl:1411.006]
RC3 mosaicking pipeline: Creating mosaics for the RC3 Catalogue

The RC3 mosaicking pipeline creates color composite images and scientifically-calibrated FITS mosaics in all SDSS imaging bands for all the RC3 galaxies that lie within the survey’s footprint and on photographic plates taken by the Digitized Palomar Observatory Sky Survey (DPOSS) for the B, R, IR bands. The pipeline uses SExtractor (ascl:1010.064) for extraction and STIFF (ascl:1110.006) to generating color images. The mosaicking program uses a recursive algorithm for positional update first to correct the positional inaccuracy inherent in the RC3 catalog, then conducts the mosaicking procedure using the Astropy (ascl:1304.002) wrapper to IPAC's Montage (ascl:1010.036) software. The program is generalized into a pipeline that can be easily extended to future survey data or other source catalogs; an online interface is available at

http://lcdm.astro.illinois.edu/data/rc3/search.html.

[ascl:2009.015]
rcosmo: Cosmic Microwave Background data analysis

rcosmo provides information processing, visualization, manipulation and spatial statistical analysis of Cosmic Microwave Background (CMB) radiation and other spherical data stored in or converted to HEALPix coordinates. The package has more than 100 different functions, and can perform spherical geometry, manipulate CMB and other spherical data, and visualize HEALPix data. rcosmo can also perform statistical analysis of CMB and spherical data, and transforme spherical data in cartesian and geographic coordinates into HEALPix format.

[ascl:1408.017]
RDGEN: Routines for data handling, display, and adjusting

RDGEN is a collection of routines for data handling, display, and adjusting, with a facility which helps to set up files for using with VPFIT (ascl:1408.015); it is included in the VPFIT distribution file. It is useful for setting region boundaries and initial guesses for VPFIT, for displaying the accumulated results, for examining by eye particular redshift systems and fits to them, testing that the error array is a true reflection of the rms scatter in the data, comparing spectra and generally examining and even modifying the data.

[ascl:2007.016]
ReadPDS: Visualization tools for PDS4 data

ReadPDS reads in and visualizes data from the Planetary Data System in PDS4 format. Tools are available in Python as PDS4Viewer and in IDL as PDS4-IDL. These tools support PDS4 data, including images, spectra, and arrays and provide multiple views of the data, including summary, image, plot, and table views in addition to easy access to metadata such as structure labels and spectral characteristics.

[ascl:1506.007]
REALMAF: Magnetic power spectra from Faraday rotation maps

REALMAF is a maximum-a-posteriori code to measure magnetic power spectra from Faraday rotation data. It uses a sophisticated model for the magnetic autocorrelation in real space, thus alleviating the need for simplifying assumptions in the processing. REALMAF treats the divergence relation of the magnetic field with a multiplicative factor in Fourier space, which allows modeling the magnetic autocorrelation as a spherically symmetric function.

[ascl:1107.009]
REAS3: Modeling Radio Emission from Cosmic Ray Air Showers

The freely available Monte Carlo code REAS for modelling radio emission from cosmic ray air showers has evolved to include the full complexity of air shower physics. REAS3 improves the calculation of the emission contributions, which was not fully consistent in earlier versions of REAS, by incorporating the missing radio emission due to the variation of the number of charged particles during the air shower evolution using an "end-point formalism". With the inclusion of these emission contributions, the structure of the simulated radio pulses changes from unipolar to bipolar, and the azimuthal emission pattern becomes nearly symmetric. Remaining asymmetries can be explained by radio emission due to the variation of the net charge excess in air showers, which is automatically taken into account in the new implementation. REAS3 constitutes the first self-consistent time-domain implementation based on single particle emission taking the full complexity of air shower physics into account, and is freely available for all interested users. REAS3 has been superseded by CoREAS (ascl:1406.003).

[ascl:1110.016]
REBOUND: Multi-purpose N-body code for collisional dynamics

REBOUND is a multi-purpose N-body code which is freely available under an open-source license. It was designed for collisional dynamics such as planetary rings but can also solve the classical N-body problem. It is highly modular and can be customized easily to work on a wide variety of different problems in astrophysics and beyond.

REBOUND comes with symplectic integrators WHFast, WHFastHelio, SEI, and LEAPFROG. It supports open, periodic and shearing-sheet boundary conditions. REBOUND can use a Barnes-Hut tree to calculate both self-gravity and collisions. These modules are fully parallelized with MPI as well as OpenMP. The former makes use of a static domain decomposition and a distributed essential tree. Two new collision detection modules based on a plane-sweep algorithm are also implemented. The performance of the plane-sweep algorithm is superior to a tree code for simulations in which one dimension is much longer than the other two and in simulations which are quasi-two dimensional with less than one million particles.

[ascl:2011.020]
REBOUNDx: Adding effects in REBOUND N-body integrations

REBOUNDx incorporates additional physics into REBOUND (ascl:1110.016) simulations. Users can add effects from a list of pre-implemented astrophysical forces or contribute new ones. The main code is written in C, and a Python wrapper is provided for interfacing with other libraries. The REBOUNDx source code is machine independent and a binary format to save REBOUNDx configurations interfaces with the SimulationArchive class in REBOUND, making it possible to share and reproduce results bit by bit.

[ascl:1106.026]
RECFAST: Calculate the Recombination History of the Universe

RECFAST calculates the recombination of H, HeI, and HeII in the early Universe; this involves a line-by-line treatment of each atomic level. It differs in comparison to previous calculations in two major ways: firstly, the ionization fraction x_e is approximately 10% smaller for redshifts <~800, due to non-equilibrium processes in the excited states of H, and secondly, HeI recombination is much slower than previously thought, and is delayed until just before H recombines. RECFAST enables fast computation of the ionization history (and quantities such as the power spectrum of CMB anisotropies which depend on it) for arbitrary cosmologies.

[ascl:2005.004]
REDFIT: Red-noise spectra directly from unevenly spaced time series

Time series are commonly unevenly spaced in time make it difficult to obtain an accurate estimate of their typical red-noise spectrum. REDFIT overcomes this problem by fitting a first-order autoregressive (AR1) process directly to unevenly spaced time series. Hence, interpolation in the time domain and its inevitable bias can be avoided. The program can be used to test if peaks in the spectrum of a time series are significant against the red-noise background from an AR1 process.

[ascl:2103.004]
redshifts: Spectroscopic redshifts search tool

redshifts collects all unique spectroscopic redshifts from online databases such as VizieR and NED. It can perform a flexible search within a radius of a given set of (RA, DEC) coordinates and uses column names and descriptions (including UCD keywords) to identify columns containing spectroscopic redshifts or velocities. It weeds out photometric redshifts and duplicates and returns a unique list of best spectroscopic redshift measurements. redshifts can be used standalone from the terminal, and can take a number of optional command line arguments, or from Python.

[ascl:1507.017]
REDSPEC: NIRSPEC data reduction

REDSPEC is an IDL based reduction package designed with NIRSPEC in mind though can be used to reduce data from other spectrographs as well. REDSPEC accomplishes spatial rectification by summing an *A+B* pair of a calibration star to produce an image with two spectra; the image is remapped on the basis of polynomial fits to the spectral traces and calculation of gaussian centroids to define their separation, producing straight spectral traces with respect to the detector rows. The raw images are remapped onto a coordinate system with uniform intervals in spatial extent along the slit and in wavelength along the dispersion axis.

[ascl:1508.003]
REDUCEME: Long-slit spectroscopic data reduction and analysis

Cardiel, N; Gorgas, J.; Pedraz, S.; Cenarro, J.; Alonso, O; Gil de Paz, A.; García-Dabó, E.; Sánchez-Blázquez, P.; Mármol-Queraltó, E.; Toloba, E.

The astronomical data reduction package REDUCEME reduces and analyzes long-slit spectroscopic data. The package uses the unformatted FORTRAN raw data format, so requires FITS files be transformed to REDUCEME format; the reverse operation (from REDUCEME to FITS format) is also available. The package is a set of programs written in FORTRAN 77 and includes shell scripts (using the C shell syntax) to perform routine tasks; it can be extended by the inclusion of external programs. REDUCEME uses PGPLOT (ascl:1103.002) for line plots and images, and a subset of subroutines, called BUTTON, enables the user to communicate interactively with the image display employing graphic buttons. One advantage of using REDUCEME is that for each image an associated error image can also be processed throughout the reduction process, allowing for a careful control of the error propagation.

[ascl:1401.004]
Reflex: Graphical workflow engine for data reduction

Reflex provides an easy and flexible way to reduce VLT/VLTI science data using the ESO pipelines. It allows graphically specifying the sequence in which the data reduction steps are executed, including conditional stops, loops and conditional branches. It eases inspection of the intermediate and final data products and allows repetition of selected processing steps to optimize the data reduction. The data organization necessary to reduce the data is built into the system and is fully automatic; advanced users can plug their own modules and steps into the data reduction sequence. Reflex supports the development of data reduction workflows based on the ESO Common Pipeline Library. Reflex is based on the concept of a scientific workflow, whereby the data reduction cascade is rendered graphically and data seamlessly flow from one processing step to the next. It is distributed with a number of complete test datasets so users can immediately start experimenting and familiarize themselves with the system.

[ascl:1206.001]
RegiStax: Alignment, stacking and processing of images

Berrevoets, Cor; DeClerq, Bart; George, Tony; Makolkin, Dmitry; Maxson, Paul; Pilz, Bob; Presnyakov, Pavel; Roel, Eric; Weiller, Sylvain

RegiStax is software for alignment/stacking/processing of images; it was released over 10 years ago and continues to be developed and improved. The current version is RegiStax 6, which supports the following formats: AVI, SER, RFL (RegiStax Framelist), BMP, JPG, TIF, and FIT. This version has a shorter and simpler processing sequence than its predecessor, and optimizing isn't necessary anymore as a new image alignment method optimizes directly. The interface of RegiStax 6 has been simplified to look more uniform in appearance and functionality, and RegiStax 6 now uses Multi-core processing, allowing the user to have up to have multiple cores(recommended to use maximally 4) working simultaneous during alignment/stacking.

[ascl:1404.012]
RegPT: Regularized cosmological power spectrum

RegPT computes the power spectrum in flat wCDM class models based on the RegPT treatment when provided with either of transfer function or matter power spectrum. It then gives the multiple-redshift outputs for power spectrum, and optionally provides correlation function data. The Fortran code has two major options for power spectrum calculations; -fast, which quickly computes the power spectrum at two-loop level (typically a few seconds) using the pre-computed data set of PT kernels for fiducial cosmological models, and -direct, in which the code first applies the fast method, and then follows the regularized expression for power spectrum to directly evaluate the multi-dimensional integrals. The output results are the power spectrum of direct calculation and difference of the results between fast and direct method. The code also gives the data set of PT diagrams necessary for power spectrum calculations from which the power spectrum can be constructed.

[ascl:1505.021]
relline: Relativistic line profiles calculation

relline calculates relativistic line profiles; it is compatible with the common X-ray data analysis software XSPEC (ascl:9910.005) and ISIS (ascl:1302.002). The two basic forms are an additive line model (RELLINE) and a convolution model to calculate relativistic smearing (RELCONV).

[ascl:2010.015]
relxill: Reflection models of black hole accretion disks

Dauser, T.; García, J.; Parker, M. L.; Fabian, A. C.; Wilms, J.; Lohfink, A.; Kallman, T. R.; Steiner, J. F.; McClintock, J. E.; Brenneman, L.; Eikmann, W.; Reynolds, C. S.; Tombesi, F.

relxill self-consistently connects an angle-dependent reflection model constructed with XILLVER (http://www.srl.caltech.edu/personnel/javier/xillver/index.html) with the relativistic blurring code RELLINE (ascl:1505.021). It calculates the proper emission angle of the radiation at each point on the accretion disk and then takes the corresponding reflection spectrum into account.

[ascl:1904.008]
repack: Repack and compress line-transition data

repack re-packs and compresses line-transition data for radiative-transfer calculations. It identifies the strong lines that dominate the spectrum from the large-majority of weaker lines, returning a binary line-by-line (LBL) file with the strong lines info (wavenumber, Elow, gf, and isotope ID), and an ASCII file with the combined contribution of the weaker lines compressed into a continuum extinction coefficient (in cm-1 amagat-1) as function of wavenumber and temperature.

[ascl:2011.023]
reproject: Python-based astronomical image reprojection

reproject implements image reprojection (resampling) methods for astronomical images using various techniques via a uniform interface. Reprojection re-grids images from one world coordinate system to another (for example changing the pixel resolution, orientation, coordinate system). reproject works on celestial images by interpolation, as well as by finding the exact overlap between pixels on the celestial sphere. It can also reproject to/from HEALPIX projections by relying on the astropy-healpix package.

[ascl:1612.022]
REPS: REscaled Power Spectra for initial conditions with massive neutrinos

Zennaro, Matteo; Bel, Julien; Villaescusa-Navarro, Francisco; Carbone, Carmelita; Sefusatti, Emiliano; Guzzo, Luigi

REPS (REscaled Power Spectra) provides accurate, one-percent level, numerical simulations of the initial conditions for massive neutrino cosmologies, rescaling the late-time linear power spectra to the simulation initial redshift.

[ascl:1809.016]
RequiSim: Variance weighted overlap calculator

RequiSim computes the Variance Weighted Overlap, which is a measure of the bias on the lensing signal from power spectrum modelling bias for any non-linear model. It assumes that the bias on the power spectrum is Gaussian with a covariance described by a user-provided knowledge matrix that describes the covariance in the bias on the power spectrum. The data from the Euclid wide-field survey are included.

[ascl:1505.028]
RESOLVE: Bayesian algorithm for aperture synthesis imaging in radio astronomy

RESOLVE is a Bayesian inference algorithm for image reconstruction in radio interferometry. It is optimized for extended and diffuse sources. Features include parameter-free Bayesian reconstruction of radio continuum data with a focus on extended and weak diffuse sources, reconstruction with uncertainty propagation dependent on measurement noise, and estimation of the spatial correlation structure of the radio astronomical source. RESOLVE provides full support for measurement sets and includes a simulation tool (if uv-coverage is provided).

[ascl:1907.023]
REVOLVER: REal-space VOid Locations from suVEy Reconstruction

REVOLVER reconstructs real space positions from redshift-space tracer data by subtracting RSD through FFT-based reconstruction (optional) and applies void-finding algorithms to create a catalogue of voids in these tracers. The tracers are normally galaxies from a redshift survey but could also be halos or dark matter particles from a simulation box. Two void-finding routines are provided. The first is based on ZOBOV (ascl:1304.005) and uses Voronoi tessellation of the tracer field to estimate the local density, followed by a watershed void-finding step. The second is a voxel-based method, which uses a particle-mesh interpolation to estimate the tracer density, and then uses a similar watershed algorithm. Input data files can be in FITS format, or ASCII- or NPY-formatted data arrays.

[ascl:2005.018]
RFCDE: Random Forests for Conditional Density Estimation

RFCDE provides an implementation of random forests designed for conditional density estimation. It computes a kernel density estimate of y with nearest neighbor weightings defined by the location of the evaluation point x relative to the leaves in the random forest.

[ascl:1710.002]
rfpipe: Radio interferometric transient search pipeline

rfpipe supports Python-based analysis of radio interferometric data (especially from the Very Large Array) and searches for fast radio transients. This extends on the rtpipe library (ascl:1706.002) with new approaches to parallelization, acceleration, and more portable data products. rfpipe can run in standalone mode or be in a cluster environment.

[ascl:1711.006]
RGW: Goodman-Weare Affine-Invariant Sampling

RGW is a lightweight R-language implementation of the affine-invariant Markov Chain Monte Carlo sampling method of Goodman & Weare (2010). The implementation is based on the description of the python package emcee (ascl:1303.002).

[ascl:1502.001]
RH 1.5D: Polarized multi-level radiative transfer with partial frequency distribution

RH 1.5D performs Zeeman multi-level non-local thermodynamical equilibrium calculations with partial frequency redistribution for an arbitrary amount of chemical species. Derived from the RH code and written in C, it calculates spectra from 3D, 2D or 1D atmospheric models on a column-by-column basis (or 1.5D). It includes optimization features to speed up or improve convergence, which are particularly useful in dynamic models of chromospheres. While one should be aware of its limitations, the calculation of spectra using the 1.5D or column-by-column is a good approximation in many cases, and generally allows for faster convergence and more flexible methods of improving convergence. RH 1.5D scales well to at least tens of thousands of CPU cores.

[ascl:1611.009]
RHOCUBE: 3D density distributions modeling code

RHOCUBE models 3D density distributions on a discrete Cartesian grid and their integrated 2D maps. It can be used for a range of applications, including modeling the electron number density in LBV shells and computing the emission measure. The RHOCUBE Python package provides several 3D density distributions, including a powerlaw shell, truncated Gaussian shell, constant-density torus, dual cones, and spiralling helical tubes, and can accept additional distributions. RHOCUBE provides convenient methods for shifts and rotations in 3D, and if necessary, an arbitrary number of density distributions can be combined into the same model cube and the integration ∫ d*z* performed through the joint density field.

[ascl:2003.005]
RHT: Rolling Hough Transform

The RHT (Rolling Hough Transform) measures linear intensity as a function of orientation in images. This machine vision algorithm works on any image-space (2D) data, and quantifies the presence of linear structure as a function of orientation. The RHT can be used to identify linear features in images, to quantify the orientation of structure in images, and to map image intensity from 2D x-y space to 3D x-y-orientation space. An option in the code allows the user to quantify intensity as a function of direction (modulo 2pi) rather than orientation (modulo pi). The RHT was first used to discover that filamentary structures in neutral hydrogen emission are aligned with the ambient magnetic field.

[ascl:1410.005]
RICH: Numerical simulation of compressible hydrodynamics on a moving Voronoi mesh

RICH (Racah Institute Computational Hydrodynamics) is a 2D hydrodynamic code based on Godunov's method. The code, largely based on AREPO (ascl:1909.010), acts on an unstructured moving mesh. It differs from AREPO in the interpolation and time advancement scheme as well as a novel parallelization scheme based on Voronoi tessellation. Though not universally true, in many cases a moving mesh gives better results than a static mesh: where matter moves one way and a sound wave is traveling in the other way (such that relative to the grid the wave is not moving), a static mesh gives better results than a moving mesh. RICH is designed in an object oriented, user friendly way that facilitates incorporation of new algorithms and physical processes.

[ascl:2005.001]
RID: Relativistic Image Doubling in water Cherenkov detectors

RID (Relativistic Image Doubling in water Cherenkov detectors) uses Monte Carlo simulations to find the relative fraction of charged, relativistic particles entering a HAWC-like Water Cherenkov Detector that can cause a Relativistic Image Doubling (RID) effect visible to at least one of the internal detectors. The technique is available in C++ and Fortran; RID also includes python code for the horizontal incidence of the muon inside the tank.

[ascl:2104.006]
RJObject: Reversible Jump Objects

RJObject provides a general approach to trans-dimensional Bayesian inference problems, using trans-dimensional MCMC embedded within a Nested Sampling algorithm. This allows exploration of the posterior distribution and calculattion of the marginal likelihood (summed over N) even if the problem contains a phase transition or other difficult features such as multimodality.

[ascl:1811.009]
RLOS: Time-resolved imaging of model astrophysical jets

RLOS (Relativistic Line Of Sight) uses hydrocode output data, such as that from PLUTO (ascl:1010.045), to create synthetic images depicting what a model relativistic astrophysical jet looks like to a stationary observer. The approximate time-delayed imaging algorithm used is implemented within existing line-of-sight code. The software has the potential to study a variety of dynamical astrophysical phenomena in collaboration with other imaging and simulation tools.

[ascl:1708.011]
RM-CLEAN: RM spectra cleaner

RM-CLEAN reads in dirty Q and U cubes, generates rmtf based on the frequencies given in an ASCII file, and cleans the RM spectra following the algorithm given by Brentjens (2007). The output cubes contain the clean model components and the CLEANed RM spectra. The input cubes must be reordered with mode=312, and the output cubes will have the same ordering and thus must be reordered after being written to disk. RM-CLEAN runs as a MIRIAD (ascl:1106.007) task and a Python wrapper is included with the code.

[ascl:2005.003]
RM-Tools: Rotation measure (RM) synthesis and Stokes QU-fitting

RM-Tools analyzes radio polarization data, specifically the use of Faraday rotation measure synthesis and Stokes QU model fitting. It contains routines for both single-pixel 1D polarized spectra as well as 3D polarization cubes. RM-Tools is intended to serve as a toolkit for studies of polarized radio sources and measurements of their Faraday rotation. RM-Tools is the core package for the pipelines used for the POlarized Sky Survey of the Universe's Magnetism (POSSUM) and the polarization component of the Very Large Array Sky Survey (VLASS). The package is maintained by the Canadian Initiative for Radio Astronomy Data Analysis (CIRADA; cirada.org).

[ascl:1806.024]
RMextract: Ionospheric Faraday Rotation calculator

RMextract calculates Ionospheric Faraday Rotation for a given epoch, location and line of sight. This Python code extracts TEC, vTEC, Earthmagnetic field and Rotation Measures from GPS and WMM data for radio interferometry observations.

[ascl:1409.011]
rmfit: Forward-folding spectral analysis software

Rmfit uses a forward-folding technique to obtain the best-fit parameters for a chosen model given user-selected source and background time intervals from data files containing observed count rates and a corresponding detector response matrix. rmfit displays lightcurves and spectra using a graphical interface that enables user-defined integrated or time-resolved spectral fits and binning in either time or energy. Originally developed for the analysis of BATSE Gamma-Ray Burst (GRB) spectroscopy, rmfit is a tool for the spectroscopy of transient sources; it accommodates the Fermi GBM and LAT data and Swift BAT.

[ascl:1403.011]
RMHB: Hierarchical Reverberation Mapping

RMHB is a hierarchical Bayesian code for reverberation mapping (RM) that combines results of a sparsely sampled broad line region (BLR) light curve and a large sample of active galactic nuclei (AGN) to infer properties of the sample of the AGN. The key idea of RM is to measure the time lag τ between variations in the continuum emission from the accretion disc and subsequent response of the broad line region (BLR). The measurement of τ is typically used to estimate the physical size of the BLR and is combined with other measurements to estimate the black hole mass MBH. A major difficulty with RM campaigns is the large amount of data needed to measure τ. RMHB allows a clear interpretation of a posterior distribution for hyperparameters describing the sample of AGN.

[ascl:1104.008]
Rmodel: Determining Stellar Population Parameters

This program determines stellar population parameters (e.g. age, metallicity, IMF slope,...), using as input a pair of line-strength indices, through the interpolation in SSP model predictions. Both linear and bivariate fits are computed to perform the interpolation.

[ascl:1603.008]
ROBAST: ROOT-based ray-tracing library for cosmic-ray telescopes

ROBAST (ROOT-based simulator for ray tracing) is a non-sequential ray-tracing simulation library developed for wide use in optical simulations of gamma-ray and cosmic-ray telescopes. The library is written in C++ and fully utilizes the geometry library of the ROOT analysis framework, and can build the complex optics geometries typically used in cosmic ray experiments and ground-based gamma-ray telescopes.

[ascl:1808.011]
Robbie: Radio transients and variables detection workflow

Robbie automates cataloging sources, finding variables, and identifying transients in the image domain. It works in a batch processing paradigm with a modular design so components can be swapped out or upgraded to adapt to different input data while retaining a consistent and coherent methodological approach. Robbie is based on commonly used and open software, including AegeanTools (ascl:1212.009) and STILS/TOPCAT (ascl:1101.010).

[ascl:1502.023]
ROBOSPECT: Width fitting program

ROBOSPECT, written in C, automatically measures and deblends line equivalent widths for absorption and emission spectra. ROBOSPECT should not be used for stars with spectra in which there is no discernible continuum over large wavelength regions, nor for the most carbon-enhanced stars for which spectral synthesis would be favored. Although ROBOSPECT was designed for metal-poor stars, it is capable of fitting absorption and emission features in a variety of astronomical sources.

[ascl:2012.006]
Robovetter: Automatic vetting of Threshold Crossing Events (TCEs)

The DR25 Kepler Robovetter is a robotic decision-making code that dispositions each Threshold Crossing Event (TCE) from the final processing (DR 25) of the Kepler data into Planet Candidates (PCs) and False Positives (FPs). The Robovetter provides four major flags to designate each FP TCE as Not Transit-Like (NTL), a Stellar Eclipse (SS), a Centroid Offset (CO), and/or an Ephemeris Match (EM). It produces a score ranging from 0.0 to 1.0 that indicates the Robovetter's disposition confidence, where 1.0 indicates strong confidence in PC, and 0.0 indicates strong confidence in FP. Finally, the Robovetter provides comments in a text string that indicate the specific tests each FP TCE fails and provides supplemental information on all TCEs as necessary.

[ascl:1201.002]
Roche: Visualization and analysis tool for Roche-lobe geometry of evolving binaries

Roche is a visualization and analysis tool for drawing the Roche-lobe geometry of evolving binaries. Roche can be used as a standalone program reading data from the command line or from a file generated by SeBa (ascl:1201.003). Eventually Roche will be able to read data from any other binary evolution program. Roche requires Starlab (ascl:1010.076) version 4.1.1 or later and the pgplot (ascl:1103.002) libraries. Roche creates a series of images, based on the SeBa output file SeBa.data, displaying the evolutionary state of a binary.

[ascl:1210.008]
Rockstar: Phase-space halo finder

Rockstar (Robust Overdensity Calculation using K-Space Topologically Adaptive Refinement) identifies dark matter halos, substructure, and tidal features. The approach is based on adaptive hierarchical refinement of friends-of-friends groups in six phase-space dimensions and one time dimension, which allows for robust (grid-independent, shape-independent, and noise-resilient) tracking of substructure. Our method is massively parallel (up to 10^5 CPUs) and runs on the largest current simulations (>10^10 particles) with high efficiency (10 CPU hours and 60 gigabytes of memory required per billion particles analyzed). Rockstar offers significant improvement in substructure recovery as compared to several other halo finders.

[ascl:1712.009]
RODRIGUES: RATT Online Deconvolved Radio Image Generation Using Esoteric Software

RODRIGUES (RATT Online Deconvolved Radio Image Generation Using Esoteric Software) is a web-based radio telescope simulation and reduction tool. From a technical perspective it is a web based parameterized docker container scheduler with a result set viewer.

[ascl:2010.011]
ROGER: Automatic classification of galaxies using phase-space information

de los Rios, Martín; Julián Martínez, Héctor; Coenda, Valeria; Muriel, Hernán; Nicolás Ruiz, Andrés; Vega-Martínez, Cristian Antonio; Cora, Sofia Alejandra

ROGER (Reconstructing Orbits of Galaxies in Extreme Regions) predicts the dynamical properties of galaxies using the projected phase-space information. Written in R, it offers a choice of machine learning methods to classify the dynamical properties of galaxies. An interface for online use of the software is available at https://mdelosrios.shinyapps.io/roger_shiny/.

[ascl:1907.028]
ROHSA: Separation of diffuse sources in hyper-spectral data

ROHSA (Regularized Optimization for Hyper-Spectral Analysis) reveals the statistical properties of interstellar gas through atomic and molecular lines. It uses a Gaussian decomposition algorithm based on a multi-resolution process from coarse to fine grid to decompose any kind of hyper-spectral observations into a sum of coherent Gaussian. Optimization is performed on the whole data cube at once to obtain a solution with spatially smooth parameters.

[ascl:2005.005]
RoLo: Calculate radius and potential of the Roche Lobe

RoLo (Roche Lobe) calculates the radius and potential of the Roche Lobe for any specified direction, and also gives some other commonly used quantities (such as the Lagrange points). The calculator is valid for any mass ratio q between 0.01 and 100. The coordinates are spherical-polar (R, theta, phi) centered on one star (M1), with the x-axis (theta=pi/2, phi=0) pointing towards the other star (M2). The mass ratio is defined as q=M2/M1. Distances are given in units of the binary separation, a. A circular orbit is assumed.

[ascl:1902.006]
RPFITS: Routines for reading and writing RPFITS files

Norris, Ray; Calabretta, Mark; Wieringa, Mark; Hunt, A.J.; May, H.; Killeen, N.; Sault, R.; Wilson, Warwick; Kesteven, Michael

The RPFITS data file format records synthesis visibility data obtained from the Australia Telescope Compact Array (ATCA) at Narrabri, NSW. It is also used for single-dish spectral line data obtained from Parkes and Mopra, including Parkes multibeam data. RPFITS superficially resembles random group FITS, but differs in important respects, making it incompatible with standard FITS software such as FITSIO (ascl:1010.001) and FTOOLS (ascl:9912.002) and, in particular, it precludes the use of fv (ascl:1205.005). The RPFITS Fortran library contains routines for reading and writing RPFITS files. A header file, RPFITS.h, is provided to facilitate usage by C and C++ applications. Also included is rpfhdr, a utility for viewing RPFITS headers (it also works for standard FITS), and rpfex for extracting selected scans from an RPFITS file.

[ascl:1905.015]
rPICARD: Radboud PIpeline for the Calibration of high Angular Resolution Data

rPICARD (Radboud PIpeline for the Calibration of high Angular Resolution Data) reduces data from different VLBI arrays, including high-frequency and low-sensitivity arrays, and supports continuum, polarization, and phase-referencing observations. Built on the CASA (ascl:1107.013) framework, it uses CASA for CLEAN imaging and self-calibration, and can be run non-interactively after only a few non-default input parameters are set. rPICARD delivers high-quality calibrated data and large bandwidth data can be processed within reasonable computing times.

[ascl:2001.013]
RPPPS: Re-analyzing Pipeline for Parkes Pulsar Survey

RPPPS (Re-analysing Pipeline for Parkes Pulsar Survey) uses Linux shell scripts, C language, and python code and two parallel strategies to reorganize the PRESTO (ascl:1107.017) pulsar search pipeline to run multiple processes in parallel, thus accelerating the search for pulsars. Though originally designed for reprocessing PMPS data, the code has also been successfully used with FAST (Five-hundred-meter Aperture Spherical radio Telescope) drift scan data. The pipeline is only CPU-based and can be easily and quickly deployed in computing nodes for testing purposes or data processes.

[ascl:2011.017]
RRATtrap: Rotating Radio Transient identifier

Karako-Argaman, C.; Kaspi, V. M.; Lynch, R. S.; Hessels, J. W. T.; Kondratiev, V. I.; McLaughlin, M. A.; Ransom, S. M.; Archibald, A. M.; Boyles, J.; Jenet, F. A.; Kaplan, D. L.; Levin, L.; Lorimer, D. R.; Madsen, E. C.; Roberts, M. S. E.; Siemens, X.; Stairs, I. H.; Stovall, K.; Swiggum, J. K.; van Leeuwen, J.

RRATtrap is a single-pulse sifting algorithm to identify Rotating Radio Transients (RRATs) and transients using output from the PRESTO (ascl:1107.017) routine single_pulse_search.py. It can be integrated into pulsar survey data analysis pipelines and, in addition to finding RRATs, it can also identify Fast Radio Bursts.

[ascl:1808.002]
rsigma: Resonant disturbance

rsigma calculates the resonant disturbing function, R(sigma), for a massless particle in an arbitrary orbit perturbed by a planet in circular orbit. This function defines the strength of the resonance (its semi-amplitude) and the location of the stable equilibrium points (the minima). It depends on the variable sigma called critical angle and on the particle's orbital elements a, e, i and the argument of the perihelion. R(sigma) is numerically calculated and the code is valid for arbitrary eccentricities and inclinations, including retrograde orbits.

[ascl:1607.015]
RT1D: 1D code for Rayleigh-Taylor instability

The parallel one-dimensional moving-mesh hydrodynamics code RT1D reproduces the multidimensional dynamics from Rayleigh-Taylor instability in supernova remnants.

[ascl:1706.002]
rtpipe: Searching for Fast Radio Transients in Interferometric Data

rtpipe (real-time pipeline) analyzes radio interferometric data with an emphasis on searching for transient or variable astrophysical sources. The package combines single-dish concepts such as dedispersion and filters with interferometric concepts, including images and the uv-plane. In contrast to time-domain data recorded with large single-dish telescopes, visibilities from interferometers can precisely localize sources anywhere in the entire field of view. rtpipe opens interferometers to the study of fast transient sky, including sources like pulsars, stellar flares, rotating radio transients, and fast radio bursts. Key portions of the search pipeline, such as image generation and dedispersion, have been accelerated. That, in combination with its multi-threaded, multi-node design, makes rtpipe capable of searching millisecond timescale data in real time on small compute clusters.

[ascl:1802.011]
runDM: Running couplings of Dark Matter to the Standard Model

runDM calculates the running of the couplings of Dark Matter (DM) to the Standard Model (SM) in simplified models with vector mediators. By specifying the mass of the mediator and the couplings of the mediator to SM fields at high energy, the code can calculate the couplings at low energy, taking into account the mixing of all dimension-6 operators. runDM can also extract the operator coefficients relevant for direct detection, namely low energy couplings to up, down and strange quarks and to protons and neutrons.

[ascl:1406.007]
RV: Radial Components of Observer's Velocity

The RV program produces a report listing the components, in a given direction, of the observer's velocity on a given date. This allows an observed radial velocity to be referred to an appropriate standard of rest -- typically either the Sun or an LSR.

As a secondary function, RV computes light time components to the Sun, thus allowing the times of phenomena observed from a terrestrial observatory to be referred to a heliocentric frame of reference. n.b. It will of course, in addition, be necessary to express the observations in the appropriate timescale as well as applying light time corrections. In particular, it is likely that an observed UTC will need to be converted to TDB as well as being corrected to the Sun.)

RV is distributed with the Starlink software collection (ascl:1110.012) and uses SLALIB (ascl:1403.025).

[ascl:1505.020]
rvfit: Radial velocity curves fitting for binary stars or exoplanets

rvfit, developed in IDL 7.0, fits non-precessing keplerian radial velocity (RV) curves for double-line and single-line binary stars or exoplanets. It fits a simple keplerian model to the observed RV and computes the seven parameters (six for a single-line system) from the model. Some parameters can be fixed beforehand if they are known, for instance, if photometric observations are available. The fit is done using an Adaptive Simulated Annealing algorithm optimized for this specific task. Simulated Annealing methods are powerful heuristic algorithms to minimize functions in multiparametric spaces.

[ascl:1210.031]
RVLIN: Fitting Keplerian curves to radial velocity data

The RVLIN package for IDL is a set of routines that quickly fits an arbitrary number of Keplerian curves to radial velocity data. It can handle data from multiple telescopes (i.e. it solves for the offset), constraints on P, e, and time of peri passage, and can incorporate transit timing data. The code handles fixed periods and circular orbits in combination and transit time constraints, including for multiple transiting planets.

[ascl:9912.003]
RVSAO 2.0: Digital Redshifts and Radial Velocities

RVSAO is a set of programs to obtain redshifts and radial velocities from digital spectra. RVSAO operates in the IRAF (Tody 1986, 1993) environment. The heart of the system is xcsao, which implements the cross-correlation method, and is a direct descendant of the system built by Tonry and Davis (1979). emsao uses intelligent heuristics to search for emission lines in spectra, then fits them to obtain a redshift. sumspec shifts and sums spectra to build templates for cross-correlation. linespec builds synthetic spectra given a list of spectral lines. bcvcorr corrects velocities for the motion of the earth. We discuss in detail the parameters necessary to run xcsao and emsao properly. We discuss the reliability and error associated with xcsao derived redshifts. We develop an internal error estimator, and we show how large, stable surveys can be used to develop more accurate error estimators. We develop a new methodology for building spectral templates for galaxy redshifts. We show how to obtain correlation velocities using emission line templates. Emission line correlations are substantially more efficient than the previous standard technique, automated emission line fitting. We compare the use of RVSAO with new methods, which use Singular Value Decomposition and $chi^2$ fitting techniques.

[ascl:1907.013]
RVSpecFit: Radial velocity and stellar atmospheric parameter fitting

RVSpecFit determines radial velocities and stellar atmospheric parameters from spectra by direct pixel fitting by interpolated stellar templates. The code doesn't require spectrum normalization and can deal with non-flux calibrated spectra. RVSpecFit is able to fit multiple spectra simultaneously.

[ascl:1606.008]
s2: Object oriented wrapper for functions on the sphere

The s2 package can represent any arbitrary function defined on the sphere. Both real space map and harmonic space spherical harmonic representations are supported. Basic sky representations have been extended to simulate full sky noise distributions and Gaussian cosmic microwave background realisations. Support for the representation and convolution of beams is also provided. The code requires HEALPix (ascl:1107.018) and CFITSIO (ascl:1010.001).

[ascl:1110.013]
S2HAT: Scalable Spherical Harmonic Transform Library

Many problems in astronomy and astrophysics require a computation of the spherical harmonic transforms. This is in particular the case whenever data to be analyzed are distributed over the sphere or a set of corresponding mock data sets has to be generated. In many of those contexts, rapidly improving resolutions of both the data and simulations puts increasingly bigger emphasis on our ability to calculate the transforms quickly and reliably.

The scalable spherical harmonic transform library S2HAT consists of a set of flexible, massively parallel, and scalable routines for calculating diverse (scalar, spin-weighted, etc) spherical harmonic transforms for a class of isolatitude sky grids or pixelizations. The library routines implement the standard algorithm with the complexity of O(n^3/2), where n is a number of pixels/grid points on the sphere, however, owing to their efficient parallelization and advanced numerical implementation, they achieve very competitive performance and near perfect scalability. S2HAT is written in Fortran 90 with a C interface. This software is a derivative of the spherical harmonic transforms included in the HEALPix package and is based on both serial and MPI routines of its version 2.01, however, since version 2.5 this software is fully autonomous of HEALPix and can be compiled and run without the HEALPix library.

[ascl:1211.001]
S2LET: Fast wavelet analysis on the sphere

S2LET provides high performance routines for fast wavelet analysis of signals on the sphere. It uses the SSHT code built on the MW sampling theorem to perform exact spherical harmonic transforms on the sphere. The resulting wavelet transform implemented in S2LET is theoretically exact, i.e. a band-limited signal can be recovered from its wavelet coefficients exactly and the wavelet coefficients capture all the information. S2LET also supports the HEALPix sampling scheme, in which case the transforms are not theoretically exact but achieve good numerical accuracy. The core routines of S2LET are written in C and have interfaces in Matlab, IDL and Java. Real signals can be written to and read from FITS files and plotted as Mollweide projections.

[ascl:1103.003]
S2PLOT: Three-dimensional (3D) Plotting Library

We present a new, three-dimensional (3D) plotting library with advanced features, and support for standard and enhanced display devices. The library - S2PLOT - is written in C and can be used by C, C++ and FORTRAN programs on GNU/Linux and Apple/OSX systems. S2PLOT draws objects in a 3D (x,y,z) Cartesian space and the user interactively controls how this space is rendered at run time. With a PGPLOT inspired interface, S2PLOT provides astronomers with elegant techniques for displaying and exploring 3D data sets directly from their program code, and the potential to use stereoscopic and dome display devices. The S2PLOT architecture supports dynamic geometry and can be used to plot time-evolving data sets, such as might be produced by simulation codes. In this paper, we introduce S2PLOT to the astronomical community, describe its potential applications, and present some example uses of the library.

[ascl:2005.009]
s3PCF: Compute the 3-point correlation function in the squeezed limit

s3PCF computes the 3-point correlation function (3PCF) in the squeezed limit given galaxy positions and pair positions. The code is currently written specifically for the Abacus simulations, but the main functionalities can be also easily adapted for other galaxy catalogs with the appropriate properties.

[ascl:1111.003]
Saada: A Generator of Astronomical Database

Saada transforms a set of heterogeneous FITS files or VOtables of various categories (images, tables, spectra, etc.) in a powerful database deployed on the Web. Databases are located on your host and stay independent of any external server. This job doesn’t require writing code. Saada can mix data of various categories in multiple collections. Data collections can be linked each to others making relevant browsing paths and allowing data-mining oriented queries. Saada supports 4 VO services (Spectra, images, sources and TAP) . Data collections can be published immediately after the deployment of the Web interface.

[ascl:1306.001]
SAC: Sheffield Advanced Code

The Sheffield Advanced Code (SAC) is a fully non-linear MHD code designed for simulations of linear and non-linear wave propagation in gravitationally strongly stratified magnetized plasma. It was developed primarily for the forward modelling of helioseismological processes and for the coupling processes in the solar interior, photosphere, and corona; it is built on the well-known VAC platform that allows robust simulation of the macroscopic processes in gravitationally stratified (non-)magnetized plasmas. The code has no limitations of simulation length in time imposed by complications originating from the upper boundary, nor does it require implementation of special procedures to treat the upper boundaries. SAC inherited its modular structure from VAC, thereby allowing modification to easily add new physics.

[submitted]
Sacc: Save All Correlations and Covariances

Zuntz, J.; Slosar, A.; Alonso, D.; Becker, M.; Broussard, A.; McClintock, T.; Nicola, A.; Miyatake, H.; Sanchez, J.; Neveu, J.

SACC (Save All Correlations and Covariances) is a format and reference library for general storage

of summary statistic measurements for the Dark Energy Science Collaboration (DESC) within and from the Large Synoptic Survey Telescope (LSST) project's Dark Energy Science Collaboration.

[ascl:1601.006]
SAGE: Semi-Analytic Galaxy Evolution

Croton, Darren J.; Stevens, Adam R. H.; Tonini, Chiara; Garel, Thibault; Bernyk, Maksym; Bibiano, Antonio; Hodkinson, Luke; Mutch, Simon J.; Poole, Gregory B.; Shattow, Genevieve M.

SAGE (Semi-Analytic Galaxy Evolution) models galaxy formation in a cosmological context. SAGE has been rebuilt to be modular and customizable. The model runs on any dark matter cosmological N-body simulation whose trees are organized in a supported format and contain a minimum set of basic halo properties.

[ascl:1203.011]
SALT2: Spectral Adaptive Lightcurve Template

SALT (Spectral Adaptive Lightcurve Template) is a package for Type Ia Supernovae light curve fitting. Its main purpose is to provide a distance estimator but it can also be used for photometric redshifts, and spectroscopic + photometric identification. This code is also known by the name snfit.

[ascl:1407.006]
SAMI: Sydney-AAO Multi-object Integral field spectrograph pipeline

Allen, J. T.; Green, A. W.; Fogarty, L. M. R.; Sharp, R.; Nielsen, J.; Konstantopoulos, I.; Taylor, E. N.; Scott, N.; Cortese, L.; Richards, S. N.; Croom, S.; Owers, M. S.; Bauer, A. E.; Sweet, S. M.; Bryant, J. J.

The SAMI (Sydney-AAO Multi-object Integral field spectrograph) pipeline reduces data from the Sydney-AAO Multi-object Integral field spectrograph (SAMI) for the SAMI Galaxy Survey. The python code organizes SAMI data and, along with the AAO 2dfdr package, carries out all steps in the data reduction, from raw data to fully calibrated datacubes. The principal steps are: data management, use of 2dfdr to produce row-stacked spectra, flux calibration, correction for telluric absorption, removal of atmospheric dispersion, alignment of dithered exposures, and drizzling onto a regular output grid. Variance and covariance information is tracked throughout the pipeline. Some quality control routines are also included.

[ascl:1504.011]
samiDB: A Prototype Data Archive for Big Science Exploration

samiDB is an archive, database, and query engine to serve the spectra, spectral hypercubes, and high-level science products that make up the SAMI Galaxy Survey. Based on the versatile Hierarchical Data Format (HDF5), samiDB does not depend on relational database structures and hence lightens the setup and maintenance load imposed on science teams by metadata tables. The code, written in Python, covers the ingestion, querying, and exporting of data as well as the automatic setup of an HTML schema browser. samiDB serves as a maintenance-light data archive for Big Science and can be adopted and adapted by science teams that lack the means to hire professional archivists to set up the data back end for their projects.

[ascl:1605.015]
SAND: Automated VLBI imaging and analyzing pipeline

The Search And Non-Destroy (SAND) is a VLBI data reduction pipeline composed of a set of Python programs based on the AIPS interface provided by ObitTalk. It is designed for the massive data reduction of multi-epoch VLBI monitoring research. It can automatically investigate calibrated visibility data, search all the radio emissions above a given noise floor and do the model fitting either on the CLEANed image or directly on the uv data. It then digests the model-fitting results, intelligently identifies the multi-epoch jet component correspondence, and recognizes the linear or non-linear proper motion patterns. The outputs including CLEANed image catalogue with polarization maps, animation cube, proper motion fitting and core light curves. For uncalibrated data, a user can easily add inline modules to do the calibration and self-calibration in a batch for a specific array.

[ascl:0003.002]
SAOImage DS9: A utility for displaying astronomical images in the X11 window environment

SAOImage DS9 is an astronomical imaging and data visualization application. DS9 supports FITS images and binary tables, multiple frame buffers, region manipulation, and many scale algorithms and colormaps. It provides for easy communication with external analysis tasks and is highly configurable and extensible via XPA and SAMP. DS9 is a stand-alone application. It requires no installation or support files. Versions of DS9 currently exist for Solaris, Linux, MacOSX, and Windows. All versions and platforms support a consistent set of GUI and functional capabilities. DS9 supports advanced features such as multiple frame buffers, mosaic images, tiling, blinking, geometric markers, colormap manipulation, scaling, arbitrary zoom, rotation, pan, and a variety of coordinate systems. DS9 also supports FTP and HTTP access. The GUI for DS9 is user configurable. GUI elements such as the coordinate display, panner, magnifier, horizontal and vertical graphs, button bar, and colorbar can be configured via menus or the command line. DS9 is a Tk/Tcl application which utilizes the SAOTk widget set. It also incorporates the X Public Access (XPA) mechanism to allow external processes to access and control its data, GUI functions, and algorithms.

[ascl:1210.029]
Sapporo: N-body simulation library for GPUs

Sapporo mimics the behavior of GRAPE hardware and uses the GPU to perform high-precision gravitational N-body simulations. It makes use of CUDA and therefore only works on NVIDIA GPUs. N-body codes currently running on GRAPE-6 can switch to Sapporo by a simple relinking of the library. Sapporo's precision is comparable to that of GRAPE-6, even though internally the GPU hardware is limited to single precision arithmetics. This limitation is effectively overcome by emulating double precision for calculating the distance between particles.

[ascl:1907.005]
SARA-PPD: Preconditioned primal-dual algorithm for radio-interferometric imaging

SARA-PPD is a proof of concept MATLAB implementation of an acceleration strategy for a recently proposed primal-dual distributed algorithm. The algorithm optimizes resolution by accounting for the correct noise statistics, leverages natural weighting in the definition of the minimization problem for image reconstruction, and optimizes sensitivity by enabling accelerated convergence through a preconditioning strategy incorporating sampling density information. This algorithm offers efficient processing of large-scale data sets that will be acquired by next generation radio-interferometers such as the Square Kilometer Array.

[ascl:1904.020]
SARAH: SUSY and non-SUSY model builder and analyzer

SARAH builds and analyzes SUSY and non-SUSY models. It calculates all vertices, mass matrices, tadpoles equations, one-loop corrections for tadpoles and self-energies, and two-loop RGEs for a given model. SARAH writes model files for a variety of other software packages for dark matter studies, includes many SUSY and non-SUSY models, and makes implementing new models efficient and straightforward. Written in Mathematica, SARAH can also use output from Vevacious (ascl:1904.019) to check for the global minimum for a given model and parameter point.

[ascl:1404.004]
SAS: Science Analysis System for XMM-Newton observatory

The Science Analysis System (SAS) is an extensive suite of software tasks developed to process the data collected by the XMM-Newton Observatory. The SAS extracts standard (spectra, light curves) and/or customized science products, and allows reproductions of the reduction pipelines run to get the PPS products from the ODFs files. SAS includes a powerful and extensive suite of FITS file manipulation packages based on the Data Access Layer library.

[ascl:1707.002]
SASRST: Semi-Analytic Solutions for 1-D Radiative Shock Tubes

SASRST, a small collection of Python scripts, attempts to reproduce the semi-analytical one-dimensional equilibrium and non-equilibrium radiative shock tube solutions of Lowrie & Rauenzahn (2007) and Lowrie & Edwards (2008), respectively. The included code calculates the solution for a given set of input parameters and also plots the results using Matplotlib. This software was written to provide validation for numerical radiative shock tube solutions produced by a radiation hydrodynamics code.

[ascl:2103.005]
satcand: Orbital stability and tidal migration constraints for KOI exomoon candidates

satcand applies theoretical constraints of orbital stability and tidal migration to KOI exomoon candidates. The package can evaluate the tidal migration within a Sun-Earth-Moon system, plot angular velocity over time, and calculate the migration time scale (T1) and the total migration time scale, among other things. In addition to the theoretical constraints, observational constraints can be applied.

[ascl:1309.005]
SATMC: SED Analysis Through Monte Carlo

SATMC is a general purpose, MCMC-based SED fitting code written for IDL and Python. Following Bayesian statistics and Monte Carlo Markov Chain algorithms, SATMC derives the best fit parameter values and returns the sampling of parameter space used to construct confidence intervals and parameter-parameter confidence contours. The fitting may cover any range of wavelengths. The code is designed to incorporate any models (and potential priors) of the user's choice. The user guide lists all the relevant details for including observations, models and usage under both IDL and Python.

[ascl:1601.012]
SavGolFilterCov: Savitzky Golay filter for data with error covariance

A Savitzky–Golay filter is often applied to data to smooth the data without greatly distorting the signal; however, almost all data inherently comes with noise, and the noise properties can differ from point to point. This python script improves upon the traditional Savitzky-Golay filter by accounting for error covariance in the data. The inputs and arguments are modeled after scipy.signal.savgol_filter.

[ascl:1904.015]
SBGAT: Small Bodies Geophysical Analysis Tool

SBGAT (Small Body Geophysical Analysis Tool) generates simulated data originating from small bodies shape models, combined with advanced shape-modification properties. It uses polyhedral shape models from which can be computed mass properties such as volume, center of mass, and inertia, synthetic observations such as lightcurves and radar, and which can be used within dynamical models, such as spherical harmonics and polyhedron gravity modeling. SBGAT can generate spherical harmonics expansions from constant-density polyhedra (and export them to JSON) and evaluate the spherical harmonics expansions. It can also generate YORP coefficients, multi-threaded Polyhedron Gravity Model gravity and potential evaluations, and synthetic light-curve and radar observations for single/primary asteroids.

SBGAT has two distinct packages: a dynamic library SBGAT Core that contains the data structure and algorithm backbone of SBGAT, and SBGAT Gui, which wraps the former inside a VTK, Qt user interface to facilitate user/data interaction. SBGAT Core can be used without the SBGAT Gui wrapper.

[ascl:1907.014]
sbpy: Small-body planetary astronomy

Mommert, Michael; Kelley, Michael S. P.; de Val-Borro, Miguel; Li, Jian-Yang; Guzman, Giannina; Sipőcz, Brigitta; Ďurech, Josef; Granvik, Mikael; Grundy, Will; Moskovitz, Nick; Penttilä, Antti; Samarasinha, Nalin

sbpy, an Astropy affiliated package, supplements functionality provided by Astropy (ascl:1304.002) with functions and methods that are frequently used for planetary astronomy with a clear focus on asteroids and comets. It offers access tools for various databases for orbital and physical data, spectroscopy analysis tools and models, photometry models for resolved and unresolved observations, ephemerides services, and other tools useful for small-body planetary astronomy.

Would you like to view a random code?