ASCL.net

Astrophysics Source Code Library

Making codes discoverable since 1999

Browsing Codes

Results 101-200 of 3450 (3361 ASCL, 89 submitted)

Order
Title Date
 
Mode
Abstract Compact
Per Page
50100250All
[ascl:2307.037] WDMWaveletTransforms: Fast forward and inverse WDM wavelet transforms

WDMWaveletTransforms implements the fast forward and inverse WDM wavelet transforms in Python from both the time and frequency domains. The frequency domain transforms are inherently faster and more accurate. The wavelet domain->frequency domain and frequency domain->wavelet domain transforms are nearly exact numerical inverses of each other for a variety of inputs tested, including Gaussian random noise. WDMWaveletTransforms has both command line and Python interfaces.

[ascl:1807.020] wdmerger: Simulate white dwarf mergers with CASTRO

wdmerger simulates binary white dwarf mergers (and related events) in CASTRO (ascl:1105.010) and provides useful information on the viability of mergers of white dwarfs as a progenitor for Type Ia supernovae.

[ascl:1806.012] WDEC: White Dwarf Evolution Code

WDEC (White Dwarf Evolution Code), written in Fortran, offers a fast and fairly easy way to produce models of white dwarfs. The code evolves hot (~100,000 K) input models down to a chosen effective temperature by relaxing the models to be solutions of the equations of stellar structure. The code can also be used to obtain g-mode oscillation modes for the models.

[ascl:2004.004] WD: Wilson-Devinney binary star modeling

Wilson-Devinney binary star modeling code (WD) is a complete package for modeling binary stars and their eclipes and consists of two main modules. The LC module generates light and radial velocity curves, spectral line profiles, images, conjunction times, and timing residuals; the DC module handles differential corrections, performing parameter adjustment of light curves, velocity curves, and eclipse timings by the Least Squares criterion. WD handles eccentric orbits and asynchronous rotation, and can compute velocity curves (with proximity and eclipse effects). It offers options for detailed reflection and nonlinear (logarithmic law) limb darkening, adjustment of spot parameters, an optional provision for spots to drift over the surface, and can follow light curve development over large numbers of orbits. Absolute flux solution allow Direct Distance Estimation (DDE) and there are improved solutions for ellipsoidal variables and for eclipsing binaries (EBs) with very shallow eclipses. Absolute flux solutions also can estimate temperatures of both EB components under suitable circumstances.

[ascl:1109.015] WCSTools: Image Astrometry Toolkit

WCSTools is a package of programs and a library of utility subroutines for setting and using the world coordinate systems (WCS) in the headers of the most common astronomical image formats, FITS and IRAF .imh, to relate image pixels to sky coordinates. In addition to dealing with image WCS information, WCSTools has extensive catalog search, image header manipulation, and coordinate and time conversion tasks. This software is all written in very portable C, so it should compile and run on any computer with a C compiler.

[ascl:1108.003] WCSLIB and PGSBOX

WCSLIB is a C library, supplied with a full set of Fortran wrappers, that implements the "World Coordinate System" (WCS) standard in FITS (Flexible Image Transport System). It also includes a PGPLOT-based routine, PGSBOX, for drawing general curvilinear coordinate graticules and a number of utility programs.

[ascl:2311.001] wcpy: Wavelength Calibrator

The graphical user interface Wavelength Calibrator facilitates wavelength calibration. Although developed for astronomical data reduction, it can also be used in any place where wavelength calibration is needed.

[ascl:2206.024] Wavetrack: Arbitrary time-evolving solar object recognition and tracking

Wavetrack recognizes and tracks CME shock waves, filaments, and other solar objects. The code creates base images by averaging а series of images a few minutes prior to the start of the eruption and constructs base difference images by subtracting base images from the current raw image of the sequence. This enhances the change in intensity caused by coronal bright fronts, omits static details, and reduces noise. Wavetrack then chooses an appropriate intensity interval and decomposes the base difference or running difference image with an A-Trous wavelet transform, where each wavelet coefficient is obtained by convolving the image array with a corresponding iteration of the wavelet kernel. When the maximum value of the wavelet coefficients for a connected set of pixels satisfies certain conditions, this region is considered as a structure on the respective wavelet coefficient. Separate stand-alone object masks are obtained with a clustering algorithm and objects are renumbered according to the number of the quadrant they belong at each iteration.

[ascl:2307.038] WarpX: Time-based electromagnetic and electrostatic Particle-In-Cell code

WarpX is an advanced electromagnetic & electrostatic Particle-In-Cell code. It supports many features including Perfectly-Matched Layers (PML), mesh refinement, and the boosted-frame technique. A highly-parallel and highly-optimized code, WarpX can run on GPUs and multi-core CPUs, includes load balancing capabilities, and scales to the largest supercomputers.

[ascl:1807.002] Warpfield: Winds And Radiation Pressure: Feedback Induced Expansion, colLapse and Dissolution

Warpfield (Winds And Radiation Pressure: Feedback Induced Expansion, colLapse and Dissolution) calculates shell dynamics and shell structure simultaneously for isolated massive clouds (≥105 M). This semi-analytic 1D feedback model scans a large range of physical parameters (gas density, star formation efficiency, and metallicity) to estimate escape fractions of ionizing radiation fesc, I, the minimum star formation efficiency ∊min required to drive an outflow, and recollapse time-scales for clouds that are not destroyed by feedback.

[ascl:2207.019] walter: Predictor for the number of resolved stars in a given observation from RST

walter calculates the number density of stars detected in a given observation aiming to resolve a stellar population. The code also calculates the exposure time needed to reach certain population features, such as the horizontal branch, and provides an estimate of the crowding limit. walter was written with the expectation that such calculations will be very useful for planning surveys with the Nancy Grace Roman Space Telescope (RST, formerly WFIRST).

[ascl:2108.004] WaldoInSky: Anomaly detection algorithms for time-domain astronomy

WaldoInSky finds anomalous astronomical light curves and their analogs. The package contains four methods: an adaptation of the Unsupervised Random Forest for anomaly detection in light curves that operates on the light curve points and their power spectra; two manifold-learning methods (the t-SNE and UMAP) that operate on the DMDT maps (image representations of the light curves), and that can be used to find analog light curves in the low-dimensional representation; and an Isolation Forest method for evaluating approaches of light curve pre-processing, before they are passed to the anomaly detectors. WaldoInSky also contain code for random sparsification of light curves.

[ascl:2301.021] WALDO: Waveform AnomaLy DetectOr

WALDO (Waveform AnomaLy DetectOr) flags possible anomalous Gravitational Waves from Numerical Relativity catalogs using deep learning. It uses a U-Net architecture to learn the waveform features of a dataset. After computing the mismatch between those waveforms and the neural predictions, WALDO isolates high mismatch evaluations for anomaly search.

[ascl:1710.001] vysmaw: Fast visibility stream muncher

The vysmaw client library facilitates the development of code for processes to tap into the fast visibility stream on the National Radio Astronomy Observatory's Very Large Array correlator back-end InfiniBand network. This uses the vys protocol to allow loose coupling to clients that need to remotely access memory over an Infiniband network.

[ascl:1704.011] VULCAN: Chemical Kinetics For Exoplanetary Atmospheres

VULCAN describes gaseous chemistry from 500 to 2500 K using a reduced C-H-O chemical network with about 300 reactions. It uses eddy diffusion to mimic atmospheric dynamics and excludes photochemistry, and can be used to examine the theoretical trends produced when the temperature-pressure profile and carbon-to-oxygen ratio are varied.

[ascl:1407.013] VStar: Variable star data visualization and analysis tool

VStar is a multi-platform, easy-to-use variable star data visualization and analysis tool. Data for a star can be read from the AAVSO (American Association of Variable Star Observers) database or from CSV and TSV files. VStar displays light curves and phase plots, can produce a mean curve, and analyzes time-frequency with Weighted Wavelet Z-Transform. It offers tools for period analysis, filtering, and other functions.

[ascl:1811.017] VPLanet: Virtual planet simulator

VPLanet (Virtual Planetary Laboratory) simulates planetary system evolution with a focus on habitability. Physical models, typically consisting of ordinary differential equations for stellar, orbital, tidal, rotational, atmospheric, internal, magnetic, climate, and galactic evolution, are coupled together to simulate evolution for the age of a system.

[ascl:1408.016] vpguess: Fitting multiple Voigt profiles to spectroscopic data

vpguess facilitates the fitting of multiple Voigt profiles to spectroscopic data. It is a graphical interface to VPFIT (ascl:1408.015). Originally meant to simplify the process of setting up first guesses for a subsequent fit with VPFIT, it has developed into a full interface to VPFIT. It may also be used independently of VPFIT for displaying data, playing around with data and models, "chi-by-eye" fits, displaying the result of a proper fit, pretty plots, etc. vpguess is written in C, and the graphics are based on PGPLOT (ascl:1103.002).

[ascl:1408.015] VPFIT: Voigt profile fitting program

The VPFIT program fits multiple Voigt profiles (convolved with the instrument profiles) to spectroscopic data that is in FITS or an ASCII file. It requires CFITSIO (ascl:1010.001) and PGPLOT (ascl:1103.002); the tarball includes RDGEN (ascl:1408.017), which can be used with VPFIT to set up the fits, fit the profiles, and examine the result in interactive mode for setting up initial guesses; vpguess (ascl:1408.016) can also be used to set up an initial file.

[ascl:1309.008] VOStat: Statistical analysis of astronomical data

VOStat allows astronomers to use both simple and sophisticated statistical routines on large datasets. This tool uses the large public-domain statistical computing package R. Datasets can be uploaded in either ASCII or VOTABLE (preferred) format. The statistical computations are performed by the VOStat and results are returned to the user.

[ascl:1205.011] VOSpec: VO Spectral Analysis Tool

VOSpec is a multi-wavelength spectral analysis tool with access to spectra, theoretical models and atomic and molecular line databases registered in the VO. The standard tools of VOSpec include line and continuum fitting, redshift and reddening correction, spectral arithmetic and convolution between spectra, equivalent width and flux calculations, and a best fitting algorithm for fitting selected SEDs to a TSAP service. VOSpec offers several display modes (tree vs table) and organising functionalities according to the available metadata for each service, including distance from the observation position.

[ascl:2206.001] vortex: Helmholtz-Hodge decomposition for an AMR velocity field

vortex performs a Helmholtz-Hodge decomposition on vector fields defined on AMR grids, decomposing a vector field in its solenoidal (divergence-less) and compressive (curl-less) parts. It works natively on vector fields defined on Adaptive Mesh Refinement (AMR) grids, so that it can perform the decomposition over large dynamical ranges; it is also applicable to particle-based simulations. As vortex is devised primarily to investigate the properties of the turbulent velocity field in the Intracluster Medium (ICM), it also includes routines for multi-scale filtering the velocity field.

[ascl:1211.006] VorBin: Voronoi binning method

VorBin (Voronoi binning method) bins two-dimensional data to a constant signal-to-noise ratio per bin. It optimally solves the problem of preserving the maximum spatial resolution of general two-dimensional data, given a constraint on the minimum signal-to-noise ratio. The method is available in both IDL and Python.

[ascl:1309.006] VOPlot: Toolkit for Scientific Discovery using VOTables

VOPlot is a tool for visualizing astronomical data. It was developed in Java and acts on data available in VOTABLE, ASCII and FITS formats. VOPlot is available as a stand alone version, which is to be installed on the user's machine, or as a web-based version fully integrated with the VizieR database.

[ascl:1309.007] VOMegaPlot: Plotting millions of points

VOMegaPlot, a Java based tool, has been developed for visualizing astronomical data that is available in VOTable format. It has been specifically optimized for handling large number of points (in the range of millions). It has the same look and feel as VOPlot (ascl:1309.006) and both these tools have certain common functionality.

[ascl:2109.003] VOLKS2: VLBI Observation for transient Localization Keen Searcher

The VOLK2 (VLBI Observation for transient Localization Keen Searcher) pipeline conducts single pulse searches and localization in regular VLBI observations as well as single pulse detections from known sources in dedicated observations. In VOLKS2, the search and localization are two independent steps. The search step takes the idea of geodetic VLBI post processing, which fully utilizes the cross spectrum fringe phase information to maximize the signal power. Compared with auto spectrum based method, it is able to extract single pulses from highly RFI contaminated data. The localization uses the geodetic VLBI solving methods, which derives the single pulse location by solving a set of linear equations given the relation between the residual delay and the offset to a priori position.

[ascl:1811.016] VoigtFit: Absorption line fitting for Voigt profiles

VoigtFit fits Voigt profiles to absorption lines. It fits multiple components for various atomic lines simultaneously, allowing parameters to be tied and fixed, and can automatically fit a polynomial continuum model together with the line profiles. A physical model can be used to constrain thermal and turbulent broadening of absorption lines as well as implementing molecular excitation models. The code uses a χ2 minimization approach to find the best solution and offers interactive features such as manual continuum placement locally around each line, manual masking of undesired fitting regions, and interactive definition of velocity components for various elements, improving the ease of estimating initial guesses.

[ascl:1411.003] voevent-parse: Parse, manipulate, and generate VOEvent XML packets

voevent-parse, written in Python, parses, manipulates, and generates VOEvent XML packets; it is built atop lxml.objectify. Details of transients detected by many projects, including Fermi, Swift, and the Catalina Sky Survey, are currently made available as VOEvents, which is also the standard alert format by future facilities such as LSST and SKA. However, working with XML and adhering to the sometimes lengthy VOEvent schema can be a tricky process. voevent-parse provides convenience routines for common tasks, while allowing the user to utilise the full power of the lxml library when required. An earlier version of voevent-parse was part of the pysovo (ascl:1411.002) library.

[ascl:1304.005] VOBOZ/ZOBOV: Halo-finding and Void-finding algorithms

VOBOZ (VOronoi BOund Zones) is an algorithm to find haloes in an N-body dark matter simulation which has little dependence on free parameters.

ZOBOV (ZOnes Bordering On Voidness) is an algorithm that finds density depressions in a set of points without any free parameters or assumptions about shape. It uses the Voronoi tessellation to estimate densities to find both voids and subvoids. It also measures probabilities that each void or subvoid arises from Poisson fluctuations.

[ascl:2009.002] vlt-sphere: Automatic VLT/SPHERE data reduction and analysis

The high-contrast imager SPHERE at the Very Large Telescope combines extreme adaptive optics and coronagraphy to directly image exoplanets in the near-infrared. The vlt-sphere package enables easy reduction of the data coming from IRDIS and IFS, the two near-infrared subsystems of SPHERE. The package relies on the official ESO pipeline (ascl:1402.010), which must be installed separately.

[ascl:1908.014] Vlasiator: Hybrid-Vlasov simulation code

Vlasiator is a 6-dimensional Vlasov theory-based simulation. It simulates the entire near-Earth space at a global scale using the kinetic hybrid-Vlasov approach, to study fundamental plasma processes (reconnection, particle acceleration, shocks), and to gain a deeper understanding of space weather.

[ascl:2207.020] vKompth: Time-dependent Comptonization model for black-hole X-ray binaries

vKompth fits the energy-dependent rms-amplitude and phase-lag spectra of low-frequency quasi-periodic oscillations in low mass black-hole X-ray binaries using a variable Comptonization model. The accretion disc is modeled as a multi-temperature blackbody source emitting soft photons which are then Compton up-scattered in a spherical corona, including feedback of Comptonized photons that return to the disc.

[ascl:1701.002] Vizic: Jupyter-based interactive visualization tool for astronomical catalogs

Vizic is a Python visualization library that builds the connection between images and catalogs through an interactive map of the sky region. The software visualizes catalog data over a custom background canvas using the shape, size and orientation of each object in the catalog and displays interactive and customizable objects in the map. Property values such as redshift and magnitude can be used to filter or apply colormaps, and objects can be selected for further analysis through standard Python functions from inside a Jupyter notebook.

Vizic allows custom overlays to be appended dynamically on top of the sky map; included are Voronoi, Delaunay, Minimum Spanning Tree and HEALPix layers, which are helpful for visualizing large-scale structure. Overlays can be generated, added or removed dynamically with one line of code. Catalog data is kept in a non-relational database. The Jupyter Notebook allows the user to create scripts to analyze and plot the data selected/displayed in the interactive map, making Vizic a powerful and flexible interactive analysis tool. Vizic be used for data inspection, clustering analysis, galaxy alignment studies, outlier identification or simply large-scale visualizations.

[ascl:1402.001] Vissage: ALMA VO Desktop Viewer

Vissage (VISualisation Software for Astronomical Gigantic data cubEs) is a FITS browser primarily targeting FITS data cubes obtained from ALMA. Vissage offers basic functionality for viewing three-dimensional data cubes, integrated intensity map, flipbook, channel map, and P-V diagram. It has several color sets and color scales available, offers panning and zooming, and can connect with the ALMA WebQL system and the JVO Subaru Image Cutout Service.

[ascl:1011.020] VisIVO: Integrated Tools and Services for Large-Scale Astrophysical Visualization

VisIVO is an integrated suite of tools and services specifically designed for the Virtual Observatory. This suite constitutes a software framework for effective visual discovery in currently available (and next-generation) very large-scale astrophysical datasets. VisIVO consists of VisiVO Desktop - a stand alone application for interactive visualization on standard PCs, VisIVO Server - a grid-enabled platform for high performance visualization and VisIVO Web - a custom designed web portal supporting services based on the VisIVO Server functionality. The main characteristic of VisIVO is support for high-performance, multidimensional visualization of very large-scale astrophysical datasets. Users can obtain meaningful visualizations rapidly while preserving full and intuitive control of the relevant visualization parameters. This paper focuses on newly developed integrated tools in VisIVO Server allowing intuitive visual discovery with 3D views being created from data tables. VisIVO Server can be installed easily on any web server with a database repository. We discuss briefly aspects of our implementation of VisiVO Server on a computational grid and also outline the functionality of the services offered by VisIVO Web. Finally we conclude with a summary of our work and pointers to future developments.

[ascl:1103.007] VisIt: Interactive Parallel Visualization and Graphical Analysis Tool

VisIt is a free interactive parallel visualization and graphical analysis tool for viewing scientific data on Unix and PC platforms. Users can quickly generate visualizations from their data, animate them through time, manipulate them, and save the resulting images for presentations. VisIt contains a rich set of visualization features so that you can view your data in a variety of ways. It can be used to visualize scalar and vector fields defined on two- and three-dimensional (2D and 3D) structured and unstructured meshes. VisIt was designed to handle very large data set sizes in the terascale range and yet can also handle small data sets in the kilobyte range. See the table below for more details about the tool’s features.

VisIt was developed by the Department of Energy (DOE) Advanced Simulation and Computing Initiative (ASCI) to visualize and analyze the results of terascale simulations. It was developed as a framework for adding custom capabilities and rapidly deploying new visualization technologies. Although the primary driving force behind the development of VisIt was for visualizing terascale data, it is also well suited for visualizing data from typical simulations on desktop systems.

[ascl:1408.010] VisiOmatic: Celestial image viewer

VisiOmatic is a web client for IIPImage (ascl:1408.009) and is used to visualize and navigate through large science images from remote locations. It requires STIFF (ascl:1110.006), is based on the Leaflet Javascript library, and works on both touch-based and mouse-based devices.

[ascl:1802.006] VISIBLE: VISIbility Based Line Extraction

VISIBLE applies approximated matched filters to interferometric data, allowing line detection directly in visibility space. The filter can be created from a FITS image or RADMC3D output image, and the weak line data can be a CASA MS or uvfits file. The filter response spectrum can be output either to a .npy file or returned back to the user for scripting.

[ascl:2102.007] viscm: Colormaps analyzer and creator

viscm is a Python tool for visualizing and designing colormaps using colorspacious and matplotlib.

[ascl:1804.019] ViSBARD: Visual System for Browsing, Analysis and Retrieval of Data

ViSBARD interactively visualizes and analyzes space physics data. It provides an interactive integrated 3-D and 2-D environment to determine correlations between measurements across many spacecraft. It supports a variety of spacecraft data products and MHD models and is easily extensible to others. ViSBARD provides a way of visualizing multiple vector and scalar quantities as measured by many spacecraft at once. The data are displayed three-dimesionally along the orbits which may be displayed either as connected lines or as points. The data display allows the rapid determination of vector configurations, correlations between many measurements at multiple points, and global relationships. With the addition of magnetohydrodynamic (MHD) model data, this environment can also be used to validate simulation results with observed data, use simulated data to provide a global context for sparse observed data, and apply feature detection techniques to the simulated data.

[ascl:2305.002] Virtual Telescope: Next-Generation Space Telescope Simulator

Virtual Telescope predicts the signal-to-noise and other parameters of imaging and/or spectroscopic observations as a function of telescope size, detector noise, and other factors for the Next-Generation Space Telescope.

[ascl:1204.012] VirGO: A Visual Browser for the ESO Science Archive Facility

VirGO is the next generation Visual Browser for the ESO Science Archive Facility developed by the Virtual Observatory (VO) Systems Department. It is a plug-in for the popular open source software Stellarium adding capabilities for browsing professional astronomical data. VirGO gives astronomers the possibility to easily discover and select data from millions of observations in a new visual and intuitive way. Its main feature is to perform real-time access and graphical display of a large number of observations by showing instrumental footprints and image previews, and to allow their selection and filtering for subsequent download from the ESO SAF web interface. It also allows the loading of external FITS files or VOTables, the superimposition of Digitized Sky Survey (DSS) background images, and the visualization of the sky in a `real life' mode as seen from the main ESO sites. All data interfaces are based on Virtual Observatory standards which allow access to images and spectra from external data centers, and interaction with the ESO SAF web interface or any other VO applications supporting the PLASTIC messaging system.

[ascl:2108.006] viper: Velocity and IP EstimatoR

viper (Velocity and IP EstimatoR) measures differential radial velocities from stellar spectra taken through iodine or other gas cells. It convolves the product of a stellar template and a gas cell spectrum with an instrumental profile. Via least square fitting, it optimizes the parameters of the instrumental profile, the wavelength solution, flux normalization, and the stellar Doppler shift. viper offers various functions to describe the instrumental profile such as Gaussian, super-Gaussian, skewed Gaussian or mixtures of Gaussians. The code is developed for echelle spectra; it can handle data from CES, CRIRES+, KECK, OES, TCES, and UVES, and additional instruments can easily be added. A graphical interface facilitates the work with numerous flexible options.

[ascl:1603.003] VIP: Vortex Image Processing pipeline for high-contrast direct imaging of exoplanets

VIP (Vortex Image Processing pipeline) provides pre- and post-processing algorithms for high-contrast direct imaging of exoplanets. Written in Python, VIP provides a very flexible framework for data exploration and image processing and supports high-contrast imaging observational techniques, including angular, reference-star and multi-spectral differential imaging. Several post-processing algorithms for PSF subtraction based on principal component analysis are available as well as the LLSG (Local Low-rank plus Sparse plus Gaussian-noise decomposition) algorithm for angular differential imaging. VIP also implements the negative fake companion technique coupled with MCMC sampling for rigorous estimation of the flux and position of potential companions.

[ascl:1010.058] VINE: A numerical code for simulating astrophysical systems using particles

VINE is a particle based astrophysical simulation code. It uses a tree structure to efficiently solve the gravitational N-body problem and Smoothed Particle Hydrodynamics (SPH) to simulate gas dynamical effects. The code has been successfully used for a number of studies on galaxy interactions, galactic dynamics, star formation and planet formation and given the implemented physics, other applications are possible as well.

[ascl:1201.006] VIM: Visual Integration and Mining

VIM (Virtual Observatory Integration and Mining) is a data retrieval and exploration application that assumes an astronomer has a list of 'sources' (positions in the sky), and wants to explore archival catalogs, images, and spectra of the sources, in order to identify, select, and mine the list. VIM does this either through web forms, building a custom 'data matrix,' or locally through downloadable Python code. Any VO-registered catalog service can be used by VIM, as well as co-registered image cutouts from VO-image services, and spectra from VO-spectrum services. The user could, for example, show together: proper motions from GSC2, name and spectral type from NED, magnitudes and colors from 2MASS, and cutouts and spectra from SDSS. VIM can compute columns across surveys and sort on these (eg. 2MASS J magnitude minus SDSS g). For larger sets of sources, VIM utilizes the asynchronous Nesssi services from NVO, that can run thousands of cone and image services overnight.

[ascl:1403.016] Viewpoints: Fast interactive linked plotting of large multivariate data sets

Viewpoints is an interactive tool for exploratory visual analysis of large high-dimensional (multivariate) data. It uses linked scatterplots to find relations in a few seconds that can take much longer with other plotting tools. Its features include linked scatter plots with brushing, dynamic histograms, normalization, and outlier detection/removal.

[ascl:1407.014] VIDE: The Void IDentification and Examination toolkit

The Void IDentification and Examination toolkit (VIDE) identifies voids using a modified version of the parameter-free void finder ZOBOV (ascl:1304.005); a Voronoi tessellation of the tracer particles is used to estimate the density field followed by a watershed algorithm to group Voronoi cells into zones and subsequently voids. Output is a summary of void properties in plain ASCII; a Python API is provided for analysis tasks, including loading and manipulating void catalogs and particle members, filtering, plotting, computing clustering statistics, stacking, comparing catalogs, and fitting density profiles.

[ascl:1404.010] VictoriaReginaModels: Stellar evolutionary tracks

The Victoria–Regina stellar models are comprised of seventy-two grids of stellar evolutionary tracks accompanied by complementary zero-age horizontal branches and are presented in “equivalent evolutionary phase” (.eep) files. This Fortran 77 software interpolates isochrones, isochrone population functions, luminosity functions, and color functions of stellar evolutionary tracks.

[ascl:1306.015] VHD: Viscous pseudo-Newtonian accretion

VHD is a numerical study of viscous fluid accretion onto a black hole. The flow is axisymmetric and uses a pseudo-Newtonian potential to model relativistic effects near the event horizon. VHD is based on ZEUS-2D (Stone & Norman 1992) with the addition of an explicit scheme for the viscosity.

[ascl:1204.007] VH-1: Multidimensional ideal compressible hydrodynamics code

VH-1 is a multidimensional ideal compressible hydrodynamics code written in FORTRAN for use on any computing platform, from desktop workstations to supercomputers. It uses a Lagrangian remap version of the Piecewise Parabolic Method developed by Paul Woodward and Phil Colella in their 1984 paper. VH-1 comes in a variety of versions, from a simple one-dimensional serial variant to a multi-dimensional version scalable to thousands of processors.

[ascl:1904.019] Vevacious: Global minima of one-loop effective potentials generator

Vevacious takes a generic expression for a one-loop effective potential energy function and finds all the tree-level extrema, which are then used as the starting points for gradient-based minimization of the one-loop effective potential. The tunneling time from a given input vacuum to the deepest minimum, if different from the input vacuum, can be calculated. The parameter points are given as files in the SLHA format (though is not restricted to supersymmetric models), and new model files can be easily generated automatically by the Mathematica package SARAH (ascl:1904.020).

[ascl:2307.017] Veusz: Scientific plotting package

Veusz produces a wide variety of publication-ready 2D and 3D plots. Plots are created by building up plotting widgets with a consistent object-based interface, and the package provides many options for customizing plots. Veusz can import data from text, CSV, HDF5 and FITS files; datasets can also be entered within the program and new datasets created via the manipulation of existing datasets using mathematical expressions and more. The program can also be extended, by adding plugins supporting importing new data formats, different types of data manipulation or for automating tasks, and it supports vector and bitmap output, including PDF, Postscript, SVG and EMF.

[ascl:2203.022] Vetting: Stand-alone tools for vetting transit signals in Kepler, K2 and TESS data

vetting contains simple, stand-alone Python tools for vetting transiting signals in NASA's Kepler, K2, and TESS data. The code performs a centroid test to look for significant changes in the centroid of a star during a transit or eclipse. vetting requires an installation of Python 3.8 or higher.

[ascl:1503.011] VESPA: False positive probabilities calculator

Validation of Exoplanet Signals using a Probabilistic Algorithm (VESPA) calculates false positive probabilities and statistically validates transiting exoplanets. Written in Python, it uses isochrones [ascl:1503.010] and the package simpledist.

[ascl:1802.005] Verne: Earth-stopping effect for heavy dark matter

Verne calculates the Earth-stopping effect for super-heavy Dark Matter (DM). The code allows you to calculate the speed distribution (and DM signal rate) at an arbitrary detector location on the Earth. The calculation takes into account the full anisotropic DM velocity distribution and the full velocity dependence of the DM-nucleus cross section. Results can be obtained for any DM mass and cross section, though the results are most reliable for very heavy DM particles.

[ascl:1802.002] venice: Mask utility

venice reads a mask file (DS9 or fits type) and a catalogue of objects (ascii or fits type) to create a pixelized mask, find objects inside/outside a mask, or generate a random catalogue of objects inside/outside a mask. The program reads the mask file and checks if a point, giving its coordinates, is inside or outside the mask, i.e. inside or outside at least one polygon of the mask.

[ascl:1911.020] VELOCIraptor-STF: Six-dimensional Friends-of-Friends phase space halo finder

VELOCIraptor-STF, formerly STructure Finder (ascl:1306.009), is a 6-Dimensional Friends-of-Friends (6D-FoF) phase space halo finder and constructs halo catalogs. The code uses using MPI and OpenMP APIs and can be compiled as a library for on-the-fly halo finding within an N-body/hydrodynamnical code. There is an associated halo merger tree code TreeFrog (ascl:1911.021).

[ascl:2308.014] velocileptors: Velocity-based Lagrangian and Eulerian PT expansions of redshift-space distortions

velocileptors computes the real- and redshift-space power spectra and correlation functions of biased tracers using 1-loop perturbation theory (with effective field theory counter terms and up to cubic biasing) as well as the real-space pairwise velocity moments. It provides simple computation of the power spectrum wedges or multipoles, and uses a reduced set of parameters for computing the most common case of the redshift-space power spectrum. In addition, velocileptors offers two "direct expansion" modules available in LPT and EPT.

[ascl:1010.021] velfit: A Code for Modeling Non-Circular Flows in Disk Galaxies

High-quality velocity maps of galaxies frequently exhibit signatures of non-circular streaming motions. velfit yields results that are more easily interpreted than the commonly used procedure. It can estimate the magnitudes of forced non-circular motions over a broad range of bar strengths from a strongly barred galaxy, through cases of mild bar-like distortions to placing bounds on the shapes of halos in galaxies having extended rotation curves.

This code is no longer maintained and has been superseded by DiskFit (ascl:1209.011).

[ascl:1610.009] velbin: radial velocity corrected for binary orbital motions

Velbin convolves the radial velocity offsets due to binary orbital motions with a Gaussian to model an observed velocity distribution. This can be used to measure the mean velocity and velocity dispersion from an observed radial velocity distribution, corrected for binary orbital motions. Velbin fits single- or multi-epoch data with any arbitrary binary orbital parameter distribution (as long as it can be sampled properly), however it always assumes that the intrinsic velocity distribution (i.e. corrected for binary orbital motions) is a Gaussian. Velbin samples (and edits) a binary orbital parameter distribution, fits an observed radial velocity distribution, and creates a mock radial velocity distribution that can be used to provide the fitted radial velocities in the single_epoch or multi_epoch methods.

[ascl:2301.020] VDA: Void Dwarf Analyzer

void-dwarf-analysis analyzes Keck Cosmic Web Imager datacubes to produce maps of kinematic properties (velocity and velocity dispersion), emission line fluxes, and gas-phase metallicities of void dwarf galaxies.

[ascl:2311.002] VCAL-SPHERE: Hybrid pipeline for reduction of VLT/SPHERE data

VCAL-SPHERE, for VIP-based Calibration of VLT/SPHERE data, is a versatile pipeline for high-contrast imaging of exoplanets and circumstellar disks. The pipeline covers all steps of data reduction, including raw calibration, pre-processing and post-processing (i.e., modeling and subtraction of the stellar halo), for the IFS, IRDIS-DBI and IRDIS-CI modes (and combinations thereof) of the VLT instrument SPHERE. The three main steps of the reduction correspond to different modules, where the first follows the recommended EsoRex (ascl:1504.003) workflow and associated recipes with occasional inclusion of VIP (ascl:1603.003) routines (e.g., for PCA-based sky subtraction), while the other two stages fully rely on the VIP toolbox. Although the default parameters of the pipeline should yield a good reduction in most cases, these can be tuned using JSON parameter files for each stage of the pipeline for optimal reduction of specific datasets.

[ascl:1809.004] VBBINARYLENSING: Microlensing light-curve computation

VBBinaryLensing forward models gravitational microlensing events using the advanced contour integration method; it supports single and binary lenses. The lens map is inverted on a collection of points on the source boundary to obtain a corresponding collection of points on the boundaries of the images from which the area of the images can be recovered by use of Green’s theorem. The code takes advantage of a number of techniques to make contour integration much more efficient, including using a parabolic correction to increase the accuracy of the summation, introducing an error estimate on each arc of the boundary to enable defining an optimal sampling, and allowing the inclusion of limb darkening. The code is written as a C++ library and wrapped as a Python package, and can be called from either C++ or Python.

[ascl:1704.005] VaST: Variability Search Toolkit

VaST (Variability Search Toolkit) finds variable objects on a series of astronomical images in FITS format. The software performs object detection and aperture photometry using SExtractor (ascl:1010.064) on each image, cross-matches lists of detected stars, performs magnitude calibration with respect to the first (reference) image and constructs a lightcurve for each object. The sigma-magnitude, Stetson's L variability index, Robust Median Statistic (RoMS) and other plots may be used to visually identify variable star candidates. The two distinguishing features of VaST are its ability to perform accurate aperture photometry of images obtained with non-linear detectors and to handle complex image distortions. VaST can be used in cases of unstable PSF (e.g., bad guiding or with digitized wide-field photographic images), and has been successfully applied to images obtained with telescopes ranging from 0.08 to 2.5m in diameter equipped with a variety of detectors including CCD, CMOS, MIC and photographic plates.

[ascl:1208.016] VARTOOLS: Light Curve Analysis Program

The VARTOOLS program is a command line utility that provides tools for analyzing time series astronomical data. It implements a number of routines for calculating variability/periodicity statistics of light curves, as well as tools for modifying and filtering light curves.

[ascl:2109.026] Varstar Detect: Variable star detection in TESS data

Varstar Detect uses several numerical and statistical methods to filter and interpret the data obtained from TESS. It performs an amplitude test to determine whether a star is variable and if so, provides the characteristics of each star through phenomenological analysis of the lightcurve.

[ascl:2208.007] VapoRock: Modeling magma ocean atmospheres and stellar nebula

VapoRock calculates the equilibrium partial pressures of metal-bearing gas species of specific elements above the magma ocean surface to determine the metal-bearing composition of the atmosphere as a function of temperature and the bulk composition of the magma ocean. It utilizes ENKI's ThermoEngine (ascl:2208.006) and combines estimates for element activities in silicate melts with thermodynamic data for metal and metal oxide vapor species.

[ascl:1111.012] VAPOR: Visualization and Analysis Platform for Ocean, Atmosphere, and Solar Researchers

VAPOR is the Visualization and Analysis Platform for Ocean, Atmosphere, and Solar Researchers. VAPOR provides an interactive 3D visualization environment that runs on most UNIX and Windows systems equipped with modern 3D graphics cards. VAPOR provides:

- A visual data discovery environment tailored towards the specialized needs of the astro and geosciences CFD community
- A desktop solution capable of handling terascale size data sets
- Advanced interactive 3D visualization tightly coupled with quantitative data analysis
- Support for multi-variate, time-varying data
- Close coupling with RSI's powerful interpretive data language, IDL
- Support for 3D visualization of WRF-ARW datasets

[ascl:1506.010] VAPID: Voigt Absorption-Profile [Interstellar] Dabbler

VAPID (Voigt Absorption Profile [Interstellar] Dabbler) models interstellar absorption lines. It predicts profiles and optimizes model parameters by least-squares fitting to observed spectra. VAPID allows cloud parameters to be optimized with respect to several different data set simultaneously; those data sets may include observations of different transitions of a given species, and may have different S/N ratios and resolutions.

[ascl:1309.002] VAPHOT: Precision differential aperture photometry package

VAPHOT is an aperture photometry package for precise time−series photometry of uncrowded fields, geared towards the extraction of target lightcurves of eclipsing or transiting systems. Its photometric main routine works within the IRAF (ascl:9911.002) environment and is built upon the standard aperture photometry task 'phot' from IRAF, using optimized aperture sizes. The associated analysis program 'VANALIZ' works in the IDL environment. It performs differential photometry with graphical and numerical output. VANALIZ produces plots indicative of photometric stability and permits the interactive evaluation and weighting of comparison stars. Also possible is the automatic or manual suppression of data-points and the output of statistical analyses. Several methods for the calculation of the reference brightness are offered. Specific routines for the analysis of transit 'on'-'off' photometry, comparing the target brightness inside against outside a transit are also available.

[ascl:1702.004] Validation: Codes to compare simulation data to various observations

Validation provides codes to compare several observations to simulated data with stellar mass and star formation rate, simulated data stellar mass function with observed stellar mass function from PRIMUS or SDSS-GALEX in several redshift bins from 0.01-1.0, and simulated data B band luminosity function with observed stellar mass function, and to create plots for various attributes, including stellar mass functions, and stellar mass to halo mass. These codes can model predictions (in some cases alongside observational data) to test other mock catalogs.

[ascl:1810.004] VaeX: Visualization and eXploration of Out-of-Core DataFrames

VaeX (Visualization and eXploration) interactively visualizes and explores big tabular datasets. It can calculate statistics such as mean, sum, count, and standard deviation on an N-dimensional grid up to a billion (109) objects/rows per second. Visualization is done using histograms, density plots, and 3d volume rendering, allowing interactive exploration of big data. VaeX uses memory mapping, zero memory copy policy and lazy computations for best performance, and integrates well with the Jupyter/IPython notebook/lab ecosystem.

[ascl:1406.009] VADER: Viscous Accretion Disk Evolution Resource

VADER is a flexible, general code that simulates the time evolution of thin axisymmetric accretion disks in time-steady potentials. VADER handles arbitrary viscosities, equations of state, boundary conditions, and source and sink terms for both mass and energy.

[ascl:1207.003] VAC: Versatile Advection Code

The Versatile Advection Code (VAC) is a freely available general hydrodynamic and magnetohydrodynamic simulation software that works in 1, 2 or 3 dimensions on Cartesian and logically Cartesian grids. VAC runs on any Unix/Linux system with a Fortran 90 (or 77) compiler and Perl interpreter. VAC can run on parallel machines using either the Message Passing Interface (MPI) library or a High Performance Fortran (HPF) compiler.

[ascl:1911.002] uvplot: Interferometric visibilities plotter

uvplot makes nice plots of deprojected interferometric visibilities (often called uvplots). It implements plotting functionality, handles MS tables with spectral windows with different number of channels, and can import visibilities from ASCII to MS Table. It also allows export of specific channels. uvplot can be installed inside the NRAO CASA package (ascl:1107.013).

[ascl:1410.004] UVOTPY: Swift UVOT grism data reduction

The two Swift UVOT grisms provide uv (170.0-500.0 nm) and visible (285.0-660.0 nm) spectra with a resolution of R~100 and 75. To reduce the grism data, UVOTPY extracts a spectrum given source sky position, and outputs a flux calibrated spectrum. UVOTPY is a replacement for the UVOTIMGRISM FTOOL (ascl:9912.002) in the HEADAS Swift package. Its extraction uses a curved aperture for the uv spectra, accounts the coincidence losses in the detector, provides more accurate anchor positions for the wavelength scale, and is valid for the whole detector.

[ascl:1402.017] UVMULTIFIT: Fitting astronomical radio interferometric data

UVMULTIFIT, written in Python, is a versatile library for fitting models directly to visibility data. These models can depend on frequency and fitting parameters in an arbitrary algebraic way. The results from the fit to the visibilities of sources with sizes smaller than the diffraction limit of the interferometer are superior to the output obtained from a mere analysis of the deconvolved images. Though UVMULTIFIT is based on the CASA package, it can be easily adapted to other analysis packages that have a Python API.

[ascl:1606.006] uvmcmcfit: Parametric models to interferometric data fitter

Uvmcmcfit fits parametric models to interferometric data. It is ideally suited to extract the maximum amount of information from marginally resolved observations with interferometers like the Atacama Large Millimeter Array (ALMA), Submillimeter Array (SMA), and Plateau de Bure Interferometer (PdBI). uvmcmcfit uses emcee (ascl:1303.002) to do Markov Chain Monte Carlo (MCMC) and can measure the goodness of fit from visibilities rather than deconvolved images, an advantage when there is strong gravitational lensing and in other situations. uvmcmcfit includes a pure-Python adaptation of Miriad’s (ascl:1106.007) uvmodel task to generate simulated visibilities given observed visibilities and a model image and a simple ray-tracing routine that allows it to account for both strongly lensed systems (where multiple images of the lensed galaxy are detected) and weakly lensed systems (where only a single image of the lensed galaxy is detected).

[ascl:2208.014] uvcombine: Combine images with different resolutions

uvcombine combines single-dish and interferometric data. It can combine high-resolution images that are missing large angular scales (Fourier-domain short-spacings) with low-resolution images containing the short/zero spacing. uvcombine includes the "feathering" technique for interferometry data, implementing a similar approach to CASA’s (ascl:1107.013) feather task but with additional options. Also included are consistency tests for the flux calibration and single-dish scale by comparing the data in the uv-overlap range.

[ascl:1412.003] UTM: Universal Transit Modeller

The Universal Transit Modeller (UTM) is a light-curve simulator for all kinds of transiting or eclipsing configurations between arbitrary numbers of several types of objects, which may be stars, planets, planetary moons, and planetary rings. A separate fitting program, UFIT (Universal Fitter) is part of the UTM distribution and may be used to derive best fits to light-curves for any set of continuously variable parameters. UTM/UFIT is written in IDL code and its source is released in the public domain under the GNU General Public License.

[ascl:1411.012] util_2comp: Planck-based two-component dust model utilities

The util_2comp software utilities generate predictions of far-infrared Galactic dust emission and reddening based on a two-component dust emission model fit to Planck HFI, DIRBE and IRAS data from 100 GHz to 3000 GHz. These predictions and the associated dust temperature map have angular resolution of 6.1 arcminutes and are available over the entire sky. Implementations in IDL and Python are included.

[ascl:2209.012] URILIGHT: Time-dependent Monte-Carlo radiative-transfer

The time dependent Monte-Carlo code URILIGHT, written in Fortran 90, assumes homologous expansion. Energy deposition resulting from the decay of radioactive isotopes is calculated by a Monte-Carlo solution of the γ-ray transport, for which interaction with matter is included through Compton scattering and photoelectric absorption. The temperature is iteratively solved for in each cell by requiring that the total emissivity equals the total absorbed energy.

[ascl:2403.013] URecon: Reconstruct initial conditions of N-Body simulations

URecon reconstructs the initial conditions of N-body simulations from late time (e.g., z=0) density fields. This simple UNET architecture is implemented in TensorFlow and requires Pylians3 (ascl:2403.012) for measuring power spectrum of density fields. The package includes weights trained on Quijote fiducial cosmology simulations.

[ascl:1412.009] URCHIN: Reverse ray tracer

URCHIN is a Smoothed Particle Hydrodynamics (SPH) reverse ray tracer (i.e. from particles to sources). It calculates the amount of shielding from a uniform background that each particle experiences. Preservation of the adaptive density field resolution present in many gas dynamics codes and uniform sampling of gas resolution elements with rays are two of the benefits of URCHIN; it also offers preservation of Galilean invariance, high spectral resolution, and preservation of the standard uniform UV background in optically thin gas.

[ascl:1512.019] UPSILoN: AUtomated Classification of Periodic Variable Stars using MachIne LearNing

UPSILoN (AUtomated Classification of Periodic Variable Stars using MachIne LearNing) classifies periodic variable stars such as Delta Scuti stars, RR Lyraes, Cepheids, Type II Cepheids, eclipsing binaries, and long-period variables (i.e. superclasses), and their subclasses (e.g. RR Lyrae ab, c, d, and e types) using well-sampled light curves from any astronomical time-series surveys in optical bands regardless of their survey-specific characteristics such as color, magnitude, and sampling rate. UPSILoN consists of two parts, one which extracts variability features from a light curve, and another which classifies a light curve, and returns extracted features, a predicted class, and a class probability. In principle, UPSILoN can classify any light curves having arbitrary number of data points, but using light curves with more than ~80 data points provides the best classification quality.

[ascl:1504.001] UPMASK: Unsupervised Photometric Membership Assignment in Stellar Clusters

UPMASK, written in R, performs membership assignment in stellar clusters. It uses photometry and spatial positions, but can take into account other types of data. UPMASK takes into account arbitrary error models; the code is unsupervised, data-driven, physical-model-free and relies on as few assumptions as possible. The approach followed for membership assessment is based on an iterative process, principal component analysis, a clustering algorithm and a kernel density estimation.

[submitted] unWISE-verse: An Integrated WiseView and Zooniverse Data Pipeline

unWISE-verse is an integrated Python pipeline for downloading sets of unWISE time-resolved coadd cutouts from the WiseView image service and uploading subjects to Zooniverse.org for use in astronomical citizen science research. This software was initially designed for the Backyard Worlds: Cool Neighbors research project and is optimized for target sets containing low luminosity brown dwarf candidates. However, unWISE-verse can be applied to other future astronomical research projects that seek to make use of unWISE infrared sky maps, such as studies of infrared variable/transient sources.

[ascl:1901.004] unwise_psf: PSF models for unWISE coadds

The unwise_psf Python module renders point spread function (PSF) models appropriate for use in modeling of unWISE coadd images. unwise_psf translates highly detailed single-exposure WISE PSF models in detector coordinates to the corresponding pixelized PSF models in coadd space, accounting for subtleties including the WISE scan direction and its considerable variation near the ecliptic poles. Applications of the unwise_psf module include performing forced photometry on unWISE coadds, constructing WISE-selected source catalogs based on unWISE coadds and masking unWISE coadd regions contaminated by bright stars.

[ascl:2211.005] unTimely_Catalog_explorer: A search and visualization tool for the unTimely Catalog

unTimely Catalog Explorer searches for and visualizes detections in the unTimely Catalog, a full-sky, time-domain catalog of detections based on WISE and NEOWISE image data acquired between 2010 and 2020. The tool searches the catalog by coordinates to create finder charts for each epoch with overplotted catalog positions and light curves using the unTimely photometry, to overplot these light curves with AllWISE multi-epoch and NEOWISE-R single exposure (L1b) photometry, and to create image blinks with overlaid catalog positions in GIF format.

[ascl:2109.015] unpopular: Using CPM detrending to obtain TESS light curves

unpopular is an implementation of the Causal Pixel Model (CPM) de-trending method to obtain TESS Full-Frame Image (FFI) light curves. The code, written in Python, models the systematics in the light curves of individual pixels as a linear combination of light curves from many other distant pixels and removes shared flux variations. unpopular is able to preserve sector-length astrophysical signals, allowing for the extraction of multi-sector light curves from the FFI data.

[ascl:1110.021] Univiewer: Visualisation Program for HEALPix Maps

Univiewer is a visualisation program for HEALPix maps. It is written in C++ and uses OpenGL and the wxWidgets library for cross-platform portability. Using it you can:

- Rotate and zoom maps on the sphere in 3D
- Create high-resolution views of square patches of the map
- Change maximum and minimum values of the colourmap interactively
- Calculate the power spectrum of the full-sky map or a patch
- Display any column of a HEALPix map FITS file on the sphere

Since Univiewer uses OpenGL for 3D graphics, its performance is dependent your video card. It has been tested successfully on computers with as little as 8Mb video memory, but it is recommended to have at least 32Mb to get good performance.

In the 3D view, a HEALPix map is projected onto a ECP pixelation to create a texture which is wrapped around the sphere. In calculating the power spectrum, the spherical harmonic transforms are computed using the same ECP pixelation. This inevitably leads to some discrepancies at small scales due to repixelation effects, but they are reasonably small.

[ascl:2302.011] UniverseMachine: Empirical model for galaxy formation

The UniverseMachine applies simple empirical models of galaxy formation to dark matter halo merger trees. For each model, it generates an entire mock universe, which it then observes in the same way as the real Universe to calculate a likelihood function. It includes an advanced MCMC algorithm to explore the allowed parameter space of empirical models that are consistent with observations.

[ascl:1503.007] UniPOPS: Unified data reduction suite

UniPOPS, a suite of programs and utilities developed at the National Radio Astronomy Observatory (NRAO), reduced data from the observatory's single-dish telescopes: the Tucson 12-m, the Green Bank 140-ft, and archived data from the Green Bank 300-ft. The primary reduction programs, 'line' (for spectral-line reduction) and 'condar' (for continuum reduction), used the People-Oriented Parsing Service (POPS) as the command line interpreter. UniPOPS unified previous analysis packages and provided new capabilities; development of UniPOPS continued within the NRAO until 2004 when the 12-m was turned over to the Arizona Radio Observatory (ARO). The submitted code is version 3.5 from 2004, the last supported by the NRAO.

[ascl:2111.014] UniMAP: Unicorn Multi-window Anomaly Detection Pipeline

The data analysis UniMAP (Unicorn Multi-window Anomaly Detection Pipeline) leverages the Temporal Outlier Factor (TOF) method to find anomalies in LVC data. The pipeline requires a target detector and a start and stop GPS time describing a time interval to analyze, and has three outputs: 1.) an array of GPS times corresponding to TOF detections; 2.) a long q-transform of the entire data interval with visualizations of the TOF detections in the time series; and 3.) q-transforms of the data windows that triggered TOF detections.

[ascl:1804.022] UniDAM: Unified tool to estimate Distances, Ages, and Masses

UniDAM obtains a homogenized set of stellar parameters from spectrophotometric data of different surveys. Parallax and extinction data can be incorporated into the isochrone fitting method used in UniDAM to reduce distance and age estimate uncertainties for TGAS stars for distances up to 1 kpc and decrease distance Gaia end-of-mission parallax uncertainties by about a factor of 20 and age uncertainties by a factor of two for stars up to 10 kpc away from the Sun.

[submitted] UMIST

Astrochemistry database of chemical species.

[ascl:2008.006] Umbrella: Asteroid detection, validation, and identification

Umbrella detects, validates, and identifies asteroids. The core of this software suite, Umbrella2, includes algorithms and interfaces for all steps of the processing pipeline, including a novel detection algorithm for faint trails. A detection pipeline accessible as a desktop program (ViaNearby) builds on the library to provide near real-time data reduction of asteroid surveys on the Wide Field Camera of the Isaac Newton Telescope. Umbrella can read and write MPC optical reports, supports SkyBoT and VizieR querying, and can be extended by user image processing functions to take advantage of the algorithms framework as a multi-threaded CPU scheduler for easy algorithm parallelization.

[ascl:1104.007] ULySS: A Full Spectrum Fitting Package

ULySS (University of Lyon Spectroscopic Analysis Software) is an open-source software package written in the GDL/IDL language to analyze astronomical data. ULySS fits a spectrum with a linear combination of non-linear components convolved with a line-of-sight velocity distribution (LOSVD) and multiplied by a polynomial continuum. ULySS is used to study stellar populations of galaxies and star clusters and atmospheric parameters of stars.

[submitted] Ulula: a lightweight 2D hydro code for teaching

Ulula is an ultra-lightweight 2D hydro code for teaching purposes. The code is written in pure python and is designed to be as short and easy to understand as possible, while not compromising on performance. The latter is achieved with a simple Godunov solver and by using numpy for all array operations.

Would you like to view a random code?