"""
Module for adding the IRTF Spectral Library to the database.
"""
import os
import tarfile
import urllib.request
import numpy as np
import pandas as pd
from astropy.io import fits
from species.util import data_util, query_util
[docs]def add_irtf(input_path,
database,
sptypes):
"""
Function for adding the IRTF Spectral Library to the database.
Parameters
----------
input_path : str
Path of the data folder.
database : h5py._hl.files.File
Database.
sptypes : tuple(str, )
Spectral types ('F', 'G', 'K', 'M', 'L', 'T').
Returns
-------
NoneType
None
"""
distance_url = 'https://people.phys.ethz.ch/~stolkert/species/distance.dat'
distance_file = os.path.join(input_path, 'distance.dat')
if not os.path.isfile(distance_file):
urllib.request.urlretrieve(distance_url, distance_file)
distance_data = pd.pandas.read_csv(distance_file,
usecols=[0, 3, 4],
names=['object', 'distance', 'distance_error'],
delimiter=',',
dtype={'object': str,
'distance': float,
'distance_error': float})
datadir = os.path.join(input_path, 'irtf')
if not os.path.exists(datadir):
os.makedirs(datadir)
data_file = {'F': os.path.join(input_path, 'irtf/F_fits_091201.tar'),
'G': os.path.join(input_path, 'irtf/G_fits_091201.tar'),
'K': os.path.join(input_path, 'irtf/K_fits_091201.tar'),
'M': os.path.join(input_path, 'irtf/M_fits_091201.tar'),
'L': os.path.join(input_path, 'irtf/L_fits_091201.tar'),
'T': os.path.join(input_path, 'irtf/T_fits_091201.tar')}
data_folder = {'F': os.path.join(input_path, 'irtf/F_fits_091201'),
'G': os.path.join(input_path, 'irtf/G_fits_091201'),
'K': os.path.join(input_path, 'irtf/K_fits_091201'),
'M': os.path.join(input_path, 'irtf/M_fits_091201'),
'L': os.path.join(input_path, 'irtf/L_fits_091201'),
'T': os.path.join(input_path, 'irtf/T_fits_091201')}
data_type = {'F': 'F stars (4.4 MB)',
'G': 'G stars (5.6 MB)',
'K': 'K stars (5.5 MB)',
'M': 'M stars (7.5 MB)',
'L': 'L dwarfs (850 kB)',
'T': 'T dwarfs (100 kB)'}
url_root = 'http://irtfweb.ifa.hawaii.edu/~spex/IRTF_Spectral_Library/Data/'
url = {'F': url_root+'F_fits_091201.tar',
'G': url_root+'G_fits_091201.tar',
'K': url_root+'K_fits_091201.tar',
'M': url_root+'M_fits_091201.tar',
'L': url_root+'L_fits_091201.tar',
'T': url_root+'T_fits_091201.tar'}
for item in sptypes:
if not os.path.isfile(data_file[item]):
print(f'Downloading IRTF Spectral Library - {data_type[item]}...', end='', flush=True)
urllib.request.urlretrieve(url[item], data_file[item])
print(' [DONE]')
print('Unpacking IRTF Spectral Library...', end='', flush=True)
for item in sptypes:
tar = tarfile.open(data_file[item])
tar.extractall(path=datadir)
tar.close()
print(' [DONE]')
database.create_group('spectra/irtf')
for item in sptypes:
for root, _, files in os.walk(data_folder[item]):
for _, filename in enumerate(files):
if filename[-9:] != '_ext.fits':
fitsfile = os.path.join(root, filename)
spdata, header = fits.getdata(fitsfile, header=True)
name = header['OBJECT']
sptype = header['SPTYPE']
if name[-2:] == 'AB':
name = name[:-2]
elif name[-3:] == 'ABC':
name = name[:-3]
spt_split = sptype.split()
if item in ['L', 'T'] or spt_split[1][0] == 'V':
print_message = f'Adding IRTF Spectral Library... {name}'
print(f'\r{print_message:<70}', end='')
simbad_id = query_util.get_simbad(name)
if simbad_id is not None:
simbad_id = simbad_id.decode('utf-8')
dist_select = distance_data.loc[distance_data['object'] == simbad_id]
if not dist_select.empty:
distance = (dist_select['distance'], dist_select['distance_error'])
else:
simbad_id, distance = query_util.get_distance(name)
else:
distance = (np.nan, np.nan)
sptype = data_util.update_sptype(np.array([sptype]))[0]
dset = database.create_dataset(f'spectra/irtf/{name}',
data=spdata)
dset.attrs['name'] = str(name).encode()
dset.attrs['sptype'] = str(sptype).encode()
dset.attrs['simbad'] = str(simbad_id).encode()
dset.attrs['distance'] = distance[0]
dset.attrs['distance_error'] = distance[1]
print_message = 'Adding IRTF Spectral Library... [DONE]'
print(f'\r{print_message:<70}')
database.close()