Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
28 commits
Select commit Hold shift + click to select a range
547b4c1
Update link to Abyss
mvdbeek Nov 20, 2019
58bb1c5
Fix kobas test output files
mvdbeek Nov 20, 2019
359234d
Fix more links
mvdbeek Nov 20, 2019
9696f86
Fix obitools homepage
mvdbeek Nov 20, 2019
bfbf8ef
Fix data_manager_mothur on python3
mvdbeek Nov 21, 2019
4ec6fce
Add python3.7 dependency for sam_fasta_index_builder.xml
mvdbeek Nov 21, 2019
6410445
Add python3.7 to data_manager_twobit_builder
mvdbeek Nov 21, 2019
7b8ccec
Fix data_manager_rsync_g2 on python3 and tests
mvdbeek Nov 21, 2019
2eed9fb
Fix salmon data manager profile tag
mvdbeek Nov 22, 2019
bd60c32
Enable dependency resolution for twobit_build data manager
mvdbeek Nov 27, 2019
1539770
Fix data_manager_snpsift_dbnsfp test and requirements
mvdbeek Nov 27, 2019
38c4817
Fix rgrnastar requirements and test
mvdbeek Nov 22, 2019
7e319df
Add exact requirements to fetch_refseq.xml
mvdbeek Nov 27, 2019
f366654
bam2bam: Fix linting error in tools/bax2bam/bax2bam.xml
mvdbeek Nov 23, 2019
f951940
Bump data_manager_refseq version
mvdbeek Nov 27, 2019
5397ad9
bump data_manager_fetch_mothur version
mvdbeek Nov 27, 2019
d662320
Bump data_manager_rsync version
mvdbeek Nov 27, 2019
ff9ce14
Bump salmon_index_builder version
mvdbeek Nov 27, 2019
e18d5f4
Bump data_mananger_sam_fasta_index_builder version
mvdbeek Nov 27, 2019
03e0242
Bump data_manager_snpsift_dbnsfp version
mvdbeek Nov 27, 2019
b2d9ff2
Bump rna_star_index_builder version
mvdbeek Nov 27, 2019
1b9fa6b
Bump twobit_builder version
mvdbeek Nov 27, 2019
305a4b2
Fix fetch_refseq test
mvdbeek Nov 27, 2019
e672fa2
Use better obitools URL
mvdbeek Nov 27, 2019
d553db7
autopep8, drop list casting in data_manager_rsync.py
mvdbeek Nov 27, 2019
b782a8d
Fix picrust test output
mvdbeek Nov 27, 2019
b40a6b9
Fix kobas_annotate tool
nsoranzo Nov 28, 2019
2030b8a
Cleanup kobas
nsoranzo Nov 28, 2019
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
@@ -1,11 +1,11 @@
<tool id="data_manager_fetch_refseq" name="RefSeq data manager" version="0.0.18" tool_type="manage_data">
<tool id="data_manager_fetch_refseq" name="RefSeq data manager" version="0.0.19" tool_type="manage_data" profile="19.05">
<description>Fetch FASTA data from NCBI RefSeq and update all_fasta data table</description>
<requirements>
<requirement type="package">python</requirement>
<requirement type="package">requests</requirement>
<requirement type="package" version="3.7">python</requirement>
<requirement type="package" version="2.22.0">requests</requirement>
</requirements>
<command detect_errors="exit_code"><![CDATA[
python $__tool_directory__/fetch_refseq.py
python '$__tool_directory__/fetch_refseq.py'
#if str( $advanced.advanced_selector ) == 'advanced':
${advanced.compress}
#end if
Expand Down Expand Up @@ -65,7 +65,6 @@
<assert_contents>
<has_text text="2018-03-14"/>
<has_text text="refseq_plastid"/>
<has_text text="/refseq_plastid."/>
</assert_contents>
</output>
</test>
Expand All @@ -79,7 +78,6 @@
<assert_contents>
<has_text text="2018-03-14"/>
<has_text text="refseq_plastid"/>
<has_text text="/refseq_plastid."/>
</assert_contents>
</output>
</test>
Expand Down
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
<?xml version="1.0"?>
<tool id="data_manager_fetch_mothur_reference_data" name="Fetch Mothur toolsuite reference data" version="0.1.3" tool_type="manage_data">
<tool id="data_manager_fetch_mothur_reference_data" name="Fetch Mothur toolsuite reference data" version="0.1.4" tool_type="manage_data" profile="19.05">
<description>Fetch and install reference data for Mothur</description>
<requirements>
<requirement type="package" version="2.7">python</requirement>
Expand Down Expand Up @@ -66,7 +66,7 @@
</outputs>
<tests>
<test>
<param name="data_sourece|ref_data" value="lookup_titanium"/>
<param name="data_source|ref_data" value="lookup_titanium"/>
<output name="out_file">
<assert_contents>
<has_text text="GS FLX Titanium" />
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -228,12 +228,12 @@ def download_file(url, target=None, wd=None):
Returns the name that the file is saved with.

"""
print "Downloading %s" % url
print("Downloading %s" % url)
if not target:
target = os.path.basename(url)
if wd:
target = os.path.join(wd, target)
print "Saving to %s" % target
print("Saving to %s" % target)
open(target, 'wb').write(urllib2.urlopen(url).read())
return target

Expand All @@ -254,35 +254,35 @@ def unpack_zip_archive(filen, wd=None):

"""
if not zipfile.is_zipfile(filen):
print "%s: not ZIP formatted file"
print("%s: not ZIP formatted file")
return [filen]
file_list = []
z = zipfile.ZipFile(filen)
for name in z.namelist():
if reduce(lambda x, y: x or name.startswith(y), IGNORE_PATHS, False):
print "Ignoring %s" % name
print("Ignoring %s" % name)
continue
if wd:
target = os.path.join(wd, name)
else:
target = name
if name.endswith('/'):
# Make directory
print "Creating dir %s" % target
print("Creating dir %s" % target)
try:
os.makedirs(target)
except OSError:
pass
else:
# Extract file
print "Extracting %s" % name
print("Extracting %s" % name)
try:
os.makedirs(os.path.dirname(target))
except OSError:
pass
open(target, 'wb').write(z.read(name))
file_list.append(target)
print "Removing %s" % filen
print("Removing %s" % filen)
os.remove(filen)
return file_list

Expand All @@ -305,23 +305,23 @@ def unpack_tar_archive(filen, wd=None):
"""
file_list = []
if not tarfile.is_tarfile(filen):
print "%s: not TAR file"
print("%s: not TAR file")
return [filen]
t = tarfile.open(filen)
for name in t.getnames():
# Check for unwanted files
if reduce(lambda x, y: x or name.startswith(y), IGNORE_PATHS, False):
print "Ignoring %s" % name
print("Ignoring %s" % name)
continue
# Extract file
print "Extracting %s" % name
print("Extracting %s" % name)
t.extract(name, wd)
if wd:
target = os.path.join(wd, name)
else:
target = name
file_list.append(target)
print "Removing %s" % filen
print("Removing %s" % filen)
os.remove(filen)
return file_list

Expand All @@ -339,9 +339,9 @@ def unpack_archive(filen, wd=None):
current working directory.

"""
print "Unpack %s" % filen
print("Unpack %s" % filen)
ext = os.path.splitext(filen)[1]
print "Extension: %s" % ext
print("Extension: %s" % ext)
if ext == ".zip":
return unpack_zip_archive(filen, wd=wd)
elif ext == ".tgz":
Expand Down Expand Up @@ -382,7 +382,7 @@ def identify_type(filen):
try:
return MOTHUR_FILE_TYPES[ext]
except KeyError:
print "WARNING: unknown file type for " + filen + ", skipping"
print("WARNING: unknown file type for " + filen + ", skipping")
return None


Expand Down Expand Up @@ -415,26 +415,26 @@ def fetch_from_mothur_website(data_tables, target_dir, datasets):
"""
# Make working dir
wd = tempfile.mkdtemp(suffix=".mothur", dir=os.getcwd())
print "Working dir %s" % wd
print("Working dir %s" % wd)
# Iterate over all requested reference data URLs
for dataset in datasets:
print "Handling dataset '%s'" % dataset
print("Handling dataset '%s'" % dataset)
for name in MOTHUR_REFERENCE_DATA[dataset]:
for f in fetch_files(MOTHUR_REFERENCE_DATA[dataset][name], wd=wd):
type_ = identify_type(f)
entry_name = "%s (%s)" % (os.path.splitext(os.path.basename(f))[0], name)
print "%s\t\'%s'\t.../%s" % (type_, entry_name, os.path.basename(f))
print("%s\t\'%s'\t.../%s" % (type_, entry_name, os.path.basename(f)))
if type_ is not None:
# Move to target dir
ref_data_file = os.path.basename(f)
f1 = os.path.join(target_dir, ref_data_file)
print "Moving %s to %s" % (f, f1)
print("Moving %s to %s" % (f, f1))
os.rename(f, f1)
# Add entry to data table
table_name = "mothur_%s" % type_
add_data_table_entry(data_tables, table_name, dict(name=entry_name, value=ref_data_file))
# Remove working dir
print "Removing %s" % wd
print("Removing %s" % wd)
shutil.rmtree(wd)


Expand All @@ -450,7 +450,7 @@ def files_from_filesystem_paths(paths):
files = []
for path in paths:
path = os.path.abspath(path)
print "Examining '%s'..." % path
print("Examining '%s'..." % path)
if os.path.isfile(path):
# Store full path for file
files.append(path)
Expand All @@ -459,7 +459,7 @@ def files_from_filesystem_paths(paths):
for f in os.listdir(path):
files.extend(files_from_filesystem_paths((os.path.join(path, f), )))
else:
print "Not a file or directory, ignored"
print("Not a file or directory, ignored")
return files


Expand Down Expand Up @@ -489,14 +489,14 @@ def import_from_server(data_tables, target_dir, paths, description, link_to_data
for f in files:
type_ = identify_type(f)
if type_ is None:
print "%s: unrecognised type, skipped" % f
print("%s: unrecognised type, skipped" % f)
continue
ref_data_file = os.path.basename(f)
target_file = os.path.join(target_dir, ref_data_file)
entry_name = "%s" % os.path.splitext(ref_data_file)[0]
if description:
entry_name += " (%s)" % description
print "%s\t\'%s'\t.../%s" % (type_, entry_name, ref_data_file)
print("%s\t\'%s'\t.../%s" % (type_, entry_name, ref_data_file))
# Link to or copy the data
if link_to_data:
os.symlink(f, target_file)
Expand All @@ -508,7 +508,7 @@ def import_from_server(data_tables, target_dir, paths, description, link_to_data


if __name__ == "__main__":
print "Starting..."
print("Starting...")

# Read command line
parser = optparse.OptionParser()
Expand All @@ -518,8 +518,8 @@ def import_from_server(data_tables, target_dir, paths, description, link_to_data
parser.add_option('--description', action='store', dest='description', default='')
parser.add_option('--link', action='store_true', dest='link_to_data')
options, args = parser.parse_args()
print "options: %s" % options
print "args : %s" % args
print("options: %s" % options)
print("args : %s" % args)

# Check for JSON file
if len(args) != 1:
Expand All @@ -532,7 +532,7 @@ def import_from_server(data_tables, target_dir, paths, description, link_to_data
params, target_dir = read_input_json(jsonfile)

# Make the target directory
print "Making %s" % target_dir
print("Making %s" % target_dir)
os.mkdir(target_dir)

# Set up data tables dictionary
Expand All @@ -554,7 +554,7 @@ def import_from_server(data_tables, target_dir, paths, description, link_to_data
paths = options.paths.replace('__cn__', '\n').replace('__cr__', '\r').split()
import_from_server(data_tables, target_dir, paths, description, link_to_data=options.link_to_data)
# Write output JSON
print "Outputting JSON"
print str(json.dumps(data_tables))
open(jsonfile, 'wb').write(json.dumps(data_tables))
print "Done."
print("Outputting JSON")
print(json.dumps(data_tables))
open(jsonfile, 'w').write(json.dumps(data_tables, sort_keys=True))
print("Done.")
Loading