You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- 16 GB of RAM is needed to build the provided influenza kraken database
134
-
135
-
# Usage
136
-
137
-
## Building example
138
-
139
-
This pipeline is built from a central set of scripts located in the `v1` directory
140
-
141
-
Build flu-kraken example with:
142
-
143
-
`build_flukraken.sh -k flukraken-$(date +"%F")`
144
-
145
-
The single script `build_flukraken.sh` functions as an outer wrapper for the influenza classification example using the Kraken classifier published in the mytax paper.
146
-
147
-
148
-
`build_flukraken.sh` can also be used as a model to build modified pipelines as desired. It is built from four main sub-modules:
149
-
```
150
-
download_IVR.sh -> download references and taxonomy from IVR
151
-
152
-
build_IVR_metadata.sh -> build tab-delimited metadata table in format for mytax
153
-
154
-
build_taxonomy.sh -> build custom taxonomy from tab-delimited table
155
-
156
-
build_krakendb.sh -> add new taxonomic IDs to reference FASTA, build kraken database, post-process database for visualization pipeline
157
-
```
158
-
159
-
`build_krakendb.sh` currently references three helper scripts, which also need to be in the PATH:
160
-
```
161
-
fix_references.sh -> adds new taxonomic IDs to reference FASTA
162
-
163
-
kraken-build -> builds kraken database
164
-
165
-
process_krakendb.sh -> post-processes database for visualization pipeline (not included in this repo yet)
166
-
```
167
-
168
-
169
-
170
-
## Running process script on kraken/kraken2 report and outfiles
171
-
172
-
### If running from Docker
173
-
174
-
docker build . -t jhuaplbio/mytax
175
-
176
-
Unix
177
-
178
-
`docker container run -it --rm -v $PWD:/data jhuaplbio/mytax bash`
179
-
180
-
Windows Powershell
181
-
182
-
`docker container run -it --rm -v $pwd:/data jhuaplbio/mytax bash`
183
-
184
-
185
-
186
-
## Run the installation script
187
-
188
-
189
-
# Activate the env, this will contain kraken2 and centrifuge scripts to build the database if needed as well as kraken2 and centrifuge dependencies
tar -xvzf databases/centrifuge.tgz --directory databases/centrifuge/
248
-
```
249
-
250
-
251
-
252
-
#### Run Centrifuge classify
253
-
254
-
```
255
-
## If you need to make a new database, see here: $CONDA_PREFIX/lib/centrifuge/centrifuge-build --taxonomy-tree taxonomy/nodes.dmp --name-table taxonomy/names.dmp sample.fastq sample
The resulting file can then imported into the sunburst plot at `server/src/sunburst/index.html` rendered with a simple `http.server` protocol like `python3 -m http.server 8080`
0 commit comments