![]() |
|
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
Pooling TruSeq (single index) and Nextera XT (dual index) in one MiSeq run? | Carosmile | Sample Prep / Library Generation | 1 | 05-16-2017 01:13 PM |
Metrix - A Server / client parser for Illumina (InterOp) run directories | Rhizosis | Bioinformatics | 29 | 10-15-2014 04:06 AM |
wrong run parameters. How can I change it after 1 cycle? | stasi | Illumina/Solexa | 4 | 10-13-2014 12:32 AM |
454 run - wrong primer - can it be salvaged? | Indiana | 454 Pyrosequencing | 7 | 10-31-2013 04:24 PM |
Aborted BWA run - what is wrong? | yaximik | Bioinformatics | 1 | 06-10-2013 02:52 AM |
![]() |
|
Thread Tools |
![]() |
#1 |
Member
Location: Stillwater Join Date: Dec 2009
Posts: 45
|
![]()
We had a client send us a set of 43 multiplexed samples. Although the NextSeq500 run went well, only 54% of the reads passed filter (PF). We normally get in the 93%+ range, so we dug but couldn't find any explanation except that there must have been a bad index (typo) in the sample sheet. Not our error but hey, we want to help.
Some of the samples had essentially zero reads after demultiplexing. We also found discarded reads where over a million had the same index. Question: Can we easily determine which samples need to be re-de-multiplexed? This must have happened before. Thanks, -pete Last edited by hoytpr; 10-24-2017 at 02:49 PM. Reason: added instrument |
![]() |
![]() |
![]() |
#2 |
Senior Member
Location: Melbourne Join Date: Jan 2013
Posts: 1,137
|
![]()
Easiest way will be to ask client the kit they have used and re-demultiplex with a new sample sheet listing all indices available for the kit. Also, if it is a duel index it might be that index2 sequences has been entered in incorrect order (might require reverse complement of index 2 if has not been done already).
|
![]() |
![]() |
![]() |
#3 | |
Senior Member
Location: East Coast USA Join Date: Feb 2008
Posts: 6,697
|
![]() Quote:
|
|
![]() |
![]() |
![]() |
#4 |
Member
Location: Stillwater Join Date: Dec 2009
Posts: 45
|
![]()
Thanks, seems so logical in retrospect. I appreciate the help and will set it up today.
There are likely some failed samples also, according to the client. But hopefully we can get most of the 43 samples straightened out. -pete |
![]() |
![]() |
![]() |
#5 |
Member
Location: Aarhus, Denmark Join Date: Mar 2008
Posts: 40
|
![]()
See the file /Reports/html/index.html in the run folder. It includes a list of top10 unknown barcodes (and the known ones). Click on "show barcodes" in the top-right corner when you have opened the html file.
/Jakob |
![]() |
![]() |
![]() |
#6 | |
Member
Location: Stillwater Join Date: Dec 2009
Posts: 45
|
![]() Quote:
At least three and probably five of the samples have no or VERY few reads. Two might just be bad libraries. Unfortunately based on the percent of each base (A,G,C,T) at each of the twelve base read positions in the 43 index sequences, it looks like the sample sheet was mixed up in several places, and I can't substitute these unknown index reads into the index reads %A, %C, %G, %T to correct for the differences. It was a long shot but I can't figure out anything else to do. I hope they can figure out which are their samples from the assemblies. -pete |
|
![]() |
![]() |
![]() |
#7 |
Member
Location: Stillwater Join Date: Dec 2009
Posts: 45
|
![]()
FYI: I wanted to post back about an error we got running 384 samples. We'd done 224 before, but going over 250 might give you an error like:
ERROR: bcl2fastq::common::Exception: 2017-Oct-25 17:44:17: Too many open files (24): The solution is found here: https://erikclarke.net/2016/03/31/op...and-bcl2fastq/ -pete |
![]() |
![]() |
![]() |
#8 |
Senior Member
Location: East Coast USA Join Date: Feb 2008
Posts: 6,697
|
![]()
@pete: Out of curiosity is this an error purely on client's part? Is it only restricted to having wrong entries in samplesheet? Were samples pooled from multiple submitters (I don't understand what you wrote in post #6)?
Last edited by GenoMax; 10-26-2017 at 08:35 AM. |
![]() |
![]() |
![]() |
#9 | |
Member
Location: Stillwater Join Date: Dec 2009
Posts: 45
|
![]() Quote:
As a shot in the dark, I looked in the SAV software, looked at all PF reads, but limited the output to the 12 reads (BIOO indecies) of "Read2" (the single index read cycle). The percent of As, Cs, Gs, and Ts for each read 146 through 157 (the twelve index reads) is shown. With 43 indecies, if the percentage of "A" at read 146 was for example 17.5%, then it suggests there were 8/43 "A". Didn't try to correct for phasing, but this was just me trying to learn. The same was done for C, G, T, and got 5 Cs, 11 Gs and 19 Ts, so in read 146 there were the expected 43 bases. Then I followed this through the rest of the reads, and ended up with a 4X12 matrix of how the As, Cs, Gs, and Ts should be distributed in the index reads 146 - 157 (worked surprisingly well). It doesn't tell me exactly which indecies were found when the machine was running, but it gives a base distribution approximation. Then I tried to match this up with the index sequences as given from the Sample Sheet. I calculated the same 4x12 matrix distribution. They didn't match up very well suggesting (to me, and old molecular geneticist) the indecies in the sample sheet and the indecies in the run were different. Then I took the top 25 "unknown" index sequences, and tried substituting them into the 4X12 matrix to see if they could get the Sample sheet index matrix to match up what the run data matrix. No joy. This probably sounds stupid considering how phasing could have screwed up things, and it still wouldn't have told me which index went to which sample, but it might have determined which of the "unknown" indecies and associated reads were the correct ones. Like I said, just trying to help. If it's still not clear I can send you a spreadsheet. ![]() -pete |
|
![]() |
![]() |
![]() |
#10 | |
Member
Location: Stillwater Join Date: Dec 2009
Posts: 45
|
![]() Quote:
-pete |
|
![]() |
![]() |
![]() |
#11 |
Senior Member
Location: East Coast USA Join Date: Feb 2008
Posts: 6,697
|
![]() |
![]() |
![]() |
![]() |
#12 |
Senior Member
Location: East Coast USA Join Date: Feb 2008
Posts: 6,697
|
![]()
Such is the life of a core facility. At least the submitter can't blame you since they did everything.
I wonder if the two sets of libraries had distinct insert sizes and one set competed the other one out (people sometimes think they are being clever and try to save money). As a last ditch effort you could just put Sample_1, Sample_2 against the indices you actually see and demux using that generic scheme. Submitter hopefully has some alternate means of figuring out what is what. |
![]() |
![]() |
![]() |
#13 |
Senior Member
Location: CT Join Date: Apr 2015
Posts: 233
|
![]()
The poor pf and wrong indices are 2 different issues. Bad libraries shouldn't lead to poor pf unless you were loosing all of a certain type of libraries and ended up with just low diversity? Might check your thumbnails to see if you were overclustered (which can still be the users fault if their libraries are much small than expected).
I've attempted to figure out the correct indices when users have told me the wrong ones and generally haven't been able to do it. But I haven't tried that hard.
__________________
Microbial ecologist, running a sequencing core. I have lots of strong opinions on how to survey communities, pretty sure some are even correct. |
![]() |
![]() |
![]() |
#14 |
Member
Location: Stillwater Join Date: Dec 2009
Posts: 45
|
![]()
Thanks. Yes, I learned a lesson today. The run was only slightly overclustered and the 384-index run had 90% PF (all the bad PF were labeled default or unknown). Our PhiX was only 1.2%.
After the re-analysis with all 384 indecies, the 43 highest numbers of indexed sequences included 7 barcodes not on the sample sheet. There wasn't an obvious falloff of sample read numbers for another 6-7 indecies down the list. It's a mess. Note to group: For reference, with 174,615,251 clusters PF, (if you factor in that approximately 43 indecies were supposed to be there), we had a total of 240,038 "bad" clusters or ~0.14 percent. -pete |
![]() |
![]() |
![]() |
Thread Tools | |
|
|