|Thread||Thread Starter||Forum||Replies||Last Post|
|Which Computer to Buy?||polsum||Bioinformatics||7||08-04-2009 10:48 PM|
|In Sequence: With 15 GB on GA II and Avantome Buy, Illumina Confident in Next-Gen Bat||Newsbot!||Illumina/Solexa||0||07-29-2008 01:30 PM|
|08-27-2008, 08:42 PM||#1|
Join Date: Aug 2008
NGS platform: To buy or not to buy?
This is a question I have been hearing lately. Malaysia has not own any NGS platform yet but it's coming (I know at least 4) and probably will arrive by end of this year. When I ask which platform, they say they're not sure (or they are not sure if they wanna let me know).
I was told that many of the NGS machines were owned by small labs rather than big genome center. My lab used to have plans to buy a 454 machine too. After doing all the mathematics and economics, we decided not to.
Yes, it's so much cheaper when you run sequencing on your own. And you can do lotsa things like barcoding your samples up to 4000+ with 6nt barcode. (If you have a software that can sort so many different samples).
There are limitations if you use the service provided by others (I should know better)
Just found an article in Genome Technology published in april 08.
"Is Next-Gen Technology Right for Your lab?"
Some of the things you might wanna consider before you purchase a machine:
The data storage will be more expensive than generating the data. Another problem is slow data transfer rate.
Extra funding like data storage upgrade, technical support and machine storage environment etc should be taken into consideration.
They found out that Solexa can perform better in a cold environment.
3) Labour (Hire someone with experience)
4) Test run the machine (Optimization could be weeks or months)
5) How long it takes before the machine is being replaced by a new technology?
Usually 5 years. Anything beyond that will be called LGS (last generation sequencing) 3rd Generation Sequencing will be here in 2010.
Then again, please correct me if I'm wrong.
|12-31-2008, 10:08 AM||#3|
Join Date: May 2008
So I don't understand where this comes from. $80 on computing for a $4K reagent run seems fine to me. More to the point many people record a 5-15% run failure rate (all reasons), if you amortize that over the life-time of the instrument it is costing you FAR more than the computing (about 6X more in fact). So it makes more sense to worry about run failure causes.
Maybe it comes from the idea that you need to keep all the data, and at around $500 per terabyte that ends up being comparable to the run costs in terms of reagents and chips ? well, you dont need to keep raw data beyond your chosen QC period; and the data that is left boils down to a few 10's gigabytes - so again around $50 to store. Will that data even be kept for 3 years ?
Data transfer - another myth. Data transfer is as fast as your network - so gigabit I'd hope. If not please upgrade it - this will cost you a few $K at most. Gigabit means around 70-120 Megabytes/s aggregate bandwidth. A system like a GAII produces anywhere up to 4 Megabytes per second of raw data (+/- 4Mbytes - lots of factors). Anyway, a typical network has plenty of bandwidth to run 1 system - and in principle quite a number. I think this myth arises because may people wait till the end of a run before moving data to where they want to process and look at it - so you add copying time to the end of the run. The solution is obvious and adopted in many places - write the data where you want it over the network in the first place - or even better and more reliable - mirror it in *real time* to where you want it. Thus the entire data set is off the instrument and on your disk within a few minutes of a run finishing. If you do this I do not see how data transfer can be a problem as you can begin the next run immediately. Some labs even process the data as they collect it (but this is tricky to implement yourself - i believe the iPar attempts something similar too).
Sanger does something very like this on ~40 GAs and has produced over 4 (6-8 raw) Terabases of data so far. (I think that compute comes out at about $40K per instrument although it is used also for other heavy compute jobs).
The other points seem very sensible and good advice.
Last edited by cgb; 12-31-2008 at 10:30 AM.
|03-05-2009, 07:15 AM||#4|
Location: 41°17'49"N / 2°4'42"E
Join Date: Oct 2008
That's a huge improvement both in disk storage and cpu cycles. If you want (and you should) you can toss the images and start your analysis from Bustard/basecalling.
Ipar will perform basecalling very soon too.