Jan 27, 2010

A Web Crawler for Automated Location of Genomic Resources

The purpose of this project is to develop a Web Crawler based software package that will locate and then download genomic data. A Web Crawler is a software package that searches the internet for particular strings of interest.The Web Crawler being developed here works by seeking out files in a set of webpages and then downloading them if they are of interest to the user. Whether the files are of interest to the user depends on the application. In this case the files of interest are Genomic Resources.

Author:-Steven James Mayocchi

Source:- The University of Queensland


Tags: , , ,

0 Responses to “A Web Crawler for Automated Location of Genomic Resources”

Subscribe Now!! final year project subscribers

Get the latest projects on your Email, 27k+ user's can't be wrong !!

Note: We promise, we don't SPAM, so don't forget to verify your Email and get FREE projects.
© 2015 FREE FINAL YEAR PROJECT'S. All rights reserved.
Designed by SpicyTricks