Font Size: a A A

Development Of Recombinant Mycobacteirum Bovis BCG Of Toxoplasma Gondii Rhomboid Gene

Posted on:2013-09-23Degree:MasterType:Thesis
Country:ChinaCandidate:Q ZhangFull Text:PDF
GTID:2248330371483416Subject:Veterinarians
Abstract/Summary:PDF Full Text Request
With the rapid development of Internet search engine can be described as inevitable.Internet is like a giant library in the network library and at all times with a lot of information.Tens of thousands of information far beyond our imagination and control, if not theemergence of search engines, maybe we could not find what we want to target.Web data capture refers to the bulk, quickly extract information from the site of acomputer software technology. Web data capture program to simulate the behavior of thebrowser, any data extracted from the program can be displayed in the browser, the ultimategoal of Web data crawled unstructured information extracted from a large number of pages ina structured stored.The following shortcomings of the traditional search technology make it difficult to meetthe needs of users:First of all, this technology is demanding for the choice of keywords, if improperlyselected keywords, so that restricted non-mature users use search engines. Second, the resultsof this search engine can be displayed on the results page is also very limited, the results of asingle, usually filled with redundant information. The reason for this result because thistechnique is a simple one-dimensional keyword-based query, the search engine does not takethe initiative to "understanding document, only a passive keyword matching. This technique aresult, users often do not have access to valuable information. This situation is particularlyevident in the timeliness, and information structured job.The information redundancy of the Internet is too large. The article is reproducedhundreds of times. Although the current technology in terms of recognition technology, butstill significant weakness.The simple point of the vertical search is relative to the general search engine forindustry-specific expert search engine, professional web repository of information fromrefinement, integration, classify, extract specific data returned to the client, the crawl isstructured data and metadata, which is the biggest difference between and the existence ofuniversal search, usually by three parts in the crawl system, index and search.With the development of vertical search engines and the problems faced in the development of this thesis, a theoretical analysis of the key technologies of the vertical searchsystem, the specific classification of vertical search engines and related knowledge. Networkoperating rules of the spider design, the framework of the educational information system ofvertical search engines, analysis of the role of the function of each module, given thearchitecture of the education information system of vertical search engines and build thesystem processes, detailed information involved in the framework of the educationalinformation system of vertical search engines to crawl, the realization of Chinese extractionand retrieval functions.Carry out the design to the management module, the page capture, data processing andindexing, and constructed perpendicular to the field of education, information searchframework. The system architecture, the set of processes, from the overall structure of thefront and back, respectively, indicate the process. Finally, UML use case analysis.
Keywords/Search Tags:Vertical search engine, Lucene, vertical crawling, clustering
PDF Full Text Request
Related items