When you set to ignore URLs case, then WDE convert all URLs to lowercase and can remove duplicate URLs like above. Set this option to avoid duplicate URLs like For example: in above case, if an external site found like then WDE will grab only base It will not visit unless you set such depth that covers also milk dir. With this option you can tell WDE to process always the Base URLs of external sites. If you tell WDE to "Stop Site on First email Found then it will not go for other pages (#3-12) For example: you set WDE to go entire site and WDE found email in #2 URL (support.htm). So sometimes you may prefer to use "Stop Site on First Email Found" option. Some websites may have only few files and some may have thousands of files. Select "Depth=0" and check "Stay within Full URL"Įach website is structured differently on the server. Only matching URL page of search ( URL #6 ) WDE is powerful and fully featured unique spider! You need to decide how deep you want WDE to look for data. A setting of "1" will process index or home page with associated files under root dir only.įor example: WDE is going to visit URL for data extraction. A setting of "0" will process and look for data in whole website. If you want WDE to stay within first page, just select "Process First Page Only". Here you need to tell WDE - how many levels to dig down within the specified website. How many deep it spiders in the matching websites depends on "Depth" setting of "External Site" tab. Next it visits those matching websites for data extraction. WDE send queries to search engines to get matching website URLs. You can add other engine sources as well. Click "Engines" button and uncheck listing that you do not want to use. You can tell WDE how many search engines to use. What WDE Does: WDE will query 18+ popular search engines, extract all matching URLs from search results, remove duplicate URLs and finally visits those websites and extract data from there. Select "Search Engines" source - Enter keyword - Click OK WDE spiders 18+ Search engines for right web sites and get data from them. You can setup different type of extraction with this UNIQUE spider, link extractor: A powerful phone, fax harvester/extraction tools for responsible tel/fax marketing. You can specify various filters to help insure that the phone, fax numbers harvested are extremely targeted to your market. It has various limiters of scanning range - url filter, page text filter, phone/fax filter, domain filter - using which you can extract only the data you actually need from web pages, instead of extracting all the phone, fax present there, as a result, you create your own custom and targeted data base of phone/fax collection. It specializes in harvesting tel, fax numbers from web. WDE can extract tel, fax numbers from website, search results, web directories, list of urls from local file. There are millions of websites on the internet today and most of which are businesses that list their telephone, fax number as a point of contact. WDE - Phone, Fax Harvester module is designed to spider the web for fresh Tel, FAX numbers targeted to the group that you want to market your product or services to.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |