Friends who know a little bit about dede collection rules can completely skip the previous part. Let’s take a look at how to solve the problem of paging collection with static braking and constant change.
2. Collect new targets
Target address:
1. http://www.tiansou.net/Html/Y_CYFW/R_Gzzj/F_Gzjh/index.html
2. http: //www.tiansou.net/Html/Y_CYFW/R_Gzzj/F_Gzjh/2007-2/9/20070209110903558.html
The reason why two target pages are selected is because one of the above two pages has pagination, and the other No, and there is a big difference in pagination and full-text sampling. The following description has been modified on the basis of collecting all links to the target address (homepage). Some places may appear to be confusing, just for the convenience of explanation.
Target text part header code 1:
Target text part header code 2:
Pass It is not difficult to find that the starting collection part of the two text parts can be determined as the black part. The beginning part is easy to say. The code is as follows:
Target tail and pagination area code 1:
= 700) window.open('/upload/200743213057312.gif');" src="http://www.bkjia.com/uploads/allimg/131016/06064K143-2.gif" onload="if(this.width >'700')this.width='700';" border=0> Target tail and paging area code 2:
Compare the two endings. Although I want to advance the ending of the first one a little further, there is no way. We have to consider everything The common part of the link has to be the part outlined in black. This will also make it a little troublesome to determine the filtering rules in the future. This is a story later. Let’s finalize the ending part first:
The code is as follows:
http: //www.bkjia.com/PHPjc/318217.html