A Comprehensive Study Of Features and Algorithms for URL Based Topic Classification

CONCLUSIONS

In this article we studied Web page topic classification from URL. Even though contentbased topic classifiers gave better results than URL-based ones, topic classification from URL is preferable when the content is not available, or when classification speed has the highest importance.
We can summarize our main findings for URL-based Web page topic classification as follows.

  • (1) We showed that dictionary-based baseline algorithms are not enough for highperformance URL-based topic classification. For the dictionary-based methods even the best-performing variant using all-grams (a combination of 4-, 5-, 6-, 7-, and 8-grams) achieved a macroaveraged F-measure of only 73. On the other hand for topic classifiers where precision is important and some recall can be sacrificed, token-based statistical dictionaries can be used as they achieved a macroaveraged precision of 92 with a macroaveraged recall of 34 on the ODP dataset.
  • (2) We showed that the features have more impact on the classifier performance than the classification algorithms. Allgrams derived from URLs was the best feature set, considerably better than tokens.
    Explicitly encoding positional information for all-grams derived from tokens performed slightly better than using only such all-grams without positional information, but still slightly worse than using all-grams from the URL. Given the same feature set, the ME and SVM algorithms showed the same performance, and the other algorithms also performed similarly.
  • (3) We obtained a macroaveraged F-measure of 83 when ME
    algorithm is used with all-grams derived directly from URLs.
  • (4) We reported a performance which improves the best previously reported URL-only performance for a small dataset of university pages by 13 points in F-measure.
  • (5) Using summaries of Web pages for training and testing led to a large improvement over using only URLs, with a macroaveraged F-measure of 94. On the other hand the performance of URLbased topic classification decreased when the summaries of Web pages are used in
    training phase in addition to URLs. The reason for this is the vocabulary difference between URLs and the summaries of Web pages.
  • (6) We achieved an additional small improvement with using inlink information.
  • (7) Applying boosting to combine different algorithms gave a small performance improvement of 1 or 2 points in F-measure.
  • (8) The challenges for URL-based topic classification are:
  • (i) data consistency as the definitions of topics differ from one dataset to another,
  • (ii) overlap between different
    topics in one dataset,
  • (iii) empty URLs consisting of only stop tokens or previously
    unseen tokens.

Dictionaries and String Matching

    1. We formed four types of dictionaries which contain a list of representative tokens or all-grams derived from URLs for topics. In the first type of dictionary, Tokens byHand, we used all words from the first two levels of the ODP hierarchy. For example, the terms “Basketball” and “Football” which are listed one level below “Sports” in ODP hierarchy, are added to the sports dictionary. Some terms, such as “Online” listed
      under “Games,” are not put into the games dictionary if they appeared nontopic-specific to a human. The average dictionary size was 19.8 words per topic.
    1. For the second type of dictionary Tokens by-Statistics we formed a list of tokens which have length greater than 2, by first merging the “set-of-tokens” from URLs of each Web page listed under ODP for each topic. Set-of-tokens is simply the list of tokens seen in the URL.

Then we obtained representative tokens for a particular topic, by looking at the percentage of “set-of-tokens” containing this token, both for the topic itself and for the other topics.

A token was added to the dictionary corresponding to the topic if:
(i) it appeared in at least five “set-of-tokens” of the topic, (ii) it had a precision of at least 80 on these “set-of-tokens”, and (iii) it had a recall of at least .01.

The average dictionary size for this approach was 987 words. We decided on these rules after manually inspecting the list of tokens and we found tokens with a recall above 0.01% and a precision of 80% or more to be tokens that humans might choose as “topically relevant”. In Tokens by-Statistics dictionary “basketball” and “mensbasketball” are some example tokens for Sports topic.

    1. The third type of token dictionary is referred to as Domains by-Statistics. For this the list of typical domains for each topic is formed in the same way as representative tokens are formed for dictionary Tokens by-Statistics. The difference is now “set-of-tokens” contains only domains of URLs. With the Domains by-Statistics dictionary a Web page is classified as “Sports” if and only if it comes from one of the typical sports domains. In Domains by-Statistics dictionary “football.co.uk” and “sportsnetwork.com” are some example domains for Sports topic.
    1. Finally, we also trained a dictionary by using all-grams instead of tokens named All-grams by-Statistics, which resulted in an average dictionary size of 13k n-grams. In All-grams by-Statistics dictionary “sports” and “sportspa” are some example grams for Sports topic.
Baseline Results for Using Dictionaries with Token Match on ODP

Table shows the performance of the baseline algorithms when the tokens (or allgrams) from the dictionary are checked with token match strategy with the tokens (or all-grams) derived from the test URLs. For all the topics, all the token dictionaries gave fairly high precision values but low recall values. The Tokens by-Hand dictionary gave the lowest recall values as their vocabulary for each topic is limited to the two first levels of ODP hierarchy. The Domains by-Statistics dictionary seems to give higher precision values than the other dictionaries. However, it has a macroaveraged recall of 22. This shows that domains are indeed good indicators but using only domain information will not be enough to achieve an adequate level of recall. When we compare the performances of all dictionaries we see that the statistical dictionaries give higher performances. We had two types of statistical dictionaries, Tokens byStatistics and All-grams by-Statistics. The Tokens by-Statistics dictionary achieved a higher precision but a much lower recall than the All-grams by-Statistics dictionary. On the ODP dataset All-grams by-Statistics gives the highest performance in terms of F-measure with a macroaveraged F-measure of 73.

For the first two token-based dictionaries we also experimented with using
substring matches, rather than token matches on the ODP dataset. This increases recall as now the “volleyball” in http://www.attackvolleyballclub.net/ is also detected. The macroaverages in this case are P = 84, R = 15, F = 24 for the Tokens by-Hand dictionary and P = 70, R = 63 and F = 62 for the Tokens by-Statistics dictionary.

From these results we can conclude that baseline algorithms are not enough for a good performance. However, results for baseline algorithms give us an insight about which topics are easier to classify. For example “Adult”, “Sports”, and “Games” seem to have higher F-measure values than the macroaverage for most of the dictionaries both in cases of token match and substring match. In practice dictionary-based baselines using tokens might be of interest if high precision but not necessarily high recall is required. For example, a topic-focused crawler might want to detect some surely relevant URLs early during a crawl and then use these pages to “bootstrap”, for example, using the link information.

?著作權(quán)歸作者所有,轉(zhuǎn)載或內(nèi)容合作請(qǐng)聯(lián)系作者
  • 序言:七十年代末悄泥,一起剝皮案震驚了整個(gè)濱河市虏冻,隨后出現(xiàn)的幾起案子,更是在濱河造成了極大的恐慌弹囚,老刑警劉巖厨相,帶你破解...
    沈念sama閱讀 211,817評(píng)論 6 492
  • 序言:濱河連續(xù)發(fā)生了三起死亡事件,死亡現(xiàn)場(chǎng)離奇詭異鸥鹉,居然都是意外死亡蛮穿,警方通過(guò)查閱死者的電腦和手機(jī),發(fā)現(xiàn)死者居然都...
    沈念sama閱讀 90,329評(píng)論 3 385
  • 文/潘曉璐 我一進(jìn)店門毁渗,熙熙樓的掌柜王于貴愁眉苦臉地迎上來(lái)践磅,“玉大人,你說(shuō)我怎么就攤上這事灸异「剩” “怎么了?”我有些...
    開(kāi)封第一講書(shū)人閱讀 157,354評(píng)論 0 348
  • 文/不壞的土叔 我叫張陵肺樟,是天一觀的道長(zhǎng)檐春。 經(jīng)常有香客問(wèn)我,道長(zhǎng)么伯,這世上最難降的妖魔是什么疟暖? 我笑而不...
    開(kāi)封第一講書(shū)人閱讀 56,498評(píng)論 1 284
  • 正文 為了忘掉前任,我火速辦了婚禮,結(jié)果婚禮上俐巴,老公的妹妹穿的比我還像新娘骨望。我一直安慰自己,他們只是感情好欣舵,可當(dāng)我...
    茶點(diǎn)故事閱讀 65,600評(píng)論 6 386
  • 文/花漫 我一把揭開(kāi)白布擎鸠。 她就那樣靜靜地躺著,像睡著了一般邻遏。 火紅的嫁衣襯著肌膚如雪糠亩。 梳的紋絲不亂的頭發(fā)上,一...
    開(kāi)封第一講書(shū)人閱讀 49,829評(píng)論 1 290
  • 那天准验,我揣著相機(jī)與錄音,去河邊找鬼廷没。 笑死糊饱,一個(gè)胖子當(dāng)著我的面吹牛,可吹牛的內(nèi)容都是我干的颠黎。 我是一名探鬼主播另锋,決...
    沈念sama閱讀 38,979評(píng)論 3 408
  • 文/蒼蘭香墨 我猛地睜開(kāi)眼,長(zhǎng)吁一口氣:“原來(lái)是場(chǎng)噩夢(mèng)啊……” “哼狭归!你這毒婦竟也來(lái)了夭坪?” 一聲冷哼從身側(cè)響起,我...
    開(kāi)封第一講書(shū)人閱讀 37,722評(píng)論 0 266
  • 序言:老撾萬(wàn)榮一對(duì)情侶失蹤过椎,失蹤者是張志新(化名)和其女友劉穎室梅,沒(méi)想到半個(gè)月后,有當(dāng)?shù)厝嗽跇?shù)林里發(fā)現(xiàn)了一具尸體疚宇,經(jīng)...
    沈念sama閱讀 44,189評(píng)論 1 303
  • 正文 獨(dú)居荒郊野嶺守林人離奇死亡亡鼠,尸身上長(zhǎng)有42處帶血的膿包…… 初始之章·張勛 以下內(nèi)容為張勛視角 年9月15日...
    茶點(diǎn)故事閱讀 36,519評(píng)論 2 327
  • 正文 我和宋清朗相戀三年,在試婚紗的時(shí)候發(fā)現(xiàn)自己被綠了敷待。 大學(xué)時(shí)的朋友給我發(fā)了我未婚夫和他白月光在一起吃飯的照片间涵。...
    茶點(diǎn)故事閱讀 38,654評(píng)論 1 340
  • 序言:一個(gè)原本活蹦亂跳的男人離奇死亡,死狀恐怖榜揖,靈堂內(nèi)的尸體忽然破棺而出勾哩,到底是詐尸還是另有隱情,我是刑警寧澤举哟,帶...
    沈念sama閱讀 34,329評(píng)論 4 330
  • 正文 年R本政府宣布思劳,位于F島的核電站,受9級(jí)特大地震影響炎滞,放射性物質(zhì)發(fā)生泄漏敢艰。R本人自食惡果不足惜,卻給世界環(huán)境...
    茶點(diǎn)故事閱讀 39,940評(píng)論 3 313
  • 文/蒙蒙 一册赛、第九天 我趴在偏房一處隱蔽的房頂上張望钠导。 院中可真熱鬧震嫉,春花似錦、人聲如沸牡属。這莊子的主人今日做“春日...
    開(kāi)封第一講書(shū)人閱讀 30,762評(píng)論 0 21
  • 文/蒼蘭香墨 我抬頭看了看天上的太陽(yáng)逮栅。三九已至悴势,卻和暖如春,著一層夾襖步出監(jiān)牢的瞬間措伐,已是汗流浹背特纤。 一陣腳步聲響...
    開(kāi)封第一講書(shū)人閱讀 31,993評(píng)論 1 266
  • 我被黑心中介騙來(lái)泰國(guó)打工, 沒(méi)想到剛下飛機(jī)就差點(diǎn)兒被人妖公主榨干…… 1. 我叫王不留侥加,地道東北人捧存。 一個(gè)月前我還...
    沈念sama閱讀 46,382評(píng)論 2 360
  • 正文 我出身青樓,卻偏偏與公主長(zhǎng)得像担败,于是被迫代替她去往敵國(guó)和親昔穴。 傳聞我的和親對(duì)象是個(gè)殘疾皇子,可洞房花燭夜當(dāng)晚...
    茶點(diǎn)故事閱讀 43,543評(píng)論 2 349

推薦閱讀更多精彩內(nèi)容