Twitter is the most popular real-time micro-blogging service and it is a platform where users provide and obtain information at rapid pace. In this scenario, one of the biggest challenge is to find a way to automatically identify the most influential users of a given topic. Currently, there are several approaches that try to address this challenge using different Twitter signals (e.g., number of followers, lists, metadata), but results are not clear and sometimes conflicting. In this paper, we propose TRank, a novel method designed to address the problem of identifying the most influential Twitter users on specific topics identified with hashtags. The novelty of our approach is that it combines different Twitter signals (that represent both the user and the user's tweets) to provide three different indicators that are intended to capture different aspects of being influent. The computation of these indicators is not based on the magnitude of the Twitter signals alone, but they are computed taking into consideration also human factors, as for example the fact that a user with many active followings might have a very noisy time lime and, thus, miss to read many tweets. The experimental assessment confirms that our approach provides results that are more reasonable than the one obtained by mechanisms based on the sole magnitude of data.
|Data di pubblicazione:||2015|
|Titolo:||TRank: Ranking Twitter Users According to Specific Topics|
|Autori:||Furini, Marco; Montangero, Manuela|
|Digital Object Identifier (DOI):||10.1109/CCNC.2015.7158074|
|Data del convegno:||9-12 January 2015|
|Nome del convegno:||2015 12th Annual IEEE Consumer Communications and Networking Conference (CCNC 2015)|
|Luogo del convegno:||Las Vegas, Nevada, USA|
|Titolo del libro:||2015 12th Annual IEEE Consumer Communications and Networking Conference (CCNC 2015)|
|Appare nelle tipologie:||Relazione in Atti di Convegno|
I documenti presenti in Iris Unimore sono rilasciati con licenza Creative Commons Attribuzione - Non commerciale - Non opere derivate 3.0 Italia, salvo diversa indicazione.
In caso di violazione di copyright, contattare Supporto Iris