Decision tree learning is one of the most popular families of machine learning algorithms. These techniques are quite intuitive and interpretable but also unstable. It is necessary to use ensemble methods that combine the output of multiple trees, to make the procedure more reliable and stable. Many approaches have been proposed for ranking data, but they are not sensitive to the importance of items. For example, changing the rank of a highly-relevant element should result in a higher penalty than changing a negligible one. Likewise, swapping two similar elements should be less penalized than swapping two dissimilar ones. This paper extends the boosting ensemble method to weighted ranking data, proposing a theoretical and computational definition of item-weighted boosting. The advantages of this procedure are shown through an example on a real data set.

Gli alberi decisionali sono una tecnica predittiva di machine learning particolarmente diffusa, utilizzata per prevedere delle variabili discrete (classificazione) o continue (regressione). Gli algoritmi alla base di queste tecniche sono intuitivi e interpretabili, ma anche instabili. Infatti, per rendere la classificazione più affidabile si `e soliti combinare l’output di più alberi. In letteratura, sono stati proposti diversi approcci per classificare ranking data attraverso gli alberi decisionali, ma nessuno di questi tiene conto ne dell’importanza, ne delle somiglianza dei singoli elementi di ogni ranking. L’obiettivo di questo articolo `e di proporre un’estensione ponderata del metodo boosting per ranking, che tenga conto della struttura di similarità e dell’importanza dei singoli elementi. I vantaggi di questa procedura sono mostrati con un esempio su un dataset reale.

Alessandro Albano, M.S. (2021). Boosting for ranking data: an extension to item weighting. In Book of short papers - SIS 2021.

Boosting for ranking data: an extension to item weighting

Alessandro Albano
;
Mariangela Sciandra;Antonella Plaia
2021-01-01

Abstract

Decision tree learning is one of the most popular families of machine learning algorithms. These techniques are quite intuitive and interpretable but also unstable. It is necessary to use ensemble methods that combine the output of multiple trees, to make the procedure more reliable and stable. Many approaches have been proposed for ranking data, but they are not sensitive to the importance of items. For example, changing the rank of a highly-relevant element should result in a higher penalty than changing a negligible one. Likewise, swapping two similar elements should be less penalized than swapping two dissimilar ones. This paper extends the boosting ensemble method to weighted ranking data, proposing a theoretical and computational definition of item-weighted boosting. The advantages of this procedure are shown through an example on a real data set.
2021
Settore SECS-S/01 - Statistica
9788891927361
Alessandro Albano, M.S. (2021). Boosting for ranking data: an extension to item weighting. In Book of short papers - SIS 2021.
File in questo prodotto:
File Dimensione Formato  
SIS_2021.pdf

Solo gestori archvio

Descrizione: Contributo convegno
Tipologia: Post-print
Dimensione 125.96 kB
Formato Adobe PDF
125.96 kB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10447/513902
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact