Please use this identifier to cite or link to this item: https://gnanaganga.inflibnet.ac.in:8443/jspui/handle/123456789/4779
Full metadata record
DC FieldValueLanguage
dc.contributor.authorSingh, Lokesh-
dc.contributor.authorSisodia, Deepti-
dc.contributor.authorTaranath, N L-
dc.date.accessioned2024-01-11T04:07:29Z-
dc.date.available2024-01-11T04:07:29Z-
dc.date.issued2023-07-22-
dc.identifier.issn1868-4238-
dc.identifier.issn1868-422X-
dc.identifier.urihttps://doi.org/10.1007/978-3-031-38296-3_22-
dc.identifier.urihttp://gnanaganga.inflibnet.ac.in:8080/jspui/handle/123456789/4779-
dc.description.abstractOnline advertising models are vulnerable to click fraud, which occurs when an individual or a group repeatedly clicks on an online advertisement with the intent to generate illegitimate clicks and make money from the advertiser. In machine learning-based approaches for detecting click fraud, the performance of the models can be affected by the presence of collinear, redundant, and least significant features in the dataset. These types of features can lead to overfitting, where the model becomes too complex and fails to generalize well to new data. Therefore, a Manifold Criterion Variable Elimination method is proposed in this work to select significant features utilizing the potential of six filter-based feature selection techniques for the discrimination of fraud and genuine publishers. Experimentations are conducted on the online advertisement user click dataset in two modes, first considering all extracted features and second considering only selected features. An extraction of 103 statistical features from the user-click dataset is performed for each class instance labelled with OK, Fraud and Observation. The Manifold Criterion Variable Elimination method selects the top 15 most relevant features. Individual and ensemble learning models are trained with selected feature-set and tuned parameter values. The performances of learners are evaluated using standard evaluation measures. The results demonstrated that, in general, the performance of all learners improved with the selected feature set. Particularly, the Gradient Tree Boosting (GTB) ensemble model performed superiorly by improving the weak learners by minimizing the model's loss via merging the weak learners into a strong one iteratively.en_US
dc.language.isoenen_US
dc.publisherInternational Conference on Computational Intelligence in Data Scienceen_US
dc.subjectPPC modelen_US
dc.subjectIllegitimate clicksen_US
dc.subjectLegitimate clicksen_US
dc.subjectEnsemble methodsen_US
dc.subjectFeature selectionen_US
dc.subjectGTBen_US
dc.titleGradient Boosting-Based Predictive Click Fraud Detection Using Manifold Criterion Variable Eliminationen_US
dc.typeArticleen_US
Appears in Collections:Journal Articles

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.