Exploring Online Ad Images Using a Deep Convolutional Neural Network Approach (original) (raw)

Deep Learning Based Modeling in Computational Advertising: A Winning Formula

Industrial Engineering & Management, 2018

Targeting the correct customers is critical to computational advertising. Once you find your potential customer, you want to know who can offer the most suitable offer at the right time. There are other ways to visualize your perfect customer using deep learning, and we developed a method to target and attract the best online customers and get to know what they want before you start selling.

ADNet: A Deep Network for Detecting Adverts

2018

Online video advertising gives content providers the ability to deliver compelling content, reach a growing audience, and generate additional revenue from online media. Recently, advertising strategies are designed to look for original advert(s) in a video frame, and replacing them with new adverts. These strategies, popularly known as product placement or embedded marketing, greatly help the marketing agencies to reach out to a wider audience. However, in the existing literature, such detection of candidate frames in a video sequence for the purpose of advert integration, is done manually. In this paper, we propose a deep-learning architecture called ADNet, that automatically detects the presence of advertisements in video frames. Our approach is the first of its kind that automatically detects the presence of adverts in a video frame, and achieves state-of-the-art results on a public dataset.

Text Advertisements Analysis using Convolutional Neural Networks

International Journal of Database Management Systems

In this paper, we describe the developed model of the Convolutional Neural Networks CNN to a classification of advertisements. The developed method has been tested on both texts (Arabic and Slovak texts).The advertisements are chosen on a classified advertisements websites as short texts. We evolved a modified model of the CNN, we have implemented it and developed next modifications. We studied their influence on the performing activity of the proposed network. The result is a functional model of the network and its implementation in Java and Python. And analysis of model results using different parameters for the network and input data. The results on experiments data show that the developed model of CNN is useful in the domains of Arabic and Slovak short texts, mainly for some classification of advertisements. This paper gives complete guidelines for authors submitting papers for the AIRCC Journals.

AD or Non-AD: A Deep Learning Approach to Detect Advertisements from Magazines

Entropy, 2018

The processing and analyzing of multimedia data has become a popular research topic due to the evolution of deep learning. Deep learning has played an important role in addressing many challenging problems, such as computer vision, image recognition, and image detection, which can be useful in many real-world applications. In this study, we analyzed visual features of images to detect advertising images from scanned images of various magazines. The aim is to identify key features of advertising images and to apply them to real-world application. The proposed work will eventually help improve marketing strategies, which requires the classification of advertising images from magazines. We employed convolutional neural networks to classify scanned images as either advertisements or non-advertisements (i.e., articles). The results show that the proposed approach outperforms other classifiers and the related work in terms of accuracy.

Enhancing Dynamic Image Advertising with Vision-Language Pre-training

Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval

In the multimedia era, image is an effective medium in search advertising. Dynamic Image Advertising (DIA), a system that matches queries with ad images and generates multimodal ads, is introduced to improve user experience and ad revenue. The core of DIA is a query-image matching module performing ad image retrieval and relevance modeling. Current query-image matching suffers from limited and inconsistent data, and insufficient cross-modal interaction. Also, the separate optimization of retrieval and relevance models affects overall performance. To address this issue, we propose a vision-language framework consisting of two parts. First, we train a base model on large-scale image-text pairs to learn general multimodal representation. Then, we fine-tune the base model

CAN: Effective cross features by global attention mechanism and neural network for ad click prediction

Tsinghua Science and Technology, 2022

Online advertising click-through rate (CTR) prediction is aimed at predicting the probability of a user clicking an ad, and it has undergone considerable development in recent years. One of the hot topics in this area is the construction of feature interactions to facilitate accurate prediction. Factorization machine provides second-order feature interactions by linearly multiplying hidden feature factors. However, real-world data present a complex and nonlinear structure. Hence, second-order feature interactions are unable to represent cross information adequately. This drawback has been addressed using deep neural networks (DNNs), which enable high-order nonlinear feature interactions. However, DNN-based feature interactions cannot easily optimize deep structures because of the absence of cross information in the original features. In this study, we propose an effective CTR prediction algorithm called CAN, which explicitly exploits the benefits of attention mechanisms and DNN models. The attention mechanism is used to provide rich and expressive low-order feature interactions and facilitate the optimization of DNN-based predictors that implicitly incorporate high-order nonlinear feature interactions. The experiments using two real datasets demonstrate that our proposed CAN model performs better than other cross feature-and DNN-based predictors.

Automatic Understanding of Image and Video Advertisements

2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)

There is more to images than their objective physical content: for example, advertisements are created to persuade a viewer to take a certain action. We propose the novel problem of automatic advertisement understanding. To enable research on this problem, we create two datasets: an image dataset of 64,832 image ads, and a video dataset of 3,477 ads. Our data contains rich annotations encompassing the topic and sentiment of the ads, questions and answers describing what actions the viewer is prompted to take and the reasoning that the ad presents to persuade the viewer ("What should I do according to this ad, and why should I do it?"), and symbolic references ads make (e.g. a dove symbolizes peace). We also analyze the most common persuasive strategies ads use, and the capabilities that computer vision systems should have to understand these strategies. We present baseline classification results for several prediction tasks, including automatically answering questions about the messages of the ads.

Viewability Prediction for Online Display Ads

As a massive industry, display advertising delivers advertis-ers' marketing messages to attract customers through graphic banners on webpages. Advertisers are charged for each view of a page that contains their ads. However, recent studies have found out that about half of the ads were actually never seen by users because they do not scroll deep enough to bring the ads in-view. Low viewability hurts financially both the advertisers and the publishers. This paper is the first to address the problem of ad viewability prediction, which can improve the performance of guaranteed ad delivery , real-time bidding, and even recommender systems. We analyze a real-life dataset from a large publisher, identify a number of features that impact the scroll depth of a given user-page pair, and propose a probabilistic latent class model that can predict the viewability of any given scroll depth for a user-page pair. The experiments demonstrate that our model outperforms comparison systems based on singular value decomposition and logistic regression. Furthermore, our model needs to be trained only once, independent of the target scroll depth, and works well in real-time.

M2FN: Multi-step modality fusion for advertisement image assessment

Applied Soft Computing, 2021

Assessing advertisements, specifically on the basis of user preferences and ad quality, is crucial to the marketing industry. Although recent studies have attempted to use deep neural networks for this purpose, these studies have not utilized image-related auxiliary attributes, which include embedded text frequently found in ad images. We, therefore, investigated the influence of these attributes on ad image preferences. First, we analyzed large-scale realworld ad log data and, based on our findings, proposed a novel multi-step modality fusion network (M2FN) that determines advertising images likely to appeal to user preferences. Our method utilizes auxiliary attributes through multiple steps in the network, which include conditional batch normalization-based low-level fusion and attention-based high-level fusion. We verified M2FN on the AVA dataset, which is widely used for aesthetic image assessment, and then demonstrated that M2FN can achieve state-of-the-art performance in preference prediction using a real-world ad dataset with rich auxiliary attributes.

Conversion Prediction Using Multi-task Conditional Attention Networks to Support the Creation of Effective Ad Creatives

Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining - KDD '19, 2019

Accurately predicting conversions in advertisements is generally a challenging task, because such conversions do not occur frequently. In this paper, we propose a new framework to support creating high-performing ad creatives, including the accurate prediction of ad creative text conversions before delivering to the consumer. The proposed framework includes three key ideas: multi-task learning, conditional attention, and attention highlighting. Multi-task learning is an idea for improving the prediction accuracy of conversion, which predicts clicks and conversions simultaneously, to solve the difficulty of data imbalance. Furthermore, conditional attention focuses attention of each ad creative with the consideration of its genre and target gender, thus improving conversion prediction accuracy. Attention highlighting visualizes important words and/or phrases based on conditional attention. We evaluated the proposed framework with actual delivery history data (14,000 creatives displayed more than a certain number of times from Gunosy Inc.), and confirmed that these ideas improve the prediction performance of conversions, and visualize noteworthy words according to the creatives' attributes. CCS CONCEPTS • Information systems → Online advertising; • Computing methodologies → Multi-task learning; Neural networks.