The Effects of Virtual AI Influencers’ Form Realism on Consumers Perception and Behavior on Social Media

Inhwa Kim, Chung-Wha ‘Chloe’ Ki, Hyunhwan “Aiden” Lee, and Youn-Kyung Kim

The existing literature examines VIs’ utilitarian, personal, and relational traits, but a gap exists in interface design understanding. This study employs avatar marketing and ambivalence theories to explore how VIs’ design elements (VI’s varying form realism and behavioral realism design) induce ambivalent emotions (eeriness and coolness) and impact VI marketing performance (follow and purchase intentions). Two experimental studies were conducted, using ANCOVA, multiple regression, and PROCESS macro analyses. In Study 1, utilizing contemporary VIs from the market, it was revealed that a VI exhibiting mid (vs. low) form realism triggered heightened perceptions of both eeriness and coolness, with these effects being magnified under the condition of high (vs. low) behavioral realism. In Study 2, using VIs generated through the StyleGAN AI technique, it was validated that those with high (vs. low) form realism were associated with decreased perceptions of eeriness and increased coolness, particularly when combined with high (vs. low) behavioral realism. Additionally, our findings emphasized the negative impact of eeriness and the positive influence of coolness on VI performance outcomes. In summary, our study reveals the significant role of VI interface designs in VI marketing performance, highlighting consumers’ mixed emotions as crucial mediators in this context. Additionally, our findings highlight the negative impact of eeriness and the positive influence of coolness on VI performance. Its primary contribution is uncovering VIs’ interface designs as significant factors in VI marketing performance, while recognizing consumers’ ambivalent emotions as vital mediators in this relationship.

Clothes Made of Pixels and Bits: An AI-driven Topic Modelling Analysis of Fashion NFTs Compared to Digital Fashion

Chenn, Ashley Stevens, Hyunhwan Aiden Lee, Sze Man Chong, Juyeun Jang, and Chung-Wha Chloe Ki. ” Clothes Made of Pixels and Bits: An AI-driven Topic Modelling Analysis of Fashion NFTs Compared to Digital Fashion.”

Non-fungible tokens (NFTs) exploded onto the global digital landscape in 2020, spurred by pandemic-related lockdowns and government stimulus (Ossinger, 2021). An NFT is a unit of data stored on a blockchain that represents or authenticates digital or physical items (Nadini, 2021). Since it resides on a blockchain, NFTs carry the benefits of decentralization, anti-tampering, and traceability (Joy et al., 2022). Fashion brands quickly capitalized on these features, launching fashion NFT collections and garnering significant profits from the sale of fashion NFTs in 2021 (Zhao, 2021). For example, Nike’s December 2021 acquisition of RTFKT (pronounced “artifact”) resulted in USD 185 million in sales less than a year after their acquisition (Marr, 2022).

A New Auditory Image for Social Media: Moving towards Correlation of Spectrographic Analysis and Interpretation with Audience Perception

Nguyen Le Thanh, Hyunhwan “Aiden” Lee, Joseph Johnson, Mitsunori Ogihara, Gang Ren, and James W. Beauchamp. (2019). “A new auditory image for social media: Moving towards correlation of spectrographic analysis and interpretation with audience perception”

Spectrogram and other time-frequency analysis methods transfer an audio file into an auditory image. When signal processing-based analysis and interpretation is performed on these auditory images instead of an audio signal, spectrographic analyses can identify interesting patterns that focus on very different aspects of the signal compared to an audio-based analysis. To facilitate an auditory image-based study, a quantitative analysis and interpretation framework is implemented for exploring the spectrographic images in multiple time and frequency scales and for automatically identifying image features that are relevant to human auditory perception. This analysis framework is applied to two social media datasets: (1) soundtracks from video commercials and “hit” music excerpts from social media platforms, and (2) soundtracks from television and film. Analysis results from social media are also compared with audience subjective evaluations to validate the perceptual relevance of the identified spectrographic patterns.

Multi-Scale Auralization for Multimedia Analytical Feature Interaction

Nguyen Le Thanh, Hyunhwan “Aiden” Lee, Joseph Johnson, Mitsunori Ogihara, Gang Ren, and James W. Beauchamp, “Multi-Scale Auralization for Multimedia Analytical Feature Interaction”

Modern human-computer interaction systems use multiple perceptual dimensions to enhance intuition and efficiency of the user by improving their situational awareness. A signal processing and interaction framework is proposed for auralizing signal patterns for augmenting the visualization-focused analysis tasks of social media content analysis and annotations, with the goal of assisting the user in analyzing, retrieving, and organizing relevant information for marketing research. Audio signals are generated from video/audio signal patterns as an auralization framework, for example, using the audio frequency modulation that follows the magnitude contours of video color saturation. The integration of visual and aural presentations will benefit the user interactions by reducing the fatigue level and sharping the users’ sensitivity, thereby improving work efficiency, confidence, and satisfaction.

Multimodal Content Analysis for Effective Advertisements on YouTube

Nikhita Vedula, Wei Sun, Hyunhwan “Aiden” Lee, Harsh Gupta, Mitsunori Ogihara, Joseph Johnson, Gang Ren, and Srinivasan Parthasarathy. “Multimodal Content Analysis for Effective Advertisements on YouTube”

The rapid advances in e-commerce and Web 2.0 technologies have greatly increased the impact of commercial advertisements on the general public. As a key enabling technology, a multitude of recommender systems exists which analyzes user features and browsing patterns to recommend appealing advertisements to users. In this work, we seek to study the characteristics or attributes that characterize an effective advertisement and recommend a useful set of features to aid the designing and production processes of commercial advertisements. We analyze the temporal patterns from multimedia content of advertisement videos including auditory, visual and textual components, and study their individual roles and synergies in the success of an advertisement. The objective of this work is then to measure the effectiveness of an advertisement, and to recommend a useful set of features to advertisement designers to make it more successful and approachable to users. Our proposed framework employs the signal processing technique of cross-modality feature learning where data streams from different components are employed to train separate neural network models and are then fused together to learn a shared representation. Subsequently, a neural network model trained on this joint feature embedding representation is utilized as a classifier to predict advertisement effectiveness. We validate our approach using subjective ratings from a dedicated user study, the sentiment strength of online viewer comments, and a viewer opinion metric of the ratio of the Likes and Views received by each advertisement from an online platform.

Sequential Pattern Based Temporal Contour Representations for Content-Based Multimedia Timeline Analysis

Gang Ren, Joseph Johnson, Hyunhwan “Aiden” Lee, Mitsunori Ogihara. “Sequential Pattern Based Temporal Contour Representations for Content-Based Multimedia Timeline Analysis”

Temporal contour shapes are closely linked to the narrative structure of multimedia content and provide important reference points in content-based multimedia timeline analysis. In this paper, multimedia timeline is extracted from content as time varying video and audio signal features. A temporal contour representation is implemented based on sequential pattern discovery algorithm for modeling the variation contours of multimedia features. The proposed contour representation extracts repetitive temporal patterns from a hierarchy of time resolutions or from synchronized video/audio feature dimensions. The statistically significant contour components, depicting the dominant timeline shapes, are utilized as a structural or analytical representation of the timeline. The modeling performance of this proposed temporal modeling framework is demonstrated through empirical validation and subjective evaluations.