How AI encourages consumers to share their secrets? The role of anthropomorphism, personalisation, and privacy concerns and avenues for future research

Bianca Kronemann (University of Hull, Hull, UK)
Hatice Kizgin (University of Twente, Enschede, The Netherlands)
Nripendra Rana (Qatar University, Doha, Qatar)
Yogesh K. Dwivedi (Swansea University, Bay Campus, Swansea, UK and Symbiosis International (Deemed University), Pune, India)

Spanish Journal of Marketing - ESIC

ISSN: 2444-9695

Article publication date: 13 January 2023

Issue publication date: 19 April 2023

10869

Abstract

Purpose

This paper aims to explore the overall research question “How can artificial intelligence (AI) influence consumer information disclosure?”. It considers how anthropomorphism of AI, personalisation and privacy concerns influence consumers’ attitudes and encourage disclosure of their private information.

Design/methodology/approach

This research draws upon the personalisation-privacy paradox (PPP) and privacy calculus theory (PCT) to address the research question and examine how AI can influence consumer information disclosure. It is proposed that anthropomorphism of AI and personalisation positively influence consumer attitudes and intentions to disclose personal information to a digital assistant, while privacy concerns negatively affect attitude and information disclosure.

Findings

This paper develops a conceptual model based on and presents seven research propositions (RPs) for future research.

Originality/value

Building upon PPP and PCT, this paper presents a view on the benefits and drawbacks of AI from a consumer perspective. This paper contributes to literature by critically reflecting upon on the question how consumer information disclosure is influenced by AI. In addition, seven RPs and future research areas are outlined in relation to privacy and consumer information disclosure in relation to AI.

¿Cómo anima la IA a los consumidores a compartir sus secretos?

El papel del antropomorfismo, la personalización y los problemas de privacidad y perspectivas para la investigación futura

Resumen

Propósito

Este artículo explora la pregunta general de investigación “¿Cómo puede influir la inteligencia artificial (IA) en la divulgación de información por parte de los consumidores? Se analiza cómo el antropomorfismo de la IA, la personalización y la preocupación por la privacidad influyen en la actitud de los consumidores y fomentan la revelación de su información privada.

Diseño/metodología/enfoque

Esta investigación se basa en la paradoja de la personalización y la privacidad y en la teoría del cálculo de la privacidad para abordar la pregunta de investigación y examinar cómo la IA puede influir en la revelación de información de los consumidores. Se propone que el antropomorfismo de la IA y la personalización influyen positivamente en las actitudes de los consumidores y en su intención de revelar información personal a un asistente digital, mientras que la preocupación por la privacidad afecta negativamente a la actitud y a la revelación de información.

Conclusiones

Este artículo desarrolla un modelo conceptual basado en siete propuestas de investigación para el futuro.

Originalidad

Basándose en la paradoja de la personalización y la privacidad y en la teoría del cálculo de la privacidad, este artículo presenta un punto de vista sobre los beneficios e inconvenientes de la IA desde la perspectiva del consumidor. Este artículo contribuye a la literatura al reflexionar de forma crítica sobre la cuestión de cómo influye la IA en la revelación de información del consumidor. Además, se esbozan siete propuestas de investigación y futuras áreas de investigación en relación con la privacidad y la divulgación de información del consumidor en relación con la IA.

人工智能如何

鼓励消费者分享他们的秘密?拟人化、个性化和隐私问题的作用以及未来研究的途径

摘要

目的

本文探讨了 “人工智能如何影响消费者的信息披露?"这一总体研究问题。它考虑了人工智能(AI)的拟人化、个性化和隐私问题是如何影响消费者的态度并鼓励他们披露私人信息的。

设计/方法/途径

本研究借鉴了个性化-隐私悖论和隐私计算理论来解决研究问题, 并研究人工智能如何影响消费者信息披露。本文提出, 人工智能的拟人化和个性化对消费者向数字助理披露个人信息的态度和意图有积极影响, 而隐私问题对态度和信息披露有消极影响。

研究结果

本文在此基础上建立了一个概念模型, 并为未来的研究提出了七个研究命题。

原创性

在个性化-隐私悖论和隐私计算理论的基础上, 本文从消费者的角度提出了对人工智能的好处和坏处的看法。本文通过对消费者信息披露如何受到人工智能影响的问题进行批判性反思, 对文献做出了贡献。此外, 本文概述了与人工智能相关的隐私和消费者信息披露方面的七个研究命题和未来研究领域。

Keywords

Citation

Kronemann, B., Kizgin, H., Rana, N. and K. Dwivedi, Y. (2023), "How AI encourages consumers to share their secrets? The role of anthropomorphism, personalisation, and privacy concerns and avenues for future research", Spanish Journal of Marketing - ESIC, Vol. 27 No. 1, pp. 3-19. https://doi.org/10.1108/SJME-10-2022-0213

Publisher

:

Emerald Publishing Limited

Copyright © 2022, Bianca Kronemann, Hatice Kizgin, Nripendra Rana and Yogesh K. Dwivedi.

License

Published in Spanish Journal of Marketing – ESIC. Published by Emerald Publishing Limited. This article is published under the Creative Commons Attribution (CC BY 4.0) licence. Anyone may reproduce, distribute, translate and create derivative works of this article (for both commercial and non-commercial purposes), subject to full attribution to the original publication and authors. The full terms of this licence maybe seen at http://creativecommons.org/licences/by/4.0/legalcode


1. Introduction

Artificial intelligence (AI) is contributing to a shift towards a more algorithmic society (Shankar, 2018). The exchange of value in the digital environment is changing as technology is changing from being a tool that is used by actors to being an actor in the value exchange itself (Novak and Hoffman, 2019). Human-less transaction mediated by intelligent technology is growing in numbers and frequency (Hofacker and Corsaro, 2020). Advancements in AI are now also being used across several consumer industries such as retail and e-commerce as well as in marketing. One popular application of AI in marketing is in form conversational agents (Rai, 2020; Thomaz et al., 2020). These are based on natural language computer programmes designed to approximate human speech and interact with people via a digital interface (Thomaz et al., 2020).

However, AI may not be welcomed favourably by all stakeholders due to the circumstance that AI implies the increased ability of technology owners to collect, analyse, combine and control consumer data (Mazurek and Malagocka, 2019). The interaction with digital assistants leaves a rich digital footprint of personal and behavioural data, which means that consumers become active participants in the erosion of their own privacy (Pomfret et al., 2020). Thus, the topic of privacy is under heightened attention across academic researchers, social critics as well as regulators (Martin and Murphy, 2019) because in our knowledge-based digital market environment, information privacy and security have become the primary consumer concerns related to the internet, algorithms and data (Brill et al., 2019; Hong and Thong, 2013; Kietzmann et al., 2018; Mazurek and Malagocka, 2019). The notion of data privacy is critical because advancements in AI are based on and have accelerated with the availability of large amounts of data and computational capabilities to analyse them. The free flow of data is critical because data is the fuel of AI success (Mazurek and Malagocka, 2019).

Protecting consumer privacy and user information is critical to the expanding integration of digital assistants into business models as well as everyday life. Businesses face a substantial risk of abandoned investment into digital assistants if consumers lack trust in either the technology or the company to protect their privacy and personal information (Pappas, 2016). Existing research finds that privacy concerns can form a barrier to consumer adoption of new technology. However, recent research suggests that consumers showcase a trade-off between benefits and risks (Vilmakumar et al., 2021). Based on this ambiguity in the existing literature, this paper will critically reflect on the topics of privacy concerns in relation to AI, benefits in terms of personalisation offered by AI, and consumer information disclosure to AI to answer the overall research question of “How can AI influence consumer information disclosure?”.

This paper contributes to the growing discussion in the marketing literature around consumers’ interaction with AI, benefits of personalisation versus drawbacks of privacy concerns, and consumer information disclosure. To contribute to existing literature, this study will draw upon the personalisation-privacy paradox (PPP) to examine the effects of AI on consumer attitudes and consumer information disclosure. The PPP is a valuable theoretical basis because it describes a situation in which consumers appreciate the value of personalisation offered by a technology (Aguirre et al., 2015; Bleier and Eisenbeiss, 2015) while being aware of marketers’ exploitation of their personal data which raises privacy concerns (Cloarec, 2020). This idea is also reflected in privacy calculus theory (PCT) (Dinev and Hart, 2006; Wang et al., 2016) which is also drawn upon to better understand how and why consumers assess benefits and drawbacks of AI and decide whether to disclose their private information. Based on PCT and PPP, this research proposes a conceptual model with seven research propositions (RPs) that address how anthropomorphism of AI, personalisation and privacy concerns affect consumers’ attitudes towards AI and their intention to disclose personal information to AI.

The remaining parts of this paper are structured into four sections: Firstly, literature relating to consumer interaction with AI is reviewed to provide background information on the increased need to better understand consumer interaction with AI, as well as consequences of such interaction. Secondly, the theoretical foundation for a research model is presented, and eight RPs introduced. This section will cover the topics of anthropomorphism of AI, privacy concerns in relation to AI, personalised interactions with AI, and the PPP. Thirdly, the paper will discuss insights and findings gained from the review of literature and development of a research model. Finally, a conclusion is drawn that included reflection on limitations and suggestions for future research.

2. Literature review: consumer interaction with AI

Interaction between AI and consumers is becoming a central topic of discussion in the marketing and service literature. Companies adopt these conversational agents as new front-facing customer service that interact with consumers during service encounters (Lariviere et al., 2017; Thomaz et al., 2020; van Doorn et al., 2017; Wilson and Daugherty, 2018). It is suggested that AI will fundamentally change the nature of interaction between companies and their customers (Haenlein and Kaplan, 2019; Ramaswamy and Ozcan, 2018).

However, research relating to the impact of AI is still developing and relatively sparse (Davenport et al., 2020; Steinhoff et al., 2019). Table 1 gives an overview of selected studies, organised by publication date, that examine AI in the marketing, service and hospitality context. The table highlights each paper’s theoretical foundation and research aim, key findings of each paper as well as limitations.

From reviewing above selected studies, four key themes emerge, which will be covered in the following paragraphs: adoption and use of AI technology; anthropomorphism of AI; benefits of AI; and potential barriers to AI success such as privacy and security concerns.

The overview of previous studies shows that most of the previous literature has focused on consumer use and adoption of AI, highlighting that AI integration and application in the consumer context is not a fad but a rising trend (Gursoy et al., 2019). Research shows that consumers are rapidly embracing AI for daily use because they enable individuals to access timely and useful information and can meet customer demand for contextually relevant and highly personalised content that is delivered in real-time (Brill et al., 2019; Canbek and Mutlu, 2016). In addition, research finds that the ease of using AI and its ability to provide relevant information are key factors in influencing consumers to use them, which also is found to influence loyalty (Moriuchi, 2019). Further studies find that a variety of factors along social, relational and functional dimensions are important for adoption of AI (Fernandes and Oliveira, 2021; Vilmakumar et al., 2021). Research shows that individuals are mainly motivated to adopt AI based on functional elements, but they also find support for social and relational elements that drive adoption (Fernandes and Oliveira, 2021). In addition, experience with technology is important to consider in the development of attitudes and adoption intentions of new technology like AI.

Anthropomorphism of AI is a second main theme across literature. The underlying idea for anthropomorphism is that it can influence consumers to think about intelligent agents more favourably. The positive influence of increased human likeness is also confirmed in a study by Belanche et al. (2021) who found that human likeness positively affects utilitarian, social, monetary and emotional value expectations. However, another study examines the uncanny valley theory and find support for their hypotheses that too humanlike features lead to negative evaluations due to perceived threats to human identity (Kim et al., 2019). Yet it remains unclear perceptions of uncanniness translate into behaviour. This is in line with another study that found anthropomorphism as an impeding determinant of consumers’ acceptance of service robots due to the reason that they pose a threat to users’ identity (Lu et al., 2019).

AI offers many benefits to both companies in terms of improved capabilities to collect, analyse and use data in real-time, which creates competitiveness but also the opportunity to co-create value with consumers through interaction. Yet, the paper does not consider consumer attitudes around the availability and provision for that data that is necessary for real-time co-creation. Companies like Amazon or Apple are exemplary for the utilisation of AI, either sold separately to consumers (e.g. Amazon’s Echo) or included in existing consumer devices (e.g. Siri). As an operating system of iOS, Siri is one of the most salient applications of Apple that has caught attention of a tech-savvy audience (Canbek and Mutlu, 2016). One of the most significant advantages of these assistants is their conversational interface (Vilmakumar et al., 2021) and the opportunity to meet the demand for contextually relevant and personalized information that is delivered to the consumer in real-time (Brill et al., 2019). Furthermore, AI has the advantage of being highly scalable, and the ability to deliver routine customer service to large numbers of people simultaneously (Chui et al., 2018; Davenport et al., 2020; Duan et al., 2019; Wilson and Daugherty, 2018).

The increased ability of AI to collect, analyse and store consumer data, however, also results in concerns relating to privacy and security of data. Privacy concerns are relevant to consider for adoption because they also influence trust (Vilmakumar et al., 2021). Existing literature shows that privacy concerns have a negative impact on intentions to use a technology, engagement and attitude towards an innovative technology (Brill et al., 2019; Grewal et al., 2021). Consumers have also privacy concerns because they do not trust AI. Overall, it becomes clear that perceived risk of AI poses challenges that need to be investigated to understand implications for AI.

Above discussed studies shed first light on the adoption, potential benefits and barriers as well as consequences of the integration of AI. However, it becomes evident that most studies have been from a use and adoption perspective. There is a lack of critical reflection around the necessity of data and personal information that is required for the successful delivery of AI in the first place. It has been neglected to fully examine consumer concerns and potential barriers in relation to data privacy and information disclosure to AI.

3. Theoretical background and research propositions

The following section will present the theoretical background of this study being PCT and the PPP and highlight how drawing upon these theories will offer new insights on how consumers may be encouraged to disclose their private information to AI. In addition, this section will propose a research model with seven RPs.

3.1 Personalisation-privacy paradox and privacy calculus theory

In Western developed societies, individuals have increasingly raised privacy and data security concerns (Foehr and Germelmann, 2020). Becoming increasingly aware “that they are being watched”, consumers are concerned about their data being misused or abused due to eroding confidence that firms are respecting data privacy (Fitzgerald, 2019). Individuals are challenged to manage the complex trade-offs that new technological innovation brings. It must be evaluated how the benefits of a new technology compare to the risks of information privacy concerns (Acquisti et al., 2015) This trade-off has been highlighted as the PPP (Aguirre et al., 2015; Bleier and Eisenbeiss, 2015) which describes the process of consumers´ critical assessment of offered benefits and added value of personalisation against marketers´ exploitation of their private information which raises privacy concerns (Cloarec, 2020). However, the evaluation can be difficult for consumers due to the circumstance that technology often functions as a black box with a lack of evidence of where consumer data is stored, or who will be able to access it (Hofacker and Corsaro, 2020).

Reflecting the idea of the PPP, and to gain a better understanding of privacy and disclosure behaviour, the concept of privacy calculus or PCT will be drawn upon in addition to the PPP. According to the privacy calculus, consumers trade off their privacy or risks to their privacy in a cognitive process of rationality against the prospect of receiving benefit or utility (Dinev and Hart, 2006; Wang et al., 2016). Benefits that personalisation offers include improved experience through curation of relevant content or discounts (Zhu et al., 2017). Central to the privacy calculus perspective is the underlying assumption that consumers are empowered to exercise control over their private information (Dinev et al., 2016), and that rational consumers are apprehensive towards revealing personal information. PCT has been adopted in previous research to gain a better understanding of how consumers evaluate information disclosure (Gutierrez et al., 2019; Pentina et al., 2016; Sun et al., 2015; Xu et al., 2011). For the context of this research, this means, how consumers evaluate benefits of anthropomorphised interaction and personalisation versus privacy risks and the question how this assessment influences their attitudes and information disclosure intentions. PPP and privacy calculus offer therefore a useful theoretical foundation to examine how consumers evaluate the benefits of personalisation against the risks of privacy concerns and its implications on consumer information disclosure.

3.2 Proposed conceptual model and research propositions

Based on the discussion of PCT and the concept of anthropomorphism, a conceptual research model is proposed that examines consumer interaction with AI-based digital assistants and their information disclosure towards AI-based digital assistants in the light of privacy concerns. The conceptual model is depicted in Figure 1 below. The model builds upon PCT and anthropomorphism to explain consumer intentions to disclose information and includes eight RPs. In the following paragraphs, the RPs will be discussed and presented.

3.2.1 Anthropomorphism of artificial intelligence.

Anthropomorphism has received considerable attention in the marketing and consumer behaviour literature because it represents an opportunity for marketers to affect consumption or affect consumer behaviour related to consumption (Epley, 2018). Anthropomorphism occurs when humans perceive other things as like themselves (Hume, 1757/1957) or “imbue real or imagined behaviour of nonhuman agents with humanlike characteristics, motivations, intentions or emotions” (Epley et al., 2007, p. 864) which will influence how humans react to those objects. The concept of anthropomorphism is relevant in the context of AI because companies try to humanise their technology by assigning them humanoid traits, e.g. looks, name and voice, which has the aim to encourage customers to converse and bond with them (Novak and Hoffman, 2019; Schweitzer et al., 2019; Steinhoff et al., 2019; van Doorn et al., 2017). Across academic literature, there is a debate to whether anthropomorphism positively affects consumer attitudes, intentions and use of technology, or whether is serves as impeding variable. For instance, previous studies find that anthropomorphism has negative effects on consumer attitudes towards AI and is an impeding determinant of consumers´ acceptance of service robots due to the reason that they pose a threat to users’ identity (Gursoy et al., 2019; Lu et al., 2019). Here, reference to the uncanny valley is made (Mori, 1970) which describes the effects of how more humanlike features lead to positive evaluations, until too humanlike features result in eeriness and anxiety in response to the technology. In contrast to that, it has been suggested that anthropomorphism can improve levels of engagement, collaboration and interaction (Steinhoff et al., 2019; Yang et al., 2022).

For this study, it is assumed that anthropomorphism of AI technology has the potential to nudge consumers to greater self-disclosure of data due to the circumstance that it will transmit social cues to consumers that activate social scripts (Thomaz et al., 2020). Previous research has proven that consumers treat computers like social actors if they display a minimum of social cues, such as asking questions or sharing information with the consumer and that consumers feel obliged to respond with also sharing information due to learned social scripts (Nass and Moon, 2000). Based on this and the insights derived from existing research, it is proposed that anthropomorphism of AI will affect consumer information disclosure positively, as summarised in the first RP as follows:

RP1.

Anthropomorphism of AI positively affects consumers’ willingness to disclose information.

Anthropomorphism is a critical construct to better understand human interactions with technology (van Doorn et al., 2017). The authors point out that a product with a smiley face is perceived as friendlier than a product that looks techy. To make AI appear friendlier, designers can assign anthropomorphic appearances, and even create an identity for AI-systems (Ene et al., 2019). Moreover, anthropomorphism can influence consumers to perceive AI as more vivid and potentially treat AI like a sentient being, which has been proven to increase psychological warmth and positively affects consumers attitudes (Kim et al., 2019). Thus, it is proposed that:

RP2.

Anthropomorphism of AI positively affects consumers’ attitudes towards the digital assistant.

3.2.2 Privacy concerns.

Privacy is defined as “the ability of the individual to control the terms under which personal information is acquired and used” (Westin, 1967, p. 7) and privacy in the online environment as “an individual’s control and awareness of the collection and usage or personal data” (Hann et al., 2007). In the marketing context, discussions around information disclosure and privacy have in the past been influenced by an understanding of privacy as personal responsibility, which requires individuals to exercise control over the collection, use and distribution of their personal data (Pomfret et al., 2020), what contrasts with a philosophical approach to data privacy and data protection which assumes that individuals have a fundamental right to “be left alone”.

A plethora of studies across a variety of academic disciplines has examined the relationship between privacy concerns and privacy-related behaviours (Acquisti et al., 2015; Baruh and Popescu, 2017; Martin and Murphy, 2019). In the e-commerce context, it has been verified that perceived privacy concerns negatively affect usage behaviour (Vilmakumar et al., 2021). Moreover, privacy concerns are a major hindrance to data sharing in an online environment and limit the potential for e-commerce to grow; Angst and Agarwal (2009) and Buchanan et al. (2007) highlight that. In addition, it is argued that when consumers have privacy concerns or perceive threats to their personal information, consumers will withhold personal information, refuse participation in the online environment or refuse to disclose their personal information.

Consumer perceptions of privacy are an important influence on his or her information disclosure when consumers interact with a website (Lowry et al., 2012; Shah et al., 2014). Doubts about consumers’ privacy result in consumer hesitancy to transact in e-commerce while privacy assurances are positively affecting consumers’ behavioural intentions towards a website (Lowry et al., 2012). It has been suggested that consumers have a more positive perception towards information disclosure if privacy risks are low (Sun et al., 2015). A meta-analysis on 166 studies between 1990 and 2016 on privacy concerns found that privacy concerns predict information sharing (Baruh and Popescu, 2017). However, researchers have yet not adequately explained how consumers’ attitudes around privacy influences the types of information disclosed and the level of disclosure (Pomfret et al., 2020). Another study finds that people who believe technology to be a potential threat to their privacy also demonstrate higher concerns about the technology itself, while people who believe it to be useful, will care less about privacy concerns (Vilmakumar et al., 2021). These findings are in line with PCT. Consumers who have higher privacy concerns will perceive greater risks associated with their personal information being compromised or misused, which will affect their intentions to disclose private information negatively. Consequently, it is proposed that:

RP3.

Privacy concerns negatively affects consumers’ intentions to disclose information.

Apart from that, privacy has been found to be an important variable regarding the adoption of innovative technologies (Kumar et al., 2016; Gutierrez et al., 2019). Privacy concerns also have a negative impact on consumers’ attitude and intention to use an innovation (Müller-Seitz et al., 2009; Mani and Chouk, 2017) due to perceived intrusiveness. It can be observed that privacy concerns are negative attitudes or beliefs that influence an individual’s attitudes, for instance, towards a technology, and usually centre around the collection of personal data, data errors, unauthorised access of personal data and unauthorised secondary use of personal data (Dinev et al., 2016). Thus, it is proposed that:

RP4.

Privacy concerns negatively affect consumers’ attitude towards the digital assistant.

3.2.3 Personalisation.

AI agents such as Siri have become rapidly embraced and integrated into daily life by consumers because they enable individuals to access timely and useful information (Canbek and Mutlu, 2016). These agents meet customer demand for contextually relevant and highly personalised content that is delivered in real time (Brill et al., 2019). It is argued that companies use AI to collect user details with the aim to improve the user experience and enhance lifetime value of customers (Shankar, 2018; Wilson and Daugherty, 2018). Due to the circumstance, that with AI, data about every individual consumer can be stored and analysed at unprecedented scale, marketers can now personalise their marketing mix for each individual consumer. With the potential offered by AI, the goal of organisations is shifting now towards personalising interactions, delighting its audience and understanding each individual customer’s unique challenges (Pansari and Kumar, 2017).

Data sharing is an opportunity for consumers to participate in an exchange of data for retail value and personalized services (Barth and de Jong, 2017) Consequently, it is suggested that consumers react positively to personalisation because of improved experience and perceptions of being better understood by a company. It is hypothesized that this understanding will lead to:

RP5.

Personalisation positively affects consumers’ intentions to disclose information.

Personalisation has also led to higher click-through rates of advertisements (Aguirre et al., 2015; Bleier and Eisenbeiss, 2015). Personalisation results in more positive attitudes and higher click through rates. However, it should be noted that these studies also found that too much personalisation will impede trust and raise privacy concerns, which again reflects and highlights the relevance of the PPP. Nonetheless, it is proposed that:

RP6.

Personalisation positively affects consumers’ attitude towards the digital assistant.

3.2.4 Attitude towards AI-based digital agents.

As suggested by the theory of planned behaviour (Ajzen, 1991), attitudes influence behavioural intentions. Above discussion has highlighted, how it is expected that privacy concerns, personalisation and anthropomorphism affect consumers´ attitude towards a conversational agent. Combining those arguments with Ajzen’s theory, it is expected that attitudes towards AI has a mediating effect on the relationship between antecedents and intentions to disclose information to the conversational agent. Thus, it is proposed:

RP7.

Attitude towards the digital assistant mediates the effect of privacy concerns, personalisation and anthropomorphism on consumers’ intention to disclose information.

4. Discussion

The review of existing literature has shown that consumers use AI technology in a variety of forms and for a variety of different tasks in their daily life (Brill et al., 2019). However, the success of AI depends on the availability of vast amounts of consumer data to offer benefits such as personalised content or discounts (Zhu et al., 2017). The PPP and PCT offer a valuable theoretical foundation in order to assess how AI can influence consumer information disclosure. The proposed research model and seven RPs suggest positive effects of anthropomorphism and personalisation on consumer attitude towards an intelligent digital assistant and positive effects on consumer information disclosure. Despite a lack of consensus across academic literature regarding the effects of anthropomorphism, it is argued that anthropomorphising AI will have positive effects such as encouraging consumers to like it more (Kim et al., 2019). Anthropomorphising AI by assigning it a name or character can nudge consumers to connect with them on a more personal level. This implies that if marketing managers are able to anthropomorphise their AI technology, consumers will evaluate them more favourably (Epley, 2018; Kim et al., 2019; Novak and Hoffman, 2019; Schweitzer et al., 2019). Anthropomorphism promises to overcome consumer hesitancy towards technology and engage with AI-based digital assistants on a personal level, offering companies opportunities for data collection and data curation for more meaningful value exchange and personalised offerings.

The next two RPs addressed the topic of privacy concerns relating to the collection, storage and analysis of consumer data through AI. Based on existing literature, it is suggested that privacy concerns negatively affect consumers evaluations of digital assistants, and negatively influence consumer intentions towards information disclosure to the digital assistant (Kumar et al., 2016; Gutierrez et al., 2019). It becomes evident that the topic of privacy can form a critical barrier to the success of AI-based digital assistants due to the circumstance that they depend on the presence and provision of personal and behavioural data in order to assist the consumer with personalised information during an interaction. Thomaz et al. (2020) content that society will witness a shift in the nature of the internet as privacy concerns and awareness about the use of private data is becoming more prevalent, which will significantly impact marketing as it benefits of wide availability of customer data as source of information for marketing strategy. Because marketing AI applications such as conversational agents are based on the use of personal information (Rai, 2020), it becomes necessary to examining the relationship of internet privacy concerns on consumer attitudes and consumer behaviour in this context of AI.

In addition to this, this study has drawn upon PPP and PCT to consider how personalisation offers benefits such as personalised content or discounts (Zhu et al., 2017) can potentially outweigh the perceived risks around personal data and privacy concerns and positively affect attitudes and intentions to disclose information. It is argued that through the provision of real-time, personalised and contextually relevant curation of information, consumers evaluate digital assistants more positively and provide their information (Brill et al., 2019; Canbek and Mutlu, 2016).

Finally, this research highlights several gaps in academic literature relating to the effects of AI and the potential of AI to encourage consumer information disclosure. A detailed presentation of avenues for future research will be presented in Table 2 in the following section.

5. Conclusions, limitations and future research

Research into the effects of AI in a consumer context is still developing and sparse (Davenport et al., 2020). This research has addressed this gap and considered consumer information disclosure in the context of AI. A conceptual model with seven RPs has been developed, which can serve as foundation for future research. The following section will summarise insights from this paper and present a table with avenues for future research.

5.1 Conclusions

This study advances our understanding of the potential that AI offers in encouraging consumers to share their private information, which is critical for further advancements in AI that can only continue if companies have access to large amounts of data and computational capabilities to analyse them (Mazurek and Malagocka, 2019). Re-considering the benefits offered through increased personalisation against privacy concerns for the purpose of information disclosure in the light of AI-based digital assistants is critical due to the increasing integration of AI into marketing strategy and peoples’ lives. Research pertaining to the effects of AI and AI-based technology is still at early stages and focuses heavily on technology adoption. Understanding the implications of AI in marketing and its impact on consumer behaviour becomes increasingly important to examine by researchers due to the circumstance that AI will systematically and effectively change the way in which companies connect and interact with their customers (Davenport et al., 2020; Grewal et al., 2020; Yadav and Pavlou, 2020). This research has developed a conceptual model and seven RPs that promise to a research gap and contribute to a better understanding of the implications of AI in the consumer sphere.

5.2 Limitations and future research

As with any research, this study is also subject to limitations that must be acknowledged and taken into consideration. Firstly, this research is conceptual in nature and has developed a research model (Ismagilova et al., 2020; Hughes et al., 2020; Kumar et al., 2018; Mathivathanan et al., 2022; Mishra et al., 2017; Rana et al., 2011, 2012) and RPs on an extant review of academic literature relating to AI, privacy calculus and consumer information disclosure. While this paper has offered new insights into the potential effects of AI on consumer attitudes and consumer information disclosure, it is even more important to test the proposed research model and its RPs. Hence, future research should consider the propositions presented in this research and test the conceptual model empirically. In addition, Table 2 presents areas for future research and potential research questions for future studies to consider.

Above table outlines three areas for future research. Firstly, future research can further investigate the PPP for the context of AI. For instance, research could ask how far personalisation can go before it becomes too intrusive and whether there are individual consumer characteristics, such as age, gender or culture, that influence this evaluation. Secondly, the European Union introduced GDPR (General Data Protection Regulation) in 2018, which aims to limit the way in which personal information can be stored and processed (Haenlein and Kaplan, 2019). The main task of this regulation is to strengthen the rights of the individual, making the expression of informed consent to disclose information mandatory as well as giving the opportunity to withdrawing consent (Mazurek and Malagocka, 2019). However, it is not clear whether consumers are actively aware of such legislation, or whether it influences their privacy and disclosure behaviour. Moreover, the potential “dark side” of AI and AI integration into marketing strategy is a research area that remains less studied (Grewal et al., 2021; Rana et al., 2022), consequently, there are opportunities for future research to investigate how AI could undermine consumer privacy. Finally, future research could differentiate between different types of AI technology. Finally, the future researchers can also review the literature (Alalwan et al., 2017; Alryalat et al., 2017; Dwivedi et al., 2017a, 2017b, 2021; Kizgin et al., 2020; Patil et al., 2017; Rana et al., 2014) on the role of AI on consumers’ information disclosure.

Figures

Proposed research model

Figure 1.

Proposed research model

Studies examining the interaction between consumers and AI

Authors, Theoretical foundation and research aim Context and key variables Findings Limitations
Canbek and Mutlu (2016) examine the potential use of intelligent personal assistants for learning Education context Use Intelligent personal assistants such as Siri, Cortana and Google Now can be used to support learning in educational context Focus on smartphone-based assistants only
Based on the technology acceptance model (TAM), Moriuchi (2019) examines the effect of AI integration into e-commerce on customer engagement and loyalty Online shopping context
Perceived ease of use Perceived usefulness
Strong support for the TAM variables perceived usefulness and perceived ease of use on loyalty. Engagement serves as mediator in this relationship Functional antecedents only
Based on expectations confirmation theory, Brill et al. (2019) examine customer satisfaction with digital assistants No context specified Expectations
Satisfaction
Expectations and confirmed expectations have a positive influence on satisfaction with digital assistants Study only included consumers who continued to use digital assistants or had no experience. Discontinuance was not considered. Impact on consumer behaviour is not examined
Buhalis and Sinarta (2019) analyse value co-creation through big data, real-time data mining and contextual data by smart technology and its effect on customer experience Tourism context
Interaction
Co-creation
Interaction based on real-time data and contextual information as well as nowness are ways to co-create value with the consumer. Real-time service adds to a firm’s competitiveness Research is based on “best practices” within tourism context
Based on the uncanny valley theory, Kim et al. (2019) examine the effects of anthropomorphism of robots on consumer judgements and attitudes Consumer robots
Anthropomorphism Attitude Uncanniness
Anthropomorphism positively influences perceptions of warmth. Study also supports the uncanny valley theory by showing that too-humanlike features lead to uncanniness and negative evaluations Study only examines the effect on attitudes, not how this attitude influences behaviour
Based on the unified theory of acceptance and use of technology (UTAUT2), Lu et al. (2019) conceptualise and test a service robot integration willingness scale to determine key dimensions in consumer willingness to adopt AI-based technology into service encounters Service context
Adoption
Anthropomorphism
Performance efficacy, intrinsic motivation, facilitating conditions and emotions are positive determinants of acceptance Anthropomorphism is a barrier to adoption of service robots due to perceived threat to human identity Study does not consider cultural or individual differences.
Study relies on “overused” adoption theory
Based on the service robot acceptance model (sRAM), Fernandes and Oliveira (2021) examine consumer motivations to adopt AI-based digital voice assistants into service encounters Service context
Adoption
Functional, social and relational elements are drivers for adoption of AI-based digital voice assistants. Experience has a moderating role on acceptance Convenience sample focused on young users of the Millennial generation.
Potential inhibitors of acceptance are not included
Based on UTAUT2 and privacy calculus, Vilmakumar et al. (2021) examine the perception of consumers towards privacy concerns and its influence on the adoption of voice based digital assistants Indian consumer context
Privacy Adoption
Consumer perceived privacy risk does not influence adoption intentions directly but indirectly through trust No individual consumer variables considered
Limited to Indian consumers
Belanche et al. (2021) validate the humanness-value-loyalty model to examine how a robot’s perceived human likeness, competence and warmth affect service value expectations and loyalty Restaurant service context
Human likeness Competence
Human likeness positively affects utilitarian, social, monetary and emotional value expectations. Competence influences utilitarian value expectations Limited to specific service robots within restaurant context

Areas for future research and potential research questions

Area for future research Potential research questions
Personalisation-privacy paradox • How much personalisation is acceptable before it becomes too intrusive?
• What variables (e.g. gender, age, culture) affect an individual’s acceptance level of personalisation?
• Does personalisation influence privacy concerns?
Privacy regulations and legislation, e.g. GDPR • How does legislation to protect privacy affect consumers’ information disclosure?
• What effect does the introduction of legislation such as GDPR have on consumers’ privacy evaluations?
• Are consumers aware of regulation to protect their privacy?
The “dark” side of AI invading privacy • How can AI encourage consumers to share their data against their intentions?
• Can AI elicit information consumers are only subconsciously aware of?
Complimentary AI technology (e.g. Siri as digital assistant) vs purchased AI technology (e.g. Amazon Echo) • How does consumer attitudes and use differ for complimentary AI technology vs purchased AI technology?
• Do consumers perceive AI as invading their privacy?

References

Acquisti, A., Brandimarte, L. and Loewenstein, G. (2015), “Privacy and human behaviour in the age of information”, Science, Vol. 347 No. 6221, pp. 509-514.

Aguirre, E., Mahr, D., Grewal, D., de Ruyter, K. and Wetzels, M. (2015), “Unravelling the personalization paradox: the effect of information collection and trust-building strategies on online advertisement effectiveness”, Journal of Retailing, Vol. 91 No. 1, pp. 34-49.

Ajzen, I. (1991), “The theory of planned behaviour”, Organizational Behavior and Human Decision Processes, Vol. 50 No. 2, pp. 179-211.

Alalwan, A.A., Rana, N.P., Dwivedi, Y.K. and Algharabat, R.S. (2017), “Social media in marketing: a review and analysis of the existing literature”, Telematics and Informatics, Vol. 34 No. 7, pp. 1177-1190.

Alryalat, M., Rana, N.P., Sahu, G.P., Dwivedi, Y.K. and Tajvidi, M. (2017), “Use of social media in citizen-centric electronic government services: a literature analysis”, International Journal of Electronic Government Research, Vol. 13 No. 3, pp. 55-79.

Angst, C. and Agarwal, R. (2009), “Adoption of electronic health records in the presence of privacy concerns: the elaboration likelihood model and individual persuasion”, MIS Quarterly, Vol. 33 No. 2, pp. 339-370.

Barth, S. and de Jong, M. (2017), “The privacy paradox: investigating discrepancies between expressed privacy concerns and actual online behavior – a systematic literature review”, Telematics and Informatics, Vol. 34 No. 7, pp. 1038-1058.

Baruh, L. and Popescu, M. (2017), “Big data analytics and the limits of privacy self-management”, New Media and Society, Vol. 19 No. 4, pp. 579-596.

Belanche, D., Casaló, L.V., Schepers, J. and Flavián, C. (2021), “Examining the effects of robots' physical appearance, warmth, and competence in frontline services: the humanness‐value‐loyalty model”, Psychology and Marketing, Vol. 38 No. 12, pp. 2357-2376.

Bleier, A. and Eisenbeiss, M. (2015), “The importance of trust for personalized online advertising”, Journal of Retailing, Vol. 91 No. 3, pp. 390-409.

Brill, T., Munoz, L. and Miller, R. (2019), “Siri, alexa, and other digital assistants: a study of customer satisfaction with artificial intelligence applications”, Journal of Marketing Management, Vol. 35 Nos 15/16, pp. 1401-1436.

Buchanan, T., Paine, C., Joinson, A. and Reips, U.-D. (2007), “Development of measures of online privacy concern and protection for use on the internet”, Journal of the American Society for Information Science and Technology, Vol. 58 No. 2, pp. 157-165.

Buhalis, D. and Sinarta, Y. (2019), “Real-time co-creation and nowness service: lessons from tourism and hospitality”, Journal of Travel and Tourism Marketing, Vol. 36 No. 5, pp. 563-582.

Canbek, N. and Mutlu, M. (2016), “On the track of artificial intelligence: learning with intelligent personal assistants”, International Journal of Human Sciences, Vol. 13 No. 1, pp. 592-601.

Chui, M., Manyika, J., Miremadi, M., Henke, N., Chung, R., Nel, P. and Malhotra, S. (2018), “Notes from the AI frontier: applications and value of deep learning”, McKinsey and Company, Discussion Paper (April 2018), available at: www.mckinsey.com/featured-insights/artificial-intelligence/notes-from-the-ai-frontier-applications-and-value-of-deep-learning (accessed 05 october 2022).

Cloarec, J. (2020), “The personalisation-privacy paradox in the attention economy”, Technological Forecasting and Social Change, Vol. 161, p. 120299.

Davenport, T., Guha, A., Grewal, D. and Bressgott, T. (2020), “How artificial intelligence will change the future of marketing”, Journal of the Academy of Marketing Science, Vol. 48 No. 1, pp. 24-42.

Dinev, T. and Hart, P. (2006), “An extended privacy calculus model for e-commerce transactions”, Information Systems Research, Vol. 17 No. 1, pp. 61-80.

Dinev, T., Albano, V., Xu, H., D´Atri, A. and Hart, P. (2016), “Individuals´ attitudes towards electronic health records: a privacy calculus perspective”, in Gupta, A., Patel, V. and Greenes, R. (Eds), Advances in Healthcare Informatics and Analytics. Annals of Information Systems, Vol 19. Springer, Cham.

Duan, Y., Edwards, J. and Dwivedi, Y. (2019), “Artificial intelligence for decision making in the era of big data – evolution, challenges and research agenda”, International Journal of Information Management, Vol. 48, pp. 63-71.

Dwivedi, Y.K., Ismagilova, E., Rana, N.P. and Raman, R. (2021), “Social media adoption, usage and impact in business-to-business (B2B) context: a state-of-the-art literature review”, Information Systems Frontiers, doi: 10.1007/s10796-021-10106-y.

Dwivedi, Y.K., Rana, N.P. and Al-Ryalat, M. (2017a), “Affiliate marketing: an overview and analysis of emerging literature”, The Marketing Review, Vol. 17 No. 1, pp. 33-50.

Dwivedi, Y.K., Rana, N.P., Jeyaraj, A., Clement, M. and Williams, M.D. (2017b), “Re-examining the unified theory of acceptance and use of technology (UTAUT): towards a revised theoretical model”, Information Systems Frontiers, Vol. 21 No. 3, pp. 719-734.

Ene, I., Pop, M. and Nistoreanu, B. (2019), “Qualitative and quantitative analysis of consumer perception regarding anthropomorphic AI design”, Proceedings of the 13th International Conference on Business Excellence.

Epley, N. (2018), “A mind like mine: the exceptionally ordinary underpinnings of anthropomorphism”, Journal of the Association for Consumer Research, Vol. 3 No. 4, pp. 591-598.

Epley, N., Waytz, A. and Cacioppo, J. (2007), “On seeing human: a three factor theory of anthropomorphism”, Psychological Review, Vol. 114 No. 4, pp. 864-886.

Fernandes, T. and Oliveira, E. (2021), “Understanding consumers´ acceptance of automated technologies in service encounters: drivers of digital assistant adoption”, Journal of Business Research, Vol. 122, pp. 180-192.

Fitzgerald, K. (2019), “In the ‘Opt-In’ data economy, consumer confidence is key”, Forbes, available at: www.forbes.com/sites/forbestechcouncil/2019/01/16/in-the-opt-in-data-economy-consumer-confidence-is-key/ (accessed 7 June 2021).

Foehr, J. and Germelmann, C. (2020), “Alexa, can I trust you? Exploring consumer paths to trust in smart voice-interaction technologies”, Journal of the Association for Consumer Research, Vol. 5 No. 2, pp. 181-205.

Grewal, D., Hulland, J., Kopalle, P. and Karahanna, E. (2020), “The future of technology and marketing: a multidisciplinary perspective”, Journal of the Academy of Marketing Science, Vol. 48 No. 1, pp. 1-8.

Grewal, D., Guha, A., Satornino, C. and Schweiger, E. (2021), “Artificial intelligence: the light and the darkness”, Journal of Business Research, Vol. 136, pp. 229-236.

Gursoy, D., Chi, O., Lu, L. and Nunkoo, R. (2019), “Consumers acceptance of artificially intelligent (AI) device use in service delivery”, International Journal of Information Management, Vol. 49, pp. 157-169.

Gutierrez, A., O'Leary, S., Rana, N., Dwivedi, Y. and Calle, T. (2019), “Using privacy calculus theory to explore entrepreneurial directions in mobile location-based advertising: identifying intrusiveness as the critical risk factor”, Computers in Human Behaviour, Vol. 95, pp. 295-306.

Haenlein, M. and Kaplan, A. (2019), “A brief history of artificial intelligence: on the past, present, and future of artificial intelligence”, California Management Review, Vol. 61 No. 4, pp. 5-14.

Hann, H., Hui, K., Lee, S. and Png, I. (2007), “Overcoming online information privacy concerns: an information-processing theory approach”, Journal of Management Information Systems, Vol. 24 No. 2, pp. 13-42.

Hofacker, C. and Corsaro, D. (2020), “Dystopia and utopia in digital services”, Journal of Marketing Management, Vol. 36 Nos 5/6, pp. 412-419.

Hong, W. and Thong, J. (2013), “Internet privacy concerns: an integrated conceptualisation and four empirical studies”, MIS Quarterly, Vol. 37 No. 1, pp. 275-298.

Hughes, D.L., Dwivedi, Y.K. and Rana, N.P. (2020), “Elucidation of IS/IT project success factors – an interpretive structural modelling approach”, Annals of Operations Research, Vol. 285 Nos 1/2, pp. 35-66.

Hume, D. (1757/1957), The Natural History of Religion, University Press, Stanford.

Ismagilova, E., Rana, N.P., Slade, E. and Dwivedi, Y.K. (2020), “A meta-analysis of the factors affecting eWOM providing behaviour”, European Journal of Marketing, Vol. 55 No. 4, pp. 1067-1102.

Kietzmann, J., Paschen, J. and Treen, E. (2018), “Artificial intelligence in advertising: how marketers can leverage artificial intelligence along the consumer journey”, Journal of Advertising Research, Vol. 58 No. 3, pp. 263-267.

Kim, S., Schmitt, B. and Thalmann, N. (2019), “Eliza in the uncanny valley: anthropomorphizing consumer robots increases their perceived warmth but decreases liking”, Marketing Letters, Vol. 30 No. 1, pp. 1-12.

Kizgin, H., Dey, B., Dwivedi, Y.K., Hughes, L., Jamal, A., Jones, P., Kronemann, B., Laroche, M., Peñaloza, L., Richard, M.-O., Romer, R., Rana, N.P., Tamilmani, K. and Williams, M.D. (2020), “The impact of social media on consumer acculturation: current challenges, opportunities, and agenda for research and practice”, International Journal of Information Management, Vol. 51, pp. 1-15.

Kumar, V., Dixit, A., Javalgi, R. and Dass, M. (2016), “Research framework, strategies and applications of intelligent agent technologies (IATs) in marketing”, Journal of the Academy of Marketing Science, Vol. 44 No. 1, pp. 24-45.

Kumar, A., Mangla, S.K., Luthra, S., Rana, N.P. and Dwivedi, Y.K. (2018), “Predicting changing pattern: building model for consumer decision making in digital market”, Journal of Enterprise Information Management, Vol. 31 No. 5, pp. 674-703.

Lariviere, B., Bowen, D., Andreassen, T., Kunz, W., Sirianni, N., Voss, C. and De Keyer, A. (2017), “Service encounter 2-0: an investigation into the roles of technology, employees and customers”, Journal of Business Research, Vol. 79, pp. 238-246.

Lowry, P., Moody, G., Vance, A., Jensen, M., Jenkins, J. and Wells, T. (2012), “Using an elaboration likelihood approach to better understand the persuasiveness of website privacy assurance cues for online consumers”, Journal of the American Society for Information Science and Technology, Vol. 63 No. 4, pp. 755-776.

Lu, L., Cai, R. and Gursoy, D. (2019), “Developing and validating a service robot integration willingness scale”, International Journal of Hospitality Management, Vol. 80, pp. 36-51.

Mani, Z. and Chouk, I. (2017), “Drivers of consumers´ resistance to smart products”, Journal of Marketing Management, Vol. 33 Nos 1/2, pp. 76-97.

Martin, K. and Murphy, P. (2019), “The role of data privacy in marketing”, Journal of the Academy of Marketing Science, Vol. 45 No. 2, pp. 135-155.

Mathivathanan, D., Mathiyazhagan, K., Khorana, S., Rana, N.P. and Arora, B. (2022), “Drivers of circular economy for small and medium enterprises: a case study on the indian state of Tamil Nadu”, Journal of Business Research, Vol. 149, pp. 997-1015.

Mazurek, G. and Malagocka, K. (2019), “Perception of privacy and data protection in the context of the development of artificial intelligence”, Journal of Management Analytics, Vol. 6 No. 4, pp. 344-364.

Mishra, N., Singh, A., Rana, N.P. and Dwivedi, Y.K. (2017), “Interpretive structural modelling and fuzzy MICMAC approaches for customer centric beef supply chain: application of a big data technique”, Production Planning and Control, Vol. 28 Nos 11/12, pp. 945-963.

Moriuchi, E. (2019), “Okay google!: An empirical study on voice assistants on consumer engagement and loyalty”, Psychology and Marketing, Vol. 36 No. 5, pp. 489-501.

Müller-Seitz, G., Kautzenberg, K., Creusen, U. and Stromereder, C. (2009), “Customer acceptance of RFID technology: evidence from the german electronic retail sector”, Journal of Retailing and Consumer Services, Vol. 16 No. 1, pp. 31-39.

Nass, C. and Moon, Y. (2000), “Machines and mindlessness: social responses to computers”, Journal of Social Issues, Vol. 56 No. 1, pp. 81-103.

Novak, T. and Hoffman, D. (2019), “Relationship journeys in the internet of things: a new framework for understanding interactions between consumers and smart objects”, Journal of the Academy of Marketing Science, Vol. 47 No. 2, pp. 216-237.

Pansari, A. and Kumar, V. (2017), “Customer engagement: the construct, antecedents and consequences”, Journal of the Academy of Marketing Science, Vol. 45 No. 3, pp. 294-311.

Pappas, N. (2016), “Marketing strategies, perceived risks, and consumer trust in online buying behaviour”, Journal of Retailing and Consumer Services, Vol. 29, pp. 92-103.

Patil, P., Dwivedi, Y.K. and Rana, N.P. (2017), “Digital payments adoption: an analysis of literature”, I3E 2017, New Delhi, India.

Pentina, I., Zhang, L., Bata, H. and Chen, Y. (2016), “Exploring privacy paradox in information-sensitive mobile app adoption: a cross-cultural comparison”, Computers in Human Behavior, Vol. 65, pp. 409-419.

Pomfret, L., Previte, J. and Coote, L. (2020), “Beyond concern: socio-demographic and attitudinal influences on privacy and disclosure choices”, Journal of Marketing Management, Vol. 36 Nos 5/6, pp. 519-549.

Rai, A. (2020), “Explainable AI: from black box to glass box”, Journal of the Academy of Marketing Science, Vol. 48 No. 1, pp. 137-141.

Ramaswamy, V. and Ozcan, K. (2018), “Offerings as digitalised interactive platforms: a conceptual framework and implications”, Journal of Marketing, Vol. 82 No. 4, pp. 19-31.

Rana, N.P., Chatterjee, S., Dwivedi, Y.K. and Akter, S. (2022), “Understanding dark side of artificial intelligence (AI) integrated business analytics: assessing firm’s operational inefficiency and competitiveness”, European Journal of Information Systems, Vol. 31 No. 3, pp. 364-387.

Rana, N.P., Dwivedi, Y.K. and Weerakkody, V. (2014), “Profiling existing research on social innovations in the public sector context”, Information Systems Management, Vol. 31 No. 3, pp. 259-273.

Rana, N.P., Williams, M.D., Dwivedi, Y.K. and Williams, J. (2011), “Reflection on e-government research: toward a taxonomy of theories and theoretical constructs”, International Journal of Electronic Government Research, Vol. 7 No. 4, pp. 64-88.

Rana, N.P., Williams, M.D., Dwivedi, Y.K. and Williams, J. (2012), “Theories and theoretical models for examining the adoption of e-government services”, E-Services Journal, Vol. 8 No. 2, pp. 26-56.

Schweitzer, F., Belk, R., Jordan, W. and Ortner, M. (2019), “Servant, friend or master? The relationships users build with voice-controlled smart devices”, Journal of Marketing Management, Vol. 35 Nos 7/8, pp. 693-715.

Shah, M., Peikari, H. and Yasin, N. (2014), “The determinants of individuals´ perceived e-security: evidence from Malaysia”, International Journal of Information Management, Vol. 34 No. 1, pp. 48-57.

Shankar, V. (2018), “How artificial intelligence (AI) is reshaping retailing”, Journal of Retailing, Vol. 94 No. 4, pp. 6-11.

Steinhoff, L., Arli, D., Weaven, S. and Kozlenkova, I. (2019), “Online relationship marketing”, Journal of the Academy of Marketing Science, Vol. 47 No. 3, pp. 369-393.

Sun, Y., Wang, N., Shen, X. and Zhang, J. (2015), “Location information disclosure in location-based social network services: privacy calculus, benefit structure, and gender differences”, Computers in Human Behavior, Vol. 52, pp. 278-292.

Thomaz, F., Salge, C., Karahanna, E. and Hulland, J. (2020), “Learning from the dark web: leveraging conversational agents in the era of hyper-privacy to enhance marketing”, Journal of the Academy of Marketing Science, Vol. 48 No. 1, pp. 43-63.

Van Doorn, J., Mende, M., Noble, S., Hulland, J., Ostrom, A., Grewal, D. and Petersen, J. (2017), “Domo Arigato Mr Roboto: emergence of automated social presence in organizational frontlines and customers’ service experiences”, Journal of Service Research, Vol. 20 No. 1, pp. 43-58.

Vilmakumar, M., Sharma, S., Singh, J. and Dwivedi, Y. (2021), “Okay Google, what about my privacy? User‘s privacy perceptions and acceptance of voice based digital assistants”, Computers in Human Behavior, Vol. 120.

Wang, T., Duong, T. and Chen, C. (2016), “Intention to disclose personal information via mobile applications: a privacy calculus perspective”, International Journal of Information Management, Vol. 36 No. 4, pp. 531-542.

Westin, A. (1967), Privacy and Freedom, Atheneum, New York, NY.

Wilson, H. and Daugherty, P. (2018), “Collaborative intelligence: humans and AI are joining forces”, Harvard Business Review, Vol. 96 No. 4, pp. 114-123.

Xu, H., Luo, X., Carroll, J. and Rosson, M. (2011), “The personalisation privacy paradox: an exploratory study of decision-making process for location-aware marketing”, Decision Support Systems, Vol. 51 No. 1, pp. 42-52.

Yadav, M. and Pavlou, P. (2020), “Technology-enabled interaction in digital environments: a conceptual foundation for current and future research”, Journal of the Academy of Marketing Science, Vol. 48 No. 1, pp. 132-136.

Yang, Y., Liu, Y., Lv, X., Ai, J. and Li, Y. (2022), “Anthropomorphism and customers’ willingness to use artificial intelligent service agents”, Journal of Hospitality Marketing & Management, Vol. 31 No. 1, pp. 1-23.

Zhu, H., Ou, C., van den Heuvel, W. and Liu, H. (2017), “Privacy calculus and its utilitiy for personalization in services in e-commerce: an analysis of consumer decision-making”, Information & Management, Vol. 54, pp. 427-437.

Corresponding author

Hatice Kizgin can be contacted at: h.kizgin@gmail.com

Related articles