Skip to main content
About HEC About HEC
Summer School Summer School
Faculty & Research Faculty & Research
Master’s programs Master’s programs
Bachelor Programs Bachelor Programs
MBA Programs MBA Programs
PhD Program PhD Program
Executive Education Executive Education
HEC Online HEC Online
About HEC
Overview Overview
Who
We Are
Who
We Are
Egalité des chances Egalité des chances
HEC Talents HEC Talents
International International
Campus
Life
Campus
Life
Sustainability Sustainability
Diversity
& Inclusion
Diversity
& Inclusion
Stories Stories
The HEC
Foundation
The HEC
Foundation
Summer School
Youth Programs Youth Programs
Summer programs Summer programs
Online Programs Online Programs
Faculty & Research
Overview Overview
Faculty Directory Faculty Directory
Departments Departments
Centers Centers
Chairs Chairs
Grants Grants
Knowledge@HEC Knowledge@HEC
Master’s programs
Master in
Management
Master in
Management
Master's
Programs
Master's
Programs
Double Degree
Programs
Double Degree
Programs
Bachelor
Programs
Bachelor
Programs
Summer
Programs
Summer
Programs
Exchange
students
Exchange
students
Student
Life
Student
Life
Our
Difference
Our
Difference
Bachelor Programs
Overview Overview
Course content Course content
Admissions Admissions
Fees and Financing Fees and Financing
MBA Programs
MBA MBA
Executive MBA Executive MBA
TRIUM EMBA TRIUM EMBA
PhD Program
Overview Overview
HEC Difference HEC Difference
Program details Program details
Research areas Research areas
HEC Community HEC Community
Placement Placement
Job Market Job Market
Admissions Admissions
Financing Financing
Executive Education
Home Home
About us About us
Management topics Management topics
Open Programs Open Programs
Custom Programs Custom Programs
Events/News Events/News
Contacts Contacts
HEC Online
Overview Overview
Degree Program Degree Program
Executive certificates Executive certificates
MOOCs MOOCs
Summer Programs Summer Programs
Youth programs Youth programs
AI_In-Depth_cover

Make AI Your Ally

Artificial Intelligence has a potentially disruptive impact on organizations, firms, and society at large. The latest mind-boggling illustration came with the discovery of chatGPT’s mesmerizing results in November 2022. This followed a fall of investments in AI last year in Silicon Valley. From analyzing data in one’s business to increasing customer engagement and replacing humans in routine tasks across industries, AI is becoming more relevant to our lives and economy every day. Everyone talks about it, but do we really understand its opportunities and threats? And how can we make the best out of it, whilst ensuring that ethical requirements are met?

 

Structure

Part 1
Make AI Your Ally: Editorial
Artificial Intelligence has a potentially disruptive impact on organizations, firms, and society at large. The latest mind-boggling illustration came with the discovery of chatGPT’s mesmerizing results in November 2022. This followed a fall of investments in AI last year in Silicon Valley. From analyzing data in one’s business to increasing customer engagement and replacing humans in routine tasks across industries, AI is becoming more relevant to our lives and economy every day. Everyone talks about it, but do we really understand its opportunities and threats? And how can we make the best out of it, whilst ensuring that ethical requirements are met?
Part 2
Taking The Help or Going Alone: Students Do Worse Using ChatGPT
How good are students at using tools like ChatGPT? In particular, can they properly evaluate and correct the responses provided by ChatGPT to enhance their performance? An experiment on HEC Paris students set to find out. Its results contribute to the debate on the consequences of the use of ChatGPT in education, and in work in general.
Part 3
How AI Can Help Figure Out When Hospital Patients Need Intensive Care
When hospital patients unexpectedly take a turn for the worse and must be transferred to the intensive care unit (ICU), they are more likely to die or must spend a longer time in the hospital. But research by Julien Grand-Clément, Assistant Professor in Information Systems and Operations Management at HEC Paris and chair holder at the Hi! PARIS Center, and colleagues in the U.S. from Columbia University and the Kaiser Permanente Northern California Division of Research, suggests that a mathematical model commonly used in Artificial Intelligence (AI) can help doctors predict patients’ decline and transfer them to the ICU proactively, before their health condition deteriorate critically.
Part 4
Nudges and Artificial Intelligence: for Better or for Worse?
The latest developments in artificial intelligence are often allied to the prospect of a better world, with more powerful, more rational algorithms straightening out human flaws. The idea often floated is that public policy will be more effective because it will be better informed and more responsive. Likewise, it is said that medicine will deliver speedier, more accurate diagnoses. But where are we when it comes to the subject of consumption? Can algorithms be used to steer consumers towards better, more enlightened – and less impulsive – choices? Why not? Online assessments about a product or service, for instance, help information flow more smoothly. But we are also heading down a new path: myriad algorithms are now the power behind “nudges”, those fine details about the shopping environment that insidiously guide the choices made by consumers.
Part 5
Future of Finance: How Are New Technologies Reshaping the Sector?
The advent of digital technologies has created a very new and vastly different financial landscape. Today's buying and selling of securities is conducted mostly by computer programs that react within nanoseconds – faster than any human could – to the subtlest market fluctuations. In a new report published by the Centre for Economic Policy Research (CEPR), Professor of Finance Thierry Foucault comes to grips with how technologies are fundamentally changing the way banks, brokers, exchanges, and dealers do their work, and what it means for investors, for privacy and income inequalities.
Part 6
HEC Research Points to AI Solutions to Prevent Miscarriages of Justice
In the United Kingdom, more than 700 Post Office workers were wrongfully convicted of theft, fraud and false accounting between 2000 and 2014. That was the result of a fault in Horizon, a Fujitsu computer system used by the UK Post Office. How can AI solutions be developed to detect and prevent such intelligent anomalies? To answer these questions and more we have turned to HEC Professor of Accounting and Management Control, Aluna Wang. She is also chairholder at Hi! PARIS Center on Data Analytics and Artificial Intelligence.
Part 7
To What Extent Do People Follow Algorithms’ Advice More Than Human Advice?
Algorithms can enable faster, more effective decision making in domains ranging from medical diagnosis to the choice of a romantic partner. But for this potential to actually translate into useful practical choices, humans must trust and follow the advice algorithms provide. Researchers in Information Systems Cathy Liu Yang and Xitong Li of HEC Paris and Sangseok You of Sungkyunkwan University, have explored the factors that influence people's reliance on algorithmic decision aid.
Part 8
How Do Algorithmic Recommendations Lead Consumers to Make Online Purchases?
Many E-commerce sites such as Amazon, YouTube, and Netflix, but also online advertisers, use recommender systems. Recommender systems are algorithms that, based on data sets, recommend to users contents and products that match their preferences. In this interview, Xitong Li of HEC Paris, Associate Professor of Information Systems and a Hi! PARIS center’s research fellowship holder, reveals new research, a joint work with two German researchers, and explains how recommender systems induce consumers to buy.
Part 9
AI in HR: How is it Really Used and What are the Risks?
Artificial intelligence has only recently been taken on board by human resources, and only after being introduced into other fields. Where do we stand in concrete terms? Although there has been a whole host of ingenious innovations (driven on by start-ups in particular), and although everything virtual is all the rage, the technology seems to run up against serious limits when it comes to HR. Based on a survey carried out among HR managers and digitalization project managers working in major companies, I recall three potential pitfalls regarding the data used, the risk of turning AI into a gimmick, and algorithmic governance.
Part 10
Understanding AI-Human Interactions to Foster Creative Effort
What comes out of our collaborations with artificial intelligence? What happens in the minds of humans who are involved in interactions with smart non-humans? Does AI change how we come up with new ideas and work together on bringing them into life? Daria Morozova, Ph.D. student at HEC Paris’s Management and Human Resources department, explains how her research on AI-human interactions can help establish best practices for creative work involving artificial intelligence. 
Part 11
Meta and Data Privacy: a Clash of Cultures and Generations?
HEC Paris Assistant Professor in Marketing, Klaus Miller, analyzes the February 3 Facebook/Meta stock market plunge. What exactly does it tell us about private data on internet and its links to the advertising world? We meet Klaus on February 8, the very day he and five co-researchers self-published “The Impact of the GDPR on the Online Advertising Market”. This book focuses on Europe’s GDPR and how it affects online publicity. In a wide-ranging discussion on personal data and the advertising industry, Klaus provides insights on ad blockers on news websites and their impact on our reading habits.
Part 12
“A $%^* Sexist Program”: Detecting and Addressing AI Bias
A major issue facing companies that use AI, algorithmic bias can perpetuate social inequalities — as well as pose legal and reputational risks to the companies in question. New research at HEC Paris offers a statistical method of tracking down and eliminating unfairness.
Part 13
Developing an Automated Compliance App to Help Firms Comply with Privacy Regulations
According to recent research published by the Harvard Business Review, a strong privacy policy can help firms save millions. If privacy policies have indeed become more important for firms and consumers today, the real challenge lying ahead is to ensure compliance within and beyond the firm. A new tool developed at HEC Paris uses artificial intelligence to help managers, data protection officers, and project leaders take privacy-preserving decisions and implement privacy by design. 
Part 14
How Should We Design the Next Generation of AI-Powered Chatbots?
Have you been enraged by a chatbot recently? Businesses from Starbucks to LinkedIn and Singapore Airlines to eBay are embracing AI-powered communication tools. But the success of using chatbots to streamline operations and deliver consistent, round-the-clock customer support hinges on gaining user trust, show recent findings from HEC Paris Professor Shirish Srivastava and his co-authors Dr. Shalini Chandra of the SP Jain School of Global Management, Singapore and Dr. Anuragini Shirish of the Institut Mines Telecom Business School, France.
Part 1

Make AI Your Ally: Editorial

Artificial Intelligence
Published on:

Artificial Intelligence has a potentially disruptive impact on organizations, firms, and society at large. The latest mind-boggling illustration came with the discovery of chatGPT’s mesmerizing results in November 2022. This followed a fall of investments in AI last year in Silicon Valley. From analyzing data in one’s business to increasing customer engagement and replacing humans in routine tasks across industries, AI is becoming more relevant to our lives and economy every day. Everyone talks about it, but do we really understand its opportunities and threats? And how can we make the best out of it, whilst ensuring that ethical requirements are met?

AI_In-Depth_cover

To be a leading actor in the AI revolution, HEC Paris is active on two fronts: on the teaching side, we have set up with École Polytechnique, the MSc Data Science for Business, which is ranked third in the world. Other programs at HEC also feature a growing number of classes on coding, data science, and on the strategic value of data in companies.

On the research front, HEC Paris has joined forces with the the Institut Polytechnique de Paris and Inria to create the Hi! PARIS Center. The Center conducts multidisciplinary research on AI and its applications to business and society, and develops groundbreaking education programs on these topics. The ambition of this newly-opened center is to become a world-class hub and a destination of choice for students and faculty. It seeks to guarantee that AI and data empower business and society, providing essential resources for companies and laboratories, both public and private, and strengthening France and Europe’s leadership positions.

In this In-Depth special, you can find the latest key findings by HEC Paris researchers, including the Hi! PARIS Center’s chair holders and fellows, from the various disciplines ranging from Economics and Decision Sciences to Finance, Accounting and Management Control, Information Systems and Operations Management, Human Resources, and Marketing.

The researchers unveil techniques and opportunities offered by complex systems, but also warn about the consequences engendered by a lack of understanding and regulation. While they are very diverse in terms of topics, HEC research studies on AI share the common goal of combining academic excellence and relevance for companies and society at large. A must-read!
 

Editors: Nicolas Vieille, HEC Paris Professor of Economics and Decision Sciences and Scientific Co-Director of the Hi! PARIS Center, Christophe Pérignon, Associate Dean for Research and Professor of Finance at HEC Paris, and member of the executive committee of the Hi! PARIS Center, and Céline Bonnet-Laquitaine, Communication & Edition Project Manager for the Faculty & Research at HEC Paris.

 

See structure
Part 2

Taking The Help or Going Alone: Students Do Worse Using ChatGPT

Decision Sciences
Published on:

How good are students at using tools like ChatGPT? In particular, can they properly evaluate and correct the responses provided by ChatGPT to enhance their performance? An experiment on HEC Paris students set to find out. Its results contribute to the debate on the consequences of the use of ChatGPT in education, and in work in general.

If, as many suggest, ChatGPT-like tools will be central to many work practices in the future, then we need to think about how to design course elements that help today’s students and tomorrow’s professionals learn how to use these tools properly. A correct use will not involve humans copying the output of these tools blindly, but rather them using it as a means to enhance their own performance. Hence the simple question: can students properly evaluate and where necessary correct the responses provided by ChatGPT, to improve their grade in an assignment, for instance? Motivated by such considerations, I designed the following assignment in a first-year Masters level course at HEC Paris. 

Answering vs. correcting

Students were randomly assigned two cases, and were asked the same question about each. For the first case, students just had to provide the answer, in the traditional way, ‘from scratch’. For the second case, they were provided with an answer to the question: they were asked whether the answer was fully correct, and told to correct or add as required to make it ‘perfect’. They were told that each provided answer had been either produced by ChatGPT or by another student. In reality, in over 60% of cases, the answer had come from ChatGPT. 

Whilst the former, answer task is arguably closer to current work practices, the second correct task may correspond more closely to many jobs in the future, if AI tools become as ubiquitous as many predict. 

However, the two tasks asked for the same thing – a full reply to the question concerning the case – and the same grading scheme was used for both. The marks for both tasks counted in equal amounts for the course grade, so students were motivated to make the same amount of effort on both. 

HEC students


On this assignment, students do better without the help of ChatGPT

Nevertheless, the students, on average, got a 28% lower grade on the correct task than on the answer task. For a given case, a student correcting an answer provided by ChatGPT got, on average, 28 marks out of 100 less than a student answering the question by themselves. Students, it turns out, did considerably worse when they were given a ChatGPT aid and asked to correct it than if they were asked to provide an answer from scratch. 

Students did considerably worse when they were given a ChatGPT aid and asked to correct it than if they were asked to provide an answer from scratch.

A behavioral bias?

Perhaps these results can be explained by postulating high student trust in ChatGPT’s answers. However, students were explicitly primed to be wary of the responses provided: they had been informed that ChatGPT had been tested on a previous, similar assignment and did pretty badly. And previous research suggests that such information typically undermines trust in algorithms. Moreover, no significant difference was found between their grades on the correct task when they thought they were correcting ChatGPT or another student.

 

Our classroom experiment suggests that the professionals of tomorrow may do a considerably worse job when aided by AI than when working alone.

 

A perhaps more promising explanation is in terms of the Confirmation Bias – the tendency to insufficiently collect and interpret information contradicting a given belief or position. Inspection of answers shows a clear tendency among many students to provide small modifications to the provided responses, even where larger corrections were in order. Moreover, there is evidence that this bias tends to persist even when people are warned that the base belief has little claim to being correct 1,2. Could the tendency to display insufficient criticism with respect to certain positions – a bias that is taught in business schools worldwide and HEC in particular – be behind potential misuses of ChatGPT and its alternatives?

Chatbots have been touted as having a future role in aiding humans in a range of areas; but this assumes that humans will be capable of using them properly. One important task for humans in such interactions will be to evaluate, and where necessary correct, the output of their chatbots. 

Our classroom experiment suggests that the professionals of tomorrow may do a considerably worse job when aided than when working alone – perhaps due to behavioral biases that have been long understood, perhaps due to some that remain to be further explored. 


One of the skills of the future, that we will need to learn to teach today, is how to ensure that ChatGPT actually help.


If anything, this argues for more, rather than less, chatbots in the classroom. One of the skills of the future, that we will need to learn to teach today, is how to ensure that they actually help.


References:
1. Kahneman, D. Thinking, fast and slow. (Macmillan, 2011). 
2. Nickerson, R. S. Confirmation Bias: A Ubiquitous Phenomenon in Many Guises. Review of General Psychology 2, 175–220 (1998).
 

Article by Brian Hill, based on his paper, “Taking the help or going alone: ChatGPT and class assignments”.
See structure
Part 3

How AI Can Help Figure Out When Hospital Patients Need Intensive Care

Artificial Intelligence
Published on:

When hospital patients unexpectedly take a turn for the worse and must be transferred to the intensive care unit (ICU), they are more likely to die or must spend a longer time in the hospital. But research by Julien Grand-Clément, Assistant Professor in Information Systems and Operations Management at HEC Paris and chair holder at the Hi! PARIS Center, and colleagues in the U.S. from Columbia University and the Kaiser Permanente Northern California Division of Research, suggests that a mathematical model commonly used in Artificial Intelligence (AI) can help doctors predict patients’ decline and transfer them to the ICU proactively, before their health condition deteriorate critically.

ICU_machine-learning_cover

In a hospital, some of the most important resources are beds and specialized nurses in the ICU, where doctors and nurses use advanced technology to save the lives of critically ill patients. But providing that sort of care is complex and expensive, and hospitals can afford to have only a limited number of ICU beds. That shortfall became tragically apparent during the COVID-19 pandemic, when some facilities ran out of capacity to deal with the large number of desperately sick patients. Even in less dire times, doctors still are compelled to make difficult choices, because they can only transfer so many patients to the ICU. At the same time, if they wait until a patient’s health deteriorates to the point where a sudden, unplanned transfer is required, patients’ chances of survival plummet.

But what if doctors had a way to reliably identify the patients whose health was most likely to take a turn for the worse, and proactively send those patients to the ICU? A 2018 research article, relying on nearly 300,000 hospitalizations in the Kaiser Permanente Northern California, provides evidence that proactively transferring patients to the ICU can significantly reduce the mortality risk and length of hospital stays. But there is a risk of going too far. Indeed, other research indicates that if doctors transfer too many patients to the ICU, the ICU may become congested and the survival rate is negatively impacted. In a worst-case scenario in which the ICU is filled to capacity, this could mean that some patients who need ICU care are not able to obtain it.

For a proactive ICU transfer policy to work, three variables are needed: to recalibrate depending upon arrival rates during the day and night, how many nurses you have in the ICU, and whether transferred patients are recovering fast enough. If these metrics are not aligned, you might not make the right transfer decisions. 

What we learned by using data to create a simulated hospital

One of my esteemed collaborator, Gabriel Escobar, served as the regional director for hospital operations research at Kaiser Permanente North California in the United States. That provided us with a great opportunity to work with anonymized hospitalization data on 300,000 patients from 21 Kaiser Permanente facilities. We had a tremendous amount of detailed data on cases, right down to things like blood cell counts taken every six hours. Based upon this data, we were able to estimate things such as how patients react to different treatments and various arrival and admission rates to the ICU and the common ward.

With all this information, we were able to build a simulation model, which mimics how an actual hospital works, such as generating arrival and departure rates, the evolution of the patients and every single interaction they have with the system. With this sort of micro-modeling, you can track the simulated patient like you would a real patient during their journey in the hospital. This enables to test different scenarios of arrivals and transfer policies.

In doing our study, we utilized a mathematical model called a Markov Decision Process model, which often is used in AI. In plain English, an MDP is a natural model for sequential decision-making, in which you can inspect a sequence of decisions, and how one choice influences the next one and the overall health outcome. The sequence is influenced only by earlier decisions, not by what you might do down the line--what happens at noon, for example, isn’t affected by what will happen at 3 p.m., because that hasn’t happened yet. But you are influenced by what happened at 9 or 10 or 11 a.m.

ICU decision tree

 

The goal is to make the best decisions, not only for each patient but also with the best overall result, that is, the best outcome over the general patient population. We are concerned about the survival of the patient, and at the same time, we want to maintain the ICU occupancy at a reasonable level, so that others do well, and so that unplanned transfers may still be accounted for. We designed an optimization method, based upon machine learning model, to estimate the impact of various transfer policies.

 

When we ran the model, one surprise was that relatively small adjustments can have a big effect on the mortality of the overall patient population.

 

When we ran the model, one surprise was that relatively small adjustments in the number of health resources that the system consumes and in the arrival and recovery rates can have a big effect on the mortality of the overall patient population. Given a certain way of transferring patients, we could see the estimated mortality could jump by 20 %.

AI won’t replace human decision making in hospitals

Some people might think that the questions is whether humans should make the decisions about ICU transfers, even though they sometimes make mistakes, or whether algorithms should do it. But as I see it, these two methods are really complementary. At least in the foreseeable future, humans will have the last word. But they are more likely to make those decisions assisted with the recommendations of an algorithm. We want it to be the best automated recommendation system, and it has to provide interpretable insights on its recommendation. If the practitioners are not able to understand where the recommendations come from, they will never use it.

We have to go back the question: what exactly is at stake here? If one patient is transferred to an ICU bed, the patient has better treatment and more nursing, and better survival chance. But at the population level, if the patient transfer is going to cause the ICU to be congested, this becomes a problem for everyone else. What we are doing in this work is to look for simple transfer decision rules, based on common health metrics summarizing the health conditions of the patients and certain thresholds. This type of threshold policy is extremely simple to deploy and readily interpretable.

 

Using micro-modeling to understand a complicated enterprise and develop algorithms to assist in decision making can lead to better outcomes.

 

How to Use the Insights: Using micro-modeling to understand a complicated enterprise and develop algorithms to assist in decision making can lead to better outcomes. But it also requires an understanding of variability, and how relatively small changes in the conditions can significantly alter outcomes. Ultimately, it is up to the doctors to make the transfer decisions, but machine learning models may help to assist in the decisions and to provide valuable information regarding the impact of patient transfers.

Methodology

The researchers performed numerical experiments based on a simulation model to compare the performance of both optimal normal and robust ICU transfer policies, using hospitalization data from approximately 300,000 patients from 21 Kaiser Permanente Northern California hospitals.

Applications

The research can help hospitals to develop policies for ICU transfers, which can help more increasing the patients’ survival chance reduce their hospital stays.
This article was based upon an interview with Julien Grand-Clément and the article “Robustness of Proactive Intensive Care Unit Transfer,” published in January 2023 in Operations Research and co-authored with Carri W. Chan and Vineet Goyal of Columbia University, and Gabriel Escobar, research scientist at the Kaiser Permanente Northern California Division of Research and director of the Division of Research Systems Research Initiative.
See structure
Part 4

Nudges and Artificial Intelligence: for Better or for Worse?

Decision Sciences
Published on:

The latest developments in artificial intelligence are often allied to the prospect of a better world, with more powerful, more rational algorithms straightening out human flaws. The idea often floated is that public policy will be more effective because it will be better informed and more responsive. Likewise, it is said that medicine will deliver speedier, more accurate diagnoses. But where are we when it comes to the subject of consumption? Can algorithms be used to steer consumers towards better, more enlightened – and less impulsive – choices? Why not? Online assessments about a product or service, for instance, help information flow more smoothly. But we are also heading down a new path: myriad algorithms are now the power behind “nudges”, those fine details about the shopping environment that insidiously guide the choices made by consumers.

nudging-cover

Nudge theory was developed by Richard Thaler (Nobel Prize in Economics 2017) and Cass Sunstein in 2008. The two authors suggested that cognitive biases (faulty reasoning or perception that distorts decision-making) might serve as an instrument of public policy. These biases could be used to nudge individuals towards decisions that are deemed good for themselves or the wider community – but which they lack the perspicacity or motivation to pursue. Subtle changes in the decision-making environment can steer behaviors in a virtuous direction. Let's say you are staying in a hotel, and you know that most of the previous guests in your bedroom have re-used the same bath towel from day to day; conformity bias will then prompt you to follow suit. This same bias may prod you to cut back on your energy consumption if you find out that it is higher than your neighbor's. Automatically registering voters on the electoral roll – like pre-filled tax returns – is another instance that draws on the virtuous simplification of the target behavior. Nudging is an insidious way of inducing people to change behaviors while safeguarding their freedom of choice. It is an alternative to the conventional tools of state action such as, for example, bans or taxes.

The same methods of influencing people are also employed in marketing based on the following idea: if reason fails to persuade consumers about the utility of a purchase, you can coax them insidiously. Let’s say you want to reserve a hotel online. The site warns you that there are not many bedrooms left in your chosen category, and that other internet users are currently looking at them… all of which nudges you to book your room at full speed so you do not miss out on what you see as a rare opportunity. Websites that display a default purchase option prominently or default acceptance of specific terms and conditions are too numerous to mention. Of course, you are free to disregard these defaults provided you have time on your hands and enjoy a good search. A free trial that you end up paying for because you forgot to cancel it is another example of nudge used for marketing. And then there are those discount vouchers with conditions so limited they never get used, or targeted ads that make well-timed offers. This approach distorts the very nature of nudging since it does not aim to improve the well-being of the consumer or society, which is why it is sometimes known as “bad nudging” or “sludging”. 

The prediction made by Noah Harari in 2018 has, worryingly, already come true in part: “As biotechnology and machine learning improve, it will become easier to manipulate people's deepest emotions and desires [...] could you still tell the difference between yourself and [the] marketing experts?” It goes without saying that influence strategies are not exclusive to artificial intelligence. Door-to-door sales reps have known and used more-or-less ethical sales techniques for many years, and it follows that nudges can be employed by humans. Just think of the barista who asks if you would like a pastry – or even a set menu – when all you do is ask for a coffee. And then there is the sales assistant who kicks off negotiations with an inflated offer before pretending to give you a generous discount. Artificial intelligence, however, has the power to swell the use of influence methods by rolling them out systematically on a grand scale. The behavioral biases underpinning standard nudges were derived from experimental research. But big data can automatically detect the tiniest weak point in the decision-making process, which can then be leveraged to influence consumers. Once a new behavioral lever has been identified, algorithms can apply it extensively. 

What are the consequences of these “bad nudges”? Consumers may feel deceived because they purchase items or services that do not match their real needs, or because attempts to hold out against the influences generate a fatigue that degrades the shopping experience. Accordingly, using nudges in the field of marketing serves to lower consumer well-being.

In more general terms, the wholesale deployment of nudges often systematizes mistakes that were formerly occasional: irrationality, in other words, becomes the norm.

 

In other words, irrationality becomes the norm.

 

In this respect, the growing use of nudges is upsetting the foundations on which liberal economics is built. In this model, it is the pressure exerted by consumers making informed choices that encourages producers to offer products that best match consumer needs at the best price. Nudges upend this process since producers can use them to influence consumer preferences. This means that consumers who have been influenced by nudges no longer exert their counter-power on producers. The possibility that consumer behavior may be swayed by nudging challenges some of the virtues of the market economy. Likewise, the idea that the public might cast their votes under influence undermines the basis of the democratic model. 

How can we stave off these damaging effects? Is regulation the answer? It would be problematic to legislate in this area, since the distinction between information and influence is so slight. At the very least, it could become a requirement that the information provided during the purchase process (such as the quantities available) be true, although this would still be difficult to enforce. Change could also be driven by consumers, who could turn their backs on platforms that employ these techniques. This is no easy task: the influences are not always conscious, and some platforms operate a quasi-monopoly. Sellers themselves could also reverse the trend by certifying that they do not use influence techniques as a way of guaranteeing quality and respect for their customers. This approach could be supported by artificial intelligence: algorithms could be used to automatically test online sales sites to detect nudges, and a certification label could be created. 

Do we need “good algorithms” for fighting “bad ones”? Although this idea is simplistic, it does remind us that machines only do what we have designed them to do (apart from mistakes in programming). This means that it is up to consumers (or their representatives or advocates) to make use of the possibilities afforded by artificial intelligence to defend their interests.

Related topics:
Decision Sciences
Artificial Intelligence
Emmanuel Kemel HEC professor
Emmanuel Kemel
CNRS Research Professor
See structure
Part 5

Future of Finance: How Are New Technologies Reshaping the Sector?

Finance
Published on:

The advent of digital technologies has created a very new and vastly different financial landscape. Today's buying and selling of securities is conducted mostly by computer programs that react within nanoseconds – faster than any human could – to the subtlest market fluctuations. In a new report published by the Centre for Economic Policy Research (CEPR), Professor of Finance Thierry Foucault comes to grips with how technologies are fundamentally changing the way banks, brokers, exchanges, and dealers do their work, and what it means for investors, for privacy and income inequalities.

stock market cover

Photo Credit: Phongphan Supphakank on Adobe Stock

The 'sell side' industry, concerned with the building up and liquidating of securities portfolios, has undergone enormous changes over the past fifty years, mainly due to the computerization of trading, and more recently the emergence of big data. Those familiar, iconic scenes in which mobs of colorfully garbed brokers tussle and bustle, waving frantically and hollering orders on stock exchange floors, though still sometimes staged for marketing purposes, are a thing of the past.

Two new types of technologies 

There are at least two types of technologies driving the so-called 'electronification' of trading. First, exchanges have automated the process by which they match buyers and sellers of securities. Imagine, for example, that you want to buy 1,000 shares of L'Oréal stock. Your bank or broker might send your order to Euronext, one of the exchanges on which L'Oréal is traded. Euronext receives, buys and sells orders like this all the time, using computers and algorithms to match them.

This is already a profound change, but now consider, Euronext is also accumulating massive amounts of data, about submitted orders, about realized transactions and so on, which it can then resell to other intermediaries and investors. In this respect, securities trading platforms are increasingly looking like other digital platforms, like Facebook, Google or Twitter, and the share of their revenues coming from the sale of data is growing very quickly (at an annual rate of about 13% since 2012).

 

Like Big Tech does, trading platforms could pay you to trade with them, just so you will use their platforms and generate more data!

 

The second type of technology involves industry participants automating their decisions on the buying or selling of securities. This use of algorithms to make portfolio decisions is what we call algorithmic trading. An asset manager can buy or sell millions of shares of a given stock in a day in response to investors' inflows and outflows in his or her fund. This is the same process of automation that we see in other industries. We are removing humans and replacing them with machines.

Moving fast

Some specialized trading firms, known as high-frequency traders, use algorithms that rely on extremely fast, less-than-a-millisecond, access to information, including to market data sold by electronic trading platforms. With extremely fast access to this kind of market data, these firms can take advantage of small differences in the price of the same stock on two different trading platforms. Some of them pay to have their computer servers housed near trading platform servers – they may even rent rack space in the same room, thus gaining some nanoseconds in the delivery of key information, which can make all the difference.

The question of what effect these developments may have on trading costs for other market participants is controversial, raising many issues that are now at the center of policy debate in the EU and North America.

Questions for regulators: stability and transparency

The European Securities and Markets Authority (ESMA) and various national bodies, such as the Autorité des Marchés Financiers (AMF) in France, are the key regulatory bodies for securities markets in the EU, while the Securities and Exchange Commission (SEC) and the Commodity Futures Trading Commission (CFTC) cover U.S. markets.

A number of issues related to the impact of new technologies are up for consideration by regulators. For instance, does the electronification of financial markets actually reduce the costs at which investors can build and liquidate their portfolios? This could mean far larger returns for investors on their savings. Does algorithmic trading make financial markets more stable, or less so? Do trading platforms have too much market power in the pricing of their market data?

 

We think the development of central bank digital currencies technology should be targeted at solving market failures.

 

Another question we address in our report is whether trading should be slowed down. The issue here is that high-frequency traders might be deriving excessive profits, at the expense of other participants. 

There is also a concern about trading on platforms with less stringent transparency requirements than the main exchanges. The volume of this so-called 'dark trading' is growing, now accounting for about 40% to 50% of equity trading in the EU, raising debate over whether these platforms should be more strictly regulated. Finally, another issue is to what extent algorithms might destabilize financial markets, resulting in large price swings. 

What the future holds

In the coming years, I expect exchanges' business models to continue to rely increasingly on the monetization of data generated by trading. This means exchange platforms will compete with one another to attract users who generate that data, very much like Big Tech does.

 

The increasing use of consumer data allows for efficiency gains but also involves potential risks in terms of privacy, diminished competition, and increased income inequality.

 

This trend accelerated during the COVID-19 pandemic, and, if it continues, this will put strong competitive pressures on securities dealers, and it will eventually reduce trading costs for investors. At some point, the data generated by trading may become more profitable than the trading itself. So, there may come a time when trading platforms start to go to greater lengths to attract users. For example, they could simply pay you to trade with them, just so you will use their platforms and generate more data!
 

Methodology

The current report is the fourth in the CEPR's 'Future of Banking' series and part of the Banking Initiative launched by the IESE Business School in October 2018 with support from Citi. The goal of the initiative is to study new developments in banking and financial markets. The Center for Economic Policy Research (CEPR) is an independent, non-partisan and non-profit organization, founded in 1983 to enhance the quality of economic policy making in Europe.

Applications

One of the broader messages in the report is about the suitability of central bank digital currencies (CBDCs). We think the development of CBDC technology should be targeted at solving market failures. We also point out that the increasing use of consumer data allows for efficiency gains but also involves potential risks in terms of privacy, diminished competition, and increased income inequality. In short, the electronification of market securities has real policy and economic consequences that we must understand and address.
Based on an interview with Professor of Finance, Thierry Foucault, regarding the CEPR report 'Technology and Finance – Future of Banking 4', co-written with Darrel Duffie from Stanford University, Laura Veldkamp from Columbia University’s Graduate School of Business, and Xavier Vives, from Spain's Instituto de Estudios Superiores de La Empresa (IESE). Find the report here and another summary on VoxEU here.
See structure
Part 6

HEC Research Points to AI Solutions to Prevent Miscarriages of Justice

Management Control - Audit
Published on:

In the United Kingdom, more than 700 Post Office workers were wrongfully convicted of theft, fraud and false accounting between 2000 and 2014. That was the result of a fault in Horizon, a Fujitsu computer system used by the UK Post Office. How can AI solutions be developed to detect and prevent such intelligent anomalies? To answer these questions and more we have turned to HEC Professor of Accounting and Management Control, Aluna Wang. She is also chairholder at Hi! PARIS Center on Data Analytics and Artificial Intelligence.

They’ve branded it the most widespread miscarriage of justice in the history of the United Kingdom – perhaps the world! The consequences on human lives have been catastrophic and inquiries are ongoing on how to repair them. Experts say the Horizon affair could end up costing the British taxpayer almost €500 million in compensation payments. Nevertheless, after 20 years, the victims have won a legal battle to reverse the verdicts which had led to criminal conviction, prison and fines. So, what does this teach us about the risks of faulty IT systems?

 

Listen to the podcast:

 

Doctor Wang, part of your research centers on developing AI tools to improve our understanding of accounting data and proposing intelligent solutions to some real-world challenges in businesses which are undergoing huge changes in the digital world. We saw one of these challenges at the heart of the UK’s Post Office computer scandal. How do you respond to this affair which is the result of a computer system called Horizon, which a High Court judge said was not “remotely robust”? 

Aluna Wang: I was certainly shocked by this miscarriage of justice. First of all, we can see that hundreds of Post Office workers were falsely accused of theft and false accounting after Horizon was introduced and incorrectly showed shortfalls on the corporate accounts. If the whole story were told as a movie, even the movie critics would think that the plot was too implausible. It’s tough for me to fathom why the UK Post Office, which is partly owned by the British government, accused so many innocent employees of theft and misreporting rather than explore the possibility that the IT system may be faulty and malfunctioning. Moreover, we don’t see a single high-placed representative from the Post Office, the IT supplier Fujitsu, or Parliament has been truly held accountable for the decisions based on the incorrect information provided by the Horizon system.

As you mentioned earlier, I have experience working with audit partners and banking executives in developing intelligent anomaly detection systems. Usually, they were highly concerned about the false positives generated by the detection systems. Because they know that if they rely on the detection system and the system gives too many false positive alarms, they waste a lot of resources investigating those false-positive cases. In this sense, false positives can be very costly. 

 

They were financially ruined, put out of work, locally shunned, driven into poor health, and saw their families destroyed.

 

But here, in this Post Office scandal, we see that without rigorous monitoring of the IT system, and serious investigations into the alarms raised by the IT system, there could be even more severe costs to society. More than 700 Post Office workers were wrongfully prosecuted. Their lives and the lives of thousands of others were torn apart. They were financially ruined, put out of work, locally shunned, driven into poor health, and saw their families destroyed. This whole incident made me think more about not only the design and deployment of IT systems and AI solutions, but also how to manage the risk of using those technological solutions and how to build accountability into those solutions.  

With hindsight, what could have been done to prevent such errors?

There are undoubtedly many things that could have been done to prevent this scandal. I would like to speak more from the risk management perspective. The UK Post Office could have set a clear tone at the top regarding the transparency and integrity of the IT systems put into place. It could have conducted a thorough investigation of any potential defects in the Horizon system before signing the contract with Fujitsu and made a robust risk management and monitoring plan of the Horizon system after implementing it. 

Moreover, the Post Office should have taken a whistleblower, Alan Bates, more seriously. Bates reported the problems linked to the Horizon system to the Post Office management team in the early 2000s. Unfortunately, his reports were not only taken seriously but his contract with the Post Office was terminated. 

Given my field of research, I actually think one of the AI solutions I developed with my collaborators can be helpful in this case. We have been working on an anomaly detection system designed for internal audit, risk management, and compliance purposes. 

 

We have been working on an anomaly detection system designed for internal audit, risk management, and compliance purposes.

 

When you put accounting data into the detection system, it can assign anomaly scores to each financial transaction. It tells you why certain transactions or patterns of transactions are anomalous based on the metadata and the accounting structure of transactions. In this case, our detection system should be able to detect the changes in transaction patterns after implementing the Horizon system and flag many of the incorrect records generated by the Horizon system as highly anomalous. Furthermore, our algorithm can also generate explanations concerning how the anomaly scores were assigned based on the algorithm’s logic. 

But still, we would need the Post Office management to take the red flags of the Horizon system seriously and investigate accordingly. After all, this miscarriage of justice is not only about a flawed IT system, but also about how the Post Office deals with it.

Also, since this scandal concerns severe legal enforcement actions, I think there is also a lesson for Fujitsu and other tech companies. Fujitsu should not only be more effective in reducing product defects but also look at how its clients are using the output of its systems. Horizon is a point-of-sale system that records transactions, but the Post Office also uses the data output for prosecutions. More attention should have been paid to the data output at that point. Perhaps, Fujitsu should not have handed over data packs to the UK Post Office as court evidence.

Finally, Dr. Wang, could you share with us some of the latest research you are conducting in the Hi!PARIS context? After all, your explorations involve developing machine learning-based tools to improve our understanding of accounting data, research that seeks intelligent solutions to real-world challenges like the ones we saw in the UK Post Office affair…

Our Hi! PARIS center is a research center for science, business, and society. It really aims to combine the expertise of people from different fields of specializations to address important questions at the intersection of science, technology, and business, while developing new education programs and fostering innovation. 

I personally would like to put AI research into three categories: the first one is about “AI solutions”, which is what you called “intelligent solutions”. For this type of research, we engineer AI solutions addressing business and societal problems. For example, my collaborators and I have designed algorithm packages for risk management of financial institutions. Our graph-based machine learning algorithms can be used for anti-money laundering, email communication monitoring, and fraud detection purposes. 

 

Our graph-based machine learning algorithms can be used for anti-money laundering, email communication monitoring, and fraud detection purposes. 


I would like to call the second category “AI for Social Science”. We can leverage AI techniques to understand better economic phenomena. For instance, my collaborators and I are currently working on using graph mining techniques to investigate the knowledge spillover in the open-source community.

And, finally, I call the third category of research “Social Science for AI”. For this type of research, we can use our social science research methods to examine how AI and digital transformation affect our human behaviors and business models. My collaborators and I are currently working on analyzing the human-algorithm interactions on the social platforms and figuring out how we can design algorithms to improve the information environment of social platforms. 
 

Related topics:
Management Control - Audit
Podcasts
Accounting
Artificial Intelligence
Aluna Wang
Assistant Professor
Accounting and Management Control
See structure
Part 7

To What Extent Do People Follow Algorithms’ Advice More Than Human Advice?

Information Systems
Published on:

Algorithms can enable faster, more effective decision making in domains ranging from medical diagnosis to the choice of a romantic partner. But for this potential to actually translate into useful practical choices, humans must trust and follow the advice algorithms provide. Researchers in Information Systems Cathy Liu Yang and Xitong Li of HEC Paris and Sangseok You of Sungkyunkwan University, have explored the factors that influence people's reliance on algorithmic decision aid.

Photo Credits: Have a nice day on Adobe Stock

Machine recommendations result in 80% of Netflix viewing decisions, while more than a third of purchase decisions on Amazon are influenced by algorithms. In other words, algorithms increasingly drive the daily decisions that people make in their lives.

It isn’t just consumer decision making that algorithms influence. As algorithms appear increasingly in different situations, people are using them more frequently to make more fundamental decisions. For example, recent field studies have shown that decision makers follow algorithmic advice when making business decisions or even providing medical diagnoses and releasing criminals on parole.

Do people prefer human or machine advice-giving?

People regularly seek the advice of others to make decisions. We turn to experts when we are not sure. This provides us with greater confidence in our choices. It is clear that AI increasingly supports real-life decision making. Algorithms are ever more intertwined with our everyday lives. What we wanted to find out is the extent to which people follow the advice offered by AI.

To investigate the matter, we conducted a series of experiments to evaluate the extent to which people follow AI advice. Our study showed that people are more likely to follow algorithmic advice than identical advice offered by a human advisor due to a higher trust in algorithms than in other humans. We call this phenomenon “algorithm appreciation”.

Higher Trust in AI… but don’t go overboard on information

We wanted to find out more, to see if people would follow AI advice even if the AI is not perfect. Our second series of experiments focused on exploring under which conditions people might be either more likely or less likely to take advice from AI. We engineered experiments that tested whether people would have greater trust in algorithms even when they were aware of prediction errors with the underlying AI.

Surprisingly, when we informed participants in our study of the algorithm prediction errors, they still showed higher trust in the AI predictions than in the human ones. In short, people are generally more comfortable trusting AI than other humans to make decisions for them, regardless of known and understood imperfections in the process.

 

People are generally more comfortable trusting AI than other humans to make decisions for them, regardless of known and understood imperfections in the process, except when there is too much information about the algorithm and its performance.

 

There was an exception to this rule. We found that when transparency about the prediction performance of the AI became very complex, algorithmic appreciation declined. We believe this is because the provision of too much information about the algorithm and its performance can lead to a person becoming overwhelmed with information (cognitive load). This impedes advice taking. This is because people may discount predictions when presented with too much information about the underpinning detail and they are unable or unwilling to internalize it. However, if we do not overwhelm people with information about AI then they are more likely to rely on it.

What could possibly go wrong?

If algorithms can generally make better decisions than people, and people trust them, why not rely on them systematically? Our research raises potential issues of over-confidence in machine decision-making. In some cases, the consequences of a bad decision recommended by an algorithm are minor: If a person chooses a boring film on Netflix they can simply stop watching and try something else instead. However, for high-stakes decisions that an algorithm might get wrong, questions about accountability come into play for human decision-makers. Remember the miscarriage of justice in the UK Post Office, when more than 700 Post Office workers were wrongfully convicted of theft, fraud and false accounting between 2000 and 2014, because of a fault in a computer system.

However, our research has important implications for medical diagnosis. Algorithmic advice can help where there is a patient data for examination. AI can predict with a level of likelihood whether the chances of the patient having cancer are 60% or 80% and the healthcare professional can include this information in decision making processes about treatment. This can help avoid a patient’s higher level of risk being overlooked by a human and it can lead to more effective treatment, with the potential for a better prognosis.

In wider society, algorithms can help judges in the court system make decisions that will drive a safer society. Judges can be given predictions from algorithms that present the chance of a criminal possibly committing the crime again and so decide for how long to put them away.

Methodology

To explore how and why transparency in performance influences algorithm appreciation, we conducted five controlled behavioral experiments, each time recruiting more than 400 participants via Amazon's Mechanical Turk. Across the five experiments, participants were asked to perform a prediction task in which they predict a target student’s standardized math score based on nine pieces of information about the student before and after being presented with advice generated by the algorithmic prediction regarding the student’s predicted score.

Applications

Where firms need to make investment decisions, employees will trust AI to help inform those choices. With good data and solid, well-thought-out underlying algorithms, this has the potential to save businesses a lot of money.
Based on an interview with HEC Paris professors of Information Systems Cathy Liu Yang and Xitong Li on their paper “Algorithmic versus Human Advice: Does Presenting Prediction Performance Matter for Algorithm Appreciation,” co-written with Sangseok You, Assistant Professor of Information Systems at Sungkyunkwan University, South Korea, and published online in the Journal of Management Information Systems, 2022. This research work is partly supported by funding from the Hi! PARIS Fellowship and the French National Research Agency (ANR)'s Investissements d'Avenir LabEx Ecodec (Grant ANR-11-Labx-0047).
See structure
Part 8

How Do Algorithmic Recommendations Lead Consumers to Make Online Purchases?

Data Science
Published on:

Many E-commerce sites such as Amazon, YouTube, and Netflix, but also online advertisers, use recommender systems. Recommender systems are algorithms that, based on data sets, recommend to users contents and products that match their preferences. In this interview, Xitong Li of HEC Paris, Associate Professor of Information Systems and a Hi! PARIS center’s research fellowship holder, reveals new research, a joint work with two German researchers, and explains how recommender systems induce consumers to buy.

e-commerce - cover

Photo Credit: NaMaKuKi on Adobe Stock

What do you study?

We examine how algorithmic recommenders could induce online consumers to buy products. For example, product recommendations nowadays are widely used by online retailers, so we would like to see how showing recommendations to consumers on the online retailers’ websites would influence consumers’ consideration set, which in turns affects their purchases. By consideration set, we refer to the assortment of alternative products that consumers consider before actually making a purchase.

What recommender systems did you study?

Different types of recommender systems exist, with algorithms based on content, on collaborative filtering, or a mix of both. Collaborative filtering recommender systems are the most common in current business practices: they recommend products based on the preferences of similar users or on the similarity between products. In our study, we employed the recently developed causal mediation approach to examine the causal pathsthe underlying channel through which the use of recommender systems eventually leads to consumer purchases.

We conducted a randomized controlled field experiment on the website of an online book retailer in Europe. In the field experiment, some visitors on the online retailer’s website were randomly assigned in the treatment group to see personalized recommendations, whereas the other visitors could not see any recommendations.

Can you explain what effects of the use of personalized recommendations you observed?

Unsurprisingly, the results show that the presence of personalized recommendations increases consumers’ propensity to buy by 12.4% and basket value by 1.7%.

 

The presence of personalized recommendations increases consumers’ propensity to buy by 12.4% and basket value by 1.7%.

 

But more importantly, we find that these positive economic effects are largely mediated through influencing consumers’ consideration sets.

As explained before, a consideration set is the set of alternative products that consumers consider before making a purchase. We distinguish two aspects of a consideration set: its breadth and its depth. The breadth (also called “size”) of a consideration set refers to the number of different choices in the set, whereas the depth refers to how intensively a consumer is engaged with the choices. The depth is measured by the average number of pages viewed or the average session duration per choice before the buyer selects his final choice and purchase.

We find that the presence of personalized recommendations increases both the size and depth of consumers’ consideration set. It is these two changes that lead to the increase in consumers’ propensity to buy and the basket value.

Furthermore, we find that the effect mediated via the size of consumers’ consideration set is much stronger and significant than the effect mediated via the depth. In other words, the more choices you have taken into consideration, the more likely you will buy, and the more money you will spend.

Your findings suggest important managerial implications for professionals and practitioners. How should online retailers use recommender systems?

Now we know that consideration sets play an important role in mediating the positive effects of recommender systems on consumer purchases, online retailers need to consider the practical strategies that facilitate the formation of their considerations sets. For example, to reduce consumers’ search costs and cognitive efforts, online retailers can display the recommended products in a descending order according to the predicted similarity of consumers’ preferences. Given such a placement arrangement, consumers can quickly screen the recommended products and add the most relevant alternatives to their consideration sets, which should facilitate the consumers’ shopping process and increase shopping satisfaction.

 

To reduce consumers’ search costs and cognitive efforts, online retailers can display the recommended products in a descending order according to the predicted closeness of consumers’ preferences.

 

As recommender systems induce consumers to take more choices into consideration, it could be difficult for them to manage many choices. In other words, the more choices there are, the more difficult it is for the consumer to choose. To facilitate consumers’ shopping process, online retailers need to consider strategies and web tools that help consumers manage the choices in a better-organized manner and facilitate their comparison.

Are your findings generalizable to other contexts or business domains?

Although the field experiment in the study was conducted on the website of an online book retailer, we believe the findings are generalizable to a broader context of online retailing that uses recommender systems (Kumar and Hosanagar 2019). For example, the results can easily be applied to online retailers like Amazon and Cdiscount, as well as merchants’ online stores (e.g., Darty.com).
 

Interview with Xitong Li, based on his paper, "How Do Recommender Systems Lead to Consumer Purchases? A Causal Mediation Analysis of a Field Experiment", co-authored by Jörn Grahl (Head of Data Science at PHOENIX and former Professor of Information Systems, Digital Transformation and Analytics at the University of Cologne) and Oliver Hinz of the Goethe University Frankfurt, and forthcoming in Information Systems Research. Xitong Li serves as an Associate Editor on the editorial board of Information Systems Research since January 2022. He also served as a Program co-Chair for International Conference on Information Systems (ICIS) 2021. Professor Xitong Li is very open to collaborate with practitioners and help them develop innovative online retailing strategies based on big data and business analytics.
See structure
Part 9

AI in HR: How is it Really Used and What are the Risks?

Human Resources
Published on:
Updated on:

Artificial intelligence has only recently been taken on board by human resources, and only after being introduced into other fields. Where do we stand in concrete terms? Although there has been a whole host of ingenious innovations (driven on by start-ups in particular), and although everything virtual is all the rage, the technology seems to run up against serious limits when it comes to HR. Based on a survey carried out among HR managers and digitalization project managers working in major companies, I recall three potential pitfalls regarding the data used, the risk of turning AI into a gimmick, and algorithmic governance.

facial recognition - cover

Photo Credit: Sittinan on Adobe Stock

What do we mean by AI?

The term artificial intelligence is polysemous, just as AI itself is polymorphic. Hidden behind AI’s vocabulary – from algorithms, conversational AI and decisional AI to machine learning, deep learning, natural language processing, chatbots, voicebots and semantic analysis  – is a wide selection of techniques, and the number of practical examples is also rising fast. There’s also a distinction to be made between weak AI (non-sensitive intelligence) and strong AI (a machine endowed with consciousness, sensitivity and intellect), also called "general artificial intelligence" (a machine that can apply intelligence to any problem rather than to a specific problem). "With AI as it is at the moment, I don’t think there’s much intelligence, and not much about it is artificial... We don’t yet have AI systems that are incredibly intelligent and that have left humans behind. It’s more about how they can help and deputize for human beings" (Project manager).

AI in practice

For HR managers, AI paves the way to time and productivity savings alongside an "enhanced employee experience" (HR managers). For start-ups (and there are 600 of them innovating in HR and digital technology, including around 100 in HR and AI), the HR function is "a promising market".

Administrative and legal support: helping save time 

AI relieves HR of its repetitive, time-consuming tasks, meaning that HR staff, as well as other teams and managers, can focus on more complex assignments.

Many administrative and legal help desks are turning to AI (via virtual assistants and chatbots) to respond automatically to questions asked by employees – "Where is my training application?" or "How many days off am I entitled to?" – in real time and regardless of where staff are based. AI refers questioners to the correct legal documentation or the right expert. EDF, for example, has elected to create a legal chatbot to improve its performance with users. The chatbot is responsible for the regulatory aspects of HR management: staff absences, leave, payroll and the wage policy: "We had the idea of devising a legal chatbot to stop having to answer the same legal and recurring questions, allowing lawyers to refocus on cases with greater added value. In the beginning, the chatbot included 200 items of legal knowledge, and then 800... Users are 75% satisfied with it" (Project manager).

 

AI can help find the correct legal documentation or the right expert, AI can check the accuracy of all declarations, AI can personalize social benefits based on employee profiles.

 

AI isn’t just employed for handling absences and leave or processing expense reports and training but also for the administrative and legal aspects of payroll and salary policy. For pay, AI systems can be used to issue and check the accuracy and consistency of all declarations. In another vein, AI offers packages of personalized social benefits based on employee profiles.

Recruitment: helping to choose candidates 

Recruitment is another field where AI can be helpful: it can be used to simplify the search for candidates, sift through and manage applications, and identify profiles that meet the selection criteria for a given position. 

Chatbots can then be used to talk to a candidate in the form of pre-recorded questions, collecting information about skills, training and previous contracts. "These bots are there to replace first-level interactions between the HR manager or employees and candidates. It frees up time so they can respond to more important issues more effectively" (Project manager). 

Algorithms analyze the content of job offers semantically, targeting the CVs of applicants who are the best match for recruiters' expectations in internal and external databases via professional social networks such as LinkedIn. CV profiles that were not previously pre-selected can then be identified. Unilever has been using AI in tandem with a cognitive neuroscience approach to recruiting since 2016. Start-ups offer a service where they highlight candidate profiles without the need for CVs, diplomas or experience. Their positioning is based on affinity and predictive matching or building smart data. 

These tools are aimed primarily at companies with high volumes of applications to process, such as banks for customer service positions or large retailers for departmental supervisors. Candidates are notified about the process and give their permission for their smartphone’s or computer’s microphone and camera to be activated.

These techniques appear to be efficient for low-skilled jobs and markets that are under pressure. At the same time, the more a position calls for high, complex skills, the greater the technology’s limitations.

While new practical examples are emerging over time, the views of corporate HR managers and project managers diverge regarding the added value of AI for recruitment. Some think that AI helps generate applications; identifies skills that would not have been taken into account in a traditional approach; and provides assistance when selecting candidates. Others are more circumspect: it’s possible as things stand, they argue, that people are "over promising" in terms of what AI can bring to the table.

Training and skills: personalized career paths

The AI approach to training involves a shift from acquiring business skills to customizing career paths. With the advent of learning analytics, training techniques are evolving. Data for tracking learning modes (the time needed to acquire knowledge and the level of understanding) can be used to represent the way people learn and individualize suggestions for skills development.

In addition, AI is used to offer employees opportunities for internal mobility based on their wishes, skills and the options available inside the company. AI start-ups put forward solutions for streamlining mobility management that combine assessment, training and suggestions about pathways, positions and programs for developing skills. These are limited in France, however, with the RGPD (General Data Protection Regulation), although they can be individualized in other countries.

 

We use AI to identify talent not detected by the HR and managerial teams; and to detect talent with a high risk of leaving the company. 

 

Saint-Gobain has decided to use the potential of machine learning to upgrade the way it manages its talents. A project team with diverse profiles (HR, data scientists, lawyers, business lines, etc.) has been set up with two objectives: to use AI to identify talent not detected by the HR and managerial teams; and to detect talent with a high risk of leaving the company. Confidentiality is guaranteed, and no decision is delegated to machines .

Motivation and social climate: towards a better understanding of commitment

AI provides opportunities for pinpointing employees who are at risk of resigning or for improving our understanding of the social phenomena in companies.

"We’re going to ask employees questions every week or two weeks in 45 seconds on dozens of engagement levers... The responses are anonymized and aggregated, and – from another perspective – will give indicators to the various stakeholders, the head of HR, the manager, an administrator and so on. We’ll be able to tell what promotes or compromises engagement in real time while offering advice" (Project manager).

Thanks to this application, HR managers and other managers and employees can have use of real-time indicators that show the strengths and weaknesses of the teams without knowing by name who said what. “As well as helping to improve your own work and personal style, these variables mean that, if a corrective action has been put in place, you can tell what its real effect is straightaway” (HR manager). 

While artificial intelligence can, as these various examples show, support human resources management, some of these new uses are still being tested.

In addition to the way they are implemented, questions and criticisms remain, including the issue of HR data, return on investment and algorithmic governance.

Data quality and restricted quantity 

Data is the key ingredient for AI, and its quality is of paramount importance. If the data injected isn’t good, the results will be vague or distorted. The example of Amazon is emblematic in this respect: after launching its first recruitment software, the company decided to withdraw it quickly from the market because it tended to favor the CVs of men. In fact, the computer program Amazon used was built on resumes the company had received over the last 10 years, most of which were based on male profiles. 

In addition, data sets in the field of HR tend to be narrower compared to other areas. Indeed, the number of people employed here – including in large companies – appears to be particularly low compared to the number of purchases made by customers. The quantity of sales observations for an item is very important, and big data applications can be easily performed. But there’s nothing of the sort for HR!

The quantity and quality of data needs to be considered, as well as how to embed it in a context and time interval that encourage analysis and decision-making. This is the case, for instance, in the field of predictive maintenance, where expert systems can detect signs of wear in a machine before human beings: by collecting the appropriate data, just-in-time interventions can be made to overhaul the machinery. It’s a different matter for human systems, however, where the actions and reactions of individuals are not tracked (and should they be?), and may turn out to be totally unpredictable.

Return on investment and the dangers of gimmickry

Investing in AI can be costly, with HR managers concerned about budgets and returns on investment. 

"Company websites have to be revamped and modernized on a regular basis... And let's be realistic about it, making a chatbot will cost you around EUR 100,000, while redoing the corporate website will be ten times more expensive... And a chatbot is the ‘in thing’ and gives you a modern image, there’s a lot of hype, and – what’s more – it’s visible for the company! It’s something that can be seen!" (Project manager).

In many organizations, the constraints imposed by the small amount of data, and the few opportunities to iterate a process, raise the question of cost effectiveness. The HR managers we met for our study asked themselves: Should they invest in AI? Should they risk undermining the trust that employees may have in HR just to be one of the early adopters? And, although AI and virtual technology are in the limelight, are they really a priority for HR? Especially since trends come and go in quick succession: version 3.0 has barely been installed before version 4.0 and soon 5.0 come on the scene.

A further danger lies in wait for AI: the descent into gimmickry, a downward spiral which, it must be emphasized, doesn’t just threaten AI but also management tools more broadly. While the major HR software is now equipped with applications that can monitor the veritable profusion of HR indicators, isn’t there still the risk that we’ll drown in a sea of information whose relevance and effectiveness raise questions? Say yes to HR tools and no to gimmicks! 

Many start-ups operating in the field of AI and HR often only have a partial view of the HR function. They suggest solutions in a specific area without always being able to integrate them into a company’s distinct ecosystem.

 

algorithme

Photo Credit: Luca Bravo on Unsplash

"I admit I’m not very interested in the idea of becoming a beta tester, blinded by the complexity of the algorithms" (a HR manager)

Algorithmic governance

Faced with the growth in AI and the manifold solutions on offer, HR managers, aware of AI's strengths, are asking themselves: "Have I found HR professionals I can discuss things with? No. Consultants? Yes, but all too often I’ve only been given limited information, repeated from well-known business cases that don’t really mean I know where I‘m setting foot... As for artificial intelligence, I’ve got to say, I haven't come across it. Based on these observations, I admit I’m not very interested in the idea of becoming a beta tester, blinded by the complexity of the algorithms and without any real ability to compare the accuracy of the outcomes that these applications would give me. Should we risk our qualities and optimal functioning, not to mention the trust our employees put in us, just to be among the early adopters?" (HR manager).

Society as a whole is defined by its infinite diversity and lack of stability. Wouldn’t automating management and decision-making in HR fly in the face of this reality, which is both psychological and sociological? The unpredictability of human behavior cannot be captured in data. With AI, don’t we risk replacing management, analysis and informed choices with an automatism that destroys the vitality of innovation? 

What room would there then be for managerial inventiveness and creativity, which are so important when dealing with specific management issues? While the questions asked in HR are universal (how can we recruit, evaluate and motivate…?), the answers are always local. The art of management cannot, in fact, be divorced from its context. 

Ultimately, the challenge is not to sacrifice managerial methods but to capitalize on advances in technology to encourage a holistic, innovative vision of the HR function where AI will definitely play a larger role in the future, especially for analyzing predictive data.

 

See also: Chevalier F (2023), “Artificial Intelligence and Human Resources Management: Practices and Questions” in Monod E, Buono AF, Yuewei Jiang (ed) Digital Transformation: Organizational Challenges and Management Transformation Methods, Research in Management, Information Age Publishing.

"AI and Human Resources Management: Practices, Questions and Challenges", AOM Academy of Management Conference, August 5-9, 2022, Seattle, USA. 
 

This article presents a summary of “Intelligence Artificielle et Management des Ressources Humaines: pratiques d’entreprises”, Enjeux numériques, No. 15 (September 2021) by Françoise Chevalier and Cécile Dejoux. This research is supported by a grant of the French National Research Agency (ANR), “Investissements d’Avenir” (LabEx Ecodec/ANR-11-LABX-0047).
See structure
Part 10

Understanding AI-Human Interactions to Foster Creative Effort

Human Resources
Published on:
10 minutes

What comes out of our collaborations with artificial intelligence? What happens in the minds of humans who are involved in interactions with smart non-humans? Does AI change how we come up with new ideas and work together on bringing them into life? Daria Morozova, Ph.D. student at HEC Paris’s Management and Human Resources department, explains how her research on AI-human interactions can help establish best practices for creative work involving artificial intelligence. 

human - robot - phonlamaiphoto

©phonlamaiphoto on Adobe Stock

While many may think creativity will be a key human capacity as work processes get increasingly automated, AI-applications get more creative as well: for example, AI-painted portraits sell at exorbitant prices, pop hits feature AI-generated music, and mainstream media ‘hire’ AI-journalists to publish thousands of articles a week. How workplace AI-application will impact creative processes, crucial for innovation and organizational development, is an important question that needs rigorous scientific attention.

 

portrait peint Edmond de Belamy - wikipedia

Portrait d'Edmond de Belamy by an Artificial intelligence software (Source: Christie's)

What can research bring to the field of creativity with AI?

Research can help find ways to foster creative work and innovation augmented by the numerous AI-provided capacities. We need research that explains how we can create and adapt artificial agents and arrange working processes such that employees do not have to be defensive against AI and be more creative.

By now, we have learned that many people believe that artificial agents – these are, for instance, AI-run apps like Siri or Alexa, or social robots you may see at an airport - are not suited for creative work or work that requires diplomacy and judgement skills. So, normally people would not care much about AI trying to compose a symphony or write a novel as long as they believe that humans are creatively superior. But we also know that when AI is involved in creative work – for example, generating image suggestions, or even creating logotypes at a design studio, people often reject its results and evaluate them lower even if the AI objectively performs just as well as a human. 

This is a problem because when employees believe that AI should not be creative and it turns out to be exactly that, they may feel threatened by it, fearing it might substitute them, and these feelings are normally not very conducive to creativity. In addition, when threatened, employees will try to protect themselves – after all, we want to feel that we, humans, the creative species, are unique. This hence may also hinder creative work.

How can your research help manage creative human-AI interactions?

Together with my advisor Professor Mathis Schulte, we investigate the differences in how people work on creative and non-creative tasks when they collaborate with AI, compete with it, or use results of its work as a benchmark for their own performance. We also investigate how these interactions are different from same situations in which a human counterpart is involved instead of an AI agent. 

Creative interactions with AI is a nascent area of managerial knowledge, and we are excited to be making first contributions to it. We show, for instance, that people use the same psychological mechanisms in comparing themselves to AI agents, even when they believe these agents can’t really compete with people creatively. We also show that what we believe about ourselves and about AI matters, and while some beliefs can make employees work harder on creative tasks, other may demotivate them. Knowing what these configurations are will help managers set up collaborations such that AI is used to its full capacity, and people blossom in creative work.

We show, for instance, that people use the same psychological mechanisms in comparing themselves to AI agents, even when they believe these agents can’t really compete with people creatively.

So, what main differences do you find between human-human and human-AI creative interactions?

First of all, we saw that people did not take as much time to work on a creative task – that is, they were not as effortful - when they collaborated with an AI as when they collaborated with a colleague. This was not the case for a non-creative task, in which people worked on average for the same time both with an AI and a human collaborator. To illustrate, in one of our experiments, the creative task was to come up with recycling ideas for a medical mask (this task could be, for example, a challenge in an engineering competition, or a topic for an entertaining article, which was the setting in our experiment). The non-creative task was to find recycling-related words in a word matrix that needed to be tested as a game for publishing online. 

word search matrix - HEC
Word search matrix

 

 

People were not as effortful to work on a creative task when they collaborated with an AI as when they collaborated with a colleague.

 

We also found that an AI’s performance did not matter for such a decrease in effort. In a different experiment, we asked the participants to propose ideas on how to motivate their colleagues to move more or to eat better. We then showed them either a list of nonsense ideas (which really was a text generated by an algorithm!) or a list of sensible ideas. We told the participants that the list they see was either created by an AI or by someone else who also participated in the experiment. As a result, some people saw a nonsense list of ideas and thought it was proposed by an AI, and others thought that same list was suggested by a human. Regarding the list of sensible ideas, a third group saw a sensible list that was ostensibly AI-written, and the last group thought it was a human suggestion. Then, we again asked the participants to propose ideas on one of the topics. We found that no matter whether the ideas made sense or not, participants who looked at ideas that were introduced as suggested by other people took almost half a minute longer on the second task than those who thought the ideas were AI-generated. 

Interestingly, this was also the case in the non-creative task, in which participants looked for a specific character on an old-Russian birch bark, looked at the results of work of an AI or a human, and looked for another character again. Although we did not expect them to, participants who saw the results of AI’s work, both done well and failed, took less time to work on the task in the second round. 
 

bark - HEC
Participants looked for a specific character on an old-Russian birch bark”. (Original birch bark from “Histrf” website)

 

bark 1 - HEC
Birch bark in which the counterpart did the task well.

 

bark 2 - HEC
Birch bark in which the counterpart failed the task.

The results of these two experiments suggest that in many cases, especially when we think AI has no place in creative work, we might be less motivated to exert effort on it than we do when working together with another human.

 

In many cases, we might be less motivated to exert effort on creative work when there is AI participation, than we do when working together with another human.

 

So what are the conditions in which AI can motivate us try harder on a creative task? 

This we tackled in a further experiment. In it, we again had participants think about creative ideas for the use of a medical mask or medical gloves. But, unlike in the first experiment, we first subtly made half of the participants remember that creativity is what makes humans unique, while the other half did not see such a reminder. We also told the participants that either an AI or a person came up either with many creative ideas, or with only a few. The results of these manipulations were astounding: while people who were reminded that creativity is a uniquely human trait and saw that AI came up with 20 ideas worked on average for almost full five minutes, those who were also reminded of creativity uniqueness but saw that it was a person that thought of 20 ideas took only 3.5 minutes to work on the task. We also saw that when high creative performance of AI was unexpected, people were much more threatened by that AI. 

What are the implications of these findings for practice?

These findings have very important implications for practice: what should we tell employees about AI-applications that can be used in creative work to not have them disengage as they have nothing to prove against supposedly non-creative AI? How do we motivate them to perform creatively without scaring them with AI? These are the questions that I am looking forward to solving in my future work.

You’re using some unusual online services to conduct your research. Can you describe how you recruit your participants and why these services are so attractive?

People who participated in these experiments were recruited either on Prolific, which is a rather popular British service for academic research, or on Yandex.Toloka, which is a Russian service for recruitment of individual workers on mini-tasks like data markup. 

Using both platforms is a great advantage. First, I can test whether the effects I hypothesize are universal across different cultures and nationalities. Second, as specialized platforms have recently been criticized for data quality, people that I recruit at Yandex.Toloka, which is primarily used by businesses for crowdsourcing, are naïve to experimental studies and, as the majority of these tasks are quite mechanical and dull, they are intrinsically motivated to participate in my experiments – I have received quite a few messages from participants saying that they enjoyed the “task” and would be willing to do it again. Although, unfortunately for these participants, that will not be possible to not compromise the integrity of the research, I think it says much about how people are unlikely to sabotage participation.

 

I can test whether the effects I hypothesize are universal across different cultures and nationalities.

 

Finally, Yandex.Toloka offers a great gender balance in the sample (51% of participants I randomly recruit there are women), people come from different walks of life and mostly have had quite a lot of working experience – their average age is almost 36 years. These factors make the use of both services a great strength of my research design, and I thank GREGHEC Laboratory for enabling me to use them by providing research support in the form of funding that got me almost 1200 participants.

See structure
Part 11

Meta and Data Privacy: a Clash of Cultures and Generations?

Marketing
Published on:

HEC Paris Assistant Professor in Marketing, Klaus Miller, analyzes the February 3 Facebook/Meta stock market plunge. What exactly does it tell us about private data on internet and its links to the advertising world? We meet Klaus on February 8, the very day he and five co-researchers self-published “The Impact of the GDPR on the Online Advertising Market”. This book focuses on Europe’s GDPR and how it affects online publicity. In a wide-ranging discussion on personal data and the advertising industry, Klaus provides insights on ad blockers on news websites and their impact on our reading habits.

klaus miller cover

Klaus Miller, Assistant Professor of Marketing at HEC Paris

Listen to the podcast:

 

In January, you published a major research paper on the impact of ad blockers on the consumption of news . Some are attributing the fall in Meta’s stock prices on February 3 partly to changes in both Europe’s and Apple’s privacy regulations. Mark Zuckerberg himself admitted: “There is a clear trend where less data is available to deliver personalized ads.” Specialists believe this trend partly explains Meta’s 23% drop in stock value. Your analysis of this argument?

Let us take a step back on this question. What we as a society (and essentially our data protection agencies and regulators) are facing is the trade-off between protecting consumer privacy, on the one hand, and fostering economic prosperity of firms, on the other. And we have seen very different regimes across the globe answering this question in very different ways. For example, in the United States, the balance has always tilted towards more economic prosperity and less data protection of the users. And in Europe we have taken a different approach, whereby we want more privacy protection and we accept higher costs or less prosperity for firms. However, we can see several changes in consumer behavior that are now affecting Meta/Facebook specifically. 

Apple allowed its iPhone users to choose whether advertisers could track them. Three-quarters said no, according to the analytics company Flurry. How is this affecting advertisers?

Indeed, with regards to Apple, every app developer within the iOS network has to ask his or her app users for consent to being tracked. And this has led to many users not consenting to this tracking, which makes it more difficult for firms such as Facebook. They rely heavily on iOS devices for their revenue, collecting personal data to target consumer with personalized advertising.

 

Advertising money is essentially shifting towards Google search advertising because firms who rely heavily on iOS devices such as Facebook are facing low consent rates and thus lower revenues as they are less able to target consumers with personalized advertising

 

So what happens if you take away online tracking? That's basically the big question. What we see is that advertising money is essentially shifting towards Google search advertising, towards Android, and away from iOS app advertising. This has a major impact on Facebook because they are highly reliant on iOS and mobile advertising revenues in general.

One of your research projects studied the activities of almost 80,000 users on a news website. The results indicate an impressive leap of consumption of news by those using ad blocks. What conclusions can you draw from this?

Basically, what we found in our research paper, “How Does the Adoption of Ad Blockers Affect News Consumption?”, is that ad-sensitive users consume up to one third more content when not seeing ads. Their behavior is transformed. There's a group of users that really hate ads, to be frank. And I think it’s very important for publishers to recognize this, especially if you consider the large number of ad blocker users are younger audiences. Our data find that, on average, 24% of all users use an ad blocker. But if you look at the younger target groups, they're up to 100% ad blocker usage. Again, I think it is very important for publishers to recognize this finding that younger target groups do not want to be monetized by seeing annoying advertising. 

 

We find that ad-sensitive users consume up to 1/3 more content when not seeing ads, and that 24% of all users use an ad blocker. I think publishers should recognize this.

 

 

Find the book online for free, or pay what you want.

 

See structure
Part 5

“A $%^* Sexist Program”: Detecting and Addressing AI Bias

Artificial Intelligence
Published on:
7 minutes

A major issue facing companies that use AI, algorithmic bias can perpetuate social inequalities — as well as pose legal and reputational risks to the companies in question. New research at HEC Paris offers a statistical method of tracking down and eliminating unfairness.

facial recognition - cover

©metamorworks on Adobe Stock

Listen to the podcast:

 

Soon after Apple issued its Apple credit card, in August 2019, urgent questions arose. A well-known software developer and author, David Heinemeier Hansson, reported in a Tweet that both he and his wife had applied for the card. “My wife and I filed joint tax returns, live in a community-property state, and have been married for a long time,” Hansson wrote. “Yet Apple’s black box algorithm thinks I deserve 20x the credit limit she does.” He called it a “sexist program,” adding an expletive for good measure.

 

AI based credit scoring - Siberian-Art-AdobeStock
"If a credit-scoring algorithm is trained on a biased dataset of past decisions by humans, the algorithm would inherit and perpetuate human biases." (Photo ©Siberian Art on Adobe Stock)

 

Goldman Sachs, the issuing bank for the credit card, defended it, saying that the AI algorithm used to make determinations of creditworthiness didn’t even take gender into account. Sounds convincing, except that this ignored the fact that even if specific gender information is removed, algorithms may still use inputs that correlate with gender (“proxies” for gender, such as where a person shops) and thus may still produce unintended cases of bias.

Even Apple’s cofounder Steve Wozniak reported that he and his wife had experienced this bias. Wozniak was judged worthy of 10 times more credit than his wife, despite the fact that he and his wife shared assets and bank accounts. The ensuing melee resulted in an investigation of the Apple Card’s algorithm by New York regulators.

Biased data leads to biased results 

AI/machine learning can process larger quantities of data more efficiently than humans. If applied properly, AI has the potential to eliminate discrimination against certain societal groups. However, in reality, cases of algorithmic bias are not uncommon, as seen in the case of Apple, above.

 

If a credit-scoring algorithm is trained on a biased dataset of past decisions by humans, the algorithm would inherit and perpetuate human biases.

 

The reasons for this bias are various. If, for example, a credit-scoring algorithm is trained on a biased dataset of past decisions by humans (racist or sexist credit officers, for example), the algorithm would inherit and perpetuate human biases. Because AI uses thousands of data points and obscure methods of decision making (sometimes described as a black box), the algorithmic biases may be entirely unintended and go undetected.

 

"When machine learning techniques, which are often difficult to interpret, are poorly applied, they can generate unintended, unseen bias toward entire populations." (Photo ©Nuthawut on Adobe Stock)

 

In credit markets — the focus of our work — this lack of fairness can place groups that are underprivileged (because of their gender, race, citizenship or religion) at a systemic disadvantage. Certain groups could be unreasonably denied loans, or offered loans at unfavorable interest rates — or given low credit limits. A lack of fairness may also expose the financial institutions using these algorithms to legal and reputational risk. 

A “traffic light” test for detecting unfair algorithms

My fellow researchers, Christophe Hurlin and Sebastien Saurin, and I established a statistics-based definition of fairness as well as a way to test for it. To ensure fairness, decisions made by an algorithm should be driven only by those attributes that are related to the target variables, such as employment duration or credit history, but should be independent of gender, for example. Using statistical theory, we derived a formula to compute fairness statistics as well as the theoretical threshold above which a decision would be considered fair. 

 


We established a statistics-based definition of fairness as well as a way to test for it.


When dealing with an actual algorithm, one can first compute the fairness statistics and compare them to the theoretical value or threshold. It is then possible to conclude whether an algorithm is “green” (when the fairness statistics are greater than our established threshold) or “red” (when the fairness statistics are less than the threshold). 

Second, if there is a problem, we offer techniques to detect the variables creating the problem — even if the algorithm’s processes are impenetrable. To do so, we developed new AI explainability tools. Third, we suggest ways to mitigate the problem by removing the offending variables.

 

 

We developed new AI explainability tools to detect the variables creating the problem of unfairness.

 

From a purely practical, business perspective, it is important that banks understand the implications — and potential unintended consequences — of the technology they are using. They may risk running afoul of both the justice system and public opinion — and it goes without saying that reputation and trust are key in the banking industry.

Application across diverse fields

While our focus has been on credit scoring, our methodology could potentially be applied in many other contexts in which machine learning algorithms are employed, such as predictive justice (sentencing, probation), hiring decisions (screening of applicants’ CVs and videos), fraud detection and pricing of insurance policies.

 

"Our methodology could potentially be applied in many contexts in which machine learning algorithms are employed, such as predictive justice, hiring decisions, fraud detection and pricing of insurance policies." (Photo ©artinspiring on Adobe Stock)

 

The use of machine learning technology raises many ethical, legal and regulatory questions. When machine learning techniques, which are often difficult to interpret, are poorly applied, they can generate unintended, unseen bias toward entire populations on the basis of ethnic, religious, sexual, racial or social criteria. The opportunities and risks that come with machine learning techniques undoubtedly call for the implementation of a new form of regulation based on the certification of the algorithms and data used by companies and institutions.
 

Applications

Focus - Application pour les marques
In the short term, we aim to help companies and institutions that use AI to better understand the decisions of their algorithms and to detect potential unintended consequences. In the longer term, we hope to contribute to the discussion about guidelines, standards and regulations that public administrators should institute.

Methodology

methodology
Drawing on work conducted over 15 years on risk model validation, we developed new statistical tests that detect a lack of fairness. This “traffic light” test statistically analyzes whether an algorithm’s decisions are fair (“green”) or unfair (“red”) against protected societal groups. If an algorithm’s decisions are found to be unfair, we suggest techniques to identify the variables responsible for the bias and to mitigate them.
Based on an interview with Christophe Pérignon and his academic article “The Fairness of Credit Scoring Models,” co-written with Christophe Hurlin and Sébastien Saurin, both from the University of Orléans.
See structure
Part 13

Developing an Automated Compliance App to Help Firms Comply with Privacy Regulations

Law
Published on:

According to recent research published by the Harvard Business Review, a strong privacy policy can help firms save millions. If privacy policies have indeed become more important for firms and consumers today, the real challenge lying ahead is to ensure compliance within and beyond the firm. A new tool developed at HEC Paris uses artificial intelligence to help managers, data protection officers, and project leaders take privacy-preserving decisions and implement privacy by design. 

data compliance - leowolfert-AdobeStock

 ©leowolfert on Adobe Stock

In the back end of privacy policies, a firm needs to ensure its own processes and those of its supply chain are in line with privacy regulations, so what is presented to consumers in the front end through privacy policies is realistic and effective. 

A new tool, using artificial intelligence methods, including machine learning, builds on a year-long effort and is currently under test with industrial partners. It was developed by a multi-stakeholder initiative led by HEC Paris Professor David Restrepo Amariles, Aurore Troussel (LL.M. HEC. 19), and Rajaa El Hamdani, data scientists at HEC Paris.

 

“By ticking this box you accept...”

Think back to the last time you signed up for a new website or online service. Did you read the terms and conditions before you clicked “accept”? If the answer to that question is an embarrassed “no”, then don’t worry, you are not alone. The length and the vocabulary used in most of privacy documents make data processing of companies time-consuming and difficult to understand. Researchers at HEC Paris developed Privatech, a new machine-learning powered application which detects breaches to General Data Protection Regulation (GDPR) in privacy documents.

 

Researchers at HEC Paris developed Privatech, a new machine-learning powered application which detects breaches to GDPR in privacy documents.

 

woman using an app on a smartphone - Golden-Sikorka AdobeStock
"This application could serve consumers, lawyers, data protection officers, legal departments, and managers in auditing the privacy documents of a company." (©Golden Sikorka on Adobe Stock)

 

This application could serve consumers, lawyers, data protection officers, legal departments, and managers in auditing the privacy documents of a company. But more importantly, it aims to further generate privacy compliance in the back end of data flows, or in other words, to ensure companies are informed of data practices so they can take privacy preserving decisions. Privatech allows managers who are not specialized in privacy protection to conduct a preliminary compliance assessment and detect potential issues requiring specialized advice.

 

Privatech allows managers who are not specialized in privacy protection to conduct a preliminary compliance assessment and detect potential issues requiring specialized advice.  

 

 

The challenge for businesses: complying with EU (and US) law

The General Data Protection Regulation came into force in 2018 and many companies saw this regulation as a challenge in terms of compliance. In 2017, 90% of executives consider GDPR to be the most difficult form of compliance to achieve1. GDPR requires companies to govern their data processing while ensuring data subject’s rights. Under GDPR, companies have to set up procedures and documents enabling users to access clear information about their personal data processing and to control this processing.

 

The cost of non-compliance is estimated to be 2.71 times the cost of compliance.

 

Two aspects of GDPR are of particular importance for businesses. First, GDPR has a very broad application’s scope, far outside EU borders. Second, GDPR sets forth fines of up to 10 million euros or up to 2% of the company’s entire global turnover, whichever is higher. It explains why the cost of non-compliance is estimated to be 2.71 times the cost of compliance2. In addition, the recent entry into force of the California Consumer Privacy Act (CCPA) shows that the data processing of companies will be more and more scrutinized by regulators. This regulatory trend makes investment in privacy compliance technologies relevant. 

An app built with a coalition of law and business

Privatech uses machine learning to automate the process of complying with legislation. With the help of several law firms and companies including Atos, one of the largest IT companies in Europe, HEC Paris researchers created a tool for automating the assessment of privacy policies. 

 

The tool can read privacy policies and detect lines that might not be compliant with the law or may leave consumers' personal data open to exploitation.

 

This means the tool can read privacy policies and detect lines that might not be compliant with the law or may leave consumers' personal data open to exploitation. To develop this tool, researchers relied on annotated privacy policies, with clauses’ labels corresponding to different data practices, and connected these annotations to corresponding GDPR articles and obligations.

Data practices are categories of data processing activities, for example “data retention” which is a data practice that refers to how long data can be stored. Each paragraph in the privacy policies was tagged to a corresponding data practice. We then trained a machine-learning algorithm to identify and label different data practices in a legal document. The app also assesses readability of privacy policies because a key aspect of GDPR requires privacy policies to be easily readable. The app is calibrated so that all text should be readable by any high-school student.

Reshaping privacy compliance from the ground up

Privatech aims to streamline privacy compliance and consumer protection by focusing on firms (data controllers and processors) rather than on consumers (data subjects). The application may help individuals to better understand the privacy policies that they would otherwise sign blindly. However, we decided to focus on companies as they could generate privacy compliance by design and are liable under GDPR.

By focusing on firms, Privatech aims to ensure companies are able to translate privacy policies disclosed to consumers into effective corporate compliance mechanisms. We expect that Privatech will eventually encourage companies to design and monitor their data processing activities, so they are legal, comprehensive and easy to understand.
 

Applications

Our work will be valuable for any companies and data handlers who have a need to comply with data protection legislation. The project will reduce the need for the repetitive and labour-intensive elements of legal assessment of privacy documents and improve compliance with legislation. Our work will also be valuable for consumers who may struggle to interpret privacy documentation such as GDPR. Ultimately, data protection authorities could also use the application to conduct audits and monitor compliance with GDPR.

 

Methodology

The project started as a class deliverable for two courses offered by Professor David Restrepo at HEC Paris, TechLaw offered in the LL.M program, and Data Law and Technology Compliance offered in the MSc Data Science for Business jointly organized by HEC Paris and Ecole Polytechnique. A first beta version relied on the students’ work and collaboration from lawyers at Baker McKenzie Paris and tech entrepreneurs, including Croatian developer Vedran Grčić. Since August 2019 the project is fully developed in-house at HEC Paris by the Smart law Hub, which integrates law scholars and data scientists.

The project has also changed its methodology and focus. The application is developed to detect unlawful or problematic sentences within privacy policies and to evaluate the complexity of privacy documents.

The algorithms have been trained on data retrieved by researchers. The training data set is composed of sentences retrieved from various privacy policies, and judicial and administrative decisions. These sentences were labelled and categorized by data practices such as “data retention” or “data sharing”. This preparatory work allowed for the creation of a machine-learning algorithm able to identify and label different data practices in a legal document.

In addition, a readability algorithm evaluates the complexity of privacy document to verify its compliance with transparency and explainability requirements. The main focus of the research today is compliance generation which seeks to monitor internal documents and documents in the data supply chain.

1Globalscape and Ponemon Institute, 2017: The true cost of compliance with GDPR
2Globalscape and Ponemon Institute, 2017: The true cost of compliance with GDPR (53 multinational firms audited)
 
Based on an interview with HEC Paris’s David Restrepo Amariles and Rajaa El Hamdani. Professor Restrepo Amariles won the Bruno Roux de Bézieux Award for Educational Initiative from the HEC Foundation for his project “Automating GDPR Compliance”. This SMART Law Hub’s work was selected to be presented in the Conference on AI and Law 2021 (more here).
See structure
Part 14

How Should We Design the Next Generation of AI-Powered Chatbots?

Information Systems
Published on:

Have you been enraged by a chatbot recently? Businesses from Starbucks to LinkedIn and Singapore Airlines to eBay are embracing AI-powered communication tools. But the success of using chatbots to streamline operations and deliver consistent, round-the-clock customer support hinges on gaining user trust, show recent findings from HEC Paris Professor Shirish Srivastava and his co-authors Dr. Shalini Chandra of the SP Jain School of Global Management, Singapore and Dr. Anuragini Shirish of the Institut Mines Telecom Business School, France.

man chating with a chatbot on his cell phone

©tippapatt on Adobe Stock

Microsoft launched its ChatGPT-powered search engine chatbot earlier this year. Within three months of its release, the Bing chatbot has engaged in over half a billion chats, has over 100 million active daily users, and app downloads have increased fourfold. Since Microsoft's paperclip assistant, Clippy, first hit our screens in 1995, we've been trying to make computers more human-like, in order to smooth the often-awkward interface between the biological brain and synthetic software.

Adoption rates of conversational AI tools could more than double over the next five years, but the success of the Bing chatbot and others like it hinges on user acceptance and engagement. Many people are still uncomfortable using this technology, find it unfit for purpose, or unable to empathize and understand their needs. When computer pioneer Alan Turing conceived the "imitation game" 70 years ago, he understood that our demand for AI to think in a way that was indistinguishable from a human would be vital to its adoption.

In a world where 86% of customers still prefer humans to chatbots for online interactions, we wanted to explore how businesses could improve user engagement with chatbots.

Testing out three chatbots for their humanness

We recruited around 200 people with little to no experience with chatbots and presented them with Mitsuku, Bus Uncle, and Woebot.

Mitsuku (now known as Kuki) is designed to befriend humans in the metaverse, taking the form of an 18-year-old female. Able to chat, play games and do magic tricks at the user's request, Mitsuku is said to have inspired the movie Her and is a five-time winner of a Turing Test competition called the Loebner Prize. Woebot is designed to be a mental health support. The chatbot describes itself as a charming robot friend and uses AI to offer emotional support through talk therapy, as a friend would. And Bus Uncle is used across Singapore to tell users about bus arrival timings and deliver public transport navigation, supposedly with the “personality of a real uncle.”

We collected data on the individual characteristics of our respondents and their general perceptions of chatbots. Next, we gave the study respondents time with one of the three chatbots. We then asked them about their experience with it. For a different perspective, we also conducted ten interviews with frequent chatbot users about their interactions and experiences.

What human-like competencies in chatbots foster user engagement?

To foster user engagement, chatbots must create a natural, human-like interpersonal environment that enables natural, engaged communication. We must learn to program AI to deal with more intricate tasks, using humanized social skills to query, reason, plan, and solve problems. But which aspects of human behavior should be prioritized when designing conversational AI programs?

 

AI designers should be aware that building believable, human-like qualities into conversational AI is as important as improving its efficiency.

 

We looked at media naturalness theory (MNT), which suggests that face-to-face communication is the most natural and preferred form of human communication. The theory suggests that three main mechanisms govern how natural, smooth, and engaging interaction with technology feels:

  • A decrease in cognitive effort, i.e., the amount of mental activity a user expends communicating with the technology,
  • A reduction in communication ambiguities, i.e., misinterpretations and confusion in how the user interprets the messages,
  • An increase in physiological arousal, i.e., the emotions that users derive from interacting with the technology.

Translating these ideas into the context of chatbots, we came up with three competencies we thought such AI would need to engage users:

  • Cognitive competency – the ability of chatbots to consider and apply their problem-solving and decision-making skills to the task at hand in creative, open-minded, and spontaneous way. 
  • Relational competency – the AI agent’s interpersonal skills, such as supporting, cooperating, and collaborating with its users, appearing cooperative and considerate.
  • Emotional competency – AI’s ability to self-manage and moderate interactions with users, accounting for their moods, feelings, and reactions by displaying human warmth and compassion. 

Through our research, we have confirmed that cognitive and emotional competency builds user trust and engagement. This was less clear for relational competency because chatbots do not yet have the skills to remember previous interactions to build relationships and readjust appropriately to the task demanded. Though the importance of relational competency was confirmed by the qualitative interview study in our mixed methods research, we learnt that most chatbots do not offer enough opportunities to develop relational competency. This is a rich area for future research.

How important is users' trust in their engagement with chatbots?

We proposed that user trust in chatbots is the central mechanism that mediates the relationship between human-like interactional competencies in AI and user engagement with conversational AI agents. 

Prior research has shown that trust in technology influences its use and take-up across contexts such as m-commerce portals, business information systems, and knowledge management systems. We also know that users are more likely to trust interactive technologies with human-like features such as voice and animation. Such trust is expected to be more pronounced in the case of AI-driven customer interactions because as they interact with it, users can constantly gauge the AI's value in terms of its human-like attributes as it responds to their different queries and responses.

Our findings show that human-like cognitive and emotional competencies serve as innate trust-building mechanisms for fostering user engagement with chatbots.
 

Methodology

We asked 200 people with little to no experience with interactional AI about their perceptions of chatbots, asked them to interact with one, and then interviewed them to ask about their experiences with it. For a different perspective, we also conducted ten interviews with frequent chatbot users about their interactional experiences.

Applications

AI designers should be aware that building believable, human-like qualities into conversational AI is as important as improving its efficiency. And to boost use customer engagement, organizations need to prioritize giving chatbots cognitive capabilities that closely align with how we as humans think. They should design these tools to understand and respond to the spectrum of human emotions, providing appropriate responses and making decisions in response, such as redirecting customers to live service agents during frustrated exchanges. AI practitioners and designers should also consider ways to increase relational competency in conversational AI agents and chatbots by introducing relatable human attributes.
Based on an interview with Professor Shirish C Srivastava on his HEC Foundation-funded research “To Be or Not to Be… Human? Theorizing the Role of Human-Like Competencies in Conversational Artificial Intelligence Agents,” co-written with Dr Shalini Chandra from the S P Jain School of Global Management in Singapore and Dr Anuragini Shirish from Institut Mines-Télécom Business School in Paris, published in the Journal of Management Information Systems, December 2022. This research has been selected to receive funding from the HEC Foundation's donors.
Shirish Srivastava
Shirish Srivastava
GS1 France Chair Professor

Related content on Artificial Intelligence

Podcasts
AI Can Level Global Playing Field
Carlos Serrano
Carlos Serrano
Associate Professor
Brian Hill GREGHEC
Brian Hill
CNRS Research Professor
Artificial Intelligence

How AI Can Help Figure Out When Hospital Patients Need Intensive Care

By Julien Grand-Clément

Artificial Intelligence
Make AI Your Ally: Editorial
Nicolas Vieille_KnowledgeHEC
Nicolas Vieille
Professor, Scientific Co-Director of Hi! PARIS
Emmanuel Kemel HEC professor
Emmanuel Kemel
CNRS Research Professor
Information Technology

Using Innovations on Social Media for More Engagement? Be Aware of The Cultural Differences

By Reza Alibakhshi, Shirish Srivastava

Subscribe button for Knowledhe@HEC newsletter

Newsletter knowledge

A monthly brief in your email box and 3 issues of the book per year.

follow us

Insights @HECParis School of #Management

Follow Us

Support Research

Our articles are produced thanks to our reader's support