Daron Acemoglu
Thank you, Cristina, for that wonderful introduction and to you and Tommaso for 
inviting me. […]
As Cristina said, I'm going to talk about something that's partly inspired by 
my book. AI and
antitrust in 10 minutes. So that's a tall order especially if I try to blend in 
ideas from the book, so
let me jump into it. I'm going to do 10 question and answers in 10 minutes, but 
since that's a very
short time I'll just give you the answers. I'll let your imagination do the job 
of the what the questions
might have been to which. These are the answers.
- Yes, generative AI has great potential, so I am completely convinced that 
this is a very
interesting technology that can bring lots of goods and has capabilities so we 
can build on
that. But I think let's move forward.
- And yes, I believe that monopoly is everywhere in the tech sector. So here, 
perhaps I differ
from many IO economists, and I subscribe to the duck test. If something looks 
like a duck,
walks like a duck, and quacks like a duck, it is a duck. So if you have 
companies that have
reached sizes that have never been in human history, and that dominate a 
particular line
of business, they are monopolies. So we have to grapple with that. And that 
means all sorts
of regulatory tools have to be considered, including antitrust. So this is 
absolutely on
target.
- And yes, in my view this is getting worse with foundation models. Because 
there is a
likelihood that we may go towards a duopoly. With Microsoft Open AI and Google 
as the
two key players, even though open source and many other competitors are going 
to try to
get into foundation models, but the current business model of foundation models 
is very
resource intensive. So that raises the possibility, does not in any way creates 
a certainty,
but it raises the possibility. These two companies and their models are going 
to be the
dominant ones on which many others will have to build, raising all of the 
issues of vertical
product creation and all sorts of other questions that are going to be central 
for
policymakers and economists to grapple with.
- But no, I actually don't think monopoly power leading to high prices is the 
main problem
that we're dealing with. You know, of course, that is a problem. But if the 
only issue was
that because the foundation models are controlled by, you know, Google and 
Microsoft,
they’re going to charge higher prices, and as a result, the apps that are 
developed on them
are going to be more expensive, that would be, of course a pity and it's 
something we can
do something about, but it wouldn't be the end of the world. So we get many new 
apps.
They cost a little bit more. We don't get quite the consumer surplus. Woe is 
us, but not the
end of the world. The problem is the direction of innovation. The problem is 
that the current
market structure is selecting a particular direction of innovation. And that 
has much more
sweeping consequences. Taking the set of products and technologies as given and 
pricing
them above marginal cost and thus losing some of the welfare triangle is not 
the main issue.
There’s the potential for doing much greater damage. No, this is not because of 
existential
risk. In fact, like Cristina was implying, when all of these tech leaders are 
talking about
existential risk I see it as either a blind spot or a ploy for making us not 
worry about the
bigger risks. The bigger risks in my mind are in the labour market. Most of us 
earn our living
in the labour market, so what happens to jobs is the most important issue. And 
the current
direction of AI looks like it is going to follow some of the trends we have 
seen with digital
technologies before. Failing to create the complementarities which human 
workers and
skills, and instead going much toward much more towards automation, hence 
generating
inequality, potential job losses, especially for workers without very 
specialised skills such
as those with postgraduate degree.
- And no, it's not just economics. There is a real danger here that the current 
direction of
generative AI could again continue existing trends that we saw in social media 
degrade
political conversations. Increase the amount of misinformation and 
disinformation, with a
much more powerful tool. Create a particular type of ecosystem in online forums 
where
people are drawn on the basis of emotion rather than engagement. And hence 
generally
act towards the exploitation of people in their capacities and duties as 
democratic citizen.
It is this twin: Inequality and Elimination of good jobs in the labour market, 
as well as
erosion of democratic capacity that I think are most problematic.
- No, I am not a Luddite. So I am not saying that this is in the nature of 
technology, nor that
we should oppose technological change. The issue here is that we are not along 
the right
path. What's great about technology in general and generative AI in particular, 
that it's a
very highly malleable type of technological platform or what some economic 
historians
used to call general purpose technology, meaning that you can use it for 
creating many
apps, many different types of sub technologies and many different directions 
are possible.
It is not complete idle talk. When people used to talk about social media and 
other online
tools creating new democratic spaces today, it looks like very naive. When 
people in the
2000s said oh, online communication and social media are going to democratise
communication. But that potential was there, and that potential is much greater 
with AI.
When some people in the tech industry talk about generative AI being useful to 
humans in
terms of getting better information, performing better tasks so generative AI… 
actually, I
think the great potential that I mentioned at the beginning is precisely in 
being a human
complementary technology. The tragedy of our current age is that we have almost 
all
information that is at least codified available in some form. But we do not 
have the
processing power to decide which one we should retrieve, how we should 
interpret how we
should process, and which types of information we should engage with in 
different forms.
Generative AI has the capability to improve human interaction with information 
and hence
generate a lot of tasks, not just for knowledge workers, but for electricians, 
for carpenters,
for educators, for healthcare workers. So that possibility is there.
- But no, we are not going in the right direction. So we do need a redirection 
of technological
change.
- And no. I don't think it is naive or unrealistic to think about the 
redirection of technological
change. One view which is common among some economists and some tech leaders 
goes
back to either to the view that technology somehow has a preordained path and 
we just
have to follow it. No, I am denying that and I think history is quite a good 
guide on showing
how malleable technology is. Goes back to a saying by Ferdinand de Lesseps of 
fame, from
the Suez and the Panama Canal, which we discussed in this book where he said, 
don't worry,
men of genius would arise and solve all problems. Who are today's men of 
genius. Maybe
some Altman or Elon Musk. But no, I don't think we should trust them. So I 
think the
direction of technology is malleable, but it's also a societal choice.
- And yes, as Cristina was hinting, antitrust has a very important role in 
this. For two reasons.
One, because if we want alternatives. They are not very likely to come from a 
duopolistic
or highly oligopolistic structure, especially one further empowered by killer 
acquisitions
and these companies being a block other types of technologies that do not fit 
well with
their business model. So if we want more alternatives that go more in a human
complementary direction or more pro democratic direction, or create a more open
competitive environment, I think we have to use antitrust tools including 
potential breakup
of the largest companies which are too big and one other reason is because this 
type of
power comes with enormous social power. And by that social power, I mean 
economic
power, and also general social power. Tech companies have an enormous sway on 
public
opinion, which I think is associated with their mega profits. And again, I 
don't think that
creates a healthy environment.
- And no, finally, I don't think antitrust is the main tool as Cristina was 
also hinting in her
introductory comment. I think antitrust is a very blunt tool and I think for 
the redirection
of technology we need a suite of tools which should include exactly how data is 
used and
accessed. We need a new interoperability type of approach as well as how do we
compensate and how do we actually encourage more creation of creative data. We 
also
need to provide explicit incentives, such as, for example, those that have been 
successful
in the field of renewable technology, where we encourage more of the socially 
valuable
types of technologies. And I think there are a number of tools for that and we 
may also
need more tax policies to discourage the worst types of business models and 
create room
for alternative business models. So I am actually quite favourable, although I 
think much
more study is needed, to a digital ad tax that creates more openness for 
alternative
business models based on things like Wikipedia or subscription models in the 
online space.
- Finally, I think we also need to rethink other tools that we have, like 
fiscal tools that
currently create a very asymmetric playing field between capital and labour and 
going back
to the job market and labour market inequality issues. I think equating 
marginal tax rate
between capital and labour are things that we should definitely revisit. Thank 
you.


https://mailchi.mp/cepr/central-bank-communication-rpn-seminar-series-516707
_______________________________________________
nexa mailing list
[email protected]
https://server-nexa.polito.it/cgi-bin/mailman/listinfo/nexa

Reply via email to