
Summer usually brings everyone's rhythm a bit down but rest assured, governments are busy figuring out how to reign in Big Tech's overwhelming grip over our everyday lives. Things are in motion and have been for a while now..
On this issue we are joined by Pedro Bizarro, cofounder and Chief Science Officer at Feedzai, to discuss responsible A.I., the evolution of the Portuguese startup scene, the struggle of hiring and the future of the country's entrepreneurial ecosystem.
- Lawrence
G7 leaders
reached an "historic" agreement
to tackle tax abuses by internet giants and to introduce a global
minimum corporate tax rate of 15 percent.
European consumer lobby group is
backing the E.U.'s antitrust case against Apple, which alleges it distorts competition in the music streaming
market while E.U.
antitrust officials started investigating Google's ads business. Germany's Federal Cartel Office is getting busy
having
launched proceedings against Apple over pre-installed apps
and in-app purchase system,
as well as against Google's News Showcase
on whether it hinders competition.
The U.K. is not staying behind with the Competition and Markets
Authority which started an
investigation into Apple and Google
over their dominant position in the mobile phone market as well as
opening a probe against Google and Amazon over fake reviews.
In the U.S. top antitrust lawmakers
introduced a legislative package
that could overhaul the nation's antitrust laws in an attempt to
rein in the power of Amazon, Apple, Facebook and Google. Also,
Democrats and Republicans came together to
confirm Lina Khan to be chair of the Federal Trade Commission.
Russia fined Facebook and Telegram
for unlawful content.
Nigeria suspended Twitter "indefinitely"
after the platform removed a post from the president.
The standoff between the Indian government and Twitter
escalated
after accusing the social media giant of not complying with local
laws.
Amazon settled for $61.7m over pocketing driver tips
but is
facing a possible €425m fine in the E.U.
related to alleged violations of Europe's General Data Protection
Regulation.
Google said it will adapt its ad technology after France
delivered a $267m fine, making it easier for competitors to use its ad tools.
A Texas Supreme Court ruled that
Facebook could be held liable if sex traffickers use the platform
to prey on children.
Since 2018,
Amazon Web Services has hired at least 66 former government
officials, most directly from government posts and more than half from the
Defense Department.
Apple's new privacy feature, designed to mask users' internet
browsing,
won't be available in China.
Facebook has suspended ex-U.S. president Trump for 2 years
Bing temporarily censored image searches for 'Tank Man', even in the U.S.
YouTube blocked videos from a human rights group of testimonials
about missing Uyghurs in China
as they contained ID cards.
"Experts Doubt Ethical AI Design Will Be Broadly Adopted as the Norm Within the Next Decade" by Pew Research Center
"Collective data rights can stop big tech from obliterating privacy" by Martin Tisne
"Algorithms already fire and 'give' loans. How to make them fair" (interview with Pedro Saleiro from Feedzai, in Portuguese) by João Tomé
"A Homeless Amazon Warehouse Worker in New York City Tells Her Story" by Lauren Kaori Gurley
To get these delivered to your inbox, subscribe to CFT's newsletter at the end of the page.
Lawrence - Welcome! For this issue we have the pleasure of
being joined by Pedro Bizarro.
Pedro
is one of the co-founders and the Chief Science Officer of
Feedzai, where he
has helped develop the company's industry-leading artificial
intelligence platform to fight fraud.
Amongst other things he's worked for CERN - the European Organization
for Nuclear Research. He's been an official member of the Forbes
technology council, a visiting professor at Carnegie Mellon university
and I could go on but I think that gives a pretty good picture of who
we are talking with today. So welcome Pedro and thanks for being here
with us.
Pedro - Thank you, thank you. My pleasure. Thanks for inviting me.
L So I'm really happy
to have you here because this project, as we discussed last time, is
about giving visibility to great examples that are acting in
everything that is ethical and responsible AI, which is precisely
what I want to talk about with you.
So to set the stage for listeners that are not familiar with
Feedzai, briefly, Feedzai is a company that provides financial crime
detection services to financial institutions using big data and
artificial intelligence.
P That's right. So we work with large financial institutions, large banks and payment processors and also we have large merchants.
L You have been one of the companies in Portugal that has been talking more visibly about responsible AI. That is definitely an important topic for you guys at Feedzai. This topic is new into the mainstream conversation I would say. Over the past five, six years, it's been heavily more discussed. So it looks that you guys in a way have been ahead of the curve, at least in Portugal, talking about this and worrying about this. And I'm very curious to understand why that's the case. Can you tell us how that came to be?
P Well it's that old
sentence "with great power comes great responsibility". And I do
feel that AI is a great power and is being used almost always for
good, in a lot of use cases throughout our lives. Sometimes we don't
even realize we are using AI. It could be a Google search, an Amazon
recommendation, Netflix, Spotify, basically all the major services
that we use today are using AI in the background. Even a job search
on LinkedIn and so on.
So in about 2016, when I read the now famous book
"Weapons of Math Destruction"
by Cathy O'Neil, it was the first time that it really sank in for me
that you can develop really good machine learning models according
to some business goal, but without you realizing you can produce
some very nasty side effects that can affect people in their daily
lives.
Sometimes the people are not aware, the companies are not aware.
Even the data scientists that were designing the systems, sometimes
they didn't have any bad intentions, but they were not even aware.
That was really the first time that I realized this can happen
basically anywhere because we can introduce bias by lots of
decisions with choices of data sets, sampling strategies, decisions
of what to do when there's missing data, model parameters. So
there's a whole number of decisions that you make when building a
model and some of those can produce biased or unfair results where
the model is hurting a group more than others or reducing the
likelihood of detection for example.
And not only it can hurt in that specific case but, even more
concerning, sometimes these could create feedback loops where the
model is affecting the reality and the reality is affecting the
model and so on. So you keep on making it worse over time.
So that's when we started at
Feedzai looking at responsible AI. We were already, even before that, looking at concerns of model
explainability, how to explain model decisions to users. That was
already a typical concern in financial institutions, because
sometimes as a model that is deciding if you allow an account to be
opened or not, if you block a card or not, and then you want to
either explain that to the end-user or internally to a data
scientist or even to an external person like a regulator.
So there was already a need for explaining, but it was not until
then that I realized that besides the need to explain, there were
other more complex needs about the responsible use of AI.
L So being
transparent is part of being responsible, right? Being able to
account for why
the algorithm decided X instead of Y, right? Because someone is going to ask "why can't I have access
to this insurance or this loan or this service and the other can?".
So, the term responsible or ethical, you know - I'm going to ask a
tougher question now - can be vague, right? It can be interpreted in
many ways. So I'm curious to know how you interpret it at Feedzai?
P That's I think a great
point. You are very right. Responsible is a broad term and I'm
pretty sure that if you ask 20 different people you have 20
different answers. In my perspective responsible AI is one of those
umbrella terms that includes many things inside of it and for me it
includes at least the following.
We are developing responsible AI systems if first they are fair so
they are not hurting one group more than others. So it's really
about the disparate impact. Are we impacting these groups more than
another? That is one component of responsible AI which is having
less biased or unbiased models.
Another component of responsible AI has to do with expandability. Do
we understand the model? Do we trust the model? There's many times a
human in the loop and the human in the loop can trust or not the
model and sometimes they are trusting too much. For example,
computer say so, so if the computer says so I'll press it [the
button] and sometimes they shouldn't and vice versa.
Sometimes they are not trusting the model and they should trust the
model. So there is a huge component of how we the users, the humans
in the loop, trust or not the model. How are they understanding the
model decisions (if they are understanding the model decisions)? So
those components also play a role.
So the first one was fair model, the second one has to do with
expandability and trust.
A third point for me of responsible AI has to do also with what
nowadays is called MLOps:
machine learning operations. Maybe not many people include these within responsible AI but I
also include it in the sense that, for example, models can degrade
over time by concept drift, by data drift, if you're using an
anniversary or if your data center has less resources or more
resources.
What happens is that in reality, the world is changing and that
causes the model to also change its performance. So I think it's a
more responsible use of AI if, when the model degrades, you discover
that. You find out that the model degrades and you automatically
retrain the model or raise some sort of alert for people to retrain
or to find out that something is incorrect.
For me that is also a dimension of responsibility. How to adjust to
the changing world and also how to adjust to reasonable resource
uses?
We know that there's a big concern with modules that are too
expensive to train that they are taking thousands, sometimes tens of
thousands of machines to train so their energy consumption is too
high sometimes for the benefit that they bring.
L Yeah, you're pondering the trade-offs regarding the gains. The effort and the drawbacks in terms of energy? That's a good point, yeah.
P And maybe a fourth
dimension of responsible AI is so not only the model must be fair
and explainable and automatically adjust to the changing world, but
it must also be used for good right?
You can have a fair model that hurts men and women in the same way,
but hurts both of them, right? Maybe you have a model that is
manipulating people into buying something or to adapt or something
like that.
So not only the model itself should be less biased and unexplainable
and use a good amount of resources but what you do with it is of
course also a big part of the responsible AI perspective.
Those are the four big dimensions of responsible AI.
L Right. You
mentioned something which also is one of the things that I want to
talk with you to the extent that you can talk about it, which is, as
you said, very well put: it is a moving target. You will never reach
it and then that's done, like "we have achieved fairness". You need
to constantly tweak it. And so you need to understand, as you said,
whether or not something should be tweaked? Is it outside of the
model that we deem being fair, being right?
Have you guys developed anything like, you mentioned alarms. Have
you developed something that tells you something is going off track?
P Yes, so we have a
number of people working on what we call "auto everything". In
reality, they are working in what's now called MLOps. We've been
developing tools to do what we call model monitoring and feature
monitoring.
So for example we are monitoring how the distributions of the scores
of the model are changing over time? So you expect some distribution
of scores so you train your model, you measure the distribution of
scores and then in runtime you can keep track of that distribution
of scores and you can realize: "Oh!, something is very off here. My
distribution of scores is way off compared to what I was expecting".
So that's one thing we do. We are automatically monitoring the
model. And that was our first step.
We are also working on what we call feature monitoring. So one thing
is monitoring the models: the end results, the decision of the
model. But the other one is monitoring the input to the model. So
the features that go into the model. And that's slightly more
complex because first there's hundreds of them. So not just
monitoring the final decision, but you are monitoring hundreds or
thousands of features. And if there are many features, it's likely
that any one of them is maybe off for some reason.
We already developed even patent pending work on feature monitoring
and not only on the part of detecting that the features change, but
also that efficient part. How can you do that efficiently? Because
in production, you are having thousands of transactions per second,
each one with hundreds of features and each one needs to be
processed in a few milliseconds.
How are you computing all of those distributions efficiently and
keeping track of that? And also from a statistical perspective,
because there are so many features and it's likely that a few of
them are off for some reason. You don't want to raise too many false
alerts, so how do you statistically decide when it's really time to
realize that something changed and to raise an alarm?
And also from a user interface perspective. How will you monitor a
stream of data with 500 features? You cannot show 500 features so
you need to be smart in how to show alerts and allow people to drill
down into the alert and then identify the features that are changing
and what's going on now compared with what normally should be going
on. And what were the records of the instances that caused that
change and so on? We are working in all of those areas.
L Very interesting.
So let me get back a bit just before. You became aware that with
great power comes great responsibility. This is something that
touched you. And then you went into an all hands meeting and you
said: "guys from now on, let's be fair". You know? Like "let's bake
fairness into our day-to-day work" right?
I don't imagine that was how it happened, but first of all, do you
have just a nucleus of people that are focused on that? And then the
rest of the company is basically " unaware", let's say? Or is it
something that comes often on a day to day conversation across
teams?
And also you have roles besides pure engineers? Do you also have
other backgrounds that can also influence how you think about the
models and how you think about controlling for bias and fairness?
It's a big question.
P It's like two or three
questions, but let me go. I think the first one, which is how did we
start working on these operationally in terms, if it was like an all
hands meeting and then I started pushing.
So I approached it as I approach almost all new things, which is
first I was going to learn about the subjects. So I spent to be
honest, a few years, initially just reading, reading, understanding,
understanding the state of the art within reading books, within
papers, getting to know the research of the top institutions from
academia side, from industry side, what they were doing, what they
were not doing, what were the opportunities and what were the risks,
what type of explanations were there and so on.
And then I was able to identify that "okay, this makes sense". So in
my mind, first I was trying to understand: is the problem complex?
Is the problem going to appear on our domain? Is the problem
relevant for our clients and for our data scientists?
I concluded that yes, that it was unavoidable. I saw it affecting
other use cases, other companies in health at first and in law also.
Very famous examples for Amazon and Goldman Sachs and other
companies with explainability issues and say: "okay, we need to
address it". And if we're going to do it, let's do it right.
So I assembled a team to just work on this and they are called
internally F.A.T.E., which stands for Fairness, Accountability,
Transparency, and Ethics. We created a team from scratch to just
focus on this. And initially I also spent a couple of years so as
you can see, this is a multi-year effort, right? We started working
on this late 2016. They spent a couple of years investing in tools
and algorithms and trying to improve things because we realized that
in our domain, if you have a good idea, but if you cannot put it in
production it doesn't matter.
For me, the question was not only finding good techniques to avoid
bias or good techniques to detect bias or produce explanations.
It's: can we actually put these in the day-to-day pipeline of data
scientists? Can you put this in production? And that's really the
challenge. Are these things easy enough to use, fast enough to use
that a data scientist under pressure, developing a model for a
client with lots of deadlines, lots of constraints: is she going to
use the tools that allow to develop fair models? Because if the
tools are not easy and good to use, they are not going to use them.
So we really spent a lot of time combining the research side of bias
detection and explainability and so on with the engineering side of
how to put these in the day-to-day usage of a data scientist.
So that was the first part of your question. The second part is if
this is a single team working on this or more people are working on
this. So initially it was a single team, a team of about five
people, but now their impact grew to basically the entire company.
The tools started to be used by other people, we have many training
sessions internally with people from the engineering side, product
side, customer success side, marketing and sales. We have blog
posts, we have different materials, websites. As everything that
makes sense, it starts small and then it grows to impact many teams
and actually the company at the global scale. And it has been
presented to our end clients, to analysts, external analysts that
are analyzing companies like ours and competitors. Now it's a global
thing in the company.
L That's great. I
would like to be like a spy, just like looking at people because in
a way you need to change a little bit the paradigm I guess of an
engineer that is enticed by a complex problem and you're adding a
somehow abstract layer of complexity on top of something that is
already hard, right?
In a way you are approaching him and saying: "besides these things,
you also need to account for this set of other things", which are
still relatively fuzzy. We're still trying to figure out what those
things are concretely so that we can push them into production, as
you were saying.
And I'm trying to think what were some sort of reactions or the
feedback, how was the adoption of it you know?
P I think it was
fantastic to be honest. I think because we also have a very strong
engineering culture, even the data science team, the research team.
We are also users of our own technology so we know that whatever we
need to develop must be easy and must be feasible for the data
scientists and users.
And I think at the end of the day, our data scientists and any data
scientists, they really want to do the right thing. So we didn't
have a single case of a person saying: "oh, I don't want to do
that". Not a single person.
Everybody was excited. Everybody was: "okay I'm so glad you guys are
doing that. This is a piece of mind. It takes some weight off of my
shoulders. It allows me to feel good with myself and to feel good
with my work". So everybody was seeing these concerns with
responsible AI and bias, the famous movie now
Coded Bias, right?
Everybody was seeing these things happening out there. Even our
clients were seeing that, our CEO was seeing those concerns. So when
we developed the ability for people to do that, there was actually a
sense of relief that: "okay, we are doing it well, we are doing the
right thing". And the other part of it was also I think, positive
feedback, because we felt that we are ahead of the state of the art.
So that is also good, right? You're doing the right thing and you
are ahead of the market. So those two things. made it that the
internal feedback and external feedback was also very, very positive
because it's like a win-win: you're ahead of the market and you're
doing the right thing and at the end of the day, everybody benefits.
The company benefits, the internal developers benefit and external
clients also benefit. So it's a win-win-win situation. Very positive
feedback.
L And I'm guessing, just to add on what you said, that in terms of values, in terms of the impact of the company, I guess that any person is happier to know that they're just not building just the product. Period. Right? They are doing it in a sort of mindful way in terms of how it will affect the client and the market and the users of it.
P And we were also lucky
to win a number of awards worldwide with our responsible AI work.
Our algorithm
Fairband
got four or five different awards, like a FinTech breakthrough
award, we were featured on Fast Company on the software side and as
an honorable mention on the AI side. So we felt that it was not
something that we were seeing only us like the market was
recognizing these developments, these algorithms, these ideas as
valuable and giving us awards for that.
So everybody felt really like it was a win for the entire company to
have started investing in this so long ago, because now it was
giving dividends.
L Yeah, for sure.
Like concrete not only with clients, but the market recognition that
it's a tendency that is valued.
One of the things that I also asked you is the composition of the
teams, right? I have discussed with people (see
CFT Issue #3) that argue that if you only have engineers then you may not reach
the best solution, let's say, because you're missing some other
points of views or ways of looking at things.
Do you guys only have engineers or do you have other consultants or
other people that help you look at things differently regarding
ethics and responsible AI?
P So on the engineering
side, we actually have a very varied set of people. We have people
from physics background, math, statistics, but also biomedical
engineers, computer scientists, so all sorts of backgrounds which is
interesting.
But we also work a lot with our internal team of lawyers and legal
people that have a concern on how to describe this, what is the
impact in terms of law and regulation and also with the marketing
team and the sales team. So it's not just the research team that is
involved. Many people outside the research team are involved. Even
in the research team we have the typical data scientists, but we
also have data engineers and visualization engineers.
So it's not only data scientists that work in this subject.
L Awesome, that's
interesting. So one thing, which is, how easy has it been to
assemble your team over the years? I know you guys are already a
considerable amount of data scientists. For our market in Portugal I
think it's pretty considerable.
So how easy or how hard has it been in terms of finding those
individuals?
P It's hard. It's hard.
So I think hiring has probably been the top one challenge. Almost
always since the beginning. As I say, we are always hiring, even
when we are not hiring, because it takes so long to find good people
in terms of their background and experience and mindset and culture.
And we are always looking for people. We are always doing
interviews. We are always trying to select, always trying to grow
the team. And the market is not gigantic. We feel that there are not
enough people with all the background that we would like to have or
with all the experience that we would like to have.
It's surely a challenge. I think it's not a challenge just for us, I
think it's a challenge across the world. I realize that the
universities are now offering some degrees that they were not
offering a few years ago. Master's in data science. Almost all major
universities are offering that, which was not the case five years
ago. We feel that there are more people in the job market and even
people that were not data scientists they kind of learned new
techniques and became data scientists as well. But it's still a
challenge, yes. Probably one of the biggest challenges is hiring.
L Yeah. Well, we are a small country compared to other countries that produce a lot of engineers, Ukraine, for instance.
P At Feedzai we have, I read some statistics, the other day, we have 47- 48 different nationalities in the company. So we hire from lots of different countries. Of course being based in Portugal the majority is still Portuguese, the ones in Portugal, but there are many people from the Eastern European side, from Brazil, from Asian countries. So it's not just Portuguese. But it's true. It's hard to hire in the market, yes.
L I'm using this
question to segue a little bit into some of the things that we spoke
the first time, which is about would it be possible for us in terms
of a country to find this sort of niche? All right, let's be very
good at producing and generating people that have this mindset, the
mindset of building responsible AI, doing great machine learning
models that account for fairness, account for societal issues.
Would that be something that you believe is possible for us in our
reality, or are you guys an outlier?
P I totally believe it's
possible. I totally believe that it is possible for a small country
like Portugal to be a leading voice in a specific area.
For example you know that New Zealand is a world leader in rugby,
right? Or Iceland in music or Israel in cybersecurity, right? Even
small countries, they can be a reference in whatever area they
choose if they really invest. For instance we are also very good at
soccer for the size that we are, we are strangely good at soccer.
But why is that? It's because we have been investing for literally
decades in youth schools for soccer teams and coaches.
And so the investment is really nationwide and across generations.
And I think if we keep on investing in a specific area we can be
good in any area that we decide to be good at. I remember when we
were first raising money in Portugal twelve years ago - and I also
told you this, when we talked the other day - we went to a number of
U.S. investors, and I remember at the time they were asking us: "you
are a Portuguese this company? What big Portuguese exits have been
there? What is the history of the country in terms of startups and
tech?".
For them it was strange. Who are these guys coming from this country
that is famous for Fado and football and "pastel de nata" and now
trying to sell a high-tech company?
But they were sort of right. I remember at the time the biggest
exits that we had in Portugal were two or three companies that
altogether they sold for like $200m.
But now if you fast forward 12 years, right? We have Farfetch being
valued at $18b or so, OutSystems $10b, Talkdesk $3b or $4b. Not even
talking about Feedzai, there are dozens of companies here today,
DefinedCrowd and Unbabel and so many companies that are valued
already at hundreds of millions of dollars and they can potentially
easily reach billions of dollars. And what happened? What happened
was time essentially, right? 12 years since then, but also a strong
investment in multiple areas: startups but also the startup
ecosystem, the VC money, the founders.
Of course there's plenty to be done still, but the picture here,
today in 2021 in Portugal, is radically different from 2010. Is
radically different. Now we have big exits that we can show in our
sleeves to the world, right? We have good examples, we have good
engineering. You have companies that are worldwide leaders in their
specific domain. OutSystems is a worldwide leader in their domain.
TalkDesk as well. Feedzai is also recognized as a leader in fraud
detection.
So I think it's completely possible, but we need to continue on
investing in multiple areas and education on the university side.
Building the ecosystems, all the startup places of incubation that
we are having across the country that didn't exist 10 years ago. All
of that creates this ecosystem, just like the youth schools of
football, right? We need to have lots of little startup incubators.
And some will die, most will die, but many people will learn and
will create a culture of entrepreneurship and investment and
risk-taking and changing the world for the better.
If we keep on investing, why not Portugal being a leader in
responsible AI? I totally think it's possible. And right now I think
we are in a good position worldwide. The field is hot, it's just
starting. There's a lot of research, but there is not yet lots of
applied products, really being responsible in the way that they were
designed from scratch like that. So I think that the opportunity is
here for the taking and I think we should invest in it.
L I appreciate your
answer. It's true that when you look back, it's been an incredible
leap in so many ways in what entrepreneurship is risk-taking, as you
were saying.
I believe Portuguese are a bit risk averse people by nature, at
least that's the idea that I have, but things have been changing.
There's one thing that you mentioned, which is the ecosystem, a lot
of bright engineers, a lot of bright people, scientists.. The market
wasn't able to absorb them, right? You didn't have Feedzais or
Unbabels or bigger companies that would justify the existence of
those people in our territory. And so they just go to the U.S.,
France, Berlin, whatever, and you lose those assets that can create
new companies and create those new categories.
So it's a whole as you said. To finalize, investing in education,
but what else? Like what is some stuff that we should be paying
attention to as a country or is the ecosystem good enough? Should we
have more mechanisms? Should we have other ways of approaching it in
your opinion?
P Yes. But before I answer that, let me just touch briefly. You mentioned that there's some brain drain that we are losing people too.
L Well, we were more now I would not be as sure of that.
P And that I think we
still lose some but it can actually be beneficial, right? Sometimes
if people go outside and learn and go to top companies and learn and
become better scientists, better engineers, even if they don't come
back, they build a good reputation or they start a company in the
U.S. for example and they have a local office in Portugal.
I've seen that happen. For example, many people in India went to the
U.S. and then one generation later, many of them returned to India
and started Microsoft research labs and Oracle and IBM research labs
in India.
If you look at that with a perspective of 10-20 years, maybe that's
not too bad. Right? We have amazing professors that graduated from
Portugal and went and became top professors worldwide. We have the
same with engineers, for example,
Diogo Monica
was an engineer in Portugal, went to Square now and he's starting
his own company Anchorage, which is doing great. And I think it's
possible for people to go outside and come back one way or another,
maybe in a few years, maybe in 10 years, maybe as a cyber
organization. But I think that the net impact can still be positive
in the longer run.
And now going back to the second part of your question, what can we
do in Portugal to improve things even more?
So one thing is to make Portugal more appealing to entrepreneurs and
to those Portuguese abroad to come, but also to other entrepreneurs
to come. Our law is still too complex. Our taxes are still too high,
many things that are common in startups like stock options and
things like that are not easily translatable in our law. They are
hard to execute.
So there's an amount of taxation and regulation that makes it hard
for a founder to start a company in Portugal or for a person to join
a startup in Portugal. And I think we can reduce that barrier of
entry if we make it more appealing from that perspective, from the
legal perspective, from the taxation perspective.
L Yeah. The legal system must be clearer for investors to navigate and they want some expectability and conflict resolution not to be hard.
P But that is in
addition to what you also said: continuing investment in innovation
and education and research. I think we should invest way more in our
universities, we should finance more our professors and students. We
should have more scholarships and more collaboration.
Education is really the future of a country and I think we should
put more money there.
L I'm a hundred percent behind you on that. And on that note I'm going to ask you, do you want to leave any message for our listeners or anything that we should be on the lookout for that Feedzai is working on or just anything that you want to throw out there?
P So we are doing a lot
of work on responsible AI, on algorithms that are able to not only
find good models, but find good models that are responsible, but we
also invest a lot in visualization, novel machine learning
techniques and in active learning.
So you should be on the lookout. Keep track of my posts on
Twitter
and
LinkedIn, I'm always posting about the development that we are doing, the
papers accepted, the papers on arXiv. We are likely going
to start a
YouTube channel
with our own presentations - research presentations.
So we are really trying to share our state-of-the-art to the world
to inspire others and to help others to also benefit from the
research that we're doing. So that's why we've been publishing a lot
on conferences, arXiv and now on this new YouTube channel
that is upcoming.
L I'll be on the
lookout. I will link your profiles in the transcript and everyone go
follow Pedro. I've been following him for a bit, he publishes
interesting content.
Pedro, thank you so much for being here. I really enjoyed this
conversation and listening to how Feedzai does things. It's super
exciting to see this happening in Portugal. And again, thank you for
sharing all of that with us.
P My pleasure, it was a lot of fun.
L I'm sure we'll be talking again. Take care.
P Take care, have fun. Bye bye, thanks.
L Bye.
If you've enjoyed this publication, consider subscribing to CFT's newsletter to get this content delivered to your inbox.