Skip to content
Who's in the Video
Ramesh Srinivasan is Professor of Information Studies and Design Media Arts at UCLA. He makes regular appearances on NPR, The Young Turks, MSNBC, and Public Radio International, and his writings[…]

RAMESH SRINIVASAN: Technology can really amplify biases because we create technologies based on who we are. Just like when we write a poem or write a book. My new book, Beyond the Valley, it reflects my biases. I'm willing to admit that. And many of us like to think of ourselves as unbiased but that's part of being a human being, is being biased. That doesn't mean we're bad. That doesn't mean we're wrong, but all of us carry unconscious biases. So we encode our ways of seeing the world based on who we are into technology. That's part one and that's a major issue and that's an issue that every business leader should be aware of and conscious of.

But the second point is, increasingly, technology companies, including the big ones, describe their companies as AI companies; artificial intelligence companies. Why is that? Well, artificial intelligence is kind of real now in a way that it wasn't before, and I, as a former artificial intelligence developer, have seen this change. And the reason why is we built faster machines, we've been able to store exponentially more data all at lower costs, and those phones in our pocket which are tracking us 24/7, 365 in ways we have no knowledge about. What I'm getting at here is technologies built upon biases are learning from data sets that are out there and they're learning from an unequal world. Because our world, we still have to try to perfect our union. We have to think about artificial intelligence in aspirational ways rather than this myth that it's somehow neutral or scientific or it's "just technology". So that's the second issue that I would encourage business leaders to be aware of. What are the data sets your technologies are learning from and what are your own values that are influencing the development of your technologies? Those are one and two.

And the third is being transparent and understanding when you're using an algorithmic or AI system. It turns out that, as Americans, and actually people across the world, we are always interacting with AI systems and we don't even know it. For example, there's a lot of discussion around 5G networks which is actually configured for things to communicate with one another. Like, imagine your sidewalk communicating with your shoes. I mean, who knows why we want that but that is basically the smart city concept, is a layered infrastructure for 5G. But what are the languages by which those things are communicating? What are the algorithms that are determining what those forms of communication are? So, basically, all the time we're interacting with AI systems it's not disclosed to us; we don't know what those systems know about us, we don't know what are the values that guide their decisions, we don't know how that might shape our lives, we don't know what alternatives we might provide. All of that is a black box and all of that should be opened up. So a lot of business leaders I've actually spoken to are kind of in the diversity and inclusion space, but they're also trying to think about alternatives at this moment where the American public—bipartisan, it's incredible—across every demographic supports doing something about big technology issues.

So here's an opportunity for an alternative business model and we've seen some examples throughout history, recent memory, actually, of technology platforms that try to kind of say 'Hey, we're going to come out but we're going to be different than Facebook.' Ello was one example of that. Everybody flocked to it but it just couldn't sustain itself. So there's interest in an alternative trajectory toward technology. That's why I'm hopeful around these issues.