Nearly 3 out of 4 businesses already use artificial intelligence, and the AI market is projected to surge to $1.4 trillion by 2030. Amidst this new tech boom, assistant professor of sociology in the School of Arts & Sciences Benjamin Shestakofsky thinks itâs more important than ever to pause and explore the social systems that shape AI growth.
âEven the definition of AI itself is ambiguous, contested, and ever-changing,â says Shestakofsky, who studies how digital technologies affect employment, organizations, and the economy. âRegardless of what AI can actually do, peopleâs ideas about what AI is can influence their behavior.â
With questions swirling around how AI will shape society, Shestakofsky and Devika Narayan of the University of Bristol set out to analyze the social systems that affect artificial intelligence. In an article published in The Journal of Applied Behavioral Science, they outline four big sociological themes that help make sense of AI.
1. Follow the money
Finance is the not-so-invisible hand that shapes technology; in 2022, the U.S. tech industry attracted more than $130 billion in venture capital funds. With the potential to turn relatively small investments into millions or billions, venture capitalists encourage technology start-upsâincluding those that create AIâto take big risks and grow quickly in the name of profit, Shestakofsky says.
âThese funding structures are setting the agenda for technology development and the goals itâs aimed at achieving,â he notes, adding that investors often care more about the skyrocketing short-term valuations of AI startups than whether these companies will be sustainable in the long run.
Shestakofsky says he believes this environment can make technology development snowball, something apparent in the past few years with the rapid release and improvement of generative AI chatbots like ChatGPT.
2. Are we playing Monopoly?
Itâs no secret that the biggest tech giants in the country have consolidated their power to near-monopoly status, and much new research and development of artificial intelligence filters through these companies, Shestakofsky says. Whether itâs Microsoft partnering with OpenAI, the creator of ChatGPT, or smaller AI start-ups using Amazonâs servers, much of AI flows through these goliath companies.
But big tech companies arenât the end all, be all, of AI either, he adds. Shestakofsky says innovation in AI is often driven by new, smaller companies, many of which build apps designed to run on tech giantsâ popular platforms, like iOS and Android. In this way, he believes itâs important to understand how AI development is a push and pull between industry titans and scrappy start-ups, both of which rely on each other to prosper.
3. Founders versus followers
Whenever new technology is developed, entire industries adapt around it. Take the example of Netflix that Shestakofsky and Narayan include in the paper; when Netflix revolutionized media by introducing streaming, its rivals had to follow suit, and now nearly every media company from HBO to Disney has its own streaming platform.
While itâs still early days for AI proliferation, Shestakofsky says itâs likely AI development will push many industries to adopt new practices. He points to one already happening in India: Companies that previously specialized in maintaining clientsâ in-house IT systems must adapt as cloud computing and digital platforms become the new norm, leading to an âunbundlingâ of technology infrastructure.
4. Itâs not all glamorous
Thinking about the AI industry often conjures luxurious Silicon Valley offices, job perks like free meals and exercise classes, and sky-high salaries. But that image excludes the low-paid workers who make AI possible.
These workersâoften located in places like India and the Philippines, the researchers noteâare responsible for tasks including labeling data, testing and evaluating models, and screening out harmful content. âAI development doesnât happen without these armies of low-status workers who are out of view and out of mind for the consumers using these systems,â Shestakofsky says.
Many of these workers are now responsible for a task called red teaming, in which they prompt generative AI systems to create harmful or offensive content to identify and patch holes in their moderation systems. This means these workers may get exposed to disturbing, violent, or inappropriate imagery and content. âThis is another emerging frontier where we donât yet fully understand the consequences for workersâ health and well-being,â he says.
Thereâs still a lot to learn about AI, but the researchers say they hope these lenses can help demystify the new technology and situate it in the broader context of existing social, cultural, and economic systems.
âAs a sociologist, the name of the game is always to understand what youâre studying in relation to broader social structures,â Shestakofsky says. âSo, there may be a lot of incentives to âmove fast and break things,â but I think organizations can benefit from paying close attention to some of the second-order effects associated with introducing new technologies.â
More information:
Devika Narayan et al, Relationships That Matter: Four Perspectives on AI, Work, and Organizations, The Journal of Applied Behavioral Science (2024). DOI: 10.1177/00218863241285456
Citation:
Exploring the social systems shaping AI development (2025, January 22)
retrieved 22 January 2025
from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.