Startups

Building better startups with responsible AI

Comment

Digital generated image of split net/turbulence structure of artificial intelligence brain on purple surface.
Image Credits: Andriy Onufriyenko (opens in a new window) / Getty Images

Tom Zick

Contributor

Tom Zick is a researcher in AI ethics at the Berkman Klein Center for Internet and Society at Harvard University, where she is also a J.D. candidate. She holds a Ph.D. from UC Berkeley and was previously a fellow at Bloomberg Beta and the City of Boston.

Founders tend to think responsible AI practices are challenging to implement and may slow the progress of their business. They often jump to mature examples like Salesforce’s Office of Ethical and Humane Use and think that the only way to avoid creating a harmful product is building a big team. The truth is much simpler.

I set out to learn how founders were thinking about responsible AI practices on the ground by speaking with a handful of successful early-stage founders and found many of them were implementing responsible AI practices.

Only they didn’t call it that. They just call it “good business.”

It turns out, simple practices that make business sense and result in better products will go a long way toward reducing the risk of unforeseen societal harms. These practices rely on the insight that people, not data, are at the heart of deploying an AI solution successfully. If you account for the reality that humans are always in the loop, you can build a better business, more responsibly.

Think of AI as a bureaucracy. Like a bureaucracy, AI relies on having some general policy to follow (“the model”) that makes reasonable decisions in most cases. However, this general policy can never account for all possible scenarios a bureaucracy will need to handle — much like an AI model cannot be trained to anticipate every possible input.

When these general policies (or models) fail, those who are already marginalized are disproportionately impacted (a classic algorithmic example is of Somali immigrants being tagged for fraud because of their atypical community shopping habits).

Bureaucracies work to solve this problem with “street-level bureaucrats” like judges, DMV agents and even teachers, who can handle unique cases or decide not to enforce the policy. For example, teachers can waive a course prerequisite given extenuating circumstances, or judges can be more or less lenient in sentencing.

If any AI will inevitably fail, then — like with a bureaucracy — we must keep humans in the loop and design with them in mind. As one founder told me, “If I were a Martian coming to Earth for the first time, I would think: Humans are processing machines — I should use them.”

Whether the humans are operators augmenting the AI system by stepping in when it’s uncertain, or users choosing whether to reject, accept or manipulate a model outcome, these people determine how well any AI-based solution will work in the real world.

Here are five practical suggestions that founders of AI companies shared for keeping, and even harnessing, humans in the loop to build a more responsible AI that’s also good for business:

Introduce only as little AI as you need

Today, many companies plan to launch some services with an end-to-end AI-driven process. When those processes struggle to function under a wide range of use cases, the people who are most harmed tend to be those already marginalized.

In trying to diagnose failures, founders subtract one component at a time, still hoping to automate as much as possible. They should consider the opposite: introducing one AI component at a time.

Many processes are — even with all the wonders of AI — still just less expensive and more reliable to run with humans in the loop. If you build an end-to-end system with many components coming online at once, you may find it hard to identify which are best suited to AI.

Many founders we spoke with view AI as a way to delegate the most time-consuming, low-stakes tasks in their system away from humans, and they started with all human-run systems to identify what these important-to-automate tasks were.

This “AI second” approach also enables founders to enter fields where data is not immediately available. The people who operate parts of a system also create the very data you’ll need to automate those tasks. One founder told us that, without the advice to introduce AI gradually, and only when it was demonstrably more accurate than an operator, they would have never gotten off the ground.

Create some friction

Many founders believe that to be successful, a product must run out of the box, with as little user input as possible.

Because AI is typically used to automate part of an existing workflow — complete with associated preconceptions on how much to trust that workflow output — a perfectly seamless approach can be catastrophic.

For example, when an ACLU audit showed that Amazon’s facial recognition tool would misidentify 28 members of Congress (a disproportionately large fraction of whom were Black) as criminals, lax default settings were at the heart of the problem. The accuracy threshold out of the box was set to only 80%, clearly the wrong setting if a user takes a positive result at face value.

Motivating users to engage with a product’s strengths and weaknesses before deploying it can offset the potential for harmful assumption mismatches. It can also make customers happier with eventual product performance.

One founder we spoke with found that customers ultimately used their product more effectively if the customer had to customize it before use. He views this as a dominant component of a “design-first” approach and found it helped users play to the strengths of the product on a context-specific basis. While this approach required more upfront time to get going, it ended up translating into revenue gains for customers.

Give context, not answers

Many AI-based solutions focus on providing an output recommendation. Once these recommendations are made, they have to be acted on by humans.

Without context, poor recommendations could be blindly followed, causing downstream harm. Similarly, great recommendations could be rejected if the humans in the loop do not trust the system and lack context.

Rather than delegating decisions away from users, consider giving them the tools to make decisions. This approach harnesses the power of humans in the loop to identify problematic model outputs while securing the user buy-in necessary for a successful product.

One founder shared that when their AI made direct recommendations, users didn’t trust it. Their customers were happy with the accuracy that their model predictions turned out to have, but individual users just ignored the recommendations. Then they nixed the recommendation feature and instead used their model to augment the resources that could inform a user’s decision (e.g., this procedure is like these five past procedures and here is what worked). This led to increased adoption rates and revenue.

Consider your not-users and not-buyers

It is a known problem in enterprise tech that products can easily serve the CEO and not the end users. This is even more problematic in the AI space, where a solution is often part of a greater system that interfaces with a few direct users and many more indirect ones.

Take, for example, the controversy that arose when Starbucks began using automated scheduling software to assign shifts. The scheduler optimized for efficiency, completely disregarding working conditions. After a successful labor petition and a high-profile New York Times article, the baristas’ input was taken under consideration, improving morale and productivity.

Instead of taking a customer literally on what they ask you to solve, consider mapping out all of the stakeholders involved and understanding their needs before you decide what your AI will help optimize. That way, you will avoid inadvertently making a product that is needlessly harmful and possibly find an even better business opportunity.

One founder we spoke with took this approach to heart, camping out next to their users to understand their needs before deciding what to optimize their product for. They followed this up by meeting with both customers and union representatives to figure out how to make a product that worked for both.

While customers originally wanted a product that would allow each user to take on a greater workload, these conversations revealed an opportunity to unlock savings for their customers by optimizing the existing workload.

This insight allowed the founder to develop a product that empowered the humans in the loop and saved management more money than the solution they thought they wanted would have.

Be clear on what’s AI theater

If you limit the degree to which you hype up what your AI can do, you can both avoid irresponsible consequences and sell your product more effectively.

Yes, the hype around AI helps sell products. However, knowing how to keep those buzzwords from getting in the way of precision is crucial. While talking up the autonomous capabilities of your product might be good for sales, it can backfire if you apply that rhetoric indiscriminately.

For example, one of the founders we spoke to found that playing up the power of their AI also increased their customers’ privacy concerns. This concern persisted even when the founders explained that the portions of the product in question did not rely on data, but rather on human judgment.

Language choice can help align expectations and build trust in a product. Rather than using the language of autonomy with their users, some of the founders we talked to found that words like “augment” and “assist” were more likely to inspire adoption. This “AI as a tool” framing was also less likely to engender the blind trust that can lead to bad outcomes down the line. Being clear can both dissuade overconfidence in AI and help you sell.

These are some practical lessons learned by real founders for mitigating the risk of unforeseen harms from AI and creating more successful products built for the long term. We also believe there’s an opportunity for new startups to build services that help make it easier to create ethical AI that’s also good for business. So here are a couple of requests for startups:

  • Engage humans in the loop: We need startups that solve the “human in the loop” attention problem. Delegating to humans requires making sure those humans notice when an AI is uncertain so that they can meaningfully intervene. If an AI is correct 95% of the time, research shows that people get complacent and are unlikely to catch the 5% of instances the AI gets wrong. The solution requires more than just technology; much like social media was more of a psychological innovation than a technical one, we think startups in this space can (and should) emerge from social insights.
  • Standard compliance for responsible AI: There’s opportunity for startups that consolidate existing standards around responsible AI and measure compliance. Publication of AI standards has been on the rise in the past two years as public pressure on AI regulation has been increasing. A recent survey showed 84% of Americans think AI should be carefully managed and rate this as a top priority. Companies want to signal they are taking this seriously and showing they are following standards put forth by IEEE, CSET and others would be useful. Meanwhile, the current draft of the EU’s expansive AI Act (AIA) strongly emphasizes industry standards. If the AIA passes, compliance will become a necessity. Given the market that formed around GDPR compliance, we think this is a space to watch.

Whether you’re trying one of these tips or starting one of these companies, simple, responsible AI practices can let you unlock immense business opportunities. To avoid creating a harmful product, you need to be thoughtful in your deployment of AI.

Luckily, this thoughtfulness will pay dividends when it comes to the long-term success of your business.

More TechCrunch

Welcome to Startups Weekly — Haje‘s weekly recap of everything you can’t miss from the world of startups. Sign up here to get it in your inbox every Friday. Well,…

Startups Weekly: Drama at Techstars. Drama in AI. Drama everywhere.

Last year’s investor dreams of a strong 2024 IPO pipeline have faded, if not fully disappeared, as we approach the halfway point of the year. 2024 delivered four venture-backed tech…

From Plaid to Figma, here are the startups that are likely — or definitely — not having IPOs this year

Federal safety regulators have discovered nine more incidents that raise questions about the safety of Waymo’s self-driving vehicles operating in Phoenix and San Francisco.  The National Highway Traffic Safety Administration…

Feds add nine more incidents to Waymo robotaxi investigation

Terra One’s pitch deck has a few wins, but also a few misses. Here’s how to fix that.

Pitch Deck Teardown: Terra One’s $7.5M Seed deck

Chinasa T. Okolo researches AI policy and governance in the Global South.

Women in AI: Chinasa T. Okolo researches AI’s impact on the Global South

TechCrunch Disrupt takes place on October 28–30 in San Francisco. While the event is a few months away, the deadline to secure your early-bird tickets and save up to $800…

Disrupt 2024 early-bird tickets fly away next Friday

Another week, and another round of crazy cash injections and valuations emerged from the AI realm. DeepL, an AI language translation startup, raised $300 million on a $2 billion valuation;…

Big tech companies are plowing money into AI startups, which could help them dodge antitrust concerns

If raised, this new fund, the firm’s third, would be its largest to date.

Harlem Capital is raising a $150 million fund

About half a million patients have been notified so far, but the number of affected individuals is likely far higher.

US pharma giant Cencora says Americans’ health information stolen in data breach

Attention, tech enthusiasts and startup supporters! The final countdown is here: Today is the last day to cast your vote for the TechCrunch Disrupt 2024 Audience Choice program. Voting closes…

Last day to vote for TC Disrupt 2024 Audience Choice program

Featured Article

Signal’s Meredith Whittaker on the Telegram security clash and the ‘edge lords’ at OpenAI 

Among other things, Whittaker is concerned about the concentration of power in the five main social media platforms.

18 hours ago
Signal’s Meredith Whittaker on the Telegram security clash and the ‘edge lords’ at OpenAI 

Lucid Motors is laying off about 400 employees, or roughly 6% of its workforce, as part of a restructuring ahead of the launch of its first electric SUV later this…

Lucid Motors slashes 400 jobs ahead of crucial SUV launch

Google is investing nearly $350 million in Flipkart, becoming the latest high-profile name to back the Walmart-owned Indian e-commerce startup. The Android-maker will also provide Flipkart with cloud offerings as…

Google invests $350 million in Indian e-commerce giant Flipkart

A Jio Financial unit plans to purchase customer premises equipment and telecom gear worth $4.32 billion from Reliance Retail.

Jio Financial unit to buy $4.32B of telecom gear from Reliance Retail

Foursquare, the location-focused outfit that in 2020 merged with Factual, another location-focused outfit, is joining the parade of companies to make cuts to one of its biggest cost centers –…

Foursquare just laid off 105 employees

“Running with scissors is a cardio exercise that can increase your heart rate and require concentration and focus,” says Google’s new AI search feature. “Some say it can also improve…

Using memes, social media users have become red teams for half-baked AI features

The European Space Agency selected two companies on Wednesday to advance designs of a cargo spacecraft that could establish the continent’s first sovereign access to space.  The two awardees, major…

ESA prepares for the post-ISS era, selects The Exploration Company, Thales Alenia to develop cargo spacecraft

Expressable is a platform that offers one-on-one virtual sessions with speech language pathologists.

Expressable brings speech therapy into the home

The French Secretary of State for the Digital Economy as of this year, Marina Ferrari, revealed this year’s laureates during VivaTech week in Paris. According to its promoters, this fifth…

The biggest French startups in 2024 according to the French government

Spotify is notifying customers who purchased its Car Thing product that the devices will stop working after December 9, 2024. The company discontinued the device back in July 2022, but…

Spotify to shut off Car Thing for good, leading users to demand refunds

Elon Musk’s X is preparing to make “likes” private on the social network, in a change that could potentially confuse users over the difference between something they’ve favorited and something…

X should bring back stars, not hide ‘likes’

The FCC has proposed a $6 million fine for the scammer who used voice-cloning tech to impersonate President Biden in a series of illegal robocalls during a New Hampshire primary…

$6M fine for robocaller who used AI to clone Biden’s voice

Welcome back to TechCrunch Mobility — your central hub for news and insights on the future of transportation. Sign up here for free — just click TechCrunch Mobility! Is it…

Tesla lobbies for Elon and Kia taps into the GenAI hype

Crowdaa is an app that allows non-developers to easily create and release apps on the mobile store. 

App developer Crowdaa raises €1.2M and plans a US expansion

Back in 2019, Canva, the wildly successful design tool, introduced what the company was calling an enterprise product, but in reality it was more geared toward teams than fulfilling true…

Canva launches a proper enterprise product — and they mean it this time

TechCrunch Disrupt 2024 isn’t just an event for innovation; it’s a platform where your voice matters. With the Disrupt 2024 Audience Choice Program, you have the power to shape the…

2 days left to vote for Disrupt Audience Choice

The United States Department of Justice and 30 state attorneys general filed a lawsuit against Live Nation Entertainment, the parent company of Ticketmaster, for alleged monopolistic practices. Live Nation and…

Ticketmaster antitrust lawsuit could give new hope to ticketing startups

The U.K. will shortly get its own rulebook for Big Tech, after peers in the House of Lords agreed Thursday afternoon to pass the Digital Markets, Competition and Consumer bill…

‘Pro-competition’ rules for Big Tech make it through UK’s pre-election wash-up

Spotify’s addition of its AI DJ feature, which introduces personalized song selections to users, was the company’s first step into an AI future. Now, Spotify is developing an alternative version…

Spotify experiments with an AI DJ that speaks Spanish

Call Arc can help answer immediate and small questions, according to the company. 

Arc Search’s new Call Arc feature lets you ask questions by ‘making a phone call’