Secret Assets Owners
  • Investing
  • World News
  • Politics
  • Stock
  • Editor’s Pick
Editor's PickInvesting

Demystifying AI Experimentation and the Startup Ecosystem

by May 1, 2024
May 1, 2024

Brian Chau

In general, the policy discussion around Artificial Intelligence (AI) underestimates the diversity and vulnerability of the AI startup ecosystem. The reality is often different than what policymakers think. Without understanding the startup ecosystem, the resulting policy could lead to unnecessary interventions into competitive markets that disrupt innovation in the free market.

So what might policymakers need to better understand about the current state of the AI startup ecosystem?

Policy analysis that focuses on the effects on industry leaders rather than on startups underestimates the damage done to AI development by harsher regulatory approaches, such as the Biden Executive Order on Artificial Intelligence. For example, the National Telecommunications and Information Administration (NTIA) is set to impose auditing and content regulation on “all stages of the value chain.” This means that instead of focusing narrowly on the areas where harms actually occur, the NTIA treats all AI developers as preemptively suspect.

Consequently, developers whose products have no risk of causing the harm the NTIA identifies are still forced to censor the output of their software, conduct time‐​consuming audits, and be delayed by inefficient regulatory approval processes.

This regulation would make the experimentation process far more time‐​consuming by increasing compliance costs, restricting what instructions AI models will respond to, and possibly harming the underlying efficacy of AI models. This results in more startups failing because they cannot acquire the capital to continue operating, there are longer timelines to complete research and development for the ones that persist, and there is a less effective end product. The NTIA’s proposed regulations would also have an anticompetitive effect, as they impose fixed costs that disproportionately affect smaller businesses with less time and capital to comply with them.

AI is often equated with its industry leaders, such as OpenAI, Anthropic, Google or Meta. However, just as assembling a car requires many suppliers and subcontractors, user‐​end AI products often consist of many specialized layers. Examples include producing the hardware required to run AI models, curating specialized data used as an input to AI models, computer intensive pre‐​training to produce a general language model such as GPT‑4, or fine‐​tuning a general model to improve performance on a specific task. The following chart is a map of these categories, which captures the complicated and competitive landscapes of each AI sub‐​industry.

Click here to see a larger image of the chart. 

Competition is increasing on each of these layers as AI becomes a more well‐​funded industry. As basic tools for every part of the AI pipeline become more widely available, startups can conduct thorough experiments on a narrower niche of the AI market.

Many startups will use a base model, such as OpenAI’s GPT‑4 or Meta’s Llama 3, as an input to their product. They experiment with a variety of base models, infrastructure, hardware, and optimization techniques to create the best product for their use case, leading to a complicated network of de‐​facto suppliers and subcontractors. Just as software companies now rely on other software for web development, databases, networking, or cloud services, AI companies will rely on increasingly specialized vendors for different components of AI models.

This process of market specialization is best demonstrated by the practical experience of startups. Proemial AI is a company that narrows down GPT‑4, sacrificing generality to make it the best at doing one specific thing: communicating scientific papers. “We were just surprised at how bad [unmodified] GPT4 was, because we were constraining it in interesting ways,” said founder Geet Khosla. Scientists often struggle to anticipate what questions an ordinary user would have about their work. By being able to respond to a user’s questions and vary its depth of explanation based on the experience of each individual user, Proemial would be able to make scientific work more accessible.

While most AI models require expensive hardware to run, Hal​tia​.AI is creating one that doesn’t. “We are taking the Mistral‐​7B model and putting a [version] on the iPhone,” said Hal​tia​.AI founder Arto Bendiken. This allows users to get an AI response much faster, at a lower cost, and even when they aren’t connected to the internet. This highlights the tradeoff between accuracy and convenience, which presents another opportunity for startups to compete.

Lastly, consider Superflows, a company that assembles a chain of Large Language Models (LLMs), ChatGPT, to explain a company’s software to a new programmer. “There’s a relatively long chain of prompts we use … get one LLM to describe endpoints … another LLM reads it,” said Superflows co‐​founder Matthew Phillips. His company experiments with AI models, their own modified models, and ways of formatting user input. “There’s a process of evaluating which LLM has the best ratio of quality to speed.”

These extensive experiments on base models and techniques carried out by startups require flexibility to operate. The NTIA’s proposed regulations impede them in several ways that are particularly impactful on smaller businesses. The most obvious issue is compliance costs for auditing or content restrictions, which each company would have to adapt their end product to. This serves as a barrier to entry for new companies, which may not have the resources to comply.

The result of content restrictions on their base‐​model suppliers would also make experimentation more difficult. Restrictions on model content can result in denying user requests that are not actually harmful. Overcoming these restrictions in non‐​harmful use cases sacrifices crucial engineering hours. Finally, there is some evidence that the techniques used to achieve content restrictions can reduce the overall performance or flexibility of AI models.

Brian Chau is the executive director of Alliance for the Future. He is a former machine learning engineer and International Olympiad in Informatics gold medalist.

previous post
Biden plan to use Leahy Law sanction on IDF was ‘political,’ expert alleges: ‘Congress should consider’ probe
next post
Senate Report Highlighting Bias in Online Services Shows How the Market Can Serve More Perspectives and Advance Expression

You may also like

Porch Pirates and Jersey Values: Why Washington Should...

December 26, 2025

How Fiscal and Economic Crises Prompted Retirement Reforms...

December 26, 2025

The Declaration, the Constitution, and America’s 250th

December 24, 2025

DHS Doesn’t List CECOT Prison Deportees in Its...

December 23, 2025

Heritage Doesn’t Make Somebody an American

December 23, 2025

Schemel v. Marco Island Brief: Urging Limits on...

December 23, 2025

Singleton v. Hamm Brief: Federal Courthouses Should Hear...

December 23, 2025

SNAP Has an Eligibility Loophole. Congress Needs to...

December 23, 2025

The Trump Executive Order Is a Good Step...

December 22, 2025

Fiscal Policy Is Raising Costs for American Families

December 22, 2025
Join The Exclusive Subscription Today And Get Premium Articles For Free


Your information is secure and your privacy is protected. By opting in you agree to receive emails from us. Remember that you can opt-out any time, we hate spam too!

Recent Posts

  • Africa’s Christian Crisis: How 2025’s deadly attacks finally drew global attention after Trump’s intervention

    December 28, 2025
  • Iranian president says his country is at ‘total war’ with the US, Israel and Europe: reports

    December 28, 2025
  • 2025 shockers: The biggest moments that rocked the campaign trail

    December 27, 2025
  • Most shocking examples of Chinese espionage uncovered by the US this year: ‘Just the tip of the iceberg’

    December 27, 2025
  • Zelenskyy says Ukraine, ahead of Trump meeting, is ‘willing to do whatever it takes’ to end war with Russia

    December 27, 2025
  • About us
  • Contact us
  • Terms & Conditions
  • Privacy Policy

Copyright © 2025 SecretAssetsOwners.com All Rights Reserved.


Back To Top
Secret Assets Owners
  • Investing
  • World News
  • Politics
  • Stock
  • Editor’s Pick