Understanding the Structure of the AI Industry
Infrastructure, Integrated Systems, and the Limits of Meaning
Abstract
This paper revisits and extends a three layer model of the artificial intelligence industry in response to recent technological and industrial changes.
The original model separates AI systems into training, design and optimization, and application and operations. This framework has been useful for explaining the rapid growth of large language models. In particular, it highlights the economic gap between learning and inference, which leads to clear divisions of labor within the industry.
The revised model addresses two issues that have become increasingly important.
First, it makes clear that computational infrastructure forms a foundational layer. Semiconductor production, accelerator design, energy supply, and execution environments do not simply support AI systems. They set the basic conditions under which training, optimization, and deployment are possible. Firms at this level shape the industry not by building models or applications, but by defining the limits within which all higher layers must operate.
Second, the framework extends beyond generative AI by including integrated systems such as robotics, medical AI, and board game AI. In these fields, the three layers are present but cannot be separated at the organizational level. Training, optimization, and application occur within a single continuous system. This structure follows from real time demands and close interaction with the physical world. It should be understood as a structural requirement rather than a sign of early development.
Finally, this paper considers the meaning of these findings for debates on symbol grounding and AI autonomy. It argues that embodied AI systems develop domain specific forms of pseudo grounding through repeated training and contextual learning. They do not acquire intrinsic meaning or independent intention. AI systems may produce insights that go beyond individual human ability, but they do not generate meaning in a lived or autonomous sense.
By placing AI capabilities within layered industrial structures and material constraints, the revised three layer model offers a clear framework for economists, policymakers, and practitioners. It helps explain both the power and the limits of contemporary artificial intelligence.
This essay lays out a clear framework for understanding how contemporary AI systems are structured, from foundational infrastructure to application. Below is a revised three-layer model that explains both economic roles and philosophical limits of AI.
1. Introduction
Generative AI has moved quickly from an experimental tool to a central part of modern economic and social systems. Language models now write software, translate between languages, create images, and support decision making in government, education, and business. Despite this visible growth, the basic structure of the AI industry remains poorly understood.
Public discussion often relies on simple images such as linear value chains, cloud services, or broad ideas of AI platforms. These explanations can be helpful at first glance. However, they tend to hide more important questions. Where is capital concentrated. How is responsibility divided. Which actors hold real power over system design.
In earlier work, I proposed a three layer model of the generative AI industry to address these questions. The model distinguishes between training, design and optimization, and application and operations. These layers are not meant to describe specific organizations. They function as an ideal type that helps clarify structural roles. This approach proved especially useful for understanding the fast industrial expansion of large language models.
More recent developments now require further refinement. Advances in computing infrastructure, stronger hardware limits, and the different paths taken by robotics and embodied AI all point to the same conclusion. The original model remains valid, but it is no longer sufficient by itself.
This paper revisits and extends the three layer model. It does not reject the earlier framework. Instead, it clarifies its assumptions, makes hidden dependencies visible, and broadens its scope to include a wider range of AI systems. In doing so, it also connects industrial structure to long standing questions about meaning, grounding, and the limits of AI autonomy.
2. Revisiting the Original Three Layer Model
The original three layer model was based on a simple observation.
Artificial intelligence systems, regardless of their application, pass through several functional stages. These stages differ clearly in cost, responsibility, and economic logic.
The first stage is the Training Layer.
This is where core capabilities are created. Large datasets are prepared and processed using large scale computing resources. Through this process, model parameters are produced. This stage requires heavy investment. Most of the value created at this level is stored in the trained parameters themselves.
The second stage is the Design and Optimization Layer.
Here, raw capability is shaped into usable form. This includes choices about model structure, fine tuning, evaluation methods, safety controls, and system constraints. Decisions made at this stage determine how a model behaves, under which conditions it is used, and which trade offs are accepted.
The third stage is the Application and Operations Layer.
At this stage, AI systems are placed within real social and economic settings. Governance, system integration, monitoring, and user trust become central concerns. This is also the stage where economic value becomes visible and where political attention tends to focus.
In the case of large language models, this three layer structure closely matches industrial practice. Training and inference differ greatly in scale, frequency, and cost. As a result, a clear separation emerges between organizations that build core models and those that deploy them. This division is not imposed by theory. It arises naturally from the economics of computation.
For this reason, the three layer model should be understood as a Weberian ideal type.
It is a conceptual tool designed to clarify structural tendencies. It does not claim that real systems must exactly follow these boundaries. Its strength lies in abstraction rather than precision.

3. The Infrastructural Foundation: Layer 0
One limitation of the original three layer model was its treatment of computing resources. These resources were mainly considered inputs to the Training Layer. At the time, this approach was sufficient. However, recent developments show that computation itself forms a separate structural foundation.
This foundation can be called Layer 0.
It includes semiconductor production, accelerator design, memory capacity, data transfer systems, power supply, cooling systems, and communication networks. These elements do not simply support AI systems. They define what kinds of systems can be built, which ones are affordable, and which ones can scale.
Actors operating at this level, especially firms such as NVIDIA, do not usually train models or run AI services. Instead, they provide the environment in which all higher layers operate. Their influence is therefore indirect. However, it is also decisive.
This distinction helps resolve a common misunderstanding.
NVIDIA is often described as an AI company. In practice, it does not train large models and does not operate AI applications. Its influence comes from control over computing standards. Tools such as CUDA, optimized libraries, and tightly integrated hardware and software systems determine how training and inference are performed. Infrastructure is therefore not neutral. It acts as a constraint that shapes industrial organization.
Open strategies at this level should be understood in this context.
The release of tools, reference models, or system frameworks does not mean that control over the core is abandoned. Instead, these actions expand the ecosystem built on top of the infrastructure. As a result, dependence on the foundation often increases. Openness at higher levels can strengthen control at the base.
By identifying this foundation as Layer 0, the revised model removes an important ambiguity. The three layer structure does not stand alone. It rests on a physical and economic base that is increasingly concentrated, capital intensive, and politically sensitive. Ignoring this base leads to misunderstandings about both the power and the limits of modern AI systems.
4. Integrated AI Systems and the Collapse of Organizational Layers
The three layer model is most clearly visible in generative AI and large language models. In these areas, training, design, and application often appear as separate industrial roles. However, this separation should not be treated as universal.
In several important fields, including robotics, medical AI, and board game AI, the three layers are implemented within a single system. These cases are sometimes described as exceptions or as signs of early development. Such interpretations are misleading.
In robotic systems, for example, sensing, learning, optimization, and action form a continuous loop. Environmental signals are captured by sensors, converted into numerical form, processed during training, and reflected directly in behavior. Feedback from action immediately affects learning. If these functions were separated across organizations, delays and failures would likely occur.
This integration does not mean that layers disappear.
Training still produces parameters. Design and optimization still impose goals and constraints. Application still appears as action in the world. What changes is the organizational form. The layers remain conceptually distinct, but they cannot be separated in practice.
This structure is not a temporary stage.
It is often required by real time conditions, safety demands, and the close link between perception and action. In such systems, dividing responsibility across independent platforms would reduce performance and reliability.
Recognizing this point allows the three layer model to be applied beyond generative AI. The model does not claim that layers must be separated. It claims that they must exist. Whether they appear as separate industries or as integrated systems depends on technical, economic, and institutional conditions.
This view also avoids a common mistake. Robotics and embodied AI are not fundamentally different kinds of intelligence. They are integrated versions of the same underlying processes.
5. Symbol Grounding, Embodiment, and Pseudo Grounding
The revised framework has important philosophical implications.
In particular, it clarifies the status of symbol grounding in artificial intelligence.
Robotics and embodied AI are often presented as counterexamples to the claim that AI lacks grounding. These systems interact directly with the physical world. They receive constant feedback and act in real time. From the outside, symbols appear connected to objects, actions seem meaningful, and learning appears experience based.
However, the three layer model suggests a different interpretation.
In robotic systems, what develops is not intrinsic meaning.
Instead, training enriches contextual representations. Sensor inputs are encoded as numbers. Large amounts of experience are accumulated. These experiences shape high dimensional parameter structures. Learning improves correlations within a limited sensor environment loop. It does not create semantic reference in the philosophical sense.
This process can be described as local or domain specific pseudo grounding.
The system becomes effective within the specific world defined by its sensors and tasks. However, this effectiveness does not generalize across contexts. It does not amount to understanding as experienced by human agents.
Even in integrated systems, the layers do not disappear.
Training produces parameters. Design sets objectives. Application results in action. The layers are merged at the organizational level, not removed at the conceptual level. Embodiment changes system structure, but it does not change the basic nature of artificial intelligence.
By distinguishing lived meaning from statistical adaptation, the framework avoids both technological mysticism and excessive skepticism. It recognizes the strength of embodied AI while maintaining a clear boundary between correlation and understanding.
6. AI, Intelligence, and the Question of Singularity
This distinction also applies to debates about AI autonomy and singularity.
AI systems already exceed individual human ability in narrow domains. They detect patterns at scale, generate unexpected solutions, and influence decision making in economic and institutional settings. In this sense, AI can support forms of knowledge beyond any single human mind.
However, the revised framework highlights a clear limit.
AI systems do not generate intention, desire, or existential concern. Their outputs remain shaped by training data, optimization rules, and application contexts. These structures are designed by humans and embedded in institutions.
What appears as autonomy is better understood as mediated computation.
Intelligence, in this framework, is not the same as meaning creation. AI systems can produce useful or surprising results. However, these results are not linked to lived motivation or moral orientation.
From this perspective, claims that computational power leads directly to autonomous meaning confuse two different ideas. AI may outperform humans as a problem solving tool. It does not generate self originating purpose. Producing good ideas is not the same as producing lived thought.
This distinction has practical importance.
It affects governance, responsibility, and trust. Treating AI as an autonomous agent hides the human and institutional structures that guide its behavior. The three layer model offers a clearer way to assign responsibility.
7. Conclusion
This paper has revisited and extended the three layer model of the AI industry.
The original model remains a useful ideal type, especially for understanding large language models. Its strength lies in explaining economic differences between training and inference.
The revised framework makes key assumptions explicit.
By identifying infrastructure as a foundational layer, it clarifies the role of hardware, energy, and execution environments. By including integrated systems such as robotics, it extends the model beyond generative AI while preserving its core logic.
Finally, the framework offers a disciplined approach to questions of meaning and autonomy. It explains why AI can appear grounded without achieving understanding, and why increased computing power does not produce autonomous purpose.
The three layer model, in this revised form, is not a prediction of future AI.
It is a lens for understanding present systems. These systems are layered, constrained by infrastructure, and shaped by institutions. Their intelligence is real and influential, but it remains fundamentally different from human thought.
関連記事:
