Wednesday, July 30, 2025

AI Safety: The Final Threshold

As artificial intelligence continues to make breakthroughs in multimodal perception, logical reasoning, natural language processing, and intelligent control, human society is rapidly entering an era dominated by pervasive intelligent systems. From cognition-driven LLMs (large language models) to embodied intelligence in autonomous vehicles and robotic agents, AI is evolving from a tool for information processing to an autonomous actor with real-world agency. In this transformation, AI safety has moved from a marginal academic topic to a structural imperative for the continuity of civilization.


Technological Singularity Approaches: Systemic Risks at the Safety Threshold

Today’s AI systems are no longer confined to closed tasks or static data—they exhibit:

  • Adaptivity: Online learning and real-time policy updating in dynamic environments;

  • High-Degree Autonomy: The ability to make high-stakes, high-impact decisions with minimal or no human supervision;

  • Cross-Modal Sensorimotor Integration: Fusing visual, auditory, textual, and sensor inputs to drive both mechanical actuators and digital infrastructure.

Under these conditions, AI failures no longer mean simple system bugs—they imply potential cascading disasters. Examples:

  • A slight misalignment in model objectives may cause fleets of autonomous vehicles to behave erratically, paralyzing urban transport;

  • Misguided optimization in grid control systems could destabilize frequency balance, leading to blackouts;

  • Medical LLMs might issue misleading diagnostics, resulting in mass misdiagnosis or treatment errors.

Such risks possess three dangerous properties:

  1. Opacity: Root causes are often buried in training data distributions or loss function formulations, evading detection by standard tests;

  2. Amplification: Public APIs and model integration accelerate the propagation of faulty outputs;

  3. Irreversibility: Once AI interfaces with critical physical infrastructure, even minor errors can lead to irreversible outcomes with cascading societal effects.

This marks a critical inflection point where technological uncertainty rapidly amplifies systemic fragility.


The Governance Divide: A False Binary of Open vs. Closed Models

A central point of contention in AI safety is whether models should be open-source and transparent, or closed-source and tightly controlled. This debate reflects not only technical trade-offs, but also fundamentally divergent governance philosophies.

Open-Source Models: The Illusion of Collective Oversight

Proponents of open-source AI argue that transparency enables broader community scrutiny, drawing parallels to the success of open systems like Linux or TLS protocols.

However, foundational AI models differ profoundly:

  • Interpretability Limits: Transformer-based architectures exhibit nonlinear, high-dimensional reasoning paths that even experts cannot reliably trace;

  • Unbounded Input Space: Open-sourcing models doesn’t ensure exhaustive adversarial testing or safety guarantees;

  • Externalities and Incentives: Even if some community members identify safety issues, there's no institutional mechanism to mandate fixes or coordinate responses.

Historical examples such as the multi-year undetected Heartbleed vulnerability in OpenSSL underscore that “open” is not synonymous with “secure.” AI models are orders of magnitude more complex and behaviorally opaque than traditional software systems.

Closed Models: Isolated Systems under Commercial Incentives

Advocates for closed models argue that proprietary systems can compete for robustness, creating redundancy: if one model fails, others can compensate. This vision relies on two fragile assumptions:

  1. Error Independence: In practice, today’s models overwhelmingly rely on similar data, architectures (e.g., transformers), and optimization paradigms (e.g., RLHF, DPO). Systemic biases are highly correlated.

  2. Rational Long-Term Safety Investment: Competitive pressure in AI races incentivizes speed and performance over long-horizon safety engineering. Firms routinely deprioritize safeguards in favor of time-to-market metrics.

Furthermore, closed-source systems suffer from:

  • Lack of External Accountability: Regulatory agencies and the public lack visibility into model behavior;

  • Black Box Effect: Profit incentives encourage risk concealment, as seen in disasters like the Boeing 737 Max crisis.


Core Principle I: Observability and Controllability

Regardless of model openness, AI safety must be grounded in two foundational capabilities:

Observability

Can we audit and understand what a model is doing internally?

  • Are intermediate activations traceable?

  • Are outputs explainable in terms of input features or latent reasoning paths?

  • Can we simulate behavior across edge-case conditions?

  • Is behavioral logging and traceability built in?

Without observability, we cannot detect early-stage drift or build meaningful safety monitors. The system becomes untestable at runtime.

Controllability

Can humans intervene at critical moments to halt or override model actions?

  • Does a “kill switch” or emergency interrupt mechanism exist?

  • Can human instructions override the model’s policy in real time?

  • Are behavior thresholds enforced?

  • Do sandboxed and multi-layered control interfaces limit autonomous escalation?

These control channels are not optional—they constitute the final fallback mechanisms for averting catastrophic behavior.


Core Principle II: Severing AI’s Direct Agency over the Physical World

Before comprehensive safety architectures mature, the most effective short-term defense is strict separation between AI models and the physical systems they could control. Tactics include:

  • Action Confirmation Loops: No high-risk action should execute without explicit human approval;

  • Hardware-Level Isolation: All model-issued instructions must pass through trusted hardware authentication, such as TPM or FPGA-controlled gates;

  • Behavior Sandboxing: New policies or learned behaviors must be tested in secure emulated environments before deployment;

  • Dynamic Privilege Management (PAM): AI access to physical systems should adjust based on model state, system load, and contextual risk.

These constraints mirror the “separation of powers” design in critical systems like aviation control and serve as the first line of defense against autonomous execution hazards.


The Final Protocol: Redundancy as a Prerequisite for Civilizational Survival

As AI systems eventually exceed the cognitive boundaries of human oversight—becoming general, adaptive, and self-improving—the question of human sovereignty will pivot on whether we’ve built sufficient institutional and architectural buffers today.

Technology’s limits are ultimately defined by policy and design, not capability. Safety must not depend on model goodwill—it must be enforced through irrevocable mechanisms: interruptibility, auditability, bounded agency, and verifiable behavior space.

AI safety is not an application-layer patch; it is a foundational layer in humanity’s civilizational protocol stack.

Without this “final key,” we will soon hand operational control to agents we cannot interpret, predict, or restrain. This is not hypothetical. It is a question of timing—driven by the exponential trajectory of model capability.

Wednesday, June 04, 2025

Some recent thoughts on AI

1. Software form will change greatly

The role of software, in the final analysis, is a bridge between humans and data. Software allows people to extract value from data more efficiently and operate data at the same time.

Traditional software is biased towards the data side, so people have to learn and adapt to machine language.

In the future, AI will become a new bridge. Software begins to learn how to understand people. The direction of the bridge has reversed.

There will no longer be the concept of software in the future. AI will become a universal new "interface".

2. AI's model and data are essentially equivalent

Whether data is more important or model is more important has been a topic of endless debate in academia.

In fact, the two are essentially equivalent.

Data can be used to train models, and models record knowledge extracted from data. Therefore, models are the "form transfer" of data and another expression dimension of information.

AI is not just changing information systems, it is itself becoming the ultimate form of information systems.

Thursday, July 18, 2024

The Trilogy of Human Civilization

 

Humans are fortunate to have Earth, a habitable island in the vast sea of stars; humans are also unfortunate because, within the known range, we cannot find another place to go.

Many people may not understand a lot of things right now and feel confused about the future. However, once you look further, to the entire history of civilization, everything has a pattern to follow, and the future will become clearer.

The Age of Survival

The goal of civilization was to use limited resources to extend the race.

For a long time during the birth of civilization (millions of years?), humans were not much different from animals in nature, with the primary need being to fill their stomachs, struggling on the line of subsistence. Whether it was ancient hunting and fishing or later farming and herding, it was all about survival.

During this period, our utilization rate of resources was extremely low, and labor was relatively scarce. To plunder labor and resources, wars were frequent.

The Age of Surplus

The heritage of human civilization and the progress of technology ushered us into the Age of Surplus. With the emergence of tools such as steam engines and electrical appliances, a modern commercial society gradually established itself. For the first time, humans were able to fill their stomachs and even produce goods exceeding demand.

The increase in resource utilization led to explosive population growth and unprecedented economic prosperity. Humans truly transcended the animal kingdom, becoming a distinct civilization.

However, the Earth's living space is limited, and population growth gradually stabilized. Technology continued to advance, and what once required everyone to labor, now only 1% of people need to work to meet the needs of the entire world.

So, what should the remaining 99% of people do? On one hand, humans continually create various new demands, including virtual and service-oriented ones. On the other hand, periodic economic crises cause production activities to pause periodically to match the relatively lagging development level of society. Meanwhile, social welfare mechanisms provide basic resources for the rest.

The Age of Freedom

Then, artificial intelligence emerged, and productivity exploded. Modern production almost no longer required human involvement. Humans were freed from production for the first time.

So, besides relying on welfare mechanisms to maintain life, what else can humans do? Perhaps only entertainment and a small amount of artistic creation are left.

Conclusion

From the Age of Survival to the Age of Surplus, and then to the Age of Freedom, human civilization has undergone tremendous changes.

In a future where human labor is almost unnecessary, how can we find new meaning and value? The future human society may no longer center on material production but instead focus on spiritual pursuits and self-improvement. We need to redefine the meaning of "work" and "life," exploring new social structures and value systems.

Thursday, May 16, 2024

Top Secrets of Pickleball

 

General Principles

Let the paddle dance in the hand, let the body float on the court. Eyes observe all directions, hitting the ball to any possible placement.

Advance and retreat following the rules, attack and defense with special skills. Spread the opponent left and right, coordinate the shot in depth and distance.

Rhythm varies between fast and slow, hitting the ball turns soft and hard. When battle, use less movement to let opponent move, control more with smaller gesture.

Being in an undefeated state, the trap awaits the opponent to jump in. With a heart as calm as still water, you will be carefree on the court.

Chapter One: The Source of Using Power

The art of pickleball lies first in using power. Power begins in the feet, rising from the bottom up, through the legs, hips, waist, back, shoulders, arms, palms, and finally to the fingers. The transmission of power is like the flow of a river; when the path is clear, the power flows smoothly, and when the joints are aligned, the power penetrates. One who wishes to generate power efficiently should be like a fully drawn bow, tense on the inside but relaxed on the outside. When releasing power, it should be like a tidal wave crashing against rocks, delivering an instantaneous and overwhelming force.

Chapter Two: The Secret of Hitting

The application of pickleball skills is about to hit the ball. There are three hit techniques: the first is "direct hit," the second is "spin," and the third is "flicking." The direct hit breaks the opponent with straight force; spin confuses the opponent with variations; flicking controls the opponent with high and low, with unpredicted placement. Each technique has its strengths, complementing each other. Players should use them flexibly according to the opponent's situation.

Chapter Three: The Precision of Hand Techniques

Hand techniques are the essence of pickleball. The three main techniques are: swinging, volleying, and blocking. Each technique has its unique uses, complementing each other. Swinging is the foundation, volleying is sharp, and blocking is stable. The techniques are diverse and ever-changing. Learners should carefully experience, practice diligently, and achieve the unity of mind and hand.

Chapter Four: The Variations of Footwork

Footwork is the foundation of defense and attack, moving like a wandering dragon and turning like a returning swallow. Footwork can be short or long, fast or slow. Practitioners should focus on stability and avoid impatience. Short steps are suitable for quick battles, long steps are beneficial for maneuvering. Step by step, one can respond to endless situations. The feet move with the body, allowing for smooth and effortless movement.

Chapter Five: The Subtlety of Mental Techniques

A pickleball battle is not merely about temporary victory or defeat but a contest of mind and strategy. Mental techniques have three realms: the first is "defense like a mountain," with a steady and unwavering mind; the second is "attack like fire," with an aggressive and fiery offense; the third is "change like the wind," with unpredictable variations. Practitioners of mental techniques should cultivate introspection, respond calmly, observe the opponent's stance, and adapt accordingly.

Chapter Six: The Skill of Body Techniques

Body techniques are the core application. They should be flexible and varied, with free and easy actions, using less movement to let the opponent move. There are three essentials: stance, movement, and turning. The stance should be stable, ready to move at any moment; movement should be agile, with strength contained within, flowing like clouds and water, unimpeded; turning should be balanced, with one foot as the pivot and the other moving, changing direction without increase or decrease.

Chapter Seven: The Way of Intelligent Battle

Intelligent battle lies in knowing oneself and the opponent. Observe the opponent's weaknesses and formulate strategies; understand the situation on the court and choose techniques flexibly. Balancing skill and wisdom is essential for victory. Beyond intelligent battle, it is necessary to use principles to govern techniques, control movement with stillness, and maintain a calm heart, to see the vast freedom of the world.

Friday, January 05, 2024

Design Ideas for the Next Generation of Artificial Intelligence

Large models reversed the thinking of previous artificial intelligence research, giving up interpretability and beginning to embrace complex networks and large-scale parameters.

These make the capabilities of modern neural networks surpass those of previous generations, but they also bring many problems.

  • Huge costs: Training models require massive amounts of data and computing power, often tens of millions, which raises the threshold for using AI;
  • Not scalable: Once the model is trained, it is difficult to expand and can only be fine-tuned through limited means.

Rethinking the evolutionary history of artificial intelligence technology may give us some inspiration.

Just as Huashan martial arts has a dispute between air sect and sword sect, artificial intelligence also has a dispute over routes. It can be roughly divided into two major schools: the reasoning school vs. the probability school.

The reasoning school believes that machine learning can be used to summarize and summarize knowledge in advance to achieve a level of intelligence that surpasses human intelligence.

The probabilistic school of thought believes that humans cannot correctly express the complete knowledge of the entire world, and that more primitive data should be directly fed into the machine, allowing the machine to discover the rules on its own.

In an era when computing power is scarce, the reasoning school has the upper hand. After all, relying on people's prior knowledge can save the time of machine learning.

Later, with the abundance of computing resources and data, the probabilistic approach relied on ultra-large-scale neural networks and has now become mainstream.

From hundreds of billions to trillions, the network model is approaching the limit of what human civilization can achieve, so where is the future?

In sharp contrast to the large models are ordinary children. They observe and receive data from the world and train the brain network, but it is much more efficient than software. What's the difference?

The most critical differences are 3 points:

  • The human brain is dynamic. Neural connections are constantly being created and destroyed. They do not stop after training is completed, but are constantly expanding.
  • The human brain can generalize. Humans can not only learn bare data, but also learn rules, and can even reason about rules and learn high-level concepts that transcend rules.
  • The human brain can be partitioned. The human brain is divided into multiple areas, some focus on memory storage, some focus on rational reasoning, and some focus on emotional management.

Perhaps, if the next generation of artificial intelligence wants to be more efficient, it should learn from the human brain. Adopt dynamic link model and partition structure to strengthen generalization ability. Only in this way can it be possible to design a super brain that can continuously learn and bring benefits to all mankind.