Building A Unified AI-centric Language System

Analysis, Framework and Future Work
Teaser animation
Harvard University, The University of Sydney
AI-Friendly Language Implementation Framework

A unified AI-centric language system that offers a more concise, unambiguous, and computationally efficient alternative to traditional human languages.

Abstract

Recent advancements in large language models have demonstrated that extended inference—through techniques can markedly improve performance, yet these gains come with increased computational costs and the propagation of inherent biases found in natural languages. This paper explores the design of a unified AI-centric language system that addresses these challenges by offering a more concise, unambiguous, and computationally efficient alternative to traditional human languages.

We analyze the limitations of natural language—such as gender bias, morphological irregularities, and contextual ambiguities—and examine how these issues are exacerbated within current Transformer architectures, where redundant attention heads and token inefficiencies prevail. Drawing on insights from emergent artificial communication systems and constructed languages like Esperanto and Lojban, we propose a framework that translates diverse natural language inputs into a streamlined AI-friendly language, enabling more efficient model training and inference while reducing memory footprints.

Finally, we outline a pathway for empirical validation through controlled experiments, paving the way for a universal interchange format that could revolutionize AI-to-AI and human-to-AI interactions by enhancing clarity, fairness, and overall performance.

Human Language vs. AI Language Requirements

Comparison of human and AI language requirements

Human communication requires speech, shared cultural context, and the capacity to negotiate meaning over centuries of gradual change. In contrast, AI systems do not inherently require spoken forms or social acceptance. The barest form of AI "language" might be a symbolic code—potentially even a single unpronounceable symbol—sufficient for exchanging precise information between machines. This radical difference in requirements underlies many of the inefficiencies in forcing AI systems to parse and produce human languages.

Proposed Implementation Framework

AI-Friendly Language Implementation Framework

Our proposed framework converts natural language into an AI-friendly language for efficient processing, then translates responses back to natural language for users. This approach reduces computational costs while maintaining or improving accuracy and fairness.

Key Findings

Language Biases and Irregularities

Artificial Intelligence models that process human language inherit many of the biases, irregularities, and ambiguities of those languages. This not only skews AI outputs but also raises fairness and interpretability concerns. These include gendered language bias, plural forms and morphological complexity, and context dependence.

Language Structure and Multi-Head Attention

Recent work on Transformers reveals how AI models learn and handle linguistic structure. Many attention heads appear redundant, with research showing that a large percentage can be removed after training without major performance drops. These findings suggest that a more structured, unambiguous language could enable even further model compression.

Toward an AI-Friendly Language

Drawing from constructed languages like Esperanto (designed for simplicity and neutrality) and Lojban (developed for logical clarity), we propose key principles for an AI-centric language:

  • Clarity and unambiguity - one parse per sentence
  • Consistency and regularity - eliminating irregular morphology
  • Conciseness and efficiency - fewer tokens for faster computation
  • Unlimited or adaptable vocabulary - avoiding polysemy
  • Reduced context dependence - minimizing ambiguous references
  • Computational efficiency - designed for fast, deterministic parsing

Future Work

To rigorously assess the benefits of an AI-centric language, future work should include constructing a small-scale "toy" language featuring a reduced grammar and systematically defined vocabulary. Two parallel neural models of equal size would be trained: one on this toy language and one on English, using identical architectures and hyperparameters.

Comparing their performance on tasks such as question answering, text classification, and summarization would illuminate whether the specialized language yields measurable efficiency gains—specifically, fewer tokens, lower inference time, and reduced memory footprint—while maintaining or improving accuracy and fairness.

Related Work

Our research builds upon several key areas in language model development and linguistic theory:

Chain of thought prompting (Wei et al., 2022) introduced methods for improving LLM reasoning through extended inference steps.

Multi-head attention analysis (Michel et al., 2019; Voita et al., 2019) demonstrated that many attention heads in Transformer models can be pruned without significant performance loss.

Research on emergent communication in AI systems has shown how artificial agents can develop their own communication protocols when working together on specific tasks.

Work on bias in large language models (Bolukbasi et al., 2016; Guo et al., 2024) has highlighted how AI systems inherit and sometimes amplify biases present in human languages.

BibTeX

@article{wang2025building,
  author    = {Wang, Edward Hong and Wen, Cynthia Xin},
  title     = {Building A Unified AI-centric Language System: analysis, framework and future work},
  journal   = {arXiv preprint arXiv:2502.04488},
  year      = {2025},
}