Tencent’s new AI method aligns preferences without retraining

Tencent’s new AI method aligns preferences without retraining

Tech in Asia·2025-05-30 20:00

🔍 In one sentence

Researchers have developed a new method for aligning large language models with multiple user preferences at once, improving performance without requiring retraining.

🏛️ Paper by:

Harbin Institute of Technology, Shenzhen, China; Nanyang Technological University, Singapore; Peng Cheng Laboratory, China; Beijing Academy of Artificial Intelligence, China; Tencent, China

Authors: Zhuo Li et al.

🧠 Key discovery

The study presents a framework called Hierarchical Mixture-of-Experts (HoE), which aligns large language models (LLMs) with diverse user preferences without substantial retraining. This is notable, as traditional methods often have difficulty balancing multiple objectives and typically require costly adjustments.

📊 Surprising results

Key stat: HoE outperformed 15 recent baselines across 14 objectives and 200 different preferences in various tasks. Breakthrough: The use of specialized experts and a routing mechanism enabled dynamic adjustments to user preferences, supporting performance across the full Pareto frontier. Comparison: HoE showed notable improvements over existing alignment methods, which often required retraining for each task or preference setting.

📌 Why this matters

This research questions the common assumption that models must be retrained to handle different tasks or objectives. The HoE framework provides a more flexible and cost-efficient alternative, with potential applications in real-world settings like personalized digital assistants that must adapt to changing user needs without ongoing retraining.

💡 What are the potential applications?

Personalized chatbots and virtual assistants that adjust to user preferences in real-time. Enhanced content moderation systems that balance helpfulness, safety, and tone (e.g., humor) based on user feedback. Multi-task learning systems that require less retraining when introducing new objectives.

⚠️ Limitations

The approach depends on access top pre-trained single-objective models, which may not always be available. Its performance also varies based on the effectiveness of the model merging techniques applied.

👉 Bottom line:

The HoE framework offers a new approach to align AI models with user preferences, providing a more adaptable and efficient method for handling diverse tasks.

📄 Read the full paper: [Multi-objective Large Language Model Alignment with Hierarchical Experts](http://arxiv.org/abs/2505.20925v1)

……

Read full article on Tech in Asia

Technology