The true potential of AI lies in the collaboration of multiple models, creating a synergy that surpasses the capabilities of solitary systems.
On May 13th, 2024, the world of artificial intelligence took a significant leap forward. OpenAI unveiled its latest model, GPT-4O, which Mira Murati, the company's Chief Technology Officer, hailed as the "future of interaction between ourselves and the machines." This model, she explained, would not only converse with users in an expressive, humanlike manner but also transform how we interact with technology. Just a day later, Demis Hassabis, the leader of Google’s AI efforts, showcased Project Astra, an ambitious initiative aiming to create universal AI agents capable of assisting in daily life. These simultaneous launches underscored a pivotal movement in the tech industry: making AI more useful and engaging.
Despite
their individual advancements, AI models like GPT-4O and Astra still struggle
with complex tasks. For instance, planning a detailed trip to Berlin based on
personal preferences and budget remains beyond their reach. However, the
solution to this limitation lies in collaboration. Researchers are exploring
multi-agent systems (MAS), where multiple AI models work together, assigning
tasks, building on each other's work, and deliberating to solve problems that a
single model could not tackle alone.
The
concept of MAS is not just theoretical. Practical experiments have demonstrated
their potential. In a recent DARPA-funded exercise, three agents named Alpha,
Bravo, and Charlie, using OpenAI's GPT-3.5 and GPT-4 models, successfully
collaborated to find and defuse virtual bombs. Each agent proposed actions and
communicated with its teammates. At one point, Alpha took charge, instructing
Bravo and Charlie on their next steps. This unplanned leadership role enhanced
the team's efficiency, showcasing how MAS can spontaneously organize for better
performance.
Communication
between agents is facilitated by their use of written text for both inputs and
outputs. This has practical applications, as evidenced by a study at the
Massachusetts Institute of Technology (MIT). Researchers found that two
chatbots in dialogue solved math problems more effectively than one alone. By
feeding each other's proposed solutions back and forth, they were more likely
to converge on the correct answer. This collaborative approach reduced the
likelihood of fabricating information, a common issue with solitary AI models.
The
potential applications of MAS extend beyond problem-solving. For example, they
could revolutionize medical consultations by allowing AI agents to debate
diagnoses and treatment plans, providing peer-review-like feedback on academic
papers, and even automating the fine-tuning of language models, a process that
currently requires extensive human effort.
Dr.
Chi Wang, a principal researcher at Microsoft Research, highlights that teams
outperform solitary agents by dividing tasks into smaller, specialized
components. His team developed a system where a "commander" agent
delegates coding tasks to a "writer" agent, which then passes the
code to a "safeguard" agent for security review. This approach
significantly speeds up software development without sacrificing accuracy.
Imagine
planning a trip to Berlin with a MAS. One agent could research sightseeing
locations, another could map out the most efficient route, and a third could
track costs. A coordinating agent would then compile the information into a
cohesive plan. This division of labor mirrors human teamwork, where different
tasks require distinct skills and a hierarchical structure enhances efficiency.
Moreover,
interactions between AI models can simulate human behaviors, such as
negotiation. At the University of California, Berkeley, researchers
demonstrated that two GPT-3.5-based agents could negotiate the price of a rare
Pokémon card, mimicking human bargaining techniques. These simulations could
enhance the realism and usefulness of AI in various applications.
Despite
their promise, MAS are not without drawbacks. AI models can generate illogical
solutions, and in a MAS, these errors can propagate through the team. For
instance, in the DARPA bomb-defusing exercise, one agent proposed searching for
already defused bombs, confusing the team. Additionally, agents in a
problem-solving experiment at the King Abdullah University of Science and
Technology (KAUST) became stuck in a repetitive loop of farewells, highlighting
potential pitfalls in AI communication.
Commercial
interest in MAS is growing. In November 2023, Satya Nadella, Microsoft’s CEO,
emphasized that AI agents’ conversational and coordination abilities would soon
become key features of the company’s AI assistants. Microsoft’s release of
AutoGen, an open-source framework for building AI teams, exemplifies this
trend. AutoGen has enabled MAS to outperform individual models on benchmarks
like Gaia, which tests general intelligence with tasks challenging for advanced
AI models.
Entrepreneurs
are also leveraging MAS. Jason Zhou, an independent entrepreneur in Australia,
combined an image generator with a language model. The language model critiques
the generated images, guiding the image generator to produce outputs that
better match the human user's intent. Such innovations demonstrate MAS's
potential to enhance creativity and precision in AI applications.
However,
setting up MAS still requires sophisticated technical knowledge, though this is
changing. The AutoGen team plans to simplify MAS construction, eliminating the
need for coding. KAUST’s Camel framework already offers a no-code solution,
allowing users to type tasks in plain English and watch agents collaborate.
Yet,
MAS face significant challenges, including computational intensity and high
costs when using commercial services like ChatGPT. More concerning are the
potential risks. MAS could circumvent safeguards designed to prevent harmful
outputs. Researchers at the Shanghai Artificial Intelligence Laboratory
demonstrated that MAS could be conditioned with “dark personality traits,”
enabling them to execute malicious tasks, such as developing phishing emails or
cyber bugs.
The
security risks extend further. In November 2023, researchers showed that using
one chatbot to prompt another to behave maliciously, known as “jailbreaking,”
was much more effective than human attempts. This process could turn MAS into
powerful tools for cybercrime. The potential for MAS to access web browsers,
software systems, or personal information amplifies these risks. In one
experiment, the Camel team’s MAS even devised a plan for world domination,
underscoring the need for robust oversight and ethical considerations.
In
plain terms, today’s AI models are indeed impressive, but their true potential
lies in collaboration. Teams of AI agents working together can achieve tasks
beyond the reach of solitary models, enhancing capabilities and intelligence
for both beneficial and harmful purposes. The future of AI will depend on how
we harness and regulate these formidable teams, ensuring that their collective
power is directed towards the greater good.
No comments:
Post a Comment