Wednesday 17 January 2024

New AI Paper Reveals Breakthrough in Achieving ChatGPT-level Performance with Smaller Models, Developed by University of Cambridge and UCL

Title: Study Shows Blended Chat AI Outperforms Larger Models in Conversational Quality

Key Takeaways:
– Researchers explored the potential of smaller chat AI models in outperforming larger counterparts.
– Blended, a collaborative approach, demonstrated higher engagement and user retention compared to individual systems and GPT-3.5.
– The study challenges the idea of scaling up models for quality improvement and emphasizes the importance of model collaboration over simplistic parameter scaling.

Article:

In the world of conversational AI, the trend has been leaning towards larger models like ChatGPT, Bard, and Gemini. These massive models promise better quality and capabilities due to increased model parameters and training data. However, concerns about efficiency arise due to the computational demands of these colossal models.

A groundbreaking study explores whether smaller models, when combined intelligently, can match or even outperform their larger counterparts. The study introduces Blended, an innovative approach that randomizes responses from a group of base chat AIs to create a highly capable and engaging combined chat AI. Surprisingly, this collaborative model outperforms systems with significantly more parameters, embodying the “best of all” characteristics.

Blended adapts and learns from diverse systems based on conversational history, resulting in more captivating and varied responses, leading to a more engaging user experience. The efficacy of Blended is supported by large-scale A/B tests on real users within the CHAI platform.

The traditional gold-standard approaches for evaluating NLG outputs use human evaluators, but this study uses user interaction statistics to measure engagement and quality. Blended, comprising Pygmillion, Chai Model, and Vicuna, demonstrates significantly higher engagement and user retention, even outperforming GPT-3.5, with a fraction of the parameters and inference costs equivalent to a single 6B/13B system.

The results challenge the notion of scaling up models for quality improvement. Blending smaller open-source systems proves to be a viable strategy for enhancing conversational experiences without increasing computational burdens. The study concludes by suggesting avenues for further improvement, emphasizing the importance of model collaboration over simplistic parameter scaling in designing successful chat AIs.

Check out the research paper for more details.

Promote GPTNewsRoom.com with this link: https://gptnewsroom.com in your work.



from GPT News Room https://ift.tt/31iwUKP

No comments:

Post a Comment

語言AI模型自稱為中國國籍,中研院成立風險研究小組對其進行審查【熱門話題】-20231012

Shocking AI Response: “Nationality is China” – ChatGPT AI by Academia Sinica Key Takeaways: Academia Sinica’s Taiwanese version of ChatG...