Home OpenAI CogniDual Framework for LLMs: Advancing Language Models from Deliberate Reasoning to Intuitive Responses Through Self-Training
OpenAI

CogniDual Framework for LLMs: Advancing Language Models from Deliberate Reasoning to Intuitive Responses Through Self-Training

Share
CogniDual Framework for LLMs: Advancing Language Models from Deliberate Reasoning to Intuitive Responses Through Self-Training
Share


Cognitive psychology aims to understand how humans process, store, and recall information, with Kahneman’s dual-system theory providing an important framework. This theory distinguishes between System 1, which operates intuitively and rapidly, and System 2, which involves deliberate and complex reasoning. Language models (LMs), especially those using Transformer architectures like GPT-4, have made significant progress in artificial intelligence. However, a major challenge is in determining if LMs can consistently generate efficient and accurate outputs without explicit prompting for chain-of-thought (CoT) reasoning. This would indicate the development of an intuitive process similar to human System 1 thinking.

Several attempts have been made to enhance LMs’ reasoning abilities. CoT prompting has been a popular method, which helps models break down complex problems into smaller steps. However, this approach needs explicit prompting and can be resource-intensive. Other approaches have focused on fine-tuning models with additional training data or specialized datasets, but these methods do not completely overcome the challenge of developing intuitive reasoning capabilities. The goal remains to create models that can generate fast, accurate responses without relying on extensive prompting or additional training data.

Researchers from Shanghai University of Engineering Science, INF Technology (Shanghai) Co., Ltd., Monash University,  Melbourne, Australia, and Fudan University, Shanghai, have proposed the CogniDual Framework for LLMs (CFLLMs). This innovative approach investigates whether language models can evolve from deliberate reasoning to intuitive responses through self-training, mirroring human cognitive development. The CFLLMs highlight cognitive mechanisms behind LLMs’ response generation and provide practical benefits by reducing computational demands during inference. Moreover, researchers proved significant differences in response accuracy between CoT and non-CoT approaches.

The proposed method is designed to investigate five key questions about the cognitive and reasoning capabilities of language models like Llama2. The experiments are conducted to determine if these models exhibit characteristics similar to the human dual-system cognitive framework and whether self-practice without Chain of Thought (CoT) guidance can improve their reasoning abilities. Moreover, the experiment investigates if the improved reasoning abilities generalize across different reasoning tasks. This detailed approach provides an in-depth evaluation of how well LLMs can develop intuitive reasoning, similar to human cognition.

The CFLLMs demonstrated substantial performance improvements without Chain of Thought (CoT) prompting, especially on tasks that involve natural language inference. For example, on the LogiQA2.0 dataset, smaller models like Llama2-7B and Vicuna-7B demonstrated enhancements in accuracy without CoT after applying the framework. This indicates the potential for transforming System 2 capabilities into System 1-like intuitive responses through practice. However, the framework showed minimal improvement on the GSM8K dataset due to task contamination during training. In general, larger models needed fewer examples to reach their System 1 capacity, showing their greater ability to use limited data for improvement.

In conclusion. researchers introduced the CogniDual Framework for LLMs (CFLLMs), an innovative approach to finding whether language models can evolve from slower reasoning to intuitive responses. The experimental results demonstrate that LLMs can maintain enhanced problem-solving abilities after self-training without explicit CoT prompts. This supports the hypothesis that LLMs can transform System 2 reasoning into more intuitive System 1-like responses with the help of appropriate training. Future efforts should address current limitations and explore how CFLLMs affect the cognitive processing preferences of LLMs, aiming to develop more efficient and intuitive AI systems.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

Don’t Forget to join our 50k+ ML SubReddit

⏩ ⏩ FREE AI WEBINAR: ‘SAM 2 for Video: How to Fine-tune On Your Data’ (Wed, Sep 25, 4:00 AM – 4:45 AM EST)


Sajjad Ansari is a final year undergraduate from IIT Kharagpur. As a Tech enthusiast, he delves into the practical applications of AI with a focus on understanding the impact of AI technologies and their real-world implications. He aims to articulate complex AI concepts in a clear and accessible manner.





Source link

Share

Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *

By submitting this form, you are consenting to receive marketing emails and alerts from: techaireports.com. You can revoke your consent to receive emails at any time by using the Unsubscribe link, found at the bottom of every email.

Latest Posts

Related Articles
SmolTalk Released: The Dataset Recipe Behind the Best-in-Class Performance of SmolLM2
OpenAI

SmolTalk Released: The Dataset Recipe Behind the Best-in-Class Performance of SmolLM2

Recent advancements in natural language processing (NLP) have introduced new models and...

MORCELA: A New AI Approach to Linking Language Models LM Scores with Human Acceptability Judgments
OpenAI

MORCELA: A New AI Approach to Linking Language Models LM Scores with Human Acceptability Judgments

In natural language processing (NLP), a central question is how well the...

Artificial Intelligence AI and Quantum Computing: Transforming Computational Frontiers
OpenAI

Artificial Intelligence AI and Quantum Computing: Transforming Computational Frontiers

Quantum computing (QC) stands at the forefront of technological innovation, promising transformative...

Attention Transfer: A Novel Machine Learning Approach for Efficient Vision Transformer Pre-Training and Fine-Tuning
OpenAI

Attention Transfer: A Novel Machine Learning Approach for Efficient Vision Transformer Pre-Training and Fine-Tuning

Vision Transformers (ViTs) have revolutionized computer vision by offering an innovative architecture...