Results 331 to 340 of about 2,849,722 (390)
Some of the next articles are maybe not open access.

Reasons for Reasons

2023
Abstract Suppose Bob says, “I want to have reasons for what I do.” The skeptic replies, “But why do you want to have reasons?” Bob responds, “It’s just a brute fact about me. I want to have reasons for what I do.” The skeptic pounces: “But then you’re unreasonable.
openaire   +1 more source

Reinforcement Learning with Verifiable Rewards Implicitly Incentivizes Correct Reasoning in Base LLMs

arXiv.org
Recent advancements in long chain-of-thought (CoT) reasoning, particularly through the Group Relative Policy Optimization algorithm used by DeepSeek-R1, have led to significant interest in the potential of Reinforcement Learning with Verifiable Rewards ...
Xumeng Wen   +11 more
semanticscholar   +1 more source

SFT or RL? An Early Investigation into Training R1-Like Reasoning Large Vision-Language Models

Trans. Mach. Learn. Res.
This work revisits the dominant supervised fine-tuning (SFT) then reinforcement learning (RL) paradigm for training Large Vision-Language Models (LVLMs), and reveals a key finding: SFT can significantly undermine subsequent RL by inducing ``pseudo ...
Guiming Hardy Chen   +7 more
semanticscholar   +1 more source

Towards Reasoning Era: A Survey of Long Chain-of-Thought for Reasoning Large Language Models

arXiv.org
Recent advancements in reasoning with large language models (RLLMs), such as OpenAI-O1 and DeepSeek-R1, have demonstrated their impressive capabilities in complex domains like mathematics and coding.
Qiguang Chen   +9 more
semanticscholar   +1 more source

LMM-R1: Empowering 3B LMMs with Strong Reasoning Abilities Through Two-Stage Rule-Based RL

arXiv.org
Enhancing reasoning in Large Multimodal Models (LMMs) faces unique challenges from the complex interplay between visual perception and logical reasoning, particularly in compact 3B-parameter architectures where architectural constraints limit reasoning ...
Yi Peng   +9 more
semanticscholar   +1 more source

Reason and Reasons

2001
Abstract Grice begins this chapter by discussing to what extent the notion of variable rationality can be derived from that of flat rationality, and thus from the concept of a rational being alone. He then draws a distinction between ‘explanatory’ (motivating) and ‘justificatory’ (normative) reasons, as well as ‘personal’ reasons that ...
openaire   +1 more source

Giving reasons and given reasons

2021
Abstract Derek Parfit, as a leader of the ‘reasons-first’ movement, says that the concept of a reason is fundamental and indefinable. But his concept of a reason differs from most philosophers’. Most philosophers take a reason to be a fact, whereas Parfit says that reasons are given by facts, not that they are facts.
openaire   +1 more source

Reason and Reasoning

2001
Abstract Grice deals with the nature of practical and non‐practical (‘alethic’) reasoning, particularly what may be called imperfect reasoning. It consists of ‘misreasoning’ (the misapplication of good principles of reasoning), ‘incomplete reasoning’ (formally invalid but nonetheless correct inferences due to a missing premise that is ...
openaire   +1 more source

Measuring Multimodal Mathematical Reasoning with MATH-Vision Dataset

arXiv.org
Recent advancements in Large Multimodal Models (LMMs) have shown promising results in mathematical reasoning within visual contexts, with models approaching human-level performance on existing benchmarks such as MathVista. However, we observe significant
Ke Wang   +5 more
semanticscholar   +1 more source

Reasons and Reason

Symposium, 1999
Katherine Mo"ison charges that in my book, Back to Reality, ffailed to make my case for the adoption of a modest realism in postmodem (na"ative) therapy, because f failed to establish the motive behind that movement's adoption of antirealism. Infact, in Back to Reality, I put forth several reasons for therapists of all stripes to favor a modest realism
openaire   +1 more source

Home - About - Disclaimer - Privacy