Results 41 to 50 of about 70,803 (186)

Agentic AutoSurvey: Let LLMs Survey LLMs

open access: yes
29 pages, 7 ...
Liu, Yixin   +3 more
openaire   +2 more sources

Evaluating research quality with Large Language Models: An analysis of ChatGPT’s effectiveness with different settings and inputs

open access: yesJournal of Data and Information Science
Evaluating the quality of academic journal articles is a time consuming but critical task for national research evaluation exercises, appointments and promotion.
Thelwall Mike
doaj   +1 more source

Foregrounding doctoral knowledge and knower in the age of Generative Artificial Intelligence

open access: yesTransformation in Higher Education
While Generative Artificial Intelligence (AI) presents new challenges for doctoral education, it also offers an opportunity to refocus doctoral programmes on their fundamental purposes: contributing to knowledge and developing critical researchers.
Sioux McKenna
doaj   +1 more source

Set-LLM: A Permutation-Invariant LLM

open access: yes
While large language models (LLMs) demonstrate impressive capabilities across numerous applications, their robustness remains a critical concern. This paper is motivated by a specific vulnerability: the order sensitivity of LLMs. This vulnerability manifests itself as the order bias observed when LLMs decide between possible options (for example, a ...
Egressy, Beni, Stühmer, Jan
openaire   +2 more sources

Digital Friends and Empathy Blindness

open access: yesOpen Philosophy
Can chatbot-based virtual relationships replace physical ones? One possible bottleneck is lack of empathy in chatbots, as well as the attraction of physical relationships.
Bangsgaard Alberte Romme   +3 more
doaj   +1 more source

Programming Assessment in E-Learning through Rule-Based Automatic Question Generation with Large Language Models

open access: yesJournal of Applied Informatics and Computing
This study develops an evaluation instrument for Python programming using a Rule-Based Automatic Question Generation (AQG) system integrated with Large Language Models (LLMs), designed based on the Revised Bloom’s Taxonomy.
Halim Teguh Saputro   +2 more
doaj   +1 more source

LLMs Judging LLMs: A Simplex Perspective

open access: yes
Given the challenge of automatically evaluating free-form outputs from large language models (LLMs), an increasingly common solution is to use LLMs themselves as the judging mechanism, without any gold-standard scores. Implicitly, this practice accounts for only sampling variability (aleatoric uncertainty) and ignores uncertainty about judge quality ...
Vossler, Patrick   +4 more
openaire   +2 more sources

Retrieval-Augmented Generation vs. Baseline LLMs: A Multi-Metric Evaluation for Knowledge-Intensive Content

open access: yesInformation
(1) Background: The development of Generative Artificial Intelligence (GenAI) is transforming knowledge-intensive domains such as Education. However, Large Language Models (LLMs), which serve as the foundational components for GenAI tools, are trained on
Aparna Vinayan Kozhipuram   +2 more
doaj   +1 more source

Down and Out in Italian towns: measuring the impact of economic downturns on crime [PDF]

open access: yes, 2013
The paper investigates the effect of local economic conditions on crime. The study focuses on Italy’s local labor markets and analyzes the short-term response of crime to the severe slump of 2007-2009.
de Blasio, Guido, Menon, Carlo
core   +1 more source

LLM-SQL-Solver: Can LLMs Determine SQL Equivalence?

open access: yes, 2023
Judging the equivalence between two SQL queries is a fundamental problem with many practical applications in data management and SQL generation (i.e., evaluating the quality of generated SQL queries in text-to-SQL task). While the research community has reasoned about SQL equivalence for decades, it poses considerable difficulties and no complete ...
Zhao, Fuheng   +5 more
openaire   +2 more sources

Home - About - Disclaimer - Privacy