an open dataset and web-based application for the study of metaphor

[Submitted on 1 Mar 2025 (v1), last revised 15 Apr 2025 (this version, v2)] Authors:Maddalena Bressler, Veronica Mangiaterra, Paolo Canal, ...
Read more
Nondeterministic Polynomial-time Problem Challenge: An Ever-Scaling Reasoning Benchmark for LLMs

arXiv:2504.11239v1 Announce Type: cross Abstract: Reasoning is the fundamental capability of large language models (LLMs). Due to the rapid progress ...
Read more
Long-Context Evaluation Beyond Literal Matching

[Submitted on 7 Feb 2025 (v1), last revised 26 Mar 2025 (this version, v2)] View a PDF of the paper ...
Read more
A Novel Approach for the Automated Evaluation of Open-Ended Question Generation

[Submitted on 16 Oct 2024 (v1), last revised 25 Mar 2025 (this version, v3)] View a PDF of the paper ...
Read more
A Floater-Free Framework for 3D Gaussian Splatting

[Submitted on 24 Mar 2025 (v1), last revised 25 Mar 2025 (this version, v2)] View a PDF of the paper ...
Read more
Exploring Training and Inference Scaling Laws in Generative Retrieval

arXiv:2503.18941v1 Announce Type: cross Abstract: Generative retrieval has emerged as a novel paradigm that leverages large language models (LLMs) to ...
Read more
A Scalable Data Synthesis Framework and Guided Tree Search for Automated Theorem Proving

[Submitted on 30 Dec 2024 (v1), last revised 21 Mar 2025 (this version, v3)] View a PDF of the paper ...
Read more
Unlocking the Multi-modal Potential of CLIP for Generalized Category Discovery

[Submitted on 15 Mar 2024 (v1), last revised 21 Mar 2025 (this version, v3)] View a PDF of the paper ...
Read more
Optimizing Data Mixtures by Predicting Language Modeling Performance

[Submitted on 25 Mar 2024 (v1), last revised 20 Mar 2025 (this version, v2)] View a PDF of the paper ...
Read more
Machine Unlearning in Hyperbolic vs. Euclidean Multimodal Contrastive Learning: Adapting Alignment Calibration to MERU

arXiv:2503.15166v1 Announce Type: cross Abstract: Machine unlearning methods have become increasingly important for selective concept removal in large pre-trained models. ...
Read more