Neuron-Aware Data Selection In Instruction Tuning For Large Language Models
NAIT framework selects efficient instruction tuning data via neuron activation patterns, enhancing LLM performance.
Xin Chen, Junchao Wu, Shu Yang et al.
NAIT framework selects efficient instruction tuning data via neuron activation patterns, enhancing LLM performance.
Xin Chen, Junchao Wu, Shu Yang et al.
ESG-Bench significantly reduces hallucinations in long-context ESG report analysis using task-specific Chain-of-Thought prompting strategies.
Siqi Sun, Ben Peng Wu, Mali Jin et al.
WALAR method enhances low-resource language translation using monolingual data, surpassing LLaMAX model.
Yifeng Liu, Siqi Ouyang, Yatish Hosmane Revanasiddappa et al.
Proposed a PCA sweep method to optimize dimension selection in SSD, enhancing interpretability and stability.
Hubert Plisiecki, Maria Leniarska, Jan Piotrowski et al.
Long-form RewardBench evaluates reward models for long-form generation, revealing current models' deficiencies in long-form reward modeling.
Hui Huang, Yancheng He, Wei Liu et al.
HMS-BERT uses hybrid multi-task self-training for multilingual, multi-label cyberbullying detection, achieving a macro F1-score of 0.9847.
Zixin Feng, Xinying Cui, Yifan Sun et al.
Idea-Catalyst framework boosts scientific creativity via interdisciplinary insights, improving novelty by 21% and insightfulness by 16%.
Priyanka Kargupta, Shuhaib Mehri, Dilek Hakkani-Tur et al.
CLASP model detects malicious tokens using XGBoost classifier, achieving 95.9% token-level F1 score.
Alexandre Le Mercier, Thomas Demeester, Chris Develder
IndexCache accelerates sparse attention by reusing cross-layer indices, reducing 75% of computations, achieving 1.82x speedup.
Yushi Bai, Qian Dong, Ting Jiang et al.
Introduced a Polish long-context encoder model handling up to 8192 tokens, significantly improving long-document task performance.
Sławomir Dadas, Rafał Poświata, Marek Kozłowski et al.
LifeSim simulates user cognition via BDI model to enhance personalized assistant evaluation.
Feiyu Duan, Xuanjing Huang, Zhongyu Wei
MDER-DR framework enhances multi-hop QA with entity-centric summaries, achieving 66% improvement.
Riccardo Campi, Nicolò Oreste Pinciroli Vago, Mathyas Giudici et al.
IsalGraph method encodes any finite simple graph as a compact string over a nine-character instruction alphabet, suitable for graph similarity search.
Ezequiel Lopez-Rubio, Mario Pascual-Gonzalez
GLM-OCR combines CogViT visual encoder and GLM language decoder to enhance document understanding efficiency.
Shuaiqi Duan, Yadong Xue, Weihan Wang et al.
We release a large bilingual library dataset for GND-based multi-label classification.
Jennifer D'Souza, Sameer Sadruddin, Maximilian Kähler et al.
LLM-assisted MIPVU rule script generation enables interpretable Chinese metaphor identification; protocol choice is the main source of variation.
Weihang Huang, Mengna Liu
Introduced DOWIS dataset to evaluate SLLMs in multilingual settings, finding text prompts outperform spoken prompts.
Maike Züfle, Sara Papi, Fabian Retkowski et al.
N-gram models predict reading time best due to sensitivity to simple statistics.
James A. Michaelov, Roger P. Levy