Providing definition and guidelines in the prompt can improve the performance and robustness of instruction-tuned language models for zero-shot named entity recognition, especially on unseen entity types.
Cosine similarity can be interpreted as the sum of semantic similarities along the interpretable axes of normalized ICA-transformed embeddings.
Language models can be fine-tuned to generate well-calibrated linguistic expressions of uncertainty that accurately reflect the likelihood of their predictions being correct.
The proposed framework leverages a two-component pipeline architecture that integrates extractive and abstractive summarization techniques to generate high-quality Vietnamese multi-document summaries.
A novel neural retriever architecture that leverages pre-trained multilingual language models to effectively match freelancer profiles and project descriptions, enabling efficient and scalable candidate retrieval in a multilingual setting.
PARAPHRASUS is a comprehensive benchmark designed to assess the performance of paraphrase detection models across a diverse range of paraphrase phenomena.
A novel training approach called TaS that enables large language models to first generate reasonable thoughts and then express corresponding responses, mimicking the human cognitive process.
A novel extract-and-abstract paradigm, EXTABS, that jointly and seamlessly performs extractive and abstractive summarization tasks within a single encoder-decoder model, reducing error accumulation and improving performance.
A novel channel-based parallel decoding approach, DUO, that equips large language models with duplex capabilities, enabling simultaneous input processing and output generation, while requiring minimal additional training.
Even the latest large language models struggle to effectively detect known knowledge errors and unknown knowledge errors when playing roles, particularly for familiar knowledge.