-
LoRA+: Efficient Low Rank Adaptation of Large Models
Paper • 2402.12354 • Published • 6 -
The FinBen: An Holistic Financial Benchmark for Large Language Models
Paper • 2402.12659 • Published • 17 -
TofuEval: Evaluating Hallucinations of LLMs on Topic-Focused Dialogue Summarization
Paper • 2402.13249 • Published • 11 -
TrustLLM: Trustworthiness in Large Language Models
Paper • 2401.05561 • Published • 66
Collections
Discover the best community collections!
Collections including paper arxiv:2104.09864
-
Attention Is All You Need
Paper • 1706.03762 • Published • 50 -
ImageNet Large Scale Visual Recognition Challenge
Paper • 1409.0575 • Published • 8 -
Sequence to Sequence Learning with Neural Networks
Paper • 1409.3215 • Published • 3 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 12
-
Improving Text Embeddings with Large Language Models
Paper • 2401.00368 • Published • 79 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 16 -
Metadata Might Make Language Models Better
Paper • 2211.10086 • Published • 4 -
DecoderLens: Layerwise Interpretation of Encoder-Decoder Transformers
Paper • 2310.03686 • Published • 3
-
The RefinedWeb Dataset for Falcon LLM: Outperforming Curated Corpora with Web Data, and Web Data Only
Paper • 2306.01116 • Published • 32 -
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
Paper • 2205.14135 • Published • 11 -
RoFormer: Enhanced Transformer with Rotary Position Embedding
Paper • 2104.09864 • Published • 11 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 12