Collections
Discover the best community collections!
Collections including paper arxiv:2304.15004
-
BioMedLM: A 2.7B Parameter Language Model Trained On Biomedical Text
Paper • 2403.18421 • Published • 22 -
Long-form factuality in large language models
Paper • 2403.18802 • Published • 24 -
stanford-crfm/BioMedLM
Text Generation • Updated • 2.86k • 395 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper • 2305.18290 • Published • 48
-
Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer
Paper • 1701.06538 • Published • 4 -
Attention Is All You Need
Paper • 1706.03762 • Published • 44 -
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks
Paper • 2005.11401 • Published • 12 -
Language Model Evaluation Beyond Perplexity
Paper • 2106.00085 • Published
-
Lost in the Middle: How Language Models Use Long Contexts
Paper • 2307.03172 • Published • 36 -
Efficient Estimation of Word Representations in Vector Space
Paper • 1301.3781 • Published • 6 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 14 -
Attention Is All You Need
Paper • 1706.03762 • Published • 44
-
Why think step by step? Reasoning emerges from the locality of experience
Paper • 2304.03843 • Published -
Are Emergent Abilities of Large Language Models a Mirage?
Paper • 2304.15004 • Published • 6 -
Knowledge Mechanisms in Large Language Models: A Survey and Perspective
Paper • 2407.15017 • Published • 33
-
Dissociating language and thought in large language models: a cognitive perspective
Paper • 2301.06627 • Published • 1 -
A Latent Space Theory for Emergent Abilities in Large Language Models
Paper • 2304.09960 • Published • 3 -
Are Emergent Abilities of Large Language Models a Mirage?
Paper • 2304.15004 • Published • 6 -
Do LLMs Really Adapt to Domains? An Ontology Learning Perspective
Paper • 2407.19998 • Published • 1