Papers - Google
updated
Lumiere: A Space-Time Diffusion Model for Video Generation
Paper
• 2401.12945
• Published
• 87
Long-form factuality in large language models
Paper
• 2403.18802
• Published
• 26
ObjectDrop: Bootstrapping Counterfactuals for Photorealistic Object
Removal and Insertion
Paper
• 2403.18818
• Published
• 28
TC4D: Trajectory-Conditioned Text-to-4D Generation
Paper
• 2403.17920
• Published
• 18
Transforming and Combining Rewards for Aligning Large Language Models
Paper
• 2402.00742
• Published
• 12
Leverage the Average: an Analysis of KL Regularization in RL
Paper
• 2003.14089
• Published
• 2
Snap-it, Tap-it, Splat-it: Tactile-Informed 3D Gaussian Splatting for
Reconstructing Challenging Surfaces
Paper
• 2403.20275
• Published
• 10
Gecko: Versatile Text Embeddings Distilled from Large Language Models
Paper
• 2403.20327
• Published
• 48
Localizing Paragraph Memorization in Language Models
Paper
• 2403.19851
• Published
• 15
Streaming Dense Video Captioning
Paper
• 2404.01297
• Published
• 13
FormNetV2: Multimodal Graph Contrastive Learning for Form Document
Information Extraction
Paper
• 2305.02549
• Published
• 7
FormNet: Structural Encoding beyond Sequential Modeling in Form Document
Information Extraction
Paper
• 2203.08411
• Published
• 1
ETC: Encoding Long and Structured Inputs in Transformers
Paper
• 2004.08483
• Published
• 1
GQA: Training Generalized Multi-Query Transformer Models from Multi-Head
Checkpoints
Paper
• 2305.13245
• Published
• 6
Sparse Upcycling: Training Mixture-of-Experts from Dense Checkpoints
Paper
• 2212.05055
• Published
• 6
Exploring the Limits of Transfer Learning with a Unified Text-to-Text
Transformer
Paper
• 1910.10683
• Published
• 16
Bigger is not Always Better: Scaling Properties of Latent Diffusion
Models
Paper
• 2404.01367
• Published
• 22
Mixture-of-Depths: Dynamically allocating compute in transformer-based
language models
Paper
• 2404.02258
• Published
• 107
Training LLMs over Neurally Compressed Text
Paper
• 2404.03626
• Published
• 23
BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions
Paper
• 1905.10044
• Published
• 2
Program Induction by Rationale Generation : Learning to Solve and
Explain Algebraic Word Problems
Paper
• 1705.04146
• Published
• 1
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
Paper
• 2201.11903
• Published
• 15
Premise Order Matters in Reasoning with Large Language Models
Paper
• 2402.08939
• Published
• 28
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language
Understanding
Paper
• 1804.07461
• Published
• 4
Program of Thoughts Prompting: Disentangling Computation from Reasoning
for Numerical Reasoning Tasks
Paper
• 2211.12588
• Published
• 3
BERT Rediscovers the Classical NLP Pipeline
Paper
• 1905.05950
• Published
• 3
Prompt-to-Prompt Image Editing with Cross Attention Control
Paper
• 2208.01626
• Published
• 3
WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual
Machine Learning
Paper
• 2103.01913
• Published
• 2
No "Zero-Shot" Without Exponential Data: Pretraining Concept Frequency
Determines Multimodal Model Performance
Paper
• 2404.04125
• Published
• 29
Leave No Context Behind: Efficient Infinite Context Transformers with
Infini-attention
Paper
• 2404.07143
• Published
• 111
RecurrentGemma: Moving Past Transformers for Efficient Open Language
Models
Paper
• 2404.07839
• Published
• 48
PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive
Summarization
Paper
• 1912.08777
• Published
• 2
Leveraging Pre-trained Checkpoints for Sequence Generation Tasks
Paper
• 1907.12461
• Published
• 1
Phenaki: Variable Length Video Generation From Open Domain Textual
Description
Paper
• 2210.02399
• Published
• 3
Scaling Up Models and Data with t5x and seqio
Paper
• 2203.17189
• Published
• 1
Reuse Your Rewards: Reward Model Transfer for Zero-Shot Cross-Lingual
Alignment
Paper
• 2404.12318
• Published
• 15
Does Gaussian Splatting need SFM Initialization?
Paper
• 2404.12547
• Published
• 9
Batch Normalization: Accelerating Deep Network Training by Reducing
Internal Covariate Shift
Paper
• 1502.03167
• Published
• 2
Hyper-X: A Unified Hypernetwork for Multi-Task Multilingual Transfer
Paper
• 2205.12148
• Published
• 2
Capabilities of Gemini Models in Medicine
Paper
• 2404.18416
• Published
• 25
Stylus: Automatic Adapter Selection for Diffusion Models
Paper
• 2404.18928
• Published
• 15
DOCCI: Descriptions of Connected and Contrasting Images
Paper
• 2404.19753
• Published
• 13
What needs to go right for an induction head? A mechanistic study of
in-context learning circuits and their formation
Paper
• 2404.07129
• Published
• 3
Greedy Growing Enables High-Resolution Pixel-Based Diffusion Models
Paper
• 2405.16759
• Published
• 8
TACT: Advancing Complex Aggregative Reasoning with Information
Extraction Tools
Paper
• 2406.03618
• Published
• 2
Neural Autoregressive Distribution Estimation
Paper
• 1605.02226
• Published
• 1
Tx-LLM: A Large Language Model for Therapeutics
Paper
• 2406.06316
• Published
• 19
Classifier-Free Diffusion Guidance
Paper
• 2207.12598
• Published
• 4
Improve Mathematical Reasoning in Language Models by Automated Process
Supervision
Paper
• 2406.06592
• Published
• 29
Revisiting Unreasonable Effectiveness of Data in Deep Learning Era
Paper
• 1707.02968
• Published
• 1
Inception-v4, Inception-ResNet and the Impact of Residual Connections on
Learning
Paper
• 1602.07261
• Published
• 1
Data curation via joint example selection further accelerates multimodal
learning
Paper
• 2406.17711
• Published
• 3
TokenFormer: Rethinking Transformer Scaling with Tokenized Model
Parameters
Paper
• 2410.23168
• Published
• 24