AutoPresent: Designing Structured Visuals from Scratch
Abstract
Designing structured visuals such as presentation slides is essential for communicative needs, necessitating both content creation and visual planning skills. In this work, we tackle the challenge of automated slide generation, where models produce slide presentations from natural language (NL) instructions. We first introduce the SlidesBench benchmark, the first benchmark for slide generation with 7k training and 585 testing examples derived from 310 slide decks across 10 domains. SlidesBench supports evaluations that are (i)reference-based to measure similarity to a target slide, and (ii)reference-free to measure the design quality of generated slides alone. We benchmark end-to-end image generation and program generation methods with a variety of models, and find that programmatic methods produce higher-quality slides in user-interactable formats. Built on the success of program generation, we create AutoPresent, an 8B Llama-based model trained on 7k pairs of instructions paired with code for slide generation, and achieve results comparable to the closed-source model GPT-4o. We further explore iterative design refinement where the model is tasked to self-refine its own output, and we found that this process improves the slide's quality. We hope that our work will provide a basis for future work on generating structured visuals.
Community
AutoPresent: Designing Structured Visuals from Scratch
GitHub: https://github.com/para-lost/AutoPresent
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- SparrowVQE: Visual Question Explanation for Course Content Understanding (2024)
- From Words to Structured Visuals: A Benchmark and Framework for Text-to-Diagram Generation and Editing (2024)
- Conditional Text-to-Image Generation with Reference Guidance (2024)
- GPTDrawer: Enhancing Visual Synthesis through ChatGPT (2024)
- ShowHowTo: Generating Scene-Conditioned Step-by-Step Visual Instructions (2024)
- Art-Free Generative Models: Art Creation Without Graphic Art Knowledge (2024)
- GATE OpenING: A Comprehensive Benchmark for Judging Open-ended Interleaved Image-Text Generation (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper