petals / utils_display.py
ybelkada's picture
Update utils_display.py
6d863c3
raw
history blame
1.89 kB
import time
import random
import streamlit as st
from example_prompts import EXAMPLE_PROMPTS
HEADER = """
# The online demo has now migrated to https://petals.ml/ and http://chat.petals.ml/ please refer to these links
"""
SIDE_BAR_TEXT = """
# *PETALS: A Collaborative Inference of Large Models*
A BigScience initiative.
- [Introduction](#introduction)
* [What is *PETALS* ?](#what-is--petals---)
* [Generation parameters](#generation-parameters)
# Introduction
This Space is an interactive Space of *PETALS* paper (Submitted in EMNLP 2022) that aims to run BLOOM-176 in a distributed manner for efficient and cost-effective inference and fine-tuning.
## What is *PETALS* ?
With the release of BLOOM-176B and OPT-175B, everyone can download pretrained models of this scale. Still, using these models requires supercomputer-grade hardware, which is unavailable to many researchers.
PETALS proposes to run BLOOM-176 in a distributed manner. The model is run on multiple computers from different users. Each user can benefit from the large model's inference by checking the official links: [petals](https://petals.ml/) | [chat-petals](http://chat.petals.ml/)
## Generation parameters
"""
def write_incremental(text, place_holder, delay=0.05):
"""
Write a text in a streamlit widget, one character at a time.
Adapted from: https://discuss.streamlit.io/t/display-several-pieces-of-strings-incrementally-on-the-same-line/9279
"""
for i in range(len(text) + 1):
place_holder.markdown("### <span style='color:grey' class='font'> %s </span>" % text[0:i].replace("\n", "<br>"), unsafe_allow_html=True)
# place_holder.markdown("#### %s" % text[0:i])
time.sleep(delay)
def i_am_feeling_lucky():
"""
Return a random prompt from EXAMPLE_PROMPT
"""
return EXAMPLE_PROMPTS[random.randint(0, len(EXAMPLE_PROMPTS) - 1)]