Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -103,7 +103,7 @@ elif tokenizer == "Qwen2.5 Tokenizer":
|
|
103 |
with st.expander("About Qwen2.5 Tokenizer"):
|
104 |
st.write('')
|
105 |
ids = qwen_tokenizer.encode(txt)
|
106 |
-
split_tokens = qwen_tokenizer.
|
107 |
st.write(stream_data)
|
108 |
if token_id == True:
|
109 |
color = itertools.cycle(colors)
|
@@ -113,7 +113,7 @@ elif tokenizer == "RuAdapt Tokenizer":
|
|
113 |
with st.expander("About RuAdapt Tokenizer"):
|
114 |
st.write('')
|
115 |
ids = ruadapt_tokenizer.encode(txt)
|
116 |
-
split_tokens = ruadapt_tokenizer.
|
117 |
st.write(stream_data)
|
118 |
if token_id == True:
|
119 |
color = itertools.cycle(colors)
|
|
|
103 |
with st.expander("About Qwen2.5 Tokenizer"):
|
104 |
st.write('')
|
105 |
ids = qwen_tokenizer.encode(txt)
|
106 |
+
split_tokens = [qwen_tokenizer.decode([t]) for t in ids]
|
107 |
st.write(stream_data)
|
108 |
if token_id == True:
|
109 |
color = itertools.cycle(colors)
|
|
|
113 |
with st.expander("About RuAdapt Tokenizer"):
|
114 |
st.write('')
|
115 |
ids = ruadapt_tokenizer.encode(txt)
|
116 |
+
split_tokens = [ruadapt_tokenizer.decode([t]) for t in ids]
|
117 |
st.write(stream_data)
|
118 |
if token_id == True:
|
119 |
color = itertools.cycle(colors)
|