parquet-converter commited on
Commit
9936443
1 Parent(s): c73fd77

Update parquet files

Browse files
Files changed (4) hide show
  1. .gitattributes +0 -42
  2. README.md +0 -150
  3. default/poem-tweets-train.parquet +3 -0
  4. poem-tweets.py +0 -73
.gitattributes DELETED
@@ -1,42 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ftz filter=lfs diff=lfs merge=lfs -text
6
- *.gz filter=lfs diff=lfs merge=lfs -text
7
- *.h5 filter=lfs diff=lfs merge=lfs -text
8
- *.joblib filter=lfs diff=lfs merge=lfs -text
9
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
- *.model filter=lfs diff=lfs merge=lfs -text
11
- *.msgpack filter=lfs diff=lfs merge=lfs -text
12
- *.npy filter=lfs diff=lfs merge=lfs -text
13
- *.npz filter=lfs diff=lfs merge=lfs -text
14
- *.onnx filter=lfs diff=lfs merge=lfs -text
15
- *.ot filter=lfs diff=lfs merge=lfs -text
16
- *.parquet filter=lfs diff=lfs merge=lfs -text
17
- *.pb filter=lfs diff=lfs merge=lfs -text
18
- *.pickle filter=lfs diff=lfs merge=lfs -text
19
- *.pkl filter=lfs diff=lfs merge=lfs -text
20
- *.pt filter=lfs diff=lfs merge=lfs -text
21
- *.pth filter=lfs diff=lfs merge=lfs -text
22
- *.rar filter=lfs diff=lfs merge=lfs -text
23
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
24
- *.tar.* filter=lfs diff=lfs merge=lfs -text
25
- *.tflite filter=lfs diff=lfs merge=lfs -text
26
- *.tgz filter=lfs diff=lfs merge=lfs -text
27
- *.wasm filter=lfs diff=lfs merge=lfs -text
28
- *.xz filter=lfs diff=lfs merge=lfs -text
29
- *.zip filter=lfs diff=lfs merge=lfs -text
30
- *.zstandard filter=lfs diff=lfs merge=lfs -text
31
- *tfevents* filter=lfs diff=lfs merge=lfs -text
32
- # Audio files - uncompressed
33
- *.pcm filter=lfs diff=lfs merge=lfs -text
34
- *.sam filter=lfs diff=lfs merge=lfs -text
35
- *.raw filter=lfs diff=lfs merge=lfs -text
36
- # Audio files - compressed
37
- *.aac filter=lfs diff=lfs merge=lfs -text
38
- *.flac filter=lfs diff=lfs merge=lfs -text
39
- *.mp3 filter=lfs diff=lfs merge=lfs -text
40
- *.ogg filter=lfs diff=lfs merge=lfs -text
41
- *.wav filter=lfs diff=lfs merge=lfs -text
42
- *.csv filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md DELETED
@@ -1,150 +0,0 @@
1
- ---
2
- annotations_creators:
3
- - no-annotation
4
- language:
5
- - id
6
- language_creators:
7
- - found
8
- license:
9
- - cc-by-4.0
10
- multilinguality:
11
- - monolingual
12
- pretty_name: poem_tweets
13
- size_categories:
14
- - 10K<n<100K
15
- source_datasets:
16
- - original
17
- tags:
18
- - poem
19
- - tweets
20
- - twitter
21
- - indonesian
22
- task_categories:
23
- - text-generation
24
- task_ids:
25
- - language-modeling
26
- ---
27
-
28
- # Dataset Card for Poem Tweets
29
-
30
- ## Table of Contents
31
- - [Table of Contents](#table-of-contents)
32
- - [Dataset Description](#dataset-description)
33
- - [Dataset Summary](#dataset-summary)
34
- - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
35
- - [Languages](#languages)
36
- - [Dataset Structure](#dataset-structure)
37
- - [Data Instances](#data-instances)
38
- - [Data Fields](#data-fields)
39
- - [Data Splits](#data-splits)
40
- - [Dataset Creation](#dataset-creation)
41
- - [Curation Rationale](#curation-rationale)
42
- - [Source Data](#source-data)
43
- - [Annotations](#annotations)
44
- - [Personal and Sensitive Information](#personal-and-sensitive-information)
45
- - [Considerations for Using the Data](#considerations-for-using-the-data)
46
- - [Social Impact of Dataset](#social-impact-of-dataset)
47
- - [Discussion of Biases](#discussion-of-biases)
48
- - [Other Known Limitations](#other-known-limitations)
49
- - [Additional Information](#additional-information)
50
- - [Dataset Curators](#dataset-curators)
51
- - [Licensing Information](#licensing-information)
52
- - [Citation Information](#citation-information)
53
- - [Contributions](#contributions)
54
-
55
- ## Dataset Description
56
-
57
- - **Homepage:**
58
- - **Repository:**
59
- - **Paper:**
60
- - **Leaderboard:**
61
- - **Point of Contact:**
62
-
63
- ### Dataset Summary
64
-
65
- The data are from Twitter. The purpose of this data is to create text generation model for short text and make sure they are all coherence and rhythmic
66
-
67
- ### Supported Tasks and Leaderboards
68
-
69
- - Text Generation
70
- - Language Model
71
-
72
- ### Languages
73
-
74
- Indonesian
75
-
76
- ## Dataset Structure
77
-
78
- ### Data Instances
79
-
80
- [More Information Needed]
81
-
82
- ### Data Fields
83
-
84
- [More Information Needed]
85
-
86
- ### Data Splits
87
-
88
- [More Information Needed]
89
-
90
- ## Dataset Creation
91
-
92
- ### Curation Rationale
93
-
94
- [More Information Needed]
95
-
96
- ### Source Data
97
-
98
- #### Initial Data Collection and Normalization
99
-
100
- [More Information Needed]
101
-
102
- #### Who are the source language producers?
103
-
104
- [More Information Needed]
105
-
106
- ### Annotations
107
-
108
- #### Annotation process
109
-
110
- [More Information Needed]
111
-
112
- #### Who are the annotators?
113
-
114
- [More Information Needed]
115
-
116
- ### Personal and Sensitive Information
117
-
118
- [More Information Needed]
119
-
120
- ## Considerations for Using the Data
121
-
122
- ### Social Impact of Dataset
123
-
124
- [More Information Needed]
125
-
126
- ### Discussion of Biases
127
-
128
- [More Information Needed]
129
-
130
- ### Other Known Limitations
131
-
132
- [More Information Needed]
133
-
134
- ## Additional Information
135
-
136
- ### Dataset Curators
137
-
138
- [More Information Needed]
139
-
140
- ### Licensing Information
141
-
142
- [More Information Needed]
143
-
144
- ### Citation Information
145
-
146
- [More Information Needed]
147
-
148
- ### Contributions
149
-
150
- Thanks to [@andreaschandra](https://github.com/andreaschandra) for adding this dataset.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
default/poem-tweets-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:183ff13fcf436ab896b58303968858152dbb953aab2b0ec85b79a060bda9164d
3
+ size 1390689
poem-tweets.py DELETED
@@ -1,73 +0,0 @@
1
- # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
2
- #
3
- # Licensed under the Apache License, Version 2.0 (the "License");
4
- # you may not use this file except in compliance with the License.
5
- # You may obtain a copy of the License at
6
- #
7
- # http://www.apache.org/licenses/LICENSE-2.0
8
- #
9
- # Unless required by applicable law or agreed to in writing, software
10
- # distributed under the License is distributed on an "AS IS" BASIS,
11
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
- # See the License for the specific language governing permissions and
13
- # limitations under the License.
14
- # TODO: Address all TODOs and remove all explanatory comments
15
- """Poem Tweets: Indonesian Poem Tweets Generation Dataset."""
16
-
17
-
18
- import csv
19
- import json
20
- import os
21
-
22
- import datasets
23
-
24
- _DESCRIPTION = """\
25
- This dataset is built for text generation task in context of poem tweets in Bahasa.
26
- """
27
-
28
- _HOMEPAGE = "https://github.com/jakartaresearch"
29
-
30
- # TODO: Add link to the official dataset URLs here
31
- # The HuggingFace Datasets library doesn't host the datasets but only points to the original files.
32
- # This can be an arbitrary nested dict/list of URLs (see below in `_split_generators` method)
33
- _TRAIN_URL = "https://media.githubusercontent.com/media/jakartaresearch/hf-datasets/main/poem-tweets/poem-tweets/train.csv"
34
-
35
-
36
- # TODO: Name of the dataset usually match the script name with CamelCase instead of snake_case
37
- class Indonews(datasets.GeneratorBasedBuilder):
38
- """Poem Tweets: Indonesian Poem Tweets Generation Dataset."""
39
-
40
- VERSION = datasets.Version("1.0.0")
41
-
42
- def _info(self):
43
-
44
- features = datasets.Features(
45
- {
46
- "id": datasets.Value("int64"),
47
- "screen_name": datasets.Value("string"),
48
- "text": datasets.Value("string"),
49
- }
50
- )
51
-
52
- return datasets.DatasetInfo(
53
- description=_DESCRIPTION,
54
- features=features,
55
- homepage=_HOMEPAGE
56
- )
57
-
58
- def _split_generators(self, dl_manager):
59
-
60
- train_path = dl_manager.download_and_extract(_TRAIN_URL)
61
- return [
62
- datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": train_path}),
63
- ]
64
-
65
- # method parameters are unpacked from `gen_kwargs` as given in `_split_generators`
66
- def _generate_examples(self, filepath):
67
- """Generate examples."""
68
- with open(filepath, encoding="utf-8") as csv_file:
69
- csv_reader = csv.reader(csv_file, delimiter=",")
70
- next(csv_reader)
71
- for id_, row in enumerate(csv_reader):
72
- id_tweet, screen_name, text = row
73
- yield id_, {"id": id_tweet, "screen_name": screen_name, "text": text}