Saving weights and logs of epoch 100
Browse files
events.out.tfevents.1625853680.t1v-n-55481057-w-0.90285.3.v2
ADDED
Binary file (1.39 MB). View file
|
|
events.out.tfevents.1625857994.t1v-n-55481057-w-0.92040.3.v2
ADDED
Binary file (2.9 MB). View file
|
|
flax_model.msgpack
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1200715307
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:04bc3c643aa06fce3f104300b758f4421eeac2a267542a29923c58ca922ef108
|
3 |
size 1200715307
|
run_translation_t5_flax.py
CHANGED
@@ -431,13 +431,13 @@ def main():
|
|
431 |
# Setting padding="max_length" as we need fixed length inputs for jitted functions
|
432 |
def preprocess_function(examples):
|
433 |
#Make it two way
|
434 |
-
|
435 |
-
inputs = examples[text_column][0:int(len(examples)/2)]
|
436 |
-
inputs.extend(examples[summary_column][int(len(examples)/2):])
|
437 |
|
438 |
-
|
439 |
-
targets = examples[summary_column][0:int(len(examples)/2)]
|
440 |
-
targets.extend(examples[text_column][int(len(examples)/2):])
|
441 |
|
442 |
inputs = [prefix + inp for inp in inputs]
|
443 |
model_inputs = tokenizer(
|
|
|
431 |
# Setting padding="max_length" as we need fixed length inputs for jitted functions
|
432 |
def preprocess_function(examples):
|
433 |
#Make it two way
|
434 |
+
inputs = examples[text_column]
|
435 |
+
#inputs = examples[text_column][0:int(len(examples)/2)]
|
436 |
+
#inputs.extend(examples[summary_column][int(len(examples)/2):])
|
437 |
|
438 |
+
targets = examples[summary_column]
|
439 |
+
#targets = examples[summary_column][0:int(len(examples)/2)]
|
440 |
+
#targets.extend(examples[text_column][int(len(examples)/2):])
|
441 |
|
442 |
inputs = [prefix + inp for inp in inputs]
|
443 |
model_inputs = tokenizer(
|