{ "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict", "featuresDict": { "features": { "steps": { "pythonClassName": "tensorflow_datasets.core.features.dataset_feature.Dataset", "sequence": { "feature": { "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict", "featuresDict": { "features": { "reward": { "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar", "tensor": { "shape": {}, "dtype": "float32", "encoding": "none" }, "description": "Reward if provided, 1 on final step for demos." }, "language_embedding": { "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", "tensor": { "shape": { "dimensions": [ "512" ] }, "dtype": "float32", "encoding": "none" }, "description": "Kona language embedding. See https://tfhub.dev/google/universal-sentence-encoder-large/5" }, "is_terminal": { "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", "tensor": { "shape": {}, "dtype": "bool", "encoding": "none" } }, "is_last": { "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", "tensor": { "shape": {}, "dtype": "bool", "encoding": "none" } }, "language_instruction": { "pythonClassName": "tensorflow_datasets.core.features.text_feature.Text", "text": {}, "description": "Language Instruction." }, "observation": { "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict", "featuresDict": { "features": { "wrist_image": { "pythonClassName": "tensorflow_datasets.core.features.image_feature.Image", "image": { "shape": { "dimensions": [ "256", "256", "3" ] }, "dtype": "uint8", "encodingFormat": "jpeg" }, "description": "Wrist camera RGB observation." }, "image": { "pythonClassName": "tensorflow_datasets.core.features.image_feature.Image", "image": { "shape": { "dimensions": [ "256", "256", "3" ] }, "dtype": "uint8", "encodingFormat": "jpeg" }, "description": "Main camera RGB observation." }, "state": { "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", "tensor": { "shape": { "dimensions": [ "13" ] }, "dtype": "float32", "encoding": "none" }, "description": "Robot joints state, consists of [6x robot joint angles, 1x gripper open status, 6x robot joint velocities]." }, "end_effector_state": { "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", "tensor": { "shape": { "dimensions": [ "7" ] }, "dtype": "float32", "encoding": "none" }, "description": "Robot gripper end effector state, consists of [x, y, z] and 4x quaternion" } } } }, "is_first": { "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", "tensor": { "shape": {}, "dtype": "bool", "encoding": "none" } }, "discount": { "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar", "tensor": { "shape": {}, "dtype": "float32", "encoding": "none" }, "description": "Discount if provided, default to 1." }, "action": { "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", "tensor": { "shape": { "dimensions": [ "6" ] }, "dtype": "float32", "encoding": "none" }, "description": "Robot action, consists of [dx, dy, dz] and [droll, dpitch, dyaw]" } } } }, "length": "-1" } }, "episode_metadata": { "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict", "featuresDict": { "features": { "file_path": { "pythonClassName": "tensorflow_datasets.core.features.text_feature.Text", "text": {}, "description": "Path to the original data file." } } } } } } }