WeiChow's picture
Upload folder using huggingface_hub
062c6c8 verified
{
"pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict",
"featuresDict": {
"features": {
"steps": {
"pythonClassName": "tensorflow_datasets.core.features.dataset_feature.Dataset",
"sequence": {
"feature": {
"pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict",
"featuresDict": {
"features": {
"observation": {
"pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict",
"featuresDict": {
"features": {
"joint_states": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {
"dimensions": [
"7"
]
},
"dtype": "float32",
"encoding": "none"
},
"description": "joint values"
},
"gripper_states": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {
"dimensions": [
"1"
]
},
"dtype": "float32",
"encoding": "none"
},
"description": "gripper_states = 0 means the gripper is fully closed. The value represents the gripper width of Franka Panda Gripper."
},
"ee_states": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {
"dimensions": [
"16"
]
},
"dtype": "float32",
"encoding": "none"
},
"description": "Pose of the end effector specified as a homogenous matrix."
},
"natural_language_instruction": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {},
"dtype": "string",
"encoding": "none"
}
},
"eye_in_hand_rgb": {
"pythonClassName": "tensorflow_datasets.core.features.image_feature.Image",
"image": {
"shape": {
"dimensions": [
"256",
"256",
"3"
]
},
"dtype": "uint8",
"encodingFormat": "jpeg"
},
"description": "RGB captured by in hand camera"
},
"agentview_rgb": {
"pythonClassName": "tensorflow_datasets.core.features.image_feature.Image",
"image": {
"shape": {
"dimensions": [
"256",
"256",
"3"
]
},
"dtype": "uint8",
"encodingFormat": "jpeg"
},
"description": "RGB captured by workspace camera"
},
"natural_language_embedding": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {
"dimensions": [
"512"
]
},
"dtype": "float32",
"encoding": "none"
}
}
}
}
},
"action": {
"pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict",
"featuresDict": {
"features": {
"world_vector": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {
"dimensions": [
"3"
]
},
"dtype": "float32",
"encoding": "none"
}
},
"gripper_closedness_action": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {},
"dtype": "float32",
"encoding": "none"
}
},
"rotation_delta": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {
"dimensions": [
"3"
]
},
"dtype": "float32",
"encoding": "none"
}
},
"terminate_episode": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {},
"dtype": "float32",
"encoding": "none"
}
}
}
}
},
"is_first": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {},
"dtype": "bool",
"encoding": "none"
}
},
"is_terminal": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {},
"dtype": "bool",
"encoding": "none"
}
},
"is_last": {
"pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
"tensor": {
"shape": {},
"dtype": "bool",
"encoding": "none"
}
},
"reward": {
"pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar",
"tensor": {
"shape": {},
"dtype": "float32",
"encoding": "none"
}
}
}
}
},
"length": "-1"
}
}
}
}
}