haixuantao commited on
Commit
37584e0
1 Parent(s): 10e96ef

Simplify planning node

Browse files
graphs/dataflow.yml CHANGED
@@ -4,13 +4,13 @@ nodes:
4
  python: ../operators/robot.py
5
  inputs:
6
  blaster:
7
- source: mistral/blaster
8
  queue_size: 1
9
  led:
10
- source: mistral/led
11
  queue_size: 1
12
  control:
13
- source: mistral/control
14
  queue_size: 1
15
  gimbal_control:
16
  source: planning/gimbal_control
@@ -59,9 +59,12 @@ nodes:
59
  python: ../operators/plot.py
60
  inputs:
61
  image: bot_webcam/image
62
- text: whisper/text
 
 
63
  bbox: object_detection/bbox
64
-
 
65
  - id: planning
66
  operator:
67
  python: ../operators/planning_op.py
@@ -78,23 +81,22 @@ nodes:
78
  ## Speech to text
79
  - id: keyboard
80
  custom:
81
- source: ../operators/keybinding_op.py
82
  outputs:
83
- - mic_on
84
- - revert
85
- - failed
86
- - error
 
 
87
  inputs:
88
- tick:
89
- source: dora/timer/millis/50
90
- queue_size: 1
91
-
92
 
93
  - id: microphone
94
  operator:
95
  python: ../operators/microphone_op.py
96
  inputs:
97
- mic_on: keyboard/mic_on
98
  outputs:
99
  - audio
100
 
@@ -111,39 +113,32 @@ nodes:
111
  operator:
112
  python: ../operators/sentence_transformers_op.py
113
  inputs:
114
- query: whisper/text
115
  saved_file: file_saver/saved_file
116
  outputs:
117
  - raw_file
118
 
119
- - id: mistral
120
  operator:
121
- python: ../operators/mistral_op.py
122
  inputs:
123
- raw_file: vectordb/raw_file
 
 
124
  outputs:
125
- - output_file
 
 
126
  - control
127
  - led
128
  - blaster
 
129
 
130
- - id: chatgpt
131
- operator:
132
- python: ../operators/chatgpt_op.py
133
- inputs:
134
- raw_file: vectordb/raw_file
135
- outputs:
136
- - output_file
137
-
138
  - id: file_saver
139
  operator:
140
  python: ../operators/file_saver_op.py
141
  inputs:
142
- chatgpt_output_file: chatgpt/output_file
143
- mistral_output_file: mistral/output_file
144
- error: keyboard/error
145
- revert: keyboard/revert
146
- failed: keyboard/failed
147
  outputs:
148
  - saved_file
149
 
@@ -151,8 +146,7 @@ nodes:
151
  custom:
152
  source: dora-record
153
  inputs:
154
- chatgpt_output_file: chatgpt/output_file
155
- mistral_output_file: mistral/output_file
156
  raw_file: vectordb/raw_file
157
  saved_file: file_saver/saved_file
158
  audio: microphone/audio
@@ -170,4 +164,4 @@ nodes:
170
  inputs:
171
  image: bot_webcam/image
172
  outputs:
173
- - encoded_image
 
4
  python: ../operators/robot.py
5
  inputs:
6
  blaster:
7
+ source: llm/blaster
8
  queue_size: 1
9
  led:
10
+ source: llm/led
11
  queue_size: 1
12
  control:
13
+ source: planning/control
14
  queue_size: 1
15
  gimbal_control:
16
  source: planning/gimbal_control
 
59
  python: ../operators/plot.py
60
  inputs:
61
  image: bot_webcam/image
62
+ keyboard_buffer: keyboard/buffer
63
+ user_message: keyboard/submitted
64
+ assistant_message: llm/assistant_message
65
  bbox: object_detection/bbox
66
+
67
+
68
  - id: planning
69
  operator:
70
  python: ../operators/planning_op.py
 
81
  ## Speech to text
82
  - id: keyboard
83
  custom:
84
+ source: ../operators/keyboard_op.py
85
  outputs:
86
+ - buffer
87
+ - submitted
88
+ - record
89
+ - ask
90
+ - send
91
+ - change
92
  inputs:
93
+ recording: whisper/text
 
 
 
94
 
95
  - id: microphone
96
  operator:
97
  python: ../operators/microphone_op.py
98
  inputs:
99
+ record: keyboard/record
100
  outputs:
101
  - audio
102
 
 
113
  operator:
114
  python: ../operators/sentence_transformers_op.py
115
  inputs:
116
+ query: keyboard/change
117
  saved_file: file_saver/saved_file
118
  outputs:
119
  - raw_file
120
 
121
+ - id: llm
122
  operator:
123
+ python: ../operators/llm_op.py
124
  inputs:
125
+ code_modifier: vectordb/raw_file
126
+ assistant: keyboard/ask
127
+ message_sender: keyboard/send
128
  outputs:
129
+ - modified_file
130
+ - assistant_message
131
+ - line
132
  - control
133
  - led
134
  - blaster
135
+ - rotation
136
 
 
 
 
 
 
 
 
 
137
  - id: file_saver
138
  operator:
139
  python: ../operators/file_saver_op.py
140
  inputs:
141
+ llm_output_file: llm/modified_file
 
 
 
 
142
  outputs:
143
  - saved_file
144
 
 
146
  custom:
147
  source: dora-record
148
  inputs:
149
+ llm_output_file: llm/modified_file
 
150
  raw_file: vectordb/raw_file
151
  saved_file: file_saver/saved_file
152
  audio: microphone/audio
 
164
  inputs:
165
  image: bot_webcam/image
166
  outputs:
167
+ - encoded_image
operators/keyboard_op.py CHANGED
@@ -2,8 +2,6 @@ from pynput import keyboard
2
  from pynput.keyboard import Key, Events
3
  import pyarrow as pa
4
  from dora import Node
5
- from tkinter import Tk
6
- import tkinter as tk
7
 
8
 
9
  node = Node()
@@ -30,35 +28,8 @@ with keyboard.Events() as events:
30
  if event is not None and isinstance(event, Events.Press):
31
  if hasattr(event.key, "char"):
32
  cursor = 0
33
- if ctrl and event.key.char == "v":
34
- r = Tk()
35
- r.update()
36
- try:
37
- selection = r.clipboard_get()
38
- r.withdraw()
39
- r.update()
40
- except tk.TclError:
41
- selection = ""
42
- r.destroy()
43
- buffer_text += selection
44
- node.send_output("buffer", pa.array([buffer_text]))
45
- elif ctrl and event.key.char == "c":
46
- r = Tk()
47
- r.clipboard_clear()
48
- r.clipboard_append(buffer_text)
49
- r.update()
50
- r.destroy()
51
- elif ctrl and event.key.char == "x":
52
- r = Tk()
53
- r.clipboard_clear()
54
- r.clipboard_append(buffer_text)
55
- r.update()
56
- r.destroy()
57
- buffer_text = ""
58
- node.send_output("buffer", pa.array([buffer_text]))
59
- else:
60
- buffer_text += event.key.char
61
- node.send_output("buffer", pa.array([buffer_text]))
62
  else:
63
  if event.key == Key.backspace:
64
  buffer_text = buffer_text[:-1]
 
2
  from pynput.keyboard import Key, Events
3
  import pyarrow as pa
4
  from dora import Node
 
 
5
 
6
 
7
  node = Node()
 
28
  if event is not None and isinstance(event, Events.Press):
29
  if hasattr(event.key, "char"):
30
  cursor = 0
31
+ buffer_text += event.key.char
32
+ node.send_output("buffer", pa.array([buffer_text]))
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  else:
34
  if event.key == Key.backspace:
35
  buffer_text = buffer_text[:-1]
operators/llm_op.py CHANGED
@@ -13,7 +13,7 @@ MODEL_NAME_OR_PATH = "TheBloke/deepseek-coder-6.7B-instruct-GPTQ"
13
 
14
  CODE_MODIFIER_TEMPLATE = """
15
  ### Instruction
16
- Respond with the whole modified code only in ```python block. No explaination.
17
 
18
  ```python
19
  {code}
@@ -318,19 +318,19 @@ if __name__ == "__main__":
318
  # Directory of the current file
319
  current_directory = os.path.dirname(current_file_path)
320
 
321
- path = current_directory + "object_detection.py"
322
  with open(path, "r", encoding="utf8") as f:
323
  raw = f.read()
324
 
325
  op.on_event(
326
  {
327
  "type": "INPUT",
328
- "id": "message_sender",
329
  "value": pa.array(
330
  [
331
  {
332
  "path": path,
333
- "user_message": "send a star ",
334
  },
335
  ]
336
  ),
 
13
 
14
  CODE_MODIFIER_TEMPLATE = """
15
  ### Instruction
16
+ Respond with the modified code only in ```python block. No explaination.
17
 
18
  ```python
19
  {code}
 
318
  # Directory of the current file
319
  current_directory = os.path.dirname(current_file_path)
320
 
321
+ path = current_directory + "/planning_op.py"
322
  with open(path, "r", encoding="utf8") as f:
323
  raw = f.read()
324
 
325
  op.on_event(
326
  {
327
  "type": "INPUT",
328
+ "id": "code_modifier",
329
  "value": pa.array(
330
  [
331
  {
332
  "path": path,
333
+ "user_message": "change planning to make gimbal follow bounding box ",
334
  },
335
  ]
336
  ),
operators/object_detection.py CHANGED
@@ -27,8 +27,7 @@ class Operator:
27
  dora_event["value"].to_numpy().reshape((CAMERA_HEIGHT, CAMERA_WIDTH, 3))
28
  )
29
  frame = frame[:, :, ::-1] # OpenCV image (BGR to RGB)
30
- results = self.model(frame, verbose=False) # includes NMS
31
- # Process results for "ABC" only
32
  boxes = np.array(results[0].boxes.xyxy.cpu())
33
  conf = np.array(results[0].boxes.conf.cpu())
34
  label = np.array(results[0].boxes.cls.cpu())
 
27
  dora_event["value"].to_numpy().reshape((CAMERA_HEIGHT, CAMERA_WIDTH, 3))
28
  )
29
  frame = frame[:, :, ::-1] # OpenCV image (BGR to RGB)
30
+ results = model(frame, verbose=False) # includes NMS
 
31
  boxes = np.array(results[0].boxes.xyxy.cpu())
32
  conf = np.array(results[0].boxes.conf.cpu())
33
  label = np.array(results[0].boxes.cls.cpu())
operators/planning_op.py CHANGED
@@ -3,22 +3,10 @@ import numpy as np
3
  import pyarrow as pa
4
  from dora import DoraStatus
5
 
6
- # front-back: [-1, 1]
7
- X = 0
8
- # left-right: [-1,1]
9
- Y = 0
10
- SPEED = 0.5
11
- Z = 0
12
- # pitch-axis angle in degrees(int): [-55, 55]
13
- PITCH = 0
14
- # yaw-axis angle in degrees(int): [-55, 55]
15
- ROTATION = 0
16
- # RGB LED(int) [0, 255]
17
- RGB = [0, 0, 0]
18
- BRIGHTNESS = [0] # [0, 128]
19
-
20
- GOAL_OBJECTIVES = [X, Y, Z]
21
- GIMBAL_POSITION_GOAL = [PITCH, ROTATION]
22
 
23
  CAMERA_WIDTH = 960
24
  CAMERA_HEIGHT = 540
@@ -26,7 +14,7 @@ CAMERA_HEIGHT = 540
26
 
27
  class Operator:
28
  def __init__(self):
29
- self.bboxs = []
30
  self.time = time.time()
31
  self.position = [0, 0, 0]
32
 
@@ -35,35 +23,34 @@ class Operator:
35
  dora_event: dict,
36
  send_output,
37
  ) -> DoraStatus:
38
- global X, Y, SPEED, PITCH, ROTATION, RGB, BRIGHTNESS, GOAL_OBJECTIVES, GIMBAL_POSITION_GOAL
39
  if dora_event["type"] == "INPUT":
40
- if dora_event["id"] == "tick":
 
41
  self.time = time.time()
42
- elif dora_event["id"] == "bbox":
43
- bboxs = dora_event["value"].to_numpy()
44
- self.bboxs = np.reshape(bboxs, (-1, 6))
45
-
46
- elif dora_event["id"] == "position":
47
- [x, y, z, gimbal_pitch, gimbal_yaw] = dora_event["value"].to_numpy()
48
- self.position = [x, y, z]
 
 
 
49
  direction = np.clip(
50
- np.array(GOAL_OBJECTIVES) - np.array(self.position), -1, 1
51
  )
52
 
53
- if any(abs(direction) > 0.1):
54
- x = direction[0]
55
- y = direction[1]
56
- z = direction[2]
57
-
58
- send_output(
59
- "control",
60
- pa.array([x, y, z]),
61
- dora_event["metadata"],
62
- )
63
 
64
  send_output(
65
  "gimbal_control",
66
- pa.array([PITCH, ROTATION]),
67
  dora_event["metadata"],
68
  )
69
 
 
3
  import pyarrow as pa
4
  from dora import DoraStatus
5
 
6
+ # front-back and left-right movement(float) with min: -1 and max: 1
7
+ POSITION_GOAL = [0, 0]
8
+ # pitch and yaw axis angle in degrees(int) with min: -55 and max: 55
9
+ GIMBAL_GOAL = [0, 0]
 
 
 
 
 
 
 
 
 
 
 
 
10
 
11
  CAMERA_WIDTH = 960
12
  CAMERA_HEIGHT = 540
 
14
 
15
  class Operator:
16
  def __init__(self):
17
+ self.bboxs = np.array([])
18
  self.time = time.time()
19
  self.position = [0, 0, 0]
20
 
 
23
  dora_event: dict,
24
  send_output,
25
  ) -> DoraStatus:
26
+ global POSITION_GOAL, GIMBAL_GOAL
27
  if dora_event["type"] == "INPUT":
28
+ id = dora_event["id"]
29
+ if id == "tick":
30
  self.time = time.time()
31
+ elif id == "bbox":
32
+ value = dora_event["value"].to_numpy()
33
+ bboxs = value
34
+ self.bboxs = np.reshape(
35
+ bboxs, (-1, 6)
36
+ ) # min_x, min_y, max_x, max_y, confidence, label
37
+ elif id == "position":
38
+ value = dora_event["value"].to_numpy()
39
+ [x, y, _pitch, _yaw] = value
40
+ self.position = [x, y]
41
  direction = np.clip(
42
+ np.array(POSITION_GOAL) - np.array(self.position), -1, 1
43
  )
44
 
45
+ send_output(
46
+ "control",
47
+ pa.array(direction),
48
+ dora_event["metadata"],
49
+ )
 
 
 
 
 
50
 
51
  send_output(
52
  "gimbal_control",
53
+ pa.array(GIMBAL_GOAL),
54
  dora_event["metadata"],
55
  )
56
 
operators/plot.py CHANGED
@@ -63,17 +63,30 @@ class Operator:
63
  (0, 255, 0),
64
  )
65
 
66
- cv2.putText(image, self.buffer, (20, 12 * 25), FONT, 1, (190, 250, 0))
 
 
67
 
68
- for i, text in enumerate(self.submitted[::-1]):
 
 
 
 
 
 
69
  cv2.putText(
70
  image,
71
  text["content"],
72
- (20, 25 + (10 - i) * 25),
 
 
 
73
  FONT,
 
 
74
  1,
75
- (0, 255, 190),
76
  )
 
77
 
78
  for line in self.lines:
79
  cv2.line(
 
63
  (0, 255, 0),
64
  )
65
 
66
+ cv2.putText(
67
+ image, self.buffer, (20, 14 + 21 * 14), FONT, 0.5, (190, 250, 0), 1
68
+ )
69
 
70
+ i = 0
71
+ for text in self.submitted[::-1]:
72
+ color = (
73
+ (0, 255, 190)
74
+ if text["role"] == "user_message"
75
+ else (0, 190, 255)
76
+ )
77
  cv2.putText(
78
  image,
79
  text["content"],
80
+ (
81
+ 20,
82
+ 14 + (19 - i) * 14,
83
+ ),
84
  FONT,
85
+ 0.5,
86
+ color,
87
  1,
 
88
  )
89
+ i += 1
90
 
91
  for line in self.lines:
92
  cv2.line(
operators/robot.py CHANGED
@@ -1,16 +1,11 @@
1
- #!/usr/bin/env python3
2
- # -*- coding: utf-8 -*-
3
-
4
  from robomaster import robot, blaster, led
5
  from typing import Callable, Optional, Union
6
 
7
- # from robot import RobotController
8
  import pyarrow as pa
9
 
10
  from dora import DoraStatus
11
 
12
- # Global variables, change it to adapt your needs
13
- FREQ = 20
14
  CONN = "ap"
15
 
16
 
@@ -24,7 +19,7 @@ class Operator:
24
  ), "Could not start video stream"
25
 
26
  self.ep_robot.gimbal.recenter().wait_for_completed()
27
- self.position = [0, 0, 0]
28
  self.gimbal_position = [0, 0]
29
  self.event = None
30
  self.brightness = 0
@@ -49,13 +44,15 @@ class Operator:
49
  self.event is not None
50
  and not (self.event._event.isSet() and self.event.is_completed)
51
  ):
52
- [x, y, z] = dora_event["value"].to_numpy()
53
- self.event = self.ep_robot.chassis.move(
54
- x=x, y=y, z=z, xy_speed=0.8, z_speed=0.8
55
- )
56
- self.position[0] += x
57
- self.position[1] += y
58
- self.position[2] += z
 
 
59
  else:
60
  print("control not completed", flush=True)
61
  print("Set: ", self.event._event.isSet(), flush=True)
@@ -75,8 +72,8 @@ class Operator:
75
  self.event = self.ep_robot.gimbal.moveto(
76
  pitch=gimbal_pitch,
77
  yaw=gimbal_yaw,
78
- pitch_speed=20,
79
- yaw_speed=20,
80
  )
81
  self.gimbal_position[0] = gimbal_pitch
82
  self.gimbal_position[1] = gimbal_yaw
 
 
 
 
1
  from robomaster import robot, blaster, led
2
  from typing import Callable, Optional, Union
3
 
4
+ import numpy as np
5
  import pyarrow as pa
6
 
7
  from dora import DoraStatus
8
 
 
 
9
  CONN = "ap"
10
 
11
 
 
19
  ), "Could not start video stream"
20
 
21
  self.ep_robot.gimbal.recenter().wait_for_completed()
22
+ self.position = [0, 0]
23
  self.gimbal_position = [0, 0]
24
  self.event = None
25
  self.brightness = 0
 
44
  self.event is not None
45
  and not (self.event._event.isSet() and self.event.is_completed)
46
  ):
47
+ [x, y] = dora_event["value"].to_numpy()
48
+
49
+ if any(abs(np.array([x, y])) > 0.1):
50
+ print("received command: ", x, y, flush=True)
51
+ self.event = self.ep_robot.chassis.move(
52
+ x=x, y=y, z=0.0, xy_speed=0.8, z_speed=0.8
53
+ )
54
+ self.position[0] += x
55
+ self.position[1] += y
56
  else:
57
  print("control not completed", flush=True)
58
  print("Set: ", self.event._event.isSet(), flush=True)
 
72
  self.event = self.ep_robot.gimbal.moveto(
73
  pitch=gimbal_pitch,
74
  yaw=gimbal_yaw,
75
+ pitch_speed=50,
76
+ yaw_speed=50,
77
  )
78
  self.gimbal_position[0] = gimbal_pitch
79
  self.gimbal_position[1] = gimbal_yaw