Tutorial
Table Of Contents
- 1. Premise
- 2. Raspberry Pi System Installation and Developmen
- 3 Log In to The Raspberry Pi and Install The App
- 4 Assembly and Precautions
- 5 Controlling Robot via WEB App
- 6 Common Problems and Solutions(Q&A)
- 7 Set The Program to Start Automatically
- 8 Remote Operation of Raspberry Pi Via MobaXterm
- 9 How to Control WS2812 RGB LED
- 10 How to Control The Servo
- 11 How to Control DC Motor
- 12 Ultrasonic Module
- 13 Line Tracking
- 14 Make A Police Light or Breathing Light
- 15 Real-Time Video Transmission
- 16 Automatic Obstacle Avoidance
- 17 Why OpenCV Uses Multi-threading to Process Vide
- 18 OpenCV Learn to Use OpenCV
- 19 Using OpenCV to Realize Color Recognition and T
- 20 Machine Line Tracking Based on OpenCV
- 21 Create A WiFi Hotspot on The Raspberry Pi
- 22 Install GUI Dependent Item under Window
- 23 How to Use GUI
- 24 Control The WS2812 LED via GUI
- 25 Real-time Video Transmission Based on OpenCV
- 26 Use OpenCV to Process Video Frames on The PC
- 27 Enable UART
- 28 Control Your AWR with An Android Device
- Conclusion
115
'''
Clear the stream in preparation for the next frame
'''
rawCapture.truncate(0)
● In the following, we explain the program on the receiving end. Since the libraries used here are
cross-platform, PC.py can be run on a Windows computer or another Linux computer.
●PC.py :
'''
First import the required libraries
'''
import cv2
import zmq
import base64
import numpy as np
'''
Here we instantiate the zmq object used to receive the frame
Note that the port number needs to be consistent with the sender's
'''
context = zmq.Context()
footage_socket = context.socket(zmq.PAIR)
footage_socket.bind('tcp://*:5555')
while True:
'''
Received video frame data
'''
frame = footage_socket.recv_string()
'''
Decode and save it to the cache
'''
img = base64.b64decode(frame)
'''
Interpret a buffer as a 1-dimensional array
'''
npimg = np.frombuffer(img, dtype=np.uint8)
'''
Decode a one-dimensional array into an image
'''
source = cv2.imdecode(npimg, 1)