999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

Design of humanoid robot shooting system based on OpenMv

2020-07-24 05:40:14GuochenNIUTongZHU
機床與液壓 2020年12期

Guo-chen NIU, Tong ZHU

(College of Electronic Information and Automation, Civil Aviation University of China, Tianjin 300300, China)

Abstract: Aiming at the shortcomings of long time and low efficiency of humanoid robot images using linear CCD as image acquisition module, a kind of humanoid robot based on OpenMv is designed. The design of the humanoid robot shooting system completes the hardware and software design firstly, then converts the acquired image into the HSV color model, uses the image segmentation technology to segment the target and the background, and then calculates the centroid coordinates. Finally, the corresponding signal is output through the expert control strategy. Adjust the movement of the robot. Experiments show that the system has high reliability and adaptability, and won the first prize in the robot competition of the five provinces (municipalities and autonomous regions) in North China, which verifies the applicability, reliability and effectiveness of the method.

Key words: OpenMv, Humanoid robot, Image segmentation, Target recognition, Expert control

1 Introduction

Humanoid robots are favored by researchers because of their wide range of needs in the fields of medical care, entertainment and services, and as an ideal platform for evaluating multiple discipline technologies such as artificial intelligence. It is one of the hot topics in the field of robotics research. At the same time, compared to other types of robots wheeled, tracked and peristaltic type, which may be better adapted to the human living environment, work space broader, richer form of action, and its higher energy efficiency. So the research has important application value[1]. This paper designs a kind of human shooting robot for the robot competition of the five provinces of North China. The robot competition venue material is made of white solid wood particle board. The project is that the robot moves the ball (10 cm in diameter) from the starting point to the basket 50 cm away from the starting point. The ball is successfully grabbed and placed in the pitching area (the basket is 0.7m or less). For the pitching area put the ball into the basket.

Environmental perception is the premise and basis for humanoid robots to have the ability to make decisions about autonomous behavior[2]. Humanoid robots generally use cameras to obtain environmental information. Due to the computing power of the processor, the size and weight of the camera can affect the performance of the camera used. At present, the color block segmentation method is often used for target recognition, and there are problems such as inaccurate target recognition and large environmental interference. In terms of target positioning, Jamzad et al. [3] proposed a geometric positioning model of the target. Since the robot is a wheeled mobile robot and the camera poses fixed, the model is simple and effective, so it is susceptible to the surrounding environment. Bandlow [4] obtains the positioning of the robot itself and the positioning of the target object by recognizing the relative positional relationship between the fixed identification object and the target object, but the positioning accuracy is not high when the fixed identification object is far away from the robot, so the target object is The positioning error is large.

Therefore, this paper designs a humanoid robot hardware and software system based on OpenMv. The image segmentation algorithm based on pixel threshold algorithm and image region based algorithm is used to calculate the target coordinates through the imaging relationship between the target and the camera, which improves the shooting accuracy of the shooting robot. At the end of the paper, the shooting experiment results of humanoid robot are given.

2 Hardware system design

The humanoid shooting robot adopts a modular design method, which is convenient for experimental testing and maintenance. The main hardware system should include power module, visual OpenMv module, servo drive module and communication module, etc. The block diagram of the structure is shown in Fig.1.

Fig.1 System block diagram

The functions of each module are as follows:

Power module: Adopts dual battery power supply mode, the servo drive module and the visual processing module are respectively powered by different batteries to ensure the stability of the control signal.

Visual OpenMv module: OpenMv is an open source, low cost, and powerful machine vision module. It is based on the STM32F427 and integrates the OV7725 camera chip.This module can processes the data through image processing, and the processing speed is faster.

Steering gear: The robot joints can realize different walking postures by combining angle and speed.

Communication module: Bluetooth is a device for short-distance wireless data exchange. The Bluetooth module uses HC-05[5], which has embedded Bluetooth protocol. The operating parameters such as the relative position and threshold of the robot and the guide line can be displayed in real time, which facilitates debugging and improvement of related algorithms.

3 Software system design

When the robot control system is running, the robot posture, OpenMv, etc. are initialized first. Then, gather the field information through OpenMv, convert the collected image into the HSV color model, use image segmentation technology to segment the target and the background, and then calculate the centroid coordinates. Finally, through the expert control strategy output, the corresponding signal is obtained to adjust the robot’s movement, so as to achieve the offensive and defensive movement of the shooting robot. The system software flow chart is shown in Fig.2.

Fig.2 System main program flow chart

3.1 Color space model conversion

The RGB color space is susceptible to changes in light intensity, and the correlation between R, G, and B values is very high. For identifying a specific color, it is difficult to determine the threshold and the distribution range in space, which is not conducive to the target object. The segmentation and identification are easy to cause misjudgment, resulting in inaccurate final identification. In order to solve this problem, it is convenient to color processing and recognition, and decided to adopt the HSV color model[6]. The HSV color model has two characteristics: First, the luminance components are independent of color information. They can be processed separately and independent of each other. The change of this component does not affect the color information of other images. Second, the HSV color space is more in line with human visual perception. The perceptual characteristics of the outside world are conducive to image processing.

Fig.3 Conical space model

Fig.4 Tone angle coordinates

The conversion formula from RGB color space to HSV color space is shown in (1).

(1)

Among them

(2)

3.2 Target search

After the robot gets the search command, it performs the in situ search first. Since the camera is mounted on the robot head, it is possible to perform a comprehensive search in the visible direction by turning the head servo. In order to ensure that the target is in the central area of the field of view, it is necessary to determine the position of the target in the field of view. When the target is on the left side of the field of view, the subtraction is performed, and the robot is turned left. When the target is on the right side of the field of view, the addition is performed, and the robot is right. When the target appears in the central area, the target is in the central area of the field of view, and the robot goes straight. In this way, the target is always in the central area of the robot’s field of view. The target search is mainly used to roughly observe the position of the target object, and realize the positioning of the target by the left turn, the right turn and the forward action of the robot, and finally determine the specific position of the target through the target recognition.

3.3 Image processing and target recognition

After the robot searches for the target, it performs target recognition. This paper mainly adopts image segmentation technology based on color features. The specific process is shown in Fig.5.

Fig.5 Target recognition flow chart

The image captured by the camera may contain some other backgrounds. In order to better identify the target, the extracted image needs to be segmented. Since the target is significantly different from the background, the threshold-based image segmentation method is mainly used here[8].

Since the illumination affects the image, the captured image is subject to various disturbances and some isolated noise points appear. Filtering is an effective means to reduce the influence of noise on the image. Median filtering is a nonlinear signal processing technique that can effectively suppress noise. The principle is to replace the current pixel with the median value of the gray value of all pixels in the neighborhood window of the point. The gray value of the point makes the surrounding pixels closer to their true value. This method has a good filtering effect on some particle noises, and can well preserve the edge details of the target image so as not to be blurred. By this method, the error segmentation can be effectively reduced and the running time can be shortened.

After the image is filtered, the image is segmented by the threshold segmentation method to extract the target image. In this paper, the maximum inter-class variance method is used to select the image threshold. The maximum inter-class variance method[9] is a global-based binarization algorithm. According to the gray-scale characteristics of the image, the image is divided into two parts: the foreground and the background. The larger the variance between the two, the more The greater the difference between the two parts that make up the image. When some targets are misclassified into backgrounds or partial backgrounds are misclassified into targets, the difference between the two parts will be reduced. When the threshold is taken to maximize the variance between classes, it means that the probability of misclassification is the smallest.

Let the acquired image beI(x,y), the image gray scale varies from 1 toL, the set is represented asS={1,2,…,L}, and the pixel of gray leveliisni, then the probability of all pixels and individual pixels of the image can be formulated (3) and formula (4).

(3)

(4)

Taking theTin the setSas the threshold, the image is divided into two sets of gray levels: two types ofC1andC2, ofS1={1,2,…,T} andS2={T+1,T+2,…,L}S2, and the ratio of the number of pixels in the foreground to the whole image is recorded asω1, and the average gray levelμ1, The ratio of background pixels to the entire image isω2, and its average gray levelμ2. The total average gray level of the image is denoted byμ, and the variance between classes is denoted asg. Then there are:

(5)

Tis the best segmentation threshold when the inter variance is the largest.

g=ω1·ω2·(μ1-μ2)2

(6)

After the data is transferred to the main control system to separate the target from the background, the target gray value is 0 and the background gray value is 1. After image segmentation, the contour information of the target object is extracted to further improve the accuracy. According to formula (7), the center point of the target region can be calculated and used as the tracking feature point, and the data is transmitted to the main control system for further determination.

(7)

Wherepiexlnumrepresents the number of pixels in the target area;xsum,ysumis the sum of the abscissa and ordinate of all pixels in the target area;centerx,centeryrepresents the position of the center point of the target area identified in the image.

3.4 Target positioning

When the camera recognizes the target object, it needs to adjust the head servo to make the target in the field of view, and realize the positioning of the robot on the target object. Since the coordinates of the target in the image coordinate system are obtained from the image, and the image coordinate system is a Cartesian coordinate system, it is necessary to analyze the image to obtain the coordinates of the target in the image coordinate system, and then convert it to the actual coordinate system. In order to control the robot’s positioning target[11].

In order to achieve this goal, the robot’s foot is the base point, and the vertical robot’s front face is oriented to the right in the positivex-axis direction, the front face of the robot is in the positive direction of they-axis, and the robot coordinate system is thez-axis positive direction along the robot’s force plane. As shown in Fig.6, the pointOis the origin of the coordinate system.

Fig.6 Geometric relationship between target and camera imaging

WhereCis the camera’s optical center position,His the height of the camera’s optical center from the ground,Dis the distance from the target to the robot, planeO′ is the camera’s imaging plane, the center coordinate is (u0,v0), and pointPis the target’s centroid space coordinate point. The projections on the imaging plane areQ(u,v),φandηare the angles of the target for the horizontal and vertical directions of the robot, respectively, andαandβare the angles of the target with respect to the horizontal and vertical directions of the optical axis of the camera, respectively. In the target positioning process, only the angle information and distance information of the target relative to the robot, that is, the angleφbetween the target and the positive direction of the robot and the distanceDfrom the target to the robot, are needed to determine the position information of the target in the robot coordinate system.The camera’s small hole model is available:

(8)

Where (u,v) is the target centroid image coordinate obtained after image segmentation;θpanandθtiltare the camera’s side yaw angle and pitch angle;fis the camera focal length; dxand dyare the physical size of the image unit pixel horizontal and vertical directions, respectively,axandayare the camera internal parameter.

According to Fig.6, combined with formula (1), the angle information of the target relative to the horizontal direction of the robot can be obtained:

φ=α+θpan

(9)

Then the distance information of the target relative to the robot is studied.

Fig.7 Side view of the imaging relationship

ξis the angle of view of the camera in the vertical direction,ωis the angle between the farthest field of view and the ground, andδis the angle between the nearest field of view and the ground. According to the geometric relationship, the following formula can be obtained:

(10)

According to the imaging relationship, you can get:

(11)

WhereNis the number of pixels in the vertical direction of the image.

The distance from the target to the robotDcan be obtained from equation (10)(11) as shown in equation (12).

(12)

WhereNis the number of pixels in the vertical direction of the image.

By real-time transmission of the distance between the robot and the ball and the position control robot to perform related motion adjustment, the robot can perform the ball grabbing action when the position is less than 2 cm. When the machine judges the ball, the front basket is processed again, and the ball is successfully put into the basket through the adjustment action.

4 Experimental results and analysis

Through the above software and hardware design, the corresponding robot hardware platform is developed, as shown in Fig.8. The robot’s attitude movement was designed[12], including: large right turn, small right turn, straight walk, large left turn and small left turn.

Fig.8 Humanoid robot platform

Experiment with the humanoid shooting robot and its visual system as the experimental platform. The target is to select a 10cm orange ball. Considering that the sensor’s visual range is limited by the height of the robot body, the distance between the robot and the robot is limited to 70 cm. Then, using the method proposed in this paper to target the acquired image, the identified ball is marked with the external contour moment, and its centroid coordinates are calculated. The result is shown in the figure. Fig.9 shows, from left to right, RGB images, HSV images, threshold-divided images, and extracted contours and coordinate pictures of the ball at different positions.

Fig.9 Target segmentation results

Table 1 gives experimental data for the positioning of the beads. In the table,Lrealrepresents the actual distance between the ball and the robot,Lcalis the theoretical distance obtained from the positioning model, ΔLis the absolute error of the distance between the two, andδis the relative error of the distance between the two. It can be seen that the absolute error of each experimental result is within 3, and the relative error is up to 5.3%, which can accurately locate the target, further verifying the effectiveness of the algorithm.

Table 1 Distance measurement results

Through the distance between the robot and the ball, and the relative position information, the expert knowledge can be introduced into the humanoid robot shooting control, and then the control rule is formed: when the robot and the ball have a relative abscissa of less than 48 cm, the robot makes a right turn action; When the relative abscissa of the robot and the ball is greater than 82cm, the robot makes a left turn; when the relative abscissa of the robot and the ball is in the range of [48, 82], the robot makes a straight walk; when the distance between the robot and the ball is less than 2cm When the robot makes a shooting action. A large number of experiments have proved that the shooting robots in this paper meet the design requirements, and can overcome the external environmental factors and achieve shooting without human intervention.

5 Conclusion

In this paper, through the design of the hardware and software system of the humanoid shooting robot, by using the OpenMv module as the main control system, the surrounding environment information is collected for image preprocessing and image segmentation methods, which can accurately identify the target. At the same time, using the geometric relationship between the target and the camera imaging to calculate the target coordinates. Experiments prove that the robot can accurately locate the target, successfully complete the task of humanoid robot autonomous shooting, and has good environmental adaptability and anti-interference. The robot was successfully applied to the competition of the 5th North China Province college robot competition humanoid robot competition project, and won the first prize of Tianjin Division and the first prize of North China District.

主站蜘蛛池模板: 波多野结衣中文字幕一区| 九九热精品视频在线| 极品尤物av美乳在线观看| 免费观看精品视频999| 国产精品专区第一页在线观看| 亚洲天堂精品视频| 成年人久久黄色网站| 亚洲第一页在线观看| 亚洲欧美激情小说另类| 国产乱人激情H在线观看| 香蕉伊思人视频| 亚洲一区毛片| 日韩 欧美 国产 精品 综合| 欧美成人亚洲综合精品欧美激情| 欧美a级在线| 欧美国产成人在线| 中日韩一区二区三区中文免费视频| 亚洲中文字幕无码爆乳| 黄色三级毛片网站| 欧美劲爆第一页| 欧美成人午夜视频| 99在线视频免费| 亚洲无码精品在线播放| 在线观看国产精品第一区免费| 亚洲欧洲美色一区二区三区| 中文字幕首页系列人妻| 色综合日本| 国产区成人精品视频| 一级爱做片免费观看久久 | 伊在人亚洲香蕉精品播放| 国产午夜一级淫片| 日韩专区第一页| 国产精品偷伦视频免费观看国产| 欧美日韩国产在线播放| 亚洲色图在线观看| 911亚洲精品| 免费视频在线2021入口| 凹凸国产熟女精品视频| 日韩无码真实干出血视频| 亚洲国产日韩一区| 国产日韩精品欧美一区灰| 国产高清在线观看| 亚洲国产成人精品一二区| 国产成人综合在线视频| 国产成人精品一区二区免费看京| 日韩第九页| 日韩天堂在线观看| 18禁影院亚洲专区| 亚洲综合18p| 国产91av在线| 婷婷伊人五月| 精品人妻AV区| 欧美精品二区| 欧美性色综合网| 国产高潮视频在线观看| 国产a v无码专区亚洲av| 国产黄色片在线看| 毛片免费高清免费| 99精品热视频这里只有精品7| 亚洲第一视频网| 2021天堂在线亚洲精品专区| 亚洲精品在线91| 欧美三級片黃色三級片黃色1| 久久综合成人| 国产精品所毛片视频| 亚洲日本www| 国国产a国产片免费麻豆| 日韩无码真实干出血视频| 国产一区三区二区中文在线| 一级看片免费视频| 国产欧美综合在线观看第七页| 亚洲熟女中文字幕男人总站| 白浆免费视频国产精品视频 | 欧美a级完整在线观看| vvvv98国产成人综合青青| 国产91成人| 国产国拍精品视频免费看| 欧美色图久久| 婷五月综合| 91精品国产情侣高潮露脸| 欧美不卡视频一区发布| 国产一在线|