远程控制slam小车
1.
把小车树莓派及pc端ubuntu通过无线路由器连接到统一局域网中
2.登入路由器查看是设备是否连接成功
http://192.168.1.1/
路由器密码
树莓派: zxcar 密码:123
ssh zxcar@树莓派的IP 然后输入密码 就可以了 确保自己的电脑和树莓派在同一个网络
远程拷贝
scp -r src zxcar@192.168.1.12:~/itheima4_ws
远程登入
ssh zxcar@192.168.1.12
密码 123
上位机单独启动驱动命令
rosrun car_driver car_driver.py
ubuntu链接串口
ls -al /dev/ttyU*
ros 控制键盘调试
rosrun teleop_twist_keyboard teleop_twist_keyboard.py
键盘控制各个键介绍:
---------------------------
UI O
JK L
M< >
K—停止
I、J、<、L—前、左、后、右
q/z : 最大速度增加/减少10%
w/x : 仅线性速度增加10%
e/c : 仅角速度增加10%
查找设备
lsusb -vvv
设置设备串口udev,定义rules文件 激活rules
1.拷贝到以下目录
sudo cp 58-robot.rules /etc/udev/rules.d/
或
sudo cp 58-robot.rules /lib/udev/rules.d/
2.加载文件
sudo service udev reload
3.启动
sudo service udev restart
启动报错提示用下面的命令:
sudo systemctl daemon-reload
4.插上设备 查看是否安装完成
ls -al /dev/skser*ls -al /dev/sulid*
命令
roslaunch zxcar bringup.launch
树莓派启动命令建图
roslaunch zxcar lidar_slam.launch
这个要在pc本机启动
roslaunch zxcar lidar_slam_rviz.launch
本机启动键盘控制
rosrun teleop_twist_keyboard teleop_twist_keyboard.py
问题:
sk@sk-PC:~/workspas/slam$ rosrun teleop_twist_keyboard teleop_twist_keyboard.py
Traceback (most recent call last):
File "/home/sk/workspas/slam/src/teleop_twist_keyboard/teleop_twist_keyboard.py", line 7, in <module>
import roslib; roslib.load_manifest('teleop_twist_keyboard')
File "/opt/ros/melodic/lib/python2.7/dist-packages/roslib/launcher.py", line 64, in load_manifest
sys.path = _generate_python_path(package_name, _rospack) + sys.path
File "/opt/ros/melodic/lib/python2.7/dist-packages/roslib/launcher.py", line 97, in _generate_python_path
m = rospack.get_manifest(pkg)
File "/usr/lib/python2.7/dist-packages/rospkg/rospack.py", line 171, in get_manifest
return self._load_manifest(name)
File "/usr/lib/python2.7/dist-packages/rospkg/rospack.py", line 215, in _load_manifest
retval = self._manifests[name] = parse_manifest_file(self.get_path(name), self._manifest_name, rospack=self)
File "/usr/lib/python2.7/dist-packages/rospkg/manifest.py", line 414, in parse_manifest_file
_static_rosdep_view = init_rospack_interface()
File "/usr/lib/python2.7/dist-packages/rosdep2/rospack.py", line 59, in init_rospack_interface
lookup = _get_default_RosdepLookup(Options())
File "/usr/lib/python2.7/dist-packages/rosdep2/main.py", line 134, in _get_default_RosdepLookup
verbose=options.verbose)
File "/usr/lib/python2.7/dist-packages/rosdep2/sources_list.py", line 594, in create_default
sources = load_cached_sources_list(sources_cache_dir=sources_cache_dir, verbose=verbose)
File "/usr/lib/python2.7/dist-packages/rosdep2/sources_list.py", line 552, in load_cached_sources_list
with open(cache_index, 'r') as f:
IOError: [Errno 13] Permission denied: '/home/sk/.ros/rosdep/sources.cache/index'
IOError: [Errno 13] 权限问题
解决:
cd /home/sk/.ros/rosdep/sources.cache/sudo chmod 777 index
问题2:链接不上或找不到节点 可能ip问题
输入
ifconfig
与
echo $ROS_HOSTNAME
查看ip是否一致
强制更改 本机ip; 把动态命令写死
export ROS_HOSTNAME=`hostname -I | awk '{print $1}'`
改
export ROS_HOSTNAME=192.168.1.10
动态命令在多网卡或网络中断会导致把本机ip发送给mast节点的ip不准确
第一种:动态多机部署ros 方法:
主机:树莓派端配置
export ROS_IP=`hostname -I | awk '{print $1}'`export ROS_HOSTNAME=`hostname -I | awk '{print $1}'`export ROS_MASTER_URI=http://`hostname -I | awk '{print $1}'`:11311
从机:
export ROS_IP=`hostname -I | awk '{print $1}'`export ROS_HOSTNAME=`hostname -I | awk '{print $1}'`export ROS_MASTER_URI=http://树莓派的IP地址:11311
第二种:多机部署ros 方法:
> 主机和从机要在同一个局域网内
1. 主机添加在hosts中添加所有节点的ip和主机名(主机配置)
```bashsudo gedit /etc/hosts```
> ip 主机名
2. 从机配置master节点服务(主机也要修改)
```bashsudo gedit ~/.bashrcexport ROS_HOSTNAME=本机ipexport ROS_MASTER_URI=http://主机ip:11311
设置树莓派连接wifi
HUAWEI-sukai1 01524922
2.
把树莓派tf卡插入电脑
1.新建一个文件名字ssh的空文件
4 ) 使 用 记 事 本 工 具 , 按 照 下 面 的 参 考 格 式 填 入 记 事 本 内 容 , 然 后 命 名 为
“wpa_supplicant.conf ”(温馨提示:可先输入下方内容,保存并退出该文件。然后将文件名
进行重命名,需删除包含 txt 在内的所有文件名,替换为上方名称。如有弹窗提示,选择“是”
34即可),再保存至“boot”盘符内。
country=CN
ctrl_interface=DIR=/var/run/wpa_supplicant GROUP=netdev
update_config=1
network={
ssid="Hiwonder"
psk="123456789"
key_mgmt=WPA-PSK
priority=1
}
弹出tf卡插入树莓派,让树莓派与电脑都连接到热点
PC与树莓派ssh链接时出现间歇性联通段开网络故障
sendmsg: No buffer space acailable
sendmsg: Network is Unreachable
From 192.168.**.*** icmp_se=183 Destination Host Unreachable
connection closed by 192.168.**.*** port 22
以下是ping的图片
故障排除
如果当你尝试连接到小派时遇到connection reset by peer错误,可能是由SSH key引起的。你可以通过下面的命令来重置它。
首先,删除旧的key文件:
sudo rm /etc/ssh/ssh_host_*
然后生成一个新的:
sudo dpkg-reconfigure openssh-server
然后再试一次。
重启
reboot
在windows 中使用putty登入试试
NameError: global name 'ser' is not defined
serial 这个报错:
查出来必须要卸载serial 不要安装:pip uninstall serial要安装:pip install pyserial
slam小车pid调试
安装 rqt_plot
rosdep install rqt_plot
启动命令:
rqt_plot
在上位机car_driver驱动中需要把速度信息发布出来:
rostopic list 查看节点
控制台输出句柄消息
rostopic echo /zxcar/get_vel
rqt_plot 中监听速度
命令
rqt_plot
调整y方向的范围:
调试:
Rosrun teleop_twist_keyboard teleop_twist_keyboard.py //启动键盘控制
$rviz //参考上节开启navigatin.rviz观察机器人仿真情况
采用键盘控制模拟机器人,模拟机器人开始移动。注意鼠标指针必须位于teleop_twist_keyboard终端页面,否则控制键盘模拟机器人无法移动。
键盘控制各个键介绍:
---------------------------
UI O
JK L
M< >
K—停止
I、J、<、L—前、左、后、右
q/z : 最大速度增加/减少10%
w/x : 仅线性速度增加10%
e/c : 仅角速度增加10%
速度设置到0.215
启动rosrun rqt_topic rqt_topic
小车反馈回来,小车收到的数据
看角速度的数据
py_slam_exploring_slam.py自动导航建模
#!/usr/bin/env python # -*- coding: utf-8 -*-import roslib;import rospy import actionlib from actionlib_msgs.msg import * from geometry_msgs.msg import Pose, PoseWithCovarianceStamped, Point, Quaternion, Twist from move_base_msgs.msg import MoveBaseAction, MoveBaseGoal from random import sample from math import pow, sqrt class NavTest(): def __init__(self): rospy.init_node('exploring_slam', anonymous=True) rospy.on_shutdown(self.shutdown) # 在每个目标位置暂停的时间 (单位:s)self.rest_time = rospy.get_param("~rest_time", 2) # 是否仿真? self.fake_test = rospy.get_param("~fake_test", True) # 到达目标的状态 goal_states = ['PENDING', 'ACTIVE', 'PREEMPTED', 'SUCCEEDED', 'ABORTED', 'REJECTED', 'PREEMPTING', 'RECALLING', 'RECALLED', 'LOST'] # 设置目标点的位置 # 在rviz中点击 2D Nav Goal 按键,然后单击地图中一点 # 在终端中就会看到该点的坐标信息 locations = dict() locations['1'] = Pose(Point(4.589, -0.376, 0.000), Quaternion(0.000, 0.000, -0.447, 0.894)) locations['2'] = Pose(Point(4.231, -6.050, 0.000), Quaternion(0.000, 0.000, -0.847, 0.532)) locations['3'] = Pose(Point(-0.674, -5.244, 0.000), Quaternion(0.000, 0.000, 0.000, 1.000)) locations['4'] = Pose(Point(-5.543, -4.779, 0.000), Quaternion(0.000, 0.000, 0.645, 0.764)) locations['5'] = Pose(Point(-4.701, -0.590, 0.000), Quaternion(0.000, 0.000, 0.340, 0.940)) locations['6'] = Pose(Point(2.924, 0.018, 0.000), Quaternion(0.000, 0.000, 0.000, 1.000)) # 发布控制机器人的消息 self.cmd_vel_pub = rospy.Publisher('cmd_vel', Twist, queue_size=5) # 订阅move_base服务器的消息 self.move_base = actionlib.SimpleActionClient("move_base", MoveBaseAction) rospy.loginfo("Waiting for move_base action server...") # 60s等待时间限制 self.move_base.wait_for_server(rospy.Duration(60)) rospy.loginfo("Connected to move base server") # 保存机器人的在rviz中的初始位置 initial_pose = PoseWithCovarianceStamped() # 保存成功率、运行时间、和距离的变量 n_locations = len(locations) n_goals = 0 n_successes = 0 i = n_locations distance_traveled = 0 start_time = rospy.Time.now() running_time = 0 location = "" last_location = "" # 确保有初始位置 while initial_pose.header.stamp == "": rospy.sleep(1) rospy.loginfo("Starting navigation test") # 开始主循环,随机导航 while not rospy.is_shutdown(): # 如果已经走完了所有点,再重新开始排序 if i == n_locations: i = 0 sequence = sample(locations, n_locations) # 如果最后一个点和第一个点相同,则跳过 if sequence[0] == last_location: i = 1 # 在当前的排序中获取下一个目标点 location = sequence[i] # 跟踪行驶距离 # 使用更新的初始位置 if initial_pose.header.stamp == "": distance = sqrt(pow(locations[location].position.x - locations[last_location].position.x, 2) + pow(locations[location].position.y - locations[last_location].position.y, 2)) else: rospy.loginfo("Updating current pose.") distance = sqrt(pow(locations[location].position.x - initial_pose.pose.pose.position.x, 2) + pow(locations[location].position.y - initial_pose.pose.pose.position.y, 2)) initial_pose.header.stamp = "" # 存储上一次的位置,计算距离 last_location = location # 计数器加1 i += 1 n_goals += 1 # 设定下一个目标点 self.goal = MoveBaseGoal() self.goal.target_pose.pose = locations[location] self.goal.target_pose.header.frame_id = 'map' self.goal.target_pose.header.stamp = rospy.Time.now() # 让用户知道下一个位置 rospy.loginfo("Going to: " + str(location)) # 向下一个位置进发 self.move_base.send_goal(self.goal) # 五分钟时间限制 finished_within_time = self.move_base.wait_for_result(rospy.Duration(300)) # 查看是否成功到达 if not finished_within_time: #取消机器人运动,不需要运动到指定点了self.move_base.cancel_goal() rospy.loginfo("Timed out achieving goal") else: state = self.move_base.get_state() if state == GoalStatus.SUCCEEDED: rospy.loginfo("Goal succeeded!") n_successes += 1 distance_traveled += distance rospy.loginfo("State:" + str(state)) else: rospy.loginfo("Goal failed with error code: " + str(goal_states[state])) # 运行所用时间 running_time = rospy.Time.now() - start_time running_time = running_time.secs / 60.0 # 输出本次导航的所有信息 rospy.loginfo("Success so far: " + str(n_successes) + "/" + str(n_goals) + " = " + str(100 * n_successes/n_goals) + "%") rospy.loginfo("Running time: " + str(trunc(running_time, 1)) + " min Distance: " + str(trunc(distance_traveled, 1)) + " m") rospy.sleep(self.rest_time) def update_initial_pose(self, initial_pose): self.initial_pose = initial_pose def shutdown(self): rospy.loginfo("Stopping the robot...") self.move_base.cancel_goal() rospy.sleep(2) self.cmd_vel_pub.publish(Twist()) rospy.sleep(1) def trunc(f, n): slen = len('%.*f' % (n, f)) return float(str(f)[:slen]) if __name__ == '__main__': try: NavTest() rospy.spin() except rospy.ROSInterruptException: rospy.loginfo("Exploring SLAM finished.")
py_opencv_motion_detector.py物体跟随
#!/usr/bin/env python# -*- coding: utf-8 -*-import rospyimport cv2import numpy as npfrom sensor_msgs.msg import Image, RegionOfInterestfrom cv_bridge import CvBridge, CvBridgeErrorclass motionDetector:def __init__(self):rospy.on_shutdown(self.cleanup);# 创建cv_bridgeself.bridge = CvBridge()self.image_pub = rospy.Publisher("cv_bridge_image", Image, queue_size=1)# 设置参数:最小区域、阈值self.minArea = rospy.get_param("~minArea", 500)self.threshold = rospy.get_param("~threshold", 25)self.firstFrame = Noneself.text = "Unoccupied"# 初始化订阅rgb格式图像数据的订阅者,此处图像topic的话题名可以在launch文件中重映射self.image_sub = rospy.Subscriber("input_rgb_image", Image, self.image_callback, queue_size=1)def image_callback(self, data):# 使用cv_bridge将ROS的图像数据转换成OpenCV的图像格式try:cv_image = self.bridge.imgmsg_to_cv2(data, "bgr8")frame = np.array(cv_image, dtype=np.uint8)except CvBridgeError, e:print e# 创建灰度图像gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)gray = cv2.GaussianBlur(gray, (21, 21), 0)# 使用两帧图像做比较,检测移动物体的区域if self.firstFrame is None:self.firstFrame = grayreturn #获取差分图 就是将两幅图像作差两个图片相减,这里用的是灰度图,类型是uint8frameDelta = cv2.absdiff(self.firstFrame, gray)thresh = cv2.threshold(frameDelta, self.threshold, 255, cv2.THRESH_BINARY)[1]thresh = cv2.dilate(thresh, None, iterations=2)binary, cnts, hierarchy= cv2.findContours(thresh.copy(), cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)for c in cnts:# 如果检测到的区域小于设置值,则忽略if cv2.contourArea(c) < self.minArea:continue # 在输出画面上框出识别到的物体(x, y, w, h) = cv2.boundingRect(c)cv2.rectangle(frame, (x, y), (x + w, y + h), (50, 255, 50), 2)self.text = "Occupied"# 在输出画面上打当前状态和时间戳信息cv2.putText(frame, "Status: {}".format(self.text), (10, 20),cv2.FONT_HERSHEY_SIMPLEX, 0.5, (0, 0, 255), 2)# 将识别后的图像转换成ROS消息并发布self.image_pub.publish(self.bridge.cv2_to_imgmsg(frame, "bgr8"))def cleanup(self):print "Shutting down vision node."cv2.destroyAllWindows()if __name__ == '__main__':try:# 初始化ros节点rospy.init_node("motion_detector")rospy.loginfo("motion_detector node is started...")rospy.loginfo("Please subscribe the ROS image.")motionDetector()rospy.spin()except KeyboardInterrupt:print "Shutting down motion detector node."cv2.destroyAllWindows()
py_slam_startros_在代码中启动ros节点roslaunch和rosrun
import subprocessimport rospyimport rosnodeclass launch_demo:def __init__(self, cmd=None):self.cmd = cmddef launch(self):self.child = subprocess.Popen(self.cmd)return Truedef shutdown(self):self.child.terminate()self.child.wait()return Trueif __name__ == "__main__":rospy.init_node('launch_demo',anonymous=True)launch_nav = launch_demo(["roslaunch", "pibot_simulator", "nav.launch"])launch_nav.launch()r = rospy.Rate(0.2)r.sleep()rospy.loginfo("switch map...")r = rospy.Rate(1)r.sleep()rosnode.kill_nodes(['map_server'])map_name = "/home/pibot/ros_ws/src/pibot_simulator/maps/blank_map_with_obstacle.yaml"map_node = subprocess.Popen(["rosrun", "map_server", "map_server", map_name, "__name:=map_server"])while not rospy.is_shutdown():r.sleep()
上面使用python
代码启动了一个PIBOT
模拟器的导航,然后5s
后切换了一个地图
使用subprocess.Popen
可以启动一个进程(roslaunch
或者rosrun
)使用rosnode.kill_nodes
可以杀死一个rosnode
py_slam_定点导航
from launch_demo import launch_demoimport rospyimport actionlibfrom actionlib_msgs.msg import *from move_base_msgs.msg import MoveBaseAction, MoveBaseGoalfrom nav_msgs.msg import Pathfrom geometry_msgs.msg import PoseWithCovarianceStampedfrom tf_conversions import transformationsfrom math import piclass navigation_demo:def __init__(self):self.set_pose_pub = rospy.Publisher('/initialpose', PoseWithCovarianceStamped, queue_size=5)self.move_base = actionlib.SimpleActionClient("move_base", MoveBaseAction)self.move_base.wait_for_server(rospy.Duration(60))def set_pose(self, p):if self.move_base is None:return Falsex, y, th = ppose = PoseWithCovarianceStamped()pose.header.stamp = rospy.Time.now()pose.header.frame_id = 'map'pose.pose.pose.position.x = xpose.pose.pose.position.y = yq = transformations.quaternion_from_euler(0.0, 0.0, th/180.0*pi)pose.pose.pose.orientation.x = q[0]pose.pose.pose.orientation.y = q[1]pose.pose.pose.orientation.z = q[2]pose.pose.pose.orientation.w = q[3]self.set_pose_pub.publish(pose)return Truedef _done_cb(self, status, result):rospy.loginfo("navigation done! status:%d result:%s"%(status, result))def _active_cb(self):rospy.loginfo("[Navi] navigation has be actived")def _feedback_cb(self, feedback):rospy.loginfo("[Navi] navigation feedback\r\n%s"%feedback)def goto(self, p):goal = MoveBaseGoal()goal.target_pose.header.frame_id = 'map'goal.target_pose.header.stamp = rospy.Time.now()goal.target_pose.pose.position.x = p[0]goal.target_pose.pose.position.y = p[1]q = transformations.quaternion_from_euler(0.0, 0.0, p[2]/180.0*pi)goal.target_pose.pose.orientation.x = q[0]goal.target_pose.pose.orientation.y = q[1]goal.target_pose.pose.orientation.z = q[2]goal.target_pose.pose.orientation.w = q[3]self.move_base.send_goal(goal, self._done_cb, self._active_cb, self._feedback_cb)return Truedef cancel(self):self.move_base.cancel_all_goals()return Trueif __name__ == "__main__":rospy.init_node('navigation_demo',anonymous=True)launch_nav = launch_demo(["roslaunch", "pibot_simulator", "nav.launch"])launch_nav.launch()r = rospy.Rate(0.2)r.sleep()rospy.loginfo("set pose...")r = rospy.Rate(1)r.sleep()navi = navigation_demo()navi.set_pose([-0.7,-0.4,0])rospy.loginfo("goto goal...")r = rospy.Rate(1)r.sleep()navi.goto([0.25,4, 90])while not rospy.is_shutdown():r.sleep()
上面完成设置机器人位置和导航到某一位置的功能
navi.set_pose([-0.7,-0.4,0])
设置机器人位于地图中位置(-0.7,-0.4) 姿态yaw=0°
navi.goto([0.25,4, 90])
该接口调用一个服务完成机器人运动至位置(0.25,4)姿态yaw=90°
有了该接口就不难完成多点导航巡逻等应用
Twiddle 调校最优解PID
为了避免人为的猜测、试错来获取比较好的PID值,我们可以设计一种自动化的调参逻辑循环,根据使指定目标函数的值达到最小时,得到更加优秀的参数列表。
一种更智能、自动的方法是使用梯度下降算法。 前提是您从三个增益的初始猜测向量开始。 通常对P使用小的非零值,对I和D使用0。 然后,分别对每个收益进行小幅更改,然后测试目标函数是否降低。 如果降低了,将沿相同方向不断更改参数,否则尝试沿相反方向调整参数。 如果增益值的增加或减少均不会降低成本函数,则减小增益增量的大小并重复。 整个循环应继续进行,直到增量大小降至某个阈值以下。
我们可以把以下过程称之为 Twiddle(旋弄、捻弄),可以帮助我们确定合适的P、I、D三个值,其原理是构建一个初始化为[0,0,0][0,0,0]的参数列表,然后循环多次,根据每次循环返回的误差均值,调大或调小参数(调整的步伐参数由另一个3个值的列表控制)。直到步伐参数之和小于一定阈值时,停止循环,把此时参数列表作为最终的PID参数。
代码实现:
拷贝依赖文件robot.py到项目目录新建文件pid_twiddle.py
并编写如下内容:
from robot import Robot, showimport numpy as npdef make_robot():"""创建并初始化机器人小车, 设置初始位置为(0, -1), 初始旋转角度为0"""robot = Robot()robot.set(0, -1, 0)robot.set_steering_drift(10 / 180 * np.pi)return robot# NOTE: We use params instead of k_p, k_d, k_idef run(robot, params, n=100, speed=1.0):x_trajectory = []y_trajectory = []err = 0prev_cte = 0 - robot.yint_cte = 0for i in range(2 * n):cte = 0 - robot.ydiff_cte = cte - prev_cteint_cte += cteprev_cte = ctesteer = params[0] * cte + params[1] * diff_cte + params[2] * int_cterobot.move(steer, speed)x_trajectory.append(robot.x)y_trajectory.append(robot.y)if i >= n:err += cte ** 2return x_trajectory, y_trajectory, err / n# Make this tolerance bigger if you are timing out!def twiddle(tol=0.2):p = [0, 0, 0]dp = [1, 1, 1]robot = make_robot()x_trajectory, y_trajectory, best_err = run(robot, p)it = 0while sum(dp) > tol:# 循环,直到系数之和小于等于阈值(默认阈值为0.2,起始值为3.0)print("Iteration {}, best error = {}".format(it, best_err))for i in range(len(p)):p[i] += dp[i]robot.reset()x_trajectory, y_trajectory, err = run(robot, p)if err < best_err:best_err = errdp[i] *= 1.1# 此值有助于减少总误差,下次可以多加点else:p[i] -= 2 * dp[i]# 此值不利于于减少总误差,直接把刚加的去掉,并且向反方向减一倍robot.reset()x_trajectory, y_trajectory, err = run(robot, p)if err < best_err:best_err = errdp[i] *= 1.1 # 如果反方向有利于减少总误差,扩大此值else:p[i] += dp[i] # 恢复为原值dp[i] *= 0.9 # 把缩小变化系数it += 1return p, best_errparams, err = twiddle()print("Final twiddle error = {} params = {}".format(err, params))robot = make_robot()x_trajectory, y_trajectory, err = run(robot, params)show(x_trajectory, y_trajectory, label="Twiddle PID")
运行结果:
可以与之前使用的PID进行对比,明显可以看出当前PID在x=50附近就开始在目标轨迹附近稳定下来。而之前的要到x=100的位置才可以稳定下来。
输出日志:
根据最后的输出日志可知:
最小的误差均值为7.940560962605189e-07
此时对应的P、D、I分别为:10.716018504541431, 18.68325573584582, 0.020275559590445292
Iteration 0, best error = 7972.071547906822Iteration 1, best error = 0.048853806107299856Iteration 2, best error = 0.03026214567061226Iteration 3, best error = 0.0077046028132098255Iteration 4, best error = 0.003222969736312333Iteration 5, best error = 0.0016693580238629137Iteration 6, best error = 0.0009763548793623677Iteration 7, best error = 0.0006143945322198215Iteration 8, best error = 0.0006143945322198215Iteration 9, best error = 0.0006143945322198215Iteration 10, best error = 0.0006143945322198215Iteration 11, best error = 0.0006143945322198215Iteration 12, best error = 0.0006143945322198215Iteration 13, best error = 0.0006143945322198215Iteration 14, best error = 0.0006143945322198215Iteration 15, best error = 0.0006143945322198215Iteration 16, best error = 0.0006143945322198215Iteration 17, best error = 0.0006143945322198215Iteration 18, best error = 0.000612580641120018Iteration 19, best error = 0.0005393890590252054Iteration 20, best error = 0.0004707073557690764Iteration 21, best error = 0.00040854195467355256Iteration 22, best error = 0.00036173191894990056Iteration 23, best error = 0.0003496794660242756Iteration 24, best error = 0.00030372592629584626Iteration 25, best error = 0.00026649497706284647Iteration 26, best error = 0.00026502850791677363Iteration 27, best error = 0.00023267758106029443Iteration 28, best error = 0.00023267758106029443Iteration 29, best error = 0.00023267758106029443Iteration 30, best error = 0.00023267758106029443Iteration 31, best error = 0.00023267758106029443Iteration 32, best error = 0.00023222333181571606Iteration 33, best error = 0.00023222333181571606Iteration 34, best error = 0.00023222333181571606Iteration 35, best error = 0.00023222333181571606Iteration 36, best error = 0.00023222333181571606Iteration 37, best error = 0.00022978905387007294Iteration 38, best error = 0.00019428921416175855Iteration 39, best error = 0.00014919774411248834Iteration 40, best error = 2.2810673236127803e-05Iteration 41, best error = 6.65198948619874e-06Iteration 42, best error = 8.957452817602662e-07Iteration 43, best error = 8.957452817602662e-07Iteration 44, best error = 8.957452817602662e-07Iteration 45, best error = 8.957452817602662e-07Iteration 46, best error = 8.957452817602662e-07Iteration 47, best error = 8.957452817602662e-07Iteration 48, best error = 8.957452817602662e-07Iteration 49, best error = 8.957452817602662e-07Iteration 50, best error = 8.957452817602662e-07Iteration 51, best error = 7.940560962605189e-07Iteration 52, best error = 7.940560962605189e-07Iteration 53, best error = 7.940560962605189e-07Iteration 54, best error = 7.940560962605189e-07Iteration 55, best error = 7.940560962605189e-07Iteration 56, best error = 7.940560962605189e-07Iteration 57, best error = 7.940560962605189e-07Final twiddle error = 7.940560962605189e-07 params = [10.716018504541431, 18.68325573
robot.py
import randomimport numpy as npimport matplotlib.pyplot as pltclass Robot(object):def __init__(self, length=20.0):"""鍒涘缓鏈哄櫒浜哄苟鍒濆鍖栦綅缃拰鏂瑰悜涓�0, 0, 0."""self.x = 0.0self.y = 0.0self.orientation = 0.0 # 涓嶺杞存鏂瑰悜鐨勫す瑙掞紙鍗曚綅涓哄姬搴︼級self.length = length# 鍓嶅悗杞瓙鐨勮酱璺�self.steering_noise = 0.0 # 鏂瑰悜鍣0self.distance_noise = 0.0 # 璺濈鍣0self.steering_drift = 0.0 # 鏂瑰悜婕傜Щself.default_state = {"x": self.x,"y": self.y,"o": self.orientation}def reset(self):self.x = self.default_state["x"]self.y = self.default_state["y"]self.orientation = self.default_state["o"]def set(self, x, y, orientation):"""璁剧疆鏈哄櫒浜虹殑鍧愭爣鍙婃柟鍚�"""self.x = xself.y = yself.orientation = orientation % (2.0 * np.pi)self.default_state = {"x": self.x,"y": self.y,"o": self.orientation}def set_noise(self, steering_noise, distance_noise):"""璁剧疆鍣0鍙傛暟:param steering_noise: 杞悜鍣0:param distance_noise: 璺濈鍣0"""# makes it possible to change the noise parameters# this is often useful in particle filtersself.steering_noise = steering_noiseself.distance_noise = distance_noisedef set_steering_drift(self, drift):"""璁剧疆绯荤粺鐨勮浆鍚戞紓绉诲弬鏁�"""self.steering_drift = driftdef move(self, steering, distance, tolerance=0.001, max_steering_angle=np.pi / 4.0):"""灏忚溅鐨勭Щ鍔ㄥ嚱鏁�:param steering: 鍓嶈疆鐨勮浆鍚戣锛屾渶澶у€间负max_steering_angle:param distance: 鎬昏椹惰窛绂伙紝涓€鑸负闈炶礋:param tolerance: 杞悜鐨勬渶灏忓樊鍊硷紙闃堝€硷級锛屽皬浜庢闃堝€兼椂锛岃灏忚溅璧扮洿绾匡紝鍗曚綅涓哄姬搴�:param max_steering_angle: 鏈€澶ц浆鍚戣锛岄粯璁や负 180 / 4.0 = 45掳"""# if steering > max_steering_angle:#steering = max_steering_angle# if steering < -max_steering_angle:#steering = -max_steering_anglesteering = np.clip(steering, -max_steering_angle, max_steering_angle)if distance < 0.0:distance = 0.0# apply noisesteering2 = random.gauss(steering, self.steering_noise)distance2 = random.gauss(distance, self.distance_noise)# apply steering driftsteering2 += self.steering_drift# 瑙掗€熷害 = 绾块€熷害 / 杞集鍗婂緞# Execute motionturn = np.tan(steering2) * distance2 / self.lengthif abs(turn) < tolerance:# approximate by straight line motion 杩戜技鐩寸嚎妯″瀷self.x += distance2 * np.cos(self.orientation)self.y += distance2 * np.sin(self.orientation)self.orientation = (self.orientation + turn) % (2.0 * np.pi)else:# approximate bicycle model for motion 杩戜技鑷杞︽ā鍨�radius = distance2 / turncx = self.x - (np.sin(self.orientation) * radius)cy = self.y + (np.cos(self.orientation) * radius)self.orientation = (self.orientation + turn) % (2.0 * np.pi)self.x = cx + (np.sin(self.orientation) * radius)self.y = cy - (np.cos(self.orientation) * radius)def __repr__(self):return '[x=%.5f y=%.5f orient=%.5f]' % (self.x, self.y, self.orientation)def show(x_trajectory, y_trajectory, p_array=[], i_array=[], d_array=[], label = 'PID'):n = len(x_trajectory)# fig, (ax1, ax2) = plt.subplots(2, 1, figsize=(8, 8))# fig, ax1 = plt.subplots(figsize=(8, 4))fig = plt.figure()ax1 = fig.add_subplot(211)ax1.plot(x_trajectory, np.zeros(n), 'pink', label='reference')ax1.plot(x_trajectory, y_trajectory, 'black', label= label + ' controller')ax1.set_xlabel('x') # Add an x-label to the axes.ax1.set_ylabel('y') # Add a y-label to the axes.ax1.set_title('Car-Position')h, l = ax1.get_legend_handles_labels()ax1.legend(h, l) # h 涓虹嚎鏉″璞″垪琛紝 l涓烘枃瀛楁弿杩板垪琛�ax2 = fig.add_subplot(212)if len(p_array) > 0:ax2.plot(x_trajectory, p_array, color='r', label='p')if len(i_array) > 0:ax2.plot(x_trajectory, i_array, color='g', label='i')if len(d_array) > 0:ax2.plot(x_trajectory, d_array, color='b', label='d')ax2.set_title('PID-Value')h, l = ax2.get_legend_handles_labels()ax2.legend(h, l) # h 涓虹嚎鏉″璞″垪琛紝 l涓烘枃瀛楁弿杩板垪琛�plt.ylim((-1.5, 1.5))plt.tight_layout()plt.show()
小车跟随颜色块
小车跟随黄色块进行运动,当然如果大家在地面上贴一根黄色的线,小车也可以跟随黄色线进行运动.
#! /usr/bin/env python# encoding: utf-8import rospyfrom geometry_msgs.msg import Twistimport cv2 as cvdef shutdown():twist = Twist()twist.linear.x = 0twist.angular.z = 0cmd_vel_Publisher.publish(twist)print "stop car..."if __name__ == '__main__':rospy.init_node("yellow_follow")# 当程序退出rospy.on_shutdown(shutdown);# ros控制的频率rate = rospy.Rate(100)# 定义publisher : cmd_velcmd_vel_Publisher = rospy.Publisher("/cmd_vel",Twist,queue_size=1)capture = cv.VideoCapture(0)print capture.isOpened()ok,frame = capture.read()lowerb = (23,43,46)upperb = (34,255,255)height,width = frame.shape[0:2]screen_center = width / 2offset = 50while not rospy.is_shutdown():# 将图像转成HSV颜色空间hsv_frame = cv.cvtColor(frame, cv.COLOR_BGR2HSV)# 基于颜色的物品提取mask = cv.inRange(hsv_frame,lowerb,upperb)# 找出面积最大的区域_,contours,_ = cv.findContours(mask,cv.RETR_EXTERNAL,cv.CHAIN_APPROX_SIMPLE)maxArea = 0maxIndex = 0for i,c in enumerate(contours):area = cv.contourArea(c)if area > maxArea:maxArea = areamaxIndex = i# 绘制cv.drawContours(frame,contours,maxIndex,(0,0,255),2)# 获取外切矩形x,y,w,h = cv.boundingRect(contours[maxIndex])cv.rectangle(frame,(x,y),(x+w,y+h),(255,255,0),2)# 获取外切矩形的中心像素点center_x = int(x + w/2)center_y = int(y + h/2)cv.circle(frame,(center_x,center_y),5,(0,0,255),-1)# 判断当前小车应该是左转还是右转还是直行twist = Twist()if center_x < screen_center - offset:twist.linear.x = 0.05twist.angular.z = 0.2print "turn left"elif center_x >= screen_center - offset and center_x <= screen_center + offset:twist.linear.x = 0.1twist.angular.z = 0.0print "go"elif center_x > screen_center + offset:twist.linear.x = 0.05twist.angular.z = -0.2print "turn right"else:twist.linear.x = 0twist.angular.z = 0print "stop"# 将速度信息发送出去cmd_vel_Publisher.publish(twist)cv.imshow("mask",mask)cv.imshow("frame",frame)cv.waitKey(1)rate.sleep()ok, frame = capture.read()
pid
intergral = 0;derivative = 0;prev_error = 0;pid = 0def pid_control(curr,target):global intergral,derivative,prev_error,piderror = target - currintergral += error;derivative = prev_error;# pid 计算公式kp = 0.2ki = 0.0kd = 0.02pid = kp*error + ki * intergral + kd* derivativeprev_error = error;print pid;return pid
视觉巡线
##基于颜色提取#车道线表现出来是黄色的,所以我们可以利用我们前面基于HSV颜色范围的方式进行提取,最终提取出来的结果#def fetch_yellow(img):hsv_img = cv.cvtColor(img,cv.COLOR_BGR2HSV)low = (25,46,46);upper = (35,255,255)binary = cv.inRange(hsv_img,low,upper)if DEBUG:cv.imshow("binary color",binary);return binary##截取ROIdef fetch_roi(img):# 提取感兴趣的而区域height,width = img.shape[0:2]roi = img[int(height/2):height,0:width]# 对提取的图像进行闭操作kernel = cv.getStructuringElement(cv.MORPH_RECT, (3, 3))roi = cv.morphologyEx(roi,cv.MORPH_CLOSE,kernel,iterations=3)_, contours, _ = cv.findContours(roi,cv.RETR_EXTERNAL,cv.CHAIN_APPROX_SIMPLE)dst = np.zeros_like(roi)for i,c in enumerate(contours):area = cv.contourArea(c)if area>200:cv.drawContours(dst,contours,i,255,-1)if DEBUG:cv.imshow("roi",dst)return dst;##提取边缘def fetch_edge(img):canny = cv.Canny(img,200,400)if DEBUG:cv.imshow("canny",canny)return canny##找出直线#先提取了黄色,然后又提取感兴趣的区域,在感兴趣的区域中,我们提取了图像的边缘信息,在这一步我们就需要#来找出视线中可能的车道线了def make_points(line_kb,height):# 求右边的平均斜率和截距lines_avg_kb = np.average(line_kb,axis=0);# y = kx + b 绘制右边的直线y1 = height*0.5;x1 = (y1 - lines_avg_kb[1]) / lines_avg_kb[0];y2 = 0x2 = (y2 - lines_avg_kb[1]) / lines_avg_kb[0];return [x1,y1+0.5*height,x2,y2+0.5*height]def fetch_lines(img,height,width):lines = cv.HoughLinesP(img,1,np.pi/180,10,minLineLength=40,maxLineGap=5);left_kb = [];right_kb = [];if lines is None:return [];boundary = 1/3left_region_boundary = width * (1 - boundary) # left lane line segment should be on left 2/3 of the screenright_region_boundary = width * boundary# 计算每一条线段的斜率和截距for line in lines:x1,y1,x2,y2 = line[0];# 计算当前线段的斜率和截距 y = kx + bif x1 == x2:continueparams = np.polyfit((x1,x2),(y1,y2),1);# 斜率k = params[0];# 截距b = params[1];if k < 0:if x1 < left_region_boundary and x2 < left_region_boundary:left_kb.append((k,b));else:if x1 > right_region_boundary and x2 > right_region_boundary:right_kb.append((k,b));lines = []if len(left_kb)>0:left_line = make_points(left_kb,480)lines.append(left_line)if len(right_kb)>0:right_line = make_points(right_kb,height)lines.append(right_line)return lines;#计算引导线def calc_head_line(lines,width,height):if len(lines) == 0:return None;x = width/2y = heightif len(lines) == 1:x1,y1,x2,y2 = lines[0]# params = np.polyfit((x1,x2),(y1,y2),1);# # 斜率# k = params[0];# # 截距# b = params[1];x_head = x2 + (x - x1)y_head = 0.5*height;# x_head = (y_head - b)/kreturn [x,y,x_head,y_head]else:_,_,line1_x2,_ = lines[0]_,_,line2_x2,_ = lines[1]x_head = (line1_x2 + line2_x2)/2y_head = 0.5*heightreturn [x,y,x_head,y_head]#为了方便观察,显示所有线的函数def show_lines(lines,frame,head_line=None):if DEBUG:try:line_img = np.zeros_like(frame)for x1,y1,x2,y2 in lines:cv.line(line_img,(int(x1),int(y1)),(int(x2),int(y2)),(0,255,0),10)if head_line:x1,y1,x2,y2 = head_linecv.line(line_img,(int(x1),int(y1)),(int(x2),int(y2)),(0,0,255),10)dst = cv.addWeighted(line_img,0.8,frame,1,0)cv.imshow("dst",dst)except Exception as e:print e##计算偏差角度#为了便于控制小车转向的角度大小,我们计算引导线偏离屏幕中央的度数. 若引导线偏离中线较大,则角速度要#增大,反之角速度要减小.这里其实也是为我们接下来视觉PID引导做铺垫def calc_angle(x_offset,y_offset):angle_to_mid_radian = math.atan(x_offset / y_offset) return angle_to_mid_radian##pid#并且在这个过程中,偏离的角度越大,小车的角速度也越大,反之越小.这个正好可以用PID来进行控制,基于这样#的设想,我们用上位机视觉PID来控制小车,我们的目标就是要把偏离角度控制在0附近intergral = 0;derivative = 0;prev_error = 0;pid = 0def pid_control(curr,target):global intergral,derivative,prev_error,piderror = target - currintergral += error;derivative = prev_error;# pid 计算公式kp = 0.2ki = 0.0kd = 0.02pid = kp*error + ki * intergral + kd* derivativeprev_error = error;print pid;return pidif __name__ == '__main__':# Give the node a namerospy.init_node('heima_linefollow', anonymous=False)# Set rospy to execute a shutdown function when terminating the scriptrospy.on_shutdown(shutdown)# How fast will we check the odometry values?rate = rospy.Rate(100)# Publisher to control the robot's speedcmd_vel = rospy.Publisher('/cmd_vel', Twist, queue_size=1)capture = cv.VideoCapture(0);ok,frame = capture.read()height,width = frame.shape[0:2]# while ok:while not rospy.is_shutdown():ok,frame = capture.read()# 提取黄色binary = fetch_yellow(frame)# 提取看兴趣的内容roi = fetch_roi(binary)# 提取边缘canny = fetch_edge(roi)# 提取线lines = fetch_lines(canny,height,width)head_line = calc_head_line(lines,width,height)if head_line:mid = width/2 ;_,_,curr,_ = head_linex_offset = curr - mid;y_offset = height*0.5angle = calc_angle(x_offset,y_offset)if angle > 60 or angle < -60:continueshow_lines(lines,frame,head_line)value = pid_control(angle,0)print "x_offset:{},{},{},angle={}".format(x_offset,mid,curr,angle)# 假设恒定线速度为0.25R = 0.1445/math.tan(value)V = 0.10W = V/R;twist = Twist()twist.linear.x = Vtwist.angular.z = Wcmd_vel.publish(twist)cv.imshow("frame",frame)cv.waitKey(1)rate.sleep()
=========================================================================================================================
远程控制slam小车及pid调试PC与树莓派ssh链接时出现间歇性联通段开网络故障acailable I Destination Host Unreachable_然后5s后切换了一个地图