一、环境介绍
Windows版本: Win10 X64
OpenCV版本: 2.4.13.6
QT版本: 5.12
二、OpenCV库下载
OpenCV官网下载地址: https://opencv.org
目前官网OpenCV最新的版本是4.2.0 ,Windows版本的OpenCV在3.X版本后就不带X86的库,只有X64的库,如果需要X86的库,需要自己下载源码去重新编译。
由于我的QT软件在安装时没有安装64位的编译器,又不想重新安装64位编译器,就选择了2.X的版本完成开发测试,3.X的版本在添加库的时候非常方便,就一个库。
库下载之后,双击运行,解压到指定的目录就行。
得到X86的库:
接下来要把OpenCV的动态库目录加到系统环境变量里,否则程序运行时找不到依赖库会崩溃。
三、程序思路说明
程序功能: 在子线程里打开摄像头,获取摄像头的数据,通过信号与槽的方式,将摄像头数据传递给主UI界面实时显示,在采用定时器每100ms取一次标签上的数据进行人脸检测处理,将处理的数据再显示到另一个标签上。人脸检测分类器采用OpenCV自带的分类器,程序主要目的是介绍OpenCV配合QT如何进行开发。
OpenCV自带的人脸检测分类器路径:C:/OpenCV_2.4/opencv/sources/data/haarcascades_GPU/haarcascade_frontalface_alt2.xml
四、实现效果
五、核心代码
xxx.pro工程文件代码:
代码语言:javascript复制QT = core gui
QT = multimediawidgets
QT = xml
QT = multimedia
QT = network
QT = widgets
QT = serialport
greaterThan(QT_MAJOR_VERSION, 4): QT = widgets
CONFIG = c 11
# The following define makes your compiler emit warnings if you use
# any Qt feature that has been marked deprecated (the exact warnings
# depend on your compiler). Please consult the documentation of the
# deprecated API in order to know how to port your code away from it.
DEFINES = QT_DEPRECATED_WARNINGS
# You can also make your code fail to compile if it uses deprecated APIs.
# In order to do so, uncomment the following line.
# You can also select to disable deprecated APIs only up to a certain version of Qt.
#DEFINES = QT_DISABLE_DEPRECATED_BEFORE=0x060000 # disables all the APIs deprecated before Qt 6.0.0
SOURCES =
main.cpp
widget.cpp
HEADERS =
widget.h
FORMS =
widget.ui
# Default rules for deployment.
qnx: target.path = /tmp/$${TARGET}/bin
else: unix:!android: target.path = /opt/$${TARGET}/bin
!isEmpty(target.path): INSTALLS = target
#linu平台的路径设置
linux {
message('运行linu版本')
#添加opencv头文件的路径,需要根据自己的头文件路径进行修改
INCLUDEPATH =/home/wbyq/work_pc/opencv-3.4.9/_install/install/include
/home/wbyq/work_pc/opencv-3.4.9/_install/install/include/opencv
/home/wbyq/work_pc/opencv-3.4.9/_install/install/include/opencv2
LIBS =/home/wbyq/work_pc/opencv-3.4.9/_install/install/lib/libopencv_*
}
win32
{
message('运行win32版本')
#添加opencv头文件的路径,需要根据自己的头文件路径进行修改
INCLUDEPATH =C:/OpenCV_2.4/opencv/build/include
C:/OpenCV_2.4/opencv/build/include/opencv
C:/OpenCV_2.4/opencv/build/include/opencv2
LIBS =C:/OpenCV_2.4/opencv/build/x86/vc14/lib/opencv_*
}
RESOURCES =
image.qrc
widget.cpp文件代码:
代码语言:javascript复制#include "widget.h"
#include "ui_widget.h"
class VideoAudioEncode videoaudioencode_0;
Widget::Widget(QWidget *parent)
: QWidget(parent)
, ui(new Ui::Widget)
{
ui->setupUi(this);
//驾驶室摄像头
//工作对象
videoRead_WorkClass_0=new VideoReadThread_0;
videoRead_Workthread_0=new QThread;
//连接摄像头采集信号,在主线程实时显示视频画面
connect(videoRead_WorkClass_0,SIGNAL(VideoDataOutput(QImage )),this,SLOT(VideoDataDisplay_0(QImage )));
//摄像头初始化函数
connect(this,SIGNAL(Init_VideoAudio_WorkClass_0()),videoRead_WorkClass_0,SLOT(run()));
//停止视频采集
connect(this,SIGNAL(Stop_AudioVideo0_work_0()),videoRead_WorkClass_0,SLOT(stop()));
//将工作对象移动到子线程里工作
videoRead_WorkClass_0->moveToThread(videoRead_Workthread_0);
//更新设备列表
UpdateVideoAudiodDevice(ui->comboBox_video_0,ui->plainTextEdit_log_0);
//
timer.start(100);
connect(&timer,SIGNAL(timeout()), this, SLOT(update()));
timer.start(100);
}
Widget::~Widget()
{
delete ui;
}
//分类器的路径
#define source_xml_addr "C:/OpenCV_2.4/opencv/sources/data/haarcascades_GPU/haarcascade_frontalface_alt2.xml"
//将要检测的图片路径
#define source_pix_addr "D:/linux-share-dir/1.jpg"
//人脸检测代码
void Widget::opencv_face(QImage qImage)
{
QTime time;
time.start();
static CvMemStorage* storage = 0;
static CvHaarClassifierCascade* cascade = 0;
const char*cascade_name =source_xml_addr;
//加载分类器
cascade = (CvHaarClassifierCascade*)cvLoad( cascade_name, 0, 0, 0 );
if( !cascade )
{
Log_Display_0("分类器加载错误.n");
return ;
}
//创建内存空间
storage = cvCreateMemStorage(0);
//加载需要检测的图片
//const char* filename =source_pix_addr;
//IplImage* img = cvLoadImage(filename, 1 );
IplImage* img = QImageToIplImage(&qImage);
if(img ==nullptr )
{
Log_Display_0("图片加载错误.n");
return;
}
double scale=1.2;
static CvScalar colors[] = {
{{0,0,255}},{{0,128,255}},{{0,255,255}},{{0,255,0}},
{{255,128,0}},{{255,255,0}},{{255,0,0}},{{255,0,255}}
};//Just some pretty colors to draw with
IplImage* gray = cvCreateImage(cvSize(img->width,img->height),8,1);
IplImage* small_img=cvCreateImage(cvSize(cvRound(img->width/scale),cvRound(img->height/scale)),8,1);
cvCvtColor(img,gray, CV_BGR2GRAY);
cvResize(gray, small_img, CV_INTER_LINEAR);
cvEqualizeHist(small_img,small_img); //直方图均衡
cvClearMemStorage(storage);
double t = (double)cvGetTickCount();
CvSeq* objects = cvHaarDetectObjects(small_img,
cascade,
storage,
1.1,
2,
0/*CV_HAAR_DO_CANNY_PRUNING*/,
cvSize(30,30));
t = (double)cvGetTickCount() - t;
//遍历找到对象和周围画盒
for(int i=0;i<(objects->total); i)
{
CvRect* r=(CvRect*)cvGetSeqElem(objects,i);
cvRectangle(img, cvPoint(r->x*scale,r->y*scale), cvPoint((r->x r->width)*scale,(r->y r->height)*scale), colors[i%8]);
}
for( int i = 0; i < (objects? objects->total : 0); i )
{
CvRect* r = (CvRect*)cvGetSeqElem( objects, i );
CvPoint center;
int radius;
center.x = cvRound((r->x r->width*0.5)*scale);
center.y = cvRound((r->y r->height*0.5)*scale);
radius = cvRound((r->width r->height)*0.25*scale);
cvCircle(img, center, radius, colors[i%8], 3, 8, 0 );
}
show_face(img); //显示检测的结果
cvReleaseImage(&gray);
cvReleaseImage(&small_img);
//释放图片
cvReleaseImage(&img);
Log_Display_0(tr("耗时:%1 msn").arg(time.elapsed()));
}
/*将QImage图片转为opecv的qimage格式*/
IplImage *Widget::QImageToIplImage(const QImage * qImage)
{
int width = qImage->width();
int height = qImage->height();
CvSize Size;
Size.height = height;
Size.width = width;
IplImage *IplImageBuffer = cvCreateImage(Size, IPL_DEPTH_8U, 3);
for (int y = 0; y < height; y)
{
for (int x = 0; x < width; x)
{
QRgb rgb = qImage->pixel(x, y);
CV_IMAGE_ELEM( IplImageBuffer, uchar, y, x*3 0 ) = qBlue(rgb);
CV_IMAGE_ELEM( IplImageBuffer, uchar, y, x*3 1 ) = qGreen(rgb);
CV_IMAGE_ELEM( IplImageBuffer, uchar, y, x*3 2 ) = qRed(rgb);
}
}
return IplImageBuffer;
}
/*将opecv的图片转为qimage格式*/
QImage *Widget::IplImageToQImage(IplImage *img)
{
QImage *qmg;
uchar *imgData=(uchar *)img->imageData;
qmg = new QImage(imgData,img->width,img->height,QImage::Format_RGB888);
*qmg=qmg->rgbSwapped(); //BGR格式转RGB
return qmg;
}
//显示检测的结果
void Widget::show_face(IplImage* img)
{
uchar *imgData=(uchar *)img->imageData;
QImage my_image=QImage(imgData,img->width,img->height,QImage::Format_RGB888);
my_image=my_image.rgbSwapped(); //BGR格式转RGB
QPixmap my_pix; //创建画图类
my_pix.convertFromImage(my_image);
/*在控件上显示*/
ui->label_display->setPixmap(my_pix);
}
//开始采集
void Widget::on_pushButton_Start_clicked()
{
//设置当前选择的摄像头
videoaudioencode_0.camera=video_dev_list.at(ui->comboBox_video_0->currentIndex());
Stop_VideoAudioEncode_0(true);
Start_VideoAudioEncode_Thread_0();
}
//析构函数
VideoReadThread_0::~VideoReadThread_0()
{
}
//停止视频采集
void VideoReadThread_0::stop()
{
qDebug()<<"停止视频采集--stop";
if(camera)
{
camera->stop();
delete camera;
camera=nullptr;
}
if(m_pProbe)
{
delete m_pProbe;
m_pProbe=nullptr;
}
}
//执行线程
void VideoReadThread_0::run()
{
stop();
Camear_Init();
qDebug()<<"摄像头开始采集数据";
}
void VideoReadThread_0::Camear_Init()
{
/*创建摄像头对象,根据选择的摄像头打开*/
camera = new QCamera(videoaudioencode_0.camera);
m_pProbe = new QVideoProbe;
if(m_pProbe != nullptr)
{
m_pProbe->setSource(camera); // Returns true, hopefully.
connect(m_pProbe, SIGNAL(videoFrameProbed(QVideoFrame)),this, SLOT(slotOnProbeFrame(QVideoFrame)), Qt::QueuedConnection);
}
/*配置摄像头捕 QCamera *camera;
QVideoProbe *m_pProbe;获模式为帧捕获模式*/
//camera->setCaptureMode(QCamera::CaptureStillImage); //如果在Linux系统下运行就这样设置
camera->setCaptureMode(QCamera::CaptureVideo);//如果在android系统下运行就这样设置
/*启动摄像头*/
camera->start();
/*设置摄像头的采集帧率和分辨率*/
QCameraViewfinderSettings settings;
settings.setPixelFormat(QVideoFrame::Format_YUYV); //设置像素格式 Android上只支持NV21格式
settings.setResolution(QSize(VIDEO_WIDTH,VIDEO_HEIGHT)); //设置摄像头的分辨率
camera->setViewfinderSettings(settings);
}
/**
* NV21是android相机默认格式
* @param data
* @param rgb
* @param width
* @param height
*/
void NV21_TO_RGB24(unsigned char *yuyv, unsigned char *rgb, int width, int height)
{
const int nv_start = width * height ;
int index = 0, rgb_index = 0;
uint8_t y, u, v;
int r, g, b, nv_index = 0,i, j;
for(i = 0; i < height; i ){
for(j = 0; j < width; j ){
//nv_index = (rgb_index / 2 - width / 2 * ((i 1) / 2)) * 2;
nv_index = i / 2 * width j - j % 2;
y = yuyv[rgb_index];
u = yuyv[nv_start nv_index ];
v = yuyv[nv_start nv_index 1];
r = y (140 * (v-128))/100; //r
g = y - (34 * (u-128))/100 - (71 * (v-128))/100; //g
b = y (177 * (u-128))/100; //b
if(r > 255) r = 255;
if(g > 255) g = 255;
if(b > 255) b = 255;
if(r < 0) r = 0;
if(g < 0) g = 0;
if(b < 0) b = 0;
index = rgb_index % width (height - i - 1) * width;
//rgb[index * 3 0] = b;
//rgb[index * 3 1] = g;
//rgb[index * 3 2] = r;
//颠倒图像
//rgb[height * width * 3 - i * width * 3 - 3 * j - 1] = b;
//rgb[height * width * 3 - i * width * 3 - 3 * j - 2] = g;
//rgb[height * width * 3 - i * width * 3 - 3 * j - 3] = r;
//正面图像
rgb[i * width * 3 3 * j 0] = b;
rgb[i * width * 3 3 * j 1] = g;
rgb[i * width * 3 3 * j 2] = r;
rgb_index ;
}
}
}
/*
函数功能: 将YUV数据转为RGB格式
函数参数:
unsigned char *yuv_buffer: YUV源数据
unsigned char *rgb_buffer: 转换之后的RGB数据
int iWidth,int iHeight : 图像的宽度和高度
*/
void yuyv_to_rgb(unsigned char *yuv_buffer,unsigned char *rgb_buffer,int iWidth,int iHeight)
{
int x;
int z=0;
unsigned char *ptr = rgb_buffer;
unsigned char *yuyv= yuv_buffer;
for (x = 0; x < iWidth*iHeight; x )
{
int r, g, b;
int y, u, v;
if (!z)
y = yuyv[0] << 8;
else
y = yuyv[2] << 8;
u = yuyv[1] - 128;
v = yuyv[3] - 128;
r = (y (359 * v)) >> 8;
g = (y - (88 * u) - (183 * v)) >> 8;
b = (y (454 * u)) >> 8;
*(ptr ) = (r > 255) ? 255 : ((r < 0) ? 0 : r);
*(ptr ) = (g > 255) ? 255 : ((g < 0) ? 0 : g);
*(ptr ) = (b > 255) ? 255 : ((b < 0) ? 0 : b);
if(z )
{
z = 0;
yuyv = 4;
}
}
}
void VideoReadThread_0::slotOnProbeFrame(const QVideoFrame &frame)
{
QVideoFrame cloneFrame(frame);
cloneFrame.map(QAbstractVideoBuffer::ReadOnly);
//qDebug()<<"height:"<<cloneFrame.height();
//qDebug()<<"width:"<<cloneFrame.width();
//qDebug()<<"bytesPerLine:"<<cloneFrame.bytesPerLine();
//qDebug()<<"mappedBytes:"<<cloneFrame.mappedBytes();
//qDebug()<<"pixelFormat:"<<cloneFrame.pixelFormat();
unsigned char rgb_buffer[VIDEO_WIDTH*VIDEO_HEIGHT*3];
if(cloneFrame.pixelFormat()==QVideoFrame::Format_NV21)
{
NV21_TO_RGB24(cloneFrame.bits(),rgb_buffer,cloneFrame.width(),cloneFrame.height());
}
else if(cloneFrame.pixelFormat()==QVideoFrame::Format_YUYV)
{
yuyv_to_rgb(cloneFrame.bits(),rgb_buffer,cloneFrame.width(),cloneFrame.height());
}
else
{
qDebug("当前格式编码为%1,暂时不支持转换.n");
}
cloneFrame.unmap();
//加载图片数据
QImage image(rgb_buffer,
cloneFrame.width(),
cloneFrame.height(),
QImage::Format_RGB888);
//绘制图片水印
QDateTime dateTime(QDateTime::currentDateTime());
//时间效果: 2020-03-05 16:25::04 周一
QString qStr="";
qStr =dateTime.toString("yyyy-MM-dd hh:mm:ss ddd");
QPainter pp(&image);
QPen pen = QPen(Qt::white);
pp.setPen(pen);
pp.drawText(QPointF(0,20),qStr);
//提取RGB数据
unsigned char *p=rgb_buffer;
for(int i=0;i<image.height();i )
{
for(int j=0;j<image.width();j )
{
QRgb rgb=image.pixel(j,i);
*p =qRed(rgb);
*p =qGreen(rgb);
*p =qBlue(rgb);
}
}
emit VideoDataOutput(image); //发送信号
}
//驾驶室:视频刷新显示
void Widget::VideoDataDisplay_0(QImage image)
{
QPixmap my_pixmap;
my_pixmap.convertFromImage(image);
/*在控件上显示*/
ui->label_ImageDisplay_0->setPixmap(my_pixmap);
}
//驾驶室:日志显示
void Widget::Log_Display_0(QString text)
{
Log_Text_Display(ui->plainTextEdit_log_0,text);
}
/*日志显示*/
void Widget::Log_Text_Display(QPlainTextEdit *plainTextEdit_log,QString text)
{
plainTextEdit_log->insertPlainText(text);
//移动滚动条到底部
QScrollBar *scrollbar = plainTextEdit_log->verticalScrollBar();
if(scrollbar)
{
scrollbar->setSliderPosition(scrollbar->maximum());
}
}
//驾驶室:开启所有采集线程
void Widget::Start_VideoAudioEncode_Thread_0()
{
videoRead_Workthread_0->start(); //开启视频采集线程
Init_VideoAudio_WorkClass_0(); //发送初始化信号
}
//驾驶室:退出所有采集线程
void Widget::Stop_VideoAudioEncode_0(bool flag)
{
if(flag==true)
{
Stop_AudioVideo0_work_0(); //发送信号停止摄像头
QThread::msleep(10);
//退出视频采集
videoRead_Workthread_0->quit(); //告诉线程的事件循环以return 0(成功)退出
videoRead_Workthread_0->wait(); //等待线程退出
}
}
void Widget::on_pushButton_update_clicked()
{
UpdateVideoAudiodDevice(ui->comboBox_video_0,ui->plainTextEdit_log_0);
}
/*刷新本机可以的音频设备列表*/
void Widget::UpdateVideoAudiodDevice(
QComboBox *comboBox_video,
QPlainTextEdit *plainTextEdit_log)
{
/*2. 获取摄像头列表*/
video_dev_list.clear();
comboBox_video->clear();
video_dev_list=QCameraInfo::availableCameras();
for(int i=0;i<video_dev_list.size();i )
{
comboBox_video->addItem(video_dev_list.at(i).deviceName());
}
/*如果没有可用的摄像头设备,按钮不可用*/
if(video_dev_list.size()==0)
{
Log_Text_Display(plainTextEdit_log,"未查询到可用的摄像头设备.n");
}
}
//停止采集
void Widget::on_pushButton_stop_clicked()
{
Stop_VideoAudioEncode_0(true);
}
void Widget::update()
{
if(ui->label_ImageDisplay_0->pixmap())
opencv_face(ui->label_ImageDisplay_0->pixmap()->toImage());
}
widget.h文件代码:
代码语言:javascript复制#ifndef WIDGET_H
#define WIDGET_H
#include <QWidget>
//opencv include
#include <cv.h>
#include <cxcore.h>
#include <highgui.h>
#include <QCamera>
#include <QVideoProbe>
#include <QAbstractVideoSurface>
#include <QVideoProbe>
#include <QThread>
#include <QApplication>
#include <QDebug>
#include <QObject>
#include <QMutex>
#include <QMutexLocker>
#include <QWaitCondition>
#include <QQueue>
#include <QCamera>
#include <QPen>
#include <QPainter>
#include <QRgb>
#include <QDateTime>
#include <QCameraInfo>
#include <QPlainTextEdit>
#include <QScrollBar>
#include <QComboBox>
#include <QTimer>
QT_BEGIN_NAMESPACE
namespace Ui { class Widget; }
QT_END_NAMESPACE
class Widget : public QWidget
{
Q_OBJECT
public:
Widget(QWidget *parent = nullptr);
void opencv_face(QImage qImage);
~Widget();
QList<QCameraInfo> video_dev_list;
void show_face(IplImage* img);
class VideoReadThread_0 *videoRead_WorkClass_0; //视频工作类
QThread *videoRead_Workthread_0; //视频线程
/*驾驶室摄像头*/
void Stop_VideoAudioEncode_0(bool flag);//停止线程
void Start_VideoAudioEncode_Thread_0(); //启动线程
void Log_Text_Display(QPlainTextEdit *plainTextEdit_log,QString text);
void UpdateVideoAudiodDevice(QComboBox *comboBox_video, QPlainTextEdit *plainTextEdit_log);
IplImage *QImageToIplImage(const QImage * qImage);
QImage *IplImageToQImage(IplImage *img);
QTimer timer;
signals:
void Init_VideoAudio_WorkClass_0();
void Stop_AudioVideo0_work_0();
private slots:
void update();
void on_pushButton_Start_clicked();
void Log_Display_0(QString text);
void VideoDataDisplay_0(QImage );
void on_pushButton_update_clicked();
void on_pushButton_stop_clicked();
private:
Ui::Widget *ui;
};
class VideoReadThread_0:public QObject
{
Q_OBJECT
public:
QCamera *camera;
QVideoProbe *m_pProbe;
VideoReadThread_0(QObject* parent=nullptr):QObject(parent){camera=nullptr;m_pProbe=nullptr;}
~VideoReadThread_0();
void Camear_Init(void);
public slots:
void stop();
void run();
void slotOnProbeFrame(const QVideoFrame &frame);
signals:
void VideoDataOutput(QImage); //输出信号
};
//视频音频编码类
class VideoAudioEncode
{
public:
QCameraInfo camera; //当前选择的摄像头
};
//视频输出尺寸
#define VIDEO_WIDTH 640
#define VIDEO_HEIGHT 480
extern class VideoAudioEncode videoaudioencode_0;
#endif // WIDGET_H
main.cpp文件代码:
代码语言:javascript复制#include "widget.h"
#include <QApplication>
int main(int argc, char *argv[])
{
QApplication a(argc, argv);
Widget w;
w.show();
return a.exec();
}
UI界面文件: