python部分
#!/usr/bin/env Python # coding=utf-8 from ctypes import * from PyQt5.QtCore import * from PyQt5.QtGui import * from PyQt5.QtWidgets import * import time import numpy as np import cv2 import struct import datetime from numba import jit import os cam_dict={} class CamInfo: def __init__(self, cam_no=0,deviceid="default",cam_name="default"): self.cam_no = cam_no self.deviceid = deviceid self.cam_name = cam_name @jit def trans(data,size,height,width): bbb = string_at(data,size) nparr = np.fromstring(bbb, np.uint8) r = nparr.reshape(height,width, 3) return r def str2char_p(str_v): pStr = c_char_p( ) pStr.value = str_v return pStr def callb_stream(data,size,cam_no,height,width): r = trans(data, size,height,width) r = cv2.cvtColor(r, cv2.COLOR_RGB2BGR) counter = datetime.datetime.now().strftime('%Y%m%d_%H%M%S_%f') # print(1, counter) cv2.imshow(str(cam_no), r) cv2.waitKey(1) def callb_camerainfo(cam_no,camera_info,camera_info_size): # print(cast(camera_info,c_char_p).value) # print(str(cast(camera_info, c_char_p).value)) bbb = string_at(camera_info, camera_info_size) info=str(bbb,encoding="utf-8").split(",") cam_dict[cam_no]= CamInfo(cam_no,info[1],info[2]) print("camerainfo",cam_dict[cam_no].cam_no,cam_dict[cam_no].cam_name,cam_dict[cam_no].deviceid) class Mythread(QThread): # 定义信号,定义参数为str类型 breakSignal = pyqtSignal(str,list) def __init__(self, parent=None): super().__init__(parent) # super(Mythread, self).__init__() def callb_error(self, err_type, cam_no, msg_no, msg_level, msg_txt, msg_txtlen): print("myerror", err_type, cam_no, msg_no, msg_level, msg_txt, msg_txtlen) def run(self): dll = CDLL(r"./hik_client.dll") width=60 height=40 dll.pre_encode.restype = c_void_p ret=dll.pre_encode(width,height) ret=cast(ret,c_void_p) for i in range(20000): n=i%200+1 img=cv2.imread("bmp/"+str(n)+".bmp") len = img.shape[0] * img.shape[1] * img.shape[2] # img=np.transpose(img,(1, 0, 2)) # b, g, r = cv2.split(img) # b = b.reshape(-1) # g = g.reshape(-1) # r = r.reshape(-1) # b = np.append(b, g) # img = np.append(b, r) img = img.reshape(-1) # b, g, r = cv2.split(img) # b = b.reshape(-1) # g = g.reshape(-1) # r = r.reshape(-1) # b = np.append(b,g) # img = np.append(b, r) INPUT = c_int * len # 实例化一个长度为2的整型数组 input = INPUT() # 为数组赋值(input这个数组是不支持迭代的) for i in range(len): input[i] = img[i] # bytes(aaaa, encoding="utf-8") a = dll.push_rtsp(input,len,ret) QCoreApplication.instance().quit() # print("encode_ok",i) # b = string_at(a, 1280*720*3) # print(b) # nparr = np.fromstring(b, np.uint8) # # print(nparr[-10:-1],min(nparr),max(nparr)) # img_decode = cv2.imdecode(nparr, cv2.IMREAD_COLOR) # # if img_decode: # cv2.imshow("sadf", img_decode) # # cv2.imwrite(str(index)+".jpg", img_decode) # cv2.waitKey(0) # # # # # ErrorCall = CFUNCTYPE(c_void_p, c_int, c_int, c_int,c_int,c_char_p,c_int) # error_callback = ErrorCall(self.callb_error) # dll.set_callback(error_callback) # CamCall = CFUNCTYPE(c_void_p, c_int, c_char_p,c_int) # caminfo_CamCall = CamCall(callb_camerainfo) # # print(b) # if not os.path.exists("video"): # os.makedirs("video") # ip = b"127.0.0.1" # port = 8888 # print("start conn") # ret=-1 # while(ret): # print("conn server...") # ret= dll.tcp_init(str2char_p(ip), port) # time.sleep(0.3) # if (ret==0): # type=1 # ret = dll.getcameralist(type, caminfo_CamCall) # if (1): # # deviceId = b"af94a68df0124d1fbf0fc2b07f3b3c3a" # cam_no=14 # else: # print("tcp error") # for i in range(2000000): # # 发出信号 # a=[i,i+1] # self.breakSignal.emit(str(i),a) # # 让程序休眠 # time.sleep(0.5) if __name__ == '__main__': app = QApplication([]) dlg = QDialog() dlg.resize(400, 300) dlg.setWindowTitle("自定义按钮测试") dlgLayout = QVBoxLayout() dlgLayout.setContentsMargins(40, 40, 40, 40) btn = QPushButton('测试按钮') dlgLayout.addWidget(btn) dlgLayout.addStretch(40) dlg.setLayout(dlgLayout) dlg.show() def chuli(a,s): # dlg.setWindowTitle(s) btn.setText(a+str(s[0]*10)) # 创建线程 thread = Mythread() # # 注册信号处理函数 thread.breakSignal.connect(chuli) # # 启动线程 thread.start() dlg.exec_() app.exit()
c++动态库部分
#include "stdafx.h" #include "CVdll.h" #include "SimpleLog.h" #include <iostream> #include<fstream> #include <sys/types.h> #include "opencv2/opencv.hpp" #include "Ws2tcpip.h" #include <winsock2.h> #include <fcntl.h> #include <cstring> #include <cstdio> #include <signal.h> #pragma comment(lib,"ws2_32.lib") #include <queue> using namespace cv; extern "C" { #include "libavcodec/avcodec.h" #include "libavformat/avformat.h" #include "libswscale/swscale.h" #include "libavdevice/avdevice.h" #include "libavutil/log.h" //#include "libavutil/imgutils.h" }; //说明,动态库需要拷贝三个文件,否则重连会出问题 char* testchar(int plus1) { char* str = "hello world111111"; return str; } char* testimg(char* data,int length) { char* str = "hello world111111"; return str; } int outbuf_size = 100000; class Rtmp_tool { public: int nWidth = 0; int nHeight = 0; AVCodecContext *c; AVFrame *m_pRGBFrame = new AVFrame[1]; //RGB帧数据 AVFrame *m_pYUVFrame = new AVFrame[1];; //YUV帧数据 uint8_t * yuv_buff;// uint8_t * outbuf; SwsContext * scxt; FILE *f = NULL; }; void* pre_encode(int width,int height) { Rtmp_tool *rtmp_tool; rtmp_tool = new Rtmp_tool(); int nLen; int fileI; rtmp_tool->nWidth = width; rtmp_tool->nHeight = height; av_register_all(); avcodec_register_all(); //AVFrame *m_pRGBFrame = new AVFrame[1]; //RGB帧数据 //AVFrame *m_pYUVFrame = new AVFrame[1];; //YUV帧数据 AVCodecContext *c = NULL; AVCodecContext *in_c = NULL; AVCodec *pCodecH264; //编码器 //查找h264编码器 pCodecH264 = avcodec_find_encoder(AV_CODEC_ID_H264); c = avcodec_alloc_context3(pCodecH264); c->bit_rate = 3000000;// put sample parameters c->width = width;// c->height = height;// // frames per second AVRational rate; rate.num = 1; rate.den = 5; c->time_base = rate;//(AVRational){1,25}; c->gop_size = 10; // emit one intra frame every ten frames c->max_b_frames = 1; c->thread_count = 1; c->pix_fmt = AV_PIX_FMT_YUV420P;//PIX_FMT_RGB24; //av_opt_set(c->priv_data, /*"preset"*/"libvpx-1080p.ffpreset", /*"slow"*/NULL, 0); //打开编码器 if (avcodec_open2(c, pCodecH264, NULL)<0) printf("不能打开编码库"); int size = c->width * c->height; rtmp_tool->yuv_buff = (uint8_t *)malloc((size * 3) / 2); // size for YUV 420 //图象编码 rtmp_tool->outbuf = (uint8_t*)malloc(outbuf_size); int u_size = 0; const char * filename = "0_Data.h264"; rtmp_tool->f = fopen(filename, "wb"); if (!rtmp_tool->f) { printf("could not open %s\n", filename); exit(1); } //初始化SwsContext rtmp_tool->scxt = sws_getContext(c->width, c->height, AV_PIX_FMT_BGR24, c->width, c->height, AV_PIX_FMT_YUV420P, SWS_POINT, NULL, NULL, NULL); rtmp_tool->c = c; return rtmp_tool; } char* push_rtsp(int* plus1,int len,void* vp) { Rtmp_tool *rtmp_tool =(Rtmp_tool *) vp; for (int i = 0; i < len; i++) { plus1[i] = (uint8_t)plus1[i]; } AVCodecContext *c = rtmp_tool->c;// (AVCodecContext*)vp; printf("2 %d %d\n", c->width, c->height); //--------------- AVPacket avpkt; AVFrame *m_pRGBFrame = rtmp_tool->m_pRGBFrame; AVFrame *m_pYUVFrame = rtmp_tool->m_pYUVFrame; /*unsigned char *pBmpBuf; pBmpBuf = new unsigned char[len];*/ //memcpy(rgb_buff, (uint8_t*)plus1, nDataLen); // avpicture_fill((AVPicture*)m_pRGBFrame, (uint8_t*)plus1, AV_PIX_FMT_RGB24, rtmp_tool->nWidth, rtmp_tool->nHeight); m_pRGBFrame->linesize[0] = c->width * 3; m_pRGBFrame->linesize[1] =0; m_pRGBFrame->linesize[2] =0; m_pRGBFrame->linesize[3] =0; m_pRGBFrame->format = AV_PIX_FMT_RGB24; m_pRGBFrame->width = rtmp_tool->nWidth; m_pRGBFrame->height = rtmp_tool->nHeight; uint8_t *p = m_pRGBFrame->data[0]; int y = 0, x = 0; for (y = 0; y < rtmp_tool->nHeight; y++) { for (x = 0; x < rtmp_tool->nWidth; x++) { *p++ = (uint8_t)plus1[(y*rtmp_tool->nWidth + x) * 3]; // R *p++ = (uint8_t)plus1[(y*rtmp_tool->nWidth + x) * 3 +1]; // G *p++ = (uint8_t)plus1[(y*rtmp_tool->nWidth + x) * 3 +2] ; // B } } printf("1 %d %d \n", rtmp_tool->nWidth, rtmp_tool->nHeight); //将YUV buffer 填充YUV Frame avpicture_fill((AVPicture*)m_pYUVFrame, (uint8_t*)rtmp_tool->yuv_buff, AV_PIX_FMT_YUV420P, rtmp_tool->nWidth, rtmp_tool->nHeight); // 翻转RGB图像 //m_pRGBFrame->data[0] += m_pRGBFrame->linesize[0] * (rtmp_tool->nHeight - 1); //m_pRGBFrame->linesize[0] *= -1; //m_pRGBFrame->data[1] += m_pRGBFrame->linesize[1] * (rtmp_tool->nHeight / 2 - 1); //m_pRGBFrame->linesize[1] *= -1; //m_pRGBFrame->data[2] += m_pRGBFrame->linesize[2] * (rtmp_tool->nHeight / 2 - 1); //m_pRGBFrame->linesize[2] *= -1; //将RGB转化为YUV sws_scale(rtmp_tool->scxt, m_pRGBFrame->data, m_pRGBFrame->linesize, 0, c->height, m_pYUVFrame->data, m_pYUVFrame->linesize); int got_packet_ptr = 0; av_init_packet(&avpkt); avpkt.data = rtmp_tool->outbuf; avpkt.size = outbuf_size; int u_size = avcodec_encode_video2(c, &avpkt, m_pYUVFrame, &got_packet_ptr); m_pYUVFrame->pts++; if (u_size == 0){ int res = fwrite(avpkt.data, 1, avpkt.size, rtmp_tool->f); if (res == 0) { printf("000"); } else { printf("1253"); } } //-------end--------- //Mat mat; ////加载图片 //mat = imread("bgs.jpg", CV_LOAD_IMAGE_COLOR); //printf("a %d %d", mat.rows, mat.cols); ////if (!mat.empty()) { //int m, n; //n = mat.cols * 3; //m = mat.rows; //unsigned char *data = (unsigned char*)malloc(sizeof(unsigned char) * m * n); //int p = 0; //for (int i = 0; i < m; i++) //{ // for (int j = 0; j < n; j++) // { // data[p] = mat.at<unsigned char>(i, j); // p++; // } //} //*plus1 = p; return NULL; //return (char*)data; } struct RecStruct //数据包 { int size; int data_type; int cam_no; int error_code; char recvbuf[1500]; }; struct SendStcuct { int size; int data_type; int cam_no; char sendbuf[1000]; }data_send; static ErrorCallBack g_errorcall = 0; static CamInfoCallBack g_caminfocall = 0; typedef struct CameraInfo { std::ofstream foutV; int timeInHour = -1; }caminfo; std::map<int, CameraInfo*> cameraMap; //static std::map<int, queue<RecStruct* namemap; static SOCKET g_sockClient; HANDLE hMutex; //char* deviceId; /**判断str1是否以str2开头 * 如果是返回1 * 不是返回0 * 出错返回-1 * */ int is_begin_with(const char * str1, char *str2) { if (str1 == NULL || str2 == NULL) return -1; int len1 = strlen(str1); int len2 = strlen(str2); if ((len1 < len2) || (len1 == 0 || len2 == 0)) return -1; char *p = str2; int i = 0; while (*p != '\0') { if (*p != str1[i]) return 0; p++; i++; } return 1; } char* Strcpy(char * a, const char * b) { int i = 0; while (*b) a[i++] = *b++; a[i] = 0; return a; } int sendcmd(char* data, int cam_no, int type, int size) { memset(data_send.sendbuf, 0, 1000); //data2send.sendbuf = new char[strlen(data)]; //memset(data2send.sendbuf, 0, strlen(data)); data_send.size = size; data_send.data_type = type; data_send.cam_no = cam_no; memcpy(data_send.sendbuf, data, sizeof(char) * (size)); printf("data_send len %d\n", sizeof(data_send)); if (g_sockClient) send(g_sockClient, (char *)&data_send, sizeof(struct SendStcuct), 0); return 0; } int set_callback(ErrorCallBack terrorcall(int error_type, int cam_no, int err_no, int msg_level, char* msg_txt, int spare)) { g_errorcall = (ErrorCallBack)terrorcall; return 0; } MYLIBDLL int getcameralist(int type, CamInfoCallBack caminfocall(int cam_no, char* cam_info, int cam_info_size)) { g_caminfocall = (CamInfoCallBack)caminfocall; SendStcuct data_send; memset(&data_send, 0, sizeof(struct SendStcuct)); data_send.size = 20; data_send.data_type = 1; data_send.cam_no = 0; char* data = "getcameralist"; memcpy(data_send.sendbuf, data, sizeof(char) * (20)); printf("data_send len %d\n", sizeof(data_send)); if (g_sockClient) send(g_sockClient, (char *)&data_send, sizeof(struct SendStcuct), 0); return 0; } DWORD WINAPI RecvThread(LPVOID lpParameter); DWORD WINAPI RecvThread(LPVOID lpParameter){ SOCKET sockClient = (SOCKET)lpParameter; while (1) { RecStruct data_recv; int ret; memset(&data_recv, '0', sizeof(struct RecStruct)); ret = recv(sockClient, (char *)&data_recv, sizeof(struct RecStruct), 0); //第二个参数使用强制类型,为一个数据包 if (ret == 0) // server调用了close { printf("server close"); break; } else if (ret == SOCKET_ERROR) // 网络错误 { int err = WSAGetLastError(); printf("get message %d %d %d \n", ret, SOCKET_ERROR, err); if (err == WSAECONNRESET || err == WSAECONNABORTED) { printf("tcp error %d %d \n", err, SOCKET_ERROR); //int n = namemap.erase(deviceId);//如果删除了会返回1,否则返回0 } break; } //printf("reve type %d %d", data_recv.data_type, data_recv.size); switch (data_recv.data_type) { case 1://摄像头列表 { g_caminfocall(data_recv.cam_no, data_recv.recvbuf, data_recv.size); } break; case 3://异常信息 { if (g_errorcall!=0) g_errorcall(1, 1, data_recv.error_code,4,NULL,0); break; } case 2: { char* recemsg = data_recv.recvbuf; int is_null = is_begin_with(recemsg, "00000"); if (is_null == 1) { printf("recv type 2 00000"); continue; } break; } default: break; } if (ret < 0) { printf("WSAStartup() failed!\n"); return -1; } Sleep(20); } return 0; } int tcpInit(char* ip, int port) { av_log_set_level(AV_LOG_PANIC); WSADATA wsaData; if (WSAStartup(MAKEWORD(2, 2), &wsaData) != 0) { printf("初始化Winsock失败"); return -1; } SOCKADDR_IN addrSrv; addrSrv.sin_family = AF_INET; addrSrv.sin_port = htons(port); SOCKET sockClient = socket(AF_INET, SOCK_STREAM, 0); int nRecvBuf = 0;//设置为32K setsockopt(sockClient, SOL_SOCKET, SO_RCVBUF, (const char*)&nRecvBuf, sizeof(int)); //setsockopt(sockClient, SOL_SOCKET, SO_RCVBUF, (char *)&nZero, sizeof(nZero)); inet_pton(AF_INET, ip, &addrSrv.sin_addr.s_addr); if (connect(sockClient, (struct sockaddr*)&addrSrv, sizeof(addrSrv)) == -1) return -2; //throw "连接失败"; if (SOCKET_ERROR == sockClient) { printf("Socket() error:%d", WSAGetLastError()); return -3; } g_sockClient = sockClient; HANDLE h_thread = CreateThread(NULL, 0, RecvThread, (LPVOID)sockClient, 0, NULL); CloseHandle(h_thread); return 0; } class DeviceInfo { public: string cam_name; int cam_no; SOCKET sockClient; }; //Callback int read_buffer(void *opaque, uint8_t *buf, int buf_size) { DeviceInfo deviceInfo = *((DeviceInfo *)opaque); int null_count=0; int display_count = 0; while (1) { RecStruct data_recv; int ret; memset(&data_recv, '0', sizeof(struct RecStruct)); ret = recv(deviceInfo.sockClient, (char *)&data_recv, sizeof(struct RecStruct), 0); //第二个参数使用强制类型,为一个数据包 if (ret == 0) // server调用了close { printf("server close"); break; } else if (ret == SOCKET_ERROR) // 网络错误 { printf("get message %d %d \n", ret, SOCKET_ERROR); int err = WSAGetLastError(); if (g_errorcall != 0) g_errorcall(1, deviceInfo.cam_no, err,4,"socket err",0); //if (err == WSAECONNRESET || err == WSAECONNABORTED) { // printf("server break %s", deviceId); // //int n = namemap.erase(deviceId);//如果删除了会返回1,否则返回0 //} break; } if (data_recv.size == 0) { null_count++; if (null_count %1000==0) { if (g_errorcall != 0) g_errorcall(1, deviceInfo.cam_no, 0, 2, "data_recv 0", 11); printf("reve len=0 type %d\n", data_recv.data_type); null_count = 0; } Sleep(2); continue; } else if (data_recv.size >1500) { if (g_errorcall != 0) g_errorcall(1, deviceInfo.cam_no, 0, 2, "data_recv too long", data_recv.size); printf("reve data too long %d\n", data_recv.size); continue; } if (data_recv.data_type == 3) { if (g_errorcall) { char err_str[10]; _itoa(data_recv.error_code, err_str, 10); //正确解法一 g_errorcall(1, deviceInfo.cam_no, data_recv.error_code, 4, err_str, 0); } } else if (data_recv.data_type == 2) { null_count=0; display_count++; char* recemsg = data_recv.recvbuf; int is_null = is_begin_with(recemsg, "00000"); if (is_null == 1) { printf("recv 00000"); continue; } //printf("cam_no %d", data_recv.cam_no); //int cam_no = data_recv.cam_no; buf_size = data_recv.size; memcpy(buf, data_recv.recvbuf, buf_size); if (g_errorcall && buf_size>1000 && display_count%20==0) { g_errorcall(2, deviceInfo.cam_no, 1, 0, "rece data", 1);//err_type, cam_no, column, msg_level, msg_txt, spare display_count = 0; } //保存流数据并分小时存储 time_t tt = time(NULL);//这句返回的只是一个时间cuo tm* t = localtime(&tt); auto iter = cameraMap.find(deviceInfo.cam_no); if (iter != cameraMap.end()){ iter->second->foutV.write(data_recv.recvbuf, data_recv.size); if (t->tm_min == 0 && (iter->second->timeInHour != t->tm_hour)) { //判断间隔一小时 iter->second->timeInHour = t->tm_hour; iter->second->foutV.close(); time_t tt = time(NULL);//这句返回的只是一个时间cuo tm* t = localtime(&tt); char ctmBegin[20]; strftime(ctmBegin, 20, "/%Y%m%d%H%M", t); char str3[80]; sprintf(str3, "create data:%s%s%s", deviceInfo.cam_name, ctmBegin, ".dat"); SLOG1(str3); printf("%s", deviceInfo.cam_name + std::string(ctmBegin) + ".dat"); iter->second->foutV.open(deviceInfo.cam_name + std::string(ctmBegin) + ".dat", ios::binary); } } return buf_size; } if (ret < 0) { printf("WSAStartup() failed!\n"); continue; //return 0; } } return 0; } int send_cmd(int cam_no,int size,int datatype,char* cam_name, SOCKET& sockClient) { SendStcuct data_send; memset(&data_send, 0, sizeof(struct SendStcuct)); data_send.size = size; data_send.data_type = datatype; data_send.cam_no = cam_no; memcpy(data_send.sendbuf, cam_name, sizeof(char) * (size)); printf("data_send len %d\n", sizeof(data_send)); send(sockClient, (char *)&data_send, sizeof(struct SendStcuct), 0); return 0; } int tcp_recv_conn(char* ip, int port, char* cam_name, int size, int cam_no, FrameFunc tcallback(char* a, int size, int cam_no, int height, int width)) { WSADATA wsaData; if (WSAStartup(MAKEWORD(2, 2), &wsaData) != 0) { printf("初始化Winsock失败"); return -1; } SOCKADDR_IN addrSrv; addrSrv.sin_family = AF_INET; addrSrv.sin_port = htons(port); SOCKET sockClient = socket(AF_INET, SOCK_STREAM, 0); int nRecvBuf = 0;//设置为32K setsockopt(sockClient, SOL_SOCKET, SO_RCVBUF, (const char*)&nRecvBuf, sizeof(int)); setsockopt(sockClient, SOL_SOCKET, SO_SNDBUF, (char *)&nRecvBuf, sizeof(int)); inet_pton(AF_INET, ip, &addrSrv.sin_addr.s_addr); if (connect(sockClient, (struct sockaddr*)&addrSrv, sizeof(addrSrv)) == -1) return -2; //throw "连接失败"; if (SOCKET_ERROR == sockClient) { printf("Socket() error:%d", WSAGetLastError()); return -3; } DeviceInfo deviceInfo; deviceInfo.cam_no = cam_no; deviceInfo.sockClient = sockClient; av_register_all(); unsigned version = avcodec_version(); //printf("FFmpeg version: %d\n", version); //初始化流文件状态 time_t tt = time(NULL);//这句返回的只是一个时间cuo tm* t = localtime(&tt); char ctmBegin[20]; strftime(ctmBegin, 20, "/%Y%m%d%H%M", t); caminfo cinfoInstance; deviceInfo.cam_name = cam_name; //std::string dataName = cam_name; cinfoInstance.foutV.open(cam_name + std::string(ctmBegin) + ".dat", ios::binary); //判断间隔一小时 cinfoInstance.timeInHour = t->tm_hour; cameraMap[cam_no] = &cinfoInstance; char str3[20]; sprintf(str3, "camno: %d start", cam_no); SLOG1(str3); AVFormatContext *pFormatCtx; int i, videoindex; AVCodecContext *pCodecCtx; AVCodec *pCodec; char filepath[] = "video.264"; //av_register_all(); avformat_network_init(); pFormatCtx = avformat_alloc_context(); //patha = "C:\\Users\\sbd01\\Pictures\\ffmpegtest\\Debug\\video.dat"; //fp_open = fopen(patha.c_str(), "rb+"); unsigned char *aviobuffer = (unsigned char *)av_malloc(1512); send_cmd(cam_no, size,2,cam_name, sockClient); AVIOContext *avio = avio_alloc_context(aviobuffer, 1512, 0, &deviceInfo, read_buffer, NULL, NULL); pFormatCtx->pb = avio; //if (avformat_open_input(&pFormatCtx, patha.c_str(), NULL, NULL) != 0) { if (avformat_open_input(&pFormatCtx, NULL, NULL, NULL) != 0) { printf("Couldn't open input stream %d\n", cam_no); return -1; } printf("camno %d find stream\n", cam_no); pFormatCtx->probesize = 1000 * 1024; pFormatCtx->max_analyze_duration = 10 * AV_TIME_BASE; pCodec = NULL; while (pCodec == NULL) { printf("%d start find stream info \n", cam_no); if (avformat_find_stream_info(pFormatCtx, NULL) < 0) { printf("Couldn't find stream info %d\n", cam_no); goto restart_stream; continue; } videoindex = -1; for (i = 0; i < pFormatCtx->nb_streams; i++) if (pFormatCtx->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO) { if (videoindex == -1) { videoindex = i; } //break; } if (videoindex == -1) { printf("%d Didn't find a video stream.\n", cam_no); goto restart_stream; } pCodecCtx = pFormatCtx->streams[videoindex]->codec; //pCodec = avcodec_find_decoder(AV_CODEC_ID_H264); pCodec = avcodec_find_decoder(pCodecCtx->codec_id); if (pCodec == NULL) { printf("%d Codec not found \n", cam_no); goto restart_stream; //return -1; } if (avcodec_open2(pCodecCtx, pCodec, NULL) < 0) { printf("%d Could not open codec.\n", cam_no); goto restart_stream; continue; //return -1; } if (pCodecCtx->width <= 0 || pCodecCtx->height <= 0 || pCodecCtx->height >2000 || pCodecCtx->width >3000) { printf("cam %d pCodecCtx error 1 width %d height %d ", cam_no, pCodecCtx->width, pCodecCtx->height); goto restart_stream; } goto ok; restart_stream: printf("%d restart 1 ", cam_no); avformat_free_context(pFormatCtx); printf("restart 2 "); //avformat_close_input(&pFormatCtx); pFormatCtx = NULL; pFormatCtx = avformat_alloc_context(); printf("restart 3 "); //av_freep(aviobuffer); //printf("restart 4"); aviobuffer = (unsigned char *)av_malloc(1512); printf("restart 4 "); AVIOContext *avio2 = avio_alloc_context(aviobuffer, 1512, 0, &deviceInfo, read_buffer, NULL, NULL); pFormatCtx->pb = avio2; pFormatCtx->probesize = 1000 * 1024; pFormatCtx->max_analyze_duration = 10 * AV_TIME_BASE; if (avformat_open_input(&pFormatCtx, NULL, NULL, NULL) != 0) { printf("2Couldn't open input stream %d\n", cam_no); //return -1; } printf("restart 5\n"); pCodec = NULL; continue; ok: break; } printf("camno:%d code name :%s width %d height %d\n",cam_no, pCodec->name, pCodecCtx->width, pCodecCtx->height); AVFrame *pFrame, *pFrameYUV; pFrame = av_frame_alloc(); pFrameYUV = av_frame_alloc(); int ret, got_picture; if (g_errorcall) { char* cc; int length = strlen(pCodec->name); cc = new char[length + 1]; strcpy(cc, pCodec->name); g_errorcall(0, cam_no, pCodecCtx->width, pCodecCtx->height, cc, 11); } AVPacket *packet = (AVPacket *)av_malloc(sizeof(AVPacket)); struct SwsContext *img_convert_ctx; img_convert_ctx = sws_getContext(pCodecCtx->width, pCodecCtx->height, pCodecCtx->pix_fmt, pCodecCtx->width, pCodecCtx->height, AV_PIX_FMT_RGB24, SWS_BICUBIC, NULL, NULL, NULL); uint8_t *out_buffer; printf("cam %d ready decode 2", cam_no); out_buffer = new uint8_t[avpicture_get_size(AV_PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height)]; avpicture_fill((AVPicture *)pFrameYUV, out_buffer, AV_PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height); //av_image_fill_arrays(pFrameYUV->data, pFrameYUV->linesize, out_buffer, AV_PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height, 1); printf("cam %d ready decode 3", cam_no); int dec_error_count = 0; int tmp_test = 0; while (av_read_frame(pFormatCtx, packet) >= 0) { if (packet->stream_index == videoindex) { //tmp_test++; if (packet->size < 50) { av_free_packet(packet); //printf("cam:%d packet is too small %d\n", cam_no, packet->size); Sleep(3); continue; } if (g_errorcall != 0) g_errorcall(2, deviceInfo.cam_no, 1, 2, "start decode",3); char str_decode[40]; sprintf(str_decode, "cam %d start decode", cam_no); SLOG1(str_decode); ret = avcodec_decode_video2(pCodecCtx, pFrame, &got_picture, packet); if (ret < 0) { dec_error_count++; char str3[80]; sprintf(str3, "%d%s decode_error:%d error_count %d", cam_no, " Decode Error", ret, dec_error_count); SLOG1(str3); if (g_errorcall != 0) g_errorcall(1, deviceInfo.cam_no, 0, 2, str3, 80); printf("cam:%d Decode Error got_picture %d decode_error_num %d\n", cam_no, got_picture, dec_error_count); if (dec_error_count > 2) { dec_error_count = 0; // restart ffmpeg av_free_packet(packet); Sleep(50); sws_freeContext(img_convert_ctx); img_convert_ctx = NULL; printf("cam %d sws_freeContext 1\n", cam_no); //av_free(out_buffer); //av_free(pFrameYUV); avcodec_close(pCodecCtx); //pCodecCtx = NULL; if (avcodec_open2(pCodecCtx, pCodec, NULL)<0) { printf("Could not open codec.\n"); return -1; } /*pFrame = av_frame_alloc(); pFrameYUV = av_frame_alloc();*/ //packet = (AVPacket *)av_malloc(sizeof(AVPacket)); printf("cam_no %d avcodec_open2 ok width:%d height:%d\n", cam_no, pCodecCtx->width, pCodecCtx->height); img_convert_ctx = sws_getContext(pCodecCtx->width, pCodecCtx->height, pCodecCtx->pix_fmt, pCodecCtx->width, pCodecCtx->height, AV_PIX_FMT_RGB24, SWS_BICUBIC, NULL, NULL, NULL); char str3[40]; sprintf(str3, "ffmpeg restart cam %d ", cam_no); SLOG1(str3); continue; } } if (got_picture) { if (g_errorcall != 0) g_errorcall(2, deviceInfo.cam_no, 1, 2, "got_picture",4); char str3[40]; sprintf(str3, "cam %d got_picture", cam_no); SLOG1(str3); sws_scale(img_convert_ctx, (const uint8_t* const*)pFrame->data, pFrame->linesize, 0, pCodecCtx->height, pFrameYUV->data, pFrameYUV->linesize); /*fwrite(pFrameYUV->data[0], (pCodecCtx->width)*(pCodecCtx->height) * 3, 1, output);*/ tcallback((char*)pFrameYUV->data[0], pCodecCtx->height * pCodecCtx->width * 3, cam_no, pCodecCtx->height, pCodecCtx->width); } } av_free_packet(packet); Sleep(10); } sws_freeContext(img_convert_ctx); //av_free(out_buffer); av_free(pFrameYUV); avcodec_close(pCodecCtx); avformat_close_input(&pFormatCtx); return 0; } int tcp_init(char* ip, int port) { int res = tcpInit(ip, port); //printf("conn server\t%d\n", res); return res; } int ffmpeg_recv(int cam_no, FrameFunc tcallback(char* a, int size, int cam_no, int height, int width)) { av_register_all(); unsigned version = avcodec_version(); printf("FFmpeg version: %d\n", version); AVFormatContext *pFormatCtx; int i, videoindex; AVCodecContext *pCodecCtx; AVCodec *pCodec; char filepath[] = "video.264"; avformat_network_init(); pFormatCtx = avformat_alloc_context(); //string patha = "C:\\Users\\sbd01\\Videos\\video.264"; //fp_open = fopen(patha.c_str(), "rb+"); unsigned char *aviobuffer = (unsigned char *)av_malloc(1512); AVIOContext *avio = avio_alloc_context(aviobuffer, 1512, 0, &cam_no, read_buffer, NULL, NULL); pFormatCtx->pb = avio; if (avformat_open_input(&pFormatCtx, NULL, NULL, NULL) != 0) { printf("Couldn't open input stream.\n"); return -1; } if (avformat_find_stream_info(pFormatCtx, NULL)<0) { printf("Couldn't find stream information.\n"); return -1; } videoindex = -1; for (i = 0; i<pFormatCtx->nb_streams; i++) if (pFormatCtx->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO) { videoindex = i; break; } if (videoindex == -1) { printf("Didn't find a video stream.\n"); return -1; } pCodecCtx = pFormatCtx->streams[videoindex]->codec; pCodec = avcodec_find_decoder(pCodecCtx->codec_id); if (pCodec == NULL) { printf("Codec not found.\n"); return -1; } if (avcodec_open2(pCodecCtx, pCodec, NULL)<0) { printf("Could not open codec.\n"); return -1; } AVFrame *pFrame, *pFrameYUV; pFrame = av_frame_alloc(); pFrameYUV = av_frame_alloc(); int ret, got_picture; AVPacket *packet = (AVPacket *)av_malloc(sizeof(AVPacket)); struct SwsContext *img_convert_ctx; img_convert_ctx = sws_getContext(pCodecCtx->width, pCodecCtx->height, pCodecCtx->pix_fmt, pCodecCtx->width, pCodecCtx->height, AV_PIX_FMT_RGB24, SWS_BICUBIC, NULL, NULL, NULL); uint8_t *out_buffer; out_buffer = new uint8_t[avpicture_get_size(AV_PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height)]; avpicture_fill((AVPicture *)pFrameYUV, out_buffer, AV_PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height); while (av_read_frame(pFormatCtx, packet) >= 0) { if (packet->stream_index == videoindex) { ret = avcodec_decode_video2(pCodecCtx, pFrame, &got_picture, packet); if (ret < 0) { printf("Decode Error.\n"); return -1; } if (got_picture) { sws_scale(img_convert_ctx, (const uint8_t* const*)pFrame->data, pFrame->linesize, 0, pCodecCtx->height, pFrameYUV->data, pFrameYUV->linesize); /*fwrite(pFrameYUV->data[0], (pCodecCtx->width)*(pCodecCtx->height) * 3, 1, output);*/ tcallback((char*)pFrameYUV->data[0], pCodecCtx->height * pCodecCtx->width * 3, cam_no, pCodecCtx->height, pCodecCtx->width); } } av_free_packet(packet); } sws_freeContext(img_convert_ctx); //fclose(fp_open); //SDL_Quit(); //av_free(out_buffer); av_free(pFrameYUV); avcodec_close(pCodecCtx); avformat_close_input(&pFormatCtx); return 0; } //Callback int file_buffer(void *opaque, uint8_t *buf, int buf_size) { FILE *fp_open = (FILE *)opaque; if (!feof(fp_open)) { int true_size = fread(buf, 1, buf_size, fp_open); return true_size; } else { return -1; } } int play_file(char* file_name, FrameFunc tcallback(char* a, int size,int num, int height, int width)) { av_register_all(); unsigned version = avcodec_version(); printf("FFmpeg version: %d\n", version); AVFormatContext *pFormatCtx; int i, videoindex; AVCodecContext *pCodecCtx; AVCodec *pCodec; char filepath[] = "video.264"; //av_register_all(); avformat_network_init(); pFormatCtx = avformat_alloc_context(); string patha = "C:\\Users\\sbd01\\Videos\\video.264"; //patha = "C:\\Users\\sbd01\\Pictures\\ffmpegtest\\Debug\\video.dat"; FILE *fp_open = fopen(file_name, "rb+"); unsigned char *aviobuffer = (unsigned char *)av_malloc(32768); //printf("avio_alloc_context %d\n", cam_no); AVIOContext *avio = avio_alloc_context(aviobuffer, 32768, 0, (void*)fp_open, file_buffer, NULL, NULL); pFormatCtx->pb = avio; //if (avformat_open_input(&pFormatCtx, patha.c_str(), NULL, NULL) != 0) { if (avformat_open_input(&pFormatCtx, NULL, NULL, NULL) != 0) { printf("Couldn't open input stream.\n"); return -1; } if (avformat_find_stream_info(pFormatCtx, NULL)<0) { printf("Couldn't find stream information.\n"); return -1; } videoindex = -1; for (i = 0; i<pFormatCtx->nb_streams; i++) if (pFormatCtx->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO) { videoindex = i; break; } if (videoindex == -1) { printf("Didn't find a video stream.\n"); return -1; } pCodecCtx = pFormatCtx->streams[videoindex]->codec; pCodec = avcodec_find_decoder(pCodecCtx->codec_id); if (pCodec == NULL) { printf("Codec not found.\n"); return -1; } if (avcodec_open2(pCodecCtx, pCodec, NULL)<0) { printf("Could not open codec.\n"); return -1; } AVFrame *pFrame, *pFrameYUV; pFrame = av_frame_alloc(); pFrameYUV = av_frame_alloc(); /*if (SDL_Init(SDL_INIT_VIDEO | SDL_INIT_AUDIO | SDL_INIT_TIMER)) { printf("Could not initialize SDL - %s\n", SDL_GetError()); return -1; }*/ /*int screen_w = 0, screen_h = 0; SDL_Surface *screen; screen_w = pCodecCtx->width; screen_h = pCodecCtx->height; screen = SDL_SetVideoMode(screen_w, screen_h, 0, 0); if (!screen) { printf("SDL: could not set video mode - exiting:%s\n", SDL_GetError()); return -1; } SDL_Overlay *bmp; bmp = SDL_CreateYUVOverlay(pCodecCtx->width, pCodecCtx->height, SDL_YV12_OVERLAY, screen); SDL_Rect rect; rect.x = 0; rect.y = 0; rect.w = screen_w; rect.h = screen_h;*/ //SDL End------------------------ int ret, got_picture; AVPacket *packet = (AVPacket *)av_malloc(sizeof(AVPacket)); struct SwsContext *img_convert_ctx; img_convert_ctx = sws_getContext(pCodecCtx->width, pCodecCtx->height, pCodecCtx->pix_fmt, pCodecCtx->width, pCodecCtx->height, AV_PIX_FMT_RGB24, SWS_BICUBIC, NULL, NULL, NULL); uint8_t *out_buffer; out_buffer = new uint8_t[avpicture_get_size(AV_PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height)]; avpicture_fill((AVPicture *)pFrameYUV, out_buffer, AV_PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height); while (av_read_frame(pFormatCtx, packet) >= 0) { if (packet->stream_index == videoindex) { ret = avcodec_decode_video2(pCodecCtx, pFrame, &got_picture, packet); if (ret < 0) { printf("Decode Error.\n"); return -1; } if (got_picture) { sws_scale(img_convert_ctx, (const uint8_t* const*)pFrame->data, pFrame->linesize, 0, pCodecCtx->height, pFrameYUV->data, pFrameYUV->linesize); /*fwrite(pFrameYUV->data[0], (pCodecCtx->width)*(pCodecCtx->height) * 3, 1, output);*/ tcallback((char*)pFrameYUV->data[0], pCodecCtx->height * pCodecCtx->width * 3, 1, pCodecCtx->height, pCodecCtx->width); } } av_free_packet(packet); } sws_freeContext(img_convert_ctx); //fclose(fp_open); //SDL_Quit(); //av_free(out_buffer); av_free(pFrameYUV); avcodec_close(pCodecCtx); avformat_close_input(&pFormatCtx); return 0; }
以上这篇python opencv图片编码为h264文件的实例就是小编分享给大家的全部内容了,希望能给大家一个参考,也希望大家多多支持。
华山资源网 Design By www.eoogi.com
广告合作:本站广告合作请联系QQ:858582 申请时备注:广告合作(否则不回)
免责声明:本站资源来自互联网收集,仅供用于学习和交流,请遵循相关法律法规,本站一切资源不代表本站立场,如有侵权、后门、不妥请联系本站删除!
免责声明:本站资源来自互联网收集,仅供用于学习和交流,请遵循相关法律法规,本站一切资源不代表本站立场,如有侵权、后门、不妥请联系本站删除!
华山资源网 Design By www.eoogi.com
暂无评论...
稳了!魔兽国服回归的3条重磅消息!官宣时间再确认!
昨天有一位朋友在大神群里分享,自己亚服账号被封号之后居然弹出了国服的封号信息对话框。
这里面让他访问的是一个国服的战网网址,com.cn和后面的zh都非常明白地表明这就是国服战网。
而他在复制这个网址并且进行登录之后,确实是网易的网址,也就是我们熟悉的停服之后国服发布的暴雪游戏产品运营到期开放退款的说明。这是一件比较奇怪的事情,因为以前都没有出现这样的情况,现在突然提示跳转到国服战网的网址,是不是说明了简体中文客户端已经开始进行更新了呢?
更新日志
2024年11月19日
2024年11月19日
- 好薇2024《兵哥哥》1:124K黄金母盘[WAV+CUE]
- 胡歌.2006-珍惜(EP)【步升大风】【FLAC分轨】
- 洪荣宏.2014-拼乎自己看【华特】【WAV+CUE】
- 伊能静.1999-从脆弱到勇敢1987-1996精选2CD【华纳】【WAV+CUE】
- 刘亮鹭《汽车DJ玩主》[WAV+CUE][1.1G]
- 张杰《最接近天堂的地方》天娱传媒[WAV+CUE][1.1G]
- 群星《2022年度发烧天碟》无损黑胶碟 2CD[WAV+CUE][1.4G]
- 罗文1983-罗文甄妮-射雕英雄传(纯银AMCD)[WAV+CUE]
- 群星《亚洲故事香港纯弦》雨果UPMAGCD2024[低速原抓WAV+CUE]
- 群星《经典咏流传》限量1:1母盘直刻[低速原抓WAV+CUE]
- 庾澄庆1993《老实情歌》福茂唱片[WAV+CUE][1G]
- 许巍《在别处》美卡首版[WAV+CUE][1G]
- 林子祥《单手拍掌》华纳香港版[WAV+CUE][1G]
- 郑秀文.1997-我们的主题曲【华纳】【WAV+CUE】
- 群星.2001-生命因爱动听电影原创音乐AVCD【MEDIA】【WAV+CUE】