结合前面的 采集 v4l2 视频, 使用 live555, 通过 rtsp 发布实时流. capture.h, capture.cpp, vcompress.h, vcompress.cpp 需要参考前面几片文章. 这里仅仅贴出 v4l2_x264_service.cpp
[cpp] view plaincopy
#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <assert.h>
#include <liveMedia.hh>
#include <BasicUsageEnvironment.hh>
#include <GroupsockHelper.hh>
#include <sys/types.h>
#include <sys/syscall.h>
#include "capture.h"
#include "vcompress.h"
static UsageEnvironment *_env = 0;
#define SINK_PORT 3030
#define VIDEO_WIDTH 320
#define VIDEO_HEIGHT 240
#define FRAME_PER_SEC 5.0
pid_t gettid()
{
return syscall(SYS_gettid);
}
// 使用 webcam + x264
class WebcamFrameSource : public FramedSource
{
void *mp_capture, *mp_compress; // v4l2 + x264 encoder
int m_started;
void *mp_token;
public:
WebcamFrameSource (UsageEnvironment &env)
: FramedSource(env)
{
fprintf(stderr, "[%d] %s .... calling\n", gettid(), __func__);
mp_capture = capture_open("/dev/video0", VIDEO_WIDTH, VIDEO_HEIGHT, PIX_FMT_YUV420P);
if (!mp_capture) {
fprintf(stderr, "%s: open /dev/video0 err\n", __func__);
exit(-1);
}
mp_compress = vc_open(VIDEO_WIDTH, VIDEO_HEIGHT, FRAME_PER_SEC);
if (!mp_compress) {
fprintf(stderr, "%s: open x264 err\n", __func__);
exit(-1);
}
m_started = 0;
mp_token = 0;
}
~WebcamFrameSource ()
{
fprintf(stderr, "[%d] %s .... calling\n", gettid(), __func__);
if (m_started) {
envir().taskScheduler().unscheduleDelayedTask(mp_token);
}
if (mp_compress)
vc_close(mp_compress);
if (mp_capture)
capture_close(mp_capture);
}
protected:
virtual void doGetNextFrame ()
{
if (m_started) return;
m_started = 1;
// 根据 fps, 计算等待时间
double delay = 1000.0 / FRAME_PER_SEC;
int to_delay = delay * 1000; // us
mp_token = envir().taskScheduler().scheduleDelayedTask(to_delay,
getNextFrame, this);
}
virtual unsigned maxFrameSize() const // 这个很重要, 如果不设置, 可能导致 getNextFrame() 出现 fMaxSize 小于实际编码帧的情况, 导致图像不完整
{ return 100*1024; }
private:
static void getNextFrame (void *ptr)
{
((WebcamFrameSource*)ptr)->getNextFrame1();
}
void getNextFrame1 ()
{
// capture:
Picture pic;
if (capture_get_picture(mp_capture, &pic) < 0) {
fprintf(stderr, "==== %s: capture_get_picture err\n", __func__);
m_started = 0;
return;
}
// compress
const void *outbuf;
int outlen;
if (vc_compress(mp_compress, pic.data, pic.stride, &outbuf, &outlen) < 0) {
fprintf(stderr, "==== %s: vc_compress err\n", __func__);
m_started = 0;
return;
}
int64_t pts, dts;
int key;
vc_get_last_frame_info(mp_compress, &key, &pts, &dts);
// save outbuf
gettimeofday(&fPresentationTime, 0);
fFrameSize = outlen;
if (fFrameSize > fMaxSize) {
fNumTruncatedBytes = fFrameSize - fMaxSize;
fFrameSize = fMaxSize;
}
else {
fNumTruncatedBytes = 0;
}
memmove(fTo, outbuf, fFrameSize);
// notify
afterGetting(this);
m_started = 0;
}
};
class WebcamOndemandMediaSubsession : public OnDemandServerMediaSubsession
{
public:
static WebcamOndemandMediaSubsession *createNew (UsageEnvironment &env, FramedSource *source)
{
return new WebcamOndemandMediaSubsession(env, source);
}
protected:
WebcamOndemandMediaSubsession (UsageEnvironment &env, FramedSource *source)
: OnDemandServerMediaSubsession(env, True) // reuse the first source
{
fprintf(stderr, "[%d] %s .... calling\n", gettid(), __func__);
mp_source = source;
mp_sdp_line = 0;
}
~WebcamOndemandMediaSubsession ()
{
fprintf(stderr, "[%d] %s .... calling\n", gettid(), __func__);
if (mp_sdp_line) free(mp_sdp_line);
}
private:
static void afterPlayingDummy (void *ptr)
{
fprintf(stderr, "[%d] %s .... calling\n", gettid(), __func__);
// ok
WebcamOndemandMediaSubsession *This = (WebcamOndemandMediaSubsession*)ptr;
This->m_done = 0xff;
}
static void chkForAuxSDPLine (void *ptr)
{
WebcamOndemandMediaSubsession *This = (WebcamOndemandMediaSubsession *)ptr;
This->chkForAuxSDPLine1();
}
void chkForAuxSDPLine1 ()
{
fprintf(stderr, "[%d] %s .... calling\n", gettid(), __func__);
if (mp_dummy_rtpsink->auxSDPLine())
m_done = 0xff;
else {
int delay = 100*1000; // 100ms
nextTask() = envir().taskScheduler().scheduleDelayedTask(delay,
chkForAuxSDPLine, this);
}
}
protected:
virtual const char *getAuxSDPLine (RTPSink *sink, FramedSource *source)
{
fprintf(stderr, "[%d] %s .... calling\n", gettid(), __func__);
if (mp_sdp_line) return mp_sdp_line;
mp_dummy_rtpsink = sink;
mp_dummy_rtpsink->startPlaying(*source, 0, 0);
//mp_dummy_rtpsink->startPlaying(*source, afterPlayingDummy, this);
chkForAuxSDPLine(this);
m_done = 0;
envir().taskScheduler().doEventLoop(&m_done);
mp_sdp_line = strdup(mp_dummy_rtpsink->auxSDPLine());
mp_dummy_rtpsink->stopPlaying();
return mp_sdp_line;
}
virtual RTPSink *createNewRTPSink(Groupsock *rtpsock, unsigned char type, FramedSource *source)
{
fprintf(stderr, "[%d] %s .... calling\n", gettid(), __func__);
return H264VideoRTPSink::createNew(envir(), rtpsock, type);
}
virtual FramedSource *createNewStreamSource (unsigned sid, unsigned &bitrate)
{
fprintf(stderr, "[%d] %s .... calling\n", gettid(), __func__);
bitrate = 500;
return H264VideoStreamFramer::createNew(envir(), new WebcamFrameSource(envir()));
}
private:
FramedSource *mp_source; // 对应 WebcamFrameSource
char *mp_sdp_line;
RTPSink *mp_dummy_rtpsink;
char m_done;
};
static void test_task (void *ptr)
{
fprintf(stderr, "test: task ....\n");
_env->taskScheduler().scheduleDelayedTask(100000, test_task, 0);
}
static void test (UsageEnvironment &env)
{
fprintf(stderr, "test: begin...\n");
char done = 0;
int delay = 100 * 1000;
env.taskScheduler().scheduleDelayedTask(delay, test_task, 0);
env.taskScheduler().doEventLoop(&done);
fprintf(stderr, "test: end..\n");
}
int main (int argc, char **argv)
{
// env
TaskScheduler *scheduler = BasicTaskScheduler::createNew();
_env = BasicUsageEnvironment::createNew(*scheduler);
// test
//test(*_env);
// rtsp server
RTSPServer *rtspServer = RTSPServer::createNew(*_env, 8554);
if (!rtspServer) {
fprintf(stderr, "ERR: create RTSPServer err\n");
::exit(-1);
}
// add live stream
do {
WebcamFrameSource *webcam_source = 0;
ServerMediaSession *sms = ServerMediaSession::createNew(*_env, "webcam", 0, "Session from /dev/video0");
sms->addSubsession(WebcamOndemandMediaSubsession::createNew(*_env, webcam_source));
rtspServer->addServerMediaSession(sms);
char *url = rtspServer->rtspURL(sms);
*_env << "using url \"" << url << "\"\n";
delete [] url;
} while (0);
// run loop
_env->taskScheduler().doEventLoop();
return 1;
}
需要 live555 + libavcodec + libswscale + libx264, client 使用 vlc, mplayer, quicktime, .....
(sunkwei) |