mirror of
https://github.com/ZLMediaKit/ZLMediaKit.git
synced 2024-11-26 12:37:09 +08:00
update: 使用双循环队列作为缓冲
This commit is contained in:
parent
72b239c002
commit
cbbd575ec6
@ -10,412 +10,10 @@
|
||||
|
||||
#include "test_video_stack.h"
|
||||
|
||||
#include "Network/TcpServer.h"
|
||||
#include "Network/TcpServer.h"
|
||||
#include <memory>
|
||||
#include <stdexcept>
|
||||
#include "Common/Device.h"
|
||||
#include "Util/MD5.h"
|
||||
#include "Util/logger.h"
|
||||
#include "Util/SSLBox.h"
|
||||
#include "Util/onceToken.h"
|
||||
#include "Network/TcpServer.h"
|
||||
#include "Poller/EventPoller.h"
|
||||
|
||||
#include "Common/config.h"
|
||||
#include "Rtsp/UDPServer.h"
|
||||
#include "Rtsp/RtspSession.h"
|
||||
#include "Rtmp/RtmpSession.h"
|
||||
#include "Shell/ShellSession.h"
|
||||
#include "Rtmp/FlvMuxer.h"
|
||||
#include "Player/PlayerProxy.h"
|
||||
#include "Http/WebSocketSession.h"
|
||||
|
||||
#include "Pusher/MediaPusher.h"
|
||||
using namespace std;
|
||||
using namespace toolkit;
|
||||
using namespace mediakit;
|
||||
|
||||
|
||||
|
||||
void VideoStack::parseParam(const std::string ¶m) {
|
||||
//auto json = nlohmann::json::parse(testJson);
|
||||
Json::Value json;
|
||||
Json::Reader reader;
|
||||
reader.parse(testJson, json);
|
||||
|
||||
_width = json["width"].asInt(); //输出宽度
|
||||
_height = json["height"].asInt(); //输出高度
|
||||
_stack_id = json["id"].asString();
|
||||
|
||||
int rows = json["rows"].asInt(); // 堆叠行数
|
||||
int cols = json["cols"].asInt(); // 堆叠列数
|
||||
float gapv = json["gapv"].asFloat(); // 垂直间距
|
||||
float gaph = json["gaph"].asFloat(); // 水平间距
|
||||
|
||||
// int gapvPix = (int)std::round(gapv * _width) % 2 ? std::round(gapv * _width)+ 1 : std::round(gapv * _width);
|
||||
// int gaphPix = (int)std::round(gaph * _height) % 2 ? std::round(gaph * _height) + 1 : std::round(gaph * _height);
|
||||
// int gridWidth = _width - ((cols-1) * gapvPix); //1920*(1-0.002*3) / 4 = 477
|
||||
// int gridHeight = _height - ((rows - 1) * gaphPix); //1080*(1-0.001*3) / 4 = 269
|
||||
|
||||
//间隔先默认都是0
|
||||
auto gridWidth = _width / cols;
|
||||
auto gridHeight = _height / rows;
|
||||
int gapvPix = 0;
|
||||
int gaphPix = 0;
|
||||
|
||||
_params = std::vector<VideoStack::Param>(rows * cols);
|
||||
|
||||
for (int row = 0; row < rows; row++) {
|
||||
for (int col = 0; col < cols; col++) {
|
||||
std::string url = json["urls"][row][col].asString();
|
||||
|
||||
VideoStack::Param param;
|
||||
param.posX = gridWidth * col + col * gaphPix;
|
||||
param.posY = gridHeight * row + row * gapvPix;
|
||||
|
||||
param.width = gridWidth;
|
||||
param.height = gridHeight;
|
||||
|
||||
param.url = url;
|
||||
_params[row * cols + col] = param;
|
||||
}
|
||||
}
|
||||
|
||||
// 判断是否需要合并格子 (焦点屏)
|
||||
if (!json["span"].empty()) {
|
||||
for (const auto& subArray : json["span"]) {
|
||||
std::array<int, 4> mergePos;
|
||||
int index = 0;
|
||||
|
||||
// 获取要合并的起始格子和终止格子下标
|
||||
for (const auto& innerArray : subArray) {
|
||||
for (const auto& number : innerArray) {
|
||||
if (index < mergePos.size()) {
|
||||
mergePos[index++] = number.asInt();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
for (int i = mergePos[0]; i <= mergePos[2]; i++) {
|
||||
for (int j = mergePos[1]; j <= mergePos[3]; j++) {
|
||||
if (i == mergePos[0] && j == mergePos[1]) {
|
||||
// 重新计算合并后格子的宽高
|
||||
_params[i * cols + j].width = (mergePos[3] - mergePos[1] + 1) * gridWidth + (mergePos[3] - mergePos[1]) * gapvPix;
|
||||
_params[i * cols + j].height = (mergePos[2] - mergePos[0] + 1) * gridHeight + (mergePos[2] - mergePos[0]) * gaphPix;
|
||||
}
|
||||
else {
|
||||
_params[i * cols + j] = {}; // 置空被合并的格子
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
void VideoStack::copyToBuf(const std::shared_ptr<AVFrame> &buf, const FFmpegFrame::Ptr &frame, const Param &p) {
|
||||
|
||||
auto sws = std::make_shared<FFmpegSws>(AV_PIX_FMT_YUV420P, p.width, p.height);
|
||||
|
||||
auto tmp = sws->inputFrame(frame);
|
||||
|
||||
/*libyuv::I420Scale(frame->get()->data[0], frame->get()->linesize[0],
|
||||
frame->get()->data[1], frame->get()->linesize[1],
|
||||
frame->get()->data[2], frame->get()->linesize[2],
|
||||
frame->get()->width, frame->get()->height,
|
||||
tmp->data[0], tmp->linesize[0],
|
||||
tmp->data[1], tmp->linesize[1],
|
||||
tmp->data[2], tmp->linesize[2],
|
||||
tmp->width, tmp->height,
|
||||
libyuv::kFilterNone);*/
|
||||
|
||||
//TODO: NV12的copy
|
||||
|
||||
//Y平面
|
||||
for (int i = 0; i < p.height; i++) {
|
||||
memcpy(buf->data[0] + buf->linesize[0] * (i + p.posY) + p.posX, tmp->get()->data[0] + tmp->get()->linesize[0] * i, tmp->get()->width);
|
||||
}
|
||||
for (int i = 0; i < p.height / 2; i++) {
|
||||
// U平面
|
||||
memcpy(buf->data[1] + buf->linesize[1] * (i + p.posY / 2) + p.posX / 2, tmp->get()->data[1] + tmp->get()->linesize[1] * i, tmp->get()->width / 2);
|
||||
|
||||
// V平面
|
||||
memcpy(buf->data[2] + buf->linesize[2] * (i + p.posY / 2) + p.posX / 2, tmp->get()->data[2] + tmp->get()->linesize[2] * i, tmp->get()->width / 2);
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
void StackPlayer::play(const std::string &url) {
|
||||
_url = url;
|
||||
// 创建拉流 解码对象
|
||||
auto player = std::make_shared<mediakit::MediaPlayer>();
|
||||
std::weak_ptr<mediakit::MediaPlayer> weakPlayer = player;
|
||||
|
||||
std::weak_ptr<StackPlayer> weakSP = shared_from_this();
|
||||
|
||||
player->setOnPlayResult([weakPlayer, weakSP, url](const toolkit::SockException &ex) mutable {
|
||||
InfoL << "Dec channel OnPlayResult:" << ex.what();
|
||||
auto strongPlayer = weakPlayer.lock();
|
||||
if (!strongPlayer) {
|
||||
return;
|
||||
}
|
||||
|
||||
if (ex) {
|
||||
InfoL << "重试: " << url;
|
||||
std::this_thread::sleep_for(std::chrono::seconds(5));
|
||||
strongPlayer->play(url);
|
||||
}
|
||||
|
||||
auto videoTrack = std::dynamic_pointer_cast<mediakit::VideoTrack>(strongPlayer->getTrack(mediakit::TrackVideo, false));
|
||||
// auto audioTrack = std::dynamic_pointer_cast<mediakit::AudioTrack>(strongPlayer->getTrack(mediakit::TrackAudio, false));
|
||||
|
||||
if (videoTrack) {
|
||||
// auto decoder = std::make_shared<FFmpegDecoder>(videoTrack, 1, std::vector<std::string>{ "hevc_cuvid", "h264_cuvid"});
|
||||
auto decoder = std::make_shared<mediakit::FFmpegDecoder>(videoTrack, 0, std::vector<std::string> { "h264" });
|
||||
// auto decoder = std::make_shared<mediakit::FFmpegDecoder>(videoTrack);
|
||||
auto strongSP = weakSP.lock();
|
||||
if (!strongSP) {
|
||||
return;
|
||||
}
|
||||
|
||||
strongSP->fps = videoTrack->getVideoFps();
|
||||
|
||||
decoder->setOnDecode([weakSP](const mediakit::FFmpegFrame::Ptr &frame) mutable {
|
||||
|
||||
auto strongSP = weakSP.lock();
|
||||
if (!strongSP) {
|
||||
return;
|
||||
}
|
||||
strongSP->onFrame(frame);
|
||||
});
|
||||
|
||||
videoTrack->addDelegate((std::function<bool(const mediakit::Frame::Ptr &)>)[decoder](const mediakit::Frame::Ptr &frame) {
|
||||
return decoder->inputFrame(frame, false, false);
|
||||
});
|
||||
}
|
||||
});
|
||||
|
||||
player->setOnShutdown([](const toolkit::SockException &ex) {
|
||||
InfoL << "Stack play shutdown: " << ex.what();
|
||||
//TODO:断线 将Param中的isDisconnected置为true,然后编码线程那边对此进行判断,填充断线图片
|
||||
});
|
||||
|
||||
(*player)[mediakit::Client::kWaitTrackReady] = false; // 不等待TrackReady
|
||||
(*player)[mediakit::Client::kRtpType] = Rtsp::RTP_TCP;
|
||||
|
||||
player->play(url);
|
||||
|
||||
_player = player;
|
||||
}
|
||||
|
||||
void StackPlayer::addStackPtr(VideoStack* that) {
|
||||
//std::unique_lock<std::shared_timed_mutex> wlock(_mx);
|
||||
if (!that) {
|
||||
return;
|
||||
}
|
||||
auto it = _stacks.find(that->_stack_id);
|
||||
if (it != _stacks.end()) {
|
||||
return;
|
||||
}
|
||||
_stacks[that->_stack_id] = that;
|
||||
}
|
||||
|
||||
void StackPlayer::delStackPtr(VideoStack *that) {
|
||||
//std::unique_lock<std::shared_timed_mutex> wlock(_mx);
|
||||
// TODO:
|
||||
}
|
||||
|
||||
|
||||
//TODO: 根据相对pts来进行同步 (单位是ms,可能得加一个pts转换时间基的步骤)
|
||||
/* void StackPlayer::syncFrameByPts(const FFmpegFrame::Ptr& frame, VideoStack::Param& p, float target_fps) {
|
||||
static std::shared_ptr<FFmpegFrame> lastFrame = nullptr;
|
||||
static int64_t lastPts = 0; // 上一帧的 PTS
|
||||
static double totalDiff = 0.0;
|
||||
|
||||
// 检查 frame 是否有效
|
||||
if (!frame) return;
|
||||
|
||||
// 首帧时给lastFrame赋值
|
||||
if (!lastFrame) {
|
||||
lastFrame = frame;
|
||||
lastPts = frame->get()->pts;
|
||||
p.write.push_back(frame);
|
||||
p.tail++;
|
||||
return;
|
||||
}
|
||||
|
||||
// 计算两帧之间的时间差(假设 PTS 是以秒为单位)
|
||||
double diff = static_cast<double>(frame->get()->pts - lastPts);
|
||||
double duration = 1000 / target_fps;
|
||||
|
||||
totalDiff += diff - duration;
|
||||
|
||||
if (totalDiff >= duration) {
|
||||
totalDiff -= duration;
|
||||
// 当累积误差达到一个完整的帧时,复用上一帧
|
||||
p.write.push_back(lastFrame);
|
||||
p.tail++;
|
||||
}
|
||||
else if (totalDiff <= -duration) {
|
||||
totalDiff += duration;
|
||||
// 累积误差小于负的目标帧持续时间时,跳过当前帧(丢弃)
|
||||
// 这里不更新 lastFrame 和 lastPts
|
||||
}
|
||||
else {
|
||||
// 保留当前帧
|
||||
p.write.push_back(frame);
|
||||
p.tail++;
|
||||
lastFrame = frame;
|
||||
lastPts = frame->get()->pts;
|
||||
}
|
||||
} */
|
||||
|
||||
//直接用fps来计算 进行补帧(复用上一帧)或丢帧
|
||||
void StackPlayer::syncFrameByFps(const FFmpegFrame::Ptr& frame, VideoStack::Param& p, float target_fps) {
|
||||
|
||||
// 检查 frame 是否有效
|
||||
if (!frame) return;
|
||||
|
||||
// 首帧时给lastFrame赋值
|
||||
if (!lastFrame) {
|
||||
lastFrame = frame;
|
||||
}
|
||||
|
||||
diff += fps - target_fps;
|
||||
|
||||
if (diff >= fps) {
|
||||
diff -= fps;
|
||||
// 当累积误差达到一个完整的帧时,复用上一帧
|
||||
p.cache.push_back(lastFrame);
|
||||
}
|
||||
else if (diff <= -fps) {
|
||||
// 累积误差小于负的fps时丢弃当前帧
|
||||
diff += fps;
|
||||
// 注意这里不更新 lastFrame,因为我们丢弃了当前帧
|
||||
}
|
||||
else {
|
||||
// 保留当前帧
|
||||
lastFrame = frame;
|
||||
p.cache.push_back(frame);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
void StackPlayer::onFrame(const FFmpegFrame::Ptr &frame) {
|
||||
//std::shared_lock<std::shared_timed_mutex> rlock(_mx);
|
||||
for (auto &vsp : _stacks) {
|
||||
auto &that = vsp.second;
|
||||
if (!that) {
|
||||
continue;
|
||||
}
|
||||
|
||||
for (auto &p : that->_params) {
|
||||
|
||||
if (p.url != _url) {
|
||||
continue;
|
||||
}
|
||||
|
||||
//p.cache.push_back(frame);
|
||||
syncFrameByFps(frame,p,that->_fps); //不同帧率的视频,通过复用上一帧或丢帧来实现帧同步
|
||||
|
||||
if (that->isReady.test(p.order)) {
|
||||
continue;
|
||||
}
|
||||
if (p.cache.size() >= MAX_FRAME_SIZE) {
|
||||
for (int i = 0; i < MAX_FRAME_SIZE; i++) {
|
||||
auto &front = p.cache.front();
|
||||
that->copyToBuf(that->_buffers[i], front, p);
|
||||
p.cache.pop_front();
|
||||
that->isReady.set(p.order);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void VideoStack::init() {
|
||||
_dev = std::make_shared<mediakit::DevChannel>(mediakit::MediaTuple{ DEFAULT_VHOST, "stack", _stack_id });
|
||||
|
||||
mediakit::VideoInfo info;
|
||||
info.codecId = mediakit::CodecH264;
|
||||
info.iWidth = _width;
|
||||
info.iHeight = _height;
|
||||
info.iFrameRate = _fps;
|
||||
info.iBitRate = _bitRate;
|
||||
|
||||
_dev->initVideo(std::move(info));
|
||||
// dev->initAudio(); //TODO:音频
|
||||
_dev->addTrackCompleted();
|
||||
|
||||
for (int i = 0; i < MAX_FRAME_SIZE; i++) {
|
||||
|
||||
std::shared_ptr<AVFrame> frame(av_frame_alloc(), [](AVFrame *frame_) { av_frame_free(&frame_); });
|
||||
|
||||
frame->width = _width;
|
||||
frame->height = _height;
|
||||
frame->format = _pixfmt;
|
||||
|
||||
av_frame_get_buffer(frame.get(), 32);
|
||||
_buffers.push_back(frame);
|
||||
}
|
||||
|
||||
// setBackground(0, 0, 0);
|
||||
|
||||
_isExit = false;
|
||||
|
||||
int i = 0;
|
||||
for (auto &p : _params) {
|
||||
if (p.url.empty()) {
|
||||
continue;
|
||||
}
|
||||
/*p.tmp.reset(av_frame_alloc(), [](AVFrame *frame_) { av_frame_free(&frame_); });
|
||||
|
||||
p.tmp->width = p.width;
|
||||
p.tmp->height = p.height;
|
||||
p.tmp->format = _pixfmt;
|
||||
|
||||
av_frame_get_buffer(p.tmp.get(), 32);*/
|
||||
p.order = i++;
|
||||
|
||||
flag.set(p.order);
|
||||
|
||||
auto it = playerMap.find(p.url);
|
||||
if (it == playerMap.end()) {
|
||||
// 创建一个
|
||||
|
||||
auto player = std::make_shared<StackPlayer>();
|
||||
player->play(p.url);
|
||||
player->addStackPtr(this);
|
||||
|
||||
playerMap[p.url] = player;
|
||||
} else {
|
||||
it->second->addStackPtr(this);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
}
|
||||
|
||||
void VideoStack::start() {
|
||||
std::thread(
|
||||
[this]() {
|
||||
int64_t pts = 0, index = 0;
|
||||
while (!_isExit) {
|
||||
if (isReady == flag) {
|
||||
for (auto &buf : _buffers) {
|
||||
_dev->inputYUV((char **)buf->data, buf->linesize, pts);
|
||||
pts += 40;
|
||||
index++;
|
||||
}
|
||||
isReady = 0;
|
||||
} else {
|
||||
std::this_thread::sleep_for(std::chrono::milliseconds(5));
|
||||
}
|
||||
}
|
||||
}).detach();
|
||||
}
|
||||
|
||||
static std::unordered_map<uint16_t, toolkit::TcpServer::Ptr> _srvMap;
|
||||
|
||||
// 播放地址 http://127.0.0.1:7089/stack/89.live.flv
|
||||
@ -428,11 +26,15 @@ int main(int argc, char *argv[]) {
|
||||
httpSrv->start<HttpSession>(7089);
|
||||
_srvMap.emplace(7089, httpSrv);
|
||||
|
||||
VideoStackManager vs;
|
||||
|
||||
for (int i = 0; i < 100;i++) {
|
||||
vs.start(testJson);
|
||||
std::this_thread::sleep_for(std::chrono::seconds(60));
|
||||
vs.stop("89");
|
||||
std::this_thread::sleep_for(std::chrono::seconds(10));
|
||||
}
|
||||
|
||||
VideoStack v;
|
||||
v.parseParam();
|
||||
v.init();
|
||||
v.start();
|
||||
|
||||
getchar();
|
||||
return 0;
|
||||
|
@ -17,46 +17,339 @@
|
||||
#include <bitset>
|
||||
#include <shared_mutex>
|
||||
|
||||
#include "Network/TcpServer.h"
|
||||
#include "Network/TcpServer.h"
|
||||
#include <memory>
|
||||
#include <stdexcept>
|
||||
#include "Common/Device.h"
|
||||
#include "Util/MD5.h"
|
||||
#include "Util/logger.h"
|
||||
#include "Util/SSLBox.h"
|
||||
#include "Util/onceToken.h"
|
||||
#include "Network/TcpServer.h"
|
||||
#include "Poller/EventPoller.h"
|
||||
|
||||
static std::string testJson
|
||||
= R"({"msg":"set_combine_source","gapv":0.002,"gaph":0.001,"width":1920,"urls":[["rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test"],["rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test"],["rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test"],["rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test"]],"id":"89","rows":4,"cols":4,"height":1080,"span":[[[0,0],[1,1]],[[2,3],[3,3]]]})";
|
||||
#include "Common/config.h"
|
||||
#include "Rtsp/UDPServer.h"
|
||||
#include "Rtsp/RtspSession.h"
|
||||
#include "Rtmp/RtmpSession.h"
|
||||
#include "Shell/ShellSession.h"
|
||||
#include "Rtmp/FlvMuxer.h"
|
||||
#include "Player/PlayerProxy.h"
|
||||
#include "Http/WebSocketSession.h"
|
||||
|
||||
#include "Pusher/MediaPusher.h"
|
||||
|
||||
|
||||
static constexpr int MAX_FRAME_SIZE = 24;
|
||||
|
||||
class VideoStack : public std::enable_shared_from_this<VideoStack> {
|
||||
static std::string testJson = R"({"msg":"set_combine_source","gapv":0.002,"gaph":0.001,"width":1920,"urls":[["rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test"],["rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test"],["rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test"],["rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test","rtsp://kkem.me:1554/live/test"]],"id":"89","rows":4,"cols":4,"height":1080,"span":[[[0,0],[1,1]],[[2,3],[3,3]]]})";
|
||||
|
||||
|
||||
class VideoStack :public std::enable_shared_from_this<VideoStack>
|
||||
{
|
||||
public:
|
||||
struct Param {
|
||||
int posX;
|
||||
int posY;
|
||||
int width;
|
||||
int height;
|
||||
std::string url;
|
||||
static constexpr int MAX_CACHE_SIZE = 100;
|
||||
|
||||
using Ptr = std::shared_ptr<VideoStack>;
|
||||
|
||||
//全对象共享,使用时拷贝一份,叠加上osd信息。
|
||||
static mediakit::FFmpegFrame::Ptr noVideoPic; //没有url时,显示的图片
|
||||
static mediakit::FFmpegFrame::Ptr disConnPic; //有url,但该流断线时显示的图片
|
||||
|
||||
struct Param
|
||||
{
|
||||
int posX = 0;
|
||||
int posY = 0;
|
||||
int width = 0;
|
||||
int height = 0;
|
||||
std::string url{};
|
||||
|
||||
//运行时参数
|
||||
std::shared_ptr<AVFrame> tmp; //用于存储缩放的临时空间
|
||||
|
||||
uint64_t head = 0;
|
||||
uint64_t tail = 0;
|
||||
std::vector<mediakit::FFmpegFrame::Ptr> write = std::vector<mediakit::FFmpegFrame::Ptr>(MAX_CACHE_SIZE);
|
||||
std::vector<mediakit::FFmpegFrame::Ptr> read = std::vector<mediakit::FFmpegFrame::Ptr>(MAX_CACHE_SIZE);
|
||||
|
||||
int count = 0;
|
||||
|
||||
bool isConnected = false;
|
||||
|
||||
//运行时需要用到的参数
|
||||
//bool isDisconnected = false; //TODO: 用于标识是否断线,对于断线的做特殊处理
|
||||
//std::shared_ptr<AVFrame> tmp; // 临时存储缩放后的frame
|
||||
int order; //标识有效流的序号
|
||||
std::list<mediakit::FFmpegFrame::Ptr> cache;
|
||||
};
|
||||
|
||||
VideoStack() = default;
|
||||
~VideoStack() { _isExit = true; }
|
||||
|
||||
|
||||
// 解析参数 存储到_param中
|
||||
void parseParam(const std::string ¶m = testJson);
|
||||
/*-------给StackPlayer用到的回调-------*/
|
||||
void onPlaySucess(Param& p)
|
||||
{
|
||||
p.tail = getMinTail(false);
|
||||
p.head = p.tail;
|
||||
p.write.clear();
|
||||
p.isConnected = true;
|
||||
}
|
||||
|
||||
// 创建推流对象
|
||||
void init();
|
||||
void onShutdown(Param& p)
|
||||
{
|
||||
p.isConnected = false;
|
||||
p.head = p.tail;
|
||||
p.write.clear();
|
||||
}
|
||||
|
||||
void start();
|
||||
void onFrame(Param& p, const mediakit::FFmpegFrame::Ptr& frame)
|
||||
{
|
||||
//TODO:syncFrameByFps syncFrameByPts
|
||||
|
||||
if (p.tail - p.head > MAX_CACHE_SIZE) {
|
||||
p.tail -= MAX_CACHE_SIZE;
|
||||
}
|
||||
p.write[p.tail % MAX_CACHE_SIZE] = frame;
|
||||
p.tail++;
|
||||
|
||||
|
||||
if (p.head < tail) {
|
||||
p.read.clear();
|
||||
|
||||
int start = p.head % MAX_CACHE_SIZE;
|
||||
int end = tail % MAX_CACHE_SIZE;
|
||||
|
||||
if (end <= start) {
|
||||
// 复制 start 到 MAX_CACHE_SIZE 之间的元素
|
||||
std::copy(p.write.begin() + start, p.write.begin() + MAX_CACHE_SIZE, std::back_inserter(p.read));
|
||||
// 复制 0 到 end 之间的元素
|
||||
std::copy(p.write.begin(), p.write.begin() + end, std::back_inserter(p.read));
|
||||
}
|
||||
else {
|
||||
// 复制 start 到 end 之间的元素
|
||||
std::copy(p.write.begin() + start, p.write.begin() + end, std::back_inserter(p.read));
|
||||
}
|
||||
|
||||
p.head += (tail - p.head);
|
||||
|
||||
std::unique_lock<std::mutex> lock(_mx);
|
||||
readyCount++;
|
||||
cv.notify_one();
|
||||
}
|
||||
}
|
||||
|
||||
//实现拼接
|
||||
void copyToBuf(const std::shared_ptr<AVFrame> &buf, const mediakit::FFmpegFrame::Ptr &frame, const Param &p);
|
||||
|
||||
public:
|
||||
std::string _stack_id;
|
||||
|
||||
VideoStack(const std::string& id, std::vector<Param>& param, int width = 1920, int height = 1080,
|
||||
AVPixelFormat pixfmt = AV_PIX_FMT_YUV420P, float fps = 25.0, int bitRate = 2 * 1024 * 1024, uint8_t r = 20, uint8_t g = 20, uint8_t b = 20)
|
||||
:_id(id), _params(std::move(param)), _width(width), _height(height), _pixfmt(pixfmt), _fps(fps), _bitRate(bitRate)
|
||||
{
|
||||
_buffer.reset(av_frame_alloc(), [](AVFrame* frame_) {
|
||||
av_frame_free(&frame_);
|
||||
});
|
||||
|
||||
_buffer->width = _width;
|
||||
_buffer->height = _height;
|
||||
_buffer->format = _pixfmt;
|
||||
|
||||
av_frame_get_buffer(_buffer.get(), 32);
|
||||
|
||||
int i = 0;
|
||||
for (auto& p : _params) {
|
||||
if (p.width == 0 || p.height == 0) continue;
|
||||
|
||||
p.tmp.reset(av_frame_alloc(), [](AVFrame* frame_) {
|
||||
av_frame_free(&frame_);
|
||||
});
|
||||
p.tmp->width = p.width;
|
||||
p.tmp->height = p.height;
|
||||
p.tmp->format = _pixfmt;
|
||||
|
||||
av_frame_get_buffer(p.tmp.get(), 32);
|
||||
|
||||
}
|
||||
|
||||
|
||||
//TODO:
|
||||
//setBackground(r, g, b);
|
||||
|
||||
|
||||
_dev = std::make_shared<mediakit::DevChannel>(std::move(mediakit::MediaTuple{ DEFAULT_VHOST, "stack", _id }));
|
||||
|
||||
mediakit::VideoInfo info;
|
||||
info.codecId = mediakit::CodecH264;
|
||||
info.iWidth = _width;
|
||||
info.iHeight = _height;
|
||||
info.iFrameRate = _fps;
|
||||
info.iBitRate = _bitRate;
|
||||
|
||||
_dev->initVideo(std::move(info));
|
||||
//dev->initAudio(); //TODO:音频
|
||||
_dev->addTrackCompleted();
|
||||
|
||||
_isExit = false;
|
||||
|
||||
}
|
||||
|
||||
|
||||
~VideoStack()
|
||||
{
|
||||
_isExit = true;
|
||||
}
|
||||
|
||||
|
||||
uint64_t getMinTail(bool isUpdateTotalCount = true)
|
||||
{
|
||||
uint64_t minTail = std::numeric_limits<uint64_t>::max();
|
||||
if (isUpdateTotalCount) {
|
||||
totalCount = 0;
|
||||
}
|
||||
for (const auto& p : _params) {
|
||||
|
||||
if (!p.url.empty() && p.isConnected) {
|
||||
if (isUpdateTotalCount) {
|
||||
totalCount++;
|
||||
}
|
||||
|
||||
if (p.tail < minTail) {
|
||||
minTail = p.tail;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
return minTail == std::numeric_limits<uint64_t>::max() ? 0 : minTail; // 如果没有找到有效的最小值,返回 0
|
||||
|
||||
}
|
||||
|
||||
void copyToBuf(const mediakit::FFmpegFrame::Ptr& frame, const Param& p)
|
||||
{
|
||||
auto sws = std::make_shared<mediakit::FFmpegSws>(AV_PIX_FMT_YUV420P, p.width, p.height);
|
||||
|
||||
auto tmp = sws->inputFrame(frame);
|
||||
|
||||
auto& buf = _buffer;
|
||||
//auto& tmp = p.tmp;
|
||||
auto&& rawFrame = frame->get();
|
||||
|
||||
// libyuv::I420Scale(rawFrame->data[0], rawFrame->linesize[0],
|
||||
// rawFrame->data[1], rawFrame->linesize[1],
|
||||
// rawFrame->data[2], rawFrame->linesize[2],
|
||||
// rawFrame->width, rawFrame->height,
|
||||
// tmp->data[0], tmp->linesize[0],
|
||||
// tmp->data[1], tmp->linesize[1],
|
||||
// tmp->data[2], tmp->linesize[2],
|
||||
// tmp->width, tmp->height,
|
||||
// libyuv::kFilterNone);
|
||||
|
||||
|
||||
for (int i = 0; i < p.height; i++) {
|
||||
memcpy(buf->data[0] + buf->linesize[0] * (i + p.posY) + p.posX,
|
||||
tmp->get()->data[0] + tmp->get()->linesize[0] * i,
|
||||
tmp->get()->width);
|
||||
}
|
||||
for (int i = 0; i < p.height / 2; i++) {
|
||||
// U平面
|
||||
memcpy(buf->data[1] + buf->linesize[1] * (i + p.posY / 2) + p.posX / 2,
|
||||
tmp->get()->data[1] + tmp->get()->linesize[1] * i,
|
||||
tmp->get()->width / 2);
|
||||
|
||||
// V平面
|
||||
memcpy(buf->data[2] + buf->linesize[2] * (i + p.posY / 2) + p.posX / 2,
|
||||
tmp->get()->data[2] + tmp->get()->linesize[2] * i,
|
||||
tmp->get()->width / 2);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
void play()
|
||||
{
|
||||
std::weak_ptr<VideoStack> weakSelf = shared_from_this();
|
||||
|
||||
std::thread([weakSelf]() {
|
||||
int64_t pts = 0;
|
||||
|
||||
while (true) {
|
||||
auto self = weakSelf.lock();
|
||||
if (!self) break;
|
||||
if (self->_isExit) break;
|
||||
|
||||
uint64_t tail = self->getMinTail();
|
||||
if (tail == 0 || tail == std::numeric_limits<int64_t>::max()) {
|
||||
std::this_thread::sleep_for(std::chrono::milliseconds(1));
|
||||
continue;
|
||||
}
|
||||
|
||||
auto count = (tail - self->head) % MAX_CACHE_SIZE;
|
||||
// LOGINFO() << "tail: " << tail << " count: " << count << " head: " << self->head;
|
||||
if (count == 0) {
|
||||
std::this_thread::sleep_for(std::chrono::milliseconds(1));
|
||||
continue;
|
||||
}
|
||||
|
||||
|
||||
self->tail = tail;
|
||||
|
||||
{
|
||||
std::unique_lock<std::mutex> lock(self->_mx);
|
||||
if (!self->cv.wait_for(lock, std::chrono::milliseconds(500), [&]() {
|
||||
return self->readyCount >= self->totalCount;
|
||||
})) {
|
||||
//LOGWARN() << "等待超时!";
|
||||
std::this_thread::sleep_for(std::chrono::milliseconds(1));
|
||||
continue;
|
||||
}
|
||||
}
|
||||
self->readyCount = 0; //重置计数器,用于下次循环
|
||||
|
||||
for (int i = 0; i < count; i++) {
|
||||
for (auto& p : self->_params) {
|
||||
if (p.width == 0 || p.height == 0) {
|
||||
continue;
|
||||
}
|
||||
|
||||
if (p.url.empty()) {
|
||||
//TODO:填充无视频图片
|
||||
//copyToBuf(novideo,p);
|
||||
continue;
|
||||
}
|
||||
|
||||
mediakit::FFmpegFrame::Ptr frame;
|
||||
|
||||
if (p.isConnected) {
|
||||
if (p.read.empty()) {
|
||||
continue;
|
||||
}
|
||||
frame = p.read[i];
|
||||
}
|
||||
else {
|
||||
//frame = self->disconnPic;
|
||||
frame = nullptr; //TODO:填充断线图片
|
||||
}
|
||||
|
||||
if (!frame) continue;
|
||||
|
||||
self->copyToBuf(frame, p);
|
||||
|
||||
}
|
||||
|
||||
|
||||
self->_dev->inputYUV((char**)self->_buffer->data, self->_buffer->linesize, pts);
|
||||
pts += 40;
|
||||
}
|
||||
|
||||
self->head += count;
|
||||
|
||||
}
|
||||
|
||||
// LOGFATAL() << "退出!!!!!!!!!!!!!!!!!!!!!!!";
|
||||
}).detach();
|
||||
}
|
||||
|
||||
|
||||
protected:
|
||||
//void setBackground(uint8_t r, uint8_t g, uint8_t b); //设置背景色 (间隔的颜色) RGB->YUV/NV12
|
||||
|
||||
|
||||
public:
|
||||
std::vector<Param> _params; //存储参数
|
||||
std::string _id;
|
||||
|
||||
private:
|
||||
|
||||
int _width;
|
||||
int _height;
|
||||
@ -64,49 +357,358 @@ public:
|
||||
float _fps = 25.0;
|
||||
int _bitRate = 2 * 1024 * 1024;
|
||||
|
||||
bool _isExit;
|
||||
|
||||
std::vector<VideoStack::Param> _params; // 存储参数
|
||||
|
||||
mediakit::DevChannel::Ptr _dev;
|
||||
|
||||
std::vector<std::shared_ptr<AVFrame>> _buffers;
|
||||
bool _isExit;
|
||||
|
||||
//这两个bit位 用于判断该拼接所需的视频流的缓存帧是否都以就绪
|
||||
std::bitset<1024> isReady;
|
||||
std::bitset<1024> flag;
|
||||
std::shared_ptr<AVFrame> _buffer;
|
||||
|
||||
//FFmpegFrame::Ptr DisconnPic; //TODO: 读取一张准备好的图片,作为断线时的frame
|
||||
//mediakit::FFmpegFrame::Ptr DisconnPic;
|
||||
|
||||
std::atomic<int64_t> head {0};
|
||||
std::atomic<int64_t> tail {0};
|
||||
|
||||
int totalCount = 0;
|
||||
int readyCount = 0;
|
||||
|
||||
std::condition_variable cv;
|
||||
std::mutex _mx;
|
||||
};
|
||||
|
||||
|
||||
class StackPlayer : public std::enable_shared_from_this<StackPlayer> {
|
||||
static int rframe = 0;
|
||||
class StackPlayer : public std::enable_shared_from_this<StackPlayer>
|
||||
{
|
||||
public:
|
||||
using Ptr = std::shared_ptr<StackPlayer>;
|
||||
|
||||
void play(const std::string &url);
|
||||
~StackPlayer() = default;
|
||||
|
||||
void addStackPtr(VideoStack *that);
|
||||
void play(const std::string& url)
|
||||
{
|
||||
_url = url;
|
||||
|
||||
void delStackPtr(VideoStack *that);
|
||||
//创建拉流 解码对象
|
||||
auto player = std::make_shared<mediakit::MediaPlayer>();
|
||||
std::weak_ptr<mediakit::MediaPlayer> weakPlayer = player;
|
||||
|
||||
void onFrame(const mediakit::FFmpegFrame::Ptr &frame);
|
||||
std::weak_ptr<StackPlayer> weakSelf = shared_from_this();
|
||||
|
||||
void syncFrameByFps(const mediakit::FFmpegFrame::Ptr& frame, VideoStack::Param& p, float target_fps);
|
||||
void syncFrameByPts(const mediakit::FFmpegFrame::Ptr& frame, VideoStack::Param& p, float target_fps);
|
||||
player->setOnPlayResult([weakPlayer, weakSelf, url](const toolkit::SockException& ex) mutable {
|
||||
//LOGTRACE() << "StackPlayer OnPlayResult:" << ex.what();
|
||||
auto strongPlayer = weakPlayer.lock();
|
||||
if (!strongPlayer) {
|
||||
return;
|
||||
}
|
||||
|
||||
if (ex) {
|
||||
//LOGERR() << "StackPlayer play failed, retry: " << url;
|
||||
strongPlayer->play(url);
|
||||
}
|
||||
else {
|
||||
auto self = weakSelf.lock();
|
||||
if (!self) {
|
||||
return;
|
||||
}
|
||||
self->dispatch(&VideoStack::onPlaySucess);
|
||||
}
|
||||
|
||||
|
||||
auto videoTrack = std::dynamic_pointer_cast<mediakit::VideoTrack>(strongPlayer->getTrack(mediakit::TrackVideo, false));
|
||||
//auto audioTrack = std::dynamic_pointer_cast<mediakit::AudioTrack>(strongPlayer->getTrack(mediakit::TrackAudio, false));
|
||||
|
||||
if (videoTrack) {
|
||||
//TODO:添加使用显卡还是cpu解码的判断逻辑
|
||||
//auto decoder = std::make_shared<FFmpegDecoder>(videoTrack, 1, std::vector<std::string>{ "hevc_cuvid", "h264_cuvid"});
|
||||
auto decoder = std::make_shared<mediakit::FFmpegDecoder>(videoTrack, 0, std::vector<std::string>{"h264", "hevc" });
|
||||
//auto decoder = std::make_shared<mediakit::FFmpegDecoder>(videoTrack);
|
||||
|
||||
/*auto self = weakSelf.lock();
|
||||
if (!self) {
|
||||
return;
|
||||
}
|
||||
self->fps = videoTrack->getVideoFps();*/
|
||||
|
||||
decoder->setOnDecode([weakSelf](const mediakit::FFmpegFrame::Ptr& frame) mutable {
|
||||
|
||||
//TODO: 回调函数(copy frame数据到待编码的buf中)
|
||||
//copy到需要的编码线程的该通道的队列中
|
||||
auto self = weakSelf.lock();
|
||||
if (!self) {
|
||||
return;
|
||||
}
|
||||
|
||||
//LOGINFO() << "收到frame: " << rframe++;
|
||||
//LOGINFO() << "Frame pts: " << frame->get()->pts;
|
||||
self->dispatch(&VideoStack::onFrame, frame);
|
||||
|
||||
});
|
||||
|
||||
|
||||
videoTrack->addDelegate((std::function<bool(const mediakit::Frame::Ptr&)>)[decoder](const mediakit::Frame::Ptr& frame) {
|
||||
return decoder->inputFrame(frame, false, true);
|
||||
});
|
||||
|
||||
}
|
||||
|
||||
});
|
||||
|
||||
player->setOnShutdown([weakPlayer, url, weakSelf](const toolkit::SockException& ex) {
|
||||
//LOGTRACE() << "StackPlayer Onshutdown: " << ex.what();
|
||||
auto strongPlayer = weakPlayer.lock();
|
||||
if (!strongPlayer) {
|
||||
return;
|
||||
}
|
||||
|
||||
if (ex) {
|
||||
|
||||
auto self = weakSelf.lock();
|
||||
if (!self) {
|
||||
return;
|
||||
}
|
||||
|
||||
self->dispatch(&VideoStack::onShutdown);
|
||||
|
||||
//LOGTRACE() << "StackPlayer try to reconnect: " << url;
|
||||
strongPlayer->play(url);
|
||||
}
|
||||
|
||||
});
|
||||
|
||||
(*player)[mediakit::Client::kWaitTrackReady] = false; //不等待TrackReady
|
||||
(*player)[mediakit::Client::kRtpType] = mediakit::Rtsp::RTP_TCP;
|
||||
|
||||
player->play(url);
|
||||
|
||||
_player = player;
|
||||
}
|
||||
|
||||
void addDispatcher(const std::weak_ptr<VideoStack>& weakPtr)
|
||||
{
|
||||
auto ptr = weakPtr.lock();
|
||||
if (!ptr) {
|
||||
return;
|
||||
}
|
||||
//wlock_(_mx);
|
||||
std::lock_guard<std::mutex> lock(_mx);
|
||||
auto it = _dispatchMap.find(ptr->_id);
|
||||
if (it != _dispatchMap.end()) {
|
||||
return;
|
||||
}
|
||||
_dispatchMap[ptr->_id] = weakPtr;
|
||||
}
|
||||
|
||||
void delDispatcher(const std::string& id)
|
||||
{
|
||||
//wlock_(_mx);
|
||||
std::lock_guard<std::mutex> lock(_mx);
|
||||
auto it = _dispatchMap.find(id);
|
||||
if (it == _dispatchMap.end()) {
|
||||
return;
|
||||
}
|
||||
_dispatchMap.erase(it);
|
||||
|
||||
}
|
||||
|
||||
|
||||
protected:
|
||||
|
||||
template<typename Func, typename... Args>
|
||||
void dispatch(Func func, Args... args) {
|
||||
//rlock_(_mx);
|
||||
std::lock_guard<std::mutex> lock(_mx);
|
||||
for (auto& [_, weakPtr] : _dispatchMap) {
|
||||
auto strongPtr = weakPtr.lock();
|
||||
if (!strongPtr) continue;
|
||||
for (auto& p : strongPtr->_params) {
|
||||
if (p.url != _url) continue;
|
||||
(strongPtr.get()->*func)(p, args...);
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
private:
|
||||
std::string _url;
|
||||
|
||||
float fps;
|
||||
mediakit::FFmpegFrame::Ptr lastFrame;
|
||||
float diff = 0;
|
||||
//std::shared_timed_mutex _mx;
|
||||
std::unordered_map<std::string, VideoStack *> _stacks; // 需要给哪些Stack对象推送帧数据
|
||||
|
||||
mediakit::MediaPlayer::Ptr _player;
|
||||
|
||||
//RWMutex _mx;
|
||||
std::mutex _mx;
|
||||
std::unordered_map<std::string, std::weak_ptr<VideoStack>> _dispatchMap;
|
||||
};
|
||||
|
||||
|
||||
|
||||
static std::mutex mx;
|
||||
static std::unordered_map<std::string, StackPlayer::Ptr> playerMap;
|
||||
|
||||
class VideoStackManager
|
||||
{
|
||||
public:
|
||||
|
||||
//解析参数,解析成功返回true,解析失败返回false, 解析出来的参数通过tmp返回
|
||||
bool parseParam(const std::string& jsonStr, std::string& id, std::vector<VideoStack::Param>& params)
|
||||
{
|
||||
//auto json = nlohmann::json::parse(testJson);
|
||||
Json::Value json;
|
||||
Json::Reader reader;
|
||||
reader.parse(testJson, json);
|
||||
|
||||
int width = json["width"].asInt(); //输出宽度
|
||||
int height = json["height"].asInt(); //输出高度
|
||||
id = json["id"].asString();
|
||||
|
||||
int rows = json["rows"].asInt(); // 堆叠行数
|
||||
int cols = json["cols"].asInt(); // 堆叠列数
|
||||
float gapv = json["gapv"].asFloat(); // 垂直间距
|
||||
float gaph = json["gaph"].asFloat(); // 水平间距
|
||||
|
||||
// int gapvPix = (int)std::round(gapv * _width) % 2 ? std::round(gapv * _width)+ 1 : std::round(gapv * _width);
|
||||
// int gaphPix = (int)std::round(gaph * _height) % 2 ? std::round(gaph * _height) + 1 : std::round(gaph * _height);
|
||||
// int gridWidth = _width - ((cols-1) * gapvPix); //1920*(1-0.002*3) / 4 = 477
|
||||
// int gridHeight = _height - ((rows - 1) * gaphPix); //1080*(1-0.001*3) / 4 = 269
|
||||
|
||||
//间隔先默认都是0
|
||||
auto gridWidth = width / cols;
|
||||
auto gridHeight = height / rows;
|
||||
int gapvPix = 0;
|
||||
int gaphPix = 0;
|
||||
|
||||
params = std::vector<VideoStack::Param>(rows * cols);
|
||||
|
||||
for (int row = 0; row < rows; row++) {
|
||||
for (int col = 0; col < cols; col++) {
|
||||
std::string url = json["urls"][row][col].asString();
|
||||
|
||||
VideoStack::Param param;
|
||||
param.posX = gridWidth * col + col * gaphPix;
|
||||
param.posY = gridHeight * row + row * gapvPix;
|
||||
|
||||
param.width = gridWidth;
|
||||
param.height = gridHeight;
|
||||
|
||||
param.url = url;
|
||||
params[row * cols + col] = param;
|
||||
}
|
||||
}
|
||||
|
||||
// 判断是否需要合并格子 (焦点屏)
|
||||
if (!json["span"].empty()) {
|
||||
for (const auto& subArray : json["span"]) {
|
||||
std::array<int, 4> mergePos;
|
||||
int index = 0;
|
||||
|
||||
// 获取要合并的起始格子和终止格子下标
|
||||
for (const auto& innerArray : subArray) {
|
||||
for (const auto& number : innerArray) {
|
||||
if (index < mergePos.size()) {
|
||||
mergePos[index++] = number.asInt();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
for (int i = mergePos[0]; i <= mergePos[2]; i++) {
|
||||
for (int j = mergePos[1]; j <= mergePos[3]; j++) {
|
||||
if (i == mergePos[0] && j == mergePos[1]) {
|
||||
// 重新计算合并后格子的宽高
|
||||
params[i * cols + j].width = (mergePos[3] - mergePos[1] + 1) * gridWidth + (mergePos[3] - mergePos[1]) * gapvPix;
|
||||
params[i * cols + j].height = (mergePos[2] - mergePos[0] + 1) * gridHeight + (mergePos[2] - mergePos[0]) * gaphPix;
|
||||
}
|
||||
else {
|
||||
params[i * cols + j] = {}; // 置空被合并的格子
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
|
||||
bool start(const std::string& json)
|
||||
{
|
||||
std::vector<VideoStack::Param> params;
|
||||
std::string id;
|
||||
|
||||
bool ret = parseParam(json, id, params);
|
||||
|
||||
if (!ret) return false;
|
||||
|
||||
auto stack = std::make_shared<VideoStack>(id, params);
|
||||
|
||||
for (auto& p : stack->_params) {
|
||||
if (p.url.empty()) continue;
|
||||
if (p.width == 0 || p.height == 0) continue;
|
||||
|
||||
|
||||
StackPlayer::Ptr player;
|
||||
|
||||
{
|
||||
std::lock_guard<std::recursive_mutex> lock(_playerMx);
|
||||
auto it = _playerMap.find(p.url);
|
||||
if (it != _playerMap.end()) {
|
||||
player = it->second;
|
||||
}
|
||||
else {
|
||||
player = std::make_shared<StackPlayer>();
|
||||
player->play(p.url);
|
||||
|
||||
_playerMap[p.url] = player;
|
||||
}
|
||||
}
|
||||
_weakPlayer = player;
|
||||
player->addDispatcher(std::weak_ptr<VideoStack>(stack));
|
||||
}
|
||||
|
||||
|
||||
stack->play();
|
||||
std::lock_guard<std::recursive_mutex> lock(_stackMx);
|
||||
_stackMap[id] = stack;
|
||||
|
||||
_weakPtr = std::weak_ptr<VideoStack>(stack);
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
|
||||
void stop(const std::string& id)
|
||||
{
|
||||
//TODO:先临时全部清空
|
||||
{
|
||||
std::lock_guard<std::recursive_mutex> lock(_playerMx);
|
||||
_playerMap.clear();
|
||||
}
|
||||
|
||||
{
|
||||
std::lock_guard<std::recursive_mutex> lock(_stackMx);
|
||||
auto it = _stackMap.find(id);
|
||||
if (it == _stackMap.end()) {
|
||||
return;
|
||||
}
|
||||
|
||||
_stackMap.erase(it);
|
||||
}
|
||||
|
||||
|
||||
|
||||
|
||||
}
|
||||
|
||||
|
||||
std::weak_ptr<StackPlayer> _weakPlayer;
|
||||
|
||||
std::weak_ptr<VideoStack> _weakPtr;
|
||||
|
||||
private:
|
||||
|
||||
|
||||
|
||||
std::recursive_mutex _playerMx;
|
||||
std::unordered_map<std::string, StackPlayer::Ptr> _playerMap;
|
||||
|
||||
std::recursive_mutex _stackMx;
|
||||
std::unordered_map<std::string, VideoStack::Ptr> _stackMap;
|
||||
};
|
||||
|
||||
|
||||
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user