You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
2753 lines
92 KiB
2753 lines
92 KiB
/*
|
|
videodevice.cpp - Kopete Video Device Low-level Support
|
|
|
|
Copyright (c) 2005-2006 by Cláudio da Silveira Pinheiro <taupter@gmail.com>
|
|
|
|
Kopete (c) 2002-2003 by the Kopete developers <kopete-devel@kde.org>
|
|
|
|
*************************************************************************
|
|
* *
|
|
* This library is free software; you can redistribute it and/or *
|
|
* modify it under the terms of the GNU Lesser General Public *
|
|
* License as published by the Free Software Foundation; either *
|
|
* version 2 of the License, or (at your option) any later version. *
|
|
* *
|
|
*************************************************************************
|
|
*/
|
|
|
|
#define ENABLE_AV
|
|
|
|
#include <cstdlib>
|
|
#include <cerrno>
|
|
#include <cstring>
|
|
|
|
#include <kdebug.h>
|
|
|
|
#include "videoinput.h"
|
|
#include "videodevice.h"
|
|
|
|
#include "bayer.h"
|
|
#include "sonix_compress.h"
|
|
|
|
#define CLEAR(x) memset (&(x), 0, sizeof (x))
|
|
|
|
namespace Kopete {
|
|
|
|
namespace AV {
|
|
|
|
VideoDevice::VideoDevice()
|
|
{
|
|
// kdDebug(14010) << "libkopete (avdevice): VideoDevice() called" << endl;
|
|
descriptor = -1;
|
|
m_streambuffers = 0;
|
|
m_current_input = 0;
|
|
// kdDebug(14010) << "libkopete (avdevice): VideoDevice() exited successfuly" << endl;
|
|
maxwidth = 32767;
|
|
maxheight = 32767;
|
|
minwidth = 1;
|
|
minheight = 1;
|
|
}
|
|
|
|
|
|
VideoDevice::~VideoDevice()
|
|
{
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
|
|
void VideoDevice::enumerateMenu (void)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << " Menu items:" << endl;
|
|
|
|
memset (&querymenu, 0, sizeof (querymenu));
|
|
querymenu.id = queryctrl.id;
|
|
|
|
for (querymenu.index = queryctrl.minimum; querymenu.index <= queryctrl.maximum; querymenu.index++)
|
|
{
|
|
if (0 == xioctl (VIDIOC_QUERYMENU, &querymenu))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << " " << QString::fromLocal8Bit((const char*)querymenu.name) << endl;
|
|
}
|
|
else
|
|
{
|
|
perror ("VIDIOC_QUERYMENU");
|
|
exit (EXIT_FAILURE);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
|
|
/*!
|
|
\fn VideoDevice::xioctl(int fd, int request, void *arg)
|
|
*/
|
|
int VideoDevice::xioctl(int request, void *arg)
|
|
{
|
|
int r;
|
|
|
|
do r = ioctl (descriptor, request, arg);
|
|
while (-1 == r && EINTR == errno);
|
|
return r;
|
|
}
|
|
|
|
/*!
|
|
\fn VideoDevice::errnoReturn(const char* s)
|
|
*/
|
|
int VideoDevice::errnoReturn(const char* s)
|
|
{
|
|
/// @todo implement me
|
|
fprintf (stderr, "%s error %d, %s\n",s, errno, strerror (errno));
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
/*!
|
|
\fn VideoDevice::setFileName(QString name)
|
|
*/
|
|
int VideoDevice::setFileName(QString filename)
|
|
{
|
|
/// @todo implement me
|
|
full_filename=filename;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
|
|
/*!
|
|
\fn VideoDevice::open()
|
|
*/
|
|
int VideoDevice::open()
|
|
{
|
|
/// @todo implement me
|
|
|
|
kdDebug(14010) << k_funcinfo << "called" << endl;
|
|
if(-1 != descriptor)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device is already open" << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
descriptor = ::open (QFile::encodeName(full_filename), O_RDWR, 0);
|
|
if(isOpen())
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "File " << full_filename << " was opened successfuly" << endl;
|
|
if(EXIT_FAILURE==checkDevice())
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "File " << full_filename << " could not be opened" << endl;
|
|
close();
|
|
return EXIT_FAILURE;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Unable to open file " << full_filename << "Err: "<< errno << endl;
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
initDevice();
|
|
selectInput(m_current_input);
|
|
kdDebug(14010) << k_funcinfo << "exited successfuly" << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
|
|
bool VideoDevice::isOpen()
|
|
{
|
|
if(-1 == descriptor)
|
|
{
|
|
// kdDebug(14010) << k_funcinfo << "VideoDevice::isOpen() File is not open" << endl;
|
|
return false;
|
|
}
|
|
// kdDebug(14010) << k_funcinfo << "VideoDevice::isOpen() File is open" << endl;
|
|
return true;
|
|
}
|
|
|
|
int VideoDevice::checkDevice()
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "checkDevice() called." << endl;
|
|
if(isOpen())
|
|
{
|
|
m_videocapture=false;
|
|
m_videochromakey=false;
|
|
m_videoscale=false;
|
|
m_videooverlay=false;
|
|
m_videoread=false;
|
|
m_videoasyncio=false;
|
|
m_videostream=false;
|
|
|
|
m_driver=VIDEODEV_DRIVER_NONE;
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
|
|
//if(!getWorkaroundBrokenDriver())
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " Trying V4L2 API." << endl;
|
|
CLEAR(V4L2_capabilities);
|
|
|
|
if (-1 != xioctl (VIDIOC_QUERYCAP, &V4L2_capabilities))
|
|
{
|
|
if (!(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_CAPTURE))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " is not a video capture device." << endl;
|
|
m_driver = VIDEODEV_DRIVER_NONE;
|
|
return EXIT_FAILURE;
|
|
}
|
|
m_videocapture=true;
|
|
kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " is a V4L2 device." << endl;
|
|
m_driver = VIDEODEV_DRIVER_V4L2;
|
|
m_model=QString::fromLocal8Bit((const char*)V4L2_capabilities.card);
|
|
|
|
|
|
// Detect maximum and minimum resolution supported by the V4L2 device
|
|
CLEAR (fmt);
|
|
fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
if (-1 == xioctl (VIDIOC_G_FMT, &fmt))
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ")." << endl;
|
|
fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
fmt.fmt.pix.width = 32767;
|
|
fmt.fmt.pix.height = 32767;
|
|
fmt.fmt.pix.field = V4L2_FIELD_ANY;
|
|
if (-1 == xioctl (VIDIOC_S_FMT, &fmt))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Detecting maximum size with VIDIOC_S_FMT failed (" << errno << ").Returned maxwidth: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
|
|
// Note VIDIOC_S_FMT may change width and height.
|
|
}
|
|
else
|
|
{
|
|
maxwidth = fmt.fmt.pix.width;
|
|
maxheight = fmt.fmt.pix.height;
|
|
}
|
|
if (-1 == xioctl (VIDIOC_G_FMT, &fmt))
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ")." << endl;
|
|
fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
fmt.fmt.pix.width = 1;
|
|
fmt.fmt.pix.height = 1;
|
|
fmt.fmt.pix.field = V4L2_FIELD_ANY;
|
|
if (-1 == xioctl (VIDIOC_S_FMT, &fmt))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Detecting minimum size with VIDIOC_S_FMT failed (" << errno << ").Returned maxwidth: " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
|
|
// Note VIDIOC_S_FMT may change width and height.
|
|
}
|
|
else
|
|
{
|
|
minwidth = fmt.fmt.pix.width;
|
|
minheight = fmt.fmt.pix.height;
|
|
}
|
|
|
|
// Buggy driver paranoia
|
|
/* min = fmt.fmt.pix.width * 2;
|
|
if (fmt.fmt.pix.bytesperline < min)
|
|
fmt.fmt.pix.bytesperline = min;
|
|
min = fmt.fmt.pix.bytesperline * fmt.fmt.pix.height;
|
|
if (fmt.fmt.pix.sizeimage < min)
|
|
fmt.fmt.pix.sizeimage = min;
|
|
m_buffer_size=fmt.fmt.pix.sizeimage ;*/
|
|
|
|
int inputisok=EXIT_SUCCESS;
|
|
m_input.clear();
|
|
for(unsigned int loop=0; inputisok==EXIT_SUCCESS; loop++)
|
|
{
|
|
struct v4l2_input videoinput;
|
|
CLEAR(videoinput);
|
|
videoinput.index = loop;
|
|
inputisok=xioctl(VIDIOC_ENUMINPUT, &videoinput);
|
|
if(inputisok==EXIT_SUCCESS)
|
|
{
|
|
VideoInput tempinput;
|
|
tempinput.name = QString::fromLocal8Bit((const char*)videoinput.name);
|
|
tempinput.hastuner = videoinput.type & V4L2_INPUT_TYPE_TUNER;
|
|
tempinput.m_standards = videoinput.std;
|
|
m_input.push_back(tempinput);
|
|
kdDebug(14010) << k_funcinfo << "Input " << loop << ": " << tempinput.name << " (tuner: " << ((videoinput.type & V4L2_INPUT_TYPE_TUNER) != 0) << ")" << endl;
|
|
if((videoinput.type & V4L2_INPUT_TYPE_TUNER) != 0)
|
|
{
|
|
// _tunerForInput[name] = desc.tuner;
|
|
// _isTuner = true;
|
|
}
|
|
else
|
|
{
|
|
// _tunerForInput[name] = -1;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
|
|
|
|
// -----------------------------------------------------------------------------------------------------------------
|
|
// This must turn up to be a proper method to check for controls' existence.
|
|
CLEAR (queryctrl);
|
|
// v4l2_queryctrl may zero the .id in some cases, even if the IOCTL returns EXIT_SUCCESS (tested with a bttv card, when testing for V4L2_CID_AUDIO_VOLUME).
|
|
// As of 6th Aug 2007, according to the V4L2 specification version 0.21, this behavior is undocumented, and the example 1-8 code found at
|
|
// http://www.linuxtv.org/downloads/video4linux/API/V4L2_API/spec/x519.htm fails because of this behavior with a bttv card.
|
|
|
|
int currentid = V4L2_CID_BASE;
|
|
|
|
kdDebug(14010) << k_funcinfo << "Checking CID controls" << endl;
|
|
|
|
for (currentid = V4L2_CID_BASE; currentid < V4L2_CID_LASTP1; currentid++)
|
|
//for (queryctrl.id = 9963776; queryctrl.id < 9963800; queryctrl.id++)
|
|
{
|
|
queryctrl.id = currentid;
|
|
//kdDebug(14010) << k_funcinfo << "Checking CID controls from " << V4L2_CID_BASE << " to " << V4L2_CID_LASTP1 << ". Current: " << queryctrl.id << ". IOCTL returns: " << resultado << endl;
|
|
if (0 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
|
|
{
|
|
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
|
|
continue;
|
|
|
|
//kdDebug(14010) << k_funcinfo << " Control: " << QString::fromLocal8Bit((const char*)queryctrl.name) << endl;
|
|
kdDebug(14010) << k_funcinfo << " Control: " << QString::fromLocal8Bit((const char*)queryctrl.name) << " Values from " << queryctrl.minimum << " to " << queryctrl.maximum << " with steps of " << queryctrl.step << ". Default: " << queryctrl.default_value << endl;
|
|
|
|
/* switch (queryctrl.type)
|
|
{
|
|
case V4L2_CTRL_TYPE_INTEGER :
|
|
}*/
|
|
if (queryctrl.type == V4L2_CTRL_TYPE_MENU)
|
|
enumerateMenu ();
|
|
}
|
|
else
|
|
{
|
|
if (errno == EINVAL)
|
|
continue;
|
|
|
|
perror ("VIDIOC_QUERYCTRL");
|
|
// exit (EXIT_FAILURE);
|
|
}
|
|
}
|
|
|
|
kdDebug(14010) << k_funcinfo << "Checking CID private controls" << endl;
|
|
|
|
for (currentid = V4L2_CID_PRIVATE_BASE;; currentid++)
|
|
//for (queryctrl.id = 9963776; queryctrl.id < 9963800; queryctrl.id++)
|
|
{
|
|
queryctrl.id = currentid;
|
|
//kdDebug(14010) << k_funcinfo << "Checking CID private controls from " << V4L2_CID_PRIVATE_BASE << ". Current: " << queryctrl.id << ". IOCTL returns: " << resultado << endl;
|
|
if ( 0 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
|
|
{
|
|
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
|
|
continue;
|
|
|
|
kdDebug(14010) << k_funcinfo << " Control: " << QString::fromLocal8Bit((const char*)queryctrl.name) << " Values from " << queryctrl.minimum << " to " << queryctrl.maximum << " with steps of " << queryctrl.step << ". Default: " << queryctrl.default_value << endl;
|
|
|
|
if (queryctrl.type == V4L2_CTRL_TYPE_MENU)
|
|
enumerateMenu ();
|
|
}
|
|
else
|
|
{
|
|
if (errno == EINVAL)
|
|
break;
|
|
|
|
perror ("VIDIOC_QUERYCTRL");
|
|
// exit (EXIT_FAILURE);
|
|
}
|
|
}
|
|
|
|
|
|
|
|
|
|
}
|
|
else
|
|
{
|
|
// V4L-only drivers should return an EINVAL in errno to indicate they cannot handle V4L2 calls. Not every driver is compliant, so
|
|
// it will try the V4L api even if the error code is different than expected.
|
|
kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " is not a V4L2 device." << endl;
|
|
}
|
|
|
|
}
|
|
#endif
|
|
|
|
CLEAR(V4L_capabilities);
|
|
|
|
if(m_driver==VIDEODEV_DRIVER_NONE)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " Trying V4L API." << endl;
|
|
if (-1 == xioctl (VIDIOCGCAP, &V4L_capabilities))
|
|
{
|
|
perror ("ioctl (VIDIOCGCAP)");
|
|
m_driver = VIDEODEV_DRIVER_NONE;
|
|
return EXIT_FAILURE;
|
|
}
|
|
else
|
|
{
|
|
kdDebug(14010) << k_funcinfo << full_filename << " is a V4L device." << endl;
|
|
m_driver = VIDEODEV_DRIVER_V4L;
|
|
m_model=QString::fromLocal8Bit((const char*)V4L_capabilities.name);
|
|
if(V4L_capabilities.type & VID_TYPE_CAPTURE)
|
|
m_videocapture=true;
|
|
if(V4L_capabilities.type & VID_TYPE_CHROMAKEY)
|
|
m_videochromakey=true;
|
|
if(V4L_capabilities.type & VID_TYPE_SCALES)
|
|
m_videoscale=true;
|
|
if(V4L_capabilities.type & VID_TYPE_OVERLAY)
|
|
m_videooverlay=true;
|
|
// kdDebug(14010) << "libkopete (avdevice): Inputs : " << V4L_capabilities.channels << endl;
|
|
// kdDebug(14010) << "libkopete (avdevice): Audios : " << V4L_capabilities.audios << endl;
|
|
minwidth = V4L_capabilities.minwidth;
|
|
maxwidth = V4L_capabilities.maxwidth;
|
|
minheight = V4L_capabilities.minheight;
|
|
maxheight = V4L_capabilities.maxheight;
|
|
|
|
|
|
int inputisok=EXIT_SUCCESS;
|
|
m_input.clear();
|
|
for(int loop=0; loop < V4L_capabilities.channels; loop++)
|
|
{
|
|
struct video_channel videoinput;
|
|
CLEAR(videoinput);
|
|
videoinput.channel = loop;
|
|
videoinput.norm = 1;
|
|
inputisok=xioctl(VIDIOCGCHAN, &videoinput);
|
|
if(inputisok==EXIT_SUCCESS)
|
|
{
|
|
VideoInput tempinput;
|
|
tempinput.name = QString::fromLocal8Bit((const char*)videoinput.name);
|
|
tempinput.hastuner=videoinput.flags & VIDEO_VC_TUNER;
|
|
// TODO: The routine to detect the appropriate video standards for V4L must be placed here
|
|
m_input.push_back(tempinput);
|
|
// kdDebug(14010) << "libkopete (avdevice): Input " << loop << ": " << tempinput.name << " (tuner: " << ((videoinput.flags & VIDEO_VC_TUNER) != 0) << ")" << endl;
|
|
/* if((input.type & V4L2_INPUT_TYPE_TUNER) != 0)
|
|
{
|
|
// _tunerForInput[name] = desc.tuner;
|
|
// _isTuner = true;
|
|
}
|
|
else
|
|
{
|
|
// _tunerForInput[name] = -1;
|
|
}
|
|
*/ }
|
|
}
|
|
|
|
}
|
|
}
|
|
#endif
|
|
m_name=m_model; // Take care about changing the name to be different from the model itself...
|
|
|
|
detectPixelFormats();
|
|
|
|
// TODO: Now we must execute the proper initialization according to the type of the driver.
|
|
kdDebug(14010) << k_funcinfo << "checkDevice() exited successfuly." << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
|
|
/*!
|
|
\fn VideoDevice::showDeviceCapabilities()
|
|
*/
|
|
int VideoDevice::showDeviceCapabilities()
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "showDeviceCapabilities() called." << endl;
|
|
if(isOpen())
|
|
{
|
|
/* kdDebug(14010) << "libkopete (avdevice): Driver: " << (const char*)V4L2_capabilities.driver << " "
|
|
<< ((V4L2_capabilities.version>>16) & 0xFF) << "."
|
|
<< ((V4L2_capabilities.version>> 8) & 0xFF) << "."
|
|
<< ((V4L2_capabilities.version ) & 0xFF) << endl;
|
|
kdDebug(14010) << "libkopete (avdevice): Card: " << name << endl;
|
|
kdDebug(14010) << "libkopete (avdevice): Capabilities:" << endl;
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_CAPTURE)
|
|
kdDebug(14010) << "libkopete (avdevice): Video capture" << endl;
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_OUTPUT)
|
|
kdDebug(14010) << "libkopete (avdevice): Video output" << endl;
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_OVERLAY)
|
|
kdDebug(14010) << "libkopete (avdevice): Video overlay" << endl;
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_VBI_CAPTURE)
|
|
kdDebug(14010) << "libkopete (avdevice): VBI capture" << endl;
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_VBI_OUTPUT)
|
|
kdDebug(14010) << "libkopete (avdevice): VBI output" << endl;
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_RDS_CAPTURE)
|
|
kdDebug(14010) << "libkopete (avdevice): RDS capture" << endl;
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_TUNER)
|
|
kdDebug(14010) << "libkopete (avdevice): Tuner IO" << endl;
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_AUDIO)
|
|
kdDebug(14010) << "libkopete (avdevice): Audio IO" << endl;
|
|
;*/
|
|
kdDebug(14010) << k_funcinfo << "Card model: " << m_model << endl;
|
|
kdDebug(14010) << k_funcinfo << "Card name : " << m_name << endl;
|
|
kdDebug(14010) << k_funcinfo << "Capabilities:" << endl;
|
|
if(canCapture())
|
|
kdDebug(14010) << k_funcinfo << " Video capture" << endl;
|
|
if(canRead())
|
|
kdDebug(14010) << k_funcinfo << " Read" << endl;
|
|
if(canAsyncIO())
|
|
kdDebug(14010) << k_funcinfo << " Asynchronous input/output" << endl;
|
|
if(canStream())
|
|
kdDebug(14010) << k_funcinfo << " Streaming" << endl;
|
|
if(canChromakey())
|
|
kdDebug(14010) << k_funcinfo << " Video chromakey" << endl;
|
|
if(canScale())
|
|
kdDebug(14010) << k_funcinfo << " Video scales" << endl;
|
|
if(canOverlay())
|
|
kdDebug(14010) << k_funcinfo << " Video overlay" << endl;
|
|
// kdDebug(14010) << "libkopete (avdevice): Audios : " << V4L_capabilities.audios << endl;
|
|
kdDebug(14010) << k_funcinfo << " Max res: " << maxWidth() << " x " << maxHeight() << endl;
|
|
kdDebug(14010) << k_funcinfo << " Min res: " << minWidth() << " x " << minHeight() << endl;
|
|
kdDebug(14010) << k_funcinfo << " Inputs : " << inputs() << endl;
|
|
for (unsigned int loop=0; loop < inputs(); loop++)
|
|
kdDebug(14010) << k_funcinfo << "Input " << loop << ": " << m_input[loop].name << " (tuner: " << m_input[loop].hastuner << ")" << endl;
|
|
kdDebug(14010) << k_funcinfo << "showDeviceCapabilities() exited successfuly." << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
/*!
|
|
\fn VideoDevicePool::initDevice()
|
|
*/
|
|
int VideoDevice::initDevice()
|
|
{
|
|
/// @todo implement me
|
|
kdDebug(14010) << k_funcinfo << "initDevice() started" << endl;
|
|
if(-1 == descriptor)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "initDevice() Device is not open" << endl;
|
|
return EXIT_FAILURE;
|
|
}
|
|
m_io_method = IO_METHOD_NONE;
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_READWRITE)
|
|
{
|
|
m_videoread=true;
|
|
m_io_method = IO_METHOD_READ;
|
|
kdDebug(14010) << k_funcinfo << " Read/Write interface" << endl;
|
|
}
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_ASYNCIO)
|
|
{
|
|
m_videoasyncio=true;
|
|
kdDebug(14010) << k_funcinfo << " Async IO interface" << endl;
|
|
}
|
|
if(V4L2_capabilities.capabilities & V4L2_CAP_STREAMING)
|
|
{
|
|
m_videostream=true;
|
|
m_io_method = IO_METHOD_MMAP;
|
|
// m_io_method = IO_METHOD_USERPTR;
|
|
kdDebug(14010) << k_funcinfo << " Streaming interface" << endl;
|
|
}
|
|
if(m_io_method==IO_METHOD_NONE)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "initDevice() Found no suitable input/output method for " << full_filename << endl;
|
|
return EXIT_FAILURE;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
m_videoread=true;
|
|
m_io_method=IO_METHOD_READ;
|
|
if(-1 != xioctl(VIDIOCGFBUF,&V4L_videobuffer))
|
|
{
|
|
// m_videostream=true;
|
|
// m_io_method = IO_METHOD_MMAP;
|
|
kdDebug(14010) << k_funcinfo << " Streaming interface" << endl;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
|
|
break;
|
|
}
|
|
|
|
// Select video input, video standard and tune here.
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
cropcap.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
if (-1 == xioctl (VIDIOC_CROPCAP, &cropcap))
|
|
{ // Errors ignored.
|
|
}
|
|
crop.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
crop.c = cropcap.defrect; // reset to default
|
|
if (-1 == xioctl (VIDIOC_S_CROP, &crop))
|
|
{
|
|
switch (errno)
|
|
{
|
|
case EINVAL: break; // Cropping not supported.
|
|
default: break; // Errors ignored.
|
|
}
|
|
}
|
|
#endif
|
|
#endif
|
|
|
|
showDeviceCapabilities();
|
|
kdDebug(14010) << k_funcinfo << "initDevice() exited successfuly" << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
|
|
unsigned int VideoDevice::inputs()
|
|
{
|
|
return m_input.size();
|
|
}
|
|
|
|
|
|
int VideoDevice::width()
|
|
{
|
|
return currentwidth;
|
|
}
|
|
|
|
int VideoDevice::minWidth()
|
|
{
|
|
return minwidth;
|
|
}
|
|
|
|
int VideoDevice::maxWidth()
|
|
{
|
|
return maxwidth;
|
|
}
|
|
|
|
int VideoDevice::height()
|
|
{
|
|
return currentheight;
|
|
}
|
|
|
|
int VideoDevice::minHeight()
|
|
{
|
|
return minheight;
|
|
}
|
|
|
|
int VideoDevice::maxHeight()
|
|
{
|
|
return maxheight;
|
|
}
|
|
|
|
int VideoDevice::setSize( int newwidth, int newheight)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "setSize(" << newwidth << ", " << newheight << ") called." << endl;
|
|
if(isOpen())
|
|
{
|
|
// It should not be there. It must remain in a completely distict place, cause this method should not change the pixelformat.
|
|
kdDebug(14010) << k_funcinfo << "Trying YUY422P" << endl;
|
|
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_YUV422P))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support YUV422P format. Trying YUYV." << endl;
|
|
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_YUYV))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support YUYV format. Trying UYVY." << endl;
|
|
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_UYVY))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support UYVY format. Trying YUV420P." << endl;
|
|
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_YUV420P))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support YUV420P format. Trying RGB24." << endl;
|
|
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_RGB24))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support RGB24 format. Trying BGR24." << endl;
|
|
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_BGR24))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support RGB24 format. Trying RGB32." << endl;
|
|
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_RGB32))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support RGB32 format. Trying BGR32." << endl;
|
|
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_BGR32))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support BGR32 format. Trying SN9C10X." << endl;
|
|
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_SN9C10X))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support SN9C10X format. Trying Bayer RGB." << endl;
|
|
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_SBGGR8))
|
|
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support SBGGR8 format. Fallback from it is not yet implemented." << endl;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if(newwidth > maxwidth ) newwidth = maxwidth;
|
|
if(newheight > maxheight) newheight = maxheight;
|
|
if(newwidth < minwidth ) newwidth = minwidth;
|
|
if(newheight < minheight) newheight = minheight;
|
|
|
|
currentwidth = newwidth;
|
|
currentheight = newheight;
|
|
|
|
//kdDebug(14010) << k_funcinfo << "width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << width() << "x" << height() << endl;
|
|
// Change resolution for the video device
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
// CLEAR (fmt);
|
|
if (-1 == xioctl (VIDIOC_G_FMT, &fmt))
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
|
|
fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
fmt.fmt.pix.width = width();
|
|
fmt.fmt.pix.height = height();
|
|
fmt.fmt.pix.field = V4L2_FIELD_ANY;
|
|
if (-1 == xioctl (VIDIOC_S_FMT, &fmt))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_S_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
|
|
// Note VIDIOC_S_FMT may change width and height.
|
|
}
|
|
else
|
|
{
|
|
// Buggy driver paranoia.
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_S_FMT worked (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
|
|
unsigned int min = fmt.fmt.pix.width * 2;
|
|
if (fmt.fmt.pix.bytesperline < min)
|
|
fmt.fmt.pix.bytesperline = min;
|
|
min = fmt.fmt.pix.bytesperline * fmt.fmt.pix.height;
|
|
if (fmt.fmt.pix.sizeimage < min)
|
|
fmt.fmt.pix.sizeimage = min;
|
|
m_buffer_size=fmt.fmt.pix.sizeimage ;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
{
|
|
struct video_window V4L_videowindow;
|
|
|
|
kdDebug(14010) << "------------- width: " << V4L_videowindow.width << " Height: " << V4L_videowindow.height << " Clipcount: " << V4L_videowindow.clipcount << " -----------------" << endl;
|
|
|
|
if (xioctl (VIDIOCGWIN, &V4L_videowindow)== -1)
|
|
{
|
|
perror ("ioctl VIDIOCGWIN");
|
|
// return (NULL);
|
|
}
|
|
V4L_videowindow.width = width();
|
|
V4L_videowindow.height = height();
|
|
V4L_videowindow.clipcount=0;
|
|
if (xioctl (VIDIOCSWIN, &V4L_videowindow)== -1)
|
|
{
|
|
perror ("ioctl VIDIOCSWIN");
|
|
// return (NULL);
|
|
}
|
|
kdDebug(14010) << "------------- width: " << V4L_videowindow.width << " Height: " << V4L_videowindow.height << " Clipcount: " << V4L_videowindow.clipcount << " -----------------" << endl;
|
|
|
|
// kdDebug(14010) << "libkopete (avdevice): V4L_picture.palette: " << V4L_picture.palette << " Depth: " << V4L_picture.depth << endl;
|
|
|
|
/* if(-1 == xioctl(VIDIOCGFBUF,&V4L_videobuffer))
|
|
kdDebug(14010) << "libkopete (avdevice): VIDIOCGFBUF failed (" << errno << "): Card cannot stream" << endl;*/
|
|
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
m_buffer_size = width() * height() * pixelFormatDepth(m_pixelformat) / 8;
|
|
kdDebug(14010) << "------------------------- ------- -- m_buffer_size: " << m_buffer_size << " !!! -- ------- -----------------------------------------" << endl;
|
|
|
|
m_currentbuffer.pixelformat=m_pixelformat;
|
|
m_currentbuffer.data.resize(m_buffer_size);
|
|
|
|
switch (m_io_method)
|
|
{
|
|
case IO_METHOD_NONE: break;
|
|
case IO_METHOD_READ: initRead (); break;
|
|
case IO_METHOD_MMAP: initMmap (); break;
|
|
case IO_METHOD_USERPTR: initUserptr (); break;
|
|
}
|
|
|
|
kdDebug(14010) << k_funcinfo << "setSize(" << newwidth << ", " << newheight << ") exited successfuly." << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
kdDebug(14010) << k_funcinfo << "setSize(" << newwidth << ", " << newheight << ") Device is not open." << endl;
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
pixel_format VideoDevice::setPixelFormat(pixel_format newformat)
|
|
{
|
|
pixel_format ret = PIXELFORMAT_NONE;
|
|
//kdDebug(14010) << k_funcinfo << "called." << endl;
|
|
// Change the pixel format for the video device
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
// CLEAR (fmt);
|
|
if (-1 == xioctl (VIDIOC_G_FMT, &fmt))
|
|
{
|
|
// return errnoReturn ("VIDIOC_S_FMT");
|
|
// kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
|
|
}
|
|
else
|
|
m_pixelformat = pixelFormatForPalette(fmt.fmt.pix.pixelformat);
|
|
|
|
fmt.fmt.pix.pixelformat = pixelFormatCode(newformat);
|
|
if (-1 == xioctl (VIDIOC_S_FMT, &fmt))
|
|
{
|
|
// kdDebug(14010) << k_funcinfo << "VIDIOC_S_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
|
|
}
|
|
else
|
|
{
|
|
if (fmt.fmt.pix.pixelformat == pixelFormatCode(newformat))
|
|
{
|
|
m_pixelformat = newformat;
|
|
ret = m_pixelformat;
|
|
}
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
{
|
|
struct video_picture V4L_picture;
|
|
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
|
|
// kdDebug(14010) << k_funcinfo << "V4L_picture.palette: " << V4L_picture.palette << " Depth: " << V4L_picture.depth << endl;
|
|
V4L_picture.palette = pixelFormatCode(newformat);
|
|
V4L_picture.depth = pixelFormatDepth(newformat);
|
|
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
|
|
{
|
|
// kdDebug(14010) << k_funcinfo << "Card seems to not support " << pixelFormatName(newformat) << " format. Fallback to it is not yet implemented." << endl;
|
|
}
|
|
|
|
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
|
|
|
|
// kdDebug(14010) << k_funcinfo << "V4L_picture.palette: " << V4L_picture.palette << " Depth: " << V4L_picture.depth << endl;
|
|
m_pixelformat=pixelFormatForPalette(V4L_picture.palette);
|
|
if (m_pixelformat == newformat)
|
|
ret = newformat;
|
|
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/*!
|
|
\fn Kopete::AV::VideoDevice::currentInput()
|
|
*/
|
|
int VideoDevice::currentInput()
|
|
{
|
|
/// @todo implement me
|
|
if(isOpen())
|
|
{
|
|
return m_current_input;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*!
|
|
\fn Kopete::AV::VideoDevice::selectInput(int input)
|
|
*/
|
|
int VideoDevice::selectInput(int newinput)
|
|
{
|
|
/// @todo implement me
|
|
if(m_current_input >= inputs())
|
|
return EXIT_FAILURE;
|
|
|
|
if(isOpen())
|
|
{
|
|
switch (m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
if (-1 == ioctl (descriptor, VIDIOC_S_INPUT, &newinput))
|
|
{
|
|
perror ("VIDIOC_S_INPUT");
|
|
return EXIT_FAILURE;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
struct video_channel V4L_input;
|
|
V4L_input.channel=newinput;
|
|
V4L_input.norm=4; // Hey, it's plain wrong! It should be input's signal standard!
|
|
if (-1 == ioctl (descriptor, VIDIOCSCHAN, &V4L_input))
|
|
{
|
|
perror ("ioctl (VIDIOCSCHAN)");
|
|
return EXIT_FAILURE;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
kdDebug(14010) << k_funcinfo << "Selected input " << newinput << " (" << m_input[newinput].name << ")" << endl;
|
|
m_current_input = newinput;
|
|
setInputParameters();
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
/*!
|
|
\fn Kopete::AV::VideoDevice::setInputParameters()
|
|
*/
|
|
int VideoDevice::setInputParameters()
|
|
{
|
|
/// @todo implement me
|
|
if( (isOpen()) && (m_current_input < inputs() ) )
|
|
{
|
|
setBrightness( getBrightness() );
|
|
setContrast( getContrast() );
|
|
setSaturation( getSaturation() );
|
|
setWhiteness( getWhiteness() );
|
|
setHue( getHue() );
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
/*!
|
|
\fn VideoDevice::startCapturing()
|
|
*/
|
|
int VideoDevice::startCapturing()
|
|
{
|
|
|
|
kdDebug(14010) << k_funcinfo << "called." << endl;
|
|
if(isOpen())
|
|
{
|
|
switch (m_io_method)
|
|
{
|
|
case IO_METHOD_NONE: // Card cannot capture frames
|
|
return EXIT_FAILURE;
|
|
break;
|
|
case IO_METHOD_READ: // Nothing to do
|
|
break;
|
|
case IO_METHOD_MMAP:
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
{
|
|
unsigned int loop;
|
|
for (loop = 0; loop < m_streambuffers; ++loop)
|
|
{
|
|
struct v4l2_buffer buf;
|
|
CLEAR (buf);
|
|
buf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
buf.memory = V4L2_MEMORY_MMAP;
|
|
buf.index = loop;
|
|
if (-1 == xioctl (VIDIOC_QBUF, &buf))
|
|
return errnoReturn ("VIDIOC_QBUF");
|
|
}
|
|
enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
if (-1 == xioctl (VIDIOC_STREAMON, &type))
|
|
return errnoReturn ("VIDIOC_STREAMON");
|
|
}
|
|
#endif
|
|
#endif
|
|
break;
|
|
case IO_METHOD_USERPTR:
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
{
|
|
unsigned int loop;
|
|
for (loop = 0; loop < m_streambuffers; ++loop)
|
|
{
|
|
struct v4l2_buffer buf;
|
|
CLEAR (buf);
|
|
buf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
buf.memory = V4L2_MEMORY_USERPTR;
|
|
buf.m.userptr = (unsigned long) m_rawbuffers[loop].start;
|
|
buf.length = m_rawbuffers[loop].length;
|
|
if (-1 == xioctl (VIDIOC_QBUF, &buf))
|
|
return errnoReturn ("VIDIOC_QBUF");
|
|
}
|
|
enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
if (-1 == xioctl (VIDIOC_STREAMON, &type))
|
|
return errnoReturn ("VIDIOC_STREAMON");
|
|
}
|
|
#endif
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
kdDebug(14010) << k_funcinfo << "exited successfuly." << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
/*!
|
|
\fn VideoDevice::getFrame()
|
|
*/
|
|
int VideoDevice::getFrame()
|
|
{
|
|
/// @todo implement me
|
|
ssize_t bytesread;
|
|
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
struct v4l2_buffer v4l2buffer;
|
|
#endif
|
|
#endif
|
|
// kdDebug(14010) << k_funcinfo << "getFrame() called." << endl;
|
|
if(isOpen())
|
|
{
|
|
switch (m_io_method)
|
|
{
|
|
case IO_METHOD_NONE: // Card cannot capture frames
|
|
return EXIT_FAILURE;
|
|
break;
|
|
case IO_METHOD_READ:
|
|
// kdDebug(14010) << k_funcinfo << "Using IO_METHOD_READ.File descriptor: " << descriptor << " Buffer address: " << &m_currentbuffer.data[0] << " Size: " << m_currentbuffer.data.size() << endl;
|
|
bytesread = read (descriptor, &m_currentbuffer.data[0], m_currentbuffer.data.size());
|
|
if (-1 == bytesread) // must verify this point with ov511 driver.
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "IO_METHOD_READ failed." << endl;
|
|
switch (errno)
|
|
{
|
|
case EAGAIN:
|
|
return EXIT_FAILURE;
|
|
case EIO: /* Could ignore EIO, see spec. fall through */
|
|
default:
|
|
return errnoReturn ("read");
|
|
}
|
|
}
|
|
if((int)m_currentbuffer.data.size() < bytesread)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "IO_METHOD_READ returned less bytes (" << bytesread << ") than it was asked for (" << m_currentbuffer.data.size() <<")." << endl;
|
|
}
|
|
break;
|
|
case IO_METHOD_MMAP:
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
CLEAR (v4l2buffer);
|
|
v4l2buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
v4l2buffer.memory = V4L2_MEMORY_MMAP;
|
|
if (-1 == xioctl (VIDIOC_DQBUF, &v4l2buffer))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << full_filename << " MMAPed getFrame failed." << endl;
|
|
switch (errno)
|
|
{
|
|
case EAGAIN:
|
|
{
|
|
kdDebug(14010) << k_funcinfo << full_filename << " MMAPed getFrame failed: EAGAIN. Pointer: " << endl;
|
|
return EXIT_FAILURE;
|
|
}
|
|
case EIO: /* Could ignore EIO, see spec. fall through */
|
|
default:
|
|
return errnoReturn ("VIDIOC_DQBUF");
|
|
}
|
|
}
|
|
/* if (v4l2buffer.index < m_streambuffers)
|
|
return EXIT_FAILURE;*/ //it was an assert()
|
|
//kdDebug(14010) << k_funcinfo << "m_rawbuffers[" << v4l2buffer.index << "].start: " << (void *)m_rawbuffers[v4l2buffer.index].start << " Size: " << m_currentbuffer.data.size() << endl;
|
|
|
|
|
|
|
|
/*{
|
|
unsigned long long result=0;
|
|
unsigned long long R=0, G=0, B=0, A=0;
|
|
int Rmax=0, Gmax=0, Bmax=0, Amax=0;
|
|
int Rmin=255, Gmin=255, Bmin=255, Amin=0;
|
|
|
|
for(unsigned int loop=0;loop < m_currentbuffer.data.size();loop+=4)
|
|
{
|
|
R+=m_rawbuffers[v4l2buffer.index].start[loop];
|
|
G+=m_rawbuffers[v4l2buffer.index].start[loop+1];
|
|
B+=m_rawbuffers[v4l2buffer.index].start[loop+2];
|
|
// A+=currentbuffer.data[loop+3];
|
|
if (m_currentbuffer.data[loop] < Rmin) Rmin = m_currentbuffer.data[loop];
|
|
if (m_currentbuffer.data[loop+1] < Gmin) Gmin = m_currentbuffer.data[loop+1];
|
|
if (m_currentbuffer.data[loop+2] < Bmin) Bmin = m_currentbuffer.data[loop+2];
|
|
// if (m_currentbuffer.data[loop+3] < Amin) Amin = m_currentbuffer.data[loop+3];
|
|
if (m_currentbuffer.data[loop] > Rmax) Rmax = m_currentbuffer.data[loop];
|
|
if (m_currentbuffer.data[loop+1] > Gmax) Gmax = m_currentbuffer.data[loop+1];
|
|
if (m_currentbuffer.data[loop+2] > Bmax) Bmax = m_currentbuffer.data[loop+2];
|
|
// if (m_currentbuffer.data[loop+3] > Amax) Amax = m_currentbuffer.data[loop+3];
|
|
}
|
|
kdDebug(14010) << " R: " << R << " G: " << G << " B: " << B << " A: " << A <<
|
|
" Rmin: " << Rmin << " Gmin: " << Gmin << " Bmin: " << Bmin << " Amin: " << Amin <<
|
|
" Rmax: " << Rmax << " Gmax: " << Gmax << " Bmax: " << Bmax << " Amax: " << Amax << endl;
|
|
}*/
|
|
|
|
|
|
memcpy(&m_currentbuffer.data[0], m_rawbuffers[v4l2buffer.index].start, m_currentbuffer.data.size());
|
|
if (-1 == xioctl (VIDIOC_QBUF, &v4l2buffer))
|
|
return errnoReturn ("VIDIOC_QBUF");
|
|
#endif
|
|
#endif
|
|
break;
|
|
case IO_METHOD_USERPTR:
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
{
|
|
unsigned int i;
|
|
CLEAR (v4l2buffer);
|
|
v4l2buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
v4l2buffer.memory = V4L2_MEMORY_USERPTR;
|
|
if (-1 == xioctl (VIDIOC_DQBUF, &v4l2buffer))
|
|
{
|
|
switch (errno)
|
|
{
|
|
case EAGAIN:
|
|
return EXIT_FAILURE;
|
|
case EIO: /* Could ignore EIO, see spec. fall through */
|
|
default:
|
|
return errnoReturn ("VIDIOC_DQBUF");
|
|
}
|
|
}
|
|
for (i = 0; i < m_streambuffers; ++i)
|
|
if (v4l2buffer.m.userptr == (unsigned long) m_rawbuffers[i].start && v4l2buffer.length == m_rawbuffers[i].length)
|
|
break;
|
|
if (i < m_streambuffers)
|
|
return EXIT_FAILURE;
|
|
if (-1 == xioctl (VIDIOC_QBUF, &v4l2buffer))
|
|
return errnoReturn ("VIDIOC_QBUF");
|
|
}
|
|
#endif
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
/* Automatic color correction. Now it just swaps R and B channels in RGB24/BGR24 modes.
|
|
if(m_input[m_current_input].getAutoColorCorrection())
|
|
{
|
|
switch(m_currentbuffer.pixelformat)
|
|
{
|
|
case PIXELFORMAT_NONE : break;
|
|
case PIXELFORMAT_GREY : break;
|
|
case PIXELFORMAT_RGB332 : break;
|
|
case PIXELFORMAT_RGB555 : break;
|
|
case PIXELFORMAT_RGB555X: break;
|
|
case PIXELFORMAT_RGB565 : break;
|
|
case PIXELFORMAT_RGB565X: break;
|
|
case PIXELFORMAT_RGB24 :
|
|
case PIXELFORMAT_BGR24 :
|
|
{
|
|
unsigned char temp;
|
|
for(unsigned int loop=0;loop < m_currentbuffer.data.size();loop+=3)
|
|
{
|
|
temp = m_currentbuffer.data[loop];
|
|
m_currentbuffer.data[loop] = m_currentbuffer.data[loop+2];
|
|
m_currentbuffer.data[loop+2] = temp;
|
|
}
|
|
}
|
|
break;
|
|
case PIXELFORMAT_RGB32 :
|
|
case PIXELFORMAT_BGR32 :
|
|
{
|
|
unsigned char temp;
|
|
for(unsigned int loop=0;loop < m_currentbuffer.data.size();loop+=4)
|
|
{
|
|
temp = m_currentbuffer.data[loop];
|
|
m_currentbuffer.data[loop] = m_currentbuffer.data[loop+2];
|
|
m_currentbuffer.data[loop+2] = temp;
|
|
}
|
|
}
|
|
break;
|
|
case PIXELFORMAT_YUYV : break;
|
|
case PIXELFORMAT_UYVY : break;
|
|
case PIXELFORMAT_YUV420P: break;
|
|
case PIXELFORMAT_YUV422P: break;
|
|
}
|
|
}*/
|
|
//kdDebug(14010) << k_funcinfo << "10 Using IO_METHOD_READ.File descriptor: " << descriptor << " Buffer address: " << &m_currentbuffer.data[0] << " Size: " << m_currentbuffer.data.size() << endl;
|
|
|
|
|
|
// put frame copy operation here
|
|
// kdDebug(14010) << k_funcinfo << "exited successfuly." << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
/*!
|
|
\fn VideoDevice::getFrame(imagebuffer *imgbuffer)
|
|
*/
|
|
int VideoDevice::getFrame(imagebuffer *imgbuffer)
|
|
{
|
|
if(imgbuffer)
|
|
{
|
|
getFrame();
|
|
imgbuffer->height = m_currentbuffer.height;
|
|
imgbuffer->width = m_currentbuffer.width;
|
|
imgbuffer->pixelformat = m_currentbuffer.pixelformat;
|
|
imgbuffer->data = m_currentbuffer.data;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
/*!
|
|
\fn Kopete::AV::VideoDevice::getImage(const QImage *qimage)
|
|
*/
|
|
int VideoDevice::getImage(QImage *qimage)
|
|
{
|
|
/// @todo implement me
|
|
|
|
// do NOT delete qimage here, as it is received as a parameter
|
|
if (qimage->width() != width() || qimage->height() != height())
|
|
qimage->create(width(), height(),32, QImage::IgnoreEndian);
|
|
|
|
uchar *bits=qimage->bits();
|
|
// kDebug() << "Capturing in " << pixelFormatName(m_currentbuffer.pixelformat);
|
|
switch(m_currentbuffer.pixelformat)
|
|
{
|
|
case PIXELFORMAT_NONE : break;
|
|
|
|
// Packed RGB formats
|
|
case PIXELFORMAT_RGB332 : break;
|
|
case PIXELFORMAT_RGB444 : break;
|
|
case PIXELFORMAT_RGB555 : break;
|
|
case PIXELFORMAT_RGB565 :
|
|
{
|
|
int step=0;
|
|
for(int loop=0;loop < qimage->numBytes();loop+=4)
|
|
{
|
|
bits[loop] = (m_currentbuffer.data[step]<<3)+(m_currentbuffer.data[step]<<3>>5);
|
|
bits[loop+1] = ((m_currentbuffer.data[step+1])<<5)|m_currentbuffer.data[step]>>5;
|
|
bits[loop+2] = ((m_currentbuffer.data[step+1])&248)+((m_currentbuffer.data[step+1])>>5);
|
|
bits[loop+3] = 255;
|
|
step+=2;
|
|
}
|
|
}
|
|
break;
|
|
case PIXELFORMAT_RGB555X: break;
|
|
case PIXELFORMAT_RGB565X: break;
|
|
case PIXELFORMAT_BGR24 :
|
|
{
|
|
int step=0;
|
|
for(int loop=0;loop < qimage->numBytes();loop+=4)
|
|
{
|
|
bits[loop] = m_currentbuffer.data[step+2];
|
|
bits[loop+1] = m_currentbuffer.data[step+1];
|
|
bits[loop+2] = m_currentbuffer.data[step];
|
|
bits[loop+3] = 255;
|
|
step+=3;
|
|
}
|
|
}
|
|
break;
|
|
case PIXELFORMAT_RGB24 :
|
|
{
|
|
int step=0;
|
|
for(int loop=0;loop < qimage->numBytes();loop+=4)
|
|
{
|
|
bits[loop] = m_currentbuffer.data[step];
|
|
bits[loop+1] = m_currentbuffer.data[step+1];
|
|
bits[loop+2] = m_currentbuffer.data[step+2];
|
|
bits[loop+3] = 255;
|
|
step+=3;
|
|
}
|
|
}
|
|
break;
|
|
case PIXELFORMAT_BGR32 : break;
|
|
case PIXELFORMAT_RGB32 : memcpy(bits,&m_currentbuffer.data[0], m_currentbuffer.data.size());
|
|
break;
|
|
|
|
// Bayer RGB format
|
|
case PIXELFORMAT_SBGGR8 :
|
|
{
|
|
unsigned char *d = (unsigned char *) malloc (width() * height() * 3);
|
|
bayer2rgb24(d, &m_currentbuffer.data.first(), width(), height());
|
|
int step=0;
|
|
for(int loop=0;loop < qimage->numBytes();loop+=4)
|
|
{
|
|
bits[loop] = d[step+2];
|
|
bits[loop+1] = d[step+1];
|
|
bits[loop+2] = d[step];
|
|
bits[loop+3] = 255;
|
|
step+=3;
|
|
}
|
|
free(d);
|
|
}
|
|
break;
|
|
|
|
// YUV formats
|
|
case PIXELFORMAT_GREY : break;
|
|
case PIXELFORMAT_YUYV:
|
|
case PIXELFORMAT_UYVY:
|
|
case PIXELFORMAT_YUV420P:
|
|
case PIXELFORMAT_YUV422P:
|
|
{
|
|
uchar *yptr, *cbptr, *crptr;
|
|
bool halfheight=false;
|
|
bool packed=false;
|
|
// Adjust algorythm to specific YUV data arrangements.
|
|
if (m_currentbuffer.pixelformat == PIXELFORMAT_YUV420P)
|
|
halfheight=true;
|
|
if (m_currentbuffer.pixelformat == PIXELFORMAT_YUYV)
|
|
{
|
|
yptr = &m_currentbuffer.data[0];
|
|
cbptr = yptr + 1;
|
|
crptr = yptr + 3;
|
|
packed=true;
|
|
}
|
|
else if (m_currentbuffer.pixelformat == PIXELFORMAT_UYVY)
|
|
{
|
|
cbptr = &m_currentbuffer.data[0];
|
|
yptr = cbptr + 1;
|
|
crptr = cbptr + 2;
|
|
packed=true;
|
|
}
|
|
else
|
|
{
|
|
yptr = &m_currentbuffer.data[0];
|
|
cbptr = yptr + (width()*height());
|
|
crptr = cbptr + (width()*height()/(halfheight ? 4:2));
|
|
}
|
|
|
|
for(int y=0; y<height(); y++)
|
|
{
|
|
// Decode scanline
|
|
for(int x=0; x<width(); x++)
|
|
{
|
|
int c,d,e;
|
|
|
|
if (packed)
|
|
{
|
|
c = (yptr[x<<1])-16;
|
|
d = (cbptr[x>>1<<2])-128;
|
|
e = (crptr[x>>1<<2])-128;
|
|
}
|
|
else
|
|
{
|
|
c = (yptr[x])-16;
|
|
d = (cbptr[x>>1])-128;
|
|
e = (crptr[x>>1])-128;
|
|
}
|
|
|
|
int r = (298 * c + 409 * e + 128)>>8;
|
|
int g = (298 * c - 100 * d - 208 * e + 128)>>8;
|
|
int b = (298 * c + 516 * d + 128)>>8;
|
|
|
|
if (r<0) r=0; if (r>255) r=255;
|
|
if (g<0) g=0; if (g>255) g=255;
|
|
if (b<0) b=0; if (b>255) b=255;
|
|
|
|
uint *p = (uint*)qimage->scanLine(y)+x;
|
|
*p = qRgba(r,g,b,255);
|
|
|
|
}
|
|
// Jump to next line
|
|
if (packed)
|
|
{
|
|
yptr+=width()*2;
|
|
cbptr+=width()*2;
|
|
crptr+=width()*2;
|
|
}
|
|
else
|
|
{
|
|
yptr+=width();
|
|
if (!halfheight || y&1)
|
|
{
|
|
cbptr+=width()/2;
|
|
crptr+=width()/2;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
|
|
// Compressed formats
|
|
case PIXELFORMAT_JPEG : break;
|
|
case PIXELFORMAT_MPEG : break;
|
|
|
|
// Reserved formats
|
|
case PIXELFORMAT_DV : break;
|
|
case PIXELFORMAT_ET61X251:break;
|
|
case PIXELFORMAT_HI240 : break;
|
|
case PIXELFORMAT_HM12 : break;
|
|
case PIXELFORMAT_MJPEG : break;
|
|
case PIXELFORMAT_PWC1 : break;
|
|
case PIXELFORMAT_PWC2 : break;
|
|
case PIXELFORMAT_SN9C10X:
|
|
{
|
|
unsigned char *s = new unsigned char [width() * height()];
|
|
unsigned char *d = new unsigned char [width() * height() * 3];
|
|
sonix_decompress_init();
|
|
sonix_decompress(width(), height(), &m_currentbuffer.data.first(), s);
|
|
bayer2rgb24(d, s, width(), height());
|
|
int step=0;
|
|
for(int loop=0;loop < qimage->numBytes();loop+=4)
|
|
{
|
|
bits[loop] = d[step+2];
|
|
bits[loop+1] = d[step+1];
|
|
bits[loop+2] = d[step];
|
|
bits[loop+3] = 255;
|
|
step+=3;
|
|
}
|
|
delete[] s;
|
|
delete[] d;
|
|
}
|
|
break;
|
|
case PIXELFORMAT_WNVA : break;
|
|
case PIXELFORMAT_YYUV : break;
|
|
}
|
|
return EXIT_SUCCESS;
|
|
}
|
|
|
|
/*!
|
|
\fn VideoDevice::stopCapturing()
|
|
*/
|
|
int VideoDevice::stopCapturing()
|
|
{
|
|
/// @todo implement me
|
|
kdDebug(14010) << k_funcinfo << "called." << endl;
|
|
if(isOpen())
|
|
{
|
|
switch (m_io_method)
|
|
{
|
|
case IO_METHOD_NONE: // Card cannot capture frames
|
|
return EXIT_FAILURE;
|
|
break;
|
|
case IO_METHOD_READ: // Nothing to do
|
|
break;
|
|
case IO_METHOD_MMAP:
|
|
case IO_METHOD_USERPTR:
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
{
|
|
enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
if (-1 == xioctl (VIDIOC_STREAMOFF, &type))
|
|
return errnoReturn ("VIDIOC_STREAMOFF");
|
|
|
|
if (m_io_method == IO_METHOD_MMAP)
|
|
{
|
|
unsigned int loop;
|
|
for (loop = 0; loop < m_streambuffers; ++loop)
|
|
{
|
|
if (munmap(m_rawbuffers[loop].start,m_rawbuffers[loop].length) != 0)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "unable to munmap." << endl;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
break;
|
|
}
|
|
kdDebug(14010) << k_funcinfo << "exited successfuly." << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
|
|
/*!
|
|
\fn VideoDevice::close()
|
|
*/
|
|
int VideoDevice::close()
|
|
{
|
|
/// @todo implement me
|
|
kdDebug(14010) << k_funcinfo << " called." << endl;
|
|
if(isOpen())
|
|
{
|
|
kdDebug(14010) << k_funcinfo << " Device is open. Trying to properly shutdown the device." << endl;
|
|
stopCapturing();
|
|
kdDebug(14010) << k_funcinfo << "::close() returns " << ::close(descriptor) << endl;
|
|
}
|
|
descriptor = -1;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
|
|
float VideoDevice::getBrightness()
|
|
{
|
|
if (m_current_input < m_input.size() )
|
|
return m_input[m_current_input].getBrightness();
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
float VideoDevice::setBrightness(float brightness)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << " called." << endl;
|
|
m_input[m_current_input].setBrightness(brightness); // Just to check bounds
|
|
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
{
|
|
struct v4l2_queryctrl queryctrl;
|
|
struct v4l2_control control;
|
|
|
|
CLEAR (queryctrl);
|
|
queryctrl.id = V4L2_CID_BRIGHTNESS;
|
|
|
|
if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
|
|
{
|
|
if (errno != EINVAL)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl;
|
|
} else
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device doesn't support the Brightness control." << endl;
|
|
}
|
|
} else
|
|
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device doesn't support the Brightness control." << endl;
|
|
} else
|
|
{
|
|
CLEAR (control);
|
|
control.id = V4L2_CID_BRIGHTNESS;
|
|
control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getBrightness());
|
|
|
|
if (-1 == xioctl (VIDIOC_S_CTRL, &control))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
{
|
|
struct video_picture V4L_picture;
|
|
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
|
|
V4L_picture.brightness = uint(65535*getBrightness());
|
|
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image brightness. Fallback to it is not yet implemented." << endl;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
return getBrightness();
|
|
}
|
|
|
|
float VideoDevice::getContrast()
|
|
{
|
|
if (m_current_input < m_input.size() )
|
|
return m_input[m_current_input].getContrast();
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
float VideoDevice::setContrast(float contrast)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << " called." << endl;
|
|
m_input[m_current_input].setContrast(contrast); // Just to check bounds
|
|
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
{
|
|
struct v4l2_queryctrl queryctrl;
|
|
struct v4l2_control control;
|
|
|
|
CLEAR (queryctrl);
|
|
queryctrl.id = V4L2_CID_CONTRAST;
|
|
|
|
if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
|
|
{
|
|
if (errno != EINVAL)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl;
|
|
} else
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device doesn't support the Contrast control." << endl;
|
|
}
|
|
} else
|
|
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device doesn't support the Contrast control." << endl;
|
|
} else
|
|
{
|
|
CLEAR (control);
|
|
control.id = V4L2_CID_CONTRAST;
|
|
control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getContrast());
|
|
|
|
if (-1 == xioctl (VIDIOC_S_CTRL, &control))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
{
|
|
struct video_picture V4L_picture;
|
|
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
|
|
V4L_picture.contrast = uint(65535*getContrast());
|
|
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image contrast. Fallback to it is not yet implemented." << endl;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
return getContrast();
|
|
}
|
|
|
|
float VideoDevice::getSaturation()
|
|
{
|
|
if (m_current_input < m_input.size() )
|
|
return m_input[m_current_input].getSaturation();
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
float VideoDevice::setSaturation(float saturation)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << " called." << endl;
|
|
m_input[m_current_input].setSaturation(saturation); // Just to check bounds
|
|
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
{
|
|
struct v4l2_queryctrl queryctrl;
|
|
struct v4l2_control control;
|
|
|
|
CLEAR (queryctrl);
|
|
queryctrl.id = V4L2_CID_SATURATION;
|
|
|
|
if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
|
|
{
|
|
if (errno != EINVAL)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl;
|
|
} else
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device doesn't support the Saturation control." << endl;
|
|
}
|
|
} else
|
|
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device doesn't support the Saturation control." << endl;
|
|
} else
|
|
{
|
|
CLEAR (control);
|
|
control.id = V4L2_CID_SATURATION;
|
|
control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getSaturation());
|
|
|
|
if (-1 == xioctl (VIDIOC_S_CTRL, &control))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
{
|
|
struct video_picture V4L_picture;
|
|
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
|
|
V4L_picture.colour = uint(65535*getSaturation());
|
|
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image saturation. Fallback to it is not yet implemented." << endl;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
return getSaturation();
|
|
}
|
|
|
|
float VideoDevice::getWhiteness()
|
|
{
|
|
if (m_current_input < m_input.size() )
|
|
return m_input[m_current_input].getWhiteness();
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
float VideoDevice::setWhiteness(float whiteness)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << " called." << endl;
|
|
m_input[m_current_input].setWhiteness(whiteness); // Just to check bounds
|
|
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
{
|
|
struct v4l2_queryctrl queryctrl;
|
|
struct v4l2_control control;
|
|
|
|
CLEAR (queryctrl);
|
|
queryctrl.id = V4L2_CID_WHITENESS;
|
|
|
|
if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
|
|
{
|
|
if (errno != EINVAL)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl;
|
|
} else
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device doesn't support the Whiteness control." << endl;
|
|
}
|
|
} else
|
|
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device doesn't support the Whiteness control." << endl;
|
|
} else
|
|
{
|
|
CLEAR (control);
|
|
control.id = V4L2_CID_WHITENESS;
|
|
control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getWhiteness());
|
|
|
|
if (-1 == xioctl (VIDIOC_S_CTRL, &control))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
{
|
|
struct video_picture V4L_picture;
|
|
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
|
|
V4L_picture.whiteness = uint(65535*getWhiteness());
|
|
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting white level. Fallback to it is not yet implemented." << endl;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
return getWhiteness();
|
|
}
|
|
|
|
float VideoDevice::getHue()
|
|
{
|
|
if (m_current_input < m_input.size() )
|
|
return m_input[m_current_input].getHue();
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
float VideoDevice::setHue(float hue)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << " called." << endl;
|
|
m_input[m_current_input].setHue(hue); // Just to check bounds
|
|
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
{
|
|
struct v4l2_queryctrl queryctrl;
|
|
struct v4l2_control control;
|
|
|
|
CLEAR (queryctrl);
|
|
queryctrl.id = V4L2_CID_HUE;
|
|
|
|
if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
|
|
{
|
|
if (errno != EINVAL)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl;
|
|
} else
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device doesn't support the Hue control." << endl;
|
|
}
|
|
} else
|
|
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Device doesn't support the Hue control." << endl;
|
|
} else
|
|
{
|
|
CLEAR (control);
|
|
control.id = V4L2_CID_HUE;
|
|
control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getHue());
|
|
|
|
if (-1 == xioctl (VIDIOC_S_CTRL, &control))
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
{
|
|
struct video_picture V4L_picture;
|
|
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
|
|
V4L_picture.hue = uint(65535*getHue());
|
|
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
|
|
kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image hue. Fallback to it is not yet implemented." << endl;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
return getHue();
|
|
}
|
|
|
|
|
|
bool VideoDevice::getAutoBrightnessContrast()
|
|
{
|
|
if (m_current_input < m_input.size() )
|
|
return m_input[m_current_input].getAutoBrightnessContrast();
|
|
else
|
|
return false;
|
|
}
|
|
|
|
bool VideoDevice::setAutoBrightnessContrast(bool brightnesscontrast)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VideoDevice::setAutoBrightnessContrast(" << brightnesscontrast << ") called." << endl;
|
|
if (m_current_input < m_input.size() )
|
|
{
|
|
m_input[m_current_input].setAutoBrightnessContrast(brightnesscontrast);
|
|
return m_input[m_current_input].getAutoBrightnessContrast();
|
|
}
|
|
else
|
|
return false;
|
|
|
|
}
|
|
|
|
bool VideoDevice::getAutoColorCorrection()
|
|
{
|
|
if (m_current_input < m_input.size() )
|
|
return m_input[m_current_input].getAutoColorCorrection();
|
|
else
|
|
return false;
|
|
}
|
|
|
|
bool VideoDevice::setAutoColorCorrection(bool colorcorrection)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VideoDevice::setAutoColorCorrection(" << colorcorrection << ") called." << endl;
|
|
if (m_current_input < m_input.size() )
|
|
{
|
|
m_input[m_current_input].setAutoColorCorrection(colorcorrection);
|
|
return m_input[m_current_input].getAutoColorCorrection();
|
|
}
|
|
else
|
|
return false;
|
|
}
|
|
|
|
bool VideoDevice::getImageAsMirror()
|
|
{
|
|
if (m_current_input < m_input.size() )
|
|
return m_input[m_current_input].getImageAsMirror();
|
|
else
|
|
return false;
|
|
}
|
|
|
|
bool VideoDevice::setImageAsMirror(bool imageasmirror)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "VideoDevice::setImageAsMirror(" << imageasmirror << ") called." << endl;
|
|
if (m_current_input < m_input.size() )
|
|
{
|
|
m_input[m_current_input].setImageAsMirror(imageasmirror);
|
|
return m_input[m_current_input].getImageAsMirror();
|
|
}
|
|
else
|
|
return false;
|
|
}
|
|
|
|
pixel_format VideoDevice::pixelFormatForPalette( int palette )
|
|
{
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
switch(palette)
|
|
{
|
|
case 0 : return PIXELFORMAT_NONE; break;
|
|
|
|
// Packed RGB formats
|
|
case V4L2_PIX_FMT_RGB332 : return PIXELFORMAT_RGB332; break;
|
|
#if defined( V4L2_PIX_FMT_RGB444 )
|
|
case V4L2_PIX_FMT_RGB444 : return PIXELFORMAT_RGB444; break;
|
|
#endif
|
|
case V4L2_PIX_FMT_RGB555 : return PIXELFORMAT_RGB555; break;
|
|
case V4L2_PIX_FMT_RGB565 : return PIXELFORMAT_RGB565; break;
|
|
case V4L2_PIX_FMT_RGB555X : return PIXELFORMAT_RGB555X; break;
|
|
case V4L2_PIX_FMT_RGB565X : return PIXELFORMAT_RGB565X; break;
|
|
case V4L2_PIX_FMT_BGR24 : return PIXELFORMAT_BGR24; break;
|
|
case V4L2_PIX_FMT_RGB24 : return PIXELFORMAT_RGB24; break;
|
|
case V4L2_PIX_FMT_BGR32 : return PIXELFORMAT_BGR32; break;
|
|
case V4L2_PIX_FMT_RGB32 : return PIXELFORMAT_RGB32; break;
|
|
|
|
// Bayer RGB format
|
|
case V4L2_PIX_FMT_SBGGR8 : return PIXELFORMAT_SBGGR8; break;
|
|
|
|
// YUV formats
|
|
case V4L2_PIX_FMT_GREY : return PIXELFORMAT_GREY; break;
|
|
case V4L2_PIX_FMT_YUYV : return PIXELFORMAT_YUYV; break;
|
|
case V4L2_PIX_FMT_UYVY : return PIXELFORMAT_UYVY; break;
|
|
case V4L2_PIX_FMT_YUV420 : return PIXELFORMAT_YUV420P; break;
|
|
case V4L2_PIX_FMT_YUV422P : return PIXELFORMAT_YUV422P; break;
|
|
|
|
// Compressed formats
|
|
case V4L2_PIX_FMT_JPEG : return PIXELFORMAT_JPEG; break;
|
|
case V4L2_PIX_FMT_MPEG : return PIXELFORMAT_MPEG; break;
|
|
|
|
// Reserved formats
|
|
case V4L2_PIX_FMT_DV : return PIXELFORMAT_DV; break;
|
|
case V4L2_PIX_FMT_ET61X251 : return PIXELFORMAT_ET61X251; break;
|
|
case V4L2_PIX_FMT_HI240 : return PIXELFORMAT_HI240; break;
|
|
#if defined( V4L2_PIX_FMT_HM12 )
|
|
case V4L2_PIX_FMT_HM12 : return PIXELFORMAT_HM12; break;
|
|
#endif
|
|
case V4L2_PIX_FMT_MJPEG : return PIXELFORMAT_MJPEG; break;
|
|
case V4L2_PIX_FMT_PWC1 : return PIXELFORMAT_PWC1; break;
|
|
case V4L2_PIX_FMT_PWC2 : return PIXELFORMAT_PWC2; break;
|
|
case V4L2_PIX_FMT_SN9C10X : return PIXELFORMAT_SN9C10X; break;
|
|
case V4L2_PIX_FMT_WNVA : return PIXELFORMAT_WNVA; break;
|
|
case V4L2_PIX_FMT_YYUV : return PIXELFORMAT_YYUV; break;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
switch(palette)
|
|
{
|
|
case 0 : return PIXELFORMAT_NONE; break;
|
|
case VIDEO_PALETTE_GREY : return PIXELFORMAT_GREY; break;
|
|
case VIDEO_PALETTE_HI240 : return PIXELFORMAT_RGB332; break;
|
|
case VIDEO_PALETTE_RGB555 : return PIXELFORMAT_RGB555; break;
|
|
case VIDEO_PALETTE_RGB565 : return PIXELFORMAT_RGB565; break;
|
|
case VIDEO_PALETTE_RGB24 : return PIXELFORMAT_RGB24; break;
|
|
case VIDEO_PALETTE_RGB32 : return PIXELFORMAT_RGB32; break;
|
|
case VIDEO_PALETTE_YUYV : return PIXELFORMAT_YUYV; break;
|
|
case VIDEO_PALETTE_UYVY : return PIXELFORMAT_UYVY; break;
|
|
case VIDEO_PALETTE_YUV420 :
|
|
case VIDEO_PALETTE_YUV420P : return PIXELFORMAT_YUV420P; break;
|
|
case VIDEO_PALETTE_YUV422P : return PIXELFORMAT_YUV422P; break;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
return PIXELFORMAT_NONE; break;
|
|
}
|
|
return PIXELFORMAT_NONE;
|
|
}
|
|
|
|
int VideoDevice::pixelFormatCode(pixel_format pixelformat)
|
|
{
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
switch(pixelformat)
|
|
{
|
|
case PIXELFORMAT_NONE : return 0; break;
|
|
|
|
// Packed RGB formats
|
|
case PIXELFORMAT_RGB332 : return V4L2_PIX_FMT_RGB332; break;
|
|
#if defined( V4L2_PIX_FMT_RGB444 )
|
|
case PIXELFORMAT_RGB444 : return V4L2_PIX_FMT_RGB444; break;
|
|
#endif
|
|
case PIXELFORMAT_RGB555 : return V4L2_PIX_FMT_RGB555; break;
|
|
case PIXELFORMAT_RGB565 : return V4L2_PIX_FMT_RGB565; break;
|
|
case PIXELFORMAT_RGB555X: return V4L2_PIX_FMT_RGB555X; break;
|
|
case PIXELFORMAT_RGB565X: return V4L2_PIX_FMT_RGB565X; break;
|
|
case PIXELFORMAT_BGR24 : return V4L2_PIX_FMT_BGR24; break;
|
|
case PIXELFORMAT_RGB24 : return V4L2_PIX_FMT_RGB24; break;
|
|
case PIXELFORMAT_BGR32 : return V4L2_PIX_FMT_BGR32; break;
|
|
case PIXELFORMAT_RGB32 : return V4L2_PIX_FMT_RGB32; break;
|
|
|
|
// Bayer RGB format
|
|
case PIXELFORMAT_SBGGR8 : return V4L2_PIX_FMT_SBGGR8; break;
|
|
|
|
// YUV formats
|
|
case PIXELFORMAT_GREY : return V4L2_PIX_FMT_GREY; break;
|
|
case PIXELFORMAT_YUYV : return V4L2_PIX_FMT_YUYV; break;
|
|
case PIXELFORMAT_UYVY : return V4L2_PIX_FMT_UYVY; break;
|
|
case PIXELFORMAT_YUV420P: return V4L2_PIX_FMT_YUV420; break;
|
|
case PIXELFORMAT_YUV422P: return V4L2_PIX_FMT_YUV422P; break;
|
|
|
|
// Compressed formats
|
|
case PIXELFORMAT_JPEG : return V4L2_PIX_FMT_JPEG; break;
|
|
case PIXELFORMAT_MPEG : return V4L2_PIX_FMT_MPEG; break;
|
|
|
|
// Reserved formats
|
|
case PIXELFORMAT_DV : return V4L2_PIX_FMT_DV; break;
|
|
case PIXELFORMAT_ET61X251:return V4L2_PIX_FMT_ET61X251;break;
|
|
case PIXELFORMAT_HI240 : return V4L2_PIX_FMT_HI240; break;
|
|
#if defined( V4L2_PIX_FMT_HM12 )
|
|
case PIXELFORMAT_HM12 : return V4L2_PIX_FMT_HM12; break;
|
|
#endif
|
|
case PIXELFORMAT_MJPEG : return V4L2_PIX_FMT_MJPEG; break;
|
|
case PIXELFORMAT_PWC1 : return V4L2_PIX_FMT_PWC1; break;
|
|
case PIXELFORMAT_PWC2 : return V4L2_PIX_FMT_PWC2; break;
|
|
case PIXELFORMAT_SN9C10X: return V4L2_PIX_FMT_SN9C10X; break;
|
|
case PIXELFORMAT_WNVA : return V4L2_PIX_FMT_WNVA; break;
|
|
case PIXELFORMAT_YYUV : return V4L2_PIX_FMT_YYUV; break;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
switch(pixelformat)
|
|
{
|
|
case PIXELFORMAT_NONE : return 0; break;
|
|
|
|
// Packed RGB formats
|
|
case PIXELFORMAT_RGB332 : return VIDEO_PALETTE_HI240; break;
|
|
case PIXELFORMAT_RGB444 : return 0; break;
|
|
case PIXELFORMAT_RGB555 : return VIDEO_PALETTE_RGB555; break;
|
|
case PIXELFORMAT_RGB565 : return VIDEO_PALETTE_RGB565; break;
|
|
case PIXELFORMAT_RGB555X: return 0; break;
|
|
case PIXELFORMAT_RGB565X: return 0; break;
|
|
case PIXELFORMAT_BGR24 : return 0; break;
|
|
case PIXELFORMAT_RGB24 : return VIDEO_PALETTE_RGB24; break;
|
|
case PIXELFORMAT_BGR32 : return 0; break;
|
|
case PIXELFORMAT_RGB32 : return VIDEO_PALETTE_RGB32; break;
|
|
|
|
// Bayer RGB format
|
|
case PIXELFORMAT_SBGGR8 : return 0; break;
|
|
|
|
// YUV formats
|
|
case PIXELFORMAT_GREY : return VIDEO_PALETTE_GREY; break;
|
|
case PIXELFORMAT_YUYV : return VIDEO_PALETTE_YUYV; break;
|
|
case PIXELFORMAT_UYVY : return VIDEO_PALETTE_UYVY; break;
|
|
case PIXELFORMAT_YUV420P: return VIDEO_PALETTE_YUV420; break;
|
|
case PIXELFORMAT_YUV422P: return VIDEO_PALETTE_YUV422P; break;
|
|
|
|
// Compressed formats
|
|
case PIXELFORMAT_JPEG : return 0; break;
|
|
case PIXELFORMAT_MPEG : return 0; break;
|
|
|
|
// Reserved formats
|
|
case PIXELFORMAT_DV : return 0; break;
|
|
case PIXELFORMAT_ET61X251:return 0; break;
|
|
case PIXELFORMAT_HI240 : return VIDEO_PALETTE_HI240; break;
|
|
case PIXELFORMAT_HM12 : return 0; break;
|
|
case PIXELFORMAT_MJPEG : return 0; break;
|
|
case PIXELFORMAT_PWC1 : return 0; break;
|
|
case PIXELFORMAT_PWC2 : return 0; break;
|
|
case PIXELFORMAT_SN9C10X: return 0; break;
|
|
case PIXELFORMAT_WNVA : return 0; break;
|
|
case PIXELFORMAT_YYUV : return 0; break;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
return PIXELFORMAT_NONE; break;
|
|
}
|
|
return PIXELFORMAT_NONE;
|
|
}
|
|
|
|
int VideoDevice::pixelFormatDepth(pixel_format pixelformat)
|
|
{
|
|
switch(pixelformat)
|
|
{
|
|
case PIXELFORMAT_NONE : return 0; break;
|
|
|
|
// Packed RGB formats
|
|
case PIXELFORMAT_RGB332 : return 8; break;
|
|
case PIXELFORMAT_RGB444 : return 16; break;
|
|
case PIXELFORMAT_RGB555 : return 16; break;
|
|
case PIXELFORMAT_RGB565 : return 16; break;
|
|
case PIXELFORMAT_RGB555X: return 16; break;
|
|
case PIXELFORMAT_RGB565X: return 16; break;
|
|
case PIXELFORMAT_BGR24 : return 24; break;
|
|
case PIXELFORMAT_RGB24 : return 24; break;
|
|
case PIXELFORMAT_BGR32 : return 32; break;
|
|
case PIXELFORMAT_RGB32 : return 32; break;
|
|
|
|
// Bayer RGB format
|
|
case PIXELFORMAT_SBGGR8 : return 0; break;
|
|
|
|
// YUV formats
|
|
case PIXELFORMAT_GREY : return 8; break;
|
|
case PIXELFORMAT_YUYV : return 16; break;
|
|
case PIXELFORMAT_UYVY : return 16; break;
|
|
case PIXELFORMAT_YUV420P: return 16; break;
|
|
case PIXELFORMAT_YUV422P: return 16; break;
|
|
|
|
// Compressed formats
|
|
case PIXELFORMAT_JPEG : return 0; break;
|
|
case PIXELFORMAT_MPEG : return 0; break;
|
|
|
|
// Reserved formats
|
|
case PIXELFORMAT_DV : return 0; break;
|
|
case PIXELFORMAT_ET61X251:return 0; break;
|
|
case PIXELFORMAT_HI240 : return 8; break;
|
|
case PIXELFORMAT_HM12 : return 0; break;
|
|
case PIXELFORMAT_MJPEG : return 0; break;
|
|
case PIXELFORMAT_PWC1 : return 0; break;
|
|
case PIXELFORMAT_PWC2 : return 0; break;
|
|
case PIXELFORMAT_SN9C10X: return 0; break;
|
|
case PIXELFORMAT_WNVA : return 0; break;
|
|
case PIXELFORMAT_YYUV : return 0; break;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
QString VideoDevice::pixelFormatName(pixel_format pixelformat)
|
|
{
|
|
QString returnvalue;
|
|
returnvalue = "None";
|
|
switch(pixelformat)
|
|
{
|
|
case PIXELFORMAT_NONE : returnvalue = "None"; break;
|
|
|
|
// Packed RGB formats
|
|
case PIXELFORMAT_RGB332 : returnvalue = "8-bit RGB332"; break;
|
|
case PIXELFORMAT_RGB444 : returnvalue = "8-bit RGB444"; break;
|
|
case PIXELFORMAT_RGB555 : returnvalue = "16-bit RGB555"; break;
|
|
case PIXELFORMAT_RGB565 : returnvalue = "16-bit RGB565"; break;
|
|
case PIXELFORMAT_RGB555X: returnvalue = "16-bit RGB555X"; break;
|
|
case PIXELFORMAT_RGB565X: returnvalue = "16-bit RGB565X"; break;
|
|
case PIXELFORMAT_BGR24 : returnvalue = "24-bit BGR24"; break;
|
|
case PIXELFORMAT_RGB24 : returnvalue = "24-bit RGB24"; break;
|
|
case PIXELFORMAT_BGR32 : returnvalue = "32-bit BGR32"; break;
|
|
case PIXELFORMAT_RGB32 : returnvalue = "32-bit RGB32"; break;
|
|
|
|
// Bayer RGB format
|
|
case PIXELFORMAT_SBGGR8 : returnvalue = "Bayer RGB format"; break;
|
|
|
|
// YUV formats
|
|
case PIXELFORMAT_GREY : returnvalue = "8-bit Grayscale"; break;
|
|
case PIXELFORMAT_YUYV : returnvalue = "Packed YUV 4:2:2"; break;
|
|
case PIXELFORMAT_UYVY : returnvalue = "Packed YVU 4:2:2"; break;
|
|
case PIXELFORMAT_YUV420P: returnvalue = "Planar YUV 4:2:0"; break;
|
|
case PIXELFORMAT_YUV422P: returnvalue = "Planar YUV 4:2:2"; break;
|
|
|
|
|
|
// Compressed formats
|
|
case PIXELFORMAT_JPEG : returnvalue = "JPEG image"; break;
|
|
case PIXELFORMAT_MPEG : returnvalue = "MPEG stream"; break;
|
|
|
|
// Reserved formats
|
|
case PIXELFORMAT_DV : returnvalue = "DV (unknown)"; break;
|
|
case PIXELFORMAT_ET61X251:returnvalue = "ET61X251"; break;
|
|
case PIXELFORMAT_HI240 : returnvalue = "8-bit HI240 (RGB332)"; break;
|
|
case PIXELFORMAT_HM12 : returnvalue = "Packed YUV 4:2:2"; break;
|
|
case PIXELFORMAT_MJPEG : returnvalue = "8-bit Grayscale"; break;
|
|
case PIXELFORMAT_PWC1 : returnvalue = "PWC1"; break;
|
|
case PIXELFORMAT_PWC2 : returnvalue = "PWC2"; break;
|
|
case PIXELFORMAT_SN9C10X: returnvalue = "SN9C102"; break;
|
|
case PIXELFORMAT_WNVA : returnvalue = "Winnov Videum"; break;
|
|
case PIXELFORMAT_YYUV : returnvalue = "YYUV (unknown)"; break;
|
|
}
|
|
return returnvalue;
|
|
}
|
|
|
|
QString VideoDevice::pixelFormatName(int pixelformat)
|
|
{
|
|
QString returnvalue;
|
|
returnvalue = "None";
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
switch(pixelformat)
|
|
{
|
|
case 0 : returnvalue = pixelFormatName(PIXELFORMAT_NONE); break;
|
|
|
|
// Packed RGB formats
|
|
case V4L2_PIX_FMT_RGB332 : returnvalue = pixelFormatName(PIXELFORMAT_RGB332); break;
|
|
#if defined( V4L2_PIX_FMT_RGB444 )
|
|
case V4L2_PIX_FMT_RGB444 : returnvalue = pixelFormatName(PIXELFORMAT_RGB444); break;
|
|
#endif
|
|
case V4L2_PIX_FMT_RGB555 : returnvalue = pixelFormatName(PIXELFORMAT_RGB555); break;
|
|
case V4L2_PIX_FMT_RGB565 : returnvalue = pixelFormatName(PIXELFORMAT_RGB565); break;
|
|
case V4L2_PIX_FMT_RGB555X : returnvalue = pixelFormatName(PIXELFORMAT_RGB555X); break;
|
|
case V4L2_PIX_FMT_RGB565X : returnvalue = pixelFormatName(PIXELFORMAT_RGB565X); break;
|
|
case V4L2_PIX_FMT_BGR24 : returnvalue = pixelFormatName(PIXELFORMAT_BGR24); break;
|
|
case V4L2_PIX_FMT_RGB24 : returnvalue = pixelFormatName(PIXELFORMAT_RGB24); break;
|
|
case V4L2_PIX_FMT_BGR32 : returnvalue = pixelFormatName(PIXELFORMAT_BGR32); break;
|
|
case V4L2_PIX_FMT_RGB32 : returnvalue = pixelFormatName(PIXELFORMAT_RGB32); break;
|
|
|
|
// Bayer RGB format
|
|
case V4L2_PIX_FMT_SBGGR8 : returnvalue = pixelFormatName(PIXELFORMAT_SBGGR8); break;
|
|
|
|
// YUV formats
|
|
case V4L2_PIX_FMT_GREY : returnvalue = pixelFormatName(PIXELFORMAT_GREY); break;
|
|
case V4L2_PIX_FMT_YUYV : returnvalue = pixelFormatName(PIXELFORMAT_YUYV); break;
|
|
case V4L2_PIX_FMT_UYVY : returnvalue = pixelFormatName(PIXELFORMAT_UYVY); break;
|
|
case V4L2_PIX_FMT_YUV420 : returnvalue = pixelFormatName(PIXELFORMAT_YUV420P); break;
|
|
case V4L2_PIX_FMT_YUV422P : returnvalue = pixelFormatName(PIXELFORMAT_YUV422P); break;
|
|
|
|
// Compressed formats
|
|
case V4L2_PIX_FMT_JPEG : returnvalue = pixelFormatName(PIXELFORMAT_JPEG); break;
|
|
case V4L2_PIX_FMT_MPEG : returnvalue = pixelFormatName(PIXELFORMAT_MPEG); break;
|
|
|
|
// Reserved formats
|
|
case V4L2_PIX_FMT_DV : returnvalue = pixelFormatName(PIXELFORMAT_DV); break;
|
|
case V4L2_PIX_FMT_ET61X251 : returnvalue = pixelFormatName(PIXELFORMAT_ET61X251); break;
|
|
case V4L2_PIX_FMT_HI240 : returnvalue = pixelFormatName(PIXELFORMAT_HI240); break;
|
|
#if defined( V4L2_PIX_FMT_HM12 )
|
|
case V4L2_PIX_FMT_HM12 : returnvalue = pixelFormatName(PIXELFORMAT_HM12); break;
|
|
#endif
|
|
case V4L2_PIX_FMT_MJPEG : returnvalue = pixelFormatName(PIXELFORMAT_MJPEG); break;
|
|
case V4L2_PIX_FMT_PWC1 : returnvalue = pixelFormatName(PIXELFORMAT_PWC1); break;
|
|
case V4L2_PIX_FMT_PWC2 : returnvalue = pixelFormatName(PIXELFORMAT_PWC2); break;
|
|
case V4L2_PIX_FMT_SN9C10X : returnvalue = pixelFormatName(PIXELFORMAT_SN9C10X); break;
|
|
case V4L2_PIX_FMT_WNVA : returnvalue = pixelFormatName(PIXELFORMAT_WNVA); break;
|
|
case V4L2_PIX_FMT_YYUV : returnvalue = pixelFormatName(PIXELFORMAT_YYUV); break;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
switch(pixelformat)
|
|
{
|
|
case VIDEO_PALETTE_GREY : returnvalue = pixelFormatName(PIXELFORMAT_GREY); break;
|
|
case VIDEO_PALETTE_HI240 : returnvalue = pixelFormatName(PIXELFORMAT_RGB332); break;
|
|
case VIDEO_PALETTE_RGB555 : returnvalue = pixelFormatName(PIXELFORMAT_RGB555); break;
|
|
case VIDEO_PALETTE_RGB565 : returnvalue = pixelFormatName(PIXELFORMAT_RGB565); break;
|
|
case VIDEO_PALETTE_RGB24 : returnvalue = pixelFormatName(PIXELFORMAT_RGB24); break;
|
|
case VIDEO_PALETTE_RGB32 : returnvalue = pixelFormatName(PIXELFORMAT_RGB32); break;
|
|
case VIDEO_PALETTE_YUYV : returnvalue = pixelFormatName(PIXELFORMAT_YUYV); break;
|
|
case VIDEO_PALETTE_UYVY : returnvalue = pixelFormatName(PIXELFORMAT_UYVY); break;
|
|
case VIDEO_PALETTE_YUV420 :
|
|
case VIDEO_PALETTE_YUV420P : returnvalue = pixelFormatName(PIXELFORMAT_YUV420P); break;
|
|
case VIDEO_PALETTE_YUV422P : returnvalue = pixelFormatName(PIXELFORMAT_YUV422P); break;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
return returnvalue;
|
|
}
|
|
|
|
int VideoDevice::detectPixelFormats()
|
|
{
|
|
int err = 0;
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
fmtdesc.index = 0;
|
|
fmtdesc.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
|
|
while ( err == 0 )
|
|
{
|
|
if (-1 == xioctl (VIDIOC_ENUM_FMT, &fmtdesc))
|
|
// if (ioctl(fd, VIDIOC_ENUM_FMT, &fmtdesc) < 0 )
|
|
{
|
|
perror("VIDIOC_ENUM_FMT");
|
|
err = errno;
|
|
}
|
|
else
|
|
{
|
|
kdDebug(14010) << k_funcinfo << fmtdesc.pixelformat << " " << pixelFormatName(fmtdesc.pixelformat) << endl; // Need a cleanup. PixelFormatForPalette is a really bad name
|
|
fmtdesc.index++;
|
|
}
|
|
}
|
|
// break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
// TODO: THis thing can be used to detec what pixel formats are supported in a API-independent way, but V4L2 has VIDIOC_ENUM_PIXFMT.
|
|
// The correct thing to do is to isolate these calls and do a proper implementation for V4L and another for V4L2 when this thing will be migrated to a plugin architecture.
|
|
|
|
// Packed RGB formats
|
|
kdDebug(14010) << k_funcinfo << "Supported pixel formats:" << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB332)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB332) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB444)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB444) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB555)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB555) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB565)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB565) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB555X)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB555X) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB565X)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB565X) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_BGR24)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_BGR24) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB24)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB24) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_BGR32)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_BGR32) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB32)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB32) << endl;
|
|
|
|
// Bayer RGB format
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_SBGGR8)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_SBGGR8) << endl;
|
|
|
|
// YUV formats
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_GREY)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_GREY) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YUYV)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YUYV) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_UYVY)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_UYVY) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YUV420P)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YUV420P) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YUV422P)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YUV422P) << endl;
|
|
|
|
// Compressed formats
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_JPEG)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_JPEG) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_MPEG)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_MPEG) << endl;
|
|
|
|
// Reserved formats
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_DV)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_DV) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_ET61X251)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_ET61X251) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_HI240)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_HI240) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_HM12)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_HM12) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_MJPEG)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_MJPEG) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_PWC1)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_PWC1) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_PWC2)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_PWC2) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_SN9C10X)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_SN9C10X) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_WNVA)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_WNVA) << endl;
|
|
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YYUV)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YYUV) << endl;
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
return PIXELFORMAT_NONE; break;
|
|
}
|
|
return PIXELFORMAT_NONE;
|
|
}
|
|
|
|
__u64 VideoDevice::signalStandardCode(signal_standard standard)
|
|
{
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
switch(standard)
|
|
{
|
|
case STANDARD_NONE : return V4L2_STD_UNKNOWN; break;
|
|
case STANDARD_PAL_B : return V4L2_STD_PAL_B; break;
|
|
case STANDARD_PAL_B1 : return V4L2_STD_PAL_B1; break;
|
|
case STANDARD_PAL_G : return V4L2_STD_PAL_G; break;
|
|
case STANDARD_PAL_H : return V4L2_STD_PAL_H; break;
|
|
case STANDARD_PAL_I : return V4L2_STD_PAL_I; break;
|
|
case STANDARD_PAL_D : return V4L2_STD_PAL_D; break;
|
|
case STANDARD_PAL_D1 : return V4L2_STD_PAL_D1; break;
|
|
case STANDARD_PAL_K : return V4L2_STD_PAL_K; break;
|
|
case STANDARD_PAL_M : return V4L2_STD_PAL_M; break;
|
|
case STANDARD_PAL_N : return V4L2_STD_PAL_N; break;
|
|
case STANDARD_PAL_Nc : return V4L2_STD_PAL_Nc; break;
|
|
case STANDARD_PAL_60 : return V4L2_STD_PAL_60; break;
|
|
case STANDARD_NTSC_M : return V4L2_STD_NTSC_M; break;
|
|
case STANDARD_NTSC_M_JP : return V4L2_STD_NTSC_M_JP; break;
|
|
case STANDARD_NTSC_443 : return V4L2_STD_NTSC; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
|
|
case STANDARD_SECAM_B : return V4L2_STD_SECAM_B; break;
|
|
case STANDARD_SECAM_D : return V4L2_STD_SECAM_D; break;
|
|
case STANDARD_SECAM_G : return V4L2_STD_SECAM_G; break;
|
|
case STANDARD_SECAM_H : return V4L2_STD_SECAM_H; break;
|
|
case STANDARD_SECAM_K : return V4L2_STD_SECAM_K; break;
|
|
case STANDARD_SECAM_K1 : return V4L2_STD_SECAM_K1; break;
|
|
case STANDARD_SECAM_L : return V4L2_STD_SECAM_L; break;
|
|
case STANDARD_SECAM_LC : return V4L2_STD_SECAM; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
|
|
case STANDARD_ATSC_8_VSB : return V4L2_STD_ATSC_8_VSB; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
|
|
case STANDARD_ATSC_16_VSB : return V4L2_STD_ATSC_16_VSB; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
|
|
case STANDARD_PAL_BG : return V4L2_STD_PAL_BG; break;
|
|
case STANDARD_PAL_DK : return V4L2_STD_PAL_DK; break;
|
|
case STANDARD_PAL : return V4L2_STD_PAL; break;
|
|
case STANDARD_NTSC : return V4L2_STD_NTSC; break;
|
|
case STANDARD_SECAM_DK : return V4L2_STD_SECAM_DK; break;
|
|
case STANDARD_SECAM : return V4L2_STD_SECAM; break;
|
|
case STANDARD_525_60 : return V4L2_STD_525_60; break;
|
|
case STANDARD_625_50 : return V4L2_STD_625_50; break;
|
|
case STANDARD_ALL : return V4L2_STD_ALL; break;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
switch(standard)
|
|
{
|
|
case STANDARD_NONE : return VIDEO_MODE_AUTO; break;
|
|
case STANDARD_PAL_B : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_PAL_B1 : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_PAL_G : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_PAL_H : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_PAL_I : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_PAL_D : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_PAL_D1 : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_PAL_K : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_PAL_M : return 5; break; // Undocumented value found to be compatible with V4L bttv driver
|
|
case STANDARD_PAL_N : return 6; break; // Undocumented value found to be compatible with V4L bttv driver
|
|
case STANDARD_PAL_Nc : return 4; break; // Undocumented value found to be compatible with V4L bttv driver
|
|
case STANDARD_PAL_60 : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_NTSC_M : return VIDEO_MODE_NTSC; break;
|
|
case STANDARD_NTSC_M_JP : return 7; break; // Undocumented value found to be compatible with V4L bttv driver
|
|
case STANDARD_NTSC_443 : return VIDEO_MODE_NTSC; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
|
|
case STANDARD_SECAM_B : return VIDEO_MODE_SECAM; break;
|
|
case STANDARD_SECAM_D : return VIDEO_MODE_SECAM; break;
|
|
case STANDARD_SECAM_G : return VIDEO_MODE_SECAM; break;
|
|
case STANDARD_SECAM_H : return VIDEO_MODE_SECAM; break;
|
|
case STANDARD_SECAM_K : return VIDEO_MODE_SECAM; break;
|
|
case STANDARD_SECAM_K1 : return VIDEO_MODE_SECAM; break;
|
|
case STANDARD_SECAM_L : return VIDEO_MODE_SECAM; break;
|
|
case STANDARD_SECAM_LC : return VIDEO_MODE_SECAM; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
|
|
case STANDARD_ATSC_8_VSB : return VIDEO_MODE_AUTO; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
|
|
case STANDARD_ATSC_16_VSB : return VIDEO_MODE_AUTO; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
|
|
case STANDARD_PAL_BG : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_PAL_DK : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_PAL : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_NTSC : return VIDEO_MODE_NTSC; break;
|
|
case STANDARD_SECAM_DK : return VIDEO_MODE_SECAM; break;
|
|
case STANDARD_SECAM : return VIDEO_MODE_SECAM; break;
|
|
case STANDARD_525_60 : return VIDEO_MODE_PAL; break;
|
|
case STANDARD_625_50 : return VIDEO_MODE_SECAM; break;
|
|
case STANDARD_ALL : return VIDEO_MODE_AUTO; break;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
return STANDARD_NONE; break;
|
|
}
|
|
return STANDARD_NONE;
|
|
}
|
|
|
|
QString VideoDevice::signalStandardName(signal_standard standard)
|
|
{
|
|
QString returnvalue;
|
|
returnvalue = "None";
|
|
switch(standard)
|
|
{
|
|
case STANDARD_NONE : returnvalue = "None"; break;
|
|
case STANDARD_PAL_B : returnvalue = "PAL-B"; break;
|
|
case STANDARD_PAL_B1 : returnvalue = "PAL-B1"; break;
|
|
case STANDARD_PAL_G : returnvalue = "PAL-G"; break;
|
|
case STANDARD_PAL_H : returnvalue = "PAL-H"; break;
|
|
case STANDARD_PAL_I : returnvalue = "PAL-I"; break;
|
|
case STANDARD_PAL_D : returnvalue = "PAL-D"; break;
|
|
case STANDARD_PAL_D1 : returnvalue = "PAL-D1"; break;
|
|
case STANDARD_PAL_K : returnvalue = "PAL-K"; break;
|
|
case STANDARD_PAL_M : returnvalue = "PAL-M"; break;
|
|
case STANDARD_PAL_N : returnvalue = "PAL-N"; break;
|
|
case STANDARD_PAL_Nc : returnvalue = "PAL-Nc"; break;
|
|
case STANDARD_PAL_60 : returnvalue = "PAL-60"; break;
|
|
case STANDARD_NTSC_M : returnvalue = "NTSC-M"; break;
|
|
case STANDARD_NTSC_M_JP : returnvalue = "NTSC-M(JP)"; break;
|
|
case STANDARD_NTSC_443 : returnvalue = "NTSC-443"; break;
|
|
case STANDARD_SECAM_B : returnvalue = "SECAM-B"; break;
|
|
case STANDARD_SECAM_D : returnvalue = "SECAM-D"; break;
|
|
case STANDARD_SECAM_G : returnvalue = "SECAM-G"; break;
|
|
case STANDARD_SECAM_H : returnvalue = "SECAM-H"; break;
|
|
case STANDARD_SECAM_K : returnvalue = "SECAM-K"; break;
|
|
case STANDARD_SECAM_K1 : returnvalue = "SECAM-K1"; break;
|
|
case STANDARD_SECAM_L : returnvalue = "SECAM-L"; break;
|
|
case STANDARD_SECAM_LC : returnvalue = "SECAM-LC"; break;
|
|
case STANDARD_ATSC_8_VSB : returnvalue = "ATSC-8-VSB"; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
|
|
case STANDARD_ATSC_16_VSB : returnvalue = "ATSC-16-VSB"; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
|
|
case STANDARD_PAL_BG : returnvalue = "PAL-BG"; break;
|
|
case STANDARD_PAL_DK : returnvalue = "PAL-DK"; break;
|
|
case STANDARD_PAL : returnvalue = "PAL"; break;
|
|
case STANDARD_NTSC : returnvalue = "NTSC"; break;
|
|
case STANDARD_SECAM_DK : returnvalue = "SECAM-DK"; break;
|
|
case STANDARD_SECAM : returnvalue = "SECAM"; break;
|
|
case STANDARD_525_60 : returnvalue = "525 lines 60Hz"; break;
|
|
case STANDARD_625_50 : returnvalue = "625 lines 50Hz"; break;
|
|
case STANDARD_ALL : returnvalue = "All"; break;
|
|
}
|
|
return returnvalue;
|
|
}
|
|
|
|
QString VideoDevice::signalStandardName(int standard)
|
|
{
|
|
QString returnvalue;
|
|
returnvalue = "None";
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
switch(standard)
|
|
{
|
|
case V4L2_STD_PAL_B : returnvalue = signalStandardName(STANDARD_PAL_B); break;
|
|
case V4L2_STD_PAL_B1 : returnvalue = signalStandardName(STANDARD_PAL_B1); break;
|
|
case V4L2_STD_PAL_G : returnvalue = signalStandardName(STANDARD_PAL_G); break;
|
|
case V4L2_STD_PAL_H : returnvalue = signalStandardName(STANDARD_PAL_H); break;
|
|
case V4L2_STD_PAL_I : returnvalue = signalStandardName(STANDARD_PAL_I); break;
|
|
case V4L2_STD_PAL_D : returnvalue = signalStandardName(STANDARD_PAL_D); break;
|
|
case V4L2_STD_PAL_D1 : returnvalue = signalStandardName(STANDARD_PAL_D1); break;
|
|
case V4L2_STD_PAL_K : returnvalue = signalStandardName(STANDARD_PAL_K); break;
|
|
case V4L2_STD_PAL_M : returnvalue = signalStandardName(STANDARD_PAL_M); break;
|
|
case V4L2_STD_PAL_N : returnvalue = signalStandardName(STANDARD_PAL_N); break;
|
|
case V4L2_STD_PAL_Nc : returnvalue = signalStandardName(STANDARD_PAL_Nc); break;
|
|
case V4L2_STD_PAL_60 : returnvalue = signalStandardName(STANDARD_PAL_60); break;
|
|
case V4L2_STD_NTSC_M : returnvalue = signalStandardName(STANDARD_NTSC_M); break;
|
|
case V4L2_STD_NTSC_M_JP : returnvalue = signalStandardName(STANDARD_NTSC_M_JP); break;
|
|
// case V4L2_STD_NTSC_443 : returnvalue = signalStandardName(STANDARD_NTSC_443); break; // Commented out because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
|
|
case V4L2_STD_SECAM_B : returnvalue = signalStandardName(STANDARD_SECAM_B); break;
|
|
case V4L2_STD_SECAM_D : returnvalue = signalStandardName(STANDARD_SECAM_D); break;
|
|
case V4L2_STD_SECAM_G : returnvalue = signalStandardName(STANDARD_SECAM_G); break;
|
|
case V4L2_STD_SECAM_H : returnvalue = signalStandardName(STANDARD_SECAM_H); break;
|
|
case V4L2_STD_SECAM_K : returnvalue = signalStandardName(STANDARD_SECAM_K); break;
|
|
case V4L2_STD_SECAM_K1 : returnvalue = signalStandardName(STANDARD_SECAM_K1); break;
|
|
case V4L2_STD_SECAM_L : returnvalue = signalStandardName(STANDARD_SECAM_L); break;
|
|
// case V4L2_STD_SECAM_LC : returnvalue = signalStandardName(STANDARD_SECAM_LC); break; // Commented out because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
|
|
case V4L2_STD_ATSC_8_VSB : returnvalue = signalStandardName(STANDARD_ATSC_8_VSB); break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
|
|
case V4L2_STD_ATSC_16_VSB : returnvalue = signalStandardName(STANDARD_ATSC_16_VSB); break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
|
|
case V4L2_STD_PAL_BG : returnvalue = signalStandardName(STANDARD_PAL_BG); break;
|
|
case V4L2_STD_PAL_DK : returnvalue = signalStandardName(STANDARD_PAL_DK); break;
|
|
case V4L2_STD_PAL : returnvalue = signalStandardName(STANDARD_PAL); break;
|
|
case V4L2_STD_NTSC : returnvalue = signalStandardName(STANDARD_NTSC); break;
|
|
case V4L2_STD_SECAM_DK : returnvalue = signalStandardName(STANDARD_SECAM_DK); break;
|
|
case V4L2_STD_SECAM : returnvalue = signalStandardName(STANDARD_SECAM); break;
|
|
case V4L2_STD_525_60 : returnvalue = signalStandardName(STANDARD_525_60); break;
|
|
case V4L2_STD_625_50 : returnvalue = signalStandardName(STANDARD_625_50); break;
|
|
case V4L2_STD_ALL : returnvalue = signalStandardName(STANDARD_ALL); break;
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
switch(standard)
|
|
{
|
|
case VIDEO_MODE_PAL : returnvalue = signalStandardName(STANDARD_PAL); break;
|
|
case VIDEO_MODE_NTSC : returnvalue = signalStandardName(STANDARD_NTSC); break;
|
|
case VIDEO_MODE_SECAM : returnvalue = signalStandardName(STANDARD_SECAM); break;
|
|
case VIDEO_MODE_AUTO : returnvalue = signalStandardName(STANDARD_ALL); break; // It must be disabled until I find a correct way to handle those non-standard bttv modes
|
|
// case VIDEO_MODE_PAL_Nc : returnvalue = signalStandardName(STANDARD_PAL_Nc); break; // Undocumented value found to be compatible with V4L bttv driver
|
|
case VIDEO_MODE_PAL_M : returnvalue = signalStandardName(STANDARD_PAL_M); break; // Undocumented value found to be compatible with V4L bttv driver
|
|
case VIDEO_MODE_PAL_N : returnvalue = signalStandardName(STANDARD_PAL_N); break; // Undocumented value found to be compatible with V4L bttv driver
|
|
case VIDEO_MODE_NTSC_JP : returnvalue = signalStandardName(STANDARD_NTSC_M_JP); break; // Undocumented value found to be compatible with V4L bttv driver
|
|
}
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
return returnvalue;
|
|
}
|
|
|
|
/*!
|
|
\fn VideoDevice::detectSignalStandards()
|
|
*/
|
|
int VideoDevice::detectSignalStandards()
|
|
{
|
|
switch(m_driver)
|
|
{
|
|
#if defined(__linux__) && defined(ENABLE_AV)
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
case VIDEODEV_DRIVER_V4L2:
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_V4L:
|
|
break;
|
|
#endif
|
|
case VIDEODEV_DRIVER_NONE:
|
|
default:
|
|
break;
|
|
}
|
|
//FIXME: return a real value
|
|
return 0;
|
|
}
|
|
|
|
/*!
|
|
\fn VideoDevice::initRead()
|
|
*/
|
|
int VideoDevice::initRead()
|
|
{
|
|
/// @todo implement me
|
|
|
|
kdDebug(14010) << k_funcinfo << "called." << endl;
|
|
if(isOpen())
|
|
{
|
|
m_rawbuffers.resize(1);
|
|
if (m_rawbuffers.size()==0)
|
|
{
|
|
fprintf (stderr, "Out of memory\n");
|
|
return EXIT_FAILURE;
|
|
}
|
|
kdDebug(14010) << k_funcinfo << "m_buffer_size: " << m_buffer_size << endl;
|
|
|
|
// m_rawbuffers[0].pixelformat=m_pixelformat;
|
|
m_rawbuffers[0].length = m_buffer_size;
|
|
m_rawbuffers[0].start = (uchar *)malloc (m_buffer_size);
|
|
|
|
if (!m_rawbuffers[0].start)
|
|
{
|
|
fprintf (stderr, "Out of memory\n");
|
|
return EXIT_FAILURE;
|
|
}
|
|
kdDebug(14010) << k_funcinfo << "exited successfuly." << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
|
|
/*!
|
|
\fn VideoDevice::initMmap()
|
|
*/
|
|
int VideoDevice::initMmap()
|
|
{
|
|
/// @todo implement me
|
|
#define BUFFERS 2
|
|
if(isOpen())
|
|
{
|
|
kdDebug(14010) << k_funcinfo << full_filename << " Trying to MMAP" << endl;
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
struct v4l2_requestbuffers req;
|
|
|
|
CLEAR (req);
|
|
|
|
req.count = BUFFERS;
|
|
req.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
req.memory = V4L2_MEMORY_MMAP;
|
|
|
|
if (-1 == xioctl (VIDIOC_REQBUFS, &req))
|
|
{
|
|
if (EINVAL == errno)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << full_filename << " does not support memory mapping" << endl;
|
|
return EXIT_FAILURE;
|
|
}
|
|
else
|
|
{
|
|
return errnoReturn ("VIDIOC_REQBUFS");
|
|
}
|
|
}
|
|
|
|
if (req.count < BUFFERS)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Insufficient buffer memory on " << full_filename << endl;
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
m_rawbuffers.resize(req.count);
|
|
|
|
if (m_rawbuffers.size()==0)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Out of memory" << endl;
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
for (m_streambuffers = 0; m_streambuffers < req.count; ++m_streambuffers)
|
|
{
|
|
struct v4l2_buffer v4l2buffer;
|
|
|
|
CLEAR (v4l2buffer);
|
|
|
|
v4l2buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
v4l2buffer.memory = V4L2_MEMORY_MMAP;
|
|
v4l2buffer.index = m_streambuffers;
|
|
|
|
if (-1 == xioctl (VIDIOC_QUERYBUF, &v4l2buffer))
|
|
return errnoReturn ("VIDIOC_QUERYBUF");
|
|
|
|
m_rawbuffers[m_streambuffers].length = v4l2buffer.length;
|
|
m_rawbuffers[m_streambuffers].start = (uchar *) mmap (NULL /* start anywhere */, v4l2buffer.length, PROT_READ | PROT_WRITE /* required */, MAP_SHARED /* recommended */, descriptor, v4l2buffer.m.offset);
|
|
|
|
if (MAP_FAILED == m_rawbuffers[m_streambuffers].start)
|
|
return errnoReturn ("mmap");
|
|
}
|
|
#endif
|
|
m_currentbuffer.data.resize(m_rawbuffers[0].length); // Makes the imagesize.data buffer size equal to the rawbuffer size
|
|
kdDebug(14010) << k_funcinfo << full_filename << " m_currentbuffer.data.size(): " << m_currentbuffer.data.size() << endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
|
|
/*!
|
|
\fn VideoDevice::initUserptr()
|
|
*/
|
|
int VideoDevice::initUserptr()
|
|
{
|
|
/// @todo implement me
|
|
if(isOpen())
|
|
{
|
|
#ifdef V4L2_CAP_VIDEO_CAPTURE
|
|
struct v4l2_requestbuffers req;
|
|
|
|
CLEAR (req);
|
|
|
|
req.count = 2;
|
|
req.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
req.memory = V4L2_MEMORY_USERPTR;
|
|
|
|
if (-1 == xioctl (VIDIOC_REQBUFS, &req))
|
|
{
|
|
if (EINVAL == errno)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << full_filename << " does not support memory mapping" << endl;
|
|
return EXIT_FAILURE;
|
|
}
|
|
else
|
|
{
|
|
return errnoReturn ("VIDIOC_REQBUFS");
|
|
}
|
|
}
|
|
|
|
m_rawbuffers.resize(4);
|
|
|
|
if (m_rawbuffers.size()==0)
|
|
{
|
|
fprintf (stderr, "Out of memory\n");
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
for (m_streambuffers = 0; m_streambuffers < 4; ++m_streambuffers)
|
|
{
|
|
m_rawbuffers[m_streambuffers].length = m_buffer_size;
|
|
m_rawbuffers[m_streambuffers].start = (uchar *) malloc (m_buffer_size);
|
|
|
|
if (!m_rawbuffers[m_streambuffers].start)
|
|
{
|
|
kdDebug(14010) << k_funcinfo << "Out of memory" << endl;
|
|
return EXIT_FAILURE;
|
|
}
|
|
}
|
|
#endif
|
|
return EXIT_SUCCESS;
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
bool VideoDevice::canCapture()
|
|
{
|
|
return m_videocapture;
|
|
}
|
|
|
|
bool VideoDevice::canChromakey()
|
|
{
|
|
return m_videochromakey;
|
|
}
|
|
|
|
bool VideoDevice::canScale()
|
|
{
|
|
return m_videoscale;
|
|
}
|
|
|
|
bool VideoDevice::canOverlay()
|
|
{
|
|
return m_videooverlay;
|
|
}
|
|
|
|
bool VideoDevice::canRead()
|
|
{
|
|
return m_videoread;
|
|
}
|
|
|
|
bool VideoDevice::canAsyncIO()
|
|
{
|
|
return m_videoasyncio;
|
|
}
|
|
|
|
bool VideoDevice::canStream()
|
|
{
|
|
return m_videostream;
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
}
|