You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
tdenetwork/kopete/libkopete/avdevice/videodevice.cpp

2751 lines
92 KiB

/*
videodevice.cpp - Kopete Video Device Low-level Support
Copyright (c) 2005-2006 by Cláudio da Silveira Pinheiro <taupter@gmail.com>
Kopete (c) 2002-2003 by the Kopete developers <kopete-devel@kde.org>
*************************************************************************
* *
* This library is free software; you can redistribute it and/or *
* modify it under the terms of the GNU Lesser General Public *
* License as published by the Free Software Foundation; either *
* version 2 of the License, or (at your option) any later version. *
* *
*************************************************************************
*/
#include <cstdlib>
#include <cerrno>
#include <cstring>
#include <kdebug.h>
#include "videoinput.h"
#include "videodevice.h"
#include "bayer.h"
#include "sonix_compress.h"
#define CLEAR(x) memset (&(x), 0, sizeof (x))
namespace Kopete {
namespace AV {
VideoDevice::VideoDevice()
{
// kdDebug(14010) << "libkopete (avdevice): VideoDevice() called" << endl;
descriptor = -1;
m_streambuffers = 0;
m_current_input = 0;
// kdDebug(14010) << "libkopete (avdevice): VideoDevice() exited successfuly" << endl;
maxwidth = 32767;
maxheight = 32767;
minwidth = 1;
minheight = 1;
}
VideoDevice::~VideoDevice()
{
}
#ifdef V4L2_CAP_VIDEO_CAPTURE
void VideoDevice::enumerateMenu (void)
{
kdDebug(14010) << k_funcinfo << " Menu items:" << endl;
memset (&querymenu, 0, sizeof (querymenu));
querymenu.id = queryctrl.id;
for (querymenu.index = queryctrl.minimum; querymenu.index <= queryctrl.maximum; querymenu.index++)
{
if (0 == xioctl (VIDIOC_QUERYMENU, &querymenu))
{
kdDebug(14010) << k_funcinfo << " " << TQString(TQString::fromLocal8Bit((const char*)querymenu.name)) << endl;
}
else
{
perror ("VIDIOC_QUERYMENU");
exit (EXIT_FAILURE);
}
}
}
#endif
/*!
\fn VideoDevice::xioctl(int fd, int request, void *arg)
*/
int VideoDevice::xioctl(int request, void *arg)
{
int r;
do r = ioctl (descriptor, request, arg);
while (-1 == r && EINTR == errno);
return r;
}
/*!
\fn VideoDevice::errnoReturn(const char* s)
*/
int VideoDevice::errnoReturn(const char* s)
{
/// @todo implement me
fprintf (stderr, "%s error %d, %s\n",s, errno, strerror (errno));
return EXIT_FAILURE;
}
/*!
\fn VideoDevice::setFileName(TQString name)
*/
int VideoDevice::setFileName(TQString filename)
{
/// @todo implement me
full_filename=filename;
return EXIT_SUCCESS;
}
/*!
\fn VideoDevice::open()
*/
int VideoDevice::open()
{
/// @todo implement me
kdDebug(14010) << k_funcinfo << "called" << endl;
if(-1 != descriptor)
{
kdDebug(14010) << k_funcinfo << "Device is already open" << endl;
return EXIT_SUCCESS;
}
descriptor = ::open (TQFile::encodeName(full_filename), O_RDWR, 0);
if(isOpen())
{
kdDebug(14010) << k_funcinfo << "File " << full_filename << " was opened successfuly" << endl;
if(EXIT_FAILURE==checkDevice())
{
kdDebug(14010) << k_funcinfo << "File " << full_filename << " could not be opened" << endl;
close();
return EXIT_FAILURE;
}
}
else
{
kdDebug(14010) << k_funcinfo << "Unable to open file " << full_filename << "Err: "<< errno << endl;
return EXIT_FAILURE;
}
initDevice();
selectInput(m_current_input);
kdDebug(14010) << k_funcinfo << "exited successfuly" << endl;
return EXIT_SUCCESS;
}
bool VideoDevice::isOpen()
{
if(-1 == descriptor)
{
// kdDebug(14010) << k_funcinfo << "VideoDevice::isOpen() File is not open" << endl;
return false;
}
// kdDebug(14010) << k_funcinfo << "VideoDevice::isOpen() File is open" << endl;
return true;
}
int VideoDevice::checkDevice()
{
kdDebug(14010) << k_funcinfo << "checkDevice() called." << endl;
if(isOpen())
{
m_videocapture=false;
m_videochromakey=false;
m_videoscale=false;
m_videooverlay=false;
m_videoread=false;
m_videoasyncio=false;
m_videostream=false;
m_driver=VIDEODEV_DRIVER_NONE;
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
//if(!getWorkaroundBrokenDriver())
{
kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " Trying V4L2 API." << endl;
CLEAR(V4L2_capabilities);
if (-1 != xioctl (VIDIOC_QUERYCAP, &V4L2_capabilities))
{
if (!(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_CAPTURE))
{
kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " is not a video capture device." << endl;
m_driver = VIDEODEV_DRIVER_NONE;
return EXIT_FAILURE;
}
m_videocapture=true;
kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " is a V4L2 device." << endl;
m_driver = VIDEODEV_DRIVER_V4L2;
m_model=TQString::fromLocal8Bit((const char*)V4L2_capabilities.card);
// Detect maximum and minimum resolution supported by the V4L2 device
CLEAR (fmt);
fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
if (-1 == xioctl (VIDIOC_G_FMT, &fmt))
kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ")." << endl;
fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
fmt.fmt.pix.width = 32767;
fmt.fmt.pix.height = 32767;
fmt.fmt.pix.field = V4L2_FIELD_ANY;
if (-1 == xioctl (VIDIOC_S_FMT, &fmt))
{
kdDebug(14010) << k_funcinfo << "Detecting maximum size with VIDIOC_S_FMT failed (" << errno << ").Returned maxwidth: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
// Note VIDIOC_S_FMT may change width and height.
}
else
{
maxwidth = fmt.fmt.pix.width;
maxheight = fmt.fmt.pix.height;
}
if (-1 == xioctl (VIDIOC_G_FMT, &fmt))
kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ")." << endl;
fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
fmt.fmt.pix.width = 1;
fmt.fmt.pix.height = 1;
fmt.fmt.pix.field = V4L2_FIELD_ANY;
if (-1 == xioctl (VIDIOC_S_FMT, &fmt))
{
kdDebug(14010) << k_funcinfo << "Detecting minimum size with VIDIOC_S_FMT failed (" << errno << ").Returned maxwidth: " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
// Note VIDIOC_S_FMT may change width and height.
}
else
{
minwidth = fmt.fmt.pix.width;
minheight = fmt.fmt.pix.height;
}
// Buggy driver paranoia
/* min = fmt.fmt.pix.width * 2;
if (fmt.fmt.pix.bytesperline < min)
fmt.fmt.pix.bytesperline = min;
min = fmt.fmt.pix.bytesperline * fmt.fmt.pix.height;
if (fmt.fmt.pix.sizeimage < min)
fmt.fmt.pix.sizeimage = min;
m_buffer_size=fmt.fmt.pix.sizeimage ;*/
int inputisok=EXIT_SUCCESS;
m_input.clear();
for(unsigned int loop=0; inputisok==EXIT_SUCCESS; loop++)
{
struct v4l2_input videoinput;
CLEAR(videoinput);
videoinput.index = loop;
inputisok=xioctl(VIDIOC_ENUMINPUT, &videoinput);
if(inputisok==EXIT_SUCCESS)
{
VideoInput tempinput;
tempinput.name = TQString::fromLocal8Bit((const char*)videoinput.name);
tempinput.hastuner = videoinput.type & V4L2_INPUT_TYPE_TUNER;
tempinput.m_standards = videoinput.std;
m_input.push_back(tempinput);
kdDebug(14010) << k_funcinfo << "Input " << loop << ": " << tempinput.name << " (tuner: " << ((videoinput.type & V4L2_INPUT_TYPE_TUNER) != 0) << ")" << endl;
if((videoinput.type & V4L2_INPUT_TYPE_TUNER) != 0)
{
// _tunerForInput[name] = desc.tuner;
// _isTuner = true;
}
else
{
// _tunerForInput[name] = -1;
}
}
}
// -----------------------------------------------------------------------------------------------------------------
// This must turn up to be a proper method to check for controls' existence.
CLEAR (queryctrl);
// v4l2_queryctrl may zero the .id in some cases, even if the IOCTL returns EXIT_SUCCESS (tested with a bttv card, when testing for V4L2_CID_AUDIO_VOLUME).
// As of 6th Aug 2007, according to the V4L2 specification version 0.21, this behavior is undocumented, and the example 1-8 code found at
// http://www.linuxtv.org/downloads/video4linux/API/V4L2_API/spec/x519.htm fails because of this behavior with a bttv card.
int currentid = V4L2_CID_BASE;
kdDebug(14010) << k_funcinfo << "Checking CID controls" << endl;
for (currentid = V4L2_CID_BASE; currentid < V4L2_CID_LASTP1; currentid++)
//for (queryctrl.id = 9963776; queryctrl.id < 9963800; queryctrl.id++)
{
queryctrl.id = currentid;
//kdDebug(14010) << k_funcinfo << "Checking CID controls from " << V4L2_CID_BASE << " to " << V4L2_CID_LASTP1 << ". Current: " << queryctrl.id << ". IOCTL returns: " << resultado << endl;
if (0 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
{
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
continue;
//kdDebug(14010) << k_funcinfo << " Control: " << TQString::fromLocal8Bit((const char*)queryctrl.name) << endl;
kdDebug(14010) << k_funcinfo << " Control: " << TQString(TQString::fromLocal8Bit((const char*)queryctrl.name)) << " Values from " << queryctrl.minimum << " to " << queryctrl.maximum << " with steps of " << queryctrl.step << ". Default: " << queryctrl.default_value << endl;
/* switch (queryctrl.type)
{
case V4L2_CTRL_TYPE_INTEGER :
}*/
if (queryctrl.type == V4L2_CTRL_TYPE_MENU)
enumerateMenu ();
}
else
{
if (errno == EINVAL)
continue;
perror ("VIDIOC_QUERYCTRL");
// exit (EXIT_FAILURE);
}
}
kdDebug(14010) << k_funcinfo << "Checking CID private controls" << endl;
for (currentid = V4L2_CID_PRIVATE_BASE;; currentid++)
//for (queryctrl.id = 9963776; queryctrl.id < 9963800; queryctrl.id++)
{
queryctrl.id = currentid;
//kdDebug(14010) << k_funcinfo << "Checking CID private controls from " << V4L2_CID_PRIVATE_BASE << ". Current: " << queryctrl.id << ". IOCTL returns: " << resultado << endl;
if ( 0 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
{
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
continue;
kdDebug(14010) << k_funcinfo << " Control: " << TQString(TQString::fromLocal8Bit((const char*)queryctrl.name)) << " Values from " << queryctrl.minimum << " to " << queryctrl.maximum << " with steps of " << queryctrl.step << ". Default: " << queryctrl.default_value << endl;
if (queryctrl.type == V4L2_CTRL_TYPE_MENU)
enumerateMenu ();
}
else
{
if (errno == EINVAL)
break;
perror ("VIDIOC_QUERYCTRL");
// exit (EXIT_FAILURE);
}
}
}
else
{
// V4L-only drivers should return an EINVAL in errno to indicate they cannot handle V4L2 calls. Not every driver is compliant, so
// it will try the V4L api even if the error code is different than expected.
kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " is not a V4L2 device." << endl;
}
}
#endif
CLEAR(V4L_capabilities);
if(m_driver==VIDEODEV_DRIVER_NONE)
{
kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " Trying V4L API." << endl;
if (-1 == xioctl (VIDIOCGCAP, &V4L_capabilities))
{
perror ("ioctl (VIDIOCGCAP)");
m_driver = VIDEODEV_DRIVER_NONE;
return EXIT_FAILURE;
}
else
{
kdDebug(14010) << k_funcinfo << full_filename << " is a V4L device." << endl;
m_driver = VIDEODEV_DRIVER_V4L;
m_model=TQString::fromLocal8Bit((const char*)V4L_capabilities.name);
if(V4L_capabilities.type & VID_TYPE_CAPTURE)
m_videocapture=true;
if(V4L_capabilities.type & VID_TYPE_CHROMAKEY)
m_videochromakey=true;
if(V4L_capabilities.type & VID_TYPE_SCALES)
m_videoscale=true;
if(V4L_capabilities.type & VID_TYPE_OVERLAY)
m_videooverlay=true;
// kdDebug(14010) << "libkopete (avdevice): Inputs : " << V4L_capabilities.channels << endl;
// kdDebug(14010) << "libkopete (avdevice): Audios : " << V4L_capabilities.audios << endl;
minwidth = V4L_capabilities.minwidth;
maxwidth = V4L_capabilities.maxwidth;
minheight = V4L_capabilities.minheight;
maxheight = V4L_capabilities.maxheight;
int inputisok=EXIT_SUCCESS;
m_input.clear();
for(int loop=0; loop < V4L_capabilities.channels; loop++)
{
struct video_channel videoinput;
CLEAR(videoinput);
videoinput.channel = loop;
videoinput.norm = 1;
inputisok=xioctl(VIDIOCGCHAN, &videoinput);
if(inputisok==EXIT_SUCCESS)
{
VideoInput tempinput;
tempinput.name = TQString::fromLocal8Bit((const char*)videoinput.name);
tempinput.hastuner=videoinput.flags & VIDEO_VC_TUNER;
// TODO: The routine to detect the appropriate video standards for V4L must be placed here
m_input.push_back(tempinput);
// kdDebug(14010) << "libkopete (avdevice): Input " << loop << ": " << tempinput.name << " (tuner: " << ((videoinput.flags & VIDEO_VC_TUNER) != 0) << ")" << endl;
/* if((input.type & V4L2_INPUT_TYPE_TUNER) != 0)
{
// _tunerForInput[name] = desc.tuner;
// _isTuner = true;
}
else
{
// _tunerForInput[name] = -1;
}
*/ }
}
}
}
#endif
m_name=m_model; // Take care about changing the name to be different from the model itself...
detectPixelFormats();
// TODO: Now we must execute the proper initialization according to the type of the driver.
kdDebug(14010) << k_funcinfo << "checkDevice() exited successfuly." << endl;
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
/*!
\fn VideoDevice::showDeviceCapabilities()
*/
int VideoDevice::showDeviceCapabilities()
{
kdDebug(14010) << k_funcinfo << "showDeviceCapabilities() called." << endl;
if(isOpen())
{
/* kdDebug(14010) << "libkopete (avdevice): Driver: " << (const char*)V4L2_capabilities.driver << " "
<< ((V4L2_capabilities.version>>16) & 0xFF) << "."
<< ((V4L2_capabilities.version>> 8) & 0xFF) << "."
<< ((V4L2_capabilities.version ) & 0xFF) << endl;
kdDebug(14010) << "libkopete (avdevice): Card: " << name << endl;
kdDebug(14010) << "libkopete (avdevice): Capabilities:" << endl;
if(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_CAPTURE)
kdDebug(14010) << "libkopete (avdevice): Video capture" << endl;
if(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_OUTPUT)
kdDebug(14010) << "libkopete (avdevice): Video output" << endl;
if(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_OVERLAY)
kdDebug(14010) << "libkopete (avdevice): Video overlay" << endl;
if(V4L2_capabilities.capabilities & V4L2_CAP_VBI_CAPTURE)
kdDebug(14010) << "libkopete (avdevice): VBI capture" << endl;
if(V4L2_capabilities.capabilities & V4L2_CAP_VBI_OUTPUT)
kdDebug(14010) << "libkopete (avdevice): VBI output" << endl;
if(V4L2_capabilities.capabilities & V4L2_CAP_RDS_CAPTURE)
kdDebug(14010) << "libkopete (avdevice): RDS capture" << endl;
if(V4L2_capabilities.capabilities & V4L2_CAP_TUNER)
kdDebug(14010) << "libkopete (avdevice): Tuner IO" << endl;
if(V4L2_capabilities.capabilities & V4L2_CAP_AUDIO)
kdDebug(14010) << "libkopete (avdevice): Audio IO" << endl;
;*/
kdDebug(14010) << k_funcinfo << "Card model: " << m_model << endl;
kdDebug(14010) << k_funcinfo << "Card name : " << m_name << endl;
kdDebug(14010) << k_funcinfo << "Capabilities:" << endl;
if(canCapture())
kdDebug(14010) << k_funcinfo << " Video capture" << endl;
if(canRead())
kdDebug(14010) << k_funcinfo << " Read" << endl;
if(canAsyncIO())
kdDebug(14010) << k_funcinfo << " Asynchronous input/output" << endl;
if(canStream())
kdDebug(14010) << k_funcinfo << " Streaming" << endl;
if(canChromakey())
kdDebug(14010) << k_funcinfo << " Video chromakey" << endl;
if(canScale())
kdDebug(14010) << k_funcinfo << " Video scales" << endl;
if(canOverlay())
kdDebug(14010) << k_funcinfo << " Video overlay" << endl;
// kdDebug(14010) << "libkopete (avdevice): Audios : " << V4L_capabilities.audios << endl;
kdDebug(14010) << k_funcinfo << " Max res: " << maxWidth() << " x " << maxHeight() << endl;
kdDebug(14010) << k_funcinfo << " Min res: " << minWidth() << " x " << minHeight() << endl;
kdDebug(14010) << k_funcinfo << " Inputs : " << inputs() << endl;
for (unsigned int loop=0; loop < inputs(); loop++)
kdDebug(14010) << k_funcinfo << "Input " << loop << ": " << m_input[loop].name << " (tuner: " << m_input[loop].hastuner << ")" << endl;
kdDebug(14010) << k_funcinfo << "showDeviceCapabilities() exited successfuly." << endl;
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
/*!
\fn VideoDevicePool::initDevice()
*/
int VideoDevice::initDevice()
{
/// @todo implement me
kdDebug(14010) << k_funcinfo << "initDevice() started" << endl;
if(-1 == descriptor)
{
kdDebug(14010) << k_funcinfo << "initDevice() Device is not open" << endl;
return EXIT_FAILURE;
}
m_io_method = IO_METHOD_NONE;
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
if(V4L2_capabilities.capabilities & V4L2_CAP_READWRITE)
{
m_videoread=true;
m_io_method = IO_METHOD_READ;
kdDebug(14010) << k_funcinfo << " Read/Write interface" << endl;
}
if(V4L2_capabilities.capabilities & V4L2_CAP_ASYNCIO)
{
m_videoasyncio=true;
kdDebug(14010) << k_funcinfo << " Async IO interface" << endl;
}
if(V4L2_capabilities.capabilities & V4L2_CAP_STREAMING)
{
m_videostream=true;
m_io_method = IO_METHOD_MMAP;
// m_io_method = IO_METHOD_USERPTR;
kdDebug(14010) << k_funcinfo << " Streaming interface" << endl;
}
if(m_io_method==IO_METHOD_NONE)
{
kdDebug(14010) << k_funcinfo << "initDevice() Found no suitable input/output method for " << full_filename << endl;
return EXIT_FAILURE;
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
m_videoread=true;
m_io_method=IO_METHOD_READ;
if(-1 != xioctl(VIDIOCGFBUF,&V4L_videobuffer))
{
// m_videostream=true;
// m_io_method = IO_METHOD_MMAP;
kdDebug(14010) << k_funcinfo << " Streaming interface" << endl;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
// Select video input, video standard and tune here.
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
cropcap.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
if (-1 == xioctl (VIDIOC_CROPCAP, &cropcap))
{ // Errors ignored.
}
crop.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
crop.c = cropcap.defrect; // reset to default
if (-1 == xioctl (VIDIOC_S_CROP, &crop))
{
switch (errno)
{
case EINVAL: break; // Cropping not supported.
default: break; // Errors ignored.
}
}
#endif
#endif
showDeviceCapabilities();
kdDebug(14010) << k_funcinfo << "initDevice() exited successfuly" << endl;
return EXIT_SUCCESS;
}
unsigned int VideoDevice::inputs()
{
return m_input.size();
}
int VideoDevice::width()
{
return currentwidth;
}
int VideoDevice::minWidth()
{
return minwidth;
}
int VideoDevice::maxWidth()
{
return maxwidth;
}
int VideoDevice::height()
{
return currentheight;
}
int VideoDevice::minHeight()
{
return minheight;
}
int VideoDevice::maxHeight()
{
return maxheight;
}
int VideoDevice::setSize( int newwidth, int newheight)
{
kdDebug(14010) << k_funcinfo << "setSize(" << newwidth << ", " << newheight << ") called." << endl;
if(isOpen())
{
// It should not be there. It must remain in a completely distict place, cause this method should not change the pixelformat.
kdDebug(14010) << k_funcinfo << "Trying YUY422P" << endl;
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_YUV422P))
{
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support YUV422P format. Trying YUYV." << endl;
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_YUYV))
{
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support YUYV format. Trying UYVY." << endl;
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_UYVY))
{
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support UYVY format. Trying YUV420P." << endl;
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_YUV420P))
{
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support YUV420P format. Trying RGB24." << endl;
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_RGB24))
{
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support RGB24 format. Trying BGR24." << endl;
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_BGR24))
{
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support RGB24 format. Trying RGB32." << endl;
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_RGB32))
{
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support RGB32 format. Trying BGR32." << endl;
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_BGR32))
{
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support BGR32 format. Trying SN9C10X." << endl;
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_SN9C10X))
{
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support SN9C10X format. Trying Bayer RGB." << endl;
if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_SBGGR8))
kdDebug(14010) << k_funcinfo << "Card doesn't seem to support SBGGR8 format. Fallback from it is not yet implemented." << endl;
}
}
}
}
}
}
}
}
}
if(newwidth > maxwidth ) newwidth = maxwidth;
if(newheight > maxheight) newheight = maxheight;
if(newwidth < minwidth ) newwidth = minwidth;
if(newheight < minheight) newheight = minheight;
currentwidth = newwidth;
currentheight = newheight;
//kdDebug(14010) << k_funcinfo << "width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << width() << "x" << height() << endl;
// Change resolution for the video device
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
// CLEAR (fmt);
if (-1 == xioctl (VIDIOC_G_FMT, &fmt))
kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
fmt.fmt.pix.width = width();
fmt.fmt.pix.height = height();
fmt.fmt.pix.field = V4L2_FIELD_ANY;
if (-1 == xioctl (VIDIOC_S_FMT, &fmt))
{
kdDebug(14010) << k_funcinfo << "VIDIOC_S_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
// Note VIDIOC_S_FMT may change width and height.
}
else
{
// Buggy driver paranoia.
kdDebug(14010) << k_funcinfo << "VIDIOC_S_FMT worked (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
unsigned int min = fmt.fmt.pix.width * 2;
if (fmt.fmt.pix.bytesperline < min)
fmt.fmt.pix.bytesperline = min;
min = fmt.fmt.pix.bytesperline * fmt.fmt.pix.height;
if (fmt.fmt.pix.sizeimage < min)
fmt.fmt.pix.sizeimage = min;
m_buffer_size=fmt.fmt.pix.sizeimage ;
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
{
struct video_window V4L_videowindow;
kdDebug(14010) << "------------- width: " << V4L_videowindow.width << " Height: " << V4L_videowindow.height << " Clipcount: " << V4L_videowindow.clipcount << " -----------------" << endl;
if (xioctl (VIDIOCGWIN, &V4L_videowindow)== -1)
{
perror ("ioctl VIDIOCGWIN");
// return (NULL);
}
V4L_videowindow.width = width();
V4L_videowindow.height = height();
V4L_videowindow.clipcount=0;
if (xioctl (VIDIOCSWIN, &V4L_videowindow)== -1)
{
perror ("ioctl VIDIOCSWIN");
// return (NULL);
}
kdDebug(14010) << "------------- width: " << V4L_videowindow.width << " Height: " << V4L_videowindow.height << " Clipcount: " << V4L_videowindow.clipcount << " -----------------" << endl;
// kdDebug(14010) << "libkopete (avdevice): V4L_picture.palette: " << V4L_picture.palette << " Depth: " << V4L_picture.depth << endl;
/* if(-1 == xioctl(VIDIOCGFBUF,&V4L_videobuffer))
kdDebug(14010) << "libkopete (avdevice): VIDIOCGFBUF failed (" << errno << "): Card cannot stream" << endl;*/
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
m_buffer_size = width() * height() * pixelFormatDepth(m_pixelformat) / 8;
kdDebug(14010) << "------------------------- ------- -- m_buffer_size: " << m_buffer_size << " !!! -- ------- -----------------------------------------" << endl;
m_currentbuffer.pixelformat=m_pixelformat;
m_currentbuffer.data.resize(m_buffer_size);
switch (m_io_method)
{
case IO_METHOD_NONE: break;
case IO_METHOD_READ: initRead (); break;
case IO_METHOD_MMAP: initMmap (); break;
case IO_METHOD_USERPTR: initUserptr (); break;
}
kdDebug(14010) << k_funcinfo << "setSize(" << newwidth << ", " << newheight << ") exited successfuly." << endl;
return EXIT_SUCCESS;
}
kdDebug(14010) << k_funcinfo << "setSize(" << newwidth << ", " << newheight << ") Device is not open." << endl;
return EXIT_FAILURE;
}
pixel_format VideoDevice::setPixelFormat(pixel_format newformat)
{
pixel_format ret = PIXELFORMAT_NONE;
//kdDebug(14010) << k_funcinfo << "called." << endl;
// Change the pixel format for the video device
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
// CLEAR (fmt);
if (-1 == xioctl (VIDIOC_G_FMT, &fmt))
{
// return errnoReturn ("VIDIOC_S_FMT");
// kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
}
else
m_pixelformat = pixelFormatForPalette(fmt.fmt.pix.pixelformat);
fmt.fmt.pix.pixelformat = pixelFormatCode(newformat);
if (-1 == xioctl (VIDIOC_S_FMT, &fmt))
{
// kdDebug(14010) << k_funcinfo << "VIDIOC_S_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl;
}
else
{
if (fmt.fmt.pix.pixelformat == pixelFormatCode(newformat))
{
m_pixelformat = newformat;
ret = m_pixelformat;
}
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
{
struct video_picture V4L_picture;
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
// kdDebug(14010) << k_funcinfo << "V4L_picture.palette: " << V4L_picture.palette << " Depth: " << V4L_picture.depth << endl;
V4L_picture.palette = pixelFormatCode(newformat);
V4L_picture.depth = pixelFormatDepth(newformat);
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
{
// kdDebug(14010) << k_funcinfo << "Card seems to not support " << pixelFormatName(newformat) << " format. Fallback to it is not yet implemented." << endl;
}
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
// kdDebug(14010) << k_funcinfo << "V4L_picture.palette: " << V4L_picture.palette << " Depth: " << V4L_picture.depth << endl;
m_pixelformat=pixelFormatForPalette(V4L_picture.palette);
if (m_pixelformat == newformat)
ret = newformat;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
return ret;
}
/*!
\fn Kopete::AV::VideoDevice::currentInput()
*/
int VideoDevice::currentInput()
{
/// @todo implement me
if(isOpen())
{
return m_current_input;
}
return 0;
}
/*!
\fn Kopete::AV::VideoDevice::selectInput(int input)
*/
int VideoDevice::selectInput(int newinput)
{
/// @todo implement me
if(m_current_input >= inputs())
return EXIT_FAILURE;
if(isOpen())
{
switch (m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
if (-1 == ioctl (descriptor, VIDIOC_S_INPUT, &newinput))
{
perror ("VIDIOC_S_INPUT");
return EXIT_FAILURE;
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
struct video_channel V4L_input;
V4L_input.channel=newinput;
V4L_input.norm=4; // Hey, it's plain wrong! It should be input's signal standard!
if (-1 == ioctl (descriptor, VIDIOCSCHAN, &V4L_input))
{
perror ("ioctl (VIDIOCSCHAN)");
return EXIT_FAILURE;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
kdDebug(14010) << k_funcinfo << "Selected input " << newinput << " (" << m_input[newinput].name << ")" << endl;
m_current_input = newinput;
setInputParameters();
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
/*!
\fn Kopete::AV::VideoDevice::setInputParameters()
*/
int VideoDevice::setInputParameters()
{
/// @todo implement me
if( (isOpen()) && (m_current_input < inputs() ) )
{
setBrightness( getBrightness() );
setContrast( getContrast() );
setSaturation( getSaturation() );
setWhiteness( getWhiteness() );
setHue( getHue() );
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
/*!
\fn VideoDevice::startCapturing()
*/
int VideoDevice::startCapturing()
{
kdDebug(14010) << k_funcinfo << "called." << endl;
if(isOpen())
{
switch (m_io_method)
{
case IO_METHOD_NONE: // Card cannot capture frames
return EXIT_FAILURE;
break;
case IO_METHOD_READ: // Nothing to do
break;
case IO_METHOD_MMAP:
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
{
unsigned int loop;
for (loop = 0; loop < m_streambuffers; ++loop)
{
struct v4l2_buffer buf;
CLEAR (buf);
buf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
buf.memory = V4L2_MEMORY_MMAP;
buf.index = loop;
if (-1 == xioctl (VIDIOC_QBUF, &buf))
return errnoReturn ("VIDIOC_QBUF");
}
enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
if (-1 == xioctl (VIDIOC_STREAMON, &type))
return errnoReturn ("VIDIOC_STREAMON");
}
#endif
#endif
break;
case IO_METHOD_USERPTR:
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
{
unsigned int loop;
for (loop = 0; loop < m_streambuffers; ++loop)
{
struct v4l2_buffer buf;
CLEAR (buf);
buf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
buf.memory = V4L2_MEMORY_USERPTR;
buf.m.userptr = (unsigned long) m_rawbuffers[loop].start;
buf.length = m_rawbuffers[loop].length;
if (-1 == xioctl (VIDIOC_QBUF, &buf))
return errnoReturn ("VIDIOC_QBUF");
}
enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
if (-1 == xioctl (VIDIOC_STREAMON, &type))
return errnoReturn ("VIDIOC_STREAMON");
}
#endif
#endif
break;
}
kdDebug(14010) << k_funcinfo << "exited successfuly." << endl;
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
/*!
\fn VideoDevice::getFrame()
*/
int VideoDevice::getFrame()
{
/// @todo implement me
ssize_t bytesread;
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
struct v4l2_buffer v4l2buffer;
#endif
#endif
// kdDebug(14010) << k_funcinfo << "getFrame() called." << endl;
if(isOpen())
{
switch (m_io_method)
{
case IO_METHOD_NONE: // Card cannot capture frames
return EXIT_FAILURE;
break;
case IO_METHOD_READ:
// kdDebug(14010) << k_funcinfo << "Using IO_METHOD_READ.File descriptor: " << descriptor << " Buffer address: " << &m_currentbuffer.data[0] << " Size: " << m_currentbuffer.data.size() << endl;
bytesread = read (descriptor, &m_currentbuffer.data[0], m_currentbuffer.data.size());
if (-1 == bytesread) // must verify this point with ov511 driver.
{
kdDebug(14010) << k_funcinfo << "IO_METHOD_READ failed." << endl;
switch (errno)
{
case EAGAIN:
return EXIT_FAILURE;
case EIO: /* Could ignore EIO, see spec. fall through */
default:
return errnoReturn ("read");
}
}
if((int)m_currentbuffer.data.size() < bytesread)
{
kdDebug(14010) << k_funcinfo << "IO_METHOD_READ returned less bytes (" << bytesread << ") than it was asked for (" << m_currentbuffer.data.size() <<")." << endl;
}
break;
case IO_METHOD_MMAP:
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
CLEAR (v4l2buffer);
v4l2buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
v4l2buffer.memory = V4L2_MEMORY_MMAP;
if (-1 == xioctl (VIDIOC_DQBUF, &v4l2buffer))
{
kdDebug(14010) << k_funcinfo << full_filename << " MMAPed getFrame failed." << endl;
switch (errno)
{
case EAGAIN:
{
kdDebug(14010) << k_funcinfo << full_filename << " MMAPed getFrame failed: EAGAIN. Pointer: " << endl;
return EXIT_FAILURE;
}
case EIO: /* Could ignore EIO, see spec. fall through */
default:
return errnoReturn ("VIDIOC_DQBUF");
}
}
/* if (v4l2buffer.index < m_streambuffers)
return EXIT_FAILURE;*/ //it was an assert()
//kdDebug(14010) << k_funcinfo << "m_rawbuffers[" << v4l2buffer.index << "].start: " << (void *)m_rawbuffers[v4l2buffer.index].start << " Size: " << m_currentbuffer.data.size() << endl;
/*{
unsigned long long result=0;
unsigned long long R=0, G=0, B=0, A=0;
int Rmax=0, Gmax=0, Bmax=0, Amax=0;
int Rmin=255, Gmin=255, Bmin=255, Amin=0;
for(unsigned int loop=0;loop < m_currentbuffer.data.size();loop+=4)
{
R+=m_rawbuffers[v4l2buffer.index].start[loop];
G+=m_rawbuffers[v4l2buffer.index].start[loop+1];
B+=m_rawbuffers[v4l2buffer.index].start[loop+2];
// A+=currentbuffer.data[loop+3];
if (m_currentbuffer.data[loop] < Rmin) Rmin = m_currentbuffer.data[loop];
if (m_currentbuffer.data[loop+1] < Gmin) Gmin = m_currentbuffer.data[loop+1];
if (m_currentbuffer.data[loop+2] < Bmin) Bmin = m_currentbuffer.data[loop+2];
// if (m_currentbuffer.data[loop+3] < Amin) Amin = m_currentbuffer.data[loop+3];
if (m_currentbuffer.data[loop] > Rmax) Rmax = m_currentbuffer.data[loop];
if (m_currentbuffer.data[loop+1] > Gmax) Gmax = m_currentbuffer.data[loop+1];
if (m_currentbuffer.data[loop+2] > Bmax) Bmax = m_currentbuffer.data[loop+2];
// if (m_currentbuffer.data[loop+3] > Amax) Amax = m_currentbuffer.data[loop+3];
}
kdDebug(14010) << " R: " << R << " G: " << G << " B: " << B << " A: " << A <<
" Rmin: " << Rmin << " Gmin: " << Gmin << " Bmin: " << Bmin << " Amin: " << Amin <<
" Rmax: " << Rmax << " Gmax: " << Gmax << " Bmax: " << Bmax << " Amax: " << Amax << endl;
}*/
memcpy(&m_currentbuffer.data[0], m_rawbuffers[v4l2buffer.index].start, m_currentbuffer.data.size());
if (-1 == xioctl (VIDIOC_QBUF, &v4l2buffer))
return errnoReturn ("VIDIOC_QBUF");
#endif
#endif
break;
case IO_METHOD_USERPTR:
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
{
unsigned int i;
CLEAR (v4l2buffer);
v4l2buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
v4l2buffer.memory = V4L2_MEMORY_USERPTR;
if (-1 == xioctl (VIDIOC_DQBUF, &v4l2buffer))
{
switch (errno)
{
case EAGAIN:
return EXIT_FAILURE;
case EIO: /* Could ignore EIO, see spec. fall through */
default:
return errnoReturn ("VIDIOC_DQBUF");
}
}
for (i = 0; i < m_streambuffers; ++i)
if (v4l2buffer.m.userptr == (unsigned long) m_rawbuffers[i].start && v4l2buffer.length == m_rawbuffers[i].length)
break;
if (i < m_streambuffers)
return EXIT_FAILURE;
if (-1 == xioctl (VIDIOC_QBUF, &v4l2buffer))
return errnoReturn ("VIDIOC_QBUF");
}
#endif
#endif
break;
}
/* Automatic color correction. Now it just swaps R and B channels in RGB24/BGR24 modes.
if(m_input[m_current_input].getAutoColorCorrection())
{
switch(m_currentbuffer.pixelformat)
{
case PIXELFORMAT_NONE : break;
case PIXELFORMAT_GREY : break;
case PIXELFORMAT_RGB332 : break;
case PIXELFORMAT_RGB555 : break;
case PIXELFORMAT_RGB555X: break;
case PIXELFORMAT_RGB565 : break;
case PIXELFORMAT_RGB565X: break;
case PIXELFORMAT_RGB24 :
case PIXELFORMAT_BGR24 :
{
unsigned char temp;
for(unsigned int loop=0;loop < m_currentbuffer.data.size();loop+=3)
{
temp = m_currentbuffer.data[loop];
m_currentbuffer.data[loop] = m_currentbuffer.data[loop+2];
m_currentbuffer.data[loop+2] = temp;
}
}
break;
case PIXELFORMAT_RGB32 :
case PIXELFORMAT_BGR32 :
{
unsigned char temp;
for(unsigned int loop=0;loop < m_currentbuffer.data.size();loop+=4)
{
temp = m_currentbuffer.data[loop];
m_currentbuffer.data[loop] = m_currentbuffer.data[loop+2];
m_currentbuffer.data[loop+2] = temp;
}
}
break;
case PIXELFORMAT_YUYV : break;
case PIXELFORMAT_UYVY : break;
case PIXELFORMAT_YUV420P: break;
case PIXELFORMAT_YUV422P: break;
}
}*/
//kdDebug(14010) << k_funcinfo << "10 Using IO_METHOD_READ.File descriptor: " << descriptor << " Buffer address: " << &m_currentbuffer.data[0] << " Size: " << m_currentbuffer.data.size() << endl;
// put frame copy operation here
// kdDebug(14010) << k_funcinfo << "exited successfuly." << endl;
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
/*!
\fn VideoDevice::getFrame(imagebuffer *imgbuffer)
*/
int VideoDevice::getFrame(imagebuffer *imgbuffer)
{
if(imgbuffer)
{
getFrame();
imgbuffer->height = m_currentbuffer.height;
imgbuffer->width = m_currentbuffer.width;
imgbuffer->pixelformat = m_currentbuffer.pixelformat;
imgbuffer->data = m_currentbuffer.data;
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
/*!
\fn Kopete::AV::VideoDevice::getImage(const TQImage *qimage)
*/
int VideoDevice::getImage(TQImage *qimage)
{
/// @todo implement me
// do NOT delete qimage here, as it is received as a parameter
if (qimage->width() != width() || qimage->height() != height())
qimage->create(width(), height(),32, TQImage::IgnoreEndian);
uchar *bits=qimage->bits();
// kDebug() << "Capturing in " << pixelFormatName(m_currentbuffer.pixelformat);
switch(m_currentbuffer.pixelformat)
{
case PIXELFORMAT_NONE : break;
// Packed RGB formats
case PIXELFORMAT_RGB332 : break;
case PIXELFORMAT_RGB444 : break;
case PIXELFORMAT_RGB555 : break;
case PIXELFORMAT_RGB565 :
{
int step=0;
for(int loop=0;loop < qimage->numBytes();loop+=4)
{
bits[loop] = (m_currentbuffer.data[step]<<3)+(m_currentbuffer.data[step]<<3>>5);
bits[loop+1] = ((m_currentbuffer.data[step+1])<<5)|m_currentbuffer.data[step]>>5;
bits[loop+2] = ((m_currentbuffer.data[step+1])&248)+((m_currentbuffer.data[step+1])>>5);
bits[loop+3] = 255;
step+=2;
}
}
break;
case PIXELFORMAT_RGB555X: break;
case PIXELFORMAT_RGB565X: break;
case PIXELFORMAT_BGR24 :
{
int step=0;
for(int loop=0;loop < qimage->numBytes();loop+=4)
{
bits[loop] = m_currentbuffer.data[step+2];
bits[loop+1] = m_currentbuffer.data[step+1];
bits[loop+2] = m_currentbuffer.data[step];
bits[loop+3] = 255;
step+=3;
}
}
break;
case PIXELFORMAT_RGB24 :
{
int step=0;
for(int loop=0;loop < qimage->numBytes();loop+=4)
{
bits[loop] = m_currentbuffer.data[step];
bits[loop+1] = m_currentbuffer.data[step+1];
bits[loop+2] = m_currentbuffer.data[step+2];
bits[loop+3] = 255;
step+=3;
}
}
break;
case PIXELFORMAT_BGR32 : break;
case PIXELFORMAT_RGB32 : memcpy(bits,&m_currentbuffer.data[0], m_currentbuffer.data.size());
break;
// Bayer RGB format
case PIXELFORMAT_SBGGR8 :
{
unsigned char *d = (unsigned char *) malloc (width() * height() * 3);
bayer2rgb24(d, &m_currentbuffer.data.first(), width(), height());
int step=0;
for(int loop=0;loop < qimage->numBytes();loop+=4)
{
bits[loop] = d[step+2];
bits[loop+1] = d[step+1];
bits[loop+2] = d[step];
bits[loop+3] = 255;
step+=3;
}
free(d);
}
break;
// YUV formats
case PIXELFORMAT_GREY : break;
case PIXELFORMAT_YUYV:
case PIXELFORMAT_UYVY:
case PIXELFORMAT_YUV420P:
case PIXELFORMAT_YUV422P:
{
uchar *yptr, *cbptr, *crptr;
bool halfheight=false;
bool packed=false;
// Adjust algorythm to specific YUV data arrangements.
if (m_currentbuffer.pixelformat == PIXELFORMAT_YUV420P)
halfheight=true;
if (m_currentbuffer.pixelformat == PIXELFORMAT_YUYV)
{
yptr = &m_currentbuffer.data[0];
cbptr = yptr + 1;
crptr = yptr + 3;
packed=true;
}
else if (m_currentbuffer.pixelformat == PIXELFORMAT_UYVY)
{
cbptr = &m_currentbuffer.data[0];
yptr = cbptr + 1;
crptr = cbptr + 2;
packed=true;
}
else
{
yptr = &m_currentbuffer.data[0];
cbptr = yptr + (width()*height());
crptr = cbptr + (width()*height()/(halfheight ? 4:2));
}
for(int y=0; y<height(); y++)
{
// Decode scanline
for(int x=0; x<width(); x++)
{
int c,d,e;
if (packed)
{
c = (yptr[x<<1])-16;
d = (cbptr[x>>1<<2])-128;
e = (crptr[x>>1<<2])-128;
}
else
{
c = (yptr[x])-16;
d = (cbptr[x>>1])-128;
e = (crptr[x>>1])-128;
}
int r = (298 * c + 409 * e + 128)>>8;
int g = (298 * c - 100 * d - 208 * e + 128)>>8;
int b = (298 * c + 516 * d + 128)>>8;
if (r<0) r=0; if (r>255) r=255;
if (g<0) g=0; if (g>255) g=255;
if (b<0) b=0; if (b>255) b=255;
uint *p = (uint*)qimage->scanLine(y)+x;
*p = tqRgba(r,g,b,255);
}
// Jump to next line
if (packed)
{
yptr+=width()*2;
cbptr+=width()*2;
crptr+=width()*2;
}
else
{
yptr+=width();
if (!halfheight || y&1)
{
cbptr+=width()/2;
crptr+=width()/2;
}
}
}
}
break;
// Compressed formats
case PIXELFORMAT_JPEG : break;
case PIXELFORMAT_MPEG : break;
// Reserved formats
case PIXELFORMAT_DV : break;
case PIXELFORMAT_ET61X251:break;
case PIXELFORMAT_HI240 : break;
case PIXELFORMAT_HM12 : break;
case PIXELFORMAT_MJPEG : break;
case PIXELFORMAT_PWC1 : break;
case PIXELFORMAT_PWC2 : break;
case PIXELFORMAT_SN9C10X:
{
unsigned char *s = new unsigned char [width() * height()];
unsigned char *d = new unsigned char [width() * height() * 3];
sonix_decompress_init();
sonix_decompress(width(), height(), &m_currentbuffer.data.first(), s);
bayer2rgb24(d, s, width(), height());
int step=0;
for(int loop=0;loop < qimage->numBytes();loop+=4)
{
bits[loop] = d[step+2];
bits[loop+1] = d[step+1];
bits[loop+2] = d[step];
bits[loop+3] = 255;
step+=3;
}
delete[] s;
delete[] d;
}
break;
case PIXELFORMAT_WNVA : break;
case PIXELFORMAT_YYUV : break;
}
return EXIT_SUCCESS;
}
/*!
\fn VideoDevice::stopCapturing()
*/
int VideoDevice::stopCapturing()
{
/// @todo implement me
kdDebug(14010) << k_funcinfo << "called." << endl;
if(isOpen())
{
switch (m_io_method)
{
case IO_METHOD_NONE: // Card cannot capture frames
return EXIT_FAILURE;
break;
case IO_METHOD_READ: // Nothing to do
break;
case IO_METHOD_MMAP:
case IO_METHOD_USERPTR:
#ifdef V4L2_CAP_VIDEO_CAPTURE
{
enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
if (-1 == xioctl (VIDIOC_STREAMOFF, &type))
return errnoReturn ("VIDIOC_STREAMOFF");
if (m_io_method == IO_METHOD_MMAP)
{
unsigned int loop;
for (loop = 0; loop < m_streambuffers; ++loop)
{
if (munmap(m_rawbuffers[loop].start,m_rawbuffers[loop].length) != 0)
{
kdDebug(14010) << k_funcinfo << "unable to munmap." << endl;
}
}
}
}
#endif
break;
}
kdDebug(14010) << k_funcinfo << "exited successfuly." << endl;
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
/*!
\fn VideoDevice::close()
*/
int VideoDevice::close()
{
/// @todo implement me
kdDebug(14010) << k_funcinfo << " called." << endl;
if(isOpen())
{
kdDebug(14010) << k_funcinfo << " Device is open. Trying to properly shutdown the device." << endl;
stopCapturing();
kdDebug(14010) << k_funcinfo << "::close() returns " << ::close(descriptor) << endl;
}
descriptor = -1;
return EXIT_SUCCESS;
}
float VideoDevice::getBrightness()
{
if (m_current_input < m_input.size() )
return m_input[m_current_input].getBrightness();
else
return 0;
}
float VideoDevice::setBrightness(float brightness)
{
kdDebug(14010) << k_funcinfo << " called." << endl;
m_input[m_current_input].setBrightness(brightness); // Just to check bounds
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
{
struct v4l2_queryctrl queryctrl;
struct v4l2_control control;
CLEAR (queryctrl);
queryctrl.id = V4L2_CID_BRIGHTNESS;
if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
{
if (errno != EINVAL)
{
kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl;
} else
{
kdDebug(14010) << k_funcinfo << "Device doesn't support the Brightness control." << endl;
}
} else
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
{
kdDebug(14010) << k_funcinfo << "Device doesn't support the Brightness control." << endl;
} else
{
CLEAR (control);
control.id = V4L2_CID_BRIGHTNESS;
control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getBrightness());
if (-1 == xioctl (VIDIOC_S_CTRL, &control))
{
kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl;
}
}
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
{
struct video_picture V4L_picture;
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
V4L_picture.brightness = uint(65535*getBrightness());
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image brightness. Fallback to it is not yet implemented." << endl;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
return getBrightness();
}
float VideoDevice::getContrast()
{
if (m_current_input < m_input.size() )
return m_input[m_current_input].getContrast();
else
return 0;
}
float VideoDevice::setContrast(float contrast)
{
kdDebug(14010) << k_funcinfo << " called." << endl;
m_input[m_current_input].setContrast(contrast); // Just to check bounds
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
{
struct v4l2_queryctrl queryctrl;
struct v4l2_control control;
CLEAR (queryctrl);
queryctrl.id = V4L2_CID_CONTRAST;
if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
{
if (errno != EINVAL)
{
kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl;
} else
{
kdDebug(14010) << k_funcinfo << "Device doesn't support the Contrast control." << endl;
}
} else
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
{
kdDebug(14010) << k_funcinfo << "Device doesn't support the Contrast control." << endl;
} else
{
CLEAR (control);
control.id = V4L2_CID_CONTRAST;
control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getContrast());
if (-1 == xioctl (VIDIOC_S_CTRL, &control))
{
kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl;
}
}
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
{
struct video_picture V4L_picture;
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
V4L_picture.contrast = uint(65535*getContrast());
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image contrast. Fallback to it is not yet implemented." << endl;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
return getContrast();
}
float VideoDevice::getSaturation()
{
if (m_current_input < m_input.size() )
return m_input[m_current_input].getSaturation();
else
return 0;
}
float VideoDevice::setSaturation(float saturation)
{
kdDebug(14010) << k_funcinfo << " called." << endl;
m_input[m_current_input].setSaturation(saturation); // Just to check bounds
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
{
struct v4l2_queryctrl queryctrl;
struct v4l2_control control;
CLEAR (queryctrl);
queryctrl.id = V4L2_CID_SATURATION;
if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
{
if (errno != EINVAL)
{
kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl;
} else
{
kdDebug(14010) << k_funcinfo << "Device doesn't support the Saturation control." << endl;
}
} else
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
{
kdDebug(14010) << k_funcinfo << "Device doesn't support the Saturation control." << endl;
} else
{
CLEAR (control);
control.id = V4L2_CID_SATURATION;
control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getSaturation());
if (-1 == xioctl (VIDIOC_S_CTRL, &control))
{
kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl;
}
}
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
{
struct video_picture V4L_picture;
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
V4L_picture.colour = uint(65535*getSaturation());
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image saturation. Fallback to it is not yet implemented." << endl;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
return getSaturation();
}
float VideoDevice::getWhiteness()
{
if (m_current_input < m_input.size() )
return m_input[m_current_input].getWhiteness();
else
return 0;
}
float VideoDevice::setWhiteness(float whiteness)
{
kdDebug(14010) << k_funcinfo << " called." << endl;
m_input[m_current_input].setWhiteness(whiteness); // Just to check bounds
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
{
struct v4l2_queryctrl queryctrl;
struct v4l2_control control;
CLEAR (queryctrl);
queryctrl.id = V4L2_CID_WHITENESS;
if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
{
if (errno != EINVAL)
{
kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl;
} else
{
kdDebug(14010) << k_funcinfo << "Device doesn't support the Whiteness control." << endl;
}
} else
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
{
kdDebug(14010) << k_funcinfo << "Device doesn't support the Whiteness control." << endl;
} else
{
CLEAR (control);
control.id = V4L2_CID_WHITENESS;
control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getWhiteness());
if (-1 == xioctl (VIDIOC_S_CTRL, &control))
{
kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl;
}
}
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
{
struct video_picture V4L_picture;
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
V4L_picture.whiteness = uint(65535*getWhiteness());
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting white level. Fallback to it is not yet implemented." << endl;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
return getWhiteness();
}
float VideoDevice::getHue()
{
if (m_current_input < m_input.size() )
return m_input[m_current_input].getHue();
else
return 0;
}
float VideoDevice::setHue(float hue)
{
kdDebug(14010) << k_funcinfo << " called." << endl;
m_input[m_current_input].setHue(hue); // Just to check bounds
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
{
struct v4l2_queryctrl queryctrl;
struct v4l2_control control;
CLEAR (queryctrl);
queryctrl.id = V4L2_CID_HUE;
if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl))
{
if (errno != EINVAL)
{
kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl;
} else
{
kdDebug(14010) << k_funcinfo << "Device doesn't support the Hue control." << endl;
}
} else
if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED)
{
kdDebug(14010) << k_funcinfo << "Device doesn't support the Hue control." << endl;
} else
{
CLEAR (control);
control.id = V4L2_CID_HUE;
control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getHue());
if (-1 == xioctl (VIDIOC_S_CTRL, &control))
{
kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl;
}
}
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
{
struct video_picture V4L_picture;
if(-1 == xioctl(VIDIOCGPICT, &V4L_picture))
kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl;
V4L_picture.hue = uint(65535*getHue());
if(-1 == xioctl(VIDIOCSPICT,&V4L_picture))
kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image hue. Fallback to it is not yet implemented." << endl;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
return getHue();
}
bool VideoDevice::getAutoBrightnessContrast()
{
if (m_current_input < m_input.size() )
return m_input[m_current_input].getAutoBrightnessContrast();
else
return false;
}
bool VideoDevice::setAutoBrightnessContrast(bool brightnesscontrast)
{
kdDebug(14010) << k_funcinfo << "VideoDevice::setAutoBrightnessContrast(" << brightnesscontrast << ") called." << endl;
if (m_current_input < m_input.size() )
{
m_input[m_current_input].setAutoBrightnessContrast(brightnesscontrast);
return m_input[m_current_input].getAutoBrightnessContrast();
}
else
return false;
}
bool VideoDevice::getAutoColorCorrection()
{
if (m_current_input < m_input.size() )
return m_input[m_current_input].getAutoColorCorrection();
else
return false;
}
bool VideoDevice::setAutoColorCorrection(bool colorcorrection)
{
kdDebug(14010) << k_funcinfo << "VideoDevice::setAutoColorCorrection(" << colorcorrection << ") called." << endl;
if (m_current_input < m_input.size() )
{
m_input[m_current_input].setAutoColorCorrection(colorcorrection);
return m_input[m_current_input].getAutoColorCorrection();
}
else
return false;
}
bool VideoDevice::getImageAsMirror()
{
if (m_current_input < m_input.size() )
return m_input[m_current_input].getImageAsMirror();
else
return false;
}
bool VideoDevice::setImageAsMirror(bool imageasmirror)
{
kdDebug(14010) << k_funcinfo << "VideoDevice::setImageAsMirror(" << imageasmirror << ") called." << endl;
if (m_current_input < m_input.size() )
{
m_input[m_current_input].setImageAsMirror(imageasmirror);
return m_input[m_current_input].getImageAsMirror();
}
else
return false;
}
pixel_format VideoDevice::pixelFormatForPalette( int palette )
{
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
switch(palette)
{
case 0 : return PIXELFORMAT_NONE; break;
// Packed RGB formats
case V4L2_PIX_FMT_RGB332 : return PIXELFORMAT_RGB332; break;
#if defined( V4L2_PIX_FMT_RGB444 )
case V4L2_PIX_FMT_RGB444 : return PIXELFORMAT_RGB444; break;
#endif
case V4L2_PIX_FMT_RGB555 : return PIXELFORMAT_RGB555; break;
case V4L2_PIX_FMT_RGB565 : return PIXELFORMAT_RGB565; break;
case V4L2_PIX_FMT_RGB555X : return PIXELFORMAT_RGB555X; break;
case V4L2_PIX_FMT_RGB565X : return PIXELFORMAT_RGB565X; break;
case V4L2_PIX_FMT_BGR24 : return PIXELFORMAT_BGR24; break;
case V4L2_PIX_FMT_RGB24 : return PIXELFORMAT_RGB24; break;
case V4L2_PIX_FMT_BGR32 : return PIXELFORMAT_BGR32; break;
case V4L2_PIX_FMT_RGB32 : return PIXELFORMAT_RGB32; break;
// Bayer RGB format
case V4L2_PIX_FMT_SBGGR8 : return PIXELFORMAT_SBGGR8; break;
// YUV formats
case V4L2_PIX_FMT_GREY : return PIXELFORMAT_GREY; break;
case V4L2_PIX_FMT_YUYV : return PIXELFORMAT_YUYV; break;
case V4L2_PIX_FMT_UYVY : return PIXELFORMAT_UYVY; break;
case V4L2_PIX_FMT_YUV420 : return PIXELFORMAT_YUV420P; break;
case V4L2_PIX_FMT_YUV422P : return PIXELFORMAT_YUV422P; break;
// Compressed formats
case V4L2_PIX_FMT_JPEG : return PIXELFORMAT_JPEG; break;
case V4L2_PIX_FMT_MPEG : return PIXELFORMAT_MPEG; break;
// Reserved formats
case V4L2_PIX_FMT_DV : return PIXELFORMAT_DV; break;
case V4L2_PIX_FMT_ET61X251 : return PIXELFORMAT_ET61X251; break;
case V4L2_PIX_FMT_HI240 : return PIXELFORMAT_HI240; break;
#if defined( V4L2_PIX_FMT_HM12 )
case V4L2_PIX_FMT_HM12 : return PIXELFORMAT_HM12; break;
#endif
case V4L2_PIX_FMT_MJPEG : return PIXELFORMAT_MJPEG; break;
case V4L2_PIX_FMT_PWC1 : return PIXELFORMAT_PWC1; break;
case V4L2_PIX_FMT_PWC2 : return PIXELFORMAT_PWC2; break;
case V4L2_PIX_FMT_SN9C10X : return PIXELFORMAT_SN9C10X; break;
case V4L2_PIX_FMT_WNVA : return PIXELFORMAT_WNVA; break;
case V4L2_PIX_FMT_YYUV : return PIXELFORMAT_YYUV; break;
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
switch(palette)
{
case 0 : return PIXELFORMAT_NONE; break;
case VIDEO_PALETTE_GREY : return PIXELFORMAT_GREY; break;
case VIDEO_PALETTE_HI240 : return PIXELFORMAT_RGB332; break;
case VIDEO_PALETTE_RGB555 : return PIXELFORMAT_RGB555; break;
case VIDEO_PALETTE_RGB565 : return PIXELFORMAT_RGB565; break;
case VIDEO_PALETTE_RGB24 : return PIXELFORMAT_RGB24; break;
case VIDEO_PALETTE_RGB32 : return PIXELFORMAT_RGB32; break;
case VIDEO_PALETTE_YUYV : return PIXELFORMAT_YUYV; break;
case VIDEO_PALETTE_UYVY : return PIXELFORMAT_UYVY; break;
case VIDEO_PALETTE_YUV420 :
case VIDEO_PALETTE_YUV420P : return PIXELFORMAT_YUV420P; break;
case VIDEO_PALETTE_YUV422P : return PIXELFORMAT_YUV422P; break;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
return PIXELFORMAT_NONE; break;
}
return PIXELFORMAT_NONE;
}
int VideoDevice::pixelFormatCode(pixel_format pixelformat)
{
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
switch(pixelformat)
{
case PIXELFORMAT_NONE : return 0; break;
// Packed RGB formats
case PIXELFORMAT_RGB332 : return V4L2_PIX_FMT_RGB332; break;
#if defined( V4L2_PIX_FMT_RGB444 )
case PIXELFORMAT_RGB444 : return V4L2_PIX_FMT_RGB444; break;
#endif
case PIXELFORMAT_RGB555 : return V4L2_PIX_FMT_RGB555; break;
case PIXELFORMAT_RGB565 : return V4L2_PIX_FMT_RGB565; break;
case PIXELFORMAT_RGB555X: return V4L2_PIX_FMT_RGB555X; break;
case PIXELFORMAT_RGB565X: return V4L2_PIX_FMT_RGB565X; break;
case PIXELFORMAT_BGR24 : return V4L2_PIX_FMT_BGR24; break;
case PIXELFORMAT_RGB24 : return V4L2_PIX_FMT_RGB24; break;
case PIXELFORMAT_BGR32 : return V4L2_PIX_FMT_BGR32; break;
case PIXELFORMAT_RGB32 : return V4L2_PIX_FMT_RGB32; break;
// Bayer RGB format
case PIXELFORMAT_SBGGR8 : return V4L2_PIX_FMT_SBGGR8; break;
// YUV formats
case PIXELFORMAT_GREY : return V4L2_PIX_FMT_GREY; break;
case PIXELFORMAT_YUYV : return V4L2_PIX_FMT_YUYV; break;
case PIXELFORMAT_UYVY : return V4L2_PIX_FMT_UYVY; break;
case PIXELFORMAT_YUV420P: return V4L2_PIX_FMT_YUV420; break;
case PIXELFORMAT_YUV422P: return V4L2_PIX_FMT_YUV422P; break;
// Compressed formats
case PIXELFORMAT_JPEG : return V4L2_PIX_FMT_JPEG; break;
case PIXELFORMAT_MPEG : return V4L2_PIX_FMT_MPEG; break;
// Reserved formats
case PIXELFORMAT_DV : return V4L2_PIX_FMT_DV; break;
case PIXELFORMAT_ET61X251:return V4L2_PIX_FMT_ET61X251;break;
case PIXELFORMAT_HI240 : return V4L2_PIX_FMT_HI240; break;
#if defined( V4L2_PIX_FMT_HM12 )
case PIXELFORMAT_HM12 : return V4L2_PIX_FMT_HM12; break;
#endif
case PIXELFORMAT_MJPEG : return V4L2_PIX_FMT_MJPEG; break;
case PIXELFORMAT_PWC1 : return V4L2_PIX_FMT_PWC1; break;
case PIXELFORMAT_PWC2 : return V4L2_PIX_FMT_PWC2; break;
case PIXELFORMAT_SN9C10X: return V4L2_PIX_FMT_SN9C10X; break;
case PIXELFORMAT_WNVA : return V4L2_PIX_FMT_WNVA; break;
case PIXELFORMAT_YYUV : return V4L2_PIX_FMT_YYUV; break;
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
switch(pixelformat)
{
case PIXELFORMAT_NONE : return 0; break;
// Packed RGB formats
case PIXELFORMAT_RGB332 : return VIDEO_PALETTE_HI240; break;
case PIXELFORMAT_RGB444 : return 0; break;
case PIXELFORMAT_RGB555 : return VIDEO_PALETTE_RGB555; break;
case PIXELFORMAT_RGB565 : return VIDEO_PALETTE_RGB565; break;
case PIXELFORMAT_RGB555X: return 0; break;
case PIXELFORMAT_RGB565X: return 0; break;
case PIXELFORMAT_BGR24 : return 0; break;
case PIXELFORMAT_RGB24 : return VIDEO_PALETTE_RGB24; break;
case PIXELFORMAT_BGR32 : return 0; break;
case PIXELFORMAT_RGB32 : return VIDEO_PALETTE_RGB32; break;
// Bayer RGB format
case PIXELFORMAT_SBGGR8 : return 0; break;
// YUV formats
case PIXELFORMAT_GREY : return VIDEO_PALETTE_GREY; break;
case PIXELFORMAT_YUYV : return VIDEO_PALETTE_YUYV; break;
case PIXELFORMAT_UYVY : return VIDEO_PALETTE_UYVY; break;
case PIXELFORMAT_YUV420P: return VIDEO_PALETTE_YUV420; break;
case PIXELFORMAT_YUV422P: return VIDEO_PALETTE_YUV422P; break;
// Compressed formats
case PIXELFORMAT_JPEG : return 0; break;
case PIXELFORMAT_MPEG : return 0; break;
// Reserved formats
case PIXELFORMAT_DV : return 0; break;
case PIXELFORMAT_ET61X251:return 0; break;
case PIXELFORMAT_HI240 : return VIDEO_PALETTE_HI240; break;
case PIXELFORMAT_HM12 : return 0; break;
case PIXELFORMAT_MJPEG : return 0; break;
case PIXELFORMAT_PWC1 : return 0; break;
case PIXELFORMAT_PWC2 : return 0; break;
case PIXELFORMAT_SN9C10X: return 0; break;
case PIXELFORMAT_WNVA : return 0; break;
case PIXELFORMAT_YYUV : return 0; break;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
return PIXELFORMAT_NONE; break;
}
return PIXELFORMAT_NONE;
}
int VideoDevice::pixelFormatDepth(pixel_format pixelformat)
{
switch(pixelformat)
{
case PIXELFORMAT_NONE : return 0; break;
// Packed RGB formats
case PIXELFORMAT_RGB332 : return 8; break;
case PIXELFORMAT_RGB444 : return 16; break;
case PIXELFORMAT_RGB555 : return 16; break;
case PIXELFORMAT_RGB565 : return 16; break;
case PIXELFORMAT_RGB555X: return 16; break;
case PIXELFORMAT_RGB565X: return 16; break;
case PIXELFORMAT_BGR24 : return 24; break;
case PIXELFORMAT_RGB24 : return 24; break;
case PIXELFORMAT_BGR32 : return 32; break;
case PIXELFORMAT_RGB32 : return 32; break;
// Bayer RGB format
case PIXELFORMAT_SBGGR8 : return 0; break;
// YUV formats
case PIXELFORMAT_GREY : return 8; break;
case PIXELFORMAT_YUYV : return 16; break;
case PIXELFORMAT_UYVY : return 16; break;
case PIXELFORMAT_YUV420P: return 16; break;
case PIXELFORMAT_YUV422P: return 16; break;
// Compressed formats
case PIXELFORMAT_JPEG : return 0; break;
case PIXELFORMAT_MPEG : return 0; break;
// Reserved formats
case PIXELFORMAT_DV : return 0; break;
case PIXELFORMAT_ET61X251:return 0; break;
case PIXELFORMAT_HI240 : return 8; break;
case PIXELFORMAT_HM12 : return 0; break;
case PIXELFORMAT_MJPEG : return 0; break;
case PIXELFORMAT_PWC1 : return 0; break;
case PIXELFORMAT_PWC2 : return 0; break;
case PIXELFORMAT_SN9C10X: return 0; break;
case PIXELFORMAT_WNVA : return 0; break;
case PIXELFORMAT_YYUV : return 0; break;
}
return 0;
}
TQString VideoDevice::pixelFormatName(pixel_format pixelformat)
{
TQString returnvalue;
returnvalue = "None";
switch(pixelformat)
{
case PIXELFORMAT_NONE : returnvalue = "None"; break;
// Packed RGB formats
case PIXELFORMAT_RGB332 : returnvalue = "8-bit RGB332"; break;
case PIXELFORMAT_RGB444 : returnvalue = "8-bit RGB444"; break;
case PIXELFORMAT_RGB555 : returnvalue = "16-bit RGB555"; break;
case PIXELFORMAT_RGB565 : returnvalue = "16-bit RGB565"; break;
case PIXELFORMAT_RGB555X: returnvalue = "16-bit RGB555X"; break;
case PIXELFORMAT_RGB565X: returnvalue = "16-bit RGB565X"; break;
case PIXELFORMAT_BGR24 : returnvalue = "24-bit BGR24"; break;
case PIXELFORMAT_RGB24 : returnvalue = "24-bit RGB24"; break;
case PIXELFORMAT_BGR32 : returnvalue = "32-bit BGR32"; break;
case PIXELFORMAT_RGB32 : returnvalue = "32-bit RGB32"; break;
// Bayer RGB format
case PIXELFORMAT_SBGGR8 : returnvalue = "Bayer RGB format"; break;
// YUV formats
case PIXELFORMAT_GREY : returnvalue = "8-bit Grayscale"; break;
case PIXELFORMAT_YUYV : returnvalue = "Packed YUV 4:2:2"; break;
case PIXELFORMAT_UYVY : returnvalue = "Packed YVU 4:2:2"; break;
case PIXELFORMAT_YUV420P: returnvalue = "Planar YUV 4:2:0"; break;
case PIXELFORMAT_YUV422P: returnvalue = "Planar YUV 4:2:2"; break;
// Compressed formats
case PIXELFORMAT_JPEG : returnvalue = "JPEG image"; break;
case PIXELFORMAT_MPEG : returnvalue = "MPEG stream"; break;
// Reserved formats
case PIXELFORMAT_DV : returnvalue = "DV (unknown)"; break;
case PIXELFORMAT_ET61X251:returnvalue = "ET61X251"; break;
case PIXELFORMAT_HI240 : returnvalue = "8-bit HI240 (RGB332)"; break;
case PIXELFORMAT_HM12 : returnvalue = "Packed YUV 4:2:2"; break;
case PIXELFORMAT_MJPEG : returnvalue = "8-bit Grayscale"; break;
case PIXELFORMAT_PWC1 : returnvalue = "PWC1"; break;
case PIXELFORMAT_PWC2 : returnvalue = "PWC2"; break;
case PIXELFORMAT_SN9C10X: returnvalue = "SN9C102"; break;
case PIXELFORMAT_WNVA : returnvalue = "Winnov Videum"; break;
case PIXELFORMAT_YYUV : returnvalue = "YYUV (unknown)"; break;
}
return returnvalue;
}
TQString VideoDevice::pixelFormatName(int pixelformat)
{
TQString returnvalue;
returnvalue = "None";
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
switch(pixelformat)
{
case 0 : returnvalue = pixelFormatName(PIXELFORMAT_NONE); break;
// Packed RGB formats
case V4L2_PIX_FMT_RGB332 : returnvalue = pixelFormatName(PIXELFORMAT_RGB332); break;
#if defined( V4L2_PIX_FMT_RGB444 )
case V4L2_PIX_FMT_RGB444 : returnvalue = pixelFormatName(PIXELFORMAT_RGB444); break;
#endif
case V4L2_PIX_FMT_RGB555 : returnvalue = pixelFormatName(PIXELFORMAT_RGB555); break;
case V4L2_PIX_FMT_RGB565 : returnvalue = pixelFormatName(PIXELFORMAT_RGB565); break;
case V4L2_PIX_FMT_RGB555X : returnvalue = pixelFormatName(PIXELFORMAT_RGB555X); break;
case V4L2_PIX_FMT_RGB565X : returnvalue = pixelFormatName(PIXELFORMAT_RGB565X); break;
case V4L2_PIX_FMT_BGR24 : returnvalue = pixelFormatName(PIXELFORMAT_BGR24); break;
case V4L2_PIX_FMT_RGB24 : returnvalue = pixelFormatName(PIXELFORMAT_RGB24); break;
case V4L2_PIX_FMT_BGR32 : returnvalue = pixelFormatName(PIXELFORMAT_BGR32); break;
case V4L2_PIX_FMT_RGB32 : returnvalue = pixelFormatName(PIXELFORMAT_RGB32); break;
// Bayer RGB format
case V4L2_PIX_FMT_SBGGR8 : returnvalue = pixelFormatName(PIXELFORMAT_SBGGR8); break;
// YUV formats
case V4L2_PIX_FMT_GREY : returnvalue = pixelFormatName(PIXELFORMAT_GREY); break;
case V4L2_PIX_FMT_YUYV : returnvalue = pixelFormatName(PIXELFORMAT_YUYV); break;
case V4L2_PIX_FMT_UYVY : returnvalue = pixelFormatName(PIXELFORMAT_UYVY); break;
case V4L2_PIX_FMT_YUV420 : returnvalue = pixelFormatName(PIXELFORMAT_YUV420P); break;
case V4L2_PIX_FMT_YUV422P : returnvalue = pixelFormatName(PIXELFORMAT_YUV422P); break;
// Compressed formats
case V4L2_PIX_FMT_JPEG : returnvalue = pixelFormatName(PIXELFORMAT_JPEG); break;
case V4L2_PIX_FMT_MPEG : returnvalue = pixelFormatName(PIXELFORMAT_MPEG); break;
// Reserved formats
case V4L2_PIX_FMT_DV : returnvalue = pixelFormatName(PIXELFORMAT_DV); break;
case V4L2_PIX_FMT_ET61X251 : returnvalue = pixelFormatName(PIXELFORMAT_ET61X251); break;
case V4L2_PIX_FMT_HI240 : returnvalue = pixelFormatName(PIXELFORMAT_HI240); break;
#if defined( V4L2_PIX_FMT_HM12 )
case V4L2_PIX_FMT_HM12 : returnvalue = pixelFormatName(PIXELFORMAT_HM12); break;
#endif
case V4L2_PIX_FMT_MJPEG : returnvalue = pixelFormatName(PIXELFORMAT_MJPEG); break;
case V4L2_PIX_FMT_PWC1 : returnvalue = pixelFormatName(PIXELFORMAT_PWC1); break;
case V4L2_PIX_FMT_PWC2 : returnvalue = pixelFormatName(PIXELFORMAT_PWC2); break;
case V4L2_PIX_FMT_SN9C10X : returnvalue = pixelFormatName(PIXELFORMAT_SN9C10X); break;
case V4L2_PIX_FMT_WNVA : returnvalue = pixelFormatName(PIXELFORMAT_WNVA); break;
case V4L2_PIX_FMT_YYUV : returnvalue = pixelFormatName(PIXELFORMAT_YYUV); break;
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
switch(pixelformat)
{
case VIDEO_PALETTE_GREY : returnvalue = pixelFormatName(PIXELFORMAT_GREY); break;
case VIDEO_PALETTE_HI240 : returnvalue = pixelFormatName(PIXELFORMAT_RGB332); break;
case VIDEO_PALETTE_RGB555 : returnvalue = pixelFormatName(PIXELFORMAT_RGB555); break;
case VIDEO_PALETTE_RGB565 : returnvalue = pixelFormatName(PIXELFORMAT_RGB565); break;
case VIDEO_PALETTE_RGB24 : returnvalue = pixelFormatName(PIXELFORMAT_RGB24); break;
case VIDEO_PALETTE_RGB32 : returnvalue = pixelFormatName(PIXELFORMAT_RGB32); break;
case VIDEO_PALETTE_YUYV : returnvalue = pixelFormatName(PIXELFORMAT_YUYV); break;
case VIDEO_PALETTE_UYVY : returnvalue = pixelFormatName(PIXELFORMAT_UYVY); break;
case VIDEO_PALETTE_YUV420 :
case VIDEO_PALETTE_YUV420P : returnvalue = pixelFormatName(PIXELFORMAT_YUV420P); break;
case VIDEO_PALETTE_YUV422P : returnvalue = pixelFormatName(PIXELFORMAT_YUV422P); break;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
return returnvalue;
}
int VideoDevice::detectPixelFormats()
{
int err = 0;
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
fmtdesc.index = 0;
fmtdesc.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
while ( err == 0 )
{
if (-1 == xioctl (VIDIOC_ENUM_FMT, &fmtdesc))
// if (ioctl(fd, VIDIOC_ENUM_FMT, &fmtdesc) < 0 )
{
perror("VIDIOC_ENUM_FMT");
err = errno;
}
else
{
kdDebug(14010) << k_funcinfo << fmtdesc.pixelformat << " " << pixelFormatName(fmtdesc.pixelformat) << endl; // Need a cleanup. PixelFormatForPalette is a really bad name
fmtdesc.index++;
}
}
// break;
#endif
case VIDEODEV_DRIVER_V4L:
// TODO: THis thing can be used to detec what pixel formats are supported in a API-independent way, but V4L2 has VIDIOC_ENUM_PIXFMT.
// The correct thing to do is to isolate these calls and do a proper implementation for V4L and another for V4L2 when this thing will be migrated to a plugin architecture.
// Packed RGB formats
kdDebug(14010) << k_funcinfo << "Supported pixel formats:" << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB332)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB332) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB444)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB444) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB555)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB555) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB565)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB565) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB555X)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB555X) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB565X)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB565X) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_BGR24)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_BGR24) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB24)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB24) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_BGR32)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_BGR32) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB32)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB32) << endl;
// Bayer RGB format
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_SBGGR8)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_SBGGR8) << endl;
// YUV formats
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_GREY)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_GREY) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YUYV)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YUYV) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_UYVY)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_UYVY) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YUV420P)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YUV420P) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YUV422P)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YUV422P) << endl;
// Compressed formats
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_JPEG)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_JPEG) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_MPEG)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_MPEG) << endl;
// Reserved formats
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_DV)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_DV) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_ET61X251)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_ET61X251) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_HI240)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_HI240) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_HM12)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_HM12) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_MJPEG)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_MJPEG) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_PWC1)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_PWC1) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_PWC2)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_PWC2) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_SN9C10X)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_SN9C10X) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_WNVA)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_WNVA) << endl;
if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YYUV)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YYUV) << endl;
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
return PIXELFORMAT_NONE; break;
}
return PIXELFORMAT_NONE;
}
__u64 VideoDevice::signalStandardCode(signal_standard standard)
{
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
switch(standard)
{
case STANDARD_NONE : return V4L2_STD_UNKNOWN; break;
case STANDARD_PAL_B : return V4L2_STD_PAL_B; break;
case STANDARD_PAL_B1 : return V4L2_STD_PAL_B1; break;
case STANDARD_PAL_G : return V4L2_STD_PAL_G; break;
case STANDARD_PAL_H : return V4L2_STD_PAL_H; break;
case STANDARD_PAL_I : return V4L2_STD_PAL_I; break;
case STANDARD_PAL_D : return V4L2_STD_PAL_D; break;
case STANDARD_PAL_D1 : return V4L2_STD_PAL_D1; break;
case STANDARD_PAL_K : return V4L2_STD_PAL_K; break;
case STANDARD_PAL_M : return V4L2_STD_PAL_M; break;
case STANDARD_PAL_N : return V4L2_STD_PAL_N; break;
case STANDARD_PAL_Nc : return V4L2_STD_PAL_Nc; break;
case STANDARD_PAL_60 : return V4L2_STD_PAL_60; break;
case STANDARD_NTSC_M : return V4L2_STD_NTSC_M; break;
case STANDARD_NTSC_M_JP : return V4L2_STD_NTSC_M_JP; break;
case STANDARD_NTSC_443 : return V4L2_STD_NTSC; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
case STANDARD_SECAM_B : return V4L2_STD_SECAM_B; break;
case STANDARD_SECAM_D : return V4L2_STD_SECAM_D; break;
case STANDARD_SECAM_G : return V4L2_STD_SECAM_G; break;
case STANDARD_SECAM_H : return V4L2_STD_SECAM_H; break;
case STANDARD_SECAM_K : return V4L2_STD_SECAM_K; break;
case STANDARD_SECAM_K1 : return V4L2_STD_SECAM_K1; break;
case STANDARD_SECAM_L : return V4L2_STD_SECAM_L; break;
case STANDARD_SECAM_LC : return V4L2_STD_SECAM; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
case STANDARD_ATSC_8_VSB : return V4L2_STD_ATSC_8_VSB; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
case STANDARD_ATSC_16_VSB : return V4L2_STD_ATSC_16_VSB; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
case STANDARD_PAL_BG : return V4L2_STD_PAL_BG; break;
case STANDARD_PAL_DK : return V4L2_STD_PAL_DK; break;
case STANDARD_PAL : return V4L2_STD_PAL; break;
case STANDARD_NTSC : return V4L2_STD_NTSC; break;
case STANDARD_SECAM_DK : return V4L2_STD_SECAM_DK; break;
case STANDARD_SECAM : return V4L2_STD_SECAM; break;
case STANDARD_525_60 : return V4L2_STD_525_60; break;
case STANDARD_625_50 : return V4L2_STD_625_50; break;
case STANDARD_ALL : return V4L2_STD_ALL; break;
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
switch(standard)
{
case STANDARD_NONE : return VIDEO_MODE_AUTO; break;
case STANDARD_PAL_B : return VIDEO_MODE_PAL; break;
case STANDARD_PAL_B1 : return VIDEO_MODE_PAL; break;
case STANDARD_PAL_G : return VIDEO_MODE_PAL; break;
case STANDARD_PAL_H : return VIDEO_MODE_PAL; break;
case STANDARD_PAL_I : return VIDEO_MODE_PAL; break;
case STANDARD_PAL_D : return VIDEO_MODE_PAL; break;
case STANDARD_PAL_D1 : return VIDEO_MODE_PAL; break;
case STANDARD_PAL_K : return VIDEO_MODE_PAL; break;
case STANDARD_PAL_M : return 5; break; // Undocumented value found to be compatible with V4L bttv driver
case STANDARD_PAL_N : return 6; break; // Undocumented value found to be compatible with V4L bttv driver
case STANDARD_PAL_Nc : return 4; break; // Undocumented value found to be compatible with V4L bttv driver
case STANDARD_PAL_60 : return VIDEO_MODE_PAL; break;
case STANDARD_NTSC_M : return VIDEO_MODE_NTSC; break;
case STANDARD_NTSC_M_JP : return 7; break; // Undocumented value found to be compatible with V4L bttv driver
case STANDARD_NTSC_443 : return VIDEO_MODE_NTSC; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
case STANDARD_SECAM_B : return VIDEO_MODE_SECAM; break;
case STANDARD_SECAM_D : return VIDEO_MODE_SECAM; break;
case STANDARD_SECAM_G : return VIDEO_MODE_SECAM; break;
case STANDARD_SECAM_H : return VIDEO_MODE_SECAM; break;
case STANDARD_SECAM_K : return VIDEO_MODE_SECAM; break;
case STANDARD_SECAM_K1 : return VIDEO_MODE_SECAM; break;
case STANDARD_SECAM_L : return VIDEO_MODE_SECAM; break;
case STANDARD_SECAM_LC : return VIDEO_MODE_SECAM; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
case STANDARD_ATSC_8_VSB : return VIDEO_MODE_AUTO; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
case STANDARD_ATSC_16_VSB : return VIDEO_MODE_AUTO; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
case STANDARD_PAL_BG : return VIDEO_MODE_PAL; break;
case STANDARD_PAL_DK : return VIDEO_MODE_PAL; break;
case STANDARD_PAL : return VIDEO_MODE_PAL; break;
case STANDARD_NTSC : return VIDEO_MODE_NTSC; break;
case STANDARD_SECAM_DK : return VIDEO_MODE_SECAM; break;
case STANDARD_SECAM : return VIDEO_MODE_SECAM; break;
case STANDARD_525_60 : return VIDEO_MODE_PAL; break;
case STANDARD_625_50 : return VIDEO_MODE_SECAM; break;
case STANDARD_ALL : return VIDEO_MODE_AUTO; break;
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
return STANDARD_NONE; break;
}
return STANDARD_NONE;
}
TQString VideoDevice::signalStandardName(signal_standard standard)
{
TQString returnvalue;
returnvalue = "None";
switch(standard)
{
case STANDARD_NONE : returnvalue = "None"; break;
case STANDARD_PAL_B : returnvalue = "PAL-B"; break;
case STANDARD_PAL_B1 : returnvalue = "PAL-B1"; break;
case STANDARD_PAL_G : returnvalue = "PAL-G"; break;
case STANDARD_PAL_H : returnvalue = "PAL-H"; break;
case STANDARD_PAL_I : returnvalue = "PAL-I"; break;
case STANDARD_PAL_D : returnvalue = "PAL-D"; break;
case STANDARD_PAL_D1 : returnvalue = "PAL-D1"; break;
case STANDARD_PAL_K : returnvalue = "PAL-K"; break;
case STANDARD_PAL_M : returnvalue = "PAL-M"; break;
case STANDARD_PAL_N : returnvalue = "PAL-N"; break;
case STANDARD_PAL_Nc : returnvalue = "PAL-Nc"; break;
case STANDARD_PAL_60 : returnvalue = "PAL-60"; break;
case STANDARD_NTSC_M : returnvalue = "NTSC-M"; break;
case STANDARD_NTSC_M_JP : returnvalue = "NTSC-M(JP)"; break;
case STANDARD_NTSC_443 : returnvalue = "NTSC-443"; break;
case STANDARD_SECAM_B : returnvalue = "SECAM-B"; break;
case STANDARD_SECAM_D : returnvalue = "SECAM-D"; break;
case STANDARD_SECAM_G : returnvalue = "SECAM-G"; break;
case STANDARD_SECAM_H : returnvalue = "SECAM-H"; break;
case STANDARD_SECAM_K : returnvalue = "SECAM-K"; break;
case STANDARD_SECAM_K1 : returnvalue = "SECAM-K1"; break;
case STANDARD_SECAM_L : returnvalue = "SECAM-L"; break;
case STANDARD_SECAM_LC : returnvalue = "SECAM-LC"; break;
case STANDARD_ATSC_8_VSB : returnvalue = "ATSC-8-VSB"; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
case STANDARD_ATSC_16_VSB : returnvalue = "ATSC-16-VSB"; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
case STANDARD_PAL_BG : returnvalue = "PAL-BG"; break;
case STANDARD_PAL_DK : returnvalue = "PAL-DK"; break;
case STANDARD_PAL : returnvalue = "PAL"; break;
case STANDARD_NTSC : returnvalue = "NTSC"; break;
case STANDARD_SECAM_DK : returnvalue = "SECAM-DK"; break;
case STANDARD_SECAM : returnvalue = "SECAM"; break;
case STANDARD_525_60 : returnvalue = "525 lines 60Hz"; break;
case STANDARD_625_50 : returnvalue = "625 lines 50Hz"; break;
case STANDARD_ALL : returnvalue = "All"; break;
}
return returnvalue;
}
TQString VideoDevice::signalStandardName(int standard)
{
TQString returnvalue;
returnvalue = "None";
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
switch(standard)
{
case V4L2_STD_PAL_B : returnvalue = signalStandardName(STANDARD_PAL_B); break;
case V4L2_STD_PAL_B1 : returnvalue = signalStandardName(STANDARD_PAL_B1); break;
case V4L2_STD_PAL_G : returnvalue = signalStandardName(STANDARD_PAL_G); break;
case V4L2_STD_PAL_H : returnvalue = signalStandardName(STANDARD_PAL_H); break;
case V4L2_STD_PAL_I : returnvalue = signalStandardName(STANDARD_PAL_I); break;
case V4L2_STD_PAL_D : returnvalue = signalStandardName(STANDARD_PAL_D); break;
case V4L2_STD_PAL_D1 : returnvalue = signalStandardName(STANDARD_PAL_D1); break;
case V4L2_STD_PAL_K : returnvalue = signalStandardName(STANDARD_PAL_K); break;
case V4L2_STD_PAL_M : returnvalue = signalStandardName(STANDARD_PAL_M); break;
case V4L2_STD_PAL_N : returnvalue = signalStandardName(STANDARD_PAL_N); break;
case V4L2_STD_PAL_Nc : returnvalue = signalStandardName(STANDARD_PAL_Nc); break;
case V4L2_STD_PAL_60 : returnvalue = signalStandardName(STANDARD_PAL_60); break;
case V4L2_STD_NTSC_M : returnvalue = signalStandardName(STANDARD_NTSC_M); break;
case V4L2_STD_NTSC_M_JP : returnvalue = signalStandardName(STANDARD_NTSC_M_JP); break;
// case V4L2_STD_NTSC_443 : returnvalue = signalStandardName(STANDARD_NTSC_443); break; // Commented out because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
case V4L2_STD_SECAM_B : returnvalue = signalStandardName(STANDARD_SECAM_B); break;
case V4L2_STD_SECAM_D : returnvalue = signalStandardName(STANDARD_SECAM_D); break;
case V4L2_STD_SECAM_G : returnvalue = signalStandardName(STANDARD_SECAM_G); break;
case V4L2_STD_SECAM_H : returnvalue = signalStandardName(STANDARD_SECAM_H); break;
case V4L2_STD_SECAM_K : returnvalue = signalStandardName(STANDARD_SECAM_K); break;
case V4L2_STD_SECAM_K1 : returnvalue = signalStandardName(STANDARD_SECAM_K1); break;
case V4L2_STD_SECAM_L : returnvalue = signalStandardName(STANDARD_SECAM_L); break;
// case V4L2_STD_SECAM_LC : returnvalue = signalStandardName(STANDARD_SECAM_LC); break; // Commented out because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id
case V4L2_STD_ATSC_8_VSB : returnvalue = signalStandardName(STANDARD_ATSC_8_VSB); break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
case V4L2_STD_ATSC_16_VSB : returnvalue = signalStandardName(STANDARD_ATSC_16_VSB); break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h
case V4L2_STD_PAL_BG : returnvalue = signalStandardName(STANDARD_PAL_BG); break;
case V4L2_STD_PAL_DK : returnvalue = signalStandardName(STANDARD_PAL_DK); break;
case V4L2_STD_PAL : returnvalue = signalStandardName(STANDARD_PAL); break;
case V4L2_STD_NTSC : returnvalue = signalStandardName(STANDARD_NTSC); break;
case V4L2_STD_SECAM_DK : returnvalue = signalStandardName(STANDARD_SECAM_DK); break;
case V4L2_STD_SECAM : returnvalue = signalStandardName(STANDARD_SECAM); break;
case V4L2_STD_525_60 : returnvalue = signalStandardName(STANDARD_525_60); break;
case V4L2_STD_625_50 : returnvalue = signalStandardName(STANDARD_625_50); break;
case V4L2_STD_ALL : returnvalue = signalStandardName(STANDARD_ALL); break;
}
break;
#endif
case VIDEODEV_DRIVER_V4L:
switch(standard)
{
case VIDEO_MODE_PAL : returnvalue = signalStandardName(STANDARD_PAL); break;
case VIDEO_MODE_NTSC : returnvalue = signalStandardName(STANDARD_NTSC); break;
case VIDEO_MODE_SECAM : returnvalue = signalStandardName(STANDARD_SECAM); break;
case VIDEO_MODE_AUTO : returnvalue = signalStandardName(STANDARD_ALL); break; // It must be disabled until I find a correct way to handle those non-standard bttv modes
// case VIDEO_MODE_PAL_Nc : returnvalue = signalStandardName(STANDARD_PAL_Nc); break; // Undocumented value found to be compatible with V4L bttv driver
case VIDEO_MODE_PAL_M : returnvalue = signalStandardName(STANDARD_PAL_M); break; // Undocumented value found to be compatible with V4L bttv driver
case VIDEO_MODE_PAL_N : returnvalue = signalStandardName(STANDARD_PAL_N); break; // Undocumented value found to be compatible with V4L bttv driver
case VIDEO_MODE_NTSC_JP : returnvalue = signalStandardName(STANDARD_NTSC_M_JP); break; // Undocumented value found to be compatible with V4L bttv driver
}
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
return returnvalue;
}
/*!
\fn VideoDevice::detectSignalStandards()
*/
int VideoDevice::detectSignalStandards()
{
switch(m_driver)
{
#if defined(__linux__) && defined(ENABLE_AV)
#ifdef V4L2_CAP_VIDEO_CAPTURE
case VIDEODEV_DRIVER_V4L2:
break;
#endif
case VIDEODEV_DRIVER_V4L:
break;
#endif
case VIDEODEV_DRIVER_NONE:
default:
break;
}
//FIXME: return a real value
return 0;
}
/*!
\fn VideoDevice::initRead()
*/
int VideoDevice::initRead()
{
/// @todo implement me
kdDebug(14010) << k_funcinfo << "called." << endl;
if(isOpen())
{
m_rawbuffers.resize(1);
if (m_rawbuffers.size()==0)
{
fprintf (stderr, "Out of memory\n");
return EXIT_FAILURE;
}
kdDebug(14010) << k_funcinfo << "m_buffer_size: " << m_buffer_size << endl;
// m_rawbuffers[0].pixelformat=m_pixelformat;
m_rawbuffers[0].length = m_buffer_size;
m_rawbuffers[0].start = (uchar *)malloc (m_buffer_size);
if (!m_rawbuffers[0].start)
{
fprintf (stderr, "Out of memory\n");
return EXIT_FAILURE;
}
kdDebug(14010) << k_funcinfo << "exited successfuly." << endl;
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
/*!
\fn VideoDevice::initMmap()
*/
int VideoDevice::initMmap()
{
/// @todo implement me
#define BUFFERS 2
if(isOpen())
{
kdDebug(14010) << k_funcinfo << full_filename << " Trying to MMAP" << endl;
#ifdef V4L2_CAP_VIDEO_CAPTURE
struct v4l2_requestbuffers req;
CLEAR (req);
req.count = BUFFERS;
req.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
req.memory = V4L2_MEMORY_MMAP;
if (-1 == xioctl (VIDIOC_REQBUFS, &req))
{
if (EINVAL == errno)
{
kdDebug(14010) << k_funcinfo << full_filename << " does not support memory mapping" << endl;
return EXIT_FAILURE;
}
else
{
return errnoReturn ("VIDIOC_REQBUFS");
}
}
if (req.count < BUFFERS)
{
kdDebug(14010) << k_funcinfo << "Insufficient buffer memory on " << full_filename << endl;
return EXIT_FAILURE;
}
m_rawbuffers.resize(req.count);
if (m_rawbuffers.size()==0)
{
kdDebug(14010) << k_funcinfo << "Out of memory" << endl;
return EXIT_FAILURE;
}
for (m_streambuffers = 0; m_streambuffers < req.count; ++m_streambuffers)
{
struct v4l2_buffer v4l2buffer;
CLEAR (v4l2buffer);
v4l2buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
v4l2buffer.memory = V4L2_MEMORY_MMAP;
v4l2buffer.index = m_streambuffers;
if (-1 == xioctl (VIDIOC_QUERYBUF, &v4l2buffer))
return errnoReturn ("VIDIOC_QUERYBUF");
m_rawbuffers[m_streambuffers].length = v4l2buffer.length;
m_rawbuffers[m_streambuffers].start = (uchar *) mmap (NULL /* start anywhere */, v4l2buffer.length, PROT_READ | PROT_WRITE /* required */, MAP_SHARED /* recommended */, descriptor, v4l2buffer.m.offset);
if (MAP_FAILED == m_rawbuffers[m_streambuffers].start)
return errnoReturn ("mmap");
}
#endif
m_currentbuffer.data.resize(m_rawbuffers[0].length); // Makes the imagesize.data buffer size equal to the rawbuffer size
kdDebug(14010) << k_funcinfo << full_filename << " m_currentbuffer.data.size(): " << m_currentbuffer.data.size() << endl;
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
/*!
\fn VideoDevice::initUserptr()
*/
int VideoDevice::initUserptr()
{
/// @todo implement me
if(isOpen())
{
#ifdef V4L2_CAP_VIDEO_CAPTURE
struct v4l2_requestbuffers req;
CLEAR (req);
req.count = 2;
req.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
req.memory = V4L2_MEMORY_USERPTR;
if (-1 == xioctl (VIDIOC_REQBUFS, &req))
{
if (EINVAL == errno)
{
kdDebug(14010) << k_funcinfo << full_filename << " does not support memory mapping" << endl;
return EXIT_FAILURE;
}
else
{
return errnoReturn ("VIDIOC_REQBUFS");
}
}
m_rawbuffers.resize(4);
if (m_rawbuffers.size()==0)
{
fprintf (stderr, "Out of memory\n");
return EXIT_FAILURE;
}
for (m_streambuffers = 0; m_streambuffers < 4; ++m_streambuffers)
{
m_rawbuffers[m_streambuffers].length = m_buffer_size;
m_rawbuffers[m_streambuffers].start = (uchar *) malloc (m_buffer_size);
if (!m_rawbuffers[m_streambuffers].start)
{
kdDebug(14010) << k_funcinfo << "Out of memory" << endl;
return EXIT_FAILURE;
}
}
#endif
return EXIT_SUCCESS;
}
return EXIT_FAILURE;
}
bool VideoDevice::canCapture()
{
return m_videocapture;
}
bool VideoDevice::canChromakey()
{
return m_videochromakey;
}
bool VideoDevice::canScale()
{
return m_videoscale;
}
bool VideoDevice::canOverlay()
{
return m_videooverlay;
}
bool VideoDevice::canRead()
{
return m_videoread;
}
bool VideoDevice::canAsyncIO()
{
return m_videoasyncio;
}
bool VideoDevice::canStream()
{
return m_videostream;
}
}
}