/* videodevice.cpp - Kopete Video Device Low-level Support Copyright (c) 2005-2006 by Cláudio da Silveira Pinheiro Kopete (c) 2002-2003 by the Kopete developers ************************************************************************* * * * This library is free software; you can redistribute it and/or * * modify it under the terms of the GNU Lesser General Public * * License as published by the Free Software Foundation; either * * version 2 of the License, or (at your option) any later version. * * * ************************************************************************* */ #include #include #include #include #include "videoinput.h" #include "videodevice.h" #include "bayer.h" #include "sonix_compress.h" #define CLEAR(x) memset (&(x), 0, sizeof (x)) namespace Kopete { namespace AV { VideoDevice::VideoDevice() { // kdDebug(14010) << "libkopete (avdevice): VideoDevice() called" << endl; descriptor = -1; m_streambuffers = 0; m_current_input = 0; // kdDebug(14010) << "libkopete (avdevice): VideoDevice() exited successfuly" << endl; maxwidth = 32767; maxheight = 32767; minwidth = 1; minheight = 1; } VideoDevice::~VideoDevice() { } #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE void VideoDevice::enumerateMenu (void) { kdDebug(14010) << k_funcinfo << " Menu items:" << endl; memset (&querymenu, 0, sizeof (querymenu)); querymenu.id = queryctrl.id; for (querymenu.index = queryctrl.minimum; querymenu.index <= queryctrl.maximum; querymenu.index++) { if (0 == xioctl (VIDIOC_QUERYMENU, &querymenu)) { kdDebug(14010) << k_funcinfo << " " << TQString(TQString::fromLocal8Bit((const char*)querymenu.name)) << endl; } else { perror ("VIDIOC_QUERYMENU"); exit (EXIT_FAILURE); } } } #endif #endif /*! \fn VideoDevice::xioctl(int fd, int request, void *arg) */ int VideoDevice::xioctl(int request, void *arg) { int r; do r = ioctl (descriptor, request, arg); while (-1 == r && EINTR == errno); return r; } /*! \fn VideoDevice::errnoReturn(const char* s) */ int VideoDevice::errnoReturn(const char* s) { /// @todo implement me fprintf (stderr, "%s error %d, %s\n",s, errno, strerror (errno)); return EXIT_FAILURE; } /*! \fn VideoDevice::setFileName(TQString name) */ int VideoDevice::setFileName(TQString filename) { /// @todo implement me full_filename=filename; return EXIT_SUCCESS; } /*! \fn VideoDevice::open() */ int VideoDevice::open() { /// @todo implement me kdDebug(14010) << k_funcinfo << "called" << endl; if(-1 != descriptor) { kdDebug(14010) << k_funcinfo << "Device is already open" << endl; return EXIT_SUCCESS; } descriptor = ::open (TQFile::encodeName(full_filename), O_RDWR, 0); if(isOpen()) { kdDebug(14010) << k_funcinfo << "File " << full_filename << " was opened successfuly" << endl; if(EXIT_FAILURE==checkDevice()) { kdDebug(14010) << k_funcinfo << "File " << full_filename << " could not be opened" << endl; close(); return EXIT_FAILURE; } } else { kdDebug(14010) << k_funcinfo << "Unable to open file " << full_filename << "Err: "<< errno << endl; return EXIT_FAILURE; } initDevice(); selectInput(m_current_input); kdDebug(14010) << k_funcinfo << "exited successfuly" << endl; return EXIT_SUCCESS; } bool VideoDevice::isOpen() { if(-1 == descriptor) { // kdDebug(14010) << k_funcinfo << "VideoDevice::isOpen() File is not open" << endl; return false; } // kdDebug(14010) << k_funcinfo << "VideoDevice::isOpen() File is open" << endl; return true; } int VideoDevice::checkDevice() { kdDebug(14010) << k_funcinfo << "checkDevice() called." << endl; if(isOpen()) { m_videocapture=false; m_videochromakey=false; m_videoscale=false; m_videooverlay=false; m_videoread=false; m_videoasyncio=false; m_videostream=false; m_driver=VIDEODEV_DRIVER_NONE; #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE //if(!getWorkaroundBrokenDriver()) { kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " Trying V4L2 API." << endl; CLEAR(V4L2_capabilities); if (-1 != xioctl (VIDIOC_QUERYCAP, &V4L2_capabilities)) { if (!(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_CAPTURE)) { kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " is not a video capture device." << endl; m_driver = VIDEODEV_DRIVER_NONE; return EXIT_FAILURE; } m_videocapture=true; kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " is a V4L2 device." << endl; m_driver = VIDEODEV_DRIVER_V4L2; m_model=TQString::fromLocal8Bit((const char*)V4L2_capabilities.card); // Detect maximum and minimum resolution supported by the V4L2 device CLEAR (fmt); fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (-1 == xioctl (VIDIOC_G_FMT, &fmt)) kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ")." << endl; fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; fmt.fmt.pix.width = 32767; fmt.fmt.pix.height = 32767; fmt.fmt.pix.field = V4L2_FIELD_ANY; if (-1 == xioctl (VIDIOC_S_FMT, &fmt)) { kdDebug(14010) << k_funcinfo << "Detecting maximum size with VIDIOC_S_FMT failed (" << errno << ").Returned maxwidth: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl; // Note VIDIOC_S_FMT may change width and height. } else { maxwidth = fmt.fmt.pix.width; maxheight = fmt.fmt.pix.height; } if (-1 == xioctl (VIDIOC_G_FMT, &fmt)) kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ")." << endl; fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; fmt.fmt.pix.width = 1; fmt.fmt.pix.height = 1; fmt.fmt.pix.field = V4L2_FIELD_ANY; if (-1 == xioctl (VIDIOC_S_FMT, &fmt)) { kdDebug(14010) << k_funcinfo << "Detecting minimum size with VIDIOC_S_FMT failed (" << errno << ").Returned maxwidth: " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl; // Note VIDIOC_S_FMT may change width and height. } else { minwidth = fmt.fmt.pix.width; minheight = fmt.fmt.pix.height; } // Buggy driver paranoia /* min = fmt.fmt.pix.width * 2; if (fmt.fmt.pix.bytesperline < min) fmt.fmt.pix.bytesperline = min; min = fmt.fmt.pix.bytesperline * fmt.fmt.pix.height; if (fmt.fmt.pix.sizeimage < min) fmt.fmt.pix.sizeimage = min; m_buffer_size=fmt.fmt.pix.sizeimage ;*/ int inputisok=EXIT_SUCCESS; m_input.clear(); for(unsigned int loop=0; inputisok==EXIT_SUCCESS; loop++) { struct v4l2_input videoinput; CLEAR(videoinput); videoinput.index = loop; inputisok=xioctl(VIDIOC_ENUMINPUT, &videoinput); if(inputisok==EXIT_SUCCESS) { VideoInput tempinput; tempinput.name = TQString::fromLocal8Bit((const char*)videoinput.name); tempinput.hastuner = videoinput.type & V4L2_INPUT_TYPE_TUNER; tempinput.m_standards = videoinput.std; m_input.push_back(tempinput); kdDebug(14010) << k_funcinfo << "Input " << loop << ": " << tempinput.name << " (tuner: " << ((videoinput.type & V4L2_INPUT_TYPE_TUNER) != 0) << ")" << endl; if((videoinput.type & V4L2_INPUT_TYPE_TUNER) != 0) { // _tunerForInput[name] = desc.tuner; // _isTuner = true; } else { // _tunerForInput[name] = -1; } } } // ----------------------------------------------------------------------------------------------------------------- // This must turn up to be a proper method to check for controls' existence. CLEAR (queryctrl); // v4l2_queryctrl may zero the .id in some cases, even if the IOCTL returns EXIT_SUCCESS (tested with a bttv card, when testing for V4L2_CID_AUDIO_VOLUME). // As of 6th Aug 2007, according to the V4L2 specification version 0.21, this behavior is undocumented, and the example 1-8 code found at // http://www.linuxtv.org/downloads/video4linux/API/V4L2_API/spec/x519.htm fails because of this behavior with a bttv card. int currentid = V4L2_CID_BASE; kdDebug(14010) << k_funcinfo << "Checking CID controls" << endl; for (currentid = V4L2_CID_BASE; currentid < V4L2_CID_LASTP1; currentid++) //for (queryctrl.id = 9963776; queryctrl.id < 9963800; queryctrl.id++) { queryctrl.id = currentid; //kdDebug(14010) << k_funcinfo << "Checking CID controls from " << V4L2_CID_BASE << " to " << V4L2_CID_LASTP1 << ". Current: " << queryctrl.id << ". IOCTL returns: " << resultado << endl; if (0 == xioctl (VIDIOC_QUERYCTRL, &queryctrl)) { if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED) continue; //kdDebug(14010) << k_funcinfo << " Control: " << TQString::fromLocal8Bit((const char*)queryctrl.name) << endl; kdDebug(14010) << k_funcinfo << " Control: " << TQString(TQString::fromLocal8Bit((const char*)queryctrl.name)) << " Values from " << queryctrl.minimum << " to " << queryctrl.maximum << " with steps of " << queryctrl.step << ". Default: " << queryctrl.default_value << endl; /* switch (queryctrl.type) { case V4L2_CTRL_TYPE_INTEGER : }*/ if (queryctrl.type == V4L2_CTRL_TYPE_MENU) enumerateMenu (); } else { if (errno == EINVAL) continue; perror ("VIDIOC_QUERYCTRL"); // exit (EXIT_FAILURE); } } kdDebug(14010) << k_funcinfo << "Checking CID private controls" << endl; for (currentid = V4L2_CID_PRIVATE_BASE;; currentid++) //for (queryctrl.id = 9963776; queryctrl.id < 9963800; queryctrl.id++) { queryctrl.id = currentid; //kdDebug(14010) << k_funcinfo << "Checking CID private controls from " << V4L2_CID_PRIVATE_BASE << ". Current: " << queryctrl.id << ". IOCTL returns: " << resultado << endl; if ( 0 == xioctl (VIDIOC_QUERYCTRL, &queryctrl)) { if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED) continue; kdDebug(14010) << k_funcinfo << " Control: " << TQString(TQString::fromLocal8Bit((const char*)queryctrl.name)) << " Values from " << queryctrl.minimum << " to " << queryctrl.maximum << " with steps of " << queryctrl.step << ". Default: " << queryctrl.default_value << endl; if (queryctrl.type == V4L2_CTRL_TYPE_MENU) enumerateMenu (); } else { if (errno == EINVAL) break; perror ("VIDIOC_QUERYCTRL"); // exit (EXIT_FAILURE); } } } else { // V4L-only drivers should return an EINVAL in errno to indicate they cannot handle V4L2 calls. Not every driver is compliant, so // it will try the V4L api even if the error code is different than expected. kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " is not a V4L2 device." << endl; } } #endif CLEAR(V4L_capabilities); if(m_driver==VIDEODEV_DRIVER_NONE) { kdDebug(14010) << k_funcinfo << "checkDevice(): " << full_filename << " Trying V4L API." << endl; if (-1 == xioctl (VIDIOCGCAP, &V4L_capabilities)) { perror ("ioctl (VIDIOCGCAP)"); m_driver = VIDEODEV_DRIVER_NONE; return EXIT_FAILURE; } else { kdDebug(14010) << k_funcinfo << full_filename << " is a V4L device." << endl; m_driver = VIDEODEV_DRIVER_V4L; m_model=TQString::fromLocal8Bit((const char*)V4L_capabilities.name); if(V4L_capabilities.type & VID_TYPE_CAPTURE) m_videocapture=true; if(V4L_capabilities.type & VID_TYPE_CHROMAKEY) m_videochromakey=true; if(V4L_capabilities.type & VID_TYPE_SCALES) m_videoscale=true; if(V4L_capabilities.type & VID_TYPE_OVERLAY) m_videooverlay=true; // kdDebug(14010) << "libkopete (avdevice): Inputs : " << V4L_capabilities.channels << endl; // kdDebug(14010) << "libkopete (avdevice): Audios : " << V4L_capabilities.audios << endl; minwidth = V4L_capabilities.minwidth; maxwidth = V4L_capabilities.maxwidth; minheight = V4L_capabilities.minheight; maxheight = V4L_capabilities.maxheight; int inputisok=EXIT_SUCCESS; m_input.clear(); for(int loop=0; loop < V4L_capabilities.channels; loop++) { struct video_channel videoinput; CLEAR(videoinput); videoinput.channel = loop; videoinput.norm = 1; inputisok=xioctl(VIDIOCGCHAN, &videoinput); if(inputisok==EXIT_SUCCESS) { VideoInput tempinput; tempinput.name = TQString::fromLocal8Bit((const char*)videoinput.name); tempinput.hastuner=videoinput.flags & VIDEO_VC_TUNER; // TODO: The routine to detect the appropriate video standards for V4L must be placed here m_input.push_back(tempinput); // kdDebug(14010) << "libkopete (avdevice): Input " << loop << ": " << tempinput.name << " (tuner: " << ((videoinput.flags & VIDEO_VC_TUNER) != 0) << ")" << endl; /* if((input.type & V4L2_INPUT_TYPE_TUNER) != 0) { // _tunerForInput[name] = desc.tuner; // _isTuner = true; } else { // _tunerForInput[name] = -1; } */ } } } } #endif m_name=m_model; // Take care about changing the name to be different from the model itself... detectPixelFormats(); // TODO: Now we must execute the proper initialization according to the type of the driver. kdDebug(14010) << k_funcinfo << "checkDevice() exited successfuly." << endl; return EXIT_SUCCESS; } return EXIT_FAILURE; } /*! \fn VideoDevice::showDeviceCapabilities() */ int VideoDevice::showDeviceCapabilities() { kdDebug(14010) << k_funcinfo << "showDeviceCapabilities() called." << endl; if(isOpen()) { /* kdDebug(14010) << "libkopete (avdevice): Driver: " << (const char*)V4L2_capabilities.driver << " " << ((V4L2_capabilities.version>>16) & 0xFF) << "." << ((V4L2_capabilities.version>> 8) & 0xFF) << "." << ((V4L2_capabilities.version ) & 0xFF) << endl; kdDebug(14010) << "libkopete (avdevice): Card: " << name << endl; kdDebug(14010) << "libkopete (avdevice): Capabilities:" << endl; if(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_CAPTURE) kdDebug(14010) << "libkopete (avdevice): Video capture" << endl; if(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_OUTPUT) kdDebug(14010) << "libkopete (avdevice): Video output" << endl; if(V4L2_capabilities.capabilities & V4L2_CAP_VIDEO_OVERLAY) kdDebug(14010) << "libkopete (avdevice): Video overlay" << endl; if(V4L2_capabilities.capabilities & V4L2_CAP_VBI_CAPTURE) kdDebug(14010) << "libkopete (avdevice): VBI capture" << endl; if(V4L2_capabilities.capabilities & V4L2_CAP_VBI_OUTPUT) kdDebug(14010) << "libkopete (avdevice): VBI output" << endl; if(V4L2_capabilities.capabilities & V4L2_CAP_RDS_CAPTURE) kdDebug(14010) << "libkopete (avdevice): RDS capture" << endl; if(V4L2_capabilities.capabilities & V4L2_CAP_TUNER) kdDebug(14010) << "libkopete (avdevice): Tuner IO" << endl; if(V4L2_capabilities.capabilities & V4L2_CAP_AUDIO) kdDebug(14010) << "libkopete (avdevice): Audio IO" << endl; ;*/ kdDebug(14010) << k_funcinfo << "Card model: " << m_model << endl; kdDebug(14010) << k_funcinfo << "Card name : " << m_name << endl; kdDebug(14010) << k_funcinfo << "Capabilities:" << endl; if(canCapture()) kdDebug(14010) << k_funcinfo << " Video capture" << endl; if(canRead()) kdDebug(14010) << k_funcinfo << " Read" << endl; if(canAsyncIO()) kdDebug(14010) << k_funcinfo << " Asynchronous input/output" << endl; if(canStream()) kdDebug(14010) << k_funcinfo << " Streaming" << endl; if(canChromakey()) kdDebug(14010) << k_funcinfo << " Video chromakey" << endl; if(canScale()) kdDebug(14010) << k_funcinfo << " Video scales" << endl; if(canOverlay()) kdDebug(14010) << k_funcinfo << " Video overlay" << endl; // kdDebug(14010) << "libkopete (avdevice): Audios : " << V4L_capabilities.audios << endl; kdDebug(14010) << k_funcinfo << " Max res: " << maxWidth() << " x " << maxHeight() << endl; kdDebug(14010) << k_funcinfo << " Min res: " << minWidth() << " x " << minHeight() << endl; kdDebug(14010) << k_funcinfo << " Inputs : " << inputs() << endl; for (unsigned int loop=0; loop < inputs(); loop++) kdDebug(14010) << k_funcinfo << "Input " << loop << ": " << m_input[loop].name << " (tuner: " << m_input[loop].hastuner << ")" << endl; kdDebug(14010) << k_funcinfo << "showDeviceCapabilities() exited successfuly." << endl; return EXIT_SUCCESS; } return EXIT_FAILURE; } /*! \fn VideoDevicePool::initDevice() */ int VideoDevice::initDevice() { /// @todo implement me kdDebug(14010) << k_funcinfo << "initDevice() started" << endl; if(-1 == descriptor) { kdDebug(14010) << k_funcinfo << "initDevice() Device is not open" << endl; return EXIT_FAILURE; } m_io_method = IO_METHOD_NONE; switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: if(V4L2_capabilities.capabilities & V4L2_CAP_READWRITE) { m_videoread=true; m_io_method = IO_METHOD_READ; kdDebug(14010) << k_funcinfo << " Read/Write interface" << endl; } if(V4L2_capabilities.capabilities & V4L2_CAP_ASYNCIO) { m_videoasyncio=true; kdDebug(14010) << k_funcinfo << " Async IO interface" << endl; } if(V4L2_capabilities.capabilities & V4L2_CAP_STREAMING) { m_videostream=true; m_io_method = IO_METHOD_MMAP; // m_io_method = IO_METHOD_USERPTR; kdDebug(14010) << k_funcinfo << " Streaming interface" << endl; } if(m_io_method==IO_METHOD_NONE) { kdDebug(14010) << k_funcinfo << "initDevice() Found no suitable input/output method for " << full_filename << endl; return EXIT_FAILURE; } break; #endif case VIDEODEV_DRIVER_V4L: m_videoread=true; m_io_method=IO_METHOD_READ; if(-1 != xioctl(VIDIOCGFBUF,&V4L_videobuffer)) { // m_videostream=true; // m_io_method = IO_METHOD_MMAP; kdDebug(14010) << k_funcinfo << " Streaming interface" << endl; } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } // Select video input, video standard and tune here. #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE cropcap.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (-1 == xioctl (VIDIOC_CROPCAP, &cropcap)) { // Errors ignored. } crop.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; crop.c = cropcap.defrect; // reset to default if (-1 == xioctl (VIDIOC_S_CROP, &crop)) { switch (errno) { case EINVAL: break; // Cropping not supported. default: break; // Errors ignored. } } #endif #endif showDeviceCapabilities(); kdDebug(14010) << k_funcinfo << "initDevice() exited successfuly" << endl; return EXIT_SUCCESS; } unsigned int VideoDevice::inputs() { return m_input.size(); } int VideoDevice::width() { return currentwidth; } int VideoDevice::minWidth() { return minwidth; } int VideoDevice::maxWidth() { return maxwidth; } int VideoDevice::height() { return currentheight; } int VideoDevice::minHeight() { return minheight; } int VideoDevice::maxHeight() { return maxheight; } int VideoDevice::setSize( int newwidth, int newheight) { kdDebug(14010) << k_funcinfo << "setSize(" << newwidth << ", " << newheight << ") called." << endl; if(isOpen()) { // It should not be there. It must remain in a completely distict place, cause this method should not change the pixelformat. kdDebug(14010) << k_funcinfo << "Trying YUY422P" << endl; if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_YUV422P)) { kdDebug(14010) << k_funcinfo << "Card doesn't seem to support YUV422P format. Trying YUYV." << endl; if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_YUYV)) { kdDebug(14010) << k_funcinfo << "Card doesn't seem to support YUYV format. Trying UYVY." << endl; if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_UYVY)) { kdDebug(14010) << k_funcinfo << "Card doesn't seem to support UYVY format. Trying YUV420P." << endl; if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_YUV420P)) { kdDebug(14010) << k_funcinfo << "Card doesn't seem to support YUV420P format. Trying RGB24." << endl; if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_RGB24)) { kdDebug(14010) << k_funcinfo << "Card doesn't seem to support RGB24 format. Trying BGR24." << endl; if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_BGR24)) { kdDebug(14010) << k_funcinfo << "Card doesn't seem to support RGB24 format. Trying RGB32." << endl; if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_RGB32)) { kdDebug(14010) << k_funcinfo << "Card doesn't seem to support RGB32 format. Trying BGR32." << endl; if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_BGR32)) { kdDebug(14010) << k_funcinfo << "Card doesn't seem to support BGR32 format. Trying SN9C10X." << endl; if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_SN9C10X)) { kdDebug(14010) << k_funcinfo << "Card doesn't seem to support SN9C10X format. Trying Bayer RGB." << endl; if(PIXELFORMAT_NONE == setPixelFormat(PIXELFORMAT_SBGGR8)) kdDebug(14010) << k_funcinfo << "Card doesn't seem to support SBGGR8 format. Fallback from it is not yet implemented." << endl; } } } } } } } } } if(newwidth > maxwidth ) newwidth = maxwidth; if(newheight > maxheight) newheight = maxheight; if(newwidth < minwidth ) newwidth = minwidth; if(newheight < minheight) newheight = minheight; currentwidth = newwidth; currentheight = newheight; //kdDebug(14010) << k_funcinfo << "width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << width() << "x" << height() << endl; // Change resolution for the video device switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: // CLEAR (fmt); if (-1 == xioctl (VIDIOC_G_FMT, &fmt)) kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl; fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; fmt.fmt.pix.width = width(); fmt.fmt.pix.height = height(); fmt.fmt.pix.field = V4L2_FIELD_ANY; if (-1 == xioctl (VIDIOC_S_FMT, &fmt)) { kdDebug(14010) << k_funcinfo << "VIDIOC_S_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl; // Note VIDIOC_S_FMT may change width and height. } else { // Buggy driver paranoia. kdDebug(14010) << k_funcinfo << "VIDIOC_S_FMT worked (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl; unsigned int min = fmt.fmt.pix.width * 2; if (fmt.fmt.pix.bytesperline < min) fmt.fmt.pix.bytesperline = min; min = fmt.fmt.pix.bytesperline * fmt.fmt.pix.height; if (fmt.fmt.pix.sizeimage < min) fmt.fmt.pix.sizeimage = min; m_buffer_size=fmt.fmt.pix.sizeimage ; } break; #endif case VIDEODEV_DRIVER_V4L: { struct video_window V4L_videowindow; kdDebug(14010) << "------------- width: " << V4L_videowindow.width << " Height: " << V4L_videowindow.height << " Clipcount: " << V4L_videowindow.clipcount << " -----------------" << endl; if (xioctl (VIDIOCGWIN, &V4L_videowindow)== -1) { perror ("ioctl VIDIOCGWIN"); // return (NULL); } V4L_videowindow.width = width(); V4L_videowindow.height = height(); V4L_videowindow.clipcount=0; if (xioctl (VIDIOCSWIN, &V4L_videowindow)== -1) { perror ("ioctl VIDIOCSWIN"); // return (NULL); } kdDebug(14010) << "------------- width: " << V4L_videowindow.width << " Height: " << V4L_videowindow.height << " Clipcount: " << V4L_videowindow.clipcount << " -----------------" << endl; // kdDebug(14010) << "libkopete (avdevice): V4L_picture.palette: " << V4L_picture.palette << " Depth: " << V4L_picture.depth << endl; /* if(-1 == xioctl(VIDIOCGFBUF,&V4L_videobuffer)) kdDebug(14010) << "libkopete (avdevice): VIDIOCGFBUF failed (" << errno << "): Card cannot stream" << endl;*/ } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } m_buffer_size = width() * height() * pixelFormatDepth(m_pixelformat) / 8; kdDebug(14010) << "------------------------- ------- -- m_buffer_size: " << m_buffer_size << " !!! -- ------- -----------------------------------------" << endl; m_currentbuffer.pixelformat=m_pixelformat; m_currentbuffer.data.resize(m_buffer_size); switch (m_io_method) { case IO_METHOD_NONE: break; case IO_METHOD_READ: initRead (); break; case IO_METHOD_MMAP: initMmap (); break; case IO_METHOD_USERPTR: initUserptr (); break; } kdDebug(14010) << k_funcinfo << "setSize(" << newwidth << ", " << newheight << ") exited successfuly." << endl; return EXIT_SUCCESS; } kdDebug(14010) << k_funcinfo << "setSize(" << newwidth << ", " << newheight << ") Device is not open." << endl; return EXIT_FAILURE; } pixel_format VideoDevice::setPixelFormat(pixel_format newformat) { pixel_format ret = PIXELFORMAT_NONE; //kdDebug(14010) << k_funcinfo << "called." << endl; // Change the pixel format for the video device switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: // CLEAR (fmt); if (-1 == xioctl (VIDIOC_G_FMT, &fmt)) { // return errnoReturn ("VIDIOC_S_FMT"); // kdDebug(14010) << k_funcinfo << "VIDIOC_G_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl; } else m_pixelformat = pixelFormatForPalette(fmt.fmt.pix.pixelformat); fmt.fmt.pix.pixelformat = pixelFormatCode(newformat); if (-1 == xioctl (VIDIOC_S_FMT, &fmt)) { // kdDebug(14010) << k_funcinfo << "VIDIOC_S_FMT failed (" << errno << ").Returned width: " << pixelFormatName(fmt.fmt.pix.pixelformat) << " " << fmt.fmt.pix.width << "x" << fmt.fmt.pix.height << endl; } else { if (fmt.fmt.pix.pixelformat == pixelFormatCode(newformat)) { m_pixelformat = newformat; ret = m_pixelformat; } } break; #endif case VIDEODEV_DRIVER_V4L: { struct video_picture V4L_picture; if(-1 == xioctl(VIDIOCGPICT, &V4L_picture)) kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl; // kdDebug(14010) << k_funcinfo << "V4L_picture.palette: " << V4L_picture.palette << " Depth: " << V4L_picture.depth << endl; V4L_picture.palette = pixelFormatCode(newformat); V4L_picture.depth = pixelFormatDepth(newformat); if(-1 == xioctl(VIDIOCSPICT,&V4L_picture)) { // kdDebug(14010) << k_funcinfo << "Card seems to not support " << pixelFormatName(newformat) << " format. Fallback to it is not yet implemented." << endl; } if(-1 == xioctl(VIDIOCGPICT, &V4L_picture)) kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl; // kdDebug(14010) << k_funcinfo << "V4L_picture.palette: " << V4L_picture.palette << " Depth: " << V4L_picture.depth << endl; m_pixelformat=pixelFormatForPalette(V4L_picture.palette); if (m_pixelformat == newformat) ret = newformat; } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } return ret; } /*! \fn Kopete::AV::VideoDevice::currentInput() */ int VideoDevice::currentInput() { /// @todo implement me if(isOpen()) { return m_current_input; } return 0; } /*! \fn Kopete::AV::VideoDevice::selectInput(int input) */ int VideoDevice::selectInput(int newinput) { /// @todo implement me if(m_current_input >= inputs()) return EXIT_FAILURE; if(isOpen()) { switch (m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: if (-1 == ioctl (descriptor, VIDIOC_S_INPUT, &newinput)) { perror ("VIDIOC_S_INPUT"); return EXIT_FAILURE; } break; #endif case VIDEODEV_DRIVER_V4L: struct video_channel V4L_input; V4L_input.channel=newinput; V4L_input.norm=4; // Hey, it's plain wrong! It should be input's signal standard! if (-1 == ioctl (descriptor, VIDIOCSCHAN, &V4L_input)) { perror ("ioctl (VIDIOCSCHAN)"); return EXIT_FAILURE; } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } kdDebug(14010) << k_funcinfo << "Selected input " << newinput << " (" << m_input[newinput].name << ")" << endl; m_current_input = newinput; setInputParameters(); return EXIT_SUCCESS; } return EXIT_FAILURE; } /*! \fn Kopete::AV::VideoDevice::setInputParameters() */ int VideoDevice::setInputParameters() { /// @todo implement me if( (isOpen()) && (m_current_input < inputs() ) ) { setBrightness( getBrightness() ); setContrast( getContrast() ); setSaturation( getSaturation() ); setWhiteness( getWhiteness() ); setHue( getHue() ); return EXIT_SUCCESS; } return EXIT_FAILURE; } /*! \fn VideoDevice::startCapturing() */ int VideoDevice::startCapturing() { kdDebug(14010) << k_funcinfo << "called." << endl; if(isOpen()) { switch (m_io_method) { case IO_METHOD_NONE: // Card cannot capture frames return EXIT_FAILURE; break; case IO_METHOD_READ: // Nothing to do break; case IO_METHOD_MMAP: #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE { unsigned int loop; for (loop = 0; loop < m_streambuffers; ++loop) { struct v4l2_buffer buf; CLEAR (buf); buf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; buf.memory = V4L2_MEMORY_MMAP; buf.index = loop; if (-1 == xioctl (VIDIOC_QBUF, &buf)) return errnoReturn ("VIDIOC_QBUF"); } enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (-1 == xioctl (VIDIOC_STREAMON, &type)) return errnoReturn ("VIDIOC_STREAMON"); } #endif #endif break; case IO_METHOD_USERPTR: #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE { unsigned int loop; for (loop = 0; loop < m_streambuffers; ++loop) { struct v4l2_buffer buf; CLEAR (buf); buf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; buf.memory = V4L2_MEMORY_USERPTR; buf.m.userptr = (unsigned long) m_rawbuffers[loop].start; buf.length = m_rawbuffers[loop].length; if (-1 == xioctl (VIDIOC_QBUF, &buf)) return errnoReturn ("VIDIOC_QBUF"); } enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (-1 == xioctl (VIDIOC_STREAMON, &type)) return errnoReturn ("VIDIOC_STREAMON"); } #endif #endif break; } kdDebug(14010) << k_funcinfo << "exited successfuly." << endl; return EXIT_SUCCESS; } return EXIT_FAILURE; } /*! \fn VideoDevice::getFrame() */ int VideoDevice::getFrame() { /// @todo implement me ssize_t bytesread; #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE struct v4l2_buffer v4l2buffer; #endif #endif // kdDebug(14010) << k_funcinfo << "getFrame() called." << endl; if(isOpen()) { switch (m_io_method) { case IO_METHOD_NONE: // Card cannot capture frames return EXIT_FAILURE; break; case IO_METHOD_READ: // kdDebug(14010) << k_funcinfo << "Using IO_METHOD_READ.File descriptor: " << descriptor << " Buffer address: " << &m_currentbuffer.data[0] << " Size: " << m_currentbuffer.data.size() << endl; bytesread = read (descriptor, &m_currentbuffer.data[0], m_currentbuffer.data.size()); if (-1 == bytesread) // must verify this point with ov511 driver. { kdDebug(14010) << k_funcinfo << "IO_METHOD_READ failed." << endl; switch (errno) { case EAGAIN: return EXIT_FAILURE; case EIO: /* Could ignore EIO, see spec. fall through */ default: return errnoReturn ("read"); } } if((int)m_currentbuffer.data.size() < bytesread) { kdDebug(14010) << k_funcinfo << "IO_METHOD_READ returned less bytes (" << bytesread << ") than it was asked for (" << m_currentbuffer.data.size() <<")." << endl; } break; case IO_METHOD_MMAP: #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE CLEAR (v4l2buffer); v4l2buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; v4l2buffer.memory = V4L2_MEMORY_MMAP; if (-1 == xioctl (VIDIOC_DQBUF, &v4l2buffer)) { kdDebug(14010) << k_funcinfo << full_filename << " MMAPed getFrame failed." << endl; switch (errno) { case EAGAIN: { kdDebug(14010) << k_funcinfo << full_filename << " MMAPed getFrame failed: EAGAIN. Pointer: " << endl; return EXIT_FAILURE; } case EIO: /* Could ignore EIO, see spec. fall through */ default: return errnoReturn ("VIDIOC_DQBUF"); } } /* if (v4l2buffer.index < m_streambuffers) return EXIT_FAILURE;*/ //it was an assert() //kdDebug(14010) << k_funcinfo << "m_rawbuffers[" << v4l2buffer.index << "].start: " << (void *)m_rawbuffers[v4l2buffer.index].start << " Size: " << m_currentbuffer.data.size() << endl; /*{ unsigned long long result=0; unsigned long long R=0, G=0, B=0, A=0; int Rmax=0, Gmax=0, Bmax=0, Amax=0; int Rmin=255, Gmin=255, Bmin=255, Amin=0; for(unsigned int loop=0;loop < m_currentbuffer.data.size();loop+=4) { R+=m_rawbuffers[v4l2buffer.index].start[loop]; G+=m_rawbuffers[v4l2buffer.index].start[loop+1]; B+=m_rawbuffers[v4l2buffer.index].start[loop+2]; // A+=currentbuffer.data[loop+3]; if (m_currentbuffer.data[loop] < Rmin) Rmin = m_currentbuffer.data[loop]; if (m_currentbuffer.data[loop+1] < Gmin) Gmin = m_currentbuffer.data[loop+1]; if (m_currentbuffer.data[loop+2] < Bmin) Bmin = m_currentbuffer.data[loop+2]; // if (m_currentbuffer.data[loop+3] < Amin) Amin = m_currentbuffer.data[loop+3]; if (m_currentbuffer.data[loop] > Rmax) Rmax = m_currentbuffer.data[loop]; if (m_currentbuffer.data[loop+1] > Gmax) Gmax = m_currentbuffer.data[loop+1]; if (m_currentbuffer.data[loop+2] > Bmax) Bmax = m_currentbuffer.data[loop+2]; // if (m_currentbuffer.data[loop+3] > Amax) Amax = m_currentbuffer.data[loop+3]; } kdDebug(14010) << " R: " << R << " G: " << G << " B: " << B << " A: " << A << " Rmin: " << Rmin << " Gmin: " << Gmin << " Bmin: " << Bmin << " Amin: " << Amin << " Rmax: " << Rmax << " Gmax: " << Gmax << " Bmax: " << Bmax << " Amax: " << Amax << endl; }*/ memcpy(&m_currentbuffer.data[0], m_rawbuffers[v4l2buffer.index].start, m_currentbuffer.data.size()); if (-1 == xioctl (VIDIOC_QBUF, &v4l2buffer)) return errnoReturn ("VIDIOC_QBUF"); #endif #endif break; case IO_METHOD_USERPTR: #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE { unsigned int i; CLEAR (v4l2buffer); v4l2buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; v4l2buffer.memory = V4L2_MEMORY_USERPTR; if (-1 == xioctl (VIDIOC_DQBUF, &v4l2buffer)) { switch (errno) { case EAGAIN: return EXIT_FAILURE; case EIO: /* Could ignore EIO, see spec. fall through */ default: return errnoReturn ("VIDIOC_DQBUF"); } } for (i = 0; i < m_streambuffers; ++i) if (v4l2buffer.m.userptr == (unsigned long) m_rawbuffers[i].start && v4l2buffer.length == m_rawbuffers[i].length) break; if (i < m_streambuffers) return EXIT_FAILURE; if (-1 == xioctl (VIDIOC_QBUF, &v4l2buffer)) return errnoReturn ("VIDIOC_QBUF"); } #endif #endif break; } /* Automatic color correction. Now it just swaps R and B channels in RGB24/BGR24 modes. if(m_input[m_current_input].getAutoColorCorrection()) { switch(m_currentbuffer.pixelformat) { case PIXELFORMAT_NONE : break; case PIXELFORMAT_GREY : break; case PIXELFORMAT_RGB332 : break; case PIXELFORMAT_RGB555 : break; case PIXELFORMAT_RGB555X: break; case PIXELFORMAT_RGB565 : break; case PIXELFORMAT_RGB565X: break; case PIXELFORMAT_RGB24 : case PIXELFORMAT_BGR24 : { unsigned char temp; for(unsigned int loop=0;loop < m_currentbuffer.data.size();loop+=3) { temp = m_currentbuffer.data[loop]; m_currentbuffer.data[loop] = m_currentbuffer.data[loop+2]; m_currentbuffer.data[loop+2] = temp; } } break; case PIXELFORMAT_RGB32 : case PIXELFORMAT_BGR32 : { unsigned char temp; for(unsigned int loop=0;loop < m_currentbuffer.data.size();loop+=4) { temp = m_currentbuffer.data[loop]; m_currentbuffer.data[loop] = m_currentbuffer.data[loop+2]; m_currentbuffer.data[loop+2] = temp; } } break; case PIXELFORMAT_YUYV : break; case PIXELFORMAT_UYVY : break; case PIXELFORMAT_YUV420P: break; case PIXELFORMAT_YUV422P: break; } }*/ //kdDebug(14010) << k_funcinfo << "10 Using IO_METHOD_READ.File descriptor: " << descriptor << " Buffer address: " << &m_currentbuffer.data[0] << " Size: " << m_currentbuffer.data.size() << endl; // put frame copy operation here // kdDebug(14010) << k_funcinfo << "exited successfuly." << endl; return EXIT_SUCCESS; } return EXIT_FAILURE; } /*! \fn VideoDevice::getFrame(imagebuffer *imgbuffer) */ int VideoDevice::getFrame(imagebuffer *imgbuffer) { if(imgbuffer) { getFrame(); imgbuffer->height = m_currentbuffer.height; imgbuffer->width = m_currentbuffer.width; imgbuffer->pixelformat = m_currentbuffer.pixelformat; imgbuffer->data = m_currentbuffer.data; return EXIT_SUCCESS; } return EXIT_FAILURE; } /*! \fn Kopete::AV::VideoDevice::getImage(const TQImage *qimage) */ int VideoDevice::getImage(TQImage *qimage) { /// @todo implement me // do NOT delete qimage here, as it is received as a parameter if (qimage->width() != width() || qimage->height() != height()) qimage->create(width(), height(),32, TQImage::IgnoreEndian); uchar *bits=qimage->bits(); // kDebug() << "Capturing in " << pixelFormatName(m_currentbuffer.pixelformat); switch(m_currentbuffer.pixelformat) { case PIXELFORMAT_NONE : break; // Packed RGB formats case PIXELFORMAT_RGB332 : break; case PIXELFORMAT_RGB444 : break; case PIXELFORMAT_RGB555 : break; case PIXELFORMAT_RGB565 : { int step=0; for(int loop=0;loop < qimage->numBytes();loop+=4) { bits[loop] = (m_currentbuffer.data[step]<<3)+(m_currentbuffer.data[step]<<3>>5); bits[loop+1] = ((m_currentbuffer.data[step+1])<<5)|m_currentbuffer.data[step]>>5; bits[loop+2] = ((m_currentbuffer.data[step+1])&248)+((m_currentbuffer.data[step+1])>>5); bits[loop+3] = 255; step+=2; } } break; case PIXELFORMAT_RGB555X: break; case PIXELFORMAT_RGB565X: break; case PIXELFORMAT_BGR24 : { int step=0; for(int loop=0;loop < qimage->numBytes();loop+=4) { bits[loop] = m_currentbuffer.data[step+2]; bits[loop+1] = m_currentbuffer.data[step+1]; bits[loop+2] = m_currentbuffer.data[step]; bits[loop+3] = 255; step+=3; } } break; case PIXELFORMAT_RGB24 : { int step=0; for(int loop=0;loop < qimage->numBytes();loop+=4) { bits[loop] = m_currentbuffer.data[step]; bits[loop+1] = m_currentbuffer.data[step+1]; bits[loop+2] = m_currentbuffer.data[step+2]; bits[loop+3] = 255; step+=3; } } break; case PIXELFORMAT_BGR32 : break; case PIXELFORMAT_RGB32 : memcpy(bits,&m_currentbuffer.data[0], m_currentbuffer.data.size()); break; // Bayer RGB format case PIXELFORMAT_SBGGR8 : { unsigned char *d = (unsigned char *) malloc (width() * height() * 3); bayer2rgb24(d, &m_currentbuffer.data.first(), width(), height()); int step=0; for(int loop=0;loop < qimage->numBytes();loop+=4) { bits[loop] = d[step+2]; bits[loop+1] = d[step+1]; bits[loop+2] = d[step]; bits[loop+3] = 255; step+=3; } free(d); } break; // YUV formats case PIXELFORMAT_GREY : break; case PIXELFORMAT_YUYV: case PIXELFORMAT_UYVY: case PIXELFORMAT_YUV420P: case PIXELFORMAT_YUV422P: { uchar *yptr, *cbptr, *crptr; bool halfheight=false; bool packed=false; // Adjust algorythm to specific YUV data arrangements. if (m_currentbuffer.pixelformat == PIXELFORMAT_YUV420P) halfheight=true; if (m_currentbuffer.pixelformat == PIXELFORMAT_YUYV) { yptr = &m_currentbuffer.data[0]; cbptr = yptr + 1; crptr = yptr + 3; packed=true; } else if (m_currentbuffer.pixelformat == PIXELFORMAT_UYVY) { cbptr = &m_currentbuffer.data[0]; yptr = cbptr + 1; crptr = cbptr + 2; packed=true; } else { yptr = &m_currentbuffer.data[0]; cbptr = yptr + (width()*height()); crptr = cbptr + (width()*height()/(halfheight ? 4:2)); } for(int y=0; y>1<<2])-128; e = (crptr[x>>1<<2])-128; } else { c = (yptr[x])-16; d = (cbptr[x>>1])-128; e = (crptr[x>>1])-128; } int r = (298 * c + 409 * e + 128)>>8; int g = (298 * c - 100 * d - 208 * e + 128)>>8; int b = (298 * c + 516 * d + 128)>>8; if (r<0) r=0; if (r>255) r=255; if (g<0) g=0; if (g>255) g=255; if (b<0) b=0; if (b>255) b=255; uint *p = (uint*)qimage->scanLine(y)+x; *p = tqRgba(r,g,b,255); } // Jump to next line if (packed) { yptr+=width()*2; cbptr+=width()*2; crptr+=width()*2; } else { yptr+=width(); if (!halfheight || y&1) { cbptr+=width()/2; crptr+=width()/2; } } } } break; // Compressed formats case PIXELFORMAT_JPEG : break; case PIXELFORMAT_MPEG : break; // Reserved formats case PIXELFORMAT_DV : break; case PIXELFORMAT_ET61X251:break; case PIXELFORMAT_HI240 : break; case PIXELFORMAT_HM12 : break; case PIXELFORMAT_MJPEG : break; case PIXELFORMAT_PWC1 : break; case PIXELFORMAT_PWC2 : break; case PIXELFORMAT_SN9C10X: { unsigned char *s = new unsigned char [width() * height()]; unsigned char *d = new unsigned char [width() * height() * 3]; sonix_decompress_init(); sonix_decompress(width(), height(), &m_currentbuffer.data.first(), s); bayer2rgb24(d, s, width(), height()); int step=0; for(int loop=0;loop < qimage->numBytes();loop+=4) { bits[loop] = d[step+2]; bits[loop+1] = d[step+1]; bits[loop+2] = d[step]; bits[loop+3] = 255; step+=3; } delete[] s; delete[] d; } break; case PIXELFORMAT_WNVA : break; case PIXELFORMAT_YYUV : break; } return EXIT_SUCCESS; } /*! \fn VideoDevice::stopCapturing() */ int VideoDevice::stopCapturing() { /// @todo implement me kdDebug(14010) << k_funcinfo << "called." << endl; if(isOpen()) { switch (m_io_method) { case IO_METHOD_NONE: // Card cannot capture frames return EXIT_FAILURE; break; case IO_METHOD_READ: // Nothing to do break; case IO_METHOD_MMAP: case IO_METHOD_USERPTR: #ifdef V4L2_CAP_VIDEO_CAPTURE { enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (-1 == xioctl (VIDIOC_STREAMOFF, &type)) return errnoReturn ("VIDIOC_STREAMOFF"); if (m_io_method == IO_METHOD_MMAP) { unsigned int loop; for (loop = 0; loop < m_streambuffers; ++loop) { if (munmap(m_rawbuffers[loop].start,m_rawbuffers[loop].length) != 0) { kdDebug(14010) << k_funcinfo << "unable to munmap." << endl; } } } } #endif break; } kdDebug(14010) << k_funcinfo << "exited successfuly." << endl; return EXIT_SUCCESS; } return EXIT_FAILURE; } /*! \fn VideoDevice::close() */ int VideoDevice::close() { /// @todo implement me kdDebug(14010) << k_funcinfo << " called." << endl; if(isOpen()) { kdDebug(14010) << k_funcinfo << " Device is open. Trying to properly shutdown the device." << endl; stopCapturing(); kdDebug(14010) << k_funcinfo << "::close() returns " << ::close(descriptor) << endl; } descriptor = -1; return EXIT_SUCCESS; } float VideoDevice::getBrightness() { if (m_current_input < m_input.size() ) return m_input[m_current_input].getBrightness(); else return 0; } float VideoDevice::setBrightness(float brightness) { kdDebug(14010) << k_funcinfo << " called." << endl; m_input[m_current_input].setBrightness(brightness); // Just to check bounds switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: { struct v4l2_queryctrl queryctrl; struct v4l2_control control; CLEAR (queryctrl); queryctrl.id = V4L2_CID_BRIGHTNESS; if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl)) { if (errno != EINVAL) { kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl; } else { kdDebug(14010) << k_funcinfo << "Device doesn't support the Brightness control." << endl; } } else if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED) { kdDebug(14010) << k_funcinfo << "Device doesn't support the Brightness control." << endl; } else { CLEAR (control); control.id = V4L2_CID_BRIGHTNESS; control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getBrightness()); if (-1 == xioctl (VIDIOC_S_CTRL, &control)) { kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl; } } } break; #endif case VIDEODEV_DRIVER_V4L: { struct video_picture V4L_picture; if(-1 == xioctl(VIDIOCGPICT, &V4L_picture)) kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl; V4L_picture.brightness = uint(65535*getBrightness()); if(-1 == xioctl(VIDIOCSPICT,&V4L_picture)) kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image brightness. Fallback to it is not yet implemented." << endl; } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } return getBrightness(); } float VideoDevice::getContrast() { if (m_current_input < m_input.size() ) return m_input[m_current_input].getContrast(); else return 0; } float VideoDevice::setContrast(float contrast) { kdDebug(14010) << k_funcinfo << " called." << endl; m_input[m_current_input].setContrast(contrast); // Just to check bounds switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: { struct v4l2_queryctrl queryctrl; struct v4l2_control control; CLEAR (queryctrl); queryctrl.id = V4L2_CID_CONTRAST; if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl)) { if (errno != EINVAL) { kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl; } else { kdDebug(14010) << k_funcinfo << "Device doesn't support the Contrast control." << endl; } } else if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED) { kdDebug(14010) << k_funcinfo << "Device doesn't support the Contrast control." << endl; } else { CLEAR (control); control.id = V4L2_CID_CONTRAST; control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getContrast()); if (-1 == xioctl (VIDIOC_S_CTRL, &control)) { kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl; } } } break; #endif case VIDEODEV_DRIVER_V4L: { struct video_picture V4L_picture; if(-1 == xioctl(VIDIOCGPICT, &V4L_picture)) kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl; V4L_picture.contrast = uint(65535*getContrast()); if(-1 == xioctl(VIDIOCSPICT,&V4L_picture)) kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image contrast. Fallback to it is not yet implemented." << endl; } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } return getContrast(); } float VideoDevice::getSaturation() { if (m_current_input < m_input.size() ) return m_input[m_current_input].getSaturation(); else return 0; } float VideoDevice::setSaturation(float saturation) { kdDebug(14010) << k_funcinfo << " called." << endl; m_input[m_current_input].setSaturation(saturation); // Just to check bounds switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: { struct v4l2_queryctrl queryctrl; struct v4l2_control control; CLEAR (queryctrl); queryctrl.id = V4L2_CID_SATURATION; if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl)) { if (errno != EINVAL) { kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl; } else { kdDebug(14010) << k_funcinfo << "Device doesn't support the Saturation control." << endl; } } else if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED) { kdDebug(14010) << k_funcinfo << "Device doesn't support the Saturation control." << endl; } else { CLEAR (control); control.id = V4L2_CID_SATURATION; control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getSaturation()); if (-1 == xioctl (VIDIOC_S_CTRL, &control)) { kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl; } } } break; #endif case VIDEODEV_DRIVER_V4L: { struct video_picture V4L_picture; if(-1 == xioctl(VIDIOCGPICT, &V4L_picture)) kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl; V4L_picture.colour = uint(65535*getSaturation()); if(-1 == xioctl(VIDIOCSPICT,&V4L_picture)) kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image saturation. Fallback to it is not yet implemented." << endl; } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } return getSaturation(); } float VideoDevice::getWhiteness() { if (m_current_input < m_input.size() ) return m_input[m_current_input].getWhiteness(); else return 0; } float VideoDevice::setWhiteness(float whiteness) { kdDebug(14010) << k_funcinfo << " called." << endl; m_input[m_current_input].setWhiteness(whiteness); // Just to check bounds switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: { struct v4l2_queryctrl queryctrl; struct v4l2_control control; CLEAR (queryctrl); queryctrl.id = V4L2_CID_WHITENESS; if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl)) { if (errno != EINVAL) { kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl; } else { kdDebug(14010) << k_funcinfo << "Device doesn't support the Whiteness control." << endl; } } else if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED) { kdDebug(14010) << k_funcinfo << "Device doesn't support the Whiteness control." << endl; } else { CLEAR (control); control.id = V4L2_CID_WHITENESS; control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getWhiteness()); if (-1 == xioctl (VIDIOC_S_CTRL, &control)) { kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl; } } } break; #endif case VIDEODEV_DRIVER_V4L: { struct video_picture V4L_picture; if(-1 == xioctl(VIDIOCGPICT, &V4L_picture)) kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl; V4L_picture.whiteness = uint(65535*getWhiteness()); if(-1 == xioctl(VIDIOCSPICT,&V4L_picture)) kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting white level. Fallback to it is not yet implemented." << endl; } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } return getWhiteness(); } float VideoDevice::getHue() { if (m_current_input < m_input.size() ) return m_input[m_current_input].getHue(); else return 0; } float VideoDevice::setHue(float hue) { kdDebug(14010) << k_funcinfo << " called." << endl; m_input[m_current_input].setHue(hue); // Just to check bounds switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: { struct v4l2_queryctrl queryctrl; struct v4l2_control control; CLEAR (queryctrl); queryctrl.id = V4L2_CID_HUE; if (-1 == xioctl (VIDIOC_QUERYCTRL, &queryctrl)) { if (errno != EINVAL) { kdDebug(14010) << k_funcinfo << "VIDIOC_QUERYCTRL failed (" << errno << ")." << endl; } else { kdDebug(14010) << k_funcinfo << "Device doesn't support the Hue control." << endl; } } else if (queryctrl.flags & V4L2_CTRL_FLAG_DISABLED) { kdDebug(14010) << k_funcinfo << "Device doesn't support the Hue control." << endl; } else { CLEAR (control); control.id = V4L2_CID_HUE; control.value = (__s32)((queryctrl.maximum - queryctrl.minimum)*getHue()); if (-1 == xioctl (VIDIOC_S_CTRL, &control)) { kdDebug(14010) << k_funcinfo << "VIDIOC_S_CTRL failed (" << errno << ")." << endl; } } } break; #endif case VIDEODEV_DRIVER_V4L: { struct video_picture V4L_picture; if(-1 == xioctl(VIDIOCGPICT, &V4L_picture)) kdDebug(14010) << k_funcinfo << "VIDIOCGPICT failed (" << errno << ")." << endl; V4L_picture.hue = uint(65535*getHue()); if(-1 == xioctl(VIDIOCSPICT,&V4L_picture)) kdDebug(14010) << k_funcinfo << "Card seems to not support adjusting image hue. Fallback to it is not yet implemented." << endl; } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } return getHue(); } bool VideoDevice::getAutoBrightnessContrast() { if (m_current_input < m_input.size() ) return m_input[m_current_input].getAutoBrightnessContrast(); else return false; } bool VideoDevice::setAutoBrightnessContrast(bool brightnesscontrast) { kdDebug(14010) << k_funcinfo << "VideoDevice::setAutoBrightnessContrast(" << brightnesscontrast << ") called." << endl; if (m_current_input < m_input.size() ) { m_input[m_current_input].setAutoBrightnessContrast(brightnesscontrast); return m_input[m_current_input].getAutoBrightnessContrast(); } else return false; } bool VideoDevice::getAutoColorCorrection() { if (m_current_input < m_input.size() ) return m_input[m_current_input].getAutoColorCorrection(); else return false; } bool VideoDevice::setAutoColorCorrection(bool colorcorrection) { kdDebug(14010) << k_funcinfo << "VideoDevice::setAutoColorCorrection(" << colorcorrection << ") called." << endl; if (m_current_input < m_input.size() ) { m_input[m_current_input].setAutoColorCorrection(colorcorrection); return m_input[m_current_input].getAutoColorCorrection(); } else return false; } bool VideoDevice::getImageAsMirror() { if (m_current_input < m_input.size() ) return m_input[m_current_input].getImageAsMirror(); else return false; } bool VideoDevice::setImageAsMirror(bool imageasmirror) { kdDebug(14010) << k_funcinfo << "VideoDevice::setImageAsMirror(" << imageasmirror << ") called." << endl; if (m_current_input < m_input.size() ) { m_input[m_current_input].setImageAsMirror(imageasmirror); return m_input[m_current_input].getImageAsMirror(); } else return false; } pixel_format VideoDevice::pixelFormatForPalette( int palette ) { switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: switch(palette) { case 0 : return PIXELFORMAT_NONE; break; // Packed RGB formats case V4L2_PIX_FMT_RGB332 : return PIXELFORMAT_RGB332; break; #if defined( V4L2_PIX_FMT_RGB444 ) case V4L2_PIX_FMT_RGB444 : return PIXELFORMAT_RGB444; break; #endif case V4L2_PIX_FMT_RGB555 : return PIXELFORMAT_RGB555; break; case V4L2_PIX_FMT_RGB565 : return PIXELFORMAT_RGB565; break; case V4L2_PIX_FMT_RGB555X : return PIXELFORMAT_RGB555X; break; case V4L2_PIX_FMT_RGB565X : return PIXELFORMAT_RGB565X; break; case V4L2_PIX_FMT_BGR24 : return PIXELFORMAT_BGR24; break; case V4L2_PIX_FMT_RGB24 : return PIXELFORMAT_RGB24; break; case V4L2_PIX_FMT_BGR32 : return PIXELFORMAT_BGR32; break; case V4L2_PIX_FMT_RGB32 : return PIXELFORMAT_RGB32; break; // Bayer RGB format case V4L2_PIX_FMT_SBGGR8 : return PIXELFORMAT_SBGGR8; break; // YUV formats case V4L2_PIX_FMT_GREY : return PIXELFORMAT_GREY; break; case V4L2_PIX_FMT_YUYV : return PIXELFORMAT_YUYV; break; case V4L2_PIX_FMT_UYVY : return PIXELFORMAT_UYVY; break; case V4L2_PIX_FMT_YUV420 : return PIXELFORMAT_YUV420P; break; case V4L2_PIX_FMT_YUV422P : return PIXELFORMAT_YUV422P; break; // Compressed formats case V4L2_PIX_FMT_JPEG : return PIXELFORMAT_JPEG; break; case V4L2_PIX_FMT_MPEG : return PIXELFORMAT_MPEG; break; // Reserved formats case V4L2_PIX_FMT_DV : return PIXELFORMAT_DV; break; case V4L2_PIX_FMT_ET61X251 : return PIXELFORMAT_ET61X251; break; case V4L2_PIX_FMT_HI240 : return PIXELFORMAT_HI240; break; #if defined( V4L2_PIX_FMT_HM12 ) case V4L2_PIX_FMT_HM12 : return PIXELFORMAT_HM12; break; #endif case V4L2_PIX_FMT_MJPEG : return PIXELFORMAT_MJPEG; break; case V4L2_PIX_FMT_PWC1 : return PIXELFORMAT_PWC1; break; case V4L2_PIX_FMT_PWC2 : return PIXELFORMAT_PWC2; break; case V4L2_PIX_FMT_SN9C10X : return PIXELFORMAT_SN9C10X; break; case V4L2_PIX_FMT_WNVA : return PIXELFORMAT_WNVA; break; case V4L2_PIX_FMT_YYUV : return PIXELFORMAT_YYUV; break; } break; #endif case VIDEODEV_DRIVER_V4L: switch(palette) { case 0 : return PIXELFORMAT_NONE; break; case VIDEO_PALETTE_GREY : return PIXELFORMAT_GREY; break; case VIDEO_PALETTE_HI240 : return PIXELFORMAT_RGB332; break; case VIDEO_PALETTE_RGB555 : return PIXELFORMAT_RGB555; break; case VIDEO_PALETTE_RGB565 : return PIXELFORMAT_RGB565; break; case VIDEO_PALETTE_RGB24 : return PIXELFORMAT_RGB24; break; case VIDEO_PALETTE_RGB32 : return PIXELFORMAT_RGB32; break; case VIDEO_PALETTE_YUYV : return PIXELFORMAT_YUYV; break; case VIDEO_PALETTE_UYVY : return PIXELFORMAT_UYVY; break; #if !defined(__FreeBSD__) case VIDEO_PALETTE_YUV420 : #endif case VIDEO_PALETTE_YUV420P : return PIXELFORMAT_YUV420P; break; case VIDEO_PALETTE_YUV422P : return PIXELFORMAT_YUV422P; break; } break; #endif case VIDEODEV_DRIVER_NONE: default: return PIXELFORMAT_NONE; break; } return PIXELFORMAT_NONE; } int VideoDevice::pixelFormatCode(pixel_format pixelformat) { switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: switch(pixelformat) { case PIXELFORMAT_NONE : return 0; break; // Packed RGB formats case PIXELFORMAT_RGB332 : return V4L2_PIX_FMT_RGB332; break; #if defined( V4L2_PIX_FMT_RGB444 ) case PIXELFORMAT_RGB444 : return V4L2_PIX_FMT_RGB444; break; #endif case PIXELFORMAT_RGB555 : return V4L2_PIX_FMT_RGB555; break; case PIXELFORMAT_RGB565 : return V4L2_PIX_FMT_RGB565; break; case PIXELFORMAT_RGB555X: return V4L2_PIX_FMT_RGB555X; break; case PIXELFORMAT_RGB565X: return V4L2_PIX_FMT_RGB565X; break; case PIXELFORMAT_BGR24 : return V4L2_PIX_FMT_BGR24; break; case PIXELFORMAT_RGB24 : return V4L2_PIX_FMT_RGB24; break; case PIXELFORMAT_BGR32 : return V4L2_PIX_FMT_BGR32; break; case PIXELFORMAT_RGB32 : return V4L2_PIX_FMT_RGB32; break; // Bayer RGB format case PIXELFORMAT_SBGGR8 : return V4L2_PIX_FMT_SBGGR8; break; // YUV formats case PIXELFORMAT_GREY : return V4L2_PIX_FMT_GREY; break; case PIXELFORMAT_YUYV : return V4L2_PIX_FMT_YUYV; break; case PIXELFORMAT_UYVY : return V4L2_PIX_FMT_UYVY; break; case PIXELFORMAT_YUV420P: return V4L2_PIX_FMT_YUV420; break; case PIXELFORMAT_YUV422P: return V4L2_PIX_FMT_YUV422P; break; // Compressed formats case PIXELFORMAT_JPEG : return V4L2_PIX_FMT_JPEG; break; case PIXELFORMAT_MPEG : return V4L2_PIX_FMT_MPEG; break; // Reserved formats case PIXELFORMAT_DV : return V4L2_PIX_FMT_DV; break; case PIXELFORMAT_ET61X251:return V4L2_PIX_FMT_ET61X251;break; case PIXELFORMAT_HI240 : return V4L2_PIX_FMT_HI240; break; #if defined( V4L2_PIX_FMT_HM12 ) case PIXELFORMAT_HM12 : return V4L2_PIX_FMT_HM12; break; #endif case PIXELFORMAT_MJPEG : return V4L2_PIX_FMT_MJPEG; break; case PIXELFORMAT_PWC1 : return V4L2_PIX_FMT_PWC1; break; case PIXELFORMAT_PWC2 : return V4L2_PIX_FMT_PWC2; break; case PIXELFORMAT_SN9C10X: return V4L2_PIX_FMT_SN9C10X; break; case PIXELFORMAT_WNVA : return V4L2_PIX_FMT_WNVA; break; case PIXELFORMAT_YYUV : return V4L2_PIX_FMT_YYUV; break; } break; #endif case VIDEODEV_DRIVER_V4L: switch(pixelformat) { case PIXELFORMAT_NONE : return 0; break; // Packed RGB formats case PIXELFORMAT_RGB332 : return VIDEO_PALETTE_HI240; break; case PIXELFORMAT_RGB444 : return 0; break; case PIXELFORMAT_RGB555 : return VIDEO_PALETTE_RGB555; break; case PIXELFORMAT_RGB565 : return VIDEO_PALETTE_RGB565; break; case PIXELFORMAT_RGB555X: return 0; break; case PIXELFORMAT_RGB565X: return 0; break; case PIXELFORMAT_BGR24 : return 0; break; case PIXELFORMAT_RGB24 : return VIDEO_PALETTE_RGB24; break; case PIXELFORMAT_BGR32 : return 0; break; case PIXELFORMAT_RGB32 : return VIDEO_PALETTE_RGB32; break; // Bayer RGB format case PIXELFORMAT_SBGGR8 : return 0; break; // YUV formats case PIXELFORMAT_GREY : return VIDEO_PALETTE_GREY; break; case PIXELFORMAT_YUYV : return VIDEO_PALETTE_YUYV; break; case PIXELFORMAT_UYVY : return VIDEO_PALETTE_UYVY; break; #if defined(__FreeBSD__) case PIXELFORMAT_YUV420P: return VIDEO_PALETTE_YUV420P; break; #else case PIXELFORMAT_YUV420P: return VIDEO_PALETTE_YUV420; break; #endif case PIXELFORMAT_YUV422P: return VIDEO_PALETTE_YUV422P; break; // Compressed formats case PIXELFORMAT_JPEG : return 0; break; case PIXELFORMAT_MPEG : return 0; break; // Reserved formats case PIXELFORMAT_DV : return 0; break; case PIXELFORMAT_ET61X251:return 0; break; case PIXELFORMAT_HI240 : return VIDEO_PALETTE_HI240; break; case PIXELFORMAT_HM12 : return 0; break; case PIXELFORMAT_MJPEG : return 0; break; case PIXELFORMAT_PWC1 : return 0; break; case PIXELFORMAT_PWC2 : return 0; break; case PIXELFORMAT_SN9C10X: return 0; break; case PIXELFORMAT_WNVA : return 0; break; case PIXELFORMAT_YYUV : return 0; break; } break; #endif case VIDEODEV_DRIVER_NONE: default: return PIXELFORMAT_NONE; break; } return PIXELFORMAT_NONE; } int VideoDevice::pixelFormatDepth(pixel_format pixelformat) { switch(pixelformat) { case PIXELFORMAT_NONE : return 0; break; // Packed RGB formats case PIXELFORMAT_RGB332 : return 8; break; case PIXELFORMAT_RGB444 : return 16; break; case PIXELFORMAT_RGB555 : return 16; break; case PIXELFORMAT_RGB565 : return 16; break; case PIXELFORMAT_RGB555X: return 16; break; case PIXELFORMAT_RGB565X: return 16; break; case PIXELFORMAT_BGR24 : return 24; break; case PIXELFORMAT_RGB24 : return 24; break; case PIXELFORMAT_BGR32 : return 32; break; case PIXELFORMAT_RGB32 : return 32; break; // Bayer RGB format case PIXELFORMAT_SBGGR8 : return 0; break; // YUV formats case PIXELFORMAT_GREY : return 8; break; case PIXELFORMAT_YUYV : return 16; break; case PIXELFORMAT_UYVY : return 16; break; #if defined(__FreeBSD__) case PIXELFORMAT_YUV420P: return 24; break; #else case PIXELFORMAT_YUV420P: return 16; break; #endif case PIXELFORMAT_YUV422P: return 16; break; // Compressed formats case PIXELFORMAT_JPEG : return 0; break; case PIXELFORMAT_MPEG : return 0; break; // Reserved formats case PIXELFORMAT_DV : return 0; break; case PIXELFORMAT_ET61X251:return 0; break; case PIXELFORMAT_HI240 : return 8; break; case PIXELFORMAT_HM12 : return 0; break; case PIXELFORMAT_MJPEG : return 0; break; case PIXELFORMAT_PWC1 : return 0; break; case PIXELFORMAT_PWC2 : return 0; break; case PIXELFORMAT_SN9C10X: return 0; break; case PIXELFORMAT_WNVA : return 0; break; case PIXELFORMAT_YYUV : return 0; break; } return 0; } TQString VideoDevice::pixelFormatName(pixel_format pixelformat) { TQString returnvalue; returnvalue = "None"; switch(pixelformat) { case PIXELFORMAT_NONE : returnvalue = "None"; break; // Packed RGB formats case PIXELFORMAT_RGB332 : returnvalue = "8-bit RGB332"; break; case PIXELFORMAT_RGB444 : returnvalue = "8-bit RGB444"; break; case PIXELFORMAT_RGB555 : returnvalue = "16-bit RGB555"; break; case PIXELFORMAT_RGB565 : returnvalue = "16-bit RGB565"; break; case PIXELFORMAT_RGB555X: returnvalue = "16-bit RGB555X"; break; case PIXELFORMAT_RGB565X: returnvalue = "16-bit RGB565X"; break; case PIXELFORMAT_BGR24 : returnvalue = "24-bit BGR24"; break; case PIXELFORMAT_RGB24 : returnvalue = "24-bit RGB24"; break; case PIXELFORMAT_BGR32 : returnvalue = "32-bit BGR32"; break; case PIXELFORMAT_RGB32 : returnvalue = "32-bit RGB32"; break; // Bayer RGB format case PIXELFORMAT_SBGGR8 : returnvalue = "Bayer RGB format"; break; // YUV formats case PIXELFORMAT_GREY : returnvalue = "8-bit Grayscale"; break; case PIXELFORMAT_YUYV : returnvalue = "Packed YUV 4:2:2"; break; case PIXELFORMAT_UYVY : returnvalue = "Packed YVU 4:2:2"; break; case PIXELFORMAT_YUV420P: returnvalue = "Planar YUV 4:2:0"; break; case PIXELFORMAT_YUV422P: returnvalue = "Planar YUV 4:2:2"; break; // Compressed formats case PIXELFORMAT_JPEG : returnvalue = "JPEG image"; break; case PIXELFORMAT_MPEG : returnvalue = "MPEG stream"; break; // Reserved formats case PIXELFORMAT_DV : returnvalue = "DV (unknown)"; break; case PIXELFORMAT_ET61X251:returnvalue = "ET61X251"; break; case PIXELFORMAT_HI240 : returnvalue = "8-bit HI240 (RGB332)"; break; case PIXELFORMAT_HM12 : returnvalue = "Packed YUV 4:2:2"; break; case PIXELFORMAT_MJPEG : returnvalue = "8-bit Grayscale"; break; case PIXELFORMAT_PWC1 : returnvalue = "PWC1"; break; case PIXELFORMAT_PWC2 : returnvalue = "PWC2"; break; case PIXELFORMAT_SN9C10X: returnvalue = "SN9C102"; break; case PIXELFORMAT_WNVA : returnvalue = "Winnov Videum"; break; case PIXELFORMAT_YYUV : returnvalue = "YYUV (unknown)"; break; } return returnvalue; } TQString VideoDevice::pixelFormatName(int pixelformat) { TQString returnvalue; returnvalue = "None"; switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: switch(pixelformat) { case 0 : returnvalue = pixelFormatName(PIXELFORMAT_NONE); break; // Packed RGB formats case V4L2_PIX_FMT_RGB332 : returnvalue = pixelFormatName(PIXELFORMAT_RGB332); break; #if defined( V4L2_PIX_FMT_RGB444 ) case V4L2_PIX_FMT_RGB444 : returnvalue = pixelFormatName(PIXELFORMAT_RGB444); break; #endif case V4L2_PIX_FMT_RGB555 : returnvalue = pixelFormatName(PIXELFORMAT_RGB555); break; case V4L2_PIX_FMT_RGB565 : returnvalue = pixelFormatName(PIXELFORMAT_RGB565); break; case V4L2_PIX_FMT_RGB555X : returnvalue = pixelFormatName(PIXELFORMAT_RGB555X); break; case V4L2_PIX_FMT_RGB565X : returnvalue = pixelFormatName(PIXELFORMAT_RGB565X); break; case V4L2_PIX_FMT_BGR24 : returnvalue = pixelFormatName(PIXELFORMAT_BGR24); break; case V4L2_PIX_FMT_RGB24 : returnvalue = pixelFormatName(PIXELFORMAT_RGB24); break; case V4L2_PIX_FMT_BGR32 : returnvalue = pixelFormatName(PIXELFORMAT_BGR32); break; case V4L2_PIX_FMT_RGB32 : returnvalue = pixelFormatName(PIXELFORMAT_RGB32); break; // Bayer RGB format case V4L2_PIX_FMT_SBGGR8 : returnvalue = pixelFormatName(PIXELFORMAT_SBGGR8); break; // YUV formats case V4L2_PIX_FMT_GREY : returnvalue = pixelFormatName(PIXELFORMAT_GREY); break; case V4L2_PIX_FMT_YUYV : returnvalue = pixelFormatName(PIXELFORMAT_YUYV); break; case V4L2_PIX_FMT_UYVY : returnvalue = pixelFormatName(PIXELFORMAT_UYVY); break; case V4L2_PIX_FMT_YUV420 : returnvalue = pixelFormatName(PIXELFORMAT_YUV420P); break; case V4L2_PIX_FMT_YUV422P : returnvalue = pixelFormatName(PIXELFORMAT_YUV422P); break; // Compressed formats case V4L2_PIX_FMT_JPEG : returnvalue = pixelFormatName(PIXELFORMAT_JPEG); break; case V4L2_PIX_FMT_MPEG : returnvalue = pixelFormatName(PIXELFORMAT_MPEG); break; // Reserved formats case V4L2_PIX_FMT_DV : returnvalue = pixelFormatName(PIXELFORMAT_DV); break; case V4L2_PIX_FMT_ET61X251 : returnvalue = pixelFormatName(PIXELFORMAT_ET61X251); break; case V4L2_PIX_FMT_HI240 : returnvalue = pixelFormatName(PIXELFORMAT_HI240); break; #if defined( V4L2_PIX_FMT_HM12 ) case V4L2_PIX_FMT_HM12 : returnvalue = pixelFormatName(PIXELFORMAT_HM12); break; #endif case V4L2_PIX_FMT_MJPEG : returnvalue = pixelFormatName(PIXELFORMAT_MJPEG); break; case V4L2_PIX_FMT_PWC1 : returnvalue = pixelFormatName(PIXELFORMAT_PWC1); break; case V4L2_PIX_FMT_PWC2 : returnvalue = pixelFormatName(PIXELFORMAT_PWC2); break; case V4L2_PIX_FMT_SN9C10X : returnvalue = pixelFormatName(PIXELFORMAT_SN9C10X); break; case V4L2_PIX_FMT_WNVA : returnvalue = pixelFormatName(PIXELFORMAT_WNVA); break; case V4L2_PIX_FMT_YYUV : returnvalue = pixelFormatName(PIXELFORMAT_YYUV); break; } break; #endif case VIDEODEV_DRIVER_V4L: switch(pixelformat) { case VIDEO_PALETTE_GREY : returnvalue = pixelFormatName(PIXELFORMAT_GREY); break; case VIDEO_PALETTE_HI240 : returnvalue = pixelFormatName(PIXELFORMAT_RGB332); break; case VIDEO_PALETTE_RGB555 : returnvalue = pixelFormatName(PIXELFORMAT_RGB555); break; case VIDEO_PALETTE_RGB565 : returnvalue = pixelFormatName(PIXELFORMAT_RGB565); break; case VIDEO_PALETTE_RGB24 : returnvalue = pixelFormatName(PIXELFORMAT_RGB24); break; case VIDEO_PALETTE_RGB32 : returnvalue = pixelFormatName(PIXELFORMAT_RGB32); break; case VIDEO_PALETTE_YUYV : returnvalue = pixelFormatName(PIXELFORMAT_YUYV); break; case VIDEO_PALETTE_UYVY : returnvalue = pixelFormatName(PIXELFORMAT_UYVY); break; #if !defined(__FreeBSD__) case VIDEO_PALETTE_YUV420 : #endif case VIDEO_PALETTE_YUV420P : returnvalue = pixelFormatName(PIXELFORMAT_YUV420P); break; case VIDEO_PALETTE_YUV422P : returnvalue = pixelFormatName(PIXELFORMAT_YUV422P); break; } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } return returnvalue; } int VideoDevice::detectPixelFormats() { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE int err = 0; #endif #endif switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: fmtdesc.index = 0; fmtdesc.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; while ( err == 0 ) { if (-1 == xioctl (VIDIOC_ENUM_FMT, &fmtdesc)) // if (ioctl(fd, VIDIOC_ENUM_FMT, &fmtdesc) < 0 ) { perror("VIDIOC_ENUM_FMT"); err = errno; } else { kdDebug(14010) << k_funcinfo << fmtdesc.pixelformat << " " << pixelFormatName(fmtdesc.pixelformat) << endl; // Need a cleanup. PixelFormatForPalette is a really bad name fmtdesc.index++; } } // break; #endif case VIDEODEV_DRIVER_V4L: // TODO: THis thing can be used to detec what pixel formats are supported in a API-independent way, but V4L2 has VIDIOC_ENUM_PIXFMT. // The correct thing to do is to isolate these calls and do a proper implementation for V4L and another for V4L2 when this thing will be migrated to a plugin architecture. // Packed RGB formats kdDebug(14010) << k_funcinfo << "Supported pixel formats:" << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB332)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB332) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB444)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB444) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB555)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB555) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB565)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB565) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB555X)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB555X) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB565X)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB565X) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_BGR24)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_BGR24) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB24)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB24) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_BGR32)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_BGR32) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_RGB32)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_RGB32) << endl; // Bayer RGB format if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_SBGGR8)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_SBGGR8) << endl; // YUV formats if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_GREY)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_GREY) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YUYV)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YUYV) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_UYVY)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_UYVY) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YUV420P)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YUV420P) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YUV422P)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YUV422P) << endl; // Compressed formats if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_JPEG)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_JPEG) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_MPEG)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_MPEG) << endl; // Reserved formats if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_DV)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_DV) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_ET61X251)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_ET61X251) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_HI240)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_HI240) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_HM12)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_HM12) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_MJPEG)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_MJPEG) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_PWC1)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_PWC1) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_PWC2)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_PWC2) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_SN9C10X)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_SN9C10X) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_WNVA)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_WNVA) << endl; if(PIXELFORMAT_NONE != setPixelFormat(PIXELFORMAT_YYUV)) kdDebug(14010) << k_funcinfo << pixelFormatName(PIXELFORMAT_YYUV) << endl; break; #endif case VIDEODEV_DRIVER_NONE: default: return PIXELFORMAT_NONE; break; } return PIXELFORMAT_NONE; } __u64 VideoDevice::signalStandardCode(signal_standard standard) { switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: switch(standard) { case STANDARD_NONE : return V4L2_STD_UNKNOWN; break; case STANDARD_PAL_B : return V4L2_STD_PAL_B; break; case STANDARD_PAL_B1 : return V4L2_STD_PAL_B1; break; case STANDARD_PAL_G : return V4L2_STD_PAL_G; break; case STANDARD_PAL_H : return V4L2_STD_PAL_H; break; case STANDARD_PAL_I : return V4L2_STD_PAL_I; break; case STANDARD_PAL_D : return V4L2_STD_PAL_D; break; case STANDARD_PAL_D1 : return V4L2_STD_PAL_D1; break; case STANDARD_PAL_K : return V4L2_STD_PAL_K; break; case STANDARD_PAL_M : return V4L2_STD_PAL_M; break; case STANDARD_PAL_N : return V4L2_STD_PAL_N; break; case STANDARD_PAL_Nc : return V4L2_STD_PAL_Nc; break; case STANDARD_PAL_60 : return V4L2_STD_PAL_60; break; case STANDARD_NTSC_M : return V4L2_STD_NTSC_M; break; case STANDARD_NTSC_M_JP : return V4L2_STD_NTSC_M_JP; break; case STANDARD_NTSC_443 : return V4L2_STD_NTSC; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id case STANDARD_SECAM_B : return V4L2_STD_SECAM_B; break; case STANDARD_SECAM_D : return V4L2_STD_SECAM_D; break; case STANDARD_SECAM_G : return V4L2_STD_SECAM_G; break; case STANDARD_SECAM_H : return V4L2_STD_SECAM_H; break; case STANDARD_SECAM_K : return V4L2_STD_SECAM_K; break; case STANDARD_SECAM_K1 : return V4L2_STD_SECAM_K1; break; case STANDARD_SECAM_L : return V4L2_STD_SECAM_L; break; case STANDARD_SECAM_LC : return V4L2_STD_SECAM; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id case STANDARD_ATSC_8_VSB : return V4L2_STD_ATSC_8_VSB; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h case STANDARD_ATSC_16_VSB : return V4L2_STD_ATSC_16_VSB; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h case STANDARD_PAL_BG : return V4L2_STD_PAL_BG; break; case STANDARD_PAL_DK : return V4L2_STD_PAL_DK; break; case STANDARD_PAL : return V4L2_STD_PAL; break; case STANDARD_NTSC : return V4L2_STD_NTSC; break; case STANDARD_SECAM_DK : return V4L2_STD_SECAM_DK; break; case STANDARD_SECAM : return V4L2_STD_SECAM; break; case STANDARD_525_60 : return V4L2_STD_525_60; break; case STANDARD_625_50 : return V4L2_STD_625_50; break; case STANDARD_ALL : return V4L2_STD_ALL; break; } break; #endif case VIDEODEV_DRIVER_V4L: switch(standard) { case STANDARD_NONE : return VIDEO_MODE_AUTO; break; case STANDARD_PAL_B : return VIDEO_MODE_PAL; break; case STANDARD_PAL_B1 : return VIDEO_MODE_PAL; break; case STANDARD_PAL_G : return VIDEO_MODE_PAL; break; case STANDARD_PAL_H : return VIDEO_MODE_PAL; break; case STANDARD_PAL_I : return VIDEO_MODE_PAL; break; case STANDARD_PAL_D : return VIDEO_MODE_PAL; break; case STANDARD_PAL_D1 : return VIDEO_MODE_PAL; break; case STANDARD_PAL_K : return VIDEO_MODE_PAL; break; case STANDARD_PAL_M : return 5; break; // Undocumented value found to be compatible with V4L bttv driver case STANDARD_PAL_N : return 6; break; // Undocumented value found to be compatible with V4L bttv driver case STANDARD_PAL_Nc : return 4; break; // Undocumented value found to be compatible with V4L bttv driver case STANDARD_PAL_60 : return VIDEO_MODE_PAL; break; case STANDARD_NTSC_M : return VIDEO_MODE_NTSC; break; case STANDARD_NTSC_M_JP : return 7; break; // Undocumented value found to be compatible with V4L bttv driver case STANDARD_NTSC_443 : return VIDEO_MODE_NTSC; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id case STANDARD_SECAM_B : return VIDEO_MODE_SECAM; break; case STANDARD_SECAM_D : return VIDEO_MODE_SECAM; break; case STANDARD_SECAM_G : return VIDEO_MODE_SECAM; break; case STANDARD_SECAM_H : return VIDEO_MODE_SECAM; break; case STANDARD_SECAM_K : return VIDEO_MODE_SECAM; break; case STANDARD_SECAM_K1 : return VIDEO_MODE_SECAM; break; case STANDARD_SECAM_L : return VIDEO_MODE_SECAM; break; case STANDARD_SECAM_LC : return VIDEO_MODE_SECAM; break; // Using workaround value because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id case STANDARD_ATSC_8_VSB : return VIDEO_MODE_AUTO; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h case STANDARD_ATSC_16_VSB : return VIDEO_MODE_AUTO; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h case STANDARD_PAL_BG : return VIDEO_MODE_PAL; break; case STANDARD_PAL_DK : return VIDEO_MODE_PAL; break; case STANDARD_PAL : return VIDEO_MODE_PAL; break; case STANDARD_NTSC : return VIDEO_MODE_NTSC; break; case STANDARD_SECAM_DK : return VIDEO_MODE_SECAM; break; case STANDARD_SECAM : return VIDEO_MODE_SECAM; break; case STANDARD_525_60 : return VIDEO_MODE_PAL; break; case STANDARD_625_50 : return VIDEO_MODE_SECAM; break; case STANDARD_ALL : return VIDEO_MODE_AUTO; break; } break; #endif case VIDEODEV_DRIVER_NONE: default: return STANDARD_NONE; break; } return STANDARD_NONE; } TQString VideoDevice::signalStandardName(signal_standard standard) { TQString returnvalue; returnvalue = "None"; switch(standard) { case STANDARD_NONE : returnvalue = "None"; break; case STANDARD_PAL_B : returnvalue = "PAL-B"; break; case STANDARD_PAL_B1 : returnvalue = "PAL-B1"; break; case STANDARD_PAL_G : returnvalue = "PAL-G"; break; case STANDARD_PAL_H : returnvalue = "PAL-H"; break; case STANDARD_PAL_I : returnvalue = "PAL-I"; break; case STANDARD_PAL_D : returnvalue = "PAL-D"; break; case STANDARD_PAL_D1 : returnvalue = "PAL-D1"; break; case STANDARD_PAL_K : returnvalue = "PAL-K"; break; case STANDARD_PAL_M : returnvalue = "PAL-M"; break; case STANDARD_PAL_N : returnvalue = "PAL-N"; break; case STANDARD_PAL_Nc : returnvalue = "PAL-Nc"; break; case STANDARD_PAL_60 : returnvalue = "PAL-60"; break; case STANDARD_NTSC_M : returnvalue = "NTSC-M"; break; case STANDARD_NTSC_M_JP : returnvalue = "NTSC-M(JP)"; break; case STANDARD_NTSC_443 : returnvalue = "NTSC-443"; break; case STANDARD_SECAM_B : returnvalue = "SECAM-B"; break; case STANDARD_SECAM_D : returnvalue = "SECAM-D"; break; case STANDARD_SECAM_G : returnvalue = "SECAM-G"; break; case STANDARD_SECAM_H : returnvalue = "SECAM-H"; break; case STANDARD_SECAM_K : returnvalue = "SECAM-K"; break; case STANDARD_SECAM_K1 : returnvalue = "SECAM-K1"; break; case STANDARD_SECAM_L : returnvalue = "SECAM-L"; break; case STANDARD_SECAM_LC : returnvalue = "SECAM-LC"; break; case STANDARD_ATSC_8_VSB : returnvalue = "ATSC-8-VSB"; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h case STANDARD_ATSC_16_VSB : returnvalue = "ATSC-16-VSB"; break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h case STANDARD_PAL_BG : returnvalue = "PAL-BG"; break; case STANDARD_PAL_DK : returnvalue = "PAL-DK"; break; case STANDARD_PAL : returnvalue = "PAL"; break; case STANDARD_NTSC : returnvalue = "NTSC"; break; case STANDARD_SECAM_DK : returnvalue = "SECAM-DK"; break; case STANDARD_SECAM : returnvalue = "SECAM"; break; case STANDARD_525_60 : returnvalue = "525 lines 60Hz"; break; case STANDARD_625_50 : returnvalue = "625 lines 50Hz"; break; case STANDARD_ALL : returnvalue = "All"; break; } return returnvalue; } TQString VideoDevice::signalStandardName(int standard) { TQString returnvalue; returnvalue = "None"; switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: switch(standard) { case V4L2_STD_PAL_B : returnvalue = signalStandardName(STANDARD_PAL_B); break; case V4L2_STD_PAL_B1 : returnvalue = signalStandardName(STANDARD_PAL_B1); break; case V4L2_STD_PAL_G : returnvalue = signalStandardName(STANDARD_PAL_G); break; case V4L2_STD_PAL_H : returnvalue = signalStandardName(STANDARD_PAL_H); break; case V4L2_STD_PAL_I : returnvalue = signalStandardName(STANDARD_PAL_I); break; case V4L2_STD_PAL_D : returnvalue = signalStandardName(STANDARD_PAL_D); break; case V4L2_STD_PAL_D1 : returnvalue = signalStandardName(STANDARD_PAL_D1); break; case V4L2_STD_PAL_K : returnvalue = signalStandardName(STANDARD_PAL_K); break; case V4L2_STD_PAL_M : returnvalue = signalStandardName(STANDARD_PAL_M); break; case V4L2_STD_PAL_N : returnvalue = signalStandardName(STANDARD_PAL_N); break; case V4L2_STD_PAL_Nc : returnvalue = signalStandardName(STANDARD_PAL_Nc); break; case V4L2_STD_PAL_60 : returnvalue = signalStandardName(STANDARD_PAL_60); break; case V4L2_STD_NTSC_M : returnvalue = signalStandardName(STANDARD_NTSC_M); break; case V4L2_STD_NTSC_M_JP : returnvalue = signalStandardName(STANDARD_NTSC_M_JP); break; // case V4L2_STD_NTSC_443 : returnvalue = signalStandardName(STANDARD_NTSC_443); break; // Commented out because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id case V4L2_STD_SECAM_B : returnvalue = signalStandardName(STANDARD_SECAM_B); break; case V4L2_STD_SECAM_D : returnvalue = signalStandardName(STANDARD_SECAM_D); break; case V4L2_STD_SECAM_G : returnvalue = signalStandardName(STANDARD_SECAM_G); break; case V4L2_STD_SECAM_H : returnvalue = signalStandardName(STANDARD_SECAM_H); break; case V4L2_STD_SECAM_K : returnvalue = signalStandardName(STANDARD_SECAM_K); break; case V4L2_STD_SECAM_K1 : returnvalue = signalStandardName(STANDARD_SECAM_K1); break; case V4L2_STD_SECAM_L : returnvalue = signalStandardName(STANDARD_SECAM_L); break; // case V4L2_STD_SECAM_LC : returnvalue = signalStandardName(STANDARD_SECAM_LC); break; // Commented out because my videodev2.h header seems to not include this standard in struct __u64 v4l2_std_id case V4L2_STD_ATSC_8_VSB : returnvalue = signalStandardName(STANDARD_ATSC_8_VSB); break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h case V4L2_STD_ATSC_16_VSB : returnvalue = signalStandardName(STANDARD_ATSC_16_VSB); break; // ATSC/HDTV Standard officially not supported by V4L2 but exists in videodev2.h case V4L2_STD_PAL_BG : returnvalue = signalStandardName(STANDARD_PAL_BG); break; case V4L2_STD_PAL_DK : returnvalue = signalStandardName(STANDARD_PAL_DK); break; case V4L2_STD_PAL : returnvalue = signalStandardName(STANDARD_PAL); break; case V4L2_STD_NTSC : returnvalue = signalStandardName(STANDARD_NTSC); break; case V4L2_STD_SECAM_DK : returnvalue = signalStandardName(STANDARD_SECAM_DK); break; case V4L2_STD_SECAM : returnvalue = signalStandardName(STANDARD_SECAM); break; case V4L2_STD_525_60 : returnvalue = signalStandardName(STANDARD_525_60); break; case V4L2_STD_625_50 : returnvalue = signalStandardName(STANDARD_625_50); break; case V4L2_STD_ALL : returnvalue = signalStandardName(STANDARD_ALL); break; } break; #endif case VIDEODEV_DRIVER_V4L: switch(standard) { case VIDEO_MODE_PAL : returnvalue = signalStandardName(STANDARD_PAL); break; case VIDEO_MODE_NTSC : returnvalue = signalStandardName(STANDARD_NTSC); break; case VIDEO_MODE_SECAM : returnvalue = signalStandardName(STANDARD_SECAM); break; case VIDEO_MODE_AUTO : returnvalue = signalStandardName(STANDARD_ALL); break; // It must be disabled until I find a correct way to handle those non-standard bttv modes // case VIDEO_MODE_PAL_Nc : returnvalue = signalStandardName(STANDARD_PAL_Nc); break; // Undocumented value found to be compatible with V4L bttv driver case VIDEO_MODE_PAL_M : returnvalue = signalStandardName(STANDARD_PAL_M); break; // Undocumented value found to be compatible with V4L bttv driver case VIDEO_MODE_PAL_N : returnvalue = signalStandardName(STANDARD_PAL_N); break; // Undocumented value found to be compatible with V4L bttv driver case VIDEO_MODE_NTSC_JP : returnvalue = signalStandardName(STANDARD_NTSC_M_JP); break; // Undocumented value found to be compatible with V4L bttv driver } break; #endif case VIDEODEV_DRIVER_NONE: default: break; } return returnvalue; } /*! \fn VideoDevice::detectSignalStandards() */ int VideoDevice::detectSignalStandards() { switch(m_driver) { #if (defined(__linux__) || defined(__FreeBSD__)) && defined(ENABLE_AV) #ifdef V4L2_CAP_VIDEO_CAPTURE case VIDEODEV_DRIVER_V4L2: break; #endif case VIDEODEV_DRIVER_V4L: break; #endif case VIDEODEV_DRIVER_NONE: default: break; } //FIXME: return a real value return 0; } /*! \fn VideoDevice::initRead() */ int VideoDevice::initRead() { /// @todo implement me kdDebug(14010) << k_funcinfo << "called." << endl; if(isOpen()) { m_rawbuffers.resize(1); if (m_rawbuffers.size()==0) { fprintf (stderr, "Out of memory\n"); return EXIT_FAILURE; } kdDebug(14010) << k_funcinfo << "m_buffer_size: " << m_buffer_size << endl; // m_rawbuffers[0].pixelformat=m_pixelformat; m_rawbuffers[0].length = m_buffer_size; m_rawbuffers[0].start = (uchar *)malloc (m_buffer_size); if (!m_rawbuffers[0].start) { fprintf (stderr, "Out of memory\n"); return EXIT_FAILURE; } kdDebug(14010) << k_funcinfo << "exited successfuly." << endl; return EXIT_SUCCESS; } return EXIT_FAILURE; } /*! \fn VideoDevice::initMmap() */ int VideoDevice::initMmap() { /// @todo implement me #define BUFFERS 2 if(isOpen()) { kdDebug(14010) << k_funcinfo << full_filename << " Trying to MMAP" << endl; #ifdef V4L2_CAP_VIDEO_CAPTURE struct v4l2_requestbuffers req; CLEAR (req); req.count = BUFFERS; req.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; req.memory = V4L2_MEMORY_MMAP; if (-1 == xioctl (VIDIOC_REQBUFS, &req)) { if (EINVAL == errno) { kdDebug(14010) << k_funcinfo << full_filename << " does not support memory mapping" << endl; return EXIT_FAILURE; } else { return errnoReturn ("VIDIOC_REQBUFS"); } } if (req.count < BUFFERS) { kdDebug(14010) << k_funcinfo << "Insufficient buffer memory on " << full_filename << endl; return EXIT_FAILURE; } m_rawbuffers.resize(req.count); if (m_rawbuffers.size()==0) { kdDebug(14010) << k_funcinfo << "Out of memory" << endl; return EXIT_FAILURE; } for (m_streambuffers = 0; m_streambuffers < req.count; ++m_streambuffers) { struct v4l2_buffer v4l2buffer; CLEAR (v4l2buffer); v4l2buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; v4l2buffer.memory = V4L2_MEMORY_MMAP; v4l2buffer.index = m_streambuffers; if (-1 == xioctl (VIDIOC_QUERYBUF, &v4l2buffer)) return errnoReturn ("VIDIOC_QUERYBUF"); m_rawbuffers[m_streambuffers].length = v4l2buffer.length; m_rawbuffers[m_streambuffers].start = (uchar *) mmap (NULL /* start anywhere */, v4l2buffer.length, PROT_READ | PROT_WRITE /* required */, MAP_SHARED /* recommended */, descriptor, v4l2buffer.m.offset); if (MAP_FAILED == m_rawbuffers[m_streambuffers].start) return errnoReturn ("mmap"); } #endif m_currentbuffer.data.resize(m_rawbuffers[0].length); // Makes the imagesize.data buffer size equal to the rawbuffer size kdDebug(14010) << k_funcinfo << full_filename << " m_currentbuffer.data.size(): " << m_currentbuffer.data.size() << endl; return EXIT_SUCCESS; } return EXIT_FAILURE; } /*! \fn VideoDevice::initUserptr() */ int VideoDevice::initUserptr() { /// @todo implement me if(isOpen()) { #ifdef V4L2_CAP_VIDEO_CAPTURE struct v4l2_requestbuffers req; CLEAR (req); req.count = 2; req.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; req.memory = V4L2_MEMORY_USERPTR; if (-1 == xioctl (VIDIOC_REQBUFS, &req)) { if (EINVAL == errno) { kdDebug(14010) << k_funcinfo << full_filename << " does not support memory mapping" << endl; return EXIT_FAILURE; } else { return errnoReturn ("VIDIOC_REQBUFS"); } } m_rawbuffers.resize(4); if (m_rawbuffers.size()==0) { fprintf (stderr, "Out of memory\n"); return EXIT_FAILURE; } for (m_streambuffers = 0; m_streambuffers < 4; ++m_streambuffers) { m_rawbuffers[m_streambuffers].length = m_buffer_size; m_rawbuffers[m_streambuffers].start = (uchar *) malloc (m_buffer_size); if (!m_rawbuffers[m_streambuffers].start) { kdDebug(14010) << k_funcinfo << "Out of memory" << endl; return EXIT_FAILURE; } } #endif return EXIT_SUCCESS; } return EXIT_FAILURE; } bool VideoDevice::canCapture() { return m_videocapture; } bool VideoDevice::canChromakey() { return m_videochromakey; } bool VideoDevice::canScale() { return m_videoscale; } bool VideoDevice::canOverlay() { return m_videooverlay; } bool VideoDevice::canRead() { return m_videoread; } bool VideoDevice::canAsyncIO() { return m_videoasyncio; } bool VideoDevice::canStream() { return m_videostream; } } }