| /* |
| * GStreamer OpenNI2 device source element |
| * Copyright (C) 2013 Miguel Casas-Sanchez <miguelecasassanchez@gmail.com> |
| |
| * This library is free software; you can |
| * redistribute it and/or modify it under the terms of the GNU Library |
| * General Public License as published by the Free Software Foundation; |
| * either version 2 of the License, or (at your option) any later version. |
| * This library is distributed in the hope that it will be useful, but |
| * WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Library |
| * General Public License for more details. You should have received a copy |
| * of the GNU Library General Public License along with this library; if |
| * not, write to the Free Software Foundation, Inc., 51 Franklin St, |
| * Fifth Floor, Boston, MA 02110-1301, USA. |
| */ |
| |
| /** |
| * SECTION:element-openni2src |
| * |
| * <refsect2> |
| * <title>Examples</title> |
| * <para> |
| * Some recorded .oni files are available at: |
| * <programlisting> |
| * http://people.cs.pitt.edu/~chang/1635/proj11/kinectRecord |
| * </programlisting> |
| * |
| * <programlisting> |
| LD_LIBRARY_PATH=/usr/lib/OpenNI2/Drivers/ gst-launch-1.0 --gst-debug=openni2src:5 openni2src location='Downloads/mr.oni' sourcetype=depth ! videoconvert ! ximagesink |
| * </programlisting> |
| * <programlisting> |
| LD_LIBRARY_PATH=/usr/lib/OpenNI2/Drivers/ gst-launch-1.0 --gst-debug=openni2src:5 openni2src location='Downloads/mr.oni' sourcetype=color ! videoconvert ! ximagesink |
| * </programlisting> |
| * </para> |
| * </refsect2> |
| */ |
| |
| #ifdef HAVE_CONFIG_H |
| #include "config.h" |
| #endif |
| |
| #include "gstopenni2src.h" |
| |
| GST_DEBUG_CATEGORY_STATIC (openni2src_debug); |
| #define GST_CAT_DEFAULT openni2src_debug |
| static GstStaticPadTemplate srctemplate = GST_STATIC_PAD_TEMPLATE ("src", |
| GST_PAD_SRC, |
| GST_PAD_ALWAYS, |
| GST_STATIC_CAPS (GST_VIDEO_CAPS_MAKE ("{RGBA, RGB, GRAY16_LE}")) |
| ); |
| |
| enum |
| { |
| PROP_0, |
| PROP_LOCATION, |
| PROP_SOURCETYPE |
| }; |
| typedef enum |
| { |
| SOURCETYPE_DEPTH, |
| SOURCETYPE_COLOR, |
| SOURCETYPE_BOTH |
| } GstOpenni2SourceType; |
| #define DEFAULT_SOURCETYPE SOURCETYPE_DEPTH |
| |
| #define SAMPLE_READ_WAIT_TIMEOUT 2000 /* 2000ms */ |
| |
| #define GST_TYPE_OPENNI2_SRC_SOURCETYPE (gst_openni2_src_sourcetype_get_type ()) |
| static GType |
| gst_openni2_src_sourcetype_get_type (void) |
| { |
| static GType etype = 0; |
| if (etype == 0) { |
| static const GEnumValue values[] = { |
| {SOURCETYPE_DEPTH, "Get depth readings", "depth"}, |
| {SOURCETYPE_COLOR, "Get color readings", "color"}, |
| {SOURCETYPE_BOTH, |
| "Get color and depth (as alpha) readings - EXPERIMENTAL", |
| "both"}, |
| {0, NULL, NULL}, |
| }; |
| etype = g_enum_register_static ("GstOpenni2SrcSourcetype", values); |
| } |
| return etype; |
| } |
| |
| /* GObject methods */ |
| static void gst_openni2_src_dispose (GObject * object); |
| static void gst_openni2_src_finalize (GObject * gobject); |
| static void gst_openni2_src_set_property (GObject * object, guint prop_id, |
| const GValue * value, GParamSpec * pspec); |
| static void gst_openni2_src_get_property (GObject * object, guint prop_id, |
| GValue * value, GParamSpec * pspec); |
| |
| /* basesrc methods */ |
| static gboolean gst_openni2_src_start (GstBaseSrc * bsrc); |
| static gboolean gst_openni2_src_stop (GstBaseSrc * bsrc); |
| static gboolean gst_openni2_src_set_caps (GstBaseSrc * src, GstCaps * caps); |
| static GstCaps *gst_openni2_src_get_caps (GstBaseSrc * src, GstCaps * filter); |
| static gboolean gst_openni2src_decide_allocation (GstBaseSrc * bsrc, |
| GstQuery * query); |
| |
| /* element methods */ |
| static GstStateChangeReturn gst_openni2_src_change_state (GstElement * element, |
| GstStateChange transition); |
| |
| /* pushsrc method */ |
| static GstFlowReturn gst_openni2src_fill (GstPushSrc * src, GstBuffer * buf); |
| |
| /* OpenNI2 interaction methods */ |
| static gboolean openni2_initialise_library (); |
| static gboolean openni2_initialise_devices (GstOpenni2Src * src); |
| static GstFlowReturn openni2_read_gstbuffer (GstOpenni2Src * src, |
| GstBuffer * buf); |
| |
| #define parent_class gst_openni2_src_parent_class |
| G_DEFINE_TYPE (GstOpenni2Src, gst_openni2_src, GST_TYPE_PUSH_SRC); |
| |
| static void |
| gst_openni2_src_class_init (GstOpenni2SrcClass * klass) |
| { |
| GObjectClass *gobject_class; |
| GstPushSrcClass *pushsrc_class; |
| GstBaseSrcClass *basesrc_class; |
| GstElementClass *element_class = GST_ELEMENT_CLASS (klass); |
| |
| gobject_class = (GObjectClass *) klass; |
| basesrc_class = (GstBaseSrcClass *) klass; |
| pushsrc_class = (GstPushSrcClass *) klass; |
| |
| gobject_class->dispose = gst_openni2_src_dispose; |
| gobject_class->finalize = gst_openni2_src_finalize; |
| gobject_class->set_property = gst_openni2_src_set_property; |
| gobject_class->get_property = gst_openni2_src_get_property; |
| g_object_class_install_property |
| (gobject_class, PROP_LOCATION, |
| g_param_spec_string ("location", "Location", |
| "Source uri, can be a file or a device.", "", (GParamFlags) |
| (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); |
| g_object_class_install_property (gobject_class, PROP_SOURCETYPE, |
| g_param_spec_enum ("sourcetype", |
| "Device source type", |
| "Type of readings to get from the source", |
| GST_TYPE_OPENNI2_SRC_SOURCETYPE, DEFAULT_SOURCETYPE, |
| (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); |
| |
| |
| basesrc_class->start = GST_DEBUG_FUNCPTR (gst_openni2_src_start); |
| basesrc_class->stop = GST_DEBUG_FUNCPTR (gst_openni2_src_stop); |
| basesrc_class->get_caps = GST_DEBUG_FUNCPTR (gst_openni2_src_get_caps); |
| basesrc_class->set_caps = GST_DEBUG_FUNCPTR (gst_openni2_src_set_caps); |
| basesrc_class->decide_allocation = |
| GST_DEBUG_FUNCPTR (gst_openni2src_decide_allocation); |
| |
| gst_element_class_add_static_pad_template (element_class, &srctemplate); |
| |
| gst_element_class_set_static_metadata (element_class, "Openni2 client source", |
| "Source/Video", |
| "Extract readings from an OpenNI supported device (Kinect etc). ", |
| "Miguel Casas-Sanchez <miguelecasassanchez@gmail.com>"); |
| |
| element_class->change_state = gst_openni2_src_change_state; |
| |
| pushsrc_class->fill = GST_DEBUG_FUNCPTR (gst_openni2src_fill); |
| |
| GST_DEBUG_CATEGORY_INIT (openni2src_debug, "openni2src", 0, |
| "OpenNI2 Device Source"); |
| |
| /* OpenNI2 initialisation inside this function */ |
| openni2_initialise_library (); |
| } |
| |
| static void |
| gst_openni2_src_init (GstOpenni2Src * ni2src) |
| { |
| gst_base_src_set_live (GST_BASE_SRC (ni2src), TRUE); |
| gst_base_src_set_format (GST_BASE_SRC (ni2src), GST_FORMAT_TIME); |
| |
| ni2src->device = new openni::Device (); |
| ni2src->depth = new openni::VideoStream (); |
| ni2src->color = new openni::VideoStream (); |
| ni2src->depthFrame = new openni::VideoFrameRef (); |
| ni2src->colorFrame = new openni::VideoFrameRef (); |
| |
| ni2src->oni_start_ts = GST_CLOCK_TIME_NONE; |
| } |
| |
| static void |
| gst_openni2_src_dispose (GObject * object) |
| { |
| GstOpenni2Src *ni2src = GST_OPENNI2_SRC (object); |
| |
| if (ni2src->gst_caps) |
| gst_caps_unref (ni2src->gst_caps); |
| |
| G_OBJECT_CLASS (parent_class)->dispose (object); |
| } |
| |
| static void |
| gst_openni2_src_finalize (GObject * gobject) |
| { |
| GstOpenni2Src *ni2src = GST_OPENNI2_SRC (gobject); |
| |
| if (ni2src->uri_name) { |
| g_free (ni2src->uri_name); |
| ni2src->uri_name = NULL; |
| } |
| |
| if (ni2src->gst_caps) { |
| gst_caps_unref (ni2src->gst_caps); |
| ni2src->gst_caps = NULL; |
| } |
| |
| if (ni2src->device) { |
| delete ni2src->device; |
| ni2src->device = NULL; |
| } |
| |
| if (ni2src->depth) { |
| delete ni2src->depth; |
| ni2src->depth = NULL; |
| } |
| |
| if (ni2src->color) { |
| delete ni2src->color; |
| ni2src->color = NULL; |
| } |
| |
| if (ni2src->depthFrame) { |
| delete ni2src->depthFrame; |
| ni2src->depthFrame = NULL; |
| } |
| |
| if (ni2src->colorFrame) { |
| delete ni2src->colorFrame; |
| ni2src->colorFrame = NULL; |
| } |
| |
| G_OBJECT_CLASS (parent_class)->finalize (gobject); |
| } |
| |
| static void |
| gst_openni2_src_set_property (GObject * object, guint prop_id, |
| const GValue * value, GParamSpec * pspec) |
| { |
| GstOpenni2Src *openni2src = GST_OPENNI2_SRC (object); |
| |
| GST_OBJECT_LOCK (openni2src); |
| switch (prop_id) { |
| case PROP_LOCATION: |
| if (!g_value_get_string (value)) { |
| GST_WARNING ("location property cannot be NULL"); |
| break; |
| } |
| |
| if (openni2src->uri_name != NULL) { |
| g_free (openni2src->uri_name); |
| openni2src->uri_name = NULL; |
| } |
| |
| openni2src->uri_name = g_value_dup_string (value); |
| break; |
| case PROP_SOURCETYPE: |
| openni2src->sourcetype = g_value_get_enum (value); |
| break; |
| default: |
| G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); |
| break; |
| } |
| |
| GST_OBJECT_UNLOCK (openni2src); |
| } |
| |
| static void |
| gst_openni2_src_get_property (GObject * object, guint prop_id, |
| GValue * value, GParamSpec * pspec) |
| { |
| GstOpenni2Src *openni2src = GST_OPENNI2_SRC (object); |
| |
| GST_OBJECT_LOCK (openni2src); |
| switch (prop_id) { |
| case PROP_LOCATION: |
| g_value_set_string (value, openni2src->uri_name); |
| break; |
| case PROP_SOURCETYPE: |
| g_value_set_enum (value, openni2src->sourcetype); |
| break; |
| default: |
| G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); |
| break; |
| } |
| GST_OBJECT_UNLOCK (openni2src); |
| } |
| |
| /* Interesting info from gstv4l2src.c: |
| * "start and stop are not symmetric -- start will open the device, but not |
| * start capture. it's setcaps that will start capture, which is called via |
| * basesrc's negotiate method. stop will both stop capture and close t device." |
| */ |
| static gboolean |
| gst_openni2_src_start (GstBaseSrc * bsrc) |
| { |
| GstOpenni2Src *src = GST_OPENNI2_SRC (bsrc); |
| openni::Status rc = openni::STATUS_OK; |
| |
| if (src->depth->isValid ()) { |
| rc = src->depth->start (); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR_OBJECT (src, "Couldn't start the depth stream\n%s\n", |
| openni::OpenNI::getExtendedError ()); |
| return FALSE; |
| } |
| } |
| |
| if (src->color->isValid ()) { |
| rc = src->color->start (); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR_OBJECT (src, "Couldn't start the color stream\n%s\n", |
| openni::OpenNI::getExtendedError ()); |
| return FALSE; |
| } |
| } |
| |
| return TRUE; |
| } |
| |
| static gboolean |
| gst_openni2_src_stop (GstBaseSrc * bsrc) |
| { |
| GstOpenni2Src *src = GST_OPENNI2_SRC (bsrc); |
| |
| if (src->depthFrame) |
| src->depthFrame->release (); |
| |
| if (src->colorFrame) |
| src->colorFrame->release (); |
| |
| if (src->depth->isValid ()) { |
| src->depth->stop (); |
| src->depth->destroy (); |
| } |
| |
| if (src->color->isValid ()) { |
| src->color->stop (); |
| src->color->destroy (); |
| } |
| |
| src->device->close (); |
| |
| return TRUE; |
| } |
| |
| static GstCaps * |
| gst_openni2_src_get_caps (GstBaseSrc * src, GstCaps * filter) |
| { |
| GstOpenni2Src *ni2src; |
| GstCaps *caps; |
| GstVideoInfo info; |
| GstVideoFormat format; |
| |
| ni2src = GST_OPENNI2_SRC (src); |
| |
| GST_OBJECT_LOCK (ni2src); |
| if (ni2src->gst_caps) |
| goto out; |
| |
| // If we are here, we need to compose the caps and return them. |
| |
| if (ni2src->depth->isValid () && ni2src->color->isValid () && |
| ni2src->sourcetype == SOURCETYPE_BOTH |
| && ni2src->colorpixfmt == openni::PIXEL_FORMAT_RGB888) { |
| format = GST_VIDEO_FORMAT_RGBA; |
| } else if (ni2src->depth->isValid () && |
| ni2src->sourcetype == SOURCETYPE_DEPTH) { |
| format = GST_VIDEO_FORMAT_GRAY16_LE; |
| } else if (ni2src->color->isValid () && ni2src->sourcetype == SOURCETYPE_COLOR |
| && ni2src->colorpixfmt == openni::PIXEL_FORMAT_RGB888) { |
| format = GST_VIDEO_FORMAT_RGB; |
| } else { |
| goto out; |
| } |
| |
| gst_video_info_init (&info); |
| gst_video_info_set_format (&info, format, ni2src->width, ni2src->height); |
| info.fps_n = ni2src->fps; |
| info.fps_d = 1; |
| caps = gst_video_info_to_caps (&info); |
| |
| GST_INFO_OBJECT (ni2src, "probed caps: %" GST_PTR_FORMAT, caps); |
| ni2src->gst_caps = caps; |
| |
| out: |
| GST_OBJECT_UNLOCK (ni2src); |
| |
| if (!ni2src->gst_caps) |
| return gst_pad_get_pad_template_caps (GST_BASE_SRC_PAD (ni2src)); |
| |
| return (filter) |
| ? gst_caps_intersect_full (filter, ni2src->gst_caps, |
| GST_CAPS_INTERSECT_FIRST) |
| : gst_caps_ref (ni2src->gst_caps); |
| } |
| |
| static gboolean |
| gst_openni2_src_set_caps (GstBaseSrc * src, GstCaps * caps) |
| { |
| GstOpenni2Src *ni2src; |
| |
| ni2src = GST_OPENNI2_SRC (src); |
| |
| return gst_video_info_from_caps (&ni2src->info, caps); |
| } |
| |
| static GstStateChangeReturn |
| gst_openni2_src_change_state (GstElement * element, GstStateChange transition) |
| { |
| GstStateChangeReturn ret = GST_STATE_CHANGE_FAILURE; |
| GstOpenni2Src *src = GST_OPENNI2_SRC (element); |
| |
| switch (transition) { |
| case GST_STATE_CHANGE_NULL_TO_READY: |
| /* Action! */ |
| if (!openni2_initialise_devices (src)) |
| return GST_STATE_CHANGE_FAILURE; |
| break; |
| case GST_STATE_CHANGE_READY_TO_PAUSED: |
| break; |
| case GST_STATE_CHANGE_PAUSED_TO_PLAYING: |
| break; |
| default: |
| break; |
| } |
| |
| ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); |
| if (ret == GST_STATE_CHANGE_FAILURE) { |
| return ret; |
| } |
| |
| switch (transition) { |
| case GST_STATE_CHANGE_READY_TO_NULL: |
| gst_openni2_src_stop (GST_BASE_SRC (src)); |
| if (src->gst_caps) { |
| gst_caps_unref (src->gst_caps); |
| src->gst_caps = NULL; |
| } |
| break; |
| case GST_STATE_CHANGE_PLAYING_TO_PAUSED: |
| break; |
| case GST_STATE_CHANGE_PAUSED_TO_READY: |
| src->oni_start_ts = GST_CLOCK_TIME_NONE; |
| break; |
| default: |
| break; |
| } |
| |
| return ret; |
| } |
| |
| |
| static GstFlowReturn |
| gst_openni2src_fill (GstPushSrc * src, GstBuffer * buf) |
| { |
| GstOpenni2Src *ni2src = GST_OPENNI2_SRC (src); |
| return openni2_read_gstbuffer (ni2src, buf); |
| } |
| |
| static gboolean |
| gst_openni2src_decide_allocation (GstBaseSrc * bsrc, GstQuery * query) |
| { |
| GstBufferPool *pool; |
| guint size, min, max; |
| gboolean update; |
| GstStructure *config; |
| GstCaps *caps; |
| GstVideoInfo info; |
| |
| gst_query_parse_allocation (query, &caps, NULL); |
| gst_video_info_from_caps (&info, caps); |
| |
| if (gst_query_get_n_allocation_pools (query) > 0) { |
| gst_query_parse_nth_allocation_pool (query, 0, &pool, &size, &min, &max); |
| update = TRUE; |
| } else { |
| pool = NULL; |
| min = max = 0; |
| size = info.size; |
| update = FALSE; |
| } |
| |
| GST_DEBUG_OBJECT (bsrc, "allocation: size:%u min:%u max:%u pool:%" |
| GST_PTR_FORMAT " caps:%" GST_PTR_FORMAT, size, min, max, pool, caps); |
| |
| if (!pool) |
| pool = gst_video_buffer_pool_new (); |
| |
| config = gst_buffer_pool_get_config (pool); |
| gst_buffer_pool_config_set_params (config, caps, size, min, max); |
| |
| if (gst_query_find_allocation_meta (query, GST_VIDEO_META_API_TYPE, NULL)) { |
| GST_DEBUG_OBJECT (pool, "activate Video Meta"); |
| gst_buffer_pool_config_add_option (config, |
| GST_BUFFER_POOL_OPTION_VIDEO_META); |
| } |
| |
| gst_buffer_pool_set_config (pool, config); |
| |
| if (update) |
| gst_query_set_nth_allocation_pool (query, 0, pool, size, min, max); |
| else |
| gst_query_add_allocation_pool (query, pool, size, min, max); |
| |
| gst_object_unref (pool); |
| |
| return GST_BASE_SRC_CLASS (parent_class)->decide_allocation (bsrc, query); |
| } |
| |
| gboolean |
| gst_openni2src_plugin_init (GstPlugin * plugin) |
| { |
| return gst_element_register (plugin, "openni2src", GST_RANK_NONE, |
| GST_TYPE_OPENNI2_SRC); |
| } |
| |
| |
| static gboolean |
| openni2_initialise_library (void) |
| { |
| openni::Status rc = openni::STATUS_OK; |
| rc = openni::OpenNI::initialize (); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR ("Initialization failed: %s", |
| openni::OpenNI::getExtendedError ()); |
| openni::OpenNI::shutdown (); |
| return GST_FLOW_ERROR; |
| } |
| return (rc == openni::STATUS_OK); |
| } |
| |
| static gboolean |
| openni2_initialise_devices (GstOpenni2Src * src) |
| { |
| openni::Status rc = openni::STATUS_OK; |
| const char *deviceURI = openni::ANY_DEVICE; |
| |
| if (src->uri_name) |
| deviceURI = src->uri_name; |
| |
| rc = src->device->open (deviceURI); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR_OBJECT (src, "Device (%s) open failed: %s", deviceURI, |
| openni::OpenNI::getExtendedError ()); |
| openni::OpenNI::shutdown (); |
| return FALSE; |
| } |
| |
| /** depth sensor **/ |
| rc = src->depth->create (*src->device, openni::SENSOR_DEPTH); |
| if (rc == openni::STATUS_OK) { |
| rc = src->depth->start (); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR_OBJECT (src, "%s", openni::OpenNI::getExtendedError ()); |
| src->depth->destroy (); |
| } |
| } else { |
| GST_WARNING_OBJECT (src, "Couldn't find depth stream: %s", |
| openni::OpenNI::getExtendedError ()); |
| } |
| |
| /** color sensor **/ |
| rc = src->color->create (*src->device, openni::SENSOR_COLOR); |
| if (rc == openni::STATUS_OK) { |
| rc = src->color->start (); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR_OBJECT (src, "Couldn't start color stream: %s ", |
| openni::OpenNI::getExtendedError ()); |
| src->color->destroy (); |
| } |
| } else { |
| GST_WARNING_OBJECT (src, "Couldn't find color stream: %s", |
| openni::OpenNI::getExtendedError ()); |
| } |
| |
| if (!src->depth->isValid () && !src->color->isValid ()) { |
| GST_ERROR_OBJECT (src, "No valid streams. Exiting\n"); |
| openni::OpenNI::shutdown (); |
| return FALSE; |
| } |
| |
| /** Get resolution and make sure is valid **/ |
| if (src->depth->isValid () && src->color->isValid ()) { |
| src->depthVideoMode = src->depth->getVideoMode (); |
| src->colorVideoMode = src->color->getVideoMode (); |
| |
| int depthWidth = src->depthVideoMode.getResolutionX (); |
| int depthHeight = src->depthVideoMode.getResolutionY (); |
| int colorWidth = src->colorVideoMode.getResolutionX (); |
| int colorHeight = src->colorVideoMode.getResolutionY (); |
| |
| if (depthWidth == colorWidth && depthHeight == colorHeight) { |
| src->width = depthWidth; |
| src->height = depthHeight; |
| src->fps = src->depthVideoMode.getFps (); |
| src->colorpixfmt = src->colorVideoMode.getPixelFormat (); |
| src->depthpixfmt = src->depthVideoMode.getPixelFormat (); |
| } else { |
| GST_ERROR_OBJECT (src, "Error - expect color and depth to be" |
| " in same resolution: D: %dx%d vs C: %dx%d", |
| depthWidth, depthHeight, colorWidth, colorHeight); |
| return FALSE; |
| } |
| GST_INFO_OBJECT (src, "DEPTH&COLOR resolution: %dx%d", |
| src->width, src->height); |
| } else if (src->depth->isValid ()) { |
| src->depthVideoMode = src->depth->getVideoMode (); |
| src->width = src->depthVideoMode.getResolutionX (); |
| src->height = src->depthVideoMode.getResolutionY (); |
| src->fps = src->depthVideoMode.getFps (); |
| src->depthpixfmt = src->depthVideoMode.getPixelFormat (); |
| GST_INFO_OBJECT (src, "DEPTH resolution: %dx%d", src->width, src->height); |
| } else if (src->color->isValid ()) { |
| src->colorVideoMode = src->color->getVideoMode (); |
| src->width = src->colorVideoMode.getResolutionX (); |
| src->height = src->colorVideoMode.getResolutionY (); |
| src->fps = src->colorVideoMode.getFps (); |
| src->colorpixfmt = src->colorVideoMode.getPixelFormat (); |
| GST_INFO_OBJECT (src, "COLOR resolution: %dx%d", src->width, src->height); |
| } else { |
| GST_ERROR_OBJECT (src, "Expected at least one of the streams to be valid."); |
| return FALSE; |
| } |
| |
| return TRUE; |
| } |
| |
| static GstFlowReturn |
| openni2_read_gstbuffer (GstOpenni2Src * src, GstBuffer * buf) |
| { |
| openni::Status rc = openni::STATUS_OK; |
| openni::VideoStream * pStream = src->depth; |
| int changedStreamDummy; |
| GstVideoFrame vframe; |
| uint64_t oni_ts; |
| |
| /* Block until we get some data */ |
| rc = openni::OpenNI::waitForAnyStream (&pStream, 1, &changedStreamDummy, |
| SAMPLE_READ_WAIT_TIMEOUT); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR_OBJECT (src, "Frame read timeout: %s", |
| openni::OpenNI::getExtendedError ()); |
| return GST_FLOW_ERROR; |
| } |
| |
| if (src->depth->isValid () && src->color->isValid () && |
| src->sourcetype == SOURCETYPE_BOTH) { |
| rc = src->depth->readFrame (src->depthFrame); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR_OBJECT (src, "Frame read error: %s", |
| openni::OpenNI::getExtendedError ()); |
| return GST_FLOW_ERROR; |
| } |
| rc = src->color->readFrame (src->colorFrame); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR_OBJECT (src, "Frame read error: %s", |
| openni::OpenNI::getExtendedError ()); |
| return GST_FLOW_ERROR; |
| } |
| |
| /* Copy colour information */ |
| gst_video_frame_map (&vframe, &src->info, buf, GST_MAP_WRITE); |
| |
| guint8 *pData = (guint8 *) GST_VIDEO_FRAME_PLANE_DATA (&vframe, 0); |
| guint8 *pColor = (guint8 *) src->colorFrame->getData (); |
| /* Add depth as 8bit alpha channel, depth is 16bit samples. */ |
| guint16 *pDepth = (guint16 *) src->depthFrame->getData (); |
| |
| for (int i = 0; i < src->colorFrame->getHeight (); ++i) { |
| for (int j = 0; j < src->colorFrame->getWidth (); ++j) { |
| pData[4 * j + 0] = pColor[3 * j + 0]; |
| pData[4 * j + 1] = pColor[3 * j + 1]; |
| pData[4 * j + 2] = pColor[3 * j + 2]; |
| pData[4 * j + 3] = pDepth[j] >> 8; |
| } |
| pData += GST_VIDEO_FRAME_PLANE_STRIDE (&vframe, 0); |
| pColor += src->colorFrame->getStrideInBytes (); |
| pDepth += src->depthFrame->getStrideInBytes () / 2; |
| } |
| gst_video_frame_unmap (&vframe); |
| |
| oni_ts = src->colorFrame->getTimestamp () * 1000; |
| |
| GST_LOG_OBJECT (src, "sending buffer (%d+%d)B", |
| src->colorFrame->getDataSize (), |
| src->depthFrame->getDataSize ()); |
| } else if (src->depth->isValid () && src->sourcetype == SOURCETYPE_DEPTH) { |
| rc = src->depth->readFrame (src->depthFrame); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR_OBJECT (src, "Frame read error: %s", |
| openni::OpenNI::getExtendedError ()); |
| return GST_FLOW_ERROR; |
| } |
| |
| /* Copy depth information */ |
| gst_video_frame_map (&vframe, &src->info, buf, GST_MAP_WRITE); |
| |
| guint16 *pData = (guint16 *) GST_VIDEO_FRAME_PLANE_DATA (&vframe, 0); |
| guint16 *pDepth = (guint16 *) src->depthFrame->getData (); |
| |
| for (int i = 0; i < src->depthFrame->getHeight (); ++i) { |
| memcpy (pData, pDepth, 2 * src->depthFrame->getWidth ()); |
| pDepth += src->depthFrame->getStrideInBytes () / 2; |
| pData += GST_VIDEO_FRAME_PLANE_STRIDE (&vframe, 0) / 2; |
| } |
| gst_video_frame_unmap (&vframe); |
| |
| oni_ts = src->depthFrame->getTimestamp () * 1000; |
| |
| GST_LOG_OBJECT (src, "sending buffer (%dx%d)=%dB", |
| src->depthFrame->getWidth (), |
| src->depthFrame->getHeight (), |
| src->depthFrame->getDataSize ()); |
| } else if (src->color->isValid () && src->sourcetype == SOURCETYPE_COLOR) { |
| rc = src->color->readFrame (src->colorFrame); |
| if (rc != openni::STATUS_OK) { |
| GST_ERROR_OBJECT (src, "Frame read error: %s", |
| openni::OpenNI::getExtendedError ()); |
| return GST_FLOW_ERROR; |
| } |
| |
| gst_video_frame_map (&vframe, &src->info, buf, GST_MAP_WRITE); |
| |
| guint8 *pData = (guint8 *) GST_VIDEO_FRAME_PLANE_DATA (&vframe, 0); |
| guint8 *pColor = (guint8 *) src->colorFrame->getData (); |
| |
| for (int i = 0; i < src->colorFrame->getHeight (); ++i) { |
| memcpy (pData, pColor, 3 * src->colorFrame->getWidth ()); |
| pColor += src->colorFrame->getStrideInBytes (); |
| pData += GST_VIDEO_FRAME_PLANE_STRIDE (&vframe, 0); |
| } |
| gst_video_frame_unmap (&vframe); |
| |
| oni_ts = src->colorFrame->getTimestamp () * 1000; |
| |
| GST_LOG_OBJECT (src, "sending buffer (%dx%d)=%dB", |
| src->colorFrame->getWidth (), |
| src->colorFrame->getHeight (), |
| src->colorFrame->getDataSize ()); |
| } else { |
| g_return_val_if_reached (GST_FLOW_ERROR); |
| return GST_FLOW_ERROR; |
| } |
| |
| if (G_UNLIKELY (src->oni_start_ts == GST_CLOCK_TIME_NONE)) |
| src->oni_start_ts = oni_ts; |
| |
| GST_BUFFER_PTS (buf) = oni_ts - src->oni_start_ts; |
| |
| GST_LOG_OBJECT (src, "Calculated PTS as %" GST_TIME_FORMAT, |
| GST_TIME_ARGS (GST_BUFFER_PTS (buf))); |
| |
| return GST_FLOW_OK; |
| } |