--- a/khronosfws/openmax_al/src/adaptation/xamediaplayeradaptctx.c Mon May 03 12:59:52 2010 +0300
+++ /dev/null Thu Jan 01 00:00:00 1970 +0000
@@ -1,1253 +0,0 @@
-/*
-* Copyright (c) 2009 Nokia Corporation and/or its subsidiary(-ies).
-* All rights reserved.
-* This component and the accompanying materials are made available
-* under the terms of "Eclipse Public License v1.0"
-* which accompanies this distribution, and is available
-* at the URL "http://www.eclipse.org/legal/epl-v10.html".
-*
-* Initial Contributors:
-* Nokia Corporation - initial contribution.
-*
-* Contributors:
-*
-* Description:
-*
-*/
-
-#include <assert.h>
-#include <stdlib.h>
-#include <gstappsrc.h>
-#include <gst.h>
-#include "XAMediaPlayerAdaptCtx.h"
-#include "XAMediaPlayerAdaptCtxMMF.h"
-#include "XAAdaptation.h"
-#include "XAObjectItf.h"
-#include "XACameraDevice.h"
-#include "XAOutputMix.h"
-#include "XAMetadataAdaptation.h"
-#include "XANGAVideoSink.h"
-
-static void need_data_for_prerolling (GstElement * pipeline, guint size, XAMediaPlayerAdaptationCtx* ctx );
-static void push_data_for_prerolling (GstElement * pipeline, GstBuffer *buffer, XAMediaPlayerAdaptationCtx* ctx);
-static void enough_data_for_prerolling (GstElement * pipeline, XAMediaPlayerAdaptationCtx* ctx);
-/* forward declarations */
-XAresult XAMediaPlayerAdapt_CreatePipeline( XAMediaPlayerAdaptationCtx* ctx );
-
-void* ngaVideoSinkPtr = NULL;
-
-
-extern XAboolean cameraRealized;
-extern XACameraAdaptationCtx_* cameraCtx;
-
-/*
- * static void XAMediaPlayerAdapt_NewPadCb (GstElement *element, GstPad *pad, gpointer data)
- * Listen to codec bin dynamic pads
- */
-static void XAMediaPlayerAdapt_NewPadCb (GstElement *element, GstPad *pad, gpointer data)
-{
- XAMediaPlayerAdaptationCtx* mCtx = (XAMediaPlayerAdaptationCtx*)data;
- gchar *name = gst_pad_get_name (pad);
- DEBUG_API_A1("->XAMediaPlayerAdapt_NewPadCb: A new pad %s was created", name);
- /*try first newly created pad to video pipe*/
- if( mCtx->videoppbin && !(gst_pad_is_linked(gst_element_get_static_pad(mCtx->videoppbin, "videopp_sink"))) )
- {
- if(gst_element_link_pads (mCtx->codecbin, name, mCtx->videoppbin, "videopp_sink"))
- {
- DEBUG_INFO_A1("Pads linked, codecbin:%s to videopp:sink", name);
- g_free (name);
- DEBUG_API("<-XAMediaPlayerAdapt_NewPadCb");
- return;
- }
- }
- /*..and then to audio pipe*/
- if( mCtx->audioppbin && !(gst_pad_is_linked(gst_element_get_static_pad(mCtx->audioppbin, "sink"))) )
- {
- if(gst_element_link_pads (mCtx->codecbin, name, mCtx->audioppbin, "sink"))
- {
- DEBUG_INFO_A1("Pads linked, codecbin:%s to audiopp:sink", name);
- g_free (name);
- DEBUG_API("<-XAMediaPlayerAdapt_NewPadCb");
- return;
- }
- }
-
- g_free (name);
- DEBUG_INFO("Warning: Could not find anything to link to new pad.");
- DEBUG_API("<-XAMediaPlayerAdapt_NewPadCb");
-}
-
-/*
- * void push_data_for_prerolling (GstElement * pipeline, GstBuffer *buffer, XAMediaPlayerAdaptationCtx* ctx)
- * Called when "push-buffer" signal is emitted
- */
-void push_data_for_prerolling (GstElement * pipeline, GstBuffer *buffer, XAMediaPlayerAdaptationCtx* ctx)
-{
- DEBUG_API("->push_data_for_prerolling");
- gst_app_src_push_buffer( GST_APP_SRC(ctx->source), GST_BUFFER(buffer) );
- /*GstPad* prerollPad = NULL;
- prerollPad = gst_element_get_static_pad(GST_ELEMENT(ctx->source),"src");
- gst_pad_push (prerollPad, buffer);
- gst_element_send_event(GST_ELEMENT(ctx->source),gst_event_new_flush_start());
- gst_element_send_event(GST_ELEMENT(ctx->source),gst_event_new_flush_stop());*/
- DEBUG_API("<-push_data_for_prerolling");
-}
-
-/*
- * void enough_data_for_prerolling (GstElement * pipeline, XAMediaPlayerAdaptationCtx* ctx)
- * Called when appsrc has enough data
- */
-void enough_data_for_prerolling (GstElement * pipeline, XAMediaPlayerAdaptationCtx* ctx)
-{
- DEBUG_API("->enough_data_for_prerolling");
- /*No any functionality yet*/
- DEBUG_API("<-enough_data_for_prerolling");
-}
-
-/*
- * void need_data_for_prerolling (GstElement * pipeline, guint size, XAMediaPlayerAdaptationCtx* ctx )
- * Called when the appsrc needs more data during prerolling
- */
-void need_data_for_prerolling (GstElement * pipeline, guint size, XAMediaPlayerAdaptationCtx* ctx )
-{
- CPresult cpRet;
- XAuint32 requestedBytes = size;
- guint readedBytes = 0;
- CP_CHECKBYTESRESULTTYPE eResult;
-
- DEBUG_API("->need_data_for_prerolling");
-
- /*Start prerolling to contentpipe data*/
- ctx->baseObj.pipeSrcThrCtx.state = CPStatePrerolling;
-
- do
- {
- gpointer cpBuffer = NULL;
- GstBuffer *buffer = NULL;
-
- cpRet = ctx->baseObj.pipeSrcThrCtx.pipe->pContentPipe->CheckAvailableBytes(&(ctx->baseObj.pipeSrcThrCtx.dataHandle), requestedBytes, &eResult);
- if ( cpRet != EXIT_SUCCESS )
- {
- DEBUG_API("ERROR");
- ctx->baseObj.pipeSrcThrCtx.state = CPStateError;
- }
-
- if ( eResult == CP_CheckBytesOk )
- {
- cpBuffer = g_malloc0(requestedBytes );
- readedBytes =+ requestedBytes;
- cpRet = ctx->baseObj.pipeSrcThrCtx.pipe->pContentPipe->Read( &(ctx->baseObj.pipeSrcThrCtx.dataHandle), (CPbyte*)cpBuffer, requestedBytes );
- if ( cpRet != EXIT_SUCCESS )
- {
- DEBUG_ERR("Could not read data from content pipe!");
- ctx->baseObj.pipeSrcThrCtx.state = CPStateError;
- }
- else
- {
- DEBUG_INFO_A1("Readed %u bytes", requestedBytes );
- /* Create gstBuffer, GStreamer frees data */
- buffer = gst_app_buffer_new( (void*)cpBuffer, requestedBytes, g_free, cpBuffer );
- if ( !buffer )
- {
- DEBUG_ERR("Could not allocate buffer for content pipe source!");
- ctx->baseObj.pipeSrcThrCtx.state = CPStateError;
- }
- }
-
- if ( cpRet == EXIT_SUCCESS )
- {
- if( buffer )
- {
- DEBUG_INFO("Pushing preroll buffer");
- /*CP code: causes some delay for appsrc but we use push-buffer signal at this point
- GstFlowReturn ret;
- g_signal_emit_by_name (ctx->source, "push-buffer", GST_BUFFER(buffer), &ret);
- if( ret != GST_FLOW_OK )
- {
- DEBUG_ERR("Some problem during preroll");
- DEBUG_API("<-need_data_for_prerolling");
- }*/
- gst_element_send_event(GST_ELEMENT(ctx->source),gst_event_new_flush_start());
- gst_app_src_push_buffer( GST_APP_SRC(ctx->source), GST_BUFFER(buffer) );
- gst_element_send_event(GST_ELEMENT(ctx->source),gst_event_new_flush_stop());
- break;
- }
- }
- }
- else if( eResult == CP_CheckBytesAtEndOfStream )
- {
- ctx->baseObj.pipeSrcThrCtx.state = CPStateEOS;
- break;
- }
- else if( eResult == CP_CheckBytesInsufficientBytes )
- {
- ctx->baseObj.pipeSrcThrCtx.state = CPStateWaitForData;
- }
- else if( eResult == CP_CheckBytesVendorStartUnused )
- {
- /*We use this enumeration when client is started caching from the beginning again*/
- DEBUG_API( "CP_CheckBytesVendorStartUnused from implementation");
- }
- }while ( ctx->baseObj.pipeSrcThrCtx.state == CPStatePrerolling || ctx->baseObj.pipeSrcThrCtx.state == CPStateWaitForData );
-
- DEBUG_API("<-need_data_for_prerolling");
-}
-
-/*
- * gboolean XAMediaPlayerAdapt_GstBusCb( GstBus *bus, GstMessage *message, gpointer data )
- * MediaPlayer Gst-bus message handler (Callback)
- */
-gboolean XAMediaPlayerAdapt_GstBusCb( GstBus *bus, GstMessage *message, gpointer data )
-{
- XAAdaptationBaseCtx* bCtx = (XAAdaptationBaseCtx*)data;
- /* only listen to bin messages */
- if(GST_MESSAGE_SRC(message)==(GstObject*)(bCtx->bin))
- {
- XAMediaPlayerAdaptationCtx* mCtx = (XAMediaPlayerAdaptationCtx*)data;
- DEBUG_API_A2("->XAMediaPlayerAdapt_GstBusCb:\"%s\" from object \"%s\"",
- GST_MESSAGE_TYPE_NAME(message), GST_OBJECT_NAME(GST_MESSAGE_SRC(message)));
-
- switch( GST_MESSAGE_TYPE(message))
- {
- case GST_MESSAGE_EOS:
- {
- if( mCtx && mCtx->loopingenabled && mCtx->loopend == GST_CLOCK_TIME_NONE)
- {
- DEBUG_INFO_A2("Restart loop from %"GST_TIME_FORMAT" to %"GST_TIME_FORMAT,
- GST_TIME_ARGS(mCtx->loopstart), GST_TIME_ARGS(mCtx->loopend));
- gst_element_seek( bCtx->bin, mCtx->playrate, GST_FORMAT_TIME, GST_SEEK_FLAG_FLUSH|GST_SEEK_FLAG_ACCURATE,
- GST_SEEK_TYPE_SET, mCtx->loopstart,
- GST_SEEK_TYPE_NONE, GST_CLOCK_TIME_NONE );
- gst_element_get_state(bCtx->bin,NULL,NULL,XA_ADAPT_ASYNC_TIMEOUT_SHORT_NSEC);
- mCtx->lastpos = mCtx->loopstart;
- if( mCtx && mCtx->trackpositionenabled )
- {
- XAmillisecond posMsec = GST_TIME_AS_MSECONDS(mCtx->lastpos);/*Warning ok due to used API specification*/
- XAAdaptEvent event = {XA_PLAYITFEVENTS, XA_ADAPT_POSITION_UPDATE_EVT, 1, NULL};
- event.data = &posMsec;
- XAAdaptationBase_SendAdaptEvents(bCtx, &event );
- }
- }
- else
- {
- /* stop position tracking */
- if(mCtx->runpositiontimer > 0)
- {
- g_source_remove(mCtx->runpositiontimer);
- mCtx->runpositiontimer=0;
- }
-
- /* complete any ongoing client async operations */
- XAAdaptationBase_CompleteAsyncWait(bCtx);
-
- /* send needed events */
- {
- XAMediaType mediatype;
- if( mCtx->baseObj.pipeSrcThrCtx.pipe )
- {
- XACommon_CheckDataSource((XADataSource*)mCtx->xaAudioSink, &mediatype);
- }
- else
- {
- XACommon_CheckDataSource(mCtx->xaSource, &mediatype);
- }
- if(mediatype!=XA_MEDIATYPE_IMAGE)
- {
- XAAdaptEvent event = { XA_PLAYITFEVENTS, XA_PLAYEVENT_HEADATEND, 0, NULL };
- XAAdaptationBase_SendAdaptEvents(bCtx, &event );
- }
- }
- if(mCtx->positionCb)
- {
- mCtx->positionCb(bCtx);
- }
- bCtx->binWantedState = GST_STATE_PAUSED;
- }
- break;
- }
-
- case GST_MESSAGE_STATE_CHANGED:
- {
- GstState oldstate, newstate, pendingstate, gsttargetstate;
- gst_message_parse_state_changed(message, &oldstate, &newstate, &pendingstate);
- gsttargetstate = GST_STATE_TARGET(bCtx->bin);
- DEBUG_INFO_A4("old %s -> new %s ( pending %s, gsttarget %s )",
- gst_element_state_get_name(oldstate),
- gst_element_state_get_name(newstate),
- gst_element_state_get_name(pendingstate),
- gst_element_state_get_name(gsttargetstate) );
- if(gsttargetstate!=bCtx->binWantedState)
- {
- DEBUG_ERR_A1("WARNING: Gst target is not wanted target [%s]!!!",
- gst_element_state_get_name(bCtx->binWantedState));
- }
- /* print out some more info */
- if( pendingstate == GST_STATE_VOID_PENDING )
- {
- if( newstate != bCtx->binWantedState )
- {
- DEBUG_INFO_A2("Gst in intermediate state transition (curr %s, target %s)",
- gst_element_state_get_name(newstate),
- gst_element_state_get_name(bCtx->binWantedState));
- break; // TL: add to avoid extra event sent below in this case...
- }
- else
- {
- DEBUG_INFO_A1("Gst in wanted target state (%s)",
- gst_element_state_get_name(newstate));
- }
- }
- if( oldstate!=GST_STATE_PLAYING && newstate==GST_STATE_PLAYING )
- {
- /* send needed events */
- XAAdaptEvent event = {XA_PLAYITFEVENTS, XA_PLAYEVENT_HEADMOVING, 0, NULL };
- XAAdaptationBase_SendAdaptEvents(bCtx, &event );
- /* enable position tracking if needed */
- XAMediaPlayerAdapt_UpdatePositionCbTimer(mCtx);
- }
- if ( oldstate== GST_STATE_READY && newstate==GST_STATE_PAUSED)
- {
- if ( mCtx->baseObj.pipeSinkThrCtx.dataHandle )
- {
- mCtx->baseObj.pipeSrcThrCtx.state = CPStateInitialized;
- }
- }
- break;
- }
-
- case GST_MESSAGE_ASYNC_DONE:
- {
- /* some async sequence ended */
- XAAdaptationBase_CompleteAsyncWait(bCtx);
- break;
- }
-
- case GST_MESSAGE_ERROR:
- {
- GError* error;
- gchar* debug;
- gst_message_parse_error(message, &error, &debug);
- DEBUG_ERR_A1("Gst reports error \"%s\"", debug);
- /* stop waiting any ongoing async operations */
- XAAdaptationBase_CompleteAsyncWait(bCtx);
- break;
- }
- case GST_MESSAGE_BUFFERING:
- {
- gint percent;
- gst_message_parse_buffering(message, &percent);
- DEBUG_INFO_A1("Gst message buffering %d", percent);
- mCtx->buffering = percent;
- {
- XAAdaptEvent event = {XA_PREFETCHITFEVENTS, XA_ADAPT_BUFFERING, 1, NULL };
- event.data = &mCtx->buffering;
- XAAdaptationBase_SendAdaptEvents(bCtx, &event );
- }
- break;
- }
- case GST_MESSAGE_ELEMENT:
- {
- DEBUG_INFO("GST_MESSAGE_ELEMENT");
- if ((gst_structure_has_name(message->structure, "graphics-surface-created")) ||
- (gst_structure_has_name(message->structure, "graphics-surface-updated")))
- {
- DEBUG_INFO("graphics-surface-created message recived");
- }
- break;
- }
- default:
- break;
- }
- }
- else //if (GST_MESSAGE_SRC(message)==(GstObject*)(bCtx->videosink))
- {
- switch( GST_MESSAGE_TYPE(message))
- {
- case GST_MESSAGE_ELEMENT:
- {
- DEBUG_INFO("GST_MESSAGE_ELEMENT");
- if ((gst_structure_has_name(message->structure, "graphics-surface-created")) ||
- (gst_structure_has_name(message->structure, "graphics-surface-updated")))
- {
- guint32 surfaceid0, surfaceid1, surfaceid2, surfaceid3;
- gint crop_rect_tl_x, crop_rect_tl_y, crop_rect_br_x, crop_rect_br_y;
- gint aspect_ratio_num, aspect_ratio_denom;
-
- GstObject *sink= GST_MESSAGE_SRC(message);
-
- g_object_get(sink, "surfaceid0", &surfaceid0, NULL);
- g_object_get(sink, "surfaceid1", &surfaceid1,NULL);
- g_object_get(sink, "surfaceid2", &surfaceid2,NULL);
- g_object_get(sink, "surfaceid3", &surfaceid3,NULL);
- g_object_get(sink, "croprect_tl_x", &crop_rect_tl_x, NULL);
- g_object_get(sink, "croprect_tl_y", &crop_rect_tl_y, NULL);
- g_object_get(sink, "croprect_br_x", &crop_rect_br_x, NULL);
- g_object_get(sink, "croprect_br_y", &crop_rect_br_y, NULL);
- g_object_get(sink, "aspectratio_num", &aspect_ratio_num, NULL);
- g_object_get(sink, "aspectratio_denom", &aspect_ratio_denom, NULL);
- surface_created(ngaVideoSinkPtr, surfaceid0,surfaceid1,surfaceid2,surfaceid3,crop_rect_tl_x,
- crop_rect_tl_y,crop_rect_br_x,crop_rect_br_y,aspect_ratio_num,aspect_ratio_denom);
- }
- break;
- }
- default:
- break;
- }
- }
- DEBUG_API("<-XAMediaPlayerAdapt_GstBusCb");
- return TRUE;
-}
-
-/*
- * XAMediaPlayerAdaptationCtx* XAMediaPlayerAdapt_Create()
- * Allocates memory for Media Player Adaptation Context and makes 1st phase initialization
- * @param XADataSource *pDataSrc - pointer to OMX-AL data source
- * @param XADataSource *pBankSrc - pointer to instrument bank structure in Mobile DLS, if NULL default will be used.
- * @param XADataSink *pAudioSnk - pointer to OMX-AL audio sink definition
- * @param XADataSink *pImageVideoSnk - pointer to OMX-AL image and video sink definition
- * @returns XAMediaPlayerAdaptationCtx* - Pointer to created context, NULL if error occurs.
- */
-XAAdaptationBaseCtx* XAMediaPlayerAdapt_Create(XADataSource *pDataSrc, XADataSource *pBankSrc,
- XADataSink *pAudioSnk, XADataSink *pImageVideoSnk,
- XADataSink *pVibra, XADataSink *pLEDArray)
-{
- XAMediaPlayerAdaptationCtx *pSelf = NULL;
- XAuint32 locType = 0;
- XADataLocator_IODevice *ioDevice;
- XAresult res = XA_RESULT_INTERNAL_ERROR;
- DEBUG_API("->XAMediaPlayerAdapt_Create");
-
- //Create NGA Video Sink class
- if(pImageVideoSnk)
- {
- ngaVideoSinkPtr = nga_video_sink_init();
- }
-
- pSelf = calloc(1, sizeof(XAMediaPlayerAdaptationCtx));
- if ( pSelf)
- {
- if( XAAdaptationBase_Init(&(pSelf->baseObj),XAMediaPlayerAdaptation)
- != XA_RESULT_SUCCESS )
- {
- DEBUG_ERR("Failed to init base context!!!");
- free(pSelf);
- pSelf = NULL;
- }
- else
- {
- pSelf->xaSource = pDataSrc;
- pSelf->xaBankSrc = pBankSrc;
- pSelf->xaAudioSink = pAudioSnk;
- pSelf->xaVideoSink = pImageVideoSnk;
- pSelf->xaLEDArray = pLEDArray;
- pSelf->xaVibra = pVibra;
- pSelf->loopstart = 0;
- pSelf->loopend = GST_CLOCK_TIME_NONE;
- pSelf->playrate = 1.0;
- pSelf->baseObj.pipeSrcThrCtx.state = CPStateNull;
- pSelf->baseObj.pipeSinkThrCtx.dataHandle = NULL;
- pSelf->rateprops = (XA_RATEPROP_SMOOTHVIDEO | XA_RATEPROP_SILENTAUDIO);
- pSelf->curMirror = XA_VIDEOMIRROR_NONE;
- pSelf->curRotation = 0;
- pSelf->isobjsrc = XA_BOOLEAN_FALSE;
- pSelf->cameraSinkSynced = XA_BOOLEAN_FALSE;
- /*pSelf->waitData = XA_BOOLEAN_FALSE;*/
- if(ngaVideoSinkPtr)
- {
- setup_native_display(ngaVideoSinkPtr, pImageVideoSnk);
- }
- }
-
- if ( pDataSrc )
- {
- locType = *((XAuint32*)(pDataSrc->pLocator));
- if ( locType == XA_DATALOCATOR_IODEVICE )
- {
- ioDevice = (XADataLocator_IODevice*)(pDataSrc->pLocator);
- if ( ioDevice->deviceType == XA_IODEVICE_CAMERA && !cameraRealized )
- {
- DEBUG_ERR("Preconditions violated - Camera object not realized");
- XAAdaptationBase_Free(&pSelf->baseObj);
- free(pSelf);
- pSelf = NULL;
- }
- }
- }
- }
-
- DEBUG_API("<-XAMediaPlayerAdapt_Create");
- return (XAAdaptationBaseCtx*)pSelf;
-}
-
-
-
-/*
- * XAresult XAMediaPlayerAdapt_PostInit()
- * 2nd phase initialization of Media Player Adaptation Context
- * @param XAMediaPlayerAdaptationCtx* ctx - pointer to Media Player adaptation context
- * @return XAresult - Success value
- */
-XAresult XAMediaPlayerAdapt_PostInit( XAAdaptationBaseCtx* bCtx )
-{
- XAresult ret = XA_RESULT_SUCCESS;
- GstStateChangeReturn gret;
- XAuint32 locType = 0;
- XAMediaPlayerAdaptationCtx* ctx = NULL;
- GstElement *videotest=NULL;
-
- DEBUG_API("->XAMediaPlayerAdapt_PostInit");
- if(bCtx == NULL || bCtx->ctxId != XAMediaPlayerAdaptation )
- {
- DEBUG_ERR("Invalid parameter!!");
- DEBUG_API("<-XAMediaPlayerAdapt_PostInit");
- return XA_RESULT_PARAMETER_INVALID;
- }
- ctx = (XAMediaPlayerAdaptationCtx*)bCtx;
- assert(ctx);
- ret = XAAdaptationBase_PostInit( bCtx );
- if( ret!=XA_RESULT_SUCCESS )
- {
- DEBUG_ERR("Base context postinit failed!!");
- return ret;
- }
-
- /* top level bin for media player */
- ctx->baseObj.bin = gst_pipeline_new("media_player");
- /* Create Gst bus listener. */
- ret = XAAdaptationBase_InitGstListener(bCtx);
- if( ret!=XA_RESULT_SUCCESS )
- {
- DEBUG_ERR("Bus listener creation failed!!");
- return ret;
- }
- /* Add Media Player specific handler */
- if(ctx->baseObj.bus)
- {
- ctx->baseObj.busCb = XAMediaPlayerAdapt_GstBusCb;
- gst_bus_add_signal_watch( ctx->baseObj.bus );
- gst_bus_enable_sync_message_emission( ctx->baseObj.bus );
- g_signal_connect(ctx->baseObj.bus, "message::eos", G_CALLBACK(bCtx->busCb), ctx );
- g_signal_connect(ctx->baseObj.bus, "message::error", G_CALLBACK(bCtx->busCb), ctx );
- g_signal_connect(ctx->baseObj.bus, "message::warning", G_CALLBACK(bCtx->busCb), ctx );
- g_signal_connect(ctx->baseObj.bus, "message::state-changed", G_CALLBACK(bCtx->busCb), ctx );
- g_signal_connect(ctx->baseObj.bus, "message::segment-done", G_CALLBACK(bCtx->busCb), ctx );
- g_signal_connect(ctx->baseObj.bus, "message::async-done", G_CALLBACK(bCtx->busCb), ctx );
- g_signal_connect(ctx->baseObj.bus, "message::element", G_CALLBACK(bCtx->busCb), ctx );
- }
- else
- {
- DEBUG_ERR("Failed to create message bus");
- return XA_RESULT_INTERNAL_ERROR;
- }
-
- XAMetadataAdapt_PreInit(bCtx);
-
- /* create pipeline */
- ret = XAMediaPlayerAdapt_CreatePipeline(ctx);
- if ( ret != XA_RESULT_SUCCESS )
- {
- DEBUG_ERR("Failed to create Media Player pipeline");
- return ret;
- }
-
- /* Init content pipe if used*/
- locType = *((XAuint32*)(ctx->xaSource->pLocator));
- if ( locType == XA_DATALOCATOR_CONTENTPIPE)
- {
- XAMediaPlayerAdapt_InitContentPipeSrc(ctx);
- }
-
-
-#ifdef XA_IMPL_MEASURE_GST_DELAY
- ctx->baseObj.startTime = clock();
-#endif /* XA_IMPL_MEASURE_GST_DELAY */
- /* roll up bin */
- ctx->baseObj.binWantedState = GST_STATE_PAUSED;
-
- XAAdaptationBase_PrepareAsyncWait(bCtx);
- gret = gst_element_set_state( GST_ELEMENT(ctx->baseObj.bin), bCtx->binWantedState);
- if( gret == GST_STATE_CHANGE_ASYNC )
- {
- DEBUG_INFO("Wait for preroll");
- XAAdaptationBase_StartAsyncWait(bCtx);
- DEBUG_INFO("Preroll ready");
- }
- else if( gret == GST_STATE_CHANGE_FAILURE )
- {
- DEBUG_ERR("Preroll FAILED");
- ret = XA_RESULT_INTERNAL_ERROR;
- }
-
- bCtx->waitingasyncop = XA_BOOLEAN_FALSE;
-
- gret = gst_element_get_state( GST_ELEMENT(bCtx->bin), NULL, NULL, XA_ADAPT_ASYNC_TIMEOUT_SHORT_NSEC);
- if(GST_STATE(bCtx->bin)<GST_STATE_PAUSED)
- {
- DEBUG_INFO("Warning! Preroll not ready");
- if( ctx->audioppbin && !(gst_pad_is_linked(gst_element_get_static_pad(ctx->audioppbin, "sink"))) )
- {/*could not find suitable pad for audiopipeline - remove it*/
- DEBUG_INFO("Warning! No suitable decodebin pad for audio pipeline!");
- gst_element_set_state( GST_ELEMENT(ctx->audioppbin), GST_STATE_NULL);
- gst_bin_remove(GST_BIN(bCtx->bin), ctx->audioppbin);
- gst_element_set_state( GST_ELEMENT(ctx->audiosink), GST_STATE_NULL);
- gst_bin_remove(GST_BIN(bCtx->bin), ctx->audiosink);
- }
- else if( ctx->videoppbin && !(gst_pad_is_linked(gst_element_get_static_pad(ctx->videoppbin, "videopp_sink"))) )
- {/*could not find suitable pad for videopipeline - remove it*/
- DEBUG_INFO("Warning! No suitable decodebin pad for video pipeline!");
- gst_element_set_state( GST_ELEMENT(ctx->videoppbin), GST_STATE_NULL);
- gst_bin_remove(GST_BIN(bCtx->bin), ctx->videoppbin);
- gst_element_set_state( GST_ELEMENT(ctx->videosink), GST_STATE_NULL);
- gst_bin_remove(GST_BIN(bCtx->bin), ctx->videosink);
- }
- gst_element_set_state( GST_ELEMENT(bCtx->bin), bCtx->binWantedState);
- gst_element_get_state( GST_ELEMENT(bCtx->bin), NULL, NULL, XA_ADAPT_ASYNC_TIMEOUT_SHORT_NSEC);
- if(GST_STATE(bCtx->bin)==GST_STATE_PAUSED)
- {
- DEBUG_INFO("Retry preroll successful!")
- ret = XA_RESULT_SUCCESS;
- }
- }
- else
- {
- DEBUG_INFO("Preroll ready");
- }
- /*
- CP code: Removed because appsrc does not call any "need-data" signal after that.
- locType = *((XAuint32*)(ctx->xaSource->pLocator));
- if( locType == XA_DATALOCATOR_CONTENTPIPE)
- {
- gboolean emit = gst_app_src_get_emit_signals( GST_APP_SRC( ctx->source));
- if( emit )
- {
- gst_app_src_set_emit_signals( GST_APP_SRC( ctx->source), FALSE );
- }
- }*/
-#ifdef XA_IMPL_MEASURE_GST_DELAY
- bCtx->endTime = clock();
- double diff = bCtx->endTime - bCtx->startTime ;
- diff = diff / CLOCKS_PER_SEC;
- DEBUG_API_A1( "Starting up bin took %.4lf secs",diff);
-#endif /* XA_IMPL_MEASURE_GST_DELAY */
- videotest = gst_bin_get_by_name(GST_BIN(bCtx->bin), "videotest");
- if ( videotest && !ctx->isobjsrc )
- {
- gst_element_set_state( GST_ELEMENT(videotest),GST_STATE_PLAYING);
- }
-
- XAMetadataAdapt_PostInit(bCtx);
-
- if ( videotest )
- {
- gst_object_unref(videotest);
- }
-
- DEBUG_API("<-XAMediaPlayerAdapt_PostInit");
- return ret;
-}
-
-/*
- * void XAMediaPlayerAdapt_Destroy( XAMediaPlayerAdaptationCtx* ctx )
- * Destroys Media Player Adaptation Context
- * @param ctx - Media Player Adaptation context to be destroyed
- */
-void XAMediaPlayerAdapt_Destroy( XAAdaptationBaseCtx* bCtx )
-{
- XAMediaPlayerAdaptationCtx* ctx = NULL;
-
- XAresult res = XA_RESULT_SUCCESS;
- DEBUG_API("->XAMediaPlayerAdapt_Destroy");
- if(bCtx == NULL || bCtx->ctxId != XAMediaPlayerAdaptation )
- {
- DEBUG_ERR("Invalid parameter!!");
- DEBUG_API("<-XAMediaPlayerAdapt_Destroy");
- return;
- }
- ctx = (XAMediaPlayerAdaptationCtx*)bCtx;
-
- if( ctx->isobjsrc )
- { /* external source, unlink and remove now */
- /*gst_object_unparent( GST_OBJECT(ctx->source) );*/
- gst_element_unlink( ctx->source, ctx->codecbin );
- /*gst_bin_remove( GST_BIN(bCtx->bin), ctx->source );*/
- GST_OBJECT_FLAG_SET(GST_OBJECT(ctx->source),GST_OBJECT_FLOATING);
- }
-#if 0
- if ( ctx->xaSource )
- {
- XAuint32 locType = *(XAuint32*)(ctx->xaSource->pLocator);
- switch (locType )
- {
- case XA_DATALOCATOR_IODEVICE:
- {
- XADataLocator_IODevice* ioDevice = (XADataLocator_IODevice*)(ctx->xaSource->pLocator);
- if ( ioDevice->deviceType == XA_IODEVICE_RADIO )
- {
- gst_object_unparent( GST_OBJECT(ctx->source) );
- gst_element_unlink( ctx->source, ctx->codecbin );
- gst_bin_remove( GST_BIN(bCtx->bin), ctx->source );
- GST_OBJECT_FLAG_SET(GST_OBJECT(ctx->source),GST_OBJECT_FLOATING);
- }
- }
- default:
- break;
- }
- }
-#endif
-
- if( ctx->baseObj.pipeSrcThrCtx.dataHandle )
- {
- XAuint32 locType = *(XAuint32*)(ctx->xaAudioSink->pLocator);
- switch ( locType )
- {
- case XA_DATALOCATOR_CONTENTPIPE:
- {
- res = ctx->baseObj.pipeSrcThrCtx.pipe->pContentPipe->Close(&(ctx->baseObj.pipeSrcThrCtx.dataHandle));
- if( res != XA_RESULT_SUCCESS)
- {
- DEBUG_ERR("Cannot close contentpipe content");
- }
- break;
- }
- default:
- break;
- }
- }
-
- if( ctx->isobjasink && ctx->xaAudioSink && ctx->xaAudioSink->pLocator )
- {
- XAuint32 locType = *(XAuint32*)(ctx->xaAudioSink->pLocator);
- switch ( locType )
- {
- case XA_DATALOCATOR_OUTPUTMIX:
- {
- XADataLocator_OutputMix* omix = (XADataLocator_OutputMix*)(ctx->xaAudioSink->pLocator);
- XAOMixImpl* omixDevice = (XAOMixImpl*)(*omix->outputMix);
- if(omixDevice)
- {
- XAOutputMixAdapt_DisconnectObject(omixDevice->adaptationCtx, bCtx);
- }
- break;
- }
- default:
- /* Vibra and LED need no handling */
- break;
- }
-
- }
-
- if( bCtx->bus )
- {
- gst_bus_remove_signal_watch( bCtx->bus );
- gst_bus_disable_sync_message_emission ( bCtx->bus );
- }
- XAAdaptationBase_CancelAsyncWait(bCtx);
-
- if( ctx->runpositiontimer )
- {
- g_source_remove(ctx->runpositiontimer);
- }
- XAMetadataAdapt_FreeVars(ctx->metadatavars);
- XAAdaptationBase_Free( bCtx );
- free(ctx);
- ctx = NULL;
-
- DEBUG_API("<-XAMediaPlayerAdapt_Destroy");
-}
-
-
-/*
- * void XAMediaPlayerAdapt_CreatePipeline( XAMediaPlayerAdaptationCtx* ctx );
- */
-XAresult XAMediaPlayerAdapt_CreatePipeline( XAMediaPlayerAdaptationCtx* ctx )
-{
- XAresult ret = XA_RESULT_SUCCESS;
- XAboolean delayedlink = XA_BOOLEAN_FALSE;
- XAboolean isPCM = XA_BOOLEAN_FALSE;
- XAboolean isRawImage = XA_BOOLEAN_FALSE;
- XAMediaType mediatype;
- XAuint32 locType = 0;
- GstCaps* encSrcCaps;
- DEBUG_API("->XAMediaPlayerAdapt_CreatePipeline");
-
- /* create and add data source */
- XACommon_CheckDataSource(ctx->xaSource, &mediatype);
- ctx->source = XAAdaptationBase_CreateGstSource( ctx->xaSource, "datasrc", &(ctx->isobjsrc), &isPCM, &isRawImage);
- if( !(ctx->source) )
- {
- DEBUG_ERR("Could not create data source!!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
- else
- {
- if(mediatype != XA_MEDIATYPE_AUDIO)
- {
- //temporary work around for video
- encSrcCaps = gst_caps_new_simple ("video/h263-2000",
- "framerate", GST_TYPE_FRACTION, 25, 1,
- "pixel-aspect-ratio", GST_TYPE_FRACTION, 16, 9,
- "width", G_TYPE_INT, 176,
- "height", G_TYPE_INT, 144,
- NULL);
- g_object_set(G_OBJECT(ctx->source), "caps", encSrcCaps, NULL);
- }
-
- //boolRetVal = gst_bin_add(GST_BIN(pipeline), appsrc);
-
-
- }
-
- if ( !ctx->isobjsrc )
- { /* Add other than camera source to media player bin */
- DEBUG_INFO("No camera source");
- gst_bin_add(GST_BIN(ctx->baseObj.bin), ctx->source);
- }
- else
- {
- GstCaps* encSrcCaps;
- encSrcCaps = gst_caps_new_simple("video/x-raw-yuv",
- "format", GST_TYPE_FOURCC,GST_MAKE_FOURCC('I','4','2','0'),
- "framerate", GST_TYPE_FRACTION, 30, 1,
- NULL);
- DEBUG_INFO_A1("new camera encoding filter: %s",gst_caps_to_string(encSrcCaps));
- g_object_set( G_OBJECT(ctx->source), "filter-caps",encSrcCaps,NULL);
- gst_caps_unref(encSrcCaps);
- }
-
- /* create and add codec bin */
- if( !(ctx->isobjsrc || isPCM) )
- {
- DEBUG_INFO("Create decodebin");
- if(mediatype == XA_MEDIATYPE_AUDIO)
- {
- ctx->codecbin = gst_element_factory_make( "decodebin" , "mpcodecbin" );
- }
- else
- {
- ctx->codecbin = gst_element_factory_make( "identity" , "mpcodecbin" );
- }
- }
- else if(ctx->isobjsrc )
- { /* object sources produce framed raw data, decodebin only causes trouble */ //shyward
- DEBUG_INFO("Create identity")
- ctx->codecbin = gst_element_factory_make( "identity" , "mpcodecbin" );
- }
- else if(isPCM)
- { /* decodebin does not know how to handle PCM files */
- DEBUG_INFO("Create audioparse")
- ctx->codecbin = gst_element_factory_make( "audioparse" , "mpcodecbin" );
- }
- else if ( isRawImage)
- { /* decodebin does not know how to handle raw unframed video data */
- DEBUG_INFO("Create videoparse")
- ctx->codecbin = gst_element_factory_make( "videoparse", "mpcodecbin" );
- }
-
- if( ctx->codecbin )
- {
- gst_bin_add(GST_BIN(ctx->baseObj.bin), ctx->codecbin);
- if ( !ctx->isobjsrc )
- {
- if(mediatype == XA_MEDIATYPE_AUDIO)
- {
- if( !gst_element_link(ctx->source, ctx->codecbin) )
- {
- DEBUG_ERR("Could not link source to decodebin!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
- }
- else
- {
- if( !gst_element_link_filtered(ctx->source, ctx->codecbin, encSrcCaps ) )
- {
- DEBUG_ERR("Could not link source to decodebin!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
- }
- }
- else
- { /* Link camera source by using ghost-pads, because elements are in different bins */
-
- GstPad *cameraBinGhostPad=NULL;
- GstPad* ghost=NULL;
- GstElement *camTee=NULL;
- GstStateChangeReturn gret;
- GstPad *mpGhostSink=NULL;
-
- /* Set external camera source to ready for pipeline manipulation */
- DEBUG_INFO("Set ext-source PAUSED for pipeline manipulation");
- gret = gst_element_set_state( GST_ELEMENT(ctx->source), GST_STATE_READY);
- gret = gst_element_get_state( GST_ELEMENT(ctx->source), NULL,NULL,XA_ADAPT_ASYNC_TIMEOUT_SHORT_NSEC);
-
- /* Add new ghost-pad to external camera source */
- camTee = gst_bin_get_by_name( GST_BIN(ctx->source), "CamTee");
- if ( !camTee )
- {
- DEBUG_ERR("Could not get tee-element from camera");
- }
- cameraBinGhostPad = gst_element_get_request_pad( camTee, "src%d" );
- if ( !cameraBinGhostPad )
- {
- DEBUG_ERR("Could not get new src-pad from CamTee element");
- }
- gst_element_add_pad(ctx->source, gst_ghost_pad_new("MPObjSrc",cameraBinGhostPad));
- ghost = gst_element_get_static_pad( GST_ELEMENT(ctx->source), "MPObjSrc" );
- DEBUG_INFO_A2("Setting element:%s pad:%s to blocking.",
- gst_element_get_name(ctx->baseObj.bin),
- gst_pad_get_name(ghost));
- /* Set newly created pad to blocking */
- gst_pad_set_blocked_async(ghost, TRUE, XAAdaptationBase_PadBlockCb, NULL);
-
-
- /* Create new ghost-pad to media player pipeline where external camera is connected */
- mpGhostSink = gst_element_get_static_pad( GST_ELEMENT(ctx->codecbin), "sink");
- gst_element_add_pad(ctx->baseObj.bin, gst_ghost_pad_new("MPObjSink",mpGhostSink));
-
- if ( !gst_element_link_pads( GST_ELEMENT(ctx->source), "MPObjSrc",
- GST_ELEMENT(ctx->baseObj.bin), "MPObjSink") )
- {
- DEBUG_ERR("Could not link camera:MPObjSrc to videofilter:MPObjSink");
- return XA_RESULT_INTERNAL_ERROR;
- }
-
- if ( cameraBinGhostPad )
- {
- gst_object_unref( cameraBinGhostPad );
- }
- if ( ghost )
- {
- gst_object_unref( ghost );
- }
- if ( mpGhostSink )
- {
- gst_object_unref( mpGhostSink );
- }
- if ( camTee )
- {
- gst_object_unref( camTee );
- }
- }
- }
- else
- {
- DEBUG_ERR("Could not create decoder bin!!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
-
- /* create and add video stream pipeline */
- if(!ctx->xaLEDArray && !ctx->xaVibra && mediatype!=XA_MEDIATYPE_AUDIO) /*no video for these*/
- {
- /* create video processing pipeline */
- ctx->videoppbin = XAAdaptationBase_CreateVideoPP( );
- if( ctx->videoppbin )
- {
- gst_bin_add(GST_BIN(ctx->baseObj.bin), ctx->videoppbin);
- //shyward ---link filtered???
- // boolRetVal = gst_element_link_filtered(appsrc, videosink, caps);
- //if(!gst_element_link(ctx->codecbin, ctx->videoppbin))
- if(!gst_element_link_filtered(ctx->codecbin, ctx->videoppbin,encSrcCaps))
- {
- /* probably dynamic pads in codecbin */
- DEBUG_INFO("Could not link codec to videopp, trying delayed link");
- delayedlink = XA_BOOLEAN_TRUE;
- }
- ctx->videoScrSrcPad = gst_element_get_static_pad(ctx->videoppbin, "videopp_src");
- }
- else
- {
- DEBUG_ERR("Could not create video pp bin!!!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
- //shyward
- /* Black screen pipeline not needed under Symbian. May need to revist for acceptance testing
- ctx->videoppBScrbin = XAAdaptationBase_CreateVideoPPBlackScr( );
- if( ctx->videoppBScrbin )
- {
- gst_bin_add(GST_BIN(ctx->baseObj.bin), ctx->videoppBScrbin);
- ctx->blackScrSrcPad = gst_element_get_static_pad(ctx->videoppBScrbin, "videoppBSrc_src");
- }
- else
- {
- DEBUG_ERR("Could not create video pp bin for black screen!!!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
- */
- ctx->inputSelector = XAAdaptationBase_CreateInputSelector( );
- if( ctx->inputSelector )
- {
- gst_bin_add(GST_BIN(ctx->baseObj.bin), ctx->inputSelector);
- ctx->videoScrSinkPad = gst_element_get_request_pad(ctx->inputSelector, "sink%d");
- ctx->blackScrSinkPad = gst_element_get_request_pad(ctx->inputSelector, "sink%d");
- gst_pad_link(ctx->blackScrSrcPad, ctx->blackScrSinkPad);
- gst_pad_link(ctx->videoScrSrcPad, ctx->videoScrSinkPad);
- }
-
- //shyward - We have no video filter at this time
- /*
- ctx->filter = gst_element_factory_make("ffmpegcolorspace", "videofilter");
- gst_bin_add( GST_BIN(ctx->baseObj.bin), ctx->filter);
- if ( !gst_element_link( ctx->inputSelector, ctx->filter ) )
- {
- DEBUG_ERR("Could not link ctx->filter <-> ctx->inputSelector");
- return XA_RESULT_INTERNAL_ERROR;
- }
- */
- /* create video pipe sink */
- ctx->videosink = XAAdaptationBase_CreateGstSink( ctx->xaVideoSink, "videosink", &(ctx->isobjvsink) );
- /* NOTE: no valid object sinks for video output available */
- if( ctx->videosink )
- {
- gst_bin_add(GST_BIN(ctx->baseObj.bin), ctx->videosink);
-
- //shyward
- //if(!gst_element_link(ctx->filter, ctx->videosink))
- if(!gst_element_link_filtered(ctx->videoppbin, ctx->videosink,encSrcCaps))
- {
- DEBUG_ERR("Could not link videopp to videosink!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
- else
- {
- gst_caps_unref(encSrcCaps);
- }
- }
- else
- {
- DEBUG_ERR("Could not create video sink!!!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
- }
- else
- {
- DEBUG_INFO("Media does not contain video!");
- }
-
- /* create and add audio stream pipeline */
-
- if(!ctx->xaLEDArray && !ctx->xaVibra && mediatype!=XA_MEDIATYPE_IMAGE) /*no audio for these*/
- {
- /* create audio post processing pipeline */
- ctx->audioppbin = XAAdaptationBase_CreateAudioPP( );
- if( ctx->audioppbin )
- {
- gst_bin_add(GST_BIN(ctx->baseObj.bin), ctx->audioppbin);
- if(!gst_element_link(ctx->codecbin, ctx->audioppbin))
- {
- DEBUG_INFO("Could not link codec to audiopp, trying delayed link");
- delayedlink = XA_BOOLEAN_TRUE;
- }
- }
- else
- {
- DEBUG_ERR("Could not create audio pp bin!!!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
- /* create audio pipe sink */
- ctx->audiosink = XAAdaptationBase_CreateGstSink( ctx->xaAudioSink, "audiosink", &(ctx->isobjasink) );
- if( ctx->audiosink )
- {
- if( ctx->isobjasink && ctx->xaAudioSink && ctx->xaAudioSink->pLocator )
- {
- locType = *(XAuint32*)(ctx->xaAudioSink->pLocator);
- switch ( locType )
- {
- case XA_DATALOCATOR_OUTPUTMIX:
- {
- XADataLocator_OutputMix* omix = (XADataLocator_OutputMix*)(ctx->xaAudioSink->pLocator);
- XAOMixImpl* omixDevice = (XAOMixImpl*)(*omix->outputMix);
- if(omixDevice)
- {
- XAOutputMixAdapt_ConnectObject(omixDevice->adaptationCtx, &(ctx->baseObj), ctx->audiosink);
- }
- break;
- }
- default:
- /* Vibra and LED need no handling */
- break;
- }
- }
- gst_bin_add(GST_BIN(ctx->baseObj.bin), ctx->audiosink);
- if(!gst_element_link(ctx->audioppbin, ctx->audiosink))
- {
- DEBUG_ERR("Could not link audiopp to audiosink!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
- }
- else
- {
- DEBUG_ERR("Could not create audio sink!!!!");
- return XA_RESULT_INTERNAL_ERROR;
- }
- }
- else
- {
- DEBUG_INFO("Media does not contain audio!");
- }
-
- if(delayedlink)
- {
- /* listen for dynamically created pads */
- g_signal_connect (ctx->codecbin, "pad-added", G_CALLBACK (XAMediaPlayerAdapt_NewPadCb), ctx);
- }
- locType = *((XAuint32*)(ctx->xaSource->pLocator));
- if( locType == XA_DATALOCATOR_CONTENTPIPE)
- {
-
- g_signal_connect (ctx->source, "need-data", G_CALLBACK (need_data_for_prerolling), ctx);
- g_signal_connect (ctx->source, "enough-data", G_CALLBACK (enough_data_for_prerolling), ctx);
- g_signal_connect (ctx->source, "push-buffer", G_CALLBACK (push_data_for_prerolling), ctx);
- }
- DEBUG_API("<-XAMediaPlayerAdapt_CreatePipeline");
- return ret;
-}
-
-/*
- * gboolean XAMediaPlayerAdapt_PositionUpdate(gpointer ctx)
- * callback.
- * If position tracking enabled, periodic timer calls this method every XA_ADAPT_PU_INTERVAL msecs
- * @return false to stop periodic calls
- */
-gboolean XAMediaPlayerAdapt_PositionUpdate(gpointer ctx)
-{
- XAAdaptationBaseCtx *bCtx = (XAAdaptationBaseCtx*) ctx;
- XAMediaPlayerAdaptationCtx* mCtx = (XAMediaPlayerAdaptationCtx*) ctx;
- gint64 position;
- XAmillisecond posMsec;
- GstFormat format = GST_FORMAT_TIME;
- XAAdaptEvent event = {XA_PLAYITFEVENTS, XA_ADAPT_POSITION_UPDATE_EVT, 1, NULL};
-
- DEBUG_API("->XAMediaPlayerAdapt_PositionUpdate");
- if ( !gst_element_query_position( GST_ELEMENT(bCtx->bin), &format, &position ) )
- {
- DEBUG_ERR("Gst: Failed to get position");
- return( mCtx->runpositiontimer );
- }
- DEBUG_INFO_A1("Current position %"GST_TIME_FORMAT, GST_TIME_ARGS(position));
- if( mCtx && mCtx->trackpositionenabled )
- {
- posMsec = GST_TIME_AS_MSECONDS(position);/*Warning ok due to used API specification*/
- DEBUG_INFO_A1("mCtx->trackpositionenabled sending update, position:&ld ", posMsec);
- /* send needed events */
- event.data=&posMsec;
- XAAdaptationBase_SendAdaptEvents(bCtx, &event );
- }
- if( mCtx && mCtx->loopingenabled)
- {
- DEBUG_INFO_A2("mCtx->loopingenabled, current position:%lu, loopend:%lu ", position, mCtx->loopend);
- if( (position >= mCtx->loopend) &&
- (mCtx->lastpos < mCtx->loopend) )
- {
- DEBUG_INFO_A2("Restart loop from %"GST_TIME_FORMAT" to %"GST_TIME_FORMAT,
- GST_TIME_ARGS(mCtx->loopstart), GST_TIME_ARGS(mCtx->loopend));
- gst_element_seek( bCtx->bin, mCtx->playrate, GST_FORMAT_TIME, GST_SEEK_FLAG_FLUSH|GST_SEEK_FLAG_ACCURATE,
- GST_SEEK_TYPE_SET, mCtx->loopstart,
- GST_SEEK_TYPE_NONE, GST_CLOCK_TIME_NONE );
- mCtx->lastpos = mCtx->loopstart;
- if( mCtx && mCtx->trackpositionenabled )
- {
- DEBUG_INFO_A1("mCtx->trackpositionenabled sending looping update, position:%&u ", posMsec);
- posMsec = GST_TIME_AS_MSECONDS(mCtx->lastpos);/*Warning ok due to used API specification*/
- /* send needed events */
- event.data=&posMsec;
- XAAdaptationBase_SendAdaptEvents(bCtx, &event );
- }
- }
- else
- {
- mCtx->lastpos = position;
- }
- }
- DEBUG_API("<-XAMediaPlayerAdapt_PositionUpdate");
- /* return false to stop timer */
- return( mCtx->runpositiontimer );
-}
-
-/*
- * XAresult XAMediaPlayerAdapt_UpdatePositionCbTimer
- * Enable/disable periodic position tracking callback timer
- */
-XAresult XAMediaPlayerAdapt_UpdatePositionCbTimer(XAMediaPlayerAdaptationCtx* mCtx)
-{
- DEBUG_API_A2("->XAMediaPlayerAdapt_UpdatePositionCbTimer: trackposition %u, tracklooping %u",
- mCtx->trackpositionenabled, mCtx->loopingenabled);
-
- if(mCtx->runpositiontimer==0 && (mCtx->trackpositionenabled || mCtx->loopingenabled))
- {
- DEBUG_INFO("Start position tracking timer");
- mCtx->positionCb = &XAMediaPlayerAdapt_PositionUpdate;
- /* if play is already on, create a timer to track position of playback */
- if( GST_STATE(mCtx->baseObj.bin) == GST_STATE_PLAYING )
- {
- mCtx->runpositiontimer = g_timeout_add(XA_ADAPT_PU_INTERVAL, mCtx->positionCb, mCtx);
- }
- }
- else if (mCtx->runpositiontimer!=0 && !(mCtx->trackpositionenabled || mCtx->loopingenabled))
- {
- DEBUG_INFO("Stop position tracking timer");
- mCtx->trackpositionenabled = XA_BOOLEAN_FALSE;
- if(mCtx->runpositiontimer > 0)
- {
- g_source_remove(mCtx->runpositiontimer);
- mCtx->runpositiontimer=0;
- }
- }
- DEBUG_API("<-XAMediaPlayerAdapt_UpdatePositionCbTimer");
- return XA_RESULT_SUCCESS;
-}
-
-/*
- * XAresult XAMediaPlayerAdapt_InitContentPipeSrc(ctx)
- * CP code: can be moved to context base
- */
-XAresult XAMediaPlayerAdapt_InitContentPipeSrc(XAMediaPlayerAdaptationCtx* ctx)
-{
- XAresult ret = XA_RESULT_SUCCESS;
- CPresult res;
- /*GstStateChangeReturn gstRet = GST_STATE_CHANGE_SUCCESS;*/
- DEBUG_API("->XAMediaPlayerAdapt_InitContentPipeSrc");
- ctx->baseObj.pipeSrcThrCtx.appSrc = GST_APP_SRC(ctx->source);
- ctx->baseObj.pipeSrcThrCtx.pipe = (XADataLocator_ContentPipe*)(ctx->xaSource->pLocator);
-
- /* Create thread for content pipe source*/
- ret = XAImpl_CreateThreadHandle( &(ctx->baseObj.pipeSrcThr) );
- if ( ret != XA_RESULT_SUCCESS )
- {
- DEBUG_ERR("Could not create thread for content pipe source!");
- DEBUG_API("<-XAMediaPlayerAdapt_InitContentPipeSrc");
- return ret;
- }
-
- /* Create semaphore for content pipe source */
- ret = XAImpl_CreateSemaphore( &(ctx->baseObj.pipeSrcThrCtx.stateSem));
- if ( ret != XA_RESULT_SUCCESS )
- {
- DEBUG_ERR("Could not create semaphore for content pipe source!");
- DEBUG_API("<-XAMediaPlayerAdapt_InitContentPipeSrc");
- return ret;
- }
-
- /* Open content pipe */
- res = ctx->baseObj.pipeSrcThrCtx.pipe->pContentPipe->Open(&(ctx->baseObj.pipeSrcThrCtx.dataHandle),
- (CPstring)(ctx->baseObj.pipeSrcThrCtx.pipe->URI),
- CP_AccessRead );
- if ( res == EXIT_FAILURE )
- {
- DEBUG_ERR("Could not open Content Pipe!")
- return XA_RESULT_INTERNAL_ERROR;
- }
-
- res = ctx->baseObj.pipeSrcThrCtx.pipe->pContentPipe->RegisterCallback( &(ctx->baseObj.pipeSrcThrCtx.dataHandle), &XAAdaptationBase_ContentPipeSrcCb);
- if ( res == EXIT_FAILURE )
- {
- DEBUG_ERR("Could not register content pipe callbacks!")
- return XA_RESULT_INTERNAL_ERROR;
- }
-
- /*gstRet = gst_element_set_state( GST_ELEMENT(ctx->source), GST_STATE_PAUSED);
- gst_element_sync_state_with_parent( GST_ELEMENT( ctx->source));*/
-
- XAImpl_StartThread( &(ctx->baseObj.pipeSrcThr), NULL, &XAAdaptationBase_ContentPipeScrThrFunc, &(ctx->baseObj.pipeSrcThrCtx) );
-
- DEBUG_API("<-XAMediaPlayerAdapt_InitContentPipeSrc");
- return ret;
-}