sourceinfo.containertype = ctVCD;
sourceinfo.is_video = TRUE;
}
- if ( (strncmp(filename, "http://", 7)) == 0 || (strncmp(filename, "udp://", 6)) == 0 || (strncmp(filename, "rtsp://", 7)) == 0 )
+ if ( (strncmp(filename, "http://", 7)) == 0 || (strncmp(filename, "udp://", 6)) == 0 || (strncmp(filename, "rtp://", 6)) == 0 || (strncmp(filename, "https://", 8)) == 0 || (strncmp(filename, "mms://", 6)) == 0 || (strncmp(filename, "rtsp://", 7)) == 0 )
sourceinfo.is_streaming = TRUE;
gchar *uri;
else
eDebug("eServiceMP3::gst_message from %s: %s (without structure)", sourceName, GST_MESSAGE_TYPE_NAME(msg));
#endif
- if ( GST_MESSAGE_TYPE (msg) == GST_MESSAGE_STATE_CHANGED )
+ switch (GST_MESSAGE_TYPE (msg))
{
- // only the pipeline message
- if(GST_MESSAGE_SRC(msg) != GST_OBJECT(m_gst_playbin))
+ case GST_MESSAGE_EOS:
+ m_event((iPlayableService*)this, evEOF);
+ break;
+ case GST_MESSAGE_STATE_CHANGED:
+ {
+ if(GST_MESSAGE_SRC(msg) != GST_OBJECT(m_gst_playbin))
return;
- GstState old_state, new_state;
- gst_message_parse_state_changed(msg, &old_state, &new_state, NULL);
+ GstState old_state, new_state;
+ gst_message_parse_state_changed(msg, &old_state, &new_state, NULL);
+
+ if(old_state == new_state)
+ return;
- if(old_state == new_state)
- return;
-
- eDebug("eServiceMP3::state transition %s -> %s", gst_element_state_get_name(old_state), gst_element_state_get_name(new_state));
-
- GstStateChange transition = (GstStateChange)GST_STATE_TRANSITION(old_state, new_state);
-
- switch(transition)
- {
- case GST_STATE_CHANGE_NULL_TO_READY:
- {
- }
- break;
- case GST_STATE_CHANGE_READY_TO_PAUSED:
- {
-
- } break;
- case GST_STATE_CHANGE_PAUSED_TO_PLAYING:
- {
-
- } break;
- case GST_STATE_CHANGE_PLAYING_TO_PAUSED:
- {
+ eDebug("eServiceMP3::state transition %s -> %s", gst_element_state_get_name(old_state), gst_element_state_get_name(new_state));
- } break;
- case GST_STATE_CHANGE_PAUSED_TO_READY:
- {
-
- } break;
- case GST_STATE_CHANGE_READY_TO_NULL:
+ GstStateChange transition = (GstStateChange)GST_STATE_TRANSITION(old_state, new_state);
+
+ switch(transition)
{
-
- } break;
- }
- }
-
- switch (GST_MESSAGE_TYPE (msg))
- {
- case GST_MESSAGE_EOS:
- m_event((iPlayableService*)this, evEOF);
+ case GST_STATE_CHANGE_NULL_TO_READY:
+ {
+ } break;
+ case GST_STATE_CHANGE_READY_TO_PAUSED:
+ {
+ } break;
+ case GST_STATE_CHANGE_PAUSED_TO_PLAYING:
+ {
+ } break;
+ case GST_STATE_CHANGE_PLAYING_TO_PAUSED:
+ {
+ } break;
+ case GST_STATE_CHANGE_PAUSED_TO_READY:
+ {
+ } break;
+ case GST_STATE_CHANGE_READY_TO_NULL:
+ {
+ } break;
+ }
break;
+ }
case GST_MESSAGE_ERROR:
{
gchar *debug;
eDebug("eServiceMP3::/tmp/.id3coverart %d bytes written ", ret);
m_event((iPlayableService*)this, evUser+13);
}
-
gst_tag_list_free(tags);
m_event((iPlayableService*)this, evUpdatedInfo);
break;
GstPad* pad = 0;
g_signal_emit_by_name (m_gst_playbin, "get-audio-pad", i, &pad);
GstCaps* caps = gst_pad_get_negotiated_caps(pad);
+ if (!caps)
+ continue;
GstStructure* str = gst_caps_get_structure(caps, 0);
gchar *g_type;
g_type = gst_structure_get_name(str);
{
gst_tag_list_get_string(tags, GST_TAG_AUDIO_CODEC, &g_codec);
gst_tag_list_get_string(tags, GST_TAG_LANGUAGE_CODE, &g_lang);
+ gst_tag_list_free(tags);
}
audio.language_code = std::string(g_lang);
audio.codec = std::string(g_codec);
m_audioStreams.push_back(audio);
g_free (g_lang);
g_free (g_codec);
+ gst_caps_unref(caps);
}
for (i = 0; i < n_text; i++)
if (strstr(eventname, "Changed"))
m_event((iPlayableService*)this, evVideoProgressiveChanged);
}
+ g_free(eventname);
}
}
}
m_subtitle_widget->setPage(page);
m_subtitle_pages.pop_front();
}
- } ;
-
+ }
gst_object_unref (clock);
+ gst_object_unref (syncsink);
}
RESULT eServiceMP3::enableSubtitles(eWidget *parent, ePyObject tuple)
PyObject *eServiceMP3::getCachedSubtitle()
{
- eDebug("eServiceMP3::getCachedSubtitle");
Py_RETURN_NONE;
}