/* * empathy-call-handler.c - Source for EmpathyCallHandler * Copyright (C) 2008-2009 Collabora Ltd. * @author Sjoerd Simons * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free Software * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA */ #include "config.h" #include #include #include #include #include #include #include "empathy-call-handler.h" #define DEBUG_FLAG EMPATHY_DEBUG_VOIP #include G_DEFINE_TYPE(EmpathyCallHandler, empathy_call_handler, G_TYPE_OBJECT) /* signal enum */ enum { CONFERENCE_ADDED, CONFERENCE_REMOVED, SRC_PAD_ADDED, CONTENT_ADDED, CONTENT_REMOVED, CLOSED, CANDIDATES_CHANGED, STATE_CHANGED, FRAMERATE_CHANGED, RESOLUTION_CHANGED, LAST_SIGNAL }; static guint signals[LAST_SIGNAL] = {0}; enum { PROP_CALL_CHANNEL = 1, PROP_GST_BUS, PROP_CONTACT, PROP_INITIAL_AUDIO, PROP_INITIAL_VIDEO, PROP_SEND_AUDIO_CODEC, PROP_SEND_VIDEO_CODEC, PROP_RECV_AUDIO_CODECS, PROP_RECV_VIDEO_CODECS, PROP_AUDIO_REMOTE_CANDIDATE, PROP_VIDEO_REMOTE_CANDIDATE, PROP_AUDIO_LOCAL_CANDIDATE, PROP_VIDEO_LOCAL_CANDIDATE, }; /* private structure */ struct _EmpathyCallHandlerPriv { TpCallChannel *call; EmpathyContact *contact; TfChannel *tfchannel; gboolean initial_audio; gboolean initial_video; FsCodec *send_audio_codec; FsCodec *send_video_codec; GList *recv_audio_codecs; GList *recv_video_codecs; FsCandidate *audio_remote_candidate; FsCandidate *video_remote_candidate; FsCandidate *audio_local_candidate; FsCandidate *video_local_candidate; gboolean accept_when_initialised; }; #define GET_PRIV(obj) EMPATHY_GET_PRIV (obj, EmpathyCallHandler) static void empathy_call_handler_dispose (GObject *object) { EmpathyCallHandlerPriv *priv = GET_PRIV (object); tp_clear_object (&priv->tfchannel); tp_clear_object (&priv->call); tp_clear_object (&priv->contact); G_OBJECT_CLASS (empathy_call_handler_parent_class)->dispose (object); } static void empathy_call_handler_finalize (GObject *object) { EmpathyCallHandlerPriv *priv = GET_PRIV (object); fs_codec_destroy (priv->send_audio_codec); fs_codec_destroy (priv->send_video_codec); fs_codec_list_destroy (priv->recv_audio_codecs); fs_codec_list_destroy (priv->recv_video_codecs); fs_candidate_destroy (priv->audio_remote_candidate); fs_candidate_destroy (priv->video_remote_candidate); fs_candidate_destroy (priv->audio_local_candidate); fs_candidate_destroy (priv->video_local_candidate); G_OBJECT_CLASS (empathy_call_handler_parent_class)->finalize (object); } static void empathy_call_handler_init (EmpathyCallHandler *obj) { EmpathyCallHandlerPriv *priv = G_TYPE_INSTANCE_GET_PRIVATE (obj, EMPATHY_TYPE_CALL_HANDLER, EmpathyCallHandlerPriv); obj->priv = priv; } static void on_call_accepted_cb (GObject *source_object, GAsyncResult *res, gpointer user_data) { TpCallChannel *call = TP_CALL_CHANNEL (source_object); GError *error = NULL; if (!tp_call_channel_accept_finish (call, res, &error)) { g_warning ("could not accept Call: %s", error->message); g_error_free (error); } } static void on_call_invalidated_cb (TpCallChannel *call, guint domain, gint code, gchar *message, EmpathyCallHandler *self) { EmpathyCallHandlerPriv *priv = self->priv; if (priv->call == call) { /* Invalidated unexpectedly? Fake call ending */ g_signal_emit (self, signals[STATE_CHANGED], 0, TP_CALL_STATE_ENDED, NULL); priv->accept_when_initialised = FALSE; tp_clear_object (&priv->call); tp_clear_object (&priv->tfchannel); } } static void on_call_state_changed_cb (TpCallChannel *call, TpCallState state, TpCallFlags flags, TpCallStateReason *reason, GHashTable *details, EmpathyCallHandler *handler) { EmpathyCallHandlerPriv *priv = handler->priv; /* Clean up the TfChannel before bubbling the state-change signal * further up. This ensures that the conference-removed signal is * emitted before state-changed so that the client gets a chance * to remove the conference from the pipeline before resetting the * pipeline itself. */ if (state == TP_CALL_STATE_ENDED) { tp_channel_close_async (TP_CHANNEL (call), NULL, NULL); priv->accept_when_initialised = FALSE; tp_clear_object (&priv->call); tp_clear_object (&priv->tfchannel); } g_signal_emit (handler, signals[STATE_CHANGED], 0, state, reason->dbus_reason); if (state == TP_CALL_STATE_INITIALISED && priv->accept_when_initialised) { tp_call_channel_accept_async (priv->call, on_call_accepted_cb, NULL); priv->accept_when_initialised = FALSE; } } static void empathy_call_handler_set_property (GObject *object, guint property_id, const GValue *value, GParamSpec *pspec) { EmpathyCallHandlerPriv *priv = GET_PRIV (object); switch (property_id) { case PROP_CONTACT: priv->contact = g_value_dup_object (value); break; case PROP_CALL_CHANNEL: g_return_if_fail (priv->call == NULL); priv->call = g_value_dup_object (value); tp_g_signal_connect_object (priv->call, "state-changed", G_CALLBACK (on_call_state_changed_cb), object, 0); tp_g_signal_connect_object (priv->call, "invalidated", G_CALLBACK (on_call_invalidated_cb), object, 0); break; case PROP_INITIAL_AUDIO: priv->initial_audio = g_value_get_boolean (value); break; case PROP_INITIAL_VIDEO: priv->initial_video = g_value_get_boolean (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec); } } static void empathy_call_handler_get_property (GObject *object, guint property_id, GValue *value, GParamSpec *pspec) { EmpathyCallHandlerPriv *priv = GET_PRIV (object); switch (property_id) { case PROP_CONTACT: g_value_set_object (value, priv->contact); break; case PROP_CALL_CHANNEL: g_value_set_object (value, priv->call); break; case PROP_INITIAL_AUDIO: g_value_set_boolean (value, priv->initial_audio); break; case PROP_INITIAL_VIDEO: g_value_set_boolean (value, priv->initial_video); break; case PROP_SEND_AUDIO_CODEC: g_value_set_boxed (value, priv->send_audio_codec); break; case PROP_SEND_VIDEO_CODEC: g_value_set_boxed (value, priv->send_video_codec); break; case PROP_RECV_AUDIO_CODECS: g_value_set_boxed (value, priv->recv_audio_codecs); break; case PROP_RECV_VIDEO_CODECS: g_value_set_boxed (value, priv->recv_video_codecs); break; case PROP_AUDIO_REMOTE_CANDIDATE: g_value_set_boxed (value, priv->audio_remote_candidate); break; case PROP_VIDEO_REMOTE_CANDIDATE: g_value_set_boxed (value, priv->video_remote_candidate); break; case PROP_AUDIO_LOCAL_CANDIDATE: g_value_set_boxed (value, priv->audio_local_candidate); break; case PROP_VIDEO_LOCAL_CANDIDATE: g_value_set_boxed (value, priv->video_local_candidate); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec); } } static void empathy_call_handler_class_init (EmpathyCallHandlerClass *klass) { GObjectClass *object_class = G_OBJECT_CLASS (klass); GParamSpec *param_spec; g_type_class_add_private (klass, sizeof (EmpathyCallHandlerPriv)); object_class->set_property = empathy_call_handler_set_property; object_class->get_property = empathy_call_handler_get_property; object_class->dispose = empathy_call_handler_dispose; object_class->finalize = empathy_call_handler_finalize; param_spec = g_param_spec_object ("target-contact", "TargetContact", "The contact", EMPATHY_TYPE_CONTACT, G_PARAM_READWRITE | G_PARAM_CONSTRUCT_ONLY | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_CONTACT, param_spec); param_spec = g_param_spec_object ("call-channel", "call channel", "The call channel", TP_TYPE_CALL_CHANNEL, G_PARAM_READWRITE | G_PARAM_CONSTRUCT_ONLY | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_CALL_CHANNEL, param_spec); param_spec = g_param_spec_boolean ("initial-audio", "initial-audio", "Whether the call should start with audio", TRUE, G_PARAM_READWRITE | G_PARAM_CONSTRUCT_ONLY | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_INITIAL_AUDIO, param_spec); param_spec = g_param_spec_boolean ("initial-video", "initial-video", "Whether the call should start with video", FALSE, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_INITIAL_VIDEO, param_spec); param_spec = g_param_spec_boxed ("send-audio-codec", "send audio codec", "Codec used to encode the outgoing video stream", FS_TYPE_CODEC, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_SEND_AUDIO_CODEC, param_spec); param_spec = g_param_spec_boxed ("send-video-codec", "send video codec", "Codec used to encode the outgoing video stream", FS_TYPE_CODEC, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_SEND_VIDEO_CODEC, param_spec); param_spec = g_param_spec_boxed ("recv-audio-codecs", "recvs audio codec", "Codecs used to decode the incoming audio stream", FS_TYPE_CODEC_LIST, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_RECV_AUDIO_CODECS, param_spec); param_spec = g_param_spec_boxed ("recv-video-codecs", "recvs video codec", "Codecs used to decode the incoming video stream", FS_TYPE_CODEC_LIST, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_RECV_VIDEO_CODECS, param_spec); param_spec = g_param_spec_boxed ("audio-remote-candidate", "audio remote candidate", "Remote candidate used for the audio stream", FS_TYPE_CANDIDATE, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_AUDIO_REMOTE_CANDIDATE, param_spec); param_spec = g_param_spec_boxed ("video-remote-candidate", "video remote candidate", "Remote candidate used for the video stream", FS_TYPE_CANDIDATE, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_VIDEO_REMOTE_CANDIDATE, param_spec); param_spec = g_param_spec_boxed ("audio-local-candidate", "audio local candidate", "Local candidate used for the audio stream", FS_TYPE_CANDIDATE, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_AUDIO_REMOTE_CANDIDATE, param_spec); param_spec = g_param_spec_boxed ("video-local-candidate", "video local candidate", "Local candidate used for the video stream", FS_TYPE_CANDIDATE, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_VIDEO_REMOTE_CANDIDATE, param_spec); signals[CONFERENCE_ADDED] = g_signal_new ("conference-added", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_LAST, 0, NULL, NULL, g_cclosure_marshal_generic, G_TYPE_NONE, 1, FS_TYPE_CONFERENCE); signals[CONFERENCE_REMOVED] = g_signal_new ("conference-removed", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_LAST, 0, NULL, NULL, g_cclosure_marshal_generic, G_TYPE_NONE, 1, FS_TYPE_CONFERENCE); signals[SRC_PAD_ADDED] = g_signal_new ("src-pad-added", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_LAST, 0, NULL, NULL, g_cclosure_marshal_generic, G_TYPE_BOOLEAN, 2, TF_TYPE_CONTENT, GST_TYPE_PAD); signals[CONTENT_ADDED] = g_signal_new ("content-added", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_LAST, 0, NULL, NULL, g_cclosure_marshal_generic, G_TYPE_BOOLEAN, 1, TF_TYPE_CONTENT); signals[CONTENT_REMOVED] = g_signal_new ("content-removed", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_LAST, 0, NULL, NULL, g_cclosure_marshal_generic, G_TYPE_BOOLEAN, 1, TF_TYPE_CONTENT); signals[CLOSED] = g_signal_new ("closed", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_LAST, 0, NULL, NULL, g_cclosure_marshal_generic, G_TYPE_NONE, 0); signals[CANDIDATES_CHANGED] = g_signal_new ("candidates-changed", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_LAST, 0, NULL, NULL, g_cclosure_marshal_generic, G_TYPE_NONE, 1, G_TYPE_UINT); signals[STATE_CHANGED] = g_signal_new ("state-changed", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_LAST, 0, NULL, NULL, g_cclosure_marshal_generic, G_TYPE_NONE, 2, G_TYPE_UINT, G_TYPE_STRING); signals[FRAMERATE_CHANGED] = g_signal_new ("framerate-changed", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_LAST, 0, NULL, NULL, g_cclosure_marshal_generic, G_TYPE_NONE, 1, G_TYPE_UINT); signals[RESOLUTION_CHANGED] = g_signal_new ("resolution-changed", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_LAST, 0, NULL, NULL, g_cclosure_marshal_generic, G_TYPE_NONE, 2, G_TYPE_UINT, G_TYPE_UINT); } EmpathyCallHandler * empathy_call_handler_new_for_channel (TpCallChannel *call, EmpathyContact *contact) { return EMPATHY_CALL_HANDLER (g_object_new (EMPATHY_TYPE_CALL_HANDLER, "call-channel", call, "initial-video", tp_call_channel_has_initial_video (call, NULL), "target-contact", contact, NULL)); } static void update_sending_codec (EmpathyCallHandler *self, FsCodec *codec, FsSession *session) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); FsMediaType type; if (codec == NULL || session == NULL) return; g_object_get (session, "media-type", &type, NULL); if (type == FS_MEDIA_TYPE_AUDIO) { priv->send_audio_codec = fs_codec_copy (codec); g_object_notify (G_OBJECT (self), "send-audio-codec"); } else if (type == FS_MEDIA_TYPE_VIDEO) { priv->send_video_codec = fs_codec_copy (codec); g_object_notify (G_OBJECT (self), "send-video-codec"); } } static void update_receiving_codec (EmpathyCallHandler *self, GList *codecs, FsStream *stream) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); FsSession *session; FsMediaType type; if (codecs == NULL || stream == NULL) return; g_object_get (stream, "session", &session, NULL); if (session == NULL) return; g_object_get (session, "media-type", &type, NULL); if (type == FS_MEDIA_TYPE_AUDIO) { priv->recv_audio_codecs = fs_codec_list_copy (codecs); g_object_notify (G_OBJECT (self), "recv-audio-codecs"); } else if (type == FS_MEDIA_TYPE_VIDEO) { priv->recv_video_codecs = fs_codec_list_copy (codecs); g_object_notify (G_OBJECT (self), "recv-video-codecs"); } g_object_unref (session); } static void update_candidates (EmpathyCallHandler *self, FsCandidate *remote_candidate, FsCandidate *local_candidate, FsStream *stream) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); FsSession *session; FsMediaType type; if (stream == NULL) return; g_object_get (stream, "session", &session, NULL); if (session == NULL) return; g_object_get (session, "media-type", &type, NULL); if (type == FS_MEDIA_TYPE_AUDIO) { if (remote_candidate != NULL) { fs_candidate_destroy (priv->audio_remote_candidate); priv->audio_remote_candidate = fs_candidate_copy (remote_candidate); g_object_notify (G_OBJECT (self), "audio-remote-candidate"); } if (local_candidate != NULL) { fs_candidate_destroy (priv->audio_local_candidate); priv->audio_local_candidate = fs_candidate_copy (local_candidate); g_object_notify (G_OBJECT (self), "audio-local-candidate"); } g_signal_emit (G_OBJECT (self), signals[CANDIDATES_CHANGED], 0, FS_MEDIA_TYPE_AUDIO); } else if (type == FS_MEDIA_TYPE_VIDEO) { if (remote_candidate != NULL) { fs_candidate_destroy (priv->video_remote_candidate); priv->video_remote_candidate = fs_candidate_copy (remote_candidate); g_object_notify (G_OBJECT (self), "video-remote-candidate"); } if (local_candidate != NULL) { fs_candidate_destroy (priv->video_local_candidate); priv->video_local_candidate = fs_candidate_copy (local_candidate); g_object_notify (G_OBJECT (self), "video-local-candidate"); } g_signal_emit (G_OBJECT (self), signals[CANDIDATES_CHANGED], 0, FS_MEDIA_TYPE_VIDEO); } g_object_unref (session); } void empathy_call_handler_bus_message (EmpathyCallHandler *handler, GstBus *bus, GstMessage *message) { EmpathyCallHandlerPriv *priv = GET_PRIV (handler); const GstStructure *s = gst_message_get_structure (message); if (priv->tfchannel == NULL) return; if (s != NULL && gst_structure_has_name (s, "farsight-send-codec-changed")) { const GValue *val; FsCodec *codec; FsSession *session; DEBUG ("farsight-send-codec-changed"); val = gst_structure_get_value (s, "codec"); codec = g_value_get_boxed (val); val = gst_structure_get_value (s, "session"); session = g_value_get_object (val); update_sending_codec (handler, codec, session); } else if (s != NULL && gst_structure_has_name (s, "farsight-recv-codecs-changed")) { const GValue *val; GList *codecs; FsStream *stream; DEBUG ("farsight-recv-codecs-changed"); val = gst_structure_get_value (s, "codecs"); codecs = g_value_get_boxed (val); val = gst_structure_get_value (s, "stream"); stream = g_value_get_object (val); update_receiving_codec (handler, codecs, stream); } else if (s != NULL && gst_structure_has_name (s, "farsight-new-active-candidate-pair")) { const GValue *val; FsCandidate *remote_candidate, *local_candidate; FsStream *stream; DEBUG ("farsight-new-active-candidate-pair"); val = gst_structure_get_value (s, "remote-candidate"); remote_candidate = g_value_get_boxed (val); val = gst_structure_get_value (s, "local-candidate"); local_candidate = g_value_get_boxed (val); val = gst_structure_get_value (s, "stream"); stream = g_value_get_object (val); update_candidates (handler, remote_candidate, local_candidate, stream); } tf_channel_bus_message (priv->tfchannel, message); } static void on_tf_channel_conference_added_cb (TfChannel *tfchannel, GstElement *conference, EmpathyCallHandler *self) { g_signal_emit (G_OBJECT (self), signals[CONFERENCE_ADDED], 0, conference); } static void on_tf_channel_conference_removed_cb (TfChannel *tfchannel, FsConference *conference, EmpathyCallHandler *self) { g_signal_emit (G_OBJECT (self), signals[CONFERENCE_REMOVED], 0, GST_ELEMENT (conference)); } static gboolean src_pad_added_error_idle (gpointer data) { TfContent *content = data; tf_content_error_literal (content, "Could not link sink"); g_object_unref (content); return FALSE; } static void on_tf_content_src_pad_added_cb (TfContent *content, guint handle, FsStream *stream, GstPad *pad, FsCodec *codec, EmpathyCallHandler *handler) { gboolean retval; g_signal_emit (G_OBJECT (handler), signals[SRC_PAD_ADDED], 0, content, pad, &retval); if (!retval) g_idle_add (src_pad_added_error_idle, g_object_ref (content)); } static void on_tf_content_framerate_changed (TfContent *content, GParamSpec *spec, EmpathyCallHandler *handler) { guint framerate; g_object_get (content, "framerate", &framerate, NULL); if (framerate != 0) g_signal_emit (G_OBJECT (handler), signals[FRAMERATE_CHANGED], 0, framerate); } static void on_tf_content_resolution_changed (TfContent *content, guint width, guint height, EmpathyCallHandler *handler) { if (width > 0 && height > 0) g_signal_emit (G_OBJECT (handler), signals[RESOLUTION_CHANGED], 0, width, height); } static void on_tf_channel_content_added_cb (TfChannel *tfchannel, TfContent *content, EmpathyCallHandler *handler) { FsMediaType mtype; FsSession *session; // FsStream *fs_stream; FsCodec *codec; // GList *codecs; gboolean retval; g_signal_connect (content, "src-pad-added", G_CALLBACK (on_tf_content_src_pad_added_cb), handler); #if 0 g_signal_connect (content, "start-sending", G_CALLBACK (on_tf_content_start_sending_cb), handler); g_signal_connect (content, "stop-sending", G_CALLBACK (on_tf_content_stop_sending_cb), handler); #endif g_signal_emit (G_OBJECT (handler), signals[CONTENT_ADDED], 0, content, &retval); if (!retval) tf_content_error_literal (content, "Could not link source"); /* Get sending codec */ g_object_get (content, "fs-session", &session, NULL); g_object_get (session, "current-send-codec", &codec, NULL); update_sending_codec (handler, codec, session); tp_clear_object (&session); tp_clear_object (&codec); /* Get receiving codec */ /* FIXME g_object_get (content, "fs-stream", &fs_stream, NULL); g_object_get (fs_stream, "current-recv-codecs", &codecs, NULL); update_receiving_codec (handler, codecs, fs_stream); fs_codec_list_destroy (codecs); tp_clear_object (&fs_stream); */ g_object_get (content, "media-type", &mtype, NULL); if (mtype == FS_MEDIA_TYPE_VIDEO) { guint framerate, width, height; g_signal_connect (content, "notify::framerate", G_CALLBACK (on_tf_content_framerate_changed), handler); g_signal_connect (content, "resolution-changed", G_CALLBACK (on_tf_content_resolution_changed), handler); g_object_get (content, "framerate", &framerate, "width", &width, "height", &height, NULL); if (framerate > 0) g_signal_emit (G_OBJECT (handler), signals[FRAMERATE_CHANGED], 0, framerate); if (width > 0 && height > 0) g_signal_emit (G_OBJECT (handler), signals[RESOLUTION_CHANGED], 0, width, height); } } static void on_tf_channel_content_removed_cb (TfChannel *tfchannel, TfContent *content, EmpathyCallHandler *handler) { gboolean retval; DEBUG ("removing content"); g_signal_emit (G_OBJECT (handler), signals[CONTENT_REMOVED], 0, content, &retval); if (!retval) { g_warning ("Could not remove content!"); tf_content_error_literal (content, "Could not link source"); } } static void on_tf_channel_closed_cb (TfChannel *tfchannel, EmpathyCallHandler *handler) { g_signal_emit (G_OBJECT (handler), signals[CLOSED], 0); } static void on_tf_channel_ready (GObject *source, GAsyncResult *result, gpointer user_data) { EmpathyCallHandler *self = EMPATHY_CALL_HANDLER (user_data); EmpathyCallHandlerPriv *priv = GET_PRIV (self); GError *error = NULL; priv->tfchannel = TF_CHANNEL (g_async_initable_new_finish ( G_ASYNC_INITABLE (source), result, NULL)); if (priv->tfchannel == NULL) { g_warning ("Failed to create Farstream channel: %s", error->message); g_error_free (error); return; } /* Set up the telepathy farstream channel */ g_signal_connect (priv->tfchannel, "closed", G_CALLBACK (on_tf_channel_closed_cb), self); g_signal_connect (priv->tfchannel, "fs-conference-added", G_CALLBACK (on_tf_channel_conference_added_cb), self); g_signal_connect (priv->tfchannel, "fs-conference-removed", G_CALLBACK (on_tf_channel_conference_removed_cb), self); g_signal_connect (priv->tfchannel, "content-added", G_CALLBACK (on_tf_channel_content_added_cb), self); g_signal_connect (priv->tfchannel, "content-removed", G_CALLBACK (on_tf_channel_content_removed_cb), self); } static void empathy_call_handler_start_tpfs (EmpathyCallHandler *self) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); tf_channel_new_async (TP_CHANNEL (priv->call), on_tf_channel_ready, self); } static void empathy_call_handler_request_cb (GObject *source, GAsyncResult *result, gpointer user_data) { EmpathyCallHandler *self = EMPATHY_CALL_HANDLER (user_data); EmpathyCallHandlerPriv *priv = GET_PRIV (self); TpChannel *channel; GError *error = NULL; TpAccountChannelRequest *req = TP_ACCOUNT_CHANNEL_REQUEST (source); channel = tp_account_channel_request_create_and_handle_channel_finish (req, result, NULL, &error); if (channel == NULL) { DEBUG ("Failed to create the channel: %s", error->message); g_error_free (error); return; } if (!TP_IS_CALL_CHANNEL (channel)) { DEBUG ("The channel is not a Call channel!"); return; } priv->call = TP_CALL_CHANNEL (channel); tp_g_signal_connect_object (priv->call, "state-changed", G_CALLBACK (on_call_state_changed_cb), self, 0); tp_g_signal_connect_object (priv->call, "invalidated", G_CALLBACK (on_call_invalidated_cb), self, 0); g_object_notify (G_OBJECT (self), "call-channel"); empathy_call_handler_start_tpfs (self); tp_call_channel_accept_async (priv->call, on_call_accepted_cb, NULL); } void empathy_call_handler_start_call (EmpathyCallHandler *handler, gint64 timestamp) { EmpathyCallHandlerPriv *priv = GET_PRIV (handler); TpAccountChannelRequest *req; TpAccount *account; GHashTable *request; if (priv->call != NULL) { empathy_call_handler_start_tpfs (handler); if (tp_channel_get_requested (TP_CHANNEL (priv->call))) { /* accept outgoing channels immediately */ tp_call_channel_accept_async (priv->call, on_call_accepted_cb, NULL); } else { /* accepting incoming channels when they are INITIALISED */ if (tp_call_channel_get_state (priv->call, NULL, NULL, NULL) == TP_CALL_STATE_INITIALISED) tp_call_channel_accept_async (priv->call, on_call_accepted_cb, NULL); else priv->accept_when_initialised = TRUE; } return; } /* No TpCallChannel (we are redialing). Request a new call channel */ g_assert (priv->contact != NULL); account = empathy_contact_get_account (priv->contact); request = empathy_call_create_call_request ( empathy_contact_get_id (priv->contact), priv->initial_audio, priv->initial_video); req = tp_account_channel_request_new (account, request, timestamp); tp_account_channel_request_create_and_handle_channel_async (req, NULL, empathy_call_handler_request_cb, handler); g_object_unref (req); g_hash_table_unref (request); } /** * empathy_call_handler_stop_call: * @handler: an #EmpathyCallHandler * * Closes the #EmpathyCallHandler's call and frees its resources. */ void empathy_call_handler_stop_call (EmpathyCallHandler *handler) { EmpathyCallHandlerPriv *priv = GET_PRIV (handler); if (priv->call != NULL) { tp_call_channel_hangup_async (priv->call, TP_CALL_STATE_CHANGE_REASON_USER_REQUESTED, "", "", NULL, NULL); } } /** * empathy_call_handler_has_initial_video: * @handler: an #EmpathyCallHandler * * Return %TRUE if the call managed by this #EmpathyCallHandler was * created with video enabled * * Return value: %TRUE if the call was created as a video conversation. */ gboolean empathy_call_handler_has_initial_video (EmpathyCallHandler *handler) { EmpathyCallHandlerPriv *priv = GET_PRIV (handler); return priv->initial_video; } FsCodec * empathy_call_handler_get_send_audio_codec (EmpathyCallHandler *self) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); return priv->send_audio_codec; } FsCodec * empathy_call_handler_get_send_video_codec (EmpathyCallHandler *self) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); return priv->send_video_codec; } GList * empathy_call_handler_get_recv_audio_codecs (EmpathyCallHandler *self) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); return priv->recv_audio_codecs; } GList * empathy_call_handler_get_recv_video_codecs (EmpathyCallHandler *self) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); return priv->recv_video_codecs; } FsCandidate * empathy_call_handler_get_audio_remote_candidate ( EmpathyCallHandler *self) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); return priv->audio_remote_candidate; } FsCandidate * empathy_call_handler_get_audio_local_candidate ( EmpathyCallHandler *self) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); return priv->audio_local_candidate; } FsCandidate * empathy_call_handler_get_video_remote_candidate ( EmpathyCallHandler *self) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); return priv->video_remote_candidate; } FsCandidate * empathy_call_handler_get_video_local_candidate ( EmpathyCallHandler *self) { EmpathyCallHandlerPriv *priv = GET_PRIV (self); return priv->video_local_candidate; } EmpathyContact * empathy_call_handler_get_contact (EmpathyCallHandler *self) { return self->priv->contact; }