Calls and online meetings bots
Bots can interact with Teams calls and meetings using real-time voice, video, and screen sharing. With Microsoft Graph APIs for calls and online meetings, Teams apps can now interact with users using voice and video to enhance the experience. These APIs allow you to add the following new features:
- Interactive voice response (IVR).
- Call control.
- Access to real-time audio and video streams, including desktop and app sharing.
To use these Graph APIs in a Teams app, you create a bot and specify some additional information and permissions.
In addition, the Real-time Media Platform enables bots to interact with Teams calls and meetings using real-time voice, video, and screen sharing. A bot that participates in audio or video calls and online meetings is a regular Microsoft Teams bot with few extra features used to register the bot.
The Teams app manifest with two more settings supportsCalling
and supportsVideo
, Graph permissions for your bot's Microsoft App ID, and tenant admin consent enable you to register the bot. In registering a calls and meetings bot for Teams, the Webhook URL is mentioned, which is the webhook endpoint for all incoming calls to your bot. An application-hosted media bot requires the Microsoft.Graph.Communications.Calls.Media .NET library to access the audio and video media streams, and the bot must be deployed on a Windows Server machine or Windows Server guest Operating System (OS) in Azure. Bots on Teams support only a specific set of media formats for audio and video content.
Functions of calls and online meeting bots
Bots can perform the following functions by calling Microsoft Graph APIs for calls and online meetings:
Capabilities | Functions |
---|---|
Core capabilities | • Initiate a one-to-one call between two users. • Initiate a group call with more than two users (up to 350 users). • Upgrade a one-to-one call with two users into a group call with more than two users. • Join a group call after it starts. • Invite a Voice over Internet Protocol (VoIP) participant to join an ongoing group call. |
Mid-call control | • Turn the video on or off. • Mute or unmute the microphone. • Switch between cameras. • Put the call on hold or resume locally. • Active speaker. • Choose speaker for calls. • Choose microphone for calls. • Show the state of a participant, such as idle, early media, connecting, connected, on hold, in lobby, or disconnected. • Show the state of a call, such as early media, incoming, connecting, ringing, connected, on hold, disconnecting, or disconnected. • Show if a participant is on mute. • Show the reason why a participant left a call. |
Screen sharing | • Share the entire screen from within the app. • Share a specific app (from the list of running apps). • Share a web browser tab from the list of open tabs. • Share system audio during screen sharing. • Participant can view remote screen share. |
Roster | • List the participants in a call. • Remove a participant from a call. |
Public Switched Telephone Network (PSTN) | • Place a one-to-one call with a PSTN participant. • Place a group call with PSTN participants. • Upgrade a one-to-one call with a PSTN participant into a group call. • Dial-out from a group call as a PSTN participant. • Support for early media. |
Now, you must understand some core concepts, terminology, and conventions.
Terminologies
The following core concepts, terminology, and conventions guide you by using calls and online meetings bots:
- Audio or video calls
- Call types
- Signals
- Calls and online meetings
- Real-time media
Audio or video calls
Calls in Teams can be purely audio or audio and video. Instead of audio or video call, the term call is used.
Call types
Calls are either peer-to-peer between a person and your bot, or multiparty between your bot and two or more people in a group call.
Following are the different call types and permissions required for the call:
A user can initiate a peer-to-peer call with your bot or invite your bot into an existing multiparty call. The multiparty call isn't enabled yet in the Teams user interface.
Note
User initiated calls to a bot aren't supported on the Teams mobile client.
Graph permissions aren't necessary for a user to initiate a peer-to-peer call with your bot. Additional permissions are needed for your bot to participate in a multiparty call, or for your bot to initiate a peer-to-peer call with a user.
A call can start as peer-to-peer and eventually become a multiparty call. Your bot can initiate multiparty calls by inviting others, provided your bot has the proper permissions. If your bot doesn't have permissions to participate in group calls and if a participant adds another participant to the call, your bot is dropped from the call.
Signals
There are two types of signals, incoming call, and in-call. Following are the different features of signals:
To receive an incoming call, you enter an endpoint in your bot settings. This endpoint receives a notification when an incoming call is initiated. You can answer the call, reject it, or redirect it to someone else.
When a bot is in a call, there are APIs for muting and unmuting the bot and to start or stop sharing video or desktop content with other participants.
The bot can also access the list of participants, invite new participants, and mute them.
Calls and online meetings
From a Teams user's perspective, there are two kinds of online meetings, unplanned and scheduled. From a bot's perspective, both online meetings are the same. To a bot, an online meeting is a multiparty call between a set of participants and includes meeting coordinates. Meeting coordinates are the metadata for the meeting including botId
, chatId
associated with the meeting, joinUrl
, startTime
or endTime
, and so on.
Real-time media
When a bot is participating in a call or online meeting, it must deal with audio and video streams. When users talk on a call, show themselves on a webcam, or present their screens in a meeting, to a bot it's shown as audio and video streams. If a bot wants to say something as simple as, press 0 to reach the operator in an interactive voice response (IVR) scenario, it requires playing a .WAV file. Collectively, this is referred to as media or real-time media.
Real-time media refers to scenarios where media must be processed in real-time, as opposed to playback of previously recorded audio or video. Dealing with media streams, real-time media streams, is complex. Microsoft has created the Real-time Media Platform to handle these scenarios and to offload as much of the traditional heavy lifting of real-time media processing as possible. When the bot answers an incoming call or joins a new or existing call, it needs to tell the Real-time Media Platform how media is handled. If you're building an IVR application, you can offload the expensive audio processing to Microsoft. Alternately, if your bot requires direct access to media streams, that scenario is also supported. There are two types of media processing:
Service-hosted media: Bots focus on managing application workflow, such as routing calls and offload audio processing to the Microsoft Real-time Media Platform. With service-hosted media, you have several options to implement and host your bot. A service-hosted media bot can be implemented as a stateless service as it doesn't process media locally. Service-hosted media bots can use the following APIs:
PlayPrompt
for playing an audio clip.Record
for recording audio clips.SubscribeToTone
for subscribing to dual tone multiple frequency (DTMF) tones.For example, knowing when a user has pressed 0 to reach the operator.
Application-hosted media: For a bot to get direct access to the media, it needs a specific Graph permission. After your bot has the permission, the Real-time Media Library, and the Graph calling SDK helps you build rich, real-time media, and calling bots. An application-hosted bot must be hosted in a Windows environment. For more information, see application-hosted media bots.
Code sample
Sample name | Description | Graph |
---|---|---|
Graph communication | Graph communications to interact with Microsoft's communications platform. | View |
Next step
See also
Platform Docs