Skip to content

Latest commit

 

History

History
320 lines (178 loc) · 35.7 KB

api_overview.md

File metadata and controls

320 lines (178 loc) · 35.7 KB

API Overview

This guide gives an overview of the API methods that you can use to create a meeting with audio and video.

1. Create a session

The MeetingSession and its AudioVideoFacade are the starting points for creating meetings. You will need to create a Logger and MeetingSessionConfiguration before creating a meeting session.

1a. Create a logger

You can utilize the ConsoleLogger to write logs with os_log. You can also implement the Logger protocol to customize the logging behavior.

let logger = ConsoleLogger(name: "test", level: LogLevel.DEBUG)

1b. Create a meeting session configuration

Create a MeetingSessionConfiguration object with the responses to chime:CreateMeeting and chime:CreateAttendee. Your server application should make these API calls and securely pass the meeting and attendee responses to the client application.

1c. Create a meeting session

Create a DefaultMeetingSession with the MeetingSessionConfiguration object created above.

let meetingSession = DefaultMeetingSession(configuration: configuration, logger: logger)

Configure the session

Before starting the meeting session, you should configure the audio device.

2a. Configure the audio device

To retrieve a list of available audio devices, call meetingSession.audioVideo.listAudioDevices().

To use the chosen audio device, call meetingSession.audioVideo.chooseAudioDevice(mediaDevice:).

2b. Register a device change observer (optional)

You can receive events about changes to available audio devices by implementing a DeviceChangeObserver and registering the observer with the audio video facade.

To add a DeviceChangeObserver, call meetingSession.audioVideo.addDeviceChangeObserver(observer:).

To remove a DeviceChangeObserver, call meetingSession.audioVideo.removeDeviceChangeObserver(observer:).

A DeviceChangeObserver has the following method:

3. Request permissions for audio and video

Before starting audio or video, you will need to request permissions from the user and verify that they are granted. In Xcode, open Info.plist and add NSCameraUsageDescription ("Privacy - Camera Usage Description") to the property list. If you are planning to have users join with AudioDeviceCapabilities.inputAndOutput (which is the default option), then also add NSMicrophoneUsageDescription ("Privacy - Microphone Usage Description") to the property list. This will allow the app to ask for microphone and camera permissions.

After doing this, you will also need to request permissions for microphone and camera access in your source code. You can either do this with AVAudioSession.recordPermission and AVCaptureDevice.authorizationStatus, handling the response synchronously and falling back to requesting permissions, or you can use requestRecordPermission and requestAccess with an asynchronous completion handler.

AVAudioSession.sharedInstance().requestRecordPermission // Only needed if joining meetings with `AudioDeviceCapabilities.inputAndOutput`

AVCaptureDevice.requestAccess(for: .video)

Calling the APIs without having the above permissions granted will result in a PermissionError.

4. Register an audio video observer

You can receive events about the audio session, video session, and connection health by implementing an AudioVideoObserver and registering the observer with the audio video facade.

To add an AudioVideoObserver, call meetingSession.audioVideo.addAudioVideoObserver(observer:).

To remove an AudioVideoObserver, call meetingSession.audioVideo.removeAudioVideoObserver(observer:).

An AudioVideoObserver has the following methods:

5. Starting and stopping the meeting session

Call this method after doing pre-requisite configuration (See previous sections).

To start the meeting session, call meetingSession.audioVideo.start(callKitEnabled:). This will start underlying media clients and will start sending and receiving audio. Alternatively, call meetingSession.audioVideo.start() if the call is not reported to CallKit so that audio interruptions will be handled by the SDK itself.

To stop the meeting session, call meetingSession.audioVideo.stop().

6. Building a roster of participants

6a. Register a realtime observer

You can use a RealtimeObserver to learn when attendees join and leave and when their volume level, mute state, or signal strength changes.

To add a RealtimeObserver, call meetingSession.audioVideo.addRealtimeObserver(observer:).

To remove a RealtimeObserver, call meetingSession.audioVideo.removeRealtimeObserver(observer:).

A RealtimeObserver has the following methods:

Note that only attendees whose volume level, mute state, or signal strength has changed will be included. All callbacks provide both the attendee ID and external user ID from chime:CreateAttendee so that you may map between the two IDs.

6b. Register an active speaker observer (optional)

If you are interested in detecting the active speaker (e.g. to display the active speaker's video as a large, central tile), implement an ActiveSpeakerObserver and register the observer with the audio video facade.

You will also need to provide an ActiveSpeakerPolicy. You can use DefaultActiveSpeakerPolicy or implement the ActiveSpeakerPolicy protocol to customize the policy.

To add an ActiveSpeakerObserver, call meetingSession.audioVideo.addActiveSpeakerObserver(policy:observer:).

To remove an ActiveSpeakerObserver, call meetingSession.audioVideo.removeActiveSpeakerObserver(observer:).

An ActiveSpeakerObserver has the following methods:

You can control onActiveSpeakerScoreChanged's interval by providing a value for scoresCallbackIntervalMs while implementing ActiveSpeakerPolicy. You can prevent this callback from being triggered by using a nil value.

7. Mute and unmute audio

To mute the local attendee's audio, call meetingSession.audioVideo.realtimeLocalMute().

To unmute the local attendee's audio, call meetingSession.audioVideo.realtimeLocalUnmute().

8. Share and display video

You can use the following methods in order to send, receive, and display video.

A VideoTile is a binding of a tile ID, an attendee ID, that attendee's video, and a video view. The VideoTileState will contain further information such as if the video tile is for the local attendee. Video tiles start without a video view bound to it.

You can view content share the same way that you view a remote attendee's video. The video tile state will contain additional information to distinguish if that video tile is for content share.

8a. Sending video

Video permissions are required for sending the local attendee's video.

To start sending the local attendee's video, call meetingSession.audioVideo.startLocalVideo().

To start sending video with a local video configuration, call meetingSession.audioVideo.startLocalVideo(config:). To start local video with a provided custom VideoSource, call meetingSession.audioVideo.startLocalVideo(source:).

To stop sending the local attendee's video, call meetingSession.audioVideo.stopLocalVideo().

8b. Getting and switching video device

When starting the local attendee's video, the underlying media client will use the active video device or the front facing camera if there is no active video device. You can use the following methods to get or switch the active video device.

To get the active video device, call meetingSession.audioVideo.getActiveCamera().

To switch the active video device, call meetingSession.audioVideo.switchCamera().

8c. Receiving video

To start receiving video from remote attendees, call meetingSession.audioVideo.startRemoteVideo().

To stop receiving video from remote attendees, call meetingSession.audioVideo.stopRemoteVideo().

8d. Adding a video tile observer

You will need to implement a VideoTileObserver and register the observer with the audio video facade to receive video tile events for displaying video.

To add a VideoTileObserver, call meetingSession.audioVideo.addVideoTileObserver(observer:).

To remove a VideoTileObserver, call meetingSession.audioVideo.removeVideoTileObserver(observer:).

A VideoTileObserver has the following methods:

A pause or resume event can occur when the underlying media client pauses the video tile for connection reasons or when the pause or resume video tile methods are called.

The video tile state is represented by a VideoPauseState that describes if and why (e.g., paused by user request, or paused for poor connection) it was paused.

8e. Binding a video tile to a video view

To display video, you will also need to bind a video view to a video tile. Create a VideoRenderView and bind that view to the video tile in VideoTileObserver's onVideoTileAdded method. You can use DefaultVideoRenderView or customize the behavior by implementing the VideoRenderView protocol.

To bind a video tile to a view, call meetingSession.audioVideo.bindVideoView(videoView:tileId:).

To unbind a video tile from a view, call meetingSession.audioVideo.unbindVideoView(tileId:).

8f. Pausing a remote video tile

To pause a remote attendee's video tile, call meetingSession.audioVideo.pauseRemoteVideoTile(tileId:).

To resume a remote attendee's video tile, call meetingSession.audioVideo.resumeRemoteVideoTile(tileId:).

8g. In-depth look and comparison between video APIs

Start/Stop Local Video

As the names suggest, startLocalVideo() and stopLocalVideo() will start and stop sending local attendee’s video respectively. startLocalVideo() will first check if the video client is initialized and the application has permission to access device camera, then it will update the service type of video client and turn on its sending mode, so that local video can be sent to server and forwarded to other peers.

You should call startLocalVideo() when you want to start sending local video to others, and call stopLocalVideo() when you want to stop sending local video to others (e.g. when the app is in background). Stopping local video will save uplink bandwidth and computation resource for encoding video.

Start/Stop Remote Video

Similar to startLocalVideo() and stopLocalVideo(), startRemoteVideo() and stopRemoteVideo() will start and stop receiving remote attendees’ videos by setting the service type of video client and turning on/off its receiving mode.

Note that when the application calls stopRemoteVideo(), all the resources in the render pipeline will be released, and it takes some time to re-initialize all these resources. You should call stopRemoteVideo() only when you want to stop receiving remote videos for a considerable amount of time. If you want to temporarily stop remote videos, consider iterating through remote videos and call pauseRemoteVideoTile(tileId:) instead.

Pause/Resume Remote Video Tile

You can call pauseRemoteVideoTile(tileId:) to temporarily pause a remote video stream. When you call pauseRemoteVideoTile(tileId:), the video client will add the tile id to a local block list, so that server will not forward video frames from the paused stream to the client. In that case, you can save downlink bandwidth and computation resource for decoding video. You can call resumeRemoteVideoTile(tileId:) to remove a video from the local block list and resume streaming.

You should call pauseRemoteVideoTile(tileId:) when you want to temporarily pause remote video (e.g. when user is not in the video view and remote videos are not actively being showed), or when you want to stop some remote videos (e.g. in the video pagination example in the following sections, where we want to render videos on the current page, but stop invisible videos on other pages).

Bind/Unbind Video View

To display video, you need to bind a video tile to a video view. You can call bindVideoView(videoView:tileId:) to bind a video tile to a view, and call unbindVideoView(tileId:) to unbind a video tile from a view.

Note that bind/unbind only impacts UI layer, which means these APIs only control when and where the video stream will be displayed in your application UI. If you unbind a video tile, the server will keep sending video frames to the client, and the client will continue consuming resources to receive and decode video. If you want to reduce unnecessary data consumption and CPU usage of your application by stopping remote videos, you should call pauseRemoteVideoTile(tileId:) first.

8h. Rendering selected remote videos

You can render some of the available remote videos instead of rendering them all at the same time.

To selectively render some remote videos, call meetingSession.audioVideo.resumeRemoteVideoTile(tileId:) on the videos you care about, and call meetingSession.audioVideo.pauseRemoteVideoTile(tileId:) to pause other videos.

See the Video Pagination with Active Speaker-Based Policy guide for more information about related APIs and sample code.

9. Receiving metrics (optional)

You can receive events about available audio and video metrics by implementing a MetricsObserver and registering the observer with the audio video facade. Events occur on a one second interval.

To add a MetricsObserver, call meetingSession.audioVideo.addMetricsObserver(observer:).

To remove a MetricsObserver, call meetingSession.audioVideo.removeMetricsObserver(observer:).

A MetricsObserver has the following method:

10. Sending and receiving data messages (optional)

Attendees can broadcast small (2KB max) data messages to other attendees. Data messages can be used to signal attendees of changes to meeting state or develop custom collaborative features. Each message is sent on a particular topic, which allows you to tag messages according to their function to make it easier to handle messages of different types.

To send a message on a given topic, meetingSession.audioVideo.realtimeSendDataMessage(topic:data:lifetimeMs:). When sending a message, the media server stores the messages for the duration specified by lifetimeMs. Up to 1024 messages may be stored for a maximum of 5 minutes. Any attendee joining late or reconnecting will automatically receive the messages in this buffer once they connect. You can use this feature to help paper over gaps in connectivity or give attendees some context into messages that were recently received.

To receive messages on a given topic, implement a DataMessageObserver and subscribe it to the topic using meetingSession.audioVideo.addRealtimeDataMessageObserver(topic:observer:). Through dataMessageDidReceived(dataMessage:) in the observer, you receive a DataMessage containing the payload of the message and other metadata about the message.

To unsubscribe all DataMessageObservers from the topic, call meetingSession.audioVideo.removeRealtimeDataMessageObserverFromTopic(topic:).

If you send too many messages at once, your messages may be returned to you with the throttled flag set. If you continue to exceed the throttle limit, the server may hang up the connection.

Note: You can only send and receive data message after calling meetingSession.audioVideo.start() or meetingSession.audioVideo.start(callKitEnabled:). To avoid missing messages, subscribe the DataMessageObserver to the topic prior to starting audio video.

11. Using Amazon Voice Focus (optional)

Amazon Voice Focus reduces the sound levels of noises that can intrude on a meeting, such as:

  • Environment noises – wind, fans, running water.
  • Background noises – lawnmowers, barking dogs.
  • Foreground noises – typing, papers shuffling.

Note: Amazon Voice Focus doesn't eliminate those types of noises; it reduces their sound levels. To ensure privacy during a meeting, call meetingSession.audioVideo.realtimeLocalMute() to silence yourself.

You must start the audio session before enabling/disabling Amazon Voice Focus or before checking if Amazon Voice Focus is enabled. To enable/disable Amazon Voice Focus, call meetingSession.audioVideo.realtimeSetVoiceFocusEnabled(enabled:). To check if Amazon Voice Focus is enabled, call meetingSession.audioVideo.realtimeIsVoiceFocusEnabled().

When Amazon Voice Focus is running, a CPU usage increase is expected, but the performance impact is small on modern devices (on average, we observed around 5-7% CPU increase). If your app will be running on resource-critical devices, you should take this into consideration before enabling Amazon Voice Focus.

Note that if you want to share music or background sounds with others in the call (e.g., in a fitness or music lesson application), you should disable Amazon Voice Focus. Otherwise, the Amazon Chime SDK will filter these sounds out.

12. Using a custom video source, sink or processing step (optional)

Builders using the Amazon Chime SDK for video can produce, modify, and consume raw video frames transmitted or received during the call. You can allow the facade to manage its own camera capture source, provide your own custom source, or use a provided SDK capture source as the first step in a video processing pipeline which modifies frames before transmission. See the Custom Video Sources, Processors, and Sinks guide for more information.

13. Share screen and other content (optional)

Builders using the Amazon Chime SDK for iOS can share a second video stream such as screen capture in a meeting without disrupting their applications existing audio/video stream. When a content share is started, another attendee with the attendee ID <attendee-id>#content joins the meeting. You can subscribe its presence event to show it in the roster and bind its video tile to a video render view the same as you would for a regular attendee.

Each attendee can share one content share in addition to their main video. Each meeting may have two simultaneous content shares. Content share does not count towards the max video tile limit.

13a. Start and stop the content share

To start the content share, call meetingSession.audioVideo.startContentShare(source:).

To stop the content share, call meetingSession.audioVideo.stopContentShare().

13b. Register a content share observer

You can receive events about the content share by implementing a ContentShareObserver.

To add a ContentShareObserver, call meetingSession.audioVideo.addContentShareObserver(observer:).

To remove a ContentShareObserver, call meetingSession.audioVideo.removeContentShareObserver(observer:).

You can implement the following callbacks:

13c. Collect content share metrics

You will receive content share metrics if you registered a metric observer by 9.Receiving metrics (optional).

Content share metrics will be prefixed by contentShare.

14. Configuring Remote Video Subscriptions

Amazon Chime SDK allows builders to have complete control over the remote videos received by each of their application’s end-users. This can be accomplished using the API AudioVideoFacade.updateVideoSourceSubscriptions. See Configuring Remote Video Subscriptions for more information.