mirror of
https://github.com/livekit/livekit.git
synced 2026-03-30 17:45:40 +00:00
* Send error response when update metadata fails. Keeping it simple for the first implementation. - Send error response only if request_id != 0 - Two kinds of errors notified o does not have permissions - NOT_ALLOWED o attributes exceeds size limits - INVALID_ARGUMENT * Check size limits on metadata and name set from client. Added a name length limit also. * check name length in service update participant path also * limit check in limit config * update protocol * longer keys
1550 lines
44 KiB
Go
1550 lines
44 KiB
Go
// Copyright 2023 LiveKit, Inc.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package rtc
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"math"
|
|
"slices"
|
|
"sort"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"go.uber.org/atomic"
|
|
"golang.org/x/exp/maps"
|
|
"google.golang.org/protobuf/proto"
|
|
|
|
"github.com/pion/sctp"
|
|
|
|
"github.com/livekit/protocol/livekit"
|
|
"github.com/livekit/protocol/logger"
|
|
"github.com/livekit/protocol/utils"
|
|
|
|
"github.com/livekit/livekit-server/pkg/agent"
|
|
"github.com/livekit/livekit-server/pkg/config"
|
|
"github.com/livekit/livekit-server/pkg/routing"
|
|
"github.com/livekit/livekit-server/pkg/rtc/types"
|
|
"github.com/livekit/livekit-server/pkg/sfu/buffer"
|
|
"github.com/livekit/livekit-server/pkg/sfu/connectionquality"
|
|
"github.com/livekit/livekit-server/pkg/telemetry"
|
|
"github.com/livekit/livekit-server/pkg/telemetry/prometheus"
|
|
sutils "github.com/livekit/livekit-server/pkg/utils"
|
|
)
|
|
|
|
const (
|
|
AudioLevelQuantization = 8 // ideally power of 2 to minimize float decimal
|
|
invAudioLevelQuantization = 1.0 / AudioLevelQuantization
|
|
subscriberUpdateInterval = 3 * time.Second
|
|
|
|
dataForwardLoadBalanceThreshold = 20
|
|
|
|
simulateDisconnectSignalTimeout = 5 * time.Second
|
|
)
|
|
|
|
var (
|
|
// var to allow unit test override
|
|
roomUpdateInterval = 5 * time.Second // frequency to update room participant counts
|
|
)
|
|
|
|
type broadcastOptions struct {
|
|
skipSource bool
|
|
immediate bool
|
|
}
|
|
|
|
type participantUpdate struct {
|
|
pi *livekit.ParticipantInfo
|
|
isSynthesizedDisconnect bool
|
|
closeReason types.ParticipantCloseReason
|
|
}
|
|
|
|
type disconnectSignalOnResumeNoMessages struct {
|
|
expiry time.Time
|
|
closedCount int
|
|
}
|
|
|
|
type Room struct {
|
|
// atomics always need to be 64bit/8byte aligned
|
|
// on 32bit arch only the beginning of the struct
|
|
// starts at such a boundary.
|
|
// time the first participant joined the room
|
|
joinedAt atomic.Int64
|
|
// time that the last participant left the room
|
|
leftAt atomic.Int64
|
|
holds atomic.Int32
|
|
|
|
lock sync.RWMutex
|
|
|
|
protoRoom *livekit.Room
|
|
internal *livekit.RoomInternal
|
|
protoProxy *utils.ProtoProxy[*livekit.Room]
|
|
Logger logger.Logger
|
|
|
|
config WebRTCConfig
|
|
audioConfig *config.AudioConfig
|
|
serverInfo *livekit.ServerInfo
|
|
telemetry telemetry.TelemetryService
|
|
egressLauncher EgressLauncher
|
|
trackManager *RoomTrackManager
|
|
|
|
// agents
|
|
agentClient agent.Client
|
|
|
|
// map of identity -> Participant
|
|
participants map[livekit.ParticipantIdentity]types.LocalParticipant
|
|
participantOpts map[livekit.ParticipantIdentity]*ParticipantOptions
|
|
participantRequestSources map[livekit.ParticipantIdentity]routing.MessageSource
|
|
hasPublished map[livekit.ParticipantIdentity]bool
|
|
bufferFactory *buffer.FactoryOfBufferFactory
|
|
|
|
// batch update participant info for non-publishers
|
|
batchedUpdates map[livekit.ParticipantIdentity]*participantUpdate
|
|
batchedUpdatesMu sync.Mutex
|
|
|
|
closed chan struct{}
|
|
|
|
trailer []byte
|
|
|
|
onParticipantChanged func(p types.LocalParticipant)
|
|
onRoomUpdated func()
|
|
onClose func()
|
|
|
|
simulationLock sync.Mutex
|
|
disconnectSignalOnResumeParticipants map[livekit.ParticipantIdentity]time.Time
|
|
disconnectSignalOnResumeNoMessagesParticipants map[livekit.ParticipantIdentity]*disconnectSignalOnResumeNoMessages
|
|
}
|
|
|
|
type ParticipantOptions struct {
|
|
AutoSubscribe bool
|
|
}
|
|
|
|
func NewRoom(
|
|
room *livekit.Room,
|
|
internal *livekit.RoomInternal,
|
|
config WebRTCConfig,
|
|
roomConfig config.RoomConfig,
|
|
audioConfig *config.AudioConfig,
|
|
serverInfo *livekit.ServerInfo,
|
|
telemetry telemetry.TelemetryService,
|
|
agentClient agent.Client,
|
|
egressLauncher EgressLauncher,
|
|
) *Room {
|
|
r := &Room{
|
|
protoRoom: proto.Clone(room).(*livekit.Room),
|
|
internal: internal,
|
|
Logger: LoggerWithRoom(
|
|
logger.GetLogger().WithComponent(sutils.ComponentRoom),
|
|
livekit.RoomName(room.Name),
|
|
livekit.RoomID(room.Sid),
|
|
),
|
|
config: config,
|
|
audioConfig: audioConfig,
|
|
telemetry: telemetry,
|
|
egressLauncher: egressLauncher,
|
|
agentClient: agentClient,
|
|
trackManager: NewRoomTrackManager(),
|
|
serverInfo: serverInfo,
|
|
participants: make(map[livekit.ParticipantIdentity]types.LocalParticipant),
|
|
participantOpts: make(map[livekit.ParticipantIdentity]*ParticipantOptions),
|
|
participantRequestSources: make(map[livekit.ParticipantIdentity]routing.MessageSource),
|
|
hasPublished: make(map[livekit.ParticipantIdentity]bool),
|
|
bufferFactory: buffer.NewFactoryOfBufferFactory(config.Receiver.PacketBufferSizeVideo, config.Receiver.PacketBufferSizeAudio),
|
|
batchedUpdates: make(map[livekit.ParticipantIdentity]*participantUpdate),
|
|
closed: make(chan struct{}),
|
|
trailer: []byte(utils.RandomSecret()),
|
|
disconnectSignalOnResumeParticipants: make(map[livekit.ParticipantIdentity]time.Time),
|
|
disconnectSignalOnResumeNoMessagesParticipants: make(map[livekit.ParticipantIdentity]*disconnectSignalOnResumeNoMessages),
|
|
}
|
|
|
|
if r.protoRoom.EmptyTimeout == 0 {
|
|
r.protoRoom.EmptyTimeout = roomConfig.EmptyTimeout
|
|
}
|
|
if r.protoRoom.DepartureTimeout == 0 {
|
|
r.protoRoom.DepartureTimeout = roomConfig.DepartureTimeout
|
|
}
|
|
if r.protoRoom.CreationTime == 0 {
|
|
r.protoRoom.CreationTime = time.Now().Unix()
|
|
}
|
|
r.protoProxy = utils.NewProtoProxy[*livekit.Room](roomUpdateInterval, r.updateProto)
|
|
|
|
go r.audioUpdateWorker()
|
|
go r.connectionQualityWorker()
|
|
go r.changeUpdateWorker()
|
|
go r.simulationCleanupWorker()
|
|
|
|
return r
|
|
}
|
|
|
|
func (r *Room) ToProto() *livekit.Room {
|
|
return r.protoProxy.Get()
|
|
}
|
|
|
|
func (r *Room) Name() livekit.RoomName {
|
|
return livekit.RoomName(r.protoRoom.Name)
|
|
}
|
|
|
|
func (r *Room) ID() livekit.RoomID {
|
|
return livekit.RoomID(r.protoRoom.Sid)
|
|
}
|
|
|
|
func (r *Room) Trailer() []byte {
|
|
r.lock.RLock()
|
|
defer r.lock.RUnlock()
|
|
|
|
trailer := make([]byte, len(r.trailer))
|
|
copy(trailer, r.trailer)
|
|
return trailer
|
|
}
|
|
|
|
func (r *Room) GetParticipant(identity livekit.ParticipantIdentity) types.LocalParticipant {
|
|
r.lock.RLock()
|
|
defer r.lock.RUnlock()
|
|
return r.participants[identity]
|
|
}
|
|
|
|
func (r *Room) GetParticipantByID(participantID livekit.ParticipantID) types.LocalParticipant {
|
|
r.lock.RLock()
|
|
defer r.lock.RUnlock()
|
|
|
|
for _, p := range r.participants {
|
|
if p.ID() == participantID {
|
|
return p
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (r *Room) GetParticipants() []types.LocalParticipant {
|
|
r.lock.RLock()
|
|
defer r.lock.RUnlock()
|
|
|
|
return maps.Values(r.participants)
|
|
}
|
|
|
|
func (r *Room) GetLocalParticipants() []types.LocalParticipant {
|
|
return r.GetParticipants()
|
|
}
|
|
|
|
func (r *Room) GetParticipantCount() int {
|
|
r.lock.RLock()
|
|
defer r.lock.RUnlock()
|
|
|
|
return len(r.participants)
|
|
}
|
|
|
|
func (r *Room) GetActiveSpeakers() []*livekit.SpeakerInfo {
|
|
participants := r.GetParticipants()
|
|
speakers := make([]*livekit.SpeakerInfo, 0, len(participants))
|
|
for _, p := range participants {
|
|
level, active := p.GetAudioLevel()
|
|
if !active {
|
|
continue
|
|
}
|
|
speakers = append(speakers, &livekit.SpeakerInfo{
|
|
Sid: string(p.ID()),
|
|
Level: float32(level),
|
|
Active: active,
|
|
})
|
|
}
|
|
|
|
sort.Slice(speakers, func(i, j int) bool {
|
|
return speakers[i].Level > speakers[j].Level
|
|
})
|
|
|
|
// quantize to smooth out small changes
|
|
for _, speaker := range speakers {
|
|
speaker.Level = float32(math.Ceil(float64(speaker.Level*AudioLevelQuantization)) * invAudioLevelQuantization)
|
|
}
|
|
|
|
return speakers
|
|
}
|
|
|
|
func (r *Room) GetBufferFactory() *buffer.Factory {
|
|
return r.bufferFactory.CreateBufferFactory()
|
|
}
|
|
|
|
func (r *Room) FirstJoinedAt() int64 {
|
|
return r.joinedAt.Load()
|
|
}
|
|
|
|
func (r *Room) LastLeftAt() int64 {
|
|
return r.leftAt.Load()
|
|
}
|
|
|
|
func (r *Room) Internal() *livekit.RoomInternal {
|
|
return r.internal
|
|
}
|
|
|
|
func (r *Room) Hold() bool {
|
|
r.lock.Lock()
|
|
defer r.lock.Unlock()
|
|
|
|
if r.IsClosed() {
|
|
return false
|
|
}
|
|
|
|
r.holds.Inc()
|
|
return true
|
|
}
|
|
|
|
func (r *Room) Release() {
|
|
r.holds.Dec()
|
|
}
|
|
|
|
func (r *Room) Join(participant types.LocalParticipant, requestSource routing.MessageSource, opts *ParticipantOptions, iceServers []*livekit.ICEServer) error {
|
|
r.lock.Lock()
|
|
defer r.lock.Unlock()
|
|
|
|
if r.IsClosed() {
|
|
return ErrRoomClosed
|
|
}
|
|
|
|
if r.participants[participant.Identity()] != nil {
|
|
return ErrAlreadyJoined
|
|
}
|
|
if r.protoRoom.MaxParticipants > 0 && !participant.IsDependent() {
|
|
numParticipants := uint32(0)
|
|
for _, p := range r.participants {
|
|
if !p.IsDependent() {
|
|
numParticipants++
|
|
}
|
|
}
|
|
if numParticipants >= r.protoRoom.MaxParticipants {
|
|
return ErrMaxParticipantsExceeded
|
|
}
|
|
}
|
|
|
|
if r.FirstJoinedAt() == 0 {
|
|
r.joinedAt.Store(time.Now().Unix())
|
|
}
|
|
|
|
participant.OnStateChange(func(p types.LocalParticipant, state livekit.ParticipantInfo_State) {
|
|
if r.onParticipantChanged != nil {
|
|
r.onParticipantChanged(p)
|
|
}
|
|
r.broadcastParticipantState(p, broadcastOptions{skipSource: true})
|
|
|
|
if state == livekit.ParticipantInfo_ACTIVE {
|
|
// subscribe participant to existing published tracks
|
|
r.subscribeToExistingTracks(p)
|
|
|
|
meta := &livekit.AnalyticsClientMeta{
|
|
ClientConnectTime: uint32(time.Since(p.ConnectedAt()).Milliseconds()),
|
|
}
|
|
cds := p.GetICEConnectionDetails()
|
|
for _, cd := range cds {
|
|
if cd.Type != types.ICEConnectionTypeUnknown {
|
|
meta.ConnectionType = string(cd.Type)
|
|
break
|
|
}
|
|
}
|
|
r.telemetry.ParticipantActive(context.Background(),
|
|
r.ToProto(),
|
|
p.ToProto(),
|
|
meta,
|
|
false,
|
|
)
|
|
|
|
p.GetLogger().Infow("participant active", connectionDetailsFields(cds)...)
|
|
} else if state == livekit.ParticipantInfo_DISCONNECTED {
|
|
// remove participant from room
|
|
// participant should already be closed and have a close reason, so NONE is fine here
|
|
go r.RemoveParticipant(p.Identity(), p.ID(), types.ParticipantCloseReasonNone)
|
|
}
|
|
})
|
|
// it's important to set this before connection, we don't want to miss out on any published tracks
|
|
participant.OnTrackPublished(r.onTrackPublished)
|
|
participant.OnTrackUpdated(r.onTrackUpdated)
|
|
participant.OnTrackUnpublished(r.onTrackUnpublished)
|
|
participant.OnParticipantUpdate(r.onParticipantUpdate)
|
|
participant.OnDataPacket(r.onDataPacket)
|
|
participant.OnSubscribeStatusChanged(func(publisherID livekit.ParticipantID, subscribed bool) {
|
|
if subscribed {
|
|
pub := r.GetParticipantByID(publisherID)
|
|
if pub != nil && pub.State() == livekit.ParticipantInfo_ACTIVE {
|
|
// when a participant subscribes to another participant,
|
|
// send speaker update if the subscribed to participant is active.
|
|
level, active := pub.GetAudioLevel()
|
|
if active {
|
|
_ = participant.SendSpeakerUpdate([]*livekit.SpeakerInfo{
|
|
{
|
|
Sid: string(pub.ID()),
|
|
Level: float32(level),
|
|
Active: active,
|
|
},
|
|
}, false)
|
|
}
|
|
|
|
if cq := pub.GetConnectionQuality(); cq != nil {
|
|
update := &livekit.ConnectionQualityUpdate{}
|
|
update.Updates = append(update.Updates, cq)
|
|
_ = participant.SendConnectionQualityUpdate(update)
|
|
}
|
|
}
|
|
} else {
|
|
// no longer subscribed to the publisher, clear speaker status
|
|
_ = participant.SendSpeakerUpdate([]*livekit.SpeakerInfo{
|
|
{
|
|
Sid: string(publisherID),
|
|
Level: 0,
|
|
Active: false,
|
|
},
|
|
}, true)
|
|
}
|
|
})
|
|
|
|
r.Logger.Debugw("new participant joined",
|
|
"pID", participant.ID(),
|
|
"participant", participant.Identity(),
|
|
"clientInfo", logger.Proto(participant.GetClientInfo()),
|
|
"options", opts,
|
|
"numParticipants", len(r.participants),
|
|
)
|
|
|
|
if participant.IsRecorder() && !r.protoRoom.ActiveRecording {
|
|
r.protoRoom.ActiveRecording = true
|
|
r.protoProxy.MarkDirty(true)
|
|
} else {
|
|
r.protoProxy.MarkDirty(false)
|
|
}
|
|
|
|
r.participants[participant.Identity()] = participant
|
|
r.participantOpts[participant.Identity()] = opts
|
|
r.participantRequestSources[participant.Identity()] = requestSource
|
|
|
|
if r.onParticipantChanged != nil {
|
|
r.onParticipantChanged(participant)
|
|
}
|
|
|
|
time.AfterFunc(time.Minute, func() {
|
|
state := participant.State()
|
|
if state == livekit.ParticipantInfo_JOINING || state == livekit.ParticipantInfo_JOINED {
|
|
r.RemoveParticipant(participant.Identity(), participant.ID(), types.ParticipantCloseReasonJoinTimeout)
|
|
}
|
|
})
|
|
|
|
joinResponse := r.createJoinResponseLocked(participant, iceServers)
|
|
if err := participant.SendJoinResponse(joinResponse); err != nil {
|
|
prometheus.ServiceOperationCounter.WithLabelValues("participant_join", "error", "send_response").Add(1)
|
|
return err
|
|
}
|
|
|
|
participant.SetMigrateState(types.MigrateStateComplete)
|
|
|
|
if participant.SubscriberAsPrimary() {
|
|
// initiates sub connection as primary
|
|
if participant.ProtocolVersion().SupportFastStart() {
|
|
go func() {
|
|
r.subscribeToExistingTracks(participant)
|
|
participant.Negotiate(true)
|
|
}()
|
|
} else {
|
|
participant.Negotiate(true)
|
|
}
|
|
}
|
|
|
|
prometheus.ServiceOperationCounter.WithLabelValues("participant_join", "success", "").Add(1)
|
|
|
|
return nil
|
|
}
|
|
|
|
func (r *Room) ReplaceParticipantRequestSource(identity livekit.ParticipantIdentity, reqSource routing.MessageSource) {
|
|
r.lock.Lock()
|
|
if rs, ok := r.participantRequestSources[identity]; ok {
|
|
rs.Close()
|
|
}
|
|
r.participantRequestSources[identity] = reqSource
|
|
r.lock.Unlock()
|
|
}
|
|
|
|
func (r *Room) GetParticipantRequestSource(identity livekit.ParticipantIdentity) routing.MessageSource {
|
|
r.lock.RLock()
|
|
defer r.lock.RUnlock()
|
|
return r.participantRequestSources[identity]
|
|
}
|
|
|
|
func (r *Room) ResumeParticipant(
|
|
p types.LocalParticipant,
|
|
requestSource routing.MessageSource,
|
|
responseSink routing.MessageSink,
|
|
iceConfig *livekit.ICEConfig,
|
|
iceServers []*livekit.ICEServer,
|
|
reason livekit.ReconnectReason,
|
|
) error {
|
|
r.ReplaceParticipantRequestSource(p.Identity(), requestSource)
|
|
// close previous sink, and link to new one
|
|
p.CloseSignalConnection(types.SignallingCloseReasonResume)
|
|
p.SetResponseSink(responseSink)
|
|
|
|
p.SetSignalSourceValid(true)
|
|
|
|
// check for simulated signal disconnect on resume before sending any signal response messages
|
|
r.simulationLock.Lock()
|
|
if state, ok := r.disconnectSignalOnResumeNoMessagesParticipants[p.Identity()]; ok {
|
|
// WARNING: this uses knowledge that service layer tries internally
|
|
simulated := false
|
|
if time.Now().Before(state.expiry) {
|
|
state.closedCount++
|
|
p.CloseSignalConnection(types.SignallingCloseReasonDisconnectOnResumeNoMessages)
|
|
simulated = true
|
|
}
|
|
if state.closedCount == 3 {
|
|
delete(r.disconnectSignalOnResumeNoMessagesParticipants, p.Identity())
|
|
}
|
|
if simulated {
|
|
r.simulationLock.Unlock()
|
|
return nil
|
|
}
|
|
}
|
|
r.simulationLock.Unlock()
|
|
|
|
if err := p.HandleReconnectAndSendResponse(reason, &livekit.ReconnectResponse{
|
|
IceServers: iceServers,
|
|
ClientConfiguration: p.GetClientConfiguration(),
|
|
}); err != nil {
|
|
return err
|
|
}
|
|
|
|
// include the local participant's info as well, since metadata could have been changed
|
|
updates := r.getOtherParticipantInfo("")
|
|
if err := p.SendParticipantUpdate(updates); err != nil {
|
|
return err
|
|
}
|
|
|
|
_ = p.SendRoomUpdate(r.ToProto())
|
|
p.ICERestart(iceConfig)
|
|
|
|
// check for simulated signal disconnect on resume
|
|
r.simulationLock.Lock()
|
|
if timeout, ok := r.disconnectSignalOnResumeParticipants[p.Identity()]; ok {
|
|
if time.Now().Before(timeout) {
|
|
p.CloseSignalConnection(types.SignallingCloseReasonDisconnectOnResume)
|
|
}
|
|
delete(r.disconnectSignalOnResumeParticipants, p.Identity())
|
|
}
|
|
r.simulationLock.Unlock()
|
|
|
|
return nil
|
|
}
|
|
|
|
func (r *Room) RemoveParticipant(identity livekit.ParticipantIdentity, pID livekit.ParticipantID, reason types.ParticipantCloseReason) {
|
|
r.lock.Lock()
|
|
p, ok := r.participants[identity]
|
|
if !ok {
|
|
r.lock.Unlock()
|
|
return
|
|
}
|
|
|
|
if pID != "" && p.ID() != pID {
|
|
// participant session has been replaced
|
|
r.lock.Unlock()
|
|
return
|
|
}
|
|
|
|
delete(r.participants, identity)
|
|
delete(r.participantOpts, identity)
|
|
delete(r.participantRequestSources, identity)
|
|
delete(r.hasPublished, identity)
|
|
if !p.Hidden() {
|
|
r.protoRoom.NumParticipants--
|
|
}
|
|
|
|
immediateChange := false
|
|
if p.IsRecorder() {
|
|
activeRecording := false
|
|
for _, op := range r.participants {
|
|
if op.IsRecorder() {
|
|
activeRecording = true
|
|
break
|
|
}
|
|
}
|
|
|
|
if r.protoRoom.ActiveRecording != activeRecording {
|
|
r.protoRoom.ActiveRecording = activeRecording
|
|
immediateChange = true
|
|
}
|
|
}
|
|
r.lock.Unlock()
|
|
r.protoProxy.MarkDirty(immediateChange)
|
|
|
|
if !p.HasConnected() {
|
|
fields := append(connectionDetailsFields(p.GetICEConnectionDetails()),
|
|
"reason", reason.String(),
|
|
)
|
|
p.GetLogger().Infow("removing participant without connection", fields...)
|
|
}
|
|
|
|
// send broadcast only if it's not already closed
|
|
sendUpdates := !p.IsDisconnected()
|
|
|
|
// remove all published tracks
|
|
for _, t := range p.GetPublishedTracks() {
|
|
r.trackManager.RemoveTrack(t)
|
|
}
|
|
|
|
p.OnTrackUpdated(nil)
|
|
p.OnTrackPublished(nil)
|
|
p.OnTrackUnpublished(nil)
|
|
p.OnStateChange(nil)
|
|
p.OnParticipantUpdate(nil)
|
|
p.OnDataPacket(nil)
|
|
p.OnSubscribeStatusChanged(nil)
|
|
|
|
// close participant as well
|
|
_ = p.Close(true, reason, false)
|
|
|
|
r.leftAt.Store(time.Now().Unix())
|
|
|
|
if sendUpdates {
|
|
if r.onParticipantChanged != nil {
|
|
r.onParticipantChanged(p)
|
|
}
|
|
r.broadcastParticipantState(p, broadcastOptions{skipSource: true})
|
|
}
|
|
}
|
|
|
|
func (r *Room) UpdateSubscriptions(
|
|
participant types.LocalParticipant,
|
|
trackIDs []livekit.TrackID,
|
|
participantTracks []*livekit.ParticipantTracks,
|
|
subscribe bool,
|
|
) {
|
|
// handle subscription changes
|
|
for _, trackID := range trackIDs {
|
|
if subscribe {
|
|
participant.SubscribeToTrack(trackID)
|
|
} else {
|
|
participant.UnsubscribeFromTrack(trackID)
|
|
}
|
|
}
|
|
|
|
for _, pt := range participantTracks {
|
|
for _, trackID := range livekit.StringsAsIDs[livekit.TrackID](pt.TrackSids) {
|
|
if subscribe {
|
|
participant.SubscribeToTrack(trackID)
|
|
} else {
|
|
participant.UnsubscribeFromTrack(trackID)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
func (r *Room) SyncState(participant types.LocalParticipant, state *livekit.SyncState) error {
|
|
pLogger := participant.GetLogger()
|
|
pLogger.Infow("setting sync state", "state", logger.Proto(state))
|
|
|
|
shouldReconnect := false
|
|
pubTracks := state.GetPublishTracks()
|
|
existingPubTracks := participant.GetPublishedTracks()
|
|
for _, pubTrack := range pubTracks {
|
|
// client may not have sent TrackInfo for each published track
|
|
ti := pubTrack.Track
|
|
if ti == nil {
|
|
pLogger.Warnw("TrackInfo not sent during resume", nil)
|
|
shouldReconnect = true
|
|
break
|
|
}
|
|
|
|
found := false
|
|
for _, existingPubTrack := range existingPubTracks {
|
|
if existingPubTrack.ID() == livekit.TrackID(ti.Sid) {
|
|
found = true
|
|
break
|
|
}
|
|
}
|
|
if !found {
|
|
// is there a pending track?
|
|
found = participant.GetPendingTrack(livekit.TrackID(ti.Sid)) != nil
|
|
}
|
|
if !found {
|
|
pLogger.Warnw("unknown track during resume", nil, "trackID", ti.Sid)
|
|
shouldReconnect = true
|
|
break
|
|
}
|
|
}
|
|
if shouldReconnect {
|
|
pLogger.Warnw("unable to resume due to missing published tracks, starting full reconnect", nil)
|
|
participant.IssueFullReconnect(types.ParticipantCloseReasonPublicationError)
|
|
return nil
|
|
}
|
|
|
|
// synthesize a track setting for each disabled track,
|
|
// can be set before addding subscriptions,
|
|
// in fact it is done before so that setting can be updated immediately upon subscription.
|
|
for _, trackSid := range state.TrackSidsDisabled {
|
|
participant.UpdateSubscribedTrackSettings(livekit.TrackID(trackSid), &livekit.UpdateTrackSettings{Disabled: true})
|
|
}
|
|
|
|
r.UpdateSubscriptions(
|
|
participant,
|
|
livekit.StringsAsIDs[livekit.TrackID](state.Subscription.TrackSids),
|
|
state.Subscription.ParticipantTracks,
|
|
state.Subscription.Subscribe,
|
|
)
|
|
return nil
|
|
}
|
|
|
|
func (r *Room) UpdateSubscriptionPermission(participant types.LocalParticipant, subscriptionPermission *livekit.SubscriptionPermission) error {
|
|
if err := participant.UpdateSubscriptionPermission(subscriptionPermission, utils.TimedVersion(0), r.GetParticipantByID); err != nil {
|
|
return err
|
|
}
|
|
for _, track := range participant.GetPublishedTracks() {
|
|
r.trackManager.NotifyTrackChanged(track.ID())
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (r *Room) ResolveMediaTrackForSubscriber(subIdentity livekit.ParticipantIdentity, trackID livekit.TrackID) types.MediaResolverResult {
|
|
res := types.MediaResolverResult{}
|
|
|
|
info := r.trackManager.GetTrackInfo(trackID)
|
|
res.TrackChangedNotifier = r.trackManager.GetOrCreateTrackChangeNotifier(trackID)
|
|
|
|
if info == nil {
|
|
return res
|
|
}
|
|
|
|
res.Track = info.Track
|
|
res.TrackRemovedNotifier = r.trackManager.GetOrCreateTrackRemoveNotifier(trackID)
|
|
res.PublisherIdentity = info.PublisherIdentity
|
|
res.PublisherID = info.PublisherID
|
|
|
|
pub := r.GetParticipantByID(info.PublisherID)
|
|
// when publisher is not found, we will assume it doesn't have permission to access
|
|
if pub != nil {
|
|
res.HasPermission = pub.HasPermission(trackID, subIdentity)
|
|
}
|
|
|
|
return res
|
|
}
|
|
|
|
func (r *Room) IsClosed() bool {
|
|
select {
|
|
case <-r.closed:
|
|
return true
|
|
default:
|
|
return false
|
|
}
|
|
}
|
|
|
|
// CloseIfEmpty closes the room if all participants had left, or it's still empty past timeout
|
|
func (r *Room) CloseIfEmpty() {
|
|
r.lock.Lock()
|
|
|
|
if r.IsClosed() || r.holds.Load() > 0 {
|
|
r.lock.Unlock()
|
|
return
|
|
}
|
|
|
|
for _, p := range r.participants {
|
|
if !p.IsDependent() {
|
|
r.lock.Unlock()
|
|
return
|
|
}
|
|
}
|
|
|
|
var timeout uint32
|
|
var elapsed int64
|
|
if r.FirstJoinedAt() > 0 && r.LastLeftAt() > 0 {
|
|
elapsed = time.Now().Unix() - r.LastLeftAt()
|
|
// need to give time in case participant is reconnecting
|
|
timeout = r.protoRoom.DepartureTimeout
|
|
} else {
|
|
elapsed = time.Now().Unix() - r.protoRoom.CreationTime
|
|
timeout = r.protoRoom.EmptyTimeout
|
|
}
|
|
r.lock.Unlock()
|
|
|
|
if elapsed >= int64(timeout) {
|
|
r.Close(types.ParticipantCloseReasonNone)
|
|
}
|
|
}
|
|
|
|
func (r *Room) Close(reason types.ParticipantCloseReason) {
|
|
r.lock.Lock()
|
|
select {
|
|
case <-r.closed:
|
|
r.lock.Unlock()
|
|
return
|
|
default:
|
|
// fall through
|
|
}
|
|
close(r.closed)
|
|
r.lock.Unlock()
|
|
|
|
r.Logger.Infow("closing room")
|
|
for _, p := range r.GetParticipants() {
|
|
_ = p.Close(true, reason, false)
|
|
}
|
|
|
|
r.protoProxy.Stop()
|
|
|
|
if r.onClose != nil {
|
|
r.onClose()
|
|
}
|
|
}
|
|
|
|
func (r *Room) OnClose(f func()) {
|
|
r.onClose = f
|
|
}
|
|
|
|
func (r *Room) OnParticipantChanged(f func(participant types.LocalParticipant)) {
|
|
r.onParticipantChanged = f
|
|
}
|
|
|
|
func (r *Room) SendDataPacket(dp *livekit.DataPacket, kind livekit.DataPacket_Kind) {
|
|
r.onDataPacket(nil, kind, dp)
|
|
}
|
|
|
|
func (r *Room) SetMetadata(metadata string) <-chan struct{} {
|
|
r.lock.Lock()
|
|
r.protoRoom.Metadata = metadata
|
|
r.lock.Unlock()
|
|
return r.protoProxy.MarkDirty(true)
|
|
}
|
|
|
|
func (r *Room) sendRoomUpdate() {
|
|
roomInfo := r.ToProto()
|
|
// Send update to participants
|
|
for _, p := range r.GetParticipants() {
|
|
if !p.IsReady() {
|
|
continue
|
|
}
|
|
|
|
err := p.SendRoomUpdate(roomInfo)
|
|
if err != nil {
|
|
r.Logger.Warnw("failed to send room update", err, "participant", p.Identity())
|
|
}
|
|
}
|
|
}
|
|
|
|
func (r *Room) OnRoomUpdated(f func()) {
|
|
r.onRoomUpdated = f
|
|
}
|
|
|
|
func (r *Room) SimulateScenario(participant types.LocalParticipant, simulateScenario *livekit.SimulateScenario) error {
|
|
switch scenario := simulateScenario.Scenario.(type) {
|
|
case *livekit.SimulateScenario_SpeakerUpdate:
|
|
r.Logger.Infow("simulating speaker update", "participant", participant.Identity(), "duration", scenario.SpeakerUpdate)
|
|
go func() {
|
|
<-time.After(time.Duration(scenario.SpeakerUpdate) * time.Second)
|
|
r.sendSpeakerChanges([]*livekit.SpeakerInfo{{
|
|
Sid: string(participant.ID()),
|
|
Active: false,
|
|
Level: 0,
|
|
}})
|
|
}()
|
|
r.sendSpeakerChanges([]*livekit.SpeakerInfo{{
|
|
Sid: string(participant.ID()),
|
|
Active: true,
|
|
Level: 0.9,
|
|
}})
|
|
case *livekit.SimulateScenario_Migration:
|
|
r.Logger.Infow("simulating migration", "participant", participant.Identity())
|
|
// drop participant without necessarily cleaning up
|
|
if err := participant.Close(false, types.ParticipantCloseReasonSimulateMigration, true); err != nil {
|
|
return err
|
|
}
|
|
case *livekit.SimulateScenario_NodeFailure:
|
|
r.Logger.Infow("simulating node failure", "participant", participant.Identity())
|
|
// drop participant without necessarily cleaning up
|
|
if err := participant.Close(false, types.ParticipantCloseReasonSimulateNodeFailure, true); err != nil {
|
|
return err
|
|
}
|
|
case *livekit.SimulateScenario_ServerLeave:
|
|
r.Logger.Infow("simulating server leave", "participant", participant.Identity())
|
|
if err := participant.Close(true, types.ParticipantCloseReasonSimulateServerLeave, false); err != nil {
|
|
return err
|
|
}
|
|
case *livekit.SimulateScenario_SwitchCandidateProtocol:
|
|
r.Logger.Infow("simulating switch candidate protocol", "participant", participant.Identity())
|
|
participant.ICERestart(&livekit.ICEConfig{
|
|
PreferenceSubscriber: livekit.ICECandidateType(scenario.SwitchCandidateProtocol),
|
|
PreferencePublisher: livekit.ICECandidateType(scenario.SwitchCandidateProtocol),
|
|
})
|
|
case *livekit.SimulateScenario_SubscriberBandwidth:
|
|
if scenario.SubscriberBandwidth > 0 {
|
|
r.Logger.Infow("simulating subscriber bandwidth start", "participant", participant.Identity(), "bandwidth", scenario.SubscriberBandwidth)
|
|
} else {
|
|
r.Logger.Infow("simulating subscriber bandwidth end", "participant", participant.Identity())
|
|
}
|
|
participant.SetSubscriberChannelCapacity(scenario.SubscriberBandwidth)
|
|
case *livekit.SimulateScenario_DisconnectSignalOnResume:
|
|
participant.GetLogger().Infow("simulating disconnect signal on resume")
|
|
r.simulationLock.Lock()
|
|
r.disconnectSignalOnResumeParticipants[participant.Identity()] = time.Now().Add(simulateDisconnectSignalTimeout)
|
|
r.simulationLock.Unlock()
|
|
case *livekit.SimulateScenario_DisconnectSignalOnResumeNoMessages:
|
|
participant.GetLogger().Infow("simulating disconnect signal on resume before sending any response messages")
|
|
r.simulationLock.Lock()
|
|
r.disconnectSignalOnResumeNoMessagesParticipants[participant.Identity()] = &disconnectSignalOnResumeNoMessages{
|
|
expiry: time.Now().Add(simulateDisconnectSignalTimeout),
|
|
}
|
|
r.simulationLock.Unlock()
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (r *Room) getOtherParticipantInfo(identity livekit.ParticipantIdentity) []*livekit.ParticipantInfo {
|
|
participants := r.GetParticipants()
|
|
pi := make([]*livekit.ParticipantInfo, 0, len(participants))
|
|
for _, p := range participants {
|
|
if !p.Hidden() && p.Identity() != identity {
|
|
pi = append(pi, p.ToProto())
|
|
}
|
|
}
|
|
|
|
return pi
|
|
}
|
|
|
|
// checks if participant should be autosubscribed to new tracks, assumes lock is already acquired
|
|
func (r *Room) autoSubscribe(participant types.LocalParticipant) bool {
|
|
opts := r.participantOpts[participant.Identity()]
|
|
// default to true if no options are set
|
|
if opts != nil && !opts.AutoSubscribe {
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
|
|
func (r *Room) createJoinResponseLocked(participant types.LocalParticipant, iceServers []*livekit.ICEServer) *livekit.JoinResponse {
|
|
// gather other participants and send join response
|
|
otherParticipants := make([]*livekit.ParticipantInfo, 0, len(r.participants))
|
|
for _, p := range r.participants {
|
|
if p.ID() != participant.ID() && !p.Hidden() {
|
|
otherParticipants = append(otherParticipants, p.ToProto())
|
|
}
|
|
}
|
|
|
|
return &livekit.JoinResponse{
|
|
Room: r.ToProto(),
|
|
Participant: participant.ToProto(),
|
|
OtherParticipants: otherParticipants,
|
|
IceServers: iceServers,
|
|
// indicates both server and client support subscriber as primary
|
|
SubscriberPrimary: participant.SubscriberAsPrimary(),
|
|
ClientConfiguration: participant.GetClientConfiguration(),
|
|
// sane defaults for ping interval & timeout
|
|
PingInterval: PingIntervalSeconds,
|
|
PingTimeout: PingTimeoutSeconds,
|
|
ServerInfo: r.serverInfo,
|
|
ServerVersion: r.serverInfo.Version,
|
|
ServerRegion: r.serverInfo.Region,
|
|
SifTrailer: r.trailer,
|
|
}
|
|
}
|
|
|
|
// a ParticipantImpl in the room added a new track, subscribe other participants to it
|
|
func (r *Room) onTrackPublished(participant types.LocalParticipant, track types.MediaTrack) {
|
|
// publish participant update, since track state is changed
|
|
r.broadcastParticipantState(participant, broadcastOptions{skipSource: true})
|
|
|
|
r.lock.RLock()
|
|
// subscribe all existing participants to this MediaTrack
|
|
for _, existingParticipant := range r.participants {
|
|
if existingParticipant == participant {
|
|
// skip publishing participant
|
|
continue
|
|
}
|
|
if existingParticipant.State() != livekit.ParticipantInfo_ACTIVE {
|
|
// not fully joined. don't subscribe yet
|
|
continue
|
|
}
|
|
if !r.autoSubscribe(existingParticipant) {
|
|
continue
|
|
}
|
|
|
|
r.Logger.Debugw("subscribing to new track",
|
|
"participant", existingParticipant.Identity(),
|
|
"pID", existingParticipant.ID(),
|
|
"publisher", participant.Identity(),
|
|
"publisherID", participant.ID(),
|
|
"trackID", track.ID())
|
|
existingParticipant.SubscribeToTrack(track.ID())
|
|
}
|
|
onParticipantChanged := r.onParticipantChanged
|
|
r.lock.RUnlock()
|
|
|
|
if onParticipantChanged != nil {
|
|
onParticipantChanged(participant)
|
|
}
|
|
|
|
r.trackManager.AddTrack(track, participant.Identity(), participant.ID())
|
|
|
|
// launch jobs
|
|
r.lock.Lock()
|
|
hasPublished := r.hasPublished[participant.Identity()]
|
|
r.hasPublished[participant.Identity()] = true
|
|
r.lock.Unlock()
|
|
|
|
if !hasPublished {
|
|
r.launchPublisherAgents(participant)
|
|
if r.internal != nil && r.internal.ParticipantEgress != nil {
|
|
go func() {
|
|
if err := StartParticipantEgress(
|
|
context.Background(),
|
|
r.egressLauncher,
|
|
r.telemetry,
|
|
r.internal.ParticipantEgress,
|
|
participant.Identity(),
|
|
r.Name(),
|
|
r.ID(),
|
|
); err != nil {
|
|
r.Logger.Errorw("failed to launch participant egress", err)
|
|
}
|
|
}()
|
|
}
|
|
}
|
|
if r.internal != nil && r.internal.TrackEgress != nil {
|
|
go func() {
|
|
if err := StartTrackEgress(
|
|
context.Background(),
|
|
r.egressLauncher,
|
|
r.telemetry,
|
|
r.internal.TrackEgress,
|
|
track,
|
|
r.Name(),
|
|
r.ID(),
|
|
); err != nil {
|
|
r.Logger.Errorw("failed to launch track egress", err)
|
|
}
|
|
}()
|
|
}
|
|
}
|
|
|
|
func (r *Room) onTrackUpdated(p types.LocalParticipant, _ types.MediaTrack) {
|
|
// send track updates to everyone, especially if track was updated by admin
|
|
r.broadcastParticipantState(p, broadcastOptions{})
|
|
if r.onParticipantChanged != nil {
|
|
r.onParticipantChanged(p)
|
|
}
|
|
}
|
|
|
|
func (r *Room) onTrackUnpublished(p types.LocalParticipant, track types.MediaTrack) {
|
|
r.trackManager.RemoveTrack(track)
|
|
if !p.IsClosed() {
|
|
r.broadcastParticipantState(p, broadcastOptions{skipSource: true})
|
|
}
|
|
if r.onParticipantChanged != nil {
|
|
r.onParticipantChanged(p)
|
|
}
|
|
}
|
|
|
|
func (r *Room) onParticipantUpdate(p types.LocalParticipant) {
|
|
r.protoProxy.MarkDirty(false)
|
|
// immediately notify when permissions or metadata changed
|
|
r.broadcastParticipantState(p, broadcastOptions{immediate: true})
|
|
if r.onParticipantChanged != nil {
|
|
r.onParticipantChanged(p)
|
|
}
|
|
}
|
|
|
|
func (r *Room) onDataPacket(source types.LocalParticipant, kind livekit.DataPacket_Kind, dp *livekit.DataPacket) {
|
|
BroadcastDataPacketForRoom(r, source, kind, dp, r.Logger)
|
|
}
|
|
|
|
func (r *Room) subscribeToExistingTracks(p types.LocalParticipant) {
|
|
r.lock.RLock()
|
|
shouldSubscribe := r.autoSubscribe(p)
|
|
r.lock.RUnlock()
|
|
if !shouldSubscribe {
|
|
return
|
|
}
|
|
|
|
var trackIDs []livekit.TrackID
|
|
for _, op := range r.GetParticipants() {
|
|
if p.ID() == op.ID() {
|
|
// don't send to itself
|
|
continue
|
|
}
|
|
|
|
// subscribe to all
|
|
for _, track := range op.GetPublishedTracks() {
|
|
trackIDs = append(trackIDs, track.ID())
|
|
p.SubscribeToTrack(track.ID())
|
|
}
|
|
}
|
|
if len(trackIDs) > 0 {
|
|
r.Logger.Debugw("subscribed participant to existing tracks", "trackID", trackIDs)
|
|
}
|
|
}
|
|
|
|
// broadcast an update about participant p
|
|
func (r *Room) broadcastParticipantState(p types.LocalParticipant, opts broadcastOptions) {
|
|
pi := p.ToProto()
|
|
|
|
if p.Hidden() {
|
|
if !opts.skipSource {
|
|
// send update only to hidden participant
|
|
err := p.SendParticipantUpdate([]*livekit.ParticipantInfo{pi})
|
|
if err != nil {
|
|
p.GetLogger().Errorw("could not send update to participant", err)
|
|
}
|
|
}
|
|
return
|
|
}
|
|
|
|
updates := r.pushAndDequeueUpdates(pi, p.CloseReason(), opts.immediate)
|
|
r.sendParticipantUpdates(updates)
|
|
}
|
|
|
|
func (r *Room) sendParticipantUpdates(updates []*participantUpdate) {
|
|
if len(updates) == 0 {
|
|
return
|
|
}
|
|
|
|
// For filtered updates, skip
|
|
// 1. synthesized DISCONNECT - this happens on SID change
|
|
// 2. close reasons of DUPLICATE_IDENTITY/STALE - A newer session for that identity exists.
|
|
//
|
|
// Filtered updates are used with clients that can handle identity based reconnect and hence those
|
|
// conditions can be skipped.
|
|
var filteredUpdates []*livekit.ParticipantInfo
|
|
for _, update := range updates {
|
|
if update.isSynthesizedDisconnect || IsCloseNotifySkippable(update.closeReason) {
|
|
continue
|
|
}
|
|
filteredUpdates = append(filteredUpdates, update.pi)
|
|
}
|
|
|
|
var fullUpdates []*livekit.ParticipantInfo
|
|
for _, update := range updates {
|
|
fullUpdates = append(fullUpdates, update.pi)
|
|
}
|
|
|
|
for _, op := range r.GetParticipants() {
|
|
var err error
|
|
if op.ProtocolVersion().SupportsIdentityBasedReconnection() {
|
|
err = op.SendParticipantUpdate(filteredUpdates)
|
|
} else {
|
|
err = op.SendParticipantUpdate(fullUpdates)
|
|
}
|
|
if err != nil {
|
|
op.GetLogger().Errorw("could not send update to participant", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// for protocol 3, send only changed updates
|
|
func (r *Room) sendSpeakerChanges(speakers []*livekit.SpeakerInfo) {
|
|
for _, p := range r.GetParticipants() {
|
|
if p.ProtocolVersion().SupportsSpeakerChanged() {
|
|
_ = p.SendSpeakerUpdate(speakers, false)
|
|
}
|
|
}
|
|
}
|
|
|
|
// push a participant update for batched broadcast, optionally returning immediate updates to broadcast.
|
|
// it handles the following scenarios
|
|
// * subscriber-only updates will be queued for batch updates
|
|
// * publisher & immediate updates will be returned without queuing
|
|
// * when the SID changes, it will return both updates, with the earlier participant set to disconnected
|
|
func (r *Room) pushAndDequeueUpdates(
|
|
pi *livekit.ParticipantInfo,
|
|
closeReason types.ParticipantCloseReason,
|
|
isImmediate bool,
|
|
) []*participantUpdate {
|
|
r.batchedUpdatesMu.Lock()
|
|
defer r.batchedUpdatesMu.Unlock()
|
|
|
|
var updates []*participantUpdate
|
|
identity := livekit.ParticipantIdentity(pi.Identity)
|
|
existing := r.batchedUpdates[identity]
|
|
shouldSend := isImmediate || pi.IsPublisher
|
|
|
|
if existing != nil {
|
|
if pi.Sid == existing.pi.Sid {
|
|
// same participant session
|
|
if pi.Version < existing.pi.Version {
|
|
// out of order update
|
|
return nil
|
|
}
|
|
} else {
|
|
// different participant sessions
|
|
if existing.pi.JoinedAt < pi.JoinedAt {
|
|
// existing is older, synthesize a DISCONNECT for older and
|
|
// send immediately along with newer session to signal switch
|
|
shouldSend = true
|
|
existing.pi.State = livekit.ParticipantInfo_DISCONNECTED
|
|
existing.isSynthesizedDisconnect = true
|
|
updates = append(updates, existing)
|
|
} else {
|
|
// older session update, newer session has already become active, so nothing to do
|
|
return nil
|
|
}
|
|
}
|
|
} else {
|
|
ep := r.GetParticipant(identity)
|
|
if ep != nil {
|
|
epi := ep.ToProto()
|
|
if epi.JoinedAt > pi.JoinedAt {
|
|
// older session update, newer session has already become active, so nothing to do
|
|
return nil
|
|
}
|
|
}
|
|
}
|
|
|
|
if shouldSend {
|
|
// include any queued update, and return
|
|
delete(r.batchedUpdates, identity)
|
|
updates = append(updates, &participantUpdate{pi: pi, closeReason: closeReason})
|
|
} else {
|
|
// enqueue for batch
|
|
r.batchedUpdates[identity] = &participantUpdate{pi: pi, closeReason: closeReason}
|
|
}
|
|
|
|
return updates
|
|
}
|
|
|
|
func (r *Room) updateProto() *livekit.Room {
|
|
r.lock.RLock()
|
|
room := proto.Clone(r.protoRoom).(*livekit.Room)
|
|
r.lock.RUnlock()
|
|
|
|
room.NumPublishers = 0
|
|
room.NumParticipants = 0
|
|
for _, p := range r.GetParticipants() {
|
|
if !p.IsDependent() {
|
|
room.NumParticipants++
|
|
}
|
|
if p.IsPublisher() {
|
|
room.NumPublishers++
|
|
}
|
|
}
|
|
|
|
return room
|
|
}
|
|
|
|
func (r *Room) changeUpdateWorker() {
|
|
subTicker := time.NewTicker(subscriberUpdateInterval)
|
|
defer subTicker.Stop()
|
|
|
|
for !r.IsClosed() {
|
|
select {
|
|
case <-r.closed:
|
|
return
|
|
case <-r.protoProxy.Updated():
|
|
if r.onRoomUpdated != nil {
|
|
r.onRoomUpdated()
|
|
}
|
|
r.sendRoomUpdate()
|
|
case <-subTicker.C:
|
|
r.batchedUpdatesMu.Lock()
|
|
updatesMap := r.batchedUpdates
|
|
r.batchedUpdates = make(map[livekit.ParticipantIdentity]*participantUpdate)
|
|
r.batchedUpdatesMu.Unlock()
|
|
|
|
if len(updatesMap) == 0 {
|
|
continue
|
|
}
|
|
|
|
r.sendParticipantUpdates(maps.Values(updatesMap))
|
|
}
|
|
}
|
|
}
|
|
|
|
func (r *Room) audioUpdateWorker() {
|
|
lastActiveMap := make(map[livekit.ParticipantID]*livekit.SpeakerInfo)
|
|
for {
|
|
if r.IsClosed() {
|
|
return
|
|
}
|
|
|
|
activeSpeakers := r.GetActiveSpeakers()
|
|
changedSpeakers := make([]*livekit.SpeakerInfo, 0, len(activeSpeakers))
|
|
nextActiveMap := make(map[livekit.ParticipantID]*livekit.SpeakerInfo, len(activeSpeakers))
|
|
for _, speaker := range activeSpeakers {
|
|
prev := lastActiveMap[livekit.ParticipantID(speaker.Sid)]
|
|
if prev == nil || prev.Level != speaker.Level {
|
|
changedSpeakers = append(changedSpeakers, speaker)
|
|
}
|
|
nextActiveMap[livekit.ParticipantID(speaker.Sid)] = speaker
|
|
}
|
|
|
|
// changedSpeakers need to include previous speakers that are no longer speaking
|
|
for sid, speaker := range lastActiveMap {
|
|
if nextActiveMap[sid] == nil {
|
|
inactiveSpeaker := proto.Clone(speaker).(*livekit.SpeakerInfo)
|
|
inactiveSpeaker.Level = 0
|
|
inactiveSpeaker.Active = false
|
|
changedSpeakers = append(changedSpeakers, inactiveSpeaker)
|
|
}
|
|
}
|
|
|
|
// see if an update is needed
|
|
if len(changedSpeakers) > 0 {
|
|
r.sendSpeakerChanges(changedSpeakers)
|
|
}
|
|
|
|
lastActiveMap = nextActiveMap
|
|
|
|
time.Sleep(time.Duration(r.audioConfig.UpdateInterval) * time.Millisecond)
|
|
}
|
|
}
|
|
|
|
func (r *Room) connectionQualityWorker() {
|
|
ticker := time.NewTicker(connectionquality.UpdateInterval)
|
|
defer ticker.Stop()
|
|
|
|
prevConnectionInfos := make(map[livekit.ParticipantID]*livekit.ConnectionQualityInfo)
|
|
// send updates to only users that are subscribed to each other
|
|
for !r.IsClosed() {
|
|
<-ticker.C
|
|
|
|
participants := r.GetParticipants()
|
|
nowConnectionInfos := make(map[livekit.ParticipantID]*livekit.ConnectionQualityInfo, len(participants))
|
|
|
|
for _, p := range participants {
|
|
if p.State() != livekit.ParticipantInfo_ACTIVE {
|
|
continue
|
|
}
|
|
|
|
if q := p.GetConnectionQuality(); q != nil {
|
|
nowConnectionInfos[p.ID()] = q
|
|
}
|
|
}
|
|
|
|
// send an update if there is a change
|
|
// - new participant
|
|
// - quality change
|
|
// NOTE: participant leaving is explicitly omitted as `leave` signal notifies that a participant is not in the room anymore
|
|
sendUpdate := false
|
|
for _, p := range participants {
|
|
pID := p.ID()
|
|
prevInfo, prevOk := prevConnectionInfos[pID]
|
|
nowInfo, nowOk := nowConnectionInfos[pID]
|
|
if !nowOk {
|
|
// participant is not ACTIVE any more
|
|
continue
|
|
}
|
|
if !prevOk || nowInfo.Quality != prevInfo.Quality {
|
|
// new entrant OR change in quality
|
|
sendUpdate = true
|
|
break
|
|
}
|
|
}
|
|
|
|
if !sendUpdate {
|
|
prevConnectionInfos = nowConnectionInfos
|
|
continue
|
|
}
|
|
|
|
maybeAddToUpdate := func(pID livekit.ParticipantID, update *livekit.ConnectionQualityUpdate) {
|
|
if nowInfo, nowOk := nowConnectionInfos[pID]; nowOk {
|
|
update.Updates = append(update.Updates, nowInfo)
|
|
}
|
|
}
|
|
|
|
for _, op := range participants {
|
|
if !op.ProtocolVersion().SupportsConnectionQuality() || op.State() != livekit.ParticipantInfo_ACTIVE {
|
|
continue
|
|
}
|
|
update := &livekit.ConnectionQualityUpdate{}
|
|
|
|
// send to user itself
|
|
maybeAddToUpdate(op.ID(), update)
|
|
|
|
// add connection quality of other participants its subscribed to
|
|
for _, sid := range op.GetSubscribedParticipants() {
|
|
maybeAddToUpdate(sid, update)
|
|
}
|
|
if len(update.Updates) == 0 {
|
|
// no change
|
|
continue
|
|
}
|
|
if err := op.SendConnectionQualityUpdate(update); err != nil {
|
|
r.Logger.Warnw("could not send connection quality update", err,
|
|
"participant", op.Identity())
|
|
}
|
|
}
|
|
|
|
prevConnectionInfos = nowConnectionInfos
|
|
}
|
|
}
|
|
|
|
func (r *Room) simulationCleanupWorker() {
|
|
for {
|
|
if r.IsClosed() {
|
|
return
|
|
}
|
|
|
|
now := time.Now()
|
|
r.simulationLock.Lock()
|
|
for identity, timeout := range r.disconnectSignalOnResumeParticipants {
|
|
if now.After(timeout) {
|
|
delete(r.disconnectSignalOnResumeParticipants, identity)
|
|
}
|
|
}
|
|
|
|
for identity, state := range r.disconnectSignalOnResumeNoMessagesParticipants {
|
|
if now.After(state.expiry) {
|
|
delete(r.disconnectSignalOnResumeNoMessagesParticipants, identity)
|
|
}
|
|
}
|
|
r.simulationLock.Unlock()
|
|
|
|
time.Sleep(10 * time.Second)
|
|
}
|
|
}
|
|
|
|
func (r *Room) launchPublisherAgents(p types.Participant) {
|
|
if p == nil || p.IsDependent() || r.agentClient == nil {
|
|
return
|
|
}
|
|
|
|
if r.internal == nil {
|
|
return
|
|
}
|
|
|
|
for _, ag := range r.internal.AgentDispatches {
|
|
go r.agentClient.LaunchJob(context.Background(), &agent.JobRequest{
|
|
JobType: livekit.JobType_JT_PUBLISHER,
|
|
Room: r.ToProto(),
|
|
Participant: p.ToProto(),
|
|
Metadata: ag.Metadata,
|
|
AgentName: ag.AgentName,
|
|
})
|
|
}
|
|
}
|
|
|
|
func (r *Room) DebugInfo() map[string]interface{} {
|
|
info := map[string]interface{}{
|
|
"Name": r.protoRoom.Name,
|
|
"Sid": r.protoRoom.Sid,
|
|
"CreatedAt": r.protoRoom.CreationTime,
|
|
}
|
|
|
|
participants := r.GetParticipants()
|
|
participantInfo := make(map[string]interface{})
|
|
for _, p := range participants {
|
|
participantInfo[string(p.Identity())] = p.DebugInfo()
|
|
}
|
|
info["Participants"] = participantInfo
|
|
|
|
return info
|
|
}
|
|
|
|
// ------------------------------------------------------------
|
|
|
|
func BroadcastDataPacketForRoom(r types.Room, source types.LocalParticipant, kind livekit.DataPacket_Kind, dp *livekit.DataPacket, logger logger.Logger) {
|
|
dp.Kind = kind // backward compatibility
|
|
dest := dp.GetUser().GetDestinationSids()
|
|
if u := dp.GetUser(); u != nil {
|
|
if len(dp.DestinationIdentities) == 0 {
|
|
dp.DestinationIdentities = u.DestinationIdentities
|
|
} else {
|
|
u.DestinationIdentities = dp.DestinationIdentities
|
|
}
|
|
if dp.ParticipantIdentity != "" {
|
|
u.ParticipantIdentity = dp.ParticipantIdentity
|
|
} else {
|
|
dp.ParticipantIdentity = u.ParticipantIdentity
|
|
}
|
|
}
|
|
destIdentities := dp.DestinationIdentities
|
|
|
|
participants := r.GetLocalParticipants()
|
|
capacity := len(destIdentities)
|
|
if capacity == 0 {
|
|
capacity = len(dest)
|
|
}
|
|
if capacity == 0 {
|
|
capacity = len(participants)
|
|
}
|
|
destParticipants := make([]types.LocalParticipant, 0, capacity)
|
|
|
|
var dpData []byte
|
|
for _, op := range participants {
|
|
if op.State() != livekit.ParticipantInfo_ACTIVE {
|
|
continue
|
|
}
|
|
if source != nil && op.ID() == source.ID() {
|
|
continue
|
|
}
|
|
if len(dest) > 0 || len(destIdentities) > 0 {
|
|
if !slices.Contains(dest, string(op.ID())) && !slices.Contains(destIdentities, string(op.Identity())) {
|
|
continue
|
|
}
|
|
}
|
|
if dpData == nil {
|
|
var err error
|
|
dpData, err = proto.Marshal(dp)
|
|
if err != nil {
|
|
logger.Errorw("failed to marshal data packet", err)
|
|
return
|
|
}
|
|
}
|
|
destParticipants = append(destParticipants, op)
|
|
}
|
|
|
|
utils.ParallelExec(destParticipants, dataForwardLoadBalanceThreshold, 1, func(op types.LocalParticipant) {
|
|
err := op.SendDataPacket(kind, dpData)
|
|
if err != nil && !errors.Is(err, io.ErrClosedPipe) && !errors.Is(err, sctp.ErrStreamClosed) &&
|
|
!errors.Is(err, ErrTransportFailure) && !errors.Is(err, ErrDataChannelBufferFull) {
|
|
op.GetLogger().Infow("send data packet error", "error", err)
|
|
}
|
|
})
|
|
}
|
|
|
|
func IsCloseNotifySkippable(closeReason types.ParticipantCloseReason) bool {
|
|
return closeReason == types.ParticipantCloseReasonDuplicateIdentity
|
|
}
|
|
|
|
func connectionDetailsFields(cds []*types.ICEConnectionDetails) []interface{} {
|
|
var fields []interface{}
|
|
connectionType := types.ICEConnectionTypeUnknown
|
|
for _, cd := range cds {
|
|
candidates := make([]string, 0, len(cd.Remote)+len(cd.Local))
|
|
for _, c := range cd.Local {
|
|
cStr := "[local]"
|
|
if c.Selected {
|
|
cStr += "[selected]"
|
|
} else if c.Filtered {
|
|
cStr += "[filtered]"
|
|
}
|
|
if c.Trickle {
|
|
cStr += "[trickle]"
|
|
}
|
|
cStr += " " + c.Local.String()
|
|
candidates = append(candidates, cStr)
|
|
}
|
|
for _, c := range cd.Remote {
|
|
cStr := "[remote]"
|
|
if c.Selected {
|
|
cStr += "[selected]"
|
|
} else if c.Filtered {
|
|
cStr += "[filtered]"
|
|
}
|
|
if c.Trickle {
|
|
cStr += "[trickle]"
|
|
}
|
|
cStr += " " + c.Remote.String()
|
|
candidates = append(candidates, cStr)
|
|
}
|
|
if len(candidates) > 0 {
|
|
fields = append(fields, fmt.Sprintf("%sCandidates", strings.ToLower(cd.Transport.String())), candidates)
|
|
}
|
|
if cd.Type != types.ICEConnectionTypeUnknown {
|
|
connectionType = cd.Type
|
|
}
|
|
}
|
|
fields = append(fields, "connectionType", connectionType)
|
|
return fields
|
|
}
|