Abstract:
A videoconferencing system has a videoconferencing unit that use portable devices as peripherals for the system. The portable devices obtain near-end audio and send the audio to the videoconferencing unit via a wireless connection. In turn, the videoconferencing unit sends the near-end audio from the loudest portable device along with near-end video to the far-end. The portable devices can control the videoconferencing unit and can initially establish the videoconference by connecting with the far-end and then transferring operations to the videoconferencing unit. To deal with acoustic coupling between the unit's loudspeaker and the portable device's microphone, the unit uses an echo canceller that is compensated for differences in the clocks used in the A/D and D/A converters of the loudspeaker and microphone.
Abstract:
A master device (e.g., a laptop) controls a videoconference connection with a far-end and acts as a controller for slave devices in a group. The master establishes the connection with the far-end using a wide area network and connects with the slaves using a Wi-Fi connection of a local area connection. To connect to the slaves, the master outputs an ultrasonic beacon encoding the master's address in the local network. The beacon also allows differences in the various clocks to be handled during audio processing. During the videoconference, the master outputs far-end audio to a loudspeaker and sends the far-end video to its display and to the slaves. Each device captures audio/video with its microphone and camera. The master determines which captured audio is a dominant talker's and may only send that audio to the far-end. The video associated with the dominant talker's device may also be sent to the far-end.
Abstract:
A conferencing system for an enterprise is disclosed. The conferencing system includes a management apparatus for managing conferencing resources in a premises. The management apparatus includes a communications link for communicating with a proximity announcing system, which indicates whether a portable communications device is near a conferencing device within the premises. The management apparatus can transmit a message to a proximate portable communication device indicating that a conferencing device is nearby. The proximity announcing system can include wireless transmitters and receivers associated with the premises and with portable communication devices used within the premises for tracking the location of the portable communication devices throughout the premises.
Abstract:
A scalable audio codec for a processing device determines first and second bit allocations for each frame of input audio. First bits are allocated for a first frequency band, and second bits are allocated for a second frequency band. The allocations are made on a frame-by-frame basis based on the energy ratio between the two bands. For each frame, the codec transform codes both frequency bands into two sets of transform coefficients, which are then packetized based on the bit allocations. The packets are then transmitted with the processing device. Additionally, the frequency regions of the transform coefficients can be arranged in order of importance determined by power levels and perceptual modeling. Should bit stripping occur, the decoder at a receiving device can produce audio of suitable quality given that bits have been allocated between the bands and the regions of transform coefficients have been ordered by importance.
Abstract:
A distributed audio signal processing system having a plurality of linked audio signal processing units is disclosed. Each audio signal processing unit has physical channels for receiving and sending local audio signals and a high bandwidth interface for exchanging audio signals with other linked audio signal processing units. Each of the physical channels of each of the audio signal processing units is mapped to a corresponding global channel. Global channels can be combined to form virtual channels that can be processed as a signal channel. Combinations of virtual channels can be grouped into logical channel groups that can be controlled as a group.
Abstract:
A conferencing endpoint uses acoustic echo cancellation with clock compensation. Receiving far-end audio to be output by a local loudspeaker, the endpoint performs acoustic echo cancellation so that the near-end audio capture by a microphone will lack echo of the far-end audio output from the loudspeaker. The converters for the local microphone and loudspeaker may have different clocks so that their sample rates differ. To assist the echo cancellation, the endpoint uses a clock compensator that cross-correlates an echo estimate of the far-end audio and the near-end audio and adjusts a sample rate conversion factor to be used for the far-end audio analyzed for echo cancellation.
Abstract:
Use of a scalable audio codec to implement distributed mixing and/or sender bit rate regulation in a multipoint conference is disclosed. The scalable audio codec allows the audio signal from each endpoint to be split into one or more frequency bands and for the transform coefficients within such bands to be prioritized such that usable audio may be decoded from a subset of the entire signal. The subset may be created by omitting certain frequency bands and/or by omitting certain coefficients within the frequency bands. By providing various rules for each endpoint in a conference, the endpoint can determine the importance of its signal to the conference and can select an appropriate bit rate, thereby conserving bandwidth and/or processing power throughout the conference.
Abstract:
Systems and methods are disclosed for controlling cropping areas of video images to match the allocated area associated with the image in a video conferencing layout. The disclosed methods can protect regions of interest from being cropped. A region of interest within a video image is identified to adjust the cropping in such a way that the region of interest is preserved within the cropped image. The region of interest may be identified based on motion detection or flesh tone detection, for example.
Abstract:
Systems and methods are disclosed to identify and generate keyword searches in real-time or near real-time for active participants in an on-going Audio, Video and Data Collaboration meeting (also referred to as “Unified Communications and Collaboration” or UCC). In one embodiment, multiple input sources are screened to detect text data and generate search strings from the deciphered keywords. Keywords are deciphered from presentation materials and other forms of data input to a UCC (e.g., documents, video, and audio). Keywords and generated search strings can then be presented to one or more participants for selection (e.g., hyperlink) to retrieve and present supporting material relative to a topic of discussion or point of interest in the UCC. Alternatively, recorded content can be search during or prior to playback to allow incorporation of disclosed embodiments and concepts.
Abstract:
The present application discloses a method and apparatus for a multipoint video conferencing system that can offer a wide variety of layouts and the ability to support a large number of conferees. A received stream of NALs from a plurality of endpoints can be processed by an exemplary NAL Processor Module (NPM) and be arranged into a stream of NALs that are referred to in this disclosure as organized NALs (ONALs). Two or more streams of compressed video organized into ONALs can be received by an exemplary NAL Editor Module (NEM), wherein each stream was created from a video image that was generated at an endpoint participating in the conference. An exemplary NEM can comprise a compressed composed frame memory (CCFM) module. Each CCFM can be divided into segments according to the layout of the continuous presence (CP) image that is to be displayed by one or more receiving endpoints.