Abstract:
A method to quickly determine whether a macroblock in H.264 or other similar standards should be encoded in skip mode in a SIMD processor environment is provided. The method exploits the fact that the processor provides enough register space to store N=4 4×4 set of 8 bits data. By performing 4 consecutive reads of 16 linear aligned values into 16 SIMD register variables each capable of storing 16 values, the entire data set for operating can be stored within the processors SIMD register variables and can be operated upon, without any memory related latency.
Abstract:
According to a general embodiment of the present invention, a computer, tablet or smart phone executable application is provided, allowing a user of the application to activate pairing of the device and a video conference session allowing e.g. for streaming of duo video content or other forms of presentations streams of the video conference session from video conference infrastructure like MCUs, conference bridges or servers.
Abstract:
A method and a device for providing for sharing encoder resources in an MCU without compromising with image quality and bandwidth adaption is provided. The method and device comprises dynamic allocation of encoders in a multi-party video conference of a number of endpoints, wherein the video conference endpoints are adapted to encode and transmit a media stream comprising the steps of creating a new encoder at a request for a new pad, determining whether a maximum number of encoders is reached, if the maximum number of encoders is not reached, then linking the new pad to the new encoder, if the maximum number of encoders is reached, then linking all existing pads to the new encoder, 15 and releasing all unused encoders.
Abstract:
A method for managing media streams including at least video data and audio data in a multi-party video conference involving one or more groups of one or more nodes respectively receiving a media stream from each of a plurality of associated endpoints is provided. In each of the one or more nodes of each of the one or more groups, the loudest audio of the media streams received from the respective plurality of associated endpoints is determined. An indication of the volume of the determined loudest audio and an identification of the node is transmitted to a decision maker. The highest received volume and the associated identification is determined. The node of the identification is instructed to transmit the received media stream of the endpoint to one of: all nodes in the one or more groups and to a proxy.
Abstract:
System and method for determining congestion of a communication link transmitting a media stream over the communication link from a sender device to a receiving device. The method comprising transmitting media packets into the media stream, each media packet having a packet identifier; transmitting recovery packets into the media stream, each recovery packet having a packet identifier and is adapted to replace a missing packet of the media stream; mapping for each packet in the media stream the packet identifier, a transmission timestamp and a packet size; - mapping for each recovery packet the packet identifier of the recovery packet and the packet identifier of the packet the recovery packet is adapted to replace; transmitting for each packet received at the receiving device a feedback message to the sender device, the feedback message comprising the packet identifier and an arrival timestamp; determining a packet as missing when none of the feedback messages comprises the packet identifier of the packet; determining a packet as recovered when a feedback message comprises the packet identifier of a recovery packet corresponding to the missing packet; and determining a packet recovery success rate by dividing a total number of packets determined as recovered within a first sliding window with a total number of packets determined as missing within the first sliding window; and determining that the communication link is congested when for the first sliding window of the N transmitted packets the packet recovery success rate is below a congestion threshold.
Abstract:
The present invention relates to an image processing device and a method of framing changes and movements in a video image divided into NxN blocks of pixel positions. The method comprises calculating a first bitmap of the video image by a DCT transform on each of the NxN blocks of pixel positions, assigning a first binary value to the pixel positions of the NxN blocks when more than an amount of change, and a second binary value to the pixel positions of the NxN blocks when less than an amount of change. Calculating a third bitmap by an OR operation between a number of bitmaps representing past time frames of the video image, calculating a fourth bitmap by performing a dilation process of the third bitmap representing the current time frame of the video image, and creating one or more frames identifying area of changes and movements in the video image based on detecting BLOBs (Binary Large Objects) in the fourth bitmaps.
Abstract:
A real time audio multi-point videoconferencing system comprising at least two videoconferencing endpoints in communication with a multipoint control unit, and a method of providing the real time audio. The method comprising capturing input lossless digital audio streams on each of the at least two videoconferencing endpoints, transmitting the input lossless digital audio streams from each of the at least two videoconferencing endpoints to the MCU using a frame size in the range 0.25 ms - 3 ms, mixing, using the MCU, the input lossless digital audio streams into a plurality of output lossless digital audio streams, transmitting a different one of the plurality of output lossless digital audio streams to each of the at least two videoconferencing endpoints using a frame size in the range 0.25 ms - 3 ms, and playback the plurality of output lossless digital audio streams on each of the at least two videoconferencing endpoints.
Abstract:
A method and a device for providing for sharing encoder resources in an MCU without compromising with image quality and bandwidth adaption is provided. The method and device comprises dynamic allocation of encoders in a multi-party video conference of a number of endpoints, wherein the video conference endpoints are adapted to encode and transmit a media stream comprising the steps of creating a new encoder at a request for a new pad, determining whether a maximum number of encoders is reached, if the maximum number of encoders is not reached, then linking the new pad to the new encoder, if the maximum number of encoders is reached, then linking all existing pads to the new encoder, 15 and releasing all unused encoders.
Abstract:
A method of speech detection and speech enhancement in a speech detection and speech enhancement unit of Multipoint Conferencing Node (MCN) and a method of training the same. The method comprising receiving input audio segments, and determining an acoustic environment based on input audio auxiliary information, extracting T-F-domain features from the received input audio segments, determining if each of the received input audio segments is speech by inputting the T-F domain features into a speech detection classifier trained for the determined acoustic environment, determining, when one of the received input audio segments is speech, if the received audio segment is noisy speech by inputting the T-F domain features into a noise classifier using a statistical generative model representing the probability distributions of the T-F domain features of noisy speech trained for the determined acoustic environment, and applying a noise reduction mask on the received input audio segments according to the determination of the received audio segment is noisy speech
Abstract:
Timeline (200) information is provided to endpoints in a video conferencing system. The timeline (200) comprises a starting time point (201), a current time point (203) and a plurality of timestamps (202) that represent respective points in time between the starting time point and the current time point. A plurality of events are detected that are associated with respective system generated activities or activities performed by participants located at the endpoints. Event information is provided, to at least one endpoint, for displaying along the timeline a respective graphical item (204) representing respective detected events. This provision of event information is performed repeatedly such that the graphical items slide along the timeline in a direction away from the current time point.