Abstract:
A method of reducing bandwidth limitations to send events to a set of interested clients within a pre-defined time period as quickly and fairly as possible. The clients can be re-distributed among the servers in a network such that the delay due to server overloading is minimized by moving clients from an overloaded server to a server with available bandwidth. In addition, the latency of client-server communications can be incorporated into an estimation of download times, and the servers can then initiate delivery to respective clients based on those download times. By staggering the send times to account for heterogeneous latencies, more clients can receive the event at the same time, and a fairness of distribution can be achieved.
Abstract:
Example embodiments provide for keeping an HTTP reply flow for a communication open such that portions of an overall response corresponding to a single request may be sent across the HTTP reply flow. As the various portions of the overall response become available at a corresponding service endpoint, the service appropriately encapsulates the messages and sends them to the requesting endpoint. The recipient or requesting endpoint of the response is then capable of reading the available portions of the response and appropriately decodes the embedded portions and is free to process these as appropriate. Accordingly, because only one request is made for several portions of a response valuable system resource are reserved for only requiring one authentication and/or validation of a requesting endpoint.
Abstract:
Mechanisms for providing requested date items in a request-driven enumeration session while retaining to control over how much inter-message context information is retained by the data provider. Upon receiving a request for the data items, the data provider identifies a portion of the data items to be provided in the first response along with context information that reflects that the first portion of information has been provided. The data provider then makes a determination of how much of the context is to be provided to the data receiver system, and then provides that appropriate context information to the data receiver. Furthermore, the data provider provides a response that includes the first portion of the data items. The data receiver includes this first provided context information in the second request for the next portion of the data items, allowing continuity in the data transfer session.
Abstract:
A method of reducing bandwidth limitations to send events to a set of interested clients within a pre-defined time period as quickly and fairly as possible. The clients can be re-distributed among the servers in a network such that the delay due to server overloading is minimized by moving clients from an overloaded server to a server with available bandwidth. In addition, the latency of client-server communications can be incorporated into an estimation of download times, and the servers can then initiate delivery to respective clients based on those download times. By staggering the send times to account for heterogeneous latencies, more clients can receive the event at the same time, and a fairness of distribution can be achieved.
Abstract:
To provide a fairness of distribution, an encrypted event, containing information not intended for release prior to a release time, can be sent to clients prior to the release time. In such a manner the bulk of the information can be transferred to the clients without concern to the duration of the transfer. At the release time, a small decryption key can be sent, either from a central sever, or from multiple server, utilizing multiple network paths to provide for the greatest likelihood that each client will receive the decryption key with a minimum of delay. Each client is thereby provided access to the information at approximately the same time, regardless of the bandwidth available to each client. Additionally, trusted edge servers, that can be trusted not to release information prior to an appropriate time, can send an unencrypted event, or decrypt the encrypted event and send the decrypted event, at a determined time, either prior to or after the release time, such that the decrypted or unencrypted event arrives at the clients that could not store and decrypt the encrypted event at approximately the same time as the key arrives at the other clients. Each client can thus receive the information at approximately the same time, regardless of the client's bandwidth or its ability to store and decrypt information.
Abstract:
Mechanisms for automatically generating code that tests capabilities of a test computing system to simulate a message exchange pattern. The code generation computing system uses a message exchange pattern definition to generate the simulation code. For each state in which the message exchange pattern definition allows valid messages to be transmitted, code is generated for that state that at least simulates the transmission of a valid transmission message. For each state in which the message exchange pattern definition allows valid messages to be received, code is generated for that state that simulates the receipt of a valid receipt message. If the transmission or receipt of the message causes a state transition to occur, code is generated that causes the appropriate state transition in the message exchange pattern.
Abstract:
Methods and systems for providing a virtual network are disclosed. At least one layer of abstraction is created between network service applications and conventional network protocols by inserting an adaptive dispatcher between applications and network transport services on each machine in a network. The message protocol in the virtual network is extensible, allowing application programs to create new headers within any message as needed. The adaptive dispatcher contains handlers that route and dispatch messages within the virtual network based on arbitrary content within each message, including any combination of headers and/or data content. Each device on the virtual network has a virtual address to which messages are directed, allowing devices to move within the network without reconfiguring routing tables. Handlers may be automatically created when an event meeting predefined criteria occurs, including the non-occurrence of a condition, making the virtual network self-healing and adaptive to reconfiguration.