Patent application number | Description | Published |
20120066400 | SYSTEM AND METHOD FOR PARALLEL MUXING BETWEEN SERVERS IN A CLUSTER - Systems and methods are provided for providing parallel muxing between servers in a cluster. One such system can include a cluster of one or more high performance computing systems, each including one or more processors and a high performance memory. The cluster communicates over an InfiniBand network. The system can also include a middleware environment, executing on the cluster, that includes one or more application server instances. The system can further include a plurality of muxers, wherein each application server instance includes at least one muxer. Each muxer can receive information from a plurality of threads to transmit to a different muxer on a different application server instance over the Infiniband network using a plurality of parallel channels. | 03-15-2012 |
20120144045 | SYSTEM AND METHOD FOR SUPPORTING ONE-WAY REMOTE METHOD INVOCATION FOR SESSION REPLICATION IN A SERVER CLUSTER - A system and method can support one-way Remote Method Invocation (RMI) session replication in a middleware machine environment, such as a server cluster. The server cluster can include a primary application server and a secondary application server. The primary application server receives a request with a session from a client, maintain the session information, and replicate the session information to a secondary application server using one-way Remote Method Invocation (RMI). The primary application server can then respond to the client based on the session information after receiving a return message of the one-way RMI from a network socket associate with the primary application server. The benefit of using one-way RMI is that the primary application server does not have to wait for an acknowledgment from the secondary application server instance, since the underlying transport protocol guarantees that the session data has been successfully transmitted to the second server instance. | 06-07-2012 |
20120239730 | SYSTEM INCLUDING A MIDDLEWARE MACHINE ENVIRONMENT - A system and method for providing a middleware machine or similar platform. In accordance with an embodiment the system (referred to herein in some implementations as “Exalogic”) comprises a combination of high performance hardware, together with an application server or middleware environment, to provide a complete Java EE application server complex which includes a massively parallel in-memory grid, can be provisioned quickly, and can scale on demand. In accordance with an embodiment, the system can be deployed as a full, half, or quarter rack, or other configuration, that provides an application server grid, storage area network, and InfiniBand network, which support the execution of an application server, middleware or other functionality such as, for example, WebLogic Server, JRockit or Hotspot JVM, Oracle Linux or Solaris, and Oracle VM. Additional features of the system can include, e.g. Zero Buffer Copies, Scatter/Gather I/O, T3 Connections, and Lazy Deserialization. | 09-20-2012 |
20130081060 | System and Method for Efficient Concurrent Queue Implementation - A method, system, and medium are disclosed for facilitating communication between multiple concurrent threads of execution using an efficient concurrent queue. The efficient concurrent queue provides an insert function usable by producer threads to insert messages concurrently. The queue also includes a consume function usable by consumer threads to read the messages from the queue concurrently. The consume function is configured to guarantee a per-producer ordering, such that, for any producer, messages inserted by the producer are read only once and in the order in which the producer inserted those messages. | 03-28-2013 |
20130081061 | Multi-Lane Concurrent Bag for Facilitating Inter-Thread Communication - A method, system, and medium are disclosed for facilitating communication between multiple concurrent threads of execution using a multi-lane concurrent bag. The bag comprises a plurality of independently-accessible concurrent intermediaries (lanes) that are each configured to store data elements. The bag provides an insert function executable to insert a given data element into the bag by selecting one of the intermediaries and inserting the data element into the selected intermediary. The bag also provides a consume function executable to consume a data element from the bag by choosing one of the intermediaries and consuming (removing and returning) a data element stored in the chosen intermediary. The bag guarantees that execution of the consume function consumes a data element if the bag is non-empty and permits multiple threads to execute the insert or consume functions concurrently. | 03-28-2013 |
20130346376 | De-Duplicating Immutable Data at Runtime - De-duplication of immutable data items at runtime may include identifying a set of potentially duplicate immutable data items in use by one or more applications. The applications may access the immutable data items through pointers of respective objects corresponding to the immutable data items. A de-duplication component executing distinctly from the applications may analyze the identified set of potentially duplicate immutable data items to determine two or more that have identical content and may then modify one or more pointers of the corresponding objects so that at least two of the pointers point to a single immutable data item. | 12-26-2013 |
20140047140 | SYSTEM AND METHOD FOR PROVIDING A LINEARIZABLE REQUEST MANAGER - Described herein are systems and methods for improving concurrency of a request manager for use in an application server or other environment. A request manager receives a request, and upon receiving the request the request manager associates a token with the request. A reference to the request is enqueued in each of a plurality of queues, wherein each queue stores a local copy of the token. A first reference to the request is dequeued from a particular queue, wherein when the first reference to the request is dequeued, the token is modified to create a modified token. Thereafter the request is processed. When other references to the request are dequeued from other queues, the other references to the request are discarded. | 02-13-2014 |
20140215475 | SYSTEM AND METHOD FOR SUPPORTING WORK SHARING MUXING IN A CLUSTER - A system and method can provide efficient low-latency muxing between servers in the cluster. One such system can include a cluster of one or more high performance computing systems, each including one or more processors and a high performance memory. The cluster communicates over an InfiniBand network. The system can also include a middleware environment, executing on the cluster, which includes one or more application server instances. The system can include one or more selectors, wherein each said selector contains a queue of read-ready file descriptors. Furthermore, the system can include a shared queue, wherein the read-ready file descriptors in each said selector can be emptied into the shared queue. Additionally, a plurality of muxer threads operates to take work from said shared queue. | 07-31-2014 |
20140245309 | SYSTEM AND METHOD FOR TRANSFORMING A QUEUE FROM NON-BLOCKING TO BLOCKING - A system and method can use continuation-passing to transform a queue from non-blocking to blocking. The non-blocking queue can maintain one or more idle workers in a thread pool that is not accessible from outside of the non-blocking queue. The continuation-passing can eliminate one or more serialization points in the non-blocking queue, and allows a caller to manage the one or more idle workers in the thread pool from outside of the non-blocking queue. | 08-28-2014 |
20140245312 | SYSTEM AND METHOD FOR SUPPORTING COOPERATIVE CONCURRENCY IN A MIDDLEWARE MACHINE ENVIRONMENT - A system and method can support cooperative concurrency in a priority queue. The priority queue, which includes a calendar ring and a fast lane, can detect one or more threads that contend to claim one or more requests in the priority queue. Then, a victim thread can place a request in the fast lane in the priority queue, and release a contending thread, which proceeds to consume the request in the fast lane. | 08-28-2014 |
20140245313 | SYSTEM AND METHOD FOR USING A SEQUENCER IN A CONCURRENT PRIORITY QUEUE - A system and method can support a concurrent priority queue. The concurrent priority queue allows a plurality of threads to interact with the priority queue. The priority queue can use a sequencer to detect and order a plurality of threads that contend for one or more requests in the priority queue. Furthermore, the priority queue operates to reduce the contention among the plurality of threads. | 08-28-2014 |
20140280988 | SYSTEM AND METHOD FOR PARALLEL MULTIPLEXING BETWEEN SERVERS IN A CLUSTER - Systems and methods are provided for providing parallel muxing between servers in a cluster. One such system can include a cluster of one or more high performance computing systems, each including one or more processors and a high performance memory. The cluster communicates over an InfiniBand network. The system can also include a middleware environment, executing on the cluster, that includes one or more application server instances. The system can further include a plurality of muxers, wherein each application server instance includes at least one muxer. Each muxer can receive information from a plurality of threads to transmit to a different muxer on a different application server instance over the Infiniband network using a plurality of parallel channels. | 09-18-2014 |