Patent application number | Description | Published |
20120102357 | METHOD AND APPARATUS FOR PROCESSING LOAD INSTRUCTIONS IN A MICROPROCESSOR HAVING AN ENHANCED INSTRUCTION DECODER AND AN ENHANCED LOAD STORE UNIT - A method and microprocessor are described for efficiently executing load instructions out-of-order (speculatively). The microprocessor includes an enhanced load store unit (LSU) and an enhanced instruction decoder. The enhanced LSU receives a plurality of out-of-order value addresses, and sends a resync signal to the enhanced instruction decoder when an execution error associated with a particular load instruction occurs. The enhanced instruction decoder stores a specific address associated with the particular load instruction, and increments a counter value that indicates how many times the resync signal was sent by the resync predictor. When the counter value reaches a predetermined threshold, subsequent load instructions from the specific address are executed in order (non-speculatively). When a future execution of the particular load instruction indicates that the probability of an execution error has been reduced, the counter value is decremented, facilitating newer load instructions associated with the same address to again be executed speculatively. | 04-26-2012 |
20140379986 | STACK ACCESS TRACKING - A processor employs a prediction table at a front end of its instruction pipeline, whereby the prediction table stores address register and offset information for store instructions; and stack offset information for stack access instructions. The stack offset information for a corresponding instruction indicates the entry of the stack accessed by the instruction stack relative to a base entry. The processor uses pattern matching to identify predicted dependencies between load/store instructions and predicted dependencies between stack access instructions. A scheduler unit of the instruction pipeline uses the predicted dependencies to perform store-to-load forwarding or other operations that increase efficiency and reduce power consumption at the processing system. | 12-25-2014 |
20140380022 | STACK ACCESS TRACKING USING DEDICATED TABLE - A processor employs a prediction table at a front end of its instruction pipeline, whereby the prediction table stores address register and offset information for store instructions; and stack offset information for stack access instructions. The stack offset information for a corresponding instruction indicates the location of the data accessed by the instruction at the processor stack relative to a base location. The processor uses pattern matching to identify predicted dependencies between load/store instructions and predicted dependencies between stack access instructions. A scheduler unit of the instruction pipeline uses the predicted dependencies to perform store-to-load forwarding or other operations that increase efficiency and reduce power consumption at the processing system. | 12-25-2014 |
Patent application number | Description | Published |
20130132954 | Image Deployment in a Cloud Environment - The invention relates to a mechanism for image deployment in a cloud environment comprising at least two hosts coupled to at least one disk, and at least one virtual machine created in at least one host of the at least two hosts by deploying an additional image to the at least one disk. The method is characterized in that the at least one host for deploying an additional image is selected according to a deployment factor representing a communality relationship between one or more existing base images in the cloud environment and the additional image. | 05-23-2013 |
20130132955 | Customizable File-Type Aware Cache Mechanism - A caching mechanism is provided in a computer server system including at least one host coupled to at least one disk, one or more virtual machines, one or more images of the one or more virtual machines created in a file system of the disk, and a virtual machine manager for assigning hardware resources of the computer server system to the one or more virtual machines. The mechanism provides a masking cache layer between the one or more virtual machines and the virtual machine manager, the masking cache layer including at least one separate entry for every cacheable data block of the file system to be referenced, collects profiling information related to one or more of the cacheable data blocks ( | 05-23-2013 |
20130159252 | Customizable File-Type Aware Cache Mechanism - A caching mechanism is provided in a computer server system including at least one host coupled to at least one disk, one or more virtual machines, one or more images of the one or more virtual machines created in a file system of the disk, and a virtual machine manager for assigning hardware resources of the computer server system to the one or more virtual machines. The mechanism provides a masking cache layer between the one or more virtual machines and the virtual machine manager, the masking cache layer including at least one separate entry for every cacheable data block of the file system to be referenced, collects profiling information related to one or more of the cacheable data blocks ( | 06-20-2013 |
20140137113 | VIRTUAL MACHINE TEMPLATE CREATION BASED ON DATA FEEDS - Data that includes information about an application is retrieved from a data feed. It is determined that the data from the data feed indicates that the application has been updated. In response to a determination that the data from the data feed indicates that the application has been updated, data descriptive of the application is extracted from the data from the data feed. One or more instructions for installing the application are determined based, at least in part, on the data descriptive of the application. In response to determination of the one or more instructions for installing the application, a first virtual machine skeleton that includes the one or more instructions for installing the application is generated. | 05-15-2014 |
20140137114 | VIRTUAL MACHINE TEMPLATE CREATION BASED ON DATA FEEDS - Data that includes information about an application is retrieved from a data feed. It is determined that the data from the data feed indicates that the application has been updated. In response to a determination that the data from the data feed indicates that the application has been updated, data descriptive of the application is extracted from the data from the data feed. One or more instructions for installing the application are determined based, at least in part, on the data descriptive of the application. In response to determination of the one or more instructions for installing the application, a first virtual machine skeleton that includes the one or more instructions for installing the application is generated. | 05-15-2014 |
20140214776 | DATA DE-DUPLICATION FOR DISK IMAGE FILES - The invention relates to a data processing system, comprising at least two disk emulators operating in parallel and emulating a disk subsystem each, the disk emulators each using a file in a file system for any data stored on the respective disk, a separate de-duplicator for de-duplicating the data stored in the files, the de-duplicator operating in parallel to the disk emulators, the de-duplicator further using an additional disk emulator emulating an additional disk subsystem by using an additional file in a file system for storing data shared between the other disk subsystems. | 07-31-2014 |
20140282537 | VIRTUAL MACHINE IMAGE DISK USAGE - The invention relates to a method for managing virtual machine image disk usage comprising a disk image emulator for a virtual machine provided by a hypervisor, comprising the steps of providing at least a first disk image comprising a sequence of data blocks for accumulating write operations to the first disk image, providing at least a second disk image comprising a sequence of data blocks for permanently storing disk image data, and providing a disk cleaning process for transferring disk image data from the first disk image to the second disk image and deleting unused data blocks in the first and/or the second disk image. | 09-18-2014 |
20140325510 | Image Deployment in a Cloud Environment - The invention relates to a mechanism for image deployment in a cloud environment comprising at least two hosts coupled to at least one disk, and at least one virtual machine created in at least one host of the at least two hosts by deploying an additional image to the at least one disk. The method is characterized in that the at least one host for deploying an additional image is selected according to a deployment factor representing a communality relationship between one or more existing base images in the cloud environment and the additional image. | 10-30-2014 |
20160055027 | VIRTUAL MACHINE IMAGE DISK USAGE - The invention relates to a method for managing virtual machine image disk usage comprising a disk image emulator for a virtual machine provided by a hypervisor, comprising the steps of providing at least a first disk image comprising a sequence of data blocks for accumulating write operations to the first disk image, providing at least a second disk image comprising a sequence of data blocks for permanently storing disk image data, and providing a disk cleaning process for transferring disk image data from the first disk image to the second disk image and deleting unused data blocks in the first and/or the second disk image. | 02-25-2016 |
Patent application number | Description | Published |
20150199204 | Controlling the Configuration of Computer Systems - Embodiments relate to controlling configuration of a computer system comprising one or more exchangeable components. The exchangeable components comprising identification means to store an identifier. A pair of a private key and a public key are generated for each accepted manufacturer of the exchangeable components and a pair of a private key and a public key for the computer system; assigning an identifier for each exchangeable component available for attachment to the system; receiving configuration data comprising a list of encrypted identity records comprising identifiers of the components together with signatures over the data generated with the private key of the respective component manufacturer for each component expected to be attached to the system; and receiving a configuration record. The configuration data of the expected components from the received configuration record is compared with the configuration data of the components attached to the system. | 07-16-2015 |
20150268973 | Controlling the Configuration of Computer Systems - Embodiments relate to controlling configuration of a computer system comprising one or more exchangeable components. The exchangeable components comprising identification means to store an identifier. A pair of a private key and a public key are generated for each accepted manufacturer of the exchangeable components and a pair of a private key and a public key for the computer system; assigning an identifier for each exchangeable component available for attachment to the system; receiving configuration data comprising a list of encrypted identity records comprising identifiers of the components together with signatures over the data generated with the private key of the respective component manufacturer for each component expected to be attached to the system; and receiving a configuration record. The configuration data of the expected components from the received configuration record is compared with the configuration data of the components attached to the system. | 09-24-2015 |
20160048404 | SUSPENDING AND RESUMING VIRTUAL MACHINES - Suspending multiple virtual machines executed simultaneously a hypervisor a data processing system, including: sending, by a control instance, a request to the hypervisor to suspend virtual machines at a given future first point in time; the hypervisor determining if its virtual machines can be suspended at the first point in time, and, if the suspending is possible, returning a suspend handle to the control instance and storing the suspend handle, otherwise returning an invalid handle; when receiving a suspend handle from the hypervisor in return to the request to suspend virtual machines, the control instance sending a suspend command to the hypervisor, wherein the respective suspend handle is provided to the corresponding hypervisor; the hypervisor determining if the received suspend handle is equal to its stored handle and, if yes, suspending its virtual machines at the first point in time. | 02-18-2016 |
20160048405 | SUSPENDING AND RESUMING VIRTUAL MACHINES - Suspending multiple virtual machines executed simultaneously a hypervisor a data processing system, including: sending, by a control instance, a request to the hypervisor to suspend virtual machines at a given future first point in time; the hypervisor determining if its virtual machines can be suspended at the first point in time, and, if the suspending is possible, returning a suspend handle to the control instance and storing the suspend handle, otherwise returning an invalid handle; when receiving a suspend handle from the hypervisor in return to the request to suspend virtual machines, the control instance sending a suspend command to the hypervisor, wherein the respective suspend handle is provided to the corresponding hypervisor; the hypervisor determining if the received suspend handle is equal to its stored handle and, if yes, suspending its virtual machines at the first point in time. | 02-18-2016 |