Patent application number | Description | Published |
20150293784 | Model Driven Optimization of Annotator Execution in Question Answering System - Mechanisms are provided for scheduling execution of pre-execution operations of an annotator of a question and answer (QA) system pipeline. A model is used to represent a system of annotators of the QA system pipeline, where the model represents each annotator as a node having one or more performance parameters indicating a performance of an execution of an annotator corresponding to the node. For each annotator in a set of annotators of the system of annotators, an effective response time for the annotator is calculated based on the performance parameters. A pre-execution start interval for a first annotator based on an effective response time of a second annotator is calculated where execution of the first annotator is sequentially after execution of the second annotator. Execution of pre-execution operations associated with the first annotator is scheduled based on the calculated pre-execution start interval for the first annotator. | 10-15-2015 |
20150347430 | TECHNIQUES FOR ENABLING COARSE-GRAINED VOLUME SNAPSHOTS FOR VIRTUAL MACHINE BACKUP AND RESTORE - A technique of backing up a workload in a virtual environment includes identifying one or more files that are associated with the workload. One or more source volumes that include the one or more files are identified. A respective target volume is provisioned for each of the one or more source volumes identified. Only dirty blocks are copied (in a snapshot mode that prevents an update to the one or more source volumes during the snapshot mode) from each of the one or more source volumes to its respective target volume. The one or more dirty blocks are then copied from each target volume to a backup medium. | 12-03-2015 |
20150363275 | TECHNIQUES FOR IMPROVING CLOUD INFRASTRUCTURE BACKUP IN A SHARED STORAGE ENVIRONMENT - A technique for cloud infrastructure backup in a virtualized environment utilizing shared storage includes obtaining a workload input/output (I/O) profile to the shared storage over a time period. An attempt to locate one or more time windows in the workload I/O profile for which a cloud infrastructure backup can be staged is initiated. In response to determining the cloud infrastructure backup can be staged during at least one of the time windows, staging of the cloud infrastructure backup is scheduled during a selected one of the time windows. In response to determining the cloud infrastructure backup cannot be staged during at least one of the time windows, an interference tolerance approach is employed for accessing the shared storage for active workloads and the cloud infrastructure backup during the staging of the cloud infrastructure backup. | 12-17-2015 |
20150363413 | TECHNIQUES FOR ENABLING COARSE-GRAINED VOLUME SNAPSHOTS FOR VIRTUAL MACHINE BACKUP AND RESTORE - A technique of backing up a workload in a virtual environment includes identifying one or more files that are associated with the workload. One or more source volumes that include the one or more files are identified. A respective target volume is provisioned for each of the one or more source volumes identified. Only dirty blocks are copied (in a snapshot mode that prevents an update to the one or more source volumes during the snapshot mode) from each of the one or more source volumes to its respective target volume. The one or more dirty blocks are then copied from each target volume to a backup medium. | 12-17-2015 |
20150365478 | CLOUD INFRASTRUCTURE BACKUP IN A SHARED STORAGE ENVIRONMENT - A technique for cloud infrastructure backup in a virtualized environment utilizing shared storage includes obtaining a workload input/output (I/O) profile to the shared storage over a time period. An attempt to locate one or more time windows in the workload I/O profile for which a cloud infrastructure backup can be staged is initiated. In response to determining the cloud infrastructure backup can be staged during at least one of the time windows, staging of the cloud infrastructure backup is scheduled during a selected one of the time windows. In response to determining the cloud infrastructure backup cannot be staged during at least one of the time windows, an interference tolerance approach is employed for accessing the shared storage for active workloads and the cloud infrastructure backup during the staging of the cloud infrastructure backup. | 12-17-2015 |
20150370588 | SELECTING OPTIMAL HYPERVISOR PLATFORMS THAT SATISFY APPLICATION WORKLOAD REQUIREMENTS - A method, system and computer program product for selecting hypervisor platforms that are best suited to process application workloads. Attribute requirements for an application workload, such as high CPU capacity, high power and low cost, are received. A ranking algorithm is then applied to a list of pools of compute nodes to identify an ordered list of pools of compute nodes that are best suited for satisfying the attribute requirements of the application workload by comparing hypervisor characteristics of the pools of compute nodes with the attribute requirements of the application workload. Each pool of compute nodes runs on a particular hypervisor platform which has a unique combination of characteristics that correspond to a combination of a set of attribute requirements (e.g., medium CPU/memory/disk capacity; high CPU and memory performance). In this manner, the hypervisor platforms that are best suited for satisfying the application workload requirements are identified. | 12-24-2015 |
20150370594 | OPTIMIZING RUNTIME PERFORMANCE OF AN APPLICATION WORKLOAD BY MINIMIZING NETWORK INPUT/OUTPUT COMMUNICATIONS BETWEEN VIRTUAL MACHINES ON DIFFERENT CLOUDS IN A HYBRID CLOUD TOPOLOGY DURING CLOUD BURSTING - A method, system and computer program product for optimizing runtime performance of an application workload. Network input/output (I/O) operations between virtual machines of a pattern of virtual machines servicing the application workload in a private cloud are measured over a period of time and depicted in a histogram. A score is generated for each virtual machine or group of virtual machines in the pattern of virtual machines based on which range in the ranges of I/O operations per seconds (IOPS) depicted in the histogram has the largest sample size and the number of virtual machines in the same pattern that are allowed to be in the public cloud. In this manner, the runtime performance of the application workload is improved by minimizing the network input/output communications between the two cloud environments by migrating those virtual machine(s) or group(s) of virtual machines with a score that exceeds a threshold value. | 12-24-2015 |
20150370601 | OPTIMIZING RUNTIME PERFORMANCE OF AN APPLICATION WORKLOAD BY MINIMIZING NETWORK INPUT/OUTPUT COMMUNICATIONS BETWEEN VIRTUAL MACHINES ON DIFFERENT CLOUDS IN A HYBRID CLOUD TOPOLOGY DURING CLOUD BURSTING - A method, system and computer program product for optimizing runtime performance of an application workload. Network input/output (I/O) operations between virtual machines of a pattern of virtual machines servicing the application workload in a private cloud are measured over a period of time and depicted in a histogram. A score is generated for each virtual machine or group of virtual machines in the pattern of virtual machines based on which range in the ranges of I/O operations per seconds (IOPS) depicted in the histogram has the largest sample size and the number of virtual machines in the same pattern that are allowed to be in the public cloud. In this manner, the runtime performance of the application workload is improved by minimizing the network input/output communications between the two cloud environments by migrating those virtual machine(s) or group(s) of virtual machines with a score that exceeds a threshold value. | 12-24-2015 |
20150373093 | SELECTING OPTIMAL HYPERVISOR PLATFORMS THAT SATISFY APPLICATION WORKLOAD REQUIREMENTS - A method, system and computer program product for selecting hypervisor platforms that are best suited to process application workloads. Attribute requirements for an application workload, such as high CPU capacity, high power and low cost, are received. A ranking algorithm is then applied to a list of pools of compute nodes to identify an ordered list of pools of compute nodes that are best suited for satisfying the attribute requirements of the application workload by comparing hypervisor characteristics of the pools of compute nodes with the attribute requirements of the application workload. Each pool of compute nodes runs on a particular hypervisor platform which has a unique combination of characteristics that correspond to a combination of a set of attribute requirements (e.g., medium CPU/memory/disk capacity; high CPU and memory performance). In this manner, the hypervisor platforms that are best suited for satisfying the application workload requirements are identified. | 12-24-2015 |