Entries |
Document | Title | Date |
20100070459 | EXTENSIBLE DATA WAREHOUSE FOR FLOW CYTOMETRY DATA - Methods, systems, and computer program products for storing, managing, querying, and displaying data are described herein. The method operates by classifying and displaying received flow cytometry data. In an embodiment, the method creates a first extensible dimension table with columns that include keywords from received data. The method creates a second extensible dimension table with columns corresponding to parameters from the received data. The method then populates dimension tables with data describing values of dimensions of the classified data. The method populates a first fact table from the classified data. The method then populates a second fact table with a summary of data from the first fact table. The method displays a user interface on a display of a computer device. The method then permits a user to select query attributes including keywords and parameters. The method builds and executes queries based upon dimensions generated from grouped selected query attributes. | 03-18-2010 |
20100070460 | SYSTEM AND METHOD FOR RULE-BASED DATA OBJECT MATCHING - An object matching system comprises a plurality of disparate databases comprising data objects and a master data server. The master data server comprises a computer programmed to provide a plurality of adapters configured to connect to said plurality of disparate databases, a matching module configured to compare data objects provided by said plurality of adapters through use of an object matching rule to determine at least two matching data objects, and a data integrator configured to merge said at least two matching data objects to produce a master data object. | 03-18-2010 |
20100082525 | APPARATUS, SYSTEM, AND METHOD FOR PROCESSING MDX POST-ORDER HIERARCHIZE EXPRESSIONS - Disclosed are an apparatus, system, and method for generating a post-order set of members referenced by an MDX query statement invoking a post-order function. The apparatus includes an input module that receives the MDX expression specifying the relevant data set. A pre-order module generates a set of the members in pre-order organization. An initialization module generates an empty post-order set and inserts the first member of the pre-order set into the post-order set. An ancestor module determines whether the next member in the pre-order set has an ancestor in the post-order set and the location of the closest ancestor of the member if it exists. An insertion module then inserts the member immediately before the closest ancestor if the ancestor exists, or puts it in the rightmost position of the post-order set if an ancestor does not exist. An iteration module invokes the ancestor and insertion modules for each member of the pre-order set until the post-order set is filled. | 04-01-2010 |
20100114814 | ONLINE MARKETPLACE FOR AUTOMATICALLY EXTRACTED DATA - A system for automatically locating and data-typing information originating from many Web pages, and then collecting that information in a database. The database is then made available via an online data marketplace which allows users from different organizations to buy and sell related data, associated advertisements, and access to the communities of end-users who may also view advertisements and make purchases. | 05-06-2010 |
20100125545 | FEDERATING BUSINESS EVENT DATA WITHIN AN ENTERPRISE NETWORK - Parsing of event queries in an enterprise system is described. The enterprise system receives queries, which are broken down into query components. The query components each represent segments of the query. The enterprise system identifies sources of event data and sends queries towards the sources of the event data. The query components are processed close to the data source. The responses are combined to generate an event query response that indicates the event data of each of the query components. | 05-20-2010 |
20100131455 | CROSS-WEBSITE MANAGEMENT INFORMATION SYSTEM - Methods, systems, apparatus, and software for providing information from one or more data feeds to a user are provided. In some embodiments, the methods provided by the invention include obtaining at least one putative feed; confirming that the putative feed is a feed; extracting at least a portion of content from the feed or data related to the content from the feed; processing the extracted content or the data to produce a synthetic feed; and presenting the synthetic feed to the user. In more specific embodiments, the putative fee is selected from the group consisting of: RSS feeds and Web syndication feeds. In still more specific embodiments, the portion of the content or data related to the content is selected from the group consisting of: blog posts, podcasts, videos, and Web pages. | 05-27-2010 |
20100131456 | SYSTEMS AND METHODS FOR MIDDLEWARE DATABASE HOSTING DIMENSIONAL TRANSFORMS - Embodiments relate to systems and methods for a middleware database hosting dimensional transforms. A set of applications can access a combined data of a set of data sources via a database system. The set of data sources can be configured with diverse dimensions and associated data. The requesting application(s) can generate a request to the database system to extract data in compatible formats for the application(s). In embodiments, the database system can transform selected dimensional data (e.g., company, year, sales dimensions) from the data sources into transformed dimensional data. The requesting application(s) can therefore perform searches, sorts, and/or other operations on the transformed dimensional results. | 05-27-2010 |
20100131457 | FLATTENING MULTI-DIMENSIONAL DATA SETS INTO DE-NORMALIZED FORM - Performance metrics data in a multi-dimensional structure such as a nested scorecard matrix is transformed into a flat structure or de-normalized for efficient querying of individual records. Each dimension and header is converted to a column and data values resolved at intersection of dimension levels through an iterative process covering all dimensions and headers of the data structure. A key corresponding to a tuple representation of each cell or a transform of the tuple may be used to identify rows corresponding to the resolved data in cells for further enhanced query capabilities. | 05-27-2010 |
20100131458 | GENERATION OF AN EXTRACTED TIMING MODEL FILE - A system, apparatus and method for generating extracted timing model files, such as macro library files, are disclosed. A user interface or data template is provided to an engineer that allows for the population of data within particular fields related to timing characteristics of an IP block, cell or core. Various types of data and fields may be provided into the user interface or data template. The location of relevant files, such as a cell or core netlist, may be provided within the template. Additionally, one or more modes may be selected by the user to define the manner in which the ETM file(s) are to be generated. An ETM file is automatically generated using the information provided in the data template. | 05-27-2010 |
20100153331 | SYSTEM AND METHOD FOR MANAGING SEMANTIC AND SYNTACTIC METADATA - A method for managing semantic and syntactic metadata. The method begins with receiving heterogeneous data from either a machine or a human. Next, the method captures the semantic metadata corresponding to the received heterogeneous data by means of a semantic template. Next, the method captures the syntactic metadata corresponding to the received heterogeneous data by means of a syntactic template. After capturing the semantic and syntactic metadata, the method logically links the received heterogeneous data with the captured semantic and syntactic metadata. Finally, the method stores the received heterogeneous data in a repository. | 06-17-2010 |
20100161548 | SYSTEM AND METHOD FOR CAPACITY PLANNING IN AN INFORMATION NETWORK - A systems and method for extracting utilization data from a plurality of system monitoring products, storing the extracted utilization data in a data warehouse, correlating the extracted utilization data extracted from the plurality of system monitoring products, retrieving utilization data for a group of systems from the data warehouse and displaying the retrieved utilization data on a display. | 06-24-2010 |
20100169266 | DATA-OVERLAP ANALYSIS FOR A DATA-WAREHOUSING SYSTEM - A computer system is used to manage content in a database system that stores data representing data elements defined by a logical data model. For at least one of multiple business functions to be carried out with the database system, the computer system (1) identifies a total amount of data elements required to carry out the business function, (2) identifies an amount of common data elements that are required to carry out both the business function and another of the multiple business functions, (3) creates information that indicates a percentage of the total amount of data elements that are also common data elements, and (4) delivers the information for presentation to a human user. | 07-01-2010 |
20100185581 | Data aggregation module supporting dynamic query responsive aggregation during the servicing of database query requests provided by one or more client machines - Improved method of and apparatus for aggregating data elements in multidimensional databases (MDDB). In the preferred embodiment, the apparatus is realized in the form of a high-performance stand-alone (i.e. external) aggregation server which can be plugged-into conventional MOLAP systems to achieve significant improments in system performance. In accordance with the principles of the present invention, the stand-alone aggregation server contains a scalable MDDB and a high-performance aggregation engine that are integrated into the modular architecture of the aggregation server. The stand-alone aggregation server of the present invention can uniformly distribute data elements among a plurality of processors, for balanced loading and processing, and therefore is highly scalable. The stand-alone aggregation server of the present invention can be used to realize (i) an improved MDDB for supporting on-line analytical processing (OLAP) operations, (ii) an improved Internet URL Directory for supporting on-line information searching operations by Web-enabled client machines, as well as (iii) diverse types of MDDB-based systems for supporting real-time control of processes in response to complex states of information reflected in the MDDB. | 07-22-2010 |
20100191700 | Communication Handler for Flex Integration with a Secure Application - Systems and methods are provided for a global communication handler to facilitate communication between an Flex application and an Enterprise Application Server (“EAS”) database. Data is obtained from the EAS database by the global communication handler, and the global communication handler instantiates at least one object that provides that data to a Flex application. The global communication handler also receives data from the Flex application, and writes that data to the EAS database. | 07-29-2010 |
20100198778 | Method and system for collecting and distributing user-created content within a data-warehouse-based computational system - Embodiments of the present invention are directed to data-warehouse-based computational systems that provide for collection and distribution of user-created content among users of the data-warehouse-based computational systems. User-created content is initially stored in a semantic-model-based data-storage facility, for efficient and rapid access to other users, and migrates to storage according to a dimensional model within the data warehouse. The combination of a semantic-model-based data-storage facility and dimensional-model-based data-storage facility within the data warehouse provides both fast access to, and rapid update of, user-created content as well as incorporation of user-created content into the data warehouse for long-term storage and access. | 08-05-2010 |
20100205141 | METHOD, SYSTEM AND COMPUTER PROGRAM PRODUCT FOR DATA COLLECTION AND RETRIEVAL FOR MEDICAL RESEARCH - A method, system and computer program product for data collection and retrieval for medical research includes selecting at least one group of medical documents, from a plurality of medical document groups, to retrieve documents from. Search criteria is supplied for a search in the at least one group and a report is generated including medical documents matching the search criteria. Medical data is extracted from source medical systems. The extracted data is transported using flat files. The flat files are transformed and loaded into a central warehouse database, wherein the central warehouse database includes a master table and a medical data table, and the search is performed on the at least a portion of the central warehouse database. Records in the source medical systems are standardized. | 08-12-2010 |
20100205142 | APPARATUS, METHOD, SYSTEM AND COMPUTER-READABLE MEDIUM FOR STORING AND MANAGING IMAGE DATA - An apparatus, method, system and computer-readable medium store and manage image data with automatic labeling of image data corresponding to body slices, such as obtained by a computed tomography scanner. The labels include a body coordinate value along the body axis. The respective body coordinate value can be determined by comparing received image data sets with reference data sets with known attached coordinate values utilizing pattern recognition techniques. Applications include medical image data management in hospitals or operating and providing medical networks. Queries for images that include particular body regions are processed more efficiently. This results in less local memory required and narrower bandwidth resources of transmission networks. | 08-12-2010 |
20100205143 | System for Ordering Acquisition of Frequency Domain Components Representing MR Image Data - A system orders acquisition of frequency domain components representing MR image data for storage in a storage array (e.g., k-space). A storage array of individual data elements stores corresponding individual frequency components comprising an MR dataset. The array of individual data elements has a designated center and individual data elements individually have a radius to the designated center. A magnetic field generator generates a magnetic field for use in acquiring multiple individual frequency components corresponding to individual data elements in the storage array. The individual frequency components are successively acquired in an order in which radius of respective corresponding individual data elements increases and decreases as the multiple individual frequency components are sequentially acquired during acquisition of an MR dataset representing an MR image. A storage processor stores individual frequency components acquired using the magnetic field in corresponding individual data elements in the array. | 08-12-2010 |
20100211539 | SYSTEM AND METHOD FOR BUILDING A DATA WAREHOUSE - There is disclosed a system and method of building a data warehouse. In an embodiment, the method comprises: receiving extract input parameters for extracting data from a source data file; building a communications channel to a selected one of a plurality of source database management systems containing the source data file in dependence upon the received extract input parameters; extracting data from the selected source data file and storing the extracted data into a staging table; receiving transform and load input parameters for transforming data stored in the staging table and loading the transformed data into a data warehouse table; and selecting pre-built modules for building the data warehouse with the data warehouse tables loaded with the transformed data in dependence upon the selected source database management system and the received extract, transform and load input parameters. In an embodiment, building the communications channel comprises dynamically creating an object linking and embedding (OLE) database server connection to the source database management system. | 08-19-2010 |
20100250484 | PROFILE SCANNER - Embodiments of the invention relate to systems, methods, and computer program products for identifying users of personal computers within a network for the purposes of electronic discovery. Embodiments herein disclosed provide for a tool that accesses the personal computers via the network, locates and queries the profile list on each personal computer, transmits for each personal computer computer-identifying data associated with the personal computer and user-identifying data associated with a user of the personal computer, and stores the computer-identifying data in a database so that it can be used in an electronic discovery system to collect data stored by the user on the computer. | 09-30-2010 |
20100250485 | System for Aggregating Data and a Method for Providing the Same - A system for aggregating data from a plurality of operational databases, and a method for providing the same, are provided. The system includes a data store storing a collection of pre-defined extraction scripts. The extraction scripts identify data available for extraction from a plurality of operational database products, and including parameters for specifying which of the data to extract. A data warehouse server executes the extraction scripts to extract, merge, transform and load the specified data from the plurality of operational databases into a multidimensional database. | 09-30-2010 |
20100274756 | MULTIDIMENSIONAL DATA ANALYSIS METHOD, MULTIDIMENSIONAL DATA ANALYSIS APPARATUS, AND PROGRAM - A highly-usable multidimensional data analysis method for performing interactive analysis on, for example, medical/administrative data stored in a hospital information system to support knowledge discovery about clinical decision-making is proposed. A multidimensional data analysis apparatus ( | 10-28-2010 |
20100280990 | ETL FOR PROCESS DATA WAREHOUSE - One embodiment is a method extract information technology (IT) events that indicate start and completion times of a business process. The method transforms the IT events into business data changes that are agnostic to multiple different ETL implementation languages and transforms the business data changes into execution data. Execution data is stored in a data warehouse. | 11-04-2010 |
20100280991 | METHOD AND SYSTEM FOR VERSIONING DATA WAREHOUSES - A method, system, and computer program product are disclosed. Exemplary embodiments of the method, system, and computer program product may include hardware, process steps, and computer program instructions for supporting versioning in a data warehouse. The data warehouse may include a data warehouse engine for creating a data warehouse including a fact table and temporary tables. Updated or new data records may be transferred into the data warehouse and bulk loaded into the temporary tables. The updated or new data records may be evaluated for attributes matching existing data records. A version number may be assigned to data records and data records may be marked as being the most current version. Updated and new data records may be bulk loaded from the temporary tables into the fact table when a version number or a version status is calculated. | 11-04-2010 |
20100293135 | HIGHCONCURRENCY QUERY OPERATOR AND METHOD - In one embodiment, a method includes concurrently executing a set of multiple queries, through a processor, to improve a resource usage within a data warehouse system. The method also includes permitting a group of users of the data warehouse system to simultaneously run a set of queries. In addition, the method includes applying a high-concurrency query operator to continuously optimize a large number of concurrent queries for a set of highly concurrent dynamic workloads. | 11-18-2010 |
20100299304 | MULTIMEDIA INSPECTION DATABASE SYSTEM (MIDAS) FOR DYNAMIC RUN-TIME DATA EVALUATION - An object-oriented system provides a comprehensive and integrated tool to manage, operate, analyze and inspect structures by providing 3-D visualization of the structures, a database interface, connectivity with one or more databases, a data display and analysis capabilities. The structural model is stored as an object-oriented, serialized file as a series of objects, including primitives, coordinates, object names, group names and other object identifiers. The system integrates various types of data, including tabulated textual data, annotated engineering drawings, photographic records, graphical plots, audio and videotaped records, from different tables and databases. Data are displayed to the user based on queries that the user submits to the system, which queries are evaluated at run-time. As a result, the user achieves unparalleled flexibility in data analysis and evaluation. | 11-25-2010 |
20100312747 | Computer Systems and Methods for Visualizing Data - A method for forming a visual plot using a hierarchical structure of a dataset. The dataset comprises a measure and a dimension. The dimension consists of a plurality of levels. The plurality of levels form a dimension hierarchy. The visual plot is constructed based on a specification. A first level from the plurality of levels is represented by a first component of the visual plot. A second level from the plurality of levels is represented by a second component of the visual plot. The dataset is queried to retrieve data in accordance with the specification. The data includes all or a portion of the dimension and all or a portion of the measure. The visual plot is populated with the retrieved data in accordance with the specification. | 12-09-2010 |
20100312748 | MULTIDIMENSIONAL CUBE FUNCTIONS - Cube functions may be used to obtain data from a multidimensional database. The cube functions may be contained within one or more cells of a spreadsheet. These cube functions behave similarly to the standard functions that may be included within a spreadsheet. Exemplary cube functions include obtaining: a cube member, a cube value, a cube set, a ranked member, a KPI, a member property and a count relating to a set. The cube functions within the spreadsheet may access the cube data from one or more multidimensional databases. Using the cube formulas in individual cells allows the user to add/delete rows and/or columns from within the spreadsheet. | 12-09-2010 |
20110010330 | Apparatus and Method for Read Optimized Bulk Data Storage - A computer readable storage medium includes executable instructions to store load data in new rows of an append only table within a data warehouse, without updating or deleting existing rows in the append only table. The new rows are added in a database transaction that guarantees that the distributed reads and writes of all transactions are processed reliably and that the transactions obey snapshot isolation rules. Data changes associated with the load data are maintained in a separate update table. Data from the append only table is merged with data changes from the update table to supply read data. | 01-13-2011 |
20110035353 | Computer-Implemented Multidimensional Database Processing Method And System - A computer-implemented system and method are provided to handle operations dealing with a multidimensional database. Information about the hierarchical structure of a multidimensional database is used when performing the operations upon the database. Operations such as update, select and others are performed using the hierarchical structural information. Keys and indexes may be used in accessing the hierarchical structural information. | 02-10-2011 |
20110035354 | METHOD AND SYSTEM FOR RECONCILING META-DATA IN A DATA WAREHOUSE - A data warehouse system comprising a warehouse database and a data warehouse manager communicatively coupled to the database. The warehouse manager is to store data to and retrieve data from the warehouse database and perform metadata reconciliation to update metadata of objects in the data warehouse to reflect a change in metadata of a first object in the warehouse by identifying a first set of objects impacted by the change to the metadata of the first object, identifying a second set of objects on which the first object depends, identifying at least one task that has to be completed to propagate the impact of the change in the metadata of the first object to objects in the first and second sets of objects, performing the task, and dynamically computing additional tasks based on whether or not definitions of objects in the first or second sets of objects changes based upon completion of the at least one task. | 02-10-2011 |
20110055147 | GENERATING EXTRACT, TRANSFORM, AND LOAD (ETL) JOBS FOR LOADING DATA INCREMENTALLY - Systems, methods and articles of manufacture are disclosed for generating an incremental ETL job from an instance of an abstract ETL operator model representing an exhaustive ETL job for loading data from a source schema to a target schema. In one embodiment, the abstract operator model may be extended to include incremental variants of operators of the abstract ETL operator model. From the instance of the abstract ETL operator model and from the extended ETL operator model, a corresponding instance of the extended ETL operator model may be generated. An incremental ETL job may be deployed from the instance of the extended ETL operator model. The incremental ETL job may be executed to load data from the source schema to the target schema. | 03-03-2011 |
20110055148 | SYSTEM AND METHOD FOR REDUCING ULTRASOUND INFORMATION STORAGE REQUIREMENTS - A system and method for storing ultrasound information are provided. The method includes storing raw medical data of a scanned object in a reference data file, storing a set of image generating parameters into a parameter-constrained file that is separate from the reference data file, linking the set of image generating parameters stored in the parameter-constrained file to the reference data file, and generating an image by applying the set of image generating parameters to the reference data file. | 03-03-2011 |
20110055149 | GENERATING QUERY PREDICATES FOR OLAP PROCESSING - Systems, methods and articles of manufacture are disclosed for generating a relational database query from an online analytical processing (OLAP) query against a cube. The cube may have dimensions, each dimension having a hierarchy of members. A level of the hierarchy of a dimension of the cube may be identified from which the OLAP query accesses a number of members exceeding a predefined threshold count of members. A relational database query may be generated that includes a predicate for at least one member from at least one level above the identified level in the hierarchy, instead of a plurality of predicates for members from the identified level. The relational database query may be generated from the OLAP query, from a definition for the hierarchy, and from the accessed members from the identified level. As a result, the relational database query may include fewer predicates and may be executed more efficiently. | 03-03-2011 |
20110060719 | Method for Transforming Setup Data in Business Applications - The disclosure describes a method for automating the transformation data for business applications. The method involves a transformation engine using a data map between a set of physical entities corresponding to business objects in a source database instance and a corresponding set of physical entities in a target database instance. The data transformation engine using the data map generates automatically scripts to extract data from the source database instance to a set of intermediate tables, transform the data in the set of intermediate tables, and load the transformed data to the target database instance. The data map may be generated based on one or more of foreign key information and constraint information. The method also considers the dependencies of various business objects on each other and the sequence in which the business objects should be transformed. | 03-10-2011 |
20110066590 | ANALYTICS INTEGRATION WORKBENCH WITHIN A COMPREHENSIVE FRAMEWORK FOR COMPOSING AND EXECUTING ANALYTICS APPLICATIONS IN BUSINESS LEVEL LANGUAGES - Systems, methods and articles of manufacture are disclosed for building and executing analytics solutions. Such a solution may provide a comprehensive analytics solution (e.g., a risk assessment, fraud detection solution, dynamic operational risk evaluations, regulatory compliance assessments, etc.). The analytics solution may perform an analytics task using operational data distributed across a variety of independently created and governed data repositories in different departments of an organization. A framework is disclosed which allows a user (e.g., a risk analyst) to compose analytical tools that can access data from a variety of sources (both internal and external to an enterprise) and perform a variety of analytic functions. | 03-17-2011 |
20110078108 | AGENTLESS DATA COLLECTION - Tools and techniques for collecting data from target systems without the need for installation of expensive and high-maintenance software agents on the target systems. In some cases, these tools can be used to collect performance data from collection targets while still allowing the operator(s) of the collection targets to govern the access of the tools, alleviating security concerns. Merely by way of example, certain tools employ data collection scripts, which may be analyzed by the collection target against security policies of that collection target, to ensure that the scripts will not take any actions that might violate such security policies. Other tools may be used to collect data in real time by transmitting a script for immediate execution by the host system, which could then return the data produced by the script immediately upon completion (and/or as part of execution of the script). | 03-31-2011 |
20110082830 | CHANGE APPLICATION FOR DATASETS USING FROZEN AND DELTA VALUES - A method includes storing a first value for one or more measures in a block of data for the cube of data. The first value is a value where the state of the cube is that no changes are applicable to the block of data. A second value is stored for the one or more measures in the block of data. The second value is a delta value representing changes made to the block of data based on the first value. A change is received to the block. A computer processor then calculates a new second value based on the change to the block. The new second value is stored in place of the second value previously stored while the first value remains the same. The second value represents the cumulative changes to the measure since the frozen value was determined. | 04-07-2011 |
20110099143 | EMBEDDING AND RETRIEVING DATA IN AN APPLICATION FILE FORMAT - Third party data embedding may be provided. Data associated with a third party provider may be saved in a native application save file. If a save file opened by the application is determined to comprise third party data, the application may determine whether an appropriate data provider is available. The application may connect to the data provider and provide on-demand, asynchronous access to the third party data upon request from a user. | 04-28-2011 |
20110106752 | System and method of extensible observation for computer network monitoring - Extensible observer system with report management, visual observer, and event management functionality. Uses a component loader for loading new functionality. Leverages data provider, data extractor, and state display components. Includes report manager, report generator, and report propagator. Displays visual observers that include states generated using state display components. Visual observers selectable through hierarchical selector. Manages data-driven events, generating event responses, such as message distribution, based on triggering criteria. Data extractor components used to extract state information. Component generator produces skeleton component code. | 05-05-2011 |
20110113005 | SUPPORTING SET-LEVEL SLICE AND DICE IN DATA WAREHOUSES - A method and system for coping with slice and dice operations in data warehouses is disclosed. An external approach may be utilized, creating queries using structured query language on a computer. An algorithm may be used to rewrite the queries. The resulting predicates may be joined to dimension tables corresponding to fact tables. An internal approach may be utilized, using aggregation functions with early aggregation for creating the queries. The results of the slice and dice operations may be outputted to a user on a computer monitor. | 05-12-2011 |
20110113006 | BUSINESS PROCESS CONTROL APPARATUS, BUSINESSES PROCESS CONTROL METHOD AND BUSINESS PROCESS CONTROL PROGRAM - A business process control apparatus, a business process control method and a business process control program are provided which can properly estimate a business process. The business process control apparatus includes: business process data storage means for storing a business process data group; copying means for copying business process data contained in said business process data group and adding new business process data to said business process data group; and estimation calculating means for calculating an estimation value indicating how much said business process data is effective. | 05-12-2011 |
20110119227 | System And Method For Graphically Distinguishing Levels Of A Multidimensional Database - In accordance with the teachings described herein, systems and methods are provided for graphically distinguishing levels from a multidimensional database. Levels from a multidimensional database are distinguished, such as by associating two or more of database's levels with a plurality of different visual indicators. | 05-19-2011 |
20110125705 | AUTO-GENERATION OF CODE FOR PERFORMING A TRANSFORM IN AN EXTRACT, TRANSFORM, AND LOAD PROCESS - A mapping is received and stored that maps elements of a data warehouse to types of a type system implemented by a data source. Program code is generated that performs a transform of data retrieved from a data source based on the mapping. Generation of the program code may include generating program code for performing a dimension transform based on the mapping, generating program code for performing a fact transform based on the mapping, and generating program code for performing an outrigger transform based on the mapping. The generated program code may then be executed to transform the data retrieved from the data source prior to loading into the data warehouse. | 05-26-2011 |
20110125706 | PROCESSOR AND METHOD CONFIGURED FOR EXECUTING DATA TRANSFER OR DATA ADJUSTMENT FUNCTIONS ON OLAP BASED DATA - A system and method for processing data between a source cell and a target cell in a OLAP database. The system comprises a component configured for selecting a source cell based on dimensions in the OLAP database and having a value derived from the OLAP database, and a component for selecting a target cell based on dimensions in the OLAP database. The system includes a component configured for receiving a modification amount and applying the modification amount to the source cell and a component configured for applying the modification amount to the target cell and generating a modified target cell value. The system further comprises a component configured for storing the modified source cell and/or the modified target cell in the OLAP database. According to an embodiment, the modification amount is applied to the target cell as a credit amount or as a debit amount. | 05-26-2011 |
20110125707 | EFFICIENT DATA BACKFLOW PROCESSING FOR DATA WAREHOUSE - Processing backflow data includes obtaining backflow data from a data warehouse, determining, according to a backflow rule for the backflow data, a respective destination table in a production system for each data item in the backflow data, and sending the data item according to the determined respective destination table in the production system. | 05-26-2011 |
20110131173 | COMPENSATING FOR UNBALANCED HIERARCHIES WHEN GENERATING OLAP QUERIES FROM REPORT SPECIFICATIONS - Systems, methods and articles of manufacture are disclosed for generating an online analytical processing (OLAP) query against a cube. The cube may have dimensions, each dimension having a hierarchy of members. A report specification may be received against a hierarchy of the cube. An OLAP query may be generated from the report specification. The OLAP query may include projected members not having any descendant at the detail level of the hierarchy, thus producing more accurate results. | 06-02-2011 |
20110167033 | ALLOCATING RESOURCES IN A DATA WAREHOUSE - Systems, methods and computer-readable storage media are provided for allocating resources in a data warehouse. A table of resources may be retrieved from the data warehouse, the table including a plurality of records, each record having data that describes a resource to be allocated. A first resource described in a first record of the table of resources may be reserved pursuant to a first rule of a first stage. The first resource may then be allocated into one or more records of a working allocation table pursuant to the first rule. | 07-07-2011 |
20110167034 | SYSTEM AND METHOD FOR METRIC BASED ALLOCATION OF COSTS - Systems, methods and computer-readable storage media are provided for automatically allocating resources based on metrics. A plurality of metric values related to a respective plurality of dimensions may be determined. A resource may be allocated to the plurality of dimensions based on the plurality of metric values. Other embodiments are described and claimed. | 07-07-2011 |
20110167035 | MULTIPLE-CLIENT CENTRALLY-HOSTED DATA WAREHOUSE AND TREND SYSTEM - A system for analyzing data for an industry may include a user that extracts source data from source systems, transforms the source data into key metrics data for a given time period, and loads the key metrics data into a file extract, a data warehouse including data tables that are customized for the industry, a web portal for accepting input of data within the file extract related to the industry, a transformation logic for mapping the data to data elements of one or more of the data tables, and a graphical dashboard that presents in graphical form information specific to the industry based on the data elements of one or more of the data tables. | 07-07-2011 |
20110184905 | METHOD OF STORING AND ANALYSING DATA PRODUCED FROM INTERACTIONS BETWEEN EXTERNAL AGENTS AND A SYSTEM - The present invention relates to a method of storing data records produced from monitoring interactions between external agents and a system. The method comprises defining specific interactions that can occur between the external agents and the system as potential events of interest. A chain of interactions occurring during respective interaction sessions between a respective external agent and the system are monitored and potential events of interest occurring in the said chain are determined. Data records from the monitored chain of interactions are produced, the respective data record including data identifying determined potential events of interest and data associated therewith. A profile identity, representative of the external agent, is assigned to each data record produced during a said interaction session. Data records of individual potential events of interest are stored in a way that is ordered according to the type of potential event of interest and data records of potential events of interest occurring during an interaction session are stored in a way that is ordered according to assigned profile identity. | 07-28-2011 |
20110191285 | MAP UPDATE DATA SUPPLY DEVICE AND METHOD - A map update data supply device and method includes an update map database of per section versions of an update data file, and a request update data extraction unit for extracting a request update section and an update data file. A safeguard update data extraction unit extracts a safeguard update section to safeguard a road network connection between adjacent sections. An integrated data generation unit integrates all versions of the update data file for each extracted request update section and generates a request update integrated data file. The integrated data generation unit integrates, per safeguard update section, versions of the update data file up to the update safeguard version for each extracted safeguard update section, and generates a safeguard update integrated data file. An integrated data supply unit supplies the generated request update integrated data file and the safeguard update integrated data file to a navigation device. | 08-04-2011 |
20110202497 | Systems and Methods for Performing Direct Reporting Access to Transaction Databases - In one embodiment, the present invention includes a computer-implemented method comprising receiving a query and generating first and second requests. The first request is sent to an analytic data store and includes a data transfer indicator representing a time period when the data record was received in the analytic data store. The second request is sent to a transactional model proxy and includes a monotonically changing value, such as a timestamp. The transactional model proxy reads data from a transactional data store based on the value in the request. Data from the analytic and transactional data stores may be aggregated and returned in response to the query. | 08-18-2011 |
20110208690 | VIEWING AN OLAP CUBE AS A VIRTUAL DATA SOURCE - Systems, methods and articles of manufacture are disclosed for generating an online analytical processing (OLAP) query. A report specification may be received against the OLAP system. The report specification may include relational constructs. An application may apply constructs in the OLAP query according to a mapping and in a specific order. Consequently, the semantics of the report specification may be preserved. | 08-25-2011 |
20110208691 | Accessing Large Collection Object Tables in a Database - The present disclosure provides a method and apparatus for accessing large object collection tables in a data warehouse, so that input-output complexities are reduced and the performance and responsiveness of the data warehouse are improved. In one aspect, a process may set up a new large object collection table by determining the object identification information of business activities occurring in a business period using the records in a business flow table. A sub-table from the original large object collection table may be generated based on the derived object identification information. The resulting sub-table may be incorporated into a new large object collection table that is partitioned according to business periods. | 08-25-2011 |
20110213751 | OPTIMIZING DATA CACHE WHEN APPLYING USER-BASED SECURITY - A secure caching system and caching method include receiving a user request for data, the request containing a security context, and searching a cache for the requested data based on the user request and the received security context. If the requested data is found in cache, returning the cached data in response to the user request. If the requested data is not found in cache, obtaining the requested data from a data source, storing the obtained data in the cache and associating the obtained data with the security context, and returning the requested data in response to the user request. The search for the requested data can include searching for a security list that has the security context as a key, the security list including an address in the cache of the requested data. | 09-01-2011 |
20110218959 | SEARCH ENGINE MARKETING ANALYZER - A method of processing search engine marketing (SEM) transaction data generated responsive to transactions associated with a search engine marketing campaign carried on the internet for an advertiser by at least one SEM channel, the method comprising: acquiring channel SEM transaction data generated by the at least one SEM channel responsive to the transactions; acquiring back office (BO) SEM transaction data generated by the advertiser responsive to the transactions; merging the SEM data and BO data associated with a same transaction into a merged record; storing the merged record in an OLAP database stored in a computer storage media storing computer executable instructions; and using the merged record in the OLAP database to analyze the transactions and generate an alert. | 09-08-2011 |
20110225116 | SYSTEMS AND METHODS FOR POLICY BASED EXECUTION OF TIME CRITICAL DATA WAREHOUSE TRIGGERS - Embodiments of the invention broadly contemplate systems, methods, apparatuses and program products that employ a policy based method of timely execution of triggers over a data warehouse. Embodiments of the invention utilize methods to convert a trigger condition into differential queries over data sources, to decide when to send queries to data sources, and determine to which data sources to send queries. | 09-15-2011 |
20110225117 | MANAGEMENT SYSTEM AND DATA ALLOCATION CONTROL METHOD FOR CONTROLLING ALLOCATION OF DATA IN STORAGE SYSTEM - A storage system includes a plurality of types of media and a controller, provides a plurality of logical volumes, and performs a reallocation process for migrating data specified in accordance with a reallocation policy from media in which this data is stored to media of different types from the media. A management system searches for a logical volume that conforms to a migration policy, which is a condition under which the reallocation process is considered ineffective, from a reallocation group, which is a group of one or more logical volumes subjected to a target of a reallocation process, and causes the storage system to transfer data in the logical volume that is discovered to outside of the reallocation group. | 09-15-2011 |
20110231360 | Persistent flow method to define transformation of metrics packages into a data store suitable for analysis by visualization - A persistent flow provides a contract for delivering certain measures in a format which can be interactively analyzed along certain dimensions. It defines how a large number of metrics packages may be transformed into one or more hypercubes within a datamart. In particular a Carrier IQ persistent flow defines the dimensions along which key performance indicators may be displayed interactively in at least one dashboard with analytic tool controls. A persistent flow is stateful to incrementally process metrics packages over multiple collection periods which are not correlated with the times the metrics are recorded at the device. A flow defines the measures to be derived from metrics and the attributes of the measures of interest in a study. A flow defines enrichments that may be determined by examining measures from apparently independent sources and uses reference files to decode status records. A persistent flow provides an up-to-date view in the datamart by being run on a regular schedule to combine the most recently received data with previous intermediate results, thereby improving performance and avoiding staleness. | 09-22-2011 |
20110231361 | CONSOLIDATED SECURITY APPLICATION DASHBOARD - A consolidated security application dashboard system is described wherein a plurality of endpoint systems include visibility agents that collect status and event attributes/metrics from a plurality of security applications and upload the information to datamarts on a backend server. The backend server aggregates and processed the security application attributes/metrics to enable configurable dashboards to present summary and detailed information to IT users about the security metrics relating to a group of endpoints. | 09-22-2011 |
20110238616 | SYSTEM AND METHOD FOR IMPROVING OUTCOMES IN ENTERPRISE LEVEL PROCESSES - A method and system for using a data warehouse to improve results of enterprise level processes are provided. The data warehouse typically includes industry-wide empirical data relating to corresponding operational practices, metrics, and outcomes. The method focuses on actual process results by taking a holistic, end-to-end view of the process in conjunction with using the data in the data warehouse to enable effective process improvements. | 09-29-2011 |
20110246415 | METHOD AND SYSTEM FOR VALIDATING DATA - A method and system for validating data. Warehouse data is generated by transforming source data via an ETL transformation model. A data cube is generated by transforming the warehouse data via an OLAP transformation model. A report dataset (MDS | 10-06-2011 |
20110264618 | SYSTEM AND METHOD FOR PROCESSING AND ANALYZING DIMENSION DATA - A processor is operated to combine a first row of a dimension table in a data warehouse with a second row in the dimension table. The result is a combined row that includes a row identification key for the first row and a row identification key for the second row. The row identification key for the first row joins the combined row to fact data from a prior time period. The second row corresponds to a current time period that is later than the prior time period. The processor is also operated to join at least a portion of the combined row to at least a portion of the corresponding row in a fact table associated with the dimension table. The fact data from the prior time period is included in the corresponding row in the fact table. | 10-27-2011 |
20110289046 | Systems and Methods for Archiving Business Objects - The invention relates to systems and methods for archiving business objects, particularly unstructured information from email and journaling agents and file systems. The systems and methods include a remotely installed archiving agent on a target server connected via VPN to a source system at a client location. The target servers are located at a remote Tier IV, SAS70 certified data center that is staffed by operators monitoring the archive process. The systems and methods include the movement of data from individual target servers to individual SQL servers to a data warehouse which provides support for a web portal. This web portal allows for enhanced search capabilities across different file types of archived business objects. | 11-24-2011 |
20110295793 | SYSTEM AND METHOD FOR PROVIDING DATA FLEXIBILITY IN A BUSINESS INTELLIGENCE SERVER USING AN ADMINISTRATION TOOL - An administration tool can be used to provide data flexibility in a business intelligence (BI) server that is associated with a data warehouse. The administration tool can display one or more data objects that are adapted to be imported from an application framework into a physical model maintained on the BI server. The administration tool can further map the one or more data objects in the physical model into a logical model, also maintained on the BI server, which corresponds to at least one target table in the data warehouse. Additionally, the administration tool can publish an extension input to an extender associated with the data warehouse based on the logical model, wherein the extender operates to generate one or more metadata extensions based on the extension input. | 12-01-2011 |
20110295794 | SYSTEM AND METHOD FOR SUPPORTING DATA WAREHOUSE METADATA EXTENSION USING AN EXTENDER - An extender associated with a data warehouse can support metadata extension for the data warehouse. The extender can create an ETL extension plan for extending the data warehouse. The ETL extension plan includes one or more ETL plan elements that indicate extensions on ETL metadata objects. The one or more ETL plan elements within the ETL extension plan can be ordered, and ETL metadata extension can be performed based on the ETL extension plan. | 12-01-2011 |
20110295795 | SYSTEM AND METHOD FOR ENABLING EXTRACT TRANSFORM AND LOAD PROCESSES IN A BUSINESS INTELLIGENCE SERVER - A business intelligence (BI) server maintains a plurality of metadata objects to support the extract, transform and load (ETL) processes. These metadata objects includes a transparent view object, which takes a joined set of source tables and represents a data shape of the joined set of source tables using a transformation, and a ETL mapping association object that maps the transformation contained in the transparent view object to a target table. The BI server can then orchestrate the movement of data from source systems into the target data warehouses in a source and target system agnostic way. | 12-01-2011 |
20110302130 | INTELLIGENT VIDEO INFORMATION RETRIEVAL APPARATUS AND METHOD CAPABLE OF MULTI-DIMENSIONAL VIDEO INDEXING AND RETRIEVAL - Provided is an intelligent video information retrieval apparatus capable of multi-dimensional video indexing and retrieval. The intelligent video information retrieval apparatus includes an event detector configured to detect pieces of event information from footage collected by a plurality of video capture devices, a data mart builder configured to generate a data cube using the detected pieces of event information, and capture time and capture location information related to the pieces of event information, and store and manage the generated data cube, and a video retriever configured to receive an event retrieval condition from a user to retrieve event information corresponding to the received event retrieval condition using the data cube, and output the retrieval result to the user. | 12-08-2011 |
20110313969 | UPDATING HISTORIC DATA AND REAL-TIME DATA IN REPORTS - Disclosed are methods and systems for updating a report with real-time data. The method and systems involve receiving a request to view both historic data and real-time data in the report, identifying one or more data objects associated with the request, identifying one or more ETL jobs associated with the identified one or more data objects, determining a delta ETL job associated with the identified one or more ETL jobs and a upload timestamp of a delta ETL job, generating a data warehouse query based on the upload timestamp of the delta ETL job, generating a real-time query based on the transformations in the one or more ETL jobs and the upload timestamp of the delta ETL job, executing queries of the data warehouse query and real-time query to obtain historic data and real-time data and updating the report with both real-time data and historic data. | 12-22-2011 |
20110320399 | ETL BUILDER - A method of building a staging database which examines the staging database metadata and constructs a schedule of operations to extract the data, transform it and load it into the staging database by using normalisation, translation blocks and filtering to create a business view of the data that can be seen using standard query languages. This is applied to products such as Navision in building an OLAP cube for use in business intelligence applications. | 12-29-2011 |
20120005152 | Merged Event Logs - A record of an event is received in a common format. The record includes a time of the event, a type of the event, and an identity of a device that created the record in the common format using an event creation module selected based on the type of the event. A collection of records including the received record is managed, and a portion of the collection of records is provided in response to a user request. The portion is configured to be sorted according to user-configurable criteria and includes records with a plurality of types. | 01-05-2012 |
20120005153 | CREATION OF A DATA STORE - A method for structuring a data store by analysing source data bases using the steps of relationship discovery, schema merging, hierarchy discovery, heuristic based attribute inclusion and optionally denormalising This is applied to products such as Navision in building an OLAP cube for use in business intelligence applications. Also disclosed is a security adapter to carry security settings from a source data base to an OLAP cube which includes creating a synthetic dimension in the OLAP cube which is a common trait related to all other dimensions in the cube and one role is created for each role in the source data base and users treated as members of those roles as defined in the source data base. | 01-05-2012 |
20120011096 | EFFICIENTLY UPDATING ROWS IN A DATA WAREHOUSE - An aspect of the present invention facilitates efficient updating of rows in a data warehouse. In one embodiment, a sequence of identifiers is allocated to rows of data elements received earlier, and then inserted along with the received data elements in a table according to the defined by the sequence of identifiers. On receiving data elements as updates to the rows having associated identifiers, the received data elements are sorted according to the order of the associated identifiers and updates of the rows are performed successively in the sorted order of identifiers. According to another aspect, the above features are performed when (target) fact/dimension tables in the data warehouse are sought to be updated based on transformations/computations performed on data in other (source) fact/dimension tables. | 01-12-2012 |
20120066174 | MULTIMEDIA INSPECTION DATABASE SYSTEM (MIDAS) FOR DYNAMIC RUN-TIME DATA EVALUATION - An object-oriented system provides a comprehensive and integrated tool to manage, operate, analyze and inspect structures by providing 3-D visualization of the structures, a database interface, connectivity with one or more databases, a data display and analysis capabilities. The structural model is stored as an object-oriented, serialized file as a series of objects, including primitives, coordinates, object names, group names and other object identifiers. The system integrates various types of data, including tabulated textual data, annotated engineering drawings, photographic records, graphical plots, audio and videotaped records, from different tables and databases. Data are displayed to the user based on queries that the user submits to the system, which queries are evaluated at run-time. As a result, the user achieves unparalleled flexibility in data analysis and evaluation. | 03-15-2012 |
20120072391 | APPARATUS AND METHOD FOR AN AUTOMATIC INFORMATION INTEGRATION FLOW OPTIMIZER - An apparatus and method provides automatic information integration flow optimization. The apparatus may include an input/output port connecting the information integration flow optimizer to extract-transform-load tools. The information integration flow optimizer includes a parser unit to create a tool-agnostic input file containing rich semantics, a converter to transform the tool-agnostic input file into an input DAG, and a QoX-driven optimizer applying one or more heuristic algorithms to the input DAG to develop an optimum information integration flow design based on the rich semantics. The method may include receiving a tool-specific input file representing a physical information integration flow, parsing and converting the tool-specific input file into an input DAG containing tool-agnostic rich semantics, and applying heuristic algorithms to the input DAG to develop an optimum information integration flow design based on the rich semantics. | 03-22-2012 |
20120089562 | Methods and Apparatus for Integrated Management of Structured Data From Various Sources and Having Various Formats - The invention provides advanced software architected to leverage the operational data volumes from information technology (IT), intelligent devices and machines supported and maintained by IT or end users for use by business groups or individuals in order to create low Total Cost of Operation (TCO) processes that enable data sources to be transferred, modified, and computations performed based on the needs of the end users operating independently or within the enterprise data assets to cost-effectively capture any data source and place that within the enterprise's standard data structures and processes. The invention provides the end user options with capability to re-structure, copy, and re-position data for ongoing corporate applications via a selection of options methodology that is supported by usage of the end user's versions of spreadsheet software. | 04-12-2012 |
20120089563 | DATA AGGREGATION MODULE SUPPORTING DYNAMIC QUERY RESPONSIVE AGGREGATION DURING THE SERVICING OF DATABASE QUERY REQUESTS PROVIDED BY ONE OR MORE CLIENT MACHINES - Improved method of and apparatus for aggregating data elements in multidimensional databases (MDDB). In the preferred embodiment, the apparatus is realized in the form of a high-performance stand-alone (i.e. external) aggregation server which can be plugged-into conventional MOLAP systems to achieve significant improvements in system performance. In accordance with the principles of the present invention, the stand-alone aggregation server contains a scalable MDDB and a high-performance aggregation engine that are integrated into the modular architecture of the aggregation server. The stand-alone aggregation server of the present invention can uniformly distribute data elements among a plurality of processors, for balanced loading and processing, and therefore is highly scalable. | 04-12-2012 |
20120089564 | DATA AGGREGATION MODULE SUPPORTING DYNAMIC QUERY RESPONSIVE AGGREGATION DURING THE SERVICING OF DATABASE QUERY REQUESTS PROVIDED BY ONE OR MORE CLIENT MACHINES - Improved method of and apparatus for aggregating data elements in multidimensional databases (MDDB). In the preferred embodiment, the apparatus is realized in the form of a high-performance stand-alone (i.e. external) aggregation server which can be plugged-into conventional MOLAP systems to achieve significant improvements in system performance. In accordance with the principles of the present invention, the stand-alone aggregation server contains a scalable MDDB and a high-performance aggregation engine that are integrated into the modular architecture of the aggregation server. The stand-alone aggregation server of the present invention can uniformly distribute data elements among a plurality of processors, for balanced loading and processing, and therefore is highly scalable. | 04-12-2012 |
20120095957 | Component Based Approach to Building Data Integration Tools - A method facilitating interoperability across the integrating platforms is provided by composing purpose specific integrating tools using conceptual modeling framework and a set of primitive foundational components. The model driven architecture provides easy migration from one solution architecture to another. Also, the present invention discloses a method to migrate from one implementation to another using set of composed purpose specific integration tools. | 04-19-2012 |
20120101978 | SYSTEM AND METHOD FOR GENERATING AN INFORMATION INTEGRATION FLOW DESIGN USING HYPERCUBES - A system, method, and computer readable medium for generating an information integration flow design (IIFD). The system includes a processor to receive a conceptual model of the IIFD, having an extract phase, a load phase, and a transformation phase, an extract unit to model an interface between a data source information object and a transformation function based on at least one extract hypercube, a load unit to specify at least one load hypercube and a data warehouse target object, a transformation unit to express one or more steps as a hypercube operation, and a translation unit to generate the IIFD based on the conceptual model. The method includes receiving a conceptual model of the IIFD having an extract phase, a load phase, and a transformation phase. The method generates logical information integration operations based on the conceptual model. A computer readable medium may include instructions to generate the IIFD. | 04-26-2012 |
20120101979 | SYSTEMS AND METHODS FOR EXTRACTING INFORMATION FROM STRUCTURED DOCUMENTS - Systems and methods for extracting information from structured documents are provided. The systems and methods relate to selecting a centroid document from a group of structured documents, selecting a subset of the group of structured documents in order to form a cluster of the subset of documents about the centroid document. The selecting the subset is preferably based on the relative similarity between each of the selected subset and the centroid document. Then, systems and methods according to the invention include marking a data element on the centroid document. The systems and elements also include identifying a data element on each of the subset of documents, the data element that corresponds to the marked data element on the centroid document. Finally, data may be extracted from the subset of documents based on the identifying step. | 04-26-2012 |
20120109878 | DEBUGGING SYSTEM FOR MULTIDIMENSIONAL DATABASE QUERY EXPRESSIONS ON A PROCESSING SERVER - The present invention relates to data processing systems and methods to facilitate debugging, development and execution of multidimensional database queries, such as MDX (Multi-Dimensional eXpressions), for a multidimensional database structure on a computer architecture running on a server, such as an OnLine Analytical Processing Server (OLAP). The expression or query can be broken down and executed incrementally and associated with contextual information which can be viewed by an end-user via an understandable user interface. The advantage is the drastic simplification of very complex computations, giving the user a perspective not available till now. This allows users to better understand multidimensional queries, find errors that occur in the execution and correct such errors. | 05-03-2012 |
20120109879 | METHOD FOR SCHEDULING A TASK IN A DATA WAREHOUSE - Provided is a method and system for scheduling a task in a data warehouse. A task in a data warehouse is scheduled based on selection of a report related to the data warehouse. | 05-03-2012 |
20120109880 | USING ORGANIZATIONAL AWARENESS IN LOCATING BUSINESS INTELLIGENCE - An illustrative embodiment of a computer-implemented method for using organizational awareness in locating business intelligence receives an identity of an individual in an organizational hierarchy of users to form an identified individual and identifies people related to the identified individual in the organizational hierarchy of users using a people information database and relationship criteria to form related people. The computer-implemented method further identifies documents associated with the related people to form identified documents, inspects gathered information of the identified documents using a subset of relationship criteria to form inspected information and creates a list of suggested documents based at least on the inspected information. | 05-03-2012 |
20120117022 | METHOD AND SYSTEM FOR VERIFYING DATA ACCURACY - Presented is a method and system for verifying data accuracy. The method includes defining various steps of a data flow process, specifying data structure details at each step of the data flow process by metadata, providing at least one column dedicated for auditing in every table of a data warehouse involved in the data flow process, logging details related to a task execution at each step of the data flow process in the at least one column dedicated for auditing, storing the details related to a task execution along with data structure details and providing a visual representation of the stored details. | 05-10-2012 |
20120123996 | METHOD AND SYSTEM FOR REPLACING DATA IN A STRUCTURED DESIGN TEMPLATE - A method and system for replacing data in a design template. An end-user may provide a dynamic, structured client data set to be adapted to a predetermined, structured design template (e.g., a map or report). A repository of design templates may be searched to discover a list of candidate templates that may be capable of matching the client data. The end-user may select a design template that may be capable of showing the client data. The server may match the client data to a data structure for a candidate design template by using a data replacement approach, such as a column/alias approach. The end-user may potentially enlarge the matching set by providing information to help the matching algorithm; for example, providing column aliases where the matching algorithm may match template data column and column alias names with those specified for the client dataset. | 05-17-2012 |
20120130942 | OLAP Execution Model Using Relational Operations - In one embodiment the present invention includes an OLAP execution model using relational operations. In one embodiment, the present invention includes, a method comprising receiving a first query in an online analytic processor (OLAP) executing on one or more computers, the OLAP generating and comprising a model specifying a graph defining a plurality of nodes and a plurality of tiers, each node corresponding to a different operation on data. A second query is generated by the OLAP. The second query includes a plurality of layered subqueries each corresponding to one of the nodes in the graph for specifying the different operations on data. The second query is received in a relational engine coupled to the datastore. The relational engine executes the second query, and in accordance therewith, retrieves data. | 05-24-2012 |
20120150792 | DATA EXTRACTION FRAMEWORK - The present disclosure involves systems, software, and computer implemented methods for providing a data extraction framework for extracting data and metadata from an application to provide additional functionality for the extracted data and metadata. One process includes operations for identifying a first application for data extraction and determining a set of data suitable for extraction from the first application using a software development kit associated with the first application. The set of data is stored in a repository without storing visualization components of the first application in the repository. The set of data is sent to a second application for further processing of the set of data. The second application is configured to bind different visualization components to the set of data for display of data elements in the set of data to a user. | 06-14-2012 |
20120150793 | SYSTEMS AND METHODS FOR FILE LOADING - The application describes systems and methods for preventing the distribution of large volumes of electronic data by loading selected sensitive files with pad data to increase the size of the files such that file transfer, distribution, or downloading to removable media storage devices is more cumbersome. In one aspect, a file loading system comprises a datastore for storing a plurality of data files where each of the plurality of data files includes information and a processor arranged to: access the plurality of data files in the datastore, and load a data pad into one or more of the plurality of data files to increase the size of the one or more of the plurality of data files. | 06-14-2012 |
20120158644 | DATA FEED HAVING CUSTOMIZABLE ANALYTIC AND VISUAL BEHAVIOR - A data marketplace infrastructure allows a data source company or a third party to associate to a data feed with a customizable preview application having analytic and visual features (e.g., business logic and a user interface). A potential customer can use discovery services of the data marketplace infrastructure to identify one or more data feeds offering sample data in a domain of interest of the potential customer. The data feed can transmit the customizable preview application to allow the potential customer to interact with the data from the data feed before purchasing the data. The potential customer may also customize the preview application so as to change the default application behavior in a manner that improves the potential customer's understanding and appreciation of the data feed. | 06-21-2012 |
20120166384 | SYSTEMS AND METHODS TO PROVIDE SERVER-SIDE CLIENT BASED CACHING - According to some embodiments, a system, method, means, and/or computer program code are provided to facilitate a display of information on a client device. For example, a server may retrieve first enterprise data from an enterprise database and store the first enterprise data into a first client based cache at the server, the first client based cache being associated with a first user. Similarly, the server may retrieve second enterprise data from the enterprise database and store the second enterprise data into a second client based cache at the server, the second client based cache being associated with a second user. Subsequent to the storing of the first enterprise data, the server may receive a display request from a first client device associated with the first user and transmit the first enterprise data to the first client device. | 06-28-2012 |
20120166385 | DATA LOADING METHOD FOR A DATA WAREHOUSE - Provided is a method of loading data in a data warehouse. Two types of work tables are generated, namely, a first work table and a second work table. The data is loaded into the first work table and prepared for loading, using the second work table as temporary table for data preparation. The prepared data is loaded into the data warehouse. | 06-28-2012 |
20120173477 | PREDICTIVE RESOURCE MANAGEMENT - Apparatus, systems, and methods may operate to monitor database system resource consumption over various time periods, in conjunction with scheduled data loading, data export, and query operations. Additional activities may include generating a database system resource consumption map based on the monitoring, and adjusting database system workload throttling to accommodate predicted database system resource consumption based on the resource consumption map and current system loading, prior to the current database resource consumption reaching a predefined critical consumption level. The current system loading may be induced by data loading, data export, or query activity. Other apparatus, systems, and methods are disclosed. | 07-05-2012 |
20120173478 | CUSTOM DATA MART CREATION - Systems, methods, and computer-readable media for creating custom data marts are provided. A selection of at least one table from a data source stored within a data warehouse is received. A selection of at least one column from the at least one table is received. Joins are created based on the selections, metadata is created and stored, and physical schema is generated to store the data mart data. The custom data mart fact table is populated with data from the selected columns. Facilities are provided for the end-user to create an abstraction layer for business intelligence reporting needs. | 07-05-2012 |
20120173479 | APPLICATION CACHE PROFILER - In an embodiment of the invention, a method for data profiling incorporating an enterprise service bus (ESB) coupling the target and source systems following an extraction, transformation, and loading (ETL) process for a target system and a source system is provided. The method includes receiving baseline data profiling results obtained during ETL from a source application to a target application, caching the updates, determining current data profiling results within the ESB for cached updates, and triggering an action if a threshold disparity is detected upon the current data profiling results and the baseline data profiling results. | 07-05-2012 |
20120179643 | Hyper-Lattice Model for Optimized Sequencing of Online Analytical Processing (OLAP) Operations on Data Warehouses - Systems and methods are provided for Online Analytical Processing of data included within data warehouses. According to one example, a method for adding a dimension to a hyper-lattice structure is provided and includes adding a new base cuboid to a hyper-lattice that describes functionality of the data warehouses. According to another example, a method for determining an optimal path through an existing hyper-lattice by which to generate a destination cuboid from a source cuboid is also provided. | 07-12-2012 |
20120191641 | CHARACTERIZING BUSINESS INTELLIGENCE WORKLOADS - One or more embodiments characterize workloads in a data warehouse system. A set of accounting data associated with a data warehouse system comprising at least one database is collected. A set of query concurrency information associated with the database is determined determining based on the set of accounting data. The set of query concurrency information identifies a plurality of queries executed on the database simultaneously and a duration of this execution. A set of processor utilization distribution information associated with the plurality of queries is determined based on the set of accounting data. A set of data access information indicating a quantity of data accessed by each query in the plurality of queries is determined. A workload profile associated with the database is generated based on the set of query concurrency information, the set of processor utilization distribution information, and the set of data access information. | 07-26-2012 |
20120191642 | USER DEFINED FUNCTION CLASSIFICATION IN ANALYTICAL DATA PROCESSING SYSTEMS - Systems and apparatuses are provided for integrating user defined functions into an analytical data processing framework. The system includes a plurality of user defined functions ( | 07-26-2012 |
20120203734 | AUTOMATIC MAPPING OF A LOCATION IDENTIFIER PATTERN OF AN OBJECT TO A SEMANTIC TYPE USING OBJECT METADATA - Systems and methods for automatic mapping of a location identifier pattern of an object to a semantic type using object metadata are disclosed. In one aspect, embodiments of the present disclosure include a method, which may be implemented on a system, of identifying a set of tags associated with a website that is hosted by a web server. The method further includes, detecting a web page in the website in which a tag of the set of tags is identified, extracting a pattern from a Universal Resource Locator (URL) of the web page, and/or storing the pattern in a database embodied in a machine-readable storage medium as being mapped to the semantic type. The tag corresponds to a semantic type with which the content embodied in the web page has a semantic relationship and the pattern corresponds to the semantic type with which the content embodied in the web page has a semantic relationship. | 08-09-2012 |
20120215737 | CENTRAL REPOSITORY FOR SEARCHES - Search systems and methods are disclosed. Content and data in support of search results are stored in a central repository. Moreover, the centrally stored content and data can be made available to one or more authorized devices. Searches can be conducted across a single device, or a community of devices. The community of devices over which a search is performed can include devices that are not publicly accessible, and that require the provision of authentication information. The community of devices over which a search is performed, and the set of authorized devices, can be determined based on various factors. | 08-23-2012 |
20120221508 | SYSTEMS AND METHODS FOR EFFICIENT DEVELOPMENT OF A RULE-BASED SYSTEM USING CROWD-SOURCING - Described herein are methods, systems, apparatuses and products for efficient development of a rule-based system. An aspect provides a method including accessing data records; converting said data records to an intermediate form; utilizing intermediate forms to compute similarity scores for said data records; and selecting as an example to be provided for rule making at least one record of said data records having a maximum dissimilarity score indicative of dissimilarity to already considered examples. | 08-30-2012 |
20120221509 | DETERMINING A DATA MAPPING RELATIONSHIP BETWEEN DATABASE TABLES - A method and apparatus for determining a data mapping relationship between a source database table and a target database table are included. The method includes obtaining attribute values of an attribute other than a primary key and corresponding primary key value sets from plural rows of data in a source database table, and obtaining attribute values of a specific attribute other than a corresponding primary key and corresponding primary key value sets from plural rows of data in the target database table. A determination is made as to whether the attribute of the source database table and the specific attribute of the target database table have a potential data mapping relationship. If the determination is affirmative, a data mapping relationship is determined therebetween. | 08-30-2012 |
20120221510 | METHOD AND SYSTEM FOR VALIDATING DATA - A method and system for validating data. Warehouse data is generated by transforming source data via an ETL transformation model. A data cube is generated by transforming the warehouse data via an OLAP transformation model. A report dataset (MDS | 08-30-2012 |
20120221511 | SYSTEM AND METHOD FOR ANALYZING DATA IN A REPORT - A method and a system for analyzing a data value of interest in a multidimensional database. The data value of interest is first identified in a structured report. The data source for the structured report is retrieved. A context for the data value of interest in the data source is collected, for example, by extracting a slicer from the structured report, by extracting members of master-detail pages, or by extracting a default measure from the structured report. An analysis specification based on the collected context is then built, and the data value of interest is analyzed using the analysis specification. The result from analyzing the data value of interest may be presented in an analysis view. | 08-30-2012 |
20120226657 | GENERATING BPEL CONTROL FLOWS - Generating Business Process Execution Language (‘BPEL’) control flows to populate a data warehouse for analysis of data of an operational system including identifying source tables in the operational system and one or more corresponding target tables in the data warehouse; creating BPEL steps to represent links between each source table in the operational system and each corresponding target table in the data warehouse; determining a sequence for the BPEL step; and adding the BPEL steps to a BPEL control flow according to the determined sequence. | 09-06-2012 |
20120233113 | METHODS, SYSTEMS, AND COMPUTER PROGRAM PRODUCTS FOR IMPLEMENTING A STANDARDIZED INTERPRETIVE ENGINE - Implementing a standardized interpretive engine includes externalizing a business requirement into a component class of a data control structure, defining a solution to the business requirement, entering the solution into metadata within the component class, and pulling data from a data warehouse into the component class based upon the business requirement. Pulling the data includes aggregating sub-tending data from the data warehouse. The aggregation is specified to a level requested via the component class. Implementing the standardized interpretive engine also includes performing a calculation on the data in accordance with the business requirement. The data is pulled away from the data warehouse and forward to the component class, such that a single data source associated with the data is usable multiple times for a plurality of component classes. | 09-13-2012 |
20120233114 | MONITORING ORGANIZATIONAL INFORMATION FOR FAST DECISION MAKING - An organizational monitoring system is described herein that continuously monitors existing data sources to identify select data to organizational decision makers more quickly. The system may pull information from existing databases and other data sources within an organization, and monitor the information for relevant changes. The system analyses the data to determine events and trends that potentially merit attention of organization leaders. The system provides a rapid organizational information dashboard that decision makers can view on a display device to continuously monitor the organization. Organizational leaders can use the information to make faster business decisions and to get on top of trends that affect the business at an early time when action can be much more effective. Thus, the organizational monitoring system allows faster response to trends and events and more effective management of organizations. | 09-13-2012 |
20120239610 | XBRL DATABASE MAPPING SYSTEM AND METHOD - XBRL instance data configured according to a given XBRL hypercube specification (or an extension thereof) may be automatically mapped to a multidimensional database configured according to the given XBRL hypercube specification. Such a multidimensional database may be configured for analytical processing of multi-dimensional analytical queries related to the XBRL instance data and/or for analytical processing by a business intelligence tool. Multidimensional data from a multidimensional database may also be automatically mapped to an automatically-generated XBRL hypercube specification structured according to the multidimensional database. | 09-20-2012 |
20120239611 | XBRL FLAT TABLE MAPPING SYSTEM AND METHOD - XBRL data may be automatically mapped back and forth between an XBRL instance an set of automatically generated flat tables, where each table represents the projection of a single hypercube. | 09-20-2012 |
20120239612 | USER DEFINED FUNCTIONS FOR DATA LOADING - Data loading with user defined functions is described in various implementations. An example system for data loading may include a structured query language (SQL) compiler to identify a call to a table valued user defined function (TVUDF) within a SQL statement that includes an insert statement; identify metadata associated with the TVUDF; validate and resolve a subclass type of the TVUDF based on the metadata and the insert statement; and generate a data loading plan to retrieve and load data from an external data source into a table of a database based on the subclass type of the TVUDF. The system may also include a data loading engine in the database to execute the data loading plan, the data loading plan including the TVUDF to retrieve data from the external data source, and load the retrieved data into the table of the database in accordance with the data loading plan. | 09-20-2012 |
20120246110 | Master Data Management in a Data Warehouse/Data Mart - A data warehouse incorporates processing for creating, managing, and otherwise maintaining master data. The data warehouse may communicate with a master data manager to obtain services for handling master data. A surrogate master data identifier may be defined by the data warehouse to reference the master data, thereby decoupling any modifications of the master data identifier that may be made by the master data manager. The data warehouse may export the master data to an application system, and conversely import master data from an application system. | 09-27-2012 |
20120254103 | EXTRACT, TRANSFORM AND LOAD USING METADATA - The extract and transformation operation performed on multiple data sources. In the extract operation, a computing system periodically performs an extraction operation on the data sources, perhaps as defined by metadata. The transformation operation is performed on at least some the extracted data, and is periodically performed, perhaps also in a manner that is defined by metadata. When performing each iteration of the transformation operation, the computing system performs a transformation operation on valid data, marks other data as suspect (due to a dependency on data that is not present), and completes transformation on yet other data that had previously been marked as suspect or perhaps reaffirms its suspect status if the needed data is still not available. | 10-04-2012 |
20120265726 | AUTOMATED DATA WAREHOUSE MIGRATION - Systems, apparatus, computer-readable storage media, and methods are disclosed for allowing analysis, migration, and validation of data from a source environment (such as an RDBMS system) to a target environment (such as a data warehouse (DW) appliance). In one example, a method comprises analyzing a source database, a source ETL environment, a target database, and a target ETL environment to produce configuration data, the configuration data being used for generating a mapping of the source database to a target database in the target database environment, a mapping of the source DDL code to target DDL code in the target database environment, and a mapping of source ETL code to target ETL code for the target database environment, and migrating at least one table from the source database, at least a portion of the source DDL code, and at least a portion of the source ETL code to the target database environment, where the migrating is based at least in part on the mapping generated using the configuration data. | 10-18-2012 |
20120265727 | DECLARATIVE AND UNIFIED DATA TRANSITION - Described are computer-based methods and apparatuses, including computer program products, for declarative and unified data transition. In some embodiments, a computer implemented method includes determining an unified configuration for a knowledge domain. The unified configuration can include one or more predicates for one or more system objects, and/or one or more relationships between the one or more system objects. The method can include generating one or more transformation rules based on the one or more predicates. The one or more transformations can enable transformation of input data. The input data can include information associated with the one or more system objects and/or the one or more relationships between the one or more system objects. The method can include generating one or more reconciliation rules based on the one or more predicates. The one or more reconciliation rules can be associated with a destination database and can enable reconciliation of the transformed data with the destination database. | 10-18-2012 |
20120290527 | DATA EXTRACTION AND TESTING METHOD AND SYSTEM - The present method and apparatus provides for automated testing of data integration and business intelligence projects using Extract, Load and Validate (ELV) architecture. The method and computer program product provides a testing framework that automates the querying, extraction and loading of test data into a test result database from plurality of data sources and application interfaces using source specific adaptors. The test data available for extraction using the adaptors include metadata such as the database query generated by the OLAP Tools that are critical to validate the changes in business intelligence systems. A validation module helps define validation rules for verifying the test data loaded into the test result database. The validation module further provides a framework for comparing the test data with previously archived test data as well as benchmark test data. | 11-15-2012 |
20120296861 | STORING EVENTS FROM A DATASTREAM - A data stream is stored in a database. An event pipe stores a sliding window of events from the data stream. A query is responded to by utilizing the sliding window of events stored in the event pipe. | 11-22-2012 |
20120296862 | METHOD AND APPARATUS FOR ANALYZING AND MIGRATING DATA INTEGRATION APPLICATIONS - An automatic data integration process analysis and conversion is disclosed. The input may be formatted as Extract Load Transform (ELT) or Extract Transform Load (ETL), or any other type of input. The processes may then be analyzed and different outputs generated in a format requested by the user. | 11-22-2012 |
20120296863 | INTERACTIVE REAL ESTATE CONTRACT AND NEGOTIATION TOOL - Term negotiation can utilize centralized systems accessed via web interfaces for purposes such as mediation of communications between buyers and sellers, maintenance of a history of negotiations, and notification of parties regarding changes suggested during negotiation. Changes to terms proposed by parties using centralized systems can be stored in a data warehouse, potentially along with timestamp and identification information. | 11-22-2012 |
20120310875 | METHOD AND SYSTEM OF GENERATING A DATA LINEAGE REPOSITORY WITH LINEAGE VISIBILITY, SNAPSHOT COMPARISON AND VERSION CONTROL IN A CLOUD-COMPUTING PLATFORM - In one exemplary embodiment, a computer-implemented method of a database management system including the step of obtaining a metadata about a data from a metadata source. The metadata is converted to an extensible markup language (XML). XML variant or text formatted file. The formatted file is uploaded to a central repository. The formatted file is parsed to acquire information about the data. A data structure that includes the information about the data is generated. The data structure can be stored in a database cluster resident in a cloud computing platform. The metadata source can be an extract, transform and load (ETL) server or a data warehouse server. A dashboard visualization of the data lineage information can be rendered for display with a graphical user interface. | 12-06-2012 |
20120310876 | Content transfer - The present invention provides a method of transferring content from a file and a database. In this case, the file includes content instances, each content instance being associated with a respective field, and each field having a respective type. The transfer is achieved by determining the type of each field, and then storing each content instance in a store in accordance with the determined field type of the associated field. Each content instance can then be transferred to the database in accordance with the determined field type. A similar procedure is provided for creating XML files based on content within the database. | 12-06-2012 |
20120317071 | Content transfer - The present invention provides a method of transferring content from a file and a database. In this case, the file includes content instances, each content instance being associated with a respective field, and each field having a respective type. The transfer is achieved by determining the type of each field, and then storing each content instance in a store in accordance with the determined field type of the associated field. Each content instance can then be transferred to the database in accordance with the determined field type. A similar procedure is provided for creating XML files based on content within the database. | 12-13-2012 |
20120323840 | DATA FLOW COST MODELING - Techniques are disclosed for cost modeling in a data flow model. In one embodiment, a data flow model is provided that includes multiple data flow objects, each data flow object including one or more data units. Each data flow object represents a respective information asset. Quality and cost components are assigned to each data unit of a first data flow object. A cost property is determined for each data unit of the first data flow object, based on the assigned components. An object cost property is also determined for the first data flow object. The determined cost properties may be used to identify information assets that are more cost-effective to upgrade, relative to other information assets. | 12-20-2012 |
20120323841 | DATA INTEGRATION APPARATUS, DATA INTEGRATION METHOD, AND COMPUTER PRODUCT - When collecting and integrating data present in separately managed information sources, data are collected from the information sources through a physical model of the information sources. Integration is executed by a process of converting a data configuration (mapping) into a logical model predefined for each utilization-side application and a process of converting and conforming values (cleansing); and the result thereof is provided as a view (logical model) for each application to the utilization-side application. | 12-20-2012 |
20130013553 | Automated Entity Verification - Some embodiments provide a verification system for automated verification of entities. The verification system automatedly verifies entities using a two part verification campaign. One part verifies that the entity is the true owner of the entity account to be verified. This verification step involves (1) the entity receiving a verification code at the entity account and returning the verification code to the verification system, (2) the entity associating an account that it has registered at a service provider to an account that the verification system has registered at the service provider, (3) both. Another part verifies the entity can respond to communications that are sent to methods of contact that have been previously verified as belonging to the entity. The verification system submits a first communication with a code using a verified method of contact. The verification system then monitors for a second communication to be returned with the code. | 01-10-2013 |
20130013554 | COMPENSATING FOR UNBALANCED HIERARCHIES WHEN GENERATING OLAP QUERIES FROM REPORT SPECIFICATIONS - Systems, methods and articles of manufacture are disclosed for generating an online analytical processing (OLAP) query against a cube. The cube may have dimensions, each dimension having a hierarchy of members. A report specification may be received against a hierarchy of the cube. An OLAP query may be generated from the report specification. The OLAP query may include projected members not having any descendant at the detail level of the hierarchy, thus producing more accurate results. | 01-10-2013 |
20130018840 | Enhanced usage of business intelligence systemsAANM Averbuch; JakobAACI NetanyaAACO ILAAGP Averbuch; Jakob Netanya IL - A collaborative Business Intelligence (BI) system comprising pages based on said BI system data, comprising a collaboration platform having means for selecting most relevant users and most relevant pages and means for enabling collaboration between users, | 01-17-2013 |
20130018841 | ENABLING COMPUTATIONAL PROCESS AS A DYNAMIC DATA SOURCE FOR BI REPORTING SYSTEMS - A system, method and computer program product for integrating at least one analytical or modeling driver into an existing business analytic environment. The system comprises a BAMS (Business Analytics and Mathematical Sciences) driver, implemented on the memory device and run by the processor, loads a real database driver and redirects a request from a client application to the real database driver. An extension driver, invoked by the BAMS driver, provides an interface to a data modeling or processing software and converts the interface from a software environment compiled by a programming language to another software environment targeted to another programming language. The extension driver extends or overrides functionalities of the BAMS driver. Implementing the BAMS driver and the extension driver in the existing business analytic environment do not require any change made on the existing business analytic environment. | 01-17-2013 |
20130054511 | DATA ENRICHMENT USING HETEROGENEOUS SOURCES - A data enrichment system may include an attribute relevance module to measure relevance of an attribute to a data object to be enriched. The data object may include the attribute including a known or an unknown value. An output value confidence module may calculate a confidence of an output value of a source used for enrichment of the data object. The output value may represent the known and/or unknown values of the attribute. The system may use the measured relevance of the attribute and the calculated confidence of the output value to determine assignment of the known or unknown values to the attribute. | 02-28-2013 |
20130054512 | SYSTEM AND METHOD FOR TEXT EXTRACTION AND CONTEXTUAL DECISION SUPPORT - A contextual analysis system that extracts data elements from an unstructured text input; determines whether the extracted data elements are relevant to a predetermined context; and determines, for the extracted elements deemed as relevant, whether the information contained in the relevant data elements complies with a guideline. | 02-28-2013 |
20130066826 | ADAPTIVE DATA MODEL AND WAREHOUSE PALETTE - Systems, methods, and other embodiments associated with and\ adaptive data model and warehouse palette are provided. In one embodiment, a method includes providing a user interface for collecting item definition attributes from a user. A database structure is generated to store item data according to collected the item definition attributes. An extract, transform, and load (ETL) layer is generated to extract item data from user data, transform the extracted data for storing in the database structure, and load the transformed extracted data into the database structure. | 03-14-2013 |
20130066827 | INFORMATION PROCESSING APPARATUS AND ELEMENT EXTRACTION METHOD - In an information processing apparatus, a first deletion unit deletes, from sets stored in a storage unit, sets having less than a threshold number of included elements and elements existing in only sets whose number is less than a threshold number of occurrences. A grouping unit generates a group of sets associated with each other in terms of commonality of elements out of the sets remaining in the storage unit. A second deletion unit deletes, with respect to each generated group, sets having less than the threshold number of included elements and elements existing in only sets whose number is less than the threshold number of occurrences, from the sets belonging to the group. An output unit outputs a list of elements included in the sets remaining in each group where there are no sets or elements that need to be deleted. | 03-14-2013 |
20130073514 | FLEXIBLE AND SCALABLE STRUCTURED WEB DATA EXTRACTION - This document describes techniques that label text nodes of a seed site for each of a plurality of verticals. Once a seed site is labeled for a given vertical, the techniques extract features from the labeled text nodes of the seed site. The techniques learn vertical knowledge for the seed site based on the human labels and the extracted features, and adapt the learned vertical knowledge to a new web site to automatically and accurately identify attributes and extract attribute values targeted within a given vertical for structured web data extraction. | 03-21-2013 |
20130073515 | COLUMN BASED DATA TRANSFER IN EXTRACT TRANSFORM AND LOAD (ETL) SYSTEMS - Executing a plurality of transform stages in an extract, transform and load (ETL) job including, for each of the transform stages, receiving a plurality of input row identifiers (RIDs) corresponding to a first plurality of source database table rows in a source database table. Data is retrieved directly from a subset of the source database table columns in the first plurality of source database table rows based on the input RIDs and transform logic. Partial row data including data from the subset of the source database table columns is generated for each of the first plurality of source database table rows. Transformed data is generated based on the partial row data and to the transform logic. Output RIDs corresponding to a second plurality of rows in the source database table that include a least a subset of the transformed data are output to a downstream stage. | 03-21-2013 |
20130073516 | Extracting Incremental Data - The present disclosure introduces a method, an apparatus, and a system for extracting incremental data. Primary key information of incremental data is obtained from a backup database. The incremental data is inquired based on the primary key information from a main database that synchronizes with the backup database. The found incremental data is then inserted into a target data warehouse. The present techniques not only save a lot of time and system resources but also improve the efficiency of incremental data extraction. | 03-21-2013 |
20130073517 | FACTORY AUTOMATION TRANSACTIONS - A system that facilitates direct communication of a transaction between an automation controller and a business system comprises a request analyzer that receives a request for data relating to the automation controller and locates a transaction definition within the automation controller based upon the request. A subscribing component subscribes the business system to the automation controller based at least in part upon the located transaction definition. | 03-21-2013 |
20130091089 | EXTENSIBLE AND GENERIC FRAMEWORK FOR MANAGING RESOURCES FOR DATA WAREHOUSING LOADS - Methods, systems, and computer program products are provided for managing resources. A request is received for a plurality of resources from a requester. A ticket is generated for the request that indicates the resources. For each resource, a resource provider for the resource is queried to determine whether the resource can be acquired. The ticket is queued if any of the resources cannot be acquired. The ticket is granted if all of the resources can be acquired. | 04-11-2013 |
20130097114 | QUERY METADATA ENGINE - In one example, a method comprises receiving, with a data access service, an input query from an enterprise software application, wherein the data access service is positioned between the enterprise software application and a multidimensional data source; parsing the input query into a parse tree; generating an abstract syntax tree based on the parse tree, the abstract syntax tree comprising input query elements and multidimensional expression query elements based on the input query; performing an evaluation of metadata comprised in the multidimensional expression query elements of the abstract syntax tree; restructuring the abstract syntax tree based on the evaluation of the metadata; executing a query based on the restructured abstract syntax tree on the multidimensional data source; and retrieving a query result set from the multidimensional data source comprising results that are responsive to the query based on the restructured abstract syntax tree. | 04-18-2013 |
20130103637 | SYSTEM AND METHODS THEREOF FOR DETECTION OF USER DEMOGRAPHIC INFORMATION - A system and method for detecting implicit information related to a user are disclosed. The system comprises a network interface for accessing one or more data sources; a plurality of agents for collecting textual content from the one or more data sources; a processing unit for extracting explicit demographic information related to the user and generating implicit demographic information related to the user respective of the explicit demographic information; and a data warehouse connected to the processing unit for storing the generated implicit demographic information. | 04-25-2013 |
20130117216 | STAR AND SNOWFLAKE SCHEMAS IN EXTRACT, TRANSFORM, LOAD PROCESSES - A computer-implemented method, computer program product and a system for supporting star and snowflake data schemas for use with an Extract, Transform, Load (ETL) process, comprising selecting a data source comprising dimensional data, where the dimensional data comprises at least one source table comprising at least one source column, importing a data model for the dimensional data into a data integration system, analyzing the imported data model to select a star or snowflake target data schema comprising target dimensions and target facts, generating a meta-model representation by mapping at least one source table or source column to each target fact and target dimension, automatically converting the meta-model representation into one or more ETL jobs, and executing the ETL jobs to extract the dimensional data from the data source and loading the dimensional data into the selected target data schema in a target data system. | 05-09-2013 |
20130117217 | STAR AND SNOWFLAKE SCHEMAS IN EXTRACT, TRANSFORM, LOAD PROCESSES - A computer-implemented method, computer program product and a system for supporting star and snowflake data schemas for use with an Extract, Transform, Load (ETL) process, comprising selecting a data source comprising dimensional data, where the dimensional data comprises at least one source table comprising at least one source column, importing a data model for the dimensional data into a data integration system, analyzing the imported data model to select a star or snowflake target data schema comprising target dimensions and target facts, generating a meta-model representation by mapping at least one source table or source column to each target fact and target dimension, automatically converting the meta-model representation into one or more ETL jobs, and executing the ETL jobs to extract the dimensional data from the data source and loading the dimensional data into the selected target data schema in a target data system. | 05-09-2013 |
20130124453 | SLOWLY CHANGING DIMENSION ATTRIBUTES IN EXTRACT, TRANSFORM, LOAD PROCESSES - A computer-implemented method, computer program product and a system for identifying and handling slowly changing dimension (SCD) attributes for use with an Extract, Transform, Load (ETL) process, comprising importing a data model for dimensional data into a data integration system, where the dimensional data comprises a plurality of attributes, identifying via a data discovery analyzer one or more attributes in the data model as SCD attributes, importing the identified SCD attributes into the data integration system, selecting a data source comprising dimensional data, automatically generating an ETL job for the dimensional data utilizing the imported SCD attributes, and executing the automatically generated ETL to extract the dimensional data from the data source and loading the dimensional data into the imported SCD attributes in a target data system. | 05-16-2013 |
20130124454 | Slowly Changing Dimension Attributes in Extract, Transform, Load Processes - A computer-implemented method, computer program product and a system for identifying and handling slowly changing dimension (SCD) attributes for use with an Extract, Transform, Load (ETL) process, comprising importing a data model for dimensional data into a data integration system, where the dimensional data comprises a plurality of attributes, identifying via a data discovery analyzer one or more attributes in the data model as SCD attributes, importing the identified SCD attributes into the data integration system, selecting a data source comprising dimensional data, automatically generating an ETL job for the dimensional data utilizing the imported SCD attributes, and executing the automatically generated ETL to extract the dimensional data from the data source and loading the dimensional data into the imported SCD attributes in a target data system. | 05-16-2013 |
20130138601 | SYSTEM AND METHOD FOR DATA AGGREGATION, INTEGRATION AND ANALYSES IN A MULTI-DIMENSIONAL DATABASE - The present invention relates to a system and method for flexibly integrating a data warehouse with one or more operational database in real-time. The system comprises of a plurality of components embedded under one or more layers. The system comprises of a staging database, an ETL (extract transform load) module, a data warehouse and a networking module. The flexible integration is enabled by invoking one or more web services from a business enterprise. The invention further provides one or more analytical modules for performing an intelligent operation to carry out comparative analyses of the data. The designing and implementation module enables a deployment of the system on any related technology platform and tools. | 05-30-2013 |
20130159244 | METHODS AND APPARATUS FOR IMPROVING DATA WAREHOUSE PERFORMANCE - Methods and apparatus for improving data warehouse performance are disclosed. An example method to transform a first database into a second database includes generating a first mapping file associated with a fact table of the first database, the first mapping file referenced by a first hierarchical directory file path, generating a second mapping file associated with a dimension table, the second mapping file referenced by a second hierarchical directory file path, and retrieving data from the second database using the first and second mapping files via sequential navigation of the first and second hierarchical directory file paths. | 06-20-2013 |
20130159245 | DATA EXTRACTION AND SPARSENESS DISCOVERY FROM MULTIDIMENSIONAL DATABASES - Techniques for a method for efficiently extracting data from multidimensional databases are disclosed herein. When extracting data from a subset of a multidimensional database, the method first extracts data from high-level cells within the subset. If there are any empty high-level cells, the method removes lower-level cells that roll-up into the empty high-level cell, without extracting data from these lower-level cells. The method improves the data extraction efficiency by reducing the overhead of extracting data from empty cells of multidimensional databases. | 06-20-2013 |
20130159246 | ON-LINE TRANSACTION PROCESSING (OLTP) COMPRESSION AND RE-COMPRESSION OF DATABASE DATA - A computer is programmed to compress data of a database in response to database modification language (DML) statements generated by on-line transaction processing (OLTP) systems. In several embodiments, data that is initially added to a database block is left uncompressed until a predetermined condition is satisfied, which happens infrequently (relative to OLTP transactions on the block). When satisfied, the computer automatically compresses all uncompressed data in the block, which increases the amount of unused space in the block. New data is thereafter added uncompressed to the partially compressed block, until satisfaction of a predetermined condition whereby the partially compressed block is again compressed, i.e. re-compressed. Adding of new data to a partially compressed block and its compression are repeated unless another predetermined condition is met, in response to which the block is not further re-compressed, thereby to recognize a limit on the benefit from compression. | 06-20-2013 |
20130166495 | GENERATING A COMPILER INFRASTRUCTURE - In an embodiment, the compiler infrastructure allows execution of multidimensional analytical metadata from various databases by providing a generic transformation. A compilation request to execute a multidimensional analytical metadata is received. A type of the compilation request is determined to identify an associated transformation and corresponding transformation rules. Based upon the type of compilation request, a database of an application server is queried to retrieve the corresponding multidimensional analytical metadata. Based upon the identified transformation rules, the multidimensional analytical metadata is transformed into a generic metadata that is executable by any desired engine. An instance of a calculation scenario is generated based upon the transformation. The compiler infrastructure is generated by deploying the instance of the calculation scenario in the desired engine (e.g. in-memory computing engine.) | 06-27-2013 |
20130166496 | EXECUTING RUNTIME CALLBACK FUNCTIONS - In an embodiment, a runtime callback function is a part of a code that is invoked upon execution of an associated function. To execute the runtime callback function associated with an in-memory computing engine, multidimensional analytical metadata associated with an application server is received and transformed into an in-memory executable metadata, to generate an instance of an in-memory executable calculation scenario. The instance of the in-memory executable calculation scenario is analyzed to determine process callbacks associated with nodes of the in-memory executable calculation scenario. Based upon the determined process callbacks, the runtime callback function is executed by executing a selection callback at the nodes and a transformation callback at part providers associated with the in-memory executable calculation scenario. | 06-27-2013 |
20130166497 | DYNAMIC RECREATION OF MULTIDIMENSIONAL ANALYTICAL DATA - According to one aspect of systems and methods for dynamic recreation of multidimensional analytical data, lost sets of calculation scenarios that provide multidimensional analytical data results after aggregations and transformations of the multidimensional analytical data are recreated in the volatile storage of an in-memory computing engine. A multidimensional analytical data view (MDAV) compiler is triggered to read the MDAV metadata stored in an intermediate buffer in the MDAV compiler. The read MDAV metadata is compiled into a calculation scenario including calculation view metadata. The calculation view metadata is stored in the intermediate buffer. The recreated set of calculation scenarios is deployed on the in-memory computing engine. | 06-27-2013 |
20130166498 | Model Based OLAP Cube Framework - Systems, methods and computer program products that provide a framework for the creation, editing, manipulation and use of model-based, multidimensional analysis services (MAS) cubes are disclosed. A method of generating a model-based MAS cube comprises creating a data source comprising a data warehouse in the memory via the processor, creating a data source view providing a dimension, a fact and an outrigger from the created data source, and creating the MAS cube comprising at least one measure group. A key performance indicator (KPI) may be calculated from the MAS cube as a scorecard of a display associated with the processor. A user of the model-based MAS cube may create a new cube by targeting a set of facts and adding dimensions to the facts. | 06-27-2013 |
20130173528 | MULTI-FACT QUERY PROCESSING IN DATA PROCESSING SYSTEM - Techniques are described for optimizing a multi-fact star schema query to retrieve data from two or more fact tables of a database. The optimization techniques may include internally transforming a received multi-fact star schema query into a group of single-fact sub-queries, thereby reducing or eliminating certain constraints during query processing. The group of sub-queries may be enumerated to determine a query execution plan. The multi-fact star schema query may be executed according to the query execution plan to retrieve the data from the two or more fact tables of the database. | 07-04-2013 |
20130173529 | AUTOMATED DATA ANALYSIS AND TRANSFORMATION - A transformation method and system is provided. The method includes generating a data hub application configured to embed extract, transform, and load (ETL) processes. The data hub application is linked to source tables and target tables. Meta data associated with the source and target tables is transferred from virtual views of the data hub application to an ETL work area of the ETL processes. An ETL job is generated and linked to the data hub application. ETL processes are executed and results are determined. | 07-04-2013 |
20130179394 | System and Method for Interpreting and Generating Integration Flows - There is provided a computer system for generating an extract, transform, and load (ETL) workflow. The computer system includes a processor configured to receive ( | 07-11-2013 |
20130185249 | EFFICIENT LOADING OF DATA IN DATABASES - A computer implemented method, system and computer program product for managing data in databases generates a control file for a database by utilizing a definition file associated with the database and predefined column definitions, organizes data within a data file in a predefined format, wherein the data is desirable to be loaded in the database, and loads the data from the data file to the database by utilizing the control file and the data file. | 07-18-2013 |
20130185250 | SYSTEM AND METHODS FOR INTEGRATING SOFTWARE LAYERS FOR OLAP - In accordance with aspects of the disclosure, a system and methods may be provided for integrating software layers in OLAP including receiving, at an OLAP engine, a dynamic result set structure passed from an application layer via one or more other software layers. The dynamic result set structure may include a description of entities for a target layer. The system and methods may include enhancing raw data, at the OLAP engine, acquired from a database associated with a source layer based on the description of entities for the dynamic result set structure and passing one or more references to one or more memory locations for data elements in the dynamic result set structure from the OLAP engine to the application layer via the one or more other software layers for acquiring each data element associated with each corresponding reference in the dynamic result set structure. | 07-18-2013 |
20130185251 | EFFICIENT LOADING OF DATA IN DATABASES - A computer implemented method, system and computer program product for managing data in databases generates a control file for a database by utilizing a definition file associated with the database and predefined column definitions, organizes data within a data file in a predefined format, wherein the data is desirable to be loaded in the database, and loads the data from the data file to the database by utilizing the control file and the data file. | 07-18-2013 |
20130191322 | DISCLOSURE RANGE DETERMINATION METHOD, DISCLOSURE RANGE DETERMINATION APPARATUS, AND MEDIUM - A disclosure range determination method includes accepting a disclosure range of data and metadata of the data, extracting attribute information of a user that belongs to the disclosure range of the data accepted in the accepting based on association information associating a disclosure range of data with attribute information of a user that belongs to the disclosure range, converting the metadata accepted in the accepting into a keyword based on association information associating a conversion rule that converts the metadata into significant information with a keyword corresponding to the conversion rule, and determining, by a computer, whether or not a user that belongs to the disclosure range accepted in the accepting qualifies for the disclosure range accepted in the accepting, based on whether or not the keyword converted in the converting is included in the attribute information extracted in the extracting. | 07-25-2013 |
20130204836 | METHOD AND APPARATUS FOR CONSTRUCTING DATA WAREHOUSE TO MANAGE MASS BUILDING ENERGY INFORMATION - A method and apparatus for constructing a data warehouse for managing mass building energy information are provided. The apparatus for constructing a data warehouse in order to manage mass building energy information includes a primitive database generator for constructing primitive databases in each building based on energy control point information, a time series data warehouse generator for generating time series data warehouses based on the constructed primitive databases, a space time series data warehouse generator for reconstructing the generated time series data warehouses based on an object construction information table including space information about objects for control points of the plurality of buildings to be controlled to generate space time series data warehouses, and a view database generator for generating view databases by a search condition in accordance with the search condition for energy usage information based on the generated space time series data warehouses. | 08-08-2013 |
20130212060 | SYSTEM AND METHOD FOR MATHEMATICS ONTOLOGY EXTRACTION AND RESEARCH - An extensive computer based online math research system (the “Research System”) having as its foundation an Ontology of mathematics, and utilizing unique and intensive computer support, coordination, data structuring, data storage, computer processing, retrieval capabilities, and data-mining capabilities, and an Ontology editing system that runs on computer software with computer processors and data storage capabilities (the “Ontology Editor System”). The Research System also includes a methodology to enable online reference and data manipulation of the Ontology, and an Internet based search of the concepts of mathematics and applications of mathematics to the sciences on the basis of the Ontology. | 08-15-2013 |
20130212061 | CONTENT ANALYTICS SYSTEM CONFIGURED TO SUPPORT MULTIPLE TENANTS - Techniques are disclosed for a software as a service (SaaS) provider to host a content analytics tool used to evaluate data collections for multiple customers (referred to as tenants) using one dedicated and expandable computing infrastructure, without requiring that the service provider obtain, install, license, and manage a separate copy of the content analytics tools for each tenant. Customers are provided access to resources dedicated to their enterprise, but do not have access, or even awareness, of data collections or analytics resources hosted for other customers. That is, embodiments presented herein allow a provider to host content analytics tools used by customers to evaluate their enterprise data in a secure and timely manner. | 08-15-2013 |
20130218828 | SYSTEMS AND METHODS FOR INTERVAL CONTROL ELEMENT CHAIN ARCHITECTURE - The present disclosure generally relates to accessing data, and more particularly, to systems and methods for improving the efficiency and quality of real-time extracting, transforming, and/or loading data using customer information control system (CICS) interval control element (ICE) chain processing. | 08-22-2013 |
20130226860 | Interest-Driven Business Intelligence Systems and Methods of Data Analysis Using Interest-Driven Data Pipelines - Interest-driven Business Intelligence (BI) systems in accordance with embodiments of the invention are illustrated. In one embodiment of the invention, a data processing system includes raw data storage containing raw data, metadata storage containing metadata that describes the raw data, and an interest-driven data pipeline that is automatically compiled to generate reporting data using the raw data, wherein the interest-driven data pipeline is compiled based upon reporting data requirements automatically derived from at least one report specification defined using the metadata. | 08-29-2013 |
20130232105 | Dynamic access of data - A heterogeneous information technology system in which compatible and incompatible client systems are able to dynamically access master data stored in a master database maintained by a master data server. An integration server communicates with the client systems and the master data server, and in response to a request for data from a client system, maps the master data in the master database to mapped data based on a set of mapping rules associated with the client system, and sends the mapped data to the client system. | 09-05-2013 |
20130238549 | Using Dimension Substitutions in OLAP Cubes - Systems, methods and computer program products that provide a framework for the creation, editing, manipulation and use of model-based, multidimensional analysis services (MAS) cubes and using substitute dimensions in such cubes are disclosed. To permit a user to obtain better and automatic access to business intelligence, a method of generating a model-based MAS cube comprises creating a data source comprising a data warehouse in the memory via the processor, creating a data source view providing a dimension, a fact and an outrigger from the created data source, and creating the MAS cube comprising at least one measure group. Using substitute dimensions comprises finding all relevant substitutions for a measure group, creating a table for the measure group in the data source view, adding a property as the primary key of the substitute dimension and generating a query containing an inner join logical link between the substitute and original dimension. | 09-12-2013 |
20130238550 | METHOD TO DETECT TRANSCODING TABLES IN ETL PROCESSES - Techniques are disclosed for identifying transcoding tables in an Extract-Transform-Load (ETL) process, by identifying, by operation of one or more computer processors, records passing through an operator configured to replace values in the records with values from at least one table linked to the operator before being sent to an output table, wherein the operator specifies an operation for extracting, transforming, or loading data stored in one or more source systems into storage by a target system, and evaluating at least a first table linked to the operator to determine whether the first table is a transcoding table by assigning a score to the first table, wherein the score is indicative of the likelihood that the first table is a transcoding table, wherein a transcoding table is used to harmonize values from a plurality of tables in the one or more source systems to a table in the target. | 09-12-2013 |
20130238551 | Interest-Driven Business Intelligence Systems and Methods of Data Analysis Using Interest-Driven Data Pipelines - Interest-driven Business Intelligence (BI) systems in accordance with embodiments of the invention are illustrated. In one embodiment of the invention, a data processing system includes raw data storage containing raw data, metadata storage containing metadata that describes the raw data, and an interest-driven data pipeline that is automatically compiled to generate reporting data using the raw data, wherein the interest-driven data pipeline is compiled based upon reporting data requirements automatically derived from at least one report specification defined using the metadata. | 09-12-2013 |
20130246338 | System and method for indexing a capture system - An indexer to index keyword and metadata input to or captured by a network capture device acting on a stream of captured content is described. A keyword is a word, phrase, name, or other alphanumeric term that exists within common textual content such as an email, Microsoft Office document, or similar content. Metadata includes properties describing the network characteristics of the content containing keywords. Examples of network characteristics include, but are not limited to, the source and destination addresses (Internet Protocol (IP) addresses), time and date of the transmission, size and name of the content, and protocol used to transmit the content. Additional descriptive properties may be used to describe the device upon which the content was captured, the user, viewer of the captured content or security settings of the captured content, etc. | 09-19-2013 |
20130246339 | SYSTEMS AND METHODS FOR CATEGORIZING AND PRESENTING PERFORMANCE ASSESSMENT DATA - The field of the invention relates to systems and methods for data mining and processing, and more particularly to systems and methods for automating content from performance assessment data. In one embodiment, an automated notes and categorization system may include a primary database comprising performance assessment data. The primary database is operatively coupled to a computer program product having a computer-usable medium having a sequence of instructions which, when executed by a processor, causes said processor to execute a process that analyzes and converts raw performance data into automated content that presents data in readable user friendly form to facilitate human understanding. | 09-19-2013 |
20130246340 | SINGLE COMMAND DATA WAREHOUSE TABLE UPDATE - Information warehouse table data is updated through a full outer join between warehouse tables and source tables in response to a single query language command statement of a relational database computer language. The single command statement comprises table mapping instructions for updating the warehouse tables that define matching conditions, delete instructions, archive instructions, and update instructions. The full outer join inserts rows in the warehouse tables for source table data rows that do not have corresponding matching rows in the warehouse table. Data rows in the warehouse tables without corresponding matching source table rows are deleted if a value of a parameter in the warehouse table system is set to delete; updated if the parameter is update; and updated with archive values and a new table row inserted with existing values if the parameter is archive. | 09-19-2013 |
20130254155 | Multi-Dimensional Query Expansion Employing Semantics and Usage Statistics - Embodiments relate to systems and methods employing personalized query expansion to suggest measures and dimensions allowing iterative building of consistent queries over a data warehouse. Embodiments may leverage one or more of: semantics defined in multi-dimensional domain models, user profiles defining preferences, and collaborative usage statistics derived from existing repositories of Business Intelligence (BI) documents (e.g. dashboards, reports). Embodiments may utilize a collaborative co-occurrence value derived from profiles of users or social network information of a user. | 09-26-2013 |
20130254156 | Algorithm and System for Automated Enterprise-wide Data Quality Improvement - Algorithm and System for Automated Enterprise-wide Data Quality Improvement by creating an infrastructure where error patterns can be stored in SQL statement format to system's local repository, in this way system can identify data errors either coming directly through keyboard entries or coming from another system through an automated feeds or manual feeds. The system automatically scans for erroneous records based on those error patterns and emails only faulty records in encrypted MS Excel format to correction agents for review and update to the production RDBMS. | 09-26-2013 |
20130262370 | Fast Predicate Table Scans Using Single Instruction, Multiple Data Architecture - An approach is provided in which a processor receives a scan request to scan data included in a data table. The processor selects a column in the data table corresponding to the scan request and retrieves column data entries from the selected column. In addition, the processor identifies the width of the selected column and selects a scan algorithm based upon the identified column width. In turn, the processor loads the column data entries into column data vectors and computes scan results from the column data vectors using the selected scan algorithm. | 10-03-2013 |
20130268478 | System and method for correct and effective transaction execution in a data center system by implementing FIFO-grouping in a data store - A data center is a computing environment focusing on the execution of data transactions for corporations and organizations. The data center has the capability of running many execution threads in parallel. | 10-10-2013 |
20130268479 | SYSTEM AND METHOD FOR PRESENTING AND MANAGING SOCIAL MEDIA - A method, system, apparatus, and computer program product provide the ability to present media content on a social media network. Live streaming data is received from multiple data sources. The data is enriched with meta information using internal systems via application programming interfaces (APIs). The live streamed data is transformed and transmitted, based on the meta information, to one or more users on a social media network. | 10-10-2013 |
20130275360 | DATA EXTRACTION, TRANSFORMATION, AND LOADING - A device receives instructions for extracting, transforming, and loading data; identifies a plurality of data processing activities based on the instructions; and determines a processing sequence for performing the plurality of data processing activities. The device further creates a plurality of processing objects corresponding to the plurality of data processing activities. Each one of the plurality of processing objects including a discrete data structure that is produced by performing a corresponding one of the plurality of data processing activities. The device performs the plurality of data processing activities in accordance with the processing sequence and by using the plurality of processing objects. | 10-17-2013 |
20130275361 | ASSOCIATING MULTIPLE DATA SOURCES INTO A WEB-ACCESSIBLE FRAMEWORK - Systems, methods, and computer-readable media for associating multiple data sources into a web-accessible framework. Health data is received from multiple data sources and is used to populate a framework comprising at least one topic focused data mart. Each topic focused data mart has a common structure and is associated with a web service providing standard features supported by each topic focused data mart and custom features specific to a topic associated with each topic focused data mart. In various embodiments, demographic information is received from a clinician and is utilized to present context-specific data derived from the topic focused data mart. | 10-17-2013 |
20130275362 | Classic to In-Memory Cube Conversion - Conversion of a snowflake-like schema to a star-like schema includes a distribution of the snowflake-like schema across several host computer systems. Each host computer system has a fragment of the snowflake-like schema, including a fragment of constituent fact tables. Dimension keys in the fact table fragments are mapped to data tables via the constituent dimension tables. Data table keys are identified and populate the fact table for the converted star-like schema, thus obviating the dimension tables in the converted star-like schema. | 10-17-2013 |
20130275363 | META-DATA DRIVEN DATA INGESTION USING MAPREDUCE FRAMEWORK - A generic approach for automatically ingesting data into an HDFS (Hadoop File System) based data warehouse includes a datahub server, a generic pipelined data loading framework, and a meta-data model that, together, address data loading efficiency, data source heterogeneities, and data warehouse schema evolvement. The loading efficiency is achieved via the MapReduce scale-out solution. The meta-data model is comprised of configuration files and a catalog. The configuration file is setup per ingestion task. The catalog manages the data warehouse schema. When a scheduled data loading task is executed, the configuration files and the catalog collaboratively drive the datahub server to load the heterogeneous data to their destination schemas automatically. | 10-17-2013 |
20130275364 | Concurrent OLAP-Oriented Database Query Processing Method - A concurrent on-line analytical processing (OLAP)-oriented database query processing method is described, for performing, on the basis of predicate vector-based memory OLAP star-join optimization, concurrent OLAP query processing based on a batch query predicate vector bit operation. The concurrent query processing optimization technology is implemented for I/O performance and parallel OLAP processing performance in a database management system, and setting of concurrent OLAP processing load in an optimized way catering to the I/O performance is supported, thereby improving predictable processing performance oriented to diversified OLAP queries and implementing concurrent query star-join bitmap filtering processing based on predicate vector arrays. | 10-17-2013 |
20130275365 | Multi-Dimensional OLAP Query Processing Method Oriented to Column Store Data Warehouse - A multi-dimensional OLAP query processing method oriented to a column store data warehouse is described. With this method, an OLAP query is divided into a bitmap filtering operation, a group-by operation and an aggregate operation. In the bitmap filtering operation, a predicate is first executed on a dimension table to generate a predicate vector bitmap, and a join operation is converted, through address mapping of a surrogate key, into a direct dimension table tuple access operation; in the group-by operation, a fact table tuple satisfying a filtering condition is pre-generated into a group-by unit according to a group-by attribute in an SQL command and is allocated with an increasing ID; and in the aggregate operation, group-by aggregate calculation is performed according to a group item of a fact table filtering group-by vector through one-pass column scan on a fact table measure attribute. | 10-17-2013 |
20130297556 | IN-MEMORY SPATIAL DATABASE FOR GEOCODING/GEOPROCESSING - The embodiments provide an in-memory database system having an extraction module configured to extract data (e.g., business data and address data) from one or more external data sources and transform the data into a standard format, a geocoder configured to geocode the address data including obtaining spatial data based on the address data using an internal reference table, and an internal database configured to store the internal reference table, the business data, the address data, and the spatial data | 11-07-2013 |
20130297557 | COLUMN BASED DATA TRANSFER IN EXTRACT, TRANSFORM AND LOAD (ETL) SYSTEMS - Executing a plurality of transform stages in an extract, transform and load (ETL) job including, for each of the transform stages, receiving a plurality of input row identifiers (RIDs) corresponding to a first plurality of source database table rows in a source database table. Data is retrieved directly from a subset of the source database table columns in the first plurality of source database table rows based on the input RIDs and transform logic. Partial row data including data from the subset of the source database table columns is generated for each of the first plurality of source database table rows. Transformed data is generated based on the partial row data and to the transform logic. Output RIDs corresponding to a second plurality of rows in the source database table that include a least a subset of the transformed data are output to a downstream stage. | 11-07-2013 |
20130304693 | Cloud Logistics - Implementations of complex software solutions are accompanied with an intrinsic conflict between requirements, such as an ever-growing complexity of enterprise software solutions versus a demand for more and more shortened implementation cycles. Various embodiments include at least one of systems methods and software to address such conflicts. One method embodiment includes receiving a request to assemble a virtual appliance from at least one deployment template, the request including characteristic data representative of at least one virtual appliance functional and deployment requirement. In response to the received request, such embodiments may query a template warehouse to identify and deployment templates from a template warehouse to a data storage device to form a virtual appliance. The virtual appliance may then be packaged into a transport data structure and stored. The transport data structure may then be transported by various means to a target virtual machine and deployed therein. | 11-14-2013 |
20130318030 | Embedded Data Marts for Central Data Warehouse - A central data warehouse includes embedded data marts. These embedded data marts, referred to as workspaces, are assigned centrally manage data by reference only but rely directly on the centrally managed data and the underlying infrastructure. Workspaces still allow departments in an enterprise to perform certain actions on their own (like adding new data and building new models) without having to instantiate copies of the centrally managed data in a locally managed data mart. | 11-28-2013 |
20130318031 | DISTRIBUTED PROCESSING OF STREAMING DATA RECORDS - Representative embodiments of a distributed processing method of facilitating interactive analytics of streaming data records by receiving the data records at a plurality of distributed computational nodes, establishing and storing dimensions corresponding to attributes of the data records, parsing the streaming data records to identify matches to at least one of the dimensions and based thereon, reducing the number of data records to create a targeted subset of the data, re-distributing the targeted subsets of the streaming data records among the distributed computational nodes in accordance with the dimensions stored on the nodes, updating a database storing measures of the dimensions in accordance with the targeted subsets of the streaming data records, and using the database to respond to a query based on measures associated with one or more of the dimensions. | 11-28-2013 |
20130318032 | EMBEDDING AND RETRIEVING DATA IN AN APPLICATION FILE FORMAT - Third party data embedding may be provided. Data associated with a third party provider may be saved in a native application save file. If a save file opened by the application is determined to comprise third party data, the application may determine whether an appropriate data provider is available. The application may connect to the data provider and provide on-demand, asynchronous access to the third party data upon request from a user. | 11-28-2013 |
20130318033 | Method for Operating an Automation Device - A method for operating an automation device, into the memory of which an automation solution has been loaded, wherein a technology-oriented control interpreter accesses a data warehouse of the automation solution, and is able to control external commands by virtue of such commands being analyzed and being implemented according to the analysis, where the technology-oriented control interpreter extracts at least one entity designation and at least one instruction from a respective command, the technology-oriented control interpreter searches for an object matching the entity designation in the data warehouse of the automation solution and, in the event of success, checks whether the instruction contained in the command has been defined for the found object, and where the technology-oriented control interpreter causes execution of the instruction for the found object. | 11-28-2013 |
20130318034 | DISTRIBUTED PROCESSING OF STREAMING DATA RECORDS - Representative embodiments of a distributed processing method of facilitating interactive analytics of streaming data records by receiving the data records at a plurality of distributed computational nodes, establishing and storing dimensions corresponding to attributes of the data records, parsing the streaming data records to identify matches to at least one of the dimensions and based thereon, reducing the number of data records to create a targeted subset of the data, re-distributing the targeted subsets of the streaming data records among the distributed computational nodes in accordance with the dimensions stored on the nodes, updating a database storing measures of the dimensions in accordance with the targeted subsets of the streaming data records, and using the database to respond to a query based on measures associated with one or more of the dimensions. | 11-28-2013 |
20130325788 | ETL DATA TRANSIT METHOD AND SYSTEM - A method and system for archiving data is provided. The method includes an extraction stage, a load stage, and a production merge stage. In the extraction stage, incoming files from a source system are extracted, where each incoming file has data records. Upon determining that the incoming files are of different data record types, the incoming files are split by record type. The files are also parsed. In a load stage, all files from a load ready directory are placed together in a single batch with a common RUNID for each data record type. In a production merge stage, all files made available from the load stage are processed. The processed files are loaded into a data warehouse. Each stage operates independently of any other stage. Each stage is also iterative and circular. | 12-05-2013 |
20130325789 | Defining and Mapping Application Interface Semantics - An example implementation includes a system having one or more processors, a definition module, a data store, and a workflow module. The definition module is executable by the one or more processors to determine a first semantic element for a first operation included in a first application interface representing a first service and a second semantic element for a second operation included in a second application interface representing a second service, such as an enterprise service bus. The definition module may further determine a mapping of the first semantic element with the second sematic element. The data store is configured to store the mapping of the first semantic element with the second sematic element by the definition module. The workflow module is executable by the one or more processors to integrate the first application interface with the second service based on the mapping. | 12-05-2013 |
20130325790 | MULTI-TENANT DATA INTEGRATION - An enterprise application integration system (EAIS) is disclosed that enables customer service applications to access and share data with enterprise information systems in real time. | 12-05-2013 |
20130325791 | Method and System for Analyzing Data Related to an Event - A system and method for analyzing data from a plurality of computer environments. A user may search for computer environments that meet a certain criterion. The computer environments are authenticated and data is copied from the computer environments to a memory location. The data may be marked so that a user may determine which computer environment provided the data. The user may add notations to the data during a review. Changes to data on the computer environments may be provided to a user using a syndication feed. | 12-05-2013 |
20130325792 | Method and System for Analyzing Data Related to an Event - A system and method for analyzing data from a plurality of computer environments. A user may search for computer environments that meet a certain criterion. The computer environments are authenticated and data is copied from the computer environments to a memory location. The data may be marked so that a user may determine which computer environment provided the data. The user may add notations to the data during a review. Changes to data on the computer environments may be provided to a user using a syndication feed. | 12-05-2013 |
20130332407 | IN-QUERYING DATA CLEANSING WITH SEMANTIC STANDARDIZATION - The present invention relates to data cleansing, and in particular performing the semantic standardization process within a database before the transform portion of the extract-transform-load (ETL) process. Provided are a method, system and computer program product for standardizing data within a database engine, configuring the standardization function to determine at least one standardized value for at least one data value by applying the standardization table in a context of at least one data value, receiving a database query identifying the standardization function, at least one database value and the context of the data, and invoking the standardization function. | 12-12-2013 |
20130332408 | IN-QUERYING DATA CLEANSING WITH SEMANTIC STANDARDIZATION - The present invention relates to data cleansing, and in particular performing the semantic standardization process within a database before the transform portion of the extract-transform-load (ETL) process. Provided are a method, system and computer program product for standardizing data within a database engine, configuring the standardization function to determine at least one standardized value for at least one data value by applying the standardization table in a context of at least one data value, receiving a database query identifying the standardization function, at least one database value and the context of the data, and invoking the standardization function. | 12-12-2013 |
20140006338 | BIG DATA ANALYTICS SYSTEM | 01-02-2014 |
20140006339 | DETECTING REFERENCE DATA TABLES IN EXTRACT-TRANSFORM-LOAD PROCESSES | 01-02-2014 |
20140025625 | AUTOMATED DATA ANALYSIS AND TRANSFORMATION - A transformation method and system is provided. The method includes generating a data hub application configured to embed extract, transform, and load (ETL) processes. The data hub application is linked to source tables and target tables. Meta data associated with the source and target tables is transferred from virtual views of the data hub application to an ETL work area of the ETL processes. An ETL job is generated and linked to the data hub application. ETL processes are executed and results are determined. | 01-23-2014 |
20140032479 | AQUATIC GEOGRAPHIC INFORMATION SYSTEM - A method of processing geo-statistical data includes preparing a data log, extracting acoustic data and coordinate data from the data log, and aligning the acoustic data and the coordinate data. The method also includes cleaning and aggregating the coordinate data, validating the coordinate data geospatially, and creating an output. | 01-30-2014 |
20140040182 | SYSTEMS AND METHODS FOR COLLECTION AND CONSOLIDATION OF HETEROGENEOUS REMOTE BUSINESS DATA USING DYNAMIC DATA HANDLING - Remote data collection systems and methods retrieve data including financial, sales, marketing, operational and the like data from a plurality of databases and database types remotely over a network in an automated, platform-agnostic manner. An Extract Transform and Load (ETL) data replication method for Chart of Account (COA) standardization includes receiving a request for remote data collection to extract data from a data source; extracting data in a non-intrusive manner from the data source, wherein the data comprises non-standard COA data; and transforming one of an entire set or a subset of the extracted data based on the request based on a template or a standardized form desired for comparisons. | 02-06-2014 |
20140067750 | Cross-Platform Compiler for Data Transforms - Techniques for automatically partitioning a multi-platform data transform flow graph to one or more target output platforms are provided. The techniques include performing type inference on a transform graph, wherein the transform graph comprises one or more data transforms, automatically partitioning the transform graph to one or more target output platforms based on one or more policies, performing an optimization of the partitioned transform graph, and generating code, from the partitioned transform graph, for each set of the one or more data transforms based on the one or more target output platforms. | 03-06-2014 |
20140067751 | COMPRESSED SET REPRESENTATION FOR SETS AS MEASURES IN OLAP CUBES - A cardinality of an incoming data stream is maintained in real time; the cardinality is maintained in a data structure that is represented by an unsorted list at low cardinalities, a linear counter at medium cardinalities, and a PCSA at high cardinalities. The conversion to the linear counter makes use of the data in the unsorted list, after which that data is discarded. The conversion to the PCSA uses only the data in the linear counter. | 03-06-2014 |
20140074771 | QUERY OPTIMIZATION - Provided are techniques for generating a relational query. Information is collected from a query specification and a model for an On-Line Analytical Processing (OLAP) query having at least a first expression and a second expression. The collected information is used to generate a relational query to retrieve report data to be used to satisfy the first expression and the second expression. | 03-13-2014 |
20140081902 | Integrating Data Transform Test with Data Transform Tool - Embodiments relate to integrating data transform test with a data transform tool. A method and system are described for creating a data transform test for a data transform job having a data transform script, the method includes determining all data transform units available in the data transform job, determining a subset of the available data transform units for a new test, and generating a subset test execution script for the subset of data transform units from the data transform script. The method further includes determining boundary test data at each boundary of the subset of data transform units, defining a data transform test, and saving the data transform test for later testing. The data transform test includes the subset of data transform units with subset test execution script and with boundary test data. | 03-20-2014 |
20140081903 | METHODS AND SYSTEMS FOR DISPLAYING AND FILTERING BUSINESS ANALYTICS DATA STORED IN THE CLOUD - In accordance with disclosed embodiments, there are provided methods, systems, and apparatuses for displaying and filtering business analytics data stored in the cloud, including, for example, means for displaying a graphical interface at a client device; communicating a business analytics query from the client device to a remote host organization via a public Internet; receiving a business analytics dataset in a complete and unfiltered form from the host organization responsive to the business analytics query; caching the business analytics dataset in its complete and unfiltered form to the memory of the client device; displaying a business analytics report at the graphical interface of the client device, the business analytics report representative of the business analytics dataset in its complete and unfiltered form; receiving filter input at the client device; applying the filter input to the business analytics dataset to yield a filtered sub-set; and updating the business analytics report displayed at the graphical interface using the filtered sub-set of the business analytics dataset. Other related embodiments are disclosed. | 03-20-2014 |
20140089251 | ENHANCING PERFORMANCE OF EXTRACT, TRANSFORM, AND LOAD (ETL) JOBS - A computer receives one or more files having configuration information that includes data that defines a plurality of stages of an extract, transform, and load (ETL) job, wherein the plurality of stages comprise a read stage that is preceded by a write stage, and wherein the read stage reads data from a source location, and wherein the data that is read or a modified version of the data that is read is being written by the write stage that writes data to the source location. The computer replaces the read stage with a decompressor stage. The computer replaces the write stage with a compressor stage. The computer executes the decompressor stage and compressor stage on a field-programmable gate array that is programmatically customized with data compression and data decompression functionality to enhance the performance of the ETL job. | 03-27-2014 |
20140089252 | ENHANCING PERFORMANCE OF EXTRACT, TRANSFORM, AND LOAD (ETL) JOBS - A computer receives one or more files having configuration information that includes data that defines a plurality of stages of an extract, transform, and load (ETL) job, wherein the plurality of stages comprise a read stage that is preceded by a write stage, and wherein the read stage reads data from a source location, and wherein the data that is read or a modified version of the data that is read is being written by the write stage that writes data to the source location. The computer replaces the read stage with a decompressor stage. The computer replaces the write stage with a compressor stage. The computer executes the decompressor stage and compressor stage on a field-programmable gate array that is programmatically customized with data compression and data decompression functionality to enhance the performance of the ETL job. | 03-27-2014 |
20140101091 | RULE-BASED EXTRACTION, TRANSFORMATION, AND LOADING OF DATA BETWEEN DISPARATE DATA SOURCES - Methods and apparatus are presented for extracting, transforming, and loading data from one database to another database. For example, an extraction, transformation, and loading (ETL) component may access an operational log of a given database in order to detect an update to the database. Upon detecting the update, the ETL component may extract a subset of data from the operational log, where the extraction of the subset of data is based on one or more rules. Once the subset of data has been extracted, the ETL component may transform the extracted subset of data from the operational log into a format for another, target database, where the data format for the other, target database is different from a data format for the given, source database. The ETL component may then load the subset of data transformed into the data format for the other, target database into the target database. | 04-10-2014 |
20140101092 | ADJUSTMENT OF MAP REDUCE EXECUTION - Disclosed herein are techniques for adjusting a map reduce execution environment. It is determined whether some operations in a sequence of operations should be implemented in a map reduce execution environment. If it is determined that some operations in a sequence of operations should be implemented in a map reduce execution environment, the map reduce execution environment is adjusted to achieve a predefined performance objective. | 04-10-2014 |
20140101093 | DISTRIBUTED, REAL-TIME ONLINE ANALYTICAL PROCESSING (OLAP) - Source data of an event stream is parsed and supplemented with additional data from reference data sources, producing an enriched event stream from the parsed event stream data. The data records of the enriched event stream are partitioned into data fields designated as a dimension partition and a metric partition, which are partitioned into sub-dimension projections mapped to a plurality of storage keys, such that each of the storage keys includes one or more placeholder wildcard values and each of the storage keys is stored into a database of the computer system by the computer processor. The stored storage keys are then aggregated onto a two-dimensional coordinate vector such that, if the computer processor identifies a permuted storage key having metric values for which a corresponding storage key already exists in the database, then the computer processor aggregates the metric values of the identified storage key with the metric values of the corresponding storage key, and if the computer processor does not identify the permuted storage key as having a corresponding storage key that already exists in the database, then the computer processor writes the metric values of the permuted storage key into the database, comprising initial values for the key combination of dimension values. | 04-10-2014 |
20140108331 | OLAP Execution Model Using Relational Operations - In one embodiment the present invention includes an OLAP execution model using relational operations. In one embodiment, the present invention includes, a method comprising receiving a first query in an online analytic processor (OLAP) executing on one or more computers, the OLAP generating and comprising a model specifying a graph defining a plurality of nodes and a plurality of tiers, each node corresponding to a different operation on data. A second query is generated by the OLAP. The second query includes a plurality of layered subqueries each corresponding to one of the nodes in the graph for specifying the different operations on data. The second query is received in a relational engine coupled to the datastore. The relational engine executes the second query, and in accordance therewith, retrieves data. | 04-17-2014 |
20140114906 | GENERIC SEMANTIC LAYER FOR IN-MEMORY DATABASE REPORTING - The disclosure generally describes computer-implemented methods, software, and systems for providing a generic semantic layer for in-memory database reporting. One computer-implemented method for combining online transactional processing and online analytical processing in an in-memory database, comprises: retrieving two or more tables from an online transaction processing system; identifying related tables among the two or more tables; determining relationships between the related tables; determining a measure based on the relationships; and outputting the measure. | 04-24-2014 |
20140114907 | DATA LINEAGE SYSTEM - A data lineage system is provided that traces a data lineage of a data warehouse. The data lineage system maps a target data element to one or more source data elements. The data lineage system further stores one or more source surrogate keys within one or more auxiliary columns of a target data record. The data lineage system further stores, for each source data element, a data lineage mapping system record within a data lineage mapping system table that represents the mapping of the target data element and the corresponding source data element. The data lineage system further maps a source data element to one or more target data elements. The system further stores, for each target data element, a shadow system record within a shadow system table that represents the mapping of the source data element and the corresponding target data element. | 04-24-2014 |
20140114908 | Method and Apparatus for Accelerated Format Translation of Data in a Delimited Data Format - Various methods and apparatuses are described for performing high speed format translations of incoming data, where the incoming data is arranged in a delimited data format. As an example, the data in the delimited data format can be translated to a fixed field format using pipelined operations. A reconfigurable logic device can be used in exemplary embodiments as a platform for the format translation. | 04-24-2014 |
20140114909 | Systems and Methods for Providing Performance Metadata in Interest-Driven Business Intelligence Systems - Interest-driven business intelligence server systems that provide performance metadata are described. In the disclosed embodiments, an interest-driven business intelligence server system receives a report specification. The report specification includes at least reporting data requirement. The interest-driven business intelligence server determines performance metadata information for an interest-driven data pipeline that is utilized to generate reporting data based on the report specification. The performance metadata information for the interest-driven data pipeline is transmitted to an interest-driven user visualization system by the interest-driven business intelligence server system. | 04-24-2014 |
20140122412 | SYSTEM AND METHOD TO PROVIDE ANALYTICAL PROCESSING OF DATA IN A DISTRIBUTED DATA STORAGE SYSTEMS - The present disclosure in general relates to technologies for processing data in a distributed data storage system, and more particularly, to a method, a system, and a computer program product for analytical processing of data by using the processing power of the distributed data storage system. In one embodiment, a system for analytical processing of data in a distributed data storage system is disclosed. The system comprises: a data extraction module configured to perform analytical operations to extract data from source databases in one or more data formats; and a processing module configured to perform data refinement operations to categorize the data while the data is being extracted. The processing module comprises: a mapping module configured to perform mapping operations of the categorized data; and a transformation module configured to perform an analytical transforming operation of the mapped categorized data to obtain a transformed categorized data. | 05-01-2014 |
20140122413 | BULK READ AND WRITE BETWEEN MULTI-DIMENSIONAL DATA STRUCTURES - A system and method for reading and writing of data values between multidimensional structures that reside in Online Analytical Processing (OLAP) databases are disclosed. Data queries may be performed and updates executed between multidimensional data structures, whether existing on the same server or separate servers. Bulk (being two or more intersections) transfers are allowed between multidimensional structures (or cubes), providing a performance gain that cannot be matched using a standard point-by-point implementation. An intersection only contains a numerical or data value if there is a value for each dimension at that intersection within the database. Multidimensional data structures naturally generate sparse intersections where no data values are found, which can greatly impact performance. Within each cube, only a small intersection of members actually contains values. The system may very quickly calculate reports which include any intersection in any very large cube. | 05-01-2014 |
20140136471 | Rapid Provisioning of Information for Business Analytics - An approach is provided in which a system creates schema terms based upon matching input data query requirements to industry terms. In turn, the system generates a query and an associative map, which includes data organized according to the schema terms. The system executes the query, which retrieves the data from the associative map and loads the data into one or more storage areas. | 05-15-2014 |
20140136472 | Methodology supported business intelligence (BI) software and system - The invention provides idealized and reusable data source interfaces. The process of idealizing includes reengineering of the original data model using a surrogate key based model. The technique emphasizes readability and performance of the resulting operational data store. | 05-15-2014 |
20140149338 | SYSTEMS AND METHODS FOR COOPERATIVE DATA EXCHANGE - Various systems and methods for financial analysis are provided. A system is provided comprising a first node comprising a public facing data store in communication with a private facing data store, wherein the first node further comprises a cooperative lookup module configured to locate a second node, a secure data connection between the first node and the second node, and wherein the first node is configured to request processed internal data from the second node. | 05-29-2014 |
20140156589 | DIVIDING AND COMBINING OPERATIONS - Disclosed herein are techniques for arranging a series of operations. It is determined whether an operation executes more efficiently when divided. It is further determined whether a plurality of operations execute more efficiently when combined. | 06-05-2014 |
20140156590 | PRODUCING AUTOMATED TERMS LISTINGS IN HTML DOCUMENT PUBLISHING WITH ENHANCED TERMS DEFINITIONS - A terms listing production system generates a terms listing associated with an electronic document. Primary terms and primary definitions may be extracted from a glossary associated with the electronic document. Secondary terms may be identified based on user searches in relation to the electronic document. Secondary definitions associated with primary terms or secondary terms may be retrieved from external content and stored in connection with the associated terms. A markup language document configured for presentation to users is generated based on the stored terms and definitions. | 06-05-2014 |
20140156591 | KNOWLEDGE CATALYSTS - A computer implemented method integrates data from remote disparate data sources by processing a non-transitory media. The non-transitory media stores instructions for detecting data sets in different formats hosted in a plurality of heterogeneous databases that are accessible through a distributed network. The method extracts schema data from the plurality of heterogeneous databases and identifies related fields in two or more of the heterogeneous databases. The method links the related fields in the two or more of the plurality of heterogeneous databases and makes the data accessible through a virtual warehouse. As schemas change, as new data sources and analysis artifacts are created, the computer implemented method and system can act as a meta-data store, a provenance tracking device, and/or a knowledge management service. | 06-05-2014 |
20140172777 | SYSTEMS AND METHODS FOR INTERVAL CONTROL ELEMENT CHAIN ARCHITECTURE - The present disclosure generally relates to accessing data, and more particularly, to systems and methods for improving the efficiency and quality of real-time extracting, transforming, and/or loading data using customer information control system (CICS) interval control element (ICE) chain processing. | 06-19-2014 |
20140181004 | Common Framework for Definition, Generation, and Management of Metadata Runtime-Loads - A system, a method, and a computer-program product for management of runtime-loads in business process applications are disclosed. Metadata associated with an application is identified. The metadata is configured to enable loading during a runtime of the application and is stored on at least one node in a metadata repository. A runtime-loading structure for loading the metadata during the runtime of the application is generated based on the identified metadata. The runtime-loading structure contains an identification of the at least one node storing the identified metadata in the metadata repository. The identified metadata is loaded based on the generated runtime-loading structure during the runtime of the application. | 06-26-2014 |
20140181005 | EFFICIENT ASSIGNMENT OF QUERY REQUESTS BETWEEN ANALYTICAL ENGINES OF DIFFERING COMPLEXITY - Runtime performance of report generation and other response to query requests can be improve using a report analyzer that can parse the query request and distribute the necessary analytical task among two or more analytical engines of varying levels of complexity. | 06-26-2014 |
20140181006 | EXTRACT-TRANSFORM-LOAD PROCESSOR CONTROLLER - A controller is coupled to an Extract-Transform-Load (ETL) processor, which is connected to data storage devices. The controller comprises: a hardware storage device; a storage control component for storing, on said hardware storage device, a set of criteria for preferredness of ETL stage placements; an I/O input device detecting component for recognizing a proposed placement of a stage on a GUI canvas on the display device; an analytical component for analyzing an eventual result of the proposed placement in an ETL activity represented on the GUI canvas; a comparator for comparing the eventual result of the proposed placement in the ETL activity with the set of criteria; and an indicator control component for, responsive to an outcome of an operation of the comparator, providing to the user an indicator of a degree of preferredness of said proposed placement according to the set of criteria. | 06-26-2014 |
20140188783 | SAMPLING TRANSACTIONS FROM MULTI-LEVEL LOG FILE RECORDS - A log file contains operation records, each operation record is of a certain type, and each operation record is associated with a transaction. A plurality of operation records is read from the log file into a record store. Records of the plurality of operation records of each operation record type are sampled at a predefined sampling rate. Operation records in the plurality of operations records are identified that are associated with completed transactions of which the sampled operation records are associated. The identified operation records are then extracted from the record store into a data store. | 07-03-2014 |
20140188784 | SYSTEMS AND METHODS FOR DATA-WAREHOUSING TO FACILITATE ADVANCED BUSINESS ANALYTIC ASSESSMENT - Certain embodiments contemplate systems and methods for improving the speed and efficiency of a data warehouse. In some embodiments, an ETL process is modified to perform a joined indexing operation which reduces the number of lookup requests required. Certain embodiments contemplate a date dimension and hierarchical data structure which improve operation speed. Still other embodiments contemplate structural organizations of biographical fact tables to better improve data access. | 07-03-2014 |
20140188785 | INFORMATION PROCESSING DEVICE, COMPUTER-READABLE RECORDING MEDIUM, AND NODE EXTRACTION METHOD - A process includes selecting a node associated with an extracted node and a node associated with the node accepted by an acceptance unit accepting a node from nodes included in a node link obtained from text relating to a specific matter, accepting a first item in which a registered content of an analysis target is registered, from items in a table relating to the specific matter, accepting the registered content of the analysis target, a second item for use in calculating degrees of association between registered contents in the first item in the table and nodes; calculating the degrees of association between each of the registered content in the first item in the table and the nodes selected, according to a registered content in the second item in the table; and extracting a node associated with the analysis target from the selected nodes, according to the calculated degrees of association. | 07-03-2014 |
20140188786 | SYSTEM AND METHOD FOR IDENTIFYING THE CONTEXT OF MULTIMEDIA CONTENT ELEMENTS DISPLAYED IN A WEB-PAGE AND PROVIDING CONTEXTUAL FILTERS RESPECTIVE THERETO - A method and system for providing contextual filters respective of an identified context of a plurality of multimedia content elements are provided. The method comprises receiving the plurality of multimedia content elements; generating at least one signature for each of the plurality of multimedia content elements; determining a context of each of the plurality of multimedia content elements based on its respective at least one signature, wherein a context is determined as the correlation among a plurality of cluster of signatures; and providing at least one contextual filter respective of the context of each of the plurality of multimedia content elements. | 07-03-2014 |
20140201129 | EFFICIENT QUERY PROCESSING USING HISTOGRAMS IN A COLUMNAR DATABASE - A probabilistic data structure is generated for efficient query processing using a histogram for unsorted data in a column of a columnar database. A bucket range size is determined for multiples buckets of a histogram of a column in a columnar database table. In at least some embodiments, the histogram may be a height-balanced histogram. A probabilistic data structure is generated to indicate for which particular buckets in the histogram there is a data value stored in the data block. When an indication of a query directed to the column for select data is received, the probabilistic data structure for each of the data blocks storing data for the column may be examined to determine particular ones of the data blocks which do not need to be read in order to service the query for the select data. | 07-17-2014 |
20140207728 | Systems for Storing Data Streams in a Distributed Environment - Systems and methods for storing and retrieving data elements transmitted via data streams received from distributed devices connected via a network. The received data elements may be stored in block stores on the distributed devices. The stored data-elements may be allocated to data-blocks of a block-store that have assigned block-identifiers and further allocated to events of the data-blocks that have assigned token-names. Stream-schema of the received data-streams may comprise a list of token-names and an index-definition for each corresponding data-stream. Indices may be generated for the event-allocated data-elements. A query may be executed in order to retrieve data-elements of the received data-streams based on the indices. | 07-24-2014 |
20140207729 | Rapid Provisioning of Information for Business Analytics - An approach is provided in which a system creates schema terms based upon matching input data query requirements to industry terms. In turn, the system generates a query and an associative map, which includes data organized according to the schema terms. The system executes the query, which retrieves the data from the associative map and loads the data into one or more storage areas. | 07-24-2014 |
20140214753 | SYSTEMS AND METHODS FOR MULTI-SOURCE DATA-WAREHOUSING - Preferred embodiments of the invention provide systems and methods for improving the speed and efficiency of a data warehouse. The invention enables the loading of data from different data sources into a common data warehouse structure. Preferred embodiments include an ETL process is modified to perform a joined indexing operation which reduces the number of lookup requests required. Further embodiments contemplate a date dimension and hierarchical data structure which improve operation speed. | 07-31-2014 |
20140236884 | ANALYSIS OF DATASETS WITHOUT PREDEFINED DIMENSIONS - Embodiments of systems, methods, and computer-readable mediums for analysis of datasets without predefined dimensions are generally described herein. In some embodiments, analysis of datasets without predefined dimensions may include receiving a selection of multiple dimensions of information from a database to be used for data analysis; receiving a selection for a type of report; and dynamically generating a query statement based on the selection of the multiple dimensions. Further embodiments may include the dynamically generated query statement including multiple iterative join clauses as a function of the selected dimensions; executing the query statement against an in-memory database; and displaying columns of information obtained from execution of the query statement. | 08-21-2014 |
20140244570 | OPTIMIZING AND MANAGING EXECUTION OF HYBRID FLOWS - Disclosed herein are techniques for optimizing and managing the execution of hybrid flows. An execution plan is generated for each hybrid flow based at least partially on attributes associated therewith. The execution of each hybrid flow is managed in accordance with the execution plan. | 08-28-2014 |
20140250054 | Methods and Systems for Calculating and Retrieving Analytic Data - A data management system for calculating and retrieving analytic data involves a computing device coupled to a database. The system includes a measure data structure within the database including a measure-data element, a collection data structure within the database including a collection-data element, and a fact data structure within the database, and configured to store a fact-data element which includes an aggregated value, a link to the measure-data element, and a link to the collection-data element. The system also includes a fact-analyzer module configured to generate the aggregated value from the transaction data utilizing at least the measure-data element and the collection-data element, wherein the measure-data element defines how the aggregated value is calculated and the collection-data element defines a categorization of the aggregated value. The fact-analyzer module is also configured to store the aggregated value and links to the measure-data element and the collection-data element in the fact-data element. | 09-04-2014 |
20140258209 | Systems and Methods for Interest-Driven Distributed Data Server Systems - Systems and methods for interest-driven business intelligence systems in accordance with embodiments of the invention are illustrated. In one embodiment, an interest-driven distributed data server system includes a processor, an interest-driven distributed data server application, aggregate data storage, and a data source directory, wherein the interest-driven distributed data server application configures the processor to obtain reporting data requirements, determine at least one data dimension using the reporting data requirements, identify at least one data source identified that can provide data associated with the data dimension, generate data source job data, transmit the data source job data to at least one data source, and receive job response data, store the received job response data in the aggregate data storage, build reporting data based on the reporting data requirements, and transmit the reporting data. | 09-11-2014 |
20140258210 | BUSINESS INTELLIGENCE REPORTS WITH NAVIGABLE REFERENCE INDICATORS - The present subject matter relates to generation of business intelligence reports with navigable reference indicators. The method comprises obtaining information from an information source for providing in a business intelligence report. Further, report parameters are obtained through a user interface, wherein the report parameters include at least notes and indicator parameters. The notes and indicator parameters indicate a plurality of navigable reference indicators to be included in the business intelligence report. The method further comprises mapping the notes and indicator parameters with the obtained information. Based on the mapping, the business intelligence report is generated with the plurality of navigable reference indicators. Further, the plurality of navigable reference indicators provides a forward and backward link between two or more sections of the business intelligence report. | 09-11-2014 |
20140279826 | DATA FEEDS PLATFORM - A data feeds platform, comprising a data feed using a data feed protocol. The data feed protocol comprises, a feed header defining at least one identification attribute, a data header defining a data structure comprising data nodes, each data node corresponding to a subset of the data, and a data section, the data section storing the data according to the data structure. A plurality of heterogeneous data protocols is stored. A data input is received in one of the plurality of heterogeneous data protocols. The data input is converted to the data feed protocol, creating the data feed. | 09-18-2014 |
20140279827 | FLOATING TIME DIMENSION DESIGN - A method, system, and computer program product for managing time dimension tables are disclosed. In a first aspect, the method comprises providing an intelligent time dimension agent application and optimizing a time dimension table by the intelligent time dimension agent based upon changing data in associated fact tables. In a second aspect, the system comprises a processor and a memory coupled to the processor, wherein the memory includes an application which, when executed by the processor, causes the processor to carry out the steps of the method. In a third aspect, the computer program product comprises a non-transitory computer readable medium having a computer readable program recorded thereon, wherein the computer readable program, when executed on a computer device, causes the computer device to carry out the steps of the method. | 09-18-2014 |
20140279828 | CONTROL DATA DRIVEN MODIFICATIONS AND GENERATION OF NEW SCHEMA DURING RUNTIME OPERATIONS - A computational device receives input data and control data, where the control data includes instructions to modify one or more operations performed during a runtime execution associated with the input data. The control data is processed to modify the one or more operations during the runtime execution associated of the input data. | 09-18-2014 |
20140279829 | GEO-Enabling of Data Sources - A method includes receiving data from a non-geo enabled data source, obtaining information related to location in the received data, converting the obtained information to a standardized form of geo-location data, and storing the geo-location data. | 09-18-2014 |
20140279830 | DATA INTEGRATION USING AUTOMATED DATA PROCESSING BASED ON TARGET METADATA - Approaches for data integration between multiple IT sources using automated data processing based on target metadata are provided. Specifically, an integration component is configured to load a mapped data set into a table with delta processing based on a configuration object containing, e.g., the source data location, target table name, and source to target mapping. The integration component uses the target metadata to validate data, identify changes, generate the necessary database programming language (e.g., structured query language (SQL)), and run the database programming language with data binding to perform the actual data updates. The integration component leverages the data target metadata to automate the processing of source data, thus providing a way to validate the data, and identify delta changes at the field level between the source and target. This significantly reduces the overall development effort, while providing consistency in record handling and error reporting. | 09-18-2014 |
20140279831 | DATA MODELING TECHNIQUES - Techniques for data modeling are provided. Enterprise data is organized into reference data for entities that an enterprise wants to track and monitor. Relationship data is created that establishes relationships among the various entities within the enterprise data. The reference data and the relationship data are published within an enterprise data warehouse for accessing the enterprise data. | 09-18-2014 |
20140279832 | SYSTEMS, DEVICES, AND METHODS FOR GENERATION OF CONTEXTUAL OBJECTS MAPPED BY DIMENSIONAL DATA TO DATA MEASURES - This disclosure in part describes systems, devices, and methods for implementing a unification mechanism that enables the unification of various data sets, which can permit greater collaboration between users of a data management platform. For example, a system can be configured to enable a user to add or associate a contextual object with a data measure. In associating the contextual object with the data measure, the system can store the contextual object with the unique dimension coordinates that are associated with data measure. Other users of the system may, in turn, access the data measure and be shown the contextual object associated with the data measure. | 09-18-2014 |
20140279833 | METHOD TO REDUCE LARGE OLAP CUBE SIZE USING CELL SELECTION RULES - Various embodiments are directed to techniques for providing one or more reduced-size rule cubes indicating cell rules. A computer-program product embodied in a machine-readable storage medium includes instructions to cause a computing device to select a cell rule to include in a rule cube based on applicability of the cell rule to a selected portion of a data cube; analyze the cell rule to identify a wildcarded dimension in a specification of cells of the data cube that are subject to the cell rule; and generate the rule cube indicating applicability of the cell rule to the selected portion of the data cube, wherein a cell of the rule cube corresponds to multiple cells of the data cube, and wherein the wildcarded dimension of the rule cube is reduced in length in comparison to a length of the wildcarded dimension of the data cube. Other embodiments are described and claimed. | 09-18-2014 |
20140279834 | Scalable Analysis Platform For Semi-Structured Data - A method of operating a data analysis system includes retrieving objects from a data source. Each of the retrieved objects includes (i) data and (ii) metadata describing the data. The method further includes dynamically creating a cumulative schema by, for each object of the retrieved objects: (i) inferring a schema from the object based on the metadata of the object and inferred data types of elements of the data of the object, (ii) creating a unified schema, wherein the unified schema describes both (a) the object described by the inferred schema and (b) a cumulative set of objects described by the cumulative schema, and (iii) storing the unified schema as the cumulative schema. The method further includes exporting the data of each of the retrieved objects to a data warehouse. | 09-18-2014 |
20140279835 | SELF-ANALYZING DATA PROCESSING JOB TO DETERMINE DATA QUALITY ISSUES - Techniques are disclosed to determine data quality issues in data processing jobs. The data processing job is received, the data processing job specifying one or more processing steps designed based on one or more data schemas and further specifies one or more desired quality metrics to measure at the one or more processing steps. One or more state machines are provided, that are generated based on the quality metrics and on the data schemas. Input data to the data process job are processed using the one or more state machines, in order to generate output data and a set of data quality records characterizing a set of data quality issues identified during the execution of the data processing job. | 09-18-2014 |
20140289183 | KEY FIGURE DATA FILTERS IN OLAP WITH HEIRARCHIES - A system and method of key figure data filters are presented. The key figure data filters are implemented in an analytical engine of a business warehouse system. The key figure data filters employ conditions, which can be expressed as a kind of selection that describe a set. A key figure data algorithm can be implemented by the analytical engine using the conditions, yet still respect hierarchies in the business warehouse database. | 09-25-2014 |
20140297583 | Determining Statistics for Cost-Based Optimization of a Workflow - Techniques, systems, and articles of manufacture for determining statistics for cost-based optimization of a workflow. A method includes generating individual sets of statistics for each intermediate relation of a workflow, wherein said intermediate relations comprise results of stages of any plan of the workflow, and wherein each individual set of statistics computes cardinality of the corresponding intermediate relation; determining a global set of statistics for the workflow, wherein said global set of statistics comprises at least one of the individual sets of statistics for each of the intermediate relations; instrumenting a given plan of the workflow to collect the global set of statistics during execution; executing the given plan to collect the global set of statistics; and determining a plan of the workflow with the lowest cost by comparing the cost of multiple plans, wherein the cost of each plan is derived from the global set of statistics. | 10-02-2014 |
20140297584 | CONTEXT TRANSFER FOR DATA STORAGE - A method for context transfer for data storage is disclosed. The method includes receiving, from an analysis environment, a request to load a data item for analysis, creating, based on the request, a task request including context information of the data item, transmitting the task request to a data manager environment, receiving, from the data manager environment, an approval of the task request based on the context information, and loading the data item in a master data store in response to the approval. | 10-02-2014 |
20140304215 | Methods and Systems for Creating and Storing Metadata - One embodiment of the present disclosure is a method of creating metadata during object development. The method comprises receiving a change to an object during its development that results in a changed version of the object, identifying information about the change, and creating metadata comprising the information about the change. The information about the change may include a unique instance identifier identifying and unique to the changed version of the object. As an object is changed multiple times during development, the created metadata may include a series of information segments each relating to a particular change and each uniquely identified by its unique instance identifier. The information about the change may also include, as examples, an identification of a unique instance identifier of a prior version of the object, the time of the change to the object, and/or identification of the software used to make the change. Other embodiments relate to compound objects, part mapping, and modification of ingredients, among other things. | 10-09-2014 |
20140304216 | SEARCHABLE SCREEN SHARING SESSIONS - Systems and methods are provided for searchable screen sharing sessions. A system records at least one image associated with a screen sharing session. The system extracts content from the at least one image. The system stores the content as searchable content associated with the screen sharing session. | 10-09-2014 |
20140304217 | METHOD AND SYSTEM FOR IMPLEMENTING AN ON-DEMAND DATA WAREHOUSE - Disclosed are methods and apparatuses for reducing the number of database calls when transferring data from a database to a server. One embodiment provides an approach for utilizing staging tables to collate and transform core data and related data in the database before transferring the data to the server, so that the number of database calls used may be reduced. | 10-09-2014 |
20140310231 | SYSTEM AND METHOD FOR AUTOMATING DATA WAREHOUSING PROCESSES - A system and computer-implemented method for automating data warehousing processes is provided. The system comprises a code generator configured to generate codes for Extract, Transform and Load (ETL) tools, wherein the codes facilitate the ETL tools in extracting, transforming and loading data read from data sources. The system further comprises a code reviewer configured to review and analyze the generated codes. Furthermore, the system comprises a data migration module configured to facilitate migrating the data read from the data sources to one or more data warehouses. Also, the system comprises a data generator configured to mask the data read from the data sources to generate processed data. In addition, the system comprises a Data Warehouse Quality Assurance module configured to facilitate testing the read and the processed data. The system further comprises a reporting module configured to provide status reports on the data warehousing processes. | 10-16-2014 |
20140310232 | AGGREGATE QUERY-CACHING IN DATABASES ARCHITECTURES WITH A DIFFERENTIAL BUFFER AND A MAIN STORE - The invention relates to a computer system for both online transaction processing and online analytical processing, comprising: a processor coupled to a database, the database comprising the database comprising: a main store ( | 10-16-2014 |
20140310233 | TRACING DATA THROUGH A TRANSFORMATION PROCESS USING TRACER CODES - Systems and methods to provide tracing of data through a transformation process using tracers added to the data during the transformation. In one embodiment, a method includes: receiving a plurality of data records; processing the data records using a plurality of transformation rules, wherein each of the transformation rules corresponds to a respective one of a plurality of tracers; and adding the tracers to the data records as needed to indicate those of the transformation rules used to process each data record. After processing the data records using the transformation rules, the data records are provided to a target system, and the transformation process is validated using the tracers added to the data records. | 10-16-2014 |
20140317046 | METHOD AND SYSTEM FOR SCORING DATA IN A DATABASE - Computer system and method are provided which include or utilize first and second databases, and a receiver for receiving online data. The first database stores received online data in a first data structure, and the second database stores the online data in a second data structure, with the second structure being obtained by applying a transformation to the online data in the first data structure. The system also has a software component for generating an analytical model using at least part of the online data stored in the second database, and a processor and memory storing a plurality of applications that are executable on the processor, where a mapping application includes instructions that, when executed, cause at runtime the processor to automatically retrieve at least part of the online data from the second database, and store in a storage of the system the retrieved online data. | 10-23-2014 |
20140330766 | Positions and Interests Map - A method to search concrete positions based on abstract interests or vice versa is disclosed. Based on user's initial inputs of interests to positions mapping (PI Map), this method extends the map further with other users' similar one. Weights are associated to each interest to position connection in the map to represent preferences. Multiple layers in the map are interconnected to have different level of abstraction. A personal PI Map database is established based on these initial inputs, extension, weights and multiple layers. When user inquires specific interests with constrain and execution order, the method use the PI Map to find the optimal route to concrete positions. Based on the users' usage and final choice, the PI Map is dynamically adjusted through connections and associated weight. Further data analytics is conducted based on the aggregated PI Map from a group of users. | 11-06-2014 |
20140337276 | VIRTUAL PERSISTENCE - A virtual persistence platform is described that provides requirement-driven persistence store orchestration. Application developers can annotate program code with persistence properties and requirements, and a compiler component of the virtual persistence platform generates a persistence service level agreement (PSLA) that becomes attached to the application. After deployment and launch, the virtual persistence platform uses the PSLA to handle persistence commands from the application and perform persistence store orchestration. | 11-13-2014 |
20140344210 | USE OF PROJECTOR AND SELECTOR COMPONENT TYPES FOR ETL MAP DESIGN - A data integration system is disclosed that incorporates one or more techniques for simplifying the design and maintenance of a mapping. As components are added or removed to an existing design, the data integration system removes the need to specify all input and output attributes. In one aspect, components types are implemented that allow assignment expressions to reference all or part of upstream components. Therefore, attributes of certain types of components can be propagated to downstream components or otherwise inherited from upstream components with minimal effort on the part of a map designer. During code generation the attributes required to be projected by any component can be derived based on the needs of the downstream components. | 11-20-2014 |
20140344211 | SUPPORTING COMBINATION OF FLOW BASED ETL AND ENTITY RELATIONSHIP BASED ETL - A data integration system is disclosed that incorporates one or more techniques for eases the design and maintenance of a mapping. As components are added to an existing design, the data integration system removes the need to specify all input and output attributes. In one aspect, components types are implement that allow entity relationships to be added and/or edited in a flow view of a logical design. Therefore, attributes of components representing datasets can be added and propagated to downstream components with minimal effort on the part of a map designer. | 11-20-2014 |
20140351200 | PIVOT ANALYSIS METHOD USING CONDITION GROUP - A pivot analysis method using a condition group is applicable to a front end processing device and includes the following steps. A file name processing interface is presented according to a plurality of file names. A file name selection value is received via a first sub interface of the file name processing interface. A range setting value is received via a second sub interface of the file name processing interface. A file name computing value is received via a third sub interface of the file name processing interface. A file name processing value generated by integrating the file name selection value, the file name range setting value and the file name computing value, is outputted. A multidimensional array is received to present a pivot analysis table. | 11-27-2014 |
20140358844 | WORKFLOW CONTROLLER COMPATIBILITY - Aspects of the disclosure relate to workflow controller compatibility for in order to format workflow data such that communication between incompatible workflow controllers is facilitated. A compatibility processing module may execute one or more processes to format a workflow such that communication between previously incompatible workflow controllers is facilitated. A first workflow controller is configured with a compatibility processing module and receives a large number of data points. The compatibility module executes one or more processes on a workflow such that the processed workflow is compatible with a second workflow controller. | 12-04-2014 |
20140358845 | DATA WAREHOUSE COMPATIBILITY - A compatibility processing module, for executing one or more processes to format and manipulate data, such that communication between previously-incompatible data warehouses is facilitated. In particular, a first warehouse is disclosed, wherein the first data warehouse is configured with a compatibility processing module, for receiving a large number of data points, and for executing one or more processes on a stored portion of the received data points such that the resulting processed data points are compatible with formatting conventions of a second data warehouse. | 12-04-2014 |
20140372365 | Optimization of Business Warehouse Queries by Calculation Engines - A query is received by a database server from a remote application server. The query is associated with a calculation scenario that defines a data flow model that includes one or more calculation nodes with each calculation node defining one or more operations to execute on the database server. Thereafter, at least one of the calculation nodes is transformed into a star schema. Next, the database server instantiates the calculation scenario with the transformed at least one calculation node. Subsequently, a calculation engine of the database server executes the operations defined by the calculation nodes of the instantiated calculation scenario other than the transformed at least one calculation node and, additionally, an OLAP engine executes the transformed at least one calculation node to collectively result in a responsive data set. | 12-18-2014 |
20140372366 | ADAPTIVE WAREHOUSE DATA VALIDATION TOOL - Techniques for data validation may include dynamically generating one or more database queries to be performed on a target data warehouse and a baseline data warehouse based on warehouse model metadata for the target data warehouse and the baseline data warehouse. The techniques may further include executing the one or more database queries against the target data warehouse and the baseline data warehouse to receive one or more data sets from the baseline data warehouse and one or more data sets from the target data warehouse. The techniques may further include comparing the one or more data sets from the baseline data warehouse and the one or more data sets from the target data warehouse to validate target data in the target data warehouse against baseline data in the baseline data warehouse. | 12-18-2014 |
20150012477 | DATA LINEAGE NOTIFICATION TOOLS - Resources for data lineage discovery, data lineage analysis, role-based security, notification. The resources may include or involve machine readable memory that is configured to store a technical data element (“TDE”); a receiver that is configured to receive a query for data lineage information corresponding to a business element identifier; and a processor configured to: register a logical association between the business element identifier and the TDE; and formulate the data lineage information of the TDE associated with the business element identifier. The receiver may be configured to receive a criterion that is required to access one or more technical data elements (“TDEs”) associated with the business element identifier. The receiver may be configured to receive an election to receive a notification of a change of data lineage. The processor may be configured to toggle between a first data lineage graph and a second data lineage graph. | 01-08-2015 |
20150012478 | DATA LINEAGE TRANSFORMATION ANALYSIS - Resources for data lineage discovery, data lineage analysis, role-based security, notification. The resources may include or involve machine readable memory that is configured to store a technical data element (“TDE”); a receiver that is configured to receive a query for data lineage information corresponding to a business element identifier; and a processor configured to: register a logical association between the business element identifier and the TDE; and formulate the data lineage information of the TDE associated with the business element identifier. The receiver may be configured to receive a criterion that is required to access one or more technical data elements (“TDEs”) associated with the business element identifier. The receiver may be configured to receive an election to receive a notification of a change of data lineage. The processor may be configured to toggle between a first data lineage graph and a second data lineage graph. | 01-08-2015 |
20150012479 | MULTIMEDIA INSPECTION DATABASE SYSTEM (MIDaS) FOR DYNAMIC RUN-TIME DATA EVALUATION - An object-oriented system provides a comprehensive and integrated tool to manage, operate, analyze and inspect structures by providing 3-D visualization of the structures, a database interface, connectivity with one or more databases, a data display and analysis capabilities. The structural model is stored as an object-oriented, serialized file as a series of objects, including primitives, coordinates, object names, group names and other object identifiers. The system integrates various types of data, including tabulated textual data, annotated engineering drawings, photographic records, graphical plots, audio and videotaped records, from different tables and databases. Data are displayed to the user based on queries that the user submits to the system, which queries are evaluated at run-time. As a result, the user achieves unparalleled flexibility in data analysis and evaluation. | 01-08-2015 |
20150012480 | FLOATING TIME DIMENSION DESIGN - A method, system, and computer program product for managing time dimension tables are disclosed. In a first aspect, the method comprises providing an intelligent time dimension agent application and optimizing a time dimension table by the intelligent time dimension agent based upon changing data in associated fact tables. In a second aspect, the system comprises a processor and a memory coupled to the processor, wherein the memory includes an application which, when executed by the processor, causes the processor to carry out the steps of the method. In a third aspect, the computer program product comprises a non-transitory computer readable medium having a computer readable program recorded thereon, wherein the computer readable program, when executed on a computer device, causes the computer device to carry out the steps of the method. | 01-08-2015 |
20150019476 | SYSTEMS AND METHODS FOR CONVERTING DATA FROM LEGACY COMPUTER SYSTEMS INTO TARGET FORMATS FOR TARGET COMPUTER SYSTEMS - A computer system for converting data from one of a plurality of different legacy systems to one of a plurality of different target systems includes one or more processors, memory, a database stored in the memory, and a software framework stored in the memory for execution by the one or more processors. Each of the plurality of different legacy systems has a legacy data format and each of the plurality of different target systems has a target data format. The software framework includes a plurality of software components callable by an output adaptor for performing a plurality of data conversion functions. The software framework is configured to interact with each of the plurality of different legacy systems having the legacy data format and/or each of the plurality of different target systems having the target data format. Other example computer systems and methods are also disclosed. | 01-15-2015 |
20150019477 | OUTPUT DRIVEN GENERATION OF A COMBINED SCHEMA FROM A PLURALITY OF INPUT DATA SCHEMAS - A computational device receives a plurality of versions of an input data schema. At least one element is selected from the plurality of versions of the input data schema based on an expected result. A combined schema is generated based on the at least one selected element. The input data is processed according to the combined schema. | 01-15-2015 |
20150026114 | SYSTEM AND METHOD OF AUTOMATICALLY EXTRACTING DATA FROM PLURALITY OF DATA SOURCES AND LOADING THE SAME TO PLURALITY OF TARGET DATABASES - The present invention discloses system and method for automatically extracting data from plurality of data sources in various formats through source channels and loading data to plurality of target databases through connectors. The system includes a data transformation module for transforming data received from the plurality of data sources, a data processing module for automatically analyzing and organising the received data for loading into the plurality of target databases, and a metadata repository for storing metadata of the processed data for future usage. The metadata regarding data structure of the data sources is automatically extracted from the data sources and used to create predefined data structures of the target databases. The data processing module includes a data input handling module for identifying mime-type, extension and the metadata of the data sources, a data structure identification module for identifying type and subtype of the data sources and a target-data-structure creation module for creating the predefined data structures of the target databases. | 01-22-2015 |
20150026115 | CREATION OF CHANGE-BASED DATA INTEGRATION JOBS - A computer software implemented method for transforming a first extract transform load (ETL) job having at least some unload transform load (UTL) portions. The method includes the following steps: (i) decomposing the first ETL job into an intermediate set of one or more jobs; and (ii) for each job of the intermediate set, transforming the job into a transactionally equivalent job to yield a final set of one or more jobs. The decomposing is performed so that each job of the intermediate jobs set is a Simple UTL job. The transforming is performed so that each job of the final set includes no UTL portions. | 01-22-2015 |
20150026116 | MANAGING SPARSITY IN AN MULTIDIMENSIONAL DATA STRUCTURE - Embodiments of the present invention relate to a method, program product, and system for managing the percentage of unpopulated cells in a multidimensional data structure during the servicing of multidimensional analytical queries. In an embodiment, a multidimensional data structure is stored in a memory location of a computing device, wherein the multidimensional data structure includes a dimension that is a subset of a plurality of dimensions included in a relational database. The computing device determines a frequency of utilization of a dimension to service a multidimensional query request. The computing device combines the dimension with the multidimensional data structure responsive to the frequency of utilization. | 01-22-2015 |
20150026117 | METHOD AND SYSTEM FOR RECONCILING META-DATA IN A DATA WAREHOUSE - A data warehouse system comprising a warehouse database and a data warehouse manager communicatively coupled to the database. The warehouse manager is to store data to and retrieve data from the warehouse database and perform metadata reconciliation to update metadata of objects in the data warehouse to reflect a change in metadata of a first object in the warehouse by identifying a first set of objects impacted by the change to the metadata of the first object, identifying a second set of objects on which the first object depends, identifying at least one task that has to be completed to propagate the impact of the change in the metadata of the first object to objects in the first and second sets of objects, performing the task, and dynamically computing additional tasks based on whether or not definitions of objects in the first or second sets of objects changes based upon completion of the at least one task. | 01-22-2015 |
20150046389 | POST-MIGRATION VALIDATION OF ETL JOBS AND EXCEPTION MANAGEMENT - Handling extract-transform-load (ETL) job mismatches as “exceptions.” Exception handling may include the following steps: (i) determining a mismatch while running an extract-transform-load job with the mismatch being a mismatch of at least one of the following types: design time information mismatch, and/or operational metadata mismatch; and (ii) responsive to determining the mismatch, handling the mismatch as an exception. | 02-12-2015 |
20150046390 | Storing Metadata in a Table-based Format from Metadata in a Serial Format - In one embodiment, a method receives a document including metadata in a serial format and parses the document to determine components in the metadata, a nested structure of the components, and an order of the components in the data. The metadata is stored in a table format where columns of the table define the nested structure of the components and the order of the components in the data. A query is received for the data. The method determines components from the metadata stored in the table format from the columns that satisfy the query where the components are determined based on the nested structure of the determined components and the order of the determined components from information in the columns of the table. A result for the query is output including the determined components based on the nested structure of the determined components and the order of the determined components. | 02-12-2015 |
20150058278 | DETERMINING RELIABILITY OF DATA REPORTS - A computer implemented system for automatically determining a reliability score of a data report. The system includes a data structure indicative of data lineage that includes data assets flowing into a data report. The system further includes a processor configured to calculate trust scores for data assets based on characteristics of the data assets as well as a data report trust density score indicative of the reliability score of the report based on the calculated trust scores. The system further includes an output device configured to output the calculated data report trust density score. | 02-26-2015 |
20150058279 | CONTROL DATA DRIVEN MODIFICATIONS AND GENERATION OF NEW SCHEMA DURING RUNTIME OPERATIONS - A computational device receives input data and control data, where the control data includes instructions to modify one or more operations performed during a runtime execution associated with the input data. The control data is processed to modify the one or more operations during the runtime execution associated of the input data. | 02-26-2015 |
20150058280 | DATA QUALITY MONITORING - A computer implemented method, computer program product and system for data quality monitoring includes measuring a data quality of loaded data relative to a predefined data quality metric. The measuring the data quality includes identifying delta changes in at least one of the loaded data and the data quality rules relative to a previous measurement of the data quality of the loaded data. Logical calculus defined in the data quality rules is applied to the identified delta changes. | 02-26-2015 |
20150066838 | Annotation of Data in an Operating Plan Data Aggregation System - In an embodiment a method is provided. The method includes receiving forecast information in a computer from a first group of users. The method also includes integrating the forecast information into a database and corresponding OLAP cube accessible by the computer. The method further includes receiving comments on specific entries of the forecast information in the computer from a second group of users. The method also includes receiving changes of the forecast information in the computer from a second group of users The method further includes integrating comments and changes into the database and OLAP cube. Moreover, the method includes providing a display of the comments and the changes to the first group of users in an interface to the computer. | 03-05-2015 |
20150074037 | In Memory Database Warehouse - An in memory database system includes a random access memory to store an in memory database table, a stored set of defined views over the database tables, a stored set of analytical services adapted to provide transaction services and database warehouse services, an engine to run the analytical services on the database table, and an interface to interact with users and provide access to the engine, analytical services, and views to perform transactions and data warehouse functions on the database table. | 03-12-2015 |
20150074038 | Historical Data for In Memory Data Warehouse - A method includes processing a transaction on an in memory database where data being processed has a validity time, updating a time dependent data view responsive to the transaction being processed to capture time validity information regarding the data, and storing the time validity information in a historization table to provide historical access to past time dependent data following expiration of the validity time. | 03-12-2015 |
20150081616 | DATABASE INSERT WITH DEFERRED MATERIALIZATION - According to one embodiment of the present invention, a system inserts data into a database object. The system associates the database object with a parameter specifying materialization of data for the database object. The system inserts data into the database object and materializes the data in accordance with the parameter to provide access to the data from the database object, wherein the parameter specifies a portion of the data to be materialized upon insertion. Embodiments of the present invention further include a method and computer program product for inserting data into a database object in substantially the same manners described above. | 03-19-2015 |
20150081617 | OUT OF HOME MEDIA MEASUREMENT - A system may include at least one data source configured to provide network usage data indicative of the existence of communications with subscriber devices, and web and application usage data indicative of data usage of the subscriber network by the subscriber devices. The system may also include a data warehouse server configured to perform operations including correlating the network usage data and web and application usage data into subscriber-level data; associating, with the subscriber-level data, subscriber attributes indicative of a preference of the subscriber for content in a particular category of content, and profile attributes indicative of demographic characteristics of the subscriber; matching the subscriber-level data with a set of subscriber profiles, each of the set of subscriber profiles including a set of subscriber attributes and profile attributes associated with the respective subscriber profile; and aggregating the subscriber-level data into aggregate-level data according to the matching subscriber profiles. | 03-19-2015 |
20150081618 | Systems and Methods for Interest-Driven Business Intelligence Systems Including Event-Oriented Data - Systems and methods for interest-driven business intelligence systems including event-oriented data in accordance with embodiments of the invention are illustrated. In one embodiment, an interest-driven business intelligence system includes raw data storage configured to contain raw data and perform ETL processes, a data mart configured to contain metadata that describes the raw data, and an intermediate processing layer, wherein the intermediate processing layer is configured to compile an interest-driven data pipeline configured to generate ETL processing jobs to generate event-oriented data from the raw data by filtering the raw data based on the metadata describing the raw data, determining ordering data based on the metadata describing the raw data, aligning the filtered raw data based on the ordering data, and generating event-oriented data based on the aligned filtered raw data, and storing the event-oriented data in the data mart. | 03-19-2015 |
20150081619 | Systems and Methods for Interest-Driven Business Intelligence Systems Including Geo-Spatial Data - Systems and methods for interest-driven business intelligence systems including geo-spatial data in accordance with embodiments of the invention are illustrated. An interest-driven business intelligence system including raw data storage and perform extract, transform, and load processes, a data mart, and an intermediate processing layer, wherein the intermediate processing layer is configured to automatically generate metadata describing the raw data, derive reporting data requirements, and compile an interest-driven data pipeline based upon the reporting data requirements, where compiling the interest-driven data pipeline includes generating ETL processing jobs to generate geo-spatial data from the raw data, determining bounding data, bounding the filtered raw data based on the bounding data, generating geo-spatial data, and storing the geo-spatial data, generating reporting data including data satisfying the reporting data requirements based on the geo-spatial data, and storing the reporting data in the data mart for exploration by an interest-driven data visualization system. | 03-19-2015 |
20150081620 | Component Based Approach to Building Data Integration Tools - A method facilitating interoperability across the integrating platforms is provided by composing purpose specific integrating tools using conceptual modeling framework and a set of primitive foundational components. The model driven architecture provides easy migration from one solution architecture to another. Also, the present invention discloses a method to migrate from one implementation to another using set of composed purpose specific integration tools. | 03-19-2015 |
20150088806 | SUPPORTING MULTI-TENANCY IN A FEDERATED DATA MANAGEMENT SYSTEM - A system and method for supporting multi-tenancy in a federated data management system are provided herein. The method includes receiving a query from a client at a data management server, wherein the query includes a tenant property condition corresponding to the client. The method also includes identifying an external data store including data specified by the query and determining whether the external data store is multi-tenant enabled. The method further includes removing the tenant property condition and executing the query to retrieve the specified data if the external data store is not multi-tenant enabled. | 03-26-2015 |
20150088807 | SYSTEM AND METHOD FOR GRANULAR SCALABILITY IN ANALYTICAL DATA PROCESSING - A method of resolving data queries in a data processing system. The method comprises receiving in the data processing system a data query, where the data processing system stores a plurality of information units describing pluralities of data elements, a first information unit having a retrieval subunit that includes information for retrieving all unique data elements in a first plurality of data elements and a summary subunit including summarized information about data elements in the first plurality of data elements. The method further includes deriving, via the data processing system, a result of the data query, wherein the result of the data query comprises a plurality of new data elements. The data processing system uses summary subunits of information units to select a set of information units describing data elements that are sufficient to resolve the data query. | 03-26-2015 |
20150100541 | AUTOMATIC GENERATION OF AN EXTRACT, TRANSFORM, LOAD (ETL) JOB - According to one embodiment of the present invention, a method automatically generates one or more Extract, Transform and Load (ETL) jobs. Input data in a source format and output data in a target format is received. The input data and output data is analyzed to determine properties and relationships thereof. One or more mapping models are automatically generated using the properties and relationships, wherein the mapping models describe the mapping and transformation of the input data to the output data. One or more ETL jobs are generated using the mapping models. Embodiments further include a system and program product apparatus for automatically generating one or more ETL jobs. | 04-09-2015 |
20150100542 | AUTOMATIC GENERATION OF AN EXTRACT, TRANSFORM, LOAD (ETL) JOB - According to one embodiment of the present invention, a method automatically generates one or more Extract, Transform and Load (ETL) jobs. Input data in a source format and output data in a target format is received. The input data and output data is analyzed to determine properties and relationships thereof. One or more mapping models are automatically generated using the properties and relationships, wherein the mapping models describe the mapping and transformation of the input data to the output data. One or more ETL jobs are generated using the mapping models. Embodiments further include a system and program product apparatus for automatically generating one or more ETL jobs. | 04-09-2015 |
20150112921 | OPTIMIZATION OF JOB SCHEDULING IN A COMPUTING ENVIRONMENT - Optimizing job scheduling in a data processing system includes determining a time factor associated with a report, determining one or more jobs that contribute to the report, and narrowing execution frequency of the one or more jobs based on the time factor of the report. The time factor provides information about timing details for data included in the report. The jobs are executed in a computing environment according to a time schedule. | 04-23-2015 |
20150120642 | REALTIME SNAPSHOT INDICES - A system and method for realtime snapshot indices is presented. A query is calculated on all target data of a data warehouse, with all variable combinations, to generate a result. The result is stored in a snapshot index associated with the data warehouse. The result is recalcualated to generate a subresult, and the snapshot index is updated with the subresult. A conversion routine is generated to recalculate the subresult into a separate table, and the separate table is then recalculated by a background job to recalculate the subresult. | 04-30-2015 |
20150120643 | TRIPLESTORE REPLICATOR - This invention relates to a system, method and computer program product for replicating triplestore data from generic data records including: a plurality of triplestore mapping operations; a record reader for listening for data events from the generic data record and for matching each data event to one or more triplestore mapping operations; a processor for generating equivalent triples for each located data event by applying the matched triplestore mapping operation; and a data sink for receiving generated equivalent triples so that the triplestore data is synchronized in real time with the generic data records. | 04-30-2015 |
20150120644 | SYSTEM AND METHOD FOR PERFORMING ANALYTICS - A data analytics system includes processing circuitry that receives one or more objects from one or more data sources, and the one or more objects are described based on a common ontology that defines the one or more objects as data objects, manipulation objects, visualization objects, and utility objects. The one or more objects are self-referencing and self-validating. Data pipelines are defined based on input from a user. The data pipelines are executed to perform a runtime instance. | 04-30-2015 |
20150127599 | STATELESS DATABASE CACHE - A cache and method of caching for a database are provided, in which the cache and the database are in the same state. Responsive to a query for data, the method extracts search terms from the query and maps filter settings to an appropriate format. The method also (a) outputs data retrieved from a cache, if there is a cache hit and (b) builds cache and retrieves the data from the database, if there is a cache miss. If there are multiple users and/or instances, the cache may be updated for all users and/or instances by determining whether all changes are to be committed, in which case the changes are pushed to all users and/or instances. | 05-07-2015 |
20150127600 | REPORT DATA MANAGEMENT SERVER AND REPORT DATA MANAGEMENT PROGRAM - A report data management server includes a report data storing unit storing position information and report data in association with each other. The report data management server receives a request for supply of information containing position information, extracts information corresponding to position information contained in the request for supply from the report data storing unit, and supplies extracted information to a user terminal. | 05-07-2015 |
20150134589 | PROCESSING DATA IN DATA MIGRATION - A computer-implemented method for processing information related to an extract-transform-load (ETL) data migration, including aggregating operational metadata and determining: a plurality of metrics, organized by business object, corresponding to the migration; a number of business object instances not successfully loaded; a first end-to-end execution time for at least one business object; relevant input metadata; load readiness status per business object; impact of a business object that is not load ready by analyzing business process hierarchies; business object load readiness by reference to incomplete development status or data defects; scope per test cycle based, at least in part, upon business object load readiness; and high-priority defects of business objects that stop testing based, at least in part, upon analysis of business process hierarchies. | 05-14-2015 |
20150134590 | NORMALIZING AMORPHOUS QUERY RESULT SETS - A method, system, and computer program product for normalizing amorphous query result sets are provided in the illustrative embodiments. A property of data in a portion of the result set is identified. the property is usable for normalizing the portion into a structured data. Based on the property, the portion is categorized into a first category as a candidate for normalization using a first structure specification. The portion is transformed, responsive to the first category being selected for normalizing the portion over a second category in an evaluation, into the structured data according to the first structure specification of the first category. The structured data and a metadata of structure specification are added to a normalized result set. The normalized result set is output to a consumer application. | 05-14-2015 |
20150134591 | METHOD, APPARATUS, AND COMPUTER PROGRAM PRODUCT FOR DATA QUALITY ANALYSIS - A system is disclosed to evaluate data quality in a big data environment. An example method performed by the system includes retrieving one or more rules from an asset catalog. The method further includes retrieving, based on the one or more rules, counter information from a data set, and generating, by a processor, one or more metrics based on the one or more rules and the counter information. In addition, the method includes evaluating, by the processor, the one or more metrics based on the one or more rules. In an instance in which evaluation of a particular metric of the one or more metrics identifies an attribute value that exceeds a predetermined threshold, the method includes causing a notification message regarding the particular metric to be output. A corresponding apparatus and computer program product are also provided. | 05-14-2015 |
20150134592 | EXTRACT TRANSFORM LOAD (ETL) APPLICATIONS FOR JOB MATCHING - Various of the disclosed embodiments present systems and methods for generating consolidated job postings from disparate originating sources and formats. Applying an Extraction Transform Load (ETL) framework to the incoming data, a parallel and asynchronous as well as scalable approach to distributing job posting information is presented. “Extraction” may involve the recognition of salient information in the disparate formats (e.g., in employment listings on company webpages). During “transformation”, the information may be reformatted into a universal format or into a format suitable for use at a given destination system. During “loading”, the reformatted data may be supplied to a suitable destination system, e.g., the Application Programming Interface (APIs) of a job board system. Applications in related domains and various optimizations are also discussed. | 05-14-2015 |
20150134593 | SYSTEM AND METHOD FOR OPTIMIZING COMPUTATION OF METRICS FACILITATING MONITORING OF SERVICE LEVEL AGREEMENTS (SLA) - A system and method for optimizing the computation of metrics facilitating monitoring of service level agreements (SLA) in an IT-enabled framework to reduce computational overhead. A receiving module receives data associated with activities performed by one or more service providers. A transforming module transforms the data into a user-defined data structure using a transforming technique. A selecting module selects an attribute and parameters associated with a key performance indicator (KPI) of the one or more service providers. A view creating module creates a materialized view of the user-defined data structure based on the selected attribute. A query generating module m generates a query to be executed on the materialized view. A metric executing module executes the query on the materialized view in order to compute the metric. | 05-14-2015 |
20150134594 | SYSTEMS AND METHODS FOR MEDICAL INFORMATION DATA WAREHOUSE MANAGEMENT - Systems and methods for data warehouse management for medical information is provided. The system receives a set of medical record documents. These documents include evidence for one or more findings which are identified using natural language processing of evidence. The data set is probabilistically transformed into a structured data set (often as a table). This structured data set includes embedded links that reference the source document where the evidence is located. If the finding is supported by multiple articles of evidence, the link will direct the user to the source document with the highest confidence ranking. Evidence in the source document is highlighted and may include associated annotations. Evidence, findings and annotations may be updated by the user. | 05-14-2015 |
20150134595 | DATA FEEDS PLATFORM - A data feeds platform, comprising a data feed using a data feed protocol. The data feed protocol comprises, a feed header defining at least one identification attribute, a data header defining a data structure comprising data nodes, each data node corresponding to a subset of the data, and a data section, the data section storing the data according to the data structure. A plurality of heterogeneous data protocols is stored. A data input is received in one of the plurality of heterogeneous data protocols. The data input is converted to the data feed protocol, creating the data feed. | 05-14-2015 |
20150142724 | METHOD AND SYSTEM FOR REPRESENTING OLAP QUERIES USING DIRECTED ACYCLIC GRAPH STRUCTURES IN A DATAGRID TO SUPPORT REAL-TIME ANALYTICAL OPERATIONS - The various embodiments herein provide a method for representing OLAP (Online Analytical Processing) queries in a data grid in real time. The method comprises of providing a new OLAP query, building a directed acyclic graph structure for the OLAP query, initializing the graph structure by generating signals for one or more business entities, processing the generated signals of one or more entities, linking entities based on matching of query filter parameters specified in the OLAP query and calculating aggregate values at one or more nodes. The nodes in the directed acyclic graph comprises: a plurality of blue nodes adapted to maintain links to one or more business entities stored in a datagrid and to hold the aggregate values and a plurality of grey nodes adapted to hold the calculated aggregate values. The aggregate values comprise at least one of SUM, ACCOUNT and AVG. | 05-21-2015 |
20150142725 | HIGH-THROUGHPUT EXTRACT-TRANSFORM-LOAD (ETL) OF PROGRAM EVENTS FOR SUBSEQUENT ANALYSIS - An event tap associated with a server, such as a Web server, at a machine can transform a server event into a tuple, select a database node for the tuple, and place the tuple in a queue for that database node, and then flush the queue periodically directly into database notes. The use of an event tap can thus reduce the computational burden on the database while keeping the server event data in the database relatively fresh. | 05-21-2015 |
20150149401 | LOCK-FREE GENERATION OF COLUMNS WITH MINIMAL DICTIONARIES AFTER PARALLEL AGGREGATION - A new dictionary can be created for a result column in a query plan operation executed on a database. The result column can be generated by multiple worker jobs running in parallel to read tasks from a shared queue as part of a query plan operation that includes a group-by column within an input set of input columns. The group-by column can include an original dictionary for all values contained within the group-by column If the new dictionary has fewer entries than the original dictionary for the group-by column such that mapping is required between old value identifiers within the group-by column and new value identifiers within the result column, the old value identifiers are renamed to the new value identifiers using a mapping vector. | 05-28-2015 |
20150149402 | EFFICIENT QUERY PROCESSING USING HISTOGRAMS IN A COLUMNAR DATABASE - A probabilistic data structure is generated for efficient query processing using a histogram for unsorted data in a column of a columnar database. A bucket range size is determined for multiples buckets of a histogram of a column in a columnar database table. In at least some embodiments, the histogram may be a height-balanced histogram. A probabilistic data structure is generated to indicate for which particular buckets in the histogram there is a data value stored in the data block. When an indication of a query directed to the column for select data is received, the probabilistic data structure for each of the data blocks storing data for the column may be examined to determine particular ones of the data blocks which do not need to be read in order to service the query for the select data. | 05-28-2015 |
20150293981 | EXTRACTION DEVICE, DATA PROCESSING SYSTEM, AND EXTRACTION METHOD - An extraction method for extracting a sub query to be converted to a program for processing stream data continuously inputted to a database, from a query including instructions, as sub queries, to be issued to a database management system. The extraction method includes receiving input of the query and a lower limit value of efficiency as processing time per unit memory increase amount. A calculating operation calculates a one memory increase amount and the efficiency using the memory increase amount and the processing time to be reduced. The method selects a sub query whose calculated efficiency is equal to or higher than the lower limit value and extracts the selected sub query as a conversion object on condition that the integrated memory increase amount is equal to or smaller than the maximum memory increase amount. | 10-15-2015 |
20150302007 | System and Methods for Migrating Data - A method of migrating data stored in a source device, the method comprising extracting one or more studies to be migrated from the source device; loading each of the one or more extracted studies into the storage device; receiving an identifier associated with each of the studies that have been loaded to the storage device. At the destination device, the one or more loaded studies are indexed using the identifiers. The method further includes transferring the storage device from a first location to a second location; and unifying the studies stored in the storage device with the indexed studies in the destination device. | 10-22-2015 |
20150302068 | Managing Activities Over Time in an Activity Graph - Systems and processes for managing a data warehouse using an activity graph are disclosed. One example process may include selectively tracking new versions of an entity based on a tracking type, removing entities based on data retention rules, removing entities based on data decay rules, and summarizing decayed data. | 10-22-2015 |
20150302076 | METHOD OF STORING AND EXPRESSING WEB PAGE IN AN ELECTRONIC DEVICE - An electronic device and method are disclosed herein. The electronic device includes a display, and a processor. The processor is configured to control operations of extracting data from at least one web page based on an identified parameter, sorting the extracted data into a particular category of a plurality of categories, generating a page populated by at least some of the extracted data, the page further including functions generated according to content of the extracted data, the functions executable on the content, and displaying a group of pages on the display unit including the generated page. | 10-22-2015 |
20150302090 | Method and System for the Structural Analysis of Websites - A method for management of websites ( | 10-22-2015 |
20150310076 | DATA LOAD PROCESS - A source data file can be loaded into a database by extracting data from the source data file at a computing device, validating the data from the source data file at the computing device, transforming the data from the source data file at the computing device, requesting the data to be transferred to a first database at a first message processor module, delivering the data to the first database, and responding that the data has been transferred to the first database at the first message processor module to acknowledge receipt of the data in the database at the computing device. | 10-29-2015 |
20150310082 | HADOOP OLAP ENGINE - In various example embodiments, systems and methods for building data cubes to be stored in a cube store are presented. In some embodiments, a metadata engine generates the cube metadata. In further embodiments, cube data is generated by a cube build engine based on the cube metadata and source data. The cube build engine performs a multi-stage MapReduce job on the source data to produce a multi-dimensional cube lattice having multiple cuboids. In further embodiments, the cube data is provided to the cube store. | 10-29-2015 |
20150317370 | Removal of Invisible Data Packages in Data Warehouses - In accordance with one embodiment of the disclosed technology, inconsistencies are detected between various records relating to data that has been associated with an identification tag. Data packages associated with the inconsistencies may then be removed. In accordance with another aspect of the disclosed technology, requests relating to data packages associated with inconsistencies in the various stored records are identified and removed. The disclosed technology may be implemented in data warehouses. | 11-05-2015 |
20150324437 | GRID FORMAT DATA VIEWING AND EDITING ENVIRONMENT - A data profiling server extracts data from multiple data sources and stores the extracted data in a target data store. Validation functions are performed on the extracted to determine whether the data conforms to validation rules associated with the target data store and/or business logic definitions. For data that does not conform to the validation rules, the data profiling server identifies recommended transformations to be applied to the data to correct any errors that cause the data to not conform with the validation rules. A data profiling application presents to a user the data extracted from the multiple data sources in a grid structure. The grid structure also includes error indicators identifying the data that does not conform to the validation rules and specifies the recommended transformations identified by the data profiling server. | 11-12-2015 |
20150331916 | COMPUTER, DATA ACCESS MANAGEMENT METHOD AND RECORDING MEDIUM - A computer system including a shared file server manages the access to file data for performing access to the file data accurately and efficiently. This computer includes a plurality of first name spaces to which is assigned an access path to data stored in a storage area, and a name space to which is assigned a path corresponding to the access path and which is different from the first name spaces. When the access paths generated in different first name spaces are the same, the corresponding paths which correspond to the same access paths are changed into mutually different paths. Moreover, by assigning a path corresponding to the data to be analyzed, it is possible to efficiently access the requested data among a large amount of data. In addition, the sorting of the corresponding paths is changed according to the load of the computer storing the data. | 11-19-2015 |
20150331917 | RECORDING MEDIUM HAVING STORED THEREIN TRANSMISSION ORDER DETERMINATION PROGRAM, TRANSMISSION ORDER DETERMINATION DEVICE, AND TRANSMISSION ORDER DETERMINATION METHOD - A computer obtains communication information that is transmitted/received between tiers in an information processing system providing a result of a process performed in an n-th (n>2) tier in reply to received request, identifies pairs of the request and corresponding response in each of the plurality of tiers from the obtained communication information, compares the pairs in each of tiers, extracts an order relationship wherein the request of a second pair is transmitted after the response of a first pair is obtained, obtains the order relationship in an (n-i)th tier on the basis of the order relationship between the pairs in an (n-i+1)th tier that respectively correspond to the pairs in the (n-i)th tier, sequentially from the n-th tier to a first tier, and determines a transmission order of the request to the information processing system on the basis of the obtained order relationship. | 11-19-2015 |
20150331918 | Business Intelligence Document - A business intelligence (BI) document preserves references to identities and formats of remote data sources and allows a local computing device to offload analytical operations to remote data sources. The BI document specifies a graph of entities connected by directed edges from the output of one entity to an input of another entity. An entity, for example, can represent without limitation a data structure, an external data source, a control element, an external event source, a visualization, or an update service. The entities of a BI document at a local computing device can reference data at an original data source—rather than extracting data from the original data source to a preferred local datastore. An entity of the BI document can direct a remote data source to execute transformations on the remote data before returning a solution to the local computing device. | 11-19-2015 |
20150331928 | USER-CREATED MEMBERS POSITIONING FOR OLAP DATABASES - The present disclosure describes methods, systems, and computer program products for positioning user-created data in an Online Analytical Processing (OLAP) database. A method may comprise receiving, at a server and from a client computing device, a query specifying one or more anchor members in a first hierarchy of the database and one or more user-created members to be positioned in the first hierarchy in relation to the anchor members according to a user-specified order, the first hierarchy of the database including members arranged in positions specified by the server; positioning, by the server, the user-created members in the first hierarchy of the database in relation to the anchor members according to the user-specified order, wherein the positioning comprises: determining a rank position for each of the anchor members; creating, for each of the user-created members, an auxiliary calculated member on an ancillary hierarchy, the auxiliary calculated member indicating the respective user-created member and the rank position of the associated anchor member. | 11-19-2015 |
20150339358 | MANAGING QUERIES IN BUSINESS INTELLIGENCE PLATFORMS - A method is provided for managing queries in business intelligence platforms. The method includes receiving, by a processor, a first query requesting data from at least one business intelligence content, the first query having a first format. The method further includes determining, by the processor, at least one requirement to complete the first query. The method further includes converting, by the processor, the first query to a second query configured to be used by an Extract, Transform, and Load (ETL) program, in response to determining that the at least one requirement exceeds a threshold, the second query having a second format. | 11-26-2015 |
20150339359 | COMPUTER SYSTEM, METADATA MANAGEMENT METHOD, AND RECORDING MEDIUM - An object of the invention is to manage correspondence between data having different schemas stored in process servers, data warehouses, or the like in the computer system having a complicated modularity configuration in an integrated manner. The computer system includes a storage portion that maintains schema correspondence information indicating correspondence of metadata described in different schema formats; a storage device that stores metadata in a second schema format different from a first schema format, which is transformed based on the schema correspondence information in a manner of being associated with corresponding actual data and a unified ID; an index portion that acquires the metadata and the unified ID in the second schema format from the storage device and maintains the metadata index transformed in a third schema format different from the first and second schema formats based on the schema correspondence information; and an update managing portion that specifies the unified ID of the metadata to be an update target by using the schema correspondence information in response to an update request of the metadata having a predetermined schema format. | 11-26-2015 |
20150339360 | PROCESSING A DATA SET - Embodiments relate to processing a data set stored in a computer system. In one aspect, a method of processing a data set stored in a computer system includes providing one or more parameters for quantifying data quality of the data set. A processor generates, for each parameter of the one or more parameters, a reference pattern indicating a dysfunctional behavior of the values of the parameter. The data set is processed to obtain values of the one or more parameters. A parameter of the one or more parameters is identified whose obtained values match a corresponding reference pattern of the generated reference patterns. The identified parameter is assigned a resource weight value indicating the amount of processing resources required to fix the dysfunctional behavior of the identified parameter. | 11-26-2015 |
20150339377 | Count Based Real Time Display of Statistics and Values for Selected Regular Expressions - Embodiments are directed towards real time display of event records and extracted values based on at least one extraction rule, such as a regular expression. A user interface may be employed to enable a user to have an extraction rule automatically generate and/or to manually enter an extraction rule. The user may be enabled to manually edit a previously provided extraction rule, which may result in real time display of updated extracted values. The extraction rule may be utilized to extract values from each of a plurality of records, including event records of unstructured machine data. Statistics may be determined for each unique extracted value, and may be displayed to the user in real time. The user interface may also enable the user to select at least one unique extracted value to display those event records that include an extracted value that matches the selected value. | 11-26-2015 |
20150347261 | PERFORMANCE CHECKING COMPONENT FOR AN ETL JOB - Generation of a performance determination report for an Extract, Transform, Load (ETL) job includes decomposing the ETL job into two or more stage instances, and identifying one or more conditions for each of the stage instances. A set of tests for each of the identified conditions are generated. A first set of test results are generated by performing the set of tests. It is determined whether a test result from the first set of test results is outside of a first range. Conditions that can be identified include a non-volatile free memory condition, a network reliability condition, a network configuration condition, an application availability condition, a database availability condition, a database performance condition, a schema validity condition, an installed libraries condition, a configuration parameter condition, a volatile free memory condition, and a third party tool condition. | 12-03-2015 |
20150347386 | AUTOMATIC AGGREGATION FOR INFRASTRUCTURE STRING MATCHING - In some examples, a computing system may receive a plurality of string matching requests with respect to a plurality of data streams, and determine one or more opportunities for aggregation of string matching. | 12-03-2015 |
20150347539 | SYSTEM AND METHOD OF CONSUMING AND INTEGRATING WITH REST-BASED CLOUD AND ENTERPRISE SERVICES - Techniques are disclosed for consuming and integrating with representational state transfer (REST) based services. In one embodiment, a multiple-step extract, transform, load (ETL) job includes one or more REST steps. Each step in the ETL job performs a particular well-defined task, and the REST step in particular constructs REST requests, invokes REST services, and retrieves REST responses. The REST step may be mapped to a vector of inputs such that multiple REST service calls are made for each of the inputs. Further, a reusable connection may be employed so that one connection can be used to make multiple REST service calls. URLs used to make REST service calls may be dynamically constructed based on input data parameters and ETL job parameters. Additionally, a graphical user interface (GUI) allows users to configure aspects of a REST step. | 12-03-2015 |
20150347540 | DATA CONSTRUCTION FOR EXTRACT, TRANSFORM AND LOAD OPERATIONS FOR A DATABASE - An extract, transform and load (ETL) data construction system provides an extensible, interim platform for transitioning data from a legacy database or other data sources to a target database. The data construction system performs data construction operations which may include creating data that can be used for data transformation in ETL operations. The data construction system provides check-in-check-out procedures for controlling multi-user data access and data editing, and also provides multi-level user access checks. | 12-03-2015 |
20150347541 | SYSTEM AND METHOD OF CONSUMING AND INTEGRATING WITH REST-BASED CLOUD AND ENTERPRISE SERVICES - Techniques are disclosed for consuming and integrating with representational state transfer (REST) based services. In one embodiment, a multiple-step extract, transform, load (ETL) job includes one or more REST steps. Each step in the ETL job performs a particular well-defined task, and the REST step in particular constructs REST requests, invokes REST services, and retrieves REST responses. The REST step may be mapped to a vector of inputs such that multiple REST service calls are made for each of the inputs. Further, a reusable connection may be employed so that one connection can be used to make multiple REST service calls. URLs used to make REST service calls may be dynamically constructed based on input data parameters and ETL job parameters. Additionally, a graphical user interface (GUI) allows users to configure aspects of a REST step. | 12-03-2015 |
20150347542 | Systems and Methods for Data Warehousing in Private Cloud Environment - A system and method are disclosed for private cloud computing and for the development and deployment of cloud applications in the private cloud. The private cloud computing system and method of the present invention include as components at least a cloud controller, a cloud stack, Service Registry, and a cloud application builder. | 12-03-2015 |
20150363451 | TIMESTAMP NORMALIZED COUNTER DATA - A method including collecting, from a cable modem termination system (CMTS), counter data corresponding to usage of a network resource, wherein the counter data includes timestamp data, converting the counter data into timestamp normalized counter data by dividing the counter data among a plurality of buckets according to the timestamp data, and storing the timestamp normalized counter data. | 12-17-2015 |
20150370868 | AUTOMATED METADATA LOOKUP FOR LEGACY SYSTEMS - A method and system for accessing data in a de-commissioned legacy system are provided. Data are automatically extracted from the legacy system, although data structure(s) of the legacy system might not be known, by finding views corresponding to a query for the data. Attributes, metadata, and/or fields (“attributes”) can be parsed from the query. Tables and/or fields including the parsed attributes are identified. Views can be then identified, where the views contain the tables and/or fields including the parsed attributes. The views can be ranked in an order from those that include the greatest number of parsed attributes to those including the least number of parsed attributes. A data request understandable by the legacy system, e.g. a packet, can then be formed using the least number of views, where the views can collectively include all of the parsed attributes. | 12-24-2015 |
20150370870 | ELASTIC SCALABILITY OF A CONTENT TRANSFORMATION CLUSTER - Content transformations can include transformation of content items in a CMS repository from a source format to a target format. Such transformations can be performed using a transformation node cluster having multiple nodes, each of which is configured for a specific content transformation type. Router nodes can receive requests for content items and route content items to transformation nodes having a proper content transformation type to either transform a requested content item to the target format or perform an intermediate transformation as part of a transformation chain. A transformation node cluster can be dynamically configurable based on estimates of expected loads for the various types of transformations. Systems, methods, and articles of manufacture are also described. | 12-24-2015 |
20150370871 | ETL TOOL INTERFACE FOR REMOTE MAINFRAMES - Embodiments of the present invention disclose a method for remote job submission. A computer receives a first job of a job stream. If the first job accesses a mainframe operating system, then the computer retrieves at least a return code threshold for the first job from a data store. If an option to build job control language is selected for the first job, then the computer generates job control language for the first job. The computer submits the first job to a job reader on the mainframe operating system using a file transfer protocol. The computer parses a retrieved output file to determine a return code for the first job. | 12-24-2015 |
20150370873 | DATA AGGREGATION AND REPORTING ENVIRONMENT FOR DATA CENTER INFRASTRUCTURE MANAGEMENT - Systems and methods that facilitate data center infrastructure management are discussed. One such system can include a plurality of data center infrastructure management (DCIM) tools. Each DCIM tool can maintain data associated with one or more data centers. Such a system can also include a data aggregation component that can extract the maintained data from at least two of the plurality of DCIM tools and can transform the extracted data into a common format. Such a system can also include an aggregated database that stores the extracted data, and can include a user interface that can provide for simultaneous access to the plurality of DCIM tools. | 12-24-2015 |
20150370880 | IDENTIFYING ITEMS ON A SECOND WEBSITE ALREADY BROWSED AT A FIRST WEBSITE - For browsing items on the worldwide web, operations are performed, including: searching for items to browse at a first website; recording attributes of each item browsed at the first website; searching for items to browse at a second website; comparing attributes of items identified during the searching of the second website with the recorded attributes of items browsed at the first website; in the event of a match between attributes of an item identified during the searching of the second website and the recorded attributes of an item browsed at the first website, indicating that the item identified at the second website is the same as the item browsed at the first website; and displaying an indication to the user that the item identified at the second website is the same as the item browsed at the first website. | 12-24-2015 |
20150379096 | SYSTEM AND METHOD FOR AUTOMATICALLY CONNECTING MULTIPLE, DISTINCT TABLES OF A DATA REPOSITORY - A system and method for automatically establishing connections to tables within a data repository may include establishing a pointer to a data repository inclusive of a plurality of tables. System-level metadata within a data repository, catalog, and tables may be accessed. Connections for the respective tables inclusive of data values and metadata associated with the data values may be automatically created. | 12-31-2015 |
20160004757 | DATA MANAGEMENT METHOD, DATA MANAGEMENT DEVICE AND STORAGE MEDIUM - A data management method employing the results of an analysis of data stored in a storage unit of a computer provided with a processor and a storage unit, wherein the computer generates an analysis data set by selecting data stored in the storage unit, subjects the analysis data set to prescribed data mining, extracts a model from the analysis data set, converts the model into a relational table, and associates the relational table with a dimension table and a history table that have been stored in advance in the storage unit. | 01-07-2016 |
20160034379 | INFORMATION TECHNOLOGY TESTING AND TESTING DATA MANAGEMENT - A testing and extract, transform and load (TETL) system is operable to interface with test platforms testing hardware or software components of information technology systems. The TETL system can execute extract, transform and load operations to load test data into a data warehouse and facilitates evaluating the test data across projects, entities and domains. | 02-04-2016 |
20160034546 | OPTIMIZING LOADING OF RELATIONAL DATA FOR APPLICATIONS - A mechanism for optimizing loading of relational data for applications is disclosed. A method includes monitoring a plurality of requests selecting entities and processes to be executed for the selected entities. The plurality of requests are generated in an application server. The method also includes associating an application context with each of the processes to be executed for the selected entities. The application context includes the process to be executed using an entity stored in a database external to the application server. The application further includes generating a fetch plan for the entity stored in the database; and modifying, the generated fetch plan in view of the selected entities and the associated application context. | 02-04-2016 |
20160034547 | SYSTEMS AND METHODS FOR AN SQL-DRIVEN DISTRIBUTED OPERATING SYSTEM - Various embodiments can include systems, methods, and non-transitory computer readable media configured to receive at least one operation to be performed using (i) first data that is managed by a first computing system and (ii) second data that is managed by a second computing system, the operation being received through an interface provided by the computing system, and wherein the operation is based at least in part on a Structured Query Language (SQL). At least one optimization can be performed based at least in part on the operation. The operation can be executed using at least the first data and the second data. A result generated can be provided upon executing the operation through the interface provided by the computing system. The computing system, the first computing system, and the second computing system are each able to concurrently process, access, and create at least a portion of the generated result. | 02-04-2016 |
20160034552 | PROCESSING DATA IN A DATA WAREHOUSE - Processing data of a data warehouse is provided and includes receiving, by a processing device, user input to create simple filter objects. Each filter object defines an ad hoc subset of a respective dimension of a dimension table of the data warehouse. User input is received to create a filtered operation object that specifies an operation and a plurality of the simple filter objects. The ad hoc subset differs from all subsets defined in the dimension table. | 02-04-2016 |
20160042019 | Columnar Storage of a Database Index - Methods, systems, and computer-readable media of columnar storage of a database index are disclosed. A particular columnar index includes a column store that stores rows of the columnar index in a column-wise fashion and a delta store that stores rows of the columnar index in a row-wise fashion. The column store also includes an absence flag array. The absence flag array includes entries that indicate whether certain rows have been logically deleted from the column store. | 02-11-2016 |
20160063078 | AUTOMATIC IDENTIFICATION AND TRACKING OF LOG ENTRY SCHEMAS CHANGES - A log analysis unit compares log entries describing an event to one or more schemas associated with the event. Each of the schemas describes a different log entry structure. When a log entry is determine to have a structure that does not match any of the structures defined by any of the schemas associated with a particular event, a new schema describing the structure of the log entry is generated. In response to the generation of the new schema, one or more entities are notified. Additionally, instructions for processing log entries adhering to the new schema are generated. A cumulative schema and an intersection schema corresponding to the event are also generated. | 03-03-2016 |
20160063079 | MANAGEMENT OF CONTENT TAILORING BY SERVICES - Embodiments of the present invention provide systems, methods, and computer program products for processing responses from services (e.g., content providers) and managing content tailoring by services and/or recommender systems used by those services. Embodiments of the present invention can afford users with the ability to control the diversity of content in responses provided by services based one or more detected themes of the responses. Furthermore, embodiments of the present invention may be used to provide users with enriched responses from services, without needing cooperation of those services. | 03-03-2016 |
20160063080 | EVENT STREAM TRANSFORMATIONS - The formulation of transformations on one or more input event streams to generation one or more output event streams. Accordingly, the transformations may be considered to be a query on the original input event stream(s). The event query includes event stream source representations representing an input event stream available in a particular execution context. The event query also includes a transformation module identifying the transformation set to be performed on the input event streams in the execution context. Once the query is properly formed, an execution module may then cause the transformations to be executed upon the designated input event stream(s) to generate output event streams. | 03-03-2016 |
20160063081 | Multidimensional Graph Analytics - Disclosed herein are system, method, and computer program product embodiments for performing ad-hoc analytical queries of graph data. An embodiment operates by receiving a graph pattern for a subgraph of interest. The facts of interest are then selected from graph data based on the received graph pattern. Dimensions are then defined based on a dimension seed pattern and a set of level expressions, and measures are defined based on a computation function and an aggregation function. A graph cube is formed based on the selected facts and the defined dimensions and measures. Because the facts, dimensions, and measures of interest are defined at the time of an analytical query, a user does not have to define such facts, dimensions, and measures, or know which analytical queries will be of interest, at the time of data collection. | 03-03-2016 |
20160063082 | DATA INTEGRATION USING AUTOMATED DATA PROCESSING BASED ON TARGET METADATA - Approaches for data integration between multiple IT sources using automated data processing based on target metadata are provided. Specifically, an integration component is configured to load a mapped data set into a table with delta processing based on a configuration object containing, e.g., the source data location, target table name, and source to target mapping. The integration component uses the target metadata to validate data, identify changes, generate the necessary database programming language (e.g., structured query language (SQL)), and run the database programming language with data binding to perform the actual data updates. The integration component leverages the data target metadata to automate the processing of source data, thus providing a way to validate the data, and identify delta changes at the field level between the source and target. This significantly reduces the overall development effort, while providing consistency in record handling and error reporting. | 03-03-2016 |
20160070766 | METHODS FOR FACILITATING BATCH ANALYTICS ON ARCHIVED DATA AND DEVICES THEREOF - A method, non-transitory computer readable medium, and archive node computing device that receives an indication of each of a plurality of archived files required to service a job from one of a plurality of compute node computing devices of an analytics tier. An optimized schedule for retrieving the archived files from one or more archive storage devices of an archive tier is generated. The optimized schedule is provided to the one of the plurality of compute node computing devices. Requests for the archived files received from the one of the plurality of compute node computing device and at least one other of the plurality of compute node computing devices, wherein the requests are sent according to the optimized schedule. | 03-10-2016 |
20160070767 | TREE DATA STRUCTURES BASED ON A PLURALITY OF LOCAL COORDINATE SYSTEMS - A system, computer readable medium, and method are disclosed for performing a tree traversal operation. The method includes the steps of executing, via a processor, a tree traversal operation for a tree data structure, receiving a transformation node that includes transformation data during the tree traversal operation, and transforming spatial data included in a query data structure based on the transformation data. Each node in the tree data structure is classified according to one of a plurality of nodesets, the plurality of nodesets corresponding to a plurality of local coordinate systems. The processor may be a parallel processing unit that includes one or more tree traversal units, which implement the tree traversal operation in hardware, software, or a combination of hardware and software. | 03-10-2016 |
20160070768 | EXTENSIBLE SURFACE FOR CONSUMING INFORMATION EXTRACTION SERIVICES - Representing structured data extracted from unstructured data in fashion allowing querying using relational database concepts. A method includes receiving user input specifying one or more database views. The method further includes receiving user input specifying an information extraction technique, such as an extraction workflow. The method further includes receiving user input specifying a corpus of data. The extraction technique is applied to the corpus of data to produce the one or more database views. These views can then be queried or operated on using database tools. | 03-10-2016 |
20160070769 | ONTOLOGY HARMONIZATION AND MEDIATION SYSTEMS AND METHODS - A method and system for harmonizing and mediating ontologies to search across large data sources is disclosed. The method comprises receiving a query targeting a first ontology. The method further comprises translating the query into one or more translated queries, each translated query targeting a respective ontology different from the first ontology. For each of the queries, issuing the query to a respective database organized according to the respective ontology of the query, and receiving a respective result set for the query, wherein the respective result set corresponds to the respective ontology of the query. The method further comprises translating the respective result set into a translated result set corresponding to the first ontology, aggregating the result sets into an aggregated result set corresponding to the first ontology, and returning the aggregated results set corresponding to the first ontology. | 03-10-2016 |
20160071135 | METHOD FOR OBTAINING BUSINESS INTELLIGENCE INFORMATION FROM A LARGE DATASET - Behavioural data relating to online interactions is collected and stored in the form of a raw dataset. A data filter created on the basis of defined characteristics of interest is applied to the raw dataset, thereby obtaining a subset of data. Business intelligence analysis is performed on the data of the subset of data, and a business intelligence report is generated, in accordance with the defined characteristics of interest. | 03-10-2016 |
20160078112 | Aggregation and Analytics for Application-Specific Optimization Based on Multiple Data Sources - Aggregating and transforming data, and performing analytics thereupon, for application-specific optimization based on multiple data sources. The data is preferably ingressed automatically, and may originate from various public and/or private data sources. Data transformation preferably aligns the data aggregated from the various sources, to thereby allow meaningful referencing. Complex and non-aligned data can therefore be consolidated, such that it is readily digestible by simulation (or other) software. In an embodiment, risk of flooding for a supply chain is computed from the aggregated and transformed data, using data analytics based on physical computation for flood risk assessment, allowing the supply chain to be optimized with regard to threat of flooding and/or actual flooding. In another embodiment, risk of wild fire may be assessed. Other types of risk may also be assessed. | 03-17-2016 |
20160078113 | VALIDATING CODE OF AN EXTRACT, TRANSFORM AND LOAD (ETL) TOOL - An approach for validating code for an extract, transform and load tool is provided. Naming, coding, and performance standards for the code are received. The code is exported to a job definition file and parsed. Violations of the standards are determined by a mismatch between the parsed code and the standards. A report identifying the violations is generated. Based on a review of the report and a rework of the code to comply with the standards, the reworked code is exported to another job definition file and parsed, and subsequently is determined to not include the violations of the standards. A second report is generated that indicates the reworked code does not include the violations. An approval of the reworked code is received based on the second report. Based on attributes of a job included in the code, a violation of one of the performance standards is determined. | 03-17-2016 |
20160085794 | DATA CONSISTENCY AND ROLLBACK FOR CLOUD ANALYTICS - An extract-transform-load (ETL) platform fetches consistent datasets in a batch for a given period of time and provides the ability to rollback that batch. The batch may be fetched for an interval of time, and the ETL platform may fetch new or changed data from different cloud/on-premise applications. It will store this data in the cloud or on-premise to build data history. As the ETL platform fetches new data, the system will not overwrite existing data, but rather will create new versions so that change history is preserved. For any reason, if businesses would like to rollback data, they could rollback to any previous batch. | 03-24-2016 |
20160085833 | TARGETED MULTI-DIMENSION DATA EXTRACTION FOR REAL-TIME ANALYSIS - Methods and systems for extracting targeted data for real-time reporting are discussed. In an example, a system can include a data store, a server, and a denormalized database. The data store can maintain data created by an application. The server can be communicatively coupled to the data store. The server can include a data extraction module and a data compression module. The data extraction module can extract a subset of the data stored in the data store according to an extraction scheme. The data compression module can compress the extracted subset of the data into a set of aggregated key value pairs. The denormalized database can store the aggregated key value pairs. | 03-24-2016 |
20160085835 | VISUALIZATION SUGGESTION APPLICATION PROGRAMMING INTERFACE - A dataset and some user selected columns of the dataset are received by a statistical analysis module for analysis. The statistical analysis module generates a score for each unselected column of the dataset based on statistical analysis of the unselected columns and all or a subset of the selected columns. A ranking of the unselected columns is presented to the user for selection of one additional column of the dataset, after which the remaining unselected columns are re-ranked according to their associated scores and once again displayed to the user. The user may continue selecting from among the ranked columns until a threshold number of columns has been selected, at which point the user may deselect a selected column in order to continue selecting additional columns. A visualization suggestion application program interface then matches the selected columns with compatible visualization configurations and presents some of these visualizations to the user. | 03-24-2016 |
20160085840 | Universal Interaction for Capturing Content to Persistent Storage - Systems and methods for enhanced content capture on a computing device are presented. In operation, a user interaction is detected on a computing device with the intent to capture content to a content store associated with the computer user operating the computing device. A content capture service is executed to capture content to the content store, comprising the following: applications executing on the computing device are notified to suspend output to display views corresponding to the applications; content to be captured to the content store is identified and obtained; the applications executing on the computing device are notified to resume output to display views; and automatically storing the obtained content in a content store associated with the computer user. | 03-24-2016 |
20160092522 | INTERACTIVE SOCIAL MEDIA ACCESS TO DATA SYSTEMS - Provided are techniques for interactive social media access to data systems. A social media gateway interface with adapters for each of multiple social media applications is provided. An event is published from a data system to at least one of the multiple social media applications. A social media bot is created to represent the data system. The social media bot and one of the adapters are used to enable the data system to interact with at least one of the multiple social media applications. | 03-31-2016 |
20160092523 | INFORMATION HANDLING SYSTEM AND COMPUTER PROGRAM PRODUCT FOR DYNAMCALLY ASSIGNING QUESTION PRIORITY BASED ON QUESTION EXTRACTION AND DOMAIN DICTIONARY - An approach is provided dynamically prioritizing question requests based on extracted question data. In the approach, performed by an information handling system, a number of question requests to a question and answering (QA) system are received from a computer network, and a plurality of question priority parameters are identified, including one or more question topics and a plurality question context parameters, by performing natural language processing (NLP) analysis of each question request. The approach determines a target priority value for each question request based on the plurality of question priority parameters identified for said question request. By evaluating the target priority values for the plurality of question requests, processing of the question requests is prioritized, such as by applying an artificial intelligence (AI) learned models and rule-based logic at the information handling system to evaluate the target priority values for the plurality of question requests. | 03-31-2016 |
20160092524 | SYSTEM AND METHOD FOR DATA TRANSFER FROM JDBC TO A DATA WAREHOUSE LAYER IN A MASSIVELY PARALLEL OR DISTRIBUTED DATABASE ENVIRONMENT - A system and method is described for database split generation in a massively parallel or distributed database environment including a plurality of databases and a data warehouse layer providing data summarization and querying functionality. A database table accessor of the system obtains, from an associated client application, a query for data in a table of the data warehouse layer, wherein the query includes a user preference. The system obtains table data representative of properties of the table, and determines a splits generator in accordance with one or more of the user preference or the properties of the table. The system generates, by the selected splits generator, table splits dividing the user query into a plurality of query splits, and outputs the plurality of query splits to an associated plurality of mappers for execution by the associated plurality of mappers of each of the plurality of query splits against the table. | 03-31-2016 |
20160092525 | INTEGRATING OBJECT-BASED DATA INTEGRATION TOOL WITH A VERSION CONTROL SYSTEM IN CENTRALIZED AND DECENTRALIZED ENVIRONMENTS - The present disclosure relates generally to a data integration system that integrate an object based data integration tool, such a GUI-based data integration tools, with version control systems using a relational database repository for persistence. Examples of distributed version control systems include Git, Mercurial, and Bazaar, and examples of centralized version control systems include Subversion, CVS etc. in centralized or distributed environments. | 03-31-2016 |
20160092526 | INTEGRATING OBJECT-BASED DATA INTEGRATION TOOL WITH A VERSION CONTROL SYSTEM IN CENTRALIZED AND DECENTRALIZED ENVIRONMENTS - The present disclosure relates generally to a data integration system that integrate an object based data integration tool, such a GUI-based data integration tools, with version control systems using a relational database repository for persistence. Examples of distributed version control systems include Git, Mercurial, and Bazaar, and examples of centralized version control systems include Subversion, CVS etc. in centralized or distributed environments. | 03-31-2016 |
20160092567 | SOCIAL MEDIA BOT TO REPRESENTATIONAL STATE TRANSFER (REST) PROXY FOR DATA SYSTEMS - Provided are techniques for social media bot to Representational State Transfer (REST) proxy for data systems. An adapter for a social media application and message to Representational State Transfer (REST) mappings are provided. A social media bot is created to represent a data system component. The social media bot and the adapter are used to enable the data system component to interact with the social media application by using the message to REST mappings to map social media messages for the social media application and REST calls that are used to access the data system component. | 03-31-2016 |
20160094506 | INTERACTIVE SOCIAL MEDIA ACCESS TO DATA SYSTEMS - Provided are techniques for interactive social media access to data systems. A social media gateway interface with adapters for each of multiple social media applications is provided. An event is published from a data system to at least one of the multiple social media applications. A social media bot is created to represent the data system. The social media bot and one of the adapters are used to enable the data system to interact with at least one of the multiple social media applications. | 03-31-2016 |
20160103897 | Database, Data Structure and Framework Transformer Apparatuses, Methods and Systems - The Database, Data Structure and Framework Transformer Apparatuses, Methods and Systems (“DDSFT”) transforms variable list request, population selection, base table transform extract data inputs via DDSFT components into transformed, merged data outputs. The DDSFT includes a database structure that stores data used in the framework operations. A macro-tool includes one or more macros that control a sequence of database queries that extract the data from the database structure and then perform transformations on the extracted data. The macro-tool includes a series of binary flags indicative of whether or not statements are executed. | 04-14-2016 |
20160103898 | DATAFLOW OPTIMIZATION FOR EXTRACTIONS FROM A DATA REPOSITORY - Dataflow optimization is described for extractions from a data depository. In one example an object-relationship graph of a data extraction definition is traversed in a first pass. The object-relationship graph has a node for each object. The steps from each node of the graph in the first pass is determined. It is determined if any of the determined steps are repeated. A single instance of each repeated step is placed before other steps in a query plan. Data is then extracted in a second pass from the object-oriented database system according to the object-relationship graph by performing the repeated steps first and then reusing the repeated steps in other determined steps. | 04-14-2016 |
20160103899 | NAVIGATION OF A DATA EXTRACTION GRAPH OF DATA AND METADATA FROM A DATA REPOSITORY - Navigation of a data extraction graph is described. The data extraction graph may be a graph representation of data and metadata in an object-oriented database. In one example, a data extraction definition of metadata from a database system is presented in a graph format. The graph format is transformed into a linear object tree, and/or an object list. Navigation commands are received from a user to navigate the transformed graph format. Selections of objects are received from the user to select objects in the transformed graph format. A summary view of the selected objects is then provided in the transformed graph format. | 04-14-2016 |
20160103900 | DATA STRUCTURING AND SEARCHING METHODS AND APPARATUS - Various computer implemented methods and data processing apparatus are described for use in structuring digital items and searching a plurality of digital items using a query item. At least one feature of a query digital item is extracted from a data file of the query digital item to form a query feature vector from a plurality of numerical data items representing the feature. It is determined which of a plurality of first clusters is most similar to the query digital item to identify a result cluster from the plurality of first clusters by calculating the aggregated similarity of a plurality of different digital items represented by a one of the first clusters to the query digital item for each of the plurality of first clusters using the query feature vector. Each of the plurality of first clusters represents a different plurality of digital items and each digital item is represented by only one of the plurality of first clusters. A search result is output comprising one or more digital items from the result cluster. | 04-14-2016 |
20160103902 | Multivariate Insight Discovery Approach - A raw dataset including measures and dimensions is processed, by a preprocessing module, using an algorithm that produces a preprocessed dataset such that at least one type of statistical analysis of the preprocessed dataset yields equal results to the same type of statistical analysis of the raw dataset. The preprocessed dataset is then analyzed by a statistical analysis module to identify subsets of the preprocessed dataset that include a non-random structure or pattern. The analysis of the preprocessed dataset includes the at least one type of statistical analysis that produces the same results for both the preprocessed and raw datasets. The identified subsets are then ranked by a statistical ranker based on the analysis of the preprocessed dataset and a subset is selected for visualization based on the rankings A visualization module then generates a visualization of the selected identified subset that highlights a non-random structure of the selected subset. | 04-14-2016 |
20160110352 | INFORMATION REDACTION FROM DOCUMENT DATA - Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for redacting data from a document collection generated for a set of documents that include personal information. The redaction of the data is based in part on a comparison of the document collection to a set of a personal documents of users for which the users have provided explicit approval to use in the processing of the document collection. | 04-21-2016 |
20160110434 | METHOD AND SYSTEM THAT DETERMINE WHETHER OR NOT TWO GRAPH-LIKE REPRESENTATIONS OF TWO SYSTEMS DESCRIBE EQUIVALENT SYSTEMS - The current document is directed to methods and systems that determine whether or not two graph-like representations of two physically or temporally distinct computer systems or computer-system configurations are equivalent. The currently described methods and systems extract a first and second ordered set of subgraphs from each of a first and second graph-like representation of a first and a second computer system. The ordered sets of subgraphs are logically aligned, forming a set of subgraph pairs. The currently described methods and systems transform the first and second subgraph of each subgraph pair into a corresponding first and second set of trees, label the trees, and then compare labels at each level of the trees to determine whether or not an isomorphic tree can be found in the second set of trees for each tree in the first set of trees. | 04-21-2016 |
20160110435 | COPYING DATASETS BETWEEN DATA INTEGRATION SYSTEMS - According to one embodiment, a method of copying a dataset associated with a first extract, transform, and load (ETL) job in a first data integration system to a second data integration system comprises copying executable code associated with the first ETL job from the first to the second system. Operating system software, integration system software, and file system structure are substantially identical between the first and second systems. The method further comprises executing the second ETL job to read the dataset from the first data integration system and write the dataset to the second data integration system. The second ETL job is associated with configuration parameters specifying storage resources in the first system associated with the dataset and destination parameters specifying storage resources in the second system. The method further comprises copying metadata generated by the second ETL job from the first to the second data integration system. | 04-21-2016 |
20160110436 | REDUNDANT DATA INTEGRATION PLATFORM - According to one embodiment, a data integration system includes a first and second head node, both operable to store and process an extract, translate, and load (ETL) job. Operating system software, integration system software, and file system structure are substantially identical between the first and second head node. A dataset, comprising user data and metadata, is associated with the ETL job. The metadata comprises virtual identifiers that identify resources of the data integration system. An identifier resolution server resolves the virtual identifiers to actual identifiers of resources associated with either the first or second head node. The first head node is operable to process the ETL job when the identifier resolution server resolves virtual identifiers to actual identifiers of resources associated with the first head node. The second head node is operable to process the ETL job when virtual identifiers are resolved to resources associated with the second head node. | 04-21-2016 |
20160117350 | COLUMN GROUP SELECTION METHOD AND APPARATUS FOR EFFICIENTLY STORING DATA IN MIXED OLAP/OLTP WORKLOAD ENVIRONMENT - Disclosed is a technology for data storage management in a database system, and more particularly, a data storage technology for a mixed OLAP/OLTP workload. A column group selection apparatus for efficiently storing data in a mixed workload processing environment includes a query processor configured to create column access information about queries that are input, a page monitoring module configured to create page-specific query pattern information using information about a page at which each of the input queries accesses and the column access information, a page layout manager configured to create page column group information in which a column group to be used to form each page is selected by applying a tree-based algorithm for selecting a column group to the page-specific query pattern information, and a data storage manager configured to create and store pages in units of column groups based on the page column group information. | 04-28-2016 |
20160117371 | JOB AUTHORING WITH DATA PREVIEW - Jobs can be authored in conjunction with a visual workspace. Upon selection of representation of a data source in the workspace, a preview of the data source can be generated within context of the visual workspace. Further, representations of one or more data transformation operations can be provided with the preview. Selection of a transformation operation results in an updated preview reflecting application of the operation as well as generation of backend code to perform the operation. Furthermore, a job comprising one or more transformation operations can be added to the workspace automatically. | 04-28-2016 |
20160132549 | Methods, Systems, and Products for Maintaining Data Consistency in a Stream Warehouse - Methods, systems, and products characterize consistency of data in a stream warehouse. A warehouse table is derived from a continuously received a stream of data. The warehouse table is stored in memory as a plurality of temporal partitions, with each temporal partition storing data within a contiguous range of time. A level of consistency is assigned to each temporal partition in the warehouse table. | 05-12-2016 |
20160140195 | CUSTOM PARALLIZATION FOR DATABASE TABLE LOADING - Systems, methods, and other embodiments associated with custom parallelization on a per load operation basis are described. In one embodiment, a method includes identifying a first load operation that loads first data into a first database table, determining a first level of parallelization with which to execute the first load operation, and storing the first level of parallelization for use in future execution of the first load operation. The method includes identifying a second load operation that loads second data into a second database table, determining a second level of parallelization, different from the first level of parallelization, with which to execute the second load operation, and storing the second level of parallelization for use in future execution of the second load operation. | 05-19-2016 |
20160140204 | COMPUTER IMPLEMENTED METHODS AND SYSTEMS FOR EFFICIENT DATA MAPPING REQUIREMENTS ESTABLISHMENT AND REFERENCE - The current invention can provide methods and systems for creating and referencing data mapping requirements in a highly efficient, effective manner by providing functionality needed at the beginning, ending, and throughout the life of a data warehousing project. This can be accomplished through the ability to: prioritize fields of interest, provide visualization of the data mapping, set the ETL rules, provide progress and filtering functionality based on current status, provide learned intelligent tips for the next needed functionality, provide source comparisons per applied learning, apply learning for product enhancement, provide data profiling, provide data lineage, and have all of this functionality work together to achieve these capabilities. | 05-19-2016 |
20160140205 | QUERIES INVOLVING MULTIPLE DATABASES AND EXECUTION ENGINES - Described herein are techniques for handling a query that uses first data from a first database and second data from a second database. The first database is accessible via a first execution engine, and the second database is accessible via a second execution engine. A sub-query of the query can be sent to the second execution engine for execution on the second database. Streaming results of the sub-query can be received by the first execution engine. | 05-19-2016 |
20160147850 | ENTERPRISE DATA WAREHOUSE MODEL FEDERATION - The present disclosure describes methods, systems, and computer program products for providing data warehouse model federation. One computer-implemented method includes receiving a request to transform a first semantically enriched model to a second semantically enriched model, wherein the first semantically enriched model operates in a first data warehouse and the second semantically enriched model operates in a second data warehouse; identifying a set of data associated with the first semantically enriched model, wherein the set of data is stored in the first data warehouse; providing access to the set of data associated with the first semantically enriched model from the second data warehouse; creating the second semantically enriched model based on the first semantically enriched model using an interface of the second data warehouse; and deploying the second semantically enriched model in the second data warehouse. | 05-26-2016 |
20160147851 | CREATION OF CHANGE-BASED DATA INTEGRATION JOBS - A computer software implemented method for transforming a first extract transform load (ETL) job having at least some unload transform load (UTL) portions. The method includes the following steps: (i) decomposing the first ETL job into an intermediate set of one or more jobs; and (ii) for each job of the intermediate set, transforming the job into a transactionally equivalent job to yield a final set of one or more jobs. The decomposing is performed so that each job of the intermediate jobs set is a Simple UTL job. The transforming is performed so that each job of the final set includes no UTL portions. | 05-26-2016 |
20160154865 | Method and Software for Retrieving Information from Big Data Systems and Analyzing the Retrieved Data | 06-02-2016 |
20160162521 | Systems and Methods for Data Ingest in Interest-Driven Business Intelligence Systems - Systems and methods for data ingest in interest-driven business intelligence systems in accordance with embodiments of the invention are illustrated. The interest-driven business intelligence system may maintain a set of registered data ingest instruction data that includes at least one registered data ingest instruction data. Each of the at least one registered data ingest instruction data includes an identifier and data ingest instruction data associated with the identifier. The system may receive a request to generate data using registered data instruction data. The request may include the identifier of the registered data instruction data. Data is generated using the data ingest instruction data associated with the requested identifier and at least one of raw data, source data, and aggregate data, and provided for use. | 06-09-2016 |
20160162533 | DATA ENVIRONMENT CHANGE NOTIFICATION - A method of and system for monitoring and providing notification of changes to a data environment. Parameters of interest for monitoring changes made to a metadata data store are stored. Instructions are stored related to notifying a recipient of a communication about the changes. The metadata is monitored for changes related to the parameters of interest, and a change notification communication is generated with transmission initiated to the recipient. The instructions related to notification of the changes may include the identity of a recipient, mechanism for notification, and format and content of notification. A computer readable memory system and a computer program product may be provided that is encoded with a data structure for notifying data consumers of changes to a data environment that includes metadata. | 06-09-2016 |
20160162550 | METHOD, APPARATUS, AND APPLICATION SYSTEM FOR REAL-TIME PROCESSING THE DATA STREAMS - Disclosed are a method, a data processing engine, and a system for real-time processing a plurality of continuously-generated data streams. The method for real-time processing the data with different schemas that transmit from heterogeneous relational databases includes steps of identifying categories the data, converting the data, and then storing the data in a non-relational data. Moreover, an architecture is provided together with the system and the method to improve the management of products, product lines or lifecycle such as the feedback of information regarding the performance analysis of an online game, or real-time alerts and recommended actions regarding the yield rate in a manufacturing stage of an industry such as the semiconductor manufacturing industry. | 06-09-2016 |
20160171026 | INTERPRETING INVALID DATA AS VALID DATA | 06-16-2016 |
20160171063 | MODELING ACTIONS, CONSEQUENCES AND GOAL ACHIEVEMENT FROM SOCIAL MEDIA AND OTHER DIGITAL TRACES | 06-16-2016 |
20160171064 | EFFICIENT QUERY PROCESSING USING HISTOGRAMS IN A COLUMNAR DATABASE | 06-16-2016 |
20160171229 | IMPLEMENTATION OF DATA PROTECTION POLICIES IN ETL LANDSCAPES | 06-16-2016 |
20160179912 | METHOD AND APPARATUS TO MAP ANALYTICS TO EDGE DEVICES | 06-23-2016 |
20160179919 | ASYNCHRONOUS DATA REPLICATION USING AN EXTERNAL BUFFER TABLE | 06-23-2016 |
20160188685 | FAN IDENTITY DATA INTEGRATION AND UNIFICATION - An analytical request for information that includes queryable attributes mapped to corresponding fields of customer records from in-memory database of a number of data sources is received. Semantic mapping of the queryable attributes to the corresponding fields of the customer records is resolved according to semantic mapping metadata. A query that includes the corresponding fields is generated and over the in-memory database views and a number fan records each representing a unique fan entity. Fan related data is retrieved from the number of fan records and the in-memory database views based on the executed at least one query. Upon executing the query, at runtime the in-memory database views are dynamically integrated with the number fan records based on a cross-reference table linking the in-memory database views with the number of fan records based on a fan identifier. | 06-30-2016 |
20160188686 | Systems and methods for creating, editing, storing and retrieving knowledge contained in specification documents - A system for storing and disseminating knowledge contained in documents includes a document annotator that creates a structured syntactic textual model of each of the documents, an ontology directed extractor that extracts properties from the textual models, a database for storing the textual models and the properties, and an interface permitting queries to the database. The document annotator includes a plurality of data transformers and a plurality of custom annotator tools. The ontology directed extractor includes an ontology based schema definition and a plurality of ontology based data transformers. The user interface includes the ability to view, search, navigate, create, and exchange documents. The creation feature includes a transclusion function. | 06-30-2016 |
20160188687 | METADATA EXTRACTION, PROCESSING, AND LOADING - Techniques for data storage are described herein. The techniques may include receiving data | 06-30-2016 |
20160188744 | DATA DETECTION METHOD, DATA DETECTION DEVICE, AND PROGRAM - The present invention enables designated data to be extracted from a structured document even when the structured document differs from others in terms of screen layout and document structure. A first structured document is read in and outputted to an output device; a first label to be extracted and first data to be extracted are acquired via an input device; an extraction pattern representing a relative relation in document structure between the first label to be extracted and the first data to be extracted is generated; and the extraction pattern is stored in a storage device. A second structured document is read in; a second label to be extracted is acquired; an extraction rule for extracting, from the second structured document and on the basis of the extraction pattern stored in the storage device and the second label to be extracted, second data to be extracted corresponding to the second label to be extracted is generated; and the second data to be extracted is extracted from the second structured document on the basis of the extraction rule. | 06-30-2016 |
20160196317 | INTEGRATED SUPPLIER INFORMATION TOOL | 07-07-2016 |
20160203197 | Method and System for Automatic Management of Dynamically Allocated Memory in a Computing Unit | 07-14-2016 |
20160203205 | Common Cell Algorithm for LRS Segment Join Operations | 07-14-2016 |
20160378842 | ENHANCED MECHANISMS FOR MANAGING MULTIDIMENSIONAL DATA - An enhanced multidimensional data management system significantly reduces the resources required to convert a spreadsheet into an enterprise collaborative system. A spreadsheet may be imported as is without modifications. Additional information that cannot be automatically inferred about the data model used in the spreadsheet is obtained. Dimensions, timescale, and measure data are clearly defined and located. The spreadsheet is kept in the system in the form of both a spreadsheet and as a multidimensional data structure such as an online analytical processing (OLAP) cube or structure. This duality allows the spreadsheet to be manipulated as if it was an OLAP cube, modifying iterations within the dimensions, and the OLAP cube to be computed as if it was a spreadsheet. | 12-29-2016 |
20160378843 | SYSTEMS AND METHODS FOR INTELLIGENT DATA PREPARATION AND VISUALIZATION - Real-time data visualization systems and methods are described. A data cube may be generated, wherein the data cube comprises a set of transforms to be applied to two or more data elements from disparate data sources, wherein processing of the data cube is to result in a data cube result having a plurality of measures and one or more hierarchies. A user may build a visualization, which produces a visualization request associated with the data cube, the visualization request specifying one or more of the measures and hierarchies of the data cube. The system identifies, based on the visualization request, one or more transforms within the data cube to remove from the data cube for the purpose of the visualization request to produce a modified data cube, and carries out the transforms of the modified data cube to produce a modified data cube result, which is exposed to the client-side visualization processor for rendering the visualization. | 12-29-2016 |
20170235801 | IDENTITY SERVICE MANAGEMENT IN LIMITED CONNECTIVITY ENVIRONMENTS | 08-17-2017 |
20180025061 | BOOTSTRAPPING THE DATA LAKE AND GLOSSARIES WITH 'DATASET JOINS' METADATA FROM EXISTING APPLICATION PATTERNS | 01-25-2018 |
20180025065 | EFFICIENT QUERY PROCESSING USING HISTOGRAMS IN A COLUMNAR DATABASE | 01-25-2018 |
20180025093 | QUERY CAPABILITIES OF TOPOLOGICAL DATA ANALYSIS GRAPHS | 01-25-2018 |
20190146781 | DYNAMIC DATA INGESTION | 05-16-2019 |
20190146949 | INTEGRATING OBJECT-BASED DATA INTEGRATION TOOL WITH A VERSION CONTROL SYSTEM IN CENTRALIZED AND DECENTRALIZED ENVIRONMENTS | 05-16-2019 |
20190146977 | METHOD AND SYSTEM FOR PERSISTING DATA | 05-16-2019 |
20220138216 | EXTRACT-TRANSFORM-LOAD SCRIPT GENERATION - One embodiment provides a computer implemented method, including: receiving, from a user, a natural language query for data contained within at least one data repository; identifying at least one concept from the natural language query, wherein the at least one concept includes an entity and an intent; identifying a plurality of datasets satisfying the natural language query by querying the at least one data repository utilizing the at least one concept; ranking the dataset based on relevance to the query; generating an extract-transform-load script that extracts, transforms, and loads a dataset selected by the user from the plurality of datasets; and retrieving data included in the dataset utilizing the extract-transform-load script, wherein the retrieving includes returning the data to the user. | 05-05-2022 |