Vous êtes sur la page 1sur 2

Data virtualization Data virtualization is the process of abstracting, transforming, federating and delivering data contained within a variety

of information sources so that they may be accessed by consuming application or users when requested without regard to their physical storage or heterogeneous structure. This concept and software is commonly used within data integration, business intelligence, service-oriented architecture data services, cloud computing, enterprise search, master data management and virtual master data management. Data Virtualization Functionality Data Virtualization software is an enabling technology which provides the following capabilities: Abstraction Abstract data the technical aspects of stored data, such as location, storage structure, API, access language, and storage technology. Virtualized Data Access Connect to different data sources and make them accessible from one logical place Transformation / Integration Transform, improve quality, and integrate data based on need across multiple sources Data Federation Combine results sets from across multiple source systems. Flexible Data Delivery Publish result sets as views and/or data services executed by consuming application or users when requested In delivering these capabilities, data virtualization also addresses requirements for data security, data quality, data governance, query optimization, caching, etc. Data virtualization software includes functions for development, operation and management. Data Virtualization Vendors

Composite Software Denodo Technologies IBM Informatica Queplix Radiant Logic Red Hat SAS

[VirtualWorks Group] The History of Data Virtualization Enterprise Information Integration (EII) and data federation have been terms used to describe a core element of data virtualization; to the capability to create relational joints in a federated view. Some forms of legacy data virtualization builds on knowledge and concepts developed within EII and Data Federation. It should be noted that, in general, Enterprise Information Integration failed in the market and some of the vendors repositioned themselves as data virtualization vendors

Data Virtualization and Abstraction Data virtualization is based on the premise of the abstraction of data contained within a variety of data sources (databases, applications, file repositories, websites, data services vendors, etc.) for the purpose of providing a single-point access to the data and its architecture is based on a shared semantic abstraction layer as opposed to limited visibility semantic metadata confined to a single data source. Practical Applications of Data Virtualization Rick van der Lans, author of multiple books on SQL, has defined data virtualization; "Data virtualization is the process of offering data consumers a data access interface that hides the technical aspects of stored data, such as location, storage structure, API, access language, and storage technology." [1] Consuming applications may include: business

intelligence, analytics, CRM, enterprise resource planning, and more across both cloud computing platforms and on-premises. Data integration Data integration involves combining data residing in different sources and providing users with a unified view of these data.]This process becomes significant in a variety of situations both commercial (when two similar companies need to merge their databases) and scientific (combining research results from different bioinformatics repositories, for example). Data integration appears with increasing frequency as the volume and the need to share existing data explodes.[2] It has become the focus of extensive theoretical work, and numerous open problems remain unsolved. In management circles, people frequently refer to data integration as "Enterprise Information Integration" (EII). Service-oriented architecture Service-oriented architecture (SOA) is a flexible set of design principles used during the phases of systems development and integration in computing. A system based on an SOA will package functionality as a suite of interoperable services that can be used within multiple, separate systems from several business domains. Cloud computing Cloud computing is the delivery of computing as a service rather than a product, whereby shared resources, software and information are provided to computers and other devices as a utility(like the electricity grid) over a network (typically the Internet). Enterprise search Enterprise search is the practice of making content from multiple enterprise-type sources, such as databases and intranets, searchable to a defined audience. Master data management In computing, master data management (MDM) comprises a set of processes and tools that consistently defines and manages the non-transactional data entities of an organization (which may include reference data). MDM has the objective of providing processes for collecting, aggregating, matching, consolidating, quality-assuring, persisting and distributing such data throughout an organization to ensure consistency and control in the ongoing maintenance and application use of this information. The term recalls the concept of a master file from an earlier computing era. MDM is similar to, and some would say the same as, virtual or federated database management.

www.earnrupees4you.com

Vous aimerez peut-être aussi