Vous êtes sur la page 1sur 38

GFT Technologies AG 2014

Big Data Uncovering Hidden Business


Value in the Financial Services Industry
Authors

Dr. Karl Rieder, Dr. Ignasi Barri and Josep Tarruella

Version 1.0
Published:
September 2014


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 2/38
Table of Contents
1 Executive Summary ..................................................................................................................................... 4
2 Introduction ................................................................................................................................................... 5
2.1 So what is big data? .................................................................................................................................... 5
2.2 Its a big data world ..................................................................................................................................... 6
2.3 The financial services industry today .......................................................................................................... 6
2.4 Big data adoption by the numbers .............................................................................................................. 8
3 Big data opportunities in the financial sector ......................................................................................... 10
3.1 Retail banking ............................................................................................................................................ 10
3.2 Investment banking ................................................................................................................................... 14
3.3 Insurance ................................................................................................................................................... 17
3.4 IT efficiency ............................................................................................................................................... 19
4 The architecture of big data solutions ..................................................................................................... 20
4.1 Big data technologies ................................................................................................................................ 20
4.2 Commercial appliances ............................................................................................................................. 23
4.3 Impact on existing systems ....................................................................................................................... 23
4.4 Cloud architectures ................................................................................................................................... 24
5 Addressing big datas challenges ............................................................................................................ 25
5.1 Technology ................................................................................................................................................ 25
5.2 Security and data protection ...................................................................................................................... 26
5.3 Data quality................................................................................................................................................ 26
5.4 Internal organisation .................................................................................................................................. 27



Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 3/38
6 First steps ................................................................................................................................................... 29
6.1 Getting started ........................................................................................................................................... 29
6.2 Establishing new roles in the company ..................................................................................................... 31
6.3 Strategies to follow lessons learned from successful big data projects ................................................. 32
7 GFTs Big Data practice ............................................................................................................................. 34
About the authors ............................................................................................................................................. 35









This report has been published based on a number of interviews with industry experts, secondary market research, and GFTs internal
expertise. The intention of the report is to render industry trends transparent and understandable within their context and to give
readers ideas for their businesses. The content has been created with the utmost diligence. Therefore, we are not liable for any
possible mistakes.
GFT Technologies AG
Executive Board: Ulrich Dietz (CEO), Jean-Franois Bodin, Marika Lulay, Dr. Jochen Ruetz.
Chairman of the Supervisory Board: Dr. Paul Lerbinger
Commercial Register of the local court (Amtsgericht): Stuttgart, Register number: HRB 727178
Copyright 2014 GFT Technologies AG. All rights reserved.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 4/38
Big data success requires new skill
sets and processes as well as
technology
Agility is key for success in the 21st
century
1 Executive Summary
Between now and 2020, the digital universe will double in size every two years. Businesses are projected to
increase their investment in hardware, software, and services by 40% between 2012 and 2020 to keep up with
this expansion
1
, and the data-driven financial sector is likely to be strongly impacted. Banks will be looking to
use new technology to remain competitive in the face of a rapidly diversifying financial services landscape.
Investment in big data technologies will grow even faster, given the massive evolutionary leap it represents in
shifting IT focus from relational databases to more open and flexible platforms that enable the management of
huge volumes of data and provide new analytical and actionable capabilities. The key challenge will be to
employ these technologies in such a way that they meet IT needs and at the same time provide new value to
the business, delivering a real return on that investment in the form of an improved bottom line.
Big data technologies will enable broader and better data analysis than ever before, leading to targeted event-
driven, customer-centric marketing, improved fraud detection, better
risk calculation, and operational efficiencies. Agility is key for business
success in the 21st century, and the effective deployment of big data
projects is likely to be the difference between success and failure in an ever more competitive market.
This blue paper is a guide to the opportunities, requirements, and challenges of the big data revolution,
revealing the value that can be tapped from big data technologies. In
making the shift to the brave new world of big data, the financial
services sector will need to consider not only technology changes, but
new use cases, processes, and skill sets. The paper also includes a
set of recommendations that will help financial services organisations to successfully implement big data
technologies.

1
IDC Study Dec 2012: The Digital Universe in 2020: Big Data, Bigger Digital Shadows, and Biggest Growth in the Far East


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 5/38
Relational databases and their associa-
ted processes were not designed to
handle huge data volumes
The financial services sector has seen
more wide-ranging change because of
its extensive use of data
2 Introduction
The computer age has brought significant changes to every industry, and the financial services sector is no
exception. It has seen wide-ranging change mainly because of its
extensive use of data: customer data, market data, and trading data
are all central to the industry. From the automation of business
processes in the 1960s through the emergence of the Internet in the
1990s, and on to the current era of mobile banking and high-frequency algorithmic trading, success in financial
services has always been about the smart use of data.
However, the technological infrastructure on which todays banking systems were built is beginning to buckle
under the strain of the sheer volume of data. Relational databases and
their associated processes cannot effectively handle such a high
volume of information; the continuous tuning of multiple environments
and the number of resources needed to keep data extraction,
transformation, and loading (ETL) processes working every day combine to make these legacy systems
extremely expensive to maintain. Insurance companies, too, are discovering that they could significantly
improve their bottom lines through more effective use of data to control fraud.
2.1 So what is big data?
Big data is the buzzword of the current technological decade. The volume of data available to organisations is
so huge that it requires entirely new technologies and processes to address it. The concept of big data is
usually described by its four dimensions of volume, variety, velocity, and value, also called the 4 Vs.
2



2
1 Zettabyte = 1,000 Exabytes = 1,000,000 Petabytes = 1,000,000,000 Terabytes = 1,000,000,000,000 Gigabytes


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 6/38
Between now and 2020, the digital
universe will double in size every two
years
IT investment in the digital universe
infrastructure is forecast to grow by
40% between 2012 and 2020
2.2 Its a big data world
The ever-decreasing cost of storage means that there is essentially no physical barrier to the amount of data
retained; by 2020, it will cost less than $0.20 to store a gigabyte of data, down from $2.00 per gigabyte today.
3

It is anticipated that, by 2020, the size of the digital universe will have increased by a factor of 300 over 2005s
volume; in hard numbers, this represents an increase from 130 exabytes to 40,000 exabytes, or 40 trillion
gigabytes thats over 5,200 gigabytes of data for every man, woman,
and child on the planet. Between now and 2020, the digital universe
will double in size every two years, fuelled by the 2.5 quintillion bytes
of information we create every day.
4
As we enter the age of the
Internet of Things, in which all types of sensors, appliances, and systems are connected and controlled
remotely from smartphones and other mobile devices, we can expect the growth to increase further.
In order to support these volumes of data, IT investment in hardware,
software, services, telecommunications, and people what we can
collectively describe as the infrastructure of the digital universe is
forecast to grow by 40% between 2012 and 2020. Targeted areas that
specifically apply to the storage and use of data, such as storage management, security, big data ETL, and
cloud computing, will likely grow at an even faster pace.
5

2.3 The financial services industry today
Driven by the adoption of technology changes around mobile devices, cloud computing, social media, and big
data, the financial services landscape is evolving from a traditional model to a digital model alongside increased
competition from both traditional and non-traditional players such as telecom carriers, retailers, and electronic
payment providers.
We see four main trends that financial institutions need to address.
Regulatory compliance: In the last 10 years, the number and variety of regulations affecting the
banking sector have greatly increased. Currently, maintaining compliance directly or indirectly affects
around 50% of the discretionary IT budgets of financial services companies. In order to react to the
constant business change in the financial services environment while also meeting compliance
regulations, increased agility is required.
Customer centricity: The way customers interact with their financial institutions is changing
fundamentally for the first time in the history of the industry. New technologies are enabling a focus on
customers that both empowers them and helps meet their individual, customised needs.
Back-office modernisation: Banks, especially in mature markets, are beginning to suffer heavily from
aging IT infrastructures that still run back-office legacy systems such as core banking software. Many
major banks have already started to replace these systems with more modern technologies.

3
IDC Study Dec 2012: The Digital Universe in 2020: Big Data, Bigger Digital Shadows, and Biggest Growth in the Far East
4
Mike Hogan, Big Data of your Own, Aug 2013.
5
IDC Study Dec 2012: The Digital Universe in 2020: Big Data, Bigger Digital Shadows, and Biggest Growth in the Far East


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 7/38
Banking and insurance activities are
now just another digital service or
shopping experience
There must be a 360 view of each
customer, with a complete
understanding of wants and needs
Cost reduction: As with almost every industry, financial institutions are pursuing cost reduction
strategies in order to reduce operational expenses. These strategies translate into fewer branches,
fewer new hires, more cloud and outsourced services, and the rationalisation of existing resources
through shared-service centres such as
payment service hubs that are either
self-run or run by a third party.
All four major trends are shaping the future of the
industry, and all are driving big data adoption.
However, customer centricity and regulatory
compliance are particularly relevant. Customer
centricity is pushing much of the modernization of
retail banking and insurance, while regulatory
compliance is driving changes in investment
banking itself.
2.3.1 Customer centricity
The trend toward customer centricity, while not
new, will greatly change the banking business
over the next few years. Banks are now focusing
on multiple initiatives, especially around customer empowerment and customer understanding.
Mobile banking means that customers rarely need to visit a physical bank branch or meet with an insurance
agent in order to conduct routine financial transactions. Almost everything can be done on the web or via a
mobile device, which means that banking and insurance activities
have essentially been commoditised as just another digital service or
shopping experience. This trend is growing faster in some countries
than others, but as more digital natives become bank customers, the
adoption of digital banking will continue apace. As devices and apps become ever-easier to use, even late
adopters will shift more of their routine banking activities online. This commoditisation has had a dramatic effect
on the way individuals regard their financial institutions; because there is little or no human interaction between
the institutions and their customers, there is also little or no loyalty. People can change their bank or insurance
provider with greater ease than ever before.
To differentiate themselves, banks and insurance firms must develop truly personalised customer service. This
requires a 360
o
view of each customer, with a complete understanding
of each individuals wants and needs a particular challenge for the
insurance industry, which continues to operate a siloed system where
different types of insurance coverage offered (life, health, cars,
houses, etc.) live in separate worlds.



Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 8/38
Increased regulation of the financial
services industry will lead to more
storage of historical data
The big data sector will grow at about
six times the rate of the overall
information technology market
2.3.2 Regulatory compliance

Over the last decade, there has been a significant push toward increased regulation of the financial services
industry, primarily in the area of risk control, but also in data
processing governance and auditing. Regulatory agencies are trying to
ensure that the industry acts responsibly and continues to provide
services regardless of market movements.
In order to meet regulatory demand, historical data must now be retained for seven years under the
requirements of the Dodd-Frank Reform Act, or for five years under the terms of the Basel Agreement.
Moreover, banks must have systems and processes in place to bring together these data to respond to
regulatory reporting requirements. In the case of a specific enquiry, the bank must be able to quickly sort
through the data to find all relevant information about a particular case; this requires data management far
beyond what the industry is currently doing.
2.4 Big data adoption by the numbers
In a study published at the end of 2013
6
, the European Information Technology Observatory (EITO) found that
financial institutions are leading the way in adopting big-data centric strategies. 92% of financial institutions
were identified as considering big data strategies (compared with only 40% across all industries); however, only
9% of financial institutions (and 4% of all industries) had actually implemented systems using these
technologies. Nevertheless, preparations do appear to be in process, with 38% of financial institutions (27%
across all industries) investing in improvements to their data storage facilities.
Clearly, the big data technology and services marketplace is set for
significant growth. At the end of 2013, IDC issued a prediction
7
that
the sector will grow at 27% CAGR to $32.4 billion by 2017 about six
times the growth rate of the overall information technology market.
Also in 2013, the IBM Institute for Business Value published an Executive Report
8
that summarised research
undertaken with Oxford Universitys Said School of Business across 1144 business and IT professionals in 95
countries, including 124 respondents from the financial services sector. The research focused on the use of big
data inside organisations and found, not unsurprisingly at this stage, that most initiatives were being
constructed around customer centricity (55%) and risk management (34%); the latter figure was significantly
higher for the financial services sector than for other business types, where operational improvements
outranked risk management.

6
EITO, Big Data in Europe: Evolution AND Revolution, December 2013
7
IDC,Worldwide Big Data Technology and Services 20132017 Forecast, December 2013
8
IBM Institute for Business Value and the Said Business School at the University of Oxford,Analytics: The real-world use of big data in financial services,
2013


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 9/38
90% of insurance firms have yet to
implement a company-wide big data
strategy
Insurance firms appear to be lagging somewhat in big data strategy adoption, according to research
undertaken by Bearing Point
9
. This survey found that 90% of insurance firms have yet to implement a
company-wide big data strategy, despite
more than two-thirds of participants
stating that big data would play an
important role in their future. The
research also revealed that, while 71%
said big data would be a top priority by
2018, less than quarter (24%) said their
companys big data maturity was
advanced or leading, and only 33% have
actually started a departmental or
enterprise implementation process.
While the financial services sector is still
in the early stages of big data strategy
adoption, there is general understanding
that the industry is at a crucial point and
has everything to gain from moving
forward with efforts to leverage big data
to improve its public image and deliver
excellence in customer service to its
customers and clients.

9
DataIQ News, May 2014


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 10/38
By bringing together disparate data
stores, organisations can begin to
derive new insights into their business
3 Big data opportunities in the financial sector
The availability of technologies and systems designed specifically for dealing with large volumes data opens
the door to the application of new approaches to common use cases, both evolutionary and revolutionary.
Evolutionary approaches might involve the development of complementary processes and architectures to
accelerate data processing performance when applied to massive stores of structured and unstructured data;
such approaches may meld elements of parallel and distributed computing to achieve the required outcomes.
A revolutionary approach, on the other hand, might see the complete restructuring of architectures and
processes to support entirely new ways of doing things. Different data
stores and warehouses could be refactored and consolidated into a
single raw golden data source, enabling the analysis of full data sets
rather than small samples or slices. By bringing together disparate
data stores, organisations could begin to derive new insights through the application of machine learning
algorithms.

3.1 Retail banking
According to the Millennial Disruption Index
10
, a survey of more than 10,000 Americans aged
18-33 conducted by Scratch (a division of Viacom), todays retail banking sector is at
exceptionally high risk for disruption. The four leading banks in the United States all appear in
the top ten lowest-rated companies in the country, with 53% of respondents seeing no difference between their
bank and any other. This disturbingly low level of loyalty is further underscored by the fact that 73% of
respondents would enthusiastically welcome offers of financial services from brands outside of the traditional
financial services marketplace such as Google, Amazon, Apple, and Paypal, among others. Retail banking is
vulnerable on many fronts.

10
Scratch-Viacom media networks, Millennial Disruption Index, 2014


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 11/38
Retail banks must put the customer at
the centre of their product and service
development initiatives
Client focus improves service,
relevance, customer satisfaction, and
loyalty
Customer-facing staff have a unified
history of interactions between the
company and each customer
3.1.1 Customer centricity replaces product centricity
Historically, banks and insurance companies have focused on developing product offerings, often at the
expense of understanding what their customers actually want. In this,
the financial services industry lags behind many other user-facing
sectors such as consumer goods. Retail banks must put the customer
at the centre of their product and service development initiatives.
This process can be made much easier through the development of a single, 360
o
view of each customer,
achieved by gathering information from across the organisation to better understand the services each
customer is using now and those that may be needed in the future. Big data technologies provide the means by
which large volumes of disparate data (accounts, consumer credit,
credit/debit cards, mortgages, etc.) can be brought together and
synthesised into a custom package of goods and services that will
best serve each individual customer.
When everything is focused on the client, service improves, relevance improves, customer satisfaction
improves, and loyalty will return. At the same time, cross-selling increases and customer churn slows.
3.1.2 Improved best offer process
Marketing campaign success rates can improve dramatically when banks put a relevant offer in front of a
customer at the optimum time for that customer to make a positive purchasing decision. Heres what the
traditional best offer marketing campaign looks like:

The problem with this type of campaign is that its effectiveness is very limited, largely because there is no real
alignment between the customers needs and the bank's offer. Here are a couple of ways this type of campaign
can be brought into the 21
st
century:
Unified vision: All staff in customer-facing positions (branches, call centres, etc.) are provided with a
unified history of interactions between the company and each customer. For instance, when a
customer enters the bank branch, the staff member meeting with that
customer can immediately see recent interactions such as reporting
the loss of a credit card or browsing the banks website to test out
different mortgage scenarios. This enables the staff member to have a
relevant discussion with the customer (about credit card protection or mortgages) rather than focus on
the product the bank has determined should be promoted to customers (for example, pension plans).
Dynamic offers: By making use of web and mobile channels, and the event log of online interactions
between the customer and the bank, dynamic and highly relevant offers can also be made to each
customer. For example, if the bank knows the customer is about to reach their credit card limit, the
system can be programmed to generate a custom-created offer for an increased credit limit or higher-


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 12/38
Big-data technologies can help by
enabling large volumes of data to be
stored and analysed rapidly
Big data also allows this type of
hindsight analysis to take place
and identify at-risk customers
level card account the next time that customer logs in. This model can be applied to any perceived
time-relevant customer need.
By aligning offerings with customer need, revenues are positively impacted. Big data makes such campaigns
possible by bringing together large amounts of data about customer behaviour, analysing them, and generating
offers that match the perceived customer need.
3.1.3 Client scoring
To effectively segment their customer base, banks apply a range of different client scoring methodologies. Of
particular interest is credit scoring, which measures the potential risk associated with lending to a particular
client. To make an accurate score, banks need to sort through large volumes of data and apply complex
algorithms to come up with a realistic risk factor for any particular individual. Some companies are even using
social network data (e.g. from Facebook) to measure potential credit risk.
Big data technologies such as in-memory databases can help with this
by enabling large volumes of data to be stored and analysed rapidly;
because the data is being held in memory rather than on a traditional
storage device (e.g. disk drive), access and processing times can be
reduced by an order of magnitude.
Client scoring methodologies can also be effective in improving up-sell and cross-sell revenues by detecting
potential opportunities and framing relevant offers for customers in the existing client base.
3.1.4 Customer retention and churn prevention
When banks understand why their customers are leaving, they can take appropriate steps to rectify the
situation and keep the customer in the fold. Big data also allows this type of hindsight analysis to take place
by providing insight into the customers behaviour prior to their ending
the relationship. Structured and unstructured bank data can be
combined with external sources (social media comments, media
coverage) to better understand brand reputation issues and identify at-
risk customers. Banks are then in a position to react in real time to these customers as they navigate through
the website or call centre and adjust their behaviours to focus on customer retention.
3.1.5 Credit card fraud detection
Credit and debit cards are the de facto worldwide standard for conducting secure payments. According to the
World Payments Report 2013
11
, the use of these payment methods increased by 15.8% for debit cards (124
billion transactions) and by 12.3% for credit cards (57 billion transactions). In the rapidly expanding field of
mobile payments, an increase of 58.5% per annum is predicted for 2014, equivalent to 28.9 billion transactions.

11
RBS & Cap Gemini, World Payments Report 2013


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 13/38
In 2013, credit card fraud represented
40% of the total number of fraud
incidents in banking
More complete data about customers
can prevent incorrect predictions and
lost customers
Banks have a great deal of aggregate
information that can be useful to other
types of businesses
3.1.5.1 Fraud control
The widespread use of any payment mechanism naturally also attracts widespread attempts to defraud that
mechanism. In 2013, credit card fraud represented 40% of the total number of fraud incidents in banking,
totalling approximately $5.5 billion.
12
Clearly, there is room for banks to
save significant amounts of money by improving fraud controls, and
big data can also help here. By sifting through the millions of payment
transactions made every day, combining these with data from other
internal and external sources, and analysing and understanding customer behaviour, investigators can
establish patterns and more accurately detect potential fraud quickly enough to minimize the damage. Big data
technologies allow this analysis to be done in real-time, as the transaction occurs, as opposed to batch
processing at the end of the day.
3.1.5.2 False positives
To prevent fraud, financial institutions frequently disable credit cards at the first sign of suspicious behaviour, a
move which is likely to lose them money and customers - if the prediction is incorrect. For example, a
customer of a bank who was traveling in China his first overseas trip
in several years activated the fraud alert system when he used his
credit card there. If the bank had had access to a more complete
picture of the customer that included, for example, recent payments (to
travel agents and airlines) or social network posts, they would easily have been able to verify the legitimate use
of the card because they would know he was travelling. Instead, the customer was embarrassed and
inconvenienced, and his relationship with his bank was negatively impacted.
3.1.6 New business models
The consolidation of customer and payment data can generate new business models and revenue
opportunities for retail banks:
The sale of non-identifiable data: Banks have a great deal of aggregate information, such as credit
card usage patterns (without identifying individual customers) that can be useful for other types of
businesses:

- Average expenditure on a given street or in a given area over a
particular period of time
- The times of day or days of the week when an area is busiest
- Where customers go when they leave a particular business location
Obviously, such data must be completely anonymous to be sellable.


12
Bank Systems & Technology, August 2013


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 14/38
Big data technologies allow banks to
streamline operational procedures and
compliance reporting
Banks can ensure a consistent view of
activity for business planning and
regulatory compliance
Creating solutions based on data: In addition to selling the data itself, banks can also create value-
added solutions based on the data. For instance, by analysing the behaviour of customers who pay
using POS devices, banks would be able to guide businesses to prioritise where they set up their
stores. This kind of service will allow thousands of small businesses to access information that was
previously only available to large corporations.

3.2 Investment banking
In investment banking, the trade is the central piece of information around which all other data
is mapped. For each of the tens of millions of trades a large investment bank has open at any
one time, the bank needs to process the trade through its lifecycle and calculate the related
profit or loss and the level of market and credit risk that attach to that
trade. With an exponential increase in the number of regulations
applied to stock trading, banks are struggling to report all the
information that is required. Big data technologies allow the bank to
manage more data and process it more quickly, thus streamlining operational procedures and compliance
reporting.
3.2.1 Consolidated view of trades
Over the decades, investment banking has generated a large number of silos for different types of products
fixed income, equities, foreign exchange, derivatives, etc. effectively preventing a 360
o
view of any aspect of
the whole business. Big data changes all that. Banks can now store huge volumes of data in a single data
warehouse, permitting the creation of a bank-wide trade repository. By taking advantage of distributed storage
techniques, banks can also store historic views of the data, enabling trade and position data to be consolidated
in a single data store.
By unifying data storage in this way, banks can centralise data
functions, rationalise storage architecture, and ensure a consistent
view of banking activity for both business planning and regulatory
compliance purposes. This also represents a huge cost savings for the
bank thanks to reduced redundant processes, systems, and personnel.
3.2.2 Flexible formats for trade repositories
When consolidating data from across a bank, IT departments will likely encounter a major challenge: how to
unify disparate data types in a single universal data model. For the consolidation to be successful, data
received from different business units using different front-office trading systems must be extracted,
transformed, and loaded (ETL) into a central data repository that uses a common model.
But whenever the upstream or downstream data format changes, which will happen because of constantly
changing requirements, the ETL system has to be modified. Using non-structured data storage, however, data
from different systems can be stored using the same approach, without the need to artificially fit any particular
record types into a universal data model.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 15/38
Processes that used to be run as
overnight batches can now be com-
pleted in minutes
Market and credit risk can be estimated
more quickly and in greater detail
This not only removes the need to develop and maintain a multitude of ETL systems, but also gives the
platform flexibility to grow and evolve over time. Such a system is less costly to maintain, as fewer changes are
required to implement new functionality.
3.2.3 Trade analytics and business intelligence
Once the data is centralised, the bank can execute trade analytics to prepare management reports, meet
regulatory requirements, and effectively measure business
performance. By harnessing the power of big datas parallel computing
architecture, data processing can be completed orders of magnitude
more rapidly, enabling operations and reporting to be concluded faster
than ever before. Processes that used to be run as overnight batches can now be completed in minutes,
producing an almost-real-time system.
Advanced analytical and business intelligence capability can also be applied to a broader, more complete data
set, thus enabling the identification of new insights into improved business performance. Having more data at
their disposal and having access to tools with which that data can be analysed, banks can better measure and
understand their trading activity.
3.2.4 Market and credit risk calculation
Risk management applies to a number of different aspects of the investment banking business, including
market risk and credit risk, and operates on multiple levels: individual trader, desk-, department-, or enterprise-
wide. As the scope increases, so too does the volume of underlying data that needs to be considered.
Market risk estimates the potential effect of adverse market movements on portfolios of financial
instruments.
Credit risk represents the measure of the effect on the banks trades due to counterparty failure.
To calculate market and credit risk, banks must make a statistical estimate of the future development of their
portfolios. This is done by generating thousands of potential scenarios and evaluating the whole portfolio based
on these market changes, a process that requires an enormous
amount of computing power. Fortunately, this is a highly parallelisable
computation and one that lends itself perfectly to distributed
computing. The results of this evaluation can then be combined to estimate the market and credit risk with a
high degree of accuracy.
Historically, these processes have been run using grid computing, an expensive and technologically complex
approach; today, these same processes can be completed far more quickly and effectively using commodity
hardware. According to the SAP report Big Data for Finance
13
, risk management is most effectively enhanced
through the adoption of big data technologies to create a system-wide trading database for oversight and
compliance, following the proposed Consolidated Audit Trail (CAT) standard.

13
BigDataForFinance.com from A-Team Group report for SAP - 2012


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 16/38
Detecting patterns in trading data is an
ideal task for big data analytics
Big data management tools are
perfectly suited for counterparty risk
monitoring
3.2.5 Rogue trade detection
Rogue trading continues to hit the front pages with depressing frequency and remains an important operational
risk for investment banks. Typically, banks place trading limits on their employees to ensure that no one
individual can increase the banks exposure (measured by its market risk) beyond an accepted level, but in the
most recent case, the trader bypassed these limits by hiding false
trades among the millions of legitimate trades.
The only effective way to prevent this is to detect patterns in the
trading data an ideal task for big data analytics. By analysing the huge volumes of data associated with
todays hyperactive trading market, it is possible to uncover the patterns and correlations that are the tell-tale
signs of rogue trading. Big data brings both the ability to manage huge volumes of data and the capacity to sift
through that data to discover problems before they can get out of hand.
3.2.6 Counterparty risk monitoring
As we all know, the financial crisis of 2008 began with the collapse of Lehman Brothers on September 15, a
collapse which was brought about by the banks over-exposure to the subprime mortgage market. The knock-
on effect, however, meant that every bank doing business with Lehman was also impacted.
Banks regularly undertake counterparty risk monitoring to protect themselves against just such a situation by
analysing trading activity and measuring both direct and indirect exposure. To do this effectively, they need to
collect a broad array of market data, potentially including information
from sources such as social networks, to take the pulse of the market
and detect potential problems before they occur. Once again, big data
management tools are perfectly suited for this kind of task; they are
able to collect data from a wide variety of sources in a wide variety of formats and combine them under a single
umbrella. Once thats done, banks can develop algorithms to detect counterparty risk in order to take timely
action.
3.2.7 Regulatory reporting
One of the most onerous tasks facing investment banks today is regulation. In the last 10 years, a number of
regulations have been passed which require investment banks to measure and report activity in a consistent
and accurate way. Markets in Financial Instruments Directive (MIFID), Sarbanes-Oxley, Basel III, FATCA, and
Dodd-Frank are just a few of the more significant regulations currently affecting bank operations in Europe and
the U.S.
These regulations require banks to report across all their asset classes, necessitating bank-wide views of
operations and activities. This entails not only the management of huge volumes of data, but the rapid and
accurate processing of that data in order to deliver timely reports to the compliance authorities. Big data
technologies enable both distributed storage and computing, which together provide an effective framework on
which to build regulatory reporting systems.
One clear example of this is the Volcker Rule. As part of the Frank Dodd Reform Act, the Volcker Rule requires
banks to report on the inventory aging (the length of time an instrument has been held by the bank) of all their


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 17/38
Big data technologies provide an
effective framework for regulatory
reporting systems
Big data can create customised
incentive programs for drivers
positions. To make this calculation correctly, the bank needs to hold a
view of its tens of millions of positions for every day up to one year
clearly a big data problem. Parallel storage and computing enable the
consolidation of these data and the computing power to process them
effectively.
3.3 Insurance
The insurance industry, as with retail and investment banking, has much to gain from big
data. By gathering and analysing data about their customers, insurance companies can gain
significant advantage over their competitors.
3.3.1 Premium calculation
Until very recently, the insurance industry had used statistics based on global, generalised variables to
calculate insurance premiums. However, this resulted in a level and quality of service which was not
customised to its customers. Big data technologies permits the use of customer data to enable a more
complex, and personalised, mode of calculation.
Under the old system, if a customer went to buy car insurance, the company would ask for age, type of car,
annual distance driven, and a few other variables in order to calculate
insurance premiums. However, new insurance models, and big data
technologies, allow for the passive collection of data from the car itself
(driving style, speed, times and places of use, fuel consumption, roads used, etc.) which in turn enables the
application of pay-as-you-drive car insurance using real-time risk analysis. Such systems can also be used to
help drivers reduce their premiums by improving their driving skills, making them a lower insurance risk. This
data is an ideal foundation for the creation of special offers that reward drivers for reaching established safety
goals.
3.3.2 Fraud detection and prevention
Fraud has always been a major challenge for the insurance industry
14
; its the second most costly white-collar
crime in America, and its easy to see its impact on the market from these numbers:
The property-casualty insurance industry pays out about $20 billion a year in fraudulent claims
At least 10 percent of all property-casualty insurance claims are either inflated or outright fraudulent
Insurance fraud raises insurance premiums on the innocent by approximately $300 per household per year,
affects every type of insurance, and takes many forms, from underwriting fraud to staged accidents to
conspiracy. Every $1 invested in workers compensation anti-fraud efforts returned $6.17, or $260.3 million in
total, in California in 2006-2007.
15


14
http://www.maif.net/site/insurance/insurance-fraud-faqs/
15
California Insurance Department annual report 2007


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 18/38
Big data analysis can uncover and stop
the activities of fraud networks
Fraud is an equal-opportunity crime, undertaken by people from all walks of life and all echelons of society, and
conducted in many different ways. There is no standard methodology
for fraud. By mining data from a multiplicity of sources, big data
analysis can uncover hitherto inaccessible patterns and connections.
For example, the relationships between claimants can reveal a network of individuals engaged in the
perpetration of a fraud. Insurance companies are able to reduce the incidence and cost of fraudulent claims by
analysing relationships among subjects and broadening the pool of source data. By applying complex anti-fraud
rules in real time and rapidly analysing and processing huge volumes of heterogeneous data, companies can
detect potential fraud early.
3.3.3 Customer Segmentation
In a business where profits are made or lost by measuring the risk associated with each customer, knowing
each customer well is absolutely critical. As with retail banking, having a 360
o
view of customer activity enables
companies to segment customers and create personalised services. In the insurance business, these might
include tools to enable drivers to adapt their driving style to save fuel or reward customers with a drop in their
premium if they achieve certain goals. By identifying preferred customers, insurance companies can target
them for up-selling as well.
These are the kinds of incentives that directly benefit the customer, more closely link the client to the company,
and will tend to improve customer loyalty over the long term.



Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 19/38
NoSQL databases make integrating
siloed data stores easier
Big data technology ensures easy
retrieval of transaction histories
3.4 IT efficiency
Across all industries, there are common use cases for big data technologies. Either by
making existing processes run more efficiently or by re-architecting processes completely,
new technologies can make important improvements to operational efficiency.
3.4.1 Improved operations
Banks and insurance companies typically operate hundreds of different systems to support hundreds of
different business processes, which spawn thousands of data integration processes. These processes sit for
the most part in siloed relational databases. IT processes have traditionally used ETL processes to transfer
data between those databases, but such a system is inefficient and expensive to maintain. Every single change
in a field in a table of a database can mean weeks of work to change the whole dependent chain of systems.
Systems that are built using big data technologies can greatly reduce
this problem by allowing the storage of data in heterogeneous formats,
thus not requiring that fields to be defined nor included in advance.
The integration process is therefore not affected by changes to the structure, and vice versa.
3.4.2 Database and system archiving
Regulatory bodies require that financial services companies maintain a history of their transactions for years
beyond the expiry of the initial transaction. They must not only keep the structured data from finance, risk, and
accounting systems, but also retain emails, instant messages, and other unstructured data to provide context
about those transactions. Historically, banks have done this by independently archiving the data and
documents from each system and retrieving them in response to a specific enquiry. However, this is a difficult
and time-consuming task when the required data is stored across dozens of different systems on different
platforms using different formats.
By using big data technology, banks can archive all their databases
and unstructured data sources in a single repository. This means that,
when an enquiry is received, the history of any particular transaction
can be quickly and easily retrieved by using the systems integral search capabilities a significant
improvement over todays fragmented processes.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 20/38
Big data architecture is associated
more with Google and Amazon than
with traditional enterprise systems
Hadoop and NoSQL are at the heart of
big data technologies
4 The architecture of big data solutions
Big data technologies originated at Internet-driven companies like Google, Yahoo, Facebook, and Amazon. So
it is not surprising that these technologies seem to more naturally address the requirements of such companies
than of traditional businesses which do not have to solve Internet-scale problems as part of their core business.
However, the trends identified previously are causing this distinction to
blur, as brick-and-mortar companies search for ways to efficiently
manage the wealth of data now at their disposal.
Big data solutions are typically based around four core technology
trends:

4.1 Big data technologies
Three key technologies underpin all major current implementations of
big data architectures: the Hadoop open source framework, NoSQL
databases, and event management platforms to support the real-time
processing of large streams of data.

Distributed
storage


The ability to store huge volumes of data across multiple servers, with
essentially no cap on the amount of data that can be managed.

Distributed
computing


The ability to distribute (and speed up) the processing of that data
across multiple servers, breaking up large jobs into multiple smaller and
more manageable ones without loss of integrity.

Unstructured
data storage


The ability to manage data from a variety of sources and in a variety of
formats: relational databases, documents, system logs, data feeds,
social media, and more.

Real-time
analysis


The ability to analyse and process data as soon as it becomes available
instead of waiting for daily batch processing schedules.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 21/38
4.1.1 Hadoop
Hadoop is an open source framework which employs distributed storage and processing using an adaptable,
reliable, and scalable programming model. It has been developed as an open source initiative under the
Apache umbrella that drives so much of the Internets infrastructure. Hadoop comprises two core modules:
HDFS the Hadoop Distributed File System, which manages the distribution and replication of data
throughout a cluster of data servers. Multiple copies of the data are held on different servers to enable
failover if any single server goes down, there is no loss of data.
MapReduce a programming model that supports distributed computing processes via a standard
Java API originally developed by Google.
Hadoop runs on commodity, low-cost hardware and offers a host of other software components, including;
HBase columnar data store for large data sets
Hive data warehousing and SQL-like query capability
Mahout machine learning and data mining capability
Pig high-level language for expressing data analysis
Flume data collection and loading
Sqoop data exchange for relational database connectivity
Zookeeper process coordination and synchronization
Oozie workflow management for Hadoop
By combining these software components, a wide array of technological solutions can be built to meet any big
data need.
4.1.2 NoSQL databases
The NoSQL name originally meant simply Not SQL, but it has evolved today to the more flexible Not Only
SQL. These databases, unlike relational databases:
generally cannot be accessed using SQL standards
do not for the most part support transactionality
are schema-less
are in some cases built on top of Hadoop in order to take advantage of distributed storage and
computing
The main NoSQL databases can be classified as follows
16
:
In-memory: provides a quick response for dynamic data. Example: Redis
Document-oriented: stores structured documents in XML/JSON formats. Examples: MongoDB,
CouchDB, MarkLogic
Key-value: stores data "key-value" pairs from data such as weblogs. Examples: Cassandra, HBase,
Redis (see in-memory databases above)

16
DB-Engines Ranking, 2014


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 22/38
Event management is an essential
component of an effective big data
solution
Search engines: databases used to generate indices that enable fast searches on the stored data.
Examples: Solr, Elasticsearch
NoSQL databases specialize in storing heterogeneous (both structured and un-structured) data side by side
and providing the tools to search and analyse them.
4.1.3 Event management
Sometimes the need to respond to events in real-time is even greater than the need to store those huge
volumes of data. In those instances, event management is an essential component of an effective big data
solution. Event management platforms such as CEP (Complex Event
Processing) and RTD (Real Time Decision) systems are able to
manage multiple types of events in a split second and determine
appropriate action based on the input received by the decision engine.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 23/38
All major vendors are offering big data
tools
The data warehouse model will in
future be complemented by big data
systems
4.2 Commercial appliances
There are many commercial implementations of big data technology on the market already, which is not
unusual considering that the original technology was open-source. Today, the technology has matured to the
point where corporate services such as security and access control
are being added to delivered solutions, along with more formal
support, version control, and release management. This growing
maturity is reflected in the inclusion for the
first time of a pure big data player, Cloudera,
in the latest Gartner Magic Quadrant
17
for
Data Warehouse and Database Management
Systems.
As we can see, all the major vendors are
already offering implementations:
IBMs InfoSphere BigInsights
Oracles Big Data Appliance, which
combines relational and NoSQL
databases
SAP HANAs in-memory, NoSQL
database
HPs Vertica
EMCs Pivotal (acquired from
Greenplum)
4.3 Impact on existing systems
Will this technology render existing data warehouse and business intelligence environments obsolete? No, but
some tasks may be accomplished in a different way, such as the elimination of intermediate transformation
processes developed with classic ETL tools in favour of ETL based on Hadoop MapReduce or none
whatsoever, instead using NoSQL to store data with heterogeneous formats.
The data warehouse model will remain an essential part of many corporate processes that involve largely
structured data, such as financial and management reporting, but it will in future be complemented by big data
systems that will enrich those models with additional, particularly non-structured, data. Hadoop will be used to
collect all the data from the enterprise and permit analysis of the wider data set; there will still be data
warehouses, but not the need to create specific data marts for every
reporting purpose.
In this scenario, the big data environment will process all the
structured and unstructured information, remove the noise, add value,

17
Gartner Magic Quadrant for Data Warehouse and Database Management Systems, 2014


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 24/38
Financial services will likely continue to
lag in adopting cloud-based services
consolidate, and deliver the clean, high-value data to the data warehouse, which will provide context to the
business analyst working with that data.
The data warehouse will also continue to respond to the needs of recurring reports and information analysis;
the big data platform will provide new data not previously addressed, and enhance the value of the output
through automated decision-making.
4.4 Cloud architectures
No discussion of big data architecture is complete without acknowledging the role of the cloud. Use of cloud
computing, which provides computer and IT services as a utility, is growing rapidly across all industries.
The first wave of cloud computing focused on improving IT department efficiency and reducing the need to
manage ever-expanding infrastructures in-house. This offloading of resources from the physical business
environment will continue with the next area of focus, Business as a Service, which will further externalise
business services.
The conservative nature of the financial services industry, however, is likely to continue to act as a drag on the
expansion of cloud-based services. Gartner predicts that, while 90% of
organisations across the board will store personally identifiable
information in the cloud by 2019, only 60% of banks will be conducting
the majority of their transactions in the cloud. Cloud service providers are now responding to bankings
hesitance with new and innovative services which more closely address data security and compliance
requirements.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 25/38
A structured, iterative approach will
deliver the best results
Big data systems become more
challenging as they grow
Specialised staffing is key for
successful deployment
5 Addressing big datas challenges
As one might expect, IT departments face a number of challenges in considering a shift to big data-centric
systems. In this section, we highlight four of the biggest issues organisations should pay close attention to in
planning their move.
5.1 Technology
The continuous evolution of technology is a challenge in and of itself; given the high rate of
innovation, many technologies never reach maturity and can quickly become obsolete. It is
important to make smart decisions about which technologies to bet on.
Big data systems are by their very nature distributed systems, normally with significant scale. This means that
software architects must be prepared to deal with issues like partial failures, unpredictable communications
latencies, concurrency, consistency, and replication in the process of
designing the system. These issues become increasingly more
challenging as systems grow to encompass the use of thousands of
processing nodes and disks, geographically distributed across multiple data centres. The probability of failure of
a hardware component, for example, increases dramatically with scale.
Scale also impacts the economics of big data projects. Big data applications can require a huge volume of
computing and storage resources. Regardless of whether these resources are covered by capital expenditure
or hosted by a commercial cloud provider, they will be a major cost factor and thus a target for budget or scope
reductions. A straightforward resource reduction approach such as data compression is a relatively simple way
to reduce storage costs. Elasticity is another way in which resource usage can be optimised, by dynamically
deploying new servers to handle increases in load and releasing them as the load decreases.
The successful deployment of big data technologies requires
specialised staffing, which can also be expensive, especially given the
immaturity of these technologies and shortage of specialist resources.
To mitigate the potential risks associated with scale and technology, organisations should adopt a systematic,
iterative approach to ensure that initial design models and technology selections can support the long-term
scalability and analysis needs of a big data application. A relatively
modest investment in upfront design can produce a major return on
investment in terms of reduced redesign, implementation, and
operational costs over the lifetime of a large-scale big data system.
Because the scale of such systems can prevent the creation of full-fidelity prototypes, a well-structured
software engineering approach is needed to frame the technical issues, identify the architecture decision
criteria, and rapidly construct and execute relevant but focused prototypes. Without this structured approach, it
is easy to fall into the trap of chasing after a deep understanding of the underlying technology instead of
answering the key go/no-go questions about a particular technology option.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 26/38
Context is key when deciding who gets
access to what data and when
5.2 Security and data protection
It is often assumed that the primary security problem with big data is access management and
control. In reality, however, the loss of context that results when large amounts of data are
aggregated becomes a bigger issue because it impacts the ease with which the bank can
preserve granular rights for accessing those data. For example, a certain user may have access to certain data
entities (or tables), but which data (or rows) within those entities can be accessed? This is true for relational
databases as well, but, by combining all the data in one place, the problem is exacerbated.
Data access is not the only challenge. For example, even the most well-intentioned users can make mistakes,
such as deleting huge stacks of data, in seconds by accidentally
executing a distributed delete. Less well-intentioned users could
decide to lower the priorities of other Hadoop jobs in order to ensure
their job is completed ahead of the others or worse, kill those other jobs.
So how can an organisation keep carefully constructed data ownership and data context rules in effect without
killing the benefits that led it to choose a big data solution in the first place?
While a highly scalable architecture like Hadoop does make it possible to store context alongside data,
checking the entire context for each piece of data is an expensive proposition. Differential privacy, which aims
to provide a means by which organisations can maximise the accuracy of queries from statistical databases
while minimising the likelihood of any records being identified, is one concept thats being explored.
The need for security hasnt been lost on the largest vendor of Hadoop software and services. Cloudera has
been constantly improving the security of the platform and this year has made further acquisitions to improve
future security.
5.3 Data quality
Data quality is imperative in the financial services sector. In order to extract information and
insight from the data, its quality has to be well understood and the data has to be trusted. When
bringing a wide array of data together from across the organisation, it is especially important to understand the
relationships and dependencies between them.
A clear methodology must be established which includes core data management principles:
Data governance true accountability and commitment to quality can only happen with an effective
governance and control framework in place which identifies data owners and stewards
Data architecture understanding data architecture (relationships between data) and standardising
data interfaces between systems will streamline data flows and help identify and resolve the root cause
of many data quality problems
Data lifecycle identifying data quality issues at each stage of their lifecycle is key to understanding
data lineage: the sourcing, transformation, aggregation, and consumption of data
Reference data establishing consistent reference data will eliminate inconsistencies and ensure a
common view across the organisation


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 27/38
An agile structure is required to take
advantage of big data
Data analytics in order to fully empower business users and enable decision-making, the data must
be fully trusted and the systems must be highly usable. Rich, graphic dashboards, multi-dimensional
analysis and drill-down capability, and what-if scenario analysis are also required.
Data technology the optimum mix of technology which meets data quality business needs will enable
data profiling, matching, merging, and cleansing processes
Without high quality data which is fully trusted by the business users, no big data initiative can be successful.
By establishing a clear data governance model and methodology, quality can be controlled and the maximum
value can be extracted from the data.

5.4 Internal organisation
The financial sector consists primarily of large organisations with complex, hierarchical
structures that do not easily (or cheaply) change direction. A more agile structure is required to
take advantage of big data, and this may prove to be the most difficult challenge to overcome. Data siloes that
are scattered across the organisation must be merged, and new
staffing roles created, to deliver the essential 360
o
view of the
customer and trade data a big change from the way most large
financial services organisations function currently.
Embracing this change in organisational structure will involve several key decisions. The business must:
determine a strategy for big data deployment
assign responsibility for the collection and ownership of data across business functions
plan how to extract useful information from the data
prioritise opportunities
allocate data scientists time appropriately
host and maintain the IT infrastructure
set privacy policy and access rights
determine accountability for compliance with local data protection mandates
Organisations must plan carefully as they navigate their way towards big data readiness. Companies will likely
implement one of the following four organisational models:
Business unit led: when business units have their own data sets and scale isnt an issue, each
business unit can make its own big data decisions with limited coordination. AT&T and Zynga are
among the companies that use this model.
Business unit led with central support: business units make their own decisions but collaborate on
selected initiatives. Google is an example of this approach.
Centre of Excellence (CoE): an independent specialised division oversees the companys big data
initiative. Each unit pursues appropriate initiatives, guided and coordinated by the CoE. Amazon and
LinkedIn rely on CoE.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 28/38
It is important to choose the right model
for your organisation
Fully centralised: the companys central operations take direct responsibility for identifying and
prioritizing big data initiatives. Netflix is an example of a company that pursues this route.
Companies must consider how they want to use its data and which
model is most appropriate for its organisation.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 29/38
Approach 1: Test the new big data
technology in non-critical IT initiatives
Approach 2: Create a CoE that focuses
on building competence in big data
6 First steps
Incorporating new technologies and processes into a large, hierarchical corporate structure is a major
challenge, particularly when the expectations are as high as they are with big data.
6.1 Getting started
Financial institutions have already begun to explore leveraging big data to improve their business. We have
identified three common approaches which are representative of how banks are incorporating these new
technologies into their organizations:
IT operations project
This approach is the most common within large organisations which do not have good coordination between
business units. Here, big data infrastructure is directly created by IT that supports differing business areas in
order to test the technology in non-business-critical initiatives. These
test initiatives may focus on streamlining IT departmental functions
and operations but would not be related to establishing company-wide
architectures or new business models. Such initiatives can be used to improve operation times (itself a direct
business benefit), to understand the value the new technology brings in a low-risk environment, and to build
staff competency by enabling the development of familiarity with new systems.
This creates a foundation on which more visible initiatives can be built in the future and can be adopted more
centrally. However, this approach tries to make order from chaos, instead of planning and organizing from the
beginning.
Establishing a Big Data CoE
An innovation department exists within the structure of the corporation to bring new ideas to the organisation. It
generally has a high-level sponsor, is provided with a technology budget, and has at its core a dedicated team
that owns technological innovation within the company.
The innovation department creates groups of technologists which concentrate on specific topics; this often
results in a Centre of Excellence (CoE) of key resources who can build
out competence in these technologies. The technological resources
used by the CoE are initially provided by the corporate IT department
and vendor support teams until such time as specialist staff are hired and can take on the mantle of knowledge
ownership.
The CoE staff is then 100% dedicated to big data initiatives, responsible for extracting maximum value from the
technology and staffing investments. The Centres head must be able to bridge the gap between business and
technology to understand, manage, and communicate all facts of the project to all levels of the organization.
The centre itself must have:
a strong team with deep knowledge of all aspects of big data and good exposure to different
technologies


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 30/38
Approach 3: Spin-off competence in
order to become more agile
sufficient resources to fully focus on the initiative
a high-level sponsor within the organisation with a clear strategy for implementing big data projects
the ability to work with business and IT departments and being able to provide support into IT
a close relationship with an external provider that can bring an additional dimension to the project
clear goals and objectives, whether for internal (developmental) or commercial purposes
the ability to bring in resources from outside while the in-house capacity is being built out.
Creating an experimental spin-off
The CoE helps build infrastructure and capability in the technology and promotes it to the various business
units in the organisation, providing them with business value. As business uses are identified, projects are
proposed to and adopted by line-of-business teams (with support of the CoE) in an organic fashion.
These may grow to a level where they may be spun off into separate business units, rather like start-ups. This
type of operation will likely include teams of data scientists who will
leverage the branding, capacity, and funding of the business outside
the confines of the corporate structure, enabling them to remain
flexible, make rapid decisions, and deploy new business models at will.
This approach of externalising the banks technological competence and available data requires a mature
organisation and a well-developed business model, but it can also provide totally new revenue streams.
In reviewing the above examples, we can identify certain common characteristics:
It is unrealistic to expect a fast return on investment; initial activities must be perceived more as R&D
than fully rounded business plans.
These exercises have potentially far-reaching effects across the organisation. It is therefore imperative
that they have a corporate sponsor in the form of someone who can make decisions at a high level and
promote and defend the project in the C-suite.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 31/38

Financial institutions are making major investments in big data, so there is clearly a strong belief within
the sector that there is value to be extracted from this data that may convey a competitive advantage to
the fastest movers.
Deriving value from big data means cutting across silos to bring together and analyse diverse data from
across the organisation and using new technology thats better suited to dealing with multiple types of
data.

6.2 Establishing new roles in the company
As these big data initiatives evolve, new roles need to evolve to promote cross-business communication and a
common focus to meet the corporations broad goals. Expect to see the following job descriptions begin to
emerge:
Chief Data Officer (CDO)
The CDO is responsible for all the data within the organisation and plays a significant role in overseeing the
whole data environment. Much emphasis is being placed on this role in investment banks today, because
they understand that good quality data and data analysis are fundamental to the business. The CDO must
also ensure that all applicable regulations regarding governance, security, and privacy are adhered to, and
that access controls are tightly managed.
Data Scientist
Data scientists play one of the most important roles in the deployment and management of an effective big
data solution; their focus is entirely on extracting optimum value from the data they manage, via
mathematical modelling to identify correlations between data, segment and generate predictive models,
mine the data, and more. Data scientists need to understand the business, the data, and the technology
although in the latter instance, excellent analytical capability is more important than the deep, detailed
architectural skills required to design and build code. Data scientists are more likely to take advantage of
end-user tools to extract meaning from the data.
Chief eXperience Officer (CXO)
When the customer is at the centre of the corporate strategy, all customer interactions with the company
through every channel must be managed globally. The CXO is responsible for developing and maintaining
the 360
o
view of the customer, much of which will be achieved through the deployment of big data
initiatives. While the CXO is not directly responsible for the data itself (thats the province of the Chief Data
Officer see above), he or she is responsible for optimising the customer experience through the
appropriate analysis and application of that data.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 32/38
6.3 Strategies to follow lessons learned from successful big data projects
This paper has highlighted key use cases for big data in the financial services industry and how banks and
insurance companies have begun to implement them within their organisations. GFT, working with clients
across retail banking, investment banking, and insurance has gained significant experience in designing and
building big data systems. From this experience, and many successful big data projects, we have identified
certain common lessons learned:


1

Bring IT and
business together

IT must understand the business need and build a
cross-functional team to support short, mid, and
long-term goals. Involve business analysts who can
effectively bridge the gap between end users and
the core IT team.

2

Control
the data


Start out with only internal data. By minimising the
use of data from external sources that you cannot
control, you minimise the risks involved in uncertain
data quality. Bring all available data into a single
location, but dont filter it initially. Build user trust in
the data by directly involving them in the data
collection / provision process.

3

Define the right
technology

Take your time in deciding on the technology to be
used both now and into the future, but dont
overthink your decision. Remember that the
technology is dynamic, so focus on stability and
flexibility of integration with your current
infrastructure in choosing your starting platform.
Consult experts for an objective opinion.

4

Install the necessary
infrastructure

Although Hadoop clusters offer a highly scalable
infrastructure, think beyond your first project. Your
use of this platform will grow rapidly and you need
to ensure it will support that growth. Consult experts
to ensure the infrastructure is configured
appropriately for your performance requirements
and expectations. Budget appropriate resource
levels for monitoring and maintenance.


Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 33/38

5

Choose your first
project carefully

The first project will have a great deal of visibility.
Work iteratively and build out functionality little by
little, but be sure to show return on value as early in
the process as possible to ensure continuous
executive sponsorship. Often, the maximum value
will come from merging data that is currently in
separate silos a clear case of 1+1 = 3. Create
mockups of a user interface and design to ensure
user acceptance.

6

Establish the right
team

Define the scope and responsibilities of your Centre
of Excellence. Complement in-house capability with
third-party support. Make sure you have a team that
encompasses both technology and business
knowledge to collectively meet the needs of the
project.

7


Gain operational
buy-in

Ensure that everyone is appropriately trained and
that IT operations are equipped to run the new
systems from both a functional and a technological
perspective. Systems with high usability which
effectively meet the business need are imperative.



Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 34/38
GFT has 25 years experience in
financial services technology projects
7 GFTs Big Data practice
GFT helps financial services firms extract optimum value from the ever-accelerating deluge of data. We help
firms meet their regulatory requirements and better serve their corporate, retail, and institutional clients by
managing data in a more efficient way, which in turn reduces cost and risk.
We offer industry and business insight as well as the latest technology, tools, and techniques - designed to help
clients grow, scale, innovate, and compete. Our breadth of industry
knowledge, along with our rigorous approach to delivering quality,
strategic planning, client collaboration, project management, and
passion for innovation, allows us to continually add insight and value throughout the engagement process.
Employing big data technologies, GFT has successfully implemented a number of solutions for financial
institutions, enabling those institutions to extract greater value from the huge volumes of data they are
processing every day:
Debit/credit postings are calculated in real-time through trade message processors, managing
hundreds of millions of daily balances in different asset classes around the world
Regulatory reports are accurately calculated on a daily basis using the age of an investment banks
trading positions, trawling through four petabytes of data in just 20 minutes
Deep financial insight into bank customers spending habits is provided via a highly usable web
interface which allows them to sort, categorize, and visualize over 10 years of transaction data
The incidence and cost of fraudulent insurance claims is reduced by analysing relationships among
subjects and broadening the pool of source data.
5.5 million trades are stored on a daily basis in a trade repository, with the ability to hold more than
seven years of data (10 billion trades or approximately six petabytes of data), facilitating cutting-edge
deep-dive analysis
Fraudulent credit card transactions are minimised by implementing a set of rule-based filters and
adaptive pattern recognition methods through a logic engine to detect suspicious payments.
Millions of daily trade events are stored and managed centrally, reducing data duplication,
inconsistency, and process redundancy.



Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 35/38
About the authors
This Blue Paper has been developed by specialists from our Big Data and Innovation practices:

Dr. Karl Rieder (GFT)

Karl Rieder is a Director at GFT and leads its Big Data Practice. He helps clients with
the challenges of business innovation, defining technology strategies and delivering
applications for retail and investment banking. He acts as the bridge between clients
who want to understand how new technologies can help them and the technology
experts who create engaging applications.
Karl has previously worked as a senior project manager and lead technical architect,
through which he has gained deep banking industry knowledge and more than 18
years IT management experience. He publishes a mobility and big data blog on
Finextra.
Karl holds a Bachelors Degree in Mechanical Engineering and a PhD in
Oceanography from the University of California, San Diego.

Dr. Ignasi Barri (GFT)

Ignasi Barri is a member of the Applied Technologies group at GFT.
His passion for business led him to study for a bachelors degree in business
administration. His fascination with technology and business helps him to
achieve his primary goal at GFT: to empower innovation across entire
organisations. His particular focus is on the identification and promotion of
outstandingly talented people with creative ideas.
He received his PhD in Computer Science from the University of Lleida,
Spain, in 2012.




Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 36/38
Josep Tarruella (freelance)

With more than 14 years of IT experience, Josep has focused his career on
the field of data management. He has expertise in the management of teams
and projects in Business Intelligence, data warehouses, and ETL.
Josep has implemented projects in large corporations in many industry
sectors such as pharmaceuticals, banking, telecommunications, and
insurance. Through his work with technology consulting firms, he has played
many roles, including technical consultant, pre-sales consultant, and services
director.
He holds a Degree in Computer Science from the Polytechnic University of
Catalonia and an Executive MBA at the IESE Business School.





Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 37/38



Big Data Uncovering Hidden Business Value in the Financial Services Industry | GFT Technologies AG 2014 September 2014 Page 38/38


Contact:
www.gft.com/bigdata
bigdata@gft.com

Vous aimerez peut-être aussi