The role of Big Data in Banks

I was listening at R. Martin Chavez, Goldman Sachs deputy CFO just last month in Harvard at the ComputeFest 2017 event, more precisely, the SYMPOSIUM ON THE FUTURE OF COMPUTATION IN SCIENCE AND ENGINEERING on “Data, Dollars, and Algorithms: The Computational Economy” held in Harvard on Thursday, January 19, 2017.

His claim was that

Banks are essentially API providers.

The entire structure and infrastructure of Goldman Sachs is being restructured for that. His case is that you should not compare a bank with a shop or store, you should compare it with Google. Just imagine that every time you want to search on Google you need to get in touch (i.e., make a phone call or submit a request) to some Google employee, who at some points comes back to you with the result. Non sense, right?  Well, but this is what actually happens with banks. It was happening with consumer-oriented banks before online banking, and it’s still largely happening for business banks.

But this is going to change. Amount of data and speed and volume of financial transaction doesn’t allow that any more.

Banks are actually among the richest (not [just] in terms of money, but in data ownership). But they are also craving for further “less official” big data sources.

c4tmizavuaa1fc3
Juri Marcucci: Importance of Big Data for Central (National) Banks.

Today at the ISTAT National Big Data Committee meeting in Rome, Juri Marcucci from Bank of Italy discussed their research activity in integration of Google Trends information in their financial predictive analytics.

Google Trends provide insights of user interests in general, as the probability that a random user is going to search for a particular keyword (normalized and scaled, also with geographical detail down to city level).

Bank of Italy is using Google Trends data for complementing their prediction of unemployment rates in short and mid term. It’s definitely a big challenge, but preliminary results are promising in terms of confidence on the obtained models. More details are available in this paper.

Paolo Giudici from University of Pavia showed how one can correlate the risk of bank defaults with their exposition on Twitter:

c4tuo4yxuae86gm
Paolo Giudici: bank risk contagion based (also) on Twitter data.

Obviously, all this must take into account the bias of the sources and the quality of the data collected. This was pointed out also by Paolo Giudici from University of Pavia. Assessment of “trustability” of online sources is crucial. In their research, they defined the T-index on Twitter accounts in a very similar way academics define the h-index for relevance of publications, as reported in the photographed slide below.

dig
Paolo Giudici: T-index describing the quality of Twitter authors in finance.

It’s very interesting to see how creative the use of (non-traditional, web based) big data is becoming, in very diverse fields, including very traditional ones like macroeconomy and finance.

And once again, I think the biggest challenges and opportunities come from the fusion of multiple data sources together: mobile phones, financial tracks, web searches, online news, social networks, and official statistics.

This is also the path that ISTAT (the official institute for Italian statistics) is pursuing. For instance, in the calculation of official national inflation rates, web scraping techniques (for ecommerce prices) upon more than 40.000 product prices are integrated in the process too.

 

 

Again on BPM posting. Responses to Sandy Kemsley

Maintaining Consistency Across BPM Initiatives’ Content(November 3, Enterprise Irregulars)

In response to a writeup by Sandy Kemsley on a user experince on MDD+BPM at Bank of America.
Dear Sandy,
Thank you so much for sharing this writeup and comment.
Being a researcher on MDD and BPM, I’m glad to see that this recipe is getting more and more attention in the industry.
Being also a (software) analyst and a partner in a tool vendor company (WebRatio) I agree with you that pushing this combination of methodologies to the customers can be painful, because customers are rather conservative in terms of innovation and do not trust tools and approaches that are not in the orthodox mainstream.
However, we also had some really good success stories in making big customers adopt these approaches (see for instance our experiences we shared at BPM 2010 in Hoboken and that will also be published in an upcoming book edited by M. Rosemann and M. zur Muehlen. Presentation here: http://www.slideshare.net/mbrambil/web-ratio-bpmindustrialcasesbpm2010 ).
I share with you the concern about setting up weird combinations of tools for addressing BPM with a MDD approach.
In WebRatio, it took us 10 years to come up with a coherent suite of modeling, transformation, and code generation facilities that integrate in a seamless way BPM, data modeling, application modeling, quick prototyping, and code generation for production applications.
Since MDD is a very neat and conceptual approach, I think appropriate toolsuites that correctly deal with the models and their relationship is crucial for granting full benefits of a virtuous model-driven lifecycle.

Enabling Smarter BPM through Business Entities with Lifecycles (November 3, Column2 blog by Sandy Kemsley)

In response to a writeup by Sandy Kemsley on Rick Hull work on integrated  BP and data modeling at IBM T.J. Watson.
Thanks for this overview on Rick’s work. I think this is a crucial change of paradigm that companies will embrace sooner or later.
It doesn’t make any sense to consider data and processes as separate realms, with different kings and armies fighting at their borders.
I also have been doing integrated modeling of different aspects of applications for a while, not limited to BP and data models, but comprising also other aspects like user role modeling, application logic modeling, user interaction modeling, and so on. In our approach we envision an integrated model driven development (MDD) approach covering all these aspects and providing automatic alignment and quick prototyping.
I strongly believe that model integration, alignment and continuous cross-validation down to the implementation is the only response to the complex needs of today’s enterprises.
Here is an overview of what we do:
http://www.slideshare.net/stefano_butti/webratio-a-mdd-approach-to-bpm-4872510
And a summary of MDD responses to current BPM trends (including MDM integration):
http://www.slideshare.net/mbrambil/webratio-bpm-trends-and-challenges
I would be glad to get feedback on this.

Iterative Development in BPM Applications Using Traceability (November 3,Column2 blog by Sandy Kemsley)

Dear Sandy,
thanks for this post too. I appreciate your work, I think it’s big value for the BPM community.
I think traceability is crucial, but may not be enough. What we really need is coherency and alignment of models. The complexity of modern enterprise applications requires several models to describe all the relevant aspects. Design tools should be aware of this and *grant* alignment by construction.
This could be fairly easy for models that deal with the same level of abstraction, while it could be much more difficult when alignment is needed between business and IT models of course.
Furthermore (maybe even more crucial), alignment should be granted between models and implementation.
In my experience, all this can be achieved only through well engineered model-driven development (MDD) approaches.

We are having an experience in a large banking scenario on integrating BPM, MDD, SOA approaches and the lesson learned is exactly in line with this presentation. We:
– collected requirements from business users (managers)
– analysts built the models
– we showed them the models — and we showed the end users only running prototypes
– and we collected feedback.
Guess what? Most of the feedback came from the prototypes.. hence, the business users understood the models but still were not able to focus on the real issues (especially related to exceptional cases).
[btw, we will present our experience and demonstrate our WebRatio BPM tool at BPM 2010, Hoboken, Sept 15th, 2010]