Vous êtes sur la page 1sur 118

SIX WEEKS SUMMER TRAINING REPORT ON

Data Analytics And Business Intelligence





UNDER THE GUIDANCE OF:-
MR. Nikesh Bajaj
Lovely professional university
Phagwara, Punjab



SUBMITTED BY:-
ABHISHEK KUMAR
REG. NO:-11104669
SECTION:-E2109

School of Electronics and Electrical Engineering
Lovely Professional University, Phagwara, Punjab
(June-July, 2014)

DECLARATION



I hereby declare that I have completed my six weeks
summer training at Lovely professional University (name
and location of organization) from 9/06/2014 (start date) to
18/07/2014(end date) under the guidance of Mr. Nikesh Bajaj
(Name of Industry coordinator). I have declare that I have
worked with full dedication during these six weeks of training
and my learning outcomes fulfill the requirements of training for
the award of degree of B.tech 3
rd
year ECE (Relevant Degree ),
Lovely Professional University, Phagwara, Punjab.






(Signature of student)


Name of student:-Abhishek Kumar
Registration no:-11104699
Date :-03/08/2014
(I)
ACKNOWLEDGEMENT


I would deeply like to express my gratitude to a few
people whose help and guidelines encouraged me
throughout the project.
Firstly, I express my deep thanks to my training
coordinator, Mr. Nikesh Bajaj, for the way he taught us.
The effort, he has made while teaching us was helpful in
project completion also.
Secondly, my DBMS teacher, Mr. Ravindra Singh Who
along with giving this interesting topic, showed faith in
me and I have completed this by burning my night oil in
the best possible way which I can. They were after me for
the whole course of this project.
Nextly , for this project, I would like to thank my friends
for helping and showing me the right path for the
completion of this project.
Last but not least, The Almighty, for everything he has
blessed me.

-Abhishek kumar
11104669





(II)

Certificate















(III)
PERIODIC APPRAISAL PERFORMA
(To be filled by External Advisor)

Name of Student:__________________
Registration No:_____________
Project/TechnologyLearnt_____________________________________
Name & Address of organization ______________________________
Name of External Project Supervisor (with Phone No)_______________

__________________________________________________________

Period of evaluation: From _________To _________


(IV)

General Remarks / Observations with regard to deficiencies /
problems / suggestions for improvements:

____________________________________________________
____________________________________________________
____________________________________________________
____________________________________________________
____________________________________________________
____________________________________________________




___________________________________________________

Signature of External Project Supervisor/Guide (With Seal, Date
& Designation)


(V)
INDEX




Chapter Contents Page No
1.
Organization Overview 1
2. Technology Learnt 7
3. Motivation In The training 31
4. Learning Outcome
5. Gantt chart
References







(VI)


Chapter One :-Organization Overview



Type Of Organization:- Cognizant Technology Solutions Corporation is
a multinational information technology, consulting and business process
outsourcing company. It is headquartered in Teaneck, New Jersey, United
States, North America.
This is the first Indian software services firm to be listed on the NASDAQ.

Cognizant provides information technology, consulting and BPO
services. These Include business & technology consulting, systems
integration, application development & maintenance, IT infrastructure services,
analytics, business intelligence, data warehousing, CRM, supply chain
management, engineering & manufacturing Solutions, ERP, R&Doutsourcing
, and testing solutions.

Cognizant is organized into several verticals and horizontal units. The vertical
units focus on specific industries such as Banking & Financial Services,
Healthcare, Manufacturing and Retail.
The horizontals focus on specific technologies or process areas such as Analytics,
mobile computing, BPO and Testing.

Both horizontal and vertical units have business consultants, who together form
the organization-wide Cognizant Business Consulting (CBC) team. Cognizant is
among the largest recruiters of MBAs in the industry; they are involved in business
development and business analysis for IT services projects.

In 2011, the company's revenue from IT services was split roughly
evenly between application development and application maintenance.
Its business process outsourcing portfolio leans towards "higher-end"
services i.e., work that involves domain knowledge and skills, such as
legal services or healthcare claims.

(1)


Like many other IT services firms, Cognizant follows a global delivery
model based on offshore software R&D and offshore outsourcing. The company
has a number of offshore development centers outside the United States and near-
shore centers in the U.S., Europe and South America.

In its early years, Cognizant gained business from a number of
American and European companies with the help of the D&B brand.
The company's senior executives envisaged the firm as a provider of
high-end customer services on-par with the six contemporary major
system integrators
(Accenture, BearingPoint, Capgemini, E&Y, Deloitte
and IBM), but at lower prices.

In the 2012 earnings announcements, the CEO Francisco D'Souza
categorized the company's service offerings in three groups: Horizon
1 (application development and maintenance), Horizon 2 (BPO, IT
Infrastructure Services & business consulting) and Horizon 3 ("SMAC"
- Social, Mobile, Analytics and Cloud).As of September 2012, the
Horizon 1 services accounted for over 75% of the company's revenues,
and Horizon 2 services about 20%.




Training Location:- Lovely Professional University Phagwara, Punjab.
Location is good for student of LPU because any student can easily reach the
location. It was mine privilege to get my training done with one of the leading
business tycoon at my own college campus. Basically, Cognizant is a grown IT
company and it has several training venue but this is the first time when they have
just came inside our campus o train us .

(2)




Specialization Of/Work culture in Organization:- I have done my six
weeks summer training in the field of Data analytics and Business Intelligence.
Working environment under the training period is good and it is also interesting
because the facility provided by the organization is so nice, they provide us a good
kind of computer lab in which we can do our practices. All the faculty members
and staffs are helpful they interact with all the students and resolve the problems.
In the training period there was a nice feedback system all the feedbacks are taken
care.
They provide us study materials with a large no. Of mini projects which are
completed under the training period, and study materials are helpful when we are
doing our projects.

External Guide :-






(3)

Chapter Two :-Technology Learnt

Data base Management System:- A database management system
(DBMS) is a collection of programs that enables you to store, modify, and
extract information from a database. There are many different types of
DBMSs, ranging from small systems that run on personal computers to
huge systems that run on mainframes. The following are examples
of database applications.
Computerized library systems
Automated teller machines
Flight reservation systems
Computerized parts inventory systems

Database:- A database is an organized collection of data. The data are
typically organized to model aspects of reality in a way that supports
processes requiring this information. For example , modeling the
availability of rooms in hotels in a way that supports finding a hotel with
vacancies.

DBMS Features:- There are a variety of DBMS database management
tools that provide different features and functionality. Here are the most
probable and must-have features of a typical DBMS:
(I) Data Structuring:- All information in a digital repository is
organized into a clear hierarchical structure with records, tables or objects.
Every piece of information that you can add to your database will be
organized in a kind of catalogue, so it will be easier for you to search and
update your records later on.

(4)


II) Database Customization :- Along with default and necessary
elements (records, tables, or objects) that compose a database's structure, there can
be created custom elements that meet specific needs of users.

III) Data Retrieval :- DBMS database management system accepts data
input from users and stores it. Users can access the database later on to retrieve
their records as a file, printed, or viewed on the screen.

IV) Query Languages:-A typical DBMS makes it possible to use query
languages for collecting, searching, sorting, modifying and other activities that
help users manipulate their records in the database.

V) Multi-user Access :-DBMS provides multiple users access to all types of
information stored in one and the same data store. There is also a security feature
that prevents some users from viewing and/or modifying certain data types.

VI) Data Integrity:-All information in a database is accessible by several or
more users but only one user can change the same piece of data at a time. This
capability lets prevent database corruptions and failure.

VII) Metadata :-DBMS database management software provides a metadata
library (or a data dictionary) that describes how the database is organized and what
elements (objects, related files, records, etc.) compose its structure.

There are some commonly found features in DBMSs :-


(5)

Advantages Of DBMSs:-
The database management system has a number of advantages as
compared to traditional computer file-based processing approach. The DBA
must keep in mind these benefits or capabilities during databases and
monitoring the DBMS.

The Main advantages of DBMS are described below.
I). Controlling Data Redundancy :- In non-database systems
each application program has its own private files. In this case, the
duplicated copies of the same data are created in many places. In DBMS,
all data of an organization is integrated into a single database file. The data
is recorded in only one place in the database and it is not duplicated.
II) Sharing of Data :- In DBMS, data can be shared by
authorized users of the organization. The database administrator manages
the data and gives rights to users to access the data. Many users can be
authorized to access the same piece of information simultaneously. The
remote users can also share same data. Similarly, the data of same
database can be shared between different application programs.
III) Data Consistency :- By controlling the data redundancy, the
data consistency is obtained. If a data item appears only once, any update
to its value has to be performed only once and the updated value is
immediately available to all users. If the DBMS has controlled redundancy,
the database system enforces consistency.
IV) Integration of Data :-In Database management system, data
in database is stored in tables. A single database contains multiple tables
and relationships can be created between tables (or associated data
entities). This makes easy to retrieve and update data.
V) Integration Constraints:-Integrity constraints or consistency
rules can be applied to database so that the correct data can be entered
into database. The constraints may be applied to data item within a single
record or the may be applied to relationships between records.

(6)


Applications Of DBMS :-

I) Banking:-For customer Information accounts, loan and other Banking
Transactions.

II) Universities:- For Student Information related to their fee, library,
accounts and result grades etc.

III) Airlines:- For customer information reservation system and scheduled
flight arrival and departure.

IV) Credit Card And Transactions:- Credit and debit card
information of customer monthly statements and ATM
transactions.

V) Telecommunication:-For keeping records of calls made and
generating monthly bill etc.










(7)
SAS BUSINESS ANALYTICS SOFTWARE-
SAS (Statistical Analysis System) is a software suite developed by SAS
Institute for advanced analytics, business intelligence, data management,
and predictive analytics. It is the largest market-share holder for advanced
analytics.
SAS was developed at North Carolina State University from 1966 until 1976, when
SAS Institute was incorporated. SAS was further developed in the 1980s and 1990s
with the addition of new statistical procedures, additional components and the
introduction of JMP. A point-and-click interface was added in version 9 in 2004. A
social media analytics product was added in 2010.
Basic Structure of SAS
There are two main components to most SAS programs - the data
step(s) and the procedure step(s).
The data step reads data from external sources, manipulates and
combines it with other data set and prints reports. The data step is
used to prepare your data for use by one of the procedures (often
called procs).
SAS is very lenient about the format of its input - statements can
be broken up across lines, multiple statements can appear on a
single line, and blank spaces and lines can be added to make the
program more readable.
The procedure steps perform analysis on the data, and produce
(often huge amounts of) output.
The most effective strategy for learning SAS is to concentrate on
the details of the data step, and learn the details of each procedure
as you have a need for them
Technical overview and terminology
SAS is a software suite that can mine, alter, manage and retrieve data from a variety of sources and
perform statistical analysis on it.
[1]
SAS provides a graphical point-and-click user interface for non-
technical users and more advanced options through the SAS programming language.
[1]
SAS
programs have a DATA step, which retrieves and manipulates data, usually creating a SAS data set,
and a PROC step, which analyzes the data.
The SAS software suite has more than 200
[9]
components
[10][11]
Some of the SAS components
include:
[1][10][12]

Base SAS - Basic procedures and data management
SAS/STAT - Statistical analysis
SAS/GRAPH - Graphics and presentation
SAS/OR - Operations research
SAS/ETS - Econometrics and Time Series Analysis
SAS/IML - Interactive matrix language
SAS/AF - Applications facility
SAS/QC - Quality control
SAS/INSIGHT - Data mining
SAS/PH - Clinical trial analysis
Enterprise Miner - data mining
Some Preliminary Concepts and Rules
SAS variable names must be 32 characters or less, constructed
of letters, digits and the underscore character. (Before version
7, the limit was 8.)
Its a good idea not to start variable names with an underscore,
because special system variables are named that way.
Data set names follow similar rules as variables, but they have
a different name space.
There are virtually no reserved keywords in SAS; its very good
at figuring things out by context.
SAS is not case sensitive, except inside of quoted strings.
Starting in Version 7, SAS will remember the case of variable
names when it displays them.
Missing values are handled consistently in SAS, and are
represented by a period (.).
Each statement in SAS must end in a semicolon (;).

Code examples[edit]
Sample code - "Hello World!"
%put "Hello World!";
DATA step
DATA distance;
Miles = 100;
Kilometers = 1.67 * Miles;
RUN;
DATA step using an existing table:
DATA distance_mi;
SET distance_km;
Miles = 0.62137*Kilometers;
RUN;

Software product
As of 2011 SAS' largest set of products is its line for customer intelligence. Numerous SAS modules
for web, social media and marketing analytics may be used to profile customers and prospects,
predict their behaviors and manage and optimize communications.
[35][53]
SAS also provides the SAS
Fraud Framework. The framework's primary functionality is to monitor transactions across different
applications, networks and partners and use analytics to identify anomalies that are indicative of
fraud.
[54][55][56][57]
SAS Enterprise GRC (Governance, Risk and Compliance) provides risk modeling,
scenario analysis and other functions
[58][59]
in order to manage and visualize risk, compliance and
corporate policies.
[60]
There is also a SAS Enterprise Risk Management product-set designed
primarily for banks and financial services organizations.
[61]

SQL SERVER LANGUAGE
SQL ( Structured Query Language) is a special-purpose programming language designed for
managing data held in a relational database management system (RDBMS).
History
SQL was initially developed at IBM by Donald D. Chamberlin and Raymond F. Boyce in the early
1970s.
[13]
This version, initially called SEQUEL (Structured English Query Language), was designed
to manipulate and retrieve data stored in IBM's original quasi-relational database management
system, System R, which a group at IBM San Jose Research Laboratory had developed during the
1970s.
After testing SQL at customer test sites to determine the usefulness and practicality of the system,
IBM began developing commercial products based on their System R prototype including
System/38, SQL/DS, and DB2, which were commercially available in 1979, 1981, and 1983,
respectively.





Syntax
Language elements

The SQL language is subdivided into several language elements, including:
Clauses, which are constituent components of statements and queries. (In some cases, these
are optional.)
[16]

Expressions, which can produce either scalar values, or tables consisting
of columns and rows of data
Predicates, which specify conditions that can be evaluated to SQL three-valued logic
(3VL)(true/false/unknown) or Boolean truth values and are used to limit the effects of statements
and queries, or to change program flow.
Queries, which retrieve the data based on specific criteria. This is an important element of SQL.
Statements, which may have a persistent effect on schemata and data, or may
control transactions, program flow, connections, sessions, or diagnostics.
SQL statements also include the semicolon (";") statement terminator. Though not required
on every platform, it is defined as a standard part of the SQL grammar.
Insignificant whitespace is generally ignored in SQL statements and queries, making it easier to
format SQL code for readability.
Conditional (CASE) expressions
CASE WHEN n > 0
THEN 'positive'
WHEN n < 0
THEN 'negative'
ELSE 'zero'
END
Subqueries
Queries can be nested so that the results of one query can be used in another query via a relational
operator or aggregation function. A nested query is also known as a subquery. While joins and other
table operations provide computationally superior (i.e. faster) alternatives in many cases, the use of
subqueries introduces a hierarchy in execution that can be useful or necessary.
SELECT isbn,
title,
price
FROM Book1111
WHERE price < (SELECT AVG(price) FROM Book)
ORDER BY title
Data manipulation[edit]
The Data Manipulation Language (DML) is the subset of SQL used to add, update and delete data:
INSERT adds rows (formally tuples) to an existing table, e.g.:
INSERT INTO example
(field1, field2, field3)
VALUES
('test', 'N', NULL);
UPDATE modifies a set of existing table rows, e.g.:
UPDATE example
SET field1 = 'updated value'
WHERE field2 = 'N';
DELETE removes existing rows from a table, e.g.:
DELETE FROM example
WHERE field2 = 'N';
MERGE is used to combine the data of multiple tables. It combines
the INSERT and UPDATE elements. It is defined in the SQL:2003 standard; prior to that, some
databases provided similar functionality via different syntax, sometimes called "upsert".
MERGE INTO TABLE_NAME USING table_reference ON (condition)
WHEN MATCHED THEN
UPDATE SET column1 = value1 [, column2 = value2 ...]
WHEN NOT MATCHED THEN
INSERT (column1 [, column2 ...]) VALUES (value1 [, value2 ...
Transaction controls[edit]
Transactions, if available, wrap DML operations:
START TRANSACTION (or BEGIN WORK, or BEGIN TRANSACTION, depending on SQL dialect)
marks the start of a database transaction, which either completes entirely or not at all.
SAVE TRANSACTION (or SAVEPOINT) saves the state of the database at the current point in
transaction
CREATE TABLE tbl_1(id INT);
INSERT INTO tbl_1(id) VALUES(1);
INSERT INTO tbl_1(id) VALUES(2);
COMMIT;
UPDATE tbl_1 SET id=200 WHERE id=1;
SAVEPOINT id_1upd;
UPDATE tbl_1 SET id=1000 WHERE id=2;
ROLLBACK TO id_1upd;
SELECT id FROM tbl_1;
COMMIT makes all data changes in a transaction permanent.
ROLLBACK discards all data changes since the last COMMIT or ROLLBACK, leaving the data as it
was prior to those changes. Once the COMMIT statement completes, the transaction's changes
cannot be rolled back.
COMMIT and ROLLBACK terminate the current transaction and release data locks. In the absence of
a START TRANSACTION or similar statement, the semantics of SQL are implementation-dependent.
The following example shows a classic transfer of funds transaction, where money is removed from
one account and added to another. If either the removal or the addition fails, the entire transaction is
rolled back.
START TRANSACTION;
UPDATE Account SET amount=amount-200 WHERE account_number=1234;
UPDATE Account SET amount=amount+200 WHERE account_number=2345;

IF ERRORS=0 COMMIT;
IF ERRORS<>0 ROLLBACK;
Data definition[edit]
The Data Definition Language (DDL) manages table and index structure. The most basic items of
DDL are the CREATE, ALTER, RENAME, DROP and TRUNCATE statements:
CREATE creates an object (a table, for example) in the database, e.g.:
CREATE TABLE example(
field1 INTEGER,
field2 VARCHAR(50),
field3 DATE NOT NULL,
PRIMARY KEY (field1, field2)
);
ALTER modifies the structure of an existing object in various ways, for example, adding a
column to an existing table or a constraint, e.g.:
ALTER TABLE example ADD field4 NUMBER(3) NOT NULL;
TRUNCATE deletes all data from a table in a very fast way, deleting the data inside the table and
not the table itself. It usually implies a subsequent COMMIT operation, i.e., it cannot be rolled
back (data is not written to the logs for rollback later, unlike DELETE).
TRUNCATE TABLE example;
DROP deletes an object in the database, usually irretrievably, i.e., it cannot be rolled back, e.g.:
DROP TABLE example;
Data types[edit]
Each column in an SQL table declares the type(s) that column may contain. ANSI SQL includes the
following data types.
[25]

Character strings[edit]
CHARACTER(n) or CHAR(n): fixed-width n-character string, padded with spaces as needed
CHARACTER VARYING(n) or VARCHAR(n): variable-width string with a maximum size
of n characters
NATIONAL CHARACTER(n) or NCHAR(n): fixed width string supporting an international
character set
NATIONAL CHARACTER VARYING(n) or NVARCHAR(n): variable-width NCHAR string
Bit strings[edit]
BIT(n): an array of n bits
BIT VARYING(n): an array of up to n bits
Numbers[edit]
INTEGER, SMALLINT and BIGINT
FLOAT, REAL and DOUBLE PRECISION
NUMERIC(precision, scale) or DECIMAL(precision, scale)
For example, the number 123.45 has a precision of 5 and a scale of 2. The precision is a positive
integer that determines the number of significant digits in a particular radix (binary or decimal).
The scale is a non-negative integer. A scale of 0 indicates that the number is an integer. For a
decimal number with scale S, the exact numeric value is the integer value of the significant digits
divided by 10
S
.
SQL provides a function to round numerics or dates, called TRUNC (in Informix, DB2, PostgreSQL,
Oracle and MySQL) or ROUND (in Informix, SQLite, Sybase, Oracle, PostgreSQL and Microsoft SQL
Server)
[26]

Date and time[edit]
DATE: for date values (e.g. 2011-05-03)
TIME: for time values (e.g. 15:51:36). The granularity of the time value is usually a tick (100
nanoseconds).
TIME WITH TIME ZONE or TIMETZ: the same as TIME, but including details about the time
zone in question.
TIMESTAMP: This is a DATE and a TIME put together in one variable (e.g. 2011-05-03
15:51:36).
TIMESTAMP WITH TIME ZONE or TIMESTAMPTZ: the same as TIMESTAMP, but including
details about the time zone in question.
SQL provides several functions for generating a date / time variable out of a date / time string
( TO_DATE, TO_TIME, TO_TIMESTAMP), as well as for extracting the respective members (seconds,
for instance) of such variables. The current system date / time of the database server can be called
by using functions like NOW. The IBM Informiximplementation provides the EXTEND and
the FRACTION functions to increase the accuracy of time, for systems requiring sub-second
precision.
[27]

Data control[edit]
The Data Control Language (DCL) authorizes users to access and manipulate data. Its two main
statements are:
GRANT authorizes one or more users to perform an operation or a set of operations on an
object.
REVOKE eliminates a grant, which may be the default grant.
Example:
GRANT SELECT, UPDATE
ON example
TO some_user, another_user;

REVOKE SELECT, UPDATE
ON example
FROM some_user, another_user;

Criticism[edit]
SQL deviates in several ways from its theoretical foundation, the relational model and its tuple
calculus. In that model, a table is a set of tuples, while in SQL, tables and query results are lists of
rows: the same row may occur multiple times, and the order of rows can be employed in queries
(e.g. in the LIMIT clause). Whether this is a common practical concern, it is also a subject of debate.
Furthermore, additional features (such as NULL and views) were introduced without founding them
directly on the relational model, which makes them more difficult to interpret.
Critics argue that SQL should be replaced with a language that strictly returns to the original
foundation: for example, see The Third Manifesto. Other critics suggest that Datalog has two
advantages over SQL: it has cleaner semantics, which facilitates program understanding and
maintenance, and it is more expressive, in particular for recursive queries.
[29]

Another criticism is that SQL implementations are incompatible between vendors. In particular date
and time syntax, string concatenation, NULLs, and comparison case sensitivityvary from vendor to
vendor. A particular exception is PostgreSQL, which strives for compliance.
[30]

Popular implementations of SQL commonly omit support for basic features of Standard SQL, such
as the DATE or TIME data types. The most obvious such examples, and incidentally the most
popular commercial and proprietary SQL DBMSs, are Oracle (whose DATE behaves
as DATETIME,
[31][32]
and lacks a TIME type)
[33]
and MS SQL Server (before the 2008 version). As a
result, SQL code can rarely be ported between database systems without modifications.

Microsoft Excel

Introduction-
Microsoft Excel is a spreadsheet application developed by Microsoft for Microsoft
Windows and Mac OS. It features calculation, graphing tools, pivot tables, and
a macro programming language called Visual Basic for Applications. It has been a very widely
applied spreadsheet for these platforms, especially since version 5 in 1993, and it has
replaced Lotus 1-2-3 as the industry standard for spreadsheets. Excel forms part of Microsoft Office.
Description-
When you look at the Excel screen (refer to the example on this page) you see a rectangular table or
grid of rows and columns. The horizontal rows are identified by numbers (1,2,3) and the vertical
columns with letters of the alphabet (A,B,C). For columns beyond 26, columns are identified by two or
more letters such as AA, AB, AC.
The intersection point between a column and a row is a small rectangular box known as a cell. A cell is
the basic unit for storing data in the spreadsheet. Because an Excel spreadsheet contains thousands of
these cells, each is given a cell reference or address to identify it. The cell reference is a combination of
the column letter and the row number such as A3, B6, AA345.



Advantages-
1. Used to store Excel and Financial Data
Spreadsheets are often used to store financial data. Formulas and functions that are used on this type
of data include:
Performing basic mathematical operations such as summing columns and rows of figures.
Finding values such as profit or loss.
Calculating repayment plans for loans or mortgages.
Finding the average, maximum, or minimum values in a specified range of data.

2. Easy and effective comparisons - With the powerful analytical tools included within Microsoft
Excel you have the ability to analyze large amounts of data to discover trends and patterns that will
influence decisions. Microsoft Excels graphing capabilities allows you to summarize your data
enhancing your ability to organize and structure your data.
3. Powerful analysis of large amounts of data - Recent upgrades to the Excel spreadsheet
enhance your ability to analyze large amounts of data. With powerful filtering, sorting and search tools
you are able to quickly and easily narrow down the criteria that will assist in your decisions. Combine
these tools with the tables, Pivot Tables and Graphs you can find the information that you want
quickly and easily even if you have hundreds of thousands of data items.
4. Microsoft Excel Mobile & iPad Apps - With the advent of the tablet and the smart phone it is now
possible to take your worksheets to a client or a meeting without having to bring along your Laptop.
The power of these mobile devices now allows you to manipulate data and update your spreadsheets
and then view the spreadsheets immediately on your phone or tablet.

Disadvantages of Microsoft Excel
Not easy to share Microsoft have made great strides with their Web applications but the
sharing functionality still has a long way to go to become world class. Part of the problem is that
often the ability for sharing a tool requires extra software to be installed on the PC and this
often becomes cumbersome. Tools like Google Documents are revolutionizing the way people
share documents online.
Non relational I use MS Access for databases and love the relational aspect of the
database. The idea that you can link data by common elements (such as employee ID, Stock
Number etc) greatly assists in the reduction of data, and the ability to analyze data. Aside from
some database functions and some lookup tools Microsoft Excel lacks any true relational
functionality.
Scalability Excel 2010 has introduced larger spreadsheets and binary formats, howver using
large amounts of data on the average PC will slow down and often freeze the PC. As some of
the competing tools are in the 'cloud' file size does not become a problem.
No Forms while you can use forms within VBA, the average user will not have much
knowledge of VBA and therefore not be able to incorporate functional forms into their
spreadsheets quickly.
Big data: Treasure hides in everyday details

Definitions

Big data is a buzzword, or catch-phrase, used to describe a massive volume of
both structured and unstructured data that is so large that it's difficult to process using
traditional database and softwaretechniques. In most enterprise scenarios the data is
too big or it moves too fast or it exceeds current processing capacity.

An Example of Big Data
Big data might be petabytes (1,024 terabytes) or exabytes (1,024 petabytes) of data
consisting of billions to trillions of records of millions of peopleall from different
sources (e.g. Web, sales, customer contact center, social media, mobile data and so
on). The data is typically loosely structured data that is often incomplete and
inaccessible.
Typically, the data come from social media such as Facebook or Twitter, or vast
numbers of connected devices, including satellites, closed-circuit television cameras,
audio and video recording equipment, payment terminals, mobile phones and sensors.

Impact-
Big data has the potential to help companies improve operations and make faster, more
intelligent decisions.
Big Data and Types of Business Datasets
When dealing with larger datasets, organizations face difficulties in being able to create, manipulate,
and manage big data. Big data is particularly a problem in business analytics because standard tools
and procedures are not designed to search and analyze massive datasets.



Chapter 3
Motivation of the Training

Students with internship experience are significantly more likely to get hired after college than
peers lacking internship experience. NACEs research also shows that students with internship
experience tend to have much higher starting salaries than students without such experience.
Here are 11 specific reasons why internship experience in college is so essential for career
success after graduation.
1. Internships demonstrate initiative. Every employer (or graduate school) wants go-
getters. Given that internships are usually not required in college, such experience will
prove that you are proactive about career development.
2. Internships help you learn more about yourself. Through an internship, you can gain
clarity on your strengths, weaknesses, and interests. Equally important, you can also
learn what you dont enjoy doing.
3. Internships enhance your resume. Worried about having a solid resume? Internship
experience can fix that concern.
4. Internships help you sell yourself in interviews. Worried about what you will discuss
in an interview for a full-time job? Internship experience helps you overcome that
obstacle also.
5. Internships increase your professional confidence. By interning, you can confront and
overcome the fear of the unknown that you may have about what its like to work in a
professional environment
6. Internships improve your communication skills. The #1 skill employers look for in
job candidates is strong communication skills. By interacting with others in a
professional environment during an internship, you will have many opportunities to
become a better communicator.
7. Internships help you prove yourself in the real-world. Academic success matters,
but achievement in the real-world is even more important when you are looking for
full-time work.
8. Internships expand your professional network. Over 70% of all jobs are filled
through personal contacts and networking. Through an internship, you will meet lots of
new people, each of whom could help you get a full-time job after college.
9. Internships help you collect professional references. By doing great work in your
internship, you can impress bosses and generate a group of people willing to endorse you
to potential hiring persons in the future.
10. Internships help you go behind-the-scenes. Through an internship, you get a chance to
learn what it is really like to work in a company, in an industry, and in various job
functions.
11. Internships help you develop better work habits. Through an internship, you will
learn how to manage tasks/projects and learn how to carry yourself in a professional
environment. You can also learn from your colleagues by observing their positive and
negative work habits.
In summary, internship experience makes you a better job candidate, helps
you decide what type of job/career you want to pursue, and increases your
chances of getting hired at a great starting salary. Get as much internship
experience as you can!

























Chapter 4
Learning Outcome of the training
1. Have learnt
how to put theory into practice. What skills are needed to fulfil a position in a (bigger or smaller size)
company? Have you obtained these skills already to the full extent?
2. Have experience the culture
of a company. Do you love to wear a suit and tie while working or would you rather wear sneakers
instead? Do you prefer large organisations to smaller companies? In what culture can you perform at your
best?


Overview
Big data, Hadoop, in-memory analytics, self-service BI, data warehouse automation, analytical database servers,
data virtualization, data vault, operational intelligence, predictive analytics, and NoSQL are just a few of the new
technologies and techniques that have become available for developing BI systems. Most of them are very powerful
and allow for development of more flexible and scalable BI systems. But which ones do you pick?
Due to this waterfall of new developments, its becoming harder and harder for organizations to select the right tools.
Which technologies are relevant? Are they mature? What are their use cases? These are all valid but difficult to
answer questions.
This seminar gives a clear and extensive overview of all the new developments and their inter-relationships.
Technologies and techniques are explained, market overviews are presented, strengths and weaknesses are
discussed, and guidelines and best practices are given.
The biggest revolution in BI is evidently big data. Therefore, considerable time in the seminar is reserved for this
intriguing topic. Hadoop, MapReduce, Hive, NoSQL, SQL-on-Hadoop are all explained. In addition, the relation with
analytics is discussed extensively.
This seminar gives you a unique opportunity to see and learn about all the new BI developments.


Learning Objectives
Learn about the trends and the technological developments related to business intelligence, analytics, data
warehousing, and big data.
Discover the value of big data and analytics for organizations
Learn which products and technologies are winners and which ones are losers.
Learn how new and existing technologies, such as Hadoop, NoSQL and NewSQL, will help you create new
opportunities in your organization.
Learn how more agile data business intelligence systems can be designed.
Learn how to embed big data and analytics in existing business intelligence architectures.
Future opportunity of Data Analytics and Business Intelligence
It is undergoing massive, disruptive changes that will radically transform the way the industry and customers think
about analytics. The exponential growth in data is a key driver of this change. In addition, the mainstream adoption of
cloud computing across the enterprise continues to put pressure on the capabilities of businesses to incorporate all
relevant data from multiple data sources to enable users to make more timely, comprehensive, insightful business
decisions.
The amount of data that is generated in the business world is doubling every year. Information from devices,
machines and social media creates an entirely new set of challenges, and reinforces the fact that data will continue to
grow exponentially for the foreseeable future.

Vous aimerez peut-être aussi