School of Electronics and Electrical Engineering Lovely Professional University, Phagwara, Punjab (June-July, 2014)
DECLARATION
I hereby declare that I have completed my six weeks summer training at Lovely professional University (name and location of organization) from 9/06/2014 (start date) to 18/07/2014(end date) under the guidance of Mr. Nikesh Bajaj (Name of Industry coordinator). I have declare that I have worked with full dedication during these six weeks of training and my learning outcomes fulfill the requirements of training for the award of degree of B.tech 3 rd year ECE (Relevant Degree ), Lovely Professional University, Phagwara, Punjab.
(Signature of student)
Name of student:-Abhishek Kumar Registration no:-11104699 Date :-03/08/2014 (I) ACKNOWLEDGEMENT
I would deeply like to express my gratitude to a few people whose help and guidelines encouraged me throughout the project. Firstly, I express my deep thanks to my training coordinator, Mr. Nikesh Bajaj, for the way he taught us. The effort, he has made while teaching us was helpful in project completion also. Secondly, my DBMS teacher, Mr. Ravindra Singh Who along with giving this interesting topic, showed faith in me and I have completed this by burning my night oil in the best possible way which I can. They were after me for the whole course of this project. Nextly , for this project, I would like to thank my friends for helping and showing me the right path for the completion of this project. Last but not least, The Almighty, for everything he has blessed me.
-Abhishek kumar 11104669
(II)
Certificate
(III) PERIODIC APPRAISAL PERFORMA (To be filled by External Advisor)
Name of Student:__________________ Registration No:_____________ Project/TechnologyLearnt_____________________________________ Name & Address of organization ______________________________ Name of External Project Supervisor (with Phone No)_______________
Signature of External Project Supervisor/Guide (With Seal, Date & Designation)
(V) INDEX
Chapter Contents Page No 1. Organization Overview 1 2. Technology Learnt 7 3. Motivation In The training 31 4. Learning Outcome 5. Gantt chart References
(VI)
Chapter One :-Organization Overview
Type Of Organization:- Cognizant Technology Solutions Corporation is a multinational information technology, consulting and business process outsourcing company. It is headquartered in Teaneck, New Jersey, United States, North America. This is the first Indian software services firm to be listed on the NASDAQ.
Cognizant provides information technology, consulting and BPO services. These Include business & technology consulting, systems integration, application development & maintenance, IT infrastructure services, analytics, business intelligence, data warehousing, CRM, supply chain management, engineering & manufacturing Solutions, ERP, R&Doutsourcing , and testing solutions.
Cognizant is organized into several verticals and horizontal units. The vertical units focus on specific industries such as Banking & Financial Services, Healthcare, Manufacturing and Retail. The horizontals focus on specific technologies or process areas such as Analytics, mobile computing, BPO and Testing.
Both horizontal and vertical units have business consultants, who together form the organization-wide Cognizant Business Consulting (CBC) team. Cognizant is among the largest recruiters of MBAs in the industry; they are involved in business development and business analysis for IT services projects.
In 2011, the company's revenue from IT services was split roughly evenly between application development and application maintenance. Its business process outsourcing portfolio leans towards "higher-end" services i.e., work that involves domain knowledge and skills, such as legal services or healthcare claims.
(1)
Like many other IT services firms, Cognizant follows a global delivery model based on offshore software R&D and offshore outsourcing. The company has a number of offshore development centers outside the United States and near- shore centers in the U.S., Europe and South America.
In its early years, Cognizant gained business from a number of American and European companies with the help of the D&B brand. The company's senior executives envisaged the firm as a provider of high-end customer services on-par with the six contemporary major system integrators (Accenture, BearingPoint, Capgemini, E&Y, Deloitte and IBM), but at lower prices.
In the 2012 earnings announcements, the CEO Francisco D'Souza categorized the company's service offerings in three groups: Horizon 1 (application development and maintenance), Horizon 2 (BPO, IT Infrastructure Services & business consulting) and Horizon 3 ("SMAC" - Social, Mobile, Analytics and Cloud).As of September 2012, the Horizon 1 services accounted for over 75% of the company's revenues, and Horizon 2 services about 20%.
Training Location:- Lovely Professional University Phagwara, Punjab. Location is good for student of LPU because any student can easily reach the location. It was mine privilege to get my training done with one of the leading business tycoon at my own college campus. Basically, Cognizant is a grown IT company and it has several training venue but this is the first time when they have just came inside our campus o train us .
(2)
Specialization Of/Work culture in Organization:- I have done my six weeks summer training in the field of Data analytics and Business Intelligence. Working environment under the training period is good and it is also interesting because the facility provided by the organization is so nice, they provide us a good kind of computer lab in which we can do our practices. All the faculty members and staffs are helpful they interact with all the students and resolve the problems. In the training period there was a nice feedback system all the feedbacks are taken care. They provide us study materials with a large no. Of mini projects which are completed under the training period, and study materials are helpful when we are doing our projects.
External Guide :-
(3)
Chapter Two :-Technology Learnt
Data base Management System:- A database management system (DBMS) is a collection of programs that enables you to store, modify, and extract information from a database. There are many different types of DBMSs, ranging from small systems that run on personal computers to huge systems that run on mainframes. The following are examples of database applications. Computerized library systems Automated teller machines Flight reservation systems Computerized parts inventory systems
Database:- A database is an organized collection of data. The data are typically organized to model aspects of reality in a way that supports processes requiring this information. For example , modeling the availability of rooms in hotels in a way that supports finding a hotel with vacancies.
DBMS Features:- There are a variety of DBMS database management tools that provide different features and functionality. Here are the most probable and must-have features of a typical DBMS: (I) Data Structuring:- All information in a digital repository is organized into a clear hierarchical structure with records, tables or objects. Every piece of information that you can add to your database will be organized in a kind of catalogue, so it will be easier for you to search and update your records later on.
(4)
II) Database Customization :- Along with default and necessary elements (records, tables, or objects) that compose a database's structure, there can be created custom elements that meet specific needs of users.
III) Data Retrieval :- DBMS database management system accepts data input from users and stores it. Users can access the database later on to retrieve their records as a file, printed, or viewed on the screen.
IV) Query Languages:-A typical DBMS makes it possible to use query languages for collecting, searching, sorting, modifying and other activities that help users manipulate their records in the database.
V) Multi-user Access :-DBMS provides multiple users access to all types of information stored in one and the same data store. There is also a security feature that prevents some users from viewing and/or modifying certain data types.
VI) Data Integrity:-All information in a database is accessible by several or more users but only one user can change the same piece of data at a time. This capability lets prevent database corruptions and failure.
VII) Metadata :-DBMS database management software provides a metadata library (or a data dictionary) that describes how the database is organized and what elements (objects, related files, records, etc.) compose its structure.
There are some commonly found features in DBMSs :-
(5)
Advantages Of DBMSs:- The database management system has a number of advantages as compared to traditional computer file-based processing approach. The DBA must keep in mind these benefits or capabilities during databases and monitoring the DBMS.
The Main advantages of DBMS are described below. I). Controlling Data Redundancy :- In non-database systems each application program has its own private files. In this case, the duplicated copies of the same data are created in many places. In DBMS, all data of an organization is integrated into a single database file. The data is recorded in only one place in the database and it is not duplicated. II) Sharing of Data :- In DBMS, data can be shared by authorized users of the organization. The database administrator manages the data and gives rights to users to access the data. Many users can be authorized to access the same piece of information simultaneously. The remote users can also share same data. Similarly, the data of same database can be shared between different application programs. III) Data Consistency :- By controlling the data redundancy, the data consistency is obtained. If a data item appears only once, any update to its value has to be performed only once and the updated value is immediately available to all users. If the DBMS has controlled redundancy, the database system enforces consistency. IV) Integration of Data :-In Database management system, data in database is stored in tables. A single database contains multiple tables and relationships can be created between tables (or associated data entities). This makes easy to retrieve and update data. V) Integration Constraints:-Integrity constraints or consistency rules can be applied to database so that the correct data can be entered into database. The constraints may be applied to data item within a single record or the may be applied to relationships between records.
(6)
Applications Of DBMS :-
I) Banking:-For customer Information accounts, loan and other Banking Transactions.
II) Universities:- For Student Information related to their fee, library, accounts and result grades etc.
III) Airlines:- For customer information reservation system and scheduled flight arrival and departure.
IV) Credit Card And Transactions:- Credit and debit card information of customer monthly statements and ATM transactions.
V) Telecommunication:-For keeping records of calls made and generating monthly bill etc.
(7) SAS BUSINESS ANALYTICS SOFTWARE- SAS (Statistical Analysis System) is a software suite developed by SAS Institute for advanced analytics, business intelligence, data management, and predictive analytics. It is the largest market-share holder for advanced analytics. SAS was developed at North Carolina State University from 1966 until 1976, when SAS Institute was incorporated. SAS was further developed in the 1980s and 1990s with the addition of new statistical procedures, additional components and the introduction of JMP. A point-and-click interface was added in version 9 in 2004. A social media analytics product was added in 2010. Basic Structure of SAS There are two main components to most SAS programs - the data step(s) and the procedure step(s). The data step reads data from external sources, manipulates and combines it with other data set and prints reports. The data step is used to prepare your data for use by one of the procedures (often called procs). SAS is very lenient about the format of its input - statements can be broken up across lines, multiple statements can appear on a single line, and blank spaces and lines can be added to make the program more readable. The procedure steps perform analysis on the data, and produce (often huge amounts of) output. The most effective strategy for learning SAS is to concentrate on the details of the data step, and learn the details of each procedure as you have a need for them Technical overview and terminology SAS is a software suite that can mine, alter, manage and retrieve data from a variety of sources and perform statistical analysis on it. [1] SAS provides a graphical point-and-click user interface for non- technical users and more advanced options through the SAS programming language. [1] SAS programs have a DATA step, which retrieves and manipulates data, usually creating a SAS data set, and a PROC step, which analyzes the data. The SAS software suite has more than 200 [9] components [10][11] Some of the SAS components include: [1][10][12]
Base SAS - Basic procedures and data management SAS/STAT - Statistical analysis SAS/GRAPH - Graphics and presentation SAS/OR - Operations research SAS/ETS - Econometrics and Time Series Analysis SAS/IML - Interactive matrix language SAS/AF - Applications facility SAS/QC - Quality control SAS/INSIGHT - Data mining SAS/PH - Clinical trial analysis Enterprise Miner - data mining Some Preliminary Concepts and Rules SAS variable names must be 32 characters or less, constructed of letters, digits and the underscore character. (Before version 7, the limit was 8.) Its a good idea not to start variable names with an underscore, because special system variables are named that way. Data set names follow similar rules as variables, but they have a different name space. There are virtually no reserved keywords in SAS; its very good at figuring things out by context. SAS is not case sensitive, except inside of quoted strings. Starting in Version 7, SAS will remember the case of variable names when it displays them. Missing values are handled consistently in SAS, and are represented by a period (.). Each statement in SAS must end in a semicolon (;).
Code examples[edit] Sample code - "Hello World!" %put "Hello World!"; DATA step DATA distance; Miles = 100; Kilometers = 1.67 * Miles; RUN; DATA step using an existing table: DATA distance_mi; SET distance_km; Miles = 0.62137*Kilometers; RUN;
Software product As of 2011 SAS' largest set of products is its line for customer intelligence. Numerous SAS modules for web, social media and marketing analytics may be used to profile customers and prospects, predict their behaviors and manage and optimize communications. [35][53] SAS also provides the SAS Fraud Framework. The framework's primary functionality is to monitor transactions across different applications, networks and partners and use analytics to identify anomalies that are indicative of fraud. [54][55][56][57] SAS Enterprise GRC (Governance, Risk and Compliance) provides risk modeling, scenario analysis and other functions [58][59] in order to manage and visualize risk, compliance and corporate policies. [60] There is also a SAS Enterprise Risk Management product-set designed primarily for banks and financial services organizations. [61]
SQL SERVER LANGUAGE SQL ( Structured Query Language) is a special-purpose programming language designed for managing data held in a relational database management system (RDBMS). History SQL was initially developed at IBM by Donald D. Chamberlin and Raymond F. Boyce in the early 1970s. [13] This version, initially called SEQUEL (Structured English Query Language), was designed to manipulate and retrieve data stored in IBM's original quasi-relational database management system, System R, which a group at IBM San Jose Research Laboratory had developed during the 1970s. After testing SQL at customer test sites to determine the usefulness and practicality of the system, IBM began developing commercial products based on their System R prototype including System/38, SQL/DS, and DB2, which were commercially available in 1979, 1981, and 1983, respectively.
Syntax Language elements
The SQL language is subdivided into several language elements, including: Clauses, which are constituent components of statements and queries. (In some cases, these are optional.) [16]
Expressions, which can produce either scalar values, or tables consisting of columns and rows of data Predicates, which specify conditions that can be evaluated to SQL three-valued logic (3VL)(true/false/unknown) or Boolean truth values and are used to limit the effects of statements and queries, or to change program flow. Queries, which retrieve the data based on specific criteria. This is an important element of SQL. Statements, which may have a persistent effect on schemata and data, or may control transactions, program flow, connections, sessions, or diagnostics. SQL statements also include the semicolon (";") statement terminator. Though not required on every platform, it is defined as a standard part of the SQL grammar. Insignificant whitespace is generally ignored in SQL statements and queries, making it easier to format SQL code for readability. Conditional (CASE) expressions CASE WHEN n > 0 THEN 'positive' WHEN n < 0 THEN 'negative' ELSE 'zero' END Subqueries Queries can be nested so that the results of one query can be used in another query via a relational operator or aggregation function. A nested query is also known as a subquery. While joins and other table operations provide computationally superior (i.e. faster) alternatives in many cases, the use of subqueries introduces a hierarchy in execution that can be useful or necessary. SELECT isbn, title, price FROM Book1111 WHERE price < (SELECT AVG(price) FROM Book) ORDER BY title Data manipulation[edit] The Data Manipulation Language (DML) is the subset of SQL used to add, update and delete data: INSERT adds rows (formally tuples) to an existing table, e.g.: INSERT INTO example (field1, field2, field3) VALUES ('test', 'N', NULL); UPDATE modifies a set of existing table rows, e.g.: UPDATE example SET field1 = 'updated value' WHERE field2 = 'N'; DELETE removes existing rows from a table, e.g.: DELETE FROM example WHERE field2 = 'N'; MERGE is used to combine the data of multiple tables. It combines the INSERT and UPDATE elements. It is defined in the SQL:2003 standard; prior to that, some databases provided similar functionality via different syntax, sometimes called "upsert". MERGE INTO TABLE_NAME USING table_reference ON (condition) WHEN MATCHED THEN UPDATE SET column1 = value1 [, column2 = value2 ...] WHEN NOT MATCHED THEN INSERT (column1 [, column2 ...]) VALUES (value1 [, value2 ... Transaction controls[edit] Transactions, if available, wrap DML operations: START TRANSACTION (or BEGIN WORK, or BEGIN TRANSACTION, depending on SQL dialect) marks the start of a database transaction, which either completes entirely or not at all. SAVE TRANSACTION (or SAVEPOINT) saves the state of the database at the current point in transaction CREATE TABLE tbl_1(id INT); INSERT INTO tbl_1(id) VALUES(1); INSERT INTO tbl_1(id) VALUES(2); COMMIT; UPDATE tbl_1 SET id=200 WHERE id=1; SAVEPOINT id_1upd; UPDATE tbl_1 SET id=1000 WHERE id=2; ROLLBACK TO id_1upd; SELECT id FROM tbl_1; COMMIT makes all data changes in a transaction permanent. ROLLBACK discards all data changes since the last COMMIT or ROLLBACK, leaving the data as it was prior to those changes. Once the COMMIT statement completes, the transaction's changes cannot be rolled back. COMMIT and ROLLBACK terminate the current transaction and release data locks. In the absence of a START TRANSACTION or similar statement, the semantics of SQL are implementation-dependent. The following example shows a classic transfer of funds transaction, where money is removed from one account and added to another. If either the removal or the addition fails, the entire transaction is rolled back. START TRANSACTION; UPDATE Account SET amount=amount-200 WHERE account_number=1234; UPDATE Account SET amount=amount+200 WHERE account_number=2345;
IF ERRORS=0 COMMIT; IF ERRORS<>0 ROLLBACK; Data definition[edit] The Data Definition Language (DDL) manages table and index structure. The most basic items of DDL are the CREATE, ALTER, RENAME, DROP and TRUNCATE statements: CREATE creates an object (a table, for example) in the database, e.g.: CREATE TABLE example( field1 INTEGER, field2 VARCHAR(50), field3 DATE NOT NULL, PRIMARY KEY (field1, field2) ); ALTER modifies the structure of an existing object in various ways, for example, adding a column to an existing table or a constraint, e.g.: ALTER TABLE example ADD field4 NUMBER(3) NOT NULL; TRUNCATE deletes all data from a table in a very fast way, deleting the data inside the table and not the table itself. It usually implies a subsequent COMMIT operation, i.e., it cannot be rolled back (data is not written to the logs for rollback later, unlike DELETE). TRUNCATE TABLE example; DROP deletes an object in the database, usually irretrievably, i.e., it cannot be rolled back, e.g.: DROP TABLE example; Data types[edit] Each column in an SQL table declares the type(s) that column may contain. ANSI SQL includes the following data types. [25]
Character strings[edit] CHARACTER(n) or CHAR(n): fixed-width n-character string, padded with spaces as needed CHARACTER VARYING(n) or VARCHAR(n): variable-width string with a maximum size of n characters NATIONAL CHARACTER(n) or NCHAR(n): fixed width string supporting an international character set NATIONAL CHARACTER VARYING(n) or NVARCHAR(n): variable-width NCHAR string Bit strings[edit] BIT(n): an array of n bits BIT VARYING(n): an array of up to n bits Numbers[edit] INTEGER, SMALLINT and BIGINT FLOAT, REAL and DOUBLE PRECISION NUMERIC(precision, scale) or DECIMAL(precision, scale) For example, the number 123.45 has a precision of 5 and a scale of 2. The precision is a positive integer that determines the number of significant digits in a particular radix (binary or decimal). The scale is a non-negative integer. A scale of 0 indicates that the number is an integer. For a decimal number with scale S, the exact numeric value is the integer value of the significant digits divided by 10 S . SQL provides a function to round numerics or dates, called TRUNC (in Informix, DB2, PostgreSQL, Oracle and MySQL) or ROUND (in Informix, SQLite, Sybase, Oracle, PostgreSQL and Microsoft SQL Server) [26]
Date and time[edit] DATE: for date values (e.g. 2011-05-03) TIME: for time values (e.g. 15:51:36). The granularity of the time value is usually a tick (100 nanoseconds). TIME WITH TIME ZONE or TIMETZ: the same as TIME, but including details about the time zone in question. TIMESTAMP: This is a DATE and a TIME put together in one variable (e.g. 2011-05-03 15:51:36). TIMESTAMP WITH TIME ZONE or TIMESTAMPTZ: the same as TIMESTAMP, but including details about the time zone in question. SQL provides several functions for generating a date / time variable out of a date / time string ( TO_DATE, TO_TIME, TO_TIMESTAMP), as well as for extracting the respective members (seconds, for instance) of such variables. The current system date / time of the database server can be called by using functions like NOW. The IBM Informiximplementation provides the EXTEND and the FRACTION functions to increase the accuracy of time, for systems requiring sub-second precision. [27]
Data control[edit] The Data Control Language (DCL) authorizes users to access and manipulate data. Its two main statements are: GRANT authorizes one or more users to perform an operation or a set of operations on an object. REVOKE eliminates a grant, which may be the default grant. Example: GRANT SELECT, UPDATE ON example TO some_user, another_user;
REVOKE SELECT, UPDATE ON example FROM some_user, another_user;
Criticism[edit] SQL deviates in several ways from its theoretical foundation, the relational model and its tuple calculus. In that model, a table is a set of tuples, while in SQL, tables and query results are lists of rows: the same row may occur multiple times, and the order of rows can be employed in queries (e.g. in the LIMIT clause). Whether this is a common practical concern, it is also a subject of debate. Furthermore, additional features (such as NULL and views) were introduced without founding them directly on the relational model, which makes them more difficult to interpret. Critics argue that SQL should be replaced with a language that strictly returns to the original foundation: for example, see The Third Manifesto. Other critics suggest that Datalog has two advantages over SQL: it has cleaner semantics, which facilitates program understanding and maintenance, and it is more expressive, in particular for recursive queries. [29]
Another criticism is that SQL implementations are incompatible between vendors. In particular date and time syntax, string concatenation, NULLs, and comparison case sensitivityvary from vendor to vendor. A particular exception is PostgreSQL, which strives for compliance. [30]
Popular implementations of SQL commonly omit support for basic features of Standard SQL, such as the DATE or TIME data types. The most obvious such examples, and incidentally the most popular commercial and proprietary SQL DBMSs, are Oracle (whose DATE behaves as DATETIME, [31][32] and lacks a TIME type) [33] and MS SQL Server (before the 2008 version). As a result, SQL code can rarely be ported between database systems without modifications.
Microsoft Excel
Introduction- Microsoft Excel is a spreadsheet application developed by Microsoft for Microsoft Windows and Mac OS. It features calculation, graphing tools, pivot tables, and a macro programming language called Visual Basic for Applications. It has been a very widely applied spreadsheet for these platforms, especially since version 5 in 1993, and it has replaced Lotus 1-2-3 as the industry standard for spreadsheets. Excel forms part of Microsoft Office. Description- When you look at the Excel screen (refer to the example on this page) you see a rectangular table or grid of rows and columns. The horizontal rows are identified by numbers (1,2,3) and the vertical columns with letters of the alphabet (A,B,C). For columns beyond 26, columns are identified by two or more letters such as AA, AB, AC. The intersection point between a column and a row is a small rectangular box known as a cell. A cell is the basic unit for storing data in the spreadsheet. Because an Excel spreadsheet contains thousands of these cells, each is given a cell reference or address to identify it. The cell reference is a combination of the column letter and the row number such as A3, B6, AA345.
Advantages- 1. Used to store Excel and Financial Data Spreadsheets are often used to store financial data. Formulas and functions that are used on this type of data include: Performing basic mathematical operations such as summing columns and rows of figures. Finding values such as profit or loss. Calculating repayment plans for loans or mortgages. Finding the average, maximum, or minimum values in a specified range of data.
2. Easy and effective comparisons - With the powerful analytical tools included within Microsoft Excel you have the ability to analyze large amounts of data to discover trends and patterns that will influence decisions. Microsoft Excels graphing capabilities allows you to summarize your data enhancing your ability to organize and structure your data. 3. Powerful analysis of large amounts of data - Recent upgrades to the Excel spreadsheet enhance your ability to analyze large amounts of data. With powerful filtering, sorting and search tools you are able to quickly and easily narrow down the criteria that will assist in your decisions. Combine these tools with the tables, Pivot Tables and Graphs you can find the information that you want quickly and easily even if you have hundreds of thousands of data items. 4. Microsoft Excel Mobile & iPad Apps - With the advent of the tablet and the smart phone it is now possible to take your worksheets to a client or a meeting without having to bring along your Laptop. The power of these mobile devices now allows you to manipulate data and update your spreadsheets and then view the spreadsheets immediately on your phone or tablet.
Disadvantages of Microsoft Excel Not easy to share Microsoft have made great strides with their Web applications but the sharing functionality still has a long way to go to become world class. Part of the problem is that often the ability for sharing a tool requires extra software to be installed on the PC and this often becomes cumbersome. Tools like Google Documents are revolutionizing the way people share documents online. Non relational I use MS Access for databases and love the relational aspect of the database. The idea that you can link data by common elements (such as employee ID, Stock Number etc) greatly assists in the reduction of data, and the ability to analyze data. Aside from some database functions and some lookup tools Microsoft Excel lacks any true relational functionality. Scalability Excel 2010 has introduced larger spreadsheets and binary formats, howver using large amounts of data on the average PC will slow down and often freeze the PC. As some of the competing tools are in the 'cloud' file size does not become a problem. No Forms while you can use forms within VBA, the average user will not have much knowledge of VBA and therefore not be able to incorporate functional forms into their spreadsheets quickly. Big data: Treasure hides in everyday details
Definitions
Big data is a buzzword, or catch-phrase, used to describe a massive volume of both structured and unstructured data that is so large that it's difficult to process using traditional database and softwaretechniques. In most enterprise scenarios the data is too big or it moves too fast or it exceeds current processing capacity.
An Example of Big Data Big data might be petabytes (1,024 terabytes) or exabytes (1,024 petabytes) of data consisting of billions to trillions of records of millions of peopleall from different sources (e.g. Web, sales, customer contact center, social media, mobile data and so on). The data is typically loosely structured data that is often incomplete and inaccessible. Typically, the data come from social media such as Facebook or Twitter, or vast numbers of connected devices, including satellites, closed-circuit television cameras, audio and video recording equipment, payment terminals, mobile phones and sensors.
Impact- Big data has the potential to help companies improve operations and make faster, more intelligent decisions. Big Data and Types of Business Datasets When dealing with larger datasets, organizations face difficulties in being able to create, manipulate, and manage big data. Big data is particularly a problem in business analytics because standard tools and procedures are not designed to search and analyze massive datasets.
Chapter 3 Motivation of the Training
Students with internship experience are significantly more likely to get hired after college than peers lacking internship experience. NACEs research also shows that students with internship experience tend to have much higher starting salaries than students without such experience. Here are 11 specific reasons why internship experience in college is so essential for career success after graduation. 1. Internships demonstrate initiative. Every employer (or graduate school) wants go- getters. Given that internships are usually not required in college, such experience will prove that you are proactive about career development. 2. Internships help you learn more about yourself. Through an internship, you can gain clarity on your strengths, weaknesses, and interests. Equally important, you can also learn what you dont enjoy doing. 3. Internships enhance your resume. Worried about having a solid resume? Internship experience can fix that concern. 4. Internships help you sell yourself in interviews. Worried about what you will discuss in an interview for a full-time job? Internship experience helps you overcome that obstacle also. 5. Internships increase your professional confidence. By interning, you can confront and overcome the fear of the unknown that you may have about what its like to work in a professional environment 6. Internships improve your communication skills. The #1 skill employers look for in job candidates is strong communication skills. By interacting with others in a professional environment during an internship, you will have many opportunities to become a better communicator. 7. Internships help you prove yourself in the real-world. Academic success matters, but achievement in the real-world is even more important when you are looking for full-time work. 8. Internships expand your professional network. Over 70% of all jobs are filled through personal contacts and networking. Through an internship, you will meet lots of new people, each of whom could help you get a full-time job after college. 9. Internships help you collect professional references. By doing great work in your internship, you can impress bosses and generate a group of people willing to endorse you to potential hiring persons in the future. 10. Internships help you go behind-the-scenes. Through an internship, you get a chance to learn what it is really like to work in a company, in an industry, and in various job functions. 11. Internships help you develop better work habits. Through an internship, you will learn how to manage tasks/projects and learn how to carry yourself in a professional environment. You can also learn from your colleagues by observing their positive and negative work habits. In summary, internship experience makes you a better job candidate, helps you decide what type of job/career you want to pursue, and increases your chances of getting hired at a great starting salary. Get as much internship experience as you can!
Chapter 4 Learning Outcome of the training 1. Have learnt how to put theory into practice. What skills are needed to fulfil a position in a (bigger or smaller size) company? Have you obtained these skills already to the full extent? 2. Have experience the culture of a company. Do you love to wear a suit and tie while working or would you rather wear sneakers instead? Do you prefer large organisations to smaller companies? In what culture can you perform at your best?
Overview Big data, Hadoop, in-memory analytics, self-service BI, data warehouse automation, analytical database servers, data virtualization, data vault, operational intelligence, predictive analytics, and NoSQL are just a few of the new technologies and techniques that have become available for developing BI systems. Most of them are very powerful and allow for development of more flexible and scalable BI systems. But which ones do you pick? Due to this waterfall of new developments, its becoming harder and harder for organizations to select the right tools. Which technologies are relevant? Are they mature? What are their use cases? These are all valid but difficult to answer questions. This seminar gives a clear and extensive overview of all the new developments and their inter-relationships. Technologies and techniques are explained, market overviews are presented, strengths and weaknesses are discussed, and guidelines and best practices are given. The biggest revolution in BI is evidently big data. Therefore, considerable time in the seminar is reserved for this intriguing topic. Hadoop, MapReduce, Hive, NoSQL, SQL-on-Hadoop are all explained. In addition, the relation with analytics is discussed extensively. This seminar gives you a unique opportunity to see and learn about all the new BI developments.
Learning Objectives Learn about the trends and the technological developments related to business intelligence, analytics, data warehousing, and big data. Discover the value of big data and analytics for organizations Learn which products and technologies are winners and which ones are losers. Learn how new and existing technologies, such as Hadoop, NoSQL and NewSQL, will help you create new opportunities in your organization. Learn how more agile data business intelligence systems can be designed. Learn how to embed big data and analytics in existing business intelligence architectures. Future opportunity of Data Analytics and Business Intelligence It is undergoing massive, disruptive changes that will radically transform the way the industry and customers think about analytics. The exponential growth in data is a key driver of this change. In addition, the mainstream adoption of cloud computing across the enterprise continues to put pressure on the capabilities of businesses to incorporate all relevant data from multiple data sources to enable users to make more timely, comprehensive, insightful business decisions. The amount of data that is generated in the business world is doubling every year. Information from devices, machines and social media creates an entirely new set of challenges, and reinforces the fact that data will continue to grow exponentially for the foreseeable future.