Académique Documents
Professionnel Documents
Culture Documents
Abstract- This paper narrate the design image tools as a data warehouse (DW) and data mining
for the implementation of the DW in an software is recommended. Organizational
educational institute. (BI) Business Knowledge can be used for strategic planning and
Intelligence build upon a set of applications improvement of the main performance indicators
and tools that enable the analysis of huge in research and Academy.
amount of information (Big Data). Data
warehouse (DW) using educational data This paper explains the use of BI and enterprise
mining (EDM) techniques will be used for architecture (EA) to take all the knowledge
knowledge discovery process to handle the dimensions in an organization. In explaining the
information for analysis of key performance process of knowledge distribution and creation a
indicator. Every year a large volume of Big knowledge management framework was
Data is handle by educational institutes. To proposed. In addition, the framework includes
improve process and decision support system Web components to view the information from
there is a strong need to use BI in these EA and BI repositories .The application of this
institute. DW is the key technology in a BI KMF can support the enhancement of different
project. To get creation, capture, transfer and processes and services in educational institutions.
digitalization of knowledge this framework
was used. The objective is to over come the In analyzing and processing of multidimensional
severe gap between existing academic data DW serves it better and there are many
potential and their unsuccessful learning in papers which talk about data mining, educational
schools and universities. In a private data mining (EDM). The design of a DW in
university this approach and the framework institutions are shown in this paper. Moreover,
are two outcome of a research project. ETL of data from operational data sources into a
Moreover, this paper suggests how to select the DW is described. It this study the paper shows the
best methodology in higher instructions or steps to design of a DW in a private university.
colleges. To analyse information using EDM The requirement for big data framework can be
techniques this study can be used for study and seen when applying any algorithm on big
practical use in studies that plan to design a database. A one core of the CPU is a local system
DW. is used to improve performance as the data size
increases. GPU are being widely used which have
I. Introduction multi cores. As GPU are not every time
economically feasible and accessible so there is a
Huge amounts of information are generated every need for the framework this uses the existing
day in educational institutions. The proper use of CPU in the local system.
this information may be essential for the creation
of knowledge. To transform information into This has been hampered by the diversity of
knowledge, the use of business intelligence (BI) storage methods information in transactional
systems, responsible for operational data implementation of data mining techniques
processing. Moreover, it's difficult to analyze (Fernandez et al. 2014). The use of big data are as
information from previous years in transactional follows: According to Hashem et al. "Large data
terms data base. Information is not constant and sources from the cloud and Web are stored in a
often redundant and not always reliable. While distributed fault-tolerant database and processed
searching for solutions to these problems the idea through a programing model for large datasets
of creating data warehouses have emerged. with a parallel distributed algorithm in a cluster."
To collect these huge amount of data the use of
The most important tools that can achieve this different educational data mining approaches and
task is Hadoop which is an open source platform techniques are used which removes the cover or
that provides best data management provisions. extracting knowledge from large dataset.
The important use of Hadoop is to facilitate the (Aghabozorgi et all. 2017).
processing of very large datasets in a distributed
computing environment using Hadoop (Cen et al. 2016) New opportunities for big data
distribution file system (HDFS). However, this analytics to get better efficiency of student
paper, explore a more efficient and robust tool, learning and maximize their knowledge retention.
Apache Spark, which was designed to work with Learning for the individual student they proposed
Hadoop to meet some of its limitations. data driven learning for identify of the patterns of
learning that could advise on the most effective
II. Literature review format. There are two main way to process the big
educational data.
The objective of this study is the main research • Hardware cost and investment of the
question:- What is the design thoughts for the computer cluster. Which can be done by
application of a DW in an educational institution? installing Hadoop as open source
The use of internet in education has new context platform.
which is known as web based education in which • To save all the hardware cost and use
huge amount of data about teaching and learning some of the open source or commercial
are generated and easily available. The education platform and cloud sources.
data are categorized to five different categories:- Above mentioned solutions have similar base.
The File System (FS) which can handle all the
• Identity data which contain student distributed and parallel processing operations
identity and boarding. automatically.
• Student activity based datasets that have
potential to improve learning results
(inferred student data, user interaction
data, inferred content data and system
wide data)