An Overview in Large Databases Using Deep Learning Algorithm

Exploring the Applications of Deep Learning in Big Data Analytics

by Satinder Pal Singh*, Dr. Mukesh Arora,

- Published in Journal of Advances and Scholarly Researches in Allied Education, E-ISSN: 2230-7540

Volume 14, Issue No. 2, Jan 2018, Pages 2053 - 2058 (6)

Published by: Ignited Minds Journals


ABSTRACT

Big Data Analytics and Deep Learning are two high-focus of data science. Big Data has become important as many organizations both public and private have been collecting massive amounts of domain-specific information, which can contain useful information about problems such as national intelligence, cyber security, fraud detection, marketing, and medical informatics. Companies such as Google and Microsoft are analyzing large volumes of data for business analysis and decisions, impacting existing and future technology. Deep Learning algorithms extract high-level, complex abstractions as data representations through a hierarchical learning process. Complex abstractions are learnt at a given level based on relatively simpler abstractions formulated in the preceding level in the hierarchy. A key benefit of Deep Learning is the analysis and learning of massive amounts of unsupervised data, making it a valuable tool for Big Data Analytics where raw data is largely unlabeled and un-categorized. In the present study, we explore how Deep Learning can be utilized for addressing some important problems in Big Data Analytics, including extracting complex patterns from massive volumes of data, semantic indexing, data tagging, fast information retrieval, and simplifying discriminative tasks. We also investigate some aspects of Deep Learning research that need further exploration to incorporate specific challenges introduced by Big Data Analytics, including streaming data, high-dimensional data, scalability of models, and distributed computing. We conclude by presenting insights into relevant future works by posing some questions, including defining data sampling criteria, domain adaptation modeling, defining criteria for obtaining useful data abstractions, improving semantic indexing, semi-supervised learning, and active learning.

KEYWORD

Big Data Analytics, Deep Learning, Data Science, High-level Abstractions, Unsupervised Data, Complex Patterns, Semantic Indexing, Data Tagging, Information Retrieval, Discriminative Tasks

INTRODUCTION

The general focus of machine learning is the representation of the input data and generalization of the learnt patterns for use on future unseen data. The goodness of the data representation has a large impact on the performance of machine learners on the data: a poor data representation is likely to reduce the performance of even an advanced, complex machine learner, while a good data representation can lead to high performance for a relatively simpler machine learner. Thus, feature engineering, which focuses on constructing features and data representations from raw data, is an important element of machine learning. Feature engineering consumes a large portion of the effort in a machine learning task, and is typically quite domain specific and involves considerable human input. For example, the Histogram of Oriented Gradients (HOG) and Scale Invariant Feature Transform (SIFT) are popular feature engineering algorithms developed specifically for the computer vision domain. Performing feature engineering in a more automated and general fashion would be a major breakthrough in machine learning as this would allow practitioners to automatically extract such features without direct human input. Deep Learning algorithms are one promising avenue of research into the automated extraction of complex data representations (features) at high levels of abstraction. Such algorithms develop a layered, hierarchical architecture of learning and representing data, where higher-level (more abstract) features are defined in terms of lower-level (less abstract) features. The hierarchical learning architecture of Deep Learning algorithms is motivated by artificial intelligence emulating the deep, layered learning process of the primary sensorial areas of the neocortex in the human brain, which automatically extracts features and abstractions from the underlying data. Deep Learning algorithms are quite beneficial when dealing with learning from large amounts of unsupervised data, and typically learn data representations in a greedy layer-wise fashion [7,8]. better machine learning results, e.g., improved classification modeling better quality of generated samples by generative probabilistic models and the invariant property of data representations. Deep Learning solutions have yielded outstanding results in different machine learning applications, including speech recognition computer vision and natural language processing A more detailed overview of Deep Learning is presented in Section ―Deep learning in data mining and machine learning‖. Big Data represents the general realm of problems and techniques used for application domains that collect and maintain massive volumes of raw data for domain-specific data analysis. Modern data-intensive technologies as well as increased computational and data storage resources have contributed heavily to the development of Big Data science Technology based companies such as Google, Yahoo, Microsoft, and Amazon have collected and maintained data that is measured in exabyte proportions or larger. Moreover, social media organizations such as Facebook, YouTube, and Twitter have billions of users that constantly generate a very large quantity of data. Various organizations have invested in developing products using Big Data Analytics to addressing their monitoring, experimentation, data analysis, simulations, and other knowledge and business needs making it a central topic in data science research.

OBJECTIVES

• To analyze the exhibitions of the different profound learning algorithms, • To Compare and Analysis of web archive extraction utilize different Algorithms, for example, Deep learning algorithm, Naive Bayes and Back Propagation Neural Network Algorithm.

Need of privacy in machine learning methods and data analytics

AI and information mining strategies must be altered in a productive manner for guaranteeing total ability of assembled information. Today AI techniques with distributed computing are assuming a significant part in large information examination connected widely for successful utilization of predictive capacity of enormous information. For instance, in clinical science, cosmology and so forth the predictive capacity of huge information is generally utilized. The outsider assets do practically all calculations on private information that lead to danger for clients protection. To give security the AI strategies are to be received in protection safeguarding way. In the current computerized world innovation headways have permitted the clients to pull out and burn-through huge information that cause security break to information in most extreme conditions. Information used for controlled with outright laws and rules. The veracity normal for large information shows the requirement for protection safeguarding and security confirmation in enormous information requiring center around weakness databases, static information, dynamic information and so forth, in huge information investigation.

OVERVIEW

Data mining, ―Extraction of hidden predictive information from large databases,” is a powerful new technology with great potential to help companies focus on the most important information in their data warehouses. Data mining tools predict future trends and behaviors, allowing businesses to make proactive, knowledge-driven decisions. The automated, prospective analyses offered by data mining move beyond the analyses of past events provided by retrospective tools typical of decision support systems. Data mining tools can answer business questions that traditionally were time consuming to resolve. They scour databases for hidden patterns, finding predictive information that experts may miss because it lies outside their expectations. Most companies already collect and refine massive quantities of data. Data mining techniques implemented rapidly on existing software and hardware platforms to enhance the value of existing information resources, and integrated with new products and systems as they brought on-line. When implemented on high performance client/server or parallel processing computers, data mining tools can analyze massive databases to deliver answers to questions such as, "Which clients are most likely to respond to my next promotional mailing, and why?"

CONTINUOUS INNOVATION

In spite of the fact that information mining is a moderately new term, the innovation isn't. Organizations have utilized amazing PCs to filter through volumes of grocery store scanner information and break down statistical surveying reports for quite a long time. In any case, ceaseless advancements in PC preparing power, plate stockpiling, and factual programming are significantly expanding the precision of examination while driving down the expense.

Data

Information are any realities, numbers, or text that prepared by a PC. Today, associations are amassing huge and developing measures of information in various arrangements and various databases. This incorporates:

deals, cost, stock, finance, and bookkeeping. • Nonoperational information, like industry deals, estimate information, and large scale financial information. • Meta information portrays information about the actual information, for example, consistent data set plan or information word reference definitions.

Information

The patterns, associations, or relationships among all this data can provide information. For example, analysis of retail point of sale transaction data can yield information on which products are selling and when.

Knowledge

Information changed over into information about chronicled examples and future patterns. For instance, synopsis information on retail grocery store deals broke down considering limited time endeavors to give information on shopper purchasing conduct. Accordingly, a maker or retailer could figure out which things are generally helpless to special endeavors.

Data Warehouses

Emotional advances in information catch, preparing power, information transmission, and capacity abilities are empowering associations to coordinate their different databases into information distribution centers. Information warehousing characterized as a cycle of unified information the board and recovery. Information warehousing, similar to information mining, is a generally new term albeit the actual idea has been around for quite a long time. Information warehousing addresses an ideal vision of keeping a focal storehouse of all authoritative information. Centralization of information expected to amplify client access and examination. Emotional mechanical advances are making this vision a reality for some organizations and, similarly sensational advances in information examination programming are permitting clients to get to this information uninhibitedly. The information examination programming is the thing that upholds information mining.

Example

For instance, one Midwest basic food item chain utilized the information mining limit of Oracle programming to dissect neighborhood purchasing behaviors. They found that when men purchased diapers on Thursdays and Saturdays, they additionally would in general purchase brew. Further investigation showed that these customers ordinarily did their week by week shopping for food on Saturdays. On couple of things. The retailer inferred that they bought the lager to have it accessible for the impending end of the week. The staple chain could utilize this newfound information in different manners to expand income. For instance, they could draw the lager show nearer to the diaper show. Also, they could ensure brew and diapers were sold at the maximum on Thursdays.

FOUNDATIONS OF DATA MINING

Information mining strategies are the aftereffect of a long interaction of exploration and item advancement. This advancement started when business information was first put away on PCs, proceeded with upgrades in information access, and all the more as of late, created innovations that permit clients to explore through their information progressively. Information mining takes this developmental interaction past review information access and route to forthcoming and proactive information conveyance. Information digging is prepared for application in the business local area since it upheld by three advancements that are presently adequately developed. • Massive information assortment, • Powerful multiprocessor PCs and • Data mining algorithms. Business databases are developing at exceptional rates. A new META Group study of information distribution center activities tracked down that 19% of respondents are past the 50-gigabyte level, while 59% hope to be there by second quarter of 1996. One in certain enterprises, like retail, these numbers can be a lot bigger. The going with need for improved computational motors met in a savvy way with equal multiprocessor PC innovation. Information mining algorithms epitomize procedures that have existed for in any event 10 years however have as of late been carried out as full grown, dependable, reasonable instruments that reliably outflank more established measurable techniques.

DATA MINING TECHNIQUES

Artificial neural networks: Non-linear predictive models that learn through training and resemble biological neural networks in structure. Decision trees: Tree-shaped structures that represent sets of decisions. These decisions generate rules for the classification of a dataset. Specific decision tree methods include Classification and Regression Trees (CART) and process such as genetic combination, mutation, and natural selection in a design based on the concepts of evolution. Nearest neighbor method: A technique that classifies each record in a dataset based on a combination of the classes of the k records most similar to it in a historical dataset (where k ³ 1) sometimes called the k-nearest neighbor technique. Rule induction: The extraction of valuable in the event that rules from data dependent on factual importance. Large numbers of these innovations have been in need for over 10 years in particular examination devices that work with generally little volumes of data. These capacities are presently advancing to coordinate straightforwardly with industry-standard data distribution center and OLAP stages. The supplement to this white paper gives a glossary of data mining terms.

What Does Data Mining Work

While enormous scope information innovation has been advancing separate exchange and scientific frameworks, data mining gives the connection between the two. Data mining programming investigates connections and examples in put away exchange data dependent on open-finished client questions. A few sorts of scientific programming are accessible: factual, AI, and neural organizations. By and large, any of four kinds of relationship required are, Classes: Put away data used to find data in foreordained gatherings. For instance, a café network could mine client buy data to decide when clients visit and what they regularly request This information increment traffic by having day by day specials. Clusters: Data items grouped according to logical relationships or consumer preferences. For example, data mined to identify market segments or consumer affinities. Associations: Data mined to identify associations. The beer-diaper example is an example of associative mining. Sequential patterns: Data mined to expect personal conduct standards and patterns. For instance, an outside hardware retailer could anticipate the probability of a knapsack bought dependent on a shopper's acquisition of camping cots and climbing shoes. Data mining comprises of five significant components: • Extract, change, and burden exchange data onto the data stockroom framework. • Provide data admittance to business investigators and information innovation experts. • Analyze the data by application programming.

APPLICATIONS OF DEEP LEARNING IN BIG DATA ANALYTICS

As stated previously, Deep Learning algorithms extract meaningful abstract representations of the raw data through the use of an hierarchical multi-level learning approach, where in a higher-level more abstract and complex representations are learnt based on the less abstract concepts and representations in the lower level(s) of the learning hierarchy. While Deep Learning can be applied to learn from labeled data if it is available in sufficiently large amounts, it is primarily attractive for learning from large amounts of unlabeled/unsupervised data making it attractive for extracting meaningful representations and patterns from Big Data. Once the hierarchical data abstractions are learnt from unsupervised data with Deep Learning, more conventional discriminative models can be trained with the aid of relatively fewer supervised/labeled data points, where the labeled data is typically obtained through human/expert input. Deep Learning algorithms are shown to perform better at extracting non-local and global relationships and patterns in the data, compared to relatively shallow learning architectures Other useful characteristics of the learnt abstract representations by Deep Learning include: (1) relatively simple linear models can work effectively with the knowledge obtained from the more complex and more abstract data representations, (2) increased automation of data representation extraction from unsupervised data enables its broad application to different data types, such as image, textural, audio, etc., and (3) relational and semantic knowledge can be obtained at the higher levels of abstraction and representation of the raw data. While there are other useful aspects of Deep Learning based representations of data, the specific characteristics mentioned above are particularly important for Big Data Analytics. Considering each of the four Vs of Big Data characteristics, i.e., Volume, Variety, Velocity, and Veracity, Deep Learning algorithms and architectures are more aptly suited to address issues related to Volume and Variety of Big Data Analytics. Deep Learning inherently exploits the availability of massive amounts of data, i.e. Volume in Big Data, where algorithms with shallow learning hierarchies fail to explore and understand the higher complexities of data patterns. Moreover, since Deep Learning deals with data abstraction and representations, it is quite likely suited for analyzing raw data presented in different formats and/or from

minimize need for input from human experts to extract features from every new data type observed in Big Data. While presenting different challenges for more conventional data analysis approaches, Big Data Analytics presents an important opportunity for developing novel algorithms and models to address specific issues related to Big Data. Deep Learning concepts provide one such solution venue for data analytics experts and practitioners. For example, the extracted representations by Deep Learning can be considered as a practical source of knowledge for decision-making, semantic indexing, information retrieval, and for other purposes in Big Data Analytics, and in addition, simple linear modeling techniques can be considered for Big Data Analytics when complex data is represented in higher forms of abstraction. In the remainder of this section, we summarize some important works that have been performed in the field of Deep Learning algorithms and architectures, including semantic indexing, discriminative tasks, and data tagging. Our focus is that by presenting these works in Deep Learning, experts can observe the novel applicability of Deep Learning techniques in Big Data Analytics, particularly since some of the application domains in the works presented involve large scale data. Deep Learning algorithms are applicable to different kinds of input data; however, in this section we focus on its application on image, textual, and audio data.

CONCLUSION

The exactness pace of Deep Learning gives a high recognition rate when contrasted with Naive Baye's Classifier and BPNN. The review pace of Naive Baye's is 65% whereas, for back engendering is 72% individually. Notwithstanding Deep Learning algorithm yields 74% review rate. Subsequently, Deep Learning gives 2% higher review rate than the BPNN. The affectability pace of Deep Learning, which gives a high recognition rate when looked at Naive Baye's Classifier, and BPNN. The determine pace of Naive Baye's is 63% though, for back proliferation is 70% individually. All things considered, Deep Learning algorithm yields 72% review rate. So Deep Learning gives 2% higher explicitness rate than the BPNN. Just as the F-Measure, worth of Deep Learning which gives a high discovery rate when contrasted with Naive Baye's Classifier and BPNN. At long last, it is inspecting that Deep Learning is quicker and precise when contrasted with Naive Baye's and Back proliferation. Profound learning frameworks are feasible to execute now in view of three reasons: High CPU power, Better Algorithms, and the accessibility of more data. Throughout the following not many years, these variables will prompt more utilizations of Deep learning frameworks. Profound learning applications are most appropriate for circumstances, which include a lot of data and complex connections between various boundaries. Tackling natural issues: Training a that: "Given an information, this is the right yield" If this done what's necessary occasions, an adequately prepared organization will imitate the capacity you are reproducing. It will likewise disregard inputs that are unessential to the arrangement.

REFRANCES

[1] Michael Azmy (2005). ―Web Content Mining Research: A Survey‖ DRAFT Version 1, Pages: 1-15. [2] Paul Viola (2005). ―Learning to Extract Information from Semi structured Text using a Discriminative Context Free Grammar‖ Draft submitted to the conference ACM-SIGIR, Pages: 1-8. [3] Georgios Sigletos, Michalis Hatzopoulos, Georgios Paliouras and Constantine D. Spyropoulos (2005). ―Combining Information Extraction Systems Using Voting and Stacked Generalization‖ Journal of Machine Learning Research, 6, Pages: 1751-1782. [4] Geoffrey E. Hinton, Simon Osindero, and Yee-Whye The (2006). ―A fast learning algorithm for deep belief nets‖ Neural Computation, Pages: 1-16. [5] Qiang Yang, and Xindong WU (2006). ―10 Challenging Problems in data mining research‖ International Journal of Information Technology & Decision Making, Vol. 5, No. 4, Pages: 597–604. [6] Ioan Pop (2006). ―An approach of the Naive Bayes classifier for the document classification‖ General Mathematics Vol. 14, No. 4, Pages: 135–138. [7] Jordi Turmo, Alicia Ageno, and Neus Catal (2006). ―Adaptive Information Extraction‖ ACM Computing Surveys, Vol. 38, No. 2, Article 4, Pages: 1-47. [8] Geoffrey E. Hinton, Simon Osindero and Yee-Whye The (2006). ―A Fast Learning Algorithm for Deep Belief Nets‖ Neural Computation (Elsevier) 18, Pages: 1527–1554. [9] Hal Daum and Daniel Marcu (2006). ―Domain Adaptation for Statistical Classifiers‖ Journal of Artificial Intelligence Research 26, Pages: 101-126. [10] A. Jebaraj Ratnakumar (2006). ―An implementation of web personalization [11] Jun Zhu, Zaiqing Nie, Ji-Rong Wen, Bo Zhang, and Hsiao-Wuen Hon (2007). ―Webpage Understanding: an Integrated Approach‖ ACM Int. Conf on KDD‘07, Pages: 1-10, August 12–15, San Jose, California, USA. [12] Tak-Lam Wong and Wai Lam (2007). ―Adapting Web Information Extraction Knowledge via Mining Site-Invariant and Site-Dependent Features‖ ACM Transactions on Internet Technology, Vol. 7, No. 1, Article 6, Pages:1-40.

Corresponding Author Satinder Pal Singh*

Research Scholar, Sunrise University, Alwar, Rajasthan