Sie sind auf Seite 1von 2

1.

the researches on the process management gets increased but the toughest job
is to coduct process analysis due to huge amount of data. So by this project we
collect the event logs called big data and analyze the collected logs with the
processes as structured and unstructured data. Usually only structured processes
are identified but we also use unstructured data for enhanced clarity results. The
system automatically discovers a process model and conducts various performance
analysis on the manufacturing processes
2. the knowledge of dynamic traffic and its usage of data services in cellular
networks is important for network resources and improving users experience. The
studies related on its behavior, device type and application is made. Here we use
service providers, to reveal the traffic characters in cellular data networks. Traffic
data collected over a five-day period from a leading mobile operator's core network
in China. We propose a Zipf-like model to characterise the distributions of the traffic
volume, subscribers, and requests among service providers. Our work differs from
previous related works in that we perform measurements on a large quantity of data
covering 2.2 billion traffic records, and we first explore the traffic patterns of
thousands of service providers. Results of our study present mobile Internet
participants with a better understanding of the traffic and usage characteristics of
service providers, which play a critical role in the mobile Internet era.
3. the finding of rules from the data set is the tougher research in which the
previous schemes sometimes fail to reduce the association rules while more
reduction is possible since they do not consider the relationship between the data
items. In this paper we propose Latent Semantic Analysis (LSA) reduction technique
for mining valuable rules at high speed regardless of the number of items. The
proposed scheme extracts the relationship such as inverse and equivalence
between a set of items. Computer simulation reveals that it significantly increases
credibility, support, processing time, reduction rate of the rules and rejection rate of
the item, compared to the existing schemes.
4. There has been an increasing interest in big data and big data security with the
development of network technology and cloud computing. However, big data is not
an entirely new technology but an extension of data mining. In this paper, we
describe the background of big data, data mining and big data features, and
propose attribute selection methodology for protecting the value of big data.
Extracting valuable information is the main goal of analyzing big data which need to
be protected. Therefore, relevance between attributes of a dataset is a very
important element for big data analysis. We focus on two things. Firstly, attribute
relevance in big data is a key element for extracting information. In this perspective,
we studied on how to secure a big data through protecting valuable information
inside. Secondly, it is impossible to protect all big data and its attributes. We
consider big data as a single object which has its own attributes. We assume that a
attribute which have a higher relevance is more important than other attributes.

5. With the beginning of new era data has grown rapidly not only in size but also in
variety. There is a difficulty in analyzing such big data. Data mining is the technique
in which useful information and hidden relationship among data is extracted. The
traditional data mining approaches could not be directly implanted on big data as it
faces difficulties to analyze big data. Clustering is one of the major techniques used
for data mining in which mining is performed by finding out clusters having similar
group of data. In this paper we have discussed some of the current big data mining
clustering techniques. Comprehensive analysis of these techniques is carried out
and appropriate clustering algorithm is provided.

Das könnte Ihnen auch gefallen