Beruflich Dokumente
Kultur Dokumente
Abstract: Network calculations make virtual organization able resources to execute these processes is significant issue in
to share distributed resources in geographical view, to achieve the field of distributed networks with grid architecture.
common goals this method lacks a place, a central controlling In this paper, we will consider a part, called decision
and trusted relation generally , in order to solve the gride
controlling management , based on reinforcement learning
problem. It is necessary to find the most suitable resource in the
shortest time. This purpose as applied in part of solving process.
in resource management unit in grids, which makes us
All devoted approaches of information retrieving try to serve all know better about the resources by passing of time and also
requests optimally in the shortest time, but they are unable to makes the resources map onto the request optimally and
match with or be flexible to grid network changes. Therefore, rapidly.
the flexibility in retrieving and resource allocation are We will utilize reinforcement learning since this method
necessary. In the presented paper, a few part is inserted into matches grid network. Reinforcement learning is the online
protocol in order to manage decision controlling based on
learning and also is applied in environment which is
reinforcement learning, that using learning patterns in grid
network, space recognition , factor recognition, a number of reportedly visible .this method is independent of long
factors and obtaining resource information in grid network ,this information mass for instructor (such as neural network and
part performs retrieving operation and resources allocation genetic – based method). But in our framework, regarding
more optical than other methods. rewards which are achieved in the of path, we obtain
recognition or generally rewards based on traversed steps
Keywords: resource retrieving, grid , reinforcement learning. repeating this approach. Our resource (node) recognition
will be stronger; therefore we can understand the network
1. Introduction better.
Network computation is a computation model that huge Following, we introduce a new approach which is proved by
computations can be processed through them, using comparison and testing, to show important and efficiency of
computation power of many netted computes in addition to our work to the previous studies.
keeping them as a unique virtual computer in view. In other
words, rid is able to solve enormous computational 2. Introduction of information retrieving in
problems, using computational power of several separated grid network
computers which are mostly connected through network
Serving request is performed in two manners: 1.real-time
(internet) [1, 3]. One of the most important current issues of
and 2. Non-real-time method. First, serving is rapidly
computer network is the distributed network with grid
performed to set out appropriate resources as soon as
architecture. Regarding of computer applications
receiving request from broker resource which has mentioned
development as well as hardware rapid, advancement ,
conditions, send a message to broker, then broker selects the
creating integrated systems ( from free heterogeneous
best resource based on different factors such as distance and
resources in order to multi purpose process along with
send the request to the resource for serving. Obviously, this
supporting of maximum- efficiency resources).
method makes the network traffic heavier. In second,
As many as some of the investigate and computational
information of resources is available at request reporting
application these networks are for bottleneck problem
time the discovered resources are saved by brokers in
solving and in order to provide users dedicating optimal
resource management part. Optimal resources are searched
180 (IJCNS) International Journal of Computer and Network Security,
Vol. 2, No. 10, 2010
and selected as soon as request report and then the request our presented method with mentioned ones are applying of
will be sent for selected resources. According to this way, reinforcement learning which makes suitable recognition by
network traffic is significantly decreased and therefore the passing of time in to allocate optimally resources.
requests are responded more rapid.
2.1 Breadth First Search (BFS) 3. Reinforcement learning
BFS is one the simplest and most functional searching Reinforcement learning [13] generally is the art of finding
method with easy process. In this method, every node what strategy to improve the status to achieve the certain goal,
has request, sends the request to all its neighbors. The node regarding of environment recognition, relation results
also searches its local information in order to find perfect behind environment and benefits-damages of performing
answer. These mentioned steps will be done for every node several task simply, Reinforcement learning is the learning
which receives request in the case of finding necessary through environmental relation to achieve specified goal
resourced , a message asks the resources through the nearest decision maker and the person who learn, are called
node which has replied to its request [8]. Making heavy “Agents”. The thing what agent make relation with it,
traffic is the disadvantage of this method. Assume that, we named environment (in fact, every foreign thing of agent is
have ‘n’ given nodes so that every node has m neighbors. In involved). This relation continuously, is occurred in this
the first step, the broker sends a message to or responding m manner that, agent makes decision and then operates an
nodes .after passing of some steps, several messages are action accordingly. After that, the environment will respond
sent. This phenomenon makes the network occupied. with granting reward. Lastly, the agent will be transferred to
new state.
2.2 Random Breadth First Search (RBFS) In detail, agent and environment have relation sequencly
This method is similar to the previous one with this through time step t=1,2,3,…. . in every step, for example in
different that corresponding node doesn’t send the message step t, the agent receives a new state from environment. In
for all its neighbors in every step, but it sends the inquiry to this paper, we suppose that the whole space of grid is s. st Є
the part of neighbors also this method has some S where S is the possible state set of allocating environment
disadvantages and advantages decreasing network traffic resources. at is the possible task set of agent whom does
and rapid operating of the search are advantages. Since them in state st. in the new step, environment grants reward
nodes haven’t any information about neighbors that the R in time t + 1, so that r(t+1) Є R. based on its previous
message is sent for them. Therefore there is no wasting time task, the agent will be transferred to the new state S(t+1). To
to verify and decide in other words. Every node selects mathematics, a policy is a mapping. For instance
randomly some of the neighbors and sends the message to
Π :S× A
→[0,1] is a policy. It means, a number in
them. Neighbor random selection is the significant
disadvantage of this method because the dead parts of [0,1] is appropriated for any pair (action, state) like (s,a)
network which are weakly connected to the network, almost which are belong to Π : S × A . This is shown by
never are made inquiry. Π ( S × A) as follow:
Pr{at = a}{st = s} = Π (s, a )
2.3 Random Breadth First Search with RND-Step
The set of states, actions and sequenced rewards is
actually , this is improvement of previous ones in this
considered in reinforcement learning as follow:
method, we start the search, using ‘n’ brokers instead of
only one broker(‘n’ is depended on given steps) and then
every node of n node searches the releasing resources.
Disadvantages of this method are like RBFS deliberations. Figure 1. The set of stats, actions and sequenced rewards.
The value of s is a function of state value:
In other words, because of being random performance in
searching step, optimal results will not e obtained. On other
hand, searching through couple of paths (and linear
increasing of number of neighbors of all nodes) makes the
efficiency higher. Calculating of real value of the whole state is known as
2.4 Searching with keeping information policy evaluation providing of following of policy Π and it
is necessary for perfect learning the value which can be
In spite of 3 mentioned methods, this method responds the
considered for action-state (s,a) is:
request in non-real status. There are some methods which
Example: when we are going to learn reinforcement
consider the status of neighbors and responses to the
learning system appropriated for the separated and best
requests, including directed breadth first search (DBFS) and
resources to serve requests. According to the mount and
hashing method the efficiency of those methods are higher
time of processing to allocate resources, we initial this unit
then random based methods. In detail, these methods
+1 : appropriate, -1 “ inappropriate , 0: middle. It is notable
decrease significantly network traffic and mount of inquires
that rewarding should be granted so that agent van satisfy
and therefore are able to rapidly the required resources to in-
us, maximizing reward and also it shouldn’t learn, how to
time responses of requests [10, 12].the notable difference of
(IJCNS) International Journal of Computer and Network Security, 181
Vol. 2, No. 10, 2010
satisfy o instance, in example 2 , score +1 ill be granted
when the best resource is selected for processing.