SlideShare una empresa de Scribd logo
1 de 8
Descargar para leer sin conexión
International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013

A Novel Framework and Policies for On-line Block
of Cores Allotment for Multiple DAGs on NoC
Uma Boregowda1 and Venugopal C R2
1

2

Malnad College of Engineering, Hassan, India
Sri Jayachamarajendra College of Engineering, Mysore, India

ABSTRACT
Computer industry has widely accepted that future performance increases must largely come from
increasing the number of processing cores on a die. This has led to NoC processors. Task scheduling is one
of the most challenging problems facing parallel programmers today which is known to be NP-complete. A
good principle is space-sharing of cores and to schedule multiple DAGs simultaneously on NoC processor.
Hence the need to find optimal number of cores for a DAG for a particular scheduling method and further
which region of cores on NoC, to be allotted for a DAG . In this work, a method is proposed to find nearoptimal minimal block of cores for a DAG on a NoC processor. Further, a time efficient framework and
three on-line block allotment policies to the submitted DAGs are experimented. The objectives of the
policies, is to improve the NoC throughput. The policies are experimented on a simulator and found to
deliver better performance than the policies found in literature..

KEYWORDS
DAG; task scheduling; NoC

1. INTRODUCTION
The advancement in VLSI has now led to the increased number of cores on a single chip. A
study at University of California, Berkeley[1] indicates that future chips will have thousands of
cores. Traditional bus based systems and cache hierarchies have given way to more flexible and
scalable Network-on-chip(NoC) architecture. Massive parallel computing performed on NoC is
the future computing, due to the large increased computing demands. With the increasing
number of cores, it is becoming further difficult and demanding to exploit the full power of all
cores on the chip.
The data flow model is gaining popularity as a programming paradigm for multi-core processors
and NoC. When the characteristics of an application is fully deterministic, including task's
execution time, size of data communicated between tasks, and task dependencies, the application
can be represented by a Directed Acyclic Graph (DAG). A DAG G=(V,E), where V = { vi |
i=1, ... V} is a set of vertices representing tasks and E={e i,j|i ,j) {1, ... V} * {1, ... V} is a set of
edges representing precedence constraints between tasks. DAG is assumed to have a single entry
and single exit task. The structure of a general DAG is shown in Fig. 1 with eleven number of
taks. The weight of a vertex v ∈ V corresponds to the execution time of the task, which can be
measured via benchmarking or can be calculated via a performance model. The weight of an
edge ei,j ∈ E, is the communication cost measured as the time taken to transfer data Di,j between
DOI:10.5121/ijcsa.2013.3606

55
International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013

vertices vi and vj, which is a function of data volume Di,j, the network characteristics, the
processors allocated to it and data distribution cost.

Fig. 1 A General DAG

Communication-to-Computation (CCR) is the ratio of average communication cost to the average
computation cost of a DAG. This characterizes the nature of DAG. The objective of scheduling is
to map tasks onto cores and order their execution so that task dependencies are satisfied and
minimum overall completion time is achieved.
Scheduling can be static or dynamic. Static schedule can be generated only when the application
behaviour is fully deterministic and this has the advantage of being more efficient and a small
overhead during runtime. The full global knowledge of application in the form of DAG will help
to generate a better schedule. Makespan is the total time required to complete a DAG. The
challenge of scheduling tasks onto cores and also maintaining precedence constraint, with an
objective of minimizing overall execution time is proved to be a NP hard problem[2].
Not all applications can effectively utilize all cores on the chip, hence it would be advantageous
to multiprogram the NoC and space-partition the cores among applications, being
multiprogrammed. The design has now shifted from scheduling single application to multiple
applications. Hence the problem of scheduling DAG on NoC is twofold : first, to allot cores for a
DAG and second, to schedule tasks of DAG onto the allotted cores. This paper deals with the first
issue.
The contributions of this paper include (a) Finding near-optimal block of cores for a given DAG
for a particular scheduling method (b) To propose and implement a framework and three on-line
block of cores allotment policies for the DAGs submitted, with an objective of improving the
total time required to complete all DAGs and hence the throughput.

2. RELATED WORK
NoC processors are more interested in how many DAGs can be completed over a period of time
rather than how quickly an individual DAG gets completed, since the cores are not a very scarce
56
International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013

resource on it. Thus the more emphasis is on throughput of DAGs. The scheduling problem has
been extensively studied for several decades. Sakellariou et. al.[3] have studied the problem of
scheduling multiple DAGs onto a set of heterogeneous resources. The main work is to optimize
overall makespan and also to achieve fairness for all DAGs. One approach is to combine several
DAGs into one and then use standard methods to schedule the task graph. Another way is to
perform task-by-task scheduling over all tasks in all task graphs. Ahmed et. al.[4] team have
proposed a game theory based scheduler on multicore processor with the objective of minimizing
energy consumption.
Cilk[5], Intel Thread Building Blocks(TBB) and OpenMP have proposed several task scheduling
techniques over multicore. These systems are optimized specifically for scheduling DAGs on
multicore but not for NoC processors. Viktor K Prasanna et. al.[6] have implemented a
hierarchical scheduler which could adapt to the input task dependency graph and demonstrated its
advantages for NoC architectures. They have divided the threads into groups, each having a
manager to perform scheduling at the group level and several workers to perform self-scheduling
for the tasks assigned by the manager. A super manager was used to dynamically adjust the group
size, so that the scheduler could adapt to the input task dependency graph.
Tang et. al.[7] proposed a two-step genetic algorithm and the related software for mapping
concurrent applications on a fixed NoC architecture. Murali et. al.[8] presented a methodology to
map multiple use-cases onto the NoC architecture, satisfying the constraints of each use-case. The
two-step mapping method [9] first finds a region on the NoC for a given application and then
maps all tasks of the application onto the region. Several strategies based on the Maximum Empty
Rectangle(MER) techniques, like BestSize, BestShape, BestSize BestShape and BestShape
BestSize are experimented. For each DAG to be scheduled, the number of cores required is
known before hand and DAG is scheduled only if specified cores are available. Hence, a
drawback of this work is that, it does not schedule a DAG if available cores is less than the
specified cores. But to increase the throughput of DAGs completion and also to maximize core
utilization, it is better to schedule a DAG even when the available cores are marginally less than
the specified cores.
This proposed work tries to schedule a DAG with less than specified cores whenever the specified
cores are not available and if it increases the throughput without much affecting that DAG's
makespan. It is generally beneficial in terms of throughput, to run more DAGs with less than
specified cores, than running few DAGs with full specified cores.

3. PROPOSED WORK
3.1. Finding Near-optimal block of cores
The reason behind the idea of allotting cores in the form of block for a DAG are (a) it would be
easier and time efficient to allot and free cores for a DAG in the form of block (b) the nodes
average distance (NAD) is minimum for a block and this minimizes the communication time
between the tasks, hence the completion time of the schedule generated. NAD is the average
distance between any two randomly selected nodes on NoC. The block can only be a square or
rectangle. To find the near-optimal block of cores, an initial schedule for the DAG is generated
using a well known list scheduling method and trying to use all cores on NoC. The scheduling
method also takes into account the number of hops required to move data between any two cores
on NoC, the amount of data to be moved and the bandwidth of the intermediate channels, hence
the generated schedule is more realistic. All the cores are not utilized to the maximum extent in
the generated schedule. Thus it would be beneficial to eliminate low utilized cores from the
schedule, without much affecting the completion time of the DAG. Since the cores allotted for a
57
International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013

DAG must be in the form of a block, the low utilized rows or columns of cores are eliminated.
This method is given in the below algorithm.
Algorithm Near_Optimal_Block()
Input : DAG ‘G’
Output : Number of rows and columns of cores
1.
2.
3.

4.

Start
with
maximum
number
of
rows
and
column
of
cores
on the given NoC
Find the schedule for the given DAG ‘G’, using any well known list scheduling method
like- ETF, HLFET[2] etc.
Repeat
3.1 Reduce row or column of cores, whichever has least total number of tasks assigned
to it.
3.2 Generate the new schedule with revised row and
column number
3.3 Find the percentage increase in core utilization and
percentage
increase
in
makespan of the new schedule in comparison to previous schedule
Until percentage increase in makespan exceeds percenage increase in core utilization
end_algorithm

The iteration at which, percentage increase in makespan exceeds percentage increase in core
utilization, determines the number of rows and columns of cores for the DAG. It is the point
where any further reduction of row/column of cores will significantly increase the makespan.
Hence it fixes the near-optimal block of cores for a given DAG.

3.2. A Framework for On-line Block allotment
A Framework for on-line submission of DAGs, allotment of block of cores for each DAG,
scheduling of DAGs and finding its completion time is developed. An efficient method to track
the status of cores on NoC and hence to allot cores for a DAG is implemented. The NoC is
viewed as a single pool of cores, while previous methods[9] have maintained it as several blocks
of free or allotted cores. The disadvantages of maintaining NoC as a collection of free or allotted
blocks are (a) the job of splitting the blocks during DAG allotment when no free block with the
specified size is available and merging the blocks when a DAG is completed, is quite
computationally expensive to be used as an on-line method, (b) there will be few situations where
a DAG can not be allotted cores, as the required number of free cores are not available as a single
block but as a collection of few adjacent blocks. However the proposed method will allot block
when the required number of free cores are available either as a single block or as few adjacent
blocks.
Fig. 2 shows an instance where two blocks are allotted to DAG1 and DAG2, and another two
blocks FB1 and FB2 are free. Using methods[9], a demand for block of cores of size 4*2 or 5*2
can not be satisfied as the required number of cores are not available as a single block. But the
required free block is present as 2 adjacent blocks. The allotment of such adjacent blocks needs
frequent merging of blocks which is computationally expensive. The proposed method will never
fail to allot, whenever required cores are available either as a single free block or as few adjacent
free blocks. A mask with a dimension of the desired block size is used to locate the presence of
free block on NoC in a very simple manner.

58
International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013

Fig. 2 An Instance of Block Allotment for DAGs

4. EXPERIMENT AND RESULTS
4.1. Experimental Methodology
A simple simulator using the proposed framework is developed to simulate the arrival,
scheduling, execution and completion of DAGs. Poisson distribution is used to simulate the
arrival time of DAGs. This software is run on multi2sim[12], a functional simulator to measure
total running time.

4.2. A Finding Near-optimal block of cores
The proposed method to find the near-optimal minimal block of cores is implemented using the
well known list scheduling method ETF[2]. This is experimented for several kinds of benchmark
DAGs taken from Task Graphs For Free[10], Standard Task Graph Set [11] and also DAGs of
real applications like Strassen matrix multiplication, FFT and complex matrix multiplication.
DAGs with CCR values of 0.1, 0.4, 1 and 5 are used in experiments. Fig. 3 shows the plot of
percentage increase in makespan and core utilization, as the row or column of cores, allotted to a
DAG, is reduced in subsequent iterations. For the DAG considered in Fig. 1, the block of cores
dimensions are 5*6, 4*5, 3*5 and 3*4 at iterations 1, 2, 3 and 4 respectively. Thus the optimal
block size is 3*5, as the cross-over of plots happens at iteration number 4, when the block
dimension is 3*4. This method can be used for any kind of DAG.

59
International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013

Fig 3. Finding Minimal Block Size

4.3. A Comparison of Block Allotment Policies
To compare the performance of the proposed novel method of tracking the status of cores on
NoC, the policies BestSize[9] and BestShape[9] are implemented, by viewing the NoC as a
collection of free or alloted blocks. The proposed BestFit chooses the best fit block like the
BestSize but views NoC as a single pool of cores. Also, it can allot block of cores in few cases
when BestSize[9] fails to allot, because of the manner in which status of cores are maintained.
The simulator is run for 4 categories of DAG sets, each with 10, 20, 30 and 40 DAGs
respectively.
The performance metric used is complete makespan, which is the time to complete all DAGs. For
comparison, normalized complete makespan is used, which normalizes the obtained complete
makespan with respect to BestShape, as it is the best policy[9] found in literature. The average
normalized complete makespan of DAGs within each category is computed, for each block
allotment policy. Fig. 4 shows the comparison of average normalized complete makespan for
different policies. Proposed BestFit exhibits better performance against the other two policies, for
all categories. The improvement in performance is further enhanced with increase in number of
DAGs in a set. This happens as the number of times block allotment is successful in proposed
method, while it is a failure in previous method is more. But the significant advantage of
proposed method is its low execution time, which was evident from the recorded runtime on
simulator and thus proving that proposed method is the right candidate for on-line block allotment
for DAGs.

Fig. 4 Comparing New Method of Tracking Core Status with previous method

60
International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013

Fig. 5 shows the performance comparison of all three proposed block allotment policies. Metric
used is normalized overall makespan with respect to BestFit policy. DAGs with both low and
high CCR, combined together with low and high arrival rates, are considered for each set of
DAGs. Each set has 50 DAGs. The set LoC_LoA includes DAG with low CCR and low arrival
rate. DAGs with high CCR and high arrival rate are included in the set HiC_HiA.
The DAGs with low CCR can better utilize more cores, as the communication cost is low. Thus
when allotted less number of cores for a DAG, resultant increase in makespan is slightly more.
For the set LoC_LoA, it is evident from Fig. 5 that performance of AvailableCore is better by
10% with reference to BestFit policy. When the arrival rate of DAGs is low, Adoptive policy
performance is close to AvailableCore. As the arrival rate is more in LoC_HiA, the adoptive
policy performs better by 9% in comparison to AvailableCore.
When CCR is high, DAGs can not effectively utilize more cores, hence a decrease in cores
allotted would not significantly affect the makespan. Thus for the set HiC_LoA, the performance
of AvailableCore is significantly better than BestFit by 18%. When CCR and arrival rate of
DAGs is high as in HiC_HiA, both Adoptive and AvailableCore show good performance
improvement of around 20%. For the set with all kinds of DAGs randomly chosen, Adoptive and
AvailableCore policies perform better by 17% in comparison to BestFit.

Fig. 5 Comparison of Block Allotment Policies

5. CONCLUSION
A framework for on-line DAGs submission, allotting block of cores, scheduling and completion
of DAGs is proposed and implemented on a simulator. Three block allotment policies – BestFit,
AvailableCores and Adoptive are compared with the policies found in literature. The simple
BestFit policy delivers improvement over the policies found in literature, because of the new
proposed method of tracking the status of cores on NoC. The novel methods – AvailableCore and
Adoptive delivers better performance than BestFit. Of the proposed policies, Adoptive one is
good, as it can adjust the allotment to the arrival rate of DAGs, thus reducing the overall
completion time. These experiments have proved that it is beneficial to schedule several DAGs
with less than specified cores, than running DAGs with full specified cores.
Future Enhancements – The experiment is now simulated using DAGs with dummy tasks. The
idea must be experimented on real NoC, with DAGs of real tasks.
61
International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013

REFERENCES
[1]

Krste Asanovic, Ras Bodik, Bryan C. Catanzaro, Joseph J. Gebis, Parry Husbands, Kurt Keutzer,
David A. Patterson, William L. Plishker, John Shalf, Samuel W. Williams, and Katherine A. Yelick.
The landscape of parallel computing research: a view from berkeley. (UCB/EECS-2006- 183),
December 2006.
[2] Y. K. Kwok and I. Ahmad. Static Scheduling Algorithms for allocating directed task graphs to
multiprocessors. ACM Computing Surveys, 31(4):406–471, 1999. izem, Aksahya & Ayese, Ozcan
(2009) Coomunications & Networks, Network Books, ABC Publishers.
[3] H. Zhao and R. Sakellariou. Scheduling Multiple DAGs onto Heterogeneous systems. In IEEE
International Symposium on Parallel and Distributed Processing (IPDPS), pages 1– 12, 2006.
[4] I. Ahmad, S. Ranka, and S. Khan. Using Game theory for Scheduling tasks on multi-core processors
for Simultaneous optimization of performance and energy. In Intl. Sym. on Parallel Dist. Proc., pages
1–6, 2008.
[5] R. D. Blumofe, C. F. Joerg, B. C. Kuszmaul, C. E. Leiserson, K. H. Randall, and Y. Zhou. Cilk: An
efficient multithreaded runtime system. Technical report, Cambridge, 1996
[6] Yinglong Xia, Viktor K. Prasanna, James Li. Hierarchical Scheduling of DAG Structured
Computations on Manycore Processors with Dynamic Thread Grouping, 15th Workshop on Job
Scheduling Strategies for Parallel Processing, In conjunction with IPDPS (JSSPP'10), 2010
[7] Tang Lei and Shashi Kumar. A two-step genetic algorithm for mapping task graphs to a network on
chip architecture. In DSD ’03: Proceedings of the Euromicro Symposium on Digital Systems Design,
page 180, Washington, DC, USA, 2003. IEEE Computer Society
[8] Srinivasan Murali, Martijn Coenen, Andrei Radulescu, Kees Goossens, and Giovanni De Micheli.
Mapping and Configuration Methods for multi use-case networks on chips. In ASP-DAC ’06:
Proceedings of the 2006 Asia and South Pacific Design Automation Conference, pages 146–151,
Piscataway, NJ, USA, 2006. IEEE Press
[9] Bo Yang, Liang Guang, Xu, T. C. Santti, T. Plosila. Multi-application Multi-step Mapping Method
for Many-core Network-on-Chip platforms, Israel, IEEEI, 2010
[10] Task Graphs for Free http://ziyang.eecs.umich.edu/~dickrp/tgff
[11] Standard Task Graph Set http;//www.kasahara.elec.waseda.ac.lp/schedule
[12] Multi2sim Simulator http://www.multi2sim.org/

Author
C.R.Venugopal received his Ph. D from IIT, Bombay. Presently serving as a Professor
in Sri Jayachamarajendra College of Engineering, Mysore, India. His main research
interests are Cloud computing, High Performance computing, VLSI Technology and
File System development. Has authored more than 50 international conference and
journal papers.
Uma B completed M. Tech in Computer Science and Engineering from IIT, Delhi.
Currently working as a Associate Professor in Malnad College of engineering, Hassan,
India. Her research interests are Parallel Programming, High Performance Computing
and Task scheduling.

62

Más contenido relacionado

La actualidad más candente

A general weighted_fuzzy_clustering_algorithm
A general weighted_fuzzy_clustering_algorithmA general weighted_fuzzy_clustering_algorithm
A general weighted_fuzzy_clustering_algorithmTA Minh Thuy
 
Scalable and Adaptive Graph Querying with MapReduce
Scalable and Adaptive Graph Querying with MapReduceScalable and Adaptive Graph Querying with MapReduce
Scalable and Adaptive Graph Querying with MapReduceKyong-Ha Lee
 
An enhanced adaptive scoring job scheduling algorithm with replication strate...
An enhanced adaptive scoring job scheduling algorithm with replication strate...An enhanced adaptive scoring job scheduling algorithm with replication strate...
An enhanced adaptive scoring job scheduling algorithm with replication strate...eSAT Publishing House
 
Task Scheduling using Tabu Search algorithm in Cloud Computing Environment us...
Task Scheduling using Tabu Search algorithm in Cloud Computing Environment us...Task Scheduling using Tabu Search algorithm in Cloud Computing Environment us...
Task Scheduling using Tabu Search algorithm in Cloud Computing Environment us...AzarulIkhwan
 
Hadoop scheduler with deadline constraint
Hadoop scheduler with deadline constraintHadoop scheduler with deadline constraint
Hadoop scheduler with deadline constraintijccsa
 
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENT
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENTDYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENT
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENTIJCNCJournal
 
SASUM: A Sharing-based Approach to Fast Approximate Subgraph Matching for Lar...
SASUM: A Sharing-based Approach to Fast Approximate Subgraph Matching for Lar...SASUM: A Sharing-based Approach to Fast Approximate Subgraph Matching for Lar...
SASUM: A Sharing-based Approach to Fast Approximate Subgraph Matching for Lar...Kyong-Ha Lee
 
A PROGRESSIVE MESH METHOD FOR PHYSICAL SIMULATIONS USING LATTICE BOLTZMANN ME...
A PROGRESSIVE MESH METHOD FOR PHYSICAL SIMULATIONS USING LATTICE BOLTZMANN ME...A PROGRESSIVE MESH METHOD FOR PHYSICAL SIMULATIONS USING LATTICE BOLTZMANN ME...
A PROGRESSIVE MESH METHOD FOR PHYSICAL SIMULATIONS USING LATTICE BOLTZMANN ME...ijdpsjournal
 
Energy-aware VM Allocation on An Opportunistic Cloud Infrastructure
Energy-aware VM Allocation on An Opportunistic Cloud InfrastructureEnergy-aware VM Allocation on An Opportunistic Cloud Infrastructure
Energy-aware VM Allocation on An Opportunistic Cloud InfrastructureMario Jose Villamizar Cano
 
Fault-Tolerance Aware Multi Objective Scheduling Algorithm for Task Schedulin...
Fault-Tolerance Aware Multi Objective Scheduling Algorithm for Task Schedulin...Fault-Tolerance Aware Multi Objective Scheduling Algorithm for Task Schedulin...
Fault-Tolerance Aware Multi Objective Scheduling Algorithm for Task Schedulin...csandit
 
Estimation of Optimized Energy and Latency Constraint for Task Allocation in ...
Estimation of Optimized Energy and Latency Constraint for Task Allocation in ...Estimation of Optimized Energy and Latency Constraint for Task Allocation in ...
Estimation of Optimized Energy and Latency Constraint for Task Allocation in ...ijcsit
 
Congestion Control through Load Balancing Technique for Mobile Networks: A Cl...
Congestion Control through Load Balancing Technique for Mobile Networks: A Cl...Congestion Control through Load Balancing Technique for Mobile Networks: A Cl...
Congestion Control through Load Balancing Technique for Mobile Networks: A Cl...IDES Editor
 
Improved Max-Min Scheduling Algorithm
Improved Max-Min Scheduling AlgorithmImproved Max-Min Scheduling Algorithm
Improved Max-Min Scheduling Algorithmiosrjce
 
Time Efficient VM Allocation using KD-Tree Approach in Cloud Server Environment
Time Efficient VM Allocation using KD-Tree Approach in Cloud Server EnvironmentTime Efficient VM Allocation using KD-Tree Approach in Cloud Server Environment
Time Efficient VM Allocation using KD-Tree Approach in Cloud Server Environmentrahulmonikasharma
 

La actualidad más candente (17)

Bulk transfer scheduling and path reservations in research networks
Bulk transfer scheduling and path reservations in research networksBulk transfer scheduling and path reservations in research networks
Bulk transfer scheduling and path reservations in research networks
 
A general weighted_fuzzy_clustering_algorithm
A general weighted_fuzzy_clustering_algorithmA general weighted_fuzzy_clustering_algorithm
A general weighted_fuzzy_clustering_algorithm
 
Scalable and Adaptive Graph Querying with MapReduce
Scalable and Adaptive Graph Querying with MapReduceScalable and Adaptive Graph Querying with MapReduce
Scalable and Adaptive Graph Querying with MapReduce
 
An enhanced adaptive scoring job scheduling algorithm with replication strate...
An enhanced adaptive scoring job scheduling algorithm with replication strate...An enhanced adaptive scoring job scheduling algorithm with replication strate...
An enhanced adaptive scoring job scheduling algorithm with replication strate...
 
Task Scheduling using Tabu Search algorithm in Cloud Computing Environment us...
Task Scheduling using Tabu Search algorithm in Cloud Computing Environment us...Task Scheduling using Tabu Search algorithm in Cloud Computing Environment us...
Task Scheduling using Tabu Search algorithm in Cloud Computing Environment us...
 
Hadoop scheduler with deadline constraint
Hadoop scheduler with deadline constraintHadoop scheduler with deadline constraint
Hadoop scheduler with deadline constraint
 
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENT
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENTDYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENT
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENT
 
SASUM: A Sharing-based Approach to Fast Approximate Subgraph Matching for Lar...
SASUM: A Sharing-based Approach to Fast Approximate Subgraph Matching for Lar...SASUM: A Sharing-based Approach to Fast Approximate Subgraph Matching for Lar...
SASUM: A Sharing-based Approach to Fast Approximate Subgraph Matching for Lar...
 
A PROGRESSIVE MESH METHOD FOR PHYSICAL SIMULATIONS USING LATTICE BOLTZMANN ME...
A PROGRESSIVE MESH METHOD FOR PHYSICAL SIMULATIONS USING LATTICE BOLTZMANN ME...A PROGRESSIVE MESH METHOD FOR PHYSICAL SIMULATIONS USING LATTICE BOLTZMANN ME...
A PROGRESSIVE MESH METHOD FOR PHYSICAL SIMULATIONS USING LATTICE BOLTZMANN ME...
 
Energy-aware VM Allocation on An Opportunistic Cloud Infrastructure
Energy-aware VM Allocation on An Opportunistic Cloud InfrastructureEnergy-aware VM Allocation on An Opportunistic Cloud Infrastructure
Energy-aware VM Allocation on An Opportunistic Cloud Infrastructure
 
Fault-Tolerance Aware Multi Objective Scheduling Algorithm for Task Schedulin...
Fault-Tolerance Aware Multi Objective Scheduling Algorithm for Task Schedulin...Fault-Tolerance Aware Multi Objective Scheduling Algorithm for Task Schedulin...
Fault-Tolerance Aware Multi Objective Scheduling Algorithm for Task Schedulin...
 
Estimation of Optimized Energy and Latency Constraint for Task Allocation in ...
Estimation of Optimized Energy and Latency Constraint for Task Allocation in ...Estimation of Optimized Energy and Latency Constraint for Task Allocation in ...
Estimation of Optimized Energy and Latency Constraint for Task Allocation in ...
 
genetic paper
genetic papergenetic paper
genetic paper
 
Congestion Control through Load Balancing Technique for Mobile Networks: A Cl...
Congestion Control through Load Balancing Technique for Mobile Networks: A Cl...Congestion Control through Load Balancing Technique for Mobile Networks: A Cl...
Congestion Control through Load Balancing Technique for Mobile Networks: A Cl...
 
Improved Max-Min Scheduling Algorithm
Improved Max-Min Scheduling AlgorithmImproved Max-Min Scheduling Algorithm
Improved Max-Min Scheduling Algorithm
 
Time Efficient VM Allocation using KD-Tree Approach in Cloud Server Environment
Time Efficient VM Allocation using KD-Tree Approach in Cloud Server EnvironmentTime Efficient VM Allocation using KD-Tree Approach in Cloud Server Environment
Time Efficient VM Allocation using KD-Tree Approach in Cloud Server Environment
 
J41046368
J41046368J41046368
J41046368
 

Destacado

การประสานประโยชน์ระหว่างประเทศ
การประสานประโยชน์ระหว่างประเทศการประสานประโยชน์ระหว่างประเทศ
การประสานประโยชน์ระหว่างประเทศPannaray Kaewmarueang
 
Cirrus insight + Alternative: From Inbox to Quote in 4 Clicks
Cirrus insight + Alternative: From Inbox to Quote in 4 ClicksCirrus insight + Alternative: From Inbox to Quote in 4 Clicks
Cirrus insight + Alternative: From Inbox to Quote in 4 ClicksCirrus Insight
 
Instituto tecnológico superior
Instituto tecnológico superiorInstituto tecnológico superior
Instituto tecnológico superior96PAOLA
 
общие сведения
общие сведенияобщие сведения
общие сведенияpolyansky
 
Actividad integradora 3
Actividad integradora 3 Actividad integradora 3
Actividad integradora 3 LuisChiOjeda1
 
Front End Data Cleaning And Transformation In Standard Printed Form Using Neu...
Front End Data Cleaning And Transformation In Standard Printed Form Using Neu...Front End Data Cleaning And Transformation In Standard Printed Form Using Neu...
Front End Data Cleaning And Transformation In Standard Printed Form Using Neu...ijcsa
 
COUPLER, POWER DIVIDER AND CIRCULATOR IN V-BAND SUBSTRATE INTEGRATED WAVEGUID...
COUPLER, POWER DIVIDER AND CIRCULATOR IN V-BAND SUBSTRATE INTEGRATED WAVEGUID...COUPLER, POWER DIVIDER AND CIRCULATOR IN V-BAND SUBSTRATE INTEGRATED WAVEGUID...
COUPLER, POWER DIVIDER AND CIRCULATOR IN V-BAND SUBSTRATE INTEGRATED WAVEGUID...ijcsa
 
Golf Course Architects
Golf Course Architects Golf Course Architects
Golf Course Architects golfdesign52
 
Bank Fund Management_Liquidity Management Theory - Dev057_FBS-DU
Bank Fund Management_Liquidity Management Theory - Dev057_FBS-DUBank Fund Management_Liquidity Management Theory - Dev057_FBS-DU
Bank Fund Management_Liquidity Management Theory - Dev057_FBS-DUBamdeb Sarker
 
Exercise unit 1 hum3032
Exercise unit 1 hum3032Exercise unit 1 hum3032
Exercise unit 1 hum3032MARA
 
Nutraceutical Study JSUMC Research Day 2012 Abstracts
Nutraceutical Study JSUMC Research Day 2012 AbstractsNutraceutical Study JSUMC Research Day 2012 Abstracts
Nutraceutical Study JSUMC Research Day 2012 AbstractsRavin Rijhsinghani
 

Destacado (18)

การประสานประโยชน์ระหว่างประเทศ
การประสานประโยชน์ระหว่างประเทศการประสานประโยชน์ระหว่างประเทศ
การประสานประโยชน์ระหว่างประเทศ
 
coverletter
coverlettercoverletter
coverletter
 
Cirrus insight + Alternative: From Inbox to Quote in 4 Clicks
Cirrus insight + Alternative: From Inbox to Quote in 4 ClicksCirrus insight + Alternative: From Inbox to Quote in 4 Clicks
Cirrus insight + Alternative: From Inbox to Quote in 4 Clicks
 
Instituto tecnológico superior
Instituto tecnológico superiorInstituto tecnológico superior
Instituto tecnológico superior
 
общие сведения
общие сведенияобщие сведения
общие сведения
 
Actividad integradora 3
Actividad integradora 3 Actividad integradora 3
Actividad integradora 3
 
Front End Data Cleaning And Transformation In Standard Printed Form Using Neu...
Front End Data Cleaning And Transformation In Standard Printed Form Using Neu...Front End Data Cleaning And Transformation In Standard Printed Form Using Neu...
Front End Data Cleaning And Transformation In Standard Printed Form Using Neu...
 
Docsend fundraising research
Docsend fundraising researchDocsend fundraising research
Docsend fundraising research
 
Guía de desarrollos preclínicos
Guía de desarrollos preclínicosGuía de desarrollos preclínicos
Guía de desarrollos preclínicos
 
Artículos de actualización bpm
Artículos de actualización bpmArtículos de actualización bpm
Artículos de actualización bpm
 
COUPLER, POWER DIVIDER AND CIRCULATOR IN V-BAND SUBSTRATE INTEGRATED WAVEGUID...
COUPLER, POWER DIVIDER AND CIRCULATOR IN V-BAND SUBSTRATE INTEGRATED WAVEGUID...COUPLER, POWER DIVIDER AND CIRCULATOR IN V-BAND SUBSTRATE INTEGRATED WAVEGUID...
COUPLER, POWER DIVIDER AND CIRCULATOR IN V-BAND SUBSTRATE INTEGRATED WAVEGUID...
 
Golf Course Architects
Golf Course Architects Golf Course Architects
Golf Course Architects
 
Bank Fund Management_Liquidity Management Theory - Dev057_FBS-DU
Bank Fund Management_Liquidity Management Theory - Dev057_FBS-DUBank Fund Management_Liquidity Management Theory - Dev057_FBS-DU
Bank Fund Management_Liquidity Management Theory - Dev057_FBS-DU
 
Exercise unit 1 hum3032
Exercise unit 1 hum3032Exercise unit 1 hum3032
Exercise unit 1 hum3032
 
Nutraceutical Study JSUMC Research Day 2012 Abstracts
Nutraceutical Study JSUMC Research Day 2012 AbstractsNutraceutical Study JSUMC Research Day 2012 Abstracts
Nutraceutical Study JSUMC Research Day 2012 Abstracts
 
Exposicion ingles OUTCOME II
Exposicion ingles OUTCOME IIExposicion ingles OUTCOME II
Exposicion ingles OUTCOME II
 
Pintura Aislante Termica
Pintura Aislante Termica
Pintura Aislante Termica
Pintura Aislante Termica
 
Coffepdf
CoffepdfCoffepdf
Coffepdf
 

Similar a A Novel Framework and Policies for On-line Block of Cores Allotment for Multiple DAGs on NoC

MULTIPLE DAG APPLICATIONS SCHEDULING ON A CLUSTER OF PROCESSORS
MULTIPLE DAG APPLICATIONS SCHEDULING ON A CLUSTER OF PROCESSORSMULTIPLE DAG APPLICATIONS SCHEDULING ON A CLUSTER OF PROCESSORS
MULTIPLE DAG APPLICATIONS SCHEDULING ON A CLUSTER OF PROCESSORScscpconf
 
A Framework and Methods for Dynamic Scheduling of a Directed Acyclic Graph on...
A Framework and Methods for Dynamic Scheduling of a Directed Acyclic Graph on...A Framework and Methods for Dynamic Scheduling of a Directed Acyclic Graph on...
A Framework and Methods for Dynamic Scheduling of a Directed Acyclic Graph on...IDES Editor
 
A MULTI-OBJECTIVE PERSPECTIVE FOR OPERATOR SCHEDULING USING FINEGRAINED DVS A...
A MULTI-OBJECTIVE PERSPECTIVE FOR OPERATOR SCHEDULING USING FINEGRAINED DVS A...A MULTI-OBJECTIVE PERSPECTIVE FOR OPERATOR SCHEDULING USING FINEGRAINED DVS A...
A MULTI-OBJECTIVE PERSPECTIVE FOR OPERATOR SCHEDULING USING FINEGRAINED DVS A...VLSICS Design
 
An efficient cloudlet scheduling via bin packing in cloud computing
An efficient cloudlet scheduling via bin packing in cloud computingAn efficient cloudlet scheduling via bin packing in cloud computing
An efficient cloudlet scheduling via bin packing in cloud computingIJECEIAES
 
Reconfiguration Strategies for Online Hardware Multitasking in Embedded Systems
Reconfiguration Strategies for Online Hardware Multitasking in Embedded SystemsReconfiguration Strategies for Online Hardware Multitasking in Embedded Systems
Reconfiguration Strategies for Online Hardware Multitasking in Embedded SystemsCSEIJJournal
 
Scheduling in cloud computing
Scheduling in cloud computingScheduling in cloud computing
Scheduling in cloud computingijccsa
 
Accelerating S3D A GPGPU Case Study
Accelerating S3D  A GPGPU Case StudyAccelerating S3D  A GPGPU Case Study
Accelerating S3D A GPGPU Case StudyMartha Brown
 
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDM O...
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDM O...PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDM O...
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDM O...ijgca
 
ENERGY AND LATENCY AWARE APPLICATION MAPPING ALGORITHM & OPTIMIZATION FOR HOM...
ENERGY AND LATENCY AWARE APPLICATION MAPPING ALGORITHM & OPTIMIZATION FOR HOM...ENERGY AND LATENCY AWARE APPLICATION MAPPING ALGORITHM & OPTIMIZATION FOR HOM...
ENERGY AND LATENCY AWARE APPLICATION MAPPING ALGORITHM & OPTIMIZATION FOR HOM...cscpconf
 
Energy and latency aware application
Energy and latency aware applicationEnergy and latency aware application
Energy and latency aware applicationcsandit
 
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...ijassn
 
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...ijassn
 
International Journal of Engineering and Science Invention (IJESI)
International Journal of Engineering and Science Invention (IJESI)International Journal of Engineering and Science Invention (IJESI)
International Journal of Engineering and Science Invention (IJESI)inventionjournals
 
Crdom cell re ordering based domino on-the-fly mapping
Crdom  cell re ordering based domino on-the-fly mappingCrdom  cell re ordering based domino on-the-fly mapping
Crdom cell re ordering based domino on-the-fly mappingVLSICS Design
 
Max Min Fair Scheduling Algorithm using In Grid Scheduling with Load Balancing
Max Min Fair Scheduling Algorithm using In Grid Scheduling with Load Balancing Max Min Fair Scheduling Algorithm using In Grid Scheduling with Load Balancing
Max Min Fair Scheduling Algorithm using In Grid Scheduling with Load Balancing IJORCS
 
(Paper) Task scheduling algorithm for multicore processor system for minimiz...
 (Paper) Task scheduling algorithm for multicore processor system for minimiz... (Paper) Task scheduling algorithm for multicore processor system for minimiz...
(Paper) Task scheduling algorithm for multicore processor system for minimiz...Naoki Shibata
 

Similar a A Novel Framework and Policies for On-line Block of Cores Allotment for Multiple DAGs on NoC (20)

MULTIPLE DAG APPLICATIONS SCHEDULING ON A CLUSTER OF PROCESSORS
MULTIPLE DAG APPLICATIONS SCHEDULING ON A CLUSTER OF PROCESSORSMULTIPLE DAG APPLICATIONS SCHEDULING ON A CLUSTER OF PROCESSORS
MULTIPLE DAG APPLICATIONS SCHEDULING ON A CLUSTER OF PROCESSORS
 
A Framework and Methods for Dynamic Scheduling of a Directed Acyclic Graph on...
A Framework and Methods for Dynamic Scheduling of a Directed Acyclic Graph on...A Framework and Methods for Dynamic Scheduling of a Directed Acyclic Graph on...
A Framework and Methods for Dynamic Scheduling of a Directed Acyclic Graph on...
 
53
5353
53
 
A MULTI-OBJECTIVE PERSPECTIVE FOR OPERATOR SCHEDULING USING FINEGRAINED DVS A...
A MULTI-OBJECTIVE PERSPECTIVE FOR OPERATOR SCHEDULING USING FINEGRAINED DVS A...A MULTI-OBJECTIVE PERSPECTIVE FOR OPERATOR SCHEDULING USING FINEGRAINED DVS A...
A MULTI-OBJECTIVE PERSPECTIVE FOR OPERATOR SCHEDULING USING FINEGRAINED DVS A...
 
N0173696106
N0173696106N0173696106
N0173696106
 
An efficient cloudlet scheduling via bin packing in cloud computing
An efficient cloudlet scheduling via bin packing in cloud computingAn efficient cloudlet scheduling via bin packing in cloud computing
An efficient cloudlet scheduling via bin packing in cloud computing
 
Reconfiguration Strategies for Online Hardware Multitasking in Embedded Systems
Reconfiguration Strategies for Online Hardware Multitasking in Embedded SystemsReconfiguration Strategies for Online Hardware Multitasking in Embedded Systems
Reconfiguration Strategies for Online Hardware Multitasking in Embedded Systems
 
Scheduling in cloud computing
Scheduling in cloud computingScheduling in cloud computing
Scheduling in cloud computing
 
Accelerating S3D A GPGPU Case Study
Accelerating S3D  A GPGPU Case StudyAccelerating S3D  A GPGPU Case Study
Accelerating S3D A GPGPU Case Study
 
IEEE CLOUD \'11
IEEE CLOUD \'11IEEE CLOUD \'11
IEEE CLOUD \'11
 
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDM O...
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDM O...PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDM O...
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDM O...
 
ENERGY AND LATENCY AWARE APPLICATION MAPPING ALGORITHM & OPTIMIZATION FOR HOM...
ENERGY AND LATENCY AWARE APPLICATION MAPPING ALGORITHM & OPTIMIZATION FOR HOM...ENERGY AND LATENCY AWARE APPLICATION MAPPING ALGORITHM & OPTIMIZATION FOR HOM...
ENERGY AND LATENCY AWARE APPLICATION MAPPING ALGORITHM & OPTIMIZATION FOR HOM...
 
Energy and latency aware application
Energy and latency aware applicationEnergy and latency aware application
Energy and latency aware application
 
Resisting skew accumulation
Resisting skew accumulationResisting skew accumulation
Resisting skew accumulation
 
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...
 
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...
 
International Journal of Engineering and Science Invention (IJESI)
International Journal of Engineering and Science Invention (IJESI)International Journal of Engineering and Science Invention (IJESI)
International Journal of Engineering and Science Invention (IJESI)
 
Crdom cell re ordering based domino on-the-fly mapping
Crdom  cell re ordering based domino on-the-fly mappingCrdom  cell re ordering based domino on-the-fly mapping
Crdom cell re ordering based domino on-the-fly mapping
 
Max Min Fair Scheduling Algorithm using In Grid Scheduling with Load Balancing
Max Min Fair Scheduling Algorithm using In Grid Scheduling with Load Balancing Max Min Fair Scheduling Algorithm using In Grid Scheduling with Load Balancing
Max Min Fair Scheduling Algorithm using In Grid Scheduling with Load Balancing
 
(Paper) Task scheduling algorithm for multicore processor system for minimiz...
 (Paper) Task scheduling algorithm for multicore processor system for minimiz... (Paper) Task scheduling algorithm for multicore processor system for minimiz...
(Paper) Task scheduling algorithm for multicore processor system for minimiz...
 

Último

"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr LapshynFwdays
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...Fwdays
 
Training state-of-the-art general text embedding
Training state-of-the-art general text embeddingTraining state-of-the-art general text embedding
Training state-of-the-art general text embeddingZilliz
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostZilliz
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piececharlottematthew16
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfRankYa
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Manik S Magar
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024Lorenzo Miniero
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Commit University
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsRizwan Syed
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 

Último (20)

"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
 
Training state-of-the-art general text embedding
Training state-of-the-art general text embeddingTraining state-of-the-art general text embedding
Training state-of-the-art general text embedding
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptxE-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piece
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdf
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL Certs
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 

A Novel Framework and Policies for On-line Block of Cores Allotment for Multiple DAGs on NoC

  • 1. International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013 A Novel Framework and Policies for On-line Block of Cores Allotment for Multiple DAGs on NoC Uma Boregowda1 and Venugopal C R2 1 2 Malnad College of Engineering, Hassan, India Sri Jayachamarajendra College of Engineering, Mysore, India ABSTRACT Computer industry has widely accepted that future performance increases must largely come from increasing the number of processing cores on a die. This has led to NoC processors. Task scheduling is one of the most challenging problems facing parallel programmers today which is known to be NP-complete. A good principle is space-sharing of cores and to schedule multiple DAGs simultaneously on NoC processor. Hence the need to find optimal number of cores for a DAG for a particular scheduling method and further which region of cores on NoC, to be allotted for a DAG . In this work, a method is proposed to find nearoptimal minimal block of cores for a DAG on a NoC processor. Further, a time efficient framework and three on-line block allotment policies to the submitted DAGs are experimented. The objectives of the policies, is to improve the NoC throughput. The policies are experimented on a simulator and found to deliver better performance than the policies found in literature.. KEYWORDS DAG; task scheduling; NoC 1. INTRODUCTION The advancement in VLSI has now led to the increased number of cores on a single chip. A study at University of California, Berkeley[1] indicates that future chips will have thousands of cores. Traditional bus based systems and cache hierarchies have given way to more flexible and scalable Network-on-chip(NoC) architecture. Massive parallel computing performed on NoC is the future computing, due to the large increased computing demands. With the increasing number of cores, it is becoming further difficult and demanding to exploit the full power of all cores on the chip. The data flow model is gaining popularity as a programming paradigm for multi-core processors and NoC. When the characteristics of an application is fully deterministic, including task's execution time, size of data communicated between tasks, and task dependencies, the application can be represented by a Directed Acyclic Graph (DAG). A DAG G=(V,E), where V = { vi | i=1, ... V} is a set of vertices representing tasks and E={e i,j|i ,j) {1, ... V} * {1, ... V} is a set of edges representing precedence constraints between tasks. DAG is assumed to have a single entry and single exit task. The structure of a general DAG is shown in Fig. 1 with eleven number of taks. The weight of a vertex v ∈ V corresponds to the execution time of the task, which can be measured via benchmarking or can be calculated via a performance model. The weight of an edge ei,j ∈ E, is the communication cost measured as the time taken to transfer data Di,j between DOI:10.5121/ijcsa.2013.3606 55
  • 2. International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013 vertices vi and vj, which is a function of data volume Di,j, the network characteristics, the processors allocated to it and data distribution cost. Fig. 1 A General DAG Communication-to-Computation (CCR) is the ratio of average communication cost to the average computation cost of a DAG. This characterizes the nature of DAG. The objective of scheduling is to map tasks onto cores and order their execution so that task dependencies are satisfied and minimum overall completion time is achieved. Scheduling can be static or dynamic. Static schedule can be generated only when the application behaviour is fully deterministic and this has the advantage of being more efficient and a small overhead during runtime. The full global knowledge of application in the form of DAG will help to generate a better schedule. Makespan is the total time required to complete a DAG. The challenge of scheduling tasks onto cores and also maintaining precedence constraint, with an objective of minimizing overall execution time is proved to be a NP hard problem[2]. Not all applications can effectively utilize all cores on the chip, hence it would be advantageous to multiprogram the NoC and space-partition the cores among applications, being multiprogrammed. The design has now shifted from scheduling single application to multiple applications. Hence the problem of scheduling DAG on NoC is twofold : first, to allot cores for a DAG and second, to schedule tasks of DAG onto the allotted cores. This paper deals with the first issue. The contributions of this paper include (a) Finding near-optimal block of cores for a given DAG for a particular scheduling method (b) To propose and implement a framework and three on-line block of cores allotment policies for the DAGs submitted, with an objective of improving the total time required to complete all DAGs and hence the throughput. 2. RELATED WORK NoC processors are more interested in how many DAGs can be completed over a period of time rather than how quickly an individual DAG gets completed, since the cores are not a very scarce 56
  • 3. International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013 resource on it. Thus the more emphasis is on throughput of DAGs. The scheduling problem has been extensively studied for several decades. Sakellariou et. al.[3] have studied the problem of scheduling multiple DAGs onto a set of heterogeneous resources. The main work is to optimize overall makespan and also to achieve fairness for all DAGs. One approach is to combine several DAGs into one and then use standard methods to schedule the task graph. Another way is to perform task-by-task scheduling over all tasks in all task graphs. Ahmed et. al.[4] team have proposed a game theory based scheduler on multicore processor with the objective of minimizing energy consumption. Cilk[5], Intel Thread Building Blocks(TBB) and OpenMP have proposed several task scheduling techniques over multicore. These systems are optimized specifically for scheduling DAGs on multicore but not for NoC processors. Viktor K Prasanna et. al.[6] have implemented a hierarchical scheduler which could adapt to the input task dependency graph and demonstrated its advantages for NoC architectures. They have divided the threads into groups, each having a manager to perform scheduling at the group level and several workers to perform self-scheduling for the tasks assigned by the manager. A super manager was used to dynamically adjust the group size, so that the scheduler could adapt to the input task dependency graph. Tang et. al.[7] proposed a two-step genetic algorithm and the related software for mapping concurrent applications on a fixed NoC architecture. Murali et. al.[8] presented a methodology to map multiple use-cases onto the NoC architecture, satisfying the constraints of each use-case. The two-step mapping method [9] first finds a region on the NoC for a given application and then maps all tasks of the application onto the region. Several strategies based on the Maximum Empty Rectangle(MER) techniques, like BestSize, BestShape, BestSize BestShape and BestShape BestSize are experimented. For each DAG to be scheduled, the number of cores required is known before hand and DAG is scheduled only if specified cores are available. Hence, a drawback of this work is that, it does not schedule a DAG if available cores is less than the specified cores. But to increase the throughput of DAGs completion and also to maximize core utilization, it is better to schedule a DAG even when the available cores are marginally less than the specified cores. This proposed work tries to schedule a DAG with less than specified cores whenever the specified cores are not available and if it increases the throughput without much affecting that DAG's makespan. It is generally beneficial in terms of throughput, to run more DAGs with less than specified cores, than running few DAGs with full specified cores. 3. PROPOSED WORK 3.1. Finding Near-optimal block of cores The reason behind the idea of allotting cores in the form of block for a DAG are (a) it would be easier and time efficient to allot and free cores for a DAG in the form of block (b) the nodes average distance (NAD) is minimum for a block and this minimizes the communication time between the tasks, hence the completion time of the schedule generated. NAD is the average distance between any two randomly selected nodes on NoC. The block can only be a square or rectangle. To find the near-optimal block of cores, an initial schedule for the DAG is generated using a well known list scheduling method and trying to use all cores on NoC. The scheduling method also takes into account the number of hops required to move data between any two cores on NoC, the amount of data to be moved and the bandwidth of the intermediate channels, hence the generated schedule is more realistic. All the cores are not utilized to the maximum extent in the generated schedule. Thus it would be beneficial to eliminate low utilized cores from the schedule, without much affecting the completion time of the DAG. Since the cores allotted for a 57
  • 4. International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013 DAG must be in the form of a block, the low utilized rows or columns of cores are eliminated. This method is given in the below algorithm. Algorithm Near_Optimal_Block() Input : DAG ‘G’ Output : Number of rows and columns of cores 1. 2. 3. 4. Start with maximum number of rows and column of cores on the given NoC Find the schedule for the given DAG ‘G’, using any well known list scheduling method like- ETF, HLFET[2] etc. Repeat 3.1 Reduce row or column of cores, whichever has least total number of tasks assigned to it. 3.2 Generate the new schedule with revised row and column number 3.3 Find the percentage increase in core utilization and percentage increase in makespan of the new schedule in comparison to previous schedule Until percentage increase in makespan exceeds percenage increase in core utilization end_algorithm The iteration at which, percentage increase in makespan exceeds percentage increase in core utilization, determines the number of rows and columns of cores for the DAG. It is the point where any further reduction of row/column of cores will significantly increase the makespan. Hence it fixes the near-optimal block of cores for a given DAG. 3.2. A Framework for On-line Block allotment A Framework for on-line submission of DAGs, allotment of block of cores for each DAG, scheduling of DAGs and finding its completion time is developed. An efficient method to track the status of cores on NoC and hence to allot cores for a DAG is implemented. The NoC is viewed as a single pool of cores, while previous methods[9] have maintained it as several blocks of free or allotted cores. The disadvantages of maintaining NoC as a collection of free or allotted blocks are (a) the job of splitting the blocks during DAG allotment when no free block with the specified size is available and merging the blocks when a DAG is completed, is quite computationally expensive to be used as an on-line method, (b) there will be few situations where a DAG can not be allotted cores, as the required number of free cores are not available as a single block but as a collection of few adjacent blocks. However the proposed method will allot block when the required number of free cores are available either as a single block or as few adjacent blocks. Fig. 2 shows an instance where two blocks are allotted to DAG1 and DAG2, and another two blocks FB1 and FB2 are free. Using methods[9], a demand for block of cores of size 4*2 or 5*2 can not be satisfied as the required number of cores are not available as a single block. But the required free block is present as 2 adjacent blocks. The allotment of such adjacent blocks needs frequent merging of blocks which is computationally expensive. The proposed method will never fail to allot, whenever required cores are available either as a single free block or as few adjacent free blocks. A mask with a dimension of the desired block size is used to locate the presence of free block on NoC in a very simple manner. 58
  • 5. International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013 Fig. 2 An Instance of Block Allotment for DAGs 4. EXPERIMENT AND RESULTS 4.1. Experimental Methodology A simple simulator using the proposed framework is developed to simulate the arrival, scheduling, execution and completion of DAGs. Poisson distribution is used to simulate the arrival time of DAGs. This software is run on multi2sim[12], a functional simulator to measure total running time. 4.2. A Finding Near-optimal block of cores The proposed method to find the near-optimal minimal block of cores is implemented using the well known list scheduling method ETF[2]. This is experimented for several kinds of benchmark DAGs taken from Task Graphs For Free[10], Standard Task Graph Set [11] and also DAGs of real applications like Strassen matrix multiplication, FFT and complex matrix multiplication. DAGs with CCR values of 0.1, 0.4, 1 and 5 are used in experiments. Fig. 3 shows the plot of percentage increase in makespan and core utilization, as the row or column of cores, allotted to a DAG, is reduced in subsequent iterations. For the DAG considered in Fig. 1, the block of cores dimensions are 5*6, 4*5, 3*5 and 3*4 at iterations 1, 2, 3 and 4 respectively. Thus the optimal block size is 3*5, as the cross-over of plots happens at iteration number 4, when the block dimension is 3*4. This method can be used for any kind of DAG. 59
  • 6. International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013 Fig 3. Finding Minimal Block Size 4.3. A Comparison of Block Allotment Policies To compare the performance of the proposed novel method of tracking the status of cores on NoC, the policies BestSize[9] and BestShape[9] are implemented, by viewing the NoC as a collection of free or alloted blocks. The proposed BestFit chooses the best fit block like the BestSize but views NoC as a single pool of cores. Also, it can allot block of cores in few cases when BestSize[9] fails to allot, because of the manner in which status of cores are maintained. The simulator is run for 4 categories of DAG sets, each with 10, 20, 30 and 40 DAGs respectively. The performance metric used is complete makespan, which is the time to complete all DAGs. For comparison, normalized complete makespan is used, which normalizes the obtained complete makespan with respect to BestShape, as it is the best policy[9] found in literature. The average normalized complete makespan of DAGs within each category is computed, for each block allotment policy. Fig. 4 shows the comparison of average normalized complete makespan for different policies. Proposed BestFit exhibits better performance against the other two policies, for all categories. The improvement in performance is further enhanced with increase in number of DAGs in a set. This happens as the number of times block allotment is successful in proposed method, while it is a failure in previous method is more. But the significant advantage of proposed method is its low execution time, which was evident from the recorded runtime on simulator and thus proving that proposed method is the right candidate for on-line block allotment for DAGs. Fig. 4 Comparing New Method of Tracking Core Status with previous method 60
  • 7. International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013 Fig. 5 shows the performance comparison of all three proposed block allotment policies. Metric used is normalized overall makespan with respect to BestFit policy. DAGs with both low and high CCR, combined together with low and high arrival rates, are considered for each set of DAGs. Each set has 50 DAGs. The set LoC_LoA includes DAG with low CCR and low arrival rate. DAGs with high CCR and high arrival rate are included in the set HiC_HiA. The DAGs with low CCR can better utilize more cores, as the communication cost is low. Thus when allotted less number of cores for a DAG, resultant increase in makespan is slightly more. For the set LoC_LoA, it is evident from Fig. 5 that performance of AvailableCore is better by 10% with reference to BestFit policy. When the arrival rate of DAGs is low, Adoptive policy performance is close to AvailableCore. As the arrival rate is more in LoC_HiA, the adoptive policy performs better by 9% in comparison to AvailableCore. When CCR is high, DAGs can not effectively utilize more cores, hence a decrease in cores allotted would not significantly affect the makespan. Thus for the set HiC_LoA, the performance of AvailableCore is significantly better than BestFit by 18%. When CCR and arrival rate of DAGs is high as in HiC_HiA, both Adoptive and AvailableCore show good performance improvement of around 20%. For the set with all kinds of DAGs randomly chosen, Adoptive and AvailableCore policies perform better by 17% in comparison to BestFit. Fig. 5 Comparison of Block Allotment Policies 5. CONCLUSION A framework for on-line DAGs submission, allotting block of cores, scheduling and completion of DAGs is proposed and implemented on a simulator. Three block allotment policies – BestFit, AvailableCores and Adoptive are compared with the policies found in literature. The simple BestFit policy delivers improvement over the policies found in literature, because of the new proposed method of tracking the status of cores on NoC. The novel methods – AvailableCore and Adoptive delivers better performance than BestFit. Of the proposed policies, Adoptive one is good, as it can adjust the allotment to the arrival rate of DAGs, thus reducing the overall completion time. These experiments have proved that it is beneficial to schedule several DAGs with less than specified cores, than running DAGs with full specified cores. Future Enhancements – The experiment is now simulated using DAGs with dummy tasks. The idea must be experimented on real NoC, with DAGs of real tasks. 61
  • 8. International Journal on Computational Sciences & Applications (IJCSA) Vol.3, No.6, December 2013 REFERENCES [1] Krste Asanovic, Ras Bodik, Bryan C. Catanzaro, Joseph J. Gebis, Parry Husbands, Kurt Keutzer, David A. Patterson, William L. Plishker, John Shalf, Samuel W. Williams, and Katherine A. Yelick. The landscape of parallel computing research: a view from berkeley. (UCB/EECS-2006- 183), December 2006. [2] Y. K. Kwok and I. Ahmad. Static Scheduling Algorithms for allocating directed task graphs to multiprocessors. ACM Computing Surveys, 31(4):406–471, 1999. izem, Aksahya & Ayese, Ozcan (2009) Coomunications & Networks, Network Books, ABC Publishers. [3] H. Zhao and R. Sakellariou. Scheduling Multiple DAGs onto Heterogeneous systems. In IEEE International Symposium on Parallel and Distributed Processing (IPDPS), pages 1– 12, 2006. [4] I. Ahmad, S. Ranka, and S. Khan. Using Game theory for Scheduling tasks on multi-core processors for Simultaneous optimization of performance and energy. In Intl. Sym. on Parallel Dist. Proc., pages 1–6, 2008. [5] R. D. Blumofe, C. F. Joerg, B. C. Kuszmaul, C. E. Leiserson, K. H. Randall, and Y. Zhou. Cilk: An efficient multithreaded runtime system. Technical report, Cambridge, 1996 [6] Yinglong Xia, Viktor K. Prasanna, James Li. Hierarchical Scheduling of DAG Structured Computations on Manycore Processors with Dynamic Thread Grouping, 15th Workshop on Job Scheduling Strategies for Parallel Processing, In conjunction with IPDPS (JSSPP'10), 2010 [7] Tang Lei and Shashi Kumar. A two-step genetic algorithm for mapping task graphs to a network on chip architecture. In DSD ’03: Proceedings of the Euromicro Symposium on Digital Systems Design, page 180, Washington, DC, USA, 2003. IEEE Computer Society [8] Srinivasan Murali, Martijn Coenen, Andrei Radulescu, Kees Goossens, and Giovanni De Micheli. Mapping and Configuration Methods for multi use-case networks on chips. In ASP-DAC ’06: Proceedings of the 2006 Asia and South Pacific Design Automation Conference, pages 146–151, Piscataway, NJ, USA, 2006. IEEE Press [9] Bo Yang, Liang Guang, Xu, T. C. Santti, T. Plosila. Multi-application Multi-step Mapping Method for Many-core Network-on-Chip platforms, Israel, IEEEI, 2010 [10] Task Graphs for Free http://ziyang.eecs.umich.edu/~dickrp/tgff [11] Standard Task Graph Set http;//www.kasahara.elec.waseda.ac.lp/schedule [12] Multi2sim Simulator http://www.multi2sim.org/ Author C.R.Venugopal received his Ph. D from IIT, Bombay. Presently serving as a Professor in Sri Jayachamarajendra College of Engineering, Mysore, India. His main research interests are Cloud computing, High Performance computing, VLSI Technology and File System development. Has authored more than 50 international conference and journal papers. Uma B completed M. Tech in Computer Science and Engineering from IIT, Delhi. Currently working as a Associate Professor in Malnad College of engineering, Hassan, India. Her research interests are Parallel Programming, High Performance Computing and Task scheduling. 62