Basic concept of data parallelism, Computer Engineering

Basic Concept of Data Parallelism

Thinking the condition where the same problem of submission of „electricity bill? is Handled as follows:

Again, three are counters. However, now each counter handles all the actions of a resident in value of submission of his/her bill. Again, we considering that time compulsory to submit

One bill form is the same as previous, i.e., 5+300+5=310 sec.

We suppose all the counters operate concurrently and each peoples at a counter takes 310 seconds to development one bill. Then, time taken to development all the 10,000 bills will be 310*(9999/3)+310*1sec

This time is relatively much less as compared to time taken in the previous situations, viz.. 3000000 sec and 3100000 sec respectively.

The condition discussed here is the idea of data parallelism. In data parallelism, the entire set of data is manipulated into many blocks and operations on the blocks are applied comparable. As is obvious from this example, data parallelism is much faster as compared to previous situations. Here, no synchronisation is necessary between processers (or counters). It is more liberal of faults. The working of one person didn't affect the other person. There is no message required between processors. Thus, interprocessor message is less. Data parallelism has certain drawbacks. These are as given:

i)The action to be performed by every processor is predefined i.e., load of assignment is static.

ii) It should be probable to split the input task into mutually exclusive tasks. In the given instance, space would be crucial counters. This requires multiple hardware which may be expensive.

The opinion of speedup achieved by using the on top of type of parallel processing is as given:

We assume the number of jobs = m

Let assume the time to do a job = p

If each job is separated into k tasks,

Assuming task is ideally separable into activities, as mentioned over then, Time taken  to complete one task = p/k

Time taken to finish n jobs without parallel processing = n.p

Time to finish  n jobs with parallel processing = n*p/k

                                   1957_Basic Concept of Data Parallelism.png

Posted Date: 3/1/2013 3:20:51 AM | Location : United States







Related Discussions:- Basic concept of data parallelism, Assignment Help, Ask Question on Basic concept of data parallelism, Get Answer, Expert's Help, Basic concept of data parallelism Discussions

Write discussion on Basic concept of data parallelism
Your posts are moderated
Related Questions
Why is Cloud Computing important? There are many implication of cloud technology, for both developers and end users. For developers, cloud computing gives increased amounts of

develop system engineering hirarcy for public health care?

What is difference between collaboration on task and event management? Web-based task management application let the user handle the multiple pieces and parts of large projects

What is write-back or copy-back protocol? For a write operation using this protocol during write-hit: the method is to update only the cache and to mark it as updated with an a

When a database developer uses Hybrid OLAP it seperates the data among relational and specialized storage. In some certain modifications a HOLAP database may store vast amounts of

Q. What are the basic components of an expert system? ANSWER: There are three components of components: Information, people, and IT components. Information kinds are domain exp

Explain the significance of different fields of an instruction An instruction is a command given to a computer to perform a particular operation on some given data and the form

Layout pages, can describe sections, which can then be overridden by particular views making use of the layout. Major and overriding sections is optional.

What is a flip-flop? What is the difference between a latch and a flip-flop? List out the application of flip-flop. Ans. Flip-Flop: It is a fundamental memory element utilize

Q. Define the Thread libraries? The most distinctive representatives of shared memory programming models are thread libraries present in most of modern operating systems. Illus