Explain the architecting for threads, Operating System

Assignment Help:

Explain the Architecting For Threads

When available, threads are an integral part of any multitasking server application program. It is important that the operating system provide an efficient mechanism for thread creation, execution, and manipulation. Ii is this increased efficiency and decreased overhead that makes the threads an invaluable resource. Threads may be widely used to make the server application more concurrent. Thread may be widely used to make the server application more concurrent. Introducing more executing agents working on behalf of the client application component increases the performance. For the developer, there is an even finer level of control over the application. In a multithreads system, each application program may have more than one thread executing.

Multithreading adds power to the developer by allowing the programmer to dispatch multiple (> 1) units of execution. The application can execute more quickly and efficiently because each thread is working concurrently with the others. In may example of multitasking, I showed how the operating system increased efficiency by running multiple applications at the same time. Application developers can add that power to their server applications by starting multiple threads. With each thread executing portions of code concurrently, application throughput is increased. Architecting code to fully utilize threads will maximize the efficiency by using the available CPU cycles. Server application developers should take this into consideration for optimizing their code on a given platform.

A common mechanism for architecting server applications using threads is the Worker Model discussed in the next section. It should be noted that resource requirements in the Worker Model may be relatively high. For example, stack allocation of 8 K FOR 250 threads alone would equal 2MB. Other designs, modeled after one thread per client, are acceptable as well. It is possible to define pools of available threads to execute application requests as they are received.


Related Discussions:- Explain the architecting for threads

Explain rehashing to handle collision, Explain Rehashing to Handle Collisio...

Explain Rehashing to Handle Collision Rehashing:  Re-hashing schemes make use of a second hashing operation while there is a collision. If there is an additional collision, we

Custom memory allocation, Some people write custom memory allocators to mee...

Some people write custom memory allocators to meet their speci?c needs. Although this is not needed for most of the applications, it is also not uncommon. The goal, of course, is t

What is segmentation?, What is segmentation? In a segmented memory mana...

What is segmentation? In a segmented memory management scheme a programmer views a program to be collection of program components called as segments. Every segment is a logical

What is the resident set and working set of a process, What is the resident...

What is the resident set and working set of a process? Resident set is that portion of the process image that is actually in real-memory at a certain instant. Working set is th

Define process swapping , Swapping : Whole process is moved from the swap...

Swapping : Whole process is moved from the swap machine to the main memory for execution. Process size must be equal or less than to the used main memory. It is easier to exe

Multiprogramming, explain how response time are reduced by allowing time sh...

explain how response time are reduced by allowing time sharing on top of multiprogramming

What are turnaround time and response time, What are turnaround time and re...

What are turnaround time and response time? Turnaround time is the interval among the submission of a job and its completion. Response time is the interval among submission

Macro, What is macro expansion

What is macro expansion

Explain components of conflict phase of dispatch latency, Explain Component...

Explain Components of conflict phase of dispatch latency The conflict phase of dispatch latency has two components 1. Preemption of any process running in the kernel. 2.

Linux, what is I/O management Linux

what is I/O management Linux

Write Your Message!

Captcha
Free Assignment Quote

Assured A++ Grade

Get guaranteed satisfaction & time on delivery in every assignment order you paid with us! We ensure premium quality solution document along with free turntin report!

All rights reserved! Copyrights ©2019-2020 ExpertsMind IT Educational Pvt Ltd