Computer processing, Basic Computer Science

Assignment Help:

Computer Processing:

Most of the earliest computer memories have been based on physical elements which can exist in just one of the two states (on or off): such an element corresponds to one bit of information. Binary symbols are used because electronic devices can store and process them rapidly and cheaply. The physical devices used have been changed greatly,, but the principles of information representation mid manipulation within digital computers have remained essentially the same. The size of a computer can be described by the number of bits (binary digits) which its memory contains, but for most purposes larger units of storage are used to characterise machines. Generally, these are denoted by a byte which normally consists of 8 bits, and is sufficient to represent one character, and a word which can be of length -8, 16, 32 bits, is the smallest unit of storage to which most of the computer's instructions can be applied. Memories are usually described as being of a size measured in Kilobytes (K bytes) or Megabytes (M bytes). The sets or patterns of bits which make up bytes and words are used to represent all information within a computer, whether it is program instructions or data items.

The exact method of representation will vary from one machine to another, particularly with respect to instructions formats. An 8 bit byte can accommodate 256 different bit patterns and so it is sufficient to allow for most of the characters which could be required to be printed, e.g., upper and lower case letters A-Z, together with the numbers 0-9 and also a range of punctuation symbols, with allowance for non-printing characters such as end-of-line. The set of bit patterns corresponding to a set of characters is called a character code. Standard codes are ASCII  and EBCDIC. Numbers are normally represented by one or more computer words. Incase of integers, the set of 8,16, or 32 bits (as per the computer's word length) is treated as a binary integer (i.e., encoded in the notation of binary arithmetic). A real number is represented by dividing a computer word into two components: a simple decimal number (i.e., the mantissa portion) together with the power to which it must be raised (i.e., the exponent).  


Related Discussions:- Computer processing

Statistics, Scores from a statistics exam are reported as deviation scores....

Scores from a statistics exam are reported as deviation scores. Which of the following deviation scores indicates a higher position in the class distribution?

ExtremeWhether.html, How to crete a web page name esp.html that allows user...

How to crete a web page name esp.html that allows users to conduct an ESP test

The skill of Assembly Language, The skill of Assembly Language Programming ...

The skill of Assembly Language Programming (AoA), Randy Hyde''s acclaimed text on assembly language programming, is the most-frequently recommended book on 80x86 assembly language

Explain application and system software, Question 1 Explanation of impact ...

Question 1 Explanation of impact of Information Technology on governments Question 2 Explain application and system software Question 3 Briefly explain real time a

Encoder, What is an encoder? Draw the logic circuit of Decimal to BCD encod...

What is an encoder? Draw the logic circuit of Decimal to BCD encoder and explain its working.?

Graphical user interface and arrays, In this program, you are going to writ...

In this program, you are going to write a program for playing the game of Hangman. In this game, the computer will pick a word and display a sequence of blanks to be filled in; one

Architecture, explain how modern computers conform to john von nuemann''s a...

explain how modern computers conform to john von nuemann''s architeture?

Programs - programming language, Programs - programming language: Prog...

Programs - programming language: Programs to implement algorithms on the computer must be written in a language that the computer can understand. It is fruitful, therefore, to

Describe Priority Scheduling algorithm of operating system?, Priority Sched...

Priority Scheduling • The fundamental idea is straightforward: each process is assigned a priority, and priority is permitted to run. Equal-Priority processes are scheduled in FCF

Write Your Message!

Captcha
Free Assignment Quote

Assured A++ Grade

Get guaranteed satisfaction & time on delivery in every assignment order you paid with us! We ensure premium quality solution document along with free turntin report!

All rights reserved! Copyrights ©2019-2020 ExpertsMind IT Educational Pvt Ltd