programming: Definition and Much More from Answers.com
- ️Tue Aug 22 2006
Within the context of information systems, the term programming is understood to mean computer programming, which is the process of writing computer programs. A computer program is a detailed, step-by-step set of instructions that is executed by a computer in order to perform a specific task or solve a specific problem. A computer can perform a wide variety of tasks, including arithmetic calculations, text formatting, and submission of documents to the printer to be printed. However, the computer hardware does not perform these tasks by itself. It needs specific instructions on how to go about performing each specific task. It is these task-specific sets of instructions that are referred to as programs.
Programming Languages
Computer programs are written in a variety of programming languages. These languages fall into two broad categories: low-level programming languages and high-level programming languages. Low-level programming languages are so named because they are closer to machine language than to human language; that is, it is easier for the machine (computer) to understand them than it is for humans. Machine language is made up of a series of 0's and 1's. Each 0 or 1 is known as a bit (short for binary digit). A group of eight bits, known as a byte, represents one character(i.e., a number or a letter). For example, the number 2 is represented as 00000010 and the letter B as 01000010 in the American National Standards Institute (ANSI) code for character representation inside a computer. There are other coding schemes besides the ANSI standard, such as the American Standard Code for Information Interchange (ASCII) and IBM's Extended Binary Coded Decimal Interchange Code (EBCDIC). Each of these standards represents characters in a slightly different way. Such binary representation of characters is the only thing that the computer can directly "understand" and execute.
In the early days of computer programming, programmers wrote their programs directly in machine language. The time-consuming and painstaking nature of this process led to the development of assembly language, which uses alphabetic mnemonics (rather than binary digits) to write programs. For example, an assembly-language instruction to load the number 5 into a computer's accumulator is: LDA 5. This is more readable than a string of 0's and 1's. A special program called an assembler translates assembly-language instructions into machine-language instructions. Assembly language is machine-specific and is used to directly manipulate activity at the hardware level. Therefore, it is still considered low-level.
Further technological advances led to the development of high-level programming languages such as COBOL (COmmon Business Oriented Language), FORTRAN (FORmula TRANslator). BASIC (Beginners' All-purpose Symbolic Instruction Code), PASCAL, PL/1, and C. These languages are described high-level because they are closer to human language than to machine language. In these languages, the number 2 and the letter B are coded in the program exactly as they are written. Similarly, the following is a valid line of program code in some high-level languages: SUM NUM1 NUM2. A special program, known as a compiler or an interpreter (depending on the programming language) translates the high-level program code into machine language before it is executed.
Categories of Programming
There are two main categories of programming, systems programming and applications programming. Systems programs are more likely to be written in low-level programming languages, while applications programs are written almost exclusively in high-level languages.
Systems Programming Systems programming involves writing programs that enable a computer to carry out its basic internal functions as well as some other specialized functions. Examples of systems programs include operating systems, device drivers, and utility programs.
An operating system, which comes as an essential and necessary component of any computer system, is a complex set of programs that coordinates activities inside a computer and ensures the proper and efficient use of all the computer's resources. Among its basic functions are scheduling and running multiple jobs inside the computer, managing storage space, enforcing security (e.g., through password verification), and detecting equipment failure. Through its actions, the operating system enables a user to access the computer's hardware and software components. Examples of operating systems include DOS, Windows 95, Windows 98, Windows NT, Macintosh System 8, UNIX, OS/2, and VAX VMS.
Device drivers are those programs that identify particular devices to a computer and enable the computer to correctly use those devices. For example, a mouse driver program helps a computer identify the mouse attached to it.
Utility programs (or utilities) are programs that perform such specialized tasks as reorganizing data on disks, recovering lost data, recovering from system crashes, and detecting and removing computer viruses.
Applications Programming Applications programming refers to the process of developing programs to be used for specific applications, such as a business application (e.g., computing benefits to be paid to different employee classifications) or an academic application (e.g., determining who qualifies for which scholarship, based on specified eligibility criteria). Programming such applications usually requires the programmer to specify the precise logic that would be required to solve the given problem. There are a number of stages in the applications programming process, including problem statement, algorithm development, program coding, program testing, and program documentation.
Problem statement: The programming process begins with a clear, written statement of the problem to be solved by the computer. The importance of this step cannot be overemphasized. A poorly articulated or poorly understood problem statement will result in the wrong solution being developed for the problem at hand. There should also be a statement of the conditions that would determine when the problem has been solved. All known and relevant facts should also be stated at this stage, as well as any necessary assumptions to be made in the program.
Algorithm development: Once the problem has been clearly stated and all the requirements have been understood, the next step is to develop the program logic necessary for accomplishing the task. An algorithm is defined as a logical sequence of steps that must be performed in order to accomplish a given task. There are some tools available to help the programmer develop the algorithm for a given problem. The two best-known and most widely used ones are the flowchart and pseudocode. Both of these are language-independent, focusing primarily on logic flow rather than the syntax of any particular language. A flowchart uses standard flowcharting symbols to visually represent the flow of program logic. Pseudocode, on the other hand, often looks like actual program code, but it is not, since it does not follow any particular language's syntax. The term pseudocode means "false code." Unlike flowcharting, in which standard, universally accepted symbols are used, there are no set standards for writing pseudocode. Figures 1 and 2 illustrate the use of a flowchart and pseudocode, respectively, to depict the logic needed to add up all the even numbers between 2 and 100, inclusive, and print the resulting total.
Program coding: When the programmer is satisfied with the efficacy of the logic developed in the preceding step, it is time to convert that logic (in either flowchart or pseudocode form) to the specific syntax of the programming language that will be used. At this stage, the programmer adheres strictly to all of the syntax requirements for coding the logic as well as other aspects of the program.
Program testing: The coded program is next checked for errors. At least two types of programming errors must be checked for, namely, syntax errors and logic errors. The presence of syntax errors indicates that some syntactic rule(s) of the programming language has (have) been violated. Syntax errors are detected when the program is compiled (the compiler identifies all such errors within the program). They must be corrected before the program can be successfully executed. Even when all the syntax errors have been corrected, there is the possibility of logic errors. Logic errors arise when the desired logic is incorrectly specified in the program, thereby resulting in an erroneous output. An example is a program that makes students with failing grades eligible for academic scholarships when, in fact, they should not be. In computer terminology, any error in a program—syntax or logic—is known as a bug. The process of correcting these errors is known as debugging.
Program documentation: The programming process is complete when the program has been fully documented. The documentation can be either incorporated into the body of the program itself (in-line documentation) or it can be a completely separate document (external documentation). Frequently, it is both. Good documentation typically includes the following: a statement of the program's objective(s); descriptions of any input or output records or files needed to run the program; a complete definition of all data names used; and an explanation of the underlying logic, preferably with an accompanying flowchart. Pro gram documentation greatly facilitates program maintenance, which is the periodic modification to, or update of, the program in order to keep it current. This is especially important if the person maintaining the program is not the same one who wrote it.
Applications Programs on the Market
There is a wide array of programs and compilers on the market today, in the form of various software packages. Compilers for all the major programming languages mentioned above are available on virtually all computing platforms. Most of these commercial packages, such as Visual Basic, Visual C, and Micro focus COBOL, have "visual" front-ends to their programming environments, which makes it easy for programmers to design user-friendly programs for their clients.
Bibliography
Berlioux, Pierre, and Bizard, Philippe. (1986). Algorithms: The Construction. Proof, and Analysis of Programs, trans. Annwyl Williams. New York: Wiley.
Dodd, Kenneth N. (1969). Computer Programming and Languages. New York: Plenum.
Flores, Ivan. (1971). Assemblers and BAL. Englewood Cliffs, NJ: Prentice-Hall.
Iliffe, J.K. (1972). Basic Machine Principles, 2d ed. New York: American Elsevier.
Knuth, Donald E. (1997). The Art of Computer Programming, 3d ed. Reading, MA: Addison-Wesley.
McCracken, Daniel, and Golden, Donald. (1990). Simplified Structured COBOL with Microsoft/Microfocus COBOL. New York: Wiley.
Parsons, June J., and Oja, Dan. (1998). Computer Concepts—Comprehensive, 3d ed. Cambridge, MA: Course Technology.
Washburn, Dale W. (1970). Computer Programming: A Total Language Approach. New York: Holt, Reinhart & Winston.
[Article by: THEOPHILUS B. A. ADDO]
Blogs: Related blogs on: programming
- Luteo.us Eddy Luten's Blog on General and Topical Programming
Add your blog to the Answers Directory.
Wikipedia: computer programming
Computer programming (often shortened to programming or coding) is the process of writing, testing, and maintaining the source code of computer programs. The source code is written in a programming language. This code may be a modification of an existing source or something completely new, the purpose being to create a program that exhibits the desired behavior (customization). The process of writing source codes requires expertise in many different subjects, including knowledge of the application domain, specialized algorithms, and formal logic.
Within software engineering, programming (the implementation) is regarded as one phase in a software development process.
In some specialist applications or extreme situations a program may be written or modified (known as patching) by directly storing the numeric values of the machine code instructions to be executed into memory.
There is an ongoing debate on the extent to which the writing of programs is an art, a craft or an engineering discipline.[1] Good programming is generally considered to be the measured application of all three: expert knowledge informing an elegant, efficient, and maintainable software solution (the criteria for "efficient" and "maintainable" vary considerably). The discipline differs from many other technical professions in that programmers generally do not need to be licensed or pass any standardized (or governmentally regulated) certification tests in order to call themselves "programmers" or even "software engineers".
Another ongoing debate is the extent to which the programming language used in writing programs affects the form that the final program takes. This debate is analogous to that surrounding the Sapir-Whorf hypothesis in linguistics.[citation needed]
Programmers
- See Computer programmer to learn more about the process of computer programming.
Computer programmers are those who write computer software. Their job usually involves:
- Requirements analysis
- Specification
- Software architecture
- Coding
- Compilation
- Software testing
- Documentation
- Integration
- Maintenance
Programming languages
Different programming languages support different styles of programming (called programming paradigms). The choice of language used is subject to many considerations, such as company policy, suitability to task, availability of third-party packages, or individual preference. Ideally, the programming language best suited for the task at hand will be selected. Trade-offs from this ideal involve finding enough programmers who know the language to build a team, the availability of compilers for that language, and the efficiency with which programs written in a given language execute.
Allen Downey, in his book "How To Think Like A Computer Scientist" [1]wrote:
The details look different in different languages, but a few basic instructions appear in just about every language: input: Get data from the keyboard, a file, or some other device. output: Display data on the screen or send data to a file or other device. math: Perform basic mathematical operations like addition and multiplication. conditional execution: Check for certain conditions and execute the appropriate sequence of statements. repetition: Perform some action repeatedly, usually with some variation.
History of programming
The earliest programmable machine (that is a machine whose behavior can be controlled by changes to a "program") was Al-Jazari's programmable humanoid robot in 1206. Al-Jazari's robot was originally a boat with four automatic musicians that floated on a lake to entertain guests at royal drinking parties. His mechanism had a programmable drum machine with pegs (cams) that bump into little levers that operate the percussion. The drummer could be made to play different rhythms and different drum patterns by moving the pegs to different locations.[2]
The Jacquard Loom, developed in 1801, is often quoted as a source of prior art. The machine used a series of pasteboard cards with holes punched in them. The hole pattern represented the pattern that the loom had to follow in weaving cloth. The loom could produce entirely different weaves using different sets of cards. The use of punched cards was also adopted by Charles Babbage around 1830, to control his Analytical Engine.
This innovation was later refined by Herman Hollerith who, in 1896 founded the Tabulating Machine Company (which became IBM). He invented the Hollerith punched card, the card reader, and the key punch machine. These inventions were the foundation of the modern information processing industry. The addition of a plug-board to his 1906 Type I Tabulator allowed it to do different jobs without having to be rebuilt (the first step toward programming). By the late 1940s there were a variety of plug-board programmable machines, called unit record equipment, to perform data processing tasks (card reading). The early computers were also programmed using plug-boards.
The invention of the Von Neumann architecture allowed programs to be stored in computer memory. Early programs had to be painstakingly crafted using the instructions of the particular machine, often in binary notation. Every model of computer would be likely to need different instructions to do the same task. Later assembly languages were developed that let the programmer specify each instruction in a text format, entering abbreviations for each operation code instead of a number and specifying addresses in symbolic form (e.g. ADD X, TOTAL). In 1954 Fortran, the first higher level programming language, was invented. This allowed programmers to specify calculations by entering a formula directly (e.g. Y = X*2 + 5*X + 9). The program text, or source, was converted into machine instructions using a special program called a compiler. Many other languages were developed, including ones for commercial programming, such as COBOL. Programs were mostly still entered using punch cards or paper tape. (See computer programming in the punch card era). By the late-60s, data storage devices and computer terminals became inexpensive enough so programs could be created by typing directly into the computers. Text editors were developed that allowed changes and corrections to be made much more easily than with punch cards.
As time has progressed computers have made giant leaps in the area of processing power. This has brought about newer programming languages that are more abstracted from the underlying hardware. Although these more abstracted languages require additional overhead, in most cases the huge increase in speed of modern computers has brought about little performance decrease compared to earlier counterparts. The benefits of these more abstracted languages is that they allow both an easier learning curve for people less familiar with the older lower-level programming languages, and they also allow a more experienced programmer to develop simple applications quickly. Despite these benefits, large complicated programs, and programs that are more dependent on speed still require the faster and relatively lower-level languages with todays hardware. (The same concerns were raised about the original Fortran language.)
Throughout the second half of the twentieth century, programming was an attractive career in most developed countries. Some forms of programming have been increasingly subject to offshore outsourcing (importing software and services from other countries, usually at a lower wage), making programming career decisions in developed countries more complicated, while increasing economic opportunities in less developed areas. It is unclear how far this trend will continue and how deeply it will impact programmer wages and opportunities. Despite the "outsourcing trend" it can be argued that some of the richest persons on the globe are programmers by profession. Examples: Larry Page and Sergey Brin (Google), Steve Wozniak (Apple Inc.), Hasso Plattner (SAP) and so on. Programming is clearly a leading-edge craftsmanship that continues to reward its practitioners not only in developed countries but also in rapidly developing countries like India.
Modern programming
Quality requirements
Whatever be the approach to the software development, the program must finally satisfy some fundamental properties; bearing them in mind while programming reduces the costs in terms of time and/or money due to debugging, further development and user support. Although quality programming can be achieved in a number of ways, following five properties are among the most relevant:
- Efficiency: it is referred to the system resource consumption (computer processor, memory, slow devices, networks and to some extent even user interaction) which must be the lowest possible.
- Reliability: the results of the program must be correct, which not only implies a correct code implementation but also reduction of error propagation (e.g. resulting from data conversion) and prevention of typical errors (overflow, underflow or zero division).
- Robustness: a program must anticipate situations of data type conflict and all other incompatibilities which result in run time errors and stop of the program. The focus of this aspect is the interaction with the user and the handling of error messages.
- Portability: it should work as it is in any software and hardware environment, or at least without relevant reprogramming.
- Readability: the purpose of the main program and of each subroutine must be clearly defined with appropriate comments and self explanatory chose of symbolic names (constants, variables, function names, classes, methods, ...).
Algorithmic Complexity
The academic field and engineering practice of computer programming are largely concerned with discovering and implementing the most efficient algorithms for a given class of problem. For this purpose, algorithms are classified into orders using so-called Big O notation, O(n), which expresses resource use, such as execution time and memory consumption, in terms of the size of an input. Expert programmers are familiar with a variety of well-established algorithms and their respective complexities and use this knowledge to choose algorithms that are best suited to their circumstances.
Research in computer programming includes investigation into the unsolved proposition that P, the class of algorithms which can be deterministically solved in polynomial time with respect to an input, is not equal to NP, the class of algorithms for which no polynomial-time solutions are known. Work has shown that many NP algorithms can be transformed, in polynomial time, into others, such as the Travelling salesman problem, thus establishing a large class of "hard" problems which are for the purposes of analysis, equivalent.
Methodologies
The first step in every software development project should be requirements analysis, followed by modeling, implementation, and failure elimination (debugging). There exist a lot of differing approaches for each of those tasks. One approach popular for requirements analysis is Use Case analysis.
Popular modeling techniques include Object-Oriented Analysis and Design (OOAD) and Model-Driven Architecture (MDA). The Unified Modeling Language (UML) is a notation used for both OOAD and MDA.
A similar technique used for database design is Entity-Relationship Modeling (ER Modeling).
Implementation techniques include imperative languages (object-oriented or procedural), functional languages, and logic languages.
Debugging is most often done with IDEs like Visual Studio, NetBeans, and Eclipse. Separate debuggers like gdb are also used.
Measuring language usage
It is very difficult to determine what are the most popular of modern programming languages. Some languages are very popular for particular kinds of applications (e.g., COBOL is still strong in the corporate data center, often on large mainframes, FORTRAN in engineering applications, and C in embedded applications), while some languages are regularly used to write many different kinds of applications.
Methods of measuring language popularity include: counting the number of job advertisements that mention the language[3], the number of books teaching the language that are sold (this overestimates the importance of newer languages), and estimates of the number of existing lines of code written in the language (this underestimates the number of users of business languages such as COBOL).
Debugging
Debugging is a very important task for every programmer, because an erroneous program is often useless. Languages like C++ and Assembler are very challenging even to expert programmers because of failure modes like buffer overruns, bad pointers or uninitialized memory. A buffer overrun can damage adjacent memory regions and cause a failure in a totally different program line. Because of those memory issues tools like Valgrind, Purify or Boundschecker are virtually a necessity for modern software development in the C++ language. Languages such as Java, C#, PHP and Python protect the programmer from most of these runtime failure modes, but this may come at the price of a dramatically lower execution speed of the resulting program. This is acceptable for applications where execution speed is determined by other considerations such as database access or file I/O. The exact cost will depend upon specific implementation details. Modern Java virtual machines and .NET Common Language Runtime, for example, use a variety of sophisticated optimizations, including runtime conversion of interpreted instructions to native machine code.
See also
References
- ^ Paul Graham (2003). "Hackers and Painters". Retrieved on 2006-08-22.
- ^ A 13th Century Programmable Robot. University of Sheffield.
- ^ Survey of Job advertisements mentioning a given language>
External links
This entry is from Wikipedia, the leading user-contributed encyclopedia. It may not have been reviewed by professional editors (see full disclaimer)