kontera

Friday, November 27, 2009

computer

3GLs - 4GLs - Active X - Ada - APL - Application Development Environment - Application Development Tools (desktop databases) - Applications Integration - Assemblers - Basic - Borland - CASE - C/C++ - Clarion - CA-Clipper - Client Server - COBOL - CORBA/ORB - Compilers - Cross-Platform Conversion - Database Administration - Database Connectivity - Database Server - Data Conversion - DB2 - Delphi - Desktop Databases - Editors - Eiffel - Fortran - Informix - Ingres - Java - JavaScript - JDBC - Lisp - M ( Mumps ) - Memory Management - Middleware - Modelling Tools - MS Access - mySQL - Object Orientation - Object Oriented Databases - Object Oriented Programming - ODBC - Oracle - Pascal - Perl - PHP - PowerBuilder - Programming Tools - Progress - Project Management - Rapid Applications Development (RAD) - Relational Database Management Systems (RDBMS) - Software Audit - Software Components - Software Development Management - Software Engineering - Software Testing - SQL Server - Sybase - Unified Modelling Language (UML) - Visual Basic - Visual dBase - Visual FoxPro - Visual Studio - Xbas

computer

Computer science education
Some universities teach computer science as a theoretical study of computation and algorithmic reasoning. These programs often feature the
theory of computation, analysis of algorithms, formal methods, concurrency theory, databases, computer graphics and systems analysis, among others. They typically also teach computer programming, but treat it as a vessel for the support of other fields of computer science rather than a central focus of high-level study.
Other colleges and universities, as well as
secondary schools and vocational programs that teach computer science, emphasize the practice of advanced programming rather than the theory of algorithms and computation in their computer science curricula. Such curricula tend to focus on those skills that are important to workers entering the software industry. The practical aspects of computer programming are often referred to as software engineering. However, there is a lot of disagreement over the meaning of the term, and whether or not it is the same thing as programming.

computer

Computer science is considered by some to have a much closer relationship with mathematics than many scientific disciplines, with some observers saying that computing is a mathematical science.[4] Early computer science was strongly influenced by the work of mathematicians such as Kurt Gödel and Alan Turing, and there continues to be a useful interchange of ideas between the two fields in areas such as mathematical logic, category theory, domain theory, and algebra.
The relationship between computer science and
software engineering is a contentious issue, which is further muddied by disputes over what the term "software engineering" means, and how computer science is defined. David Parnas, taking a cue from the relationship between other engineering and science disciplines, has claimed that the principal focus of computer science is studying the properties of computation in general, while the principal focus of software engineering is the design of specific computations to achieve practical goals, making the two separate but complementary disciplines.[24]
The academic, political, and funding aspects of computer science tend to depend on whether a department formed with a mathematical emphasis or with an engineering emphasis. Computer science departments with a mathematics emphasis and with a numerical orientation consider alignment

computer

Relationship with other fields
Despite its name, a significant amount of computer science does not involve the study of computers themselves. Because of this, several alternative names have been proposed. Certain departments of major universities prefer the term computing science, to emphasize precisely that difference. Danish scientist
Peter Naur suggested the term datalogy, to reflect the fact that the scientific discipline revolves around data and data treatment, while not necessarily involving computers. The first scientific institution to use the term was the Department of Datalogy at the University of Copenhagen, founded in 1969, with Peter Naur being the first professor in datalogy. The term is used mainly in the Scandinavian countries. Also, in the early days of computing, a number of terms for the practitioners of the field of computing were suggested in the Communications of the ACM – turingineer, turologist, flow-charts-man, applied meta-mathematician, and applied epistemologist.[21] Three months later in the same journal, comptologist was suggested, followed next year by hypologist.[22] The term computics has also been suggested.[23] Informatik was a term used in Europe with more frequency.
The renowned computer scientist
Edsger Dijkstra stated, "Computer science is no more about computers than astronomy is about telescopes." The design and deployment of computers and computer systems is generally considered the province of disciplines other than computer science. For example, the study of computer hardware is usually considered part of computer engineering, while the study of commercial computer systems and their deployment is often called information technology or information systems. However, there has been much cross-fertilization of ideas between the various computer-related disciplines. Computer science research has also often crossed into other disciplines, such as philosophy, cognitive science, economics, mathematics, physics, and linguistics.

computer

Theory of computation
The study of the
theory of computation is focused on answering fundamental questions about what can be computed and what amount of resources are required to perform those computations. In an effort to answer the first question, computability theory examines which computational problems are solvable on various theoretical models of computation. The second question is addressed by computational complexity theory, which studies the time and space costs associated with different approaches to solving a computational problem.
The famous "
P=NP?" problem, one of the Millennium Prize Problems,[20] is an open problem in the theory of computation.

computer

Fields of computer science
As a discipline, computer science spans a range of topics from theoretical studies of algorithms and the limits of computation to the practical issues of implementing computing systems in hardware and software.
[17][18] The Computer Sciences Accreditation Board (CSAB) [19] – which is made up of representatives of the Association for Computing Machinery (ACM), the Institute of Electrical and Electronics Engineers Computer Society, and the Association for Information Systems – identifies four areas that it considers crucial to the discipline of computer science: theory of computation, algorithms and data structures, programming methodology and languages, and computer elements and architecture. In addition to these four areas, CSAB also identifies fields such as software engineering, artificial intelligence, computer networking and communication, database systems, parallel computation, distributed computation, computer-human interaction, computer graphics, operating systems, and numerical and symbolic computation as being important areas of computer science.[17]

computer

Major achievements
The start of the "digital revolution," which includes the current Information Age and the Internet.[13]
A formal definition of
computation and computability, and proof that there are computationally unsolvable and intractable problems.[14]
The concept of a
programming language, a tool for the precise expression of methodological information at various levels of abstraction.[15]
In
cryptography, breaking the Enigma machine was an important factor contributing to the Allied victory in World War II.[12]
Scientific computing enabled advanced study of the mind, and mapping the human genome became possible with Human Genome Project.[13] Distributed computing projects such as Folding@home explore protein folding.
Algorithmic trading has increased the efficiency and liquidity of financial markets by using artificial intelligence, machine learning, and other statistical and numerical techniques on a large scale.[16]

computer

history
The early foundations of what would become computer science predate the invention of the modern digital computer. Machines for calculating fixed numerical tasks, such as the abacus, have existed since antiquity. Wilhelm Schickard built the first mechanical calculator in 1623.[6] Charles Babbage designed a difference engine in Victorian times[7] helped by Ada Lovelace.[8] Around 1900, punch-card machines[9] were introduced. However, all of these machines were constrained to perform a single task, or at best some subset of all possible tasks.
During the 1940s, as newer and more powerful computing machines were developed, the term computer came to refer to the machines rather than their human predecessors. As it became clear that computers could be used for more than just mathematical calculations, the field of computer science broadened to study
computation in general. Computer science began to be established as a distinct academic discipline in the 1950s and early 1960s, with the creation of the first computer science departments and degree programs.[4][10] Since practical computers became available, many applications of computing have become distinct areas of study in their own right.
Although many initially believed it impossible that computers themselves could actually be a scientific field of study, in the late fifties it gradually became accepted among the greater academic population.
[11] It is the now well-known IBM brand that formed part of the computer science revolution during this time. IBM (short for International Business Machines) released the IBM 704 and later the IBM 709 computers, which were widely used during the exploration period of such devices. "Still, working with the IBM [computer] was frustrating...if you had misplaced as much as one letter in one instruction, the program would crash, and you would have to start the whole process over again".[11] During the late 1950s, the computer science discipline was very much in its developmental stages, and such issues were commonplace.
Time has seen significant improvements in the usability and effectiveness of computer science technology. Modern society has seen a significant shift from computers being used solely by experts or professionals to a more widespread user base.

computer

Computer science
From Wikipedia, the free encyclopedia
Jump to:
navigation, search
Computer science or computing science (CS) is the study of the theoretical foundations of information and computation, and of practical techniques for their implementation and application in computer systems.[1][2][3] It is frequently described as the systematic study of algorithmic processes that create, describe and transform information. According to Peter J. Denning, the fundamental question underlying computer science is, "What can be (efficiently) automated?"[4] Computer science has many sub-fields; some, such as computer graphics, emphasize the computation of specific results, while others, such as computational complexity theory, study the properties of computational problems. Still others focus on the challenges in implementing computations. For example, programming language theory studies approaches to describing computations, while computer programming applies specific programming languages to solve specific computational problems, and human-computer interaction focuses on the challenges in making computers and computations useful, usable, and universally accessible to people.
The general public sometimes confuses computer science with vocational areas that deal with computers (such as
information technology), or think that it relates to their own experience of computers, which typically involves activities such as gaming, web-browsing, and word-processing. However, the focus of computer science is more on understanding the properties of the programs used to implement software such as games and web-browsers, and using that understanding to create new programs or improve existing ones.[5]

computer

Professions and organizations
As the use of computers has spread throughout society, there are an increasing number of careers involving computers.
Computer-related professions
Hardware-related
Electrical engineering, Electronic engineering, Computer engineering, Telecommunications engineering, Optical engineering, Nanoengineering
Software-related
Computer science, Desktop publishing, Human–computer interaction, Information technology, Computational science, Software engineering, Video game industry, Web design
The need for computers to work well together and to be able to exchange information has spawned the need for many standards organizations, clubs and societies of both a formal and informal nature.
Organizations
Standards groups
ANSI, IEC, IEEE, IETF, ISO, W3C
Professional Societies
ACM, ACM Special Interest Groups, IET, IFIP, BCS
Free/Open source software groups
Free Software Foundation, Mozilla Foundation, Apache Software Foundation

computer

Further topics
Hardware
Main article:
Personal computer hardware
The term hardware covers all of those parts of a computer that are tangible objects. Circuits, displays, power supplies, cables, keyboards, printers and mice are all hardware.
History of computing hardware
First Generation (Mechanical/Electromechanical)
Calculators
Antikythera mechanism, Difference engine, Norden bombsight
Programmable Devices
Jacquard loom, Analytical engine, Harvard Mark I, Z3
Second Generation (Vacuum Tubes)
Calculators
Atanasoff–Berry Computer, IBM 604, UNIVAC 60, UNIVAC 120
Programmable Devices
Colossus, ENIAC, Manchester Small-Scale Experimental Machine, EDSAC, Manchester Mark 1, Ferranti Pegasus, Ferranti Mercury, CSIRAC, EDVAC, UNIVAC I, IBM 701, IBM 702, IBM 650, Z22
Third Generation (Discrete transistors and SSI, MSI, LSI Integrated circuits)
Mainframes
IBM 7090, IBM 7080, IBM System/360, BUNCH
Minicomputer
PDP-8, PDP-11, IBM System/32, IBM System/36
Fourth Generation (VLSI integrated circuits)
Minicomputer
VAX, IBM System i
4-bit microcomputer
Intel 4004, Intel 4040
8-bit microcomputer
Intel 8008, Intel 8080, Motorola 6800, Motorola 6809, MOS Technology 6502, Zilog Z80
16-bit microcomputer
Intel 8088, Zilog Z8000, WDC 65816/65802
32-bit microcomputer
Intel 80386, Pentium, Motorola 68000, ARM architecture
64-bit microcomputer[34]
Alpha, MIPS, PA-RISC, PowerPC, SPARC, x86-64
Embedded computer
Intel 8048, Intel 8051
Personal computer
Desktop computer, Home computer, Laptop computer, Personal digital assistant (PDA), Portable computer, Tablet PC, Wearable computer
Theoretical/experimental
Quantum computer, Chemical computer, DNA computing, Optical computer, Spintronics based computer
Other Hardware Topics
Peripheral device (Input/output)
Input
Mouse, Keyboard, Joystick, Image scanner, Webcam, Graphics tablet, Microphone
Output
Monitor, Printer, Loudspeaker
Both
Floppy disk drive, Hard disk drive, Optical disc drive, Teleprinter
Computer busses
Short range
RS-232, SCSI, PCI, USB
Long range (Computer networking)
Ethernet, ATM, FDDI

computer

software
aaSoftware refers to parts of the computer which do not have a material form, such as programs, data, protocols, etc. When software is stored in hardware that cannot easily be modified (such as BIOS ROM in an IBM PC compatible), it is sometimes called "firmware" to indicate that it falls into an uncertain area somewhere between hardware and software.
Computer software
Operating system
Unix and BSD
UNIX System V, IBM AIX, HP-UX, Solaris (SunOS), IRIX, List of BSD operating systems
GNU/Linux
List of Linux distributions, Comparison of Linux distributions
Microsoft Windows
Windows 95, Windows 98, Windows NT, Windows 2000, Windows XP, Windows Vista, Windows 7, Windows CE
DOS
86-DOS (QDOS), PC-DOS, MS-DOS, FreeDOS
Mac OS
Mac OS classic, Mac OS X
Embedded and real-time
List of embedded operating systems
Experimental
Amoeba, Oberon/Bluebottle, Plan 9 from Bell Labs
Library
Multimedia
DirectX, OpenGL, OpenAL
Programming library
C standard library, Standard Template Library
Data
Protocol
TCP/IP, Kermit, FTP, HTTP, SMTP
File format
HTML, XML, JPEG, MPEG, PNG
User interface
Graphical user interface (WIMP)
Microsoft Windows, GNOME, KDE, QNX Photon, CDE, GEM
Text-based user interface
Command-line interface, Text user interface
Application
Office suite
Word processing, Desktop publishing, Presentation program, Database management system, Scheduling & Time management, Spreadsheet, Accounting software
Internet Access
Browser, E-mail client, Web server, Mail transfer agent, Instant messaging
Design and manufacturing
Computer-aided design, Computer-aided manufacturing, Plant management, Robotic manufacturing, Supply chain management
Graphics
Raster graphics editor, Vector graphics editor, 3D modeler, Animation editor, 3D computer graphics, Video editing, Image processing
Audio
Digital audio editor, Audio playback, Mixing, Audio synthesis, Computer music
Software engineering
Compiler, Assembler, Interpreter, Debugger, Text editor, Integrated development environment, Software performance analysis, Revision control, Software configuration management
Educational
Edutainment, Educational game, Serious game, Flight simulator
Games
Strategy, Arcade, Puzzle, Simulation, First-person shooter, Platform, Massively multiplayer, Interactive fiction
Misc
Artificial intelligence, Antivirus software, Malware scanner, Installer/Package management systems, File manager

computer

network and the internet
Computers have been used to coordinate information between multiple locations since the 1950s. The U.S. military's SAGE system was the first large-scale example of such a system, which led to a number of special-purpose commercial systems like Sabre.[32]
In the 1970s, computer engineers at research institutions throughout the United States began to link their computers together using telecommunications technology. This effort was funded by ARPA (now DARPA), and the computer network that it produced was called the ARPANET.[33] The technologies that made the Arpanet possible spread and evolved.
In time, the network spread beyond academic and military institutions and became known as the
Internet. The emergence of networking involved a redefinition of the nature and boundaries of the computer. Computer operating systems and applications were modified to include the ability to define and access the resources of other computers on the network, such as peripheral devices, stored information, and the like, as extensions of the resources of an individual computer. Initially these facilities were available primarily to people working in high-tech environments, but in the 1990s the spread of applications like e-mail and the World Wide Web, combined with the development of cheap, fast networking technologies like Ethernet and ADSL saw computer networking become almost ubiquitous. In fact, the number of computers that are networked is growing phenomenally. A very large proportion of personal computers regularly connect to the Internet to communicate and receive information. "Wireless" networking, often utilizing mobile phone networks, has meant networking is becoming increasingly ubiquitous even in mobile computing environments.

computer

multiprocessing
Some computers are designed to distribute their work across several CPUs in a multiprocessing configuration, a technique once employed only in large and powerful machines such as supercomputers, mainframe computers and servers. Multiprocessor and multi-core (multiple CPUs on a single integrated circuit) personal and laptop computers are now widely available, and are being increasingly used in lower-end markets as a result.
Supercomputers in particular often have highly unique architectures that differ significantly from the basic stored-program architecture and from general purpose computers.
[31] They often feature thousands of CPUs, customized high-speed interconnects, and specialized computing hardware. Such designs tend to be useful only for specialized tasks due to the large scale of program organization required to successfully utilize most of the available resources at once. Supercomputers usually see usage in large-scale simulation, graphics rendering, and cryptography applications, as well as with other so-called "embarrassingly parallel" tasks.

computer

Multitasking
Main article:
Computer multitasking
While a computer may be viewed as running one gigantic program stored in its main memory, in some systems it is necessary to give the appearance of running several programs simultaneously. This is achieved by multitasking i.e. having the computer switch rapidly between running each program in turn.[29]
One means by which this is done is with a special signal called an interrupt which can periodically cause the computer to stop executing instructions where it was and do something else instead. By remembering where it was executing prior to the interrupt, the computer can return to that task later. If several programs are running "at the same time", then the interrupt generator might be causing several hundred interrupts per second, causing a program switch each time. Since modern computers typically execute instructions several orders of magnitude faster than human perception, it may appear that many programs are running at the same time even though only one is ever executing in any given instant. This method of multitasking is sometimes termed "time-sharing" since each program is allocated a "slice" of time in turn.[30]
Before the era of cheap computers, the principle use for multitasking was to allow many people to share the same computer.
Seemingly, multitasking would cause a computer that is switching between several programs to run more slowly — in direct proportion to the number of programs it is running. However, most programs spend much of their time waiting for slow input/output devices to complete their tasks. If a program is waiting for the user to click on the mouse or press a key on the keyboard, then it will not take a "time slice" until the event it is waiting for has occurred. This frees up time for other programs to execute so that many programs may be run at the same time without unacceptable speed loss.

computer

input/output
I/O is the means by which a computer exchanges information with the outside world.[27] Devices that provide input or output to the computer are called peripherals.[28] On a typical personal computer, peripherals include input devices like the keyboard and mouse, and output devices such as the display and printer. Hard disk drives, floppy disk drives and optical disc drives serve as both input and output devices. Computer networking is another form of I/O.
Often, I/O devices are complex computers in their own right with their own CPU and memory. A
graphics processing unit might contain fifty or more tiny computers that perform the calculations necessary to display 3D graphics[citation needed]. Modern desktop computers contain many smaller computers that assist the main CPU in performing I/O.