Information: Difference between revisions

From Citizendium
Jump to navigation Jump to search
imported>Larry Sanger
("It can be sad that information is a form of data processing" -- sad maybe, but better said)
imported>Greg Woodhouse
(information science is broader than information theory)
Line 1: Line 1:
The objective of '''information science''' is to determinate the informative content of determined object or event and to obtain a better measure of that content. The word “information” is derived of the Latin expression inform that signifies “to give a form to”, meaning the imposition of a structure over an indefinite set of facts. Given the fact that information is not matter neither energy, then it should not be considered an absolute greatness, instead it should be faced in a relative way, such as [[entropy]]. The expression “information quantity” is a metaphor without any physical or numerical property. The information isolated from a system and out of context does not have any meaning; it is the contextualization that brings information to live. Information is transformed in knowledge only when it has a practical utility and without this the information does not pass of mere abstract data, physical or not.
The objective of '''information theory''' is to determinate the informative content of determined object or event and to obtain a better measure of that content. The word “information” is derived of the Latin expression inform that signifies “to give a form to”, meaning the imposition of a structure over an indefinite set of facts. Given the fact that information is not matter neither energy, then it should not be considered an absolute greatness, instead it should be faced in a relative way, such as [[entropy]]. The expression “information quantity” is a metaphor without any physical or numerical property. The information isolated from a system and out of context does not have any meaning; it is the contextualization that brings information to live. Information is transformed in knowledge only when it has a practical utility and without this the information does not pass of mere abstract data, physical or not.


According to Shannon and Weaver (1949) the information can be measured as negative entropy being a measure of the order in a system. It can be said that information is a form of data processing concerning objects, events or persons that only has meaning for a receiver if the increase of knowledge reduces the uncertainly in decision making processes.
According to Shannon and Weaver (1949) the information can be measured as negative entropy being a measure of the order in a system. It can be said that information is a form of data processing concerning objects, events or persons that only has meaning for a receiver if the increase of knowledge reduces the uncertainly in decision making processes.

Revision as of 15:30, 22 May 2007

The objective of information theory is to determinate the informative content of determined object or event and to obtain a better measure of that content. The word “information” is derived of the Latin expression inform that signifies “to give a form to”, meaning the imposition of a structure over an indefinite set of facts. Given the fact that information is not matter neither energy, then it should not be considered an absolute greatness, instead it should be faced in a relative way, such as entropy. The expression “information quantity” is a metaphor without any physical or numerical property. The information isolated from a system and out of context does not have any meaning; it is the contextualization that brings information to live. Information is transformed in knowledge only when it has a practical utility and without this the information does not pass of mere abstract data, physical or not.

According to Shannon and Weaver (1949) the information can be measured as negative entropy being a measure of the order in a system. It can be said that information is a form of data processing concerning objects, events or persons that only has meaning for a receiver if the increase of knowledge reduces the uncertainly in decision making processes.

Langford’s equation defines information according to the following formula:

I = i (D, S, t)

In this formula, I represents the information obtained by a determined interpretative mechanism i which in turn acts on data set D, relatively to a previous knowledge stage S, during a period of time t.

Whatever the definition, information is an invisible agent that acts like an agglutinant in all the processes of decision-making, regulation and control. In an economically dynamic and technical situation information has the fundamental role of being the support for efficient decision making (Pinheiro, 1986). The information must be screened, condensed, stored, transmitted, received, aggregated and integrated. All these actions are based in the fact of that despite its infinity the information, for human consumption, can only be organized in a limited number of forms (Senn, 1989; Blethyn, 1990).

Organization is the result of the interaction of information with matter and energy. When it is added to the matter she shows up always some kind of structure or organization, i.e., the organization can be viewed as information stored. Beyond organizing matter and energy information is still able to self structure. The more organized is a process or a structure, less information is necessary to fully describe it. By the contrary, disorganization is always associated to an increase of system entropy (Stonier, 1990). The more disorganized is a system; more information is possible to extract from him hence there is no certainty that there aren’t hidden patterns in it (Rodrigues, 1991).