MINISTÉRIO DA EDUCAÇÃO
UNIVERSIDADE FEDERAL RURAL DE PERNAMBUCO
THE LOGIC OF THE TRANSDISCIPLINARITY
Luiz de Carvalho
Recife – PE 2007
MINISTÉRIO DA EDUCAÇÃO
UNIVERSIDADE FEDERAL RURAL DE PERNAMBUCO
THE LOGIC OF THE TRANSDISCIPLINARITY
Proposta de projeto de tese para
doutoramento a ser apresentada
ao Dr. Basarab Nicolescu com coorientação dos professores Dr.
Romildo Nogueira e Dr. George
C. Jimenez.
Luiz de Carvalho
Recife – PE 2007
Category Theory Language
1 Category theory and local and nonlocal toposes
1.1 Local theory
1.2 Local language interpretation
1.3 Toposes and levels of reality
1.4 Local toposes sheaves
1.5 Logic fibrillation
2 Logic of transdicisplinarity/complexity
2.1 Intuitive characteristics, language, and formalization
2.2 Set theory paradoxes and type theory
2.3 Ascending hierarchy of types
2.4 Higher-order logic
2.4 Complex plurality and algorithmic complexity
2.6 Law of excluded middle: local and nonlocal semantic toposes
2.6 The nonlocal and the inconsistency
3 Transdisciplinary logic´s fibrillation and algebraization
3.1 Possible translations semantics
3.2 Morphism between two signature systems in Hilbert style H 1 ⊕ H 2
Conclusion
Presentation: a brief historic-scientific introduction
The aim of this work is to offer a logic-formal and axiomatic
characterization to the Transdisciplinarity Theory. However, an introduction to
transdisciplinarity requires a brief reading of the historic-scientific events that
justifies the need for a new epistemic approach, convenient to the changes
occurred during almost two centuries. These changes proceeded in order to
break up some or almost every classical logic principles, which have motivated
or motivate the scientific concepts up to now. Those principles are: the principle
of identity, which supports that anything is itself, or formally, A is A; the
noncontradiction law, which supports that, between two contradictory
prepositions, one is false, or formally: ¬( P ∧ ¬P ); and the law of excluded
middle, or formally, P ∨ ¬ P.
The history of the classical logic-deductive method starts with Aristotle
and, since then it is improving, up to the beginning of modernity. Nevertheless,
from nineteenth century, cracks in the classical Aristotelian-Euclidean thinking
begin to appear, regarding the called non-Euclidean geometries. Euclid (330277), with his Elements, gives a systematic form to Greek knowledge gathering
proposition and demonstration in a deductive way, according to Aristotle lesson,
which poses that, in a T theory, a proposition (P) and its denial, named non-P
( ¬ P), are not deductive, and that only P or non-P are demonstrable.
Logic is the pratical study of the valid arguments and, for centuries,
according to the Elements, the geometric model theory remained till the
emergence of the completeness and movement problems, linked to the
postulate five (that, if a straight line falling on two straight lines makes the
interior angles on the same side less than two right angles, the two straight
lines, if produced indefinitely, meet on that side on which are the angles less
than the two right angles), and to the congruence axiom, that is:
a) (Completeness): The Elements axiomatization is not enough to deduce the
theorems.
b) (Movement): the congruence axioms depend on the movement concept.
For many years, two tendencies have been trying to solve these
problems: adding more axioms to complete geometry or showing that this
attempt would be impossible. Archimedes has stood up for the first tendency
and others, for different versions of the postulate five (given any straight line
and a point not on it, there "exists one and only one straight line which passes"
through that point). Nonetheless, Lobachevski, Gaus and Bolyai have argued
that geometry does not represent the physical reality, and that its postulates
and theorems are not necessarily true in the physical world.
The postulate five has been attacked, culminating in the Lobachevski´s
postulate (Lp): there exist two lines parallel to a given line through a given point
not on the line. A completely new geometry emerged from this novel axiom,
presenting no contradictions regarding the Euclidean geometry. From these
progresses, we could not infer that the axioms of mathematics are intuitive
constructions based on some kind of non-mathematical evidence.
This way, the evolutive process of the deductive axiomatic system leaves
the aristotelic grammatical and discursive scope to feature itself as a reasoning
algebra with G. Boole (1815-1864). After that, how is it possible to set up the
consistency of the axioms in a confident way? It is necessary to build up
mathematics from itself. On the other hand, the congruence axioms in the
Element´s book V, from where Eudoxus´ method of exhaustion and CauchyRiemann integral emerged (408 B.C. – 355 B.C.), were initially inspiring to the
conception of real numbers as a geometric abstration (Axiom XI – every right
angle are similar); it was necessary to give consistency to the real, free from the
infinite approach of the infinitesimals, which Berkeley used to call absent
magnitude ghost; this was performed by Weierstrass, with his formal theory of
limits, and by Dedekind/Cantor, with his construction of real numbers based on
natural numbers, named arithmetization of analysis. Such constructions involve
some use of mathematical infinity. In this context, David Hilbert´s program have
arisen: he wanted mathematics to be formulated on a solid and complete logical
foundation.
The axiomatic method has been encouraged with these developments,
proceeding with this perspective, which was also adopted by Isaac Newton
(1642-1727), who classified, in a axiomatic and deductive way, the classical
dynamics, in his PRINCIPIA, by three axioms or laws of motion. Newton´s
axiomatic counted on the infinitesimal calculus method. However, the presumed
self-evidence of the three axiomatic principles, which validated Newtonian
principles, particularly the fundamental assertion regarding the principle of
inertia and its parameter, absolute time and space, ended up as a naive
pretension. Despite its basis problems, Newtonian model prevailed during the
1700s with a scientific revolution that was born to fight Aristotle authority, but
that turned to be a closed and powerful model, whose principles evoke a
privileged observer, Laplace´s demon, that, according to Basarab Nicolescu,
validated the consistency of the newborn science postulates:
(I) The existence of universal laws, mathematically characterized
(II)The discovery of these laws by means of scientific experience
(III) The perfect reproducibility of the experimental data
According to what would assure the consistency of mathematics as an
instrument for a new research program, science, Cantor´s theory of sets, in
which elementary arithmetic is based, sustained the aristotelian classical logic
as na underlying logic (D’OTTAVIANO,1994).
Classical physics successess, on the other hand, kept confirming the
three postulates, even if the first postulate had never been definitively proved.
The arithmetization of analysis moviment, that, according sets theory would led
to the named sets paradoxes, or Cantor-Russell paradox, if such a set is not a
member of itself, it would qualify as a member of itself by the same definition,
culminated with Hilbert program that aimed to prove that such problems could
be surpassed through a suitable formalization that would allow the metatheoretical demonstration of the arithmetical consistency and, as a
consequence, of mathematics ( HILBERT & BERNAYS, 1934).
During the first half of 20th century, Kurt Gödel publishes his
incompleteness theorems, which would limit Hilbert program:
A: (First theorem) To every ω-consistent recursive class κ of formulas there
correspond recursive class signs r, such that neither vGenr nor Neg(vGenr)
belongs to Flg(κ) (where v is the free variable of r).
B: (Second theorem) For any formal recursively enumerable (i.e. effectively
generated) theory T including basic arithmetical truths and also certain truths
about formal provability, T includes a statement of its own consistency if and
only if T is inconsistent.
One of the consequences of these theorems is that any formal system
that envolves elementary arithmetic, accepts true and nondemonstrable
sentences. Newton C. A. da Costa and Francisco A. Dória (da COSTA and
DORIA 1991) have extended this result to classical physics. Nevertheless, what
does Gödel theorems mean to science and, particularly, to Physics? The
science philosopher Maria Luiza Dalla Chiara understands that, from the point
of view of the development of Logic theories in its non-independent relation to
any research content, Gödel theorems have produced a plurality situation in the
treatment of some physical theory (DALLA CHIARA, 1980).
However, if Gödel theorems regards only to the formal aspect, in what
sense the theorems could be extended to the physical theories? Before we
continue, let’s take a look to another result of the Gödel theorems, the
undecidable, which will make our question clearer:
C: For any consistent formal theory that proves basic arithmetical truths, it is
possible to construct an arithmetical statement that is true but not provable in
the theory. That is, any consistent theory of a certain expressive strength is
incomplete. The meaning of "it is possible to construct" is that there is some
mechanical procedure that produces another statement.
Albert Einstein, concerning the debate about Quantum Physic’s
incompleteness in the famous EPR argument (Einstein-Podolski-Rosen), which
we are going to consider soon, give us the necessary variables to understand
Gödel’s results regarding physical theories, that Lokenath Debnarh synthesized
in these words:
The totality of concepts and propositions constitute a physical theory
provided it satisfied two essential equirements: external justification, which
means the agreement between theory and experiment, and internal perfection,
which
implies that the theory can be inferred from the most general
principles without addition concepts and assumptions. (DEBNATH, 1982).
What Einstein comprehends as a theory’s inner perfection is its
consistency as a mathematical structure, what would lead us to Gödel’s A
theorem. However, D´s external justificative would demand an axiomatic
characterization of physical theories, which we can obtain from Patrick Suppes
structure (P SUPPES - 1967), which links mathematic theory and physical
phenomena:
A physical theory T is a triple T = 〈 M,D,R 〉 , where M is the theory’s
mathematical structure; D is the application domain and R is the set of rules that
links M and D, according to da Costa and Sant’anna (da COSTA and
SANT’ANNA, 2006):
On the other hand, mathematical concepts in principle cannot be
considered as a faithful picture of the physical world; but can, through
axiomatization, be susceptible to the use of logic-mathematical tools as
indecidable propositions like C one hold physical meaning. In fact, according to
Lycurgo (LYGURGO 2004), there is a connection between these results and
what we can observe in the physical world; but what is relevant in completeness
theory, according to the definition that “a physical theory is complete if every
element of the physical reality has a counterpart in the physical theory”, that is,
there is more in the world than the specific theory can anticipate, according to
Gödel’s results. But theses results do not provide us hope to get a complete
axiomatization, as incompleteness is structural. According to Maria Luiza Dalla
Chiara, the notion of an independent Logic is, historically connected to some
kind of single logic (Dalla Chiara, 1980).
Nevertheless, in front of current demands, logic is no more characterized
regarding the valid formula concept and is perceived as consequence relations
according to Tarski, allowing a plural and interrelation treatment between logics.
At the same time, another revolution was happening in scientific research
world, which would definitely place us in another perspective: Quantum
Mechanics. This theory is the best example of semi-complete theories´ blend.
Historically, we first have the quantum theories of Planck 1900), Einstein
(1905), Bohr (1913) e De Broglie (1924); in a posterior moment, Heisenberg’s
matrix mechanics (1925) and Schrödinger’s undulatory mechanics (1926) and
Pauli´s and Dirac´s theories. All these formulations incorporate, somehow, the
controversial phenomena that turn quantum mechanics in a singularity in
science history: wave-particle duality, its probabilistic aspect, and the
inseparability concept.
Quantum theory and its fundamental experiment, Young´s double-slit
experiment
(http://www.upscale.utoronto.ca/GeneralInterest/Harrison/DoubleSlit/DoubleSlit.
html#TwoSlitsElectrons), which distinguishes quantum mechanics from
classical mechanics through atomic phenomena explanation. This experiment
provide us with some odd results: if we call the first experiment open slit A and
the closed one, B, we have the following results, represented as logical
circuits: A ∧ ¬B ↔ ¬( A → B) ↔ ¬¬A ∧ ¬B , that is, particles remain particles with
peculiar intensity distribution, and the same happens when B is open and A is
closed, B ∧ ¬A ↔ ¬( B → A) ↔ ¬¬B ∧ ¬A . But what we note when we open
both A and B slits at the same time? Based upon previous experiments and on
the fact that the third one would keep distribution consistency, we would get the
natural sum of both beams projected in the screen, as each electron moving in
its and well-drawn and fixed trajectory, which pass through the slit without
influencing other electrons that pass through the other slit. Therefore, we could
have that, if this is valid for ∀ n experiment, so it is valid for n+1, as follows:
!) A ∧ B [ hipótese]
2) A ∧ B → A [ separação ]
3) A ∧ B → B [ separação]
4) A [ ModusPonens ]1e2
5) B [ Modusponens ]1e3
6) A ∨ B [ Introdução ] 4e5
However, this is not what happens exactly. Electronic diffraction
phenomenon shows a picture on the screen that occurs due to interference and
does not correspond to the sum of the pictures produced by each of the slits
individually, as the above hypothesis presents. Now let’s imagine that this
phenomenon is consistent with the two first experiences, i.e., the third
experience maintains the properties of electrons pertaining to the previous
experiments. This could mean that the undulatory aspect observed is motivated
by the particles individually, suggesting some hypothesis:
a) electrons should irradiate electromagnetic waves continuously;
b) they would need a inexhaustible energy source in order to do not deteriorate
their nuclei;
c) their speed should be higher than the speed of light C.
As these hypotheses are senseless, quantum mechanics is contradictory
in comparison with classical electrodynamics, which proposes that the
undulatory aspect is consequence of the electron trajectory.
We can conclude that light presents particle and wave features at the
same time. The first two experiments obviously suggest the continuity concept,
according to Nicolescu: it is not possible to pass from one extreme to another
without passing through the mid-space. However, the third experience breaks
continuity and places non-classical probability as a suitable mathematical tool to
deal with diffraction interferences.
Diffraction phenomenon demonstrates that the continuity perceived in the
previous experiments with individual particles is not valid; because diffraction
phenomenon presents bindings where particles would not be splitted, according
to the hypotheses shown above; this phenomenon is known as nonseparability
or nonlocality. This condition made many researchers react, resulting in EPR
Paradox, or Einstein-Podolsky-Rosen Paradox. Essentially, EPR is a thought
experiment that demonstrates that the result of a measurement performed on
one part A of a quantum system has a non-local effect on the physical reality of
another distant part B. As far as one can judge, this situation is not possible
because special relativity asserts that no information can be trasmitted faster
than light. Nevertheless, Jonh Bell, with his theorem, has proved an absolute
distinction between quantum and classical mechanics. Assuming separability
and sustaining a, b, and c hypotheses, we would need hidden variables that
would continuously determine experiments of diffraction, regarding a given
interpretation of the quantum phenomena.
Bell proved that the hypothesis of local separability (a particle holds
defined values that do not depend on a contextualization; and physical effects
speed of propagation is finite) is not in accordance with quantum physics, that
is, the correlation coefficient C(θ)<N for local hidden variable theories and to
quantum mechanics C(θ)>N for some θ. It was born a new kind of causality:
non-local causality. In our macrophysical world, if two objects interact in a
certain moment in time and then stand back, they interact less. In quantum
world, entities keep interacting independently of the distant (NICOLESCU,
2001).
Non-local causality idea questioned another classical science principle:
the existence of only one level of reality, by which Laplace´s demon, a
privileged observer, could be an invariant system from a number of general
laws (NICOLESCU,2001). Under this context, in one hundred years of scientific
development, which meets a perspective pluralization, Transdisciplinarity and
Complexity have emerged, as well as several perspectives with the same
objective: adjust themselves to complexity and to the current disciplinary boom.
TRANSDISCIPLINARITY: COMPLEXITY, LOGIC OF THE INCLUDED THIRD
AND LEVELS OF REALITY
What is transdisciplinarity/complexity as defined by Basarab Nicolescu? The
transdisciplinarity is founded upon three pillars: levels of reality, logic of the
included third and complexity. The revolution of quantum physics with its
problems, especially the known measurement problem, plays central role in the
concept of transdisciplinarity and complexity. Avoiding go into metaphysical
subtleties, which is as rigorous as calculation, for us is sufficient the idea of
levels of reality as an ensemble of invariant systems under the action of a
number of local general laws, or as something that offers resistance to our
experiences. But we can use the principle that "the non-local is inconsistent with
the locality (NL)" and from this say that a level of reality is a structurally welldefined class, and is local, which plurality constitutes non-local classes that
made the complexity. The NL principle means that two levels of reality are
different, if, passing from one to the other, there was a breaking in the laws and
a break in fundamental concepts, (NICOLESCU, 2001).
To deal with levels of reality from a logic point of view, that is our goal, is
necessary the locality theory, studied in the ambit of categories theory, the local
topos, which Lawvere gave an axiomatic characterization similar to the category
of the sets, the called toposes, in reference to the toposes introduced by
Grothrndieck for algebraic geometry. (LAWVERE, 1975). Then, levels of reality
and levels of perception make a local topos. A topos form a kind of category
that has the following properties similar to the category of sets:
- Should have a terminal object corresponding to the unitary element, the
set 〈 ∗ 〉 ;
- Should be able to form the Cartesian product A × B of two objects and the
exponential object CB of functions of C on B;
- Should be able to form the Cartesian product A × B of two objects and the
exponential object CB of functions of B on C ;
- Must have an object of values of truth Ω with a distinguished element ┬ and a
corresponding element sub-object one-by-one B of A and its characteristic
function χ B: A → Ω such that, for any element a of A, a belongs to B if and
only if χ B(a)=┬, and must have an object of naturals N equipped with an
element 0 ∈ N and the functions S:N→N defined by recursion such that: for each
a ∈ A and each function h:A→A, We can build a single function φ:N→A such
that φ(n)=hn(a) for any natural number n (LAMBEK, 2004).
Also according to Lawvere, these properties can be expressed in the language
of categories theory, not referring to elements, in a way that, ┬ is seen as an
application 1 → Ω and 0 as 1 → N. Thus, each topos ∆ is associated with a
type theory λ (∆) → ρ as its internal language. Inversely, for every type theory ρ
is associated a topos π(ρ), the topos generated by ρ, known as TarskiLindenbaum’s category of ρ. (LAMBEK, 2004). If we take the levels of reality
and its perception as local topos, what relationship exists between complexity
and transdisciplinarity? We can say that the possibility of dealing with multiple
levels of knowledge reality, which constitutes as a complex plurality in itself, is
given by the logic of the included third without the need of included fourth, fifth,
n.
What is supposed regarding levels of reality to meet the above requirements,
while local topos, is that he keeps doing some kind of coherence asserting the
axiom of regularity of ZF (Zermelo-Frankel), or, that any not empty set X
contains
an
element
Y
that
is
disjunctive
of
X,
formally:
X ≠ ø→ ∃Y ( Y ∈ X ∩ Y I X ≈ ø); that axiom forbids circularities as the set of all
sets that do not belongs to himself as a member, formally: M= { A / A ∉A} and
therefore belongs to itself, that is, the paradox of Russell. With the results of
the theorem of Pierce of: toute polyade supérieure à une triade peut être
analysée en terme de triades, mais une triade ne peut pas être généralement
analysée en termes de dyades" como faz Nicolescu (Nicolescu, 1998) and
those of N. Resch for polyvalent logic (MANY-VALUED LOGIC 1969),
according to that, any system of logics Ln valents will contain tautological
formulas as theorems that are not necessarily in Ln+1 and by theorem of CL
deduction (classical logic) all tautology is a theorem of CL, we have the local
topos complete.
However, the transdisciplinary local toposes are generally incomplete, not
worthing the generalized axiom of choice and could be formalized as intuitive
logic (J), classic while local levels of reality. What constitutes transdisciplinary
logic complexity, as we understand it, is the non-local interaction of local
toposes in accordance with NL principle. That is the sense of quantum
discontinuity as Nicolescu means (NICOLESCU, 1998)
From this point, the issues rose by Nicolescu about the sense of opened unity
of the world, is essential to understand what he calls the Gödelian structure, the
correlation of levels of reality, for example: What is the nature of the theory
which can describe the passage from one level of reality to another? Is there
coherence or even a unity in the ensemble of levels of reality? What should be
the role of the subject-observer in the existence of an eventual unity of all the
levels of reality? Is there a privileged level of reality? Would be the unity of
knowledge, if any, of subjective or objective nature? What is the role of reason
in the eventual unity of knowledge? What would be, in the field of reflection and
action, the predictive power of the new model of reality? Would be possible to
understand this present world? In the same text, Nicolescu refers to levels of
reality ensemble as an evolutive process that would possess a self-consistency
which, by the second theorem of Gödel, this self-reference would imply in a
contradiction and consequent separation in levels of reality undoing the same
self-consistency in local collapses in which worth the classical logics. As we
said above, what is subjacent to this process of self-consistency is an
interpretation of measurement problem in quantum physics. Then in what
consists the measurement problem?
OSVALDO (1992, pp. 177-217) in its study of measurement problem,
characterized the quantum mechanics (QM) as a structure as follows: A closed
system is described by a state that evolves over time in a determinist way,
according Schrödinger equation, Ψ(r, t). As we mentioned, this state provides
only the probability of obtaining different results from a measurement. By now,
after measurement, the system is in another state, which depends on the
outcome, thus, in the course of measuring the system evolves in an
indeterminist way, described in Von Neumann projection postulate or wave
packet collapse or state reduction. Therefore, the problem of measurement
arises from this opposition between determinist and indeterminists evolution
associated to projection postulate. And if we add two more hypotheses:
α) the metric device belongs to the quantum state;
β) The composed system, object/device may be closed on the environment, so,
its evolution should be deterministic, but, state reductions will continue to occur
during measurements.
It is not necessary to say that this paradox is linked to the double slit experiment
and to its paradox wave/particle. To summarize, the question is to know how
during measurement, a quantum superposition may become in states that do
not superpose. Von Neumann answer is that projection postulate accompanying
any act of measurement, formally:
∧
Ζ): φ (τ ) ≈ Ω φ (τ ) Where omega is the unitary operator and the state vector
is expressed by superposition φ ≈ α 1 φ1 +α 2 φ2 where φ1 and φ2
state of an observable which would reduce to
η): φ1 ∨ φ2
are self-
depending on
observation/measurement result, and that the probability of each outcome is θ)
2
a1 ∧ a2
2
∧
∧•
The unitary operator and their inverse adjunct is equal to Ω∧ Ω 1
From that solution, two problems appear: one of CHARACTERIZATION, i is,
what are the conditions for apply the projection postulate, and being it a
inherent condition for measurement, what will characterize a measurement
and/or observation? The other problem would be COMPLETENESS: Could
projection postulate be derived from other quantum physics principles with a
physical model suitable for the measurement process? Without going into
details of various currents that proposed solve this problem, we see that a new
thermodynamic axiom, linking a large number of particles and incoherent states,
would have been introduced to the problem of completeness.
SUCH ADJUNCTION, CALLED THERMODYNAMIC AMPLIFICATION, COULD
CHARACTERIZE MORE SPECIFICALLY THE ATOMIC PHENOMENAS
OBSERVATION
BASED
ON
RECORDS
OBTAINED
THROUGH
AMPLIFICATION DEVICES WITH IRREVERSIBLE WORK. This program
would be later unsolved, so that, the problem of completeness is unsolvable
and that the thermodynamic models do not provide exact solution to the
measurement problem. If the completeness problem was solvable, it would
provide an example of measurement with certain conditions:
- a) a quantum state can be attributed to the macroscopic measuring
equipment, (device state in a Hilbert space of finite dimension)
- b) The composite quantum system (device/object) can be considered closed
in relation to environment and the rest of the universe, (the composed system
evolution is unitary)
- c) The different measurement results correspond to distinct final states of the
device; (pointer states)
- d) A more precise specification of this pointers supposition constitutes a
condition of solvability;
- e) Should be precisely defined what is meant by measuring, (a measurement
is a subclass of interactions governing the evolution of a composed system,
device/object, that should be able to distinguish between classes of object
states for which the medium values of a self-adjunct operator are different. In
addition to this definition, there are the measurement classes of repeatable and
predictable. A repeatable measurement is such that, shortly after being made
and repeated, the result certainly is the same, or, the final state of the object
after the measurement is the self-state corresponding to self-value obtained as
a result. The predictable ones are that for each possible outcome there is a
state of the object that leads to a predictable result.) A measurement satisfying
these conditions, would give a solution for the completeness problem.
Meanwhile, even as (OSVALDO 1992, pp. 177-217), the evidence is that such
measurement cannot be defined, and then, the completeness problem being as
unsolved.
Being:
Σ – Unitary condition
Λ – Measurement
Ξ – Solvability
Φ – The a, b and c hypothesis
The evidence of insolvability in general can be represented as:
Φ,Σ,Ξ├ ┐Λ
Accepting that does not exist ΦeΣ then does not exist measurements that
meets the condition of solvability, Φ,Σ,Λ├ ┐Ξ.
The problem of measurement received several proposals of solution, and one of
them is the theory of open systems, in which the compound system, that
includes object/device, cannot be completely isolated from the environment. In
this conception, there is an approach in which the reduction process would be
an objective physical phenomenon that would transcend the measuring act,
transforming itself in the collapse or decoherence problem, or the problem of
potentialities update, ontologies of powers that could demand a
supplementation. In this perspective is settled the Gödelian self-consistent
universe of Nicolescu. Thus, the question for us is: Is the inconsistency
associated to reality levels ensemble evolution self-referentialty, the same
inconsistency of the correlation between locality and non-locality? If the answer
is yes, we have from the second theorem of Gödel, that levels of reality
ensemble are inconsistent. If the answer is no, we have the following
hypotheses:
- Or the correlation local/ non-local would of be of extension and therefore will
worth ψ : x ∈ σ ↔ x ∈ τ ,ψ ├ σ ≈ τ , or a conservative interpretation of the logical
structure would be extended to other structures or toposes, violating the NL
principle, and the ensemble of levels of reality would be inconsistent;
- Or, self-referentialty not infringes the axiom of regularity and the correlation
local/non-local would be inconsistent.
So, in the positive response view, the system broadly would be equal to the
correlation local/non-local which collapses from itself-reproduction in consistent
and incomplete local toposes. Answerig to questions above, we could say that
transdisciplinarity is consistent and inconsistent making worth the contradictory
pairs that experience and scientific theory, after quantum physics, had saw
appear: local causality and non-local; reversibility and irreversibility, separability
and non-separability; wave and particle, continuity and discontinuity, symmetry
and broken symmetry, and so on. But Nicolescu himself answered such
questions, particularly the first, about the nature of the theory that could
describe the transition from one level to another, saying that no one managed to
find a mathematical formalism that allows the passage of a strict world to
another. But we must differentiate logical epistemic formalization from formal
mathematical structuring of models theories. A logical epistemic formalization
has no compromise with the stander algebraic model, such as the formal
mathematical structure. Logical epistemic formalization works with theory of
categories advanced features, creating logical-topological spaces of possible
worlds from the Tarskian logic conception as relations of consequence. In this
way, we believe that logic subjacent to transdisciplinarity, is a combination of
classical and inconsistent logics, kind of not Hegelian temporal dialectic,
because we are always in some reality, and that subject self-position, as Hegel
believes, is derived. However, criticism to Hegelian succession of contradiction
seems to lack sense, because, if there is a contradiction, at some lapse of time
this contradiction is simultaneous. In any way, transdisciplinarity self-reference
is not closed in the immanence of self- position of idealistic subject of self, that
puts the non-self and understand it. The complexity of transdisciplinary selfreference opens to the multiplicity of historical-existential experience as a
subject pertain to the worlds that it is included. Two other properties are at the
core of transdisciplinarity:
# The opening of reflexivity in infinite resource becomes totality indeterminacy:
the whole, the same, just like this while incomplete (inexhaustible);
# The opening of dialectic to difference, as an outcome of relative
incommensurability of different languages and/or (levels of reality) and meaning
horizon.
But, is the sense of self-referential of Nicolescu the same as of Morin?
In a revealing dialogue we get the difference of self-referential sense involved in
the "logic" of included third between B. Nicolescu and E. Morin, at Stéphane
Lupasco international conference, where Nicolescu asks:
- B.N: E, so, you were quite attracted by theorem of Gödel. Because the
essential point is there: this organizational structure of included third logic is
Gödelian type. It never closes.
- E.M: Of course. But Gödel reaches to a principle of uncertainty. That is, he
says the system cannot give account of itself, with its own resources. It is not
known whether it is consistent and it is not known if it contains a contradiction.
First point: the uncertainty. Second point: is possible, however, conceive a
richer meta-system, able to understanding the system, and that, of course, it
opens to infinite itself. But then, at this moment, I do not see the expression of a
included third specific logic.
Nicolescu then retakes aspects of Gödel’s theorems, almost always overlooked
in the analytical approaches, introducing the meta-theoric affirmation of T inside
of T consistency, and therefore, their self-reference as a contradiction that,
according to E. Morin, is a priori: Le théorème de Gödel nous dit qu'un système
d'axiomes suffisamment riche conduit inévitablement à des résultats soit
indécidables, soit contradictoires. Cette dernière assertion est souvent oubliée
dans les ouvrages de vulgarisation de ce théorème.
So let's see: La is the arithmetic language, T an axiomatizable recursive
extension of PA, Peano arithmetic, PrT(x) a proof predicate for T and Com(T)
the sentence
¬PrT( ┌0=1┐).
We assume that T is consistent and T├ Com(T) and being σ a sentence in La
which meets T ├ σ ↔ ¬PrT(┌ σ ┐). We say that Λ is a model of T,and is
positive if Λ ╞ σ and negative in other cases. If T is consistent, there is a model
Λ1 of T. If Λ1 is positive, then Λ1 ╞ Com(T+¬σ), then there is a Λ2 negative
model of T which is a definable extension of Λ1. If not, Λ1 = Λ2. For the
hypothesis that Λ2 ╞ Com(T), we have a model Λ3 of T which is a definable
extension of Λ2. Λ2 if is negative, Λ2╞ PrT(┌ σ ┐), so Λ3 ╞ σ. Meanwhile, Λ3
╞ ¬ σ and Λ2 ╞ ¬ σ and Λ3 are an extension of Λ2, which is a contradiction
where T theory proves its own consistency. (This evidence is due Jech, 1994)
Once self-consistency reintroduces the subject in the system by itself
formulated, Nicolescu complete its approach of included third logic with the
ontological conception of Lupasco: the subject of knowledge is implied in the
logic by itself formulated. The experience is the subject itself experience. The
circular character of logic assertion, as its own logical experience, occurs from
circular character of subject: to define the subject we must take into account all
phenomena, elements, events, states and propositions that relate to our world
and, furthermore, the affectivity. An obviously impossible task: in the ontology of
Lupasco, the subject never would be defined. This Gödelian structure always
opens, beyond the levels of reality, a NOT RESISTANCE ZONE, that we rather
to call, SEMIOTICALLY NOT FORMED FIELD, which is a field not finitely
algebraizable by AAL method (abstract algebraic logic) (W. Blok and D. Pigozzi,
1989) that we will describe soon. Here there is a problem on the complexity
measurement in the algorithmic complexity point of view. Will Kolmogorov
complexity measurement apply to this semiotically not formed field? Is possible
consider the transdisciplinary not-separability as a totality whose complement is
not computable? Intuitively speaking, Kolmogorov complexity of a number n,
denoted by K (n), is the size of a program that generates n ; and n is called
random if n ≤ K (n). Kolmogorov has proved that not random numbers set is
recursively enumerable, but is not recursive. Odifreddi (1989) considered this a
version of Gödel’s first theorem of incompleteness. Second Kikuchi (1997),
complexity of Kolmogorov also leads to the second theorem of incompleteness
of Gödel, extracted from the paradox of Berry, (“What is the "first not named
number with less than ten words?”). Since Gödel has proved that the first
theorem leads to the second. Then, to apply Kolmogorov complexity to levels of
reality ensemble and semiotically not formed field, we would have to consider
local/non-local inconsistency null or levels of reality ensemble self-referential as
a recursive extension, which would give to levels of reality ensemble a local
character as such is the Kolmogorov complexity, valid for the local topos and
the incompletenesses of Gödel. The question turns on know what type of
contextuality is involved the non-resistance area or the transdisciplinary
semiotically not formed field: quantum or classical?
In classical contextuality, according to Aert,(2002), the result is affected by
various environmental aspects, but not by irreducible and specific non
predictable from system and experimental disturbance interaction. In classical
contextuality, Kolmogorov axioms are met, and a classic model of probability
can be used. However, when contextuality is intrinsic, the system and the
disturbance, both, have an internal relation of constitution, as such its interface
creates a concrescence of emergency. The presence of intrinsic contextuality
means that Kolmogorov axioms are not met. The Entanglement can be tested
by determining whether correlation experiments in entities which being together
infringe the inequality of Bell (Bell 1964) Pitowsky,(1989) had proved that if the
inequality of Bell are met for a range of probabilities concerning to considered
experiment response, then there is a Kolmogorovan classical probability that
describes this probability. In this case, the probability can be described as a
lack of knowledge about the precise status of the system. In other hand, if the
inequality of Bell is infringed, such classical Kolmogorovian probability is not
valuable; rather, Bell’s inequalities infringement proves that the probabilities
involved are not classical. In transdisciplinary contextuality case, we believe
being this classical and non-classical at the same time, what seems to confirm
Lupasco’s proposition cied by Nicolescu: quantic dialetic is a doubt expansion,
in this case is not a case of lack of knowledge, but context generation as a
potential, at the same time that some knowledge is updated. So, algorithmical
complexity pertains to transdisciplinary complexity as a local topos.
In the chapter "Transdisciplinarity and open unity of the world", which Nicolescu
calls levels of reality ensemble, is in the Gödelian conditions of its structure, a
local topos in which is valuable Kolmogorovian classical probability. But, in
other hand, in the "Included third. From quantum physics to ontology”. The
problem, if is not from translation, the relationship between a level of reality and
another, does not preserves the coherence, and then, the application of
Kolmogorovian complexity to structurally incoherent levels of reality ensemble is
not valuable. So we can say that:
Being Cn: the levels of reality ensemble;
and Cn ∪ Zr: the levels of reality ensemble union zone of non-resistance. Cn
has a coherent structure and worth the algorithmic complexity ∧ Cn ∪ Zr is
structurally incoherent and generally does not worth the algorithmic complexity.
The question that immediately arises is whether is possible, in general, a
coherent level of reality ensemble Cn. In this case, the expression levels of
reality ensemble, in general, mean a local extension. Then, the self-consistency
refers to Cn ∪ Zr as irreducibility of contradiction and relativity of consistency.
The union operator ( ∪ ) here is being taken in Lupasco sense, of
distinguishable non-separability ∧ A ( is actualization) ¬ ∧ P ( is not
potentiation), ie, Cn ∪ Zr = ∧ A ¬ ∧ P. This formulation presents us a problem
about the use of denial and implication in included third logical. For example: Is
the included third or not a negation of the principle of the excluded third (ETP)?
Apparently, the denial of the excluded third, is not equivalent to intuitionist
denial which is based on openings of a topological space X , A( X ) in which
implication definition ( → ) does not follows (∧, ∨, ¬) as in classical logic ;
¬( A ∧ ¬B) but not worth the double negation, ¬¬A ├ A.
In transdisciplinary logic, the denial of the excluded third when generates a
contradiction makes use of classical denial by the principles of De Morgan
¬( A ∨ ¬A) ≡ (¬A ∧ ¬¬A) , what suggests us that in A(X) there is induced
operations PX, the set of subsets of X, of finites intersection and union which
are closed and form a Boole algebra. So there is not a necessary link between
incompleteness, that we will call ¬C and A(X), the opening of a topological
space, i is ¬ (¬C → A(X)). Thus, we believe that the transdisciplinary local
toposes form incomplete closed spaces in which worth the classical logic with
open complement in which worth a weaker denial, possibly of intuitive type
where □(¬C→ A(X)). These spaces or logic fibrillation, ¬ ( ¬C→A(X)) ⊕ (
¬C→ A(X)) ⊕ ϖ ⊕ ( ¬C → A(X)) ⊕ ¬ (¬C → A(X)) = Ω , Where ϖ is the
intrinsic contextualization, forming a logic of contradictory. So, we have to:
1) Ω ├ ¬ ( ¬C → A(X)) ⊕ ( ¬C→ A(X)) ⊕ ϖ
2) Ω ├ ϖ ⊕
( ¬C → A(X)) ⊕ ¬ (¬C → A(X))
From 1) e 2) is made Ω1 and so on.
OBJECTIVE
This work aims to investigate the possibility of a logical construction of
transdisciplinarity in the Hilbert style LT 1 . In particular, we believe that a
transdisciplinary logical is formed from an intuitionist classical logical
combination and non-classicals temporal inconsistent. Give an algebraic
formalization of local/ temporal inconsistent interaction whose subjacent logic is
intuitionist and non-local, making use of W. Carnielli’s possible translations
semantics and Tarski’s concept of logic as consequence relations.
METHODOLOGY
The methodology used in the logics combination to form a transdisciplinary logic
is the categorical fibrillation. The process of logics composition can be synthetic
(Splicing logic), for example: fibrillation of D. Gabbay.
L= L 1 L 2 if L is our incognite, since it is known L 1 and L 2 , or can be
analytical (Splitting logic), for example: W. Carnielli’s possible translations
semantics:
L=L 1 ... L n in the case where L is known and its decomposition in more
simple logic.
The question that arises before combining logics is: what kind of structure we
want to achieve with the concept of logical system: proof system as tablê,
axiomatic system, natural deduction, or semantic methods as logical matrices,
valuations and Kripke’s semantics? As usual in this kind of methodology, A.
Tarski’s concept of logic enables more resourcefulness in the combined
treatment of logics, which is independent of the concept of valid formulas.
What's common in any system of logic is the concept of logical consequence
denoted by ├ defined from an ensemble of sentences or formulas of L.
Definition:
A system of Hilbert is a pair H = ‹ C, R› as such C it is a signature and R one
set of pairs of the formula ‹ Γ,Ψ › as such Γ U { Ψ} it is a formula set of L( C ).
The elements ‹ Γ,Ψ› of R, the elements are called rules if Γ= is called axiom.
We can make a fibrillation of logic in Hilbert’s type to form transdisciplinarity
logic in such way that they are morphed in the systems of categories of Hilbert
Hil. Therefore, the logic of the transdisciplinarity forms a composition of classic
logics and inconsistent in Hilbert’s style.
JUSTIFICATION
There is no doubt that the twentieth century was the center of great
changes in all areas of knowledge, particularly in the scientific universe,
resulting in remarkable victories in every technique and production fields. In
spite of that, it could be noted that knowledge splitting and compartmentation,
inherited from a tradition shaped by the thoughts generated during the interstice
between fifteenth and nineteenth centuries, was no longer enough to create the
epistemological references required to solve the features of knowledge itself.
Many problems seemed to be out of the entangled theory systems, and a
kind of blindness hovered over the attempts to understand many nature’s
fundamental problems, as well as its more common problems. However,
debates about topics like the structure of matter, the objectivity or the possible
relationships between the observer and the levels of reality, brought to light
unsurpassable paradoxes. Therefore, new ideas emerged, as the Theory of
Relativity, the Quantum Mechanics, and the Complexity Theory, among others
(Nicolescu, 2003).
Anyway, the straighten up of the contents referring the Complexity
Theory suggests a revaluation of the systems for selecting and determining
conceptualization, as well as of the systems that configure the logical
operations. This intends even to affect the designation structures of the
intelligibility fundamental categories, like the mechanisms that operate their
application control.
It is accepted, therefore, that thinking about the complex systems implies
accepting the need to overcome major challenges, starting from the implications
of the multidimensional and hologramatic structures, or even the
interconnectivity and inseparability of the “one” and the “multiple,” which is
pertinent to it.
Despite the expectation that open up new possibilities to understanding
and the sociocultural transformations that this could raise, it is known that it
would be impossible to embrace the study of the Complex Systems only starting
from the established contemporary disciplinarities.
During the First World Congress of Transdisciplinarity, held in Convento
de Arrábida, Portugal, from 2-7 November, 1994, a Letter of Intent was
prepared, outlining a set of fundamental principles, which exalted the need to
consolidate the transdisciplinar and transcultural thinking as the best way to
approach the different aspects of complexity in distinct systems (Nicolescu,
2001).
Transdisciplinarity would then represent a conception of the research
based on a new comprehension milestone, shared among several disciplines,
and which is followed by mutual interpretation of the disciplinary epistemologies.
The cooperation, in this case, would be headed for problems solving, where
transdisciplinarity emerges to build a new model to bring the realities of the
object of study closer (Hernández, .....).
Turning to the Article 14 of the Transdisciplinarity Letter, whose terms
refer to argument rigidity as the best limit to the possible conceptual deviations
from the articulation of the problem situation data, we can justify the need for
unfolding a logic-formal and axiomatic characterization of the Transdisciplinarity
Theory, mainly when we consider the need for another epistemic perspective
suitable to the changes in the way of thinking, inherent of the last 200 years.
Further, considering that a logic system is always useful to choose the
prevailing operations, relevant and evident under its domain (exclusioninclusion, disjunction-conjunction, implication-negation), it is intended, through
this logical constructions, to foster the inclusion of other features to the
transdisciplinary way of thinking, like the “Opening”, comprehending the
unknown, unexpected, and unpredictable acceptance; and “Tolerance”, as an
exercise to recognize the right of the ideas and truths to be the opposite of ours.
Finally considering that logic keeps close connections with metaphysics,
mathmatics, philosophy and linguistics, we can appraise the impacts from this
project, especially regarding the reiteration of almost every classical logic
principles disruption, which sustain the ground theory in all the study fields
mentioned above.
In Brazil, many research groups are involved in studing logic, standing
out Prof. Newton da Costa group, which I am involved in, becoming specialized
in non-classical logics. Nevertheless, regarding the indication of a wide work
field outlining, I intend, through this proposition, to look for deeper
Transdisciplinarity subsidies from a more authentic source, hoping that it will
constitute, in a near future, an important reference to the development of
theoretical and technological studies on semiotics. Therefore, we expect that
the outcomes from this project accomplishment allow a contextualization in this
promising scenario, and serve as a back up to exploration approaches of new
horizons, particularly those which find themselves immersed in the paradigms
invisible zone, so that we can overcome the determinism of the current
explanatory models, which are associated to the convictions and beliefs
systems, in every scope. This must contribute to the collapse of the cognitive
and intellectual conformisms.
Bibliography
Alain de Benoist, L'œuvre de Stéphane Lupasco, Le Club Français de la
Médaille, n° 53, Paris, 1976.
Aerts, D. and Aerts, S., 1994, Applications of quantum statistics in psychological
studies of decision processes, Foundations of Science 1: 85-97.
Aerts, D., Aerts, S., Broekaert, J. and Gabora, L., 2000a, The violation of Bell
inequalities in the macroworld. Foundations of Physics 30: 1387-1414.
Aerts, D., 1985, The physical origin of the EPR paradox and how to violate Bell
inequalities by macroscopical systems. In P. Lathi and P. Mittelstaedt (eds) On
the Foundations of Modern Physics (World Scientific: Singapore), pp. 305-320.
Aerts, D. and Aerts, S., 1994, Applications of quantum statistics in psychological
studies of decision processes, Foundations of Science 1: 85-97.
Aerts, D., Broekaert, J. and Gabora, L., 1999, Nonclassical contextuality in
cognition: borrowing from quantum mechanical approaches to indeterminism
and observer dependence. In R. Campbell (ed.) Dialogues Proceedings of Mind
IV Conference, Dublin, Ireland.
Aerts, D., Broekaert, J. and Gabora, L., 2000b, Intrinsic contextuality as the crux
of consciousness. In K. Yasue (ed.) Fundamental Approaches to
Consciousness. (Amsterdam: John Benjamins Publishing Company), pp.173181.
Atsushi Takahashi, Du logique et de l'être chez Stéphane Lupasco, The Review
of Liberal Arts, n° 83, Otaru University of Commerce, Otaru, Hokkaido, 1992.
Bernard Dugué, Utilisation de la logique dynamique du contradictoire pour
formaliser les systèmes : vers un paradigme ondulatoire en biologie ?, Revue
Internationale de Systémique, vol. 5, n° 4, Paris, 1991.
Basarab Nicolescu, Lupasco et la génèse de la Réalité, 3e Millénaire, n° 3,
Paris, 1982.
Trialectique et structure absolue, 3e Millénaire, n° 12, Paris, 1984
Nous, la particule et le monde, ch. "La génèse trialectique de la Réalité", Le
Mail, Paris, 1985
Regrettables oublis, 3e Millénaire, n° 18, janvier-février 1985.
Préface à L'expérience microphysique et la pensée humaine, op. cit., 1989.
La transdisciplinarité , manifeste, Le Rocher, 1996
Levels of Complexity and Levels of Reality, in "The Emergence of Complexity in
Mathematics, Physics, Chemistry, and Biology" , Proceedings of the Plenary
Session of the Pontifical Academy of Sciences, 27-31 October 1992, Casina Pio
IV, Vatican, Ed.Pontificia Academia Scientiarum, Vatican City, 1996 (distributed
by Princeton University Press), edited by Bernard Pullman.
Bernard Morel, Dialectiques du mystère, La Colombe, Paris, 1962
Constantin Noïca, Préface à Logica dinamica a contradictoriului, op. cit., 1982.
Ed. Morot-Sir, S. Lupasco, Logique et contradiction , Revue des Sciences
Humaines, n° 49, Faculté de Lettres de Lille, janvier-mars 1948.
Françoise Garoche, L'analyse paradoxale : Trialectique et système Méthodologie de formation et d'évaluation, Revue Française de Pédagogie, n°
75, Paris, avril-mai-juin 1986.
Jean-Jacques Wunenburger, La raison contradictoire - Sciences et philosophie
modernes : la pensée du complexe, Albin Michel, Paris, 1990.
HERNÁNDEZ, Fernando. Transgressão e mudança na educação: os projetos
de trabalho; trad. Jussara Haubert Rodrigues. Porto Alegre: Artes Médicas,
1998.
Marc Beigbeder, Contradiction et nouvel entendement, Bordas, Paris, 1972
(thèse de doctorat).
Pierre Solié, Ouverture sur l'unité du monde, Cahiers de Psychologie
Jungienne, n° 28 - "Synchronicité - Correspondance du psychique et du
physique", Paris, 1er trimestre 1981.
Robert Amadou, "Le principe d'antagonisme et la logique de l'énergie" de
Stéphane Lupasco, La Gazette des Lettres, 15 août 1951.
George Melhuish, The Paradoxical Universe, Rankin Bros Ltd. Bristol, 1959.
Gérard Moury, Stéphane Lupasco - Pour une nouvelle logique : la logique
dynamique du contradictoire, Institut National de Recherche et de
Documentation Pédagogiques, Paris, 1976.
Yves Barel, Le paradoxe et le système, Presses Universitaires de Grenoble,
Grenoble, 1979.
W. A. Carnielli, J. Marcos e S. de Amo. Formal inconsistency and evolutionary
databases. Logic and Logical Philosophy, 8:115–152, 2000.
M. E. Coniglio. Combina¸c˜oes de sistemas de conseq¨uˆencia. Atas do XI
Encontro Nacional de Filosofia (ANPOF). A aparecer, 2004.
W. A. Carnielli e C. Sernadas. Preservation of Interpolation by Fibring. In F.
Miguel Dion´ısio W. A. Carnielli e Paulo Mateus, editores, Proceedings of
CombLog’04–Workshop on Combination of Logics Theory and Applications, p.
151–158, Lisboa, 2004. Center for Logic and Computation–IST.
Carnielli, W. A. Possible - translation semantics for paraconsistent logics.
Proceedings of the 1st World Congress on Paraconsistency. (D. Batens, C.
Mortensen, G.
Priest, J. P.Van Bendegem eds.). Research Studies Press. Baldock, UK: 149 –
163. (2000).
Goldblatt, R. Topoi. The categorial analysis of Logic. Nort Holland, Amsterdam.
N. C. A. da Costa. Sistemas Formais Inconsistentes. Tese de doutorado,
Universidade Federal do Paraná, Curitiba, PR, Brasil, 1963. Editado pela
Editora UFPR, Curitiba, 1993.
C. Hifume. Uma Teoria da Verdade Pragmática: A Quase–Verdade de Newton
da Costa. Dissertação de mestrado, Universidade Estadual de Campinas,
Campinas, SP, 2003
R. A. Lewin, I. F. Mikenberg e M. G. Schwarze. Algebraization of paraconsistent
logic P1. The Journal of Non-Classical Logic, 7(1/2):79–88, 1990.
J. Marcos. Semânticas de Traduções Possíveis. Dissertação de mestrado,
IFCH-UNICAMP,
Campinas,
Brasil,
1999.
URL
=
http://www.cle.unicamp.br/pub/thesis/J.Marcos/.
Pigozzi, 0?] Blok, W., Pigozzi, D. Abstract algebraic logic and the deduction
theorem. The Bulletin of Symbolic Logic. (A aparecer).
[Blok – Pigozzi, 1986] Blok, W., Pigozzi, D. Protoalgebraic logics. Studia Logica,
45:337– 369. (1986).
[Blok – Pigozzi, 1989] Blok, W., Pigozzi, D. Algebraizable Logics, volume 77
(396) of Memoirs of the American Mathematical Society. AMS, Providence,
Rhode Island. (1989).
Bueno, J., Coniglio, M. E., Carnielli, W. A. Finite algebraizability via possibletranslations semantics. Proceedings of CombLog’04 - Workshop on
Combination of Logics: Theory and Applications. ( W. A. Carnielli, F. M.
Dion´ısio, P. Mateus eds.): 79–86. (2004).
“Range Theorems for Quantum Probability and Entanglement”, in Khrennikov,
A. (editor) Quantum Theory: Reconsideration of Foundations, 299-308 Vaxjo,
Vaxjo University press (2002).
(With Oron Shagrir) “The Church-Turing Thesis and Hyper Computation”, Minds
and Machines 13, 87-101 (2003).
(With M. Hemmo) "Quantum Probability and Many Worlds" (Forthcoming in a
special issue of Studies in the History and Philosophy of Modern Physics,
edited by R. Frigg and S. Hartmann).
Le principe d'antagonisme et la logique de l'énergie - Prolégomènes à une
science de la contradiction, Coll. "Actualités scientifiques et industrielles" , n°
1133, Paris, 1951 ; 2ème édition : Le Rocher, Coll. "L'esprit et la matière" ,
Paris, 1987, préface de Basarab Nicolescu.
.
•
Download

The Logic of Transdisciplinarity (2)