Académique Documents
Professionnel Documents
Culture Documents
Peter Andreasen
De
ember 17, 2000
e
r
a
m
e
t
It is obvious, that very small
rosswords are easily
onstru
ted. Espe
ially
rosswords whi
h have
only one row or one
olumn. It is also easy to
reate a few very big but very dull
rosswords: If you
keep alternating the rows between 'I I I ' and
' I I ', you
ertainly get a valid (and as big
as you like) 'Classi
English'
rossword. So we want
to not only
onsider the existen
e of big
rosswords,
but also
he
k if there are many dierent of them.
We are going to
al
ulate the number of big
rosswords now.
Assume we have
hosen an alphabet A and a language L over whi
h the
rosswords must be made.
We use the notation jAj for the number of letters
and symbols in the alphabet. Let us introdu
e the
following number as well,
h
e
Nn = jAjn
jAL (j n)n2
2
2(
(n)2n
= L n2 :
jAj
2n2
;
log jAj
just grows towards innity as n does the same. The
se
ond fra
tion,
L (n)n
jAjn2 :
n =
logjAj L (n)
2 This
(n) nd
L
(n)n
= L nd :
jAj
1
H~ (L) = nlim
!1
3
logjAj L (n)
Entropy
Something should probably be said about why entropy is so
entral to information theory. But where
to start and where to end! We will look at only two
aspe
ts, one of somewhat philosophi
al nature and
the other of very pra
ti
al nature.
Entropy is often said to be a measure of how
'
omplex' or even '
aoti
' things are. This
orresponds ni
ely with the observations given above: A
language made up of every possible integer is devoid
of any form or stru
ture. Anything goes. It is impossible to distinguish between the a sequen
e from
the language and a sequen
e of
ompletely random
digits. This language, as explained above, has the
entropy 1. On the other hand, a language made up
of sequen
es of only one letter, say 'a', is
ompletely
stru
tured. No room for
hoi
es. The fun
tion
d
1
1
H~ (L) = 0
and d0 =
;
(
n) is
onstantly 1 regardless of the value of n.
L
d0
1 H~ (L)
This
orresponds to the
ase where H~ (L) = 0.
where d0 is exa
tly the largest dimension where it
The other use of entropy whi
h we will tou
h upis possible to
reate (many)
rosswords over L.
on is data
ompression. We mention the following
4
Referen es
Notes
Ryabko, B. Y. Noiseless
oding of
ombinatorial sour
es, hausdor dimensoin, and kolmogorov
This se
tion
ontains some notes about the history
omplexity. Problems of Inform. Trans., 22(3):
of the results. It is probably most interesting to
170{179, 1986.
readers already familiar with the
on
epts in this
paper. The idea of linking
rossword puzzles and
entropy is, in fa
t, as old as [Shannon, 1948, from Shannon, C.E. A mathemati
al theory of
ommuni
ation. Te
hni
al report, Bell System, 1948.
whi
h we quote the last paragraph of se
tion 7:
The redundan
y of a language is related
to the existen
e of
rossword puzzles. If
the redundan
y is zero any sequen
e of letters is a reasonable text in the language
and any two-dimensional array of letters
forms a
rossword puzzle. I fthe redundan
y is too high the language imposes
too many
onstraints for large
rosswords puzzles to be possible. A more detailed
analysis shows that if we assume the
onstraints imposed by the language are of a
rather
haoti
and random nature, large
rossword puzzles are just pussible when
the redundan
y is 50%. If the redundan
y