Jigsaw Puzzles: Problem B

paper
Authorship
  1. 1. Michael Levison

    Queen's University

  2. 2. Craig Thomas

    Queen's University

Work text
This plain text was ingested for the purpose of full-text search, not to preserve original formatting or readability. For the most complete copy, refer to the original conference program.


Jigsaw Puzzles: Problem B

Michael
Levison

Queen's University, Ontario
levison@cs.queensu.ca

Craig
Thomas

Queen's University, Ontario
thomas@cs.queensu.ca

2002

University of Tübingen

Tübingen

ALLC/ACH 2002

editor

Harald
Fuchs

encoder

Sara
A.
Schmidt

A jigsaw puzzle problem arises whenever an archaeologist finds an old
manuscript broken into fragments and wishes to reconstruct it. The best
known examples are the Dead Sea Scrolls, reconstructed by hand during the
middle of the 20th century.
In fact, there are two different types of problem, each requiring a different
strategy. If the fragments come from a familiar text, the manuscript can be
reconstructed by locating each fragment in an existing copy without
reference to other pieces of the puzzle. If the text is unfamiliar, we must
compare the shapes of fragments to one another to discover good matches with
reasonable letter sequences formed across the boundaries. We haved called
the former Jigsaw Problem A, the latter Problem B.
The purpose is to find computer techniques which substantially reduce the
human work involved in reconstruction.

Previous Work
Very little has been published on this topic since it was brought to the
first author's attention 35 years ago. A previous paper ( Levison and Wu,
1999) reported on some experiments involving artificial fragments derived
from a paragraph of Alice in Wonderland. This work
essentially provided a computer solution for Problem A, predicting whether a
single site is likely for a given fragment based on letter-sequence
frequencies and, for our fragments, finding such a unique location in Alice
whenever it was predicted.
Experiments on problem B proved less successful. A program matched each of
the 53 fragments to each of the others with all possible vertical shifts,
computing a score for each juxtaposition. The "best" matches were listed for
each fragment, different scoring formulae being tried without materially
affecting the outcome. The correct match often occurred among the best five,
but only a few times in top or second place. In effect, two or three of 52
competitors often scored better than the correct one by chance alone.
Extrapolation suggests that, given a set of a few thousand fragments, a
human might have to examine a few hundred matches per fragment to determine
the correct one -- a ten-fold improvement over a human-only approach, but
still very time-consuming. For a satisfactory outcome, we want to narrow the
field, allowing a human researcher to inspect only a few matches per
fragment.
Incidentally, these experiments did suggest that letter-sequences play only a
secondary role in the scoring, shape being the most important feature. This
is perhaps not surprising since letter-sequences can be assessed only on
lines which fit together.

Refinement
In Levison and Wu (1999), the left- and right-profiles of each fragment are
represented by a sequence of measurements from an arbitrary vertical datum
line. At first sight, it is the step-like nature of the edges which causes
the problem, allowing too many good fits between unrelated fragments. In
fact, however, any shape represented by a finite sequence of measurements
can be viewed as step-like. The true problem is the granularity of the
measurements: a single measure per line of text, with a (fixed) character
width, about 2.5 mm, as the horizontal unit.
In new experiments, a further paragraph of Alice was
fragmented and three measurements were made for each line of text to the
nearest 0.5 mm. Once again each fragment was compared with every other in
all different vertical positions. The scoring algorithm measured the total
gap between lines when two fragments were "just touching". This initial
measure was adjusted according to the number of lines which contributed to
it, and the final score biased to reward juxtapositions where several lines
fitted closely but not exactly. In the set of 61 new fragments from the
paragraph, 41 correct matches occurred in first place, with five each in
second and third, a major improvement on the earlier results.

Tear and Wear
Do these improvements persist if the number of fragments is much higher? With
many correct fits now occurring in first place, the rough extrapolation used
earlier no longer applies. Although few good shape matches may appear by
chance among 61 fragments, they may well arise among thousands. To answer
the question we need to experiment with much larger sets. Unfortunately no
such sets are available, while tearing and measuring thousands of artificial
fragments is not an enticing prospect. We have therefore created sets of
"virtual fragments".
The comparison process uses only the left- and right-profiles of each
fragment. We therefore generate vertical "tears" which form the right-
profiles of one group of fragments and the left-profiles of an adjacent one.
A tear is simply a sequence of measurements, each randomly displaced from
the one above. The displacements are chosen from a non- uniform
distribution, small changes being common, larger ones less frequent. In
fact, the actual distribution determines how rough or smooth the edges of
the virtual fragments are.
In practice, the shapes of these virtual fragments may be too perfect. Since
the right-profile of a fragment and the left-profile of its neighbour come
from the same tear, the correct match will involve no gap at all, helping to
ensure that it is always the best. We therefore submit the virtual fragments
to a process of "wear". Each profile measurement is altered by a random
amount to reduce the quality of the fit between correctly matching
profiles.

Results
Many sets of fragments ranging from about 60 to 5400 in number were simulated
and compared using the matching process. Typical results are shown in Table
1. For sets of around 60 fragments, the average number of correct matches
occurring in the first three places is around 88%, very close to the results
obtained for the paragraph from Alice. As expected, this percentage
diminishes as the size of the set increases. For 3600 fragments, it is still
about 66%. In other words, if we have a set of 3600 fragments whose profiles
do not deviate substantially from our virtual sets, we expect to find the
correct match among our three top choices about two-thirds of the time. This
meets the objective set earlier.
In fact, we can expect even better results if we apply the process in both
directions, and use it interactively, so that the human scholar confirms
some matches and thereby eliminates some profiles from later comparisons.
The comparison program is written in C, and was run on a 500Mhz PC. Table 2
shows the time taken to carry out the comparisons for sets of different
sizes. In principle, the time should be proportional to the square of the
number of fragments, and this is closely borne out by the last three
observations. (The processing portion of the two smaller cases is swamped by
the initialization and output portions of the program.) Even for very large
sets the process is feasible on current computers.

Table 1. Typical results of the comparison process.

number offragments
number ofcorrect matchesamong top three
percentage

62
50
80.7%

62
56
90.3%

58
54
93.1%

average

87.9%

199
152
76.4%

181
157
86.7%

193
146
75.6%

average

79.4%

406
303
74.6%

407
284
69.8%

383
297
77.5%

average

73.9%

3636
2443
67.2%

3667
2448
66.8%

3627
2412
66.5%

average

66.8%

Table 2. Timings for sets of different sizes

fragments
time (seconds)

60
1

111
1.5

272
4

1084
48

5381
1100 (= 18 minutes 20 seconds)

Further Stages
The comparison process is the central component of the reconstruction task.
It permits many of the fragments to be combined into pairs and, by
extension, into horizontal bands, from which scholars can easily complete
the reconstruction. A complete program suite might go further, making use of
digital photographs to display best fits for human judgement, applying the
comparison process in the vertical direction to the horizontal bands, and so
on. One time-consuming activity remains -- that of accurately measuring
fragments to obtain their profiles. We have therefore investigated the
possibility of deriving measurements directly from digital photographs.
Some of our artificial fragments were photographed horizontally against a
bright red ground using a digital camera. For this discussion we assume that
the text is black on white. Our purpose is simply to distinguish three kinds
of pixel: text, "paper" and background. If the text is faded or the material
dirty, there are well-known digital processes for "cleaning" the image.
The horizontal rows of pixels are scanned, and the numbers of black pixels in
each row are counted. In principle, the numbers will be near zero for the
inter-line gaps, higher for the bodies of text lines, and intermediate for
rows corresponding to risers and descenders if any. This lets us locate the
pixel rows which delimit the body of each text line; after which, scanning
the top, middle and bottom rows of each line to find where red pixels stop
and start again allows us to compute the desired measurements. Experiments
confirm that this holds true in practice.
In essence, then, the processes described here provide a computer solution to
Problem B.

References

M.
Levison

The Siting of Fragments

Computer Journal

7

275-277
1965

M.
Levison

The Computer in Literary Studies

A.
D.
Booth

Machine Translation

Amsterdam
North-Holland Publishing Company
1967
173-194

M.
Levison

J.
Wu

The Jigsaw Puzzle Problem Revisited

ACH-ALLC 1999, Charlottesville

1999

(Conference abstracts, pp 106-111.)

J.
A.
Ogden

The siting of papyrus fragments: an experimental
application of digital computers

PhD Thesis 2195

University of Glasgow
1969

If this content appears in violation of your intellectual property rights, or you see errors or omissions, please reach out to Scott B. Weingart to discuss removing or amending the materials.

Conference Info

In review

ACH/ALLC / ACH/ICCH / ALLC/EADH - 2002
"New Directions in Humanities Computing"

Hosted at Universität Tübingen (University of Tubingen / Tuebingen)

Tübingen, Germany

July 23, 2002 - July 28, 2008

72 works by 136 authors indexed

Affiliations need to be double-checked.

Conference website: http://web.archive.org/web/20041117094331/http://www.uni-tuebingen.de/allcach2002/

Series: ALLC/EADH (29), ACH/ICCH (22), ACH/ALLC (14)

Organizers: ACH, ALLC

Tags
  • Keywords: None
  • Language: English
  • Topics: None