1 /* PSPP - a program for statistical analysis.
2 Copyright (C) 2009, 2010, 2011, 2013, 2014 Free Software Foundation, Inc.
4 This program is free software: you can redistribute it and/or modify
5 it under the terms of the GNU General Public License as published by
6 the Free Software Foundation, either version 3 of the License, or
7 (at your option) any later version.
9 This program is distributed in the hope that it will be useful,
10 but WITHOUT ANY WARRANTY; without even the implied warranty of
11 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 GNU General Public License for more details.
14 You should have received a copy of the GNU General Public License
15 along with this program. If not, see <http://www.gnu.org/licenses/>. */
25 #include "libpspp/assertion.h"
26 #include "libpspp/hash-functions.h"
27 #include "libpspp/hmap.h"
28 #include "output/render.h"
29 #include "output/tab.h"
30 #include "output/table-item.h"
31 #include "output/table.h"
33 #include "gl/minmax.h"
34 #include "gl/xalloc.h"
37 #define _(msgid) gettext (msgid)
39 /* This file uses TABLE_HORZ and TABLE_VERT enough to warrant abbreviating. */
43 /* A layout for rendering a specific table on a specific device.
45 May represent the layout of an entire table presented to
46 render_page_create(), or a rectangular subregion of a table broken out using
47 render_break_next() to allow a table to be broken across multiple pages.
49 A page's size is not limited to the size passed in as part of render_params.
50 render_pager breaks a render_page into smaller render_pages that will fit in
51 the available space. */
54 const struct render_params *params; /* Parameters of the target device. */
55 struct table *table; /* Table rendered. */
58 /* Local copies of table->n and table->h, for convenience. */
60 int h[TABLE_N_AXES][2];
62 /* cp[H] represents x positions within the table.
64 cp[H][1] = the width of the leftmost vertical rule.
65 cp[H][2] = cp[H][1] + the width of the leftmost column.
66 cp[H][3] = cp[H][2] + the width of the second-from-left vertical rule.
68 cp[H][2 * nc] = x position of the rightmost vertical rule.
69 cp[H][2 * nc + 1] = total table width including all rules.
71 Similarly, cp[V] represents y positions within the table.
73 cp[V][1] = the height of the topmost horizontal rule.
74 cp[V][2] = cp[V][1] + the height of the topmost row.
75 cp[V][3] = cp[V][2] + the height of the second-from-top horizontal rule.
77 cp[V][2 * nr] = y position of the bottommost horizontal rule.
78 cp[V][2 * nr + 1] = total table height including all rules.
80 Rules and columns can have width or height 0, in which case consecutive
81 values in this array are equal. */
82 int *cp[TABLE_N_AXES];
84 /* render_break_next() can break a table such that some cells are not fully
85 contained within a render_page. This will happen if a cell is too wide
86 or two tall to fit on a single page, or if a cell spans multiple rows or
87 columns and the page only includes some of those rows or columns.
89 This hash table contains "struct render_overflow"s that represents each
90 such cell that doesn't completely fit on this page.
92 Each overflow cell borders at least one header edge of the table and may
93 border more. (A single table cell that is so large that it fills the
94 entire page can overflow on all four sides!) */
95 struct hmap overflows;
97 /* Contains "struct render_footnote"s, one for each cell with one or more
100 'n_footnotes' is the number of footnotes in the table. There might be
101 more than hmap_count(&page->footnotes) because there can be more than
102 one footnote in a cell. */
103 struct hmap footnotes;
106 /* If a single column (or row) is too wide (or tall) to fit on a page
107 reasonably, then render_break_next() will split a single row or column
108 across multiple render_pages. This member indicates when this has
111 is_edge_cutoff[H][0] is true if pixels have been cut off the left side
112 of the leftmost column in this page, and false otherwise.
114 is_edge_cutoff[H][1] is true if pixels have been cut off the right side
115 of the rightmost column in this page, and false otherwise.
117 is_edge_cutoff[V][0] and is_edge_cutoff[V][1] are similar for the top
118 and bottom of the table.
120 The effect of is_edge_cutoff is to prevent rules along the edge in
121 question from being rendered.
123 When is_edge_cutoff is true for a given edge, the 'overflows' hmap will
124 contain a node for each cell along that edge. */
125 bool is_edge_cutoff[TABLE_N_AXES][2];
127 /* If part of a joined cell would be cut off by breaking a table along
128 'axis' at the rule with offset 'z' (where 0 <= z <= n[axis]), then
129 join_crossing[axis][z] is the thickness of the rule that would be cut
132 This is used to know to allocate extra space for breaking at such a
133 position, so that part of the cell's content is not lost.
135 This affects breaking a table only when headers are present. When
136 headers are not present, the rule's thickness is used for cell content,
137 so no part of the cell's content is lost (and in fact it is duplicated
138 across both pages). */
139 int *join_crossing[TABLE_N_AXES];
142 static struct render_page *render_page_create (const struct render_params *,
143 const struct table *);
145 struct render_page *render_page_ref (const struct render_page *page_);
146 static void render_page_unref (struct render_page *);
148 /* Returns the offset in struct render_page's cp[axis] array of the rule with
149 index RULE_IDX. That is, if RULE_IDX is 0, then the offset is that of the
150 leftmost or topmost rule; if RULE_IDX is 1, then the offset is that of the
151 next rule to the right (or below); and so on. */
153 rule_ofs (int rule_idx)
158 /* Returns the offset in struct render_page's cp[axis] array of the rule with
159 index RULE_IDX_R, which counts from the right side (or bottom) of the page
160 left (or up), according to whether AXIS is H or V, respectively. That is,
161 if RULE_IDX_R is 0, then the offset is that of the rightmost or bottommost
162 rule; if RULE_IDX is 1, then the offset is that of the next rule to the left
163 (or above); and so on. */
165 rule_ofs_r (const struct render_page *page, int axis, int rule_idx_r)
167 return (page->n[axis] - rule_idx_r) * 2;
170 /* Returns the offset in struct render_page's cp[axis] array of the cell with
171 index CELL_IDX. That is, if CELL_IDX is 0, then the offset is that of the
172 leftmost or topmost cell; if CELL_IDX is 1, then the offset is that of the
173 next cell to the right (or below); and so on. */
175 cell_ofs (int cell_idx)
177 return cell_idx * 2 + 1;
180 /* Returns the width of PAGE along AXIS from OFS0 to OFS1, exclusive. */
182 axis_width (const struct render_page *page, int axis, int ofs0, int ofs1)
184 return page->cp[axis][ofs1] - page->cp[axis][ofs0];
187 /* Returns the width of the headers in PAGE along AXIS. */
189 headers_width (const struct render_page *page, int axis)
191 int h0 = page->h[axis][0];
192 int w0 = axis_width (page, axis, rule_ofs (0), cell_ofs (h0));
193 int n = page->n[axis];
194 int h1 = page->h[axis][1];
195 int w1 = axis_width (page, axis, rule_ofs_r (page, axis, h1), cell_ofs (n));
199 /* Returns the width of cell X along AXIS in PAGE. */
201 cell_width (const struct render_page *page, int axis, int x)
203 return axis_width (page, axis, cell_ofs (x), cell_ofs (x) + 1);
206 /* Returns the width of rule X along AXIS in PAGE. */
208 rule_width (const struct render_page *page, int axis, int x)
210 return axis_width (page, axis, rule_ofs (x), rule_ofs (x) + 1);
213 /* Returns the width of rule X along AXIS in PAGE. */
215 rule_width_r (const struct render_page *page, int axis, int x)
217 int ofs = rule_ofs_r (page, axis, x);
218 return axis_width (page, axis, ofs, ofs + 1);
221 /* Returns the width of cells X0 through X1, exclusive, along AXIS in PAGE. */
223 joined_width (const struct render_page *page, int axis, int x0, int x1)
225 return axis_width (page, axis, cell_ofs (x0), cell_ofs (x1) - 1);
228 /* Returns the width of the widest cell, excluding headers, along AXIS in
231 max_cell_width (const struct render_page *page, int axis)
233 int n = page->n[axis];
234 int x0 = page->h[axis][0];
235 int x1 = n - page->h[axis][1];
239 for (x = x0; x < x1; x++)
241 int w = cell_width (page, axis, x);
248 /* A cell that doesn't completely fit on the render_page. */
249 struct render_overflow
251 struct hmap_node node; /* In render_page's 'overflows' hmap. */
253 /* Occupied region of page.
255 d[H][0] is the leftmost column.
256 d[H][1] is the rightmost column, plus 1.
257 d[V][0] is the top row.
258 d[V][1] is the bottom row, plus 1.
260 The cell in its original table might occupy a larger region. This
261 member reflects the size of the cell in the current render_page, after
262 trimming off any rows or columns due to page-breaking. */
265 /* The space that has been trimmed off the cell:
267 overflow[H][0]: space trimmed off its left side.
268 overflow[H][1]: space trimmed off its right side.
269 overflow[V][0]: space trimmed off its top.
270 overflow[V][1]: space trimmed off its bottom.
272 During rendering, this information is used to position the rendered
273 portion of the cell within the available space.
275 When a cell is rendered, sometimes it is permitted to spill over into
276 space that is ordinarily reserved for rules. Either way, this space is
277 still included in overflow values.
279 Suppose, for example, that a cell that joins 2 columns has a width of 60
280 pixels and content "abcdef", that the 2 columns that it joins have
281 widths of 20 and 30 pixels, respectively, and that therefore the rule
282 between the two joined columns has a width of 10 (20 + 10 + 30 = 60).
283 It might render like this, if each character is 10x10, and showing a few
284 extra table cells for context:
292 If this render_page is broken at the rule that separates "gh" from
293 "ijk", then the page that contains the left side of the "abcdef" cell
294 will have overflow[H][1] of 10 + 30 = 40 for its portion of the cell,
295 and the page that contains the right side of the cell will have
296 overflow[H][0] of 20 + 10 = 30. The two resulting pages would look like
314 int overflow[TABLE_N_AXES][2];
317 /* Returns a hash value for (,Y). */
319 hash_cell (int x, int y)
321 return hash_int (x + (y << 16), 0);
324 /* Searches PAGE's set of render_overflow for one whose top-left cell is
325 (X,Y). Returns it, if there is one, otherwise a null pointer. */
326 static const struct render_overflow *
327 find_overflow (const struct render_page *page, int x, int y)
329 if (!hmap_is_empty (&page->overflows))
331 const struct render_overflow *of;
333 HMAP_FOR_EACH_WITH_HASH (of, struct render_overflow, node,
334 hash_cell (x, y), &page->overflows)
335 if (x == of->d[H] && y == of->d[V])
343 struct render_footnote
345 struct hmap_node node;
347 /* The area of the table covered by the cell that has the footnote.
349 d[H][0] is the leftmost column.
350 d[H][1] is the rightmost column, plus 1.
351 d[V][0] is the top row.
352 d[V][1] is the bottom row, plus 1.
354 The cell in its original table might occupy a larger region. This
355 member reflects the size of the cell in the current render_page, after
356 trimming off any rows or columns due to page-breaking. */
357 int d[TABLE_N_AXES][2];
359 /* The index of the first footnote in the cell. */
364 count_footnotes (const struct table_cell *cell)
370 for (i = 0; i < cell->n_contents; i++)
371 n += cell->contents[i].n_footnotes;
376 find_footnote_idx (const struct table_cell *cell, const struct hmap *footnotes)
378 const struct render_footnote *f;
380 if (!count_footnotes (cell))
383 HMAP_FOR_EACH_WITH_HASH (f, struct render_footnote, node,
384 hash_cell (cell->d[H][0], cell->d[V][0]), footnotes)
385 if (f->d[H][0] == cell->d[H][0] && f->d[V][0] == cell->d[V][0])
391 /* Row or column dimensions. Used to figure the size of a table in
392 render_page_create() and discarded after that. */
395 /* Width without considering rows (or columns) that span more than one (or
399 /* Width taking spanned rows (or columns) into consideration. */
403 /* Modifies the 'width' members of the N elements of ROWS so that their sum,
404 when added to rule widths RULES[1] through RULES[N - 1] inclusive, is at
407 distribute_spanned_width (int width,
408 struct render_row *rows, const int *rules, int n)
414 /* Sum up the unspanned widths of the N rows for use as weights. */
416 for (x = 0; x < n; x++)
417 total_unspanned += rows[x].unspanned;
418 for (x = 0; x < n - 1; x++)
419 total_unspanned += rules[x + 1];
420 if (total_unspanned >= width)
423 /* The algorithm used here is based on the following description from HTML 4:
425 For cells that span multiple columns, a simple approach consists of
426 apportioning the min/max widths evenly to each of the constituent
427 columns. A slightly more complex approach is to use the min/max
428 widths of unspanned cells to weight how spanned widths are
429 apportioned. Experiments suggest that a blend of the two approaches
430 gives good results for a wide range of tables.
432 We blend the two approaches half-and-half, except that we cannot use the
433 unspanned weights when 'total_unspanned' is 0 (because that would cause a
436 This implementation uses floating-point types and operators, but all the
437 values involved are integers. For integers smaller than 53 bits, this
438 should not lose any precision, and it should degrade gracefully for larger
441 The calculation we want to do is this:
444 w1 = width * (column's unspanned width) / (total unspanned width)
445 (column's width) = (w0 + w1) / 2
447 We implement it as a precise calculation in integers by multiplying w0 and
448 w1 by the common denominator of all three calculations (d), dividing that
449 out in the column width calculation, and then keeping the remainder for
452 (We actually compute the unspanned width of a column as twice the
453 unspanned width, plus the width of the rule on the left, plus the width of
454 the rule on the right. That way each rule contributes to both the cell on
455 its left and on its right.)
458 d1 = 2.0 * (total_unspanned > 0 ? total_unspanned : 1.0);
460 if (total_unspanned > 0)
463 for (x = 0; x < n; x++)
466 if (total_unspanned > 0)
468 double unspanned = rows[x].unspanned * 2.0;
470 unspanned += rules[x + 1];
472 unspanned += rules[x];
473 w += width * unspanned * d0;
476 rows[x].width = MAX (rows[x].width, w / d);
477 w -= rows[x].width * d;
481 /* Initializes PAGE->cp[AXIS] from the row widths in ROWS and the rule widths
484 accumulate_row_widths (const struct render_page *page, enum table_axis axis,
485 const struct render_row *rows, const int *rules)
487 int n = page->n[axis];
493 for (z = 0; z < n; z++)
495 cp[1] = cp[0] + rules[z];
496 cp[2] = cp[1] + rows[z].width;
499 cp[1] = cp[0] + rules[n];
502 /* Returns the sum of widths of the N ROWS and N+1 RULES. */
504 calculate_table_width (int n, const struct render_row *rows, int *rules)
510 for (x = 0; x < n; x++)
511 width += rows[x].width;
512 for (x = 0; x <= n; x++)
518 /* Rendering utility functions. */
520 /* Returns the line style to use for drawing a rule of the given TYPE. */
521 static enum render_line_style
522 rule_to_render_type (unsigned char type)
528 return RENDER_LINE_NONE;
530 return RENDER_LINE_SINGLE;
532 return RENDER_LINE_DOUBLE;
538 /* Returns the width of the rule in TABLE that is at offset Z along axis A, if
539 rendered with PARAMS. */
541 measure_rule (const struct render_params *params, const struct table *table,
542 enum table_axis a, int z)
544 enum table_axis b = !a;
549 /* Determine all types of rules that are present, as a bitmap in 'rules'
550 where rule type 't' is present if bit 2**t is set. */
553 for (d[b] = 0; d[b] < table->n[b]; d[b]++)
554 rules |= 1u << table_get_rule (table, a, d[H], d[V]);
556 /* Calculate maximum width of the rules that are present. */
558 if (rules & (1u << TAL_1)
559 || (z > 0 && z < table->n[a] && rules & (1u << TAL_GAP)))
560 width = params->line_widths[a][RENDER_LINE_SINGLE];
561 if (rules & (1u << TAL_2))
562 width = MAX (width, params->line_widths[a][RENDER_LINE_DOUBLE]);
566 /* Allocates and returns a new render_page using PARAMS and TABLE. Allocates
567 space for all of the members of the new page, but the caller must initialize
568 the 'cp' member itself. */
569 static struct render_page *
570 render_page_allocate (const struct render_params *params,
573 struct render_page *page;
576 page = xmalloc (sizeof *page);
577 page->params = params;
580 page->n[H] = table->n[H];
581 page->n[V] = table->n[V];
582 page->h[H][0] = table->h[H][0];
583 page->h[H][1] = table->h[H][1];
584 page->h[V][0] = table->h[V][0];
585 page->h[V][1] = table->h[V][1];
587 for (i = 0; i < TABLE_N_AXES; i++)
589 page->cp[i] = xmalloc ((2 * page->n[i] + 2) * sizeof *page->cp[i]);
590 page->join_crossing[i] = xzalloc ((page->n[i] + 1) * sizeof *page->join_crossing[i]);
593 hmap_init (&page->overflows);
594 hmap_init (&page->footnotes);
595 page->n_footnotes = 0;
596 memset (page->is_edge_cutoff, 0, sizeof page->is_edge_cutoff);
601 /* Allocates and returns a new render_page for PARAMS and TABLE, initializing
602 cp[H] in the new page from ROWS and RULES. The caller must still initialize
604 static struct render_page *
605 create_page_with_exact_widths (const struct render_params *params,
607 const struct render_row *rows, int *rules)
609 struct render_page *page = render_page_allocate (params, table);
610 accumulate_row_widths (page, H, rows, rules);
614 /* Allocates and returns a new render_page for PARAMS and TABLE.
616 Initializes cp[H] in the new page by setting the width of each row 'i' to
617 somewhere between the minimum cell width ROW_MIN[i].width and the maximum
618 ROW_MAX[i].width. Sets the width of rules to those in RULES.
620 W_MIN is the sum of ROWS_MIN[].width.
622 W_MAX is the sum of ROWS_MAX[].width.
624 The caller must still initialize cp[V]. */
625 static struct render_page *
626 create_page_with_interpolated_widths (const struct render_params *params,
628 const struct render_row *rows_min,
629 const struct render_row *rows_max,
630 int w_min, int w_max, const int *rules)
632 /* This implementation uses floating-point types and operators, but all the
633 values involved are integers. For integers smaller than 53 bits, this
634 should not lose any precision, and it should degrade gracefully for larger
636 const int n = table->n[H];
637 const double avail = params->size[H] - w_min;
638 const double wanted = w_max - w_min;
639 struct render_page *page;
646 page = render_page_allocate (params, table);
650 w = (int) wanted / 2;
651 for (x = 0; x < n; x++)
655 w += avail * (rows_max[x].width - rows_min[x].width);
659 cph[1] = cph[0] + rules[x];
660 cph[2] = cph[1] + rows_min[x].width + extra;
663 cph[1] = cph[0] + rules[n];
665 assert (page->cp[H][n * 2 + 1] == params->size[H]);
671 set_join_crossings (struct render_page *page, enum table_axis axis,
672 const struct table_cell *cell, int *rules)
676 for (z = cell->d[axis][0] + 1; z <= cell->d[axis][1] - 1; z++)
677 page->join_crossing[axis][z] = rules[z];
680 /* Creates and returns a new render_page for rendering TABLE on a device
683 The new render_page will be suitable for rendering on a device whose page
684 size is PARAMS->size, but the caller is responsible for actually breaking it
685 up to fit on such a device, using the render_break abstraction. */
686 static struct render_page *
687 render_page_create (const struct render_params *params,
688 const struct table *table_)
690 struct render_page *page;
693 struct render_row *columns[2];
694 struct render_row *rows;
696 int *rules[TABLE_N_AXES];
697 struct hmap footnotes;
702 enum table_axis axis;
704 table = table_ref (table_);
705 nc = table_nc (table);
706 nr = table_nr (table);
708 /* Figure out rule widths. */
709 for (axis = 0; axis < TABLE_N_AXES; axis++)
711 int n = table->n[axis] + 1;
714 rules[axis] = xnmalloc (n, sizeof *rules);
715 for (z = 0; z < n; z++)
716 rules[axis][z] = measure_rule (params, table, axis, z);
719 /* Calculate minimum and maximum widths of cells that do not
720 span multiple columns. Assign footnote markers. */
721 hmap_init (&footnotes);
723 for (i = 0; i < 2; i++)
724 columns[i] = xzalloc (nc * sizeof *columns[i]);
725 for (y = 0; y < nr; y++)
726 for (x = 0; x < nc; )
728 struct table_cell cell;
730 table_get_cell (table, x, y, &cell);
731 if (y == cell.d[V][0])
735 if (table_cell_colspan (&cell) == 1)
740 params->measure_cell_width (params->aux, &cell, footnote_idx,
742 for (i = 0; i < 2; i++)
743 if (columns[i][x].unspanned < w[i])
744 columns[i][x].unspanned = w[i];
747 n = count_footnotes (&cell);
750 struct render_footnote *f = xmalloc (sizeof *f);
751 f->d[H][0] = cell.d[H][0];
752 f->d[H][1] = cell.d[H][1];
753 f->d[V][0] = cell.d[V][0];
754 f->d[V][1] = cell.d[V][1];
755 f->idx = footnote_idx;
756 hmap_insert (&footnotes, &f->node, hash_cell (x, y));
762 table_cell_free (&cell);
765 /* Distribute widths of spanned columns. */
766 for (i = 0; i < 2; i++)
767 for (x = 0; x < nc; x++)
768 columns[i][x].width = columns[i][x].unspanned;
769 for (y = 0; y < nr; y++)
770 for (x = 0; x < nc; )
772 struct table_cell cell;
774 table_get_cell (table, x, y, &cell);
775 if (y == cell.d[V][0] && table_cell_colspan (&cell) > 1)
779 params->measure_cell_width (params->aux, &cell,
780 find_footnote_idx (&cell, &footnotes),
782 for (i = 0; i < 2; i++)
783 distribute_spanned_width (w[i], &columns[i][cell.d[H][0]],
784 rules[H], table_cell_colspan (&cell));
787 table_cell_free (&cell);
790 /* Decide final column widths. */
791 for (i = 0; i < 2; i++)
792 table_widths[i] = calculate_table_width (table_nc (table),
793 columns[i], rules[H]);
794 if (table_widths[MAX] <= params->size[H])
796 /* Fits even with maximum widths. Use them. */
797 page = create_page_with_exact_widths (params, table, columns[MAX],
800 else if (table_widths[MIN] <= params->size[H])
802 /* Fits with minimum widths, so distribute the leftover space. */
803 page = create_page_with_interpolated_widths (
804 params, table, columns[MIN], columns[MAX],
805 table_widths[MIN], table_widths[MAX], rules[H]);
809 /* Doesn't fit even with minimum widths. Assign minimums for now, and
810 later we can break it horizontally into multiple pages. */
811 page = create_page_with_exact_widths (params, table, columns[MIN],
815 /* Calculate heights of cells that do not span multiple rows. */
816 rows = xzalloc (nr * sizeof *rows);
817 for (y = 0; y < nr; y++)
819 for (x = 0; x < nc; )
821 struct render_row *r = &rows[y];
822 struct table_cell cell;
824 table_get_cell (table, x, y, &cell);
825 if (y == cell.d[V][0])
827 if (table_cell_rowspan (&cell) == 1)
829 int w = joined_width (page, H, cell.d[H][0], cell.d[H][1]);
830 int h = params->measure_cell_height (
831 params->aux, &cell, find_footnote_idx (&cell, &footnotes), w);
832 if (h > r->unspanned)
833 r->unspanned = r->width = h;
836 set_join_crossings (page, V, &cell, rules[V]);
838 if (table_cell_colspan (&cell) > 1)
839 set_join_crossings (page, H, &cell, rules[H]);
842 table_cell_free (&cell);
845 for (i = 0; i < 2; i++)
848 /* Distribute heights of spanned rows. */
849 for (y = 0; y < nr; y++)
850 for (x = 0; x < nc; )
852 struct table_cell cell;
854 table_get_cell (table, x, y, &cell);
855 if (y == cell.d[V][0] && table_cell_rowspan (&cell) > 1)
857 int w = joined_width (page, H, cell.d[H][0], cell.d[H][1]);
858 int h = params->measure_cell_height (
859 params->aux, &cell, find_footnote_idx (&cell, &footnotes), w);
860 distribute_spanned_width (h, &rows[cell.d[V][0]], rules[V],
861 table_cell_rowspan (&cell));
864 table_cell_free (&cell);
867 /* Decide final row heights. */
868 accumulate_row_widths (page, V, rows, rules[V]);
871 /* Measure headers. If they are "too big", get rid of them. */
872 for (axis = 0; axis < TABLE_N_AXES; axis++)
874 int hw = headers_width (page, axis);
875 if (hw * 2 >= page->params->size[axis]
876 || hw + max_cell_width (page, axis) > page->params->size[axis])
878 page->table = table_unshare (page->table);
879 page->table->h[axis][0] = page->table->h[axis][1] = 0;
880 page->h[axis][0] = page->h[axis][1] = 0;
884 hmap_swap (&page->footnotes, &footnotes);
885 hmap_destroy (&footnotes);
886 page->n_footnotes = footnote_idx;
894 /* Increases PAGE's reference count. */
896 render_page_ref (const struct render_page *page_)
898 struct render_page *page = CONST_CAST (struct render_page *, page_);
903 /* Decreases PAGE's reference count and destroys PAGE if this causes the
904 reference count to fall to zero. */
906 render_page_unref (struct render_page *page)
908 if (page != NULL && --page->ref_cnt == 0)
911 struct render_overflow *overflow, *next;
913 HMAP_FOR_EACH_SAFE (overflow, next, struct render_overflow, node,
916 hmap_destroy (&page->overflows);
918 table_unref (page->table);
920 for (i = 0; i < TABLE_N_AXES; ++i)
922 free (page->join_crossing[i]);
930 /* Returns the size of PAGE along AXIS. (This might be larger than the page
931 size specified in the parameters passed to render_page_create(). Use a
932 render_break to break up a render_page into page-sized chunks.) */
934 render_page_get_size (const struct render_page *page, enum table_axis axis)
936 return page->cp[axis][page->n[axis] * 2 + 1];
940 render_page_get_best_breakpoint (const struct render_page *page, int height)
944 /* If there's no room for at least the top row and the rules above and below
945 it, don't include any of the table. */
946 if (page->cp[V][3] > height)
949 /* Otherwise include as many rows and rules as we can. */
950 for (y = 5; y <= 2 * page->n[V] + 1; y += 2)
951 if (page->cp[V][y] > height)
952 return page->cp[V][y - 2];
956 /* Drawing render_pages. */
958 static inline enum render_line_style
959 get_rule (const struct render_page *page, enum table_axis axis,
960 const int d[TABLE_N_AXES])
962 return rule_to_render_type (table_get_rule (page->table,
963 axis, d[H] / 2, d[V] / 2));
973 render_direction_rtl (void)
975 /* TRANSLATORS: Do not translate this string. If the script of your language
976 reads from right to left (eg Persian, Arabic, Hebrew etc), then replace
977 this string with "output-direction-rtl". Otherwise either leave it
978 untranslated or copy it verbatim. */
979 const char *dir = _("output-direction-ltr");
980 if ( 0 == strcmp ("output-direction-rtl", dir))
983 if ( 0 != strcmp ("output-direction-ltr", dir))
984 fprintf (stderr, "This localisation has been incorrectly translated. Complain to the translator.\n");
990 render_rule (const struct render_page *page, const int ofs[TABLE_N_AXES],
991 const int d[TABLE_N_AXES])
993 enum render_line_style styles[TABLE_N_AXES][2];
996 for (a = 0; a < TABLE_N_AXES; a++)
998 enum table_axis b = !a;
1000 styles[a][0] = styles[a][1] = RENDER_LINE_NONE;
1003 || (page->is_edge_cutoff[a][0] && d[a] == 0)
1004 || (page->is_edge_cutoff[a][1] && d[a] == page->n[a] * 2))
1011 int e[TABLE_N_AXES];
1015 styles[a][0] = get_rule (page, a, e);
1018 if (d[b] / 2 < page->table->n[b])
1019 styles[a][1] = get_rule (page, a, d);
1022 styles[a][0] = styles[a][1] = get_rule (page, a, d);
1025 if (styles[H][0] != RENDER_LINE_NONE || styles[H][1] != RENDER_LINE_NONE
1026 || styles[V][0] != RENDER_LINE_NONE || styles[V][1] != RENDER_LINE_NONE)
1028 int bb[TABLE_N_AXES][2];
1030 bb[H][0] = ofs[H] + page->cp[H][d[H]];
1031 bb[H][1] = ofs[H] + page->cp[H][d[H] + 1];
1032 if (render_direction_rtl ())
1034 int temp = bb[H][0];
1035 bb[H][0] = render_page_get_size (page, H) - bb[H][1];
1036 bb[H][1] = render_page_get_size (page, H) - temp;
1038 bb[V][0] = ofs[V] + page->cp[V][d[V]];
1039 bb[V][1] = ofs[V] + page->cp[V][d[V] + 1];
1040 page->params->draw_line (page->params->aux, bb, styles);
1045 render_cell (const struct render_page *page, const int ofs[TABLE_N_AXES],
1046 const struct table_cell *cell)
1048 const struct render_overflow *of;
1049 int bb[TABLE_N_AXES][2];
1050 int clip[TABLE_N_AXES][2];
1052 bb[H][0] = clip[H][0] = ofs[H] + page->cp[H][cell->d[H][0] * 2 + 1];
1053 bb[H][1] = clip[H][1] = ofs[H] + page->cp[H][cell->d[H][1] * 2];
1054 if (render_direction_rtl ())
1056 int temp = bb[H][0];
1057 bb[H][0] = clip[H][0] = render_page_get_size (page, H) - bb[H][1];
1058 bb[H][1] = clip[H][1] = render_page_get_size (page, H) - temp;
1060 bb[V][0] = clip[V][0] = ofs[V] + page->cp[V][cell->d[V][0] * 2 + 1];
1061 bb[V][1] = clip[V][1] = ofs[V] + page->cp[V][cell->d[V][1] * 2];
1063 of = find_overflow (page, cell->d[H][0], cell->d[V][0]);
1066 enum table_axis axis;
1068 for (axis = 0; axis < TABLE_N_AXES; axis++)
1070 if (of->overflow[axis][0])
1072 bb[axis][0] -= of->overflow[axis][0];
1073 if (cell->d[axis][0] == 0 && !page->is_edge_cutoff[axis][0])
1074 clip[axis][0] = ofs[axis] + page->cp[axis][cell->d[axis][0] * 2];
1076 if (of->overflow[axis][1])
1078 bb[axis][1] += of->overflow[axis][1];
1079 if (cell->d[axis][1] == page->n[axis] && !page->is_edge_cutoff[axis][1])
1080 clip[axis][1] = ofs[axis] + page->cp[axis][cell->d[axis][1] * 2 + 1];
1085 page->params->draw_cell (page->params->aux, cell,
1086 find_footnote_idx (cell, &page->footnotes), bb, clip);
1089 /* Draws the cells of PAGE indicated in BB. */
1091 render_page_draw_cells (const struct render_page *page,
1092 int ofs[TABLE_N_AXES], int bb[TABLE_N_AXES][2])
1096 for (y = bb[V][0]; y < bb[V][1]; y++)
1097 for (x = bb[H][0]; x < bb[H][1]; )
1098 if (is_rule (x) || is_rule (y))
1100 int d[TABLE_N_AXES];
1103 render_rule (page, ofs, d);
1108 struct table_cell cell;
1110 table_get_cell (page->table, x / 2, y / 2, &cell);
1111 if (y / 2 == bb[V][0] / 2 || y / 2 == cell.d[V][0])
1112 render_cell (page, ofs, &cell);
1113 x = rule_ofs (cell.d[H][1]);
1114 table_cell_free (&cell);
1118 /* Renders PAGE, by calling the 'draw_line' and 'draw_cell' functions from the
1119 render_params provided to render_page_create(). */
1121 render_page_draw (const struct render_page *page, int ofs[TABLE_N_AXES])
1123 int bb[TABLE_N_AXES][2];
1126 bb[H][1] = page->n[H] * 2 + 1;
1128 bb[V][1] = page->n[V] * 2 + 1;
1130 render_page_draw_cells (page, ofs, bb);
1133 /* Returns the greatest value i, 0 <= i < n, such that cp[i] <= x0. */
1135 get_clip_min_extent (int x0, const int cp[], int n)
1137 int low, high, best;
1144 int middle = low + (high - low) / 2;
1146 if (cp[middle] <= x0)
1158 /* Returns the least value i, 0 <= i < n, such that cp[i] >= x1. */
1160 get_clip_max_extent (int x1, const int cp[], int n)
1162 int low, high, best;
1169 int middle = low + (high - low) / 2;
1171 if (cp[middle] >= x1)
1172 best = high = middle;
1177 while (best > 0 && cp[best - 1] == cp[best])
1183 /* Renders the cells of PAGE that intersect (X,Y)-(X+W,Y+H), by calling the
1184 'draw_line' and 'draw_cell' functions from the render_params provided to
1185 render_page_create(). */
1187 render_page_draw_region (const struct render_page *page,
1188 int ofs[TABLE_N_AXES], int clip[TABLE_N_AXES][2])
1190 int bb[TABLE_N_AXES][2];
1192 bb[H][0] = get_clip_min_extent (clip[H][0], page->cp[H], page->n[H] * 2 + 1);
1193 bb[H][1] = get_clip_max_extent (clip[H][1], page->cp[H], page->n[H] * 2 + 1);
1194 bb[V][0] = get_clip_min_extent (clip[V][0], page->cp[V], page->n[V] * 2 + 1);
1195 bb[V][1] = get_clip_max_extent (clip[V][1], page->cp[V], page->n[V] * 2 + 1);
1197 render_page_draw_cells (page, ofs, bb);
1200 /* Breaking up tables to fit on a page. */
1202 /* An iterator for breaking render_pages into smaller chunks. */
1205 struct render_page *page; /* Page being broken up. */
1206 enum table_axis axis; /* Axis along which 'page' is being broken. */
1207 int z; /* Next cell along 'axis'. */
1208 int pixel; /* Pixel offset within cell 'z' (usually 0). */
1209 int hw; /* Width of headers of 'page' along 'axis'. */
1212 static int needed_size (const struct render_break *, int cell);
1213 static bool cell_is_breakable (const struct render_break *, int cell);
1214 static struct render_page *render_page_select (const struct render_page *,
1219 /* Initializes render_break B for breaking PAGE along AXIS. */
1221 render_break_init (struct render_break *b, const struct render_page *page,
1222 enum table_axis axis)
1224 b->page = render_page_ref (page);
1226 b->z = page->h[axis][0];
1228 b->hw = headers_width (page, axis);
1231 /* Initializes B as a render_break structure for which
1232 render_break_has_next() always returns false. */
1234 render_break_init_empty (struct render_break *b)
1237 b->axis = TABLE_HORZ;
1243 /* Frees B and unrefs the render_page that it owns. */
1245 render_break_destroy (struct render_break *b)
1249 render_page_unref (b->page);
1254 /* Returns true if B still has cells that are yet to be returned,
1255 false if all of B's page has been processed. */
1257 render_break_has_next (const struct render_break *b)
1259 const struct render_page *page = b->page;
1260 enum table_axis axis = b->axis;
1262 return page != NULL && b->z < page->n[axis] - page->h[axis][1];
1265 /* Returns a new render_page that is up to SIZE pixels wide along B's axis.
1266 Returns a null pointer if B has already been completely broken up, or if
1267 SIZE is too small to reasonably render any cells. The latter will never
1268 happen if SIZE is at least as large as the page size passed to
1269 render_page_create() along B's axis. */
1270 static struct render_page *
1271 render_break_next (struct render_break *b, int size)
1273 const struct render_page *page = b->page;
1274 enum table_axis axis = b->axis;
1275 struct render_page *subpage;
1278 if (!render_break_has_next (b))
1282 for (z = b->z; z < page->n[axis] - page->h[axis][1]; z++)
1284 int needed = needed_size (b, z + 1);
1287 if (cell_is_breakable (b, z))
1289 /* If there is no right header and we render a partial cell on
1290 the right side of the body, then we omit the rightmost rule of
1291 the body. Otherwise the rendering is deceptive because it
1292 looks like the whole cell is present instead of a partial
1295 This is similar to code for the left side in needed_size(). */
1296 int rule_allowance = (page->h[axis][1]
1298 : rule_width (page, axis, z));
1300 /* The amount that, if we added cell 'z', the rendering would
1301 overfill the allocated 'size'. */
1302 int overhang = needed - size - rule_allowance;
1304 /* The width of cell 'z'. */
1305 int cell_size = cell_width (page, axis, z);
1307 /* The amount trimmed off the left side of 'z',
1308 and the amount left to render. */
1309 int cell_ofs = z == b->z ? b->pixel : 0;
1310 int cell_left = cell_size - cell_ofs;
1312 /* A small but visible width. */
1313 int em = page->params->font_size[axis];
1315 /* If some of the cell remains to render,
1316 and there would still be some of the cell left afterward,
1317 then partially render that much of the cell. */
1318 pixel = (cell_left && cell_left > overhang
1319 ? cell_left - overhang + cell_ofs
1322 /* If there would be only a tiny amount of the cell left after
1323 rendering it partially, reduce the amount rendered slightly
1324 to make the output look a little better. */
1325 if (pixel + em > cell_size)
1326 pixel = MAX (pixel - em, 0);
1328 /* If we're breaking vertically, then consider whether the cells
1329 being broken have a better internal breakpoint than the exact
1330 number of pixels available, which might look bad e.g. because
1331 it breaks in the middle of a line of text. */
1332 if (axis == TABLE_VERT && page->params->adjust_break)
1336 for (x = 0; x < page->n[H]; )
1338 struct table_cell cell;
1342 table_get_cell (page->table, x, z, &cell);
1343 w = joined_width (page, H, cell.d[H][0], cell.d[H][1]);
1344 better_pixel = page->params->adjust_break (
1345 page->params->aux, &cell,
1346 find_footnote_idx (&cell, &page->footnotes), w, pixel);
1348 table_cell_free (&cell);
1350 if (better_pixel < pixel)
1352 if (better_pixel > (z == b->z ? b->pixel : 0))
1354 pixel = better_pixel;
1357 else if (better_pixel == 0 && z != b->z)
1370 if (z == b->z && !pixel)
1373 subpage = render_page_select (page, axis, b->z, b->pixel,
1375 pixel ? cell_width (page, axis, z) - pixel
1382 /* Returns the width that would be required along B's axis to render a page
1383 from B's current position up to but not including CELL. */
1385 needed_size (const struct render_break *b, int cell)
1387 const struct render_page *page = b->page;
1388 enum table_axis axis = b->axis;
1391 /* Width of left header not including its rightmost rule. */
1392 size = axis_width (page, axis, 0, rule_ofs (page->h[axis][0]));
1394 /* If we have a pixel offset and there is no left header, then we omit the
1395 leftmost rule of the body. Otherwise the rendering is deceptive because
1396 it looks like the whole cell is present instead of a partial cell.
1398 Otherwise (if there are headers) we will be merging two rules: the
1399 rightmost rule in the header and the leftmost rule in the body. We assume
1400 that the width of a merged rule is the larger of the widths of either rule
1402 if (b->pixel == 0 || page->h[axis][0])
1403 size += MAX (rule_width (page, axis, page->h[axis][0]),
1404 rule_width (page, axis, b->z));
1406 /* Width of body, minus any pixel offset in the leftmost cell. */
1407 size += joined_width (page, axis, b->z, cell) - b->pixel;
1409 /* Width of rightmost rule in body merged with leftmost rule in headers. */
1410 size += MAX (rule_width_r (page, axis, page->h[axis][1]),
1411 rule_width (page, axis, cell));
1413 /* Width of right header not including its leftmost rule. */
1414 size += axis_width (page, axis, rule_ofs_r (page, axis, page->h[axis][1]),
1415 rule_ofs_r (page, axis, 0));
1417 /* Join crossing. */
1418 if (page->h[axis][0] && page->h[axis][1])
1419 size += page->join_crossing[axis][b->z];
1424 /* Returns true if CELL along B's axis may be broken across a page boundary.
1426 This is just a heuristic. Breaking cells across page boundaries can save
1427 space, but it looks ugly. */
1429 cell_is_breakable (const struct render_break *b, int cell)
1431 const struct render_page *page = b->page;
1432 enum table_axis axis = b->axis;
1434 return cell_width (page, axis, cell) >= page->params->min_break[axis];
1441 const struct render_params *params;
1443 struct render_page **pages;
1444 size_t n_pages, allocated_pages;
1447 struct render_break x_break;
1448 struct render_break y_break;
1451 static const struct render_page *
1452 render_pager_add_table (struct render_pager *p, struct table *table)
1454 struct render_page *page;
1456 if (p->n_pages >= p->allocated_pages)
1457 p->pages = x2nrealloc (p->pages, &p->allocated_pages, sizeof *p->pages);
1458 page = p->pages[p->n_pages++] = render_page_create (p->params, table);
1463 render_pager_start_page (struct render_pager *p)
1465 render_break_init (&p->x_break, p->pages[p->cur_page++], H);
1466 render_break_init_empty (&p->y_break);
1470 add_footnote_page (struct render_pager *p, const struct render_page *body)
1472 const struct table *table = body->table;
1473 int nc = table_nc (table);
1474 int nr = table_nr (table);
1475 int footnote_idx = 0;
1476 struct tab_table *t;
1479 if (!body->n_footnotes)
1482 t = tab_create (2, body->n_footnotes);
1483 for (y = 0; y < nr; y++)
1484 for (x = 0; x < nc; )
1486 struct table_cell cell;
1488 table_get_cell (table, x, y, &cell);
1489 if (y == cell.d[V][0])
1493 for (i = 0; i < cell.n_contents; i++)
1495 const struct cell_contents *cc = &cell.contents[i];
1498 for (j = 0; j < cc->n_footnotes; j++)
1500 const char *f = cc->footnotes[j];
1502 tab_text (t, 0, footnote_idx, TAB_LEFT, "");
1503 tab_footnote (t, 0, footnote_idx, "(none)");
1504 tab_text (t, 1, footnote_idx, TAB_LEFT, f);
1510 table_cell_free (&cell);
1512 render_pager_add_table (p, &t->table);
1515 /* Creates and returns a new render_pager for rendering TABLE_ITEM on the
1516 device with the given PARAMS. */
1517 struct render_pager *
1518 render_pager_create (const struct render_params *params,
1519 const struct table_item *table_item)
1521 const char *caption = table_item_get_caption (table_item);
1522 const char *title = table_item_get_title (table_item);
1523 const struct render_page *body_page;
1524 struct render_pager *p;
1526 p = xzalloc (sizeof *p);
1531 render_pager_add_table (p, table_from_string (TAB_LEFT, title));
1534 body_page = render_pager_add_table (p, table_ref (table_item_get_table (
1539 render_pager_add_table (p, table_from_string (TAB_LEFT, caption));
1542 add_footnote_page (p, body_page);
1544 render_pager_start_page (p);
1551 render_pager_destroy (struct render_pager *p)
1557 render_break_destroy (&p->x_break);
1558 render_break_destroy (&p->y_break);
1559 for (i = 0; i < p->n_pages; i++)
1560 render_page_unref (p->pages[i]);
1566 /* Returns true if P has content remaining to render, false if rendering is
1569 render_pager_has_next (const struct render_pager *p_)
1571 struct render_pager *p = CONST_CAST (struct render_pager *, p_);
1573 while (!render_break_has_next (&p->y_break))
1575 render_break_destroy (&p->y_break);
1576 if (!render_break_has_next (&p->x_break))
1578 render_break_destroy (&p->x_break);
1579 if (p->cur_page >= p->n_pages)
1581 render_break_init_empty (&p->x_break);
1582 render_break_init_empty (&p->y_break);
1585 render_pager_start_page (p);
1588 render_break_init (&p->y_break,
1589 render_break_next (&p->x_break, p->params->size[H]), V);
1594 /* Draws a chunk of content from P to fit in a space that has vertical size
1595 SPACE and the horizontal size specified in the render_params passed to
1596 render_page_create(). Returns the amount of space actually used by the
1597 rendered chunk, which will be 0 if SPACE is too small to render anything or
1598 if no content remains (use render_pager_has_next() to distinguish these
1601 render_pager_draw_next (struct render_pager *p, int space)
1603 int ofs[TABLE_N_AXES] = { 0, 0 };
1604 size_t start_page = SIZE_MAX;
1606 while (render_pager_has_next (p))
1608 struct render_page *page;
1610 if (start_page == p->cur_page)
1612 start_page = p->cur_page;
1614 page = render_break_next (&p->y_break, space - ofs[V]);
1618 render_page_draw (page, ofs);
1619 ofs[V] += render_page_get_size (page, V);
1620 render_page_unref (page);
1625 /* Draws all of P's content. */
1627 render_pager_draw (const struct render_pager *p)
1629 render_pager_draw_region (p, 0, 0, INT_MAX, INT_MAX);
1632 /* Draws the region of P's content that lies in the region (X,Y)-(X+W,Y+H).
1633 Some extra content might be drawn; the device should perform clipping as
1636 render_pager_draw_region (const struct render_pager *p,
1637 int x, int y, int w, int h)
1639 int ofs[TABLE_N_AXES] = { 0, 0 };
1640 int clip[TABLE_N_AXES][2];
1645 for (i = 0; i < p->n_pages; i++)
1647 const struct render_page *page = p->pages[i];
1648 int size = render_page_get_size (page, V);
1650 clip[V][0] = MAX (y, ofs[V]) - ofs[V];
1651 clip[V][1] = MIN (y + h, ofs[V] + size) - ofs[V];
1652 if (clip[V][1] > clip[V][0])
1653 render_page_draw_region (page, ofs, clip);
1659 /* Returns the size of P's content along AXIS; i.e. the content's width if AXIS
1660 is TABLE_HORZ and its length if AXIS is TABLE_VERT. */
1662 render_pager_get_size (const struct render_pager *p, enum table_axis axis)
1667 for (i = 0; i < p->n_pages; i++)
1669 int subsize = render_page_get_size (p->pages[i], axis);
1670 size = axis == H ? MAX (size, subsize) : size + subsize;
1677 render_pager_get_best_breakpoint (const struct render_pager *p, int height)
1682 for (i = 0; i < p->n_pages; i++)
1684 int size = render_page_get_size (p->pages[i], V);
1685 if (y + size >= height)
1686 return render_page_get_best_breakpoint (p->pages[i], height - y) + y;
1693 /* render_page_select() and helpers. */
1695 struct render_page_selection
1697 const struct render_page *page; /* Page whose slice we are selecting. */
1698 struct render_page *subpage; /* New page under construction. */
1699 enum table_axis a; /* Axis of 'page' along which 'subpage' is a slice. */
1700 enum table_axis b; /* The opposite of 'a'. */
1701 int z0; /* First cell along 'a' being selected. */
1702 int z1; /* Last cell being selected, plus 1. */
1703 int p0; /* Number of pixels to trim off left side of z0. */
1704 int p1; /* Number of pixels to trim off right side of z1-1. */
1707 static void cell_to_subpage (struct render_page_selection *,
1708 const struct table_cell *,
1709 int subcell[TABLE_N_AXES]);
1710 static const struct render_overflow *find_overflow_for_cell (
1711 struct render_page_selection *, const struct table_cell *);
1712 static struct render_overflow *insert_overflow (struct render_page_selection *,
1713 const struct table_cell *);
1715 /* Creates and returns a new render_page whose contents are a subregion of
1716 PAGE's contents. The new render_page includes cells Z0 through Z1
1717 (exclusive) along AXIS, plus any headers on AXIS.
1719 If P0 is nonzero, then it is a number of pixels to exclude from the left or
1720 top (according to AXIS) of cell Z0. Similarly, P1 is a number of pixels to
1721 exclude from the right or bottom of cell Z1 - 1. (P0 and P1 are used to
1722 render cells that are too large to fit on a single page.)
1724 The whole of axis !AXIS is included. (The caller may follow up with another
1725 call to render_page_select() to select on !AXIS to select on that axis as
1728 The caller retains ownership of PAGE, which is not modified. */
1729 static struct render_page *
1730 render_page_select (const struct render_page *page, enum table_axis axis,
1731 int z0, int p0, int z1, int p1)
1733 const struct render_footnote *f;
1734 struct render_page_selection s;
1735 enum table_axis a = axis;
1736 enum table_axis b = !a;
1737 struct render_page *subpage;
1738 struct render_overflow *ro;
1744 /* Optimize case where all of PAGE is selected by just incrementing the
1746 if (z0 == page->h[a][0] && p0 == 0
1747 && z1 == page->n[a] - page->h[a][1] && p1 == 0)
1749 struct render_page *page_rw = CONST_CAST (struct render_page *, page);
1754 /* Allocate subpage. */
1755 subpage = render_page_allocate (page->params,
1756 table_select_slice (
1757 table_ref (page->table),
1760 /* An edge is cut off if it was cut off in PAGE or if we're trimming pixels
1761 off that side of the page and there are no headers. */
1762 subpage->is_edge_cutoff[a][0] =
1763 subpage->h[a][0] == 0 && (p0 || (z0 == 0 && page->is_edge_cutoff[a][0]));
1764 subpage->is_edge_cutoff[a][1] =
1765 subpage->h[a][1] == 0 && (p1 || (z1 == page->n[a]
1766 && page->is_edge_cutoff[a][1]));
1767 subpage->is_edge_cutoff[b][0] = page->is_edge_cutoff[b][0];
1768 subpage->is_edge_cutoff[b][1] = page->is_edge_cutoff[b][1];
1770 /* Select join crossings from PAGE into subpage. */
1771 jc = subpage->join_crossing[a];
1772 for (z = 0; z < page->h[a][0]; z++)
1773 *jc++ = page->join_crossing[a][z];
1774 for (z = z0; z <= z1; z++)
1775 *jc++ = page->join_crossing[a][z];
1776 for (z = page->n[a] - page->h[a][1]; z < page->n[a]; z++)
1777 *jc++ = page->join_crossing[a][z];
1778 assert (jc == &subpage->join_crossing[a][subpage->n[a] + 1]);
1780 memcpy (subpage->join_crossing[b], page->join_crossing[b],
1781 (subpage->n[b] + 1) * sizeof **subpage->join_crossing);
1783 /* Select widths from PAGE into subpage. */
1785 dcp = subpage->cp[a];
1787 for (z = 0; z <= rule_ofs (subpage->h[a][0]); z++, dcp++)
1789 if (z == 0 && subpage->is_edge_cutoff[a][0])
1792 dcp[1] = dcp[0] + (scp[z + 1] - scp[z]);
1794 for (z = cell_ofs (z0); z <= cell_ofs (z1 - 1); z++, dcp++)
1796 dcp[1] = dcp[0] + (scp[z + 1] - scp[z]);
1797 if (z == cell_ofs (z0))
1800 if (page->h[a][0] && page->h[a][1])
1801 dcp[1] += page->join_crossing[a][z / 2];
1803 if (z == cell_ofs (z1 - 1))
1806 for (z = rule_ofs_r (page, a, subpage->h[a][1]);
1807 z <= rule_ofs_r (page, a, 0); z++, dcp++)
1809 if (z == rule_ofs_r (page, a, 0) && subpage->is_edge_cutoff[a][1])
1812 dcp[1] = dcp[0] + (scp[z + 1] - scp[z]);
1814 assert (dcp == &subpage->cp[a][2 * subpage->n[a] + 1]);
1816 for (z = 0; z < page->n[b] * 2 + 2; z++)
1817 subpage->cp[b][z] = page->cp[b][z];
1819 /* Add new overflows. */
1827 s.subpage = subpage;
1829 if (!page->h[a][0] || z0 > page->h[a][0] || p0)
1830 for (z = 0; z < page->n[b]; )
1832 struct table_cell cell;
1833 int d[TABLE_N_AXES];
1840 table_get_cell (page->table, d[H], d[V], &cell);
1841 overflow0 = p0 || cell.d[a][0] < z0;
1842 overflow1 = cell.d[a][1] > z1 || (cell.d[a][1] == z1 && p1);
1843 if (overflow0 || overflow1)
1845 ro = insert_overflow (&s, &cell);
1849 ro->overflow[a][0] += p0 + axis_width (
1850 page, a, cell_ofs (cell.d[a][0]), cell_ofs (z0));
1851 if (page->h[a][0] && page->h[a][1])
1852 ro->overflow[a][0] -= page->join_crossing[a][cell.d[a][0]
1858 ro->overflow[a][1] += p1 + axis_width (
1859 page, a, cell_ofs (z1), cell_ofs (cell.d[a][1]));
1860 if (page->h[a][0] && page->h[a][1])
1861 ro->overflow[a][1] -= page->join_crossing[a][cell.d[a][1]];
1865 table_cell_free (&cell);
1868 if (!page->h[a][1] || z1 < page->n[a] - page->h[a][1] || p1)
1869 for (z = 0; z < page->n[b]; )
1871 struct table_cell cell;
1872 int d[TABLE_N_AXES];
1876 table_get_cell (page->table, d[H], d[V], &cell);
1877 if ((cell.d[a][1] > z1 || (cell.d[a][1] == z1 && p1))
1878 && find_overflow_for_cell (&s, &cell) == NULL)
1880 ro = insert_overflow (&s, &cell);
1881 ro->overflow[a][1] += p1 + axis_width (page, a, cell_ofs (z1),
1882 cell_ofs (cell.d[a][1]));
1885 table_cell_free (&cell);
1888 /* Copy overflows from PAGE into subpage. */
1889 HMAP_FOR_EACH (ro, struct render_overflow, node, &page->overflows)
1891 struct table_cell cell;
1893 table_get_cell (page->table, ro->d[H], ro->d[V], &cell);
1894 if (cell.d[a][1] > z0 && cell.d[a][0] < z1
1895 && find_overflow_for_cell (&s, &cell) == NULL)
1896 insert_overflow (&s, &cell);
1897 table_cell_free (&cell);
1900 /* Copy footnotes from PAGE into subpage. */
1901 HMAP_FOR_EACH (f, struct render_footnote, node, &page->footnotes)
1902 if ((f->d[a][0] >= z0 && f->d[a][0] < z1)
1903 || (f->d[a][1] - 1 >= z0 && f->d[a][1] - 1 < z1))
1905 struct render_footnote *nf = xmalloc (sizeof *nf);
1906 nf->d[a][0] = MAX (z0, f->d[a][0]) - z0 + page->h[a][0];
1907 nf->d[a][1] = MIN (z1, f->d[a][1]) - z0 + page->h[a][0];
1908 nf->d[b][0] = f->d[b][0];
1909 nf->d[b][1] = f->d[b][1];
1911 hmap_insert (&subpage->footnotes, &nf->node,
1912 hash_cell (nf->d[H][0], nf->d[V][0]));
1918 /* Given CELL, a table_cell within S->page, stores in SUBCELL the (x,y)
1919 coordinates of the top-left cell as it will appear in S->subpage.
1921 CELL must actually intersect the region of S->page that is being selected
1922 by render_page_select() or the results will not make any sense. */
1924 cell_to_subpage (struct render_page_selection *s,
1925 const struct table_cell *cell, int subcell[TABLE_N_AXES])
1927 enum table_axis a = s->a;
1928 enum table_axis b = s->b;
1929 int ha0 = s->subpage->h[a][0];
1931 subcell[a] = MAX (cell->d[a][0] - s->z0 + ha0, ha0);
1932 subcell[b] = cell->d[b][0];
1935 /* Given CELL, a table_cell within S->page, returns the render_overflow for
1936 that cell in S->subpage, if there is one, and a null pointer otherwise.
1938 CELL must actually intersect the region of S->page that is being selected
1939 by render_page_select() or the results will not make any sense. */
1940 static const struct render_overflow *
1941 find_overflow_for_cell (struct render_page_selection *s,
1942 const struct table_cell *cell)
1946 cell_to_subpage (s, cell, subcell);
1947 return find_overflow (s->subpage, subcell[H], subcell[V]);
1950 /* Given CELL, a table_cell within S->page, inserts a render_overflow for that
1951 cell in S->subpage (which must not already exist). Initializes the new
1952 render_overflow's 'overflow' member from the overflow for CELL in S->page,
1955 CELL must actually intersect the region of S->page that is being selected
1956 by render_page_select() or the results will not make any sense. */
1957 static struct render_overflow *
1958 insert_overflow (struct render_page_selection *s,
1959 const struct table_cell *cell)
1961 const struct render_overflow *old;
1962 struct render_overflow *of;
1964 of = xzalloc (sizeof *of);
1965 cell_to_subpage (s, cell, of->d);
1966 hmap_insert (&s->subpage->overflows, &of->node,
1967 hash_cell (of->d[H], of->d[V]));
1969 old = find_overflow (s->page, cell->d[H][0], cell->d[V][0]);
1971 memcpy (of->overflow, old->overflow, sizeof of->overflow);