xdiffi.c (28743B)
1 /* 2 * LibXDiff by Davide Libenzi ( File Differential Library ) 3 * Copyright (C) 2003 Davide Libenzi 4 * 5 * This library is free software; you can redistribute it and/or 6 * modify it under the terms of the GNU Lesser General Public 7 * License as published by the Free Software Foundation; either 8 * version 2.1 of the License, or (at your option) any later version. 9 * 10 * This library is distributed in the hope that it will be useful, 11 * but WITHOUT ANY WARRANTY; without even the implied warranty of 12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 13 * Lesser General Public License for more details. 14 * 15 * You should have received a copy of the GNU Lesser General Public 16 * License along with this library; if not, see 17 * <http://www.gnu.org/licenses/>. 18 * 19 * Davide Libenzi <davidel@xmailserver.org> 20 * 21 */ 22 23 #include "xinclude.h" 24 25 #define XDL_MAX_COST_MIN 256 26 #define XDL_HEUR_MIN_COST 256 27 #define XDL_LINE_MAX (long)((1UL << (CHAR_BIT * sizeof(long) - 1)) - 1) 28 #define XDL_SNAKE_CNT 20 29 #define XDL_K_HEUR 4 30 31 typedef struct s_xdpsplit { 32 long i1, i2; 33 int min_lo, min_hi; 34 } xdpsplit_t; 35 36 /* 37 * See "An O(ND) Difference Algorithm and its Variations", by Eugene Myers. 38 * Basically considers a "box" (off1, off2, lim1, lim2) and scan from both 39 * the forward diagonal starting from (off1, off2) and the backward diagonal 40 * starting from (lim1, lim2). If the K values on the same diagonal crosses 41 * returns the furthest point of reach. We might encounter expensive edge cases 42 * using this algorithm, so a little bit of heuristic is needed to cut the 43 * search and to return a suboptimal point. 44 */ 45 static long xdl_split(unsigned long const *ha1, long off1, long lim1, 46 unsigned long const *ha2, long off2, long lim2, 47 long *kvdf, long *kvdb, int need_min, xdpsplit_t *spl, 48 xdalgoenv_t *xenv) { 49 long dmin = off1 - lim2, dmax = lim1 - off2; 50 long fmid = off1 - off2, bmid = lim1 - lim2; 51 long odd = (fmid - bmid) & 1; 52 long fmin = fmid, fmax = fmid; 53 long bmin = bmid, bmax = bmid; 54 long ec, d, i1, i2, prev1, best, dd, v, k; 55 56 /* 57 * Set initial diagonal values for both forward and backward path. 58 */ 59 kvdf[fmid] = off1; 60 kvdb[bmid] = lim1; 61 62 for (ec = 1;; ec++) { 63 int got_snake = 0; 64 65 /* 66 * We need to extend the diagonal "domain" by one. If the next 67 * values exits the box boundaries we need to change it in the 68 * opposite direction because (max - min) must be a power of 69 * two. 70 * 71 * Also we initialize the external K value to -1 so that we can 72 * avoid extra conditions in the check inside the core loop. 73 */ 74 if (fmin > dmin) 75 kvdf[--fmin - 1] = -1; 76 else 77 ++fmin; 78 if (fmax < dmax) 79 kvdf[++fmax + 1] = -1; 80 else 81 --fmax; 82 83 for (d = fmax; d >= fmin; d -= 2) { 84 if (kvdf[d - 1] >= kvdf[d + 1]) 85 i1 = kvdf[d - 1] + 1; 86 else 87 i1 = kvdf[d + 1]; 88 prev1 = i1; 89 i2 = i1 - d; 90 for (; i1 < lim1 && i2 < lim2 && ha1[i1] == ha2[i2]; i1++, i2++); 91 if (i1 - prev1 > xenv->snake_cnt) 92 got_snake = 1; 93 kvdf[d] = i1; 94 if (odd && bmin <= d && d <= bmax && kvdb[d] <= i1) { 95 spl->i1 = i1; 96 spl->i2 = i2; 97 spl->min_lo = spl->min_hi = 1; 98 return ec; 99 } 100 } 101 102 /* 103 * We need to extend the diagonal "domain" by one. If the next 104 * values exits the box boundaries we need to change it in the 105 * opposite direction because (max - min) must be a power of 106 * two. 107 * 108 * Also we initialize the external K value to -1 so that we can 109 * avoid extra conditions in the check inside the core loop. 110 */ 111 if (bmin > dmin) 112 kvdb[--bmin - 1] = XDL_LINE_MAX; 113 else 114 ++bmin; 115 if (bmax < dmax) 116 kvdb[++bmax + 1] = XDL_LINE_MAX; 117 else 118 --bmax; 119 120 for (d = bmax; d >= bmin; d -= 2) { 121 if (kvdb[d - 1] < kvdb[d + 1]) 122 i1 = kvdb[d - 1]; 123 else 124 i1 = kvdb[d + 1] - 1; 125 prev1 = i1; 126 i2 = i1 - d; 127 for (; i1 > off1 && i2 > off2 && ha1[i1 - 1] == ha2[i2 - 1]; i1--, i2--); 128 if (prev1 - i1 > xenv->snake_cnt) 129 got_snake = 1; 130 kvdb[d] = i1; 131 if (!odd && fmin <= d && d <= fmax && i1 <= kvdf[d]) { 132 spl->i1 = i1; 133 spl->i2 = i2; 134 spl->min_lo = spl->min_hi = 1; 135 return ec; 136 } 137 } 138 139 if (need_min) 140 continue; 141 142 /* 143 * If the edit cost is above the heuristic trigger and if 144 * we got a good snake, we sample current diagonals to see 145 * if some of them have reached an "interesting" path. Our 146 * measure is a function of the distance from the diagonal 147 * corner (i1 + i2) penalized with the distance from the 148 * mid diagonal itself. If this value is above the current 149 * edit cost times a magic factor (XDL_K_HEUR) we consider 150 * it interesting. 151 */ 152 if (got_snake && ec > xenv->heur_min) { 153 for (best = 0, d = fmax; d >= fmin; d -= 2) { 154 dd = d > fmid ? d - fmid: fmid - d; 155 i1 = kvdf[d]; 156 i2 = i1 - d; 157 v = (i1 - off1) + (i2 - off2) - dd; 158 159 if (v > XDL_K_HEUR * ec && v > best && 160 off1 + xenv->snake_cnt <= i1 && i1 < lim1 && 161 off2 + xenv->snake_cnt <= i2 && i2 < lim2) { 162 for (k = 1; ha1[i1 - k] == ha2[i2 - k]; k++) 163 if (k == xenv->snake_cnt) { 164 best = v; 165 spl->i1 = i1; 166 spl->i2 = i2; 167 break; 168 } 169 } 170 } 171 if (best > 0) { 172 spl->min_lo = 1; 173 spl->min_hi = 0; 174 return ec; 175 } 176 177 for (best = 0, d = bmax; d >= bmin; d -= 2) { 178 dd = d > bmid ? d - bmid: bmid - d; 179 i1 = kvdb[d]; 180 i2 = i1 - d; 181 v = (lim1 - i1) + (lim2 - i2) - dd; 182 183 if (v > XDL_K_HEUR * ec && v > best && 184 off1 < i1 && i1 <= lim1 - xenv->snake_cnt && 185 off2 < i2 && i2 <= lim2 - xenv->snake_cnt) { 186 for (k = 0; ha1[i1 + k] == ha2[i2 + k]; k++) 187 if (k == xenv->snake_cnt - 1) { 188 best = v; 189 spl->i1 = i1; 190 spl->i2 = i2; 191 break; 192 } 193 } 194 } 195 if (best > 0) { 196 spl->min_lo = 0; 197 spl->min_hi = 1; 198 return ec; 199 } 200 } 201 202 /* 203 * Enough is enough. We spent too much time here and now we 204 * collect the furthest reaching path using the (i1 + i2) 205 * measure. 206 */ 207 if (ec >= xenv->mxcost) { 208 long fbest, fbest1, bbest, bbest1; 209 210 fbest = fbest1 = -1; 211 for (d = fmax; d >= fmin; d -= 2) { 212 i1 = XDL_MIN(kvdf[d], lim1); 213 i2 = i1 - d; 214 if (lim2 < i2) 215 i1 = lim2 + d, i2 = lim2; 216 if (fbest < i1 + i2) { 217 fbest = i1 + i2; 218 fbest1 = i1; 219 } 220 } 221 222 bbest = bbest1 = XDL_LINE_MAX; 223 for (d = bmax; d >= bmin; d -= 2) { 224 i1 = XDL_MAX(off1, kvdb[d]); 225 i2 = i1 - d; 226 if (i2 < off2) 227 i1 = off2 + d, i2 = off2; 228 if (i1 + i2 < bbest) { 229 bbest = i1 + i2; 230 bbest1 = i1; 231 } 232 } 233 234 if ((lim1 + lim2) - bbest < fbest - (off1 + off2)) { 235 spl->i1 = fbest1; 236 spl->i2 = fbest - fbest1; 237 spl->min_lo = 1; 238 spl->min_hi = 0; 239 } else { 240 spl->i1 = bbest1; 241 spl->i2 = bbest - bbest1; 242 spl->min_lo = 0; 243 spl->min_hi = 1; 244 } 245 return ec; 246 } 247 } 248 } 249 250 251 /* 252 * Rule: "Divide et Impera" (divide & conquer). Recursively split the box in 253 * sub-boxes by calling the box splitting function. Note that the real job 254 * (marking changed lines) is done in the two boundary reaching checks. 255 */ 256 int xdl_recs_cmp(diffdata_t *dd1, long off1, long lim1, 257 diffdata_t *dd2, long off2, long lim2, 258 long *kvdf, long *kvdb, int need_min, xdalgoenv_t *xenv) { 259 unsigned long const *ha1 = dd1->ha, *ha2 = dd2->ha; 260 261 /* 262 * Shrink the box by walking through each diagonal snake (SW and NE). 263 */ 264 for (; off1 < lim1 && off2 < lim2 && ha1[off1] == ha2[off2]; off1++, off2++); 265 for (; off1 < lim1 && off2 < lim2 && ha1[lim1 - 1] == ha2[lim2 - 1]; lim1--, lim2--); 266 267 /* 268 * If one dimension is empty, then all records on the other one must 269 * be obviously changed. 270 */ 271 if (off1 == lim1) { 272 char *rchg2 = dd2->rchg; 273 long *rindex2 = dd2->rindex; 274 275 for (; off2 < lim2; off2++) 276 rchg2[rindex2[off2]] = 1; 277 } else if (off2 == lim2) { 278 char *rchg1 = dd1->rchg; 279 long *rindex1 = dd1->rindex; 280 281 for (; off1 < lim1; off1++) 282 rchg1[rindex1[off1]] = 1; 283 } else { 284 xdpsplit_t spl; 285 spl.i1 = spl.i2 = 0; 286 287 /* 288 * Divide ... 289 */ 290 if (xdl_split(ha1, off1, lim1, ha2, off2, lim2, kvdf, kvdb, 291 need_min, &spl, xenv) < 0) { 292 293 return -1; 294 } 295 296 /* 297 * ... et Impera. 298 */ 299 if (xdl_recs_cmp(dd1, off1, spl.i1, dd2, off2, spl.i2, 300 kvdf, kvdb, spl.min_lo, xenv) < 0 || 301 xdl_recs_cmp(dd1, spl.i1, lim1, dd2, spl.i2, lim2, 302 kvdf, kvdb, spl.min_hi, xenv) < 0) { 303 304 return -1; 305 } 306 } 307 308 return 0; 309 } 310 311 312 int xdl_do_diff(mmfile_t *mf1, mmfile_t *mf2, xpparam_t const *xpp, 313 xdfenv_t *xe) { 314 long ndiags; 315 long *kvd, *kvdf, *kvdb; 316 xdalgoenv_t xenv; 317 diffdata_t dd1, dd2; 318 319 if (XDF_DIFF_ALG(xpp->flags) == XDF_PATIENCE_DIFF) 320 return xdl_do_patience_diff(mf1, mf2, xpp, xe); 321 322 if (XDF_DIFF_ALG(xpp->flags) == XDF_HISTOGRAM_DIFF) 323 return xdl_do_histogram_diff(mf1, mf2, xpp, xe); 324 325 if (xdl_prepare_env(mf1, mf2, xpp, xe) < 0) { 326 327 return -1; 328 } 329 330 /* 331 * Allocate and setup K vectors to be used by the differential 332 * algorithm. 333 * 334 * One is to store the forward path and one to store the backward path. 335 */ 336 ndiags = xe->xdf1.nreff + xe->xdf2.nreff + 3; 337 if (!(kvd = (long *) xdl_malloc((2 * ndiags + 2) * sizeof(long)))) { 338 339 xdl_free_env(xe); 340 return -1; 341 } 342 kvdf = kvd; 343 kvdb = kvdf + ndiags; 344 kvdf += xe->xdf2.nreff + 1; 345 kvdb += xe->xdf2.nreff + 1; 346 347 xenv.mxcost = xdl_bogosqrt(ndiags); 348 if (xenv.mxcost < XDL_MAX_COST_MIN) 349 xenv.mxcost = XDL_MAX_COST_MIN; 350 xenv.snake_cnt = XDL_SNAKE_CNT; 351 xenv.heur_min = XDL_HEUR_MIN_COST; 352 353 dd1.nrec = xe->xdf1.nreff; 354 dd1.ha = xe->xdf1.ha; 355 dd1.rchg = xe->xdf1.rchg; 356 dd1.rindex = xe->xdf1.rindex; 357 dd2.nrec = xe->xdf2.nreff; 358 dd2.ha = xe->xdf2.ha; 359 dd2.rchg = xe->xdf2.rchg; 360 dd2.rindex = xe->xdf2.rindex; 361 362 if (xdl_recs_cmp(&dd1, 0, dd1.nrec, &dd2, 0, dd2.nrec, 363 kvdf, kvdb, (xpp->flags & XDF_NEED_MINIMAL) != 0, &xenv) < 0) { 364 365 xdl_free(kvd); 366 xdl_free_env(xe); 367 return -1; 368 } 369 370 xdl_free(kvd); 371 372 return 0; 373 } 374 375 376 static xdchange_t *xdl_add_change(xdchange_t *xscr, long i1, long i2, long chg1, long chg2) { 377 xdchange_t *xch; 378 379 if (!(xch = (xdchange_t *) xdl_malloc(sizeof(xdchange_t)))) 380 return NULL; 381 382 xch->next = xscr; 383 xch->i1 = i1; 384 xch->i2 = i2; 385 xch->chg1 = chg1; 386 xch->chg2 = chg2; 387 xch->ignore = 0; 388 389 return xch; 390 } 391 392 393 static int recs_match(xrecord_t *rec1, xrecord_t *rec2, long flags) 394 { 395 return (rec1->ha == rec2->ha && 396 xdl_recmatch(rec1->ptr, rec1->size, 397 rec2->ptr, rec2->size, 398 flags)); 399 } 400 401 /* 402 * If a line is indented more than this, xget_indent() just returns this value. 403 * This avoids having to do absurd amounts of work for data that are not 404 * human-readable text, and also ensures that the output of xget_indent fits within 405 * an int. 406 */ 407 #define MAX_INDENT 200 408 409 /* 410 * Return the amount of indentation of the specified line, treating TAB as 8 411 * columns. Return -1 if line is empty or contains only whitespace. Clamp the 412 * output value at MAX_INDENT. 413 */ 414 static int xget_indent(xrecord_t *rec) 415 { 416 long i; 417 int ret = 0; 418 419 for (i = 0; i < rec->size; i++) { 420 char c = rec->ptr[i]; 421 422 if (!XDL_ISSPACE(c)) 423 return ret; 424 else if (c == ' ') 425 ret += 1; 426 else if (c == '\t') 427 ret += 8 - ret % 8; 428 /* ignore other whitespace characters */ 429 430 if (ret >= MAX_INDENT) 431 return MAX_INDENT; 432 } 433 434 /* The line contains only whitespace. */ 435 return -1; 436 } 437 438 /* 439 * If more than this number of consecutive blank rows are found, just return 440 * this value. This avoids requiring O(N^2) work for pathological cases, and 441 * also ensures that the output of score_split fits in an int. 442 */ 443 #define MAX_BLANKS 20 444 445 /* Characteristics measured about a hypothetical split position. */ 446 struct split_measurement { 447 /* 448 * Is the split at the end of the file (aside from any blank lines)? 449 */ 450 int end_of_file; 451 452 /* 453 * How much is the line immediately following the split indented (or -1 454 * if the line is blank): 455 */ 456 int indent; 457 458 /* 459 * How many consecutive lines above the split are blank? 460 */ 461 int pre_blank; 462 463 /* 464 * How much is the nearest non-blank line above the split indented (or 465 * -1 if there is no such line)? 466 */ 467 int pre_indent; 468 469 /* 470 * How many lines after the line following the split are blank? 471 */ 472 int post_blank; 473 474 /* 475 * How much is the nearest non-blank line after the line following the 476 * split indented (or -1 if there is no such line)? 477 */ 478 int post_indent; 479 }; 480 481 struct split_score { 482 /* The effective indent of this split (smaller is preferred). */ 483 int effective_indent; 484 485 /* Penalty for this split (smaller is preferred). */ 486 int penalty; 487 }; 488 489 /* 490 * Fill m with information about a hypothetical split of xdf above line split. 491 */ 492 static void measure_split(const xdfile_t *xdf, long split, 493 struct split_measurement *m) 494 { 495 long i; 496 497 if (split >= xdf->nrec) { 498 m->end_of_file = 1; 499 m->indent = -1; 500 } else { 501 m->end_of_file = 0; 502 m->indent = xget_indent(xdf->recs[split]); 503 } 504 505 m->pre_blank = 0; 506 m->pre_indent = -1; 507 for (i = split - 1; i >= 0; i--) { 508 m->pre_indent = xget_indent(xdf->recs[i]); 509 if (m->pre_indent != -1) 510 break; 511 m->pre_blank += 1; 512 if (m->pre_blank == MAX_BLANKS) { 513 m->pre_indent = 0; 514 break; 515 } 516 } 517 518 m->post_blank = 0; 519 m->post_indent = -1; 520 for (i = split + 1; i < xdf->nrec; i++) { 521 m->post_indent = xget_indent(xdf->recs[i]); 522 if (m->post_indent != -1) 523 break; 524 m->post_blank += 1; 525 if (m->post_blank == MAX_BLANKS) { 526 m->post_indent = 0; 527 break; 528 } 529 } 530 } 531 532 /* 533 * The empirically-determined weight factors used by score_split() below. 534 * Larger values means that the position is a less favorable place to split. 535 * 536 * Note that scores are only ever compared against each other, so multiplying 537 * all of these weight/penalty values by the same factor wouldn't change the 538 * heuristic's behavior. Still, we need to set that arbitrary scale *somehow*. 539 * In practice, these numbers are chosen to be large enough that they can be 540 * adjusted relative to each other with sufficient precision despite using 541 * integer math. 542 */ 543 544 /* Penalty if there are no non-blank lines before the split */ 545 #define START_OF_FILE_PENALTY 1 546 547 /* Penalty if there are no non-blank lines after the split */ 548 #define END_OF_FILE_PENALTY 21 549 550 /* Multiplier for the number of blank lines around the split */ 551 #define TOTAL_BLANK_WEIGHT (-30) 552 553 /* Multiplier for the number of blank lines after the split */ 554 #define POST_BLANK_WEIGHT 6 555 556 /* 557 * Penalties applied if the line is indented more than its predecessor 558 */ 559 #define RELATIVE_INDENT_PENALTY (-4) 560 #define RELATIVE_INDENT_WITH_BLANK_PENALTY 10 561 562 /* 563 * Penalties applied if the line is indented less than both its predecessor and 564 * its successor 565 */ 566 #define RELATIVE_OUTDENT_PENALTY 24 567 #define RELATIVE_OUTDENT_WITH_BLANK_PENALTY 17 568 569 /* 570 * Penalties applied if the line is indented less than its predecessor but not 571 * less than its successor 572 */ 573 #define RELATIVE_DEDENT_PENALTY 23 574 #define RELATIVE_DEDENT_WITH_BLANK_PENALTY 17 575 576 /* 577 * We only consider whether the sum of the effective indents for splits are 578 * less than (-1), equal to (0), or greater than (+1) each other. The resulting 579 * value is multiplied by the following weight and combined with the penalty to 580 * determine the better of two scores. 581 */ 582 #define INDENT_WEIGHT 60 583 584 /* 585 * How far do we slide a hunk at most? 586 */ 587 #define INDENT_HEURISTIC_MAX_SLIDING 100 588 589 /* 590 * Compute a badness score for the hypothetical split whose measurements are 591 * stored in m. The weight factors were determined empirically using the tools 592 * and corpus described in 593 * 594 * https://github.com/mhagger/diff-slider-tools 595 * 596 * Also see that project if you want to improve the weights based on, for 597 * example, a larger or more diverse corpus. 598 */ 599 static void score_add_split(const struct split_measurement *m, struct split_score *s) 600 { 601 /* 602 * A place to accumulate penalty factors (positive makes this index more 603 * favored): 604 */ 605 int post_blank, total_blank, indent, any_blanks; 606 607 if (m->pre_indent == -1 && m->pre_blank == 0) 608 s->penalty += START_OF_FILE_PENALTY; 609 610 if (m->end_of_file) 611 s->penalty += END_OF_FILE_PENALTY; 612 613 /* 614 * Set post_blank to the number of blank lines following the split, 615 * including the line immediately after the split: 616 */ 617 post_blank = (m->indent == -1) ? 1 + m->post_blank : 0; 618 total_blank = m->pre_blank + post_blank; 619 620 /* Penalties based on nearby blank lines: */ 621 s->penalty += TOTAL_BLANK_WEIGHT * total_blank; 622 s->penalty += POST_BLANK_WEIGHT * post_blank; 623 624 if (m->indent != -1) 625 indent = m->indent; 626 else 627 indent = m->post_indent; 628 629 any_blanks = (total_blank != 0); 630 631 /* Note that the effective indent is -1 at the end of the file: */ 632 s->effective_indent += indent; 633 634 if (indent == -1) { 635 /* No additional adjustments needed. */ 636 } else if (m->pre_indent == -1) { 637 /* No additional adjustments needed. */ 638 } else if (indent > m->pre_indent) { 639 /* 640 * The line is indented more than its predecessor. 641 */ 642 s->penalty += any_blanks ? 643 RELATIVE_INDENT_WITH_BLANK_PENALTY : 644 RELATIVE_INDENT_PENALTY; 645 } else if (indent == m->pre_indent) { 646 /* 647 * The line has the same indentation level as its predecessor. 648 * No additional adjustments needed. 649 */ 650 } else { 651 /* 652 * The line is indented less than its predecessor. It could be 653 * the block terminator of the previous block, but it could 654 * also be the start of a new block (e.g., an "else" block, or 655 * maybe the previous block didn't have a block terminator). 656 * Try to distinguish those cases based on what comes next: 657 */ 658 if (m->post_indent != -1 && m->post_indent > indent) { 659 /* 660 * The following line is indented more. So it is likely 661 * that this line is the start of a block. 662 */ 663 s->penalty += any_blanks ? 664 RELATIVE_OUTDENT_WITH_BLANK_PENALTY : 665 RELATIVE_OUTDENT_PENALTY; 666 } else { 667 /* 668 * That was probably the end of a block. 669 */ 670 s->penalty += any_blanks ? 671 RELATIVE_DEDENT_WITH_BLANK_PENALTY : 672 RELATIVE_DEDENT_PENALTY; 673 } 674 } 675 } 676 677 static int score_cmp(struct split_score *s1, struct split_score *s2) 678 { 679 /* -1 if s1.effective_indent < s2->effective_indent, etc. */ 680 int cmp_indents = ((s1->effective_indent > s2->effective_indent) - 681 (s1->effective_indent < s2->effective_indent)); 682 683 return INDENT_WEIGHT * cmp_indents + (s1->penalty - s2->penalty); 684 } 685 686 /* 687 * Represent a group of changed lines in an xdfile_t (i.e., a contiguous group 688 * of lines that was inserted or deleted from the corresponding version of the 689 * file). We consider there to be such a group at the beginning of the file, at 690 * the end of the file, and between any two unchanged lines, though most such 691 * groups will usually be empty. 692 * 693 * If the first line in a group is equal to the line following the group, then 694 * the group can be slid down. Similarly, if the last line in a group is equal 695 * to the line preceding the group, then the group can be slid up. See 696 * group_slide_down() and group_slide_up(). 697 * 698 * Note that loops that are testing for changed lines in xdf->rchg do not need 699 * index bounding since the array is prepared with a zero at position -1 and N. 700 */ 701 struct xdlgroup { 702 /* 703 * The index of the first changed line in the group, or the index of 704 * the unchanged line above which the (empty) group is located. 705 */ 706 long start; 707 708 /* 709 * The index of the first unchanged line after the group. For an empty 710 * group, end is equal to start. 711 */ 712 long end; 713 }; 714 715 /* 716 * Initialize g to point at the first group in xdf. 717 */ 718 static void group_init(xdfile_t *xdf, struct xdlgroup *g) 719 { 720 g->start = g->end = 0; 721 while (xdf->rchg[g->end]) 722 g->end++; 723 } 724 725 /* 726 * Move g to describe the next (possibly empty) group in xdf and return 0. If g 727 * is already at the end of the file, do nothing and return -1. 728 */ 729 static inline int group_next(xdfile_t *xdf, struct xdlgroup *g) 730 { 731 if (g->end == xdf->nrec) 732 return -1; 733 734 g->start = g->end + 1; 735 for (g->end = g->start; xdf->rchg[g->end]; g->end++) 736 ; 737 738 return 0; 739 } 740 741 /* 742 * Move g to describe the previous (possibly empty) group in xdf and return 0. 743 * If g is already at the beginning of the file, do nothing and return -1. 744 */ 745 static inline int group_previous(xdfile_t *xdf, struct xdlgroup *g) 746 { 747 if (g->start == 0) 748 return -1; 749 750 g->end = g->start - 1; 751 for (g->start = g->end; xdf->rchg[g->start - 1]; g->start--) 752 ; 753 754 return 0; 755 } 756 757 /* 758 * If g can be slid toward the end of the file, do so, and if it bumps into a 759 * following group, expand this group to include it. Return 0 on success or -1 760 * if g cannot be slid down. 761 */ 762 static int group_slide_down(xdfile_t *xdf, struct xdlgroup *g, long flags) 763 { 764 if (g->end < xdf->nrec && 765 recs_match(xdf->recs[g->start], xdf->recs[g->end], flags)) { 766 xdf->rchg[g->start++] = 0; 767 xdf->rchg[g->end++] = 1; 768 769 while (xdf->rchg[g->end]) 770 g->end++; 771 772 return 0; 773 } else { 774 return -1; 775 } 776 } 777 778 /* 779 * If g can be slid toward the beginning of the file, do so, and if it bumps 780 * into a previous group, expand this group to include it. Return 0 on success 781 * or -1 if g cannot be slid up. 782 */ 783 static int group_slide_up(xdfile_t *xdf, struct xdlgroup *g, long flags) 784 { 785 if (g->start > 0 && 786 recs_match(xdf->recs[g->start - 1], xdf->recs[g->end - 1], flags)) { 787 xdf->rchg[--g->start] = 1; 788 xdf->rchg[--g->end] = 0; 789 790 while (xdf->rchg[g->start - 1]) 791 g->start--; 792 793 return 0; 794 } else { 795 return -1; 796 } 797 } 798 799 static void xdl_bug(const char *msg) 800 { 801 fprintf(stderr, "BUG: %s\n", msg); 802 exit(1); 803 } 804 805 /* 806 * Move back and forward change groups for a consistent and pretty diff output. 807 * This also helps in finding joinable change groups and reducing the diff 808 * size. 809 */ 810 int xdl_change_compact(xdfile_t *xdf, xdfile_t *xdfo, long flags) { 811 struct xdlgroup g, go; 812 long earliest_end, end_matching_other; 813 long groupsize; 814 815 group_init(xdf, &g); 816 group_init(xdfo, &go); 817 818 while (1) { 819 /* 820 * If the group is empty in the to-be-compacted file, skip it: 821 */ 822 if (g.end == g.start) 823 goto next; 824 825 /* 826 * Now shift the change up and then down as far as possible in 827 * each direction. If it bumps into any other changes, merge 828 * them. 829 */ 830 do { 831 groupsize = g.end - g.start; 832 833 /* 834 * Keep track of the last "end" index that causes this 835 * group to align with a group of changed lines in the 836 * other file. -1 indicates that we haven't found such 837 * a match yet: 838 */ 839 end_matching_other = -1; 840 841 /* Shift the group backward as much as possible: */ 842 while (!group_slide_up(xdf, &g, flags)) 843 if (group_previous(xdfo, &go)) 844 xdl_bug("group sync broken sliding up"); 845 846 /* 847 * This is this highest that this group can be shifted. 848 * Record its end index: 849 */ 850 earliest_end = g.end; 851 852 if (go.end > go.start) 853 end_matching_other = g.end; 854 855 /* Now shift the group forward as far as possible: */ 856 while (1) { 857 if (group_slide_down(xdf, &g, flags)) 858 break; 859 if (group_next(xdfo, &go)) 860 xdl_bug("group sync broken sliding down"); 861 862 if (go.end > go.start) 863 end_matching_other = g.end; 864 } 865 } while (groupsize != g.end - g.start); 866 867 /* 868 * If the group can be shifted, then we can possibly use this 869 * freedom to produce a more intuitive diff. 870 * 871 * The group is currently shifted as far down as possible, so 872 * the heuristics below only have to handle upwards shifts. 873 */ 874 875 if (g.end == earliest_end) { 876 /* no shifting was possible */ 877 } else if (end_matching_other != -1) { 878 /* 879 * Move the possibly merged group of changes back to 880 * line up with the last group of changes from the 881 * other file that it can align with. 882 */ 883 while (go.end == go.start) { 884 if (group_slide_up(xdf, &g, flags)) 885 xdl_bug("match disappeared"); 886 if (group_previous(xdfo, &go)) 887 xdl_bug("group sync broken sliding to match"); 888 } 889 } else if (flags & XDF_INDENT_HEURISTIC) { 890 /* 891 * Indent heuristic: a group of pure add/delete lines 892 * implies two splits, one between the end of the 893 * "before" context and the start of the group, and 894 * another between the end of the group and the 895 * beginning of the "after" context. Some splits are 896 * aesthetically better and some are worse. We compute 897 * a badness "score" for each split, and add the scores 898 * for the two splits to define a "score" for each 899 * position that the group can be shifted to. Then we 900 * pick the shift with the lowest score. 901 */ 902 long shift, best_shift = -1; 903 struct split_score best_score; 904 905 shift = earliest_end; 906 if (g.end - groupsize - 1 > shift) 907 shift = g.end - groupsize - 1; 908 if (g.end - INDENT_HEURISTIC_MAX_SLIDING > shift) 909 shift = g.end - INDENT_HEURISTIC_MAX_SLIDING; 910 for (; shift <= g.end; shift++) { 911 struct split_measurement m; 912 struct split_score score = {0, 0}; 913 914 measure_split(xdf, shift, &m); 915 score_add_split(&m, &score); 916 measure_split(xdf, shift - groupsize, &m); 917 score_add_split(&m, &score); 918 if (best_shift == -1 || 919 score_cmp(&score, &best_score) <= 0) { 920 best_score.effective_indent = score.effective_indent; 921 best_score.penalty = score.penalty; 922 best_shift = shift; 923 } 924 } 925 926 while (g.end > best_shift) { 927 if (group_slide_up(xdf, &g, flags)) 928 xdl_bug("best shift unreached"); 929 if (group_previous(xdfo, &go)) 930 xdl_bug("group sync broken sliding to blank line"); 931 } 932 } 933 934 next: 935 /* Move past the just-processed group: */ 936 if (group_next(xdf, &g)) 937 break; 938 if (group_next(xdfo, &go)) 939 xdl_bug("group sync broken moving to next group"); 940 } 941 942 if (!group_next(xdfo, &go)) 943 xdl_bug("group sync broken at end of file"); 944 945 return 0; 946 } 947 948 949 int xdl_build_script(xdfenv_t *xe, xdchange_t **xscr) { 950 xdchange_t *cscr = NULL, *xch; 951 char *rchg1 = xe->xdf1.rchg, *rchg2 = xe->xdf2.rchg; 952 long i1, i2, l1, l2; 953 954 /* 955 * Trivial. Collects "groups" of changes and creates an edit script. 956 */ 957 for (i1 = xe->xdf1.nrec, i2 = xe->xdf2.nrec; i1 >= 0 || i2 >= 0; i1--, i2--) 958 if (rchg1[i1 - 1] || rchg2[i2 - 1]) { 959 for (l1 = i1; rchg1[i1 - 1]; i1--); 960 for (l2 = i2; rchg2[i2 - 1]; i2--); 961 962 if (!(xch = xdl_add_change(cscr, i1, i2, l1 - i1, l2 - i2))) { 963 xdl_free_script(cscr); 964 return -1; 965 } 966 cscr = xch; 967 } 968 969 *xscr = cscr; 970 971 return 0; 972 } 973 974 975 void xdl_free_script(xdchange_t *xscr) { 976 xdchange_t *xch; 977 978 while ((xch = xscr) != NULL) { 979 xscr = xscr->next; 980 xdl_free(xch); 981 } 982 } 983 984 static int xdl_call_hunk_func(xdfenv_t *xe UNUSED, xdchange_t *xscr, xdemitcb_t *ecb, 985 xdemitconf_t const *xecfg) 986 { 987 xdchange_t *xch, *xche; 988 989 for (xch = xscr; xch; xch = xche->next) { 990 xche = xdl_get_hunk(&xch, xecfg); 991 if (!xch) 992 break; 993 if (xecfg->hunk_func(xch->i1, xche->i1 + xche->chg1 - xch->i1, 994 xch->i2, xche->i2 + xche->chg2 - xch->i2, 995 ecb->priv) < 0) 996 return -1; 997 } 998 return 0; 999 } 1000 1001 static void xdl_mark_ignorable_lines(xdchange_t *xscr, xdfenv_t *xe, long flags) 1002 { 1003 xdchange_t *xch; 1004 1005 for (xch = xscr; xch; xch = xch->next) { 1006 int ignore = 1; 1007 xrecord_t **rec; 1008 long i; 1009 1010 rec = &xe->xdf1.recs[xch->i1]; 1011 for (i = 0; i < xch->chg1 && ignore; i++) 1012 ignore = xdl_blankline(rec[i]->ptr, rec[i]->size, flags); 1013 1014 rec = &xe->xdf2.recs[xch->i2]; 1015 for (i = 0; i < xch->chg2 && ignore; i++) 1016 ignore = xdl_blankline(rec[i]->ptr, rec[i]->size, flags); 1017 1018 xch->ignore = ignore; 1019 } 1020 } 1021 1022 #if 0 // unused by Vim 1023 static int record_matches_regex(xrecord_t *rec, xpparam_t const *xpp) { 1024 regmatch_t regmatch; 1025 int i; 1026 1027 for (i = 0; i < xpp->ignore_regex_nr; i++) 1028 if (!regexec_buf(xpp->ignore_regex[i], rec->ptr, rec->size, 1, 1029 ®match, 0)) 1030 return 1; 1031 1032 return 0; 1033 } 1034 1035 static void xdl_mark_ignorable_regex(xdchange_t *xscr, const xdfenv_t *xe, 1036 xpparam_t const *xpp) 1037 { 1038 xdchange_t *xch; 1039 1040 for (xch = xscr; xch; xch = xch->next) { 1041 xrecord_t **rec; 1042 int ignore = 1; 1043 long i; 1044 1045 /* 1046 * Do not override --ignore-blank-lines. 1047 */ 1048 if (xch->ignore) 1049 continue; 1050 1051 rec = &xe->xdf1.recs[xch->i1]; 1052 for (i = 0; i < xch->chg1 && ignore; i++) 1053 ignore = record_matches_regex(rec[i], xpp); 1054 1055 rec = &xe->xdf2.recs[xch->i2]; 1056 for (i = 0; i < xch->chg2 && ignore; i++) 1057 ignore = record_matches_regex(rec[i], xpp); 1058 1059 xch->ignore = ignore; 1060 } 1061 } 1062 #endif 1063 1064 int xdl_diff(mmfile_t *mf1, mmfile_t *mf2, xpparam_t const *xpp, 1065 xdemitconf_t const *xecfg, xdemitcb_t *ecb) { 1066 xdchange_t *xscr; 1067 xdfenv_t xe; 1068 emit_func_t ef = xecfg->hunk_func ? xdl_call_hunk_func : xdl_emit_diff; 1069 1070 if (xdl_do_diff(mf1, mf2, xpp, &xe) < 0) { 1071 1072 return -1; 1073 } 1074 if (xdl_change_compact(&xe.xdf1, &xe.xdf2, xpp->flags) < 0 || 1075 xdl_change_compact(&xe.xdf2, &xe.xdf1, xpp->flags) < 0 || 1076 xdl_build_script(&xe, &xscr) < 0) { 1077 1078 xdl_free_env(&xe); 1079 return -1; 1080 } 1081 if (xscr) { 1082 if (xpp->flags & XDF_IGNORE_BLANK_LINES) 1083 xdl_mark_ignorable_lines(xscr, &xe, xpp->flags); 1084 1085 #if 0 1086 if (xpp->ignore_regex) 1087 xdl_mark_ignorable_regex(xscr, &xe, xpp); 1088 #endif 1089 1090 if (ef(&xe, xscr, ecb, xecfg) < 0) { 1091 1092 xdl_free_script(xscr); 1093 xdl_free_env(&xe); 1094 return -1; 1095 } 1096 xdl_free_script(xscr); 1097 } 1098 xdl_free_env(&xe); 1099 1100 return 0; 1101 }