1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821 1822 1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 1893 1894 1895 1896 1897 1898 1899 1900 1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912 1913 1914 1915 1916 1917 1918 1919 1920 1921 1922 1923 1924 1925 1926 1927 1928 1929 1930 1931 1932 1933 1934 1935 1936 1937 1938 1939 1940 1941 1942 1943 1944 1945 1946 1947 1948 1949 1950 1951 1952 1953 1954 1955 1956 1957 1958 1959 1960 1961 1962 1963 1964 1965 1966 1967 1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 2018 2019 2020 2021 2022 2023 2024 2025 2026 2027 2028 2029 2030 2031 2032 2033 2034 2035 2036 2037 2038 2039 2040 2041 2042 2043 2044 2045 2046 2047 2048 2049 2050 2051 2052 2053 2054 2055 2056 2057 2058 2059 2060 2061 2062 2063 2064 2065 2066 2067 2068 2069 2070 2071 2072 2073 2074 2075 2076 2077 2078 2079 2080 2081 2082 2083 2084 2085 2086 2087 2088 2089 2090 2091 2092 2093 2094 2095 2096 2097 2098 2099 2100 2101 2102 2103 2104 2105 2106 2107 2108 2109 2110 2111 2112 2113 2114 2115 2116 2117 2118 2119 2120 2121 2122 2123 2124 2125 2126 2127 2128 2129 2130 2131 2132 2133 2134 2135 2136 2137 2138 2139 2140 2141 2142 2143 2144 2145 2146 2147 2148 2149 2150 2151 2152 2153 2154 2155 2156 2157 2158 2159 2160 2161 2162 2163 2164 2165 2166 2167 2168 2169 2170 2171 2172 2173 2174 2175 2176 2177 2178 2179 2180 2181 2182 2183 2184 2185 2186 2187 2188 2189 2190 2191 2192 2193 2194 2195 2196 2197 2198 2199 2200 2201 2202 2203 2204 2205 2206 2207 2208 2209 2210 2211 2212 2213 2214 2215 2216 2217 2218 2219 2220 2221 2222 2223 2224 2225 2226 2227 2228 2229 2230 2231 2232 2233 2234 2235 2236 2237 2238 2239 2240 2241 2242 2243 2244 2245 2246 2247 2248 2249 2250 2251 2252 2253 2254 2255 2256 2257 2258 2259 2260 2261 2262 2263 2264 2265 2266 2267 2268 2269 2270 2271 2272 2273 2274 2275 2276 2277 2278 2279 2280 2281 2282 2283 2284 2285 2286 2287 2288 2289 2290 2291 2292 2293 2294 2295 2296 2297 2298 2299 2300 2301 2302 2303 2304 2305 2306 2307 2308 2309 2310 2311 2312 2313 2314 2315 2316 2317 2318 2319 2320 2321 2322 2323 2324 2325 2326 2327 2328 2329 2330 2331 2332 2333 2334 2335 2336 2337 2338 2339 2340 2341 2342 2343 2344 2345 2346 2347 2348 2349 2350 2351 2352 2353 2354 2355 2356 2357 2358 2359 2360 2361 2362 2363 2364 2365 2366 2367 2368 2369 2370 2371 2372 2373 2374 2375 2376 2377 2378 2379 2380 2381 2382 2383 2384 2385 2386 2387 2388 2389 2390 2391 2392 2393 2394 2395 2396 2397 2398 2399 2400 2401 2402 2403 2404 2405 2406 2407 2408 2409 2410 2411 2412 2413 2414 2415 2416 2417 2418 2419 2420 2421 2422 2423 2424 2425 2426 2427 2428 2429 2430 2431 2432 2433 2434 2435 2436 2437 2438 2439 2440 2441 2442 2443 2444 2445 2446 2447 2448 2449 2450 2451 2452 2453 2454 2455 2456 2457 2458 2459 2460 2461 2462 2463 2464 2465 2466 2467 2468 2469 2470 2471 2472 2473 2474 2475 2476 2477 2478 2479 2480 2481 2482 2483 2484 2485 2486 2487 2488 2489 2490 2491 2492 2493 2494 2495 2496 2497 2498 2499 2500 2501 2502 2503 2504 2505 2506 2507 2508 2509 2510 2511 2512 2513 2514 2515 2516 2517 2518 2519 2520 2521 2522 2523 2524 2525 2526 2527 2528 2529 2530 2531 2532 2533 2534 2535 2536 2537 2538 2539 2540 2541 2542 2543 2544 2545 2546 2547 2548 2549 2550 2551 2552 2553 2554 2555 2556 2557 2558 2559 2560 2561 2562 2563 2564 2565 2566 2567 2568 2569 2570 2571 2572 2573 2574 2575 2576 2577 2578 2579 2580 2581 2582 2583 2584 2585 2586 2587 2588 2589 2590 2591 2592 2593 2594 2595 2596 2597 2598 2599 2600 2601 2602 2603 2604 2605 2606 2607 2608 2609 2610 2611 2612 2613 2614 2615 2616 2617 2618 2619 2620 2621 2622 2623 2624 2625 2626 2627 2628 2629 2630 2631 2632 2633 2634 2635 2636 2637 2638 2639 2640 2641 2642 2643 2644 2645 2646 2647 2648 2649 2650 2651 2652 2653 2654 2655 2656 2657 2658 2659 2660 2661 2662 2663 2664 2665 2666 2667 2668 2669 2670 2671 2672 2673 2674 2675 2676 2677 2678 2679 2680 2681 2682 2683 2684 2685 2686 2687 2688 2689 2690 2691 2692 2693 2694 2695 2696 2697 2698 2699 2700 2701 2702 2703 2704 2705 2706 2707 2708 2709 2710 2711 2712 2713 2714 2715 2716 2717 2718 2719 2720 2721 2722 2723 2724 2725 2726 2727 2728 2729 2730 2731 2732 2733 2734 2735 2736 2737 2738 2739 2740 2741 2742 2743 2744 2745 2746 2747 2748 2749 2750 2751 2752 2753 2754 2755 2756 2757 2758 2759 2760 2761 2762 2763 2764 2765 2766 2767 2768 2769 2770 2771 2772 2773 2774 2775 2776 2777 2778 2779 2780 2781 2782 2783 2784 2785 2786 2787 2788 2789 2790 2791 2792 2793 2794 2795 2796 2797 2798 2799 2800 2801 2802 2803 2804 2805 2806 2807 2808 2809 2810 2811 2812 2813 2814 2815 2816 2817 2818 2819 2820 2821 2822 2823 2824 2825 2826 2827 2828 2829 2830 2831 2832 2833 2834 2835 2836 2837 2838 2839 2840 2841 2842 2843 2844 2845 2846 2847 2848 2849 2850 2851 2852 2853 2854 2855 2856 2857 2858 2859 2860 2861 2862 2863 2864 2865 2866 2867 2868 2869 2870 2871 2872 2873 2874 2875 2876 2877 2878 2879 2880 2881 2882 2883 2884 2885 2886 2887 2888 2889 2890 2891 2892 2893 2894 2895 2896 2897 2898 2899 2900 2901 2902 2903 2904 2905 2906 2907 2908 2909 2910 2911 2912 2913 2914 2915 2916 2917 2918 2919 2920 2921 2922 2923 2924 2925 2926 2927 2928 2929 2930 2931 2932 2933 2934 2935 2936 2937 2938 2939 2940 2941 2942 2943 2944 2945 2946 2947 2948 2949 2950 2951 2952 2953 2954 2955 2956 2957 2958 2959 2960 2961 2962 2963 2964 2965 2966 2967 2968 2969 2970 2971 2972 2973 2974 2975 2976 2977 2978 2979 2980 2981 2982 2983 2984 2985 2986 2987 2988 2989 2990 2991 2992 2993 2994 2995 2996 2997 2998 2999 3000 3001 3002 3003 3004 3005 3006 3007 3008 3009 3010 3011 3012 3013 3014 3015 3016 3017 3018 3019 3020 3021 3022 3023 3024 3025 3026 3027 3028 3029 3030 3031 3032 3033 3034 3035 3036 3037 3038 3039 3040 3041 3042 3043 3044 3045 3046 3047 3048 3049 3050 3051 3052 3053 3054 3055 3056 3057 3058 3059 3060 3061 3062 3063 3064 3065 3066 3067 3068 3069 3070 3071 3072 3073 3074 3075 3076 3077 3078 3079 3080 3081 3082 3083 3084 3085 3086 3087 3088 3089 3090 3091 3092 3093 3094 3095 3096 3097 3098 3099 3100 3101 3102 3103 3104 3105 3106 3107 3108 3109 3110 3111 3112 3113 3114 3115 3116 3117 3118 3119 3120 3121 3122 3123 3124 3125 3126 3127 3128 3129 3130 3131 3132 3133 3134 3135 3136 3137 3138 3139 3140 3141 3142 3143 3144 3145 3146 3147 3148 3149 3150 3151 3152 3153 3154 3155 3156 3157 3158 3159 3160 3161 3162 3163 3164 3165 3166 3167 3168 3169 3170 3171 3172 3173 3174 3175 3176 3177 3178 3179 3180 3181 3182 3183 3184 3185 3186 3187 3188 3189 3190 3191 3192 3193 3194 3195 3196 3197 3198 3199 3200 3201 3202 3203 3204 3205 3206 3207 3208 3209 3210 3211 3212 3213 3214 3215 3216 3217 3218 3219 3220 3221 3222 3223 3224 3225 3226 3227 3228 3229 3230 3231 3232 3233 3234 3235 3236 3237 3238 3239 3240 3241 3242 3243 3244 3245 3246 3247 3248 3249 3250 3251 3252 3253 3254 3255 3256 3257 3258 3259 3260 3261 3262 3263 3264 3265 3266 3267 3268 3269 3270 3271 3272 3273 3274 3275 3276 3277 3278 3279 3280 3281 3282 3283 3284 3285 3286 3287 3288 3289 3290 3291 3292 3293 3294 3295 3296 3297 3298 3299 3300 3301 3302 3303 3304 3305 3306 3307 3308 3309 3310 3311 3312 3313 3314 3315 3316 3317 3318 3319 3320 3321 3322 3323 3324 3325 3326 3327 3328 3329 3330 3331 3332 3333 3334 3335 3336 3337 3338 3339 3340 3341 3342 3343 3344 3345 3346 3347 3348 3349 3350 3351 3352 3353 3354 3355 3356 3357 3358 3359 3360 3361 3362 3363 3364 3365 3366 3367 3368 3369 3370 3371 3372 3373 3374 3375 3376 3377 3378 3379 3380 3381 3382 3383 3384 3385 3386 3387 3388 3389 3390 3391 3392 3393 3394 3395 3396 3397 3398 3399 3400 3401 3402 3403 3404 3405 3406 3407 3408 3409 3410 3411 3412 3413 3414 3415 3416 3417 3418 3419 3420 3421 3422 3423 3424 3425 3426 3427 3428 3429 3430 3431 3432 3433 3434 3435 3436 3437 3438 3439 3440 3441 3442 3443 3444 3445 3446 3447 3448 3449 3450 3451 3452 3453 3454 3455 3456 3457 3458 3459 3460 3461 3462 3463 3464 3465 3466 3467 3468 3469 3470 3471 3472 3473 3474 3475 3476 3477 3478 3479 3480 3481 3482 3483 3484 3485 3486 3487 3488 3489 3490 3491 3492 3493 3494 3495 3496 3497 3498 3499 3500 3501 3502 3503 3504 3505 3506 3507 3508 3509 3510 3511 3512 3513 3514 3515 3516 3517 3518 3519 3520 3521 3522 3523 3524 3525 3526 3527 3528 3529 3530 3531 3532 3533 3534 3535 3536 3537 3538 3539 3540 3541 3542 3543 3544 3545 3546 3547 3548 3549 3550 3551 3552 3553 3554 3555 3556 3557 3558 3559 3560 3561 3562 3563 3564 3565 3566 3567 3568 3569 3570 3571 3572 3573 3574 3575 3576 3577 3578 3579 3580 3581 3582 3583 3584 3585 3586 3587 3588 3589 3590 3591 3592 3593 3594 3595 3596 3597 3598 3599 3600 3601 3602 3603 3604 3605 3606 3607 3608 3609 3610 3611 3612 3613 3614 3615 3616 3617 3618 3619 3620 3621 3622 3623 3624 3625 3626 3627 3628 3629 3630 3631 3632 3633 3634 3635 3636 3637 3638 3639 3640 3641 3642 3643 3644 3645 3646 3647 3648 3649 3650 3651 3652 3653 3654 3655 3656 3657 3658 3659 3660 3661 3662 3663 3664 3665 3666 3667 3668 3669 3670 3671 3672 3673 3674 3675 3676 3677 3678 3679 3680 3681 3682 3683 3684 3685 3686 3687 3688 3689 3690 3691 3692 3693 3694 3695 3696 3697 3698 3699 3700 3701 3702 3703 3704 3705 3706 3707 3708 3709 3710 3711 3712 3713 3714 3715 3716 3717 3718 3719 3720 3721 3722 3723 3724 3725 3726 3727 3728 3729 3730 3731 3732 3733 3734 3735 3736 3737 3738 3739 3740 3741 3742 3743 3744 3745 3746 3747 3748 3749 3750 3751 3752 3753 3754 3755 3756 3757 3758 3759 3760 3761 3762 3763 3764 3765 3766 3767 3768 3769 3770 3771 3772 3773 3774 3775 3776 3777 3778 3779 3780 3781 3782 3783 3784 3785 3786 3787 3788 3789 3790 3791 3792 3793 3794 3795 3796 3797 3798 3799 3800 3801 3802 3803 3804 3805 3806 3807 3808 3809 3810 3811 3812 3813 3814 3815 3816 3817 3818 3819 3820 3821 3822 3823 3824 3825 3826 3827 3828 3829 3830 3831 3832 3833 3834 3835 3836 3837 3838 3839 3840 3841 3842 3843 3844 3845 3846 3847 3848 3849 3850 3851 3852 3853 3854 3855 3856 3857 3858 3859 3860 3861 3862 3863 3864 3865 3866 3867 3868 3869 3870 3871 3872 3873 3874 3875 3876 3877 3878 3879 3880 3881 3882 3883 3884 3885 3886 3887 3888 3889 3890 3891 3892 3893 3894 3895 3896 3897 3898 3899 3900 3901 3902 3903 3904 3905 3906 3907 3908 3909 3910 3911 3912 3913 3914 3915 3916 3917 3918 3919 3920 3921 3922 3923 3924 3925 3926 3927 3928 3929 3930 3931 3932 3933 3934 3935 3936 3937 3938 3939 3940 3941 3942 3943 3944 3945 3946 3947 3948 3949 3950 3951 3952 3953 3954 3955 3956 3957 3958 3959 3960 3961 3962 3963 3964 3965 3966 3967 3968 3969 3970 3971 3972 3973 3974 3975 3976 3977 3978 3979 3980 3981 3982 3983 3984 3985 3986 3987 3988 3989 3990 3991 3992 3993 3994 3995 3996 3997 3998 3999 4000 4001 4002 4003 4004 4005 4006 4007 4008 4009 4010 4011 4012 4013 4014 4015 4016 4017 4018 4019 4020 4021 4022 4023 4024 4025 4026 4027 4028 4029 4030 4031 4032 4033 4034 4035 4036 4037 4038 4039 4040 4041 4042 4043 4044 4045 4046 4047 4048 4049 4050 4051 4052 4053 4054 4055 4056 4057 4058 4059 4060 4061 4062 4063 4064 4065 4066 4067 4068 4069 4070 4071 4072 4073 4074 4075 4076 4077 4078 4079 4080 4081 4082 4083 4084 4085 4086 4087 4088 4089 4090 4091 4092 4093 4094 4095 4096 4097 4098 4099 4100 4101 4102 4103 4104 4105 4106 4107 4108 4109 4110 4111 4112 4113 4114 4115 4116 4117 4118 4119 4120 4121 4122 4123 4124 4125 4126 4127 4128 4129 4130 4131 4132 4133 4134 4135 4136 4137 4138 4139 4140 4141 4142 4143 4144 4145 4146 4147 4148 4149 4150 4151 4152 4153 4154 4155 4156 4157 4158 4159 4160 4161 4162 4163 4164 4165 4166 4167 4168 4169 4170 4171 4172 4173 4174 4175 4176 4177 4178 4179 4180 4181 4182 4183 4184 4185 4186 4187 4188 4189 4190 4191 4192 4193 4194 4195 4196 4197 4198 4199 4200 4201 4202 4203 4204 4205 4206 4207 4208 4209 4210 4211 4212 4213 4214 4215 4216 4217 4218 4219 4220 4221 4222 4223 4224 4225 4226 4227 4228 4229 4230 4231 4232 4233 4234 4235 4236 4237 4238 4239 4240 4241 4242 4243 4244 4245 4246 4247 4248 4249 4250 4251 4252 4253 4254 4255 4256 4257 4258 4259 4260 4261 4262 4263 4264 4265 4266 4267 4268 4269 4270 4271 4272 4273 4274 4275 4276 4277 4278 4279 4280 4281 4282 4283 4284 4285 4286 4287 4288 4289 4290 4291 4292 4293 4294 4295 4296 4297 4298 4299 4300 4301 4302 4303 4304 4305 4306 4307 4308 4309 4310 4311 4312 4313 4314 4315 4316 4317 4318 4319 4320 4321 4322 4323 4324 4325 4326 4327 4328 4329 4330 4331 4332 4333 4334 4335 4336 4337 4338 4339 4340 4341 4342 4343 4344 4345 4346 4347 4348 4349 4350 4351 4352 4353 4354 4355 4356 4357 4358 4359 4360 4361 4362 4363 4364 4365 4366 4367 4368 4369 4370 4371 4372 4373 4374 4375 4376 4377 4378 4379 4380 4381 4382 4383 4384 4385 4386 4387 4388 4389 4390 4391 4392 4393 4394 4395 4396 4397 4398 4399 4400 4401 4402 4403 4404 4405 4406 4407 4408 4409 4410 4411 4412 4413 4414 4415 4416 4417 4418 4419 4420 4421 4422 4423 4424 4425 4426 4427 4428 4429 4430 4431 4432 4433 4434 4435 4436 4437 4438 4439 4440 4441 4442 4443 4444 4445 4446 4447 4448 4449 4450 4451 4452 4453 4454 4455 4456 4457 4458 4459 4460 4461 4462 4463 4464 4465 4466 4467 4468 4469 4470 4471 4472 4473 4474 4475 4476 4477 4478 4479 4480 4481 4482 4483 4484 4485 4486 4487 4488 4489 4490 4491 4492 4493 4494 4495 4496 4497 4498 4499 4500 4501 4502 4503 4504 4505 4506 4507 4508 4509 4510 4511 4512 4513 4514 4515 4516 4517 4518 4519 4520 4521 4522 4523 4524 4525 4526 4527 4528 4529 4530 4531 4532 4533 4534 4535 4536 4537 4538 4539 4540 4541 4542 4543 4544 4545 4546 4547 4548 4549 4550 4551 4552 4553 4554 4555 4556 4557 4558 4559 4560 4561 4562 4563 4564 4565 4566 4567 4568 4569 4570 4571 4572 4573 4574 4575 4576 4577 4578 4579 4580 4581 4582 4583 4584 4585 4586 4587 4588 4589 4590 4591 4592 4593 4594 4595 4596 4597 4598 4599 4600 4601 4602 4603 4604 4605 4606 4607 4608 4609 4610 4611 4612 4613 4614 4615 4616 4617 4618 4619 4620 4621 4622 4623 4624 4625 4626 4627 4628 4629 4630 4631 4632 4633 4634 4635 4636 4637 4638 4639 4640 4641 4642 4643 4644 4645 4646 4647 4648 4649 4650 4651 4652 4653 4654 4655 4656 4657 4658 4659 4660 4661 4662 4663 4664 4665 4666 4667 4668 4669 4670 4671 4672 4673 4674 4675 4676 4677 4678 4679 4680 4681 4682 4683 4684 4685 4686 4687 4688 4689 4690 4691 4692 4693 4694 4695 4696 4697 4698 4699 4700 4701 4702 4703 4704 4705 4706 4707 4708 4709 4710 4711 4712 4713 4714 4715 4716 4717 4718 4719 4720 4721 4722 4723 4724 4725 4726 4727 4728 4729 4730 4731 4732 4733 4734 4735 4736 4737 4738 4739 4740 4741 4742 4743 4744 4745 4746 4747 4748 4749 4750 4751 4752 4753 4754 4755 4756 4757 4758 4759 4760 4761 4762 4763 4764 4765 4766 4767 4768 4769 4770 4771 4772 4773 4774 4775 4776 4777 4778 4779 4780 4781 4782 4783 4784 4785 4786 4787 4788 4789 4790 4791 4792 4793 4794 4795 4796 4797 4798 4799 4800 4801 4802 4803 4804 4805 4806 4807 4808 4809 4810 4811 4812 4813 4814 4815 4816 4817 4818 4819 4820 4821 4822 4823 4824 4825 4826 4827 4828 4829 4830 4831 4832 4833 4834 4835 4836 4837 4838 4839 4840 4841 4842 4843 4844 4845 4846 4847 4848 4849 4850 4851 4852 4853 4854 4855 4856 4857 4858 4859 4860 4861 4862 4863 4864 4865 4866 4867 4868 4869 4870 4871 4872 4873 4874 4875 4876 4877 4878 4879 4880 4881 4882 4883 4884 4885 4886 4887 4888 4889 4890 4891 4892 4893 4894 4895 4896 4897 4898 4899 4900 4901 4902 4903 4904 4905 4906 4907 4908 4909 4910 4911 4912 4913 4914 4915 4916 4917 4918 4919 4920 4921 4922 4923 4924 4925 4926 4927 4928 4929 4930 4931 4932 4933 4934 4935 4936 4937 4938 4939 4940 4941 4942 4943 4944 4945 4946 4947 4948 4949 4950 4951 4952 4953 4954 4955 4956 4957 4958 4959 4960 4961 4962 4963 4964 4965 4966 4967 4968 4969 4970 4971 4972 4973 4974 4975 4976 4977 4978 4979 4980 4981 4982 4983 4984 4985 4986 4987 4988 4989 4990 4991 4992 4993 4994 4995 4996 4997 4998 4999 5000 5001 5002 5003 5004 5005 5006 5007 5008 5009 5010 5011 5012 5013 5014 5015 5016 5017 5018 5019 5020 5021 5022 5023 5024 5025 5026 5027 5028 5029 5030 5031 5032 5033 5034 5035 5036 5037 5038 5039 5040 5041 5042 5043 5044 5045 5046 5047 5048 5049 5050 5051 5052 5053 5054 5055 5056 5057 5058 5059 5060 5061 5062 5063 5064 5065 5066 5067 5068 5069 5070 5071 5072 5073 5074 5075 5076 5077 5078 5079 5080 5081 5082 5083 5084 5085 5086 5087 5088 5089 5090 5091 5092 5093 5094 5095 5096 5097 5098 5099 5100 5101 5102 5103 5104 5105 5106 5107 5108 5109 5110 5111 5112 5113 5114 5115 5116 5117 5118 5119 5120 5121 5122 5123 5124 5125 5126 5127 5128 5129 5130 5131 5132 5133 5134 5135 5136 5137 5138 5139 5140 5141 5142 5143 5144 5145 5146 5147 5148 5149 5150 5151 5152 5153 5154 5155 5156 5157 5158 5159 5160 5161 5162 5163 5164 5165 5166 5167 5168 5169 5170 5171 5172 5173 5174 5175 5176 5177 5178 5179 5180 5181 5182 5183 5184 5185 5186 5187 5188 5189 5190 5191 5192 5193 5194 5195 5196 5197 5198 5199 5200 5201 5202 5203 5204 5205 5206 5207 5208 5209 5210 5211 5212 5213 5214 5215 5216 5217 5218 5219 5220 5221 5222 5223 5224 5225 5226 5227 5228 5229 5230 5231 5232 5233 5234 5235 5236 5237 5238 5239 5240 5241 5242 5243 5244 5245 5246 5247 5248 5249 5250 5251 5252 5253 5254 5255 5256 5257 5258 5259 5260 5261 5262 5263 5264 5265 5266 5267 5268 5269 5270 5271 5272 5273 5274 5275 5276 5277 5278 5279 5280 5281 5282 5283 5284 5285 5286 5287 5288 5289 5290 5291 5292 5293 5294 5295 5296 5297 5298 5299 5300 5301 5302 5303 5304 5305 5306 5307 5308 5309 5310 5311 5312 5313 5314 5315 5316 5317 5318 5319 5320 5321 5322 5323 5324 5325 5326 5327 5328 5329 5330 5331 5332 5333 5334 5335 5336 5337 5338 5339 5340 5341 5342 5343 5344 5345 5346 5347 5348 5349 5350 5351 5352 5353 5354 5355 5356 5357 5358 5359 5360 5361 5362 5363 5364 5365 5366 5367 5368 5369 5370 5371 5372 5373 5374 5375 5376 5377 5378 5379 5380 5381 5382 5383 5384 5385 5386 5387 5388 5389 5390 5391 5392 5393 5394 5395 5396 5397 5398 5399 5400 5401 5402 5403 5404 5405 5406 5407 5408 5409 5410 5411 5412 5413 5414 5415 5416 5417 5418 5419 5420 5421 5422 5423 5424 5425 5426 5427 5428 5429 5430 5431 5432 5433 5434 5435 5436 5437 5438 5439 5440 5441 5442 5443 5444 5445 5446 5447 5448 5449 5450 5451 5452 5453 5454 5455 5456 5457 5458 5459 5460 5461 5462 5463 5464 5465 5466 5467 5468 5469 5470 5471 5472 5473 5474 5475 5476 5477 5478 5479 5480 5481 5482 5483 5484 5485 5486 5487 5488 5489 5490 5491 5492 5493 5494 5495 5496 5497 5498 5499 5500 5501 5502 5503 5504 5505 5506 5507 5508 5509 5510 5511 5512 5513 5514 5515 5516 5517 5518 5519 5520 5521 5522 5523 5524 5525 5526 5527 5528 5529 5530 5531 5532 5533 5534 5535 5536 5537 5538 5539 5540 5541 5542 5543 5544 5545 5546 5547 5548 5549 5550 5551 5552 5553 5554 5555 5556 5557 5558 5559 5560 5561 5562 5563 5564 5565 5566 5567 5568 5569 5570 5571 5572 5573 5574 5575 5576 5577 5578 5579 5580 5581 5582 5583 5584 5585 5586 5587 5588 5589 5590 5591 5592 5593 5594 5595 5596 5597 5598 5599 5600 5601 5602 5603 5604 5605 5606 5607 5608 5609 5610 5611 5612 5613 5614 5615 5616 5617 5618 5619 5620 5621 5622 5623 5624 5625 5626 5627 5628 5629 5630 5631 5632 5633 5634 5635 5636 5637 5638 5639 5640 5641 5642 5643 5644 5645 5646 5647 5648 5649 5650 5651 5652 5653 5654 5655 5656 5657 5658 5659 5660 5661 5662 5663 5664 5665 5666 5667 5668 5669 5670 5671 5672 5673 5674 5675 5676 5677 5678 5679 5680 5681 5682 5683 5684 5685 5686 5687 5688 5689 5690 5691 5692 5693 5694 5695 5696 5697 5698 5699 5700 5701 5702 5703 5704 5705 5706 5707 5708 5709 5710 5711 5712 5713 5714 5715 5716 5717 5718 5719 5720 5721 5722 5723 5724 5725 5726 5727 5728 5729 5730 5731 5732 5733 5734 5735 5736 5737 5738 5739 5740 5741 5742 5743 5744 5745 5746 5747 5748 5749 5750 5751 5752 5753 5754 5755 5756 5757 5758 5759 5760 5761 5762 5763 5764 5765 5766 5767 5768 5769 5770 5771 5772 5773 5774 5775 5776 5777 5778 5779 5780 5781 5782 5783 5784 5785 5786 5787 5788 5789 5790 5791 5792 5793 5794 5795 5796 5797 5798 5799 5800 5801 5802 5803 5804 5805 5806 5807 5808 5809 5810 5811 5812 5813 5814 5815 5816 5817 5818 5819 5820 5821 5822 5823 5824 5825 5826 5827 5828 5829 5830 5831 5832 5833 5834 5835 5836 5837 5838 5839 5840 5841 5842 5843 5844 5845 5846 5847 5848 5849 5850 5851 5852 5853 5854 5855 5856 5857 5858 5859 5860 5861 5862 5863 5864 5865 5866 5867 5868 5869 5870 5871 5872 5873 5874 5875 5876 5877 5878 5879 5880 5881 5882 5883 5884 5885 5886 5887 5888 5889 5890 5891 5892 5893 5894 5895 5896 5897 5898 5899 5900 5901 5902 5903 5904 5905 5906 5907 5908 5909 5910 5911 5912 5913 5914 5915 5916 5917 5918 5919 5920 5921 5922 5923 5924 5925 5926 5927 5928 5929 5930 5931 5932 5933 5934 5935 5936 5937 5938 5939 5940 5941 5942 5943 5944 5945 5946 5947 5948 5949 5950 5951 5952 5953 5954 5955 5956 5957 5958 5959 5960 5961 5962 5963 5964 5965 5966 5967 5968 5969 5970 5971 5972 5973 5974 5975 5976 5977 5978 5979 5980 5981 5982 5983 5984 5985 5986 5987 5988 5989 5990 5991 5992 5993 5994 5995 5996 5997 5998 5999 6000 6001 6002 6003 6004 6005 6006 6007 6008 6009 6010 6011 6012 6013 6014 6015 6016 6017 6018 6019 6020 6021 6022 6023 6024 6025 6026 6027 6028 6029 6030 6031 6032 6033 6034 6035 6036 6037 6038 6039 6040 6041 6042 6043 6044 6045 6046 6047 6048 6049 6050 6051 6052 6053 6054 6055 6056 6057 6058 6059 6060 6061 6062 6063 6064 6065 6066 6067 6068 6069 6070 6071 6072 6073 6074 6075 6076 6077 6078 6079 6080 6081 6082 6083 6084 6085 6086 6087 6088 6089 6090 6091 6092 6093 6094 6095 6096 6097 6098 6099 6100 6101 6102 6103 6104 6105 6106 6107 6108 6109 6110 6111 6112 6113 6114 6115 6116 6117 6118 6119 6120 6121 6122 6123 6124 6125 6126 6127 6128 6129 6130 6131 6132 6133 6134 6135 6136 6137 6138 6139 6140 6141 6142 6143 6144 6145 6146 6147 6148 6149 6150 6151 6152 6153 6154 6155 6156 6157 6158 6159 6160 6161 6162 6163 6164 6165 6166 6167 6168 6169 6170 6171 6172 6173 6174 6175 6176 6177 6178 6179 6180 6181 6182 6183 6184 6185 6186 6187 6188 6189 6190 6191 6192 6193 6194 6195 6196 6197 6198 6199 6200 6201 6202 6203 6204 6205 6206 6207 6208 6209 6210 6211 6212 6213 6214 6215 6216 6217 6218 6219 6220 6221 6222 6223 6224 6225 6226 6227 6228 6229 6230 6231 6232 6233 6234 6235 6236 6237 6238 6239 6240 6241 6242 6243 6244 6245 6246 6247 6248 6249 6250 6251 6252 6253 6254 6255 6256 6257 6258 6259 6260 6261 6262 6263 6264 6265 6266 6267 6268 6269 6270 6271 6272 6273 6274 6275 6276 6277 6278 6279 6280 6281 6282 6283 6284 6285 6286 6287 6288 6289 6290 6291 6292 6293 6294 6295 6296 6297 6298 6299 6300 6301 6302 6303 6304 6305 6306 6307 6308 6309 6310 6311 6312 6313 6314 6315 6316 6317 6318 6319 6320 6321 6322 6323 6324 6325 6326 6327 6328 6329 6330 6331 6332 6333 6334 6335 6336 6337 6338 6339 6340 6341 6342 6343 6344 6345 6346 6347 6348 6349 6350 6351 6352 6353 6354 6355 6356 6357 6358 6359 6360 6361 6362 6363 6364 6365 6366 6367 6368 6369 6370 6371 6372 6373 6374 6375 6376 6377 6378 6379 6380 6381 6382 6383 6384 6385 6386 6387 6388 6389 6390 6391 6392 6393 6394 6395 6396 6397 6398 6399 6400 6401 6402 6403 6404 6405 6406 6407 6408 6409 6410 6411 6412 6413 6414 6415 6416 6417 6418 6419 6420 6421 6422 6423 6424 6425 6426 6427 6428 6429 6430 6431 6432 6433 6434 6435 6436 6437 6438 6439 6440 6441 6442 6443 6444 6445 6446 6447 6448 6449 6450 6451 6452 6453 6454 6455 6456 6457 6458 6459 6460 6461 6462 6463 6464 6465 6466 6467 6468 6469 6470 6471 6472 6473 6474 6475 6476 6477 6478 6479 6480 6481 6482 6483 6484 6485 6486 6487 6488 6489 6490 6491 6492 6493 6494 6495 6496 6497 6498 6499 6500 6501 6502 6503 6504 6505 6506 6507 6508 6509 6510 6511 6512 6513 6514 6515 6516 6517 6518 6519 6520 6521 6522 6523 6524 6525 6526 6527 6528 6529 6530 6531 6532 6533 6534 6535 6536 6537 6538 6539 6540 6541 6542 6543 6544 6545 6546 6547 6548 6549 6550 6551 6552 6553 6554 6555 6556 6557 6558 6559 6560 6561 6562 6563 6564 6565 6566 6567 6568 6569 6570 6571 6572 6573 6574 6575 6576 6577 6578 6579 6580 6581 6582 6583 6584 6585 6586 6587 6588 6589 6590 6591 6592 6593 6594 6595 6596 6597 6598 6599 6600 6601 6602 6603 6604 6605 6606 6607 6608 6609 6610 6611 6612 6613 6614 6615 6616 6617 6618 6619 6620 6621 6622 6623 6624 6625 6626 6627 6628 6629 6630 6631 6632 6633 6634 6635 6636 6637 6638 6639 6640 6641 6642 6643 6644 6645 6646 6647 6648 6649 6650 6651 6652 6653 6654 6655 6656 6657 6658 6659 6660 6661 6662 6663 6664 6665 6666 6667 6668 6669 6670 6671 6672 6673 6674 6675 6676 6677 6678 6679 6680 6681 6682 6683 6684 6685 6686 6687 6688 6689 6690 6691 6692 6693 6694 6695 6696 6697 6698 6699 6700 6701 6702 6703 6704 6705 6706 6707 6708 6709 6710 6711 6712 6713 6714 6715 6716 6717 6718 6719 6720 6721 6722 6723 6724 6725 6726 6727 6728 6729 6730 6731 6732 6733 6734 6735 6736 6737 6738 6739 6740 6741 6742 6743 6744 6745 6746 6747 6748 6749 6750 6751 6752 6753 6754 6755 6756 6757 6758 6759 6760 6761 6762 6763 6764 6765 6766 6767 6768 6769 6770 6771 6772 6773 6774 6775 6776 6777 6778 6779 6780 6781 6782 6783 6784 6785 6786 6787 6788 6789 6790 6791 6792 6793 6794 6795 6796 6797 6798 6799 6800 6801 6802 6803 6804 6805 6806 6807 6808 6809 6810 6811 6812 6813 6814 6815 6816 6817 6818 6819 6820 6821 6822 6823 6824 6825 6826 6827 6828 6829 6830 6831 6832 6833 6834 6835 6836 6837 6838 6839 6840 6841 6842 6843 6844 6845 6846 6847 6848 6849 6850 6851 6852 6853 6854 6855 6856 6857 6858 6859 6860 6861 6862 6863 6864 6865 6866 6867 6868 6869 6870 6871 6872 6873 6874 6875 6876 6877 6878 6879 6880 6881 6882 6883 6884 6885 6886 6887 6888 6889 6890 6891 6892 6893 6894 6895 6896 6897 6898 6899 6900 6901 6902 6903 6904 6905 6906 6907 6908 6909 6910 6911 6912 6913 6914 6915 6916 6917 6918 6919 6920 6921 6922 6923 6924 6925 6926 6927 6928 6929 6930 6931 6932 6933 6934 6935 6936 6937 6938 6939 6940 6941 6942 6943 6944 6945 6946 6947 6948 6949 6950 6951 6952 6953 6954 6955 6956 6957 6958 6959 6960 6961 6962 6963 6964 6965 6966 6967 6968 6969 6970 6971 6972 6973 6974 6975 6976 6977 6978 6979 6980 6981 6982 6983 6984 6985 6986 6987 6988 6989 6990 6991 6992 6993 6994 6995 6996 6997 6998 6999 7000 7001 7002 7003 7004 7005 7006 7007 7008 7009 7010 7011 7012 7013 7014 7015 7016 7017 7018 7019 7020 7021 7022 7023 7024 7025 7026 7027 7028 7029 7030 7031 7032 7033 7034 7035 7036 7037 7038 7039 7040 7041 7042 7043 7044 7045 7046 7047 7048 7049 7050 7051 7052 7053 7054 7055 7056 7057 7058 7059 7060 7061 7062 7063 7064 7065 7066 7067 7068 7069 7070 7071 7072 7073 7074 7075 7076 7077 7078 7079 7080 7081 7082 7083 7084 7085 7086 7087 7088 7089 7090 7091 7092 7093 7094 7095 7096 7097 7098 7099 7100 7101 7102 7103 7104 7105 7106 7107 7108 7109 7110 7111 7112 7113 7114 7115 7116 7117 7118 7119 7120 7121 7122 7123 7124 7125 7126 7127 7128 7129 7130 7131 7132 7133 7134 7135 7136 7137 7138 7139 7140 7141 7142 7143 7144 7145 7146 7147 7148 7149 7150 7151 7152 7153 7154 7155 7156 7157 7158 7159 7160 7161 7162 7163 7164 7165 7166 7167 7168 7169 7170 7171 7172 7173 7174 7175 7176 7177 7178 7179 7180 7181 7182 7183 7184 7185 7186 7187 7188 7189 7190 7191 7192 7193 7194 7195 7196 7197 7198 7199 7200 7201 7202 7203 7204 7205 7206 7207 7208 7209 7210 7211 7212 7213 7214 7215 7216 7217 7218 7219 7220 7221 7222 7223 7224 7225 7226 7227 7228 7229 7230 7231 7232 7233 7234 7235 7236 7237 7238 7239 7240 7241 7242 7243 7244 7245 7246 7247 7248 7249 7250 7251 7252 7253 7254 7255 7256 7257 7258 7259 7260 7261 7262 7263 7264 7265 7266 7267 7268 7269 7270 7271 7272 7273 7274 7275 7276 7277 7278 7279 7280 7281 7282 7283 7284 7285 7286 7287 7288 7289 7290 7291 7292 7293 7294 7295 7296 7297 7298 7299 7300 7301 7302 7303 7304 7305 7306 7307 7308 7309 7310 7311 7312 7313 7314 7315 7316 7317 7318 7319 7320 7321 7322 7323 7324 7325 7326 7327 7328 7329 7330 7331 7332 7333 7334 7335 7336 7337 7338 7339 7340 7341 7342 7343 7344 7345 7346 7347 7348 7349 7350 7351 7352 7353 7354 7355 7356 7357 7358 7359 7360 7361 7362 7363 7364 7365 7366 7367 7368 7369 7370 7371 7372 7373 7374 7375 7376 7377 7378 7379 7380 7381 7382 7383 7384 7385 7386 7387 7388 7389 7390 7391 7392 7393 7394 7395 7396 7397 7398 7399 7400 7401 7402 7403 7404 7405 7406 7407 7408 7409 7410 7411 7412 7413 7414 7415 7416 7417 7418 7419 7420 7421 7422 7423 7424 7425 7426 7427 7428 7429 7430 7431 7432 7433 7434 7435 7436 7437 7438 7439 7440 7441 7442 7443 7444 7445 7446 7447 7448 7449 7450 7451 7452 7453 7454 7455 7456 7457 7458 7459 7460 7461 7462 7463 7464 7465 7466 7467 7468 7469 7470 7471 7472 7473 7474 7475 7476 7477 7478 7479 7480 7481 7482 7483 7484 7485 7486 7487 7488 7489 7490 7491 7492 7493 7494 7495 7496 7497 7498 7499 7500 7501 7502 7503 7504 7505 7506 7507 7508 7509 7510 7511 7512 7513 7514 7515 7516 7517 7518 7519 7520 7521 7522 7523 7524 7525 7526 7527 7528 7529 7530 7531 7532 7533 7534 7535 7536 7537 7538 7539 7540 7541 7542 7543 7544 7545 7546 7547 7548 7549 7550 7551 7552 7553 7554 7555 7556 7557 7558 7559 7560 7561 7562 7563 7564 7565 7566 7567 7568 7569 7570 7571 7572 7573 7574 7575 7576 7577 7578 7579 7580 7581 7582 7583 7584 7585 7586 7587 7588 7589 7590 7591 7592 7593 7594 7595 7596 7597 7598 7599 7600 7601 7602 7603 7604 7605 7606 7607 7608 7609 7610 7611 7612 7613 7614 7615 7616 7617 7618 7619 7620 7621 7622 7623 7624 7625 7626 7627 7628 7629 7630 7631 7632 7633 7634 7635 7636 7637 7638 7639 7640 7641 7642 7643 7644 7645 7646 7647 7648 7649 7650 7651 7652 7653 7654 7655 7656 7657 7658 7659 7660 7661 7662 7663 7664 7665 7666 7667 7668 7669 7670 7671 7672 7673 7674 7675 7676 7677 7678 7679 7680 7681 7682 7683 7684 7685 7686 7687 7688 7689 7690 7691 7692 7693 7694 7695 7696 7697 7698 7699 7700 7701 7702 7703 7704 7705 7706 7707 7708 7709 7710 7711 7712 7713 7714 7715 7716 7717 7718 7719 7720 7721 7722 7723 7724 7725 7726 7727 7728 7729 7730 7731 7732 7733 7734 7735 7736 7737 7738 7739 7740 7741 7742 7743 7744 7745 7746 7747 7748 7749 7750 7751 7752 7753 7754 7755 7756 7757 7758 7759 7760 7761 7762 7763 7764 7765 7766 7767 7768 7769 7770 7771 7772 7773 7774 7775 7776 7777 7778 7779 7780 7781 7782 7783 7784 7785 7786 7787 7788 7789 7790 7791 7792 7793 7794 7795 7796 7797 7798 7799 7800 7801 7802 7803 7804 7805 7806 7807 7808 7809 7810 7811 7812 7813 7814 7815 7816 7817 7818 7819 7820 7821 7822 7823 7824 7825 7826 7827 7828 7829 7830 7831 7832 7833 7834 7835 7836 7837 7838 7839 7840 7841 7842 7843 7844 7845 7846 7847 7848 7849 7850 7851 7852 7853 7854 7855 7856 7857 7858 7859 7860 7861 7862 7863 7864 7865 7866 7867 7868 7869 7870 7871 7872 7873 7874 7875 7876 7877 7878 7879 7880 7881 7882 7883 7884 7885 7886 7887 7888 7889 7890 7891 7892 7893 7894 7895 7896 7897 7898 7899 7900 7901 7902 7903 7904 7905 7906 7907 7908 7909 7910 7911 7912 7913 7914 7915 7916 7917 7918 7919 7920 7921 7922 7923 7924 7925 7926 7927 7928 7929 7930 7931 7932 7933 7934 7935 7936 7937 7938 7939 7940 7941 7942 7943 7944 7945 7946 7947 7948 7949 7950 7951 7952 7953 7954 7955 7956 7957 7958 7959 7960 7961 7962 7963 7964 7965 7966 7967 7968 7969 7970 7971 7972 7973 7974 7975 7976 7977 7978 7979 7980 7981 7982 7983 7984 7985 7986 7987 7988 7989 7990 7991 7992 7993 7994 7995 7996 7997 7998 7999 8000 8001 8002 8003 8004 8005 8006 8007 8008 8009 8010 8011 8012 8013 8014 8015 8016 8017 8018 8019 8020 8021 8022 8023 8024 8025 8026 8027 8028 8029 8030 8031 8032 8033 8034 8035 8036 8037 8038 8039 8040 8041 8042 8043 8044 8045 8046 8047 8048 8049 8050 8051 8052 8053 8054 8055 8056 8057 8058 8059 8060 8061 8062 8063 8064 8065 8066 8067 8068 8069 8070 8071 8072 8073 8074 8075 8076 8077 8078 8079 8080 8081 8082 8083 8084 8085 8086 8087 8088 8089 8090 8091 8092 8093 8094 8095 8096 8097 8098 8099 8100 8101 8102 8103 8104 8105 8106 8107 8108 8109 8110 8111 8112 8113 8114 8115 8116 8117 8118 8119 8120 8121 8122 8123 8124 8125 8126 8127 8128 8129 8130 8131 8132 8133 8134 8135 8136 8137 8138 8139 8140 8141 8142 8143 8144 8145 8146 8147 8148 8149 8150 8151 8152 8153 8154 8155 8156 8157 8158 8159 8160 8161 8162 8163 8164 8165 8166 8167 8168 8169 8170 8171 8172 8173 8174 8175 8176 8177 8178 8179 8180 8181 8182 8183 8184 8185 8186 8187 8188 8189 8190 8191 8192 8193 8194 8195 8196 8197 8198 8199 8200 8201 8202 8203 8204 8205 8206 8207 8208 8209 8210 8211 8212 8213 8214 8215 8216 8217 8218 8219 8220 8221 8222 8223 8224 8225 8226 8227 8228 8229 8230 8231 8232 8233 8234 8235 8236 8237 8238 8239 8240 8241 8242 8243 8244 8245 8246 8247 8248 8249 8250 8251 8252 8253 8254 8255 8256 8257 8258 8259 8260 8261 8262 8263 8264 8265 8266 8267 8268 8269 8270 8271 8272 8273 8274 8275 8276 8277 8278 8279 8280 8281 8282 8283 8284 8285 8286 8287 8288 8289 8290 8291 8292 8293 8294 8295 8296 8297 8298 8299 8300 8301 8302 8303 8304 8305 8306 8307 8308 8309 8310 8311 8312 8313 8314 8315 8316 8317 8318 8319 8320 8321 8322 8323 8324 8325 8326 8327 8328 8329 8330 8331 8332 8333 8334 8335 8336 8337 8338 8339 8340 8341 8342 8343 8344 8345 8346 8347 8348 8349 8350 8351 8352 8353 8354 8355 8356 8357 8358 8359 8360 8361 8362 8363 8364 8365 8366 8367 8368 8369 8370 8371 8372 8373 8374 8375 8376 8377 8378 8379 8380 8381 8382 8383 8384 8385 8386 8387 8388 8389 8390 8391 8392 8393 8394 8395 8396 8397 8398 8399 8400 8401 8402 8403 8404 8405 8406 8407 8408 8409 8410 8411 8412 8413 8414 8415 8416 8417 8418 8419 8420 8421 8422 8423 8424 8425 8426 8427 8428 8429 8430 8431 8432 8433 8434 8435 8436 8437 8438 8439 8440 8441 8442 8443 8444 8445 8446 8447 8448 8449 8450 8451 8452 8453 8454 8455 8456 8457 8458 8459 8460 8461 8462 8463 8464 8465 8466 8467 8468 8469 8470 8471 8472 8473 8474 8475 8476 8477 8478 8479 8480 8481 8482 8483 8484 8485 8486 8487 8488 8489 8490 8491 8492 8493 8494 8495 8496 8497 8498 8499 8500 8501 8502 8503 8504 8505 8506 8507 8508 8509 8510 8511 8512 8513 8514 8515 8516 8517 8518 8519 8520 8521 8522 8523 8524 8525 8526 8527 8528 8529 8530 8531 8532 8533 8534 8535 8536 8537 8538 8539 8540 8541 8542 8543 8544 8545 8546 8547 8548 8549 8550 8551 8552 8553 8554 8555 8556 8557 8558 8559 8560 8561 8562 8563 8564 8565 8566 8567 8568 8569 8570 8571 8572 8573 8574 8575 8576 8577 8578 8579 8580 8581 8582 8583 8584 8585 8586 8587 8588 8589 8590 8591 8592 8593 8594 8595 8596 8597 8598 8599 8600 8601 8602 8603 8604 8605 8606 8607 8608 8609 8610 8611 8612 8613 8614 8615 8616 8617 8618 8619 8620 8621 8622 8623 8624 8625 8626 8627 8628 8629 8630 8631 8632 8633 8634 8635 8636 8637 8638 8639 8640 8641 8642 8643 8644 8645 8646 8647 8648 8649 8650 8651 8652 8653 8654 8655 8656 8657 8658 8659 8660 8661 8662 8663 8664 8665 8666 8667 8668 8669 8670 8671 8672 8673 8674 8675 8676 8677 8678 8679 8680 8681 8682 8683 8684 8685 8686 8687 8688 8689 8690 8691 8692 8693 8694 8695 8696 8697 8698 8699 8700 8701 8702 8703 8704 8705 8706 8707 8708 8709 8710 8711 8712 8713 8714 8715 8716 8717 8718 8719 8720 8721 8722 8723 8724 8725 8726 8727 8728 8729 8730 8731 8732 8733 8734 8735 8736 8737 8738 8739 8740 8741 8742 8743 8744 8745 8746 8747 8748 8749 8750 8751 8752 8753 8754 8755 8756 8757 8758 8759 8760 8761 8762 8763 8764 8765 8766 8767 8768 8769 8770 8771 8772 8773 8774 8775 8776 8777 8778 8779 8780 8781 8782 8783 8784 8785 8786 8787 8788 8789 8790 8791 8792 8793 8794 8795 8796 8797 8798 8799 8800 8801 8802 8803 8804 8805 8806 8807 8808 8809 8810 8811 8812 8813 8814 8815 8816 8817 8818 8819 8820 8821 8822 8823 8824 8825 8826 8827 8828 8829 8830 8831 8832 8833 8834 8835 8836 8837 8838 8839 8840 8841 8842 8843 8844 8845 8846 8847 8848 8849 8850 8851 8852 8853 8854 8855 8856 8857 8858 8859 8860 8861 8862 8863 8864 8865 8866 8867 8868 8869 8870 8871 8872 8873 8874 8875 8876 8877 8878 8879 8880 8881 8882 8883 8884 8885 8886 8887 8888 8889 8890 8891 8892 8893 8894 8895 8896 8897 8898 8899 8900 8901 8902 8903 8904 8905 8906 8907 8908 8909 8910 8911 8912 8913 8914 8915 8916 8917 8918 8919 8920 8921 8922 8923 8924 8925 8926 8927 8928 8929 8930 8931 8932 8933 8934 8935 8936 8937 8938 8939 8940 8941 8942 8943 8944 8945 8946 8947 8948 8949 8950 8951 8952 8953 8954 8955 8956 8957 8958 8959 8960 8961 8962 8963 8964 8965 8966 8967 8968 8969 8970 8971 8972 8973 8974 8975 8976 8977 8978 8979 8980 8981 8982 8983 8984 8985 8986 8987 8988 8989 8990 8991 8992 8993 8994 8995 8996 8997 8998 8999 9000 9001 9002 9003 9004 9005 9006 9007 9008 9009 9010 9011 9012 9013 9014 9015 9016 9017 9018 9019 9020 9021 9022 9023 9024 9025 9026 9027 9028 9029 9030 9031 9032 9033 9034 9035 9036 9037 9038 9039 9040 9041 9042 9043 9044 9045 9046 9047 9048 9049 9050 9051 9052 9053 9054 9055 9056 9057 9058 9059 9060 9061 9062 9063 9064 9065 9066 9067 9068 9069 9070 9071 9072 9073 9074 9075 9076 9077 9078 9079 9080 9081 9082 9083 9084 9085 9086 9087 9088 9089 9090 9091 9092 9093 9094 9095 9096 9097 9098 9099 9100 9101 9102 9103 9104 9105 9106 9107 9108 9109 9110 9111 9112 9113 9114 9115 9116 9117 9118 9119 9120 9121 9122 9123 9124 9125 9126 9127 9128 9129 9130 9131 9132 9133 9134 9135 9136 9137 9138 9139 9140 9141 9142 9143 9144 9145 9146 9147 9148 9149 9150 9151 9152 9153 9154 9155 9156 9157 9158 9159 9160 9161 9162 9163 9164 9165 9166 9167 9168 9169 9170 9171 9172 9173 9174 9175 9176 9177 9178 9179 9180 9181 9182 9183 9184 9185 9186 9187 9188 9189 9190 9191 9192 9193 9194 9195 9196 9197 9198 9199 9200 9201 9202 9203 9204 9205 9206 9207 9208 9209 9210 9211 9212 9213 9214 9215 9216 9217 9218 9219 9220 9221 9222 9223 9224 9225 9226 9227 9228 9229 9230 9231 9232 9233 9234 9235 9236 9237 9238 9239 9240 9241 9242 9243 9244 9245 9246 9247 9248 9249 9250 9251 9252 9253 9254 9255 9256 9257 9258 9259 9260 9261 9262 9263 9264 9265 9266 9267 9268 9269 9270 9271 9272 9273 9274 9275 9276 9277 9278 9279 9280 9281 9282 9283 9284 9285 9286 9287 9288 9289 9290 9291 9292 9293 9294 9295 9296 9297 9298 9299 9300 9301 9302 9303 9304 9305 9306 9307 9308 9309 9310 9311 9312 9313 9314 9315 9316 9317 9318 9319 9320 9321 9322 9323 9324 9325 9326 9327 9328 9329 9330 9331 9332 9333 9334 9335 9336 9337 9338 9339 9340 9341 9342 9343 9344 9345 9346 9347 9348 9349 9350 9351 9352 9353 9354 9355 9356 9357 9358 9359 9360 9361 9362 9363 9364 9365 9366 9367 9368 9369 9370 9371 9372 9373 9374 9375 9376 9377 9378 9379 9380 9381 9382 9383 9384 9385 9386 9387 9388 9389 9390 9391 9392 9393 9394 9395 9396 9397 9398 9399 9400 9401 9402 9403 9404 9405 9406 9407 9408 9409 9410 9411 9412 9413 9414 9415 9416 9417 9418 9419 9420 9421 9422 9423 9424 9425 9426 9427 9428 9429 9430 9431 9432 9433 9434 9435 9436 9437 9438 9439 9440 9441 9442 9443 9444 9445 9446 9447 9448 9449 9450 9451 9452 9453 9454 9455 9456 9457 9458 9459 9460 9461 9462 9463 9464 9465 9466 9467 9468 9469 9470 9471 9472 9473 9474 9475 9476 9477 9478 9479 9480 9481 9482 9483 9484 9485 9486 9487 9488 9489 9490 9491 9492 9493 9494 9495 9496 9497 9498 9499 9500 9501 9502 9503 9504 9505 9506 9507 9508 9509 9510 9511 9512 9513 9514 9515 9516 9517 9518 9519 9520 9521 9522 9523 9524 9525 9526 9527 9528 9529 9530 9531 9532 9533 9534 9535 9536 9537 9538 9539 9540 9541 9542 9543 9544 9545 9546 9547 9548 9549 9550 9551 9552 9553 9554 9555 9556 9557 9558 9559 9560 9561 9562 9563 9564 9565 9566 9567 9568 9569 9570 9571 9572 9573 9574 9575 9576 9577 9578 9579 9580 9581 9582 9583 9584 9585 9586 9587 9588 9589 9590 9591 9592 9593 9594 9595 9596 9597 9598 9599 9600 9601 9602 9603 9604 9605 9606 9607 9608 9609 9610 9611 9612 9613 9614 9615 9616 9617 9618 9619 9620 9621 9622 9623 9624 9625 9626 9627 9628 9629 9630 9631 9632 9633 9634 9635 9636 9637 9638 9639 9640 9641 9642 9643 9644 9645 9646 9647 9648 9649 9650 9651 9652 9653 9654 9655 9656 9657 9658 9659 9660 9661 9662 9663 9664 9665 9666 9667 9668 9669 9670 9671 9672 9673 9674 9675 9676 9677 9678 9679 9680 9681 9682 9683 9684 9685 9686 9687 9688 9689 9690 9691 9692 9693 9694 9695 9696 9697 9698 9699 9700 9701 9702 9703 9704 9705 9706 9707 9708 9709 9710 9711 9712 9713 9714 9715 9716 9717 9718 9719 9720 9721 9722 9723 9724 9725 9726 9727 9728 9729 9730 9731 9732 9733 9734 9735 9736 9737 9738 9739 9740 9741 9742 9743 9744 9745 9746 9747 9748 9749 9750 9751 9752 9753 9754 9755 9756 9757 9758 9759 9760 9761 9762 9763 9764 9765 9766 9767 9768 9769 9770 9771 9772 9773 9774 9775 9776 9777 9778 9779 9780 9781 9782 9783 9784 9785 9786 9787 9788 9789 9790 9791 9792 9793 9794 9795 9796 9797 9798 9799 9800 9801 9802 9803 9804 9805 9806 9807 9808 9809 9810 9811 9812 9813 9814 9815 9816 9817 9818 9819 9820 9821 9822 9823 9824 9825 9826 9827 9828 9829 9830 9831 9832 9833 9834 9835 9836 9837 9838 9839 9840 9841 9842 9843 9844 9845 9846 9847 9848 9849 9850 9851 9852 9853 9854 9855 9856 9857 9858 9859 9860 9861 9862 9863 9864 9865 9866 9867 9868 9869 9870 9871 9872 9873 9874 9875 9876 9877 9878 9879 9880 9881 9882 9883 9884 9885 9886 9887 9888 9889 9890 9891 9892 9893 9894 9895 9896 9897 9898 9899 9900 9901 9902 9903 9904 9905 9906 9907 9908 9909 9910 9911 9912 9913 9914 9915 9916 9917 9918 9919 9920 9921 9922 9923 9924 9925 9926 9927 9928 9929 9930 9931 9932 9933 9934 9935 9936 9937 9938 9939 9940 9941 9942 9943 9944 9945 9946 9947 9948 9949 9950 9951 9952 9953 9954 9955 9956 9957 9958 9959 9960 9961 9962 9963 9964 9965 9966 9967 9968 9969 9970 9971 9972 9973 9974 9975 9976 9977 9978 9979 9980 9981 9982 9983 9984 9985 9986 9987 9988 9989 9990 9991 9992 9993 9994 9995 9996 9997 9998 9999 10000 10001 10002 10003 10004 10005 10006 10007 10008 10009 10010 10011 10012 10013 10014 10015 10016 10017 10018 10019 10020 10021 10022 10023 10024 10025 10026 10027 10028 10029 10030 10031 10032 10033 10034 10035 10036 10037 10038 10039 10040 10041 10042 10043 10044 10045 10046 10047 10048 10049 10050 10051 10052 10053 10054 10055 10056 10057 10058 10059 10060 10061 10062 10063 10064 10065 10066 10067 10068 10069 10070 10071 10072 10073 10074 10075 10076 10077 10078 10079 10080 10081 10082 10083 10084 10085 10086 10087 10088 10089 10090 10091 10092 10093 10094 10095 10096 10097 10098 10099 10100 10101 10102 10103 10104 10105 10106 10107 10108 10109 10110 10111 10112 10113 10114 10115 10116 10117 10118 10119 10120 10121 10122 10123 10124 10125 10126 10127 10128 10129 10130 10131 10132 10133 10134 10135 10136 10137 10138 10139 10140 10141 10142 10143 10144 10145 10146 10147 10148 10149 10150 10151 10152 10153 10154 10155 10156 10157 10158 10159 10160 10161 10162 10163 10164 10165 10166 10167 10168 10169 10170 10171 10172 10173 10174 10175 10176 10177 10178 10179 10180 10181 10182 10183 10184 10185 10186 10187 10188 10189 10190 10191 10192 10193 10194 10195 10196 10197 10198 10199 10200 10201 10202 10203 10204 10205 10206 10207 10208 10209 10210 10211 10212 10213 10214 10215 10216 10217 10218 10219 10220 10221 10222 10223 10224 10225 10226 10227 10228 10229 10230 10231 10232 10233 10234 10235 10236 10237 10238 10239 10240 10241 10242 10243 10244 10245 10246 10247 10248 10249 10250 10251 10252 10253 10254 10255 10256 10257 10258 10259 10260 10261 10262 10263 10264 10265 10266 10267 10268 10269 10270 10271 10272 10273 10274 10275 10276 10277 10278 10279 10280 10281 10282 10283 10284 10285 10286 10287 10288 10289 10290 10291 10292 10293 10294 10295 10296 10297 10298 10299 10300 10301 10302 10303 10304 10305 10306 10307 10308 10309 10310 10311 10312 10313 10314 10315 10316 10317 10318 10319 10320 10321 10322 10323 10324 10325 10326 10327 10328 10329 10330 10331 10332 10333 10334 10335 10336 10337 10338 10339 10340 10341 10342 10343 10344 10345 10346 10347 10348 10349 10350 10351 10352 10353 10354 10355 10356 10357 10358 10359 10360 10361 10362 10363 10364 10365 10366 10367 10368 10369 10370 10371 10372 10373 10374 10375 10376 10377 10378 10379 10380 10381 10382 10383 10384 10385 10386 10387 10388 10389 10390 10391 10392 10393 10394 10395 10396 10397 10398 10399 10400 10401 10402 10403 10404 10405 10406 10407 10408 10409 10410 10411 10412 10413 10414 10415 10416 10417 10418 10419 10420 10421 10422 10423 10424 10425 10426 10427 10428 10429 10430 10431 10432 10433 10434 10435 10436 10437 10438 10439 10440 10441 10442 10443 10444 10445 10446 10447 10448 10449 10450 10451 10452 10453 10454 10455 10456 10457 10458 10459 10460 10461 10462 10463 10464 10465 10466 10467 10468 10469 10470 10471 10472 10473 10474 10475 10476 10477 10478 10479 10480 10481 10482 10483 10484 10485 10486 10487 10488 10489 10490 10491 10492 10493 10494 10495 10496 10497 10498 10499 10500 10501 10502 10503 10504 10505 10506 10507 10508 10509 10510 10511 10512 10513 10514 10515 10516 10517 10518 10519 10520 10521 10522 10523 10524 10525 10526 10527 10528 10529
|
/* Copyright (c) 2000, 2025, Oracle and/or its affiliates.
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License, version 2.0,
as published by the Free Software Foundation.
This program is designed to work with certain software (including
but not limited to OpenSSL) that is licensed under separate terms,
as designated in a particular file or component or in included license
documentation. The authors of MySQL hereby grant you an additional
permission to link the program and your derivative works with the
separately licensed software that they have either included with
the program or referenced in the documentation.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License, version 2.0, for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA */
/* Basic functions needed by many modules */
#include "sql/sql_base.h"
#include <fcntl.h>
#include <limits.h>
#include <stdio.h>
#include <string.h>
#include <time.h>
#include <atomic>
#include <functional>
#include <memory>
#include <unordered_map>
#include <utility>
#include "ft_global.h"
#include "libbinlogevents/include/table_id.h"
#include "m_ctype.h"
#include "m_string.h"
#include "map_helpers.h"
#include "mf_wcomp.h" // wild_one, wild_many
#include "mutex_lock.h"
#include "my_alloc.h"
#include "my_bitmap.h"
#include "my_byteorder.h"
#include "my_compiler.h"
#include "my_dbug.h"
#include "my_dir.h"
#include "my_io.h"
#include "my_loglevel.h"
#include "my_macros.h"
#include "my_psi_config.h"
#include "my_sqlcommand.h"
#include "my_sys.h"
#include "my_systime.h"
#include "my_table_map.h"
#include "my_thread_local.h"
#include "mysql/components/services/bits/mysql_cond_bits.h"
#include "mysql/components/services/bits/psi_bits.h"
#include "mysql/components/services/bits/psi_cond_bits.h"
#include "mysql/components/services/bits/psi_mutex_bits.h"
#include "mysql/components/services/log_builtins.h"
#include "mysql/plugin.h"
#include "mysql/psi/mysql_cond.h"
#include "mysql/psi/mysql_file.h"
#include "mysql/psi/mysql_mutex.h"
#include "mysql/psi/mysql_statement.h"
#include "mysql/psi/mysql_table.h"
#include "mysql/psi/mysql_thread.h"
#include "mysql/psi/psi_table.h"
#include "mysql/service_mysql_alloc.h"
#include "mysql/thread_type.h"
#include "mysql_com.h"
#include "mysqld_error.h"
#include "sql/auth/auth_acls.h"
#include "sql/auth/auth_common.h" // check_table_access
#include "sql/auth/sql_security_ctx.h"
#include "sql/binlog.h" // mysql_bin_log
#include "sql/check_stack.h"
#include "sql/dd/cache/dictionary_client.h"
#include "sql/dd/dd_schema.h"
#include "sql/dd/dd_table.h" // dd::table_exists
#include "sql/dd/dd_tablespace.h" // dd::fill_table_and_parts_tablespace_name
#include "sql/dd/string_type.h"
#include "sql/dd/types/abstract_table.h"
#include "sql/dd/types/column.h"
#include "sql/dd/types/column_statistics.h"
#include "sql/dd/types/foreign_key.h" // dd::Foreign_key
#include "sql/dd/types/function.h"
#include "sql/dd/types/procedure.h"
#include "sql/dd/types/schema.h"
#include "sql/dd/types/table.h" // dd::Table
#include "sql/dd/types/view.h"
#include "sql/dd_table_share.h" // open_table_def
#include "sql/debug_sync.h" // DEBUG_SYNC
#include "sql/derror.h" // ER_THD
#include "sql/error_handler.h" // Internal_error_handler
#include "sql/field.h"
#include "sql/handler.h"
#include "sql/histograms/histogram.h"
#include "sql/item.h"
#include "sql/item_cmpfunc.h" // Item_func_eq
#include "sql/item_func.h"
#include "sql/item_subselect.h"
#include "sql/lock.h" // mysql_lock_remove
#include "sql/log.h"
#include "sql/log_event.h" // Query_log_event
#include "sql/mysqld.h" // replica_open_temp_tables
#include "sql/mysqld_thd_manager.h" // Global_THD_manage
#include "sql/nested_join.h"
#include "sql/partition_info.h" // partition_info
#include "sql/psi_memory_key.h" // key_memory_TABLE
#include "sql/query_options.h"
#include "sql/rpl_gtid.h"
#include "sql/rpl_handler.h" // RUN_HOOK
#include "sql/rpl_replica_commit_order_manager.h" // has_commit_order_manager
#include "sql/rpl_rli.h" //Relay_log_information
#include "sql/session_tracker.h"
#include "sql/sp.h" // Sroutine_hash_entry
#include "sql/sp_cache.h" // sp_cache_version
#include "sql/sp_head.h" // sp_head
#include "sql/sql_audit.h" // mysql_audit_table_access_notify
#include "sql/sql_backup_lock.h" // acquire_shared_backup_lock
#include "sql/sql_class.h" // THD
#include "sql/sql_const.h"
#include "sql/sql_data_change.h"
#include "sql/sql_db.h" // check_schema_readonly
#include "sql/sql_error.h" // Sql_condition
#include "sql/sql_executor.h" // unwrap_rollup_group
#include "sql/sql_handler.h" // mysql_ha_flush_tables
#include "sql/sql_lex.h"
#include "sql/sql_list.h"
#include "sql/sql_parse.h" // is_update_query
#include "sql/sql_prepare.h" // Reprepare_observer
#include "sql/sql_select.h" // reset_statement_timer
#include "sql/sql_show.h" // append_identifier
#include "sql/sql_sort.h"
#include "sql/sql_table.h" // build_table_filename
#include "sql/sql_update.h" // records_are_comparable
#include "sql/sql_view.h" // mysql_make_view
#include "sql/strfunc.h"
#include "sql/system_variables.h"
#include "sql/table.h" // Table_ref
#include "sql/table_cache.h" // table_cache_manager
#include "sql/table_trigger_dispatcher.h" // Table_trigger_dispatcher
#include "sql/thd_raii.h"
#include "sql/transaction.h" // trans_rollback_stmt
#include "sql/transaction_info.h"
#include "sql/trigger_chain.h" // Trigger_chain
#include "sql/xa.h"
#include "sql_string.h"
#include "template_utils.h"
#include "thr_mutex.h"
using std::equal_to;
using std::hash;
using std::pair;
using std::string;
using std::unique_ptr;
using std::unordered_map;
/**
The maximum length of a key in the table definition cache.
The key consists of the schema name, a '\0' character, the table
name and a '\0' character. Hence NAME_LEN * 2 + 1 + 1.
Additionally, the key can be suffixed with either 4 + 4 extra bytes
for slave tmp tables, or with a single extra byte for tables in a
secondary storage engine. Add 4 + 4 to account for either of these
suffixes.
*/
static constexpr const size_t MAX_DBKEY_LENGTH{NAME_LEN * 2 + 1 + 1 + 4 + 4};
static constexpr long STACK_MIN_SIZE_FOR_OPEN{1024 * 80};
/**
This internal handler is used to trap ER_NO_SUCH_TABLE and
ER_WRONG_MRG_TABLE errors during CHECK/REPAIR TABLE for MERGE
tables.
*/
class Repair_mrg_table_error_handler : public Internal_error_handler {
public:
Repair_mrg_table_error_handler()
: m_handled_errors(false), m_unhandled_errors(false) {}
bool handle_condition(THD *, uint sql_errno, const char *,
Sql_condition::enum_severity_level *,
const char *) override {
if (sql_errno == ER_NO_SUCH_TABLE || sql_errno == ER_WRONG_MRG_TABLE) {
m_handled_errors = true;
return true;
}
m_unhandled_errors = true;
return false;
}
/**
Returns true if there were ER_NO_SUCH_/WRONG_MRG_TABLE and there
were no unhandled errors. false otherwise.
*/
bool safely_trapped_errors() {
/*
Check for m_handled_errors is here for extra safety.
It can be useful in situation when call to open_table()
fails because some error which was suppressed by another
error handler (e.g. in case of MDL deadlock which we
decided to solve by back-off and retry).
*/
return (m_handled_errors && (!m_unhandled_errors));
}
private:
bool m_handled_errors;
bool m_unhandled_errors;
};
/**
@defgroup Data_Dictionary Data Dictionary
@{
*/
/**
LOCK_open protects the following variables/objects:
1) The table_def_cache
This is the hash table mapping table name to a table
share object. The hash table can only be manipulated
while holding LOCK_open.
2) last_table_id
Generation of a new unique table_map_id for a table
share is done through incrementing last_table_id, a
global variable used for this purpose.
3) LOCK_open protects the initialisation of the table share
object and all its members, however, it does not protect
reading the .frm file from where the table share is
initialised. In get_table_share, the lock is temporarily
released while opening the table definition in order to
allow a higher degree of concurrency. Concurrent access
to the same share is controlled by introducing a condition
variable for signaling when opening the share is completed.
4) In particular the share->ref_count is updated each time
a new table object is created that refers to a table share.
This update is protected by LOCK_open.
5) oldest_unused_share, end_of_unused_share and share->next
and share->prev are variables to handle the lists of table
share objects, these can only be read and manipulated while
holding the LOCK_open mutex.
6) table_def_shutdown_in_progress can be updated only while
holding LOCK_open and ALL table cache mutexes.
7) refresh_version
This variable can only be updated while holding LOCK_open AND
all table cache mutexes.
8) share->version
This variable is initialised while holding LOCK_open. It can only
be updated while holding LOCK_open AND all table cache mutexes.
So if a table share is found through a reference its version won't
change if any of those mutexes are held.
9) share->m_flush_tickets
*/
mysql_mutex_t LOCK_open;
/**
COND_open synchronizes concurrent opening of the same share:
If a thread calls get_table_share, it releases the LOCK_open
mutex while reading the definition from file. If a different
thread calls get_table_share for the same share at this point
in time, it will find the share in the TDC, but with the
m_open_in_progress flag set to true. This will make the
(second) thread wait for the COND_open condition, while the
first thread completes opening the table definition.
When the first thread is done reading the table definition,
it will set m_open_in_progress to false and broadcast the
COND_open condition. Then, all threads waiting for COND_open
will wake up and, re-search the TDC for the share, and:
1) If the share is gone, the thread will continue to allocate
and open the table definition. This happens, e.g., if the
first thread failed when opening the table definition and
had to destroy the share.
2) If the share is still in the cache, and m_open_in_progress
is still true, the thread will wait for the condition again.
This happens if a different thread finished opening a
different share.
3) If the share is still in the cache, and m_open_in_progress
has become false, the thread will check if the share is ok
(no error), increment the ref counter, and return the share.
*/
mysql_cond_t COND_open;
#ifdef HAVE_PSI_INTERFACE
static PSI_mutex_key key_LOCK_open;
static PSI_cond_key key_COND_open;
static PSI_mutex_info all_tdc_mutexes[] = {
{&key_LOCK_open, "LOCK_open", PSI_FLAG_SINGLETON, 0, PSI_DOCUMENT_ME}};
static PSI_cond_info all_tdc_conds[] = {
{&key_COND_open, "COND_open", 0, 0, PSI_DOCUMENT_ME}};
/**
Initialize performance schema instrumentation points
used by the table cache.
*/
static void init_tdc_psi_keys(void) {
const char *category = "sql";
int count;
count = static_cast<int>(array_elements(all_tdc_mutexes));
mysql_mutex_register(category, all_tdc_mutexes, count);
count = static_cast<int>(array_elements(all_tdc_conds));
mysql_cond_register(category, all_tdc_conds, count);
}
#endif /* HAVE_PSI_INTERFACE */
using Table_definition_cache =
malloc_unordered_map<std::string,
std::unique_ptr<TABLE_SHARE, Table_share_deleter>>;
Table_definition_cache *table_def_cache;
static TABLE_SHARE *oldest_unused_share, end_of_unused_share;
static bool table_def_shutdown_in_progress = false;
static bool check_and_update_table_version(THD *thd, Table_ref *tables,
TABLE_SHARE *table_share);
static bool open_table_entry_fini(THD *thd, TABLE_SHARE *share,
const dd::Table *table, TABLE *entry);
static bool auto_repair_table(THD *thd, Table_ref *table_list);
static TABLE *find_temporary_table(THD *thd, const char *table_key,
size_t table_key_length);
static bool tdc_open_view(THD *thd, Table_ref *table_list,
const char *cache_key, size_t cache_key_length);
static bool add_view_place_holder(THD *thd, Table_ref *table_list);
/**
Create a table cache/table definition cache key for a table. The
table is neither a temporary table nor a table in a secondary
storage engine.
@note
The table cache_key is created from:
db_name + \0
table_name + \0
@param[in] db_name the database name
@param[in] table_name the table name
@param[out] key buffer for the key to be created (must be of
size MAX_DBKEY_LENGTH)
@return the length of the key
*/
static size_t create_table_def_key(const char *db_name, const char *table_name,
char *key) {
/*
In theory caller should ensure that both db and table_name are
not longer than NAME_LEN bytes. In practice we play safe to avoid
buffer overruns.
*/
assert(strlen(db_name) <= NAME_LEN && strlen(table_name) <= NAME_LEN);
return strmake(strmake(key, db_name, NAME_LEN) + 1, table_name, NAME_LEN) -
key + 1;
}
/**
Create a table cache/table definition cache key for a temporary table.
The key is constructed by appending the following to the key
generated by #create_table_def_key():
- 4 bytes for master thread id
- 4 bytes pseudo thread id
@param[in] thd thread context
@param[in] db_name the database name
@param[in] table_name the table name
@param[out] key buffer for the key to be created (must be of
size MAX_DBKEY_LENGTH)
@return the length of the key
*/
static size_t create_table_def_key_tmp(const THD *thd, const char *db_name,
const char *table_name, char *key) {
size_t key_length = create_table_def_key(db_name, table_name, key);
int4store(key + key_length, thd->server_id);
int4store(key + key_length + 4, thd->variables.pseudo_thread_id);
return key_length + TMP_TABLE_KEY_EXTRA;
}
/**
Create a table cache/table definition cache key for a table in a
secondary storage engine.
The key is constructed by appending a single byte with the value 1
to the key generated by #create_table_def_key().
@param db_name the database name
@param table_name the table name
@return the key
*/
static std::string create_table_def_key_secondary(const char *db_name,
const char *table_name) {
char key[MAX_DBKEY_LENGTH];
size_t key_length = create_table_def_key(db_name, table_name, key);
// Add a single byte to distinguish the secondary table from the
// primary table. Their db name and table name are identical.
key[key_length++] = 1;
return {key, key_length};
}
/**
Get table cache key for a table list element.
@param [in] table_list Table list element.
@param [out] key On return points to table cache key for the table.
@note Unlike create_table_def_key() call this function doesn't construct
key in a buffer provider by caller. Instead it relies on the fact
that table list element for which key is requested has properly
initialized MDL_request object and the fact that table definition
cache key is suffix of key used in MDL subsystem. So to get table
definition key it simply needs to return pointer to appropriate
part of MDL_key object nested in this table list element.
Indeed, this means that lifetime of key produced by this call is
limited by the lifetime of table list element which it got as
parameter.
@return Length of key.
*/
size_t get_table_def_key(const Table_ref *table_list, const char **key) {
/*
This call relies on the fact that Table_ref::mdl_request::key object
is properly initialized, so table definition cache can be produced
from key used by MDL subsystem.
strcase is converted to strcasecmp because information_schema tables
can be accessed with lower case and upper case table names.
*/
assert(!my_strcasecmp(system_charset_info, table_list->get_db_name(),
table_list->mdl_request.key.db_name()) &&
!my_strcasecmp(system_charset_info, table_list->get_table_name(),
table_list->mdl_request.key.name()));
*key = (const char *)table_list->mdl_request.key.ptr() + 1;
return table_list->mdl_request.key.length() - 1;
}
/*****************************************************************************
Functions to handle table definition cache (TABLE_SHARE)
*****************************************************************************/
void Table_share_deleter::operator()(TABLE_SHARE *share) const {
DBUG_TRACE;
mysql_mutex_assert_owner(&LOCK_open);
if (share->prev) {
/* remove from old_unused_share list */
*share->prev = share->next;
share->next->prev = share->prev;
}
free_table_share(share);
}
bool table_def_init(void) {
#ifdef HAVE_PSI_INTERFACE
init_tdc_psi_keys();
#endif
mysql_mutex_init(key_LOCK_open, &LOCK_open, MY_MUTEX_INIT_FAST);
mysql_cond_init(key_COND_open, &COND_open);
oldest_unused_share = &end_of_unused_share;
end_of_unused_share.prev = &oldest_unused_share;
if (table_cache_manager.init()) {
mysql_cond_destroy(&COND_open);
mysql_mutex_destroy(&LOCK_open);
return true;
}
table_def_cache = new Table_definition_cache(key_memory_table_share);
return false;
}
/**
Notify table definition cache that process of shutting down server
has started so it has to keep number of TABLE and TABLE_SHARE objects
minimal in order to reduce number of references to pluggable engines.
*/
void table_def_start_shutdown(void) {
if (table_def_cache != nullptr) {
table_cache_manager.lock_all_and_tdc();
/*
Ensure that TABLE and TABLE_SHARE objects which are created for
tables that are open during process of plugins' shutdown are
immediately released. This keeps number of references to engine
plugins minimal and allows shutdown to proceed smoothly.
*/
table_def_shutdown_in_progress = true;
table_cache_manager.unlock_all_and_tdc();
/* Free all cached but unused TABLEs and TABLE_SHAREs. */
close_cached_tables(nullptr, nullptr, false, LONG_TIMEOUT);
}
}
void table_def_free(void) {
DBUG_TRACE;
if (table_def_cache != nullptr) {
/* Free table definitions. */
delete table_def_cache;
table_def_cache = nullptr;
table_cache_manager.destroy();
mysql_cond_destroy(&COND_open);
mysql_mutex_destroy(&LOCK_open);
}
}
uint cached_table_definitions(void) { return table_def_cache->size(); }
static TABLE_SHARE *process_found_table_share(THD *thd [[maybe_unused]],
TABLE_SHARE *share,
bool open_view) {
DBUG_TRACE;
mysql_mutex_assert_owner(&LOCK_open);
#if defined(ENABLED_DEBUG_SYNC)
if (!thd->is_attachable_ro_transaction_active())
DEBUG_SYNC(thd, "get_share_found_share");
#endif
/*
We found an existing table definition. Return it if we didn't get
an error when reading the table definition from file.
*/
if (share->error) {
/*
Table definition contained an error.
Note that we report ER_NO_SUCH_TABLE regardless of which error occurred
when the other thread tried to open the table definition (e.g. OOM).
*/
my_error(ER_NO_SUCH_TABLE, MYF(0), share->db.str, share->table_name.str);
return nullptr;
}
if (share->is_view && !open_view) {
my_error(ER_NO_SUCH_TABLE, MYF(0), share->db.str, share->table_name.str);
return nullptr;
}
share->increment_ref_count();
if (share->ref_count() == 1 && share->prev) {
/*
Share was not used before and it was in the old_unused_share list
Unlink share from this list
*/
DBUG_PRINT("info", ("Unlinking from not used list"));
*share->prev = share->next;
share->next->prev = share->prev;
share->next = nullptr;
share->prev = nullptr;
}
/* Free cache if too big */
while (table_def_cache->size() > table_def_size && oldest_unused_share->next)
table_def_cache->erase(to_string(oldest_unused_share->table_cache_key));
DBUG_PRINT("exit", ("share: %p ref_count: %u", share, share->ref_count()));
return share;
}
/**
Read any existing histogram statistics from the data dictionary and
store a copy of them in the TABLE_SHARE.
@param thd Thread handler
@param share The table share where to store the histograms
@param schema Schema definition
@param table_def Table definition
@retval true on error
@retval false on success
*/
static bool read_histograms(THD *thd, TABLE_SHARE *share,
const dd::Schema *schema,
const dd::Abstract_table *table_def) {
dd::cache::Dictionary_client::Auto_releaser releaser(thd->dd_client());
MDL_request_list mdl_requests;
for (const auto column : table_def->columns()) {
if (column->is_se_hidden()) continue;
MDL_key mdl_key;
dd::Column_statistics::create_mdl_key(schema->name(), table_def->name(),
column->name(), &mdl_key);
MDL_request *request = new (thd->mem_root) MDL_request;
MDL_REQUEST_INIT_BY_KEY(request, &mdl_key, MDL_SHARED_READ, MDL_STATEMENT);
mdl_requests.push_front(request);
}
if (thd->mdl_context.acquire_locks(&mdl_requests,
thd->variables.lock_wait_timeout))
return true; /* purecov: deadcode */
for (const auto column : table_def->columns()) {
if (column->is_se_hidden()) continue;
const histograms::Histogram *histogram = nullptr;
if (histograms::find_histogram(thd, schema->name().c_str(),
table_def->name().c_str(),
column->name().c_str(), &histogram)) {
// Any error is reported by the dictionary subsystem.
return true; /* purecov: deadcode */
}
if (histogram != nullptr) {
/*
Make a clone of the histogram so it survives together with the
TABLE_SHARE in case the original histogram is thrown out of the
dictionary cache.
*/
const histograms::Histogram *histogram_copy =
histogram->clone(&share->mem_root);
share->m_histograms->emplace(column->ordinal_position() - 1,
histogram_copy);
}
}
return false;
}
/** Update TABLE_SHARE with options from dd::Schema object */
static void update_schema_options(const dd::Schema *sch_obj,
TABLE_SHARE *share) {
assert(sch_obj != nullptr);
if (sch_obj != nullptr) {
if (sch_obj->read_only())
share->schema_read_only = TABLE_SHARE::Schema_read_only::RO_ON;
else
share->schema_read_only = TABLE_SHARE::Schema_read_only::RO_OFF;
}
}
/**
Get the TABLE_SHARE for a table.
Get a table definition from the table definition cache. If the share
does not exist, create a new one from the persistently stored table
definition, and temporarily release LOCK_open while retrieving it.
Re-lock LOCK_open when the table definition has been retrieved, and
broadcast this to other threads waiting for the share to become opened.
If the share exists, and is in the process of being opened, wait for
opening to complete before continuing.
@pre It is a precondition that the caller must own LOCK_open before
calling this function.
@note Callers of this function cannot rely on LOCK_open being
held for the duration of the call. It may be temporarily
released while the table definition is opened, and it may be
temporarily released while the thread is waiting for a different
thread to finish opening it.
@note After share->m_open_in_progress is set, there should be no wait
for resources like row- or metadata locks, table flushes, etc.
Otherwise, we may end up in deadlocks that will not be detected.
@param thd thread handle
@param db schema name
@param table_name table name
@param key table cache key
@param key_length length of key
@param open_view allow open of view
@param open_secondary get the share for a table in a secondary
storage engine
@return Pointer to the new TABLE_SHARE, or NULL if there was an error
*/
TABLE_SHARE *get_table_share(THD *thd, const char *db, const char *table_name,
const char *key, size_t key_length, bool open_view,
bool open_secondary) {
TABLE_SHARE *share;
bool open_table_err = false;
DBUG_TRACE;
/* Make sure we own LOCK_open */
mysql_mutex_assert_owner(&LOCK_open);
/*
To be able perform any operation on table we should own
some kind of metadata lock on it.
*/
assert(thd->mdl_context.owns_equal_or_stronger_lock(MDL_key::TABLE, db,
table_name, MDL_SHARED));
/*
Read table definition from the cache. If the share is being opened,
wait for the appropriate condition. The share may be destroyed if
open fails, so after cond_wait, we must repeat searching the
hash table.
*/
for (;;) {
auto it = table_def_cache->find(string(key, key_length));
if (it == table_def_cache->end()) {
if (thd->mdl_context.owns_equal_or_stronger_lock(
MDL_key::SCHEMA, db, "", MDL_INTENTION_EXCLUSIVE)) {
break;
}
mysql_mutex_unlock(&LOCK_open);
if (dd::mdl_lock_schema(thd, db, MDL_TRANSACTION)) {
// Lock LOCK_open again to preserve function contract
mysql_mutex_lock(&LOCK_open);
return nullptr;
}
mysql_mutex_lock(&LOCK_open);
// Need to re-try the find after getting the mutex again
continue;
}
share = it->second.get();
if (!share->m_open_in_progress)
return process_found_table_share(thd, share, open_view);
DEBUG_SYNC(thd, "get_share_before_COND_open_wait");
mysql_cond_wait(&COND_open, &LOCK_open);
}
/*
If alloc fails, the share object will not be present in the TDC, so no
thread will be waiting for m_open_in_progress. Hence, a broadcast is
not necessary.
*/
if (!(share = alloc_table_share(db, table_name, key, key_length,
open_secondary))) {
return nullptr;
}
/*
We assign a new table id under the protection of LOCK_open.
We do this instead of creating a new mutex
and using it for the sole purpose of serializing accesses to a
static variable, we assign the table id here. We assign it to the
share before inserting it into the table_def_cache to be really
sure that it cannot be read from the cache without having a table
id assigned.
CAVEAT. This means that the table cannot be used for
binlogging/replication purposes, unless get_table_share() has been
called directly or indirectly.
*/
assign_new_table_id(share);
table_def_cache->emplace(to_string(share->table_cache_key),
unique_ptr<TABLE_SHARE, Table_share_deleter>(share));
/*
We must increase ref_count prior to releasing LOCK_open
to keep the share from being deleted in tdc_remove_table()
and TABLE_SHARE::wait_for_old_version. We must also set
m_open_in_progress to indicate allocated but incomplete share.
*/
share->increment_ref_count(); // Mark in use
share->m_open_in_progress = true; // Mark being opened
/*
Temporarily release LOCK_open before opening the table definition,
which can be done without mutex protection.
*/
mysql_mutex_unlock(&LOCK_open);
#if defined(ENABLED_DEBUG_SYNC)
if (!thd->is_attachable_ro_transaction_active())
DEBUG_SYNC(thd, "get_share_before_open");
#endif
{
// We must make sure the schema is released and unlocked in the right order.
dd::cache::Dictionary_client::Auto_releaser releaser(thd->dd_client());
const dd::Schema *sch = nullptr;
const dd::Abstract_table *abstract_table = nullptr;
open_table_err = true; // Assume error to simplify code below.
if (thd->dd_client()->acquire(share->db.str, &sch) ||
thd->dd_client()->acquire(share->db.str, share->table_name.str,
&abstract_table)) {
} else if (sch == nullptr)
my_error(ER_BAD_DB_ERROR, MYF(0), share->db.str);
else if (abstract_table == nullptr)
my_error(ER_NO_SUCH_TABLE, MYF(0), share->db.str, share->table_name.str);
else if (abstract_table->type() == dd::enum_table_type::USER_VIEW ||
abstract_table->type() == dd::enum_table_type::SYSTEM_VIEW) {
if (!open_view) // We found a view but were trying to open table only.
my_error(ER_NO_SUCH_TABLE, MYF(0), share->db.str,
share->table_name.str);
else {
/*
Clone the view reference object and hold it in
TABLE_SHARE member view_object.
*/
share->is_view = true;
const dd::View *tmp_view =
dynamic_cast<const dd::View *>(abstract_table);
share->view_object = tmp_view->clone();
share->table_category =
get_table_category(share->db, share->table_name);
thd->status_var.opened_shares++;
open_table_err = false;
}
} else {
assert(abstract_table->type() == dd::enum_table_type::BASE_TABLE);
open_table_err = open_table_def(
thd, share, *dynamic_cast<const dd::Table *>(abstract_table));
/*
Update the table share with meta data from the schema object to
have it readily available to avoid performance degradation.
*/
if (!open_table_err) update_schema_options(sch, share);
/*
Read any existing histogram statistics from the data dictionary and
store a copy of them in the TABLE_SHARE.
We need to do this outside the protection of LOCK_open, since the data
dictionary might have to open tables in order to read histogram data
(such recursion will not work).
*/
if (!open_table_err && read_histograms(thd, share, sch, abstract_table))
open_table_err = true; /* purecov: deadcode */
}
}
/*
Get back LOCK_open before continuing. Notify all waiters that the
opening is finished, even if there was a failure while opening.
*/
mysql_mutex_lock(&LOCK_open);
share->m_open_in_progress = false;
mysql_cond_broadcast(&COND_open);
/*
Fake an open_table_def error in debug build, resulting in
ER_NO_SUCH_TABLE.
*/
DBUG_EXECUTE_IF("set_open_table_err", {
open_table_err = true;
my_error(ER_NO_SUCH_TABLE, MYF(0), share->db.str, share->table_name.str);
});
/*
If there was an error while opening the definition, delete the
share from the TDC, and (implicitly) destroy the share. Waiters
will detect that the share is gone, and repeat the attempt at
opening the table definition. The ref counter must be stepped
down to allow the share to be destroyed.
*/
if (open_table_err) {
share->error = true; // Allow waiters to detect the error
share->decrement_ref_count();
table_def_cache->erase(to_string(share->table_cache_key));
#if defined(ENABLED_DEBUG_SYNC)
if (!thd->is_attachable_ro_transaction_active())
DEBUG_SYNC(thd, "get_share_after_destroy");
#endif
return nullptr;
}
#ifdef HAVE_PSI_TABLE_INTERFACE
share->m_psi = PSI_TABLE_CALL(get_table_share)(
(share->tmp_table != NO_TMP_TABLE), share);
#else
share->m_psi = NULL;
#endif
DBUG_PRINT("exit", ("share: %p ref_count: %u", share, share->ref_count()));
/* If debug, assert that the share is actually present in the cache */
#ifndef NDEBUG
assert(table_def_cache->count(string(key, key_length)) != 0);
#endif
return share;
}
/**
Get a table share. If it didn't exist, try creating it from engine
For arguments and return values, see get_table_share()
*/
static TABLE_SHARE *get_table_share_with_discover(
THD *thd, Table_ref *table_list, const char *key, size_t key_length,
bool open_secondary, int *error)
{
TABLE_SHARE *share;
bool exists;
DBUG_TRACE;
share = get_table_share(thd, table_list->db, table_list->table_name, key,
key_length, true, open_secondary);
/*
If share is not NULL, we found an existing share.
If share is NULL, and there is no error, we're inside
pre-locking, which silences 'ER_NO_SUCH_TABLE' errors
with the intention to silently drop non-existing tables
from the pre-locking list. In this case we still need to try
auto-discover before returning a NULL share.
Or, we're inside SHOW CREATE VIEW, which
also installs a silencer for ER_NO_SUCH_TABLE error.
If share is NULL and the error is ER_NO_SUCH_TABLE, this is
the same as above, only that the error was not silenced by
pre-locking or SHOW CREATE VIEW.
In both these cases it won't harm to try to discover the
table.
Finally, if share is still NULL, it's a real error and we need
to abort.
@todo Rework alternative ways to deal with ER_NO_SUCH TABLE.
*/
if (share || (thd->is_error() &&
thd->get_stmt_da()->mysql_errno() != ER_NO_SUCH_TABLE)) {
return share;
}
*error = 0;
/* Table didn't exist. Check if some engine can provide it */
if (ha_check_if_table_exists(thd, table_list->db, table_list->table_name,
&exists)) {
thd->clear_error();
thd->get_stmt_da()->reset_condition_info(thd);
/* Conventionally, the storage engine API does not report errors. */
my_error(ER_OUT_OF_RESOURCES, MYF(0));
} else if (!exists) {
/*
No such table in any engine.
Hide "Table doesn't exist" errors if the table belongs to a view.
The check for thd->is_error() is necessary to not push an
unwanted error in case the error was already silenced.
@todo Rework the alternative ways to deal with ER_NO_SUCH TABLE.
*/
if (thd->is_error()) {
if (table_list->parent_l) {
thd->clear_error();
thd->get_stmt_da()->reset_condition_info(thd);
my_error(ER_WRONG_MRG_TABLE, MYF(0));
} else if (table_list->belong_to_view) {
// Mention the top view in message, to not reveal underlying views.
Table_ref *view = table_list->belong_to_view;
thd->clear_error();
thd->get_stmt_da()->reset_condition_info(thd);
my_error(ER_VIEW_INVALID, MYF(0), view->db, view->table_name);
}
}
} else {
thd->clear_error();
thd->get_stmt_da()->reset_condition_info(thd);
*error = 7; /* Run auto-discover. */
}
return nullptr;
}
/**
Mark that we are not using table share anymore.
@param share Table share
If the share has no open tables and (we have done a refresh or
if we have already too many open table shares) then delete the
definition.
*/
void release_table_share(TABLE_SHARE *share) {
DBUG_TRACE;
DBUG_PRINT("enter", ("share: %p table: %s.%s ref_count: %u version: %lu",
share, share->db.str, share->table_name.str,
share->ref_count(), share->version()));
mysql_mutex_assert_owner(&LOCK_open);
assert(share->ref_count() != 0);
if (share->decrement_ref_count() == 0) {
if (share->has_old_version() || table_def_shutdown_in_progress)
table_def_cache->erase(to_string(share->table_cache_key));
else {
/* Link share last in used_table_share list */
DBUG_PRINT("info", ("moving share to unused list"));
assert(share->next == nullptr);
share->prev = end_of_unused_share.prev;
*end_of_unused_share.prev = share;
end_of_unused_share.prev = &share->next;
share->next = &end_of_unused_share;
if (table_def_cache->size() > table_def_size) {
/* Delete the least used share to preserve LRU order. */
table_def_cache->erase(to_string(oldest_unused_share->table_cache_key));
}
}
}
}
/**
Get an existing table definition from the table definition cache.
*/
TABLE_SHARE *get_cached_table_share(const char *db, const char *table_name) {
char key[MAX_DBKEY_LENGTH];
size_t key_length;
mysql_mutex_assert_owner(&LOCK_open);
key_length = create_table_def_key(db, table_name, key);
return find_or_nullptr(*table_def_cache, string(key, key_length));
}
/*
Create a list for all open tables matching SQL expression
SYNOPSIS
list_open_tables()
thd Thread THD
wild SQL like expression
NOTES
One gets only a list of tables for which one has any kind of privilege.
db and table names are allocated in result struct, so one doesn't need
a lock on LOCK_open when traversing the return list.
RETURN VALUES
NULL Error (Probably OOM)
# Pointer to list of names of open tables.
*/
OPEN_TABLE_LIST *list_open_tables(THD *thd, const char *db, const char *wild) {
OPEN_TABLE_LIST **start_list, *open_list, *start, *prev;
Table_ref table_list;
DBUG_TRACE;
start_list = &open_list;
open_list = nullptr;
/*
This is done in two parts:
1. First, we will make OPEN_TABLE_LIST under LOCK_open
2. Second, we will check permission and unlink OPEN_TABLE_LIST
entries if permission check fails
*/
table_cache_manager.lock_all_and_tdc();
for (const auto &key_and_value : *table_def_cache) {
TABLE_SHARE *share = key_and_value.second.get();
/* Skip shares that are being opened */
if (share->m_open_in_progress) continue;
if (db && my_strcasecmp(system_charset_info, db, share->db.str)) continue;
if (wild && wild_compare(share->table_name.str, share->table_name.length,
wild, strlen(wild), false))
continue;
if (!(*start_list = (OPEN_TABLE_LIST *)(*THR_MALLOC)
->Alloc(sizeof(**start_list) +
share->table_cache_key.length))) {
open_list = nullptr; // Out of memory
break;
}
my_stpcpy((*start_list)->table =
my_stpcpy(((*start_list)->db = (char *)((*start_list) + 1)),
share->db.str) +
1,
share->table_name.str);
(*start_list)->in_use = 0;
Table_cache_iterator it(share);
while (it++) ++(*start_list)->in_use;
(*start_list)->locked = 0; /* Obsolete. */
start_list = &(*start_list)->next;
*start_list = nullptr;
}
table_cache_manager.unlock_all_and_tdc();
start = open_list;
prev = start;
while (start) {
/* Check if user has SELECT privilege for any column in the table */
table_list.db = start->db;
table_list.table_name = start->table;
table_list.grant.privilege = 0;
if (check_table_access(thd, SELECT_ACL, &table_list, true, 1, true)) {
/* Unlink OPEN_TABLE_LIST */
if (start == open_list) {
open_list = start->next;
prev = open_list;
} else
prev->next = start->next;
} else
prev = start;
start = start->next;
}
return open_list;
}
/*****************************************************************************
* Functions to free open table cache
****************************************************************************/
void intern_close_table(TABLE *table) { // Free all structures
DBUG_TRACE;
DBUG_PRINT("tcache",
("table: '%s'.'%s' %p", table->s ? table->s->db.str : "?",
table->s ? table->s->table_name.str : "?", table));
free_io_cache(table);
destroy(table->triggers);
if (table->file) // Not true if placeholder
(void)closefrm(table, true); // close file
destroy(table);
my_free(table);
}
/* Free resources allocated by filesort() and read_record() */
void free_io_cache(TABLE *table) {
DBUG_TRACE;
if (table->unique_result.io_cache) {
close_cached_file(table->unique_result.io_cache);
my_free(table->unique_result.io_cache);
table->unique_result.io_cache = nullptr;
}
}
/*
Close all tables which aren't in use by any thread
@param thd Thread context
@param tables List of tables to remove from the cache
@param wait_for_refresh Wait for a impending flush
@param timeout Timeout for waiting for flush to be completed.
@note THD can be NULL, but then wait_for_refresh must be false
and tables must be NULL.
@note When called as part of FLUSH TABLES WITH READ LOCK this function
ignores metadata locks held by other threads. In order to avoid
situation when FLUSH TABLES WITH READ LOCK sneaks in at the moment
when some write-locked table is being reopened (by FLUSH TABLES or
ALTER TABLE) we have to rely on additional global shared metadata
lock taken by thread trying to obtain global read lock.
*/
bool close_cached_tables(THD *thd, Table_ref *tables, bool wait_for_refresh,
ulong timeout) {
bool result = false;
bool found = true;
struct timespec abstime;
DBUG_TRACE;
assert(thd || (!wait_for_refresh && !tables));
table_cache_manager.lock_all_and_tdc();
if (!tables) {
/*
Force close of all open tables.
Note that code in TABLE_SHARE::wait_for_old_version() assumes that
incrementing of refresh_version and removal of unused tables and
shares from TDC happens atomically under protection of LOCK_open,
or putting it another way that TDC does not contain old shares
which don't have any tables used.
*/
refresh_version++;
DBUG_PRINT("tcache",
("incremented global refresh_version to: %lu", refresh_version));
/*
Get rid of all unused TABLE and TABLE_SHARE instances. By doing
this we automatically close all tables which were marked as "old".
*/
table_cache_manager.free_all_unused_tables();
/* Free table shares which were not freed implicitly by loop above. */
while (oldest_unused_share->next)
table_def_cache->erase(to_string(oldest_unused_share->table_cache_key));
} else {
bool share_found = false;
for (Table_ref *table = tables; table; table = table->next_local) {
TABLE_SHARE *share = get_cached_table_share(table->db, table->table_name);
if (share) {
/*
tdc_remove_table() also sets TABLE_SHARE::version to 0. Note that
it will work correctly even if m_open_in_progress flag is true.
*/
tdc_remove_table(thd, TDC_RT_REMOVE_UNUSED, table->db,
table->table_name, true);
share_found = true;
}
}
if (!share_found) wait_for_refresh = false; // Nothing to wait for
}
table_cache_manager.unlock_all_and_tdc();
if (!wait_for_refresh) return result;
set_timespec(&abstime, timeout);
if (thd->locked_tables_mode) {
/*
If we are under LOCK TABLES, we need to reopen the tables without
opening a door for any concurrent threads to sneak in and get
lock on our tables. To achieve this we use exclusive metadata
locks.
*/
Table_ref *tables_to_reopen =
(tables ? tables : thd->locked_tables_list.locked_tables());
/* Close open HANLER instances to avoid self-deadlock. */
mysql_ha_flush_tables(thd, tables_to_reopen);
for (Table_ref *table_list = tables_to_reopen; table_list;
table_list = table_list->next_global) {
/* A check that the table was locked for write is done by the caller. */
TABLE *table = find_table_for_mdl_upgrade(thd, table_list->db,
table_list->table_name, true);
/* May return NULL if this table has already been closed via an alias. */
if (!table) continue;
if (wait_while_table_is_used(thd, table, HA_EXTRA_FORCE_REOPEN)) {
result = true;
goto err_with_reopen;
}
close_all_tables_for_name(thd, table->s, false, nullptr);
}
}
/* Wait until all threads have closed all the tables we are flushing. */
DBUG_PRINT("info", ("Waiting for other threads to close their open tables"));
while (found && !thd->killed) {
TABLE_SHARE *share = nullptr;
found = false;
/*
To a self-deadlock or deadlocks with other FLUSH threads
waiting on our open HANDLERs, we have to flush them.
*/
mysql_ha_flush(thd);
DEBUG_SYNC(thd, "after_flush_unlock");
mysql_mutex_lock(&LOCK_open);
if (!tables) {
for (const auto &key_and_value : *table_def_cache) {
share = key_and_value.second.get();
if (share->has_old_version()) {
found = true;
break;
}
}
} else {
for (Table_ref *table = tables; table; table = table->next_local) {
share = get_cached_table_share(table->db, table->table_name);
if (share && share->has_old_version()) {
found = true;
break;
}
}
}
if (found) {
/*
The method below temporarily unlocks LOCK_open and frees
share's memory. Note that it works correctly even for
shares with m_open_in_progress flag set.
*/
if (share->wait_for_old_version(
thd, &abstime, MDL_wait_for_subgraph::DEADLOCK_WEIGHT_DDL)) {
mysql_mutex_unlock(&LOCK_open);
result = true;
goto err_with_reopen;
}
}
mysql_mutex_unlock(&LOCK_open);
}
err_with_reopen:
if (thd->locked_tables_mode) {
/*
No other thread has the locked tables open; reopen them and get the
old locks. This should succeed unless any dictionary operations fail
(e.g. when opening a dictionary table on cache miss).
*/
result |= thd->locked_tables_list.reopen_tables(thd);
/*
Since downgrade_lock() won't do anything with shared
metadata lock it is much simpler to go through all open tables rather
than picking only those tables that were flushed.
*/
for (TABLE *tab = thd->open_tables; tab; tab = tab->next)
tab->mdl_ticket->downgrade_lock(MDL_SHARED_NO_READ_WRITE);
}
return result || thd->killed;
}
/**
Mark all temporary tables which were used by the current statement or
substatement as free for reuse, but only if the query_id can be cleared.
@param thd thread context
@remark For temp tables associated with a open SQL HANDLER the query_id
is not reset until the HANDLER is closed.
*/
static void mark_temp_tables_as_free_for_reuse(THD *thd) {
for (TABLE *table = thd->temporary_tables; table; table = table->next) {
if ((table->query_id == thd->query_id) && !table->open_by_handler) {
mark_tmp_table_for_reuse(table);
table->cleanup_value_generator_items();
table->cleanup_partial_update();
}
}
}
/**
Reset a single temporary table.
Effectively this "closes" one temporary table,
in a session.
@param table Temporary table.
*/
void mark_tmp_table_for_reuse(TABLE *table) {
assert(table->s->tmp_table);
table->query_id = 0;
table->file->ha_reset();
/* Detach temporary MERGE children from temporary parent. */
assert(table->file);
table->file->ha_extra(HA_EXTRA_DETACH_CHILDREN);
/*
Reset temporary table lock type to it's default value (TL_WRITE).
Statements such as INSERT INTO .. SELECT FROM tmp, CREATE TABLE
.. SELECT FROM tmp and UPDATE may under some circumstances modify
the lock type of the tables participating in the statement. This
isn't a problem for non-temporary tables since their lock type is
reset at every open, but the same does not occur for temporary
tables for historical reasons.
Furthermore, the lock type of temporary tables is not really that
important because they can only be used by one query at a time and
not even twice in a query -- a temporary table is represented by
only one TABLE object. Nonetheless, it's safer from a maintenance
point of view to reset the lock type of this singleton TABLE object
as to not cause problems when the table is reused.
Even under LOCK TABLES mode its okay to reset the lock type as
LOCK TABLES is allowed (but ignored) for a temporary table.
*/
table->reginfo.lock_type = TL_WRITE;
}
/*
Mark all tables in the list which were used by current substatement
as free for reuse.
SYNOPSIS
mark_used_tables_as_free_for_reuse()
thd - thread context
table - head of the list of tables
DESCRIPTION
Marks all tables in the list which were used by current substatement
(they are marked by its query_id) as free for reuse.
NOTE
The reason we reset query_id is that it's not enough to just test
if table->query_id != thd->query_id to know if a table is in use.
For example
SELECT f1_that_uses_t1() FROM t1;
In f1_that_uses_t1() we will see one instance of t1 where query_id is
set to query_id of original query.
*/
static void mark_used_tables_as_free_for_reuse(THD *thd, TABLE *table) {
for (; table; table = table->next) {
assert(table->pos_in_locked_tables == nullptr ||
table->pos_in_locked_tables->table == table);
if (table->query_id == thd->query_id) {
table->query_id = 0;
table->file->ha_reset();
}
}
}
/**
Auxiliary function to close all tables in the open_tables list.
@param thd Thread context.
@remark It should not ordinarily be called directly.
*/
static void close_open_tables(THD *thd) {
mysql_mutex_assert_not_owner(&LOCK_open);
DBUG_PRINT("info", ("thd->open_tables: %p", thd->open_tables));
while (thd->open_tables) close_thread_table(thd, &thd->open_tables);
}
/**
Close all open instances of the table but keep the MDL lock.
Works both under LOCK TABLES and in the normal mode.
Removes all closed instances of the table from the table cache.
@param thd Thread context.
@param key TC/TDC key identifying the table.
@param key_length Length of TC/TDC key identifying the table.
@param db Database name.
@param table_name Table name.
@param remove_from_locked_tables
True if the table is being dropped.
In that case the documented behaviour is to
implicitly remove the table from LOCK TABLES list.
@param skip_table TABLE instance that should be kept open.
@pre Must be called with an X MDL lock on the table.
*/
static void close_all_tables_for_name(THD *thd, const char *key,
size_t key_length, const char *db,
const char *table_name,
bool remove_from_locked_tables,
TABLE *skip_table) {
mysql_mutex_assert_not_owner(&LOCK_open);
for (TABLE **prev = &thd->open_tables; *prev;) {
TABLE *table = *prev;
if (table->s->table_cache_key.length == key_length &&
!memcmp(table->s->table_cache_key.str, key, key_length) &&
table != skip_table) {
thd->locked_tables_list.unlink_from_list(thd, table->pos_in_locked_tables,
remove_from_locked_tables);
/*
Does nothing if the table is not locked.
This allows one to use this function after a table
has been unlocked, e.g. in partition management.
*/
mysql_lock_remove(thd, thd->lock, table);
/* Inform handler that table will be dropped after close */
if (table->db_stat && /* Not true for partitioned tables. */
skip_table == nullptr)
table->file->ha_extra(HA_EXTRA_PREPARE_FOR_DROP);
close_thread_table(thd, prev);
} else {
/* Step to next entry in open_tables list. */
prev = &table->next;
}
}
if (skip_table == nullptr) {
/* Remove the table share from the cache. */
tdc_remove_table(thd, TDC_RT_REMOVE_ALL, db, table_name, false);
}
}
void close_all_tables_for_name(THD *thd, TABLE_SHARE *share,
bool remove_from_locked_tables,
TABLE *skip_table) {
char key[MAX_DBKEY_LENGTH];
size_t key_length = share->table_cache_key.length;
memcpy(key, share->table_cache_key.str, key_length);
close_all_tables_for_name(thd, key, key_length,
key, // db
key + share->db.length + 1, // table_name
remove_from_locked_tables, skip_table);
}
void close_all_tables_for_name(THD *thd, const char *db, const char *table_name,
bool remove_from_locked_tables) {
char key[MAX_DBKEY_LENGTH];
size_t key_length = create_table_def_key(db, table_name, key);
close_all_tables_for_name(thd, key, key_length, db, table_name,
remove_from_locked_tables, nullptr);
}
// Check if we are under LOCK TABLE mode, and not prelocking.
static inline bool in_LTM(THD *thd) {
return (thd->locked_tables_mode == LTM_LOCK_TABLES ||
thd->locked_tables_mode == LTM_PRELOCKED_UNDER_LOCK_TABLES);
}
/**
Check if the given Table_ref belongs to a DD table.
The function checks whether the table is a DD table being used in the
context of a DD transaction, or whether it is referred by a system view.
Then, it implies that if either of these two conditions hold, then this
is a DD table. If in case this is a DD table being used in some other
situation, then this function does not return 'true'. We do not know if
there is such a situation right now.
This function ignores Table_ref's that is created by optimizer
when processing a system view.
@param tl Table_ref point to the table.
@retval true If table belongs to a DD table.
@retval false If table does not.
*/
static bool belongs_to_dd_table(const Table_ref *tl) {
return (tl->is_dd_ctx_table ||
(!tl->is_internal() && !tl->uses_materialization() &&
tl->referencing_view && tl->referencing_view->is_system_view));
}
/**
Close all tables used by the current substatement, or all tables
used by this thread if we are on the outer-most level.
@param thd Thread handler
@details
Unlocks all open persistent and temporary base tables.
Put all persistent base tables used by thread in free list.
It will only close/mark as free for reuse tables opened by this
substatement, it will also check if we are closing tables after
execution of complete query (i.e. we are on outer-most level) and will
leave prelocked mode if needed.
*/
void close_thread_tables(THD *thd) {
DBUG_TRACE;
#ifdef EXTRA_DEBUG
DBUG_PRINT("tcache", ("open tables:"));
for (TABLE *table = thd->open_tables; table; table = table->next)
DBUG_PRINT("tcache", ("table: '%s'.'%s' %p", table->s->db.str,
table->s->table_name.str, table));
#endif
#if defined(ENABLED_DEBUG_SYNC)
/* debug_sync may not be initialized for some slave threads */
if (thd->debug_sync_control) DEBUG_SYNC(thd, "before_close_thread_tables");
#endif
// TODO: dd::Transaction_impl::end() does merge DD transaction into
// thd->transaction.stmt. Later the can be second DD transaction
// which would call close_thread_tables(). In this case, the
// condition thd->transaction.stmt.is_empty() does not hold good.
// So we comment this assert for now.
//
// We should consider retaining this assert if we plan to commit
// DD RW transaction just before next close_thread_tables().
// We are not sure if this is doable and needs to be explored.
// Alik and myself plan to comment this assert for now temporarily
// and address this TODO asap.
//
// assert(thd->get_transaction()->is_empty(Transaction_ctx::STMT) ||
// thd->in_sub_stmt ||
// (thd->state_flags & Open_tables_state::BACKUPS_AVAIL));
/* Detach MERGE children after every statement. Even under LOCK TABLES. */
for (TABLE *table = thd->open_tables; table; table = table->next) {
/* Table might be in use by some outer statement. */
DBUG_PRINT("tcache", ("table: '%s' query_id: %lu",
table->s->table_name.str, (ulong)table->query_id));
if (thd->locked_tables_mode <= LTM_LOCK_TABLES ||
table->query_id == thd->query_id) {
assert(table->file);
if (table->db_stat) table->file->ha_extra(HA_EXTRA_DETACH_CHILDREN);
table->cleanup_value_generator_items();
table->cleanup_partial_update();
}
}
/*
Mark all temporary tables used by this statement as free for reuse.
*/
mark_temp_tables_as_free_for_reuse(thd);
if (thd->locked_tables_mode) {
/*
If we have
1) Implicitly opened some DD tables that belong to IS system
view executed in LOCK TABLE mode, then we should close them now.
2) Close P_S tables opened implicitly under LOCK TABLE mode.
*/
if (in_LTM(thd)) {
for (TABLE **prev = &thd->open_tables; *prev;) {
TABLE *table = *prev;
/* Ignore tables locked explicitly by LOCK TABLE. */
if (!table->pos_in_locked_tables) {
/*
We close tables only when all of following conditions satisfy,
- The table is not locked explicitly by user using LOCK TABLE
command.
- We are not executing a IS queries as part of SF/Trigger.
- The table belongs to a new DD table.
OR
- Close P_S tables unless the query is inside of a SP/trigger.
*/
Table_ref *tbl_list = table->pos_in_table_list;
if (!thd->in_sub_stmt && (belongs_to_dd_table(tbl_list) ||
belongs_to_p_s(table->pos_in_table_list))) {
if (!table->s->tmp_table) {
table->file->ha_index_or_rnd_end();
table->set_keyread(false);
table->open_by_handler = false;
/*
In case we have opened the DD table but the statement
fails before calling ha_external_lock() requesting
read lock in open_tables(), then we need to check
if we have really requested lock and then unlock.
*/
if (table->file->get_lock_type() != F_UNLCK)
table->file->ha_external_lock(thd, F_UNLCK);
close_thread_table(thd, prev);
continue;
}
}
}
prev = &table->next;
} // End of for
}
/* Ensure we are calling ha_reset() for all used tables */
mark_used_tables_as_free_for_reuse(thd, thd->open_tables);
/*
Mark this statement as one that has "unlocked" its tables.
For purposes of Query_tables_list::lock_tables_state we treat
any statement which passed through close_thread_tables() as
such.
*/
thd->lex->lock_tables_state = Query_tables_list::LTS_NOT_LOCKED;
/*
We are under simple LOCK TABLES or we're inside a sub-statement
of a prelocked statement, so should not do anything else.
Note that even if we are in LTM_LOCK_TABLES mode and statement
requires prelocking (e.g. when we are closing tables after
failing to "open" all tables required for statement execution)
we will exit this function a few lines below.
*/
if (!thd->lex->requires_prelocking()) return;
/*
We are in the top-level statement of a prelocked statement,
so we have to leave the prelocked mode now with doing implicit
UNLOCK TABLES if needed.
*/
if (thd->locked_tables_mode == LTM_PRELOCKED_UNDER_LOCK_TABLES)
thd->locked_tables_mode = LTM_LOCK_TABLES;
if (thd->locked_tables_mode == LTM_LOCK_TABLES) return;
thd->leave_locked_tables_mode();
/* Fallthrough */
}
if (thd->lock) {
/*
For RBR we flush the pending event just before we unlock all the
tables. This means that we are at the end of a topmost
statement, so we ensure that the STMT_END_F flag is set on the
pending event. For statements that are *inside* stored
functions, the pending event will not be flushed: that will be
handled either before writing a query log event (inside
binlog_query()) or when preparing a pending event.
*/
(void)thd->binlog_flush_pending_rows_event(true);
mysql_unlock_tables(thd, thd->lock);
thd->lock = nullptr;
}
thd->lex->lock_tables_state = Query_tables_list::LTS_NOT_LOCKED;
/*
Closing a MERGE child before the parent would be fatal if the
other thread tries to abort the MERGE lock in between.
*/
if (thd->open_tables) close_open_tables(thd);
}
/**
Helper function which returns TABLE to Table Cache or closes if
table is marked as needing re-open.
*/
static void release_or_close_table(THD *thd, TABLE *table) {
Table_cache *tc = table_cache_manager.get_cache(thd);
tc->lock();
if (table->s->has_old_version() || table->has_invalid_dict() ||
table->has_invalid_stats() || table_def_shutdown_in_progress) {
tc->remove_table(table);
mysql_mutex_lock(&LOCK_open);
intern_close_table(table);
mysql_mutex_unlock(&LOCK_open);
} else
tc->release_table(thd, table);
tc->unlock();
}
/* move one table to free list */
void close_thread_table(THD *thd, TABLE **table_ptr) {
TABLE *table = *table_ptr;
DBUG_TRACE;
assert(table->key_read == 0);
assert(!table->file || table->file->inited == handler::NONE);
mysql_mutex_assert_not_owner(&LOCK_open);
/*
The metadata lock must be released after giving back
the table to the table cache.
*/
assert(thd->mdl_context.owns_equal_or_stronger_lock(
MDL_key::TABLE, table->s->db.str, table->s->table_name.str, MDL_SHARED));
table->mdl_ticket = nullptr;
table->pos_in_table_list = nullptr;
mysql_mutex_lock(&thd->LOCK_thd_data);
*table_ptr = table->next;
mysql_mutex_unlock(&thd->LOCK_thd_data);
/*
It is not safe to call the below code for TABLE objects for which
handler::open() has not been called (for example, we use such objects
while updating information about views which depend on table being
ALTERed). Another possibly unsafe case is when TABLE/handler object
has been marked as invalid (for example, it is unsafe to call
handler::reset() for partitioned InnoDB tables after in-place ALTER
TABLE API commit phase).
*/
if (!table->has_invalid_dict()) {
/* Avoid having MERGE tables with attached children in unused_tables. */
table->file->ha_extra(HA_EXTRA_DETACH_CHILDREN);
/* Free memory and reset for next loop. */
free_blob_buffers_and_reset(table, MAX_TDC_BLOB_SIZE);
table->file->ha_reset();
}
/* Do this *before* entering the LOCK_open critical section. */
if (table->file != nullptr) table->file->unbind_psi();
release_or_close_table(thd, table);
}
/* close_temporary_tables' internal, 4 is due to uint4korr definition */
static inline uint tmpkeyval(TABLE *table) {
return uint4korr(table->s->table_cache_key.str +
table->s->table_cache_key.length - 4);
}
/*
Close all temporary tables created by 'CREATE TEMPORARY TABLE' for thread
creates one DROP TEMPORARY TABLE binlog event for each pseudo-thread.
TODO: In future, we should have temporary_table= 0 and
replica_open_temp_tables.fetch_add() at one place instead of repeating
it all across the function. An alternative would be to use
close_temporary_table() instead of close_temporary() that maintains
the correct invariant regarding empty list of temporary tables
and zero replica_open_temp_tables already.
*/
bool close_temporary_tables(THD *thd) {
DBUG_TRACE;
TABLE *table;
TABLE *next = nullptr;
TABLE *prev_table;
/* Assume thd->variables.option_bits has OPTION_QUOTE_SHOW_CREATE */
bool was_quote_show = true;
bool error = false;
int slave_closed_temp_tables = 0;
if (!thd->temporary_tables) return false;
assert(!thd->slave_thread ||
thd->system_thread != SYSTEM_THREAD_SLAVE_WORKER);
/*
Ensure we don't have open HANDLERs for tables we are about to close.
This is necessary when close_temporary_tables() is called as part
of execution of BINLOG statement (e.g. for format description event).
*/
mysql_ha_rm_temporary_tables(thd);
if (!mysql_bin_log.is_open()) {
TABLE *tmp_next;
for (TABLE *t = thd->temporary_tables; t; t = tmp_next) {
tmp_next = t->next;
mysql_lock_remove(thd, thd->lock, t);
/*
We should not meet temporary tables created by ALTER TABLE here.
It is responsibility of ALTER statement to close them. Otherwise
it might be necessary to remove them from DD as well.
*/
assert(t->s->tmp_table_def);
close_temporary(thd, t, true, true);
slave_closed_temp_tables++;
}
thd->temporary_tables = nullptr;
if (thd->slave_thread) {
atomic_replica_open_temp_tables -= slave_closed_temp_tables;
thd->rli_slave->get_c_rli()->atomic_channel_open_temp_tables -=
slave_closed_temp_tables;
}
return false;
}
/*
We are about to generate DROP TEMPORARY TABLE statements for all
the left out temporary tables. If GTID_NEXT is set (e.g. if user
did SET GTID_NEXT just before disconnecting the client), we must
ensure that it will be able to generate GTIDs for the statements
with this server's UUID. Therefore we set gtid_next to
AUTOMATIC_GTID.
*/
gtid_state->update_on_rollback(thd);
thd->variables.gtid_next.set_automatic();
/*
We must separate transactional temp tables and
non-transactional temp tables in two distinct DROP statements
to avoid the splitting if a slave server reads from this binlog.
*/
/* Better add "if exists", in case a RESET MASTER has been done */
const char stub[] = "DROP /*!40005 TEMPORARY */ TABLE IF EXISTS ";
uint stub_len = sizeof(stub) - 1;
char buf_trans[256], buf_non_trans[256];
String s_query_trans =
String(buf_trans, sizeof(buf_trans), system_charset_info);
String s_query_non_trans =
String(buf_non_trans, sizeof(buf_non_trans), system_charset_info);
bool found_user_tables = false;
bool found_trans_table = false;
bool found_non_trans_table = false;
memcpy(buf_trans, stub, stub_len);
memcpy(buf_non_trans, stub, stub_len);
/*
Insertion sort of temp tables by pseudo_thread_id to build ordered list
of sublists of equal pseudo_thread_id
*/
for (prev_table = thd->temporary_tables, table = prev_table->next; table;
prev_table = table, table = table->next) {
TABLE *prev_sorted /* same as for prev_table */, *sorted;
/*
We should not meet temporary tables created by ALTER TABLE here.
It is responsibility of ALTER statement to close them. Otherwise
it might be necessary to remove them from DD as well.
*/
assert(table->s->tmp_table_def);
if (is_user_table(table)) {
if (!found_user_tables) found_user_tables = true;
for (prev_sorted = nullptr, sorted = thd->temporary_tables;
sorted != table; prev_sorted = sorted, sorted = sorted->next) {
if (!is_user_table(sorted) || tmpkeyval(sorted) > tmpkeyval(table)) {
/* move into the sorted part of the list from the unsorted */
prev_table->next = table->next;
table->next = sorted;
if (prev_sorted) {
prev_sorted->next = table;
} else {
thd->temporary_tables = table;
}
table = prev_table;
break;
}
}
}
}
/* We always quote db,table names though it is slight overkill */
if (found_user_tables && !(was_quote_show = (thd->variables.option_bits &
OPTION_QUOTE_SHOW_CREATE))) {
thd->variables.option_bits |= OPTION_QUOTE_SHOW_CREATE;
}
/*
Make LEX consistent with DROP TEMPORARY TABLES statement which we
are going to log. This is important for the binary logging code.
*/
LEX *lex = thd->lex;
enum_sql_command sav_sql_command = lex->sql_command;
bool sav_drop_temp = lex->drop_temporary;
lex->sql_command = SQLCOM_DROP_TABLE;
lex->drop_temporary = true;
/* scan sorted tmps to generate sequence of DROP */
for (table = thd->temporary_tables; table; table = next) {
if (is_user_table(table) && table->should_binlog_drop_if_temp()) {
bool save_thread_specific_used = thd->thread_specific_used;
my_thread_id save_pseudo_thread_id = thd->variables.pseudo_thread_id;
/* Set pseudo_thread_id to be that of the processed table */
thd->variables.pseudo_thread_id = tmpkeyval(table);
String db;
db.append(table->s->db.str);
/* Loop forward through all tables that belong to a common database
within the sublist of common pseudo_thread_id to create single
DROP query
*/
for (s_query_trans.length(stub_len), s_query_non_trans.length(stub_len),
found_trans_table = false, found_non_trans_table = false;
table && is_user_table(table) &&
tmpkeyval(table) == thd->variables.pseudo_thread_id &&
table->s->db.length == db.length() &&
strcmp(table->s->db.str, db.ptr()) == 0;
table = next) {
/* Separate transactional from non-transactional temp tables */
if (table->should_binlog_drop_if_temp()) {
/* Separate transactional from non-transactional temp tables */
if (table->s->tmp_table == TRANSACTIONAL_TMP_TABLE) {
found_trans_table = true;
/*
We are going to add ` around the table names and possible more
due to special characters
*/
append_identifier(thd, &s_query_trans, table->s->table_name.str,
strlen(table->s->table_name.str));
s_query_trans.append(',');
} else if (table->s->tmp_table == NON_TRANSACTIONAL_TMP_TABLE) {
found_non_trans_table = true;
/*
We are going to add ` around the table names and possible more
due to special characters
*/
append_identifier(thd, &s_query_non_trans, table->s->table_name.str,
strlen(table->s->table_name.str));
s_query_non_trans.append(',');
}
}
next = table->next;
mysql_lock_remove(thd, thd->lock, table);
close_temporary(thd, table, true, true);
slave_closed_temp_tables++;
}
thd->clear_error();
const CHARSET_INFO *cs_save = thd->variables.character_set_client;
thd->variables.character_set_client = system_charset_info;
thd->thread_specific_used = true;
if (found_trans_table) {
Query_log_event qinfo(thd, s_query_trans.ptr(),
s_query_trans.length() - 1, false, true, false,
0);
qinfo.db = db.ptr();
qinfo.db_len = db.length();
thd->variables.character_set_client = cs_save;
thd->get_stmt_da()->set_overwrite_status(true);
if ((error = (mysql_bin_log.write_event(&qinfo) ||
mysql_bin_log.commit(thd, true) || error))) {
/*
If we're here following THD::cleanup, thence the connection
has been closed already. So lets print a message to the
error log instead of pushing yet another error into the
Diagnostics_area.
Also, we keep the error flag so that we propagate the error
up in the stack. This way, if we're the SQL thread we notice
that close_temporary_tables failed. (Actually, the SQL
thread only calls close_temporary_tables while applying old
Start_log_event_v3 events.)
*/
LogErr(ERROR_LEVEL, ER_BINLOG_FAILED_TO_WRITE_DROP_FOR_TEMP_TABLES);
}
thd->get_stmt_da()->set_overwrite_status(false);
}
if (found_non_trans_table) {
Query_log_event qinfo(thd, s_query_non_trans.ptr(),
s_query_non_trans.length() - 1, false, true,
false, 0);
qinfo.db = db.ptr();
qinfo.db_len = db.length();
thd->variables.character_set_client = cs_save;
thd->get_stmt_da()->set_overwrite_status(true);
if ((error = (mysql_bin_log.write_event(&qinfo) ||
mysql_bin_log.commit(thd, true) || error))) {
/*
If we're here following THD::cleanup, thence the connection
has been closed already. So lets print a message to the
error log instead of pushing yet another error into the
Diagnostics_area.
Also, we keep the error flag so that we propagate the error
up in the stack. This way, if we're the SQL thread we notice
that close_temporary_tables failed. (Actually, the SQL
thread only calls close_temporary_tables while applying old
Start_log_event_v3 events.)
*/
LogErr(ERROR_LEVEL, ER_BINLOG_FAILED_TO_WRITE_DROP_FOR_TEMP_TABLES);
}
thd->get_stmt_da()->set_overwrite_status(false);
}
thd->variables.pseudo_thread_id = save_pseudo_thread_id;
thd->thread_specific_used = save_thread_specific_used;
} else {
next = table->next;
/*
This is for those cases when we have acquired lock but drop temporary
table will not be logged.
*/
mysql_lock_remove(thd, thd->lock, table);
close_temporary(thd, table, true, true);
slave_closed_temp_tables++;
}
}
lex->drop_temporary = sav_drop_temp;
lex->sql_command = sav_sql_command;
if (!was_quote_show)
thd->variables.option_bits &=
~OPTION_QUOTE_SHOW_CREATE; /* restore option */
thd->temporary_tables = nullptr;
if (thd->slave_thread) {
atomic_replica_open_temp_tables -= slave_closed_temp_tables;
thd->rli_slave->get_c_rli()->atomic_channel_open_temp_tables -=
slave_closed_temp_tables;
}
return error;
}
/**
Find table in global list.
@param table Pointer to table list
@param db_name Data base name
@param table_name Table name
@returns Pointer to found table.
@retval NULL Table not found
*/
Table_ref *find_table_in_global_list(Table_ref *table, const char *db_name,
const char *table_name) {
for (; table; table = table->next_global) {
if ((table->table == nullptr ||
table->table->s->tmp_table == NO_TMP_TABLE) &&
strcmp(table->db, db_name) == 0 &&
strcmp(table->table_name, table_name) == 0)
break;
}
return table;
}
/**
Test that table is unique (It's only exists once in the table list)
@param table table to be checked (must be updatable base table)
@param table_list list of tables
@param check_alias whether to check tables' aliases
NOTE: to exclude derived tables from check we use following mechanism:
a) during derived table processing set THD::derived_tables_processing
b) Query_block::prepare set SELECT::exclude_from_table_unique_test if
THD::derived_tables_processing set. (we can't use JOIN::execute
because for PS we perform only Query_block::prepare, but we can't set
this flag in Query_block::prepare if we are not sure that we are in
derived table processing loop, because multi-update call fix_fields()
for some its items (which mean Query_block::prepare for subqueries)
before unique_table call to detect which tables should be locked for
write).
c) find_dup_table skip all tables which belong to SELECT with
SELECT::exclude_from_table_unique_test set.
Also SELECT::exclude_from_table_unique_test used to exclude from check
tables of main SELECT of multi-delete and multi-update
We also skip tables with Table_ref::prelocking_placeholder set,
because we want to allow SELECTs from them, and their modification
will rise the error anyway.
TODO: when we will have table/view change detection we can do this check
only once for PS/SP
@retval !=0 found duplicate
@retval 0 if table is unique
*/
static Table_ref *find_dup_table(const Table_ref *table, Table_ref *table_list,
bool check_alias) {
Table_ref *res;
const char *d_name, *t_name, *t_alias;
DBUG_TRACE;
DBUG_PRINT("enter", ("table alias: %s", table->alias));
assert(table == table->updatable_base_table());
/*
If this function called for CREATE command that we have not opened table
(table->table equal to 0) and right names is in current Table_ref
object.
*/
if (table->table) {
/* All MyISAMMRG children are plain MyISAM tables. */
assert(table->table->file->ht->db_type != DB_TYPE_MRG_MYISAM);
/* temporary table is always unique */
if (table->table->s->tmp_table != NO_TMP_TABLE) return nullptr;
}
d_name = table->db;
t_name = table->table_name;
t_alias = table->alias;
DBUG_PRINT("info", ("real table: %s.%s", d_name, t_name));
for (;;) {
/*
Table is unique if it is present only once in the global list
of tables and once in the list of table locks.
*/
if (!(res = find_table_in_global_list(table_list, d_name, t_name))) break;
/* Skip if same underlying table. */
if (res->table && (res->table == table->table)) goto next;
/* Skip if table alias does not match. */
if (check_alias) {
if (lower_case_table_names
? my_strcasecmp(files_charset_info, t_alias, res->alias)
: strcmp(t_alias, res->alias))
goto next;
}
/*
Skip if marked to be excluded (could be a derived table) or if
entry is a prelocking placeholder.
*/
if (res->query_block && !res->query_block->exclude_from_table_unique_test &&
!res->prelocking_placeholder)
break;
/*
If we found entry of this table or table of SELECT which already
processed in derived table or top select of multi-update/multi-delete
(exclude_from_table_unique_test) or prelocking placeholder.
*/
next:
table_list = res->next_global;
DBUG_PRINT("info",
("found same copy of table or table which we should skip"));
}
return res;
}
/**
Test that the subject table of INSERT/UPDATE/DELETE/CREATE
or (in case of MyISAMMRG) one of its children are not used later
in the query.
For MyISAMMRG tables, it is assumed that all the underlying
tables of @c table (if any) are listed right after it and that
their @c parent_l field points at the main table.
@param table table to be checked (must be updatable base table)
@param table_list List of tables to check against
@param check_alias whether to check tables' aliases
@retval non-NULL The table list element for the table that
represents the duplicate.
@retval NULL No duplicates found.
*/
Table_ref *unique_table(const Table_ref *table, Table_ref *table_list,
bool check_alias) {
assert(table == table->updatable_base_table());
Table_ref *dup;
if (table->table && table->table->file->ht->db_type == DB_TYPE_MRG_MYISAM) {
Table_ref *child;
dup = nullptr;
/* Check duplicates of all merge children. */
for (child = table->next_global; child && child->parent_l == table;
child = child->next_global) {
if ((dup = find_dup_table(child, child->next_global, check_alias))) break;
}
} else
dup = find_dup_table(table, table_list, check_alias);
return dup;
}
/**
Issue correct error message in case we found 2 duplicate tables which
prevent some update operation
@param update table which we try to update
@param operation name of update operation
@param duplicate duplicate table which we found
@note here we hide view underlying tables if we have them.
*/
void update_non_unique_table_error(Table_ref *update, const char *operation,
Table_ref *duplicate) {
update = update->top_table();
duplicate = duplicate->top_table();
if (!update->is_view() || !duplicate->is_view() ||
update->view_query() == duplicate->view_query() ||
update->table_name_length != duplicate->table_name_length ||
update->db_length != duplicate->db_length ||
my_strcasecmp(table_alias_charset, update->table_name,
duplicate->table_name) != 0 ||
my_strcasecmp(table_alias_charset, update->db, duplicate->db) != 0) {
/*
it is not the same view repeated (but it can be parts of the same copy
of view), so we have to hide underlying tables.
*/
if (update->is_view()) {
// Issue the ER_NON_INSERTABLE_TABLE error for an INSERT
if (duplicate->is_view() &&
update->view_query() == duplicate->view_query())
my_error(!strncmp(operation, "INSERT", 6) ? ER_NON_INSERTABLE_TABLE
: ER_NON_UPDATABLE_TABLE,
MYF(0), update->alias, operation);
else
my_error(ER_VIEW_PREVENT_UPDATE, MYF(0),
(duplicate->is_view() ? duplicate->alias : update->alias),
operation, update->alias);
return;
}
if (duplicate->is_view()) {
my_error(ER_VIEW_PREVENT_UPDATE, MYF(0), duplicate->alias, operation,
update->alias);
return;
}
}
my_error(ER_UPDATE_TABLE_USED, MYF(0), update->alias);
}
/**
Find temporary table specified by database and table names in the
THD::temporary_tables list.
@return TABLE instance if a temporary table has been found; NULL otherwise.
*/
TABLE *find_temporary_table(THD *thd, const char *db, const char *table_name) {
char key[MAX_DBKEY_LENGTH];
size_t key_length = create_table_def_key_tmp(thd, db, table_name, key);
return find_temporary_table(thd, key, key_length);
}
/**
Find a temporary table specified by Table_ref instance in the
THD::temporary_tables list.
@return TABLE instance if a temporary table has been found; NULL otherwise.
*/
TABLE *find_temporary_table(THD *thd, const Table_ref *tl) {
const char *key;
size_t key_length;
char key_suffix[TMP_TABLE_KEY_EXTRA];
TABLE *table;
key_length = get_table_def_key(tl, &key);
int4store(key_suffix, thd->server_id);
int4store(key_suffix + 4, thd->variables.pseudo_thread_id);
for (table = thd->temporary_tables; table; table = table->next) {
if ((table->s->table_cache_key.length ==
key_length + TMP_TABLE_KEY_EXTRA) &&
!memcmp(table->s->table_cache_key.str, key, key_length) &&
!memcmp(table->s->table_cache_key.str + key_length, key_suffix,
TMP_TABLE_KEY_EXTRA))
return table;
}
return nullptr;
}
/**
Find a temporary table specified by a key in the THD::temporary_tables list.
@return TABLE instance if a temporary table has been found; NULL otherwise.
*/
static TABLE *find_temporary_table(THD *thd, const char *table_key,
size_t table_key_length) {
for (TABLE *table = thd->temporary_tables; table; table = table->next) {
if (table->s->table_cache_key.length == table_key_length &&
!memcmp(table->s->table_cache_key.str, table_key, table_key_length)) {
return table;
}
}
return nullptr;
}
/**
Drop a temporary table.
- If the table is locked with LOCK TABLES or by prelocking,
unlock it and remove it from the list of locked tables
(THD::lock). Currently only transactional temporary tables
are locked.
- Close the temporary table.
- Remove the table from the list of temporary tables.
*/
void drop_temporary_table(THD *thd, Table_ref *table_list) {
DBUG_TRACE;
DBUG_PRINT("tmptable", ("closing table: '%s'.'%s'", table_list->db,
table_list->table_name));
assert(is_temporary_table(table_list));
TABLE *table = table_list->table;
assert(!table->query_id || table->query_id == thd->query_id);
/*
If LOCK TABLES list is not empty and contains this table,
unlock the table and remove the table from this list.
*/
mysql_lock_remove(thd, thd->lock, table);
close_temporary_table(thd, table, true, true);
table_list->table = nullptr;
}
/*
unlink from thd->temporary tables and close temporary table
*/
void close_temporary_table(THD *thd, TABLE *table, bool free_share,
bool delete_table) {
DBUG_TRACE;
DBUG_PRINT("tmptable",
("closing table: '%s'.'%s' %p alias: '%s'", table->s->db.str,
table->s->table_name.str, table, table->alias));
if (table->prev) {
table->prev->next = table->next;
if (table->prev->next) table->next->prev = table->prev;
} else {
/* removing the item from the list */
assert(table == thd->temporary_tables);
/*
slave must reset its temporary list pointer to zero to exclude
passing non-zero value to end_slave via rli->save_temporary_tables
when no temp tables opened, see an invariant below.
*/
thd->temporary_tables = table->next;
if (thd->temporary_tables) table->next->prev = nullptr;
}
if (thd->slave_thread) {
/* natural invariant of temporary_tables */
assert(thd->rli_slave->get_c_rli()->atomic_channel_open_temp_tables ||
!thd->temporary_tables);
--atomic_replica_open_temp_tables;
--thd->rli_slave->get_c_rli()->atomic_channel_open_temp_tables;
}
close_temporary(thd, table, free_share, delete_table);
}
/*
Close and delete a temporary table
NOTE
This doesn't unlink table from thd->temporary
If this is needed, use close_temporary_table()
*/
void close_temporary(THD *thd, TABLE *table, bool free_share,
bool delete_table) {
handlerton *table_type = table->s->db_type();
DBUG_TRACE;
DBUG_PRINT("tmptable", ("closing table: '%s'.'%s'", table->s->db.str,
table->s->table_name.str));
free_io_cache(table);
closefrm(table, false);
if (delete_table) {
assert(thd);
rm_temporary_table(thd, table_type, table->s->path.str,
table->s->tmp_table_def);
}
if (free_share) {
free_table_share(table->s);
destroy(table);
my_free(table);
}
}
/*
Used by ALTER TABLE when the table is a temporary one. It changes something
only if the ALTER contained a RENAME clause (otherwise, table_name is the old
name).
Prepares a table cache key, which is the concatenation of db, table_name and
thd->slave_proxy_id, separated by '\0'.
*/
bool rename_temporary_table(THD *thd, TABLE *table, const char *db,
const char *table_name) {
char *key;
size_t key_length;
TABLE_SHARE *share = table->s;
DBUG_TRACE;
if (!(key = (char *)share->mem_root.Alloc(MAX_DBKEY_LENGTH)))
return true; /* purecov: inspected */
key_length = create_table_def_key_tmp(thd, db, table_name, key);
share->set_table_cache_key(key, key_length);
/* Also update table name in DD object. Database name is kept reset. */
share->tmp_table_def->set_name(table_name);
return false;
}
/**
Force all other threads to stop using the table by upgrading
metadata lock on it and remove unused TABLE instances from cache.
@param thd Thread handler
@param table Table to remove from cache
@param function HA_EXTRA_PREPARE_FOR_DROP if table is to be deleted
HA_EXTRA_FORCE_REOPEN if table is not be used
HA_EXTRA_PREPARE_FOR_RENAME if table is to be renamed
@note When returning, the table will be unusable for other threads
until metadata lock is downgraded.
@retval false Success.
@retval true Failure (e.g. because thread was killed).
*/
bool wait_while_table_is_used(THD *thd, TABLE *table,
enum ha_extra_function function) {
DBUG_TRACE;
DBUG_PRINT("enter", ("table: '%s' share: %p db_stat: %u version: %lu",
table->s->table_name.str, table->s, table->db_stat,
table->s->version()));
if (thd->mdl_context.upgrade_shared_lock(table->mdl_ticket, MDL_EXCLUSIVE,
thd->variables.lock_wait_timeout))
return true;
tdc_remove_table(thd, TDC_RT_REMOVE_NOT_OWN, table->s->db.str,
table->s->table_name.str, false);
/* extra() call must come only after all instances above are closed */
(void)table->file->ha_extra(function);
return false;
}
/**
Check that table exists in data-dictionary or in some storage engine.
@param thd Thread context
@param table Table list element
@param[out] exists Out parameter which is set to true if table
exists and to false otherwise.
@note If there is no table in data-dictionary but it exists in one
of engines (e.g. it was created on another node of NDB cluster)
this function will fetch and add proper table description to
the data-dictionary.
@retval true Some error occurred
@retval false No error. 'exists' out parameter set accordingly.
*/
static bool check_if_table_exists(THD *thd, Table_ref *table, bool *exists) {
DBUG_TRACE;
*exists = true;
assert(thd->mdl_context.owns_equal_or_stronger_lock(
MDL_key::TABLE, table->db, table->table_name, MDL_SHARED));
if (dd::table_exists(thd->dd_client(), table->db, table->table_name, exists))
return true; // Error is already reported.
if (*exists) goto end;
/* Table doesn't exist. Check if some engine can provide it. */
if (ha_check_if_table_exists(thd, table->db, table->table_name, exists)) {
my_printf_error(ER_OUT_OF_RESOURCES,
"Failed to open '%-.64s', error while "
"unpacking from engine",
MYF(0), table->table_name);
return true;
}
end:
return false;
}
/**
An error handler which converts, if possible, ER_LOCK_DEADLOCK error
that can occur when we are trying to acquire a metadata lock to
a request for back-off and re-start of open_tables() process.
*/
class MDL_deadlock_handler : public Internal_error_handler {
public:
MDL_deadlock_handler(Open_table_context *ot_ctx_arg)
: m_ot_ctx(ot_ctx_arg), m_is_active(false) {}
bool handle_condition(THD *, uint sql_errno, const char *,
Sql_condition::enum_severity_level *,
const char *) override {
if (!m_is_active && sql_errno == ER_LOCK_DEADLOCK) {
/* Disable the handler to avoid infinite recursion. */
m_is_active = true;
(void)m_ot_ctx->request_backoff_action(
Open_table_context::OT_BACKOFF_AND_RETRY, nullptr);
m_is_active = false;
/*
If the above back-off request failed, a new instance of
ER_LOCK_DEADLOCK error was emitted. Thus the current
instance of error condition can be treated as handled.
*/
return true;
}
return false;
}
private:
/** Open table context to be used for back-off request. */
Open_table_context *m_ot_ctx;
/**
Indicates that we are already in the process of handling
ER_LOCK_DEADLOCK error. Allows to re-emit the error from
the error handler without falling into infinite recursion.
*/
bool m_is_active;
};
/**
Try to acquire an MDL lock for a table being opened.
@param[in,out] thd Session context, to report errors.
@param[out] ot_ctx Open table context, to hold the back off
state. If we failed to acquire a lock
due to a lock conflict, we add the
failed request to the open table context.
@param[in,out] table_list Table list element for the table being opened.
Its "mdl_request" member specifies the MDL lock
to be requested. If we managed to acquire a
ticket (no errors or lock conflicts occurred),
Table_ref::mdl_request contains a reference
to it on return. However, is not modified if
MDL lock type- modifying flags were provided.
We also use Table_ref::lock_type member to
detect cases when MDL_SHARED_WRITE_LOW_PRIO
lock should be acquired instead of the normal
MDL_SHARED_WRITE lock.
@param[in] flags flags MYSQL_OPEN_FORCE_SHARED_MDL,
MYSQL_OPEN_FORCE_SHARED_HIGH_PRIO_MDL or
MYSQL_OPEN_FAIL_ON_MDL_CONFLICT
@sa open_table().
@param[out] mdl_ticket Only modified if there was no error.
If we managed to acquire an MDL
lock, contains a reference to the
ticket, otherwise is set to NULL.
@retval true An error occurred.
@retval false No error, but perhaps a lock conflict, check mdl_ticket.
*/
static bool open_table_get_mdl_lock(THD *thd, Open_table_context *ot_ctx,
Table_ref *table_list, uint flags,
MDL_ticket **mdl_ticket) {
MDL_request *mdl_request = &table_list->mdl_request;
MDL_request new_mdl_request;
if (flags &
(MYSQL_OPEN_FORCE_SHARED_MDL | MYSQL_OPEN_FORCE_SHARED_HIGH_PRIO_MDL)) {
/*
MYSQL_OPEN_FORCE_SHARED_MDL flag means that we are executing
PREPARE for a prepared statement and want to override
the type-of-operation aware metadata lock which was set
in the parser/during view opening with a simple shared
metadata lock.
This is necessary to allow concurrent execution of PREPARE
and LOCK TABLES WRITE statement against the same table.
MYSQL_OPEN_FORCE_SHARED_HIGH_PRIO_MDL flag means that we open
the table in order to get information about it for one of I_S
queries and also want to override the type-of-operation aware
shared metadata lock which was set earlier (e.g. during view
opening) with a high-priority shared metadata lock.
This is necessary to avoid unnecessary waiting and extra
ER_WARN_I_S_SKIPPED_TABLE warnings when accessing I_S tables.
These two flags are mutually exclusive.
*/
assert(!(flags & MYSQL_OPEN_FORCE_SHARED_MDL) ||
!(flags & MYSQL_OPEN_FORCE_SHARED_HIGH_PRIO_MDL));
MDL_REQUEST_INIT_BY_KEY(&new_mdl_request, &mdl_request->key,
(flags & MYSQL_OPEN_FORCE_SHARED_MDL)
? MDL_SHARED
: MDL_SHARED_HIGH_PRIO,
MDL_TRANSACTION);
mdl_request = &new_mdl_request;
} else if (thd->variables.low_priority_updates &&
mdl_request->type == MDL_SHARED_WRITE &&
(table_list->lock_descriptor().type == TL_WRITE_DEFAULT ||
table_list->lock_descriptor().type ==
TL_WRITE_CONCURRENT_DEFAULT)) {
/*
We are in @@low_priority_updates=1 mode and are going to acquire
SW metadata lock on a table which for which neither LOW_PRIORITY nor
HIGH_PRIORITY clauses were used explicitly.
To keep compatibility with THR_LOCK locks and to avoid starving out
concurrent LOCK TABLES READ statements, we need to acquire the low-prio
version of SW lock instead of a normal SW lock in this case.
*/
MDL_REQUEST_INIT_BY_KEY(&new_mdl_request, &mdl_request->key,
MDL_SHARED_WRITE_LOW_PRIO, MDL_TRANSACTION);
mdl_request = &new_mdl_request;
}
if (flags & MYSQL_OPEN_FAIL_ON_MDL_CONFLICT) {
/*
When table is being open in order to get data for I_S table,
we might have some tables not only open but also locked (e.g. when
this happens under LOCK TABLES or in a stored function).
As a result by waiting on a conflicting metadata lock to go away
we may create a deadlock which won't entirely belong to the
MDL subsystem and thus won't be detectable by this subsystem's
deadlock detector.
To avoid such situation we skip the trouble-making table if
there is a conflicting lock.
*/
if (thd->mdl_context.try_acquire_lock(mdl_request)) return true;
if (mdl_request->ticket == nullptr) {
my_error(ER_WARN_I_S_SKIPPED_TABLE, MYF(0), mdl_request->key.db_name(),
mdl_request->key.name());
return true;
}
} else {
/*
We are doing a normal table open. Let us try to acquire a metadata
lock on the table. If there is a conflicting lock, acquire_lock()
will wait for it to go away. Sometimes this waiting may lead to a
deadlock, with the following results:
1) If a deadlock is entirely within MDL subsystem, it is
detected by the deadlock detector of this subsystem.
ER_LOCK_DEADLOCK error is produced. Then, the error handler
that is installed prior to the call to acquire_lock() attempts
to request a back-off and retry. Upon success, ER_LOCK_DEADLOCK
error is suppressed, otherwise propagated up the calling stack.
2) Otherwise, a deadlock may occur when the wait-for graph
includes edges not visible to the MDL deadlock detector.
One such example is a wait on an InnoDB row lock, e.g. when:
conn C1 gets SR MDL lock on t1 with SELECT * FROM t1
conn C2 gets a row lock on t2 with SELECT * FROM t2 FOR UPDATE
conn C3 gets in and waits on C1 with DROP TABLE t0, t1
conn C2 continues and blocks on C3 with SELECT * FROM t0
conn C1 deadlocks by waiting on C2 by issuing SELECT * FROM
t2 LOCK IN SHARE MODE.
Such circular waits are currently only resolved by timeouts,
e.g. @@innodb_lock_wait_timeout or @@lock_wait_timeout.
Note that we want to force DML deadlock weight for our context
when acquiring locks in this place. This is done to avoid situation
when LOCK TABLES statement, which acquires strong SNRW and SRO locks
on implicitly used tables, deadlocks with a concurrent DDL statement
and the DDL statement is aborted since it is chosen as a deadlock
victim. It is better to choose LOCK TABLES as a victim in this case
as a deadlock can be easily caught here and handled by back-off and retry,
without reporting any error to the user.
We still have a few weird cases, like FLUSH TABLES <table-list> WITH
READ LOCK, where we use "strong" metadata locks and open_tables() is
called with some metadata locks pre-acquired. In these cases we still
want to use DDL deadlock weight as back-off is not possible.
*/
MDL_deadlock_handler mdl_deadlock_handler(ot_ctx);
thd->push_internal_handler(&mdl_deadlock_handler);
thd->mdl_context.set_force_dml_deadlock_weight(ot_ctx->can_back_off());
bool result =
thd->mdl_context.acquire_lock(mdl_request, ot_ctx->get_timeout());
thd->mdl_context.set_force_dml_deadlock_weight(false);
thd->pop_internal_handler();
if (result && !ot_ctx->can_recover_from_failed_open()) return true;
}
*mdl_ticket = mdl_request->ticket;
return false;
}
/**
Check if table's share is being removed from the table definition
cache and, if yes, wait until the flush is complete.
@param thd Thread context.
@param db Database name.
@param table_name Table name.
@param wait_timeout Timeout for waiting.
@param deadlock_weight Weight of this wait for deadlock detector.
@retval false Success. Share is up to date or has been flushed.
@retval true Error (OOM, our was killed, the wait resulted
in a deadlock or timeout). Reported.
*/
static bool tdc_wait_for_old_version(THD *thd, const char *db,
const char *table_name, ulong wait_timeout,
uint deadlock_weight) {
TABLE_SHARE *share;
bool res = false;
mysql_mutex_lock(&LOCK_open);
if ((share = get_cached_table_share(db, table_name)) &&
share->has_old_version()) {
struct timespec abstime;
set_timespec(&abstime, wait_timeout);
res = share->wait_for_old_version(thd, &abstime, deadlock_weight);
}
mysql_mutex_unlock(&LOCK_open);
return res;
}
/**
Add a dummy LEX object for a view.
@param thd Thread context
@param table_list The list of tables in the view
@retval true error occurred
@retval false view place holder successfully added
*/
bool add_view_place_holder(THD *thd, Table_ref *table_list) {
Prepared_stmt_arena_holder ps_arena_holder(thd);
LEX *lex_obj = new (thd->mem_root) st_lex_local;
if (lex_obj == nullptr) return true;
table_list->set_view_query(lex_obj);
// Create empty list of view_tables.
table_list->view_tables =
new (thd->mem_root) mem_root_deque<Table_ref *>(thd->mem_root);
if (table_list->view_tables == nullptr) return true;
return false;
}
/**
Open a base table.
@param thd Thread context.
@param table_list Open first table in list.
@param ot_ctx Context with flags which modify how open works
and which is used to recover from a failed
open_table() attempt.
Some examples of flags:
MYSQL_OPEN_IGNORE_FLUSH - Open table even if
someone has done a flush. No version number
checking is done.
MYSQL_OPEN_HAS_MDL_LOCK - instead of acquiring
metadata locks rely on that caller already has
appropriate ones.
Uses a cache of open tables to find a TABLE instance not in use.
If Table_ref::open_strategy is set to OPEN_IF_EXISTS, the table is
opened only if it exists. If the open strategy is OPEN_STUB, the
underlying table is never opened. In both cases, metadata locks are
always taken according to the lock strategy.
@retval true Open failed. "action" parameter may contain type of action
needed to remedy problem before retrying again.
@retval false Success. Members of Table_ref structure are filled
properly (e.g. Table_ref::table is set for real tables and
Table_ref::view is set for views).
*/
bool open_table(THD *thd, Table_ref *table_list, Open_table_context *ot_ctx) {
TABLE *table = nullptr;
TABLE_SHARE *share = nullptr;
const char *key;
size_t key_length;
const char *alias = table_list->alias;
uint flags = ot_ctx->get_flags();
MDL_ticket *mdl_ticket = nullptr;
int error = 0;
DBUG_TRACE;
// Temporary tables and derived tables are not allowed:
assert(!is_temporary_table(table_list) && !table_list->is_derived());
/*
The table must not be opened already. The table can be pre-opened for
some statements if it is a temporary table.
open_temporary_table() must be used to open temporary tables.
A derived table cannot be opened with this.
*/
assert(table_list->is_view() || table_list->table == nullptr);
/* an open table operation needs a lot of the stack space */
if (check_stack_overrun(thd, STACK_MIN_SIZE_FOR_OPEN, (uchar *)&alias))
return true;
// New DD- In current_thd->is_strict_mode() mode we call open_table
// on new DD tables like mysql.tables/* when CREATE fails and we
// try to abort the operation and invoke quick_rm_table().
// Currently, we ignore deleting table in strict mode. Need to fix this.
// TODO.
DBUG_EXECUTE_IF("kill_query_on_open_table_from_tz_find", {
/*
When on calling my_tz_find the following
tables are opened in specified order: time_zone_name,
time_zone, time_zone_transition_type,
time_zone_transition. Emulate killing a query
on opening the second table in the list.
*/
if (!strcmp("time_zone", table_list->table_name))
thd->killed = THD::KILL_QUERY;
});
if (!(flags & MYSQL_OPEN_IGNORE_KILLED) && thd->killed) return true;
/*
Check if we're trying to take a write lock in a read only transaction.
Note that we allow write locks on log tables as otherwise logging
to general/slow log would be disabled in read only transactions.
*/
if (table_list->mdl_request.is_write_lock_request() && thd->tx_read_only &&
!(flags & (MYSQL_LOCK_LOG_TABLE | MYSQL_OPEN_HAS_MDL_LOCK))) {
my_error(ER_CANT_EXECUTE_IN_READ_ONLY_TRANSACTION, MYF(0));
return true;
}
/*
FLUSH TABLES is ignored for DD, I_S and P_S tables/views.
Hence setting MYSQL_OPEN_IGNORE_FLUSH flag.
*/
if (table_list->is_system_view || belongs_to_dd_table(table_list) ||
belongs_to_p_s(table_list))
flags |= MYSQL_OPEN_IGNORE_FLUSH;
key_length = get_table_def_key(table_list, &key);
// If a table in a secondary storage engine has been requested,
// adjust the key to refer to the secondary table.
std::string secondary_key;
if ((flags & MYSQL_OPEN_SECONDARY_ENGINE) != 0) {
secondary_key = create_table_def_key_secondary(
table_list->get_db_name(), table_list->get_table_name());
key = secondary_key.data();
key_length = secondary_key.length();
}
/*
If we're in pre-locked or LOCK TABLES mode, let's try to find the
requested table in the list of pre-opened and locked tables. If the
table is not there, return an error - we can't open not pre-opened
tables in pre-locked/LOCK TABLES mode.
There is a special case where we allow opening not pre-opened tables
in LOCK TABLES mode for new DD tables. The reason is as following.
With new DD, IS system views need to be accessible in LOCK TABLE
mode without user explicitly calling LOCK TABLE on IS view or its
underlying DD tables. This is required to keep the old behavior the
MySQL server had without new DD.
In case user executes IS system view under LOCK TABLE mode
(LTM and not prelocking), then MySQL server implicitly opens system
view and related DD tables. Such DD tables are then implicitly closed
upon end of statement execution.
Our goal is to hide DD tables from users, so there is no possibility of
explicit locking DD table using LOCK TABLE. In case user does LOCK TABLE
on IS system view explicitly, MySQL server throws a error.
TODO: move this block into a separate function.
*/
if (thd->locked_tables_mode && !(flags & MYSQL_OPEN_GET_NEW_TABLE) &&
!(in_LTM(thd) &&
(table_list->is_system_view || belongs_to_dd_table(table_list) ||
belongs_to_p_s(table_list)))) { // Using table locks
TABLE *best_table = nullptr;
int best_distance = INT_MIN;
for (table = thd->open_tables; table; table = table->next) {
if (table->s->table_cache_key.length == key_length &&
!memcmp(table->s->table_cache_key.str, key, key_length)) {
if (!my_strcasecmp(system_charset_info, table->alias, alias) &&
table->query_id != thd->query_id && /* skip tables already used */
(thd->locked_tables_mode == LTM_LOCK_TABLES ||
table->query_id == 0)) {
int distance = ((int)table->reginfo.lock_type -
(int)table_list->lock_descriptor().type);
/*
Find a table that either has the exact lock type requested,
or has the best suitable lock. In case there is no locked
table that has an equal or higher lock than requested,
we us the closest matching lock to be able to produce an error
message about wrong lock mode on the table. The best_table
is changed if bd < 0 <= d or bd < d < 0 or 0 <= d < bd.
distance < 0 - No suitable lock found
distance > 0 - we have lock mode higher then we require
distance == 0 - we have lock mode exactly which we need
*/
if ((best_distance < 0 && distance > best_distance) ||
(distance >= 0 && distance < best_distance)) {
best_distance = distance;
best_table = table;
if (best_distance == 0) {
/*
We have found a perfect match and can finish iterating
through open tables list. Check for table use conflict
between calling statement and SP/trigger is done in
lock_tables().
*/
break;
}
}
}
}
}
if (best_table) {
table = best_table;
table->query_id = thd->query_id;
DBUG_PRINT("info", ("Using locked table"));
goto reset;
}
/*
Is this table a view and not a base table?
(it is work around to allow to open view with locked tables,
real fix will be made after definition cache will be made)
Since opening of view which was not explicitly locked by LOCK
TABLES breaks metadata locking protocol (potentially can lead
to deadlocks) it should be disallowed.
*/
if (thd->mdl_context.owns_equal_or_stronger_lock(
MDL_key::TABLE, table_list->db, table_list->table_name,
MDL_SHARED)) {
/*
Note that we can't be 100% sure that it is a view since it's
possible that we either simply have not found unused TABLE
instance in THD::open_tables list or were unable to open table
during prelocking process (in this case in theory we still
should hold shared metadata lock on it).
*/
dd::cache::Dictionary_client::Auto_releaser releaser(thd->dd_client());
const dd::View *view = nullptr;
if (!thd->dd_client()->acquire(table_list->db, table_list->table_name,
&view) &&
view != nullptr) {
/*
If parent_l of the table_list is non null then a merge table
has this view as child table, which is not supported.
*/
if (table_list->parent_l) {
my_error(ER_WRONG_MRG_TABLE, MYF(0));
return true;
}
/*
In the case of a CREATE, add a dummy LEX object to
indicate the presence of a view amd skip processing the
existing view.
*/
if (table_list->open_strategy == Table_ref::OPEN_FOR_CREATE)
return add_view_place_holder(thd, table_list);
if (!tdc_open_view(thd, table_list, key, key_length)) {
assert(table_list->is_view());
return false; // VIEW
}
}
}
/*
No table in the locked tables list. In case of explicit LOCK TABLES
this can happen if a user did not include the table into the list.
In case of pre-locked mode locked tables list is generated automatically,
so we may only end up here if the table did not exist when
locked tables list was created.
*/
if (thd->locked_tables_mode == LTM_PRELOCKED)
my_error(ER_NO_SUCH_TABLE, MYF(0), table_list->db, table_list->alias);
else
my_error(ER_TABLE_NOT_LOCKED, MYF(0), alias);
return true;
}
// Non pre-locked/LOCK TABLES mode, and not using secondary storage engine.
// This is the normal use case.
if ((flags & (MYSQL_OPEN_HAS_MDL_LOCK | MYSQL_OPEN_SECONDARY_ENGINE)) == 0) {
/*
We are not under LOCK TABLES and going to acquire write-lock/
modify the base table. We need to acquire protection against
global read lock until end of this statement in order to have
this statement blocked by active FLUSH TABLES WITH READ LOCK.
We don't block acquire this protection under LOCK TABLES as
such protection already acquired at LOCK TABLES time and
not released until UNLOCK TABLES.
We don't block statements which modify only temporary tables
as these tables are not preserved by backup by any form of
backup which uses FLUSH TABLES WITH READ LOCK.
TODO: The fact that we sometimes acquire protection against
GRL only when we encounter table to be write-locked
slightly increases probability of deadlock.
This problem will be solved once Alik pushes his
temporary table refactoring patch and we can start
pre-acquiring metadata locks at the beginning of
open_tables() call.
*/
if (table_list->mdl_request.is_write_lock_request() &&
!(flags &
(MYSQL_OPEN_IGNORE_GLOBAL_READ_LOCK | MYSQL_OPEN_FORCE_SHARED_MDL |
MYSQL_OPEN_FORCE_SHARED_HIGH_PRIO_MDL |
MYSQL_OPEN_SKIP_SCOPED_MDL_LOCK)) &&
!ot_ctx->has_protection_against_grl()) {
MDL_request protection_request;
MDL_deadlock_handler mdl_deadlock_handler(ot_ctx);
if (thd->global_read_lock.can_acquire_protection()) return true;
MDL_REQUEST_INIT(&protection_request, MDL_key::GLOBAL, "", "",
MDL_INTENTION_EXCLUSIVE, MDL_STATEMENT);
/*
Install error handler which if possible will convert deadlock error
into request to back-off and restart process of opening tables.
Prefer this context as a victim in a deadlock when such a deadlock
can be easily handled by back-off and retry.
*/
thd->push_internal_handler(&mdl_deadlock_handler);
thd->mdl_context.set_force_dml_deadlock_weight(ot_ctx->can_back_off());
bool result = thd->mdl_context.acquire_lock(&protection_request,
ot_ctx->get_timeout());
/*
Unlike in other places where we acquire protection against global read
lock, the read_only state is not checked here since we check its state
later in mysql_lock_tables()
*/
thd->mdl_context.set_force_dml_deadlock_weight(false);
thd->pop_internal_handler();
if (result) return true;
ot_ctx->set_has_protection_against_grl();
}
if (open_table_get_mdl_lock(thd, ot_ctx, table_list, flags, &mdl_ticket) ||
mdl_ticket == nullptr) {
DEBUG_SYNC(thd, "before_open_table_wait_refresh");
return true;
}
DEBUG_SYNC(thd, "after_open_table_mdl_shared");
} else {
/*
Grab reference to the MDL lock ticket that was acquired
by the caller.
*/
mdl_ticket = table_list->mdl_request.ticket;
}
if (table_list->open_strategy == Table_ref::OPEN_IF_EXISTS ||
table_list->open_strategy == Table_ref::OPEN_FOR_CREATE) {
bool exists;
if (check_if_table_exists(thd, table_list, &exists)) return true;
/*
If the table does not exist then upgrade the lock to the EXCLUSIVE MDL
lock.
*/
if (!exists) {
if (table_list->open_strategy == Table_ref::OPEN_FOR_CREATE &&
!(flags & (MYSQL_OPEN_FORCE_SHARED_MDL |
MYSQL_OPEN_FORCE_SHARED_HIGH_PRIO_MDL))) {
MDL_deadlock_handler mdl_deadlock_handler(ot_ctx);
thd->push_internal_handler(&mdl_deadlock_handler);
DEBUG_SYNC(thd, "before_upgrading_lock_from_S_to_X_for_create_table");
bool wait_result = thd->mdl_context.upgrade_shared_lock(
table_list->mdl_request.ticket, MDL_EXCLUSIVE,
thd->variables.lock_wait_timeout);
thd->pop_internal_handler();
DEBUG_SYNC(thd, "after_upgrading_lock_from_S_to_X_for_create_table");
/* Deadlock or timeout occurred while upgrading the lock. */
if (wait_result) return true;
}
return false;
}
/* Table exists. Let us try to open it. */
} else if (table_list->open_strategy == Table_ref::OPEN_STUB)
return false;
retry_share : {
Table_cache *tc = table_cache_manager.get_cache(thd);
tc->lock();
/*
Try to get unused TABLE object or at least pointer to
TABLE_SHARE from the table cache.
*/
if (!table_list->is_view())
table = tc->get_table(thd, key, key_length, &share);
if (table) {
/* We have found an unused TABLE object. */
if (!(flags & MYSQL_OPEN_IGNORE_FLUSH)) {
/*
TABLE_SHARE::version can only be initialised while holding the
LOCK_open and in this case no one has a reference to the share
object, if a reference exists to the share object it is necessary
to lock both LOCK_open AND all table caches in order to update
TABLE_SHARE::version. The same locks are required to increment
refresh_version global variable.
As result it is safe to compare TABLE_SHARE::version and
refresh_version values while having only lock on the table
cache for this thread.
Table cache should not contain any unused TABLE objects with
old versions.
*/
assert(!share->has_old_version());
/*
Still some of already opened might become outdated (e.g. due to
concurrent table flush). So we need to compare version of opened
tables with version of TABLE object we just have got.
*/
if (thd->open_tables &&
thd->open_tables->s->version() != share->version()) {
tc->release_table(thd, table);
tc->unlock();
(void)ot_ctx->request_backoff_action(
Open_table_context::OT_REOPEN_TABLES, nullptr);
return true;
}
}
tc->unlock();
/* Call rebind_psi outside of the critical section. */
assert(table->file != nullptr);
table->file->rebind_psi();
table->file->ha_extra(HA_EXTRA_RESET_STATE);
thd->status_var.table_open_cache_hits++;
goto table_found;
} else if (share) {
/*
We weren't able to get an unused TABLE object. Still we have
found TABLE_SHARE for it. So let us try to create new TABLE
for it. We start by incrementing share's reference count and
checking its version.
*/
mysql_mutex_lock(&LOCK_open);
tc->unlock();
share->increment_ref_count();
goto share_found;
} else {
/*
We have not found neither TABLE nor TABLE_SHARE object in
table cache (this means that there are no TABLE objects for
it in it).
Let us try to get TABLE_SHARE from table definition cache or
from disk and then to create TABLE object for it.
*/
tc->unlock();
}
}
mysql_mutex_lock(&LOCK_open);
if (!(share = get_table_share_with_discover(
thd, table_list, key, key_length,
flags & MYSQL_OPEN_SECONDARY_ENGINE, &error))) {
mysql_mutex_unlock(&LOCK_open);
/*
If thd->is_error() is not set, we either need discover
(error == 7), or the error was silenced by the prelocking
handler (error == 0), in which case we should skip this
table.
*/
if (error == 7 && !thd->is_error()) {
(void)ot_ctx->request_backoff_action(Open_table_context::OT_DISCOVER,
table_list);
}
return true;
}
/*
If a view is anticipated or the TABLE_SHARE object is a view, perform
a version check for it without creating a TABLE object.
Note that there is no need to call TABLE_SHARE::has_old_version() as we
do for regular tables, because view shares are always up to date.
*/
if (table_list->is_view() || share->is_view) {
bool view_open_result = true;
/*
If parent_l of the table_list is non null then a merge table
has this view as child table, which is not supported.
*/
if (table_list->parent_l) my_error(ER_WRONG_MRG_TABLE, MYF(0));
/*
Validate metadata version: in particular, that a view is opened when
it is expected, or that a table is opened when it is expected.
*/
else if (check_and_update_table_version(thd, table_list, share))
;
else if (table_list->open_strategy == Table_ref::OPEN_FOR_CREATE) {
/*
Skip reading the view definition if the open is for a table to be
created. This scenario will happen only when there exists a view and
the current CREATE TABLE request is with the same name.
*/
release_table_share(share);
mysql_mutex_unlock(&LOCK_open);
/*
The LEX object is used by the executor and other parts of the
code to detect the presence of a view. As this is
OPEN_FOR_CREATE we skip the call to open_and_read_view(),
which creates the LEX object, and create a dummy LEX object.
For SP and PS, LEX objects are created at the time of
statement prepare and open_table() is called for every execute
after that. Skip creation of LEX objects if it is already
present.
*/
if (!table_list->is_view()) return add_view_place_holder(thd, table_list);
return false;
} else {
/*
Read definition of existing view.
*/
view_open_result = open_and_read_view(thd, share, table_list);
}
/* TODO: Don't free this */
release_table_share(share);
mysql_mutex_unlock(&LOCK_open);
if (view_open_result) return true;
if (parse_view_definition(thd, table_list)) return true;
assert(table_list->is_view());
return false;
}
share_found:
if (!(flags & MYSQL_OPEN_IGNORE_FLUSH)) {
if (share->has_old_version()) {
/*
We already have an MDL lock. But we have encountered an old
version of table in the table definition cache which is possible
when someone changes the table version directly in the cache
without acquiring a metadata lock (e.g. this can happen during
"rolling" FLUSH TABLE(S)).
Release our reference to share, wait until old version of
share goes away and then try to get new version of table share.
*/
release_table_share(share);
mysql_mutex_unlock(&LOCK_open);
MDL_deadlock_handler mdl_deadlock_handler(ot_ctx);
bool wait_result;
thd->push_internal_handler(&mdl_deadlock_handler);
/*
In case of deadlock we would like this thread to be preferred as
a deadlock victim when this deadlock can be nicely handled by
back-off and retry. We still have a few weird cases, like
FLUSH TABLES <table-list> WITH READ LOCK, where we use strong
metadata locks and open_tables() is called with some metadata
locks pre-acquired. In these cases we still want to use DDL
deadlock weight.
*/
uint deadlock_weight = ot_ctx->can_back_off()
? MDL_wait_for_subgraph::DEADLOCK_WEIGHT_DML
: mdl_ticket->get_deadlock_weight();
wait_result =
tdc_wait_for_old_version(thd, table_list->db, table_list->table_name,
ot_ctx->get_timeout(), deadlock_weight);
thd->pop_internal_handler();
if (wait_result) return true;
DEBUG_SYNC(thd, "open_table_before_retry");
goto retry_share;
}
if (thd->open_tables &&
thd->open_tables->s->version() != share->version()) {
/*
If the version changes while we're opening the tables,
we have to back off, close all the tables opened-so-far,
and try to reopen them. Note: refresh_version is currently
changed only during FLUSH TABLES.
*/
release_table_share(share);
mysql_mutex_unlock(&LOCK_open);
(void)ot_ctx->request_backoff_action(Open_table_context::OT_REOPEN_TABLES,
nullptr);
return true;
}
}
mysql_mutex_unlock(&LOCK_open);
DEBUG_SYNC(thd, "open_table_found_share");
{
dd::cache::Dictionary_client::Auto_releaser releaser(thd->dd_client());
const dd::Table *table_def = nullptr;
if (!(flags & MYSQL_OPEN_NO_NEW_TABLE_IN_SE) &&
thd->dd_client()->acquire(share->db.str, share->table_name.str,
&table_def)) {
// Error is reported by the dictionary subsystem.
goto err_lock;
}
if (table_def && table_def->hidden() == dd::Abstract_table::HT_HIDDEN_SE) {
my_error(ER_NO_SUCH_TABLE, MYF(0), table_list->db,
table_list->table_name);
goto err_lock;
}
/* make a new table */
if (!(table = (TABLE *)my_malloc(key_memory_TABLE, sizeof(*table),
MYF(MY_WME))))
goto err_lock;
error = open_table_from_share(
thd, share, alias,
((flags & MYSQL_OPEN_NO_NEW_TABLE_IN_SE)
? 0
: ((uint)(HA_OPEN_KEYFILE | HA_OPEN_RNDFILE | HA_GET_INDEX |
HA_TRY_READ_ONLY))),
EXTRA_RECORD, thd->open_options, table, false, table_def);
if (error) {
destroy(table);
my_free(table);
if (error == 7)
(void)ot_ctx->request_backoff_action(Open_table_context::OT_DISCOVER,
table_list);
else if (error == 8)
(void)ot_ctx->request_backoff_action(
Open_table_context::OT_FIX_ROW_TYPE, table_list);
else if (share->crashed)
(void)ot_ctx->request_backoff_action(Open_table_context::OT_REPAIR,
table_list);
goto err_lock;
} else if (share->crashed) {
switch (thd->lex->sql_command) {
case SQLCOM_ALTER_TABLE:
case SQLCOM_REPAIR:
case SQLCOM_CHECK:
case SQLCOM_SHOW_CREATE:
break;
default:
closefrm(table, false);
destroy(table);
my_free(table);
my_error(ER_CRASHED_ON_USAGE, MYF(0), share->table_name.str);
goto err_lock;
}
}
if (open_table_entry_fini(thd, share, table_def, table)) {
closefrm(table, false);
destroy(table);
my_free(table);
goto err_lock;
}
}
{
/* Add new TABLE object to table cache for this connection. */
Table_cache *tc = table_cache_manager.get_cache(thd);
tc->lock();
if (tc->add_used_table(thd, table)) {
tc->unlock();
goto err_lock;
}
tc->unlock();
}
thd->status_var.table_open_cache_misses++;
table_found:
table->mdl_ticket = mdl_ticket;
table->next = thd->open_tables; /* Link into simple list */
thd->set_open_tables(table);
table->reginfo.lock_type = TL_READ; /* Assume read */
reset:
table->reset();
table->set_created();
/*
Check that there is no reference to a condition from an earlier query
(cf. Bug#58553).
*/
assert(table->file->pushed_cond == nullptr);
// Table is not a derived table and not a non-updatable view:
table_list->set_updatable();
table_list->set_insertable();
table_list->table = table;
/*
Position for each partition in the bitmap is read from the Handler_share
instance of the table. In MYSQL_OPEN_NO_NEW_TABLE_IN_SE mode, table is not
opened in the SE and Handler_share instance for it is not created. Hence
skipping partitions bitmap setting in the MYSQL_OPEN_NO_NEW_TABLE_IN_SE
mode.
*/
if (!(flags & MYSQL_OPEN_NO_NEW_TABLE_IN_SE)) {
if (table->part_info) {
/* Set all [named] partitions as used. */
if (table->part_info->set_partition_bitmaps(table_list)) return true;
} else if (table_list->partition_names) {
/* Don't allow PARTITION () clause on a nonpartitioned table */
my_error(ER_PARTITION_CLAUSE_ON_NONPARTITIONED, MYF(0));
return true;
}
}
table->init(thd, table_list);
/* Request a read lock for implicitly opened P_S tables. */
if (in_LTM(thd) && table_list->table->file->get_lock_type() == F_UNLCK &&
belongs_to_p_s(table_list)) {
table_list->table->file->ha_external_lock(thd, F_RDLCK);
}
return false;
err_lock:
mysql_mutex_lock(&LOCK_open);
release_table_share(share);
mysql_mutex_unlock(&LOCK_open);
return true;
}
/**
Find table in the list of open tables.
@param list List of TABLE objects to be inspected.
@param db Database name
@param table_name Table name
@return Pointer to the TABLE object found, 0 if no table found.
*/
TABLE *find_locked_table(TABLE *list, const char *db, const char *table_name) {
char key[MAX_DBKEY_LENGTH];
size_t key_length = create_table_def_key(db, table_name, key);
for (TABLE *table = list; table; table = table->next) {
if (table->s->table_cache_key.length == key_length &&
!memcmp(table->s->table_cache_key.str, key, key_length))
return table;
}
return (nullptr);
}
/**
Find instance of TABLE with upgradable or exclusive metadata
lock from the list of open tables, emit error if no such table
found.
@param thd Thread context
@param db Database name.
@param table_name Name of table.
@param no_error Don't emit error if no suitable TABLE
instance were found.
@note This function checks if the connection holds a global IX
metadata lock. If no such lock is found, it is not safe to
upgrade the lock and ER_TABLE_NOT_LOCKED_FOR_WRITE will be
reported.
@return Pointer to TABLE instance with MDL_SHARED_UPGRADABLE
MDL_SHARED_NO_WRITE, MDL_SHARED_NO_READ_WRITE, or
MDL_EXCLUSIVE metadata lock, NULL otherwise.
*/
TABLE *find_table_for_mdl_upgrade(THD *thd, const char *db,
const char *table_name, bool no_error) {
TABLE *tab = find_locked_table(thd->open_tables, db, table_name);
if (!tab) {
if (!no_error) my_error(ER_TABLE_NOT_LOCKED, MYF(0), table_name);
return nullptr;
}
/*
It is not safe to upgrade the metadata lock without a global IX lock.
This can happen with FLUSH TABLES <list> WITH READ LOCK as we in these
cases don't take a global IX lock in order to be compatible with
global read lock.
*/
if (!thd->mdl_context.owns_equal_or_stronger_lock(MDL_key::GLOBAL, "", "",
MDL_INTENTION_EXCLUSIVE)) {
if (!no_error) my_error(ER_TABLE_NOT_LOCKED_FOR_WRITE, MYF(0), table_name);
return nullptr;
}
while (tab->mdl_ticket != nullptr &&
!tab->mdl_ticket->is_upgradable_or_exclusive() &&
(tab = find_locked_table(tab->next, db, table_name)))
continue;
if (!tab && !no_error)
my_error(ER_TABLE_NOT_LOCKED_FOR_WRITE, MYF(0), table_name);
return tab;
}
/*
Function to assign a new table map id to a table share.
PARAMETERS
share - Pointer to table share structure
DESCRIPTION
We are intentionally not checking that share->mutex is locked
since this function should only be called when opening a table
share and before it is entered into the table_def_cache (meaning
that it cannot be fetched by another thread, even accidentally).
PRE-CONDITION(S)
share is non-NULL
The LOCK_open mutex is locked.
POST-CONDITION(S)
share->table_map_id is given a value that with a high certainty is
not used by any other table (the only case where a table id can be
reused is on wrap-around, which means more than 2^48 table
share opens have been executed while one table was open all the
time).
*/
static Table_id last_table_id;
void assign_new_table_id(TABLE_SHARE *share) {
DBUG_TRACE;
/* Preconditions */
assert(share != nullptr);
mysql_mutex_assert_owner(&LOCK_open);
DBUG_EXECUTE_IF("dbug_table_map_id_500", last_table_id = 500;);
DBUG_EXECUTE_IF("dbug_table_map_id_4B_UINT_MAX+501",
last_table_id = 501ULL + UINT_MAX;);
DBUG_EXECUTE_IF("dbug_table_map_id_6B_UINT_MAX",
last_table_id = (~0ULL >> 16););
share->table_map_id = last_table_id++;
DBUG_PRINT("info", ("table_id=%llu", share->table_map_id.id()));
}
/**
Compare metadata versions of an element obtained from the table
definition cache and its corresponding node in the parse tree.
@details If the new and the old values mismatch, invoke
Metadata_version_observer.
At prepared statement prepare, all Table_ref version values are
NULL and we always have a mismatch. But there is no observer set
in THD, and therefore no error is reported. Instead, we update
the value in the parse tree, effectively recording the original
version.
At prepared statement execute, an observer may be installed. If
there is a version mismatch, we push an error and return true.
For conventional execution (no prepared statements), the
observer is never installed.
@sa Execute_observer
@sa check_prepared_statement() to see cases when an observer is installed
@sa Table_ref::is_table_ref_id_equal()
@sa TABLE_SHARE::get_table_ref_id()
@param[in] thd used to report errors
@param[in,out] tables Table_ref instance created by the parser
Metadata version information in this object
is updated upon success.
@param[in] table_share an element from the table definition cache
@retval true an error, which has been reported
@retval false success, version in Table_ref has been updated
*/
static bool check_and_update_table_version(THD *thd, Table_ref *tables,
TABLE_SHARE *table_share) {
if (!tables->is_table_ref_id_equal(table_share)) {
/*
Version of the table share is different from the
previous execution of the prepared statement, and it is
unacceptable for this SQLCOM.
*/
if (ask_to_reprepare(thd)) return true;
/* Always maintain the latest version and type */
tables->set_table_ref_id(table_share);
}
return false;
}
/**
Compares versions of a stored routine obtained from the sp cache
and the version used at prepare.
@details If the new and the old values mismatch, invoke
Metadata_version_observer.
At prepared statement prepare, all Sroutine_hash_entry version values
are NULL and we always have a mismatch. But there is no observer set
in THD, and therefore no error is reported. Instead, we update
the value in Sroutine_hash_entry, effectively recording the original
version.
At prepared statement execute, an observer may be installed. If
there is a version mismatch, we push an error and return true.
For conventional execution (no prepared statements), the
observer is never installed.
@param[in] thd used to report errors
@param[in,out] rt pointer to stored routine entry in the
parse tree
@param[in] sp pointer to stored routine cache entry.
Can be NULL if there is no such routine.
@retval true an error, which has been reported
@retval false success, version in Sroutine_hash_entry has been updated
*/
static bool check_and_update_routine_version(THD *thd, Sroutine_hash_entry *rt,
sp_head *sp) {
int64 spc_version = sp_cache_version();
/* sp is NULL if there is no such routine. */
int64 version = sp ? sp->sp_cache_version() : spc_version;
/*
If the version in the parse tree is stale,
or the version in the cache is stale and sp is not used,
we need to reprepare.
Sic: version != spc_version <--> sp is not NULL.
*/
if (rt->m_cache_version != version ||
(version != spc_version && !sp->is_invoked())) {
/*
Version of the sp cache is different from the
previous execution of the prepared statement, and it is
unacceptable for this SQLCOM.
*/
if (ask_to_reprepare(thd)) return true;
/* Always maintain the latest cache version. */
rt->m_cache_version = version;
}
return false;
}
/**
Open view by getting its definition from disk (and table cache in future).
@param thd Thread handle
@param table_list Table_ref with db, table_name & belong_to_view
@param cache_key Key for table definition cache
@param cache_key_length Length of cache_key
@todo This function is needed for special handling of views under
LOCK TABLES. We probably should get rid of it in long term.
@return false if success, true - otherwise.
*/
static bool tdc_open_view(THD *thd, Table_ref *table_list,
const char *cache_key, size_t cache_key_length) {
TABLE_SHARE *share;
mysql_mutex_lock(&LOCK_open);
if (!(share = get_table_share(thd, table_list->db, table_list->table_name,
cache_key, cache_key_length, true))) {
mysql_mutex_unlock(&LOCK_open);
return true;
}
/*
Check TABLE_SHARE-version of view only if we have been instructed to do
so. We do not need to check the version if we're executing CREATE VIEW or
ALTER VIEW statements.
In the future, this functionality should be moved out from
tdc_open_view(), and tdc_open_view() should became a part of a clean
table-definition-cache interface.
*/
if (check_and_update_table_version(thd, table_list, share)) {
release_table_share(share);
mysql_mutex_unlock(&LOCK_open);
return true;
}
if (share->is_view) {
bool view_open_result = open_and_read_view(thd, share, table_list);
release_table_share(share);
mysql_mutex_unlock(&LOCK_open);
if (view_open_result) return true;
return parse_view_definition(thd, table_list);
}
my_error(ER_WRONG_OBJECT, MYF(0), share->db.str, share->table_name.str,
"VIEW");
release_table_share(share);
mysql_mutex_unlock(&LOCK_open);
return true;
}
/**
Finalize the process of TABLE creation by loading table triggers
and taking action if a HEAP table content was emptied implicitly.
*/
static bool open_table_entry_fini(THD *thd, TABLE_SHARE *share,
const dd::Table *table, TABLE *entry) {
if (table != nullptr && table->has_trigger()) {
Table_trigger_dispatcher *d = Table_trigger_dispatcher::create(entry);
if (!d || d->check_n_load(thd, *table)) {
destroy(d);
return true;
}
entry->triggers = d;
}
/*
If we are here, there was no fatal error (but error may be still
uninitialized).
Ignore handling implicit_emptied property (which is only for heap
tables) when I_S query is opening this table to read table statistics.
The reason for avoiding this is that the
mysql_bin_log.write_dml_directly() invokes a commit(). And this commit
is not expected to be invoked when fetching I_S table statistics.
*/
if (unlikely(entry->file->implicit_emptied) &&
(!thd->lex || !thd->lex->m_IS_table_stats.is_reading_stats_by_open())) {
entry->file->implicit_emptied = false;
if (mysql_bin_log.is_open()) {
bool result = false;
String temp_buf;
result = temp_buf.append("TRUNCATE TABLE ");
append_identifier(thd, &temp_buf, share->db.str, strlen(share->db.str));
result = temp_buf.append(".");
append_identifier(thd, &temp_buf, share->table_name.str,
strlen(share->table_name.str));
result = temp_buf.append(
" /* generated by server, implicitly emptying in-memory table */");
if (result) {
/*
As replication is maybe going to be corrupted, we need to warn the
DBA on top of warning the client (which will automatically be done
because of MYF(MY_WME) in my_malloc() above).
*/
LogErr(ERROR_LEVEL,
ER_BINLOG_OOM_WRITING_DELETE_WHILE_OPENING_HEAP_TABLE,
share->db.str, share->table_name.str);
destroy(entry->triggers);
return true;
}
/*
Create a new THD object for binary logging the statement which
implicitly empties the in-memory table.
*/
THD new_thd;
new_thd.thread_stack = (char *)&thd;
new_thd.set_new_thread_id();
new_thd.store_globals();
new_thd.set_db(thd->db());
new_thd.variables.gtid_next.set_automatic();
Global_THD_manager *thd_manager = Global_THD_manager::get_instance();
thd_manager->add_thd(&new_thd);
result = mysql_bin_log.write_stmt_directly(
&new_thd, temp_buf.c_ptr_safe(), temp_buf.length(), SQLCOM_TRUNCATE);
new_thd.restore_globals();
thd->store_globals();
new_thd.release_resources();
thd_manager->remove_thd(&new_thd);
return result;
}
}
return false;
}
/**
Auxiliary routine which is used for performing automatic table repair.
*/
static bool auto_repair_table(THD *thd, Table_ref *table_list) {
const char *cache_key;
size_t cache_key_length;
TABLE_SHARE *share;
TABLE *entry;
bool result = true;
cache_key_length = get_table_def_key(table_list, &cache_key);
thd->clear_error();
mysql_mutex_lock(&LOCK_open);
if (!(share = get_table_share(thd, table_list->db, table_list->table_name,
cache_key, cache_key_length, true)))
goto end_unlock;
if (share->is_view) {
release_table_share(share);
goto end_unlock;
}
if (!(entry =
(TABLE *)my_malloc(key_memory_TABLE, sizeof(TABLE), MYF(MY_WME)))) {
release_table_share(share);
goto end_unlock;
}
mysql_mutex_unlock(&LOCK_open);
if (open_table_from_share(thd, share, table_list->alias,
(uint)(HA_OPEN_KEYFILE | HA_OPEN_RNDFILE |
HA_GET_INDEX | HA_TRY_READ_ONLY),
EXTRA_RECORD, ha_open_options | HA_OPEN_FOR_REPAIR,
entry, false, nullptr) ||
!entry->file ||
(entry->file->is_crashed() && entry->file->ha_check_and_repair(thd))) {
/* Give right error message */
thd->clear_error();
my_error(ER_NOT_KEYFILE, MYF(0), share->table_name.str);
LogErr(ERROR_LEVEL, ER_FAILED_TO_REPAIR_TABLE, share->db.str,
share->table_name.str);
if (entry->file) closefrm(entry, false);
} else {
thd->clear_error(); // Clear error message
closefrm(entry, false);
result = false;
}
my_free(entry);
table_cache_manager.lock_all_and_tdc();
release_table_share(share);
/* Remove the repaired share from the table cache. */
tdc_remove_table(thd, TDC_RT_REMOVE_ALL, table_list->db,
table_list->table_name, true);
table_cache_manager.unlock_all_and_tdc();
return result;
end_unlock:
mysql_mutex_unlock(&LOCK_open);
return result;
}
/**
Error handler class for suppressing HA_ERR_ROW_FORMAT_CHANGED errors from SE.
*/
class Fix_row_type_error_handler : public Internal_error_handler {
public:
bool handle_condition(THD *, uint sql_errno, const char *,
Sql_condition::enum_severity_level *,
const char *) override {
return sql_errno == ER_GET_ERRNO && my_errno() == HA_ERR_ROW_FORMAT_CHANGED;
}
};
/**
Auxiliary routine for automatically updating row format for the table.
*/
static bool fix_row_type(THD *thd, Table_ref *table_list) {
const char *cache_key;
size_t cache_key_length = get_table_def_key(table_list, &cache_key);
thd->clear_error();
TABLE_SHARE *share;
{
/*
Hold LOCK_open until we can keep it and are likely to
release TABLE_SHARE on return.
*/
MUTEX_LOCK(lock_open_guard, &LOCK_open);
No_such_table_error_handler no_such_table_handler;
thd->push_internal_handler(&no_such_table_handler);
share = get_table_share(thd, table_list->db, table_list->table_name,
cache_key, cache_key_length, true);
thd->pop_internal_handler();
if (!share) {
/*
Somebody managed to drop table after we have performed back-off
before trying to fix row format for the table. Such situation is
quite unlikely but theoretically possible. Do not report error
(silence it using error handler), let the caller try to reopen
tables and handle missing table in appropriate way (e.g. ignore
this fact it if the table comes from prelocking list).
*/
if (no_such_table_handler.safely_trapped_errors()) return false;
return true;
}
if (share->is_view) {
/*
Somebody managed to replace our table with a view after we
have performed back-off before trying to fix row format for
the table. Such situation is quite unlikely but is OK.
Do not report error, let the caller try to reopen tables.
*/
release_table_share(share);
return false;
}
}
int error = 0;
dd::cache::Dictionary_client::Auto_releaser releaser(thd->dd_client());
dd::Table *table_def = nullptr;
if (thd->dd_client()->acquire_for_modification(
share->db.str, share->table_name.str, &table_def))
error = 1;
assert(table_def != nullptr);
/*
Silence expected HA_ERR_ROW_FORMAT_CHANGED errors.
*/
Fix_row_type_error_handler err_handler;
thd->push_internal_handler(&err_handler);
TABLE tmp_table;
if (error == 0)
error = open_table_from_share(thd, share, table_list->alias,
(uint)(HA_OPEN_KEYFILE | HA_OPEN_RNDFILE |
HA_GET_INDEX | HA_TRY_READ_ONLY),
EXTRA_RECORD, ha_open_options, &tmp_table,
false, table_def);
thd->pop_internal_handler();
if (error == 8) {
Disable_autocommit_guard autocommit_guard(thd);
HA_CREATE_INFO create_info;
create_info.row_type = share->row_type;
create_info.table_options = share->db_options_in_use;
handler *file = get_new_handler(share, share->m_part_info != nullptr,
thd->mem_root, share->db_type());
if (file != nullptr) {
row_type correct_row_type = file->get_real_row_type(&create_info);
bool result = dd::fix_row_type(thd, table_def, correct_row_type);
destroy(file);
if (result) {
trans_rollback_stmt(thd);
trans_rollback(thd);
} else {
result = trans_commit_stmt(thd) || trans_commit(thd);
if (!result) error = 0;
}
}
} else if (error == 0)
closefrm(&tmp_table, false);
table_cache_manager.lock_all_and_tdc();
release_table_share(share);
/*
Remove the share from the table cache. So attempt to reopen table
will construct its new version with correct real_row_type value.
*/
tdc_remove_table(thd, TDC_RT_REMOVE_ALL, table_list->db,
table_list->table_name, true);
table_cache_manager.unlock_all_and_tdc();
return error != 0;
}
/** Open_table_context */
Open_table_context::Open_table_context(THD *thd, uint flags)
: m_thd(thd),
m_failed_table(nullptr),
m_start_of_statement_svp(thd->mdl_context.mdl_savepoint()),
m_timeout(flags & MYSQL_LOCK_IGNORE_TIMEOUT
? LONG_TIMEOUT
: thd->variables.lock_wait_timeout),
m_flags(flags),
m_action(OT_NO_ACTION),
m_has_locks(thd->mdl_context.has_locks()),
m_has_protection_against_grl(false) {}
/**
Check if we can back-off and set back off action if we can.
Otherwise report and return error.
@retval true if back-off is impossible.
@retval false if we can back off. Back off action has been set.
*/
bool Open_table_context::request_backoff_action(
enum_open_table_action action_arg, Table_ref *table) {
/*
A back off action may be one of four kinds:
* We met a broken table that needs repair, or a table that
is not present on this MySQL server and needs re-discovery.
To perform the action, we need an exclusive metadata lock on
the table. Acquiring X lock while holding other shared
locks can easily lead to deadlocks. We rely on MDL deadlock
detector to discover them. If this is a multi-statement
transaction that holds metadata locks for completed statements,
we should keep these locks after discovery/repair.
The action type in this case is OT_DISCOVER or OT_REPAIR.
* We met a table that has outdated value in ROW_FORMAT column
in the data-dictionary/value of TABLE_SHARE::real_row_type
attribute, which need to be updated. To update the
data-dictionary we not only need to acquire X lock on the
table, but also need to commit the transaction. If there
is an ongoing transaction (and some metadata locks acquired)
we cannot proceed and report an error. The action type for
this case is OT_FIX_ROW_TYPE.
* Our attempt to acquire an MDL lock lead to a deadlock,
detected by the MDL deadlock detector. The current
session was chosen a victim. If this is a multi-statement
transaction that holds metadata locks taken by completed
statements, restarting locking for the current statement
may lead to a livelock. Releasing locks of completed
statements can not be done as will lead to violation
of ACID. Thus, again, if m_has_locks is set,
we report an error. Otherwise, when there are no metadata
locks other than which belong to this statement, we can
try to recover from error by releasing all locks and
restarting the pre-locking.
Similarly, a deadlock error can occur when the
pre-locking process met a TABLE_SHARE that is being
flushed, and unsuccessfully waited for the flush to
complete. A deadlock in this case can happen, e.g.,
when our session is holding a metadata lock that
is being waited on by a session which is using
the table which is being flushed. The only way
to recover from this error is, again, to close all
open tables, release all locks, and retry pre-locking.
Action type name is OT_REOPEN_TABLES. Re-trying
while holding some locks may lead to a livelock,
and thus we don't do it.
* Finally, this session has open TABLEs from different
"generations" of the table cache. This can happen, e.g.,
when, after this session has successfully opened one
table used for a statement, FLUSH TABLES interfered and
expelled another table used in it. FLUSH TABLES then
blocks and waits on the table already opened by this
statement.
We detect this situation by ensuring that table cache
version of all tables used in a statement is the same.
If it isn't, all tables needs to be reopened.
Note, that we can always perform a reopen in this case,
even if we already have metadata locks, since we don't
keep tables open between statements and a livelock
is not possible.
*/
if ((action_arg == OT_BACKOFF_AND_RETRY || action_arg == OT_FIX_ROW_TYPE) &&
(has_commit_order_manager(m_thd) || m_has_locks)) {
my_error(ER_LOCK_DEADLOCK, MYF(0));
m_thd->mark_transaction_to_rollback(true);
return true;
}
/*
If auto-repair or discovery are requested, a pointer to table
list element must be provided.
*/
if (table) {
assert(action_arg == OT_DISCOVER || action_arg == OT_REPAIR ||
action_arg == OT_FIX_ROW_TYPE);
m_failed_table = new (m_thd->mem_root)
Table_ref(table->db, table->db_length, table->table_name,
table->table_name_length, table->alias, TL_WRITE);
if (m_failed_table == nullptr) return true;
m_failed_table->mdl_request.set_type(MDL_EXCLUSIVE);
}
m_action = action_arg;
return false;
}
/**
An error handler to mark transaction to rollback on DEADLOCK error
during DISCOVER / REPAIR.
*/
class MDL_deadlock_discovery_repair_handler : public Internal_error_handler {
public:
bool handle_condition(THD *thd, uint sql_errno, const char *,
Sql_condition::enum_severity_level *,
const char *) override {
if (sql_errno == ER_LOCK_DEADLOCK) {
thd->mark_transaction_to_rollback(true);
}
/*
We have marked this transaction to rollback. Return false to allow
error to be reported or handled by other handlers.
*/
return false;
}
};
/**
Recover from failed attempt of open table by performing requested action.
@pre This function should be called only with "action" != OT_NO_ACTION
and after having called @sa close_tables_for_reopen().
@retval false - Success. One should try to open tables once again.
@retval true - Error
*/
bool Open_table_context::recover_from_failed_open() {
if (m_action == OT_REPAIR) {
DEBUG_SYNC(m_thd, "recover_ot_repair");
}
/*
Skip repair and discovery in IS-queries as they require X lock
which could lead to delays or deadlock. Instead set
ER_WARN_I_S_SKIPPED_TABLE which will be converted to a warning
later.
*/
if ((m_action == OT_REPAIR || m_action == OT_DISCOVER ||
m_action == OT_FIX_ROW_TYPE) &&
(m_flags & MYSQL_OPEN_FAIL_ON_MDL_CONFLICT)) {
my_error(ER_WARN_I_S_SKIPPED_TABLE, MYF(0),
m_failed_table->mdl_request.key.db_name(),
m_failed_table->mdl_request.key.name());
return true;
}
bool result = false;
MDL_deadlock_discovery_repair_handler handler;
/*
Install error handler to mark transaction to rollback on DEADLOCK error.
*/
m_thd->push_internal_handler(&handler);
/* Execute the action. */
switch (m_action) {
case OT_BACKOFF_AND_RETRY:
break;
case OT_REOPEN_TABLES:
break;
case OT_DISCOVER: {
if ((result = lock_table_names(m_thd, m_failed_table, nullptr,
get_timeout(), 0)))
break;
tdc_remove_table(m_thd, TDC_RT_REMOVE_ALL, m_failed_table->db,
m_failed_table->table_name, false);
if (ha_create_table_from_engine(m_thd, m_failed_table->db,
m_failed_table->table_name)) {
result = true;
break;
}
m_thd->get_stmt_da()->reset_condition_info(m_thd);
m_thd->clear_error(); // Clear error message
/*
Rollback to start of the current statement to release exclusive lock
on table which was discovered but preserve locks from previous
statements in current transaction.
*/
m_thd->mdl_context.rollback_to_savepoint(start_of_statement_svp());
break;
}
case OT_REPAIR: {
if ((result = lock_table_names(m_thd, m_failed_table, nullptr,
get_timeout(), 0)))
break;
tdc_remove_table(m_thd, TDC_RT_REMOVE_ALL, m_failed_table->db,
m_failed_table->table_name, false);
result = auto_repair_table(m_thd, m_failed_table);
/*
Rollback to start of the current statement to release exclusive lock
on table which was discovered but preserve locks from previous
statements in current transaction.
*/
m_thd->mdl_context.rollback_to_savepoint(start_of_statement_svp());
break;
}
case OT_FIX_ROW_TYPE: {
/*
Since we are going to commit changes to the data-dictionary there
should not be any ongoing transaction.
We already have checked that the connection holds no metadata locks
earlier.
Still there can be transaction started by START TRANSACTION, which
we don't have right to implicitly finish (even more interesting case
is START TRANSACTION WITH CONSISTENT SNAPSHOT). Hence explicit check
for active transaction.
*/
assert(!m_thd->mdl_context.has_locks());
if (m_thd->in_active_multi_stmt_transaction()) {
my_error(ER_LOCK_OR_ACTIVE_TRANSACTION, MYF(0));
result = true;
break;
}
if ((result = lock_table_names(m_thd, m_failed_table, nullptr,
get_timeout(), 0)))
break;
result = fix_row_type(m_thd, m_failed_table);
m_thd->mdl_context.release_transactional_locks();
break;
}
default:
assert(0);
}
m_thd->pop_internal_handler();
/*
Reset the pointers to conflicting MDL request and the
Table_ref element, set when we need auto-discovery or repair,
for safety.
*/
m_failed_table = nullptr;
/*
Reset flag indicating that we have already acquired protection
against GRL. It is no longer valid as the corresponding lock was
released by close_tables_for_reopen().
*/
m_has_protection_against_grl = false;
/* Prepare for possible another back-off. */
m_action = OT_NO_ACTION;
return result;
}
/*
Return a appropriate read lock type given a table object.
@param thd Thread context
@param prelocking_ctx Prelocking context.
@param table_list Table list element for table to be locked.
@param routine_modifies_data
Some routine that is invoked by statement
modifies data.
@remark Due to a statement-based replication limitation, statements such as
INSERT INTO .. SELECT FROM .. and CREATE TABLE .. SELECT FROM need
to grab a TL_READ_NO_INSERT lock on the source table in order to
prevent the replication of a concurrent statement that modifies the
source table. If such a statement gets applied on the slave before
the INSERT .. SELECT statement finishes, data on the master could
differ from data on the slave and end-up with a discrepancy between
the binary log and table state.
This also applies to SELECT/SET/DO statements which use stored
functions. Calls to such functions are going to be logged as a
whole and thus should be serialized against concurrent changes
to tables used by those functions. This is avoided when functions
do not modify data but only read it, since in this case nothing is
written to the binary log. Argument routine_modifies_data
denotes the same. So effectively, if the statement is not a
LOCK TABLE, not a update query and routine_modifies_data is false
then prelocking_placeholder does not take importance.
Furthermore, this does not apply to I_S and log tables as it's
always unsafe to replicate such tables under statement-based
replication as the table on the slave might contain other data
(ie: general_log is enabled on the slave). The statement will
be marked as unsafe for SBR in decide_logging_format().
@remark Note that even in prelocked mode it is important to correctly
determine lock type value. In this mode lock type is passed to
handler::start_stmt() method and can be used by storage engine,
for example, to determine what kind of row locks it should acquire
when reading data from the table.
*/
thr_lock_type read_lock_type_for_table(THD *thd,
Query_tables_list *prelocking_ctx,
Table_ref *table_list,
bool routine_modifies_data) {
/*
In cases when this function is called for a sub-statement executed in
prelocked mode we can't rely on OPTION_BIN_LOG flag in THD::options
bitmap to determine that binary logging is turned on as this bit can
be cleared before executing sub-statement. So instead we have to look
at THD::variables::sql_log_bin member.
*/
bool log_on = mysql_bin_log.is_open() && thd->variables.sql_log_bin;
/*
When we do not write to binlog or when we use row based replication,
it is safe to use a weaker lock.
*/
if (log_on == false || thd->variables.binlog_format == BINLOG_FORMAT_ROW)
return TL_READ;
if ((table_list->table->s->table_category == TABLE_CATEGORY_LOG) ||
(table_list->table->s->table_category == TABLE_CATEGORY_RPL_INFO) ||
(table_list->table->s->table_category == TABLE_CATEGORY_GTID) ||
(table_list->table->s->table_category == TABLE_CATEGORY_PERFORMANCE))
return TL_READ;
// SQL queries which updates data need a stronger lock.
if (is_update_query(prelocking_ctx->sql_command)) return TL_READ_NO_INSERT;
/*
table_list is placeholder for prelocking.
Ignore prelocking_placeholder status for non "LOCK TABLE" statement's
table_list objects when routine_modifies_data is false.
*/
if (table_list->prelocking_placeholder &&
(routine_modifies_data || thd->in_lock_tables))
return TL_READ_NO_INSERT;
if (thd->locked_tables_mode > LTM_LOCK_TABLES) return TL_READ_NO_INSERT;
return TL_READ;
}
/**
Process table's foreign keys (if any) by prelocking algorithm.
@param thd Thread context.
@param prelocking_ctx Prelocking context of the statement.
@param share Table's share.
@param is_insert Indicates whether statement is going to INSERT
into the table.
@param is_update Indicates whether statement is going to UPDATE
the table.
@param is_delete Indicates whether statement is going to DELETE
from the table.
@param belong_to_view Uppermost view which uses this table element
(nullptr - if it is not used by a view).
@param[out] need_prelocking Set to true if method detects that prelocking
required, not changed otherwise.
*/
static void process_table_fks(THD *thd, Query_tables_list *prelocking_ctx,
TABLE_SHARE *share, bool is_insert,
bool is_update, bool is_delete,
Table_ref *belong_to_view,
bool *need_prelocking) {
if (!share->foreign_keys && !share->foreign_key_parents) {
/*
This table doesn't participate in any foreign keys, so nothing to
process.
*/
return;
}
*need_prelocking = true;
/*
In lower-case-table-names == 2 mode we store original versions of db
and table names for tables participating in FK relationship, even
though their comparison is performed in case insensitive fashion.
Therefore we need to normalize/lowercase these names while prelocking
set key is constructing from them.
*/
bool normalize_db_names = (lower_case_table_names == 2);
Sp_name_normalize_type name_normalize_type =
(lower_case_table_names == 2) ? Sp_name_normalize_type::LOWERCASE_NAME
: Sp_name_normalize_type::LEAVE_AS_IS;
if (is_insert || is_update) {
for (TABLE_SHARE_FOREIGN_KEY_INFO *fk = share->foreign_key;
fk < share->foreign_key + share->foreign_keys; ++fk) {
(void)sp_add_used_routine(
prelocking_ctx, thd->stmt_arena,
Sroutine_hash_entry::FK_TABLE_ROLE_PARENT_CHECK,
fk->referenced_table_db.str, fk->referenced_table_db.length,
fk->referenced_table_name.str, fk->referenced_table_name.length,
normalize_db_names, name_normalize_type, false, belong_to_view);
}
}
if (is_update || is_delete) {
for (TABLE_SHARE_FOREIGN_KEY_PARENT_INFO *fk_p = share->foreign_key_parent;
fk_p < share->foreign_key_parent + share->foreign_key_parents;
++fk_p) {
if ((is_update &&
(fk_p->update_rule == dd::Foreign_key::RULE_NO_ACTION ||
fk_p->update_rule == dd::Foreign_key::RULE_RESTRICT)) ||
(is_delete &&
(fk_p->delete_rule == dd::Foreign_key::RULE_NO_ACTION ||
fk_p->delete_rule == dd::Foreign_key::RULE_RESTRICT))) {
(void)sp_add_used_routine(
prelocking_ctx, thd->stmt_arena,
Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_CHECK,
fk_p->referencing_table_db.str, fk_p->referencing_table_db.length,
fk_p->referencing_table_name.str,
fk_p->referencing_table_name.length, normalize_db_names,
name_normalize_type, false, belong_to_view);
}
if ((is_update &&
(fk_p->update_rule == dd::Foreign_key::RULE_CASCADE ||
fk_p->update_rule == dd::Foreign_key::RULE_SET_NULL ||
fk_p->update_rule == dd::Foreign_key::RULE_SET_DEFAULT)) ||
(is_delete &&
(fk_p->delete_rule == dd::Foreign_key::RULE_SET_NULL ||
fk_p->delete_rule == dd::Foreign_key::RULE_SET_DEFAULT))) {
(void)sp_add_used_routine(
prelocking_ctx, thd->stmt_arena,
Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_UPDATE,
fk_p->referencing_table_db.str, fk_p->referencing_table_db.length,
fk_p->referencing_table_name.str,
fk_p->referencing_table_name.length, normalize_db_names,
name_normalize_type, false, belong_to_view);
}
if (is_delete && fk_p->delete_rule == dd::Foreign_key::RULE_CASCADE) {
(void)sp_add_used_routine(
prelocking_ctx, thd->stmt_arena,
Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_DELETE,
fk_p->referencing_table_db.str, fk_p->referencing_table_db.length,
fk_p->referencing_table_name.str,
fk_p->referencing_table_name.length, normalize_db_names,
name_normalize_type, false, belong_to_view);
}
}
}
}
/**
Handle element of prelocking set other than table. E.g. cache routine
and, if prelocking strategy prescribes so, extend the prelocking set
with tables and routines used by it.
@param[in] thd Thread context.
@param[in] prelocking_ctx Prelocking context.
@param[in] rt Element of prelocking set to be processed.
@param[in] prelocking_strategy Strategy which specifies how the
prelocking set should be extended when
one of its elements is processed.
@param[in] has_prelocking_list Indicates that prelocking set/list for
this statement has already been built.
@param[in] ot_ctx Context of open_table used to recover from
locking failures.
@param[out] need_prelocking Set to true if it was detected that this
statement will require prelocked mode for
its execution, not touched otherwise.
@param[out] routine_modifies_data Set to true if it was detected that this
routine does modify table data.
@retval false Success.
@retval true Failure (Conflicting metadata lock, OOM, other errors).
*/
static bool open_and_process_routine(
THD *thd, Query_tables_list *prelocking_ctx, Sroutine_hash_entry *rt,
Prelocking_strategy *prelocking_strategy, bool has_prelocking_list,
Open_table_context *ot_ctx, bool *need_prelocking,
bool *routine_modifies_data) {
*routine_modifies_data = false;
DBUG_TRACE;
switch (rt->type()) {
case Sroutine_hash_entry::FUNCTION:
case Sroutine_hash_entry::PROCEDURE: {
sp_head *sp;
/*
Try to get MDL lock on the routine.
Note that we do not take locks on top-level CALLs as this can
lead to a deadlock. Not locking top-level CALLs does not break
the binlog as only the statements in the called procedure show
up there, not the CALL itself.
*/
if (rt != prelocking_ctx->sroutines_list.first ||
rt->type() != Sroutine_hash_entry::PROCEDURE) {
MDL_request mdl_request;
MDL_key mdl_key;
if (rt->type() == Sroutine_hash_entry::FUNCTION)
dd::Function::create_mdl_key(rt->db(), rt->name(), &mdl_key);
else
dd::Procedure::create_mdl_key(rt->db(), rt->name(), &mdl_key);
MDL_REQUEST_INIT_BY_KEY(&mdl_request, &mdl_key, MDL_SHARED,
MDL_TRANSACTION);
/*
Waiting for a conflicting metadata lock to go away may
lead to a deadlock, detected by MDL subsystem.
If possible, we try to resolve such deadlocks by releasing all
metadata locks and restarting the pre-locking process.
To prevent the error from polluting the Diagnostics Area
in case of successful resolution, install a special error
handler for ER_LOCK_DEADLOCK error.
*/
MDL_deadlock_handler mdl_deadlock_handler(ot_ctx);
thd->push_internal_handler(&mdl_deadlock_handler);
bool result =
thd->mdl_context.acquire_lock(&mdl_request, ot_ctx->get_timeout());
thd->pop_internal_handler();
if (result) return true;
DEBUG_SYNC(thd, "after_shared_lock_pname");
/* Ensures the routine is up-to-date and cached, if exists. */
if (sp_cache_routine(thd, rt, has_prelocking_list, &sp)) return true;
/* Remember the version of the routine in the parse tree. */
if (check_and_update_routine_version(thd, rt, sp)) return true;
/* 'sp' is NULL when there is no such routine. */
if (sp) {
*routine_modifies_data = sp->modifies_data();
if (!has_prelocking_list)
prelocking_strategy->handle_routine(thd, prelocking_ctx, rt, sp,
need_prelocking);
}
} else {
/*
If it's a top level call, just make sure we have a recent
version of the routine, if it exists.
Validating routine version is unnecessary, since CALL
does not affect the prepared statement prelocked list.
*/
if (sp_cache_routine(thd, rt, false, &sp)) return true;
}
} break;
case Sroutine_hash_entry::TRIGGER:
/**
We add trigger entries to lex->sroutines_list, but we don't
load them here. The trigger entry is only used when building
a transitive closure of objects used in a statement, to avoid
adding to this closure objects that are used in the trigger more
than once.
E.g. if a trigger trg refers to table t2, and the trigger table t1
is used multiple times in the statement (say, because it's used in
function f1() twice), we will only add t2 once to the list of
tables to prelock.
We don't take metadata locks on triggers either: they are protected
by a respective lock on the table, on which the trigger is defined.
The only two cases which give "trouble" are SHOW CREATE TRIGGER
and DROP TRIGGER statements. For these, statement syntax doesn't
specify the table on which this trigger is defined, so we have
to make a "dirty" read in the data dictionary to find out the
table name. Once we discover the table name, we take a metadata
lock on it, and this protects all trigger operations.
Of course the table, in theory, may disappear between the dirty
read and metadata lock acquisition, but in that case we just return
a run-time error.
Grammar of other trigger DDL statements (CREATE, DROP) requires
the table to be specified explicitly, so we use the table metadata
lock to protect trigger metadata in these statements. Similarly, in
DML we always use triggers together with their tables, and thus don't
need to take separate metadata locks on them.
*/
break;
case Sroutine_hash_entry::FK_TABLE_ROLE_PARENT_CHECK:
case Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_CHECK:
case Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_UPDATE:
case Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_DELETE: {
if (thd->locked_tables_mode == LTM_NONE) {
MDL_request mdl_request;
/*
Adjust metadata lock type according to the table's role in the
FK relationship. Also acquire stronger locks when we are locking
on behalf of LOCK TABLES.
*/
enum_mdl_type mdl_lock_type;
bool executing_LT = (prelocking_ctx->sql_command == SQLCOM_LOCK_TABLES);
if (rt->type() == Sroutine_hash_entry::FK_TABLE_ROLE_PARENT_CHECK ||
rt->type() == Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_CHECK) {
mdl_lock_type =
(executing_LT ? MDL_SHARED_READ_ONLY : MDL_SHARED_READ);
} else {
mdl_lock_type =
(executing_LT ? MDL_SHARED_NO_READ_WRITE : MDL_SHARED_WRITE);
}
MDL_REQUEST_INIT_BY_PART_KEY(&mdl_request, MDL_key::TABLE,
rt->part_mdl_key(),
rt->part_mdl_key_length(), rt->db_length(),
mdl_lock_type, MDL_TRANSACTION);
MDL_deadlock_handler mdl_deadlock_handler(ot_ctx);
thd->push_internal_handler(&mdl_deadlock_handler);
bool result =
thd->mdl_context.acquire_lock(&mdl_request, ot_ctx->get_timeout());
thd->pop_internal_handler();
if (result) return true;
} else {
/*
This function is called only if we are not in prelocked mode
already. So we must be handling statement executed under
LOCK TABLES in this case.
*/
assert(thd->locked_tables_mode == LTM_LOCK_TABLES);
/*
Even though LOCK TABLES tries to automatically lock parent and child
tables which might be necessary for foreign key checks/actions, there
are some cases when we might miss them. So it is better to check that
we have appropriate metadata lock explicitly and error out if not.
Some examples of problematic cases are:
*) We are executing DELETE FROM t1 under LOCK TABLES t1 READ
and table t1 is a parent in a foreign key.
In this case error about inappropriate lock on t1 will be
reported at later stage than prelocking set is built.
So we can't assume/assert that we have proper lock on the
corresponding child table here.
*) Table t1 has a trigger, which contains DELETE FROM t2 and
t2 is participating in FK as parent. In such situation
LOCK TABLE t1 WRITE will lock t2 for write implicitly
so both updates and delete on t2 will be allowed. However,
t3 will be locked only in a way as if only deletes from
t2 were allowed.
*) Prelocking list has been built earlier. Both child and parent
definitions might have changed since this time so at LOCK TABLES
time FK which corresponds to this element of prelocked set
might be no longer around. In theory, we might be processing
statement which is not marked as requiring prelocked set invalidation
(and thus ignoring table version mismatches) or tables might be missing
and this error can be suppressed. In such case we might not have
appropriate metadata lock on our child/parent table.
*/
if (rt->type() == Sroutine_hash_entry::FK_TABLE_ROLE_PARENT_CHECK ||
rt->type() == Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_CHECK) {
if (!thd->mdl_context.owns_equal_or_stronger_lock(
MDL_key::TABLE, rt->db(), rt->name(), MDL_SHARED_READ_ONLY)) {
my_error(ER_TABLE_NOT_LOCKED, MYF(0), rt->name());
return true;
}
} else {
if (!thd->mdl_context.owns_equal_or_stronger_lock(
MDL_key::TABLE, rt->db(), rt->name(),
MDL_SHARED_NO_READ_WRITE)) {
my_error(ER_TABLE_NOT_LOCKED_FOR_WRITE, MYF(0), rt->name());
return true;
}
}
}
if (rt->type() == Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_UPDATE ||
rt->type() == Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_DELETE) {
/*
In order to continue building prelocked set or validating
prelocked set which already has been built we need to get
access to table's TABLE_SHARE.
Getting unused TABLE object is more scalable that going
directly for the TABLE_SHARE. If there are no unused TABLE
object we might get at least pointer to the TABLE_SHARE
from the table cache.
Note that under LOCK TABLES we can't rely on that table is
going to be in THD::open_tables list, as LOCK TABLES only
pre-acquires metadata locks on FK tables but doesn't
pre-open them.
TODO: Perhaps we should give it a try as it can be more
scalability friendly.
*/
Table_cache *tc = table_cache_manager.get_cache(thd);
TABLE *table;
TABLE_SHARE *share;
tc->lock();
table = tc->get_table(thd, rt->part_mdl_key(),
rt->part_mdl_key_length(), &share);
if (table) {
assert(table->s == share);
/*
Don't check if TABLE_SHARE::version matches version of tables
previously opened by this statement. It might be problematic
under LOCK TABLES and possible version difference can't affect
FK-related part of prelocking set.
*/
tc->unlock();
} else if (share) {
/*
TODO: If we constantly hit this case it would harm scalability...
Perhaps we need to create new unused TABLE instance in this
case.
*/
mysql_mutex_lock(&LOCK_open);
tc->unlock();
share->increment_ref_count();
mysql_mutex_unlock(&LOCK_open);
/*
Again, when building part of prelocking set related to foreign keys
we can ignore fact that TABLE_SHARE::version is old.
*/
} else {
tc->unlock();
/*
If we are validating existing prelocking set then the table
might have been dropped. We suppress this error in this case.
Prelocking set will be either invalidated, or error will be
reported the parent table is accessed.
TODO: Perhaps we need to use get_table_share_with_discover()
here but it gets complicated under LOCK TABLES.
*/
No_such_table_error_handler no_such_table_handler;
thd->push_internal_handler(&no_such_table_handler);
mysql_mutex_lock(&LOCK_open);
share = get_table_share(thd, rt->db(), rt->name(), rt->part_mdl_key(),
rt->part_mdl_key_length(), true);
mysql_mutex_unlock(&LOCK_open);
thd->pop_internal_handler();
if (!share && no_such_table_handler.safely_trapped_errors()) {
break; // Jump out switch without error.
}
if (!share) {
return true;
}
if (share->is_view) {
/*
Eeek! Somebody replaced the child table with a view. This can
happen only when we are validating existing prelocked set.
Parent either have been dropped or its definition has been
changed. In either case our child table won't be accessed
through the foreign key.
*/
assert(has_prelocking_list);
mysql_mutex_lock(&LOCK_open);
release_table_share(share);
mysql_mutex_unlock(&LOCK_open);
if (ask_to_reprepare(thd)) return true;
break; // Jump out switch without error.
}
}
auto release_table_lambda = [thd](TABLE *tab) {
release_or_close_table(thd, tab);
};
std::unique_ptr<TABLE, decltype(release_table_lambda)>
release_table_guard(table, release_table_lambda);
/*
We need to explicitly release TABLE_SHARE only if we don't
have TABLE object.
*/
auto release_share_lambda = [](TABLE_SHARE *tsh) {
mysql_mutex_lock(&LOCK_open);
release_table_share(tsh);
mysql_mutex_unlock(&LOCK_open);
};
std::unique_ptr<TABLE_SHARE, decltype(release_share_lambda)>
release_share_guard((table ? nullptr : share),
release_share_lambda);
/*
We need to maintain versioning of the prelocked tables since this
is needed for correct handling of prepared statements to catch
situations where a prelocked table (which is added to the prelocked
set during PREPARE) is changed between repeated executions of the
prepared statement.
*/
int64 share_version = share->get_table_ref_version();
if (rt->m_cache_version != share_version) {
/*
Version of the cached table share is different from the
previous execution of the prepared statement, and it is
unacceptable for this SQLCOM.
*/
if (ask_to_reprepare(thd)) return true;
/* Always maintain the latest cache version. */
rt->m_cache_version = share_version;
}
/*
If the child may be affected by update/delete and is in a read only
schema, we must reject the statement.
*/
if (check_schema_readonly(thd, rt->db())) {
my_error(ER_SCHEMA_READ_ONLY, MYF(0), rt->db());
return true;
}
if (!has_prelocking_list) {
bool is_update =
(rt->type() == Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_UPDATE);
bool is_delete =
(rt->type() == Sroutine_hash_entry::FK_TABLE_ROLE_CHILD_DELETE);
process_table_fks(thd, prelocking_ctx, share, false, is_update,
is_delete, rt->belong_to_view, need_prelocking);
}
}
} break;
default:
/* Impossible type value. */
assert(0);
}
return false;
}
/**
Handle table list element by obtaining metadata lock, opening table or view
and, if prelocking strategy prescribes so, extending the prelocking set with
tables and routines used by it.
@param[in] thd Thread context.
@param[in] lex LEX structure for statement.
@param[in] tables Table list element to be processed.
@param[in,out] counter Number of tables which are open.
@param[in] prelocking_strategy Strategy which specifies how the
prelocking set should be extended
when table or view is processed.
@param[in] has_prelocking_list Indicates that prelocking set/list for
this statement has already been built.
@param[in] ot_ctx Context used to recover from a failed
open_table() attempt.
@retval false Success.
@retval true Error, reported unless there is a chance to recover from it.
*/
static bool open_and_process_table(THD *thd, LEX *lex, Table_ref *const tables,
uint *counter,
Prelocking_strategy *prelocking_strategy,
bool has_prelocking_list,
Open_table_context *ot_ctx) {
bool error = false;
bool safe_to_ignore_table = false;
DBUG_TRACE;
DEBUG_SYNC(thd, "open_and_process_table");
/*
Ignore placeholders for unnamed derived tables, as they are fully resolved
by the optimizer.
*/
if (tables->is_derived() || tables->is_table_function() ||
tables->is_recursive_reference())
goto end;
assert(!tables->common_table_expr());
/*
If this Table_ref object is a placeholder for an information_schema
table, create a temporary table to represent the information_schema
table in the query. Do not fill it yet - will be filled during
execution.
*/
if (tables->schema_table) {
/*
Since we no longer set Table_ref::schema_table/table for table
list elements representing mergeable view, we can't meet a table
list element which represent information_schema table and a view
at the same time. Otherwise, acquiring metadata lock om the view
would have been necessary.
*/
assert(!tables->is_view());
if (!mysql_schema_table(thd, lex, tables) &&
!check_and_update_table_version(thd, tables, tables->table->s)) {
goto end;
}
error = true;
goto end;
}
DBUG_PRINT("tcache", ("opening table: '%s'.'%s' item: %p", tables->db,
tables->table_name, tables));
(*counter)++;
/*
Not a placeholder so this must be a base/temporary table or view.
Open it:
*/
/*
A Table_ref object may have an associated open TABLE object
(Table_ref::table is not NULL) if it represents a pre-opened temporary
table, or is a materialized view. (Derived tables are not handled here).
*/
assert(tables->table == nullptr || is_temporary_table(tables) ||
(tables->is_view() && tables->uses_materialization()));
/*
OT_TEMPORARY_ONLY means that we are in CREATE TEMPORARY TABLE statement.
Also such table list element can't correspond to prelocking placeholder
or to underlying table of merge table.
So existing temporary table should have been preopened by this moment
and we can simply continue without trying to open temporary or base table.
*/
assert(tables->open_type != OT_TEMPORARY_ONLY ||
(tables->open_strategy && !tables->prelocking_placeholder &&
tables->parent_l == nullptr));
if (tables->open_type == OT_TEMPORARY_ONLY || is_temporary_table(tables)) {
// Already "open", no action required
} else if (tables->prelocking_placeholder) {
/*
For the tables added by the pre-locking code, attempt to open
the table but fail silently if the table does not exist.
The real failure will occur when/if a statement attempts to use
that table.
*/
No_such_table_error_handler no_such_table_handler;
thd->push_internal_handler(&no_such_table_handler);
/*
We're opening a table from the prelocking list.
Since this table list element might have been added after pre-opening
of temporary tables we have to try to open temporary table for it.
We can't simply skip this table list element and postpone opening of
temporary tabletill the execution of substatement for several reasons:
- Temporary table can be a MERGE table with base underlying tables,
so its underlying tables has to be properly open and locked at
prelocking stage.
- Temporary table can be a MERGE table and we might be in PREPARE
phase for a prepared statement. In this case it is important to call
HA_ATTACH_CHILDREN for all merge children.
This is necessary because merge children remember "TABLE_SHARE ref type"
and "TABLE_SHARE def version" in the HA_ATTACH_CHILDREN operation.
If HA_ATTACH_CHILDREN is not called, these attributes are not set.
Then, during the first EXECUTE, those attributes need to be updated.
That would cause statement re-preparing (because changing those
attributes during EXECUTE is caught by THD::m_reprepare_observers).
The problem is that since those attributes are not set in merge
children, another round of PREPARE will not help.
*/
error = open_temporary_table(thd, tables);
if (!error && !tables->table) error = open_table(thd, tables, ot_ctx);
thd->pop_internal_handler();
safe_to_ignore_table = no_such_table_handler.safely_trapped_errors();
} else if (tables->parent_l && (thd->open_options & HA_OPEN_FOR_REPAIR)) {
/*
Also fail silently for underlying tables of a MERGE table if this
table is opened for CHECK/REPAIR TABLE statement. This is needed
to provide complete list of problematic underlying tables in
CHECK/REPAIR TABLE output.
*/
Repair_mrg_table_error_handler repair_mrg_table_handler;
thd->push_internal_handler(&repair_mrg_table_handler);
error = open_temporary_table(thd, tables);
if (!error && !tables->table) error = open_table(thd, tables, ot_ctx);
thd->pop_internal_handler();
safe_to_ignore_table = repair_mrg_table_handler.safely_trapped_errors();
} else {
if (tables->parent_l) {
/*
Even if we are opening table not from the prelocking list we
still might need to look for a temporary table if this table
list element corresponds to underlying table of a merge table.
*/
error = open_temporary_table(thd, tables);
}
if (!error && (tables->is_view() || tables->table == nullptr))
error = open_table(thd, tables, ot_ctx);
}
if (error) {
if (!ot_ctx->can_recover_from_failed_open() && safe_to_ignore_table) {
DBUG_PRINT("info", ("open_table: ignoring table '%s'.'%s'", tables->db,
tables->alias));
error = false;
}
goto end;
}
// Do specific processing for a view, and skip actions that apply to tables
if (tables->is_view()) {
// Views do not count as tables
(*counter)--;
/*
tables->next_global list consists of two parts:
1) Query tables and underlying tables of views.
2) Tables used by all stored routines that this statement invokes on
execution.
We need to know where the bound between these two parts is. If we've
just opened a view, which was the last table in part #1, and it
has added its base tables after itself, adjust the boundary pointer
accordingly.
*/
if (lex->query_tables_own_last == &(tables->next_global) &&
tables->view_query()->query_tables)
lex->query_tables_own_last = tables->view_query()->query_tables_last;
/*
Let us free memory used by 'sroutines' hash here since we never
call destructor for this LEX.
*/
tables->view_query()->sroutines.reset();
goto process_view_routines;
}
/*
Special types of open can succeed but still don't set
Table_ref::table to anything.
*/
if (tables->open_strategy && !tables->table) goto end;
/*
If we are not already in prelocked mode and extended table list is not
yet built we might have to build the prelocking set for this statement.
Since currently no prelocking strategy prescribes doing anything for
tables which are only read, we do below checks only if table is going
to be changed.
*/
if (thd->locked_tables_mode <= LTM_LOCK_TABLES && !has_prelocking_list &&
tables->lock_descriptor().type >= TL_WRITE_ALLOW_WRITE) {
bool need_prelocking = false;
Table_ref **save_query_tables_last = lex->query_tables_last;
/*
Extend statement's table list and the prelocking set with
tables and routines according to the current prelocking
strategy.
For example, for DML statements we need to add tables and routines
used by triggers which are going to be invoked for this element of
table list and also add tables required for handling of foreign keys.
*/
error =
prelocking_strategy->handle_table(thd, lex, tables, &need_prelocking);
if (need_prelocking && !lex->requires_prelocking())
lex->mark_as_requiring_prelocking(save_query_tables_last);
if (error) goto end;
}
/* Check and update metadata version of a base table. */
error = check_and_update_table_version(thd, tables, tables->table->s);
if (error) goto end;
/*
After opening a MERGE table add the children to the query list of
tables, so that they are opened too.
Note that placeholders don't have the handler open.
*/
/* MERGE tables need to access parent and child TABLE_LISTs. */
assert(tables->table->pos_in_table_list == tables);
/* Non-MERGE tables ignore this call. */
if (tables->table->db_stat &&
tables->table->file->ha_extra(HA_EXTRA_ADD_CHILDREN_LIST)) {
error = true;
goto end;
}
process_view_routines:
assert((tables->is_view() &&
(tables->uses_materialization() || tables->table == nullptr)) ||
(!tables->is_view()));
/*
Again we may need cache all routines used by this view and add
tables used by them to table list.
*/
if (tables->is_view() && thd->locked_tables_mode <= LTM_LOCK_TABLES &&
!has_prelocking_list) {
bool need_prelocking = false;
Table_ref **save_query_tables_last = lex->query_tables_last;
error =
prelocking_strategy->handle_view(thd, lex, tables, &need_prelocking);
if (need_prelocking && !lex->requires_prelocking())
lex->mark_as_requiring_prelocking(save_query_tables_last);
if (error) goto end;
}
end:
return error;
}
namespace {
struct schema_hash {
size_t operator()(const Table_ref *table) const {
return std::hash<std::string>()(std::string(table->db, table->db_length));
}
};
struct schema_key_equal {
bool operator()(const Table_ref *a, const Table_ref *b) const {
return a->db_length == b->db_length &&
memcmp(a->db, b->db, a->db_length) == 0;
}
};
} // namespace
/**
Run the server hook called "before_dml". This is a hook originated from
replication that allow server plugins to execute code before any DML
instruction is executed.
In case of negative outcome, it will set my_error to
ER_BEFORE_DML_VALIDATION_ERROR
@param thd Thread context
@return hook outcome
@retval 0 Everything is fine
@retval !=0 Error in the outcome of the hook.
*/
int run_before_dml_hook(THD *thd) {
int out_value = 0;
TX_TRACKER_GET(tst);
tst->add_trx_state(thd, TX_STMT_DML);
(void)RUN_HOOK(transaction, before_dml, (thd, out_value));
if (out_value) {
tst->clear_trx_state(thd, TX_STMT_DML);
my_error(ER_BEFORE_DML_VALIDATION_ERROR, MYF(0));
}
return out_value;
}
/**
Check whether a table being opened is a temporary table.
@param table table being opened
@return true if a table is temporary table, else false
*/
static inline bool is_temporary_table_being_opened(const Table_ref *table) {
return table->open_type == OT_TEMPORARY_ONLY ||
(table->open_type == OT_TEMPORARY_OR_BASE &&
is_temporary_table(table));
}
/**
Acquire IX metadata locks on tablespace names used by LOCK
TABLES or by a DDL statement.
@note That the tablespace MDL locks are taken only after locks
on tables are acquired. So it is recommended to maintain this
same lock order across the server. It is very easy to break the
this lock order if we invoke acquire_locks() with list of MDL
requests which contain both MDL_key::TABLE and
MDL_key::TABLESPACE. We would end-up in deadlock then.
@param thd Thread context.
@param tables_start Start of list of tables on which locks
should be acquired.
@param tables_end End of list of tables.
@param lock_wait_timeout Seconds to wait before timeout.
@param flags Bitmap of flags to modify how the
tables will be open, see open_table()
description for details.
@retval true Failure (e.g. connection was killed)
@retval false Success.
*/
bool get_and_lock_tablespace_names(THD *thd, Table_ref *tables_start,
Table_ref *tables_end,
ulong lock_wait_timeout, uint flags) {
// If this is a DISCARD or IMPORT TABLESPACE command (indicated by
// the THD:: tablespace_op flag), we skip this phase, because these
// commands are only used for file-per-table tablespaces, which we
// do not lock. We also skip this phase if we are within the
// context of a FLUSH TABLE WITH READ LOCK or FLUSH TABLE FOR EXPORT
// statement, indicated by the MYSQL_OPEN_SKIP_SCOPED_MDL_LOCK flag.
if (flags & MYSQL_OPEN_SKIP_SCOPED_MDL_LOCK || thd_tablespace_op(thd))
return false;
// Add tablespace names used under partition/subpartition definitions.
Tablespace_hash_set tablespace_set(PSI_INSTRUMENT_ME);
if ((thd->lex->sql_command == SQLCOM_CREATE_TABLE ||
thd->lex->sql_command == SQLCOM_ALTER_TABLE) &&
fill_partition_tablespace_names(thd->work_part_info, &tablespace_set))
return true;
// The first step is to loop over the tables, make sure we have
// locked the names, and then get hold of the tablespace names from
// the data dictionary.
Table_ref *table;
for (table = tables_start; table && table != tables_end;
table = table->next_global) {
// Consider only non-temporary tables. The if clauses below have the
// following meaning:
//
// !MDL_SHARED_READ_ONLY Not a LOCK TABLE ... READ.
// In that case, tables will not
// be altered, created or dropped,
// so no need to IX lock the
// tablespace.
// is_ddl_or...request() || ...FOR_CREATE Request for a strong DDL or
// LOCK TABLES type lock, or a
// table to be created.
// !OT_TEMPORARY_ONLY Not a user defined tmp table.
// !(OT_TEMPORARY_OR_BASE && is_temp...()) Not a pre-opened tmp table.
if (table->mdl_request.type != MDL_SHARED_READ_ONLY &&
(table->mdl_request.is_ddl_or_lock_tables_lock_request() ||
table->open_strategy == Table_ref::OPEN_FOR_CREATE) &&
!is_temporary_table_being_opened(table) && !table->is_system_view) {
// We have basically three situations here:
//
// 1. Lock only the target tablespace name and tablespace
// names that are used by partitions (e.g. CREATE TABLE
// explicitly specifying the tablespace names).
// 2. Lock only the existing tablespace name and tablespace
// names that are used by partitions (e.g. ALTER TABLE t
// ADD COLUMN ... where t is defined in some tablespace s.
// 3. Lock both the target and the existing tablespace names
// along with tablespace names used by partitions. (e.g.
// ALTER TABLE t TABLESPACE s2, where t is defined in
// some tablespace s)
if (table->target_tablespace_name.length > 0) {
tablespace_set.insert(table->target_tablespace_name.str);
}
// No need to try this for tables to be created since they are not
// yet present in the dictionary.
if (table->open_strategy != Table_ref::OPEN_FOR_CREATE) {
// Assert that we have an MDL lock on the table name. Needed to read
// the dictionary safely.
assert(thd->mdl_context.owns_equal_or_stronger_lock(
MDL_key::TABLE, table->db, table->table_name, MDL_SHARED));
/*
Add names of tablespaces used by table or by its
partitions/subpartitions. Lookup data dictionary to get
the information.
*/
if (dd::fill_table_and_parts_tablespace_names(
thd, table->db, table->table_name, &tablespace_set))
return true;
}
}
} // End of for(;;)
/*
After we have identified the tablespace names, we iterate
over the names and acquire IX locks on each of them.
*/
if (thd->lex->sql_command == SQLCOM_DROP_DB) {
/*
In case of DROP DATABASE we might have to lock many thousands of
tablespaces in extreme cases. Ensure that we don't hold memory used
by corresponding MDL_requests after locks have been acquired to
reduce memory usage by DROP DATABASE in such cases.
*/
MEM_ROOT mdl_reqs_root(key_memory_rm_db_mdl_reqs_root, MEM_ROOT_BLOCK_SIZE);
if (lock_tablespace_names(thd, &tablespace_set, lock_wait_timeout,
&mdl_reqs_root))
return true;
} else {
if (lock_tablespace_names(thd, &tablespace_set, lock_wait_timeout,
thd->mem_root))
return true;
}
return false;
}
/**
Acquire "strong" (SRO, SNW, SNRW) metadata locks on tables used by
LOCK TABLES or by a DDL statement.
Acquire lock "S" on table being created in CREATE TABLE statement.
@note Under LOCK TABLES, we can't take new locks, so use
open_tables_check_upgradable_mdl() instead.
@param thd Thread context.
@param tables_start Start of list of tables on which locks
should be acquired.
@param tables_end End of list of tables.
@param lock_wait_timeout Seconds to wait before timeout.
@param flags Bitmap of flags to modify how the tables will be
open, see open_table() description for details.
@param schema_reqs When non-nullptr, pointer to array in which
pointers to MDL requests for acquired schema
locks to be stored. It is guaranteed that
each schema will be present in this array
only once.
@retval false Success.
@retval true Failure (e.g. connection was killed)
*/
bool lock_table_names(THD *thd, Table_ref *tables_start, Table_ref *tables_end,
ulong lock_wait_timeout, uint flags,
Prealloced_array<MDL_request *, 1> *schema_reqs) {
MDL_request_list mdl_requests;
Table_ref *table;
MDL_request global_request;
MDL_request backup_lock_request;
malloc_unordered_set<Table_ref *, schema_hash, schema_key_equal> schema_set(
PSI_INSTRUMENT_ME);
bool need_global_read_lock_protection = false;
bool acquire_backup_lock = false;
/*
This function is not supposed to be used under LOCK TABLES normally.
Instead open_tables_check_upgradable_mdl() or some other function
checking if we have tables locked in proper mode should be used.
The exception to this rule is RENAME TABLES code which uses this call
to "upgrade" metadata lock on tables renamed along with acquiring
exclusive locks on target table names, after checking that tables
renamed are properly locked.
*/
assert(!thd->locked_tables_mode ||
thd->lex->sql_command == SQLCOM_RENAME_TABLE);
// Phase 1: Iterate over tables, collect set of unique schema names, and
// construct a list of requests for table MDL locks.
for (table = tables_start; table && table != tables_end;
table = table->next_global) {
if (is_temporary_table_being_opened(table)) {
continue;
}
if (!table->mdl_request.is_ddl_or_lock_tables_lock_request() &&
table->open_strategy != Table_ref::OPEN_FOR_CREATE) {
continue;
} else {
/*
MDL_request::is_ddl_or_lock_tables_lock_request() returns true for
DDL and LOCK TABLES statements. Since there isn't a way on MDL API level
to determine whether a lock being acquired is requested as part of
handling the statement LOCK TABLES, such check will be done by comparing
a value of lex->sql_command against the constant SQLCOM_LOCK_TABLES.
Also we shouldn't acquire IX backup lock in case a table being opened
with requested MDL_SHARED_READ_ONLY lock. For example, such use case
takes place when FLUSH PRIVILEGES executed.
*/
if (thd->lex->sql_command != SQLCOM_LOCK_TABLES &&
table->mdl_request.type != MDL_SHARED_READ_ONLY)
acquire_backup_lock = true;
}
if (table->mdl_request.type != MDL_SHARED_READ_ONLY) {
/* Write lock on normal tables is not allowed in a read only transaction.
*/
if (thd->tx_read_only) {
my_error(ER_CANT_EXECUTE_IN_READ_ONLY_TRANSACTION, MYF(0));
return true;
}
if (!(flags & MYSQL_OPEN_SKIP_SCOPED_MDL_LOCK)) {
schema_set.insert(table);
}
need_global_read_lock_protection = true;
}
mdl_requests.push_front(&table->mdl_request);
}
// Phase 2: Iterate over the schema set, add an IX lock for each
// schema name.
if (!(flags & MYSQL_OPEN_SKIP_SCOPED_MDL_LOCK) && !mdl_requests.is_empty()) {
/*
Scoped locks: Take intention exclusive locks on all involved
schemas.
*/
for (const Table_ref *table_l : schema_set) {
MDL_request *schema_request = new (thd->mem_root) MDL_request;
if (schema_request == nullptr) return true;
MDL_REQUEST_INIT(schema_request, MDL_key::SCHEMA, table_l->db, "",
MDL_INTENTION_EXCLUSIVE, MDL_TRANSACTION);
mdl_requests.push_front(schema_request);
if (schema_reqs) schema_reqs->push_back(schema_request);
}
if (need_global_read_lock_protection) {
/*
Protect this statement against concurrent global read lock
by acquiring global intention exclusive lock with statement
duration.
*/
if (thd->global_read_lock.can_acquire_protection()) return true;
MDL_REQUEST_INIT(&global_request, MDL_key::GLOBAL, "", "",
MDL_INTENTION_EXCLUSIVE, MDL_STATEMENT);
mdl_requests.push_front(&global_request);
}
}
if (acquire_backup_lock) {
MDL_REQUEST_INIT(&backup_lock_request, MDL_key::BACKUP_LOCK, "", "",
MDL_INTENTION_EXCLUSIVE, MDL_TRANSACTION);
mdl_requests.push_front(&backup_lock_request);
}
// Phase 3: Acquire the locks which have been requested so far.
if (thd->mdl_context.acquire_locks(&mdl_requests, lock_wait_timeout))
return true;
/*
Now when we have protection against concurrent change of read_only
option we can safely re-check its value. Skip the check for
FLUSH TABLES ... WITH READ LOCK and FLUSH TABLES ... FOR EXPORT
as they are not supposed to be affected by read_only modes.
*/
if (need_global_read_lock_protection &&
!(flags & MYSQL_OPEN_SKIP_SCOPED_MDL_LOCK) &&
!(flags & MYSQL_LOCK_IGNORE_GLOBAL_READ_ONLY) &&
check_readonly(thd, true))
return true;
// Check schema read only for all schemas.
for (const Table_ref *table_l : schema_set)
if (check_schema_readonly(thd, table_l->db)) return true;
/*
Phase 4: Lock tablespace names. This cannot be done as part
of the previous phases, because we need to read the
dictionary to get hold of the tablespace name, and in order
to do this, we must have acquired a lock on the table.
*/
return get_and_lock_tablespace_names(thd, tables_start, tables_end,
lock_wait_timeout, flags);
}
/**
Check for upgradable (SNW, SNRW) metadata locks on tables to be opened
for a DDL statement. Under LOCK TABLES, we can't take new locks, so we
must check if appropriate locks were pre-acquired.
@param thd Thread context.
@param tables_start Start of list of tables on which upgradable locks
should be searched for.
@param tables_end End of list of tables.
@retval false Success.
@retval true Failure (e.g. connection was killed)
*/
static bool open_tables_check_upgradable_mdl(THD *thd, Table_ref *tables_start,
Table_ref *tables_end) {
Table_ref *table;
assert(thd->locked_tables_mode);
for (table = tables_start; table && table != tables_end;
table = table->next_global) {
if (!table->mdl_request.is_ddl_or_lock_tables_lock_request() ||
is_temporary_table_being_opened(table)) {
continue;
}
if (table->mdl_request.type == MDL_SHARED_READ_ONLY) {
if (!thd->mdl_context.owns_equal_or_stronger_lock(
MDL_key::TABLE, table->db, table->table_name,
MDL_SHARED_READ_ONLY)) {
my_error(ER_TABLE_NOT_LOCKED, MYF(0), table->table_name);
return true;
}
} else {
/*
We don't need to do anything about the found TABLE instance as it
will be handled later in open_tables(), we only need to check that
an upgradable lock is already acquired. When we enter LOCK TABLES
mode, SNRW locks are acquired before all other locks. So if under
LOCK TABLES we find that there is TABLE instance with upgradeable
lock, all other instances of TABLE for the same table will have the
same ticket.
Note that this works OK even for CREATE TABLE statements which
request X type of metadata lock. This is because under LOCK TABLES
such statements don't create the table but only check if it exists
or, in most complex case, only insert into it.
Thus SNRW lock should be enough.
Note that find_table_for_mdl_upgrade() will report an error if
no suitable ticket is found.
*/
if (!find_table_for_mdl_upgrade(thd, table->db, table->table_name, false))
return true;
}
}
return false;
}
/**
Iterate along a list of tables and acquire BACKUP LOCK in shared mode
in case a strong MDL request (DDL/LOCK TABLES-type) was specified
for a table.
@param[in] thd Thread context.
@param[in] tables_start Pointer to a start of a list of tables to iterate
@param[in] tables_end Pointer to a end of a list of tables where to stop
@return false on success, true on error.
*/
static bool acquire_backup_lock_in_lock_tables_mode(THD *thd,
Table_ref *tables_start,
Table_ref *tables_end) {
Table_ref *table;
assert(thd->locked_tables_mode);
for (table = tables_start; table && table != tables_end;
table = table->next_global) {
if (is_temporary_table_being_opened(table)) continue;
if (table->mdl_request.is_ddl_or_lock_tables_lock_request() &&
table->mdl_request.type != MDL_SHARED_READ_ONLY)
return acquire_shared_backup_lock(thd, thd->variables.lock_wait_timeout);
}
return false;
}
/**
Check if this is a DD table used under a I_S view then request InnoDB to
do non-locking reads on the table.
@param thd Thread
@param[in] tl Table_ref pointing to table being checked.
@return false on success, true on error.
*/
static bool set_non_locking_read_for_IS_view(THD *thd, Table_ref *tl) {
TABLE *tbl = tl->table;
// Not a system view.
if (!(tbl && tbl->file && tl->referencing_view &&
tl->referencing_view->is_system_view))
return false;
// Allow I_S system views to be locked by LOCK TABLE command.
if (thd->lex->sql_command != SQLCOM_LOCK_TABLES &&
tl->lock_descriptor().type >= TL_READ_NO_INSERT) {
my_error(ER_IS_QUERY_INVALID_CLAUSE, MYF(0), "FOR UPDATE");
return true;
}
/* Convey to InnoDB (the DD table's engine) to do non-locking reads.
It is assumed that all the tables used by I_S views are
always a DD table. If this is not true, then we might
need to invoke dd::Dictionary::is_dd_tablename() to make sure.
*/
if (tbl->db_stat && tbl->file->ha_extra(HA_EXTRA_NO_READ_LOCKING)) {
// Handler->ha_extra() for innodb does not fail ever as of now.
// In case it is made to fail sometime later, we need to think
// about the kind of error to be report to user.
assert(0);
return true;
}
return false;
}
// Check if given Table_ref is a acl table and is being read and not
bool is_acl_table_in_non_LTM(const Table_ref *tl,
enum enum_locked_tables_mode ltm) {
TABLE *table = tl->table;
/**
We ignore use of ACL table,
- Under LOCK TABLE modes.
- Under system view. E.g., I_S.ROLE_* uses CTE where they use
TL_READ_DEFAULT for ACL tables. We ignore them.
- If the Table_ref is used by optimizer as placeholder.
*/
return (!tl->is_placeholder() && table->db_stat &&
table->s->table_category == TABLE_CATEGORY_ACL_TABLE &&
ltm != LTM_LOCK_TABLES && ltm != LTM_PRELOCKED_UNDER_LOCK_TABLES);
}
/**
Check if this is a ACL table is requested for read and
then request InnoDB to do non-locking reads on the
table.
@param thd Thread
@param[in] tl Table_ref pointing to table being checked.
@param[in] issue_warning If true, issue warning irrespective of
isolation level.
@return false on success, true on error.
*/
static bool set_non_locking_read_for_ACL_table(THD *thd, Table_ref *tl,
const bool &issue_warning) {
TABLE *tbl = tl->table;
/*
Request InnoDB to skip SE row locks if:
- We have a ACL table name.
- Lock type is TL_READ_DEFAULT or
- Lock type is TL_READ_HIGH_PRIORITY.
Note:
- We do this for all isolation modes as InnoDB sometimes acquires row
locks even for modes other than serializable, e.g. to ensure correct
binlogging or just to play safe.
- Checking only for TL_READ_DEFAULT and TL_READ_HIGH_PRIORITY allows to
filter out all special non-SELECT cases which require locking like
ALTER TABLE, ACL DDL and so on
*/
if (is_acl_table_in_non_LTM(tl, thd->locked_tables_mode) &&
(tl->lock_descriptor().type == TL_READ_DEFAULT ||
tl->lock_descriptor().type == TL_READ_HIGH_PRIORITY)) {
if (tbl->file->ha_extra(HA_EXTRA_NO_READ_LOCKING)) {
/*
Handler->ha_extra() for InnoDB does not fail ever as of now. In
case it is made to fail sometime later, we need to think about the
kind of error to be report to user.
*/
assert(0);
return true;
}
/**
Issue a warning when,
- We are skipping the SE locks in serializable
- We are skipping the SE locks for SELECT IN SHARE MODE in all
isolation mode.
- When ACL table is not used under I_S system view.
*/
if ((thd->tx_isolation == ISO_SERIALIZABLE || issue_warning) &&
!(tl->referencing_view && tl->referencing_view->is_system_view))
push_warning(thd, Sql_condition::SL_WARNING,
WARN_UNSUPPORTED_ACL_TABLES_READ,
ER_THD(thd, WARN_UNSUPPORTED_ACL_TABLES_READ));
}
return false;
}
/**
Open all tables in list
@param[in] thd Thread context.
@param[in,out] start List of tables to be open (it can be adjusted for
statement that uses tables only implicitly, e.g.
for "SELECT f1()").
@param[out] counter Number of tables which were open.
@param[in] flags Bitmap of flags to modify how the tables will be
open, see open_table() description for details.
@param[in] prelocking_strategy Strategy which specifies how prelocking
algorithm should work for this statement.
@note
Unless we are already in prelocked mode and prelocking strategy prescribes
so this function will also precache all SP/SFs explicitly or implicitly
(via views and triggers) used by the query and add tables needed for their
execution to table list. Statement that uses SFs, invokes triggers or
requires foreign key checks will be marked as requiring prelocking.
Prelocked mode will be enabled for such query during lock_tables() call.
If query for which we are opening tables is already marked as requiring
prelocking it won't do such precaching and will simply reuse table list
which is already built.
@retval false Success.
@retval true Error, reported.
*/
bool open_tables(THD *thd, Table_ref **start, uint *counter, uint flags,
Prelocking_strategy *prelocking_strategy) {
/*
We use pointers to "next_global" member in the last processed
Table_ref element and to the "next" member in the last processed
Sroutine_hash_entry element as iterators over, correspondingly, the table
list and stored routines list which stay valid and allow to continue
iteration when new elements are added to the tail of the lists.
*/
Table_ref **table_to_open;
TABLE *old_table;
Sroutine_hash_entry **sroutine_to_open;
Table_ref *tables;
Open_table_context ot_ctx(thd, flags);
bool error = false;
bool some_routine_modifies_data = false;
bool has_prelocking_list;
DBUG_TRACE;
bool audit_notified = false;
// Property of having external tables is always set in this function:
thd->lex->reset_has_external_tables();
restart:
/*
Close HANDLER tables which are marked for flush or against which there
are pending exclusive metadata locks. This is needed both in order to
avoid deadlocks and to have a point during statement execution at
which such HANDLERs are closed even if they don't create problems for
the current session (i.e. to avoid having a DDL blocked by HANDLERs
opened for a long time).
*/
if (!thd->handler_tables_hash.empty()) mysql_ha_flush(thd);
has_prelocking_list = thd->lex->requires_prelocking();
table_to_open = start;
old_table = *table_to_open ? (*table_to_open)->table : nullptr;
sroutine_to_open = &thd->lex->sroutines_list.first;
*counter = 0;
if (!(thd->state_flags & Open_tables_state::SYSTEM_TABLES))
THD_STAGE_INFO(thd, stage_opening_tables);
/*
If we are executing LOCK TABLES statement or a DDL statement
(in non-LOCK TABLES mode) we might have to acquire upgradable
semi-exclusive metadata locks (SNW or SNRW) on some of the
tables to be opened.
When executing CREATE TABLE .. If NOT EXISTS .. SELECT, the
table may not yet exist, in which case we acquire an exclusive
lock.
We acquire all such locks at once here as doing this in one
by one fashion may lead to deadlocks or starvation. Later when
we will be opening corresponding table pre-acquired metadata
lock will be reused (thanks to the fact that in recursive case
metadata locks are acquired without waiting).
*/
if (!(flags & (MYSQL_OPEN_HAS_MDL_LOCK | MYSQL_OPEN_FORCE_SHARED_MDL |
MYSQL_OPEN_FORCE_SHARED_HIGH_PRIO_MDL))) {
if (thd->locked_tables_mode) {
/*
Under LOCK TABLES, we can't acquire new locks, so we instead
need to check if appropriate locks were pre-acquired.
*/
Table_ref *end_table = thd->lex->first_not_own_table();
if (open_tables_check_upgradable_mdl(thd, *start, end_table) ||
acquire_backup_lock_in_lock_tables_mode(thd, *start, end_table)) {
error = true;
goto err;
}
} else {
Table_ref *table;
if (lock_table_names(thd, *start, thd->lex->first_not_own_table(),
ot_ctx.get_timeout(), flags)) {
error = true;
goto err;
}
for (table = *start; table && table != thd->lex->first_not_own_table();
table = table->next_global) {
if (table->mdl_request.is_ddl_or_lock_tables_lock_request() ||
table->open_strategy == Table_ref::OPEN_FOR_CREATE)
table->mdl_request.ticket = nullptr;
}
}
}
/*
Perform steps of prelocking algorithm until there are unprocessed
elements in prelocking list/set.
*/
while (*table_to_open ||
(thd->locked_tables_mode <= LTM_LOCK_TABLES && *sroutine_to_open)) {
/*
For every table in the list of tables to open, try to find or open
a table.
*/
for (tables = *table_to_open; tables;
table_to_open = &tables->next_global, tables = tables->next_global) {
old_table = (*table_to_open)->table;
error = open_and_process_table(thd, thd->lex, tables, counter,
prelocking_strategy, has_prelocking_list,
&ot_ctx);
if (error) {
if (ot_ctx.can_recover_from_failed_open()) {
/*
We have met exclusive metadata lock or old version of table.
Now we have to close all tables and release metadata locks.
We also have to throw away set of prelocked tables (and thus
close tables from this set that were open by now) since it
is possible that one of tables which determined its content
was changed.
Instead of implementing complex/non-robust logic mentioned
above we simply close and then reopen all tables.
We have to save pointer to table list element for table which we
have failed to open since closing tables can trigger removal of
elements from the table list (if MERGE tables are involved),
*/
close_tables_for_reopen(thd, start, ot_ctx.start_of_statement_svp());
/*
Here we rely on the fact that 'tables' still points to the valid
Table_ref element. Although currently this assumption is valid
it may change in future.
*/
if (ot_ctx.recover_from_failed_open()) goto err;
/* Re-open temporary tables after close_tables_for_reopen(). */
if (open_temporary_tables(thd, *start)) goto err;
error = false;
goto restart;
}
goto err;
}
DEBUG_SYNC(thd, "open_tables_after_open_and_process_table");
}
/*
Iterate through set of tables and generate table access audit events.
*/
if (!audit_notified && mysql_audit_table_access_notify(thd, *start)) {
error = true;
goto err;
}
/*
Event is not generated in the next loop. It may contain duplicated
table entries as well as new tables discovered for stored procedures.
Events for these tables will be generated during the queries of these
stored procedures.
*/
audit_notified = true;
/*
If we are not already in prelocked mode and extended table list is
not yet built for our statement we need to cache routines it uses
and build the prelocking list for it.
If we are not in prelocked mode but have built the extended table
list, we still need to call open_and_process_routine() to take
MDL locks on the routines.
*/
if (thd->locked_tables_mode <= LTM_LOCK_TABLES) {
bool routine_modifies_data;
/*
Process elements of the prelocking set which are present there
since parsing stage or were added to it by invocations of
Prelocking_strategy methods in the above loop over tables.
For example, if element is a routine, cache it and then,
if prelocking strategy prescribes so, add tables it uses to the
table list and routines it might invoke to the prelocking set.
*/
for (Sroutine_hash_entry *rt = *sroutine_to_open; rt;
sroutine_to_open = &rt->next, rt = rt->next) {
bool need_prelocking = false;
Table_ref **save_query_tables_last = thd->lex->query_tables_last;
error = open_and_process_routine(
thd, thd->lex, rt, prelocking_strategy, has_prelocking_list,
&ot_ctx, &need_prelocking, &routine_modifies_data);
if (need_prelocking && !thd->lex->requires_prelocking())
thd->lex->mark_as_requiring_prelocking(save_query_tables_last);
if (need_prelocking && !*start) *start = thd->lex->query_tables;
if (error) {
if (ot_ctx.can_recover_from_failed_open()) {
close_tables_for_reopen(thd, start,
ot_ctx.start_of_statement_svp());
if (ot_ctx.recover_from_failed_open()) goto err;
/* Re-open temporary tables after close_tables_for_reopen(). */
if (open_temporary_tables(thd, *start)) goto err;
error = false;
goto restart;
}
/*
Serious error during reading stored routines from mysql.proc table.
Something is wrong with the table or its contents, and an error has
been emitted; we must abort.
*/
goto err;
}
// Remember if any of SF modifies data.
some_routine_modifies_data |= routine_modifies_data;
}
}
}
/* Accessing data in XA_IDLE or XA_PREPARED is not allowed. */
if (*start &&
(thd->get_transaction()->xid_state()->check_xa_idle_or_prepared(true) ||
thd->get_transaction()->xid_state()->xa_trans_rolled_back()))
return true;
/*
If some routine is modifying the table then the statement is not read only.
If timer is enabled then resetting the timer in this case.
*/
if (thd->timer && some_routine_modifies_data) {
reset_statement_timer(thd);
push_warning(thd, Sql_condition::SL_NOTE, ER_NON_RO_SELECT_DISABLE_TIMER,
ER_THD(thd, ER_NON_RO_SELECT_DISABLE_TIMER));
}
/*
After successful open of all tables, including MERGE parents and
children, attach the children to their parents. At end of statement,
the children are detached. Attaching and detaching are always done,
even under LOCK TABLES.
We also convert all TL_WRITE_DEFAULT and TL_READ_DEFAULT locks to
appropriate "real" lock types to be used for locking and to be passed
to storage engine.
*/
for (tables = *start; tables; tables = tables->next_global) {
TABLE *tbl = tables->table;
/*
NOTE: temporary merge tables should be processed here too, because
a temporary merge table can be based on non-temporary tables.
*/
/* Schema tables may not have a TABLE object here. */
if (tbl && tbl->file && tbl->file->ht->db_type == DB_TYPE_MRG_MYISAM) {
/* MERGE tables need to access parent and child TABLE_LISTs. */
assert(tbl->pos_in_table_list == tables);
if (tbl->db_stat && tbl->file->ha_extra(HA_EXTRA_ATTACH_CHILDREN)) {
error = true;
goto err;
}
}
// Remember if an external table has been opened in this statement.
if (tbl != nullptr && tbl->s->has_secondary_engine() &&
ha_check_storage_engine_flag(tbl->s->db_type(),
HTON_SUPPORTS_EXTERNAL_SOURCE)) {
thd->lex->set_has_external_tables();
}
/*
Access to ACL table in a SELECT ... LOCK IN SHARE MODE are required
to skip acquiring row locks. So, we use TL_READ_DEFAULT lock on ACL
tables. This allows concurrent ACL DDL's.
Do not request SE to skip row lock if 'flags' has
MYSQL_OPEN_FORCE_SHARED_MDL, which indicates that this is PREPARE
phase. It is OK to do so since during this phase no rows will be read
anyway. And by doing this we avoid generation of extra warnings.
EXECUTION phase will request SE to skip row locks if necessary.
*/
bool issue_warning_on_skipping_row_lock = false;
if (tables->lock_descriptor().type == TL_READ_WITH_SHARED_LOCKS &&
!(flags & MYSQL_OPEN_FORCE_SHARED_MDL) &&
is_acl_table_in_non_LTM(tables, thd->locked_tables_mode)) {
tables->set_lock({TL_READ_DEFAULT, THR_DEFAULT});
issue_warning_on_skipping_row_lock = true;
}
/* Set appropriate TABLE::lock_type. */
if (tbl && tables->lock_descriptor().type != TL_UNLOCK &&
!thd->locked_tables_mode) {
if (tables->lock_descriptor().type == TL_WRITE_DEFAULT)
tbl->reginfo.lock_type = thd->update_lock_default;
else if (tables->lock_descriptor().type == TL_WRITE_CONCURRENT_DEFAULT)
tables->table->reginfo.lock_type = thd->insert_lock_default;
else if (tables->lock_descriptor().type == TL_READ_DEFAULT)
tbl->reginfo.lock_type = read_lock_type_for_table(
thd, thd->lex, tables, some_routine_modifies_data);
else
tbl->reginfo.lock_type = tables->lock_descriptor().type;
}
/*
SELECT using a I_S system view with 'FOR UPDATE' and
'LOCK IN SHARED MODE' clause is not allowed.
*/
if (tables->is_system_view &&
tables->lock_descriptor().type == TL_READ_WITH_SHARED_LOCKS) {
my_error(ER_IS_QUERY_INVALID_CLAUSE, MYF(0), "LOCK IN SHARE MODE");
error = true;
goto err;
}
// Setup lock type for DD tables used under I_S view.
if (set_non_locking_read_for_IS_view(thd, tables)) {
error = true;
goto err;
}
/**
Setup lock type for read requests for ACL table in SQL statements.
Do not request SE to skip row lock if 'flags' has
MYSQL_OPEN_FORCE_SHARED_MDL, which indicates that this is PREPARE
phase. It is OK to do so since during this phase no rows will be read
anyway. And by doing this we avoid generation of extra warnings.
EXECUTION phase will request SE to skip row locks if necessary.
*/
if (!(flags & MYSQL_OPEN_FORCE_SHARED_MDL) &&
set_non_locking_read_for_ACL_table(
thd, tables, issue_warning_on_skipping_row_lock)) {
error = true;
goto err;
}
} // End of for(;;)
err:
// If a new TABLE was introduced, it's garbage, don't link to it:
if (error && *table_to_open && old_table != (*table_to_open)->table) {
(*table_to_open)->table = nullptr;
}
DBUG_PRINT("open_tables", ("returning: %d", (int)error));
return error;
}
/**
Defines how prelocking algorithm for DML statements should handle routines:
- For CALL statements we do unrolling (i.e. open and lock tables for each
sub-statement individually). So for such statements prelocking is enabled
only if stored functions are used in parameter list and only for period
during which we calculate values of parameters. Thus in this strategy we
ignore procedure which is directly called by such statement and extend
the prelocking set only with tables/functions used by SF called from the
parameter list.
- For any other statement any routine which is directly or indirectly called
by statement is going to be executed in prelocked mode. So in this case we
simply add all tables and routines used by it to the prelocking set.
@param[in] thd Thread context.
@param[in] prelocking_ctx Prelocking context of the statement.
@param[in] rt Prelocking set element describing routine.
@param[in] sp Routine body.
@param[out] need_prelocking Set to true if method detects that prelocking
required, not changed otherwise.
@retval false Success.
@retval true Failure (OOM).
*/
bool DML_prelocking_strategy::handle_routine(THD *thd,
Query_tables_list *prelocking_ctx,
Sroutine_hash_entry *rt,
sp_head *sp,
bool *need_prelocking) {
/*
We assume that for any "CALL proc(...)" statement sroutines_list will
have 'proc' as first element (it may have several, consider e.g.
"proc(sp_func(...)))". This property is currently guaranteed by the
parser.
*/
if (rt != prelocking_ctx->sroutines_list.first ||
rt->type() != Sroutine_hash_entry::PROCEDURE) {
*need_prelocking = true;
sp_update_stmt_used_routines(thd, prelocking_ctx, &sp->m_sroutines,
rt->belong_to_view);
sp->add_used_tables_to_table_list(thd, &prelocking_ctx->query_tables_last,
prelocking_ctx->sql_command,
rt->belong_to_view);
}
sp->propagate_attributes(prelocking_ctx);
return false;
}
/**
Defines how prelocking algorithm for DML statements should handle table list
elements:
- If table has triggers we should add all tables and routines
used by them to the prelocking set.
- If table participates in a foreign key we should add another
table from it to the prelocking set with an appropriate metadata
lock.
We do not need to acquire metadata locks on trigger names
in DML statements, since all DDL statements
that change trigger metadata always lock their
subject tables.
@param[in] thd Thread context.
@param[in] prelocking_ctx Prelocking context of the statement.
@param[in] table_list Table list element for table.
@param[out] need_prelocking Set to true if method detects that prelocking
required, not changed otherwise.
@retval false Success.
@retval true Failure (OOM).
*/
bool DML_prelocking_strategy::handle_table(THD *thd,
Query_tables_list *prelocking_ctx,
Table_ref *table_list,
bool *need_prelocking) {
/* We rely on a caller to check that table is going to be changed. */
assert(table_list->lock_descriptor().type >= TL_WRITE_ALLOW_WRITE);
if (table_list->trg_event_map) {
if (table_list->table->triggers) {
*need_prelocking = true;
if (table_list->table->triggers->add_tables_and_routines_for_triggers(
thd, prelocking_ctx, table_list))
return true;
}
/*
When FOREIGN_KEY_CHECKS is 0 we are not going to do any foreign key checks
so we don't need to add child and parent tables to the prelocking list.
However, since trigger or stored function might change this variable for
their duration (it is, actually, advisable to do so in some scenarios),
we can apply this optimization only to tables which are directly used by
the top-level statement.
While processing LOCK TABLES, we must disregard F_K_C too, since the
prelocking set will be used while in LTM mode, and F_K_C may be turned
on later, after the set has been established.
*/
if ((!(thd->variables.option_bits & OPTION_NO_FOREIGN_KEY_CHECKS) ||
prelocking_ctx->sql_command == SQLCOM_LOCK_TABLES ||
table_list->prelocking_placeholder) &&
!(table_list->table->s->tmp_table)) {
bool is_insert =
(table_list->trg_event_map &
static_cast<uint8>(1 << static_cast<int>(TRG_EVENT_INSERT)));
bool is_update =
(table_list->trg_event_map &
static_cast<uint8>(1 << static_cast<int>(TRG_EVENT_UPDATE)));
bool is_delete =
(table_list->trg_event_map &
static_cast<uint8>(1 << static_cast<int>(TRG_EVENT_DELETE)));
process_table_fks(thd, prelocking_ctx, table_list->table->s, is_insert,
is_update, is_delete, table_list->belong_to_view,
need_prelocking);
}
}
return false;
}
/**
Defines how prelocking algorithm for DML statements should handle view -
all view routines should be added to the prelocking set.
@param[in] thd Thread context.
@param[in] prelocking_ctx Prelocking context of the statement.
@param[in] table_list Table list element for view.
@param[out] need_prelocking Set to true if method detects that prelocking
required, not changed otherwise.
@retval false Success.
@retval true Failure (OOM).
*/
bool DML_prelocking_strategy::handle_view(THD *thd,
Query_tables_list *prelocking_ctx,
Table_ref *table_list,
bool *need_prelocking) {
if (table_list->view_query()->uses_stored_routines()) {
*need_prelocking = true;
sp_update_stmt_used_routines(thd, prelocking_ctx,
&table_list->view_query()->sroutines_list,
table_list->top_table());
}
/*
If a trigger was defined on one of the associated tables then assign the
'trg_event_map' value of the view to the next table in table_list. When a
Stored function is invoked, all the associated tables including the tables
associated with the trigger are prelocked.
*/
if (table_list->trg_event_map && table_list->next_global)
table_list->next_global->trg_event_map = table_list->trg_event_map;
return false;
}
/**
Defines how prelocking algorithm for LOCK TABLES statement should handle
table list elements.
@param[in] thd Thread context.
@param[in] prelocking_ctx Prelocking context of the statement.
@param[in] table_list Table list element for table.
@param[out] need_prelocking Set to true if method detects that prelocking
required, not changed otherwise.
@retval false Success.
@retval true Failure (OOM).
*/
bool Lock_tables_prelocking_strategy::handle_table(
THD *thd, Query_tables_list *prelocking_ctx, Table_ref *table_list,
bool *need_prelocking) {
if (DML_prelocking_strategy::handle_table(thd, prelocking_ctx, table_list,
need_prelocking))
return true;
/* We rely on a caller to check that table is going to be changed. */
assert(table_list->lock_descriptor().type >= TL_WRITE_ALLOW_WRITE);
return false;
}
/**
Defines how prelocking algorithm for ALTER TABLE statement should handle
routines - do nothing as this statement is not supposed to call routines.
We still can end up in this method when someone tries
to define a foreign key referencing a view, and not just
a simple view, but one that uses stored routines.
*/
bool Alter_table_prelocking_strategy::handle_routine(THD *, Query_tables_list *,
Sroutine_hash_entry *,
sp_head *, bool *) {
return false;
}
/**
Defines how prelocking algorithm for ALTER TABLE statement should handle
table list elements.
Unlike in DML, we do not process triggers here.
*/
bool Alter_table_prelocking_strategy::handle_table(THD *, Query_tables_list *,
Table_ref *, bool *) {
return false;
}
/**
Defines how prelocking algorithm for ALTER TABLE statement
should handle view - do nothing. We don't need to add view
routines to the prelocking set in this case as view is not going
to be materialized.
*/
bool Alter_table_prelocking_strategy::handle_view(THD *, Query_tables_list *,
Table_ref *, bool *) {
return false;
}
/**
Check that lock is ok for tables; Call start stmt if ok
@param thd Thread handle.
@param prelocking_ctx Prelocking context.
@param table_list Table list element for table to be checked.
@retval false - Ok.
@retval true - Error.
*/
static bool check_lock_and_start_stmt(THD *thd,
Query_tables_list *prelocking_ctx,
Table_ref *table_list) {
int error;
thr_lock_type lock_type;
DBUG_TRACE;
/*
Prelocking placeholder is not set for Table_ref that
are directly used by TOP level statement.
*/
assert(table_list->prelocking_placeholder == false);
/*
TL_WRITE_DEFAULT, TL_READ_DEFAULT and TL_WRITE_CONCURRENT_DEFAULT
are supposed to be parser only types of locks so they should be
converted to appropriate other types to be passed to storage engine.
The exact lock type passed to the engine is important as, for example,
InnoDB uses it to determine what kind of row locks should be acquired
when executing statement in prelocked mode or under LOCK TABLES with
@@innodb_table_locks = 0.
Last argument routine_modifies_data for read_lock_type_for_table()
is ignored, as prelocking placeholder will never be set here.
*/
if (table_list->lock_descriptor().type == TL_WRITE_DEFAULT)
lock_type = thd->update_lock_default;
else if (table_list->lock_descriptor().type == TL_WRITE_CONCURRENT_DEFAULT)
lock_type = thd->insert_lock_default;
else if (table_list->lock_descriptor().type == TL_READ_DEFAULT)
lock_type = read_lock_type_for_table(thd, prelocking_ctx, table_list, true);
else
lock_type = table_list->lock_descriptor().type;
if ((int)lock_type > (int)TL_WRITE_ALLOW_WRITE &&
(int)table_list->table->reginfo.lock_type <= (int)TL_WRITE_ALLOW_WRITE) {
my_error(ER_TABLE_NOT_LOCKED_FOR_WRITE, MYF(0), table_list->alias);
return true;
}
if ((error = table_list->table->file->start_stmt(thd, lock_type))) {
table_list->table->file->print_error(error, MYF(0));
return true;
}
/*
Record in transaction state tracking
*/
if (thd->variables.session_track_transaction_info > TX_TRACK_NONE) {
TX_TRACKER_GET(tst);
enum enum_tx_state s;
s = tst->calc_trx_state(lock_type,
table_list->table->file->has_transactions());
tst->add_trx_state(thd, s);
}
return false;
}
/**
@brief Open and lock one table
@param[in] thd thread handle
@param[in] table_l table to open is first table in this list
@param[in] lock_type lock to use for table
@param[in] flags options to be used while opening and locking
table (see open_table(), mysql_lock_tables())
@param[in] prelocking_strategy Strategy which specifies how prelocking
algorithm should work for this statement.
@return table
@retval != NULL OK, opened table returned
@retval NULL Error
@note
If ok, the following are also set:
table_list->lock_type lock_type
table_list->table table
@note
If table_l is a list, not a single table, the list is temporarily
broken.
@details
This function is meant as a replacement for open_ltable() when
MERGE tables can be opened. open_ltable() cannot open MERGE tables.
There may be more differences between open_n_lock_single_table() and
open_ltable(). One known difference is that open_ltable() does
neither call thd->decide_logging_format() nor handle some other logging
and locking issues because it does not call lock_tables().
*/
TABLE *open_n_lock_single_table(THD *thd, Table_ref *table_l,
thr_lock_type lock_type, uint flags,
Prelocking_strategy *prelocking_strategy) {
Table_ref *save_next_global;
DBUG_TRACE;
/* Remember old 'next' pointer. */
save_next_global = table_l->next_global;
/* Break list. */
table_l->next_global = nullptr;
/* Set requested lock type. */
table_l->set_lock({lock_type, THR_DEFAULT});
/* Allow to open real tables only. */
table_l->required_type = dd::enum_table_type::BASE_TABLE;
/* Open the table. */
if (open_and_lock_tables(thd, table_l, flags, prelocking_strategy))
table_l->table = nullptr; /* Just to be sure. */
/* Restore list. */
table_l->next_global = save_next_global;
return table_l->table;
}
/*
Open and lock one table
SYNOPSIS
open_ltable()
thd Thread handler
table_list Table to open is first table in this list
lock_type Lock to use for open
lock_flags Flags passed to mysql_lock_table
NOTE
This function doesn't do anything like SP/SF/views/triggers analysis done
in open_table()/lock_tables(). It is intended for opening of only one
concrete table. And used only in special contexts.
RETURN VALUES
table Opened table
0 Error
If ok, the following are also set:
table_list->lock_type lock_type
table_list->table table
*/
TABLE *open_ltable(THD *thd, Table_ref *table_list, thr_lock_type lock_type,
uint lock_flags) {
TABLE *table;
Open_table_context ot_ctx(thd, lock_flags);
bool error;
DBUG_TRACE;
/* should not be used in a prelocked_mode context, see NOTE above */
assert(thd->locked_tables_mode < LTM_PRELOCKED);
if (!(thd->state_flags & Open_tables_state::SYSTEM_TABLES))
THD_STAGE_INFO(thd, stage_opening_tables);
/* open_ltable can be used only for BASIC TABLEs */
table_list->required_type = dd::enum_table_type::BASE_TABLE;
/* This function can't properly handle requests for such metadata locks. */
assert(!table_list->mdl_request.is_ddl_or_lock_tables_lock_request());
while ((error = open_table(thd, table_list, &ot_ctx)) &&
ot_ctx.can_recover_from_failed_open()) {
/*
Even though we have failed to open table we still need to
call release_transactional_locks() to release metadata locks which
might have been acquired successfully.
*/
thd->mdl_context.rollback_to_savepoint(ot_ctx.start_of_statement_svp());
table_list->mdl_request.ticket = nullptr;
if (ot_ctx.recover_from_failed_open()) break;
}
if (!error) {
/*
We can't have a view or some special "open_strategy" in this function
so there should be a TABLE instance.
*/
assert(table_list->table);
table = table_list->table;
if (table->file->ht->db_type == DB_TYPE_MRG_MYISAM) {
/* A MERGE table must not come here. */
/* purecov: begin tested */
my_error(ER_WRONG_OBJECT, MYF(0), table->s->db.str,
table->s->table_name.str, "BASE TABLE");
table = nullptr;
goto end;
/* purecov: end */
}
table_list->set_lock({lock_type, THR_DEFAULT});
if (thd->locked_tables_mode) {
if (check_lock_and_start_stmt(thd, thd->lex, table_list)) table = nullptr;
} else {
assert(thd->lock == nullptr); // You must lock everything at once
if ((table->reginfo.lock_type = lock_type) != TL_UNLOCK)
if (!(thd->lock =
mysql_lock_tables(thd, &table_list->table, 1, lock_flags))) {
table = nullptr;
}
}
} else
table = nullptr;
end:
if (table == nullptr) {
if (!thd->in_sub_stmt) trans_rollback_stmt(thd);
close_thread_tables(thd);
}
return table;
}
/**
Open all tables in list, locks them and optionally process derived tables.
@param thd Thread context.
@param tables List of tables for open and locking.
@param flags Bitmap of options to be used to open and lock
tables (see open_tables() and mysql_lock_tables()
for details).
@param prelocking_strategy Strategy which specifies how prelocking algorithm
should work for this statement.
@note
The thr_lock locks will automatically be freed by close_thread_tables().
@note
open_and_lock_tables() is not intended for open-and-locking system tables
in those cases when execution of statement has started already and other
tables have been opened. Use open_trans_system_tables_for_read() instead.
@retval false OK.
@retval true Error
*/
bool open_and_lock_tables(THD *thd, Table_ref *tables, uint flags,
Prelocking_strategy *prelocking_strategy) {
uint counter;
MDL_savepoint mdl_savepoint = thd->mdl_context.mdl_savepoint();
DBUG_TRACE;
/*
open_and_lock_tables() must not be used to open system tables. There must
be no active attachable transaction when open_and_lock_tables() is called.
Exception is made to the read-write attachables with explicitly specified
in the assert table.
Callers in the read-write case must make sure no side effect to
the global transaction state is inflicted when the attachable one
will commit.
*/
assert(!thd->is_attachable_ro_transaction_active() &&
(!thd->is_attachable_rw_transaction_active() ||
!strcmp(tables->table_name, "gtid_executed")));
if (open_tables(thd, &tables, &counter, flags, prelocking_strategy)) goto err;
DBUG_EXECUTE_IF("sleep_open_and_lock_after_open", {
const char *old_proc_info = thd->proc_info();
thd->set_proc_info("DBUG sleep");
my_sleep(6000000);
thd->set_proc_info(old_proc_info);
});
if (lock_tables(thd, tables, counter, flags)) goto err;
return false;
err:
// Rollback the statement execution done so far
if (!thd->in_sub_stmt) trans_rollback_stmt(thd);
close_thread_tables(thd);
/* Don't keep locks for a failed statement. */
thd->mdl_context.rollback_to_savepoint(mdl_savepoint);
return true;
}
/**
Check if a secondary engine can be used to execute the current
statement, and if so, replace the opened tables with their secondary
counterparts.
@param thd thread handler
@param flags bitmap of flags to pass to open_table
@return true if an error is raised, false otherwise
*/
static bool open_secondary_engine_tables(THD *thd, uint flags) {
LEX *const lex = thd->lex;
Sql_cmd *const sql_cmd = lex->m_sql_cmd;
// The previous execution context should have been destroyed.
assert(lex->secondary_engine_execution_context() == nullptr);
// If use of secondary engines has been disabled for the statement,
// there is nothing to do.
if (sql_cmd == nullptr || sql_cmd->secondary_storage_engine_disabled())
return false;
// If the user has requested the use of a secondary storage engine
// for this statement, skip past the initial optimization for the
// primary storage engine and go straight to the secondary engine.
if (thd->secondary_engine_optimization() ==
Secondary_engine_optimization::PRIMARY_TENTATIVELY &&
thd->variables.use_secondary_engine == SECONDARY_ENGINE_FORCED) {
thd->set_secondary_engine_optimization(
Secondary_engine_optimization::SECONDARY);
mysql_thread_set_secondary_engine(true);
mysql_statement_set_secondary_engine(thd->m_statement_psi, true);
}
// Only open secondary engine tables if use of a secondary engine
// has been requested.
if (thd->secondary_engine_optimization() !=
Secondary_engine_optimization::SECONDARY)
return false;
// If the statement cannot be executed in a secondary engine because
// of a property of the statement, do not attempt to open the
// secondary tables. Also disable use of secondary engines for
// future executions of the statement, since these properties will
// not change between executions.
const LEX_CSTRING *secondary_engine =
sql_cmd->eligible_secondary_storage_engine();
const plugin_ref secondary_engine_plugin =
secondary_engine == nullptr
? nullptr
: ha_resolve_by_name(thd, secondary_engine, false);
if ((secondary_engine_plugin == nullptr) ||
!plugin_is_ready(*secondary_engine, MYSQL_STORAGE_ENGINE_PLUGIN)) {
// Didn't find a secondary storage engine to use for the query.
sql_cmd->disable_secondary_storage_engine();
return false;
}
// If the statement cannot be executed in a secondary engine because
// of a property of the environment, do not attempt to open the
// secondary tables. However, do not disable use of secondary
// storage engines for future executions of the statement, since the
// environment may change before the next execution.
if (!thd->is_secondary_storage_engine_eligible()) return false;
auto hton = plugin_data<const handlerton *>(secondary_engine_plugin);
sql_cmd->use_secondary_storage_engine(hton);
// Replace the TABLE objects in the Table_ref with secondary tables.
Open_table_context ot_ctx(thd, flags | MYSQL_OPEN_SECONDARY_ENGINE);
Table_ref *tl = lex->query_tables;
// For INSERT INTO SELECT and CTAS statements, the table to insert into does
// not have to have a secondary engine. This table is always first in the list
if ((lex->sql_command == SQLCOM_INSERT_SELECT ||
lex->sql_command == SQLCOM_CREATE_TABLE) &&
tl != nullptr)
tl = tl->next_global;
for (; tl != nullptr; tl = tl->next_global) {
if (tl->is_placeholder()) continue;
TABLE *primary_table = tl->table;
tl->table = nullptr;
if (open_table(thd, tl, &ot_ctx)) {
if (!thd->is_error()) {
/*
open_table() has not registered any error, implying that we can
retry the failed open; but it is complicated to do so reliably, so we
prefer to simply fail and re-prepare the statement in the primary
engine, as an exceptional case. So we register an error.
*/
my_error(ER_SECONDARY_ENGINE_PLUGIN, MYF(0),
"Transient error when opening tables in RAPID");
}
return true;
}
assert(tl->table->s->is_secondary_engine());
tl->table->file->ha_set_primary_handler(primary_table->file);
}
// Prepare the secondary engine for executing the statement.
return hton->prepare_secondary_engine != nullptr &&
hton->prepare_secondary_engine(thd, lex);
}
/**
Open all tables for a query or statement, in list started by "tables"
@param thd thread handler
@param tables list of tables for open
@param flags bitmap of flags to modify how the tables will be open:
MYSQL_LOCK_IGNORE_FLUSH - open table even if someone has
done a flush on it.
@retval false - ok
@retval true - error
@note
This is to be used on prepare stage when you don't read any
data from the tables.
@note
Updates Query_tables_list::table_count as side-effect.
*/
bool open_tables_for_query(THD *thd, Table_ref *tables, uint flags) {
DML_prelocking_strategy prelocking_strategy;
MDL_savepoint mdl_savepoint = thd->mdl_context.mdl_savepoint();
DBUG_TRACE;
assert(tables == thd->lex->query_tables);
if (open_tables(thd, &tables, &thd->lex->table_count, flags,
&prelocking_strategy))
goto end;
if (open_secondary_engine_tables(thd, flags)) goto end;
return false;
end:
/*
No need to commit/rollback the statement transaction: it's
either not started or we're filling in an INFORMATION_SCHEMA
table on the fly, and thus mustn't manipulate with the
transaction of the enclosing statement.
*/
assert(thd->get_transaction()->is_empty(Transaction_ctx::STMT) ||
(thd->state_flags & Open_tables_state::BACKUPS_AVAIL) ||
thd->in_sub_stmt);
close_thread_tables(thd);
/* Don't keep locks for a failed statement. */
thd->mdl_context.rollback_to_savepoint(mdl_savepoint);
return true; /* purecov: inspected */
}
/*
Mark all real tables in the list as free for reuse.
SYNOPSIS
mark_real_tables_as_free_for_reuse()
thd - thread context
table - head of the list of tables
DESCRIPTION
Marks all real tables in the list (i.e. not views, derived
or schema tables) as free for reuse.
*/
static void mark_real_tables_as_free_for_reuse(Table_ref *table_list) {
Table_ref *table;
for (table = table_list; table; table = table->next_global)
if (!table->is_placeholder()) {
table->table->query_id = 0;
}
for (table = table_list; table; table = table->next_global)
if (!table->is_placeholder() && table->table->db_stat) {
/*
Detach children of MyISAMMRG tables used in
sub-statements, they will be reattached at open.
This has to be done in a separate loop to make sure
that children have had their query_id cleared.
*/
table->table->file->ha_extra(HA_EXTRA_DETACH_CHILDREN);
}
}
/**
Lock all tables in a list.
@param thd Thread handler
@param tables Tables to lock
@param count Number of opened tables
@param flags Options (see mysql_lock_tables() for details)
You can't call lock_tables() while holding thr_lock locks, as
this would break the dead-lock-free handling thr_lock gives us.
You must always get all needed locks at once.
If the query for which we are calling this function is marked as
requiring prelocking, this function will change
locked_tables_mode to LTM_PRELOCKED.
@retval false Success.
@retval true A lock wait timeout, deadlock or out of memory.
*/
bool lock_tables(THD *thd, Table_ref *tables, uint count, uint flags) {
Table_ref *table;
DBUG_TRACE;
/*
We can't meet statement requiring prelocking if we already
in prelocked mode.
*/
assert(thd->locked_tables_mode <= LTM_LOCK_TABLES ||
!thd->lex->requires_prelocking());
/*
lock_tables() should not be called if this statement has
already locked its tables.
*/
assert(thd->lex->lock_tables_state == Query_tables_list::LTS_NOT_LOCKED);
if (!tables && !thd->lex->requires_prelocking()) {
/*
Even though we are not really locking any tables mark this
statement as one that has locked its tables, so we won't
call this function second time for the same execution of
the same statement.
*/
thd->lex->lock_tables_state = Query_tables_list::LTS_LOCKED;
int ret = thd->decide_logging_format(tables);
return ret;
}
/*
Check for thd->locked_tables_mode to avoid a redundant
and harmful attempt to lock the already locked tables again.
Checking for thd->lock is not enough in some situations. For example,
if a stored function contains
"drop table t3; create temporary t3 ..; insert into t3 ...;"
thd->lock may be 0 after drop tables, whereas locked_tables_mode
is still on. In this situation an attempt to lock temporary
table t3 will lead to a memory leak.
*/
if (!thd->locked_tables_mode) {
assert(thd->lock == nullptr); // You must lock everything at once
TABLE **start, **ptr;
if (!(ptr = start = (TABLE **)thd->alloc(sizeof(TABLE *) * count)))
return true;
for (table = tables; table; table = table->next_global) {
if (!table->is_placeholder() &&
/*
Do not call handler::store_lock()/external_lock() for temporary
tables from prelocking list.
Prelocking algorithm does not add element for a table to the
prelocking list if it finds that the routine that uses the table can
create it as a temporary during its execution. Note that such
routine actually can use existing temporary table if its CREATE
TEMPORARY TABLE has IF NOT EXISTS clause. For such tables we rely on
calls to handler::start_stmt() done by routine's substatement when
it accesses the table to inform storage engine about table
participation in transaction and type of operation carried out,
instead of calls to handler::store_lock()/external_lock() done at
prelocking stage.
In cases when statement uses two routines one of which can create
temporary table and modifies it, while another only reads from this
table, storage engine might be confused about real operation type
performed by the whole statement. Calls to
handler::store_lock()/external_lock() done at prelocking stage will
inform SE only about read part, while information about modification
will be delayed until handler::start_stmt() call during execution of
the routine doing modification. InnoDB considers this breaking of
promise about operation type and fails on assertion.
To avoid this problem we try to handle both the cases when temporary
table can be created by routine and the case when it is created
outside of routine and only accessed by it, uniformly. We don't call
handler::store_lock()/external_lock() for temporary tables used by
routines at prelocking stage and rely on calls to
handler::start_stmt(), which happen during substatement execution,
to pass correct information about operation type instead.
*/
!(table->prelocking_placeholder &&
table->table->s->tmp_table != NO_TMP_TABLE)) {
*(ptr++) = table->table;
}
}
DEBUG_SYNC(thd, "before_lock_tables_takes_lock");
if (!(thd->lock =
mysql_lock_tables(thd, start, (uint)(ptr - start), flags)))
return true;
DEBUG_SYNC(thd, "after_lock_tables_takes_lock");
if (thd->lex->requires_prelocking() &&
thd->lex->sql_command != SQLCOM_LOCK_TABLES) {
Table_ref *first_not_own = thd->lex->first_not_own_table();
/*
We just have done implicit LOCK TABLES, and now we have
to emulate first open_and_lock_tables() after it.
When open_and_lock_tables() is called for a single table out of
a table list, the 'next_global' chain is temporarily broken. We
may not find 'first_not_own' before the end of the "list".
Look for example at those places where open_n_lock_single_table()
is called. That function implements the temporary breaking of
a table list for opening a single table.
*/
for (table = tables; table && table != first_not_own;
table = table->next_global) {
if (!table->is_placeholder()) {
table->table->query_id = thd->query_id;
if (check_lock_and_start_stmt(thd, thd->lex, table)) {
mysql_unlock_tables(thd, thd->lock);
thd->lock = nullptr;
return true;
}
}
}
/*
Let us mark all tables which don't belong to the statement itself,
and was marked as occupied during open_tables() as free for reuse.
*/
mark_real_tables_as_free_for_reuse(first_not_own);
DBUG_PRINT("info", ("locked_tables_mode= LTM_PRELOCKED"));
thd->enter_locked_tables_mode(LTM_PRELOCKED);
}
} else {
/*
When we implicitly open DD tables used by a IS query in LOCK TABLE mode,
we do not go through mysql_lock_tables(), which sets lock type to use
by SE. Here, we request SE to use read lock for these implicitly opened
DD tables using ha_external_lock().
TODO: In PRELOCKED under LOCKED TABLE mode, if sub-statement is a IS
query then for DD table ha_external_lock is called more than once.
This works for now as in this mode each sub-statement gets its own
brand new TABLE instances for each table.
Allocating a brand new TABLE instances for each sub-statement is
a resources wastage. Once this issue is fixed, following code
should be adjusted to not to call ha_external_lock in sub-statement
mode (similar to how code in close_thread_table() behaves).
*/
if (in_LTM(thd)) {
for (table = tables; table; table = table->next_global) {
TABLE *tbl = table->table;
if (tbl && belongs_to_dd_table(table)) {
assert(tbl->file->get_lock_type() == F_UNLCK);
tbl->file->init_table_handle_for_HANDLER();
tbl->file->ha_external_lock(thd, F_RDLCK);
}
}
}
Table_ref *first_not_own = thd->lex->first_not_own_table();
/*
When open_and_lock_tables() is called for a single table out of
a table list, the 'next_global' chain is temporarily broken. We
may not find 'first_not_own' before the end of the "list".
Look for example at those places where open_n_lock_single_table()
is called. That function implements the temporary breaking of
a table list for opening a single table.
*/
for (table = tables; table && table != first_not_own;
table = table->next_global) {
if (table->is_placeholder()) continue;
/*
In a stored function or trigger we should ensure that we won't change
a table that is already used by the calling statement.
*/
if (thd->locked_tables_mode >= LTM_PRELOCKED &&
table->lock_descriptor().type >= TL_WRITE_ALLOW_WRITE) {
for (TABLE *opentab = thd->open_tables; opentab;
opentab = opentab->next) {
if (table->table->s == opentab->s && opentab->query_id &&
table->table->query_id != opentab->query_id) {
my_error(ER_CANT_UPDATE_USED_TABLE_IN_SF_OR_TRG, MYF(0),
table->table->s->table_name.str);
return true;
}
}
}
if (check_lock_and_start_stmt(thd, thd->lex, table)) {
return true;
}
}
/*
If we are under explicit LOCK TABLES and our statement requires
prelocking, we should mark all "additional" tables as free for use
and enter prelocked mode.
*/
if (thd->lex->requires_prelocking()) {
mark_real_tables_as_free_for_reuse(first_not_own);
DBUG_PRINT("info",
("thd->locked_tables_mode= LTM_PRELOCKED_UNDER_LOCK_TABLES"));
thd->locked_tables_mode = LTM_PRELOCKED_UNDER_LOCK_TABLES;
}
}
/*
Mark the statement as having tables locked. For purposes
of Query_tables_list::lock_tables_state we treat any
statement which passes through lock_tables() as such.
*/
thd->lex->lock_tables_state = Query_tables_list::LTS_LOCKED;
int ret = thd->decide_logging_format(tables);
return ret;
}
/**
Simplified version of lock_tables() call to be used for locking
data-dictionary tables when reading or storing data-dictionary
objects.
@note The main reason why this function exists is that it avoids
allocating temporary buffer on memory root of statement.
As result it can be called many times (e.g. thousands)
during DDL statement execution without hogging memory.
*/
bool lock_dictionary_tables(THD *thd, Table_ref *tables, uint count,
uint flags) {
DBUG_TRACE;
// We always open at least one DD table.
assert(tables);
/*
This function is supposed to be called after backing up and resetting
to clean state Open_tables_state and Query_table_lists contexts.
*/
assert(thd->locked_tables_mode == LTM_NONE);
assert(!thd->lex->requires_prelocking());
assert(thd->lex->lock_tables_state == Query_tables_list::LTS_NOT_LOCKED);
assert(thd->lock == nullptr);
TABLE **start, **ptr;
if (!(ptr = start = (TABLE **)my_alloca(sizeof(TABLE *) * count)))
return true;
for (Table_ref *table = tables; table; table = table->next_global) {
// Data-dictionary tables must be base tables.
assert(!table->is_placeholder());
assert(table->table->s->tmp_table == NO_TMP_TABLE);
// There should be no prelocking when DD code uses this call.
assert(!table->prelocking_placeholder);
*(ptr++) = table->table;
}
DEBUG_SYNC(thd, "before_lock_dictionary_tables_takes_lock");
if (!(thd->lock = mysql_lock_tables(thd, start, (uint)(ptr - start), flags)))
return true;
thd->lex->lock_tables_state = Query_tables_list::LTS_LOCKED;
return false;
}
/**
Prepare statement for reopening of tables and recalculation of set of
prelocked tables.
@param[in] thd Thread context.
@param[in,out] tables List of tables which we were trying to open
and lock.
@param[in] start_of_statement_svp MDL savepoint which represents the set
of metadata locks which the current transaction
managed to acquire before execution of the current
statement and to which we should revert before
trying to reopen tables. NULL if no metadata locks
were held and thus all metadata locks should be
released.
*/
void close_tables_for_reopen(THD *thd, Table_ref **tables,
const MDL_savepoint &start_of_statement_svp) {
Table_ref *first_not_own_table = thd->lex->first_not_own_table();
/*
If table list consists only from tables from prelocking set, table list
for new attempt should be empty, so we have to update list's root pointer.
*/
if (first_not_own_table == *tables) *tables = nullptr;
thd->lex->chop_off_not_own_tables();
sp_remove_not_own_routines(thd->lex);
for (Table_ref *tr = *tables; tr != nullptr; tr = tr->next_global) {
if (tr->is_derived() || tr->is_table_function() ||
tr->is_recursive_reference())
continue;
if (!tr->is_view()) tr->table = nullptr;
tr->mdl_request.ticket = nullptr;
}
/*
No need to commit/rollback the statement transaction: it's
either not started or we're filling in an INFORMATION_SCHEMA
table on the fly, and thus mustn't manipulate with the
transaction of the enclosing statement.
*/
assert(thd->get_transaction()->is_empty(Transaction_ctx::STMT) ||
(thd->state_flags & Open_tables_state::BACKUPS_AVAIL));
close_thread_tables(thd);
thd->mdl_context.rollback_to_savepoint(start_of_statement_svp);
}
/**
Open a single table without table caching and don't add it to
THD::open_tables. Depending on the 'add_to_temporary_tables_list' value,
the opened TABLE instance will be added to THD::temporary_tables list.
@param thd Thread context.
@param path Path (without .frm)
@param db Database name.
@param table_name Table name.
@param add_to_temporary_tables_list Specifies if the opened TABLE
instance should be linked into
THD::temporary_tables list.
@param open_in_engine Indicates that we need to open table
in storage engine in addition to
constructing TABLE object for it.
@param table_def A data-dictionary Table-object describing
table to be used for opening.
@note This function is used:
- by alter_table() to open a temporary table;
- when creating a temporary table with CREATE TEMPORARY TABLE.
@return TABLE instance for opened table.
@retval NULL on error.
*/
TABLE *open_table_uncached(THD *thd, const char *path, const char *db,
const char *table_name,
bool add_to_temporary_tables_list,
bool open_in_engine, const dd::Table &table_def) {
TABLE *tmp_table;
TABLE_SHARE *share;
char cache_key[MAX_DBKEY_LENGTH], *saved_cache_key, *tmp_path;
size_t key_length;
DBUG_TRACE;
DBUG_PRINT("enter", ("table: '%s'.'%s' path: '%s' server_id: %u "
"pseudo_thread_id: %lu",
db, table_name, path, (uint)thd->server_id,
(ulong)thd->variables.pseudo_thread_id));
/* Create the cache_key for temporary tables */
key_length = create_table_def_key_tmp(thd, db, table_name, cache_key);
if (!(tmp_table = (TABLE *)my_malloc(
key_memory_TABLE,
sizeof(*tmp_table) + sizeof(*share) + strlen(path) + 1 + key_length,
MYF(MY_WME))))
return nullptr; /* purecov: inspected */
#ifndef NDEBUG
// In order to let purge thread callback call open_table_uncached()
// we cannot grab LOCK_open here, as that will cause a deadlock.
// The assert below safeguards against opening a table which is
// already found in the table definition cache. Iff the table will
// be opened in the SE below, we may get two conflicting copies of
// SE private data in the two table_shares.
// By only grabbing LOCK_open and check the assert only when
// open_in_engine is true, we safeguard the engine private data while
// also allowing the purge threads callbacks since they always call
// with open_in_engine=false.
if (open_in_engine) {
mysql_mutex_lock(&LOCK_open);
assert(table_def_cache->count(string(cache_key, key_length)) == 0);
mysql_mutex_unlock(&LOCK_open);
}
#endif
share = (TABLE_SHARE *)(tmp_table + 1);
tmp_path = (char *)(share + 1);
saved_cache_key = my_stpcpy(tmp_path, path) + 1;
memcpy(saved_cache_key, cache_key, key_length);
init_tmp_table_share(thd, share, saved_cache_key, key_length,
strend(saved_cache_key) + 1, tmp_path, nullptr);
if (open_table_def(thd, share, table_def)) {
/* No need to lock share->mutex as this is not needed for tmp tables */
free_table_share(share);
destroy(tmp_table);
my_free(tmp_table);
return nullptr;
}
#ifdef HAVE_PSI_TABLE_INTERFACE
share->m_psi = PSI_TABLE_CALL(get_table_share)(true, share);
#else
share->m_psi = NULL;
#endif
if (open_table_from_share(
thd, share, table_name,
open_in_engine
? (uint)(HA_OPEN_KEYFILE | HA_OPEN_RNDFILE | HA_GET_INDEX)
: 0,
EXTRA_RECORD, ha_open_options, tmp_table,
/*
Set "is_create_table" if the table does not
exist in SE
*/
(open_in_engine ? false : true), &table_def)) {
/* No need to lock share->mutex as this is not needed for tmp tables */
free_table_share(share);
destroy(tmp_table);
my_free(tmp_table);
return nullptr;
}
tmp_table->reginfo.lock_type = TL_WRITE; // Simulate locked
share->tmp_table =
(tmp_table->file->has_transactions() ? TRANSACTIONAL_TMP_TABLE
: NON_TRANSACTIONAL_TMP_TABLE);
if (add_to_temporary_tables_list) {
tmp_table->set_binlog_drop_if_temp(
!thd->is_current_stmt_binlog_disabled() &&
!thd->is_current_stmt_binlog_format_row());
/* growing temp list at the head */
tmp_table->next = thd->temporary_tables;
if (tmp_table->next) tmp_table->next->prev = tmp_table;
thd->temporary_tables = tmp_table;
thd->temporary_tables->prev = nullptr;
if (thd->slave_thread) {
++atomic_replica_open_temp_tables;
++thd->rli_slave->get_c_rli()->atomic_channel_open_temp_tables;
}
}
tmp_table->pos_in_table_list = nullptr;
tmp_table->set_created();
DBUG_PRINT("tmptable", ("opened table: '%s'.'%s' %p", tmp_table->s->db.str,
tmp_table->s->table_name.str, tmp_table));
return tmp_table;
}
/**
Delete a temporary table.
@param thd Thread handle
@param base Handlerton for table to be deleted.
@param path Path to the table to be deleted (without
an extension).
@param table_def dd::Table object describing temporary table
to be deleted.
@retval false - success.
@retval true - failure.
*/
bool rm_temporary_table(THD *thd, handlerton *base, const char *path,
const dd::Table *table_def) {
bool error = false;
handler *file;
DBUG_TRACE;
file = get_new_handler((TABLE_SHARE *)nullptr,
table_def->partition_type() != dd::Table::PT_NONE,
thd->mem_root, base);
if (file && file->ha_delete_table(path, table_def)) {
error = true;
LogErr(WARNING_LEVEL, ER_FAILED_TO_REMOVE_TEMP_TABLE, path, my_errno());
}
destroy(file);
return error;
}
/*****************************************************************************
* The following find_field_in_XXX procedures implement the core of the
* name resolution functionality. The entry point to resolve a column name in a
* list of tables is 'find_field_in_tables'. It calls 'find_field_in_table_ref'
* for each table reference. In turn, depending on the type of table reference,
* 'find_field_in_table_ref' calls one of the 'find_field_in_XXX' procedures
* below specific for the type of table reference.
*
* @todo: Refactor the error handling system used by these functions, so that
* it is clear when an error is reported and when an empty reference
* is returned.
*
******************************************************************************/
/* Special Field pointers as return values of find_field_in_XXX functions. */
Field *not_found_field = (Field *)0x1;
Field *view_ref_found = (Field *)0x2;
#define WRONG_GRANT (Field *)-1
/**
Find a temporary table specified by Table_ref instance in the cache and
prepare its TABLE instance for use.
This function tries to resolve this table in the list of temporary tables
of this thread. Temporary tables are thread-local and "shadow" base
tables with the same name.
@note In most cases one should use open_temporary_tables() instead
of this call.
@note One should finalize process of opening temporary table for table
list element by calling open_and_process_table(). This function
is responsible for table version checking and handling of merge
tables.
@note We used to check global_read_lock before opening temporary tables.
However, that limitation was artificial and is removed now.
@return Error status.
@retval false On success. If a temporary table exists for the given
key, tl->table is set.
@retval true On error. my_error() has been called.
*/
bool open_temporary_table(THD *thd, Table_ref *tl) {
DBUG_TRACE;
DBUG_PRINT("enter", ("table: '%s'.'%s'", tl->db, tl->table_name));
/*
Code in open_table() assumes that Table_ref::table can
be non-zero only for pre-opened temporary tables.
*/
assert(tl->table == nullptr);
/*
This function should not be called for cases when derived or I_S
tables can be met since table list elements for such tables can
have invalid db or table name.
Instead open_temporary_tables() should be used.
*/
assert(!tl->is_view_or_derived() && !tl->schema_table);
if (tl->open_type == OT_BASE_ONLY) {
DBUG_PRINT("info", ("skip_temporary is set"));
return false;
}
TABLE *table = find_temporary_table(thd, tl);
// Access to temporary tables is disallowed in XA transactions in
// xa_detach_on_prepare=ON mode.
if ((tl->open_type == OT_TEMPORARY_ONLY ||
(table && table->s->tmp_table != NO_TMP_TABLE)) &&
is_xa_tran_detached_on_prepare(thd) &&
thd->get_transaction()->xid_state()->check_in_xa(false)) {
my_error(ER_XA_TEMP_TABLE, MYF(0));
return true;
}
if (!table) {
if (tl->open_type == OT_TEMPORARY_ONLY &&
tl->open_strategy == Table_ref::OPEN_NORMAL) {
my_error(ER_NO_SUCH_TABLE, MYF(0), tl->db, tl->table_name);
return true;
}
return false;
}
if (tl->partition_names) {
/* Partitioned temporary tables is not supported. */
assert(!table->part_info);
my_error(ER_PARTITION_CLAUSE_ON_NONPARTITIONED, MYF(0));
return true;
}
if (table->query_id) {
/*
We're trying to use the same temporary table twice in a query.
Right now we don't support this because a temporary table is always
represented by only one TABLE object in THD, and it can not be
cloned. Emit an error for an unsupported behaviour.
*/
DBUG_PRINT("error", ("query_id: %lu server_id: %u pseudo_thread_id: %lu",
(ulong)table->query_id, (uint)thd->server_id,
(ulong)thd->variables.pseudo_thread_id));
my_error(ER_CANT_REOPEN_TABLE, MYF(0), table->alias);
return true;
}
table->query_id = thd->query_id;
thd->thread_specific_used = true;
tl->set_updatable(); // It is not derived table nor non-updatable VIEW.
tl->set_insertable();
table->reset();
table->init(thd, tl);
DBUG_PRINT("info", ("Using temporary table"));
return false;
}
/**
Pre-open temporary tables corresponding to table list elements.
@note One should finalize process of opening temporary tables
by calling open_tables(). This function is responsible
for table version checking and handling of merge tables.
@return Error status.
@retval false On success. If a temporary tables exists for the
given element, tl->table is set.
@retval true On error. my_error() has been called.
*/
bool open_temporary_tables(THD *thd, Table_ref *tl_list) {
Table_ref *first_not_own = thd->lex->first_not_own_table();
DBUG_TRACE;
for (Table_ref *tl = tl_list; tl && tl != first_not_own;
tl = tl->next_global) {
// Placeholder tables are processed during query execution
if (tl->is_view_or_derived() || tl->is_table_function() ||
tl->schema_table != nullptr || tl->is_recursive_reference())
continue;
if (open_temporary_table(thd, tl)) return true;
}
return false;
}
/*
Find a field by name in a view that uses merge algorithm.
SYNOPSIS
find_field_in_view()
thd thread handler
table_list view to search for 'name'
name name of field
ref expression substituted in VIEW should be passed
using this reference (return view_ref_found)
register_tree_change true if ref is not stack variable and we
need register changes in item tree
RETURN
0 field is not found
view_ref_found found value in VIEW (real result is in *ref)
# pointer to field - only for schema table fields
*/
static Field *find_field_in_view(THD *thd, Table_ref *table_list,
const char *name, Item **ref,
bool register_tree_change) {
DBUG_TRACE;
DBUG_PRINT("enter", ("view: '%s', field name: '%s', ref %p",
table_list->alias, name, ref));
Field_iterator_view field_it;
field_it.set(table_list);
assert(table_list->schema_table_reformed ||
(ref != nullptr && table_list->is_merged()));
for (; !field_it.end_of_fields(); field_it.next()) {
if (!my_strcasecmp(system_charset_info, field_it.name(), name)) {
Item *item;
{
/*
Use own arena for Prepared Statements or data will be freed after
PREPARE.
*/
Prepared_stmt_arena_holder ps_arena_holder(
thd, register_tree_change &&
thd->stmt_arena->is_stmt_prepare_or_first_stmt_execute());
/*
create_item() may, or may not create a new Item, depending on
the column reference. See create_view_field() for details.
*/
item = field_it.create_item(thd);
if (!item) return nullptr;
}
/*
*ref != NULL means that *ref contains the item that we need to
replace. If the item was aliased by the user, set the alias to
the replacing item.
We need to set alias on both ref itself and on ref real item.
*/
if (*ref && !(*ref)->item_name.is_autogenerated()) {
item->item_name = (*ref)->item_name;
item->real_item()->item_name = (*ref)->item_name;
}
*ref = item;
// WL#6570 remove-after-qa
assert(thd->stmt_arena->is_regular() || !thd->lex->is_exec_started());
return view_ref_found;
}
}
return nullptr;
}
/**
Find field by name in a NATURAL/USING join table reference.
@param thd thread handler
@param table_ref table reference to search
@param name name of field
@param [in,out] ref if 'name' is resolved to a view field, ref is
set to point to the found view field
@param register_tree_change true if ref is not stack variable and we
need register changes in item tree
@param [out] actual_table The original table reference where the field
belongs - differs from 'table_list' only for
NATURAL/USING joins
DESCRIPTION
Search for a field among the result fields of a NATURAL/USING join.
Notice that this procedure is called only for non-qualified field
names. In the case of qualified fields, we search directly the base
tables of a natural join.
Sometimes when a field is found, it is checked for privileges according to
THD::want_privilege and marked according to THD::mark_used_columns.
But it is unclear when, so caller generally has to do the same.
RETURN
NULL if the field was not found
WRONG_GRANT if no access rights to the found field
# Pointer to the found Field
*/
static Field *find_field_in_natural_join(THD *thd, Table_ref *table_ref,
const char *name, Item **ref,
bool register_tree_change,
Table_ref **actual_table) {
List_iterator_fast<Natural_join_column> field_it(*(table_ref->join_columns));
Natural_join_column *nj_col, *curr_nj_col;
Field *found_field = nullptr;
DBUG_TRACE;
DBUG_PRINT("enter", ("field name: '%s', ref %p", name, ref));
assert(table_ref->is_natural_join && table_ref->join_columns);
assert(*actual_table == nullptr);
for (nj_col = nullptr, curr_nj_col = field_it++; curr_nj_col;
curr_nj_col = field_it++) {
if (!my_strcasecmp(system_charset_info, curr_nj_col->name(), name)) {
if (nj_col) {
my_error(ER_NON_UNIQ_ERROR, MYF(0), name, thd->where);
return nullptr;
}
nj_col = curr_nj_col;
}
}
if (!nj_col) return nullptr;
if (nj_col->view_field) {
Item *item;
{
Prepared_stmt_arena_holder ps_arena_holder(thd, register_tree_change);
/*
create_item() may, or may not create a new Item, depending on the
column reference. See create_view_field() for details.
*/
item = nj_col->create_item(thd);
if (!item) return nullptr;
}
/*
*ref != NULL means that *ref contains the item that we need to
replace. If the item was aliased by the user, set the alias to
the replacing item.
We need to set alias on both ref itself and on ref real item.
*/
if (*ref && !(*ref)->item_name.is_autogenerated()) {
item->item_name = (*ref)->item_name;
item->real_item()->item_name = (*ref)->item_name;
}
assert(nj_col->table_field == nullptr);
if (nj_col->table_ref->schema_table_reformed) {
/*
Translation table items are always Item_fields and fixed
already('mysql_schema_table' function). So we can return
->field. It is used only for 'show & where' commands.
*/
return ((Item_field *)(nj_col->view_field->item))->field;
}
*ref = item;
// WL#6570 remove-after-qa
assert(thd->stmt_arena->is_regular() || !thd->lex->is_exec_started());
found_field = view_ref_found;
} else {
/* This is a base table. */
assert(nj_col->view_field == nullptr);
/*
This fix_fields is not necessary (initially this item is fixed by
the Item_field constructor; after reopen_tables the Item_func_eq
calls fix_fields on that item), it's just a check during table
reopening for columns that was dropped by the concurrent connection.
*/
if (!nj_col->table_field->fixed &&
nj_col->table_field->fix_fields(thd, (Item **)&nj_col->table_field)) {
DBUG_PRINT("info",
("column '%s' was dropped by the concurrent connection",
nj_col->table_field->item_name.ptr()));
return nullptr;
}
assert(nj_col->table_ref->table == nj_col->table_field->field->table);
found_field = nj_col->table_field->field;
}
*actual_table = nj_col->table_ref;
return found_field;
}
/**
Find field by name in a base table.
No privileges are checked, and the column is not marked in read_set/write_set.
@param table table where to search for the field
@param name name of field
@param allow_rowid do allow finding of "_rowid" field?
@param[out] field_index_ptr position in field list (used to speedup
lookup for fields in prepared tables)
@retval NULL field is not found
@retval != NULL pointer to field
*/
Field *find_field_in_table(TABLE *table, const char *name, bool allow_rowid,
uint *field_index_ptr) {
DBUG_TRACE;
DBUG_PRINT("enter", ("table: '%s', field name: '%s'", table->alias, name));
Field **field_ptr = nullptr, *field;
if (!(field_ptr = table->field)) return nullptr;
for (; *field_ptr; ++field_ptr) {
// NOTE: This should probably be strncollsp() instead of my_strcasecmp();
// in particular, Ñ != N for my_strcasecmp(), which is not according to the
// usual ai_ci rules. However, changing it would risk breaking existing
// table definitions (which don't distinguish between N and Ñ), so we can
// only do this when actually changing the system collation.
if (!my_strcasecmp(system_charset_info, (*field_ptr)->field_name, name))
break;
}
if (field_ptr && *field_ptr) {
*field_index_ptr = field_ptr - table->field;
field = *field_ptr;
} else {
if (!allow_rowid || my_strcasecmp(system_charset_info, name, "_rowid") ||
table->s->rowid_field_offset == 0)
return (Field *)nullptr;
field = table->field[table->s->rowid_field_offset - 1];
}
return field;
}
/**
Find field in a table reference.
@param thd thread handler
@param table_list table reference to search
@param name name of field
@param length length of field name
@param item_name name of item if it will be created (VIEW)
@param db_name optional database name that qualifies the field
@param table_name optional table name that qualifies the field
@param[in,out] ref if 'name' is resolved to a view field, ref
is set to point to the found view field
@param want_privilege privileges to check for column
= 0: no privilege checking is needed
@param allow_rowid do allow finding of "_rowid" field?
@param field_index_ptr position in field list (used to
speedup lookup for fields in prepared tables)
@param register_tree_change TRUE if ref is not stack variable and we
need register changes in item tree
@param[out] actual_table the original table reference where the field
belongs - differs from 'table_list' only for
NATURAL_USING joins.
Find a field in a table reference depending on the type of table
reference. There are three types of table references with respect
to the representation of their result columns:
- an array of Field_translator objects for MERGE views and some
information_schema tables,
- an array of Field objects (and possibly a name hash) for stored
tables,
- a list of Natural_join_column objects for NATURAL/USING joins.
This procedure detects the type of the table reference 'table_list'
and calls the corresponding search routine.
The function checks column-level privileges for the found field
according to argument want_privilege.
The function marks the column in corresponding table's read set or
write set according to THD::mark_used_columns.
@retval NULL field is not found
@retval view_ref_found found value in VIEW (real result is in *ref)
@retval otherwise pointer to field
*/
Field *find_field_in_table_ref(THD *thd, Table_ref *table_list,
const char *name, size_t length,
const char *item_name, const char *db_name,
const char *table_name, Item **ref,
Access_bitmask want_privilege, bool allow_rowid,
uint *field_index_ptr, bool register_tree_change,
Table_ref **actual_table) {
Field *fld;
DBUG_TRACE;
assert(table_list->alias);
assert(name);
assert(item_name);
DBUG_PRINT("enter", ("table: '%s' field name: '%s' item name: '%s' ref %p",
table_list->alias, name, item_name, ref));
/*
Check that the table and database that qualify the current field name
are the same as the table reference we are going to search for the field.
Exclude from the test below nested joins because the columns in a
nested join generally originate from different tables. Nested joins
also have no table name, except when a nested join is a merge view
or an information schema table.
We include explicitly table references with a 'field_translation' table,
because if there are views over natural joins we don't want to search
inside the view, but we want to search directly in the view columns
which are represented as a 'field_translation'.
TODO: Ensure that table_name, db_name and tables->db always points to
something !
*/
if (/* Exclude nested joins. */
(!table_list->nested_join ||
/* Include merge views and information schema tables. */
table_list->field_translation) &&
/*
Test if the field qualifiers match the table reference we plan
to search.
*/
table_name && table_name[0] &&
(my_strcasecmp(table_alias_charset, table_list->alias, table_name) ||
(db_name && db_name[0] && table_list->db && table_list->db[0] &&
(table_list->schema_table
? my_strcasecmp(system_charset_info, db_name, table_list->db)
: strcmp(db_name, table_list->db)))))
return nullptr;
*actual_table = nullptr;
if (table_list->field_translation) {
/* 'table_list' is a view or an information schema table. */
if ((fld = find_field_in_view(thd, table_list, name, ref,
register_tree_change)))
*actual_table = table_list;
} else if (!table_list->nested_join) {
/* 'table_list' is a stored table. */
assert(table_list->table);
if ((fld = find_field_in_table(table_list->table, name, allow_rowid,
field_index_ptr)))
*actual_table = table_list;
} else {
/*
'table_list' is a NATURAL/USING join, or an operand of such join that
is a nested join itself.
If the field name we search for is qualified, then search for the field
in the table references used by NATURAL/USING the join.
*/
if (table_name && table_name[0]) {
for (Table_ref *table : table_list->nested_join->m_tables) {
if ((fld = find_field_in_table_ref(
thd, table, name, length, item_name, db_name, table_name, ref,
want_privilege, allow_rowid, field_index_ptr,
register_tree_change, actual_table)))
return fld;
}
return nullptr;
}
/*
Non-qualified field, search directly in the result columns of the
natural join. The condition of the outer IF is true for the top-most
natural join, thus if the field is not qualified, we will search
directly the top-most NATURAL/USING join.
*/
fld = find_field_in_natural_join(thd, table_list, name, ref,
register_tree_change, actual_table);
}
if (fld) {
// Check if there are sufficient privileges to the found field.
if (want_privilege) {
if (fld != view_ref_found) {
if (check_column_grant_in_table_ref(thd, *actual_table, name, length,
want_privilege))
return WRONG_GRANT;
} else {
assert(ref && *ref && (*ref)->fixed);
assert(*actual_table == (down_cast<Item_ident *>(*ref))->cached_table);
Column_privilege_tracker tracker(thd, want_privilege);
if ((*ref)->walk(&Item::check_column_privileges, enum_walk::PREFIX,
(uchar *)thd))
return WRONG_GRANT;
}
}
/*
Get read_set correct for this field so that the handler knows that
this field is involved in the query and gets retrieved.
*/
if (fld == view_ref_found) {
Mark_field mf(thd->mark_used_columns);
(*ref)->walk(&Item::mark_field_in_map, enum_walk::SUBQUERY_POSTFIX,
(uchar *)&mf);
} else // surely fld != NULL (see outer if())
fld->table->mark_column_used(fld, thd->mark_used_columns);
}
return fld;
}
/*
Find field in table, no side effects, only purpose is to check for field
in table object and get reference to the field if found.
SYNOPSIS
find_field_in_table_sef()
table table where to find
name Name of field searched for
RETURN
0 field is not found
# pointer to field
*/
Field *find_field_in_table_sef(TABLE *table, const char *name) {
Field **field_ptr = nullptr;
if (!(field_ptr = table->field)) return nullptr;
for (; *field_ptr; ++field_ptr) {
// NOTE: See comment on the same call in find_field_in_table().
if (!my_strcasecmp(system_charset_info, (*field_ptr)->field_name, name))
break;
}
if (field_ptr)
return *field_ptr;
else
return (Field *)nullptr;
}
/*
Find field in table list.
SYNOPSIS
find_field_in_tables()
thd pointer to current thread structure
item field item that should be found
first_table list of tables to be searched for item
last_table end of the list of tables to search for item. If NULL
then search to the end of the list 'first_table'.
ref if 'item' is resolved to a view field, ref is set to
point to the found view field
report_error Degree of error reporting:
- IGNORE_ERRORS then do not report any error
- IGNORE_EXCEPT_NON_UNIQUE report only non-unique
fields, suppress all other errors
- REPORT_EXCEPT_NON_UNIQUE report all other errors
except when non-unique fields were found
- REPORT_ALL_ERRORS
want_privilege column privileges to check
= 0: no need to check privileges
register_tree_change true if ref is not a stack variable and we
to need register changes in item tree
RETURN VALUES
0 If error: the found field is not unique, or there are
no sufficient access privileges for the found field,
or the field is qualified with non-existing table.
not_found_field The function was called with report_error ==
(IGNORE_ERRORS || IGNORE_EXCEPT_NON_UNIQUE) and a
field was not found.
view_ref_found View field is found, item passed through ref parameter
found field If a item was resolved to some field
*/
Field *find_field_in_tables(THD *thd, Item_ident *item, Table_ref *first_table,
Table_ref *last_table, Item **ref,
find_item_error_report_type report_error,
Access_bitmask want_privilege,
bool register_tree_change) {
Field *found = nullptr;
const char *db = item->db_name;
const char *table_name = item->table_name;
const char *name = item->field_name;
size_t length = strlen(name);
uint field_index;
char name_buff[NAME_LEN + 1];
Table_ref *actual_table;
bool allow_rowid;
if (!table_name || !table_name[0]) {
table_name = nullptr; // For easier test
db = nullptr;
}
allow_rowid = table_name || (first_table && !first_table->next_local);
if (item->cached_table) {
/*
This shortcut is used by prepared statements. We assume that
Table_ref *first_table is not changed during query execution (which
is true for all queries except RENAME but luckily RENAME doesn't
use fields...) so we can rely on reusing pointer to its member.
With this optimization we also miss case when addition of one more
field makes some prepared query ambiguous and so erroneous, but we
accept this trade off.
*/
Table_ref *table_ref = item->cached_table;
/*
@todo WL#6570 - is this reasonable???
Also refactor this code to replace "cached_table" with "table_ref" -
as there is no longer need for more than one resolving, hence
no "caching" as well.
*/
if (item->type() == Item::FIELD_ITEM)
field_index = down_cast<Item_field *>(item)->field_index;
/*
The condition (table_ref->view == NULL) ensures that we will call
find_field_in_table even in the case of information schema tables
when table_ref->field_translation != NULL.
*/
if (table_ref->table && !table_ref->is_view()) {
found = find_field_in_table(table_ref->table, name, true, &field_index);
// Check if there are sufficient privileges to the found field.
if (found && want_privilege &&
check_column_grant_in_table_ref(thd, table_ref, name, length,
want_privilege))
found = WRONG_GRANT;
if (found && found != WRONG_GRANT)
table_ref->table->mark_column_used(found, thd->mark_used_columns);
} else {
found = find_field_in_table_ref(thd, table_ref, name, length,
item->item_name.ptr(), nullptr, nullptr,
ref, want_privilege, true, &field_index,
register_tree_change, &actual_table);
}
if (found) {
if (found == WRONG_GRANT) return nullptr;
// @todo WL#6570 move this assignment to a more strategic place?
if (item->type() == Item::FIELD_ITEM)
down_cast<Item_field *>(item)->field_index = field_index;
return found;
}
}
if (db && (lower_case_table_names || is_infoschema_db(db, strlen(db)))) {
/*
convert database to lower case for comparison.
We can't do this in Item_field as this would change the
'name' of the item which may be used in the select list
The 'information_schema' name is treated as case-insensitive
identifier when specified in FROM clause even in
lower_case_table_names=0. We lowercase the 'information_schema' name
below to treat it as case-insensitive even when it is referred in WHERE
or SELECT clause.
*/
strmake(name_buff, db, sizeof(name_buff) - 1);
my_casedn_str(files_charset_info, name_buff);
db = name_buff;
}
if (first_table && first_table->query_block &&
first_table->query_block->end_lateral_table)
last_table = first_table->query_block->end_lateral_table;
else if (last_table)
last_table = last_table->next_name_resolution_table;
Table_ref *cur_table;
for (cur_table = first_table; cur_table != last_table;
cur_table = cur_table->next_name_resolution_table) {
Field *cur_field = find_field_in_table_ref(
thd, cur_table, name, length, item->item_name.ptr(), db, table_name,
ref, want_privilege, allow_rowid, &field_index, register_tree_change,
&actual_table);
if ((cur_field == nullptr && thd->is_error()) || cur_field == WRONG_GRANT)
return nullptr;
if (cur_field) {
/*
Store the original table of the field, which may be different from
cur_table in the case of NATURAL/USING join.
*/
item->cached_table =
(!actual_table->cacheable_table || found) ? nullptr : actual_table;
// @todo WL#6570 move this assignment to a more strategic place?
if (item->type() == Item::FIELD_ITEM)
down_cast<Item_field *>(item)->field_index = field_index;
assert(thd->where);
/*
If we found a fully qualified field we return it directly as it can't
have duplicates.
*/
if (db) return cur_field;
if (found) {
if (report_error == REPORT_ALL_ERRORS ||
report_error == IGNORE_EXCEPT_NON_UNIQUE)
my_error(ER_NON_UNIQ_ERROR, MYF(0),
table_name ? item->full_name() : name, thd->where);
return (Field *)nullptr;
}
found = cur_field;
}
}
if (found) return found;
/*
If the field was qualified and there were no tables to search, issue
an error that an unknown table was given. The situation is detected
as follows: if there were no tables we wouldn't go through the loop
and cur_table wouldn't be updated by the loop increment part, so it
will be equal to the first table.
@todo revisit this logic. If the first table is a table function or lateral
derived table and contains an inner column reference in it which is not
found, cur_table==first_table, even though there _were_ tables to search.
*/
if (table_name && (cur_table == first_table) &&
(report_error == REPORT_ALL_ERRORS ||
report_error == REPORT_EXCEPT_NON_UNIQUE)) {
char buff[NAME_LEN * 2 + 2];
if (db && db[0]) {
strxnmov(buff, sizeof(buff) - 1, db, ".", table_name, NullS);
table_name = buff;
}
my_error(ER_UNKNOWN_TABLE, MYF(0), table_name, thd->where);
} else {
if (report_error == REPORT_ALL_ERRORS ||
report_error == REPORT_EXCEPT_NON_UNIQUE) {
/* We now know that this column does not exist in any table_list
of the query. If user does not have grant, then we should throw
error stating 'access denied'. If user does have right then we can
give proper error like column does not exist. Following is check
to see if column has wrong grants and avoids error like 'bad field'
and throw column access error.
*/
if (!first_table || (want_privilege == 0) ||
!check_column_grant_in_table_ref(thd, first_table, name, length,
want_privilege))
my_error(ER_BAD_FIELD_ERROR, MYF(0), item->full_name(), thd->where);
} else
found = not_found_field;
}
return found;
}
/**
Find Item in list of items (find_field_in_tables analog)
@param thd pointer to current thread
@param find Item to find
@param items List of items to search
@param[out] found Returns the pointer to the found item, or nullptr
if the item was not found
@param[out] counter Returns number of found item in list
@param[out] resolution Set to the resolution type if the item is found
(it says whether the item is resolved against an
alias, or as a field name without alias, or as a
field hidden by alias, or ignoring alias)
@note "counter" and "resolution" are undefined unless "found" identifies
an item.
@returns true if error, false if success
*/
bool find_item_in_list(THD *thd, Item *find, mem_root_deque<Item *> *items,
Item ***found, uint *counter,
enum_resolution_type *resolution) {
*found = nullptr;
*resolution = NOT_RESOLVED;
Item **found_unaliased = nullptr;
bool found_unaliased_non_uniq = false;
uint unaliased_counter = 0;
Item_ident *const find_ident =
find->type() == Item::FIELD_ITEM || find->type() == Item::REF_ITEM
? down_cast<Item_ident *>(find)
: nullptr;
/*
Some items, such as Item_aggregate_ref, do not have a name and hence
can never be found.
*/
assert(find_ident == nullptr || find_ident->field_name != nullptr);
int i = 0;
for (auto it = VisibleFields(*items).begin();
it != VisibleFields(*items).end(); ++it, ++i) {
Item *item = *it;
if (find_ident != nullptr &&
item->real_item()->type() == Item::FIELD_ITEM) {
Item_ident *item_field = down_cast<Item_ident *>(item);
/*
In case of group_concat() with ORDER BY condition in the QUERY
item_field can be field of temporary table without item name
(if this field created from expression argument of group_concat()),
=> we have to check presence of name before compare
*/
if (!item_field->item_name.is_set()) continue;
if (find_ident->table_name != nullptr) {
/*
If table name is specified we should find field 'field_name' in
table 'table_name'. According to SQL-standard we should ignore
aliases in this case.
Since we should NOT prefer fields from the select list over
other fields from the tables participating in this select in
case of ambiguity we have to do extra check outside this function.
*/
if (!my_strcasecmp(system_charset_info, item_field->field_name,
find_ident->field_name) &&
(item_field->table_name != nullptr &&
!my_strcasecmp(table_alias_charset, item_field->table_name,
find_ident->table_name)) &&
(find_ident->db_name == nullptr ||
(item_field->db_name != nullptr &&
!strcmp(item_field->db_name, find_ident->db_name)))) {
if (found_unaliased) {
if ((*found_unaliased)->eq(item, false)) continue;
/*
Two matching fields in select list.
We already can bail out because we are searching through
unaliased names only and will have duplicate error anyway.
*/
my_error(ER_NON_UNIQ_ERROR, MYF(0), find->full_name(), thd->where);
return true;
}
found_unaliased = &*it;
unaliased_counter = i;
*resolution = RESOLVED_IGNORING_ALIAS;
if (find_ident->db_name != nullptr) break; // Perfect match
}
} else {
int fname_cmp =
my_strcasecmp(system_charset_info, item_field->field_name,
find_ident->field_name);
if (item_field->item_name.eq_safe(find_ident->field_name)) {
/*
If table name was not given we should scan through aliases
and non-aliased fields first. We are also checking unaliased
name of the field in then next else-if, to be able to find
instantly field (hidden by alias) if no suitable alias or
non-aliased field was found.
*/
if (*found != nullptr) {
if ((**found)->eq(item, false)) continue; // Same field twice
my_error(ER_NON_UNIQ_ERROR, MYF(0), find->full_name(), thd->where);
return true;
}
*found = &*it;
*counter = i;
*resolution =
fname_cmp ? RESOLVED_AGAINST_ALIAS : RESOLVED_WITH_NO_ALIAS;
} else if (!fname_cmp) {
/*
We will use non-aliased field or react on such ambiguities only if
we won't be able to find aliased field.
Again if we have ambiguity with field outside of select list
we should prefer fields from select list.
*/
if (found_unaliased) {
if ((*found_unaliased)->eq(item, false))
continue; // Same field twice
found_unaliased_non_uniq = true;
}
found_unaliased = &*it;
unaliased_counter = i;
}
}
} else if (find_ident == nullptr || find_ident->table_name == nullptr ||
is_rollup_group_wrapper(item)) {
// Unwrap rollup wrappers, if any
item = unwrap_rollup_group(item);
find = unwrap_rollup_group(find);
if (find_ident != nullptr && item->item_name.eq_safe(find->item_name)) {
*found = &*it;
*counter = i;
*resolution = RESOLVED_AGAINST_ALIAS;
break;
} else if (find->eq(item, false)) {
*found = &*it;
*counter = i;
*resolution = RESOLVED_IGNORING_ALIAS;
break;
}
} else if (find_ident != nullptr && find_ident->table_name != nullptr &&
item->type() == Item::REF_ITEM &&
down_cast<Item_ref *>(item)->ref_type() == Item_ref::VIEW_REF) {
/*
TODO:Here we process prefixed view references only. What we should
really do is process all types of Item_refs. But this will currently
lead to a clash with the way references to outer SELECTs (from the
HAVING clause) are handled in e.g. :
SELECT 1 FROM t1 AS t1_o GROUP BY a
HAVING (SELECT t1_o.a FROM t1 AS t1_i GROUP BY t1_i.a LIMIT 1).
Processing all Item_refs here will cause t1_o.a to resolve to itself.
We still need to process the special case of Item_view_ref
because in the context of views they have the same meaning as
Item_field for tables.
*/
Item_ident *item_ref = down_cast<Item_ident *>(item);
if (!my_strcasecmp(system_charset_info, item_ref->field_name,
find_ident->field_name) &&
item_ref->table_name != nullptr &&
!my_strcasecmp(table_alias_charset, item_ref->table_name,
find_ident->table_name) &&
(find_ident->db_name == nullptr ||
(item_ref->db_name != nullptr &&
!strcmp(item_ref->db_name, find_ident->db_name)))) {
*found = &*it;
*counter = i;
*resolution = RESOLVED_IGNORING_ALIAS;
break;
}
}
}
if (*found == nullptr) {
if (found_unaliased_non_uniq) {
my_error(ER_NON_UNIQ_ERROR, MYF(0), find->full_name(), thd->where);
return true;
}
if (found_unaliased) {
*found = found_unaliased;
*counter = unaliased_counter;
*resolution = RESOLVED_BEHIND_ALIAS;
}
}
return false;
}
/*
Test if a string is a member of a list of strings.
SYNOPSIS
test_if_string_in_list()
find the string to look for
str_list a list of strings to be searched
DESCRIPTION
Sequentially search a list of strings for a string, and test whether
the list contains the same string.
RETURN
true if find is in str_list
false otherwise
*/
static bool test_if_string_in_list(const char *find, List<String> *str_list) {
List_iterator<String> str_list_it(*str_list);
String *curr_str;
size_t find_length = strlen(find);
while ((curr_str = str_list_it++)) {
if (find_length != curr_str->length()) continue;
if (!my_strcasecmp(system_charset_info, find, curr_str->ptr())) return true;
}
return false;
}
/*
Create a new name resolution context for an item so that it is
being resolved in a specific table reference.
SYNOPSIS
set_new_item_local_context()
thd pointer to current thread
item item for which new context is created and set
table_ref table ref where an item showld be resolved
DESCRIPTION
Create a new name resolution context for an item, so that the item
is resolved only the supplied 'table_ref'.
RETURN
false if all OK
true otherwise
*/
static bool set_new_item_local_context(THD *thd, Item_ident *item,
Table_ref *table_ref) {
Name_resolution_context *context;
if (!(context = new (thd->mem_root) Name_resolution_context))
return true; /* purecov: inspected */
context->init();
context->first_name_resolution_table = context->last_name_resolution_table =
table_ref;
context->query_block = table_ref->query_block;
context->next_context = table_ref->query_block->first_context;
table_ref->query_block->first_context = context;
item->context = context;
return false;
}
/*
Find and mark the common columns of two table references.
SYNOPSIS
mark_common_columns()
thd [in] current thread
table_ref_1 [in] the first (left) join operand
table_ref_2 [in] the second (right) join operand
using_fields [in] if the join is JOIN...USING - the join columns,
if NATURAL join, then NULL
found_using_fields [out] number of fields from the USING clause that were
found among the common fields
DESCRIPTION
The procedure finds the common columns of two relations (either
tables or intermediate join results), and adds an equi-join condition
to the ON clause of 'table_ref_2' for each pair of matching columns.
If some of table_ref_XXX represents a base table or view, then we
create new 'Natural_join_column' instances for each column
reference and store them in the 'join_columns' of the table
reference.
IMPLEMENTATION
The procedure assumes that store_natural_using_join_columns() was
called for the previous level of NATURAL/USING joins.
RETURN
true error when some common column is non-unique, or out of memory
false OK
*/
static bool mark_common_columns(THD *thd, Table_ref *table_ref_1,
Table_ref *table_ref_2,
List<String> *using_fields,
uint *found_using_fields) {
Field_iterator_table_ref it_1, it_2;
Natural_join_column *nj_col_1, *nj_col_2;
bool first_outer_loop = true;
List<Field> fields;
/*
Leaf table references to which new natural join columns are added
if the leaves are != NULL.
*/
Table_ref *leaf_1 =
(table_ref_1->nested_join && !table_ref_1->is_natural_join) ? nullptr
: table_ref_1;
Table_ref *leaf_2 =
(table_ref_2->nested_join && !table_ref_2->is_natural_join) ? nullptr
: table_ref_2;
DBUG_TRACE;
DBUG_PRINT("info", ("operand_1: %s operand_2: %s", table_ref_1->alias,
table_ref_2->alias));
/*
Some hidden columns cannot be participants in NATURAL JOIN / JOIN USING:
- No system-generated hidden columns (columns defined for functional
indexes or used as keys for materialized derived tables) can be used.
- User-defined hidden columns (invisible columns) can be used in JOIN
USING .
(we need to go through get_or_create_column_ref() before calling
this method).
*/
auto is_non_participant_column = [using_fields](Field *field) {
return (field != nullptr &&
(field->is_hidden_by_system() ||
((using_fields == nullptr) && field->is_hidden_by_user())));
};
Prepared_stmt_arena_holder ps_arena_holder(thd);
*found_using_fields = 0;
for (it_1.set(table_ref_1); !it_1.end_of_fields(); it_1.next()) {
bool found = false;
const char *field_name_1;
/* true if field_name_1 is a member of using_fields */
bool is_using_column_1;
if (!(nj_col_1 = it_1.get_or_create_column_ref(thd, leaf_1))) return true;
if (is_non_participant_column(it_1.field())) continue;
field_name_1 = nj_col_1->name();
is_using_column_1 =
using_fields && test_if_string_in_list(field_name_1, using_fields);
DBUG_PRINT("info", ("field_name_1=%s.%s",
nj_col_1->table_name() ? nj_col_1->table_name() : "",
field_name_1));
/*
Find a field with the same name in table_ref_2.
Note that for the second loop, it_2.set() will iterate over
table_ref_2->join_columns and not generate any new elements or
lists.
*/
nj_col_2 = nullptr;
for (it_2.set(table_ref_2); !it_2.end_of_fields(); it_2.next()) {
Natural_join_column *cur_nj_col_2;
const char *cur_field_name_2;
if (!(cur_nj_col_2 = it_2.get_or_create_column_ref(thd, leaf_2)))
return true;
if (is_non_participant_column(it_2.field())) continue;
cur_field_name_2 = cur_nj_col_2->name();
DBUG_PRINT("info",
("cur_field_name_2=%s.%s",
cur_nj_col_2->table_name() ? cur_nj_col_2->table_name() : "",
cur_field_name_2));
/*
Compare the two columns and check for duplicate common fields.
A common field is duplicate either if it was already found in
table_ref_2 (then found == true), or if a field in table_ref_2
was already matched by some previous field in table_ref_1
(then cur_nj_col_2->is_common == true).
Note that it is too early to check the columns outside of the
USING list for ambiguity because they are not actually "referenced"
here. These columns must be checked only on unqualified reference
by name (e.g. in SELECT list).
*/
if (!my_strcasecmp(system_charset_info, field_name_1, cur_field_name_2)) {
DBUG_PRINT("info", ("match c1.is_common=%d", nj_col_1->is_common));
if (cur_nj_col_2->is_common ||
(found && (!using_fields || is_using_column_1))) {
my_error(ER_NON_UNIQ_ERROR, MYF(0), field_name_1, thd->where);
return true;
}
nj_col_2 = cur_nj_col_2;
found = true;
}
}
if (first_outer_loop && leaf_2) {
/*
Make sure that the next inner loop "knows" that all columns
are materialized already.
*/
leaf_2->is_join_columns_complete = true;
first_outer_loop = false;
}
if (!found) continue; // No matching field
/*
field_1 and field_2 have the same names. Check if they are in the USING
clause (if present), mark them as common fields, and add a new
equi-join condition to the ON clause.
*/
if (nj_col_2 && (!using_fields || is_using_column_1)) {
Item *item_1 = nj_col_1->create_item(thd);
if (!item_1) return true;
Item *item_2 = nj_col_2->create_item(thd);
if (!item_2) return true;
Field *field_1 = nj_col_1->field();
Field *field_2 = nj_col_2->field();
Item_ident *item_ident_1, *item_ident_2;
Item_func_eq *eq_cond;
fields.push_back(field_1);
fields.push_back(field_2);
/*
The created items must be of sub-classes of Item_ident.
*/
assert(item_1->type() == Item::FIELD_ITEM ||
item_1->type() == Item::REF_ITEM);
assert(item_2->type() == Item::FIELD_ITEM ||
item_2->type() == Item::REF_ITEM);
/*
We need to cast item_1,2 to Item_ident, because we need to hook name
resolution contexts specific to each item.
*/
item_ident_1 = (Item_ident *)item_1;
item_ident_2 = (Item_ident *)item_2;
/*
Create and hook special name resolution contexts to each item in the
new join condition . We need this to both speed-up subsequent name
resolution of these items, and to enable proper name resolution of
the items during the execute phase of PS.
*/
if (set_new_item_local_context(thd, item_ident_1, nj_col_1->table_ref) ||
set_new_item_local_context(thd, item_ident_2, nj_col_2->table_ref))
return true;
if (!(eq_cond = new Item_func_eq(item_ident_1, item_ident_2)))
return true; // Out of memory.
/*
Add the new equi-join condition to the ON clause. Notice that
fix_fields() is applied to all ON conditions in setup_conds()
so we don't do it here.
*/
add_join_on(table_ref_2, eq_cond);
nj_col_1->is_common = nj_col_2->is_common = true;
DBUG_PRINT("info", ("%s.%s and %s.%s are common",
nj_col_1->table_name() ? nj_col_1->table_name() : "",
nj_col_1->name(),
nj_col_2->table_name() ? nj_col_2->table_name() : "",
nj_col_2->name()));
// Mark fields in the read set
if (field_1) {
nj_col_1->table_ref->table->mark_column_used(field_1,
MARK_COLUMNS_READ);
} else {
Mark_field mf(MARK_COLUMNS_READ);
item_1->walk(&Item::mark_field_in_map, enum_walk::SUBQUERY_POSTFIX,
(uchar *)&mf);
}
if (field_2) {
nj_col_2->table_ref->table->mark_column_used(field_2,
MARK_COLUMNS_READ);
} else {
Mark_field mf(MARK_COLUMNS_READ);
item_2->walk(&Item::mark_field_in_map, enum_walk::SUBQUERY_POSTFIX,
(uchar *)&mf);
}
if (using_fields != nullptr) ++(*found_using_fields);
}
}
if (leaf_1) leaf_1->is_join_columns_complete = true;
/*
Everything is OK.
Notice that at this point there may be some column names in the USING
clause that are not among the common columns. This is an SQL error and
we check for this error in store_natural_using_join_columns() when
(found_using_fields < length(join_using_fields)).
*/
return false;
}
/*
Materialize and store the row type of NATURAL/USING join.
SYNOPSIS
store_natural_using_join_columns()
thd current thread
natural_using_join the table reference of the NATURAL/USING join
table_ref_1 the first (left) operand (of a NATURAL/USING join).
table_ref_2 the second (right) operand (of a NATURAL/USING join).
using_fields if the join is JOIN...USING - the join columns,
if NATURAL join, then NULL
found_using_fields number of fields from the USING clause that were
found among the common fields
DESCRIPTION
Iterate over the columns of both join operands and sort and store
all columns into the 'join_columns' list of natural_using_join
where the list is formed by three parts:
part1: The coalesced columns of table_ref_1 and table_ref_2,
sorted according to the column order of the first table.
part2: The other columns of the first table, in the order in
which they were defined in CREATE TABLE.
part3: The other columns of the second table, in the order in
which they were defined in CREATE TABLE.
Time complexity - O(N1+N2), where Ni = length(table_ref_i).
IMPLEMENTATION
The procedure assumes that mark_common_columns() has been called
for the join that is being processed.
RETURN
true error: Some common column is ambiguous
false OK
*/
static bool store_natural_using_join_columns(THD *thd,
Table_ref *natural_using_join,
Table_ref *table_ref_1,
Table_ref *table_ref_2,
List<String> *using_fields,
uint found_using_fields) {
Field_iterator_table_ref it_1, it_2;
Natural_join_column *nj_col_1, *nj_col_2;
List<Natural_join_column> *non_join_columns;
DBUG_TRACE;
assert(!natural_using_join->join_columns);
Prepared_stmt_arena_holder ps_arena_holder(thd);
if (!(non_join_columns = new (thd->mem_root) List<Natural_join_column>) ||
!(natural_using_join->join_columns =
new (thd->mem_root) List<Natural_join_column>))
return true;
/* Append the columns of the first join operand. */
for (it_1.set(table_ref_1); !it_1.end_of_fields(); it_1.next()) {
nj_col_1 = it_1.get_natural_column_ref();
if (nj_col_1->is_common) {
natural_using_join->join_columns->push_back(nj_col_1);
/* Reset the common columns for the next call to mark_common_columns. */
nj_col_1->is_common = false;
} else
non_join_columns->push_back(nj_col_1);
}
/*
Check that all columns in the USING clause are among the common
columns. If this is not the case, report the first one that was
not found in an error.
*/
if (using_fields && found_using_fields < using_fields->elements) {
String *using_field_name;
List_iterator_fast<String> using_fields_it(*using_fields);
while ((using_field_name = using_fields_it++)) {
const char *using_field_name_ptr = using_field_name->c_ptr();
List_iterator_fast<Natural_join_column> it(
*(natural_using_join->join_columns));
Natural_join_column *common_field;
for (;;) {
/* If reached the end of fields, and none was found, report error. */
if (!(common_field = it++)) {
my_error(ER_BAD_FIELD_ERROR, MYF(0), using_field_name_ptr,
thd->where);
return true;
}
if (!my_strcasecmp(system_charset_info, common_field->name(),
using_field_name_ptr))
break; // Found match
}
}
}
/* Append the non-equi-join columns of the second join operand. */
for (it_2.set(table_ref_2); !it_2.end_of_fields(); it_2.next()) {
nj_col_2 = it_2.get_natural_column_ref();
if (!nj_col_2->is_common)
non_join_columns->push_back(nj_col_2);
else {
/* Reset the common columns for the next call to mark_common_columns. */
nj_col_2->is_common = false;
}
}
if (non_join_columns->elements > 0)
natural_using_join->join_columns->concat(non_join_columns);
natural_using_join->is_join_columns_complete = true;
return false;
}
/*
Precompute and store the row types of the top-most NATURAL/USING joins.
SYNOPSIS
store_top_level_join_columns()
thd current thread
table_ref nested join or table in a FROM clause
left_neighbor neighbor table reference to the left of table_ref at the
same level in the join tree
right_neighbor neighbor table reference to the right of table_ref at the
same level in the join tree
DESCRIPTION
The procedure performs a post-order traversal of a nested join tree
and materializes the row types of NATURAL/USING joins in a
bottom-up manner until it reaches the Table_ref elements that
represent the top-most NATURAL/USING joins. The procedure should be
applied to each element of Query_block::m_table_nest (i.e. to each
top-level element of the FROM clause).
IMPLEMENTATION
Notice that the table references in the list nested_join->join_list
are in reverse order, thus when we iterate over it, we are moving
from the right to the left in the FROM clause.
RETURN
true Error
false OK
*/
static bool store_top_level_join_columns(THD *thd, Table_ref *table_ref,
Table_ref *left_neighbor,
Table_ref *right_neighbor) {
DBUG_TRACE;
assert(!table_ref->nested_join->natural_join_processed);
Prepared_stmt_arena_holder ps_arena_holder(thd);
/* Call the procedure recursively for each nested table reference. */
if (table_ref->nested_join && !table_ref->nested_join->m_tables.empty()) {
auto nested_it = table_ref->nested_join->m_tables.begin();
Table_ref *same_level_left_neighbor = *nested_it++;
Table_ref *same_level_right_neighbor = nullptr;
/* Left/right-most neighbors, possibly at higher levels in the join tree. */
Table_ref *real_left_neighbor, *real_right_neighbor;
while (same_level_left_neighbor) {
Table_ref *cur_table_ref = same_level_left_neighbor;
same_level_left_neighbor =
(nested_it == table_ref->nested_join->m_tables.end()) ? nullptr
: *nested_it++;
/*
Pick the parent's left and right neighbors if there are no immediate
neighbors at the same level.
*/
real_left_neighbor =
(same_level_left_neighbor) ? same_level_left_neighbor : left_neighbor;
real_right_neighbor = (same_level_right_neighbor)
? same_level_right_neighbor
: right_neighbor;
if (cur_table_ref->nested_join &&
!cur_table_ref->nested_join->natural_join_processed &&
store_top_level_join_columns(thd, cur_table_ref, real_left_neighbor,
real_right_neighbor))
return true;
same_level_right_neighbor = cur_table_ref;
}
}
/*
If this is a NATURAL/USING join, materialize its result columns and
convert to a JOIN ... ON.
*/
if (table_ref->is_natural_join) {
assert(table_ref->nested_join &&
table_ref->nested_join->m_tables.size() == 2);
auto operand_it = table_ref->nested_join->m_tables.begin();
/*
Notice that the order of join operands depends on whether table_ref
represents a LEFT or a RIGHT join. In a RIGHT join, the operands are
in inverted order.
*/
Table_ref *table_ref_2 = *operand_it++; /* Second NATURAL join operand.*/
Table_ref *table_ref_1 = *operand_it++; /* First NATURAL join operand. */
List<String> *using_fields = table_ref->join_using_fields;
uint found_using_fields;
if (mark_common_columns(thd, table_ref_1, table_ref_2, using_fields,
&found_using_fields))
return true;
if (store_natural_using_join_columns(thd, table_ref, table_ref_1,
table_ref_2, using_fields,
found_using_fields))
return true;
/*
Change NATURAL JOIN to JOIN ... ON. We do this for both operands
because either one of them or the other is the one with the
natural join flag because RIGHT joins are transformed into LEFT,
and the two tables may be reordered.
*/
table_ref_1->natural_join = table_ref_2->natural_join = nullptr;
/* Add a true condition to outer joins that have no common columns. */
if (table_ref_2->outer_join && !table_ref_2->join_cond())
table_ref_2->set_join_cond(new Item_func_true());
/* Change this table reference to become a leaf for name resolution. */
if (left_neighbor) {
Table_ref *last_leaf_on_the_left;
last_leaf_on_the_left = left_neighbor->last_leaf_for_name_resolution();
last_leaf_on_the_left->next_name_resolution_table = table_ref;
}
if (right_neighbor) {
Table_ref *first_leaf_on_the_right;
first_leaf_on_the_right =
right_neighbor->first_leaf_for_name_resolution();
table_ref->next_name_resolution_table = first_leaf_on_the_right;
} else
table_ref->next_name_resolution_table = nullptr;
}
table_ref->nested_join->natural_join_processed = true;
return false;
}
/*
Compute and store the row types of the top-most NATURAL/USING joins
in a FROM clause.
SYNOPSIS
setup_natural_join_row_types()
thd current thread
from_clause list of top-level table references in a FROM clause
DESCRIPTION
Apply the procedure 'store_top_level_join_columns' to each of the
top-level table references of the FROM clause. Adjust the list of tables
for name resolution - context->first_name_resolution_table to the
top-most, lef-most NATURAL/USING join.
IMPLEMENTATION
Notice that the table references in 'from_clause' are in reverse
order, thus when we iterate over it, we are moving from the right
to the left in the FROM clause.
RETURN
true Error
false OK
*/
bool setup_natural_join_row_types(THD *thd,
mem_root_deque<Table_ref *> *from_clause,
Name_resolution_context *context) {
DBUG_TRACE;
thd->where = "from clause";
if (from_clause->empty())
return false; /* We come here in the case of UNIONs. */
auto table_ref_it = from_clause->begin();
/* Table reference to the left of the current. */
Table_ref *left_neighbor = *table_ref_it++;
/* Table reference to the right of the current. */
Table_ref *right_neighbor = nullptr;
/* Note that tables in the list are in reversed order */
while (left_neighbor) {
/* Current table reference. */
Table_ref *table_ref = left_neighbor;
left_neighbor =
(table_ref_it == from_clause->end()) ? nullptr : *table_ref_it++;
/*
Do not redo work if already done:
- for prepared statements and stored procedures,
- if already processed inside a derived table/view.
*/
if (table_ref->nested_join &&
!table_ref->nested_join->natural_join_processed) {
if (store_top_level_join_columns(thd, table_ref, left_neighbor,
right_neighbor))
return true;
}
if (left_neighbor && context->query_block->first_execution) {
left_neighbor->next_name_resolution_table =
table_ref->first_leaf_for_name_resolution();
}
right_neighbor = table_ref;
}
/*
Store the top-most, left-most NATURAL/USING join, so that we start
the search from that one instead of context->table_list. At this point
right_neighbor points to the left-most top-level table reference in the
FROM clause.
*/
assert(right_neighbor);
context->first_name_resolution_table =
right_neighbor->first_leaf_for_name_resolution();
return false;
}
/**
Resolve variable assignments from LEX object
@param thd Thread handler
@param lex Lex object containing variable assignments
@returns false if success, true if error
@note
set_entry() must be called before fix_fields() of the whole list of
field items because:
1) the list of field items has same order as in the query, and the
Item_func_get_user_var item may go before the Item_func_set_user_var:
@verbatim
SELECT @a, @a := 10 FROM t;
@endverbatim
2) The entry->update_query_id value controls constantness of
Item_func_get_user_var items, so in presence of Item_func_set_user_var
items we have to refresh their entries before fixing of
Item_func_get_user_var items.
*/
bool resolve_var_assignments(THD *thd, LEX *lex) {
List_iterator<Item_func_set_user_var> li(lex->set_var_list);
Item_func_set_user_var *var;
while ((var = li++)) var->set_entry(thd, false);
return false;
}
/****************************************************************************
** Check that all given fields exists and fill struct with current data
****************************************************************************/
/**
Resolve a list of expressions and setup appropriate data
@param thd thread handler
@param want_privilege privilege representing desired operation.
whether the expressions are selected, inserted
or updated, or no operation is done.
will also decide inclusion in read/write maps.
@param allow_sum_func true if set operations are allowed in context.
@param column_update if true, reject expressions that do not resolve
to a base table column
@param split_sum_funcs If true, Item::split_sum_func will add hidden
items to "fields". See also description in its
helper method Item::split_sum_func2.
@param typed_items List of reference items for type derivation
May be nullptr.
@param[in,out] fields list of expressions, populated with resolved
data about expressions.
@param[out] ref_item_array filled in with references to items.
@retval false if success
@retval true if error
@note The function checks updatability/insertability for the table before
checking column privileges, for consistent error reporting.
This has consequences for columns that are specified to be updated:
The column is first resolved without privilege check.
This check is followed by an updatablity/insertability check.
Finally, a column privilege check is run, and the column is marked
for update.
*/
bool setup_fields(THD *thd, Access_bitmask want_privilege, bool allow_sum_func,
bool split_sum_funcs, bool column_update,
const mem_root_deque<Item *> *typed_items,
mem_root_deque<Item *> *fields,
Ref_item_array ref_item_array) {
DBUG_TRACE;
Query_block *const select = thd->lex->current_query_block();
const enum_mark_columns save_mark_used_columns = thd->mark_used_columns;
nesting_map save_allow_sum_func = thd->lex->allow_sum_func;
Column_privilege_tracker column_privilege(thd,
column_update ? 0 : want_privilege);
// Function can only be used to set up one specific operation:
assert(want_privilege == 0 || want_privilege == SELECT_ACL ||
want_privilege == INSERT_ACL || want_privilege == UPDATE_ACL);
assert(!(column_update && (want_privilege & SELECT_ACL)));
if (want_privilege & SELECT_ACL)
thd->mark_used_columns = MARK_COLUMNS_READ;
else if (want_privilege & (INSERT_ACL | UPDATE_ACL) && !column_update)
thd->mark_used_columns = MARK_COLUMNS_WRITE;
else
thd->mark_used_columns = MARK_COLUMNS_NONE;
DBUG_PRINT("info", ("thd->mark_used_columns: %d", thd->mark_used_columns));
if (allow_sum_func)
thd->lex->allow_sum_func |= (nesting_map)1 << select->nest_level;
thd->where = THD::DEFAULT_WHERE;
bool save_is_item_list_lookup = select->is_item_list_lookup;
select->is_item_list_lookup = false;
/*
To prevent fail on forward lookup we fill it with zerows,
then if we got pointer on zero after find_item_in_list we will know
that it is forward lookup.
There is other way to solve problem: fill array with pointers to list,
but it will be slower.
*/
if (!ref_item_array.is_null()) {
size_t num_visible_fields = CountVisibleFields(*fields);
assert(ref_item_array.size() >= num_visible_fields);
memset(ref_item_array.array(), 0, sizeof(Item *) * num_visible_fields);
}
Ref_item_array ref = ref_item_array;
mem_root_deque<Item *>::const_iterator typed_it;
if (typed_items != nullptr) {
typed_it = typed_items->begin();
}
for (auto it = fields->begin(); it != fields->end(); ++it) {
const size_t old_size = fields->size();
Item *item = *it;
assert(!item->hidden);
Item **item_pos = &*it;
if ((!item->fixed && item->fix_fields(thd, item_pos)) ||
(item = *item_pos)->check_cols(1)) {
DBUG_PRINT("info",
("thd->mark_used_columns: %d", thd->mark_used_columns));
return true; /* purecov: inspected */
}
// Check that we don't have a field that is hidden system field. This should
// be caught in Item_field::fix_fields.
assert(
item->type() != Item::FIELD_ITEM ||
!static_cast<const Item_field *>(item)->field->is_hidden_by_system());
if (!ref.is_null()) {
ref[0] = item;
ref.pop_front();
/*
Items present in ref_array have a positive reference count since
removal of unused columns from derived tables depends on this.
*/
item->increment_ref_count();
}
Item *typed_item = nullptr;
if (typed_items != nullptr && typed_it != typed_items->end()) {
typed_item = *typed_it++;
assert(!typed_item->hidden);
}
if (column_update) {
Item_field *const field = item->field_for_view_update();
if (field == nullptr) {
my_error(ER_NONUPDATEABLE_COLUMN, MYF(0), item->item_name.ptr());
return true;
}
if (item->type() == Item::TRIGGER_FIELD_ITEM) {
char buff[NAME_LEN * 2];
String str(buff, sizeof(buff), &my_charset_bin);
str.length(0);
item->print(thd, &str, QT_ORDINARY);
my_error(ER_INVALID_ASSIGNMENT_TARGET, MYF(0), str.c_ptr());
return true;
}
Table_ref *tr = field->table_ref;
if ((want_privilege & UPDATE_ACL) && !tr->is_updatable()) {
/*
The base table of the column may have beeen referenced through a view
or derived table. If so, print the name of the upper-most view
referring to this table in order to print the error message with the
alias of the view as written in the original query instead of the
alias of the base table.
*/
my_error(ER_NON_UPDATABLE_TABLE, MYF(0), tr->top_table()->alias,
"UPDATE");
return true;
}
if ((want_privilege & INSERT_ACL) && !tr->is_insertable()) {
/* purecov: begin inspected */
/*
Generally unused as long as INSERT only can be applied against
one base table, for which the INSERT privileges are checked in
Sql_cmd_insert_base::prepare_inner()
*/
my_error(ER_NON_INSERTABLE_TABLE, MYF(0), tr->top_table()->alias,
"INSERT");
return true;
/* purecov: end */
}
if (want_privilege & (INSERT_ACL | UPDATE_ACL)) {
Column_privilege_tracker column_privilege_tr(thd, want_privilege);
if (item->walk(&Item::check_column_privileges, enum_walk::PREFIX,
pointer_cast<uchar *>(thd)))
return true;
}
Mark_field mf(MARK_COLUMNS_WRITE);
item->walk(&Item::mark_field_in_map, enum_walk::POSTFIX,
pointer_cast<uchar *>(&mf));
} else if (item->data_type() == MYSQL_TYPE_INVALID) {
if (typed_item != nullptr) {
if (item->propagate_type(thd, Type_properties(*typed_item)))
return true;
} else {
if (item->propagate_type(thd, item->default_data_type())) return true;
}
}
if (split_sum_funcs) {
/*
(1) Contains a grouped aggregate but is not one. If it is one, we do
not split, but in create_tmp_table() we look at its arguments and add
them to the tmp table, which achieves the same result as for window
functions in (2) but differently.
@todo: unify this (do like (2), probably).
(2) Contains a window function. Even if it is a window function, we
have to collect its arguments and add them to the hidden list of
items, as those arguments have to be stored in the first tmp tables,
and carried forward up to the tmp table where the WF can be
evaluated.
*/
if ((item->has_aggregation() && !(item->type() == Item::SUM_FUNC_ITEM &&
!item->m_is_window_function)) || //(1)
item->has_wf()) // (2)
item->split_sum_func(thd, ref_item_array, fields);
}
select->select_list_tables |= item->used_tables();
if (old_size != fields->size()) {
// Items have been added (either by fix_fields or by split_sum_func), so
// our iterator is invalidated. Reconstruct it.
it = std::find(fields->begin(), fields->end(), item);
}
}
select->is_item_list_lookup = save_is_item_list_lookup;
thd->lex->allow_sum_func = save_allow_sum_func;
thd->mark_used_columns = save_mark_used_columns;
DBUG_PRINT("info", ("thd->mark_used_columns: %d", thd->mark_used_columns));
assert(!thd->is_error());
return false;
}
/**
This is an iterator which emits leaf Table_ref nodes in an order
suitable for expansion of 'table_name.*' (qualified asterisk) or '*'
(unqualified), fur use by insert_fields().
First here is some background.
1.
SELECT T1.*, T2.* FROM T1 NATURAL JOIN T2;
has to return all columns of T1 and then all of T2's;
whereas
SELECT * FROM T1 NATURAL JOIN T2;
has to return all columns of T1 and then only those of T2's which are not
common with T1.
Thus, in the first case a NATURAL JOIN isn't considered a leaf (we have to
see through it to find T1.* and T2.*), in the second case it is (we have
to ask it for its column set).
In the first case, the place to search for tables is thus the
Table_ref::next_local list; in the second case it is
Table_ref::next_name_resolution_table.
2.
SELECT * FROM T1 RIGHT JOIN T2 ON < cond >;
is converted, during contextualization, to:
SELECT * FROM T2 LEFT JOIN T1 ON < cond >;
however the former has to return columns of T1 then of T2,
while the latter has to return T2's then T1's.
The conversion has been complete: the lists 'next_local',
'next_name_resolution_table' and Query_block::m_current_table_nest are as if
the user had typed the second query.
Now to the behaviour of this iterator.
A. If qualified asterisk, the emission order is irrelevant as the caller
tests the table's name; and a NATURAL JOIN isn't a leaf. So, we just follow
the Table_ref::next_local pointers.
B. If non-qualified asterisk, the order must be the left-to-right order
as it was in the query entered by the user. And a NATURAL JOIN is a leaf. So:
B.i. if there was no RIGHT JOIN, then the user-input order is just that of
the 'next_name_resolution' pointers.
B.ii. otherwise, then the order has to be found by a more complex procedure
(function build_vec()):
- first we traverse the join operators, taking into account operators
which had a conversion from RIGHT to LEFT JOIN, we recreate the user-input
order and store leaf TABLE_LISTs in that order in a vector.
- then, in the emission phase, we just emit tables from the vector.
Sequence of calls: constructor, init(), [get_next() N times], destructor.
*/
class Tables_in_user_order_iterator {
public:
void init(Query_block *query_block, bool qualified) {
assert(query_block && !m_query_block);
m_query_block = query_block;
m_qualified = qualified;
// Vector is needed only if '*' is not qualified and there were RIGHT JOINs
if (m_qualified) {
m_next = m_query_block->context.table_list;
return;
}
if (!m_query_block->right_joins()) {
m_next = m_query_block->context.first_name_resolution_table;
return;
}
m_next = nullptr;
m_vec = new std::vector<Table_ref *>;
fill_vec(*m_query_block->m_current_table_nest);
}
~Tables_in_user_order_iterator() {
delete m_vec;
m_vec = nullptr;
}
Table_ref *get_next() {
if (m_vec == nullptr) {
auto cur = m_next;
if (cur)
m_next =
m_qualified ? cur->next_local : cur->next_name_resolution_table;
return cur;
}
if (m_next_vec_pos == m_vec->size()) return nullptr;
return (*m_vec)[m_next_vec_pos++];
}
private:
/// Fills the vector
/// @param tables list of tables and join operators
void fill_vec(const mem_root_deque<Table_ref *> &tables) {
if (tables.size() != 0 && tables.front()->join_order_swapped) {
assert(tables.size() == 2 && !tables.back()->join_order_swapped);
add_table(tables.front());
add_table(tables.back());
return;
}
// Walk from end to beginning, as join_list is always "reversed"
// (e.g. T1 INNER JOIN T2 leads to join_list = (T2,T1)):
for (auto it = tables.rbegin(); it != tables.rend(); ++it) add_table(*it);
}
void add_table(Table_ref *tr) {
if (tr->is_leaf_for_name_resolution()) // stop diving here
return m_vec->push_back(tr);
if (tr->nested_join != nullptr) // do dive
fill_vec(tr->nested_join->m_tables);
}
// Query block which owns the FROM clause to search in
Query_block *m_query_block{nullptr};
/// True/false if we want to expand 'table_name.*' / '*'.
bool m_qualified;
/// If not using the vector: next table to emit
Table_ref *m_next;
/// Vector for the complex case. As the complex case is expected to be rare,
/// we allocate the vector only if needed. nullptr otherwise.
std::vector<Table_ref *> *m_vec{nullptr};
/// If using the vector: position in vector, of next table to emit
uint m_next_vec_pos{0};
};
/*
Drops in all fields instead of current '*' field
SYNOPSIS
insert_fields()
thd Thread handler
query_block Query block
db_name Database name in case of 'database_name.table_name.*'
table_name Table name in case of 'table_name.*'
it Pointer to '*'
any_privileges 0 If we should ensure that we have SELECT privileges
for all columns
1 If any privilege is ok
RETURN
0 ok 'it' is updated to point at last inserted
1 error. Error message is generated but not sent to client
*/
bool insert_fields(THD *thd, Query_block *query_block, const char *db_name,
const char *table_name, mem_root_deque<Item *> *fields,
mem_root_deque<Item *>::iterator *it, bool any_privileges) {
char name_buff[NAME_LEN + 1];
DBUG_TRACE;
DBUG_PRINT("arena", ("stmt arena: %p", thd->stmt_arena));
// No need to expand '*' multiple times:
assert(query_block->first_execution);
if (db_name &&
(lower_case_table_names || is_infoschema_db(db_name, strlen(db_name)))) {
/*
convert database to lower case for comparison
We can't do this in Item_field as this would change the
'name' of the item which may be used in the select list
We lowercase the 'information_schema' name below to treat it as
case-insensitive even in lower_case_table_names=0.
*/
strmake(name_buff, db_name, sizeof(name_buff) - 1);
my_casedn_str(files_charset_info, name_buff);
db_name = name_buff;
}
bool found = false;
Table_ref *tables;
Tables_in_user_order_iterator user_it;
user_it.init(query_block, table_name != nullptr);
while (true) {
tables = user_it.get_next();
if (tables == nullptr) break;
Field_iterator_table_ref field_iterator;
TABLE *const table = tables->table;
assert(tables->is_leaf_for_name_resolution());
if ((table_name &&
my_strcasecmp(table_alias_charset, table_name, tables->alias)) ||
(db_name && strcmp(tables->db, db_name)))
continue;
/*
Ensure that we have access rights to all fields to be inserted. Under
some circumstances, this check may be skipped.
- If any_privileges is true, skip the check.
- If the SELECT privilege has been found as fulfilled already,
the check is skipped.
NOTE: This check is not sufficient: If a user has SELECT_ACL privileges
for a view, it does not mean having the same privileges for the
underlying tables/view. Thus, we have to perform individual column
privilege checks below (or recurse down to all underlying tables here).
*/
if (!any_privileges && !(tables->grant.privilege & SELECT_ACL)) {
field_iterator.set(tables);
if (check_grant_all_columns(thd, SELECT_ACL, &field_iterator))
return true;
}
/*
Update the tables used in the query based on the referenced fields. For
views and natural joins this update is performed inside the loop below.
*/
if (table) {
thd->lex->current_query_block()->select_list_tables |= tables->map();
}
/*
Initialize a generic field iterator for the current table reference.
Notice that it is guaranteed that this iterator will iterate over the
fields of a single table reference, because 'tables' is a leaf (for
name resolution purposes).
*/
field_iterator.set(tables);
for (; !field_iterator.end_of_fields(); field_iterator.next()) {
Item *const item = field_iterator.create_item(thd);
if (!item) return true; /* purecov: inspected */
assert(item->fixed);
if (item->type() == Item::FIELD_ITEM) {
Item_field *field = down_cast<Item_field *>(item);
/*
If the column is hidden from users and not used in USING clause of
a join, do not add this column in place of '*'.
*/
bool is_hidden = field->field->is_hidden();
is_hidden &= (tables->join_using_fields == nullptr ||
!test_if_string_in_list(field->field_name,
tables->join_using_fields));
if (is_hidden) continue;
/* cache the table for the Item_fields inserted by expanding stars */
if (tables->cacheable_table) field->cached_table = tables;
}
if (!found) {
found = true;
**it = item; /* Replace '*' with the first found item. */
} else {
/* Add 'item' to the SELECT list, after the current one. */
*it = fields->insert(*it + 1, item);
}
/*
Set privilege information for the fields of newly created views.
We have that (any_priviliges == true) if and only if we are creating
a view. In the time of view creation we can't use the MERGE algorithm,
therefore if 'tables' is itself a view, it is represented by a
temporary table. Thus in this case we can be sure that 'item' is an
Item_field.
*/
if (any_privileges) {
assert((tables->field_translation == nullptr && table) ||
tables->is_natural_join);
assert(item->type() == Item::FIELD_ITEM);
Item_field *const fld = (Item_field *)item;
const char *field_table_name = field_iterator.get_table_name();
if (!tables->schema_table && !tables->is_internal() &&
!(fld->have_privileges =
(get_column_grant(thd, field_iterator.grant(),
field_iterator.get_db_name(),
field_table_name, fld->field_name) &
VIEW_ANY_ACL))) {
my_error(ER_TABLEACCESS_DENIED_ERROR, MYF(0), "ANY",
thd->security_context()->priv_user().str,
thd->security_context()->host_or_ip().str, field_table_name);
return true;
}
}
thd->lex->current_query_block()->select_list_tables |=
item->used_tables();
Field *const field = field_iterator.field();
if (field) {
// Register underlying fields in read map if wanted.
field->table->mark_column_used(field, thd->mark_used_columns);
} else {
if (thd->want_privilege && tables->is_view_or_derived()) {
if (item->walk(&Item::check_column_privileges, enum_walk::PREFIX,
(uchar *)thd))
return true;
}
// Register underlying fields in read map if wanted.
Mark_field mf(thd->mark_used_columns);
item->walk(&Item::mark_field_in_map, enum_walk::SUBQUERY_POSTFIX,
(uchar *)&mf);
}
}
}
if (found) return false;
/*
TODO: in the case when we skipped all columns because there was a
qualified '*', and all columns were coalesced, we have to give a more
meaningful message than ER_BAD_TABLE_ERROR.
*/
if (!table_name || !*table_name)
my_error(ER_NO_TABLES_USED, MYF(0));
else {
String tbl_name;
if (db_name) {
tbl_name.append(String(db_name, system_charset_info));
tbl_name.append('.');
}
tbl_name.append(String(table_name, system_charset_info));
my_error(ER_BAD_TABLE_ERROR, MYF(0), tbl_name.c_ptr_safe());
}
return true;
}
/******************************************************************************
** Fill a record with data (for INSERT or UPDATE)
** Returns : 1 if some field has wrong type
******************************************************************************/
/**
Fill fields with given items.
@param thd Thread handler.
@param table Table reference.
@param fields Item_fields list to be filled
@param values Values to fill with.
@param bitmap Bitmap over fields to fill.
@param insert_into_fields_bitmap Bitmap for fields that is set
in fill_record.
@param raise_autoinc_has_expl_non_null_val Set corresponding flag in TABLE
object to true if non-NULL value
is explicitly assigned to
auto-increment field.
@note fill_record() may set TABLE::autoinc_field_has_explicit_non_null_value
to true (even in case of failure!) and its caller should make sure that
it is reset before next call to this function (i.e. before processing
next row) and/or before TABLE instance is returned to table cache.
One can use helper Auto_increment_field_not_null_reset_guard class
to do this.
@note In order to simplify implementation this call is allowed to reset
TABLE::autoinc_field_has_explicit_non_null_value flag even in case
when raise_autoinc_has_expl_non_null_val is false. However, this
should be fine since this flag is supposed to be reset already in
such cases.
@return Operation status
@retval false OK
@retval true Error occurred
*/
bool fill_record(THD *thd, TABLE *table, const mem_root_deque<Item *> &fields,
const mem_root_deque<Item *> &values, MY_BITMAP *bitmap,
MY_BITMAP *insert_into_fields_bitmap,
bool raise_autoinc_has_expl_non_null_val) {
DBUG_TRACE;
assert(CountVisibleFields(fields) == CountVisibleFields(values));
/*
In case when TABLE object comes to fill_record() from Table Cache it
should have autoinc_field_has_explicit_non_null_value flag set to false.
In case when TABLE object comes to fill_record() after processing
previous row this flag should be reset to false by caller.
Code which implements LOAD DATA is the exception to the above rule
as it calls fill_record() to handle SET clause, after values for
the columns directly coming from loaded from file are set and thus
autoinc_field_has_explicit_non_null_value possibly set to true.
*/
assert(table->autoinc_field_has_explicit_non_null_value == false ||
(raise_autoinc_has_expl_non_null_val &&
thd->lex->sql_command == SQLCOM_LOAD));
auto value_it = VisibleFields(values).begin();
for (Item *fld : VisibleFields(fields)) {
Item_field *const field = fld->field_for_view_update();
assert(field != nullptr && field->table_ref->table == table);
Field *const rfield = field->field;
Item *value = *value_it++;
/* If bitmap over wanted fields are set, skip non marked fields. */
if (bitmap && !bitmap_is_set(bitmap, rfield->field_index())) continue;
bitmap_set_bit(table->fields_set_during_insert, rfield->field_index());
if (insert_into_fields_bitmap)
bitmap_set_bit(insert_into_fields_bitmap, rfield->field_index());
/* Generated columns will be filled after all base columns are done. */
if (rfield->is_gcol()) continue;
if (raise_autoinc_has_expl_non_null_val &&
rfield == table->next_number_field)
table->autoinc_field_has_explicit_non_null_value = true;
/*
We handle errors from save_in_field() by first checking the return
value and then testing thd->is_error(). thd->is_error() can be set
even when save_in_field() does not return a negative value.
@todo save_in_field returns an enum which should never be a negative
value. We should change this test to check for correct enum value.
The below call can reset TABLE::autoinc_field_has_explicit_non_null_value
flag depending on value provided (for details please see
set_field_to_null_with_conversions()). So evaluation of this flag can't
be moved outside of fill_record(), to be done once per statement.
*/
if (value->save_in_field(rfield, false) < 0) {
my_error(ER_UNKNOWN_ERROR, MYF(0));
return true;
}
if (thd->is_error()) return true;
}
if (table->has_gcol() &&
update_generated_write_fields(bitmap ? bitmap : table->write_set, table))
return true;
/*
TABLE::autoinc_field_has_explicit_non_null_value should not be set to
true in raise_autoinc_has_expl_non_null_val == false mode.
*/
assert(table->autoinc_field_has_explicit_non_null_value == false ||
raise_autoinc_has_expl_non_null_val);
return thd->is_error();
}
/**
Check the NOT NULL constraint on all the fields of the current record.
@param thd Thread context.
@param fields Collection of fields.
@return Error status.
*/
static bool check_record(THD *thd, const mem_root_deque<Item *> &fields) {
for (Item *fld : VisibleFields(fields)) {
Item_field *field = fld->field_for_view_update();
if (field &&
field->field->check_constraints(ER_BAD_NULL_ERROR) != TYPE_OK) {
my_error(ER_UNKNOWN_ERROR, MYF(0));
return true;
}
}
return thd->is_error();
}
/**
Check the NOT NULL constraint on all the fields of the current record.
@param thd Thread context.
@param ptr Fields.
@return Error status.
*/
bool check_record(THD *thd, Field **ptr) {
Field *field;
while ((field = *ptr++) && !thd->is_error()) {
if (field->check_constraints(ER_BAD_NULL_ERROR) != TYPE_OK) return true;
}
return thd->is_error();
}
/**
Check the NOT NULL constraint on all the fields explicitly set
in INSERT INTO statement or implicitly set in BEFORE trigger.
@param thd Thread context.
@param ptr Fields.
@return Error status.
*/
static bool check_inserting_record(THD *thd, Field **ptr) {
Field *field;
while ((field = *ptr++) && !thd->is_error()) {
if (bitmap_is_set(field->table->fields_set_during_insert,
field->field_index()) &&
field->check_constraints(ER_BAD_NULL_ERROR) != TYPE_OK)
return true;
}
return thd->is_error();
}
/**
Invoke check constraints defined on the table.
@param thd Thread handle.
@param table Instance of TABLE.
@retval false If all enforced check constraints are satisfied.
@retval true Otherwise. THD::is_error() may be "true" in this case.
*/
bool invoke_table_check_constraints(THD *thd, const TABLE *table) {
if (table->table_check_constraint_list != nullptr) {
for (auto &table_cc : *table->table_check_constraint_list) {
if (table_cc.is_enforced()) {
/*
Invoke check constraints only if column(s) used by check constraint is
updated.
*/
if ((thd->lex->sql_command == SQLCOM_UPDATE ||
thd->lex->sql_command == SQLCOM_UPDATE_MULTI) &&
!bitmap_is_overlapping(
&table_cc.value_generator()->base_columns_map,
table->write_set)) {
DEBUG_SYNC(thd, "skip_check_constraints_on_unaffected_columns");
continue;
}
// Validate check constraint.
Item *check_const_expr_item = table_cc.value_generator()->expr_item;
check_const_expr_item->m_in_check_constraint_exec_ctx = true;
bool is_constraint_violated = (!check_const_expr_item->val_bool() &&
!check_const_expr_item->null_value);
check_const_expr_item->m_in_check_constraint_exec_ctx = false;
/*
If check constraint is violated then report an error. If expression
operand types are incompatible and reported error in conversion even
then report a more user friendly error. Sql_conditions of DA still has
a conversion(actual reported) error in the error stack.
*/
if (is_constraint_violated || thd->is_error()) {
if (thd->is_error()) thd->clear_error();
my_error(ER_CHECK_CONSTRAINT_VIOLATED, MYF(0), table_cc.name().str);
return true;
}
}
}
}
return false;
}
/**
Check if SQL-statement is INSERT/INSERT SELECT/REPLACE/REPLACE SELECT
and trigger event is ON INSERT. When this condition is true that means
that the statement basically can invoke BEFORE INSERT trigger if it
was created before.
@param event event type for triggers to be invoked
@param sql_command Type of SQL statement
@return Test result
@retval true SQL-statement is
INSERT/INSERT SELECT/REPLACE/REPLACE SELECT
and trigger event is ON INSERT
@retval false Either SQL-statement is not
INSERT/INSERT SELECT/REPLACE/REPLACE SELECT
or trigger event is not ON INSERT
*/
static inline bool command_can_invoke_insert_triggers(
enum enum_trigger_event_type event, enum_sql_command sql_command) {
/*
If it's 'INSERT INTO ... ON DUPLICATE KEY UPDATE ...' statement
the event is TRG_EVENT_UPDATE and the SQL-command is SQLCOM_INSERT.
*/
return event == TRG_EVENT_INSERT &&
(sql_command == SQLCOM_INSERT || sql_command == SQLCOM_INSERT_SELECT ||
sql_command == SQLCOM_REPLACE ||
sql_command == SQLCOM_REPLACE_SELECT);
}
/**
Execute BEFORE INSERT trigger.
@param thd thread context
@param table TABLE-object holding list of triggers
to be invoked
@param event event type for triggers to be invoked
@param insert_into_fields_bitmap Bitmap for fields that is set
in fill_record
@return Operation status
@retval false OK
@retval true Error occurred
*/
inline bool call_before_insert_triggers(THD *thd, TABLE *table,
enum enum_trigger_event_type event,
MY_BITMAP *insert_into_fields_bitmap) {
for (Field **f = table->field; *f; ++f) {
if ((*f)->is_flag_set(NO_DEFAULT_VALUE_FLAG) &&
!bitmap_is_set(insert_into_fields_bitmap, (*f)->field_index())) {
(*f)->set_tmp_null();
}
}
return table->triggers->process_triggers(thd, event, TRG_ACTION_BEFORE, true);
}
/**
Fill fields in list with values from the list of items and invoke
before triggers.
@param thd Thread context.
@param optype_info COPY_INFO structure used for
default values handling.
@param fields Item_fields list to be filled.
@param values Values to fill with.
@param table TABLE-object for the table.
@param event Event type for triggers to be
invoked.
@param num_fields Number of fields in table.
@param raise_autoinc_has_expl_non_null_val Set corresponding flag in
TABLE to true if non-NULL
value is explicitly assigned
to auto-increment field.
@param[out] is_row_changed Set to true if a row is
changed after filling record
and invoking before triggers
for UPDATE operation.
Otherwise set to false.
@note This function assumes that fields which values will be set and
triggers to be invoked belong to the same table, and that
TABLE::record[0] and record[1] buffers correspond to new and old
versions of row respectively.
@note This call may set TABLE::autoinc_field_has_explicit_non_null_value to
true (even in case of failure!) and its caller should make sure that
it is reset appropriately (@sa fill_record()).
@return Operation status
@retval false OK
@retval true Error occurred
*/
bool fill_record_n_invoke_before_triggers(
THD *thd, COPY_INFO *optype_info, const mem_root_deque<Item *> &fields,
const mem_root_deque<Item *> &values, TABLE *table,
enum enum_trigger_event_type event, int num_fields,
bool raise_autoinc_has_expl_non_null_val, bool *is_row_changed) {
// is_row_changed is used by UPDATE operation to set compare_record() result.
assert(is_row_changed == nullptr ||
optype_info->get_operation_type() == COPY_INFO::UPDATE_OPERATION);
/*
Fill DEFAULT functions (like CURRENT_TIMESTAMP) and DEFAULT expressions on
the columns that are not on the list of assigned columns.
*/
auto fill_function_defaults = [table, optype_info, is_row_changed]() {
/*
Unlike INSERT and LOAD, UPDATE operation requires comparison of old
and new records to determine whether function defaults have to be
evaluated.
*/
if (optype_info->get_operation_type() == COPY_INFO::UPDATE_OPERATION) {
*is_row_changed =
(!records_are_comparable(table) || compare_records(table));
/*
Evaluate function defaults for columns with ON UPDATE clause only
if any other column of the row is updated.
*/
if (*is_row_changed &&
(optype_info->function_defaults_apply_on_columns(table->write_set))) {
if (optype_info->set_function_defaults(table)) return true;
}
} else if (optype_info->function_defaults_apply_on_columns(
table->write_set)) {
if (optype_info->set_function_defaults(table)) return true;
}
return false;
};
/*
If it's 'INSERT INTO ... ON DUPLICATE KEY UPDATE ...' statement
the event is TRG_EVENT_UPDATE and the SQL-command is SQLCOM_INSERT.
*/
Trigger_chain *tc =
table->triggers != nullptr
? table->triggers->get_triggers(event, TRG_ACTION_BEFORE)
: nullptr;
if (tc != nullptr) {
bool rc;
table->triggers->enable_fields_temporary_nullability(thd);
if (command_can_invoke_insert_triggers(event, thd->lex->sql_command)) {
assert(num_fields);
MY_BITMAP insert_into_fields_bitmap;
bitmap_init(&insert_into_fields_bitmap, nullptr, num_fields);
rc = fill_function_defaults();
if (!rc)
rc = fill_record(thd, table, fields, values, nullptr,
&insert_into_fields_bitmap,
raise_autoinc_has_expl_non_null_val);
if (!rc)
rc = call_before_insert_triggers(thd, table, event,
&insert_into_fields_bitmap);
bitmap_free(&insert_into_fields_bitmap);
} else {
rc = fill_record(thd, table, fields, values, nullptr, nullptr,
raise_autoinc_has_expl_non_null_val);
if (!rc) {
rc = fill_function_defaults();
if (!rc)
rc = table->triggers->process_triggers(thd, event, TRG_ACTION_BEFORE,
true);
// For UPDATE operation, check if row is updated by the triggers.
if (!rc &&
optype_info->get_operation_type() == COPY_INFO::UPDATE_OPERATION &&
!(*is_row_changed))
*is_row_changed =
(!records_are_comparable(table) || compare_records(table));
}
}
/*
Re-calculate generated fields to cater for cases when base columns are
updated by the triggers.
*/
assert(table->pos_in_table_list && !table->pos_in_table_list->is_view());
if (!rc && table->has_gcol() &&
tc->has_updated_trigger_fields(table->write_set)) {
// Dont save old value while re-calculating generated fields.
// Before image will already be saved in the first calculation.
table->blobs_need_not_keep_old_value();
rc = update_generated_write_fields(table->write_set, table);
}
table->triggers->disable_fields_temporary_nullability();
rc = rc || check_inserting_record(thd, table->field);
table->triggers->reset_field_nulls();
return rc;
} else {
if (fill_record(thd, table, fields, values, nullptr, nullptr,
raise_autoinc_has_expl_non_null_val))
return true;
if (fill_function_defaults()) return true;
return check_record(thd, fields);
}
}
/**
Fill field buffer with values from Field list.
@param thd Thread handler.
@param table Table reference.
@param ptr Array of fields to fill in.
@param values List of values to fill with.
@param bitmap Bitmap over fields to fill.
@param insert_into_fields_bitmap Bitmap for fields that is set
in fill_record.
@param raise_autoinc_has_expl_non_null_val Set corresponding flag in TABLE
object to true if non-NULL value
is explicitly assigned to
auto-increment field.
@note fill_record() may set TABLE::autoinc_field_has_explicit_non_null_value
to true (even in case of failure!) and its caller should make sure that
it is reset before next call to this function (i.e. before processing
next row) and/or before TABLE instance is returned to table cache.
One can use helper Auto_increment_field_not_null_reset_guard class
to do this.
@note In order to simplify implementation this call is allowed to reset
TABLE::autoinc_field_has_explicit_non_null_value flag even in case
when raise_autoinc_has_expl_non_null_val is false. However, this
should be fine since this flag is supposed to be reset already in
such cases.
@return Operation status
@retval false OK
@retval true Error occurred
*/
bool fill_record(THD *thd, TABLE *table, Field **ptr,
const mem_root_deque<Item *> &values, MY_BITMAP *bitmap,
MY_BITMAP *insert_into_fields_bitmap,
bool raise_autoinc_has_expl_non_null_val) {
DBUG_TRACE;
/*
In case when TABLE object comes to fill_record() from Table Cache it
should have autoinc_field_has_explicit_non_null_value flag set to false.
In case when TABLE object comes to fill_record() after processing
previous row this flag should be reset to false by caller.
*/
assert(table->autoinc_field_has_explicit_non_null_value == false);
Field *field;
auto value_it = VisibleFields(values).begin();
while ((field = *ptr++) && !thd->is_error()) {
// Skip hidden system field.
if (field->is_hidden_by_system()) continue;
Item *value = *value_it++;
assert(field->table == table);
/* If bitmap over wanted fields are set, skip non marked fields. */
if (bitmap && !bitmap_is_set(bitmap, field->field_index())) continue;
/*
fill_record could be called as part of multi update and therefore
table->fields_set_during_insert could be NULL.
*/
if (table->fields_set_during_insert)
bitmap_set_bit(table->fields_set_during_insert, field->field_index());
if (insert_into_fields_bitmap)
bitmap_set_bit(insert_into_fields_bitmap, field->field_index());
/* Generated columns will be filled after all base columns are done. */
if (field->is_gcol()) continue;
if (raise_autoinc_has_expl_non_null_val &&
field == table->next_number_field)
table->autoinc_field_has_explicit_non_null_value = true;
/*
@todo We should evaluate what other return values from save_in_field()
that should be treated as errors instead of checking thd->is_error().
The below call can reset TABLE::autoinc_field_has_explicit_non_null_value
flag depending on value provided (for details please see
set_field_to_null_with_conversions()). So evaluation of this flag can't
be moved outside of fill_record(), to be done once per statement.
*/
if (value->save_in_field(field, false) ==
TYPE_ERR_NULL_CONSTRAINT_VIOLATION ||
thd->is_error())
return true;
}
if (table->has_gcol() &&
update_generated_write_fields(bitmap ? bitmap : table->write_set, table))
return true;
assert(thd->is_error() ||
value_it == VisibleFields(values).end()); // No extra value!
/*
TABLE::autoinc_field_has_explicit_non_null_value should not be set to
true in raise_autoinc_has_expl_non_null_val == false mode.
*/
assert(table->autoinc_field_has_explicit_non_null_value == false ||
raise_autoinc_has_expl_non_null_val);
return thd->is_error();
}
/**
Fill fields in array with values from the list of items and invoke
before triggers.
@param thd Thread context.
@param ptr NULL-ended array of fields to be filled.
@param values Values to fill with.
@param table TABLE-object holding list of triggers to be invoked.
@param event Event type for triggers to be invoked.
@param num_fields Number of fields in table.
@note This function assumes that fields which values will be set and triggers
to be invoked belong to the same table, and that TABLE::record[0] and
record[1] buffers correspond to new and old versions of row
respectively.
@note This function is called during handling of statements INSERT/
INSERT SELECT/CREATE SELECT. It means that the only trigger's type
that can be invoked when this function is called is a BEFORE INSERT
trigger so we don't need to make branching based on the result of
execution function command_can_invoke_insert_triggers().
@note Unlike another version of fill_record_n_invoke_before_triggers() this
call tries to set TABLE::autoinc_field_has_explicit_non_null_value to
correct value unconditionally. So this flag can be set to true (even
in case of failure!) and the caller should make sure that it is reset
appropriately (@sa fill_record()).
@retval false OK
@retval true Error occurred.
*/
bool fill_record_n_invoke_before_triggers(THD *thd, Field **ptr,
const mem_root_deque<Item *> &values,
TABLE *table,
enum enum_trigger_event_type event,
int num_fields) {
bool rc;
Trigger_chain *tc =
table->triggers != nullptr
? table->triggers->get_triggers(event, TRG_ACTION_BEFORE)
: nullptr;
if (tc != nullptr) {
assert(command_can_invoke_insert_triggers(event, thd->lex->sql_command));
assert(num_fields);
table->triggers->enable_fields_temporary_nullability(thd);
MY_BITMAP insert_into_fields_bitmap;
bitmap_init(&insert_into_fields_bitmap, nullptr, num_fields);
rc = fill_record(thd, table, ptr, values, nullptr,
&insert_into_fields_bitmap, true);
if (!rc)
rc = call_before_insert_triggers(thd, table, event,
&insert_into_fields_bitmap);
/*
Re-calculate generated fields to cater for cases when base columns are
updated by the triggers.
*/
if (!rc && *ptr) {
TABLE *table_p = (*ptr)->table;
if (table_p->has_gcol() &&
tc->has_updated_trigger_fields(table_p->write_set)) {
// Dont save old value while re-calculating generated fields.
// Before image will already be saved in the first calculation.
table_p->blobs_need_not_keep_old_value();
rc = update_generated_write_fields(table_p->write_set, table_p);
}
}
bitmap_free(&insert_into_fields_bitmap);
table->triggers->disable_fields_temporary_nullability();
} else
rc = fill_record(thd, table, ptr, values, nullptr, nullptr, true);
if (rc) return true;
return check_inserting_record(thd, ptr);
}
/**
Drop all temporary tables which have been left from previous server run.
Used on server start-up.
@return False on success, true on error.
*/
bool mysql_rm_tmp_tables(void) {
uint i, idx;
char filePath[FN_REFLEN], *tmpdir;
MY_DIR *dirp;
FILEINFO *file;
THD *thd;
List<LEX_STRING> files;
List_iterator<LEX_STRING> files_it;
LEX_STRING *file_str;
bool result = true;
DBUG_TRACE;
if (!(thd = new THD)) return true; /* purecov: inspected */
thd->thread_stack = (char *)&thd;
thd->store_globals();
MEM_ROOT files_root(PSI_NOT_INSTRUMENTED, 32768);
for (i = 0; i <= mysql_tmpdir_list.max; i++) {
tmpdir = mysql_tmpdir_list.list[i];
/* See if the directory exists */
if (!(dirp = my_dir(tmpdir, MYF(MY_WME | MY_DONT_SORT)))) continue;
/* Find all SQLxxx files in the directory. */
for (idx = 0; idx < dirp->number_off_files; idx++) {
file = dirp->dir_entry + idx;
/* skipping . and .. */
if (file->name[0] == '.' &&
(!file->name[1] || (file->name[1] == '.' && !file->name[2])))
continue;
if (strlen(file->name) > tmp_file_prefix_length &&
!memcmp(file->name, tmp_file_prefix, tmp_file_prefix_length)) {
size_t filePath_len = snprintf(filePath, sizeof(filePath), "%s%c%s",
tmpdir, FN_LIBCHAR, file->name);
file_str = make_lex_string_root(&files_root, filePath, filePath_len);
if (file_str == nullptr || files.push_back(file_str, &files_root)) {
/* purecov: begin inspected */
my_dirend(dirp);
goto err;
/* purecov: end */
}
}
}
my_dirend(dirp);
}
/*
Ask SEs to delete temporary tables.
Pass list of SQLxxx files as a reference.
*/
result = ha_rm_tmp_tables(thd, &files);
/* Mimic old behavior, remove suspicious files if SE have not done this. */
files_it.init(files);
while ((file_str = files_it++))
(void)mysql_file_delete(key_file_misc, file_str->str, MYF(0));
err:
files_root.Clear();
delete thd;
return result;
}
/*****************************************************************************
unireg support functions
*****************************************************************************/
/*
free all unused tables
NOTE
This is called by 'handle_manager' when one wants to periodicly flush
all not used tables.
*/
void tdc_flush_unused_tables() {
table_cache_manager.lock_all_and_tdc();
table_cache_manager.free_all_unused_tables();
table_cache_manager.unlock_all_and_tdc();
}
/**
Remove all or some (depending on parameter) instances of TABLE and
TABLE_SHARE from the table definition cache.
@param thd Thread context
@param remove_type Type of removal:
TDC_RT_REMOVE_ALL - remove all TABLE instances and
TABLE_SHARE instance. There
should be no used TABLE objects
and caller should have exclusive
metadata lock on the table.
TDC_RT_REMOVE_NOT_OWN - remove all TABLE instances
except those that belong to
this thread. There should be
no TABLE objects used by other
threads and caller should have
exclusive metadata lock on the
table.
TDC_RT_REMOVE_UNUSED - remove all unused TABLE
instances (if there are no
used instances will also
remove TABLE_SHARE).
TDC_RT_REMOVE_NOT_OWN_KEEP_SHARE -
remove all TABLE instances
except those that belong to
this thread, but don't mark
TABLE_SHARE as old. There
should be no TABLE objects
used by other threads and
caller should have exclusive
metadata lock on the table.
TDC_RT_MARK_FOR_REOPEN - remove all unused TABLE
instances, mark used TABLE
instances as needing reopen.
@param db Name of database
@param table_name Name of table
@param has_lock If true, LOCK_open is already acquired
@note It assumes that table instances are already not used by any
(other) thread (this should be achieved by using meta-data locks).
*/
void tdc_remove_table(THD *thd, enum_tdc_remove_table_type remove_type,
const char *db, const char *table_name, bool has_lock) {
char key[MAX_DBKEY_LENGTH];
size_t key_length;
if (!has_lock)
table_cache_manager.lock_all_and_tdc();
else
table_cache_manager.assert_owner_all_and_tdc();
assert(remove_type == TDC_RT_REMOVE_UNUSED ||
remove_type == TDC_RT_MARK_FOR_REOPEN ||
thd->mdl_context.owns_equal_or_stronger_lock(
MDL_key::TABLE, db, table_name, MDL_EXCLUSIVE));
key_length = create_table_def_key(db, table_name, key);
auto it = table_def_cache->find(string(key, key_length));
// If the table has a shadow copy in a secondary storage engine, or
// if we don't know if the table has a shadow copy, we must also
// attempt to evict the secondary table from the cache.
const bool remove_secondary =
it == table_def_cache->end() || it->second->has_secondary_engine();
// Helper function that evicts the TABLE_SHARE pointed to by an iterator.
auto remove_table = [&](Table_definition_cache::iterator my_it) {
if (my_it == table_def_cache->end()) return;
TABLE_SHARE *share = my_it->second.get();
/*
Since share->ref_count is incremented when a table share is opened
in get_table_share(), before LOCK_open is temporarily released, it
is sufficient to check this condition alone and ignore the
share->m_open_in_progress flag.
Note that it is safe to call table_cache_manager.free_table() for
shares with m_open_in_progress == true, since such shares don't
have any TABLE objects associated.
*/
if (share->ref_count() > 0) {
/*
Set share's version to zero in order to ensure that it gets
automatically deleted once it is no longer referenced.
Note that code in TABLE_SHARE::wait_for_old_version() assumes
that marking share as old and removal of its unused tables
and of the share itself from TDC happens atomically under
protection of LOCK_open, or, putting it another way, that
TDC does not contain old shares which don't have any tables
used.
*/
if (remove_type != TDC_RT_REMOVE_NOT_OWN_KEEP_SHARE &&
remove_type != TDC_RT_MARK_FOR_REOPEN)
share->clear_version();
table_cache_manager.free_table(thd, remove_type, share);
} else if (remove_type != TDC_RT_MARK_FOR_REOPEN) {
// There are no TABLE objects associated, so just remove the
// share immediately. (Assert: When called with
// TDC_RT_REMOVE_NOT_OWN_KEEP_SHARE, there should always be a
// TABLE object associated with the primary TABLE_SHARE.)
assert(remove_type != TDC_RT_REMOVE_NOT_OWN_KEEP_SHARE ||
share->is_secondary_engine());
table_def_cache->erase(to_string(share->table_cache_key));
}
};
remove_table(it);
if (remove_secondary)
remove_table(
table_def_cache->find(create_table_def_key_secondary(db, table_name)));
if (!has_lock) table_cache_manager.unlock_all_and_tdc();
}
int setup_ftfuncs(const THD *thd, Query_block *query_block) {
assert(query_block->has_ft_funcs());
List_iterator<Item_func_match> li(*(query_block->ftfunc_list)),
lj(*(query_block->ftfunc_list));
Item_func_match *ftf, *ftf2;
while ((ftf = li++)) {
if (ftf->table_ref && ftf->fix_index(thd)) return 1;
lj.rewind();
/*
Notice that expressions added late (e.g. in ORDER BY) may be deleted
during resolving. It is therefore important that an "early" expression
is used as master for a "late" one, and not the other way around.
*/
while ((ftf2 = lj++) != ftf) {
if (ftf->eq(ftf2, true) && !ftf->master) ftf2->set_master(ftf);
}
}
return 0;
}
bool init_ftfuncs(THD *thd, Query_block *query_block) {
assert(query_block->has_ft_funcs());
DBUG_PRINT("info", ("Performing FULLTEXT search"));
THD_STAGE_INFO(thd, stage_fulltext_initialization);
if (thd->lex->using_hypergraph_optimizer()) {
// Set the no_ranking hint if ranking of the results is not required. The
// old optimizer does this when it determines which scan to use. The
// hypergraph optimizer doesn't know until the full plan is built, so do it
// here, just before the full-text search is performed.
for (Item_func_match &ifm : *query_block->ftfunc_list) {
if (ifm.master == nullptr && ifm.can_skip_ranking()) {
ifm.get_hints()->set_hint_flag(FT_NO_RANKING);
}
}
}
for (Item_func_match &ifm : *query_block->ftfunc_list) {
if (ifm.init_search(thd)) {
return true;
}
}
return false;
}
/**
Open and lock transactional system tables for read.
One must call close_trans_system_tables() to close systems tables opened
with this call.
@param thd Thread context.
@param table_list List of tables to open.
@note THR_LOCK deadlocks are not possible here because of the
restrictions we put on opening and locking of system tables for writing.
Thus, the system tables can be opened and locked for reading even if some
other tables have already been opened and locked.
@note MDL-deadlocks are possible, but they are properly detected and
reported.
@note Row-level deadlocks should be either avoided altogether using
non-locking reads (as it is done now for InnoDB), or should be correctly
detected and reported (in case of other transactional SE).
@note It is now technically possible to open non-transactional tables
(MyISAM system tables) using this function. That situation might still happen
if the user run the server on the elder data-directory or manually alters the
system tables to reside in MyISAM instead of InnoDB. It will be forbidden in
the future.
@return Error status.
*/
bool open_trans_system_tables_for_read(THD *thd, Table_ref *table_list) {
uint counter;
uint flags = MYSQL_OPEN_IGNORE_FLUSH | MYSQL_LOCK_IGNORE_TIMEOUT;
DBUG_TRACE;
assert(!thd->is_attachable_ro_transaction_active());
// Begin attachable transaction.
thd->begin_attachable_ro_transaction();
// Open tables.
if (open_tables(thd, &table_list, &counter, flags)) {
thd->end_attachable_transaction();
return true;
}
// Check the tables.
for (Table_ref *t = table_list; t; t = t->next_global) {
// Ensure the t are in storage engines, which are compatible with the
// attachable transaction requirements.
if ((t->table->file->ha_table_flags() & HA_ATTACHABLE_TRX_COMPATIBLE) ==
0) {
// Crash in the debug build ...
assert(!"HA_ATTACHABLE_TRX_COMPATIBLE is not set");
// ... or report an error in the release build.
my_error(ER_UNKNOWN_ERROR, MYF(0));
thd->end_attachable_transaction();
return true;
}
// The table should be in a transaction SE. This is not strict requirement
// however. It will be make more strict in the future.
if (!t->table->file->has_transactions())
LogErr(WARNING_LEVEL, ER_SYSTEM_TABLE_NOT_TRANSACTIONAL,
static_cast<int>(t->table_name_length), t->table_name);
}
// Lock the tables.
if (lock_tables(thd, table_list, counter, flags)) {
thd->end_attachable_transaction();
return true;
}
// Mark the table columns for use.
for (Table_ref *tables = table_list; tables; tables = tables->next_global)
tables->table->use_all_columns();
return false;
}
/**
Close transactional system tables, opened with
open_trans_system_tables_for_read().
@param thd Thread context.
*/
void close_trans_system_tables(THD *thd) { thd->end_attachable_transaction(); }
/**
A helper function to close a mysql.* table opened
in an auxiliary THD during bootstrap or in the main
connection, when we know that there are no locks
held by the connection due to a preceding implicit
commit.
This function assumes that there is no
statement transaction started for the operation
itself, since mysql.* tables are not transactional
and when they are used the binlog is off (DDL
binlogging is always statement-based.
We need this function since we'd like to not
just close the system table, but also release
the metadata lock on it.
Note, that in LOCK TABLES mode this function
does not release the metadata lock. But in this
mode the table can be opened only if it is locked
explicitly with LOCK TABLES.
*/
void close_mysql_tables(THD *thd) {
/* No need to commit/rollback statement transaction, it's not started. */
assert(thd->get_transaction()->is_empty(Transaction_ctx::STMT));
close_thread_tables(thd);
thd->mdl_context.release_transactional_locks();
}
/**
Open a log table.
Opening such tables is performed internally in the server
implementation, and is a 'nested' open, since some tables
might be already opened by the current thread.
The thread context before this call is saved, and is restored
when calling close_log_table().
@param thd The current thread
@param one_table Log table to open
@param [out] backup Temporary storage used to save the thread context
*/
TABLE *open_log_table(THD *thd, Table_ref *one_table,
Open_tables_backup *backup) {
uint flags = (MYSQL_OPEN_IGNORE_GLOBAL_READ_LOCK |
MYSQL_LOCK_IGNORE_GLOBAL_READ_ONLY | MYSQL_OPEN_IGNORE_FLUSH |
MYSQL_LOCK_IGNORE_TIMEOUT | MYSQL_LOCK_LOG_TABLE);
TABLE *table;
DBUG_TRACE;
thd->reset_n_backup_open_tables_state(backup,
Open_tables_state::SYSTEM_TABLES);
if ((table = open_ltable(thd, one_table, one_table->lock_descriptor().type,
flags))) {
assert(table->s->table_category == TABLE_CATEGORY_LOG);
/* Make sure all columns get assigned to a default value */
table->use_all_columns();
assert(table->no_replicate);
} else
thd->restore_backup_open_tables_state(backup);
return table;
}
/**
Close a log table.
The last table opened by open_log_table()
is closed, then the thread context is restored.
@param thd The current thread
@param backup The context to restore.
*/
void close_log_table(THD *thd, Open_tables_backup *backup) {
Query_tables_list query_tables_list_backup;
/*
In order not affect execution of current statement we have to
backup/reset/restore Query_tables_list part of LEX, which is
accessed and updated in the process of closing tables.
*/
thd->lex->reset_n_backup_query_tables_list(&query_tables_list_backup);
close_thread_tables(thd);
thd->lex->restore_backup_query_tables_list(&query_tables_list_backup);
thd->restore_backup_open_tables_state(backup);
}
/**
@} (end of group Data_Dictionary)
*/
|