{"diffoscope-json-version": 1, "source1": "/input1", "source2": "/input2", "unified_diff": null, "details": [{"source1": "zipinfo -v {}", "source2": "zipinfo -v {}", "unified_diff": "@@ -1376,15 +1376,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         29e31ece\n+  32-bit CRC value (hex):                         aca8b8ae\n   compressed size:                                46862 bytes\n   uncompressed size:                              46862 bytes\n   length of filename:                             47 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1448,15 +1448,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         8b11200c\n+  32-bit CRC value (hex):                         08ad488d\n   compressed size:                                29524 bytes\n   uncompressed size:                              29524 bytes\n   length of filename:                             46 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1556,15 +1556,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         9a475049\n+  32-bit CRC value (hex):                         8d09d58d\n   compressed size:                                28692 bytes\n   uncompressed size:                              28692 bytes\n   length of filename:                             54 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1592,15 +1592,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         4b7f0e44\n+  32-bit CRC value (hex):                         eabe2bfd\n   compressed size:                                375400 bytes\n   uncompressed size:                              375400 bytes\n   length of filename:                             70 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1628,15 +1628,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         fa1259ef\n+  32-bit CRC value (hex):                         0eafddc6\n   compressed size:                                376811 bytes\n   uncompressed size:                              376811 bytes\n   length of filename:                             62 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1700,15 +1700,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         ca82185b\n+  32-bit CRC value (hex):                         1358283e\n   compressed size:                                2043496 bytes\n   uncompressed size:                              2043496 bytes\n   length of filename:                             47 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1736,15 +1736,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         080ae529\n+  32-bit CRC value (hex):                         808ad0f1\n   compressed size:                                450559 bytes\n   uncompressed size:                              450559 bytes\n   length of filename:                             65 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1772,15 +1772,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         df9a119d\n+  32-bit CRC value (hex):                         2467d0ab\n   compressed size:                                148961 bytes\n   uncompressed size:                              148961 bytes\n   length of filename:                             63 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1808,15 +1808,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         b6f25e5e\n+  32-bit CRC value (hex):                         0b139ced\n   compressed size:                                664225 bytes\n   uncompressed size:                              664225 bytes\n   length of filename:                             64 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1844,15 +1844,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         a0b6a161\n+  32-bit CRC value (hex):                         c932b0e0\n   compressed size:                                1404907 bytes\n   uncompressed size:                              1404907 bytes\n   length of filename:                             65 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1880,15 +1880,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         b8fbde5d\n+  32-bit CRC value (hex):                         7652e079\n   compressed size:                                55426 bytes\n   uncompressed size:                              55426 bytes\n   length of filename:                             70 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1916,15 +1916,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         c55b515d\n+  32-bit CRC value (hex):                         07f83f9c\n   compressed size:                                285954 bytes\n   uncompressed size:                              285954 bytes\n   length of filename:                             66 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1952,15 +1952,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         77f8e89f\n+  32-bit CRC value (hex):                         12f8271b\n   compressed size:                                284613 bytes\n   uncompressed size:                              284613 bytes\n   length of filename:                             66 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n"}, {"source1": "zipdetails --redact --scan --utc {}", "source2": "zipdetails --redact --scan --utc {}", "unified_diff": "@@ -1013,15 +1013,15 @@\n 0444AC Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 0444AE   Length              0005 (5)\n 0444B0   Flags               01 (1) 'Modification'\n 0444B1   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 0444B5 PAYLOAD\n \n 04FBC3 DATA DESCRIPTOR       08074B50 (134695760)\n-04FBC7 CRC                   29E31ECE (702750414)\n+04FBC7 CRC                   ACA8B8AE (2896738478)\n 04FBCB Compressed Size       0000B70E (46862)\n 04FBCF Uncompressed Size     0000B70E (46862)\n \n 04FBD3 LOCAL HEADER #40      04034B50 (67324752)\n 04FBD7 Extract Zip Spec      14 (20) '2.0'\n 04FBD8 Extract OS            00 (0) 'MS-DOS'\n 04FBD9 General Purpose Flag  0008 (8)\n@@ -1069,15 +1069,15 @@\n 050184 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 050186   Length              0005 (5)\n 050188   Flags               01 (1) 'Modification'\n 050189   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 05018D PAYLOAD\n \n 0574E1 DATA DESCRIPTOR       08074B50 (134695760)\n-0574E5 CRC                   8B11200C (2333155340)\n+0574E5 CRC                   08AD488D (145574029)\n 0574E9 Compressed Size       00007354 (29524)\n 0574ED Uncompressed Size     00007354 (29524)\n \n 0574F1 LOCAL HEADER #42      04034B50 (67324752)\n 0574F5 Extract Zip Spec      14 (20) '2.0'\n 0574F6 Extract OS            00 (0) 'MS-DOS'\n 0574F7 General Purpose Flag  0008 (8)\n@@ -1153,15 +1153,15 @@\n 05A952 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 05A954   Length              0005 (5)\n 05A956   Flags               01 (1) 'Modification'\n 05A957   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 05A95B PAYLOAD\n \n 06196F DATA DESCRIPTOR       08074B50 (134695760)\n-061973 CRC                   9A475049 (2588364873)\n+061973 CRC                   8D09D58D (2366231949)\n 061977 Compressed Size       00007014 (28692)\n 06197B Uncompressed Size     00007014 (28692)\n \n 06197F LOCAL HEADER #45      04034B50 (67324752)\n 061983 Extract Zip Spec      14 (20) '2.0'\n 061984 Extract OS            00 (0) 'MS-DOS'\n 061985 General Purpose Flag  0008 (8)\n@@ -1181,15 +1181,15 @@\n 0619E3 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 0619E5   Length              0005 (5)\n 0619E7   Flags               01 (1) 'Modification'\n 0619E8   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 0619EC PAYLOAD\n \n 0BD454 DATA DESCRIPTOR       08074B50 (134695760)\n-0BD458 CRC                   4B7F0E44 (1266617924)\n+0BD458 CRC                   EABE2BFD (3938331645)\n 0BD45C Compressed Size       0005BA68 (375400)\n 0BD460 Uncompressed Size     0005BA68 (375400)\n \n 0BD464 LOCAL HEADER #46      04034B50 (67324752)\n 0BD468 Extract Zip Spec      14 (20) '2.0'\n 0BD469 Extract OS            00 (0) 'MS-DOS'\n 0BD46A General Purpose Flag  0008 (8)\n@@ -1209,15 +1209,15 @@\n 0BD4C0 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 0BD4C2   Length              0005 (5)\n 0BD4C4   Flags               01 (1) 'Modification'\n 0BD4C5   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 0BD4C9 PAYLOAD\n \n 1194B4 DATA DESCRIPTOR       08074B50 (134695760)\n-1194B8 CRC                   FA1259EF (4195506671)\n+1194B8 CRC                   0EAFDDC6 (246406598)\n 1194BC Compressed Size       0005BFEB (376811)\n 1194C0 Uncompressed Size     0005BFEB (376811)\n \n 1194C4 LOCAL HEADER #47      04034B50 (67324752)\n 1194C8 Extract Zip Spec      14 (20) '2.0'\n 1194C9 Extract OS            00 (0) 'MS-DOS'\n 1194CA General Purpose Flag  0008 (8)\n@@ -1265,15 +1265,15 @@\n 12F4AC Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 12F4AE   Length              0005 (5)\n 12F4B0   Flags               01 (1) 'Modification'\n 12F4B1   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 12F4B5 PAYLOAD\n \n 32231D DATA DESCRIPTOR       08074B50 (134695760)\n-322321 CRC                   CA82185B (3397523547)\n+322321 CRC                   1358283E (324544574)\n 322325 Compressed Size       001F2E68 (2043496)\n 322329 Uncompressed Size     001F2E68 (2043496)\n \n 32232D LOCAL HEADER #49      04034B50 (67324752)\n 322331 Extract Zip Spec      14 (20) '2.0'\n 322332 Extract OS            00 (0) 'MS-DOS'\n 322333 General Purpose Flag  0008 (8)\n@@ -1293,15 +1293,15 @@\n 32238C Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 32238E   Length              0005 (5)\n 322390   Flags               01 (1) 'Modification'\n 322391   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 322395 PAYLOAD\n \n 390394 DATA DESCRIPTOR       08074B50 (134695760)\n-390398 CRC                   080AE529 (134931753)\n+390398 CRC                   808AD0F1 (2156581105)\n 39039C Compressed Size       0006DFFF (450559)\n 3903A0 Uncompressed Size     0006DFFF (450559)\n \n 3903A4 LOCAL HEADER #50      04034B50 (67324752)\n 3903A8 Extract Zip Spec      14 (20) '2.0'\n 3903A9 Extract OS            00 (0) 'MS-DOS'\n 3903AA General Purpose Flag  0008 (8)\n@@ -1321,15 +1321,15 @@\n 390401 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 390403   Length              0005 (5)\n 390405   Flags               01 (1) 'Modification'\n 390406   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 39040A PAYLOAD\n \n 3B49EB DATA DESCRIPTOR       08074B50 (134695760)\n-3B49EF CRC                   DF9A119D (3751416221)\n+3B49EF CRC                   2467D0AB (610783403)\n 3B49F3 Compressed Size       000245E1 (148961)\n 3B49F7 Uncompressed Size     000245E1 (148961)\n \n 3B49FB LOCAL HEADER #51      04034B50 (67324752)\n 3B49FF Extract Zip Spec      14 (20) '2.0'\n 3B4A00 Extract OS            00 (0) 'MS-DOS'\n 3B4A01 General Purpose Flag  0008 (8)\n@@ -1349,15 +1349,15 @@\n 3B4A59 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 3B4A5B   Length              0005 (5)\n 3B4A5D   Flags               01 (1) 'Modification'\n 3B4A5E   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 3B4A62 PAYLOAD\n \n 456D03 DATA DESCRIPTOR       08074B50 (134695760)\n-456D07 CRC                   B6F25E5E (3069337182)\n+456D07 CRC                   0B139CED (185834733)\n 456D0B Compressed Size       000A22A1 (664225)\n 456D0F Uncompressed Size     000A22A1 (664225)\n \n 456D13 LOCAL HEADER #52      04034B50 (67324752)\n 456D17 Extract Zip Spec      14 (20) '2.0'\n 456D18 Extract OS            00 (0) 'MS-DOS'\n 456D19 General Purpose Flag  0008 (8)\n@@ -1377,15 +1377,15 @@\n 456D72 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 456D74   Length              0005 (5)\n 456D76   Flags               01 (1) 'Modification'\n 456D77   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 456D7B PAYLOAD\n \n 5ADD66 DATA DESCRIPTOR       08074B50 (134695760)\n-5ADD6A CRC                   A0B6A161 (2696323425)\n+5ADD6A CRC                   C932B0E0 (3375542496)\n 5ADD6E Compressed Size       00156FEB (1404907)\n 5ADD72 Uncompressed Size     00156FEB (1404907)\n \n 5ADD76 LOCAL HEADER #53      04034B50 (67324752)\n 5ADD7A Extract Zip Spec      14 (20) '2.0'\n 5ADD7B Extract OS            00 (0) 'MS-DOS'\n 5ADD7C General Purpose Flag  0008 (8)\n@@ -1405,15 +1405,15 @@\n 5ADDDA Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 5ADDDC   Length              0005 (5)\n 5ADDDE   Flags               01 (1) 'Modification'\n 5ADDDF   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 5ADDE3 PAYLOAD\n \n 5BB665 DATA DESCRIPTOR       08074B50 (134695760)\n-5BB669 CRC                   B8FBDE5D (3103514205)\n+5BB669 CRC                   7652E079 (1985142905)\n 5BB66D Compressed Size       0000D882 (55426)\n 5BB671 Uncompressed Size     0000D882 (55426)\n \n 5BB675 LOCAL HEADER #54      04034B50 (67324752)\n 5BB679 Extract Zip Spec      14 (20) '2.0'\n 5BB67A Extract OS            00 (0) 'MS-DOS'\n 5BB67B General Purpose Flag  0008 (8)\n@@ -1433,15 +1433,15 @@\n 5BB6D5 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 5BB6D7   Length              0005 (5)\n 5BB6D9   Flags               01 (1) 'Modification'\n 5BB6DA   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 5BB6DE PAYLOAD\n \n 6013E0 DATA DESCRIPTOR       08074B50 (134695760)\n-6013E4 CRC                   C55B515D (3311096157)\n+6013E4 CRC                   07F83F9C (133709724)\n 6013E8 Compressed Size       00045D02 (285954)\n 6013EC Uncompressed Size     00045D02 (285954)\n \n 6013F0 LOCAL HEADER #55      04034B50 (67324752)\n 6013F4 Extract Zip Spec      14 (20) '2.0'\n 6013F5 Extract OS            00 (0) 'MS-DOS'\n 6013F6 General Purpose Flag  0008 (8)\n@@ -1461,15 +1461,15 @@\n 601450 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 601452   Length              0005 (5)\n 601454   Flags               01 (1) 'Modification'\n 601455   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 601459 PAYLOAD\n \n 646C1E DATA DESCRIPTOR       08074B50 (134695760)\n-646C22 CRC                   77F8E89F (2012801183)\n+646C22 CRC                   12F8271B (318252827)\n 646C26 Compressed Size       000457C5 (284613)\n 646C2A Uncompressed Size     000457C5 (284613)\n \n 646C2E LOCAL HEADER #56      04034B50 (67324752)\n 646C32 Extract Zip Spec      14 (20) '2.0'\n 646C33 Extract OS            00 (0) 'MS-DOS'\n 646C34 General Purpose Flag  0008 (8)\n@@ -2884,15 +2884,15 @@\n 658402 Created OS            00 (0) 'MS-DOS'\n 658403 Extract Zip Spec      14 (20) '2.0'\n 658404 Extract OS            00 (0) 'MS-DOS'\n 658405 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 658407 Compression Method    0000 (0) 'Stored'\n 658409 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-65840D CRC                   29E31ECE (702750414)\n+65840D CRC                   ACA8B8AE (2896738478)\n 658411 Compressed Size       0000B70E (46862)\n 658415 Uncompressed Size     0000B70E (46862)\n 658419 Filename Length       002F (47)\n 65841B Extra Length          0009 (9)\n 65841D Comment Length        0000 (0)\n 65841F Disk Start            0000 (0)\n 658421 Int File Attributes   0000 (0)\n@@ -2944,15 +2944,15 @@\n 6584CA Created OS            00 (0) 'MS-DOS'\n 6584CB Extract Zip Spec      14 (20) '2.0'\n 6584CC Extract OS            00 (0) 'MS-DOS'\n 6584CD General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 6584CF Compression Method    0000 (0) 'Stored'\n 6584D1 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-6584D5 CRC                   8B11200C (2333155340)\n+6584D5 CRC                   08AD488D (145574029)\n 6584D9 Compressed Size       00007354 (29524)\n 6584DD Uncompressed Size     00007354 (29524)\n 6584E1 Filename Length       002E (46)\n 6584E3 Extra Length          0009 (9)\n 6584E5 Comment Length        0000 (0)\n 6584E7 Disk Start            0000 (0)\n 6584E9 Int File Attributes   0000 (0)\n@@ -3034,15 +3034,15 @@\n 658601 Created OS            00 (0) 'MS-DOS'\n 658602 Extract Zip Spec      14 (20) '2.0'\n 658603 Extract OS            00 (0) 'MS-DOS'\n 658604 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 658606 Compression Method    0000 (0) 'Stored'\n 658608 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-65860C CRC                   9A475049 (2588364873)\n+65860C CRC                   8D09D58D (2366231949)\n 658610 Compressed Size       00007014 (28692)\n 658614 Uncompressed Size     00007014 (28692)\n 658618 Filename Length       0036 (54)\n 65861A Extra Length          0009 (9)\n 65861C Comment Length        0000 (0)\n 65861E Disk Start            0000 (0)\n 658620 Int File Attributes   0000 (0)\n@@ -3064,15 +3064,15 @@\n 65866E Created OS            00 (0) 'MS-DOS'\n 65866F Extract Zip Spec      14 (20) '2.0'\n 658670 Extract OS            00 (0) 'MS-DOS'\n 658671 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 658673 Compression Method    0000 (0) 'Stored'\n 658675 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-658679 CRC                   4B7F0E44 (1266617924)\n+658679 CRC                   EABE2BFD (3938331645)\n 65867D Compressed Size       0005BA68 (375400)\n 658681 Uncompressed Size     0005BA68 (375400)\n 658685 Filename Length       0046 (70)\n 658687 Extra Length          0009 (9)\n 658689 Comment Length        0000 (0)\n 65868B Disk Start            0000 (0)\n 65868D Int File Attributes   0000 (0)\n@@ -3094,15 +3094,15 @@\n 6586EB Created OS            00 (0) 'MS-DOS'\n 6586EC Extract Zip Spec      14 (20) '2.0'\n 6586ED Extract OS            00 (0) 'MS-DOS'\n 6586EE General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 6586F0 Compression Method    0000 (0) 'Stored'\n 6586F2 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-6586F6 CRC                   FA1259EF (4195506671)\n+6586F6 CRC                   0EAFDDC6 (246406598)\n 6586FA Compressed Size       0005BFEB (376811)\n 6586FE Uncompressed Size     0005BFEB (376811)\n 658702 Filename Length       003E (62)\n 658704 Extra Length          0009 (9)\n 658706 Comment Length        0000 (0)\n 658708 Disk Start            0000 (0)\n 65870A Int File Attributes   0000 (0)\n@@ -3154,15 +3154,15 @@\n 6587C3 Created OS            00 (0) 'MS-DOS'\n 6587C4 Extract Zip Spec      14 (20) '2.0'\n 6587C5 Extract OS            00 (0) 'MS-DOS'\n 6587C6 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 6587C8 Compression Method    0000 (0) 'Stored'\n 6587CA Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-6587CE CRC                   CA82185B (3397523547)\n+6587CE CRC                   1358283E (324544574)\n 6587D2 Compressed Size       001F2E68 (2043496)\n 6587D6 Uncompressed Size     001F2E68 (2043496)\n 6587DA Filename Length       002F (47)\n 6587DC Extra Length          0009 (9)\n 6587DE Comment Length        0000 (0)\n 6587E0 Disk Start            0000 (0)\n 6587E2 Int File Attributes   0000 (0)\n@@ -3184,15 +3184,15 @@\n 658829 Created OS            00 (0) 'MS-DOS'\n 65882A Extract Zip Spec      14 (20) '2.0'\n 65882B Extract OS            00 (0) 'MS-DOS'\n 65882C General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 65882E Compression Method    0000 (0) 'Stored'\n 658830 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-658834 CRC                   080AE529 (134931753)\n+658834 CRC                   808AD0F1 (2156581105)\n 658838 Compressed Size       0006DFFF (450559)\n 65883C Uncompressed Size     0006DFFF (450559)\n 658840 Filename Length       0041 (65)\n 658842 Extra Length          0009 (9)\n 658844 Comment Length        0000 (0)\n 658846 Disk Start            0000 (0)\n 658848 Int File Attributes   0000 (0)\n@@ -3214,15 +3214,15 @@\n 6588A1 Created OS            00 (0) 'MS-DOS'\n 6588A2 Extract Zip Spec      14 (20) '2.0'\n 6588A3 Extract OS            00 (0) 'MS-DOS'\n 6588A4 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 6588A6 Compression Method    0000 (0) 'Stored'\n 6588A8 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-6588AC CRC                   DF9A119D (3751416221)\n+6588AC CRC                   2467D0AB (610783403)\n 6588B0 Compressed Size       000245E1 (148961)\n 6588B4 Uncompressed Size     000245E1 (148961)\n 6588B8 Filename Length       003F (63)\n 6588BA Extra Length          0009 (9)\n 6588BC Comment Length        0000 (0)\n 6588BE Disk Start            0000 (0)\n 6588C0 Int File Attributes   0000 (0)\n@@ -3244,15 +3244,15 @@\n 658917 Created OS            00 (0) 'MS-DOS'\n 658918 Extract Zip Spec      14 (20) '2.0'\n 658919 Extract OS            00 (0) 'MS-DOS'\n 65891A General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 65891C Compression Method    0000 (0) 'Stored'\n 65891E Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-658922 CRC                   B6F25E5E (3069337182)\n+658922 CRC                   0B139CED (185834733)\n 658926 Compressed Size       000A22A1 (664225)\n 65892A Uncompressed Size     000A22A1 (664225)\n 65892E Filename Length       0040 (64)\n 658930 Extra Length          0009 (9)\n 658932 Comment Length        0000 (0)\n 658934 Disk Start            0000 (0)\n 658936 Int File Attributes   0000 (0)\n@@ -3274,15 +3274,15 @@\n 65898E Created OS            00 (0) 'MS-DOS'\n 65898F Extract Zip Spec      14 (20) '2.0'\n 658990 Extract OS            00 (0) 'MS-DOS'\n 658991 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 658993 Compression Method    0000 (0) 'Stored'\n 658995 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-658999 CRC                   A0B6A161 (2696323425)\n+658999 CRC                   C932B0E0 (3375542496)\n 65899D Compressed Size       00156FEB (1404907)\n 6589A1 Uncompressed Size     00156FEB (1404907)\n 6589A5 Filename Length       0041 (65)\n 6589A7 Extra Length          0009 (9)\n 6589A9 Comment Length        0000 (0)\n 6589AB Disk Start            0000 (0)\n 6589AD Int File Attributes   0000 (0)\n@@ -3304,15 +3304,15 @@\n 658A06 Created OS            00 (0) 'MS-DOS'\n 658A07 Extract Zip Spec      14 (20) '2.0'\n 658A08 Extract OS            00 (0) 'MS-DOS'\n 658A09 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 658A0B Compression Method    0000 (0) 'Stored'\n 658A0D Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-658A11 CRC                   B8FBDE5D (3103514205)\n+658A11 CRC                   7652E079 (1985142905)\n 658A15 Compressed Size       0000D882 (55426)\n 658A19 Uncompressed Size     0000D882 (55426)\n 658A1D Filename Length       0046 (70)\n 658A1F Extra Length          0009 (9)\n 658A21 Comment Length        0000 (0)\n 658A23 Disk Start            0000 (0)\n 658A25 Int File Attributes   0000 (0)\n@@ -3334,15 +3334,15 @@\n 658A83 Created OS            00 (0) 'MS-DOS'\n 658A84 Extract Zip Spec      14 (20) '2.0'\n 658A85 Extract OS            00 (0) 'MS-DOS'\n 658A86 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 658A88 Compression Method    0000 (0) 'Stored'\n 658A8A Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-658A8E CRC                   C55B515D (3311096157)\n+658A8E CRC                   07F83F9C (133709724)\n 658A92 Compressed Size       00045D02 (285954)\n 658A96 Uncompressed Size     00045D02 (285954)\n 658A9A Filename Length       0042 (66)\n 658A9C Extra Length          0009 (9)\n 658A9E Comment Length        0000 (0)\n 658AA0 Disk Start            0000 (0)\n 658AA2 Int File Attributes   0000 (0)\n@@ -3364,15 +3364,15 @@\n 658AFC Created OS            00 (0) 'MS-DOS'\n 658AFD Extract Zip Spec      14 (20) '2.0'\n 658AFE Extract OS            00 (0) 'MS-DOS'\n 658AFF General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 658B01 Compression Method    0000 (0) 'Stored'\n 658B03 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-658B07 CRC                   77F8E89F (2012801183)\n+658B07 CRC                   12F8271B (318252827)\n 658B0B Compressed Size       000457C5 (284613)\n 658B0F Uncompressed Size     000457C5 (284613)\n 658B13 Filename Length       0042 (66)\n 658B15 Extra Length          0009 (9)\n 658B17 Comment Length        0000 (0)\n 658B19 Disk Start            0000 (0)\n 658B1B Int File Attributes   0000 (0)\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HintParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HintParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HintParser.g 2022-11-07 18:56:54\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HintParser.g 2025-01-31 11:18:20\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.hadoop.conf.Configuration;\n import org.apache.hadoop.hive.conf.HiveConf;\n \n \n@@ -823,15 +823,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: hintName, hintArgs\n+\t\t\t// elements: hintArgs, hintName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexer.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexer.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexer.g 2022-11-07 18:56:51\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexer.g 2025-01-31 11:18:17\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexerStandard.g 2022-11-07 18:56:54\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexerStandard.g 2025-01-31 11:18:20\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard_HiveLexerParent.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard_HiveLexerParent.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 HiveLexerParent.g 2022-11-07 18:56:54\n+// $ANTLR 3.5.2 HiveLexerParent.g 2025-01-31 11:18:20\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexer_HiveLexerParent.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexer_HiveLexerParent.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 HiveLexerParent.g 2022-11-07 18:56:51\n+// $ANTLR 3.5.2 HiveLexerParent.g 2025-01-31 11:18:17\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveParser.g 2022-11-07 19:16:39\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveParser.g 2025-01-31 11:18:17\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -3082,15 +3082,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_inputFileFormat.add(inputFileFormat50.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: inputFileFormat, path, islocal, isoverwrite, tab\n+\t\t\t// elements: islocal, tab, inputFileFormat, isoverwrite, path\n \t\t\t// token labels: islocal, path, isoverwrite\n \t\t\t// rule labels: tab, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3562,15 +3562,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableLocation.add(tableLocation62.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: ext, path, tab, tableLocation\n+\t\t\t// elements: path, ext, tab, tableLocation\n \t\t\t// token labels: ext, path\n \t\t\t// rule labels: tab, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4027,15 +4027,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replConfigs.add(replConf.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: dbName, replConf, sourceDbPolicy\n+\t\t\t// elements: replConf, sourceDbPolicy, dbName\n \t\t\t// token labels: \n \t\t\t// rule labels: dbName, sourceDbPolicy, retval, replConf\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4513,15 +4513,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replConfigs.add(replConf.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: replConf, dbName\n+\t\t\t// elements: dbName, replConf\n \t\t\t// token labels: \n \t\t\t// rule labels: dbName, retval, replConf\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -6270,15 +6270,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_dbProperties.add(dbprops.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: dbLocation, dbManagedLocation, dbprops, databaseComment, name, ifNotExists\n+\t\t\t\t\t// elements: databaseComment, ifNotExists, name, dbLocation, dbprops, dbManagedLocation\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, dbprops, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -6454,15 +6454,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_dbProperties.add(dbprops.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, ifNotExists, databaseComment, dbprops, dbConnectorName\n+\t\t\t\t\t// elements: databaseComment, dbprops, ifNotExists, dbConnectorName, name\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, dbprops, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -7214,15 +7214,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_restrictOrCascade.add(restrictOrCascade177.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: restrictOrCascade, ifExists, identifier\n+\t\t\t// elements: ifExists, identifier, restrictOrCascade\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -7483,15 +7483,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_force.add(force186.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: force, tablePartitionPrefix, columnNameList\n+\t\t\t// elements: tablePartitionPrefix, force, columnNameList\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -7653,15 +7653,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replicationClause.add(replicationClause192.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, KW_PURGE, replicationClause, ifExists\n+\t\t\t// elements: KW_PURGE, ifExists, replicationClause, tableName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -8735,15 +8735,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_extColumnName.add(extColumnName208.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: extColumnName, partitionSpec, tableName\n+\t\t\t// elements: tableName, partitionSpec, extColumnName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -9453,15 +9453,15 @@\n \t\t\t\t\tname=descFuncNames();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_descFuncNames.add(name.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, KW_EXTENDED\n+\t\t\t\t\t// elements: KW_EXTENDED, name\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9540,15 +9540,15 @@\n \t\t\t\t\tparttype=tabPartColTypeExpr();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tabPartColTypeExpr.add(parttype.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: parttype, descOptions\n+\t\t\t\t\t// elements: descOptions, parttype\n \t\t\t\t\t// token labels: descOptions\n \t\t\t\t\t// rule labels: parttype, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9790,15 +9790,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_COLUMNS, parttype, statsColumnName, noscan\n+\t\t\t\t\t// elements: statsColumnName, parttype, KW_COLUMNS, noscan\n \t\t\t\t\t// token labels: noscan\n \t\t\t\t\t// rule labels: statsColumnName, parttype, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10455,15 +10455,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showTablesFilterExpr.add(filter.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: db_name, filter, isExtended\n+\t\t\t\t\t// elements: isExtended, filter, db_name\n \t\t\t\t\t// token labels: isExtended\n \t\t\t\t\t// rule labels: filter, db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10608,15 +10608,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showStmtIdentifier.add(showStmtIdentifier241.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: db_name, showStmtIdentifier\n+\t\t\t\t\t// elements: showStmtIdentifier, db_name\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10756,15 +10756,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showStmtIdentifier.add(showStmtIdentifier249.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: db_name, showStmtIdentifier\n+\t\t\t\t\t// elements: showStmtIdentifier, db_name\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11043,15 +11043,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showFunctionIdentifier.add(showFunctionIdentifier264.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: showFunctionIdentifier, KW_LIKE\n+\t\t\t\t\t// elements: KW_LIKE, showFunctionIdentifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11178,15 +11178,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause270.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: limitClause, tabName, partitionSpec, orderByClause, whereClause\n+\t\t\t\t\t// elements: orderByClause, limitClause, whereClause, tabName, partitionSpec\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: tabName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11483,15 +11483,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec283.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: showStmtIdentifier, partitionSpec, db_name\n+\t\t\t\t\t// elements: db_name, showStmtIdentifier, partitionSpec\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11564,15 +11564,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableName, prptyName\n+\t\t\t\t\t// elements: prptyName, tableName\n \t\t\t\t\t// token labels: prptyName\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11972,15 +11972,15 @@\n \n \t\t\t\t\t\t\t\t\t}\n \t\t\t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t\t\t}\n \n \t\t\t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t\t\t// elements: dbName, isExtended\n+\t\t\t\t\t\t\t// elements: isExtended, dbName\n \t\t\t\t\t\t\t// token labels: isExtended\n \t\t\t\t\t\t\t// rule labels: dbName, retval\n \t\t\t\t\t\t\t// token list labels: \n \t\t\t\t\t\t\t// rule list labels: \n \t\t\t\t\t\t\t// wildcard labels: \n \t\t\t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\t\t\tretval.tree = root_0;\n@@ -12051,15 +12051,15 @@\n \n \t\t\t\t\t\t\t\t\t}\n \t\t\t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t\t\t}\n \n \t\t\t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t\t\t// elements: isExtended, parttype\n+\t\t\t\t\t\t\t// elements: parttype, isExtended\n \t\t\t\t\t\t\t// token labels: isExtended\n \t\t\t\t\t\t\t// rule labels: parttype, retval\n \t\t\t\t\t\t\t// token list labels: \n \t\t\t\t\t\t\t// rule list labels: \n \t\t\t\t\t\t\t// wildcard labels: \n \t\t\t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\t\t\tretval.tree = root_0;\n@@ -12300,15 +12300,15 @@\n \t\t\t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause302.getTree());\n \t\t\t\t\t\t\t\t\t}\n \t\t\t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t\t\t}\n \n \t\t\t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t\t\t// elements: compactionStatus, dbName, compactionPool, orderByClause, limitClause, compactionType\n+\t\t\t\t\t\t\t// elements: compactionPool, orderByClause, compactionType, compactionStatus, limitClause, dbName\n \t\t\t\t\t\t\t// token labels: \n \t\t\t\t\t\t\t// rule labels: dbName, retval\n \t\t\t\t\t\t\t// token list labels: \n \t\t\t\t\t\t\t// rule list labels: \n \t\t\t\t\t\t\t// wildcard labels: \n \t\t\t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\t\t\tretval.tree = root_0;\n@@ -12745,15 +12745,15 @@\n \t\t\t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause307.getTree());\n \t\t\t\t\t\t\t\t\t}\n \t\t\t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t\t\t}\n \n \t\t\t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t\t\t// elements: parttype, orderByClause, compactionType, compactionStatus, compactionPool, limitClause\n+\t\t\t\t\t\t\t// elements: parttype, limitClause, compactionType, compactionStatus, orderByClause, compactionPool\n \t\t\t\t\t\t\t// token labels: \n \t\t\t\t\t\t\t// rule labels: parttype, retval\n \t\t\t\t\t\t\t// token list labels: \n \t\t\t\t\t\t\t// rule list labels: \n \t\t\t\t\t\t\t// wildcard labels: \n \t\t\t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\t\t\tretval.tree = root_0;\n@@ -13390,15 +13390,15 @@\n \t\t\t\t\tEQUAL321=(Token)match(input,EQUAL,FOLLOW_EQUAL_in_showTablesFilterExpr5998); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_EQUAL.add(EQUAL321);\n \n \t\t\t\t\tStringLiteral322=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_showTablesFilterExpr6000); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(StringLiteral322);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: identifier, StringLiteral\n+\t\t\t\t\t// elements: StringLiteral, identifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -13566,15 +13566,15 @@\n \n \t\t\tpushFollow(FOLLOW_lockMode_in_lockStatement6070);\n \t\t\tlockMode330=lockMode();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_lockMode.add(lockMode330.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: lockMode, partitionSpec, tableName\n+\t\t\t// elements: tableName, lockMode, partitionSpec\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -13715,15 +13715,15 @@\n \n \t\t\tpushFollow(FOLLOW_lockMode_in_lockDatabase6124);\n \t\t\tlockMode334=lockMode();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_lockMode.add(lockMode334.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: lockMode, dbName\n+\t\t\t// elements: dbName, lockMode\n \t\t\t// token labels: \n \t\t\t// rule labels: dbName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -13895,15 +13895,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec339.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, partitionSpec\n+\t\t\t// elements: partitionSpec, tableName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14369,15 +14369,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_withGrantOption.add(withGrantOption351.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: privList, privilegeObject, principalSpecification, withGrantOption\n+\t\t\t// elements: privilegeObject, withGrantOption, privList, principalSpecification\n \t\t\t// token labels: \n \t\t\t// rule labels: privList, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14527,15 +14527,15 @@\n \n \t\t\tpushFollow(FOLLOW_principalSpecification_in_revokePrivileges6471);\n \t\t\tprincipalSpecification357=principalSpecification();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_principalSpecification.add(principalSpecification357.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: privilegeList, principalSpecification, privilegeObject, grantOptionFor\n+\t\t\t// elements: grantOptionFor, privilegeList, privilegeObject, principalSpecification\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14914,15 +14914,15 @@\n \n \t\t\tpushFollow(FOLLOW_principalSpecification_in_revokeRole6601);\n \t\t\tprincipalSpecification373=principalSpecification();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_principalSpecification.add(principalSpecification373.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: adminOptionFor, identifier, principalSpecification\n+\t\t\t// elements: adminOptionFor, principalSpecification, identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -15549,15 +15549,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_privilegeIncludeColObject.add(privilegeIncludeColObject390.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: principalName, privilegeIncludeColObject\n+\t\t\t// elements: privilegeIncludeColObject, principalName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -16419,15 +16419,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec402.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: partitionSpec, tableName\n+\t\t\t\t\t// elements: tableName, partitionSpec\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -17085,15 +17085,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec413.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableName, cols, partitionSpec\n+\t\t\t\t\t// elements: tableName, partitionSpec, cols\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: cols, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -18825,15 +18825,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: repair, opt, partitionSelectorSpec, tableName\n+\t\t\t// elements: opt, partitionSelectorSpec, repair, tableName\n \t\t\t// token labels: repair, opt\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -19058,15 +19058,15 @@\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_resourceType.add(resType.getTree());\n \t\t\tresPath=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_resource7978); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(resPath);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: resType, resPath\n+\t\t\t// elements: resPath, resType\n \t\t\t// token labels: resPath\n \t\t\t// rule labels: resType, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -19397,15 +19397,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_resourceList.add(rList.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: StringLiteral, rList, StringLiteral, rList, functionIdentifier, functionIdentifier\n+\t\t\t// elements: rList, functionIdentifier, rList, StringLiteral, functionIdentifier, StringLiteral\n \t\t\t// token labels: \n \t\t\t// rule labels: rList, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -19561,15 +19561,15 @@\n \n \t\t\tpushFollow(FOLLOW_functionIdentifier_in_dropFunctionStatement8202);\n \t\t\tfunctionIdentifier473=functionIdentifier();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_functionIdentifier.add(functionIdentifier473.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: functionIdentifier, ifExists, functionIdentifier, ifExists\n+\t\t\t// elements: functionIdentifier, functionIdentifier, ifExists, ifExists\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -19847,15 +19847,15 @@\n \n \t\t\tpushFollow(FOLLOW_expression_in_createMacroStatement8335);\n \t\t\texpression484=expression();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_expression.add(expression484.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: Identifier, columnNameTypeList, expression\n+\t\t\t// elements: columnNameTypeList, expression, Identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -19974,15 +19974,15 @@\n \n \t\t\t}\n \n \t\t\tIdentifier489=(Token)match(input,Identifier,FOLLOW_Identifier_in_dropMacroStatement8388); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_Identifier.add(Identifier489);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: Identifier, ifExists\n+\t\t\t// elements: ifExists, Identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -20232,15 +20232,15 @@\n \n \t\t\tpushFollow(FOLLOW_selectStatementWithCTE_in_createViewStatement8500);\n \t\t\tselectStatementWithCTE501=selectStatementWithCTE();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_selectStatementWithCTE.add(selectStatementWithCTE501.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: columnNameCommentList, orReplace, ifNotExists, name, selectStatementWithCTE, tableComment, tablePropertiesPrefixed, viewPartition\n+\t\t\t// elements: ifNotExists, tablePropertiesPrefixed, tableComment, selectStatementWithCTE, name, viewPartition, orReplace, columnNameCommentList\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -20961,15 +20961,15 @@\n \n \t\t\tpushFollow(FOLLOW_viewName_in_dropViewStatement8888);\n \t\t\tviewName527=viewName();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_viewName.add(viewName527.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: viewName, ifExists\n+\t\t\t// elements: ifExists, viewName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -21279,15 +21279,15 @@\n \n \t\t\tpushFollow(FOLLOW_selectStatementWithCTE_in_createMaterializedViewStatement8991);\n \t\t\tselectStatementWithCTE541=selectStatementWithCTE();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_selectStatementWithCTE.add(selectStatementWithCTE541.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableLocation, ifNotExists, tableRowFormat, tableFileFormat, viewOrganization, name, viewPartition, rewriteDisabled, selectStatementWithCTE, tablePropertiesPrefixed, tableComment\n+\t\t\t// elements: rewriteDisabled, tableRowFormat, viewOrganization, ifNotExists, tableFileFormat, viewPartition, name, tableLocation, selectStatementWithCTE, tableComment, tablePropertiesPrefixed\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -21456,15 +21456,15 @@\n \n \t\t\tpushFollow(FOLLOW_viewName_in_dropMaterializedViewStatement9168);\n \t\t\tviewName546=viewName();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_viewName.add(viewName546.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: viewName, ifExists\n+\t\t\t// elements: ifExists, viewName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -21619,15 +21619,15 @@\n \n \t\t\tpushFollow(FOLLOW_definedAsSpec_in_createScheduledQueryStatement9256);\n \t\t\tdefinedAsSpec553=definedAsSpec();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_definedAsSpec.add(definedAsSpec553.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: name, scheduleSpec, enableSpecification, definedAsSpec, executedAsSpec\n+\t\t\t// elements: name, executedAsSpec, scheduleSpec, enableSpecification, definedAsSpec\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -22263,15 +22263,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: qualifier, value, offsetTs\n+\t\t\t\t\t// elements: qualifier, offsetTs, value\n \t\t\t\t\t// token labels: offsetTs, value\n \t\t\t\t\t// rule labels: qualifier, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -24282,15 +24282,15 @@\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_columnName.add(columnName619.getTree());\n \t\t\t\t\tRPAREN620=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_partitionTransformType10413); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN620);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: columnName, value\n+\t\t\t\t\t// elements: value, columnName\n \t\t\t\t\t// token labels: value\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -24342,15 +24342,15 @@\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_columnName.add(columnName624.getTree());\n \t\t\t\t\tRPAREN625=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_partitionTransformType10461); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN625);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: columnName, value\n+\t\t\t\t\t// elements: value, columnName\n \t\t\t\t\t// token labels: value\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -24517,15 +24517,15 @@\n \t\t\tnum=(Token)match(input,Number,FOLLOW_Number_in_tableBuckets10551); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_Number.add(num);\n \n \t\t\tKW_BUCKETS635=(Token)match(input,KW_BUCKETS,FOLLOW_KW_BUCKETS_in_tableBuckets10553); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_KW_BUCKETS.add(KW_BUCKETS635);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: num, sortCols, bucketCols\n+\t\t\t// elements: sortCols, num, bucketCols\n \t\t\t// token labels: num\n \t\t\t// rule labels: bucketCols, sortCols, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -24691,15 +24691,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_storedAsDirs.add(storedAsDirs643.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: skewedCols, skewedValues, storedAsDirs\n+\t\t\t// elements: skewedValues, skewedCols, storedAsDirs\n \t\t\t// token labels: \n \t\t\t// rule labels: skewedCols, skewedValues, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -25527,15 +25527,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableRowNullFormat.add(tableRowNullFormat662.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableRowFormatMapKeysIdentifier, tableRowFormatCollItemsIdentifier, tableRowFormatLinesIdentifier, tableRowNullFormat, tableRowFormatFieldIdentifier\n+\t\t\t// elements: tableRowFormatFieldIdentifier, tableRowFormatLinesIdentifier, tableRowFormatCollItemsIdentifier, tableRowFormatMapKeysIdentifier, tableRowNullFormat\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -27113,15 +27113,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: outFmt, inDriver, inFmt, outDriver\n+\t\t\t\t\t// elements: outFmt, outDriver, inDriver, inFmt\n \t\t\t\t\t// token labels: inFmt, inDriver, outDriver, outFmt\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -27224,15 +27224,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(fileformat.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: fileformat, storageHandler, serdeprops\n+\t\t\t\t\t// elements: fileformat, serdeprops, storageHandler\n \t\t\t\t\t// token labels: storageHandler\n \t\t\t\t\t// rule labels: serdeprops, fileformat, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -27342,25 +27342,25 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(fileformat.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: genericSpec, fileformat, serdeprops\n+\t\t\t\t\t// elements: serdeprops, fileformat, genericSpec\n \t\t\t\t\t// token labels: \n-\t\t\t\t\t// rule labels: serdeprops, genericSpec, fileformat, retval\n+\t\t\t\t\t// rule labels: serdeprops, fileformat, genericSpec, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_serdeprops=new RewriteRuleSubtreeStream(adaptor,\"rule serdeprops\",serdeprops!=null?serdeprops.getTree():null);\n-\t\t\t\t\tRewriteRuleSubtreeStream stream_genericSpec=new RewriteRuleSubtreeStream(adaptor,\"rule genericSpec\",genericSpec!=null?genericSpec.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_fileformat=new RewriteRuleSubtreeStream(adaptor,\"rule fileformat\",fileformat!=null?fileformat.getTree():null);\n+\t\t\t\t\tRewriteRuleSubtreeStream stream_genericSpec=new RewriteRuleSubtreeStream(adaptor,\"rule genericSpec\",genericSpec!=null?genericSpec.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t\t\t// 2064:7: -> ^( TOK_STORAGEHANDLER $genericSpec ( $serdeprops)? ( ^( TOK_FILEFORMAT_GENERIC $fileformat) )? )\n \t\t\t\t\t{\n \t\t\t\t\t\t// org/apache/hadoop/hive/ql/parse/HiveParser.g:2064:10: ^( TOK_STORAGEHANDLER $genericSpec ( $serdeprops)? ( ^( TOK_FILEFORMAT_GENERIC $fileformat) )? )\n \t\t\t\t\t\t{\n@@ -29439,15 +29439,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate756.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintName, constraintOptsCreate, constraintOptsCreate\n+\t\t\t// elements: constraintOptsCreate, constraintOptsCreate, constraintName\n \t\t\t// token labels: \n \t\t\t// rule labels: constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -30039,24 +30039,24 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate771.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintName, fkCols, tabName, parCols, constraintOptsCreate, tabName, parCols, fkCols, constraintOptsCreate\n+\t\t\t// elements: parCols, parCols, constraintOptsCreate, tabName, constraintOptsCreate, constraintName, tabName, fkCols, fkCols\n \t\t\t// token labels: \n-\t\t\t// rule labels: tabName, parCols, fkCols, constraintName, retval\n+\t\t\t// rule labels: parCols, tabName, fkCols, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n-\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_parCols=new RewriteRuleSubtreeStream(adaptor,\"rule parCols\",parCols!=null?parCols.getTree():null);\n+\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_fkCols=new RewriteRuleSubtreeStream(adaptor,\"rule fkCols\",fkCols!=null?fkCols.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_constraintName=new RewriteRuleSubtreeStream(adaptor,\"rule constraintName\",constraintName!=null?constraintName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t// 2200:5: -> {$constraintName.tree != null}? ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) $fkCols $tabName $parCols ( constraintOptsCreate )? )\n \t\t\tif ((constraintName!=null?((ASTNode)constraintName.getTree()):null) != null) {\n@@ -30226,15 +30226,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsAlter.add(constraintOptsAlter776.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tabName, constraintName, constraintOptsAlter, parCols, fkCols\n+\t\t\t// elements: tabName, parCols, fkCols, constraintOptsAlter, constraintName\n \t\t\t// token labels: \n \t\t\t// rule labels: tabName, parCols, fkCols, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -31173,15 +31173,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_nullOrdering.add(nullSpec.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: nullSpec, identifier, nullSpec, identifier, nullSpec, identifier, identifier, identifier, identifier\n+\t\t\t// elements: identifier, identifier, nullSpec, identifier, identifier, nullSpec, identifier, nullSpec, identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: nullSpec, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -31505,15 +31505,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colName, comment\n+\t\t\t// elements: comment, colName\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -31776,15 +31776,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_nullOrdering.add(nullSpec.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, orderSpec, orderSpec, expression, nullSpec, orderSpec, expression, expression, orderSpec, orderSpec, expression, expression, expression, nullSpec, expression\n+\t\t\t// elements: expression, expression, expression, orderSpec, expression, orderSpec, expression, expression, orderSpec, expression, orderSpec, expression, nullSpec, orderSpec, nullSpec\n \t\t\t// token labels: \n \t\t\t// rule labels: orderSpec, nullSpec, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -32033,15 +32033,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colType, colName, colType, comment, colName\n+\t\t\t// elements: colType, comment, colName, colType, colName\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -32726,15 +32726,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: comment, colType, colName, columnConstraint\n+\t\t\t// elements: colType, columnConstraint, colName, comment\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -33078,15 +33078,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate818.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintOptsCreate, colName, tabName, tabName, constraintName, colName, constraintOptsCreate\n+\t\t\t// elements: colName, constraintName, constraintOptsCreate, tabName, colName, constraintOptsCreate, tabName\n \t\t\t// token labels: \n \t\t\t// rule labels: colName, tabName, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -33276,15 +33276,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate821.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintOptsCreate, constraintName, constraintOptsCreate\n+\t\t\t// elements: constraintName, constraintOptsCreate, constraintOptsCreate\n \t\t\t// token labels: \n \t\t\t// rule labels: constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -33639,24 +33639,24 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsAlter.add(constraintOptsAlter828.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintName, constraintOptsAlter, tabName, colName, colName, constraintOptsAlter, tabName\n+\t\t\t// elements: colName, tabName, colName, constraintName, constraintOptsAlter, constraintOptsAlter, tabName\n \t\t\t// token labels: \n-\t\t\t// rule labels: tabName, colName, constraintName, retval\n+\t\t\t// rule labels: colName, tabName, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n-\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_colName=new RewriteRuleSubtreeStream(adaptor,\"rule colName\",colName!=null?colName.getTree():null);\n+\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_constraintName=new RewriteRuleSubtreeStream(adaptor,\"rule constraintName\",constraintName!=null?constraintName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t// 2400:5: -> {$constraintName.tree != null}? ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) ^( TOK_TABCOLNAME ) $tabName ^( TOK_TABCOLNAME $colName) ( constraintOptsAlter )? )\n \t\t\tif ((constraintName!=null?((ASTNode)constraintName.getTree()):null) != null) {\n \t\t\t\t// org/apache/hadoop/hive/ql/parse/HiveParser.g:2401:13: ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) ^( TOK_TABCOLNAME ) $tabName ^( TOK_TABCOLNAME $colName) ( constraintOptsAlter )? )\n@@ -34918,15 +34918,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colName, comment, colName, colType, colType\n+\t\t\t// elements: colName, colType, colName, colType, comment\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -37736,15 +37736,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_setOperator.add(u.getTree());\n \t\t\t\t\tpushFollow(FOLLOW_singleFromStatement_in_fromStatement16156);\n \t\t\t\t\tr=singleFromStatement();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_singleFromStatement.add(r.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: u, r\n+\t\t\t\t\t// elements: r, u\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: r, u, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -38318,15 +38318,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_qualifyClause.add(q.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: w, s, win, q, f, g, h\n+\t\t\t\t\t// elements: q, win, w, h, s, f, g\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: q, s, f, w, g, h, win, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -38635,15 +38635,15 @@\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((c!=null?((ASTNode)c.getTree()):null));\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((d!=null?((ASTNode)d.getTree()):null));\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((sort!=null?((ASTNode)sort.getTree()):null));\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((l!=null?((ASTNode)l.getTree()):null));\n \t\t\t   }\n \t\t\t   }\n \t\t\t// AST REWRITE\n-\t\t\t// elements: l, c, o, d, sort\n+\t\t\t// elements: c, d, l, o, sort\n \t\t\t// token labels: \n \t\t\t// rule labels: c, d, sort, l, retval, o\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -38832,15 +38832,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_setOperator.add(u.getTree());\n \t\t\t\t\tpushFollow(FOLLOW_atomSelectStatement_in_setOpSelectStatement16951);\n \t\t\t\t\tb=atomSelectStatement();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_atomSelectStatement.add(b.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: u, b, b, b, u, b\n+\t\t\t\t\t// elements: b, u, b, u, b, b\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: b, u, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -39550,15 +39550,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause956.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: selectClause, limitClause, orderByClause, groupByClause, window_clause, insertClause, lateralView, sortByClause, qualifyClause, clusterByClause, distributeByClause, whereClause, havingClause\n+\t\t\t\t\t// elements: selectClause, lateralView, whereClause, groupByClause, havingClause, qualifyClause, insertClause, sortByClause, orderByClause, window_clause, clusterByClause, limitClause, distributeByClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -39875,15 +39875,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause968.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: groupByClause, window_clause, limitClause, distributeByClause, havingClause, clusterByClause, selectClause, qualifyClause, sortByClause, orderByClause, lateralView, whereClause\n+\t\t\t\t\t// elements: sortByClause, clusterByClause, havingClause, distributeByClause, window_clause, limitClause, lateralView, selectClause, groupByClause, qualifyClause, orderByClause, whereClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40125,15 +40125,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_ifNotExists.add(ifNotExists972.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ifNotExists, destination\n+\t\t\t\t\t// elements: destination, ifNotExists\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40222,15 +40222,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableOrPartition, targetCols\n+\t\t\t\t\t// elements: targetCols, tableOrPartition\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: targetCols, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40639,15 +40639,15 @@\n \n \t\t\t\t\tnum=(Token)match(input,Number,FOLLOW_Number_in_limitClause18168); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_Number.add(num);\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: num, offset\n+\t\t\t\t\t// elements: offset, num\n \t\t\t\t\t// token labels: offset, num\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40692,15 +40692,15 @@\n \t\t\t\t\tKW_OFFSET988=(Token)match(input,KW_OFFSET,FOLLOW_KW_OFFSET_in_limitClause18197); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_OFFSET.add(KW_OFFSET988);\n \n \t\t\t\t\toffset=(Token)match(input,Number,FOLLOW_Number_in_limitClause18201); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_Number.add(offset);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: offset, num\n+\t\t\t\t\t// elements: num, offset\n \t\t\t\t\t// token labels: offset, num\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40817,15 +40817,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_whereClause.add(whereClause992.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: whereClause, tableName\n+\t\t\t// elements: tableName, whereClause\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -41255,15 +41255,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_whereClause.add(whereClause1005.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, setColumnsClause, whereClause\n+\t\t\t// elements: setColumnsClause, whereClause, tableName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -42556,15 +42556,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_expression.add(expression1044.getTree());\n \t\t\tpushFollow(FOLLOW_whenClauses_in_mergeStatement18802);\n \t\t\twhenClauses1045=whenClauses();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_whenClauses.add(whenClauses1045.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, joinSourcePart, whenClauses, identifier, expression, QUERY_HINT\n+\t\t\t// elements: tableName, expression, identifier, QUERY_HINT, whenClauses, joinSourcePart\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -42997,15 +42997,15 @@\n \n \t\t\tpushFollow(FOLLOW_updateOrDelete_in_whenMatchedAndClause18967);\n \t\t\tupdateOrDelete1063=updateOrDelete();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_updateOrDelete.add(updateOrDelete1063.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, updateOrDelete\n+\t\t\t// elements: updateOrDelete, expression\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -44107,529 +44107,529 @@\n \t\t}\n \t\t}\n \n \t}\n \t// $ANTLR end synpred24_HiveParser\n \n \t// Delegated rules\n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropConstraint_return alterStatementSuffixDropConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropConstraint(); }\n-\n-\tpublic HiveParser_FromClauseParser.valuesTableConstructor_return valuesTableConstructor() throws RecognitionException { return gFromClauseParser.valuesTableConstructor(); }\n+\tpublic HiveParser_SelectClauseParser.window_clause_return window_clause() throws RecognitionException { return gSelectClauseParser.window_clause(); }\n \n-\tpublic HiveParser_AlterClauseParser.skewedLocationsList_return skewedLocationsList() throws RecognitionException { return gAlterClauseParser.skewedLocationsList(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceStarOperator_return precedenceStarOperator() throws RecognitionException { return gIdentifiersParser.precedenceStarOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.timestampLocalTZLiteral_return timestampLocalTZLiteral() throws RecognitionException { return gIdentifiersParser.timestampLocalTZLiteral(); }\n+\tpublic HiveParser_ResourcePlanParser.withReplace_return withReplace() throws RecognitionException { return gResourcePlanParser.withReplace(); }\n \n-\tpublic HiveParser_CreateDDLParser.dataConnectorUrl_return dataConnectorUrl() throws RecognitionException { return gCreateDDLParser.dataConnectorUrl(); }\n+\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation_return alterTblPartitionStatementSuffixSkewedLocation() throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectItem_return selectItem() throws RecognitionException { return gSelectClauseParser.selectItem(); }\n+\tpublic HiveParser_IdentifiersParser.function_return function() throws RecognitionException { return gIdentifiersParser.function(); }\n \n-\tpublic HiveParser_IdentifiersParser.functionName_return functionName() throws RecognitionException { return gIdentifiersParser.functionName(); }\n+\tpublic HiveParser_IdentifiersParser.expression_return expression() throws RecognitionException { return gIdentifiersParser.expression(); }\n \n-\tpublic HiveParser_IdentifiersParser.quantifierType_return quantifierType() throws RecognitionException { return gIdentifiersParser.quantifierType(); }\n+\tpublic HiveParser_IdentifiersParser.expressions_return expressions() throws RecognitionException { return gIdentifiersParser.expressions(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSelectorSpec_return partitionSelectorSpec() throws RecognitionException { return gIdentifiersParser.partitionSelectorSpec(); }\n+\tpublic HiveParser_IdentifiersParser.groupByEmpty_return groupByEmpty() throws RecognitionException { return gIdentifiersParser.groupByEmpty(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitionsElement_return alterStatementSuffixAddPartitionsElement() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitionsElement(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUnArchive_return alterStatementSuffixUnArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUnArchive(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceUnarySuffixExpression_return precedenceUnarySuffixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnarySuffixExpression(); }\n+\tpublic HiveParser_IdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName_return sql11ReservedKeywordsUsedAsFunctionName() throws RecognitionException { return gIdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName(); }\n \n-\tpublic HiveParser_IdentifiersParser.dateLiteral_return dateLiteral() throws RecognitionException { return gIdentifiersParser.dateLiteral(); }\n+\tpublic HiveParser_ResourcePlanParser.enable_return enable() throws RecognitionException { return gResourcePlanParser.enable(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_specification_return window_specification(CommonTree nullTreatment) throws RecognitionException { return gSelectClauseParser.window_specification(nullTreatment); }\n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetOwner_return alterDataConnectorSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetOwner(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceNotOperator_return precedenceNotOperator() throws RecognitionException { return gIdentifiersParser.precedenceNotOperator(); }\n+\tpublic HiveParser_IdentifiersParser.functionIdentifier_return functionIdentifier() throws RecognitionException { return gIdentifiersParser.functionIdentifier(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_value_expression_return window_value_expression() throws RecognitionException { return gSelectClauseParser.window_value_expression(); }\n+\tpublic HiveParser_CreateDDLParser.dropDataConnectorStatement_return dropDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.dropDataConnectorStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.splitSample_return splitSample() throws RecognitionException { return gFromClauseParser.splitSample(); }\n+\tpublic HiveParser_FromClauseParser.fromClause_return fromClause() throws RecognitionException { return gFromClauseParser.fromClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterViewSuffixProperties_return alterViewSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterViewSuffixProperties(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenameCol_return alterStatementSuffixRenameCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenameCol(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAndOperator_return precedenceAndOperator() throws RecognitionException { return gIdentifiersParser.precedenceAndOperator(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSelectorVal_return partitionSelectorVal() throws RecognitionException { return gIdentifiersParser.partitionSelectorVal(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpUnassign_return rpUnassign() throws RecognitionException { return gResourcePlanParser.rpUnassign(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrOperator_return precedenceBitwiseOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrOperator(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_frame_start_boundary_return window_frame_start_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_start_boundary(); }\n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixProperties_return alterDataConnectorSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixProperties(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_clause_return window_clause() throws RecognitionException { return gSelectClauseParser.window_clause(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixProperties_return alterDatabaseSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixProperties(); }\n \n-\tpublic HiveParser_IdentifiersParser.functionIdentifier_return functionIdentifier() throws RecognitionException { return gIdentifiersParser.functionIdentifier(); }\n+\tpublic HiveParser_IdentifiersParser.havingClause_return havingClause() throws RecognitionException { return gIdentifiersParser.havingClause(); }\n \n-\tpublic HiveParser_ResourcePlanParser.poolPath_return poolPath() throws RecognitionException { return gResourcePlanParser.poolPath(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSelectorSpec_return partitionSelectorSpec() throws RecognitionException { return gIdentifiersParser.partitionSelectorSpec(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetOwner_return alterStatementSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetOwner(); }\n+\tpublic HiveParser_IdentifiersParser.expressionOrDefault_return expressionOrDefault() throws RecognitionException { return gIdentifiersParser.expressionOrDefault(); }\n \n-\tpublic HiveParser_FromClauseParser.expressionList_return expressionList() throws RecognitionException { return gFromClauseParser.expressionList(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceOrExpression_return precedenceOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceOrExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.createTriggerStatement_return createTriggerStatement() throws RecognitionException { return gResourcePlanParser.createTriggerStatement(); }\n+\tpublic HiveParser_SelectClauseParser.selectTrfmClause_return selectTrfmClause() throws RecognitionException { return gSelectClauseParser.selectTrfmClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.stringLiteralSequence_return stringLiteralSequence() throws RecognitionException { return gIdentifiersParser.stringLiteralSequence(); }\n+\tpublic HiveParser_IdentifiersParser.dateLiteral_return dateLiteral() throws RecognitionException { return gIdentifiersParser.dateLiteral(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_frame_boundary_return window_frame_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_boundary(); }\n+\tpublic HiveParser_FromClauseParser.atomjoinSource_return atomjoinSource() throws RecognitionException { return gFromClauseParser.atomjoinSource(); }\n \n-\tpublic HiveParser_IdentifiersParser.extractExpression_return extractExpression() throws RecognitionException { return gIdentifiersParser.extractExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddConstraint_return alterStatementSuffixAddConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddConstraint(); }\n \n-\tpublic HiveParser_IdentifiersParser.timestampLiteral_return timestampLiteral() throws RecognitionException { return gIdentifiersParser.timestampLiteral(); }\n+\tpublic HiveParser_PrepareStatementParser.executeStatement_return executeStatement() throws RecognitionException { return gPrepareStatementParser.executeStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.columnRefOrderNotInParenthesis_return columnRefOrderNotInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderNotInParenthesis(); }\n+\tpublic HiveParser_ResourcePlanParser.createPoolStatement_return createPoolStatement() throws RecognitionException { return gResourcePlanParser.createPoolStatement(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateColumns_return alterStatementSuffixUpdateColumns() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateColumns(); }\n+\tpublic HiveParser_PrepareStatementParser.executeParamList_return executeParamList() throws RecognitionException { return gPrepareStatementParser.executeParamList(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionPart_return expressionPart(CommonTree firstExprTree, boolean isStruct) throws RecognitionException { return gIdentifiersParser.expressionPart(firstExprTree, isStruct); }\n+\tpublic HiveParser_AlterClauseParser.fileFormat_return fileFormat() throws RecognitionException { return gAlterClauseParser.fileFormat(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddCol_return alterStatementSuffixAddCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddCol(); }\n+\tpublic HiveParser_IdentifiersParser.identifier_return identifier() throws RecognitionException { return gIdentifiersParser.identifier(); }\n \n-\tpublic HiveParser_FromClauseParser.atomjoinSource_return atomjoinSource() throws RecognitionException { return gFromClauseParser.atomjoinSource(); }\n+\tpublic HiveParser_FromClauseParser.viewName_return viewName() throws RecognitionException { return gFromClauseParser.viewName(); }\n \n-\tpublic HiveParser_AlterClauseParser.compactPool_return compactPool() throws RecognitionException { return gAlterClauseParser.compactPool(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceRegexpOperator_return precedenceRegexpOperator() throws RecognitionException { return gIdentifiersParser.precedenceRegexpOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionOrDefault_return expressionOrDefault() throws RecognitionException { return gIdentifiersParser.expressionOrDefault(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceEqualOperator_return precedenceEqualOperator() throws RecognitionException { return gIdentifiersParser.precedenceEqualOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.columnRefOrderInParenthesis_return columnRefOrderInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderInParenthesis(); }\n+\tpublic HiveParser_FromClauseParser.tableSource_return tableSource() throws RecognitionException { return gFromClauseParser.tableSource(); }\n \n-\tpublic HiveParser_IdentifiersParser.expression_return expression() throws RecognitionException { return gIdentifiersParser.expression(); }\n+\tpublic HiveParser_IdentifiersParser.floorDateQualifiers_return floorDateQualifiers() throws RecognitionException { return gIdentifiersParser.floorDateQualifiers(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetLocation_return alterDatabaseSuffixSetLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetLocation(); }\n+\tpublic HiveParser_FromClauseParser.tableOrColumn_return tableOrColumn() throws RecognitionException { return gFromClauseParser.tableOrColumn(); }\n \n-\tpublic HiveParser_IdentifiersParser.havingCondition_return havingCondition() throws RecognitionException { return gIdentifiersParser.havingCondition(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenamePart_return alterStatementSuffixRenamePart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenamePart(); }\n \n-\tpublic HiveParser_ResourcePlanParser.comparisionOperator_return comparisionOperator() throws RecognitionException { return gResourcePlanParser.comparisionOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropConstraint_return alterStatementSuffixDropConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropConstraint(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionWithAlias_return expressionWithAlias() throws RecognitionException { return gIdentifiersParser.expressionWithAlias(); }\n+\tpublic HiveParser_FromClauseParser.valueRowConstructor_return valueRowConstructor() throws RecognitionException { return gFromClauseParser.valueRowConstructor(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropResourcePlanStatement_return dropResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.dropResourcePlanStatement(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatement_return alterStatement() throws RecognitionException { return gAlterClauseParser.alterStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceRegexpOperator_return precedenceRegexpOperator() throws RecognitionException { return gIdentifiersParser.precedenceRegexpOperator(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorExpression_return precedenceBitwiseXorExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAmpersandExpression_return precedenceAmpersandExpression() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandExpression(); }\n+\tpublic HiveParser_FromClauseParser.fromSource_return fromSource() throws RecognitionException { return gFromClauseParser.fromSource(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixTouch_return alterStatementSuffixTouch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixTouch(); }\n+\tpublic HiveParser_AlterClauseParser.skewedLocations_return skewedLocations() throws RecognitionException { return gAlterClauseParser.skewedLocations(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterMaterializedViewStatementSuffix_return alterMaterializedViewStatementSuffix(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewStatementSuffix(tableNameTree); }\n+\tpublic HiveParser_IdentifiersParser.floorExpression_return floorExpression() throws RecognitionException { return gIdentifiersParser.floorExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.sysFuncNames_return sysFuncNames() throws RecognitionException { return gIdentifiersParser.sysFuncNames(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerActionExpression_return triggerActionExpression() throws RecognitionException { return gResourcePlanParser.triggerActionExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.charSetStringLiteral_return charSetStringLiteral() throws RecognitionException { return gIdentifiersParser.charSetStringLiteral(); }\n+\tpublic HiveParser_SelectClauseParser.window_defn_return window_defn() throws RecognitionException { return gSelectClauseParser.window_defn(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixProperties_return alterDataConnectorSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixProperties(); }\n+\tpublic HiveParser_IdentifiersParser.partitionByClause_return partitionByClause() throws RecognitionException { return gIdentifiersParser.partitionByClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetOwner_return alterDatabaseSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetOwner(); }\n+\tpublic HiveParser_IdentifiersParser.havingCondition_return havingCondition() throws RecognitionException { return gIdentifiersParser.havingCondition(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementChangeColPosition_return alterStatementChangeColPosition() throws RecognitionException { return gAlterClauseParser.alterStatementChangeColPosition(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceDistinctOperator_return precedenceDistinctOperator() throws RecognitionException { return gIdentifiersParser.precedenceDistinctOperator(); }\n \n-\tpublic HiveParser_FromClauseParser.tableBucketSample_return tableBucketSample() throws RecognitionException { return gFromClauseParser.tableBucketSample(); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinExpr_return uniqueJoinExpr() throws RecognitionException { return gFromClauseParser.uniqueJoinExpr(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupByClause_return groupByClause() throws RecognitionException { return gIdentifiersParser.groupByClause(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetLocation_return alterDatabaseSuffixSetLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetLocation(); }\n \n-\tpublic HiveParser_IdentifiersParser.principalIdentifier_return principalIdentifier() throws RecognitionException { return gIdentifiersParser.principalIdentifier(); }\n+\tpublic HiveParser_IdentifiersParser.clusterByClause_return clusterByClause() throws RecognitionException { return gIdentifiersParser.clusterByClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.havingClause_return havingClause() throws RecognitionException { return gIdentifiersParser.havingClause(); }\n+\tpublic HiveParser_IdentifiersParser.expressionsNotInParenthesis_return expressionsNotInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsNotInParenthesis(isStruct, forceStruct); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceFieldExpression_return precedenceFieldExpression() throws RecognitionException { return gIdentifiersParser.precedenceFieldExpression(); }\n+\tpublic HiveParser_FromClauseParser.lateralView_return lateralView() throws RecognitionException { return gFromClauseParser.lateralView(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementPartitionKeyType_return alterStatementPartitionKeyType() throws RecognitionException { return gAlterClauseParser.alterStatementPartitionKeyType(); }\n+\tpublic HiveParser_IdentifiersParser.functionName_return functionName() throws RecognitionException { return gIdentifiersParser.functionName(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRename_return alterStatementSuffixRename(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRename(table); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixArchive_return alterStatementSuffixArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixArchive(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixProperties_return alterDatabaseSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixProperties(); }\n+\tpublic HiveParser_IdentifiersParser.qualifyClause_return qualifyClause() throws RecognitionException { return gIdentifiersParser.qualifyClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.caseExpression_return caseExpression() throws RecognitionException { return gIdentifiersParser.caseExpression(); }\n+\tpublic HiveParser_IdentifiersParser.quantifierType_return quantifierType() throws RecognitionException { return gIdentifiersParser.quantifierType(); }\n \n-\tpublic HiveParser_AlterClauseParser.skewedLocationMap_return skewedLocationMap() throws RecognitionException { return gAlterClauseParser.skewedLocationMap(); }\n+\tpublic HiveParser_IdentifiersParser.intervalLiteral_return intervalLiteral() throws RecognitionException { return gIdentifiersParser.intervalLiteral(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStats_return alterStatementSuffixUpdateStats(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStats(partition); }\n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetUrl_return alterDataConnectorSuffixSetUrl() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetUrl(); }\n \n \tpublic HiveParser_IdentifiersParser.descFuncNames_return descFuncNames() throws RecognitionException { return gIdentifiersParser.descFuncNames(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerOrExpression_return triggerOrExpression() throws RecognitionException { return gResourcePlanParser.triggerOrExpression(); }\n-\n-\tpublic HiveParser_IdentifiersParser.orderByClause_return orderByClause() throws RecognitionException { return gIdentifiersParser.orderByClause(); }\n+\tpublic HiveParser_PrepareStatementParser.prepareStatement_return prepareStatement() throws RecognitionException { return gPrepareStatementParser.prepareStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinExpr_return uniqueJoinExpr() throws RecognitionException { return gFromClauseParser.uniqueJoinExpr(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAndExpression_return precedenceAndExpression() throws RecognitionException { return gIdentifiersParser.precedenceAndExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarOperator_return precedenceSimilarOperator() throws RecognitionException { return gIdentifiersParser.precedenceSimilarOperator(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerExpression_return triggerExpression() throws RecognitionException { return gResourcePlanParser.triggerExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.whenExpression_return whenExpression() throws RecognitionException { return gIdentifiersParser.whenExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceUnaryOperator_return precedenceUnaryOperator() throws RecognitionException { return gIdentifiersParser.precedenceUnaryOperator(); }\n \n-\tpublic HiveParser_FromClauseParser.aliasList_return aliasList() throws RecognitionException { return gFromClauseParser.aliasList(); }\n+\tpublic HiveParser_ResourcePlanParser.alterPoolStatement_return alterPoolStatement() throws RecognitionException { return gResourcePlanParser.alterPoolStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.tableOrPartition_return tableOrPartition() throws RecognitionException { return gIdentifiersParser.tableOrPartition(); }\n+\tpublic HiveParser_ResourcePlanParser.resourcePlanDdlStatements_return resourcePlanDdlStatements() throws RecognitionException { return gResourcePlanParser.resourcePlanDdlStatements(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatement_return alterStatement() throws RecognitionException { return gAlterClauseParser.alterStatement(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetManagedLocation_return alterDatabaseSuffixSetManagedLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetManagedLocation(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_defn_return window_defn() throws RecognitionException { return gSelectClauseParser.window_defn(); }\n+\tpublic HiveParser_IdentifiersParser.charSetStringLiteral_return charSetStringLiteral() throws RecognitionException { return gIdentifiersParser.charSetStringLiteral(); }\n \n-\tpublic HiveParser_IdentifiersParser.atomExpression_return atomExpression() throws RecognitionException { return gIdentifiersParser.atomExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceUnarySuffixExpression_return precedenceUnarySuffixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnarySuffixExpression(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectExpression_return selectExpression() throws RecognitionException { return gSelectClauseParser.selectExpression(); }\n+\tpublic HiveParser_IdentifiersParser.groupby_expression_return groupby_expression() throws RecognitionException { return gIdentifiersParser.groupby_expression(); }\n \n-\tpublic HiveParser_FromClauseParser.partitionedTableFunction_return partitionedTableFunction() throws RecognitionException { return gFromClauseParser.partitionedTableFunction(); }\n+\tpublic HiveParser_IdentifiersParser.stringLiteralSequence_return stringLiteralSequence() throws RecognitionException { return gIdentifiersParser.stringLiteralSequence(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerLiteral_return triggerLiteral() throws RecognitionException { return gResourcePlanParser.triggerLiteral(); }\n+\tpublic HiveParser_IdentifiersParser.expressionsInParenthesis_return expressionsInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsInParenthesis(isStruct, forceStruct); }\n \n-\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRebuild_return alterMaterializedViewSuffixRebuild(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRebuild(tableNameTree); }\n+\tpublic HiveParser_FromClauseParser.partitionedTableFunction_return partitionedTableFunction() throws RecognitionException { return gFromClauseParser.partitionedTableFunction(); }\n \n-\tpublic HiveParser_FromClauseParser.tableSample_return tableSample() throws RecognitionException { return gFromClauseParser.tableSample(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixTouch_return alterStatementSuffixTouch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixTouch(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionMain_return precedenceSimilarExpressionMain() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionMain(); }\n+\tpublic HiveParser_IdentifiersParser.intervalQualifiers_return intervalQualifiers() throws RecognitionException { return gIdentifiersParser.intervalQualifiers(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixArchive_return alterStatementSuffixArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixArchive(); }\n+\tpublic HiveParser_IdentifiersParser.expressionPart_return expressionPart(CommonTree firstExprTree, boolean isStruct) throws RecognitionException { return gIdentifiersParser.expressionPart(firstExprTree, isStruct); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceStarOperator_return precedenceStarOperator() throws RecognitionException { return gIdentifiersParser.precedenceStarOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRewrite_return alterMaterializedViewSuffixRewrite(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRewrite(tableNameTree); }\n \n-\tpublic HiveParser_AlterClauseParser.partitionLocation_return partitionLocation() throws RecognitionException { return gAlterClauseParser.partitionLocation(); }\n+\tpublic HiveParser_AlterClauseParser.skewedLocationMap_return skewedLocationMap() throws RecognitionException { return gAlterClauseParser.skewedLocationMap(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerActionExpressionStandalone_return triggerActionExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerActionExpressionStandalone(); }\n+\tpublic HiveParser_SelectClauseParser.selectExpressionList_return selectExpressionList() throws RecognitionException { return gSelectClauseParser.selectExpressionList(); }\n \n-\tpublic HiveParser_FromClauseParser.fromSource_return fromSource() throws RecognitionException { return gFromClauseParser.fromSource(); }\n+\tpublic HiveParser_AlterClauseParser.alterTableStatementSuffix_return alterTableStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterTableStatementSuffix(); }\n \n-\tpublic HiveParser_PrepareStatementParser.executeParamList_return executeParamList() throws RecognitionException { return gPrepareStatementParser.executeParamList(); }\n+\tpublic HiveParser_SelectClauseParser.window_range_expression_return window_range_expression() throws RecognitionException { return gSelectClauseParser.window_range_expression(); }\n \n-\tpublic HiveParser_FromClauseParser.tableName_return tableName() throws RecognitionException { return gFromClauseParser.tableName(); }\n+\tpublic HiveParser_IdentifiersParser.caseExpression_return caseExpression() throws RecognitionException { return gIdentifiersParser.caseExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSpec_return partitionSpec() throws RecognitionException { return gIdentifiersParser.partitionSpec(); }\n+\tpublic HiveParser_FromClauseParser.tableAlias_return tableAlias() throws RecognitionException { return gFromClauseParser.tableAlias(); }\n \n-\tpublic HiveParser_ResourcePlanParser.replaceResourcePlanStatement_return replaceResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.replaceResourcePlanStatement(); }\n+\tpublic HiveParser_ResourcePlanParser.dropResourcePlanStatement_return dropResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.dropResourcePlanStatement(); }\n \n-\tpublic HiveParser_CreateDDLParser.dcProperties_return dcProperties() throws RecognitionException { return gCreateDDLParser.dcProperties(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceConcatenateOperator_return precedenceConcatenateOperator() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateOperator(); }\n \n-\tpublic HiveParser_ResourcePlanParser.alterPoolStatement_return alterPoolStatement() throws RecognitionException { return gResourcePlanParser.alterPoolStatement(); }\n+\tpublic HiveParser_FromClauseParser.partitionTableFunctionSource_return partitionTableFunctionSource() throws RecognitionException { return gFromClauseParser.partitionTableFunctionSource(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddConstraint_return alterStatementSuffixAddConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddConstraint(); }\n+\tpublic HiveParser_FromClauseParser.firstValueRowConstructor_return firstValueRowConstructor() throws RecognitionException { return gFromClauseParser.firstValueRowConstructor(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalLiteral_return intervalLiteral() throws RecognitionException { return gIdentifiersParser.intervalLiteral(); }\n+\tpublic HiveParser_IdentifiersParser.distributeByClause_return distributeByClause() throws RecognitionException { return gIdentifiersParser.distributeByClause(); }\n \n-\tpublic HiveParser_FromClauseParser.joinSourcePart_return joinSourcePart() throws RecognitionException { return gFromClauseParser.joinSourcePart(); }\n+\tpublic HiveParser_FromClauseParser.valuesSource_return valuesSource() throws RecognitionException { return gFromClauseParser.valuesSource(); }\n \n-\tpublic HiveParser_ResourcePlanParser.resourcePlanDdlStatements_return resourcePlanDdlStatements() throws RecognitionException { return gResourcePlanParser.resourcePlanDdlStatements(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseStatementSuffix_return alterDatabaseStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDatabaseStatementSuffix(); }\n \n-\tpublic HiveParser_FromClauseParser.virtualTableSource_return virtualTableSource() throws RecognitionException { return gFromClauseParser.virtualTableSource(); }\n+\tpublic HiveParser_FromClauseParser.partitioningSpec_return partitioningSpec() throws RecognitionException { return gFromClauseParser.partitioningSpec(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUnArchive_return alterStatementSuffixUnArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUnArchive(); }\n+\tpublic HiveParser_IdentifiersParser.timestampLiteral_return timestampLiteral() throws RecognitionException { return gIdentifiersParser.timestampLiteral(); }\n \n-\tpublic HiveParser_IdentifiersParser.null_treatment_return null_treatment() throws RecognitionException { return gIdentifiersParser.null_treatment(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorOperator_return precedenceBitwiseXorOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorOperator(); }\n \n-\tpublic HiveParser_ResourcePlanParser.withReplace_return withReplace() throws RecognitionException { return gResourcePlanParser.withReplace(); }\n+\tpublic HiveParser_CreateDDLParser.dcProperties_return dcProperties() throws RecognitionException { return gCreateDDLParser.dcProperties(); }\n \n-\tpublic HiveParser_FromClauseParser.valuesSource_return valuesSource() throws RecognitionException { return gFromClauseParser.valuesSource(); }\n+\tpublic HiveParser_IdentifiersParser.extractExpression_return extractExpression() throws RecognitionException { return gIdentifiersParser.extractExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerActionExpression_return triggerActionExpression() throws RecognitionException { return gResourcePlanParser.triggerActionExpression(); }\n+\tpublic HiveParser_ResourcePlanParser.poolAssignList_return poolAssignList() throws RecognitionException { return gResourcePlanParser.poolAssignList(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixMergeFiles_return alterStatementSuffixMergeFiles(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixMergeFiles(partition); }\n+\tpublic HiveParser_ResourcePlanParser.triggerActionExpressionStandalone_return triggerActionExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerActionExpressionStandalone(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerAtomExpression_return triggerAtomExpression() throws RecognitionException { return gResourcePlanParser.triggerAtomExpression(); }\n+\tpublic HiveParser_ResourcePlanParser.dropPoolStatement_return dropPoolStatement() throws RecognitionException { return gResourcePlanParser.dropPoolStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressions_return expressions() throws RecognitionException { return gIdentifiersParser.expressions(); }\n+\tpublic HiveParser_IdentifiersParser.subQueryExpression_return subQueryExpression() throws RecognitionException { return gIdentifiersParser.subQueryExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStatsCol_return alterStatementSuffixUpdateStatsCol(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStatsCol(partition); }\n+\tpublic HiveParser_ResourcePlanParser.createMappingStatement_return createMappingStatement() throws RecognitionException { return gResourcePlanParser.createMappingStatement(); }\n \n \tpublic HiveParser_ResourcePlanParser.alterTriggerStatement_return alterTriggerStatement() throws RecognitionException { return gResourcePlanParser.alterTriggerStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionByClause_return partitionByClause() throws RecognitionException { return gIdentifiersParser.partitionByClause(); }\n+\tpublic HiveParser_IdentifiersParser.precedencePlusOperator_return precedencePlusOperator() throws RecognitionException { return gIdentifiersParser.precedencePlusOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixLocation_return alterStatementSuffixLocation(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixLocation(partition); }\n+\tpublic HiveParser_ResourcePlanParser.rpUnassignList_return rpUnassignList() throws RecognitionException { return gResourcePlanParser.rpUnassignList(); }\n \n-\tpublic HiveParser_AlterClauseParser.blocking_return blocking() throws RecognitionException { return gAlterClauseParser.blocking(); }\n+\tpublic HiveParser_IdentifiersParser.prepareStmtParam_return prepareStmtParam() throws RecognitionException { return gIdentifiersParser.prepareStmtParam(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetUrl_return alterDataConnectorSuffixSetUrl() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetUrl(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceEqualExpression_return precedenceEqualExpression() throws RecognitionException { return gIdentifiersParser.precedenceEqualExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupByEmpty_return groupByEmpty() throws RecognitionException { return gIdentifiersParser.groupByEmpty(); }\n+\tpublic HiveParser_FromClauseParser.tableName_return tableName() throws RecognitionException { return gFromClauseParser.tableName(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalExpression_return intervalExpression() throws RecognitionException { return gIdentifiersParser.intervalExpression(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSelectorOperator_return partitionSelectorOperator() throws RecognitionException { return gIdentifiersParser.partitionSelectorOperator(); }\n \n-\tpublic HiveParser_FromClauseParser.firstValueRowConstructor_return firstValueRowConstructor() throws RecognitionException { return gFromClauseParser.firstValueRowConstructor(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRename_return alterStatementSuffixRename(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRename(table); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPart_return precedenceSimilarExpressionPart(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPart(t); }\n+\tpublic HiveParser_SelectClauseParser.selectItem_return selectItem() throws RecognitionException { return gSelectClauseParser.selectItem(); }\n \n-\tpublic HiveParser_IdentifiersParser.castExpression_return castExpression() throws RecognitionException { return gIdentifiersParser.castExpression(); }\n+\tpublic HiveParser_IdentifiersParser.timestampLocalTZLiteral_return timestampLocalTZLiteral() throws RecognitionException { return gIdentifiersParser.timestampLocalTZLiteral(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropTriggerStatement_return dropTriggerStatement() throws RecognitionException { return gResourcePlanParser.dropTriggerStatement(); }\n+\tpublic HiveParser_IdentifiersParser.atomExpression_return atomExpression() throws RecognitionException { return gIdentifiersParser.atomExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.qualifyClause_return qualifyClause() throws RecognitionException { return gIdentifiersParser.qualifyClause(); }\n+\tpublic HiveParser_IdentifiersParser.trimFunction_return trimFunction() throws RecognitionException { return gIdentifiersParser.trimFunction(); }\n \n-\tpublic HiveParser_IdentifiersParser.function_return function() throws RecognitionException { return gIdentifiersParser.function(); }\n+\tpublic HiveParser_IdentifiersParser.columnRefOrderNotInParenthesis_return columnRefOrderNotInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderNotInParenthesis(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_range_expression_return window_range_expression() throws RecognitionException { return gSelectClauseParser.window_range_expression(); }\n+\tpublic HiveParser_ResourcePlanParser.unmanaged_return unmanaged() throws RecognitionException { return gResourcePlanParser.unmanaged(); }\n \n-\tpublic HiveParser_FromClauseParser.subQuerySource_return subQuerySource() throws RecognitionException { return gFromClauseParser.subQuerySource(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceFieldExpression_return precedenceFieldExpression() throws RecognitionException { return gIdentifiersParser.precedenceFieldExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceConcatenateOperator_return precedenceConcatenateOperator() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateOperator(); }\n+\tpublic HiveParser_IdentifiersParser.sortByClause_return sortByClause() throws RecognitionException { return gIdentifiersParser.sortByClause(); }\n \n-\tpublic HiveParser_FromClauseParser.fromClause_return fromClause() throws RecognitionException { return gFromClauseParser.fromClause(); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinToken_return uniqueJoinToken() throws RecognitionException { return gFromClauseParser.uniqueJoinToken(); }\n \n-\tpublic HiveParser_ResourcePlanParser.unmanaged_return unmanaged() throws RecognitionException { return gResourcePlanParser.unmanaged(); }\n+\tpublic HiveParser_ResourcePlanParser.rpAssignList_return rpAssignList() throws RecognitionException { return gResourcePlanParser.rpAssignList(); }\n \n-\tpublic HiveParser_IdentifiersParser.floorDateQualifiers_return floorDateQualifiers() throws RecognitionException { return gIdentifiersParser.floorDateQualifiers(); }\n+\tpublic HiveParser_FromClauseParser.searchCondition_return searchCondition() throws RecognitionException { return gFromClauseParser.searchCondition(); }\n \n-\tpublic HiveParser_ResourcePlanParser.poolAssignList_return poolAssignList() throws RecognitionException { return gResourcePlanParser.poolAssignList(); }\n+\tpublic HiveParser_ResourcePlanParser.createResourcePlanStatement_return createResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.createResourcePlanStatement(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixFileFormat_return alterStatementSuffixFileFormat(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixFileFormat(partition); }\n+\tpublic HiveParser_IdentifiersParser.isCondition_return isCondition() throws RecognitionException { return gIdentifiersParser.isCondition(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinSource_return uniqueJoinSource() throws RecognitionException { return gFromClauseParser.uniqueJoinSource(); }\n+\tpublic HiveParser_ResourcePlanParser.dropMappingStatement_return dropMappingStatement() throws RecognitionException { return gResourcePlanParser.dropMappingStatement(); }\n \n-\tpublic HiveParser_ResourcePlanParser.alterMappingStatement_return alterMappingStatement() throws RecognitionException { return gResourcePlanParser.alterMappingStatement(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceNotExpression_return precedenceNotExpression() throws RecognitionException { return gIdentifiersParser.precedenceNotExpression(); }\n \n-\tpublic HiveParser_FromClauseParser.tableOrColumn_return tableOrColumn() throws RecognitionException { return gFromClauseParser.tableOrColumn(); }\n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorStatementSuffix_return alterDataConnectorStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDataConnectorStatementSuffix(); }\n \n-\tpublic HiveParser_FromClauseParser.tableAlias_return tableAlias() throws RecognitionException { return gFromClauseParser.tableAlias(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixMergeFiles_return alterStatementSuffixMergeFiles(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixMergeFiles(partition); }\n \n-\tpublic HiveParser_IdentifiersParser.distributeByClause_return distributeByClause() throws RecognitionException { return gIdentifiersParser.distributeByClause(); }\n+\tpublic HiveParser_IdentifiersParser.groupingSetExpressionMultiple_return groupingSetExpressionMultiple() throws RecognitionException { return gIdentifiersParser.groupingSetExpressionMultiple(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRewrite_return alterMaterializedViewSuffixRewrite(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRewrite(tableNameTree); }\n+\tpublic HiveParser_FromClauseParser.joinToken_return joinToken() throws RecognitionException { return gFromClauseParser.joinToken(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionIn_return precedenceSimilarExpressionIn(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionIn(t); }\n+\tpublic HiveParser_SelectClauseParser.window_frame_boundary_return window_frame_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_boundary(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetPartSpec_return alterStatementSuffixSetPartSpec() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetPartSpec(); }\n+\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffix_return alterTblPartitionStatementSuffix(boolean partition) throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffix(partition); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceUnaryOperator_return precedenceUnaryOperator() throws RecognitionException { return gIdentifiersParser.precedenceUnaryOperator(); }\n+\tpublic HiveParser_ResourcePlanParser.createTriggerStatement_return createTriggerStatement() throws RecognitionException { return gResourcePlanParser.createTriggerStatement(); }\n \n-\tpublic HiveParser_ResourcePlanParser.createPoolStatement_return createPoolStatement() throws RecognitionException { return gResourcePlanParser.createPoolStatement(); }\n+\tpublic HiveParser_ResourcePlanParser.alterMappingStatement_return alterMappingStatement() throws RecognitionException { return gResourcePlanParser.alterMappingStatement(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectClause_return selectClause() throws RecognitionException { return gSelectClauseParser.selectClause(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitionsElement_return alterStatementSuffixAddPartitionsElement() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitionsElement(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionQuantifierPredicate_return precedenceSimilarExpressionQuantifierPredicate(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionQuantifierPredicate(t); }\n+\tpublic HiveParser_ResourcePlanParser.globalWmStatement_return globalWmStatement() throws RecognitionException { return gResourcePlanParser.globalWmStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpression_return precedenceSimilarExpression() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpression(); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinSource_return uniqueJoinSource() throws RecognitionException { return gFromClauseParser.uniqueJoinSource(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinToken_return uniqueJoinToken() throws RecognitionException { return gFromClauseParser.uniqueJoinToken(); }\n+\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRebuild_return alterMaterializedViewSuffixRebuild(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRebuild(tableNameTree); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixExecute_return alterStatementSuffixExecute() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExecute(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixLocation_return alterStatementSuffixLocation(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixLocation(partition); }\n \n-\tpublic HiveParser_SelectClauseParser.selectExpressionList_return selectExpressionList() throws RecognitionException { return gSelectClauseParser.selectExpressionList(); }\n+\tpublic HiveParser_SelectClauseParser.trfmClause_return trfmClause() throws RecognitionException { return gSelectClauseParser.trfmClause(); }\n \n-\tpublic HiveParser_ResourcePlanParser.createMappingStatement_return createMappingStatement() throws RecognitionException { return gResourcePlanParser.createMappingStatement(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementChangeColPosition_return alterStatementChangeColPosition() throws RecognitionException { return gAlterClauseParser.alterStatementChangeColPosition(); }\n \n-\tpublic HiveParser_ResourcePlanParser.createResourcePlanStatement_return createResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.createResourcePlanStatement(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSpec_return partitionSpec() throws RecognitionException { return gIdentifiersParser.partitionSpec(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectList_return selectList() throws RecognitionException { return gSelectClauseParser.selectList(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixExecute_return alterStatementSuffixExecute() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExecute(); }\n \n-\tpublic HiveParser_IdentifiersParser.constant_return constant() throws RecognitionException { return gIdentifiersParser.constant(); }\n+\tpublic HiveParser_FromClauseParser.joinSource_return joinSource() throws RecognitionException { return gFromClauseParser.joinSource(); }\n \n-\tpublic HiveParser_FromClauseParser.defaultValue_return defaultValue() throws RecognitionException { return gFromClauseParser.defaultValue(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerOrExpression_return triggerOrExpression() throws RecognitionException { return gResourcePlanParser.triggerOrExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerExpressionStandalone_return triggerExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerExpressionStandalone(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetPartSpec_return alterStatementSuffixSetPartSpec() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetPartSpec(); }\n \n-\tpublic HiveParser_PrepareStatementParser.executeStatement_return executeStatement() throws RecognitionException { return gPrepareStatementParser.executeStatement(); }\n+\tpublic HiveParser_FromClauseParser.valuesTableConstructor_return valuesTableConstructor() throws RecognitionException { return gFromClauseParser.valuesTableConstructor(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSelectorVal_return partitionSelectorVal() throws RecognitionException { return gIdentifiersParser.partitionSelectorVal(); }\n+\tpublic HiveParser_ResourcePlanParser.poolAssign_return poolAssign() throws RecognitionException { return gResourcePlanParser.poolAssign(); }\n \n-\tpublic HiveParser_CreateDDLParser.dataConnectorType_return dataConnectorType() throws RecognitionException { return gCreateDDLParser.dataConnectorType(); }\n+\tpublic HiveParser_ResourcePlanParser.replaceResourcePlanStatement_return replaceResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.replaceResourcePlanStatement(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffix_return alterTblPartitionStatementSuffix(boolean partition) throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffix(partition); }\n+\tpublic HiveParser_IdentifiersParser.groupingExpressionSingle_return groupingExpressionSingle() throws RecognitionException { return gIdentifiersParser.groupingExpressionSingle(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixClusterbySortby_return alterStatementSuffixClusterbySortby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixClusterbySortby(); }\n+\tpublic HiveParser_FromClauseParser.aliasList_return aliasList() throws RecognitionException { return gFromClauseParser.aliasList(); }\n \n-\tpublic HiveParser_AlterClauseParser.skewedLocations_return skewedLocations() throws RecognitionException { return gAlterClauseParser.skewedLocations(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSkewedby_return alterStatementSuffixSkewedby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSkewedby(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionVal_return partitionVal() throws RecognitionException { return gIdentifiersParser.partitionVal(); }\n+\tpublic HiveParser_IdentifiersParser.columnRefOrderInParenthesis_return columnRefOrderInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderInParenthesis(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedencePlusExpression_return precedencePlusExpression() throws RecognitionException { return gIdentifiersParser.precedencePlusExpression(); }\n+\tpublic HiveParser_FromClauseParser.tableAllColumns_return tableAllColumns() throws RecognitionException { return gFromClauseParser.tableAllColumns(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPartNot_return precedenceSimilarExpressionPartNot(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPartNot(t); }\n+\tpublic HiveParser_FromClauseParser.valuesClause_return valuesClause() throws RecognitionException { return gFromClauseParser.valuesClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.fileFormat_return fileFormat() throws RecognitionException { return gAlterClauseParser.fileFormat(); }\n+\tpublic HiveParser_FromClauseParser.subQuerySource_return subQuerySource() throws RecognitionException { return gFromClauseParser.subQuerySource(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceNotExpression_return precedenceNotExpression() throws RecognitionException { return gIdentifiersParser.precedenceNotExpression(); }\n+\tpublic HiveParser_IdentifiersParser.groupingSetExpression_return groupingSetExpression() throws RecognitionException { return gIdentifiersParser.groupingSetExpression(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_frame_return window_frame() throws RecognitionException { return gSelectClauseParser.window_frame(); }\n+\tpublic HiveParser_IdentifiersParser.partitionVal_return partitionVal() throws RecognitionException { return gIdentifiersParser.partitionVal(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerExpression_return triggerExpression() throws RecognitionException { return gResourcePlanParser.triggerExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrExpression_return precedenceBitwiseOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropPoolStatement_return dropPoolStatement() throws RecognitionException { return gResourcePlanParser.dropPoolStatement(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionQuantifierPredicate_return precedenceSimilarExpressionQuantifierPredicate(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionQuantifierPredicate(t); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSelectorOperator_return partitionSelectorOperator() throws RecognitionException { return gIdentifiersParser.partitionSelectorOperator(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerAndExpression_return triggerAndExpression() throws RecognitionException { return gResourcePlanParser.triggerAndExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.poolAssign_return poolAssign() throws RecognitionException { return gResourcePlanParser.poolAssign(); }\n+\tpublic HiveParser_AlterClauseParser.alterViewSuffixProperties_return alterViewSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterViewSuffixProperties(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenamePart_return alterStatementSuffixRenamePart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenamePart(); }\n+\tpublic HiveParser_IdentifiersParser.rollupStandard_return rollupStandard() throws RecognitionException { return gIdentifiersParser.rollupStandard(); }\n \n-\tpublic HiveParser_FromClauseParser.valuesClause_return valuesClause() throws RecognitionException { return gFromClauseParser.valuesClause(); }\n+\tpublic HiveParser_IdentifiersParser.castExpression_return castExpression() throws RecognitionException { return gIdentifiersParser.castExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixStatsPart_return alterStatementSuffixStatsPart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixStatsPart(); }\n+\tpublic HiveParser_SelectClauseParser.window_value_expression_return window_value_expression() throws RecognitionException { return gSelectClauseParser.window_value_expression(); }\n \n-\tpublic HiveParser_IdentifiersParser.trimFunction_return trimFunction() throws RecognitionException { return gIdentifiersParser.trimFunction(); }\n+\tpublic HiveParser_FromClauseParser.expressionList_return expressionList() throws RecognitionException { return gFromClauseParser.expressionList(); }\n \n-\tpublic HiveParser_ResourcePlanParser.globalWmStatement_return globalWmStatement() throws RecognitionException { return gResourcePlanParser.globalWmStatement(); }\n+\tpublic HiveParser_SelectClauseParser.selectList_return selectList() throws RecognitionException { return gSelectClauseParser.selectList(); }\n \n-\tpublic HiveParser_IdentifiersParser.prepareStmtParam_return prepareStmtParam() throws RecognitionException { return gIdentifiersParser.prepareStmtParam(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceUnaryPrefixExpression_return precedenceUnaryPrefixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnaryPrefixExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixExchangePartition_return alterStatementSuffixExchangePartition() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExchangePartition(); }\n+\tpublic HiveParser_IdentifiersParser.intervalValue_return intervalValue() throws RecognitionException { return gIdentifiersParser.intervalValue(); }\n \n-\tpublic HiveParser_IdentifiersParser.subQueryExpression_return subQueryExpression() throws RecognitionException { return gIdentifiersParser.subQueryExpression(); }\n+\tpublic HiveParser_AlterClauseParser.skewedLocationsList_return skewedLocationsList() throws RecognitionException { return gAlterClauseParser.skewedLocationsList(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetManagedLocation_return alterDatabaseSuffixSetManagedLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetManagedLocation(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitions_return alterStatementSuffixAddPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitions(table); }\n \n-\tpublic HiveParser_IdentifiersParser.sortByClause_return sortByClause() throws RecognitionException { return gIdentifiersParser.sortByClause(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerExpressionStandalone_return triggerExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerExpressionStandalone(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalQualifiers_return intervalQualifiers() throws RecognitionException { return gIdentifiersParser.intervalQualifiers(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAndOperator_return precedenceAndOperator() throws RecognitionException { return gIdentifiersParser.precedenceAndOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCompact_return alterStatementSuffixCompact() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCompact(); }\n+\tpublic HiveParser_ResourcePlanParser.dropTriggerStatement_return dropTriggerStatement() throws RecognitionException { return gResourcePlanParser.dropTriggerStatement(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpAssignList_return rpAssignList() throws RecognitionException { return gResourcePlanParser.rpAssignList(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPartNot_return precedenceSimilarExpressionPartNot(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPartNot(t); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalValue_return intervalValue() throws RecognitionException { return gIdentifiersParser.intervalValue(); }\n+\tpublic HiveParser_IdentifiersParser.orderByClause_return orderByClause() throws RecognitionException { return gIdentifiersParser.orderByClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenameCol_return alterStatementSuffixRenameCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenameCol(); }\n+\tpublic HiveParser_ResourcePlanParser.disable_return disable() throws RecognitionException { return gResourcePlanParser.disable(); }\n \n-\tpublic HiveParser_CreateDDLParser.dataConnectorComment_return dataConnectorComment() throws RecognitionException { return gCreateDDLParser.dataConnectorComment(); }\n+\tpublic HiveParser_IdentifiersParser.booleanValue_return booleanValue() throws RecognitionException { return gIdentifiersParser.booleanValue(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSkewedby_return alterStatementSuffixSkewedby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSkewedby(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixProperties_return alterStatementSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixProperties(); }\n \n-\tpublic HiveParser_FromClauseParser.tableSource_return tableSource() throws RecognitionException { return gFromClauseParser.tableSource(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpression_return precedenceSimilarExpression() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionsNotInParenthesis_return expressionsNotInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsNotInParenthesis(isStruct, forceStruct); }\n+\tpublic HiveParser_IdentifiersParser.precedenceOrOperator_return precedenceOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceOrOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupingExpressionSingle_return groupingExpressionSingle() throws RecognitionException { return gIdentifiersParser.groupingExpressionSingle(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStats_return alterStatementSuffixUpdateStats(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStats(partition); }\n \n-\tpublic HiveParser_IdentifiersParser.clusterByClause_return clusterByClause() throws RecognitionException { return gIdentifiersParser.clusterByClause(); }\n+\tpublic HiveParser_FromClauseParser.asOfClause_return asOfClause() throws RecognitionException { return gFromClauseParser.asOfClause(); }\n \n-\tpublic HiveParser_CreateDDLParser.dropDataConnectorStatement_return dropDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.dropDataConnectorStatement(); }\n+\tpublic HiveParser_IdentifiersParser.intervalExpression_return intervalExpression() throws RecognitionException { return gIdentifiersParser.intervalExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAndExpression_return precedenceAndExpression() throws RecognitionException { return gIdentifiersParser.precedenceAndExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedencePlusExpression_return precedencePlusExpression() throws RecognitionException { return gIdentifiersParser.precedencePlusExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupingSetExpression_return groupingSetExpression() throws RecognitionException { return gIdentifiersParser.groupingSetExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceNotOperator_return precedenceNotOperator() throws RecognitionException { return gIdentifiersParser.precedenceNotOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrExpression_return precedenceBitwiseOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrExpression(); }\n+\tpublic HiveParser_SelectClauseParser.selectClause_return selectClause() throws RecognitionException { return gSelectClauseParser.selectClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseStatementSuffix_return alterDatabaseStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDatabaseStatementSuffix(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSerdeProperties_return alterStatementSuffixSerdeProperties(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSerdeProperties(partition); }\n \n-\tpublic HiveParser_FromClauseParser.partitioningSpec_return partitioningSpec() throws RecognitionException { return gFromClauseParser.partitioningSpec(); }\n+\tpublic HiveParser_IdentifiersParser.null_treatment_return null_treatment() throws RecognitionException { return gIdentifiersParser.null_treatment(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectTrfmClause_return selectTrfmClause() throws RecognitionException { return gSelectClauseParser.selectTrfmClause(); }\n+\tpublic HiveParser_IdentifiersParser.tableOrPartition_return tableOrPartition() throws RecognitionException { return gIdentifiersParser.tableOrPartition(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceEqualExpression_return precedenceEqualExpression() throws RecognitionException { return gIdentifiersParser.precedenceEqualExpression(); }\n+\tpublic HiveParser_AlterClauseParser.partitionLocation_return partitionLocation() throws RecognitionException { return gAlterClauseParser.partitionLocation(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionsInParenthesis_return expressionsInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsInParenthesis(isStruct, forceStruct); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionIn_return precedenceSimilarExpressionIn(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionIn(t); }\n \n-\tpublic HiveParser_IdentifiersParser.timeQualifiers_return timeQualifiers() throws RecognitionException { return gIdentifiersParser.timeQualifiers(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetOwner_return alterDatabaseSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetOwner(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceOrExpression_return precedenceOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceOrExpression(); }\n+\tpublic HiveParser_IdentifiersParser.firstExpressionsWithAlias_return firstExpressionsWithAlias() throws RecognitionException { return gIdentifiersParser.firstExpressionsWithAlias(); }\n \n-\tpublic HiveParser_CreateDDLParser.likeTableOrFile_return likeTableOrFile() throws RecognitionException { return gCreateDDLParser.likeTableOrFile(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAmpersandExpression_return precedenceAmpersandExpression() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandExpression(); }\n \n-\tpublic HiveParser_SelectClauseParser.trfmClause_return trfmClause() throws RecognitionException { return gSelectClauseParser.trfmClause(); }\n+\tpublic HiveParser_IdentifiersParser.parameterIdx_return parameterIdx() throws RecognitionException { return gIdentifiersParser.parameterIdx(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrOperator_return precedenceBitwiseOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropPartitions_return alterStatementSuffixDropPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropPartitions(table); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSerdeProperties_return alterStatementSuffixSerdeProperties(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSerdeProperties(partition); }\n+\tpublic HiveParser_IdentifiersParser.groupByClause_return groupByClause() throws RecognitionException { return gIdentifiersParser.groupByClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedencePlusOperator_return precedencePlusOperator() throws RecognitionException { return gIdentifiersParser.precedencePlusOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStatsCol_return alterStatementSuffixUpdateStatsCol(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStatsCol(partition); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceDistinctOperator_return precedenceDistinctOperator() throws RecognitionException { return gIdentifiersParser.precedenceDistinctOperator(); }\n+\tpublic HiveParser_ResourcePlanParser.poolPath_return poolPath() throws RecognitionException { return gResourcePlanParser.poolPath(); }\n \n-\tpublic HiveParser_IdentifiersParser.subQuerySelectorOperator_return subQuerySelectorOperator() throws RecognitionException { return gIdentifiersParser.subQuerySelectorOperator(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerAtomExpression_return triggerAtomExpression() throws RecognitionException { return gResourcePlanParser.triggerAtomExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.identifier_return identifier() throws RecognitionException { return gIdentifiersParser.identifier(); }\n+\tpublic HiveParser_IdentifiersParser.expressionWithAlias_return expressionWithAlias() throws RecognitionException { return gIdentifiersParser.expressionWithAlias(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceStarExpression_return precedenceStarExpression() throws RecognitionException { return gIdentifiersParser.precedenceStarExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionAtom_return precedenceSimilarExpressionAtom(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionAtom(t); }\n \n-\tpublic HiveParser_FromClauseParser.valueRowConstructor_return valueRowConstructor() throws RecognitionException { return gFromClauseParser.valueRowConstructor(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPart_return precedenceSimilarExpressionPart(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPart(t); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceEqualOperator_return precedenceEqualOperator() throws RecognitionException { return gIdentifiersParser.precedenceEqualOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterViewStatementSuffix_return alterViewStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterViewStatementSuffix(); }\n \n-\tpublic HiveParser_CreateDDLParser.createTableStatement_return createTableStatement() throws RecognitionException { return gCreateDDLParser.createTableStatement(); }\n+\tpublic HiveParser_AlterClauseParser.blocking_return blocking() throws RecognitionException { return gAlterClauseParser.blocking(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixBucketNum_return alterStatementSuffixBucketNum(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixBucketNum(partition); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixStatsPart_return alterStatementSuffixStatsPart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixStatsPart(); }\n \n \tpublic HiveParser_ResourcePlanParser.alterResourcePlanStatement_return alterResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.alterResourcePlanStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.parameterIdx_return parameterIdx() throws RecognitionException { return gIdentifiersParser.parameterIdx(); }\n-\n-\tpublic HiveParser_IdentifiersParser.precedenceUnaryPrefixExpression_return precedenceUnaryPrefixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnaryPrefixExpression(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerLiteral_return triggerLiteral() throws RecognitionException { return gResourcePlanParser.triggerLiteral(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAmpersandOperator_return precedenceAmpersandOperator() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandOperator(); }\n+\tpublic HiveParser_SelectClauseParser.selectExpression_return selectExpression() throws RecognitionException { return gSelectClauseParser.selectExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.nonReserved_return nonReserved() throws RecognitionException { return gIdentifiersParser.nonReserved(); }\n+\tpublic HiveParser_FromClauseParser.tableSample_return tableSample() throws RecognitionException { return gFromClauseParser.tableSample(); }\n \n-\tpublic HiveParser_FromClauseParser.partitionTableFunctionSource_return partitionTableFunctionSource() throws RecognitionException { return gFromClauseParser.partitionTableFunctionSource(); }\n+\tpublic HiveParser_IdentifiersParser.whenExpression_return whenExpression() throws RecognitionException { return gIdentifiersParser.whenExpression(); }\n \n-\tpublic HiveParser_FromClauseParser.asOfClause_return asOfClause() throws RecognitionException { return gFromClauseParser.asOfClause(); }\n+\tpublic HiveParser_FromClauseParser.virtualTableSource_return virtualTableSource() throws RecognitionException { return gFromClauseParser.virtualTableSource(); }\n \n-\tpublic HiveParser_FromClauseParser.joinSource_return joinSource() throws RecognitionException { return gFromClauseParser.joinSource(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixExchangePartition_return alterStatementSuffixExchangePartition() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExchangePartition(); }\n \n \tpublic HiveParser_FromClauseParser.whereClause_return whereClause() throws RecognitionException { return gFromClauseParser.whereClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropPartitions_return alterStatementSuffixDropPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropPartitions(table); }\n+\tpublic HiveParser_SelectClauseParser.window_frame_return window_frame() throws RecognitionException { return gSelectClauseParser.window_frame(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinTableSource_return uniqueJoinTableSource() throws RecognitionException { return gFromClauseParser.uniqueJoinTableSource(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionMain_return precedenceSimilarExpressionMain() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionMain(); }\n \n-\tpublic HiveParser_ResourcePlanParser.disable_return disable() throws RecognitionException { return gResourcePlanParser.disable(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAmpersandOperator_return precedenceAmpersandOperator() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorExpression_return precedenceBitwiseXorExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorExpression(); }\n+\tpublic HiveParser_SelectClauseParser.window_specification_return window_specification(CommonTree nullTreatment) throws RecognitionException { return gSelectClauseParser.window_specification(nullTreatment); }\n \n-\tpublic HiveParser_IdentifiersParser.isCondition_return isCondition() throws RecognitionException { return gIdentifiersParser.isCondition(); }\n+\tpublic HiveParser_CreateDDLParser.likeTableOrFile_return likeTableOrFile() throws RecognitionException { return gCreateDDLParser.likeTableOrFile(); }\n \n-\tpublic HiveParser_PrepareStatementParser.prepareStatement_return prepareStatement() throws RecognitionException { return gPrepareStatementParser.prepareStatement(); }\n+\tpublic HiveParser_FromClauseParser.tableBucketSample_return tableBucketSample() throws RecognitionException { return gFromClauseParser.tableBucketSample(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorStatementSuffix_return alterDataConnectorStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDataConnectorStatementSuffix(); }\n+\tpublic HiveParser_FromClauseParser.splitSample_return splitSample() throws RecognitionException { return gFromClauseParser.splitSample(); }\n \n-\tpublic HiveParser_FromClauseParser.viewName_return viewName() throws RecognitionException { return gFromClauseParser.viewName(); }\n+\tpublic HiveParser_IdentifiersParser.nonReserved_return nonReserved() throws RecognitionException { return gIdentifiersParser.nonReserved(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceOrOperator_return precedenceOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceOrOperator(); }\n+\tpublic HiveParser_ResourcePlanParser.rpUnassign_return rpUnassign() throws RecognitionException { return gResourcePlanParser.rpUnassign(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupby_expression_return groupby_expression() throws RecognitionException { return gIdentifiersParser.groupby_expression(); }\n+\tpublic HiveParser_CreateDDLParser.createTableStatement_return createTableStatement() throws RecognitionException { return gCreateDDLParser.createTableStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.searchCondition_return searchCondition() throws RecognitionException { return gFromClauseParser.searchCondition(); }\n+\tpublic HiveParser_IdentifiersParser.rollupOldSyntax_return rollupOldSyntax() throws RecognitionException { return gIdentifiersParser.rollupOldSyntax(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorOperator_return precedenceBitwiseXorOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorOperator(); }\n+\tpublic HiveParser_CreateDDLParser.createDataConnectorStatement_return createDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.createDataConnectorStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.rollupStandard_return rollupStandard() throws RecognitionException { return gIdentifiersParser.rollupStandard(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixClusterbySortby_return alterStatementSuffixClusterbySortby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixClusterbySortby(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupingSetExpressionMultiple_return groupingSetExpressionMultiple() throws RecognitionException { return gIdentifiersParser.groupingSetExpressionMultiple(); }\n+\tpublic HiveParser_IdentifiersParser.sysFuncNames_return sysFuncNames() throws RecognitionException { return gIdentifiersParser.sysFuncNames(); }\n \n-\tpublic HiveParser_IdentifiersParser.booleanValue_return booleanValue() throws RecognitionException { return gIdentifiersParser.booleanValue(); }\n+\tpublic HiveParser_CreateDDLParser.dataConnectorUrl_return dataConnectorUrl() throws RecognitionException { return gCreateDDLParser.dataConnectorUrl(); }\n \n-\tpublic HiveParser_IdentifiersParser.floorExpression_return floorExpression() throws RecognitionException { return gIdentifiersParser.floorExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceStarExpression_return precedenceStarExpression() throws RecognitionException { return gIdentifiersParser.precedenceStarExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName_return sql11ReservedKeywordsUsedAsFunctionName() throws RecognitionException { return gIdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName(); }\n+\tpublic HiveParser_CreateDDLParser.dataConnectorType_return dataConnectorType() throws RecognitionException { return gCreateDDLParser.dataConnectorType(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerAndExpression_return triggerAndExpression() throws RecognitionException { return gResourcePlanParser.triggerAndExpression(); }\n+\tpublic HiveParser_AlterClauseParser.tablePartitionPrefix_return tablePartitionPrefix() throws RecognitionException { return gAlterClauseParser.tablePartitionPrefix(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitions_return alterStatementSuffixAddPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitions(table); }\n+\tpublic HiveParser_AlterClauseParser.alterMaterializedViewStatementSuffix_return alterMaterializedViewStatementSuffix(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewStatementSuffix(tableNameTree); }\n \n-\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation_return alterTblPartitionStatementSuffixSkewedLocation() throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation(); }\n+\tpublic HiveParser_FromClauseParser.joinSourcePart_return joinSourcePart() throws RecognitionException { return gFromClauseParser.joinSourcePart(); }\n \n-\tpublic HiveParser_AlterClauseParser.tablePartitionPrefix_return tablePartitionPrefix() throws RecognitionException { return gAlterClauseParser.tablePartitionPrefix(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixBucketNum_return alterStatementSuffixBucketNum(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixBucketNum(partition); }\n \n-\tpublic HiveParser_AlterClauseParser.alterViewStatementSuffix_return alterViewStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterViewStatementSuffix(); }\n+\tpublic HiveParser_FromClauseParser.defaultValue_return defaultValue() throws RecognitionException { return gFromClauseParser.defaultValue(); }\n \n-\tpublic HiveParser_IdentifiersParser.firstExpressionsWithAlias_return firstExpressionsWithAlias() throws RecognitionException { return gIdentifiersParser.firstExpressionsWithAlias(); }\n+\tpublic HiveParser_SelectClauseParser.window_frame_start_boundary_return window_frame_start_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_start_boundary(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixProperties_return alterStatementSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixProperties(); }\n+\tpublic HiveParser_CreateDDLParser.dataConnectorComment_return dataConnectorComment() throws RecognitionException { return gCreateDDLParser.dataConnectorComment(); }\n \n-\tpublic HiveParser_FromClauseParser.tableAllColumns_return tableAllColumns() throws RecognitionException { return gFromClauseParser.tableAllColumns(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCompact_return alterStatementSuffixCompact() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCompact(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetOwner_return alterDataConnectorSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetOwner(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetOwner_return alterStatementSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetOwner(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpAssign_return rpAssign() throws RecognitionException { return gResourcePlanParser.rpAssign(); }\n+\tpublic HiveParser_ResourcePlanParser.comparisionOperator_return comparisionOperator() throws RecognitionException { return gResourcePlanParser.comparisionOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterTableStatementSuffix_return alterTableStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterTableStatementSuffix(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarOperator_return precedenceSimilarOperator() throws RecognitionException { return gIdentifiersParser.precedenceSimilarOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.rollupOldSyntax_return rollupOldSyntax() throws RecognitionException { return gIdentifiersParser.rollupOldSyntax(); }\n+\tpublic HiveParser_IdentifiersParser.subQuerySelectorOperator_return subQuerySelectorOperator() throws RecognitionException { return gIdentifiersParser.subQuerySelectorOperator(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropMappingStatement_return dropMappingStatement() throws RecognitionException { return gResourcePlanParser.dropMappingStatement(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddCol_return alterStatementSuffixAddCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddCol(); }\n \n-\tpublic HiveParser_CreateDDLParser.createDataConnectorStatement_return createDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.createDataConnectorStatement(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceConcatenateExpression_return precedenceConcatenateExpression() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.booleanValueTok_return booleanValueTok() throws RecognitionException { return gIdentifiersParser.booleanValueTok(); }\n+\tpublic HiveParser_IdentifiersParser.principalIdentifier_return principalIdentifier() throws RecognitionException { return gIdentifiersParser.principalIdentifier(); }\n \n-\tpublic HiveParser_FromClauseParser.joinToken_return joinToken() throws RecognitionException { return gFromClauseParser.joinToken(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixFileFormat_return alterStatementSuffixFileFormat(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixFileFormat(partition); }\n \n-\tpublic HiveParser_ResourcePlanParser.enable_return enable() throws RecognitionException { return gResourcePlanParser.enable(); }\n+\tpublic HiveParser_IdentifiersParser.timeQualifiers_return timeQualifiers() throws RecognitionException { return gIdentifiersParser.timeQualifiers(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpUnassignList_return rpUnassignList() throws RecognitionException { return gResourcePlanParser.rpUnassignList(); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinTableSource_return uniqueJoinTableSource() throws RecognitionException { return gFromClauseParser.uniqueJoinTableSource(); }\n \n-\tpublic HiveParser_ResourcePlanParser.activate_return activate() throws RecognitionException { return gResourcePlanParser.activate(); }\n+\tpublic HiveParser_AlterClauseParser.compactPool_return compactPool() throws RecognitionException { return gAlterClauseParser.compactPool(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceConcatenateExpression_return precedenceConcatenateExpression() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateExpression(); }\n+\tpublic HiveParser_IdentifiersParser.constant_return constant() throws RecognitionException { return gIdentifiersParser.constant(); }\n \n-\tpublic HiveParser_FromClauseParser.lateralView_return lateralView() throws RecognitionException { return gFromClauseParser.lateralView(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementPartitionKeyType_return alterStatementPartitionKeyType() throws RecognitionException { return gAlterClauseParser.alterStatementPartitionKeyType(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionAtom_return precedenceSimilarExpressionAtom(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionAtom(t); }\n+\tpublic HiveParser_IdentifiersParser.booleanValueTok_return booleanValueTok() throws RecognitionException { return gIdentifiersParser.booleanValueTok(); }\n+\n+\tpublic HiveParser_ResourcePlanParser.activate_return activate() throws RecognitionException { return gResourcePlanParser.activate(); }\n+\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateColumns_return alterStatementSuffixUpdateColumns() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateColumns(); }\n+\n+\tpublic HiveParser_ResourcePlanParser.rpAssign_return rpAssign() throws RecognitionException { return gResourcePlanParser.rpAssign(); }\n \n \tpublic final boolean synpred18_HiveParser() {\n \t\tstate.backtracking++;\n \t\tint start = input.mark();\n \t\ttry {\n \t\t\tsynpred18_HiveParser_fragment(); // can never throw exception\n \t\t} catch (RecognitionException re) {\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_AlterClauseParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_AlterClauseParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 AlterClauseParser.g 2022-11-07 19:16:39\n+// $ANTLR 3.5.2 AlterClauseParser.g 2025-01-31 11:18:18\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -3208,15 +3208,15 @@\n \n \t\t\tpushFollow(FOLLOW_dbProperties_in_alterDatabaseSuffixProperties765);\n \t\t\tdbProperties71=gHiveParser.dbProperties();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_dbProperties.add(dbProperties71.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: name, dbProperties\n+\t\t\t// elements: dbProperties, name\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3309,15 +3309,15 @@\n \n \t\t\tpushFollow(FOLLOW_principalName_in_alterDatabaseSuffixSetOwner815);\n \t\t\tprincipalName74=gHiveParser.principalName();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_principalName.add(principalName74.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: principalName, dbName\n+\t\t\t// elements: dbName, principalName\n \t\t\t// token labels: \n \t\t\t// rule labels: dbName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3510,15 +3510,15 @@\n \t\t\t\t\tKW_LOCATION76=(Token)match(input,KW_LOCATION,FOLLOW_KW_LOCATION_in_alterDatabaseSuffixSetLocation863); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_LOCATION.add(KW_LOCATION76);\n \n \t\t\t\t\tnewLocation=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_alterDatabaseSuffixSetLocation867); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(newLocation);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: dbName, newLocation\n+\t\t\t\t\t// elements: newLocation, dbName\n \t\t\t\t\t// token labels: newLocation\n \t\t\t\t\t// rule labels: dbName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3933,15 +3933,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_restrictOrCascade.add(restrictOrCascade88.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: columnNameTypeList, restrictOrCascade, columnNameTypeList, restrictOrCascade\n+\t\t\t// elements: columnNameTypeList, restrictOrCascade, restrictOrCascade, columnNameTypeList\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4572,15 +4572,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_restrictOrCascade.add(restrictOrCascade102.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: comment, colType, alterColumnConstraint, newName, restrictOrCascade, oldName, alterStatementChangeColPosition\n+\t\t\t// elements: oldName, colType, alterStatementChangeColPosition, newName, alterColumnConstraint, comment, restrictOrCascade\n \t\t\t// token labels: comment\n \t\t\t// rule labels: newName, oldName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4760,15 +4760,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableProperties, colName, comment, colName, comment, tableProperties\n+\t\t\t// elements: comment, tableProperties, comment, colName, colName, tableProperties\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5164,15 +5164,15 @@\n \t\t\t\t\tEarlyExitException eee = new EarlyExitException(24, input);\n \t\t\t\t\tthrow eee;\n \t\t\t\t}\n \t\t\t\tcnt24++;\n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: ifNotExists, ifNotExists, alterStatementSuffixAddPartitionsElement, alterStatementSuffixAddPartitionsElement\n+\t\t\t// elements: ifNotExists, alterStatementSuffixAddPartitionsElement, alterStatementSuffixAddPartitionsElement, ifNotExists\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5894,15 +5894,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replicationClause.add(replicationClause135.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: partitionSelectorSpec, partitionSelectorSpec, replicationClause, KW_PURGE, ifExists, replicationClause, ifExists\n+\t\t\t// elements: partitionSelectorSpec, KW_PURGE, ifExists, partitionSelectorSpec, replicationClause, replicationClause, ifExists\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -6132,15 +6132,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_tableProperties_in_alterStatementSuffixProperties2061);\n \t\t\t\t\ttableProperties142=gHiveParser.tableProperties();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableProperties.add(tableProperties142.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableProperties, ifExists\n+\t\t\t\t\t// elements: ifExists, tableProperties\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -6322,15 +6322,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_tableProperties_in_alterViewSuffixProperties2134);\n \t\t\t\t\ttableProperties149=gHiveParser.tableProperties();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableProperties.add(tableProperties149.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ifExists, tableProperties\n+\t\t\t\t\t// elements: tableProperties, ifExists\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -6504,15 +6504,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_tableProperties.add(tableProperties154.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: serdeName, tableProperties, serdeName, tableProperties\n+\t\t\t\t\t// elements: serdeName, tableProperties, tableProperties, serdeName\n \t\t\t\t\t// token labels: serdeName\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -6745,15 +6745,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec162.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: partitionSpec, tableName\n+\t\t\t// elements: tableName, partitionSpec\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -7466,15 +7466,15 @@\n \t\t\tEQUAL181=(Token)match(input,EQUAL,FOLLOW_EQUAL_in_skewedLocationMap2675); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_EQUAL.add(EQUAL181);\n \n \t\t\tvalue=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_skewedLocationMap2679); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(value);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: key, value\n+\t\t\t// elements: value, key\n \t\t\t// token labels: value\n \t\t\t// rule labels: key, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -7887,15 +7887,15 @@\n \n \t\t\tpushFollow(FOLLOW_tableName_in_alterStatementSuffixExchangePartition2855);\n \t\t\texchangename=gHiveParser.tableName();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_tableName.add(exchangename.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: partitionSpec, exchangename\n+\t\t\t// elements: exchangename, partitionSpec\n \t\t\t// token labels: \n \t\t\t// rule labels: exchangename, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -8140,15 +8140,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: comment, colName, tableProperties\n+\t\t\t// elements: tableProperties, colName, comment\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -8671,15 +8671,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableProperties.add(tableProperties215.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: compactType, blocking, compactPool, tableProperties\n+\t\t\t// elements: compactType, blocking, tableProperties, compactPool\n \t\t\t// token labels: compactType\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -9087,15 +9087,15 @@\n \n \t\t\t\t\t}\n \n \t\t\t\t\tRPAREN227=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_alterStatementSuffixExecute3383); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN227);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: rollbackParam, KW_ROLLBACK\n+\t\t\t\t\t// elements: KW_ROLLBACK, rollbackParam\n \t\t\t\t\t// token labels: rollbackParam\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9142,15 +9142,15 @@\n \n \t\t\t\t\t}\n \n \t\t\t\t\tRPAREN231=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_alterStatementSuffixExecute3418); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN231);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_EXPIRE_SNAPSHOTS, expireParam\n+\t\t\t\t\t// elements: expireParam, KW_EXPIRE_SNAPSHOTS\n \t\t\t\t\t// token labels: expireParam\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9330,26 +9330,26 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: outDriver, inFmt, serdeCls, outFmt, inDriver\n-\t\t\t\t\t// token labels: inFmt, inDriver, outDriver, serdeCls, outFmt\n+\t\t\t\t\t// elements: inFmt, outFmt, serdeCls, inDriver, outDriver\n+\t\t\t\t\t// token labels: inFmt, inDriver, serdeCls, outDriver, outFmt\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleTokenStream stream_inFmt=new RewriteRuleTokenStream(adaptor,\"token inFmt\",inFmt);\n \t\t\t\t\tRewriteRuleTokenStream stream_inDriver=new RewriteRuleTokenStream(adaptor,\"token inDriver\",inDriver);\n-\t\t\t\t\tRewriteRuleTokenStream stream_outDriver=new RewriteRuleTokenStream(adaptor,\"token outDriver\",outDriver);\n \t\t\t\t\tRewriteRuleTokenStream stream_serdeCls=new RewriteRuleTokenStream(adaptor,\"token serdeCls\",serdeCls);\n+\t\t\t\t\tRewriteRuleTokenStream stream_outDriver=new RewriteRuleTokenStream(adaptor,\"token outDriver\",outDriver);\n \t\t\t\t\tRewriteRuleTokenStream stream_outFmt=new RewriteRuleTokenStream(adaptor,\"token outFmt\",outFmt);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t\t\t// 474:7: -> ^( TOK_TABLEFILEFORMAT $inFmt $outFmt $serdeCls ( $inDriver)? ( $outDriver)? )\n \t\t\t\t\t{\n \t\t\t\t\t\t// AlterClauseParser.g:474:10: ^( TOK_TABLEFILEFORMAT $inFmt $outFmt $serdeCls ( $inDriver)? ( $outDriver)? )\n@@ -9686,15 +9686,15 @@\n \n \t\t\tpushFollow(FOLLOW_dcProperties_in_alterDataConnectorSuffixProperties3617);\n \t\t\tdcProperties242=gHiveParser.dcProperties();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_dcProperties.add(dcProperties242.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: dcProperties, name\n+\t\t\t// elements: name, dcProperties\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_CreateDDLParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_CreateDDLParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 CreateDDLParser.g 2022-11-07 19:16:42\n+// $ANTLR 3.5.2 CreateDDLParser.g 2025-01-31 11:18:20\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1785,15 +1785,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: likeTableOrFile, temp, columnNameTypeOrConstraintList, tableLocation, tableComment, tableFileFormat, trans, selectStatementWithCTE, tableSkewed, tableRowFormat, createTablePartitionSpec, ifNotExists, name, tableBuckets, ext, tablePropertiesPrefixed\n+\t\t\t\t\t// elements: ifNotExists, tableLocation, createTablePartitionSpec, trans, name, tableFileFormat, likeTableOrFile, columnNameTypeOrConstraintList, tableBuckets, tableSkewed, tablePropertiesPrefixed, temp, ext, tableRowFormat, tableComment, selectStatementWithCTE\n \t\t\t\t\t// token labels: ext, temp, trans\n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -2271,15 +2271,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableLocation, selectStatementWithCTE, name, tableSkewed, tablePropertiesPrefixed, tableBuckets, ifNotExists, tableFileFormat, mgd, tableComment, tableRowFormat, createTablePartitionSpec, columnNameTypeOrConstraintList, likeTableOrFile\n+\t\t\t\t\t// elements: tableSkewed, columnNameTypeOrConstraintList, ifNotExists, likeTableOrFile, mgd, name, createTablePartitionSpec, selectStatementWithCTE, tableLocation, tableComment, tableRowFormat, tableFileFormat, tableBuckets, tablePropertiesPrefixed\n \t\t\t\t\t// token labels: mgd\n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -2532,15 +2532,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_dcProperties.add(dcprops.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: dataConnectorUrl, name, ifNotExists, dataConnectorComment, dcprops, dataConnectorType\n+\t\t\t// elements: dataConnectorComment, ifNotExists, name, dataConnectorUrl, dcprops, dataConnectorType\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval, dcprops\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3023,15 +3023,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_dropDataConnectorStatement1229);\n \t\t\tidentifier66=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_identifier.add(identifier66.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: ifExists, identifier\n+\t\t\t// elements: identifier, ifExists\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_FromClauseParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_FromClauseParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 FromClauseParser.g 2022-11-07 19:16:40\n+// $ANTLR 3.5.2 FromClauseParser.g 2025-01-31 11:18:18\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -3674,15 +3674,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableAlias, identifier, valuesClause\n+\t\t\t\t\t// elements: tableAlias, valuesClause, identifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3979,15 +3979,15 @@\n \n \t\t\t}\n \n \t\t\tRPAREN114=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_tableBucketSample1292); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN114);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: expr, denominator, numerator\n+\t\t\t// elements: denominator, numerator, expr\n \t\t\t// token labels: denominator, numerator\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: expr\n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4575,28 +4575,28 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(alias.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tabname, alias, props, ts, asOf\n+\t\t\t// elements: ts, alias, tabname, props, asOf\n \t\t\t// token labels: \n-\t\t\t// rule labels: tabname, asOf, alias, retval, props, ts\n+\t\t\t// rule labels: tabname, asOf, alias, retval, ts, props\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_tabname=new RewriteRuleSubtreeStream(adaptor,\"rule tabname\",tabname!=null?tabname.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_asOf=new RewriteRuleSubtreeStream(adaptor,\"rule asOf\",asOf!=null?asOf.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_alias=new RewriteRuleSubtreeStream(adaptor,\"rule alias\",alias!=null?alias.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n-\t\t\tRewriteRuleSubtreeStream stream_props=new RewriteRuleSubtreeStream(adaptor,\"rule props\",props!=null?props.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_ts=new RewriteRuleSubtreeStream(adaptor,\"rule ts\",ts!=null?ts.getTree():null);\n+\t\t\tRewriteRuleSubtreeStream stream_props=new RewriteRuleSubtreeStream(adaptor,\"rule props\",props!=null?props.getTree():null);\n \n \t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t// 213:5: -> ^( TOK_TABREF $tabname ( $props)? ( $ts)? ( $alias)? ( $asOf)? )\n \t\t\t{\n \t\t\t\t// FromClauseParser.g:213:8: ^( TOK_TABREF $tabname ( $props)? ( $ts)? ( $alias)? ( $asOf)? )\n \t\t\t\t{\n \t\t\t\tASTNode root_1 = (ASTNode)adaptor.nil();\n@@ -4961,15 +4961,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(alias.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tabname, ts, alias\n+\t\t\t// elements: alias, tabname, ts\n \t\t\t// token labels: \n \t\t\t// rule labels: tabname, alias, retval, ts\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5096,15 +5096,15 @@\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\tif ( state.backtracking==0 ) {tables.add(new ImmutablePair<>((db!=null?input.toString(db.start,db.stop):null), (tab!=null?input.toString(tab.start,tab.stop):null)));}\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: meta, tab, db\n+\t\t\t\t\t// elements: tab, meta, db\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: tab, meta, db, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -5264,15 +5264,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_viewName1859);\n \t\t\tview=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_identifier.add(view.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: view, db\n+\t\t\t// elements: db, view\n \t\t\t// token labels: \n \t\t\t// rule labels: view, db, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5538,15 +5538,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_orderByClause.add(orderByClause143.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: orderByClause, partitionByClause\n+\t\t\t\t\t// elements: partitionByClause, orderByClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -5640,15 +5640,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_sortByClause.add(sortByClause146.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: distributeByClause, sortByClause\n+\t\t\t\t\t// elements: sortByClause, distributeByClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -6400,15 +6400,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(alias.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, ptfsrc, alias, name, spec\n+\t\t\t// elements: spec, name, alias, expression, ptfsrc\n \t\t\t// token labels: \n \t\t\t// rule labels: ptfsrc, name, alias, spec, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_IdentifiersParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_IdentifiersParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 IdentifiersParser.g 2022-11-07 19:16:41\n+// $ANTLR 3.5.2 IdentifiersParser.g 2025-01-31 11:18:19\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -2710,15 +2710,15 @@\n \t\t\t\tdefault :\n \t\t\t\t\tbreak loop15;\n \t\t\t\t}\n \t\t\t}\n \n \t\t\tif ( state.backtracking==0 ) { incAliasCounter(); }\n \t\t\t// AST REWRITE\n-\t\t\t// elements: expressionWithAlias, identifier, expressionWithAlias\n+\t\t\t// elements: identifier, expressionWithAlias, expressionWithAlias\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3943,15 +3943,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_selectExpression.add(str.getTree());\n \t\t\t}\n \n \t\t\tRPAREN79=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_trimFunction1438); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN79);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: trim_characters, trim_characters, str, str, trim_characters, str\n+\t\t\t// elements: trim_characters, str, str, str, trim_characters, trim_characters\n \t\t\t// token labels: \n \t\t\t// rule labels: str, trim_characters, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4762,15 +4762,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ws, selectExpression, ordBy, ws, selectExpression, selectExpression, functionName, functionName, ws, functionName, functionName\n+\t\t\t\t\t// elements: ws, functionName, selectExpression, functionName, ordBy, ws, ws, selectExpression, functionName, functionName, selectExpression\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: ordBy, ws, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -5248,15 +5248,15 @@\n \n \t\t\t}\n \n \t\t\tRPAREN107=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_castExpression2132); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN107);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, StringLiteral, toType, StringLiteral, expression, expression\n+\t\t\t// elements: expression, expression, expression, StringLiteral, toType, StringLiteral\n \t\t\t// token labels: \n \t\t\t// rule labels: toType, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5619,15 +5619,15 @@\n \n \t\t\t}\n \n \t\t\tKW_END124=(Token)match(input,KW_END,FOLLOW_KW_END_in_whenExpression2357); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_KW_END.add(KW_END124);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: KW_WHEN, expression\n+\t\t\t// elements: expression, KW_WHEN\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -10154,15 +10154,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_subQueryExpression_in_precedenceSimilarExpression4333);\n \t\t\t\t\tsubQueryExpression250=subQueryExpression();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_subQueryExpression.add(subQueryExpression250.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: subQueryExpression, KW_EXISTS\n+\t\t\t\t\t// elements: KW_EXISTS, subQueryExpression\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10357,15 +10357,15 @@\n \t\t\t\t\tequalExpr=precedenceBitwiseOrExpression();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_precedenceBitwiseOrExpression.add(equalExpr.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: precedenceSimilarOperator, equalExpr\n+\t\t\t\t\t// elements: equalExpr, precedenceSimilarOperator\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval, equalExpr\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10593,15 +10593,15 @@\n \t\t\t\t\tmax=precedenceBitwiseOrExpression();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_precedenceBitwiseOrExpression.add(max.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: max, min\n+\t\t\t\t\t// elements: min, max\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: min, max, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10809,15 +10809,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_quantifierType.add(quantifierType265.getTree());\n \t\t\tpushFollow(FOLLOW_subQueryExpression_in_precedenceSimilarExpressionQuantifierPredicate4666);\n \t\t\tsubQueryExpression266=subQueryExpression();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_subQueryExpression.add(subQueryExpression266.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: subQueryExpression, subQuerySelectorOperator, quantifierType\n+\t\t\t// elements: subQueryExpression, quantifierType, subQuerySelectorOperator\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -11254,15 +11254,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_precedenceRegexpOperator.add(precedenceRegexpOperator271.getTree());\n \t\t\t\t\tpushFollow(FOLLOW_precedenceBitwiseOrExpression_in_precedenceSimilarExpressionPartNot4848);\n \t\t\t\t\tnotExpr=precedenceBitwiseOrExpression();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_precedenceBitwiseOrExpression.add(notExpr.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: precedenceRegexpOperator, notExpr\n+\t\t\t\t\t// elements: notExpr, precedenceRegexpOperator\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: notExpr, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -12208,15 +12208,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_isCondition.add(isCondition297.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: isCondition, precedenceEqualExpression, precedenceEqualExpression\n+\t\t\t// elements: precedenceEqualExpression, precedenceEqualExpression, isCondition\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -12942,15 +12942,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec314.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, partitionSpec\n+\t\t\t// elements: partitionSpec, tableName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -13418,15 +13418,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_partitionSelectorOperator.add(partitionSelectorOperator330.getTree());\n \t\t\tpushFollow(FOLLOW_constant_in_partitionSelectorVal5657);\n \t\t\tconstant331=constant();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_constant.add(constant331.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: partitionSelectorOperator, constant, identifier\n+\t\t\t// elements: identifier, constant, partitionSelectorOperator\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_PrepareStatementParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_PrepareStatementParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 PrepareStatementParser.g 2022-11-07 19:16:42\n+// $ANTLR 3.5.2 PrepareStatementParser.g 2025-01-31 11:18:20\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1131,15 +1131,15 @@\n \n \t\t\tpushFollow(FOLLOW_executeParamList_in_executeStatement119);\n \t\t\texecuteParamList8=executeParamList();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_executeParamList.add(executeParamList8.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: executeParamList, identifier\n+\t\t\t// elements: identifier, executeParamList\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_ResourcePlanParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_ResourcePlanParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 ResourcePlanParser.g 2022-11-07 19:16:41\n+// $ANTLR 3.5.2 ResourcePlanParser.g 2025-01-31 11:18:20\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -2068,15 +2068,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_identifier_in_createResourcePlanStatement435);\n \t\t\t\t\tlikeName=gHiveParser.identifier();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_identifier.add(likeName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ifNotExists, name, likeName\n+\t\t\t\t\t// elements: name, ifNotExists, likeName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: likeName, name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_likeName=new RewriteRuleSubtreeStream(adaptor,\"rule likeName\",likeName!=null?likeName.getTree():null);\n@@ -2148,15 +2148,15 @@\n \t\t\t\t\t\t\tstream_rpAssignList.add(rpAssignList36.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, ifNotExists, rpAssignList\n+\t\t\t\t\t// elements: ifNotExists, name, rpAssignList\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -2825,15 +2825,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_rpAssignList_in_alterResourcePlanStatement675);\n \t\t\t\t\trpAssignList50=rpAssignList();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_rpAssignList.add(rpAssignList50.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, rpAssignList\n+\t\t\t\t\t// elements: rpAssignList, name\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -2871,15 +2871,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_rpUnassignList_in_alterResourcePlanStatement702);\n \t\t\t\t\trpUnassignList52=rpUnassignList();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_rpUnassignList.add(rpUnassignList52.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: rpUnassignList, name\n+\t\t\t\t\t// elements: name, rpUnassignList\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -3039,15 +3039,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: activate, name, enable\n+\t\t\t\t\t// elements: name, enable, activate\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -3404,15 +3404,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_identifier_in_replaceResourcePlanStatement939);\n \t\t\t\t\tsrc=gHiveParser.identifier();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_identifier.add(src.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: src, dest\n+\t\t\t\t\t// elements: dest, src\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: src, dest, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_src=new RewriteRuleSubtreeStream(adaptor,\"rule src\",src!=null?src.getTree():null);\n@@ -3531,15 +3531,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_dropResourcePlanStatement998);\n \t\t\tname=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_identifier.add(name.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: name, ifExists\n+\t\t\t// elements: ifExists, name\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -4670,15 +4670,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_triggerActionExpression_in_alterTriggerStatement1440);\n \t\t\t\t\ttriggerActionExpression113=triggerActionExpression();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_triggerActionExpression.add(triggerActionExpression113.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: triggerExpression, triggerName, rpName, triggerActionExpression\n+\t\t\t\t\t// elements: rpName, triggerExpression, triggerName, triggerActionExpression\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -4725,15 +4725,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_poolPath_in_alterTriggerStatement1486);\n \t\t\t\t\tpoolName=poolPath();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_poolPath.add(poolName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: poolName, triggerName, rpName\n+\t\t\t\t\t// elements: triggerName, rpName, poolName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval, poolName\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -4780,15 +4780,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_poolPath_in_alterTriggerStatement1521);\n \t\t\t\t\tpoolName=poolPath();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_poolPath.add(poolName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: triggerName, poolName, rpName\n+\t\t\t\t\t// elements: poolName, rpName, triggerName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval, poolName\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -4991,15 +4991,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_dropTriggerStatement1644);\n \t\t\ttriggerName=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_identifier.add(triggerName.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: rpName, triggerName\n+\t\t\t// elements: triggerName, rpName\n \t\t\t// token labels: \n \t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -5512,15 +5512,15 @@\n \n \t\t\tpushFollow(FOLLOW_poolAssignList_in_createPoolStatement1889);\n \t\t\tpoolAssignList145=poolAssignList();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_poolAssignList.add(poolAssignList145.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: poolPath, rpName, poolAssignList\n+\t\t\t// elements: rpName, poolAssignList, poolPath\n \t\t\t// token labels: \n \t\t\t// rule labels: rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -5683,15 +5683,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_poolAssignList_in_alterPoolStatement1958);\n \t\t\t\t\tpoolAssignList151=poolAssignList();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_poolAssignList.add(poolAssignList151.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: rpName, poolAssignList, poolPath\n+\t\t\t\t\t// elements: rpName, poolPath, poolAssignList\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -5728,15 +5728,15 @@\n \t\t\t\t\tKW_UNSET152=(Token)match(input,KW_UNSET,FOLLOW_KW_UNSET_in_alterPoolStatement1985);  \n \t\t\t\t\tstream_KW_UNSET.add(KW_UNSET152);\n \n \t\t\t\t\tKW_SCHEDULING_POLICY153=(Token)match(input,KW_SCHEDULING_POLICY,FOLLOW_KW_SCHEDULING_POLICY_in_alterPoolStatement1987);  \n \t\t\t\t\tstream_KW_SCHEDULING_POLICY.add(KW_SCHEDULING_POLICY153);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: poolPath, rpName\n+\t\t\t\t\t// elements: rpName, poolPath\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -5785,15 +5785,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_identifier_in_alterPoolStatement2024);\n \t\t\t\t\ttriggerName=gHiveParser.identifier();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_identifier.add(triggerName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: triggerName, poolPath, rpName\n+\t\t\t\t\t// elements: poolPath, triggerName, rpName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -5836,15 +5836,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_identifier_in_alterPoolStatement2058);\n \t\t\t\t\ttriggerName=gHiveParser.identifier();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_identifier.add(triggerName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: rpName, poolPath, triggerName\n+\t\t\t\t\t// elements: triggerName, rpName, poolPath\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -5949,15 +5949,15 @@\n \n \t\t\tpushFollow(FOLLOW_poolPath_in_dropPoolStatement2118);\n \t\t\tpoolPath161=poolPath();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_poolPath.add(poolPath161.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: rpName, poolPath\n+\t\t\t// elements: poolPath, rpName\n \t\t\t// token labels: \n \t\t\t// rule labels: rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -6201,15 +6201,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: mappingType, path, order, rpName, name, unmanaged\n+\t\t\t// elements: name, mappingType, unmanaged, order, rpName, path\n \t\t\t// token labels: mappingType, name, order\n \t\t\t// rule labels: path, rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleTokenStream stream_mappingType=new RewriteRuleTokenStream(adaptor,\"token mappingType\",mappingType);\n@@ -6476,15 +6476,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: mappingType, unmanaged, rpName, order, path, name\n+\t\t\t// elements: mappingType, name, rpName, order, path, unmanaged\n \t\t\t// token labels: mappingType, name, order\n \t\t\t// rule labels: path, rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleTokenStream stream_mappingType=new RewriteRuleTokenStream(adaptor,\"token mappingType\",mappingType);\n@@ -6658,15 +6658,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_dropMappingStatement2469);\n \t\t\trpName=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_identifier.add(rpName.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: rpName, mappingType, name\n+\t\t\t// elements: name, rpName, mappingType\n \t\t\t// token labels: mappingType, name\n \t\t\t// rule labels: rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleTokenStream stream_mappingType=new RewriteRuleTokenStream(adaptor,\"token mappingType\",mappingType);\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_SelectClauseParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_SelectClauseParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 SelectClauseParser.g 2022-11-07 19:16:39\n+// $ANTLR 3.5.2 SelectClauseParser.g 2025-01-31 11:18:18\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1153,15 +1153,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: selectList, QUERY_HINT, selectTrfmClause, QUERY_HINT, QUERY_HINT, selectList\n+\t\t\t\t\t// elements: QUERY_HINT, selectTrfmClause, QUERY_HINT, selectList, selectList, QUERY_HINT\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -1678,15 +1678,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_rowFormat.add(outSerde.getTree());\n \t\t\tpushFollow(FOLLOW_recordReader_in_selectTrfmClause357);\n \t\t\toutRec=gHiveParser.recordReader();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_recordReader.add(outRec.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: inSerde, inRec, outSerde, aliasList, selectExpressionList, columnNameTypeList, StringLiteral, outRec\n+\t\t\t// elements: StringLiteral, outSerde, outRec, inSerde, columnNameTypeList, selectExpressionList, aliasList, inRec\n \t\t\t// token labels: \n \t\t\t// rule labels: inRec, outRec, inSerde, outSerde, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -1939,15 +1939,15 @@\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: identifier, expression\n+\t\t\t\t\t// elements: expression, identifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -2298,15 +2298,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_rowFormat.add(outSerde.getTree());\n \t\t\tpushFollow(FOLLOW_recordReader_in_trfmClause622);\n \t\t\toutRec=gHiveParser.recordReader();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_recordReader.add(outRec.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: columnNameTypeList, inRec, outSerde, selectExpressionList, inSerde, StringLiteral, outRec, aliasList\n+\t\t\t// elements: inSerde, columnNameTypeList, inRec, outSerde, aliasList, outRec, StringLiteral, selectExpressionList\n \t\t\t// token labels: \n \t\t\t// rule labels: inRec, outRec, inSerde, outSerde, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3663,15 +3663,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_window_frame_boundary_in_window_value_expression1029);\n \t\t\t\t\tend=window_frame_boundary();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_window_frame_boundary.add(end.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: end, s\n+\t\t\t\t\t// elements: s, end\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: s, end, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3789,15 +3789,15 @@\n \t\t\t\t\tKW_UNBOUNDED73=(Token)match(input,KW_UNBOUNDED,FOLLOW_KW_UNBOUNDED_in_window_frame_start_boundary1064); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_UNBOUNDED.add(KW_UNBOUNDED73);\n \n \t\t\t\t\tKW_PRECEDING74=(Token)match(input,KW_PRECEDING,FOLLOW_KW_PRECEDING_in_window_frame_start_boundary1066); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_PRECEDING.add(KW_PRECEDING74);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_PRECEDING, KW_UNBOUNDED\n+\t\t\t\t\t// elements: KW_UNBOUNDED, KW_PRECEDING\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n"}]}
