{"diffoscope-json-version": 1, "source1": "/input1", "source2": "/input2", "unified_diff": null, "details": [{"source1": "zipinfo -v {}", "source2": "zipinfo -v {}", "unified_diff": "@@ -1376,15 +1376,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         95207a70\n+  32-bit CRC value (hex):                         fcd220db\n   compressed size:                                47644 bytes\n   uncompressed size:                              47644 bytes\n   length of filename:                             47 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1448,15 +1448,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         9536a20f\n+  32-bit CRC value (hex):                         5c86d616\n   compressed size:                                30210 bytes\n   uncompressed size:                              30210 bytes\n   length of filename:                             46 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1556,15 +1556,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         384f6380\n+  32-bit CRC value (hex):                         b3d3c0f4\n   compressed size:                                29518 bytes\n   uncompressed size:                              29518 bytes\n   length of filename:                             54 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1592,15 +1592,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         2b643e30\n+  32-bit CRC value (hex):                         9a85b6a1\n   compressed size:                                386217 bytes\n   uncompressed size:                              386217 bytes\n   length of filename:                             70 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1628,15 +1628,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         8b9b3621\n+  32-bit CRC value (hex):                         31237c9a\n   compressed size:                                387628 bytes\n   uncompressed size:                              387628 bytes\n   length of filename:                             62 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1700,15 +1700,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         c1e420a7\n+  32-bit CRC value (hex):                         41d8cd26\n   compressed size:                                2068462 bytes\n   uncompressed size:                              2068462 bytes\n   length of filename:                             47 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1736,15 +1736,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         df54657a\n+  32-bit CRC value (hex):                         d018e1c3\n   compressed size:                                554655 bytes\n   uncompressed size:                              554655 bytes\n   length of filename:                             65 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1772,15 +1772,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         70e8aba4\n+  32-bit CRC value (hex):                         63c92183\n   compressed size:                                150310 bytes\n   uncompressed size:                              150310 bytes\n   length of filename:                             63 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1808,15 +1808,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         a3bcbe1d\n+  32-bit CRC value (hex):                         60ba145d\n   compressed size:                                669381 bytes\n   uncompressed size:                              669381 bytes\n   length of filename:                             64 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1844,15 +1844,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         d5cd5f15\n+  32-bit CRC value (hex):                         3b437530\n   compressed size:                                1418080 bytes\n   uncompressed size:                              1418080 bytes\n   length of filename:                             65 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1880,15 +1880,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         2bd84363\n+  32-bit CRC value (hex):                         d332cb71\n   compressed size:                                56565 bytes\n   uncompressed size:                              56565 bytes\n   length of filename:                             70 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1916,15 +1916,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         c62697ee\n+  32-bit CRC value (hex):                         d0ae4dfe\n   compressed size:                                287498 bytes\n   uncompressed size:                              287498 bytes\n   length of filename:                             66 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1952,15 +1952,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         3baca14d\n+  32-bit CRC value (hex):                         0325c463\n   compressed size:                                288697 bytes\n   uncompressed size:                              288697 bytes\n   length of filename:                             66 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n"}, {"source1": "zipdetails --redact --scan --utc {}", "source2": "zipdetails --redact --scan --utc {}", "unified_diff": "@@ -1013,15 +1013,15 @@\n 0471BF Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 0471C1   Length              0005 (5)\n 0471C3   Flags               01 (1) 'Modification'\n 0471C4   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 0471C8 PAYLOAD\n \n 052BE4 DATA DESCRIPTOR       08074B50 (134695760)\n-052BE8 CRC                   95207A70 (2501933680)\n+052BE8 CRC                   FCD220DB (4241629403)\n 052BEC Compressed Size       0000BA1C (47644)\n 052BF0 Uncompressed Size     0000BA1C (47644)\n \n 052BF4 LOCAL HEADER #40      04034B50 (67324752)\n 052BF8 Extract Zip Spec      14 (20) '2.0'\n 052BF9 Extract OS            00 (0) 'MS-DOS'\n 052BFA General Purpose Flag  0008 (8)\n@@ -1069,15 +1069,15 @@\n 0531A5 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 0531A7   Length              0005 (5)\n 0531A9   Flags               01 (1) 'Modification'\n 0531AA   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 0531AE PAYLOAD\n \n 05A7B0 DATA DESCRIPTOR       08074B50 (134695760)\n-05A7B4 CRC                   9536A20F (2503385615)\n+05A7B4 CRC                   5C86D616 (1552340502)\n 05A7B8 Compressed Size       00007602 (30210)\n 05A7BC Uncompressed Size     00007602 (30210)\n \n 05A7C0 LOCAL HEADER #42      04034B50 (67324752)\n 05A7C4 Extract Zip Spec      14 (20) '2.0'\n 05A7C5 Extract OS            00 (0) 'MS-DOS'\n 05A7C6 General Purpose Flag  0008 (8)\n@@ -1153,15 +1153,15 @@\n 05DD87 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 05DD89   Length              0005 (5)\n 05DD8B   Flags               01 (1) 'Modification'\n 05DD8C   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 05DD90 PAYLOAD\n \n 0650DE DATA DESCRIPTOR       08074B50 (134695760)\n-0650E2 CRC                   384F6380 (944726912)\n+0650E2 CRC                   B3D3C0F4 (3016999156)\n 0650E6 Compressed Size       0000734E (29518)\n 0650EA Uncompressed Size     0000734E (29518)\n \n 0650EE LOCAL HEADER #45      04034B50 (67324752)\n 0650F2 Extract Zip Spec      14 (20) '2.0'\n 0650F3 Extract OS            00 (0) 'MS-DOS'\n 0650F4 General Purpose Flag  0008 (8)\n@@ -1181,15 +1181,15 @@\n 065152 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 065154   Length              0005 (5)\n 065156   Flags               01 (1) 'Modification'\n 065157   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 06515B PAYLOAD\n \n 0C3604 DATA DESCRIPTOR       08074B50 (134695760)\n-0C3608 CRC                   2B643E30 (727989808)\n+0C3608 CRC                   9A85B6A1 (2592454305)\n 0C360C Compressed Size       0005E4A9 (386217)\n 0C3610 Uncompressed Size     0005E4A9 (386217)\n \n 0C3614 LOCAL HEADER #46      04034B50 (67324752)\n 0C3618 Extract Zip Spec      14 (20) '2.0'\n 0C3619 Extract OS            00 (0) 'MS-DOS'\n 0C361A General Purpose Flag  0008 (8)\n@@ -1209,15 +1209,15 @@\n 0C3670 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 0C3672   Length              0005 (5)\n 0C3674   Flags               01 (1) 'Modification'\n 0C3675   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 0C3679 PAYLOAD\n \n 1220A5 DATA DESCRIPTOR       08074B50 (134695760)\n-1220A9 CRC                   8B9B3621 (2342204961)\n+1220A9 CRC                   31237C9A (824409242)\n 1220AD Compressed Size       0005EA2C (387628)\n 1220B1 Uncompressed Size     0005EA2C (387628)\n \n 1220B5 LOCAL HEADER #47      04034B50 (67324752)\n 1220B9 Extract Zip Spec      14 (20) '2.0'\n 1220BA Extract OS            00 (0) 'MS-DOS'\n 1220BB General Purpose Flag  0008 (8)\n@@ -1265,15 +1265,15 @@\n 138403 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 138405   Length              0005 (5)\n 138407   Flags               01 (1) 'Modification'\n 138408   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 13840C PAYLOAD\n \n 3313FA DATA DESCRIPTOR       08074B50 (134695760)\n-3313FE CRC                   C1E420A7 (3252953255)\n+3313FE CRC                   41D8CD26 (1104727334)\n 331402 Compressed Size       001F8FEE (2068462)\n 331406 Uncompressed Size     001F8FEE (2068462)\n \n 33140A LOCAL HEADER #49      04034B50 (67324752)\n 33140E Extract Zip Spec      14 (20) '2.0'\n 33140F Extract OS            00 (0) 'MS-DOS'\n 331410 General Purpose Flag  0008 (8)\n@@ -1293,15 +1293,15 @@\n 331469 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 33146B   Length              0005 (5)\n 33146D   Flags               01 (1) 'Modification'\n 33146E   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 331472 PAYLOAD\n \n 3B8B11 DATA DESCRIPTOR       08074B50 (134695760)\n-3B8B15 CRC                   DF54657A (3746850170)\n+3B8B15 CRC                   D018E1C3 (3491291587)\n 3B8B19 Compressed Size       0008769F (554655)\n 3B8B1D Uncompressed Size     0008769F (554655)\n \n 3B8B21 LOCAL HEADER #50      04034B50 (67324752)\n 3B8B25 Extract Zip Spec      14 (20) '2.0'\n 3B8B26 Extract OS            00 (0) 'MS-DOS'\n 3B8B27 General Purpose Flag  0008 (8)\n@@ -1321,15 +1321,15 @@\n 3B8B7E Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 3B8B80   Length              0005 (5)\n 3B8B82   Flags               01 (1) 'Modification'\n 3B8B83   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 3B8B87 PAYLOAD\n \n 3DD6AD DATA DESCRIPTOR       08074B50 (134695760)\n-3DD6B1 CRC                   70E8ABA4 (1894296484)\n+3DD6B1 CRC                   63C92183 (1674125699)\n 3DD6B5 Compressed Size       00024B26 (150310)\n 3DD6B9 Uncompressed Size     00024B26 (150310)\n \n 3DD6BD LOCAL HEADER #51      04034B50 (67324752)\n 3DD6C1 Extract Zip Spec      14 (20) '2.0'\n 3DD6C2 Extract OS            00 (0) 'MS-DOS'\n 3DD6C3 General Purpose Flag  0008 (8)\n@@ -1349,15 +1349,15 @@\n 3DD71B Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 3DD71D   Length              0005 (5)\n 3DD71F   Flags               01 (1) 'Modification'\n 3DD720   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 3DD724 PAYLOAD\n \n 480DE9 DATA DESCRIPTOR       08074B50 (134695760)\n-480DED CRC                   A3BCBE1D (2747055645)\n+480DED CRC                   60BA145D (1622807645)\n 480DF1 Compressed Size       000A36C5 (669381)\n 480DF5 Uncompressed Size     000A36C5 (669381)\n \n 480DF9 LOCAL HEADER #52      04034B50 (67324752)\n 480DFD Extract Zip Spec      14 (20) '2.0'\n 480DFE Extract OS            00 (0) 'MS-DOS'\n 480DFF General Purpose Flag  0008 (8)\n@@ -1377,15 +1377,15 @@\n 480E58 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 480E5A   Length              0005 (5)\n 480E5C   Flags               01 (1) 'Modification'\n 480E5D   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 480E61 PAYLOAD\n \n 5DB1C1 DATA DESCRIPTOR       08074B50 (134695760)\n-5DB1C5 CRC                   D5CD5F15 (3587006229)\n+5DB1C5 CRC                   3B437530 (994276656)\n 5DB1C9 Compressed Size       0015A360 (1418080)\n 5DB1CD Uncompressed Size     0015A360 (1418080)\n \n 5DB1D1 LOCAL HEADER #53      04034B50 (67324752)\n 5DB1D5 Extract Zip Spec      14 (20) '2.0'\n 5DB1D6 Extract OS            00 (0) 'MS-DOS'\n 5DB1D7 General Purpose Flag  0008 (8)\n@@ -1405,15 +1405,15 @@\n 5DB235 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 5DB237   Length              0005 (5)\n 5DB239   Flags               01 (1) 'Modification'\n 5DB23A   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 5DB23E PAYLOAD\n \n 5E8F33 DATA DESCRIPTOR       08074B50 (134695760)\n-5E8F37 CRC                   2BD84363 (735593315)\n+5E8F37 CRC                   D332CB71 (3543321457)\n 5E8F3B Compressed Size       0000DCF5 (56565)\n 5E8F3F Uncompressed Size     0000DCF5 (56565)\n \n 5E8F43 LOCAL HEADER #54      04034B50 (67324752)\n 5E8F47 Extract Zip Spec      14 (20) '2.0'\n 5E8F48 Extract OS            00 (0) 'MS-DOS'\n 5E8F49 General Purpose Flag  0008 (8)\n@@ -1433,15 +1433,15 @@\n 5E8FA3 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 5E8FA5   Length              0005 (5)\n 5E8FA7   Flags               01 (1) 'Modification'\n 5E8FA8   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 5E8FAC PAYLOAD\n \n 62F2B6 DATA DESCRIPTOR       08074B50 (134695760)\n-62F2BA CRC                   C62697EE (3324418030)\n+62F2BA CRC                   D0AE4DFE (3501084158)\n 62F2BE Compressed Size       0004630A (287498)\n 62F2C2 Uncompressed Size     0004630A (287498)\n \n 62F2C6 LOCAL HEADER #55      04034B50 (67324752)\n 62F2CA Extract Zip Spec      14 (20) '2.0'\n 62F2CB Extract OS            00 (0) 'MS-DOS'\n 62F2CC General Purpose Flag  0008 (8)\n@@ -1461,15 +1461,15 @@\n 62F326 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 62F328   Length              0005 (5)\n 62F32A   Flags               01 (1) 'Modification'\n 62F32B   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 62F32F PAYLOAD\n \n 675AE8 DATA DESCRIPTOR       08074B50 (134695760)\n-675AEC CRC                   3BACA14D (1001169229)\n+675AEC CRC                   0325C463 (52806755)\n 675AF0 Compressed Size       000467B9 (288697)\n 675AF4 Uncompressed Size     000467B9 (288697)\n \n 675AF8 LOCAL HEADER #56      04034B50 (67324752)\n 675AFC Extract Zip Spec      14 (20) '2.0'\n 675AFD Extract OS            00 (0) 'MS-DOS'\n 675AFE General Purpose Flag  0008 (8)\n@@ -2884,15 +2884,15 @@\n 687504 Created OS            00 (0) 'MS-DOS'\n 687505 Extract Zip Spec      14 (20) '2.0'\n 687506 Extract OS            00 (0) 'MS-DOS'\n 687507 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687509 Compression Method    0000 (0) 'Stored'\n 68750B Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-68750F CRC                   95207A70 (2501933680)\n+68750F CRC                   FCD220DB (4241629403)\n 687513 Compressed Size       0000BA1C (47644)\n 687517 Uncompressed Size     0000BA1C (47644)\n 68751B Filename Length       002F (47)\n 68751D Extra Length          0009 (9)\n 68751F Comment Length        0000 (0)\n 687521 Disk Start            0000 (0)\n 687523 Int File Attributes   0000 (0)\n@@ -2944,15 +2944,15 @@\n 6875CC Created OS            00 (0) 'MS-DOS'\n 6875CD Extract Zip Spec      14 (20) '2.0'\n 6875CE Extract OS            00 (0) 'MS-DOS'\n 6875CF General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 6875D1 Compression Method    0000 (0) 'Stored'\n 6875D3 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-6875D7 CRC                   9536A20F (2503385615)\n+6875D7 CRC                   5C86D616 (1552340502)\n 6875DB Compressed Size       00007602 (30210)\n 6875DF Uncompressed Size     00007602 (30210)\n 6875E3 Filename Length       002E (46)\n 6875E5 Extra Length          0009 (9)\n 6875E7 Comment Length        0000 (0)\n 6875E9 Disk Start            0000 (0)\n 6875EB Int File Attributes   0000 (0)\n@@ -3034,15 +3034,15 @@\n 687703 Created OS            00 (0) 'MS-DOS'\n 687704 Extract Zip Spec      14 (20) '2.0'\n 687705 Extract OS            00 (0) 'MS-DOS'\n 687706 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687708 Compression Method    0000 (0) 'Stored'\n 68770A Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-68770E CRC                   384F6380 (944726912)\n+68770E CRC                   B3D3C0F4 (3016999156)\n 687712 Compressed Size       0000734E (29518)\n 687716 Uncompressed Size     0000734E (29518)\n 68771A Filename Length       0036 (54)\n 68771C Extra Length          0009 (9)\n 68771E Comment Length        0000 (0)\n 687720 Disk Start            0000 (0)\n 687722 Int File Attributes   0000 (0)\n@@ -3064,15 +3064,15 @@\n 687770 Created OS            00 (0) 'MS-DOS'\n 687771 Extract Zip Spec      14 (20) '2.0'\n 687772 Extract OS            00 (0) 'MS-DOS'\n 687773 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687775 Compression Method    0000 (0) 'Stored'\n 687777 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-68777B CRC                   2B643E30 (727989808)\n+68777B CRC                   9A85B6A1 (2592454305)\n 68777F Compressed Size       0005E4A9 (386217)\n 687783 Uncompressed Size     0005E4A9 (386217)\n 687787 Filename Length       0046 (70)\n 687789 Extra Length          0009 (9)\n 68778B Comment Length        0000 (0)\n 68778D Disk Start            0000 (0)\n 68778F Int File Attributes   0000 (0)\n@@ -3094,15 +3094,15 @@\n 6877ED Created OS            00 (0) 'MS-DOS'\n 6877EE Extract Zip Spec      14 (20) '2.0'\n 6877EF Extract OS            00 (0) 'MS-DOS'\n 6877F0 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 6877F2 Compression Method    0000 (0) 'Stored'\n 6877F4 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-6877F8 CRC                   8B9B3621 (2342204961)\n+6877F8 CRC                   31237C9A (824409242)\n 6877FC Compressed Size       0005EA2C (387628)\n 687800 Uncompressed Size     0005EA2C (387628)\n 687804 Filename Length       003E (62)\n 687806 Extra Length          0009 (9)\n 687808 Comment Length        0000 (0)\n 68780A Disk Start            0000 (0)\n 68780C Int File Attributes   0000 (0)\n@@ -3154,15 +3154,15 @@\n 6878C5 Created OS            00 (0) 'MS-DOS'\n 6878C6 Extract Zip Spec      14 (20) '2.0'\n 6878C7 Extract OS            00 (0) 'MS-DOS'\n 6878C8 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 6878CA Compression Method    0000 (0) 'Stored'\n 6878CC Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-6878D0 CRC                   C1E420A7 (3252953255)\n+6878D0 CRC                   41D8CD26 (1104727334)\n 6878D4 Compressed Size       001F8FEE (2068462)\n 6878D8 Uncompressed Size     001F8FEE (2068462)\n 6878DC Filename Length       002F (47)\n 6878DE Extra Length          0009 (9)\n 6878E0 Comment Length        0000 (0)\n 6878E2 Disk Start            0000 (0)\n 6878E4 Int File Attributes   0000 (0)\n@@ -3184,15 +3184,15 @@\n 68792B Created OS            00 (0) 'MS-DOS'\n 68792C Extract Zip Spec      14 (20) '2.0'\n 68792D Extract OS            00 (0) 'MS-DOS'\n 68792E General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687930 Compression Method    0000 (0) 'Stored'\n 687932 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687936 CRC                   DF54657A (3746850170)\n+687936 CRC                   D018E1C3 (3491291587)\n 68793A Compressed Size       0008769F (554655)\n 68793E Uncompressed Size     0008769F (554655)\n 687942 Filename Length       0041 (65)\n 687944 Extra Length          0009 (9)\n 687946 Comment Length        0000 (0)\n 687948 Disk Start            0000 (0)\n 68794A Int File Attributes   0000 (0)\n@@ -3214,15 +3214,15 @@\n 6879A3 Created OS            00 (0) 'MS-DOS'\n 6879A4 Extract Zip Spec      14 (20) '2.0'\n 6879A5 Extract OS            00 (0) 'MS-DOS'\n 6879A6 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 6879A8 Compression Method    0000 (0) 'Stored'\n 6879AA Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-6879AE CRC                   70E8ABA4 (1894296484)\n+6879AE CRC                   63C92183 (1674125699)\n 6879B2 Compressed Size       00024B26 (150310)\n 6879B6 Uncompressed Size     00024B26 (150310)\n 6879BA Filename Length       003F (63)\n 6879BC Extra Length          0009 (9)\n 6879BE Comment Length        0000 (0)\n 6879C0 Disk Start            0000 (0)\n 6879C2 Int File Attributes   0000 (0)\n@@ -3244,15 +3244,15 @@\n 687A19 Created OS            00 (0) 'MS-DOS'\n 687A1A Extract Zip Spec      14 (20) '2.0'\n 687A1B Extract OS            00 (0) 'MS-DOS'\n 687A1C General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687A1E Compression Method    0000 (0) 'Stored'\n 687A20 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687A24 CRC                   A3BCBE1D (2747055645)\n+687A24 CRC                   60BA145D (1622807645)\n 687A28 Compressed Size       000A36C5 (669381)\n 687A2C Uncompressed Size     000A36C5 (669381)\n 687A30 Filename Length       0040 (64)\n 687A32 Extra Length          0009 (9)\n 687A34 Comment Length        0000 (0)\n 687A36 Disk Start            0000 (0)\n 687A38 Int File Attributes   0000 (0)\n@@ -3274,15 +3274,15 @@\n 687A90 Created OS            00 (0) 'MS-DOS'\n 687A91 Extract Zip Spec      14 (20) '2.0'\n 687A92 Extract OS            00 (0) 'MS-DOS'\n 687A93 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687A95 Compression Method    0000 (0) 'Stored'\n 687A97 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687A9B CRC                   D5CD5F15 (3587006229)\n+687A9B CRC                   3B437530 (994276656)\n 687A9F Compressed Size       0015A360 (1418080)\n 687AA3 Uncompressed Size     0015A360 (1418080)\n 687AA7 Filename Length       0041 (65)\n 687AA9 Extra Length          0009 (9)\n 687AAB Comment Length        0000 (0)\n 687AAD Disk Start            0000 (0)\n 687AAF Int File Attributes   0000 (0)\n@@ -3304,15 +3304,15 @@\n 687B08 Created OS            00 (0) 'MS-DOS'\n 687B09 Extract Zip Spec      14 (20) '2.0'\n 687B0A Extract OS            00 (0) 'MS-DOS'\n 687B0B General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687B0D Compression Method    0000 (0) 'Stored'\n 687B0F Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687B13 CRC                   2BD84363 (735593315)\n+687B13 CRC                   D332CB71 (3543321457)\n 687B17 Compressed Size       0000DCF5 (56565)\n 687B1B Uncompressed Size     0000DCF5 (56565)\n 687B1F Filename Length       0046 (70)\n 687B21 Extra Length          0009 (9)\n 687B23 Comment Length        0000 (0)\n 687B25 Disk Start            0000 (0)\n 687B27 Int File Attributes   0000 (0)\n@@ -3334,15 +3334,15 @@\n 687B85 Created OS            00 (0) 'MS-DOS'\n 687B86 Extract Zip Spec      14 (20) '2.0'\n 687B87 Extract OS            00 (0) 'MS-DOS'\n 687B88 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687B8A Compression Method    0000 (0) 'Stored'\n 687B8C Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687B90 CRC                   C62697EE (3324418030)\n+687B90 CRC                   D0AE4DFE (3501084158)\n 687B94 Compressed Size       0004630A (287498)\n 687B98 Uncompressed Size     0004630A (287498)\n 687B9C Filename Length       0042 (66)\n 687B9E Extra Length          0009 (9)\n 687BA0 Comment Length        0000 (0)\n 687BA2 Disk Start            0000 (0)\n 687BA4 Int File Attributes   0000 (0)\n@@ -3364,15 +3364,15 @@\n 687BFE Created OS            00 (0) 'MS-DOS'\n 687BFF Extract Zip Spec      14 (20) '2.0'\n 687C00 Extract OS            00 (0) 'MS-DOS'\n 687C01 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687C03 Compression Method    0000 (0) 'Stored'\n 687C05 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687C09 CRC                   3BACA14D (1001169229)\n+687C09 CRC                   0325C463 (52806755)\n 687C0D Compressed Size       000467B9 (288697)\n 687C11 Uncompressed Size     000467B9 (288697)\n 687C15 Filename Length       0042 (66)\n 687C17 Extra Length          0009 (9)\n 687C19 Comment Length        0000 (0)\n 687C1B Disk Start            0000 (0)\n 687C1D Int File Attributes   0000 (0)\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HintParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HintParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HintParser.g 2024-09-26 10:28:54\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HintParser.g 2025-01-31 11:23:54\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.hadoop.conf.Configuration;\n import org.apache.hadoop.hive.conf.HiveConf;\n \n \n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexer.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexer.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexer.g 2024-09-26 10:28:49\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexer.g 2025-01-31 11:23:51\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexerStandard.g 2024-09-26 10:28:54\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexerStandard.g 2025-01-31 11:23:54\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard_HiveLexerParent.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard_HiveLexerParent.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 HiveLexerParent.g 2024-09-26 10:28:54\n+// $ANTLR 3.5.2 HiveLexerParent.g 2025-01-31 11:23:54\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexer_HiveLexerParent.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexer_HiveLexerParent.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 HiveLexerParent.g 2024-09-26 10:28:50\n+// $ANTLR 3.5.2 HiveLexerParent.g 2025-01-31 11:23:51\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveParser.g 2024-09-26 14:10:38\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveParser.g 2025-01-31 11:23:52\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -3110,15 +3110,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_inputFileFormat.add(inputFileFormat50.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tab, islocal, isoverwrite, inputFileFormat, path\n+\t\t\t// elements: inputFileFormat, islocal, path, isoverwrite, tab\n \t\t\t// token labels: islocal, path, isoverwrite\n \t\t\t// rule labels: tab, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3590,15 +3590,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableLocation.add(tableLocation62.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: path, ext, tableLocation, tab\n+\t\t\t// elements: tab, ext, tableLocation, path\n \t\t\t// token labels: ext, path\n \t\t\t// rule labels: tab, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3764,15 +3764,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replConfigs.add(replConf.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: oldDbPolicy, replConf, dbPolicy\n+\t\t\t// elements: dbPolicy, oldDbPolicy, replConf\n \t\t\t// token labels: \n \t\t\t// rule labels: oldDbPolicy, dbPolicy, retval, replConf\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3900,15 +3900,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replTableLevelPolicy.add(tablePolicy.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tablePolicy, dbName\n+\t\t\t// elements: dbName, tablePolicy\n \t\t\t// token labels: \n \t\t\t// rule labels: dbName, tablePolicy, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4055,15 +4055,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replConfigs.add(replConf.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: sourceDbPolicy, dbName, replConf\n+\t\t\t// elements: dbName, sourceDbPolicy, replConf\n \t\t\t// token labels: \n \t\t\t// rule labels: dbName, sourceDbPolicy, retval, replConf\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4407,15 +4407,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: replTablesExcludeList, replTablesIncludeList\n+\t\t\t// elements: replTablesIncludeList, replTablesExcludeList\n \t\t\t// token labels: replTablesExcludeList, replTablesIncludeList\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4541,15 +4541,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replConfigs.add(replConf.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: dbName, replConf\n+\t\t\t// elements: replConf, dbName\n \t\t\t// token labels: \n \t\t\t// rule labels: dbName, retval, replConf\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -6313,15 +6313,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_dbProperties.add(dbprops.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, dbprops, dbLocation, databaseComment, dbManagedLocation, ifNotExists\n+\t\t\t\t\t// elements: dbLocation, dbManagedLocation, dbprops, ifNotExists, name, databaseComment\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, dbprops, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -6497,15 +6497,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_dbProperties.add(dbprops.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, dbConnectorName, ifNotExists, databaseComment, dbprops\n+\t\t\t\t\t// elements: name, dbprops, dbConnectorName, ifNotExists, databaseComment\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, dbprops, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -7257,15 +7257,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_restrictOrCascade.add(restrictOrCascade178.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: ifExists, identifier, restrictOrCascade\n+\t\t\t// elements: restrictOrCascade, identifier, ifExists\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -7526,15 +7526,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_force.add(force187.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: force, columnNameList, tablePartitionPrefix\n+\t\t\t// elements: tablePartitionPrefix, columnNameList, force\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -7696,15 +7696,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replicationClause.add(replicationClause193.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, ifExists, KW_PURGE, replicationClause\n+\t\t\t// elements: KW_PURGE, replicationClause, tableName, ifExists\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -8661,15 +8661,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec206.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: partitionSpec, tabTypeExpr\n+\t\t\t// elements: tabTypeExpr, partitionSpec\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -8792,15 +8792,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_extColumnName.add(extColumnName209.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, extColumnName, partitionSpec\n+\t\t\t// elements: partitionSpec, extColumnName, tableName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -9370,15 +9370,15 @@\n \t\t\t\t\tdbName=identifier();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(dbName.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_EXTENDED, dbName\n+\t\t\t\t\t// elements: dbName, KW_EXTENDED\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: dbName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9446,15 +9446,15 @@\n \t\t\t\t\tdcName=identifier();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(dcName.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_EXTENDED, dcName\n+\t\t\t\t\t// elements: dcName, KW_EXTENDED\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: dcName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9517,15 +9517,15 @@\n \t\t\t\t\tname=descFuncNames();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_descFuncNames.add(name.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_EXTENDED, name\n+\t\t\t\t\t// elements: name, KW_EXTENDED\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9604,15 +9604,15 @@\n \t\t\t\t\tparttype=tabPartColTypeExpr();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tabPartColTypeExpr.add(parttype.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: parttype, descOptions\n+\t\t\t\t\t// elements: descOptions, parttype\n \t\t\t\t\t// token labels: descOptions\n \t\t\t\t\t// rule labels: parttype, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9854,15 +9854,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_COLUMNS, statsColumnName, noscan, parttype\n+\t\t\t\t\t// elements: parttype, KW_COLUMNS, noscan, statsColumnName\n \t\t\t\t\t// token labels: noscan\n \t\t\t\t\t// rule labels: statsColumnName, parttype, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10519,15 +10519,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showTablesFilterExpr.add(filter.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: db_name, filter, isExtended\n+\t\t\t\t\t// elements: filter, db_name, isExtended\n \t\t\t\t\t// token labels: isExtended\n \t\t\t\t\t// rule labels: filter, db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10672,15 +10672,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showStmtIdentifier.add(showStmtIdentifier242.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: showStmtIdentifier, db_name\n+\t\t\t\t\t// elements: db_name, showStmtIdentifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11025,15 +11025,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showStmtIdentifier.add(showStmtIdentifier261.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: showStmtIdentifier, tableName, db_name, KW_SORTED\n+\t\t\t\t\t// elements: showStmtIdentifier, db_name, tableName, KW_SORTED\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11242,15 +11242,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause271.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: orderByClause, limitClause, whereClause, tabName, partitionSpec\n+\t\t\t\t\t// elements: partitionSpec, tabName, whereClause, orderByClause, limitClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: tabName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11547,15 +11547,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec284.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: db_name, showStmtIdentifier, partitionSpec\n+\t\t\t\t\t// elements: showStmtIdentifier, db_name, partitionSpec\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11628,15 +11628,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: prptyName, tableName\n+\t\t\t\t\t// elements: tableName, prptyName\n \t\t\t\t\t// token labels: prptyName\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -12122,15 +12122,15 @@\n \n \t\t\t\t\t\t\t\t\t}\n \t\t\t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t\t\t}\n \n \t\t\t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t\t\t// elements: parttype, isExtended\n+\t\t\t\t\t\t\t// elements: isExtended, parttype\n \t\t\t\t\t\t\t// token labels: isExtended\n \t\t\t\t\t\t\t// rule labels: parttype, retval\n \t\t\t\t\t\t\t// token list labels: \n \t\t\t\t\t\t\t// rule list labels: \n \t\t\t\t\t\t\t// wildcard labels: \n \t\t\t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\t\t\tretval.tree = root_0;\n@@ -12371,15 +12371,15 @@\n \t\t\t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause303.getTree());\n \t\t\t\t\t\t\t\t\t}\n \t\t\t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t\t\t}\n \n \t\t\t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t\t\t// elements: limitClause, compactionType, dbName, compactionStatus, orderByClause, compactionPool\n+\t\t\t\t\t\t\t// elements: limitClause, compactionPool, compactionType, compactionStatus, orderByClause, dbName\n \t\t\t\t\t\t\t// token labels: \n \t\t\t\t\t\t\t// rule labels: dbName, retval\n \t\t\t\t\t\t\t// token list labels: \n \t\t\t\t\t\t\t// rule list labels: \n \t\t\t\t\t\t\t// wildcard labels: \n \t\t\t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\t\t\tretval.tree = root_0;\n@@ -12823,15 +12823,15 @@\n \t\t\t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause308.getTree());\n \t\t\t\t\t\t\t\t\t}\n \t\t\t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t\t\t}\n \n \t\t\t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t\t\t// elements: compactionType, orderByClause, parttype, compactionPool, compactionStatus, limitClause\n+\t\t\t\t\t\t\t// elements: parttype, orderByClause, limitClause, compactionPool, compactionType, compactionStatus\n \t\t\t\t\t\t\t// token labels: \n \t\t\t\t\t\t\t// rule labels: parttype, retval\n \t\t\t\t\t\t\t// token list labels: \n \t\t\t\t\t\t\t// rule list labels: \n \t\t\t\t\t\t\t// wildcard labels: \n \t\t\t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\t\t\tretval.tree = root_0;\n@@ -13475,15 +13475,15 @@\n \t\t\t\t\tEQUAL322=(Token)match(input,EQUAL,FOLLOW_EQUAL_in_showTablesFilterExpr6036); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_EQUAL.add(EQUAL322);\n \n \t\t\t\t\tStringLiteral323=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_showTablesFilterExpr6038); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(StringLiteral323);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: StringLiteral, identifier\n+\t\t\t\t\t// elements: identifier, StringLiteral\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -13651,15 +13651,15 @@\n \n \t\t\tpushFollow(FOLLOW_lockMode_in_lockStatement6108);\n \t\t\tlockMode331=lockMode();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_lockMode.add(lockMode331.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, lockMode, partitionSpec\n+\t\t\t// elements: lockMode, partitionSpec, tableName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14454,15 +14454,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_withGrantOption.add(withGrantOption352.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: withGrantOption, privilegeObject, principalSpecification, privList\n+\t\t\t// elements: privList, principalSpecification, privilegeObject, withGrantOption\n \t\t\t// token labels: \n \t\t\t// rule labels: privList, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14612,15 +14612,15 @@\n \n \t\t\tpushFollow(FOLLOW_principalSpecification_in_revokePrivileges6509);\n \t\t\tprincipalSpecification358=principalSpecification();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_principalSpecification.add(principalSpecification358.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: grantOptionFor, principalSpecification, privilegeList, privilegeObject\n+\t\t\t// elements: principalSpecification, privilegeObject, privilegeList, grantOptionFor\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14803,15 +14803,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_withAdminOption.add(withAdminOption366.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: principalSpecification, identifier, withAdminOption\n+\t\t\t// elements: principalSpecification, withAdminOption, identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14999,15 +14999,15 @@\n \n \t\t\tpushFollow(FOLLOW_principalSpecification_in_revokeRole6639);\n \t\t\tprincipalSpecification374=principalSpecification();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_principalSpecification.add(principalSpecification374.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: identifier, adminOptionFor, principalSpecification\n+\t\t\t// elements: adminOptionFor, identifier, principalSpecification\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -16511,15 +16511,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec403.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: partitionSpec, tableName\n+\t\t\t\t\t// elements: tableName, partitionSpec\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -17184,15 +17184,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec414.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: partitionSpec, cols, tableName\n+\t\t\t\t\t// elements: tableName, cols, partitionSpec\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: cols, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -17520,15 +17520,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: privilegeType, cols\n+\t\t\t// elements: cols, privilegeType\n \t\t\t// token labels: \n \t\t\t// rule labels: cols, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -18924,15 +18924,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, opt, repair, partitionSelectorSpec\n+\t\t\t// elements: opt, partitionSelectorSpec, repair, tableName\n \t\t\t// token labels: repair, opt\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -19496,15 +19496,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_resourceList.add(rList.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: functionIdentifier, StringLiteral, rList, rList, StringLiteral, functionIdentifier\n+\t\t\t// elements: StringLiteral, functionIdentifier, StringLiteral, rList, rList, functionIdentifier\n \t\t\t// token labels: \n \t\t\t// rule labels: rList, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -19946,15 +19946,15 @@\n \n \t\t\tpushFollow(FOLLOW_expression_in_createMacroStatement8373);\n \t\t\texpression485=expression();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_expression.add(expression485.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, Identifier, columnNameTypeList\n+\t\t\t// elements: Identifier, columnNameTypeList, expression\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -20331,15 +20331,15 @@\n \n \t\t\tpushFollow(FOLLOW_selectStatementWithCTE_in_createViewStatement8538);\n \t\t\tselectStatementWithCTE502=selectStatementWithCTE();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_selectStatementWithCTE.add(selectStatementWithCTE502.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: orReplace, columnNameCommentList, selectStatementWithCTE, ifNotExists, name, tableComment, viewPartition, tablePropertiesPrefixed\n+\t\t\t// elements: name, orReplace, viewPartition, columnNameCommentList, selectStatementWithCTE, tableComment, ifNotExists, tablePropertiesPrefixed\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -21506,15 +21506,15 @@\n \n \t\t\tpushFollow(FOLLOW_selectStatementWithCTE_in_createMaterializedViewStatement9066);\n \t\t\tselectStatementWithCTE547=selectStatementWithCTE();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_selectStatementWithCTE.add(selectStatementWithCTE547.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableComment, tableRowFormat, tableLocation, rewriteDisabled, viewPartition, ifNotExists, tableFileFormat, selectStatementWithCTE, name, tablePropertiesPrefixed, viewOrganization\n+\t\t\t// elements: tableRowFormat, tableFileFormat, tablePropertiesPrefixed, selectStatementWithCTE, tableComment, tableLocation, name, rewriteDisabled, viewPartition, viewOrganization, ifNotExists\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -21846,15 +21846,15 @@\n \n \t\t\tpushFollow(FOLLOW_definedAsSpec_in_createScheduledQueryStatement9331);\n \t\t\tdefinedAsSpec559=definedAsSpec();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_definedAsSpec.add(definedAsSpec559.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: definedAsSpec, name, executedAsSpec, scheduleSpec, enableSpecification\n+\t\t\t// elements: enableSpecification, executedAsSpec, definedAsSpec, scheduleSpec, name\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -22490,15 +22490,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: offsetTs, value, qualifier\n+\t\t\t\t\t// elements: offsetTs, qualifier, value\n \t\t\t\t\t// token labels: offsetTs, value\n \t\t\t\t\t// rule labels: qualifier, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -23236,15 +23236,15 @@\n \n \t\t\t\t\t}\n \n \t\t\t\t\tRPAREN589=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_createTablePartitionSpec10035); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN589);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: opt2, opt1\n+\t\t\t\t\t// elements: opt1, opt2\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: opt1, opt2, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -24516,15 +24516,15 @@\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_columnName.add(columnName625.getTree());\n \t\t\t\t\tRPAREN626=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_partitionTransformType10488); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN626);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: value, columnName\n+\t\t\t\t\t// elements: columnName, value\n \t\t\t\t\t// token labels: value\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -25024,15 +25024,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_storedAsDirs.add(storedAsDirs652.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: storedAsDirs, skewedCols, skewedValues\n+\t\t\t// elements: skewedCols, skewedValues, storedAsDirs\n \t\t\t// token labels: \n \t\t\t// rule labels: skewedCols, skewedValues, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -25653,15 +25653,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableProperties.add(serdeprops.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: serdeprops, name\n+\t\t\t// elements: name, serdeprops\n \t\t\t// token labels: name\n \t\t\t// rule labels: serdeprops, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -25860,15 +25860,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableRowNullFormat.add(tableRowNullFormat671.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableRowNullFormat, tableRowFormatMapKeysIdentifier, tableRowFormatFieldIdentifier, tableRowFormatCollItemsIdentifier, tableRowFormatLinesIdentifier\n+\t\t\t// elements: tableRowNullFormat, tableRowFormatLinesIdentifier, tableRowFormatMapKeysIdentifier, tableRowFormatCollItemsIdentifier, tableRowFormatFieldIdentifier\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -26546,15 +26546,15 @@\n \t\t\tEQUAL685=(Token)match(input,EQUAL,FOLLOW_EQUAL_in_keyValueProperty11357); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_EQUAL.add(EQUAL685);\n \n \t\t\tvalue=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_keyValueProperty11361); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(value);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: key, value\n+\t\t\t// elements: value, key\n \t\t\t// token labels: value, key\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -26760,15 +26760,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: fldEscape, fldIdnt\n+\t\t\t// elements: fldIdnt, fldEscape\n \t\t\t// token labels: fldIdnt, fldEscape\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -27446,15 +27446,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: inFmt, outDriver, outFmt, inDriver\n+\t\t\t\t\t// elements: inDriver, inFmt, outFmt, outDriver\n \t\t\t\t\t// token labels: inFmt, inDriver, outDriver, outFmt\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -27557,15 +27557,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(fileformat.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: serdeprops, fileformat, storageHandler\n+\t\t\t\t\t// elements: storageHandler, fileformat, serdeprops\n \t\t\t\t\t// token labels: storageHandler\n \t\t\t\t\t// rule labels: serdeprops, fileformat, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -27675,25 +27675,25 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(fileformat.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: serdeprops, genericSpec, fileformat\n+\t\t\t\t\t// elements: serdeprops, fileformat, genericSpec\n \t\t\t\t\t// token labels: \n-\t\t\t\t\t// rule labels: serdeprops, genericSpec, fileformat, retval\n+\t\t\t\t\t// rule labels: serdeprops, fileformat, genericSpec, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_serdeprops=new RewriteRuleSubtreeStream(adaptor,\"rule serdeprops\",serdeprops!=null?serdeprops.getTree():null);\n-\t\t\t\t\tRewriteRuleSubtreeStream stream_genericSpec=new RewriteRuleSubtreeStream(adaptor,\"rule genericSpec\",genericSpec!=null?genericSpec.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_fileformat=new RewriteRuleSubtreeStream(adaptor,\"rule fileformat\",fileformat!=null?fileformat.getTree():null);\n+\t\t\t\t\tRewriteRuleSubtreeStream stream_genericSpec=new RewriteRuleSubtreeStream(adaptor,\"rule genericSpec\",genericSpec!=null?genericSpec.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t\t\t// 2086:7: -> ^( TOK_STORAGEHANDLER $genericSpec ( $serdeprops)? ( ^( TOK_FILEFORMAT_GENERIC $fileformat) )? )\n \t\t\t\t\t{\n \t\t\t\t\t\t// org/apache/hadoop/hive/ql/parse/HiveParser.g:2086:10: ^( TOK_STORAGEHANDLER $genericSpec ( $serdeprops)? ( ^( TOK_FILEFORMAT_GENERIC $fileformat) )? )\n \t\t\t\t\t\t{\n@@ -29926,15 +29926,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsAlter.add(constraintOptsAlter768.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintName, constraintOptsAlter\n+\t\t\t// elements: constraintOptsAlter, constraintName\n \t\t\t// token labels: \n \t\t\t// rule labels: constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -30379,24 +30379,24 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate780.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: fkCols, tabName, constraintOptsCreate, tabName, constraintOptsCreate, constraintName, parCols, parCols, fkCols\n+\t\t\t// elements: fkCols, fkCols, parCols, tabName, parCols, constraintName, constraintOptsCreate, constraintOptsCreate, tabName\n \t\t\t// token labels: \n-\t\t\t// rule labels: tabName, parCols, fkCols, constraintName, retval\n+\t\t\t// rule labels: parCols, tabName, fkCols, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n-\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_parCols=new RewriteRuleSubtreeStream(adaptor,\"rule parCols\",parCols!=null?parCols.getTree():null);\n+\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_fkCols=new RewriteRuleSubtreeStream(adaptor,\"rule fkCols\",fkCols!=null?fkCols.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_constraintName=new RewriteRuleSubtreeStream(adaptor,\"rule constraintName\",constraintName!=null?constraintName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t// 2222:5: -> {$constraintName.tree != null}? ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) $fkCols $tabName $parCols ( constraintOptsCreate )? )\n \t\t\tif ((constraintName!=null?((ASTNode)constraintName.getTree()):null) != null) {\n@@ -30566,15 +30566,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsAlter.add(constraintOptsAlter785.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintOptsAlter, fkCols, parCols, constraintName, tabName\n+\t\t\t// elements: parCols, constraintName, fkCols, tabName, constraintOptsAlter\n \t\t\t// token labels: \n \t\t\t// rule labels: parCols, tabName, fkCols, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -31513,15 +31513,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_nullOrdering.add(nullSpec.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: identifier, nullSpec, identifier, identifier, nullSpec, identifier, nullSpec, identifier, identifier\n+\t\t\t// elements: identifier, nullSpec, identifier, nullSpec, identifier, identifier, nullSpec, identifier, identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: nullSpec, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -31845,15 +31845,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colName, comment\n+\t\t\t// elements: comment, colName\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -32116,15 +32116,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_nullOrdering.add(nullSpec.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: orderSpec, expression, expression, expression, nullSpec, expression, orderSpec, orderSpec, expression, orderSpec, expression, nullSpec, orderSpec, expression, expression\n+\t\t\t// elements: expression, expression, expression, orderSpec, orderSpec, nullSpec, expression, expression, expression, expression, orderSpec, orderSpec, expression, nullSpec, orderSpec\n \t\t\t// token labels: \n \t\t\t// rule labels: orderSpec, nullSpec, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -32373,15 +32373,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colType, colName, comment, colType, colName\n+\t\t\t// elements: colName, colType, comment, colName, colType\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -33073,15 +33073,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: columnConstraint, colName, colType, comment\n+\t\t\t// elements: comment, columnConstraint, colName, colType\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -33425,24 +33425,24 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate827.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintOptsCreate, tabName, colName, constraintOptsCreate, tabName, colName, constraintName\n+\t\t\t// elements: constraintOptsCreate, colName, tabName, tabName, constraintName, colName, constraintOptsCreate\n \t\t\t// token labels: \n-\t\t\t// rule labels: tabName, colName, constraintName, retval\n+\t\t\t// rule labels: colName, tabName, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n-\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_colName=new RewriteRuleSubtreeStream(adaptor,\"rule colName\",colName!=null?colName.getTree():null);\n+\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_constraintName=new RewriteRuleSubtreeStream(adaptor,\"rule constraintName\",constraintName!=null?constraintName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t// 2397:5: -> {$constraintName.tree != null}? ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) ^( TOK_TABCOLNAME ) $tabName ^( TOK_TABCOLNAME $colName) ( constraintOptsCreate )? )\n \t\t\tif ((constraintName!=null?((ASTNode)constraintName.getTree()):null) != null) {\n \t\t\t\t// org/apache/hadoop/hive/ql/parse/HiveParser.g:2398:13: ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) ^( TOK_TABCOLNAME ) $tabName ^( TOK_TABCOLNAME $colName) ( constraintOptsCreate )? )\n@@ -33623,15 +33623,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate830.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintName, constraintOptsCreate, constraintOptsCreate\n+\t\t\t// elements: constraintOptsCreate, constraintOptsCreate, constraintName\n \t\t\t// token labels: \n \t\t\t// rule labels: constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -33986,15 +33986,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsAlter.add(constraintOptsAlter837.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintOptsAlter, constraintOptsAlter, constraintName, tabName, colName, colName, tabName\n+\t\t\t// elements: constraintOptsAlter, tabName, colName, tabName, constraintName, colName, constraintOptsAlter\n \t\t\t// token labels: \n \t\t\t// rule labels: tabName, colName, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -35272,15 +35272,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: comment, colName, colName, colType, colType\n+\t\t\t// elements: comment, colType, colName, colName, colType\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -36844,15 +36844,15 @@\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_type.add(right.getTree());\n \t\t\tGREATERTHAN910=(Token)match(input,GREATERTHAN,FOLLOW_GREATERTHAN_in_mapType15887); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_GREATERTHAN.add(GREATERTHAN910);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: left, right\n+\t\t\t// elements: right, left\n \t\t\t// token labels: \n \t\t\t// rule labels: left, right, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -37951,15 +37951,15 @@\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_queryStatementExpression.add(queryStatementExpression943.getTree());\n \t\t\tRPAREN944=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_cteStatement16242); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN944);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colAliases, queryStatementExpression, identifier\n+\t\t\t// elements: queryStatementExpression, identifier, colAliases\n \t\t\t// token labels: \n \t\t\t// rule labels: colAliases, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -38090,15 +38090,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_setOperator.add(u.getTree());\n \t\t\t\t\tpushFollow(FOLLOW_singleFromStatement_in_fromStatement16285);\n \t\t\t\t\tr=singleFromStatement();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_singleFromStatement.add(r.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: r, u\n+\t\t\t\t\t// elements: u, r\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: r, u, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -38672,27 +38672,27 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_qualifyClause.add(q.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: s, w, win, q, f, h, g\n+\t\t\t\t\t// elements: s, q, g, h, f, win, w\n \t\t\t\t\t// token labels: \n-\t\t\t\t\t// rule labels: q, s, f, w, g, h, win, retval\n+\t\t\t\t\t// rule labels: q, s, f, g, w, h, win, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_q=new RewriteRuleSubtreeStream(adaptor,\"rule q\",q!=null?q.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_s=new RewriteRuleSubtreeStream(adaptor,\"rule s\",s!=null?s.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_f=new RewriteRuleSubtreeStream(adaptor,\"rule f\",f!=null?f.getTree():null);\n-\t\t\t\t\tRewriteRuleSubtreeStream stream_w=new RewriteRuleSubtreeStream(adaptor,\"rule w\",w!=null?w.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_g=new RewriteRuleSubtreeStream(adaptor,\"rule g\",g!=null?g.getTree():null);\n+\t\t\t\t\tRewriteRuleSubtreeStream stream_w=new RewriteRuleSubtreeStream(adaptor,\"rule w\",w!=null?w.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_h=new RewriteRuleSubtreeStream(adaptor,\"rule h\",h!=null?h.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_win=new RewriteRuleSubtreeStream(adaptor,\"rule win\",win!=null?win.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t\t\t// 2640:4: -> ^( TOK_QUERY ( $f)? ^( TOK_INSERT ^( TOK_DESTINATION ^( TOK_DIR TOK_TMP_FILE ) ) $s ( $w)? ( $g)? ( $h)? ( $win)? ( $q)? ) )\n \t\t\t\t\t{\n@@ -38989,15 +38989,15 @@\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((c!=null?((ASTNode)c.getTree()):null));\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((d!=null?((ASTNode)d.getTree()):null));\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((sort!=null?((ASTNode)sort.getTree()):null));\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((l!=null?((ASTNode)l.getTree()):null));\n \t\t\t   }\n \t\t\t   }\n \t\t\t// AST REWRITE\n-\t\t\t// elements: d, c, o, sort, l\n+\t\t\t// elements: d, l, sort, o, c\n \t\t\t// token labels: \n \t\t\t// rule labels: c, d, sort, l, retval, o\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -39186,15 +39186,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_setOperator.add(u.getTree());\n \t\t\t\t\tpushFollow(FOLLOW_atomSelectStatement_in_setOpSelectStatement17080);\n \t\t\t\t\tb=atomSelectStatement();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_atomSelectStatement.add(b.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: b, b, u, u, b, b\n+\t\t\t\t\t// elements: b, b, u, b, b, u\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: b, u, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -39904,15 +39904,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause965.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: qualifyClause, distributeByClause, selectClause, insertClause, window_clause, whereClause, groupByClause, havingClause, lateralView, sortByClause, orderByClause, limitClause, clusterByClause\n+\t\t\t\t\t// elements: whereClause, havingClause, insertClause, clusterByClause, window_clause, orderByClause, sortByClause, groupByClause, qualifyClause, limitClause, selectClause, distributeByClause, lateralView\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40229,15 +40229,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause977.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: sortByClause, clusterByClause, limitClause, groupByClause, orderByClause, lateralView, selectClause, whereClause, distributeByClause, havingClause, qualifyClause, window_clause\n+\t\t\t\t\t// elements: window_clause, limitClause, clusterByClause, selectClause, lateralView, sortByClause, distributeByClause, qualifyClause, havingClause, whereClause, groupByClause, orderByClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40479,15 +40479,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_ifNotExists.add(ifNotExists981.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ifNotExists, destination\n+\t\t\t\t\t// elements: destination, ifNotExists\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40759,15 +40759,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_tableFileFormat.add(tableFileFormat991.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableFileFormat, tableRowFormat, local, StringLiteral\n+\t\t\t\t\t// elements: StringLiteral, tableFileFormat, tableRowFormat, local\n \t\t\t\t\t// token labels: local\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40993,15 +40993,15 @@\n \n \t\t\t\t\tnum=(Token)match(input,Number,FOLLOW_Number_in_limitClause18297); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_Number.add(num);\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: num, offset\n+\t\t\t\t\t// elements: offset, num\n \t\t\t\t\t// token labels: offset, num\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -41046,15 +41046,15 @@\n \t\t\t\t\tKW_OFFSET997=(Token)match(input,KW_OFFSET,FOLLOW_KW_OFFSET_in_limitClause18326); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_OFFSET.add(KW_OFFSET997);\n \n \t\t\t\t\toffset=(Token)match(input,Number,FOLLOW_Number_in_limitClause18330); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_Number.add(offset);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: num, offset\n+\t\t\t\t\t// elements: offset, num\n \t\t\t\t\t// token labels: offset, num\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -41609,15 +41609,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_whereClause.add(whereClause1014.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: setColumnsClause, tableName, whereClause\n+\t\t\t// elements: whereClause, tableName, setColumnsClause\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -43035,15 +43035,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_expression.add(expression1056.getTree());\n \t\t\tpushFollow(FOLLOW_whenClauses_in_mergeStatement18982);\n \t\t\twhenClauses1057=whenClauses();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_whenClauses.add(whenClauses1057.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, QUERY_HINT, identifier, joinSourcePart, whenClauses, expression\n+\t\t\t// elements: whenClauses, QUERY_HINT, tableName, identifier, expression, joinSourcePart\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -43344,15 +43344,15 @@\n \n \t\t\tpushFollow(FOLLOW_valueRowConstructor_in_whenNotMatchedClause19090);\n \t\t\tvalueRowConstructor1069=valueRowConstructor();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_valueRowConstructor.add(valueRowConstructor1069.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: targetCols, valueRowConstructor, expression\n+\t\t\t// elements: targetCols, expression, valueRowConstructor\n \t\t\t// token labels: \n \t\t\t// rule labels: targetCols, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -43476,15 +43476,15 @@\n \n \t\t\tpushFollow(FOLLOW_updateOrDelete_in_whenMatchedAndClause19147);\n \t\t\tupdateOrDelete1075=updateOrDelete();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_updateOrDelete.add(updateOrDelete1075.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: updateOrDelete, expression\n+\t\t\t// elements: expression, updateOrDelete\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -44586,551 +44586,551 @@\n \t\t}\n \t\t}\n \n \t}\n \t// $ANTLR end synpred24_HiveParser\n \n \t// Delegated rules\n-\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation_return alterTblPartitionStatementSuffixSkewedLocation() throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation(); }\n+\tpublic HiveParser_IdentifiersParser.booleanValueTok_return booleanValueTok() throws RecognitionException { return gIdentifiersParser.booleanValueTok(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementPartitionKeyType_return alterStatementPartitionKeyType() throws RecognitionException { return gAlterClauseParser.alterStatementPartitionKeyType(); }\n+\tpublic HiveParser_IdentifiersParser.havingCondition_return havingCondition() throws RecognitionException { return gIdentifiersParser.havingCondition(); }\n \n-\tpublic HiveParser_IdentifiersParser.subQueryExpression_return subQueryExpression() throws RecognitionException { return gIdentifiersParser.subQueryExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUnArchive_return alterStatementSuffixUnArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUnArchive(); }\n \n-\tpublic HiveParser_IdentifiersParser.identifier_return identifier() throws RecognitionException { return gIdentifiersParser.identifier(); }\n+\tpublic HiveParser_ResourcePlanParser.disable_return disable() throws RecognitionException { return gResourcePlanParser.disable(); }\n \n-\tpublic HiveParser_IdentifiersParser.null_treatment_return null_treatment() throws RecognitionException { return gIdentifiersParser.null_treatment(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixMergeFiles_return alterStatementSuffixMergeFiles(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixMergeFiles(partition); }\n \n-\tpublic HiveParser_AlterClauseParser.alterViewSuffixProperties_return alterViewSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterViewSuffixProperties(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionAtom_return precedenceSimilarExpressionAtom(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionAtom(t); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddCol_return alterStatementSuffixAddCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddCol(); }\n+\tpublic HiveParser_FromClauseParser.joinToken_return joinToken() throws RecognitionException { return gFromClauseParser.joinToken(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropConstraint_return alterStatementSuffixDropConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropConstraint(); }\n+\tpublic HiveParser_IdentifiersParser.qualifyClause_return qualifyClause() throws RecognitionException { return gIdentifiersParser.qualifyClause(); }\n \n-\tpublic HiveParser_PrepareStatementParser.prepareStatement_return prepareStatement() throws RecognitionException { return gPrepareStatementParser.prepareStatement(); }\n+\tpublic HiveParser_SelectClauseParser.selectClause_return selectClause() throws RecognitionException { return gSelectClauseParser.selectClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenamePart_return alterStatementSuffixRenamePart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenamePart(); }\n+\tpublic HiveParser_IdentifiersParser.timeUnitQualifiers_return timeUnitQualifiers() throws RecognitionException { return gIdentifiersParser.timeUnitQualifiers(); }\n+\n+\tpublic HiveParser_SelectClauseParser.window_frame_start_boundary_return window_frame_start_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_start_boundary(); }\n+\n+\tpublic HiveParser_ResourcePlanParser.triggerOrExpression_return triggerOrExpression() throws RecognitionException { return gResourcePlanParser.triggerOrExpression(); }\n+\n+\tpublic HiveParser_FromClauseParser.valuesTableConstructor_return valuesTableConstructor() throws RecognitionException { return gFromClauseParser.valuesTableConstructor(); }\n+\n+\tpublic HiveParser_IdentifiersParser.expressionsNotInParenthesis_return expressionsNotInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsNotInParenthesis(isStruct, forceStruct); }\n \n \tpublic HiveParser_ResourcePlanParser.alterResourcePlanStatement_return alterResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.alterResourcePlanStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.aliasList_return aliasList() throws RecognitionException { return gFromClauseParser.aliasList(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateColumns_return alterStatementSuffixUpdateColumns() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateColumns(); }\n \n \tpublic HiveParser_FromClauseParser.searchCondition_return searchCondition() throws RecognitionException { return gFromClauseParser.searchCondition(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitions_return alterStatementSuffixAddPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitions(table); }\n+\tpublic HiveParser_IdentifiersParser.precedenceDistinctOperator_return precedenceDistinctOperator() throws RecognitionException { return gIdentifiersParser.precedenceDistinctOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.skewedLocationsList_return skewedLocationsList() throws RecognitionException { return gAlterClauseParser.skewedLocationsList(); }\n+\tpublic HiveParser_ResourcePlanParser.alterTriggerStatement_return alterTriggerStatement() throws RecognitionException { return gResourcePlanParser.alterTriggerStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.rollupOldSyntax_return rollupOldSyntax() throws RecognitionException { return gIdentifiersParser.rollupOldSyntax(); }\n+\tpublic HiveParser_IdentifiersParser.sortByClause_return sortByClause() throws RecognitionException { return gIdentifiersParser.sortByClause(); }\n \n-\tpublic HiveParser_CreateDDLParser.dropDataConnectorStatement_return dropDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.dropDataConnectorStatement(); }\n+\tpublic HiveParser_AlterClauseParser.skewedLocationMap_return skewedLocationMap() throws RecognitionException { return gAlterClauseParser.skewedLocationMap(); }\n \n-\tpublic HiveParser_CreateDDLParser.createDataConnectorStatement_return createDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.createDataConnectorStatement(); }\n+\tpublic HiveParser_SelectClauseParser.window_clause_return window_clause() throws RecognitionException { return gSelectClauseParser.window_clause(); }\n \n-\tpublic HiveParser_IdentifiersParser.nonReserved_return nonReserved() throws RecognitionException { return gIdentifiersParser.nonReserved(); }\n+\tpublic HiveParser_FromClauseParser.joinSourcePart_return joinSourcePart() throws RecognitionException { return gFromClauseParser.joinSourcePart(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupingSetExpressionMultiple_return groupingSetExpressionMultiple() throws RecognitionException { return gIdentifiersParser.groupingSetExpressionMultiple(); }\n+\tpublic HiveParser_IdentifiersParser.floorExpression_return floorExpression() throws RecognitionException { return gIdentifiersParser.floorExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetLocation_return alterDatabaseSuffixSetLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetLocation(); }\n+\tpublic HiveParser_SelectClauseParser.window_range_expression_return window_range_expression() throws RecognitionException { return gSelectClauseParser.window_range_expression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropMappingStatement_return dropMappingStatement() throws RecognitionException { return gResourcePlanParser.dropMappingStatement(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixExchangePartition_return alterStatementSuffixExchangePartition() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExchangePartition(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixConvert_return alterStatementSuffixConvert() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixConvert(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddConstraint_return alterStatementSuffixAddConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddConstraint(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetManagedLocation_return alterDatabaseSuffixSetManagedLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetManagedLocation(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetLocation_return alterDatabaseSuffixSetLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetLocation(); }\n \n-\tpublic HiveParser_FromClauseParser.valuesTableConstructor_return valuesTableConstructor() throws RecognitionException { return gFromClauseParser.valuesTableConstructor(); }\n+\tpublic HiveParser_AlterClauseParser.alterTableStatementSuffix_return alterTableStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterTableStatementSuffix(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropResourcePlanStatement_return dropResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.dropResourcePlanStatement(); }\n+\tpublic HiveParser_AlterClauseParser.refRetain_return refRetain() throws RecognitionException { return gAlterClauseParser.refRetain(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAmpersandOperator_return precedenceAmpersandOperator() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandOperator(); }\n+\tpublic HiveParser_ResourcePlanParser.activate_return activate() throws RecognitionException { return gResourcePlanParser.activate(); }\n \n-\tpublic HiveParser_FromClauseParser.subQuerySource_return subQuerySource() throws RecognitionException { return gFromClauseParser.subQuerySource(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorExpression_return precedenceBitwiseXorExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseStatementSuffix_return alterDatabaseStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDatabaseStatementSuffix(); }\n+\tpublic HiveParser_ResourcePlanParser.resourcePlanDdlStatements_return resourcePlanDdlStatements() throws RecognitionException { return gResourcePlanParser.resourcePlanDdlStatements(); }\n \n-\tpublic HiveParser_FromClauseParser.virtualTableSource_return virtualTableSource() throws RecognitionException { return gFromClauseParser.virtualTableSource(); }\n+\tpublic HiveParser_AlterClauseParser.snapshotIdOfRef_return snapshotIdOfRef() throws RecognitionException { return gAlterClauseParser.snapshotIdOfRef(); }\n \n-\tpublic HiveParser_ResourcePlanParser.globalWmStatement_return globalWmStatement() throws RecognitionException { return gResourcePlanParser.globalWmStatement(); }\n+\tpublic HiveParser_FromClauseParser.valuesClause_return valuesClause() throws RecognitionException { return gFromClauseParser.valuesClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPart_return precedenceSimilarExpressionPart(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPart(t); }\n+\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRewrite_return alterMaterializedViewSuffixRewrite(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRewrite(tableNameTree); }\n \n-\tpublic HiveParser_ResourcePlanParser.createMappingStatement_return createMappingStatement() throws RecognitionException { return gResourcePlanParser.createMappingStatement(); }\n+\tpublic HiveParser_IdentifiersParser.null_treatment_return null_treatment() throws RecognitionException { return gIdentifiersParser.null_treatment(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpAssign_return rpAssign() throws RecognitionException { return gResourcePlanParser.rpAssign(); }\n+\tpublic HiveParser_IdentifiersParser.precedencePlusExpression_return precedencePlusExpression() throws RecognitionException { return gIdentifiersParser.precedencePlusExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName_return sql11ReservedKeywordsUsedAsFunctionName() throws RecognitionException { return gIdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCompact_return alterStatementSuffixCompact() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCompact(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_frame_start_boundary_return window_frame_start_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_start_boundary(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixClusterbySortby_return alterStatementSuffixClusterbySortby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixClusterbySortby(); }\n \n-\tpublic HiveParser_FromClauseParser.valuesClause_return valuesClause() throws RecognitionException { return gFromClauseParser.valuesClause(); }\n+\tpublic HiveParser_FromClauseParser.tableOrColumn_return tableOrColumn() throws RecognitionException { return gFromClauseParser.tableOrColumn(); }\n \n-\tpublic HiveParser_IdentifiersParser.timestampLiteral_return timestampLiteral() throws RecognitionException { return gIdentifiersParser.timestampLiteral(); }\n+\tpublic HiveParser_FromClauseParser.partitioningSpec_return partitioningSpec() throws RecognitionException { return gFromClauseParser.partitioningSpec(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSelectorVal_return partitionSelectorVal() throws RecognitionException { return gIdentifiersParser.partitionSelectorVal(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerActionExpressionStandalone_return triggerActionExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerActionExpressionStandalone(); }\n \n-\tpublic HiveParser_CreateDDLParser.dcProperties_return dcProperties() throws RecognitionException { return gCreateDDLParser.dcProperties(); }\n+\tpublic HiveParser_PrepareStatementParser.executeParamList_return executeParamList() throws RecognitionException { return gPrepareStatementParser.executeParamList(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSelectorSpec_return partitionSelectorSpec() throws RecognitionException { return gIdentifiersParser.partitionSelectorSpec(); }\n+\tpublic HiveParser_ResourcePlanParser.withReplace_return withReplace() throws RecognitionException { return gResourcePlanParser.withReplace(); }\n+\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrOperator_return precedenceBitwiseOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrOperator(); }\n \n \tpublic HiveParser_AlterClauseParser.compactPool_return compactPool() throws RecognitionException { return gAlterClauseParser.compactPool(); }\n \n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixProperties_return alterDataConnectorSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixProperties(); }\n+\n \tpublic HiveParser_IdentifiersParser.caseExpression_return caseExpression() throws RecognitionException { return gIdentifiersParser.caseExpression(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_clause_return window_clause() throws RecognitionException { return gSelectClauseParser.window_clause(); }\n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorStatementSuffix_return alterDataConnectorStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDataConnectorStatementSuffix(); }\n \n-\tpublic HiveParser_IdentifiersParser.parameterIdx_return parameterIdx() throws RecognitionException { return gIdentifiersParser.parameterIdx(); }\n+\tpublic HiveParser_SelectClauseParser.window_value_expression_return window_value_expression() throws RecognitionException { return gSelectClauseParser.window_value_expression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrExpression_return precedenceBitwiseOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAmpersandOperator_return precedenceAmpersandOperator() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.columnRefOrderNotInParenthesis_return columnRefOrderNotInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderNotInParenthesis(); }\n+\tpublic HiveParser_AlterClauseParser.alterViewStatementSuffix_return alterViewStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterViewStatementSuffix(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceUnaryOperator_return precedenceUnaryOperator() throws RecognitionException { return gIdentifiersParser.precedenceUnaryOperator(); }\n+\tpublic HiveParser_PrepareStatementParser.executeStatement_return executeStatement() throws RecognitionException { return gPrepareStatementParser.executeStatement(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectItem_return selectItem() throws RecognitionException { return gSelectClauseParser.selectItem(); }\n+\tpublic HiveParser_CreateDDLParser.dropDataConnectorStatement_return dropDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.dropDataConnectorStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionsNotInParenthesis_return expressionsNotInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsNotInParenthesis(isStruct, forceStruct); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixArchive_return alterStatementSuffixArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixArchive(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetOwner_return alterStatementSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetOwner(); }\n+\tpublic HiveParser_ResourcePlanParser.dropMappingStatement_return dropMappingStatement() throws RecognitionException { return gResourcePlanParser.dropMappingStatement(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStats_return alterStatementSuffixUpdateStats(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStats(partition); }\n+\tpublic HiveParser_IdentifiersParser.groupingSetExpressionMultiple_return groupingSetExpressionMultiple() throws RecognitionException { return gIdentifiersParser.groupingSetExpressionMultiple(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalLiteral_return intervalLiteral() throws RecognitionException { return gIdentifiersParser.intervalLiteral(); }\n+\tpublic HiveParser_AlterClauseParser.fileFormat_return fileFormat() throws RecognitionException { return gAlterClauseParser.fileFormat(); }\n \n-\tpublic HiveParser_ResourcePlanParser.withReplace_return withReplace() throws RecognitionException { return gResourcePlanParser.withReplace(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceUnaryOperator_return precedenceUnaryOperator() throws RecognitionException { return gIdentifiersParser.precedenceUnaryOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixProperties_return alterDataConnectorSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixProperties(); }\n+\tpublic HiveParser_IdentifiersParser.expression_return expression() throws RecognitionException { return gIdentifiersParser.expression(); }\n \n-\tpublic HiveParser_FromClauseParser.viewName_return viewName() throws RecognitionException { return gFromClauseParser.viewName(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStatsCol_return alterStatementSuffixUpdateStatsCol(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStatsCol(partition); }\n \n-\tpublic HiveParser_SelectClauseParser.selectClause_return selectClause() throws RecognitionException { return gSelectClauseParser.selectClause(); }\n+\tpublic HiveParser_FromClauseParser.tableAllColumns_return tableAllColumns() throws RecognitionException { return gFromClauseParser.tableAllColumns(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterMaterializedViewStatementSuffix_return alterMaterializedViewStatementSuffix(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewStatementSuffix(tableNameTree); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixConvert_return alterStatementSuffixConvert() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixConvert(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_frame_return window_frame() throws RecognitionException { return gSelectClauseParser.window_frame(); }\n+\tpublic HiveParser_ResourcePlanParser.rpAssignList_return rpAssignList() throws RecognitionException { return gResourcePlanParser.rpAssignList(); }\n \n-\tpublic HiveParser_AlterClauseParser.blocking_return blocking() throws RecognitionException { return gAlterClauseParser.blocking(); }\n+\tpublic HiveParser_IdentifiersParser.charSetStringLiteral_return charSetStringLiteral() throws RecognitionException { return gIdentifiersParser.charSetStringLiteral(); }\n \n-\tpublic HiveParser_PrepareStatementParser.executeParamList_return executeParamList() throws RecognitionException { return gPrepareStatementParser.executeParamList(); }\n+\tpublic HiveParser_ResourcePlanParser.replaceResourcePlanStatement_return replaceResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.replaceResourcePlanStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceRegexpOperator_return precedenceRegexpOperator() throws RecognitionException { return gIdentifiersParser.precedenceRegexpOperator(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionQuantifierPredicate_return precedenceSimilarExpressionQuantifierPredicate(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionQuantifierPredicate(t); }\n \n-\tpublic HiveParser_ResourcePlanParser.createResourcePlanStatement_return createResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.createResourcePlanStatement(); }\n+\tpublic HiveParser_IdentifiersParser.descFuncNames_return descFuncNames() throws RecognitionException { return gIdentifiersParser.descFuncNames(); }\n \n-\tpublic HiveParser_IdentifiersParser.castExpression_return castExpression() throws RecognitionException { return gIdentifiersParser.castExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRebuild_return alterMaterializedViewSuffixRebuild(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRebuild(tableNameTree); }\n \n-\tpublic HiveParser_IdentifiersParser.timeUnitQualifiers_return timeUnitQualifiers() throws RecognitionException { return gIdentifiersParser.timeUnitQualifiers(); }\n+\tpublic HiveParser_FromClauseParser.valueRowConstructor_return valueRowConstructor() throws RecognitionException { return gFromClauseParser.valueRowConstructor(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionByClause_return partitionByClause() throws RecognitionException { return gIdentifiersParser.partitionByClause(); }\n+\tpublic HiveParser_IdentifiersParser.principalIdentifier_return principalIdentifier() throws RecognitionException { return gIdentifiersParser.principalIdentifier(); }\n \n-\tpublic HiveParser_IdentifiersParser.prepareStmtParam_return prepareStmtParam() throws RecognitionException { return gIdentifiersParser.prepareStmtParam(); }\n+\tpublic HiveParser_FromClauseParser.partitionTableFunctionSource_return partitionTableFunctionSource() throws RecognitionException { return gFromClauseParser.partitionTableFunctionSource(); }\n \n-\tpublic HiveParser_FromClauseParser.tableOrColumn_return tableOrColumn() throws RecognitionException { return gFromClauseParser.tableOrColumn(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenameCol_return alterStatementSuffixRenameCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenameCol(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetPartSpec_return alterStatementSuffixSetPartSpec() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetPartSpec(); }\n+\tpublic HiveParser_IdentifiersParser.expressions_return expressions() throws RecognitionException { return gIdentifiersParser.expressions(); }\n \n-\tpublic HiveParser_IdentifiersParser.expression_return expression() throws RecognitionException { return gIdentifiersParser.expression(); }\n+\tpublic HiveParser_IdentifiersParser.trimFunction_return trimFunction() throws RecognitionException { return gIdentifiersParser.trimFunction(); }\n \n-\tpublic HiveParser_IdentifiersParser.rollupStandard_return rollupStandard() throws RecognitionException { return gIdentifiersParser.rollupStandard(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSkewedby_return alterStatementSuffixSkewedby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSkewedby(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerOrExpression_return triggerOrExpression() throws RecognitionException { return gResourcePlanParser.triggerOrExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropTag_return alterStatementSuffixDropTag() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropTag(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceDistinctOperator_return precedenceDistinctOperator() throws RecognitionException { return gIdentifiersParser.precedenceDistinctOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseStatementSuffix_return alterDatabaseStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDatabaseStatementSuffix(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_value_expression_return window_value_expression() throws RecognitionException { return gSelectClauseParser.window_value_expression(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSpec_return partitionSpec() throws RecognitionException { return gIdentifiersParser.partitionSpec(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalExpression_return intervalExpression() throws RecognitionException { return gIdentifiersParser.intervalExpression(); }\n+\tpublic HiveParser_IdentifiersParser.extractExpression_return extractExpression() throws RecognitionException { return gIdentifiersParser.extractExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.whenExpression_return whenExpression() throws RecognitionException { return gIdentifiersParser.whenExpression(); }\n+\tpublic HiveParser_AlterClauseParser.skewedLocations_return skewedLocations() throws RecognitionException { return gAlterClauseParser.skewedLocations(); }\n \n-\tpublic HiveParser_IdentifiersParser.dateLiteral_return dateLiteral() throws RecognitionException { return gIdentifiersParser.dateLiteral(); }\n+\tpublic HiveParser_ResourcePlanParser.dropResourcePlanStatement_return dropResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.dropResourcePlanStatement(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerAndExpression_return triggerAndExpression() throws RecognitionException { return gResourcePlanParser.triggerAndExpression(); }\n+\tpublic HiveParser_ResourcePlanParser.comparisionOperator_return comparisionOperator() throws RecognitionException { return gResourcePlanParser.comparisionOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.havingClause_return havingClause() throws RecognitionException { return gIdentifiersParser.havingClause(); }\n+\tpublic HiveParser_AlterClauseParser.partitionLocation_return partitionLocation() throws RecognitionException { return gAlterClauseParser.partitionLocation(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalValue_return intervalValue() throws RecognitionException { return gIdentifiersParser.intervalValue(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCreateBranch_return alterStatementSuffixCreateBranch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCreateBranch(); }\n \n-\tpublic HiveParser_AlterClauseParser.tablePartitionPrefix_return tablePartitionPrefix() throws RecognitionException { return gAlterClauseParser.tablePartitionPrefix(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCreateTag_return alterStatementSuffixCreateTag() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCreateTag(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpUnassign_return rpUnassign() throws RecognitionException { return gResourcePlanParser.rpUnassign(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceEqualExpression_return precedenceEqualExpression() throws RecognitionException { return gIdentifiersParser.precedenceEqualExpression(); }\n \n-\tpublic HiveParser_FromClauseParser.atomjoinSource_return atomjoinSource() throws RecognitionException { return gFromClauseParser.atomjoinSource(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixBucketNum_return alterStatementSuffixBucketNum(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixBucketNum(partition); }\n \n-\tpublic HiveParser_IdentifiersParser.functionName_return functionName() throws RecognitionException { return gIdentifiersParser.functionName(); }\n+\tpublic HiveParser_IdentifiersParser.booleanValue_return booleanValue() throws RecognitionException { return gIdentifiersParser.booleanValue(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSpec_return partitionSpec() throws RecognitionException { return gIdentifiersParser.partitionSpec(); }\n+\tpublic HiveParser_FromClauseParser.splitSample_return splitSample() throws RecognitionException { return gFromClauseParser.splitSample(); }\n \n-\tpublic HiveParser_FromClauseParser.valueRowConstructor_return valueRowConstructor() throws RecognitionException { return gFromClauseParser.valueRowConstructor(); }\n+\tpublic HiveParser_CreateDDLParser.dcProperties_return dcProperties() throws RecognitionException { return gCreateDDLParser.dcProperties(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddConstraint_return alterStatementSuffixAddConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddConstraint(); }\n+\tpublic HiveParser_IdentifiersParser.intervalQualifiers_return intervalQualifiers() throws RecognitionException { return gIdentifiersParser.intervalQualifiers(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixClusterbySortby_return alterStatementSuffixClusterbySortby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixClusterbySortby(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerExpressionStandalone_return triggerExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerExpressionStandalone(); }\n \n-\tpublic HiveParser_FromClauseParser.tableBucketSample_return tableBucketSample() throws RecognitionException { return gFromClauseParser.tableBucketSample(); }\n+\tpublic HiveParser_IdentifiersParser.functionIdentifier_return functionIdentifier() throws RecognitionException { return gIdentifiersParser.functionIdentifier(); }\n \n-\tpublic HiveParser_IdentifiersParser.subQuerySelectorOperator_return subQuerySelectorOperator() throws RecognitionException { return gIdentifiersParser.subQuerySelectorOperator(); }\n+\tpublic HiveParser_IdentifiersParser.expressionPart_return expressionPart(CommonTree firstExprTree, boolean isStruct) throws RecognitionException { return gIdentifiersParser.expressionPart(firstExprTree, isStruct); }\n \n-\tpublic HiveParser_FromClauseParser.splitSample_return splitSample() throws RecognitionException { return gFromClauseParser.splitSample(); }\n+\tpublic HiveParser_ResourcePlanParser.unmanaged_return unmanaged() throws RecognitionException { return gResourcePlanParser.unmanaged(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_frame_boundary_return window_frame_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_boundary(); }\n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetUrl_return alterDataConnectorSuffixSetUrl() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetUrl(); }\n \n-\tpublic HiveParser_IdentifiersParser.descFuncNames_return descFuncNames() throws RecognitionException { return gIdentifiersParser.descFuncNames(); }\n+\tpublic HiveParser_IdentifiersParser.rollupStandard_return rollupStandard() throws RecognitionException { return gIdentifiersParser.rollupStandard(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPartNot_return precedenceSimilarExpressionPartNot(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPartNot(t); }\n+\tpublic HiveParser_SelectClauseParser.window_frame_return window_frame() throws RecognitionException { return gSelectClauseParser.window_frame(); }\n \n-\tpublic HiveParser_FromClauseParser.asOfClause_return asOfClause() throws RecognitionException { return gFromClauseParser.asOfClause(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceEqualOperator_return precedenceEqualOperator() throws RecognitionException { return gIdentifiersParser.precedenceEqualOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterTableStatementSuffix_return alterTableStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterTableStatementSuffix(); }\n+\tpublic HiveParser_IdentifiersParser.sysFuncNames_return sysFuncNames() throws RecognitionException { return gIdentifiersParser.sysFuncNames(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorOperator_return precedenceBitwiseXorOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorOperator(); }\n+\tpublic HiveParser_CreateDDLParser.likeTableOrFile_return likeTableOrFile() throws RecognitionException { return gCreateDDLParser.likeTableOrFile(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceStarOperator_return precedenceStarOperator() throws RecognitionException { return gIdentifiersParser.precedenceStarOperator(); }\n+\tpublic HiveParser_IdentifiersParser.intervalValue_return intervalValue() throws RecognitionException { return gIdentifiersParser.intervalValue(); }\n \n-\tpublic HiveParser_AlterClauseParser.skewedLocationMap_return skewedLocationMap() throws RecognitionException { return gAlterClauseParser.skewedLocationMap(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerAtomExpression_return triggerAtomExpression() throws RecognitionException { return gResourcePlanParser.triggerAtomExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedencePlusExpression_return precedencePlusExpression() throws RecognitionException { return gIdentifiersParser.precedencePlusExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAndOperator_return precedenceAndOperator() throws RecognitionException { return gIdentifiersParser.precedenceAndOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.qualifyClause_return qualifyClause() throws RecognitionException { return gIdentifiersParser.qualifyClause(); }\n+\tpublic HiveParser_IdentifiersParser.distributeByClause_return distributeByClause() throws RecognitionException { return gIdentifiersParser.distributeByClause(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerExpression_return triggerExpression() throws RecognitionException { return gResourcePlanParser.triggerExpression(); }\n+\tpublic HiveParser_FromClauseParser.aliasList_return aliasList() throws RecognitionException { return gFromClauseParser.aliasList(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetOwner_return alterDatabaseSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetOwner(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceStarExpression_return precedenceStarExpression() throws RecognitionException { return gIdentifiersParser.precedenceStarExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitionsElement_return alterStatementSuffixAddPartitionsElement() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitionsElement(); }\n+\tpublic HiveParser_IdentifiersParser.groupingExpressionSingle_return groupingExpressionSingle() throws RecognitionException { return gIdentifiersParser.groupingExpressionSingle(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCompact_return alterStatementSuffixCompact() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCompact(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceUnarySuffixExpression_return precedenceUnarySuffixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnarySuffixExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.quantifierType_return quantifierType() throws RecognitionException { return gIdentifiersParser.quantifierType(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceOrOperator_return precedenceOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceOrOperator(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerExpressionStandalone_return triggerExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerExpressionStandalone(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitions_return alterStatementSuffixAddPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitions(table); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAndOperator_return precedenceAndOperator() throws RecognitionException { return gIdentifiersParser.precedenceAndOperator(); }\n+\tpublic HiveParser_ResourcePlanParser.rpUnassign_return rpUnassign() throws RecognitionException { return gResourcePlanParser.rpUnassign(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixLocation_return alterStatementSuffixLocation(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixLocation(partition); }\n+\tpublic HiveParser_IdentifiersParser.columnRefOrderNotInParenthesis_return columnRefOrderNotInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderNotInParenthesis(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixExchangePartition_return alterStatementSuffixExchangePartition() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExchangePartition(); }\n+\tpublic HiveParser_IdentifiersParser.partitionByClause_return partitionByClause() throws RecognitionException { return gIdentifiersParser.partitionByClause(); }\n \n \tpublic HiveParser_ResourcePlanParser.dropTriggerStatement_return dropTriggerStatement() throws RecognitionException { return gResourcePlanParser.dropTriggerStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.columnRefOrderInParenthesis_return columnRefOrderInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderInParenthesis(); }\n-\n-\tpublic HiveParser_AlterClauseParser.optimizeTableStatementSuffix_return optimizeTableStatementSuffix() throws RecognitionException { return gAlterClauseParser.optimizeTableStatementSuffix(); }\n+\tpublic HiveParser_IdentifiersParser.isCondition_return isCondition() throws RecognitionException { return gIdentifiersParser.isCondition(); }\n \n-\tpublic HiveParser_IdentifiersParser.havingCondition_return havingCondition() throws RecognitionException { return gIdentifiersParser.havingCondition(); }\n+\tpublic HiveParser_IdentifiersParser.timeQualifiers_return timeQualifiers() throws RecognitionException { return gIdentifiersParser.timeQualifiers(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpUnassignList_return rpUnassignList() throws RecognitionException { return gResourcePlanParser.rpUnassignList(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSelectorVal_return partitionSelectorVal() throws RecognitionException { return gIdentifiersParser.partitionSelectorVal(); }\n \n-\tpublic HiveParser_ResourcePlanParser.activate_return activate() throws RecognitionException { return gResourcePlanParser.activate(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenamePart_return alterStatementSuffixRenamePart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenamePart(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupby_expression_return groupby_expression() throws RecognitionException { return gIdentifiersParser.groupby_expression(); }\n+\tpublic HiveParser_FromClauseParser.tableAlias_return tableAlias() throws RecognitionException { return gFromClauseParser.tableAlias(); }\n \n-\tpublic HiveParser_ResourcePlanParser.alterPoolStatement_return alterPoolStatement() throws RecognitionException { return gResourcePlanParser.alterPoolStatement(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionIn_return precedenceSimilarExpressionIn(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionIn(t); }\n \n-\tpublic HiveParser_IdentifiersParser.booleanValue_return booleanValue() throws RecognitionException { return gIdentifiersParser.booleanValue(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceConcatenateExpression_return precedenceConcatenateExpression() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.snapshotIdOfRef_return snapshotIdOfRef() throws RecognitionException { return gAlterClauseParser.snapshotIdOfRef(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSerdeProperties_return alterStatementSuffixSerdeProperties(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSerdeProperties(partition); }\n \n-\tpublic HiveParser_SelectClauseParser.selectList_return selectList() throws RecognitionException { return gSelectClauseParser.selectList(); }\n+\tpublic HiveParser_SelectClauseParser.selectExpressionList_return selectExpressionList() throws RecognitionException { return gSelectClauseParser.selectExpressionList(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropPoolStatement_return dropPoolStatement() throws RecognitionException { return gResourcePlanParser.dropPoolStatement(); }\n+\tpublic HiveParser_IdentifiersParser.tableOrPartition_return tableOrPartition() throws RecognitionException { return gIdentifiersParser.tableOrPartition(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionQuantifierPredicate_return precedenceSimilarExpressionQuantifierPredicate(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionQuantifierPredicate(t); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpression_return precedenceSimilarExpression() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.stringLiteralSequence_return stringLiteralSequence() throws RecognitionException { return gIdentifiersParser.stringLiteralSequence(); }\n+\tpublic HiveParser_ResourcePlanParser.createTriggerStatement_return createTriggerStatement() throws RecognitionException { return gResourcePlanParser.createTriggerStatement(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRebuild_return alterMaterializedViewSuffixRebuild(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRebuild(tableNameTree); }\n+\tpublic HiveParser_FromClauseParser.asOfClause_return asOfClause() throws RecognitionException { return gFromClauseParser.asOfClause(); }\n \n-\tpublic HiveParser_FromClauseParser.tableAllColumns_return tableAllColumns() throws RecognitionException { return gFromClauseParser.tableAllColumns(); }\n+\tpublic HiveParser_AlterClauseParser.optimizeTblRewriteDataSuffix_return optimizeTblRewriteDataSuffix() throws RecognitionException { return gAlterClauseParser.optimizeTblRewriteDataSuffix(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalQualifiers_return intervalQualifiers() throws RecognitionException { return gIdentifiersParser.intervalQualifiers(); }\n+\tpublic HiveParser_IdentifiersParser.parameterIdx_return parameterIdx() throws RecognitionException { return gIdentifiersParser.parameterIdx(); }\n \n \tpublic HiveParser_IdentifiersParser.precedenceConcatenateOperator_return precedenceConcatenateOperator() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateOperator(); }\n \n-\tpublic HiveParser_ResourcePlanParser.createPoolStatement_return createPoolStatement() throws RecognitionException { return gResourcePlanParser.createPoolStatement(); }\n-\n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixMergeFiles_return alterStatementSuffixMergeFiles(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixMergeFiles(partition); }\n+\tpublic HiveParser_ResourcePlanParser.alterMappingStatement_return alterMappingStatement() throws RecognitionException { return gResourcePlanParser.alterMappingStatement(); }\n \n-\tpublic HiveParser_CreateDDLParser.dataConnectorType_return dataConnectorType() throws RecognitionException { return gCreateDDLParser.dataConnectorType(); }\n+\tpublic HiveParser_SelectClauseParser.selectTrfmClause_return selectTrfmClause() throws RecognitionException { return gSelectClauseParser.selectTrfmClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateColumns_return alterStatementSuffixUpdateColumns() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateColumns(); }\n+\tpublic HiveParser_SelectClauseParser.trfmClause_return trfmClause() throws RecognitionException { return gSelectClauseParser.trfmClause(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerActionExpression_return triggerActionExpression() throws RecognitionException { return gResourcePlanParser.triggerActionExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceNotOperator_return precedenceNotOperator() throws RecognitionException { return gIdentifiersParser.precedenceNotOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceUnarySuffixExpression_return precedenceUnarySuffixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnarySuffixExpression(); }\n+\tpublic HiveParser_IdentifiersParser.subQuerySelectorOperator_return subQuerySelectorOperator() throws RecognitionException { return gIdentifiersParser.subQuerySelectorOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterViewStatementSuffix_return alterViewStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterViewStatementSuffix(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAmpersandExpression_return precedenceAmpersandExpression() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerAtomExpression_return triggerAtomExpression() throws RecognitionException { return gResourcePlanParser.triggerAtomExpression(); }\n+\tpublic HiveParser_IdentifiersParser.intervalExpression_return intervalExpression() throws RecognitionException { return gIdentifiersParser.intervalExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupingExpressionSingle_return groupingExpressionSingle() throws RecognitionException { return gIdentifiersParser.groupingExpressionSingle(); }\n+\tpublic HiveParser_AlterClauseParser.skewedLocationsList_return skewedLocationsList() throws RecognitionException { return gAlterClauseParser.skewedLocationsList(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceEqualExpression_return precedenceEqualExpression() throws RecognitionException { return gIdentifiersParser.precedenceEqualExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceStarOperator_return precedenceStarOperator() throws RecognitionException { return gIdentifiersParser.precedenceStarOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixStatsPart_return alterStatementSuffixStatsPart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixStatsPart(); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinToken_return uniqueJoinToken() throws RecognitionException { return gFromClauseParser.uniqueJoinToken(); }\n \n \tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetOwner_return alterDataConnectorSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetOwner(); }\n \n-\tpublic HiveParser_IdentifiersParser.timestampLocalTZLiteral_return timestampLocalTZLiteral() throws RecognitionException { return gIdentifiersParser.timestampLocalTZLiteral(); }\n-\n-\tpublic HiveParser_IdentifiersParser.precedenceUnaryPrefixExpression_return precedenceUnaryPrefixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnaryPrefixExpression(); }\n+\tpublic HiveParser_SelectClauseParser.window_frame_boundary_return window_frame_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_boundary(); }\n \n-\tpublic HiveParser_ResourcePlanParser.alterTriggerStatement_return alterTriggerStatement() throws RecognitionException { return gResourcePlanParser.alterTriggerStatement(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddCol_return alterStatementSuffixAddCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddCol(); }\n \n-\tpublic HiveParser_IdentifiersParser.charSetStringLiteral_return charSetStringLiteral() throws RecognitionException { return gIdentifiersParser.charSetStringLiteral(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropBranch_return alterStatementSuffixDropBranch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropBranch(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinExpr_return uniqueJoinExpr() throws RecognitionException { return gFromClauseParser.uniqueJoinExpr(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStats_return alterStatementSuffixUpdateStats(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStats(partition); }\n \n-\tpublic HiveParser_IdentifiersParser.groupByEmpty_return groupByEmpty() throws RecognitionException { return gIdentifiersParser.groupByEmpty(); }\n+\tpublic HiveParser_SelectClauseParser.selectExpression_return selectExpression() throws RecognitionException { return gSelectClauseParser.selectExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceOrOperator_return precedenceOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceOrOperator(); }\n+\tpublic HiveParser_SelectClauseParser.window_defn_return window_defn() throws RecognitionException { return gSelectClauseParser.window_defn(); }\n \n-\tpublic HiveParser_FromClauseParser.partitionTableFunctionSource_return partitionTableFunctionSource() throws RecognitionException { return gFromClauseParser.partitionTableFunctionSource(); }\n+\tpublic HiveParser_IdentifiersParser.nonReserved_return nonReserved() throws RecognitionException { return gIdentifiersParser.nonReserved(); }\n \n-\tpublic HiveParser_FromClauseParser.firstValueRowConstructor_return firstValueRowConstructor() throws RecognitionException { return gFromClauseParser.firstValueRowConstructor(); }\n+\tpublic HiveParser_FromClauseParser.tableSample_return tableSample() throws RecognitionException { return gFromClauseParser.tableSample(); }\n \n-\tpublic HiveParser_FromClauseParser.fromSource_return fromSource() throws RecognitionException { return gFromClauseParser.fromSource(); }\n+\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation_return alterTblPartitionStatementSuffixSkewedLocation() throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_defn_return window_defn() throws RecognitionException { return gSelectClauseParser.window_defn(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionMain_return precedenceSimilarExpressionMain() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionMain(); }\n \n-\tpublic HiveParser_ResourcePlanParser.enable_return enable() throws RecognitionException { return gResourcePlanParser.enable(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixTouch_return alterStatementSuffixTouch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixTouch(); }\n \n-\tpublic HiveParser_FromClauseParser.lateralView_return lateralView() throws RecognitionException { return gFromClauseParser.lateralView(); }\n+\tpublic HiveParser_IdentifiersParser.prepareStmtParam_return prepareStmtParam() throws RecognitionException { return gIdentifiersParser.prepareStmtParam(); }\n \n-\tpublic HiveParser_CreateDDLParser.dataConnectorComment_return dataConnectorComment() throws RecognitionException { return gCreateDDLParser.dataConnectorComment(); }\n+\tpublic HiveParser_IdentifiersParser.stringLiteralSequence_return stringLiteralSequence() throws RecognitionException { return gIdentifiersParser.stringLiteralSequence(); }\n \n-\tpublic HiveParser_FromClauseParser.joinToken_return joinToken() throws RecognitionException { return gFromClauseParser.joinToken(); }\n+\tpublic HiveParser_IdentifiersParser.groupingSetExpression_return groupingSetExpression() throws RecognitionException { return gIdentifiersParser.groupingSetExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionWithAlias_return expressionWithAlias() throws RecognitionException { return gIdentifiersParser.expressionWithAlias(); }\n+\tpublic HiveParser_IdentifiersParser.quantifierType_return quantifierType() throws RecognitionException { return gIdentifiersParser.quantifierType(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSkewedby_return alterStatementSuffixSkewedby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSkewedby(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorOperator_return precedenceBitwiseXorOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceOrExpression_return precedenceOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceOrExpression(); }\n+\tpublic HiveParser_CreateDDLParser.createTableStatement_return createTableStatement() throws RecognitionException { return gCreateDDLParser.createTableStatement(); }\n \n-\tpublic HiveParser_CreateDDLParser.dataConnectorUrl_return dataConnectorUrl() throws RecognitionException { return gCreateDDLParser.dataConnectorUrl(); }\n+\tpublic HiveParser_AlterClauseParser.optimizeTableStatementSuffix_return optimizeTableStatementSuffix() throws RecognitionException { return gAlterClauseParser.optimizeTableStatementSuffix(); }\n \n-\tpublic HiveParser_FromClauseParser.tableAlias_return tableAlias() throws RecognitionException { return gFromClauseParser.tableAlias(); }\n+\tpublic HiveParser_FromClauseParser.tableName_return tableName() throws RecognitionException { return gFromClauseParser.tableName(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorExpression_return precedenceBitwiseXorExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorExpression(); }\n+\tpublic HiveParser_FromClauseParser.tableSource_return tableSource() throws RecognitionException { return gFromClauseParser.tableSource(); }\n \n-\tpublic HiveParser_FromClauseParser.defaultValue_return defaultValue() throws RecognitionException { return gFromClauseParser.defaultValue(); }\n+\tpublic HiveParser_FromClauseParser.whereClause_return whereClause() throws RecognitionException { return gFromClauseParser.whereClause(); }\n \n-\tpublic HiveParser_FromClauseParser.expressionList_return expressionList() throws RecognitionException { return gFromClauseParser.expressionList(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAndExpression_return precedenceAndExpression() throws RecognitionException { return gIdentifiersParser.precedenceAndExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenameCol_return alterStatementSuffixRenameCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenameCol(); }\n+\tpublic HiveParser_IdentifiersParser.expressionWithAlias_return expressionWithAlias() throws RecognitionException { return gIdentifiersParser.expressionWithAlias(); }\n \n \tpublic HiveParser_ResourcePlanParser.poolPath_return poolPath() throws RecognitionException { return gResourcePlanParser.poolPath(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAndExpression_return precedenceAndExpression() throws RecognitionException { return gIdentifiersParser.precedenceAndExpression(); }\n+\tpublic HiveParser_IdentifiersParser.identifier_return identifier() throws RecognitionException { return gIdentifiersParser.identifier(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpression_return precedenceSimilarExpression() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterViewSuffixProperties_return alterViewSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterViewSuffixProperties(); }\n \n-\tpublic HiveParser_FromClauseParser.fromClause_return fromClause() throws RecognitionException { return gFromClauseParser.fromClause(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPart_return precedenceSimilarExpressionPart(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPart(t); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementChangeColPosition_return alterStatementChangeColPosition() throws RecognitionException { return gAlterClauseParser.alterStatementChangeColPosition(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSelectorSpec_return partitionSelectorSpec() throws RecognitionException { return gIdentifiersParser.partitionSelectorSpec(); }\n \n-\tpublic HiveParser_ResourcePlanParser.poolAssignList_return poolAssignList() throws RecognitionException { return gResourcePlanParser.poolAssignList(); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinTableSource_return uniqueJoinTableSource() throws RecognitionException { return gFromClauseParser.uniqueJoinTableSource(); }\n \n-\tpublic HiveParser_ResourcePlanParser.alterMappingStatement_return alterMappingStatement() throws RecognitionException { return gResourcePlanParser.alterMappingStatement(); }\n+\tpublic HiveParser_CreateDDLParser.dataConnectorComment_return dataConnectorComment() throws RecognitionException { return gCreateDDLParser.dataConnectorComment(); }\n \n-\tpublic HiveParser_ResourcePlanParser.comparisionOperator_return comparisionOperator() throws RecognitionException { return gResourcePlanParser.comparisionOperator(); }\n+\tpublic HiveParser_FromClauseParser.defaultValue_return defaultValue() throws RecognitionException { return gFromClauseParser.defaultValue(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceFieldExpression_return precedenceFieldExpression() throws RecognitionException { return gIdentifiersParser.precedenceFieldExpression(); }\n+\tpublic HiveParser_IdentifiersParser.intervalLiteral_return intervalLiteral() throws RecognitionException { return gIdentifiersParser.intervalLiteral(); }\n \n-\tpublic HiveParser_ResourcePlanParser.unmanaged_return unmanaged() throws RecognitionException { return gResourcePlanParser.unmanaged(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixProperties_return alterStatementSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixProperties(); }\n \n-\tpublic HiveParser_FromClauseParser.valuesSource_return valuesSource() throws RecognitionException { return gFromClauseParser.valuesSource(); }\n+\tpublic HiveParser_ResourcePlanParser.rpUnassignList_return rpUnassignList() throws RecognitionException { return gResourcePlanParser.rpUnassignList(); }\n \n-\tpublic HiveParser_CreateDDLParser.createTableStatement_return createTableStatement() throws RecognitionException { return gCreateDDLParser.createTableStatement(); }\n+\tpublic HiveParser_ResourcePlanParser.createResourcePlanStatement_return createResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.createResourcePlanStatement(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUnArchive_return alterStatementSuffixUnArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUnArchive(); }\n+\tpublic HiveParser_ResourcePlanParser.globalWmStatement_return globalWmStatement() throws RecognitionException { return gResourcePlanParser.globalWmStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.function_return function() throws RecognitionException { return gIdentifiersParser.function(); }\n+\tpublic HiveParser_IdentifiersParser.subQueryExpression_return subQueryExpression() throws RecognitionException { return gIdentifiersParser.subQueryExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionMain_return precedenceSimilarExpressionMain() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionMain(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerAndExpression_return triggerAndExpression() throws RecognitionException { return gResourcePlanParser.triggerAndExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.constant_return constant() throws RecognitionException { return gIdentifiersParser.constant(); }\n+\tpublic HiveParser_ResourcePlanParser.createPoolStatement_return createPoolStatement() throws RecognitionException { return gResourcePlanParser.createPoolStatement(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCreateTag_return alterStatementSuffixCreateTag() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCreateTag(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetPartSpec_return alterStatementSuffixSetPartSpec() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetPartSpec(); }\n \n-\tpublic HiveParser_IdentifiersParser.isCondition_return isCondition() throws RecognitionException { return gIdentifiersParser.isCondition(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceRegexpOperator_return precedenceRegexpOperator() throws RecognitionException { return gIdentifiersParser.precedenceRegexpOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceEqualOperator_return precedenceEqualOperator() throws RecognitionException { return gIdentifiersParser.precedenceEqualOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffix_return alterTblPartitionStatementSuffix(boolean partition) throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffix(partition); }\n \n-\tpublic HiveParser_IdentifiersParser.sysFuncNames_return sysFuncNames() throws RecognitionException { return gIdentifiersParser.sysFuncNames(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerExpression_return triggerExpression() throws RecognitionException { return gResourcePlanParser.triggerExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.floorExpression_return floorExpression() throws RecognitionException { return gIdentifiersParser.floorExpression(); }\n+\tpublic HiveParser_PrepareStatementParser.prepareStatement_return prepareStatement() throws RecognitionException { return gPrepareStatementParser.prepareStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.firstExpressionsWithAlias_return firstExpressionsWithAlias() throws RecognitionException { return gIdentifiersParser.firstExpressionsWithAlias(); }\n+\tpublic HiveParser_FromClauseParser.fromClause_return fromClause() throws RecognitionException { return gFromClauseParser.fromClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.fileFormat_return fileFormat() throws RecognitionException { return gAlterClauseParser.fileFormat(); }\n+\tpublic HiveParser_IdentifiersParser.constant_return constant() throws RecognitionException { return gIdentifiersParser.constant(); }\n \n-\tpublic HiveParser_AlterClauseParser.retentionOfSnapshots_return retentionOfSnapshots() throws RecognitionException { return gAlterClauseParser.retentionOfSnapshots(); }\n+\tpublic HiveParser_SelectClauseParser.window_specification_return window_specification(CommonTree nullTreatment) throws RecognitionException { return gSelectClauseParser.window_specification(nullTreatment); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStatsCol_return alterStatementSuffixUpdateStatsCol(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStatsCol(partition); }\n+\tpublic HiveParser_AlterClauseParser.alterStatement_return alterStatement() throws RecognitionException { return gAlterClauseParser.alterStatement(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectExpressionList_return selectExpressionList() throws RecognitionException { return gSelectClauseParser.selectExpressionList(); }\n+\tpublic HiveParser_IdentifiersParser.functionName_return functionName() throws RecognitionException { return gIdentifiersParser.functionName(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCreateBranch_return alterStatementSuffixCreateBranch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCreateBranch(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetManagedLocation_return alterDatabaseSuffixSetManagedLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetManagedLocation(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupByClause_return groupByClause() throws RecognitionException { return gIdentifiersParser.groupByClause(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropPartitions_return alterStatementSuffixDropPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropPartitions(table); }\n \n-\tpublic HiveParser_FromClauseParser.partitioningSpec_return partitioningSpec() throws RecognitionException { return gFromClauseParser.partitioningSpec(); }\n+\tpublic HiveParser_IdentifiersParser.havingClause_return havingClause() throws RecognitionException { return gIdentifiersParser.havingClause(); }\n \n-\tpublic HiveParser_FromClauseParser.tableSample_return tableSample() throws RecognitionException { return gFromClauseParser.tableSample(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitionsElement_return alterStatementSuffixAddPartitionsElement() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitionsElement(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSelectorOperator_return partitionSelectorOperator() throws RecognitionException { return gIdentifiersParser.partitionSelectorOperator(); }\n+\tpublic HiveParser_IdentifiersParser.clusterByClause_return clusterByClause() throws RecognitionException { return gIdentifiersParser.clusterByClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.booleanValueTok_return booleanValueTok() throws RecognitionException { return gIdentifiersParser.booleanValueTok(); }\n+\tpublic HiveParser_SelectClauseParser.selectItem_return selectItem() throws RecognitionException { return gSelectClauseParser.selectItem(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixExecute_return alterStatementSuffixExecute() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExecute(); }\n+\tpublic HiveParser_FromClauseParser.atomjoinSource_return atomjoinSource() throws RecognitionException { return gFromClauseParser.atomjoinSource(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceNotOperator_return precedenceNotOperator() throws RecognitionException { return gIdentifiersParser.precedenceNotOperator(); }\n+\tpublic HiveParser_SelectClauseParser.selectList_return selectList() throws RecognitionException { return gSelectClauseParser.selectList(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionPart_return expressionPart(CommonTree firstExprTree, boolean isStruct) throws RecognitionException { return gIdentifiersParser.expressionPart(firstExprTree, isStruct); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropConstraint_return alterStatementSuffixDropConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropConstraint(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionIn_return precedenceSimilarExpressionIn(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionIn(t); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinSource_return uniqueJoinSource() throws RecognitionException { return gFromClauseParser.uniqueJoinSource(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectTrfmClause_return selectTrfmClause() throws RecognitionException { return gSelectClauseParser.selectTrfmClause(); }\n+\tpublic HiveParser_FromClauseParser.valuesSource_return valuesSource() throws RecognitionException { return gFromClauseParser.valuesSource(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRewrite_return alterMaterializedViewSuffixRewrite(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRewrite(tableNameTree); }\n+\tpublic HiveParser_IdentifiersParser.groupby_expression_return groupby_expression() throws RecognitionException { return gIdentifiersParser.groupby_expression(); }\n \n-\tpublic HiveParser_IdentifiersParser.clusterByClause_return clusterByClause() throws RecognitionException { return gIdentifiersParser.clusterByClause(); }\n+\tpublic HiveParser_FromClauseParser.virtualTableSource_return virtualTableSource() throws RecognitionException { return gFromClauseParser.virtualTableSource(); }\n \n-\tpublic HiveParser_IdentifiersParser.distributeByClause_return distributeByClause() throws RecognitionException { return gIdentifiersParser.distributeByClause(); }\n+\tpublic HiveParser_CreateDDLParser.dataConnectorType_return dataConnectorType() throws RecognitionException { return gCreateDDLParser.dataConnectorType(); }\n \n-\tpublic HiveParser_PrepareStatementParser.executeStatement_return executeStatement() throws RecognitionException { return gPrepareStatementParser.executeStatement(); }\n+\tpublic HiveParser_AlterClauseParser.alterMaterializedViewStatementSuffix_return alterMaterializedViewStatementSuffix(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewStatementSuffix(tableNameTree); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerActionExpressionStandalone_return triggerActionExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerActionExpressionStandalone(); }\n+\tpublic HiveParser_IdentifiersParser.dateLiteral_return dateLiteral() throws RecognitionException { return gIdentifiersParser.dateLiteral(); }\n \n-\tpublic HiveParser_FromClauseParser.tableSource_return tableSource() throws RecognitionException { return gFromClauseParser.tableSource(); }\n+\tpublic HiveParser_IdentifiersParser.orderByClause_return orderByClause() throws RecognitionException { return gIdentifiersParser.orderByClause(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinTableSource_return uniqueJoinTableSource() throws RecognitionException { return gFromClauseParser.uniqueJoinTableSource(); }\n+\tpublic HiveParser_ResourcePlanParser.dropPoolStatement_return dropPoolStatement() throws RecognitionException { return gResourcePlanParser.dropPoolStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.principalIdentifier_return principalIdentifier() throws RecognitionException { return gIdentifiersParser.principalIdentifier(); }\n+\tpublic HiveParser_ResourcePlanParser.enable_return enable() throws RecognitionException { return gResourcePlanParser.enable(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectExpression_return selectExpression() throws RecognitionException { return gSelectClauseParser.selectExpression(); }\n+\tpublic HiveParser_IdentifiersParser.firstExpressionsWithAlias_return firstExpressionsWithAlias() throws RecognitionException { return gIdentifiersParser.firstExpressionsWithAlias(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixBucketNum_return alterStatementSuffixBucketNum(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixBucketNum(partition); }\n+\tpublic HiveParser_IdentifiersParser.atomExpression_return atomExpression() throws RecognitionException { return gIdentifiersParser.atomExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.disable_return disable() throws RecognitionException { return gResourcePlanParser.disable(); }\n+\tpublic HiveParser_ResourcePlanParser.rpAssign_return rpAssign() throws RecognitionException { return gResourcePlanParser.rpAssign(); }\n \n-\tpublic HiveParser_FromClauseParser.joinSource_return joinSource() throws RecognitionException { return gFromClauseParser.joinSource(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPartNot_return precedenceSimilarExpressionPartNot(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPartNot(t); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixProperties_return alterDatabaseSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixProperties(); }\n+\tpublic HiveParser_AlterClauseParser.tablePartitionPrefix_return tablePartitionPrefix() throws RecognitionException { return gAlterClauseParser.tablePartitionPrefix(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionsInParenthesis_return expressionsInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsInParenthesis(isStruct, forceStruct); }\n+\tpublic HiveParser_ResourcePlanParser.poolAssign_return poolAssign() throws RecognitionException { return gResourcePlanParser.poolAssign(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropTag_return alterStatementSuffixDropTag() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropTag(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixFileFormat_return alterStatementSuffixFileFormat(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixFileFormat(partition); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetUrl_return alterDataConnectorSuffixSetUrl() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetUrl(); }\n+\tpublic HiveParser_IdentifiersParser.precedencePlusOperator_return precedencePlusOperator() throws RecognitionException { return gIdentifiersParser.precedencePlusOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.tableOrPartition_return tableOrPartition() throws RecognitionException { return gIdentifiersParser.tableOrPartition(); }\n+\tpublic HiveParser_IdentifiersParser.whenExpression_return whenExpression() throws RecognitionException { return gIdentifiersParser.whenExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixTouch_return alterStatementSuffixTouch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixTouch(); }\n+\tpublic HiveParser_CreateDDLParser.dataConnectorUrl_return dataConnectorUrl() throws RecognitionException { return gCreateDDLParser.dataConnectorUrl(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerLiteral_return triggerLiteral() throws RecognitionException { return gResourcePlanParser.triggerLiteral(); }\n+\tpublic HiveParser_AlterClauseParser.blocking_return blocking() throws RecognitionException { return gAlterClauseParser.blocking(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedencePlusOperator_return precedencePlusOperator() throws RecognitionException { return gIdentifiersParser.precedencePlusOperator(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSelectorOperator_return partitionSelectorOperator() throws RecognitionException { return gIdentifiersParser.partitionSelectorOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceNotExpression_return precedenceNotExpression() throws RecognitionException { return gIdentifiersParser.precedenceNotExpression(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerActionExpression_return triggerActionExpression() throws RecognitionException { return gResourcePlanParser.triggerActionExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionAtom_return precedenceSimilarExpressionAtom(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionAtom(t); }\n+\tpublic HiveParser_IdentifiersParser.timestampLocalTZLiteral_return timestampLocalTZLiteral() throws RecognitionException { return gIdentifiersParser.timestampLocalTZLiteral(); }\n \n-\tpublic HiveParser_AlterClauseParser.skewedLocations_return skewedLocations() throws RecognitionException { return gAlterClauseParser.skewedLocations(); }\n+\tpublic HiveParser_IdentifiersParser.columnRefOrderInParenthesis_return columnRefOrderInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderInParenthesis(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionVal_return partitionVal() throws RecognitionException { return gIdentifiersParser.partitionVal(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixStatsPart_return alterStatementSuffixStatsPart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixStatsPart(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixFileFormat_return alterStatementSuffixFileFormat(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixFileFormat(partition); }\n+\tpublic HiveParser_FromClauseParser.firstValueRowConstructor_return firstValueRowConstructor() throws RecognitionException { return gFromClauseParser.firstValueRowConstructor(); }\n \n-\tpublic HiveParser_SelectClauseParser.trfmClause_return trfmClause() throws RecognitionException { return gSelectClauseParser.trfmClause(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetOwner_return alterDatabaseSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetOwner(); }\n \n-\tpublic HiveParser_AlterClauseParser.partitionLocation_return partitionLocation() throws RecognitionException { return gAlterClauseParser.partitionLocation(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceNotExpression_return precedenceNotExpression() throws RecognitionException { return gIdentifiersParser.precedenceNotExpression(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinToken_return uniqueJoinToken() throws RecognitionException { return gFromClauseParser.uniqueJoinToken(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerLiteral_return triggerLiteral() throws RecognitionException { return gResourcePlanParser.triggerLiteral(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatement_return alterStatement() throws RecognitionException { return gAlterClauseParser.alterStatement(); }\n+\tpublic HiveParser_IdentifiersParser.rollupOldSyntax_return rollupOldSyntax() throws RecognitionException { return gIdentifiersParser.rollupOldSyntax(); }\n \n-\tpublic HiveParser_IdentifiersParser.sortByClause_return sortByClause() throws RecognitionException { return gIdentifiersParser.sortByClause(); }\n+\tpublic HiveParser_IdentifiersParser.floorDateQualifiers_return floorDateQualifiers() throws RecognitionException { return gIdentifiersParser.floorDateQualifiers(); }\n \n-\tpublic HiveParser_IdentifiersParser.trimFunction_return trimFunction() throws RecognitionException { return gIdentifiersParser.trimFunction(); }\n+\tpublic HiveParser_FromClauseParser.expressionList_return expressionList() throws RecognitionException { return gFromClauseParser.expressionList(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRename_return alterStatementSuffixRename(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRename(table); }\n+\tpublic HiveParser_IdentifiersParser.groupByClause_return groupByClause() throws RecognitionException { return gIdentifiersParser.groupByClause(); }\n \n-\tpublic HiveParser_ResourcePlanParser.createTriggerStatement_return createTriggerStatement() throws RecognitionException { return gResourcePlanParser.createTriggerStatement(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceFieldExpression_return precedenceFieldExpression() throws RecognitionException { return gIdentifiersParser.precedenceFieldExpression(); }\n \n-\tpublic HiveParser_CreateDDLParser.likeTableOrFile_return likeTableOrFile() throws RecognitionException { return gCreateDDLParser.likeTableOrFile(); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinExpr_return uniqueJoinExpr() throws RecognitionException { return gFromClauseParser.uniqueJoinExpr(); }\n \n-\tpublic HiveParser_ResourcePlanParser.replaceResourcePlanStatement_return replaceResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.replaceResourcePlanStatement(); }\n+\tpublic HiveParser_FromClauseParser.lateralView_return lateralView() throws RecognitionException { return gFromClauseParser.lateralView(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinSource_return uniqueJoinSource() throws RecognitionException { return gFromClauseParser.uniqueJoinSource(); }\n+\tpublic HiveParser_IdentifiersParser.expressionOrDefault_return expressionOrDefault() throws RecognitionException { return gIdentifiersParser.expressionOrDefault(); }\n \n-\tpublic HiveParser_IdentifiersParser.functionIdentifier_return functionIdentifier() throws RecognitionException { return gIdentifiersParser.functionIdentifier(); }\n+\tpublic HiveParser_FromClauseParser.tableBucketSample_return tableBucketSample() throws RecognitionException { return gFromClauseParser.tableBucketSample(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAmpersandExpression_return precedenceAmpersandExpression() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixLocation_return alterStatementSuffixLocation(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixLocation(partition); }\n \n-\tpublic HiveParser_ResourcePlanParser.resourcePlanDdlStatements_return resourcePlanDdlStatements() throws RecognitionException { return gResourcePlanParser.resourcePlanDdlStatements(); }\n+\tpublic HiveParser_IdentifiersParser.castExpression_return castExpression() throws RecognitionException { return gIdentifiersParser.castExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.atomExpression_return atomExpression() throws RecognitionException { return gIdentifiersParser.atomExpression(); }\n+\tpublic HiveParser_IdentifiersParser.groupByEmpty_return groupByEmpty() throws RecognitionException { return gIdentifiersParser.groupByEmpty(); }\n \n-\tpublic HiveParser_FromClauseParser.joinSourcePart_return joinSourcePart() throws RecognitionException { return gFromClauseParser.joinSourcePart(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetOwner_return alterStatementSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetOwner(); }\n \n-\tpublic HiveParser_AlterClauseParser.refRetain_return refRetain() throws RecognitionException { return gAlterClauseParser.refRetain(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixExecute_return alterStatementSuffixExecute() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExecute(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarOperator_return precedenceSimilarOperator() throws RecognitionException { return gIdentifiersParser.precedenceSimilarOperator(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceUnaryPrefixExpression_return precedenceUnaryPrefixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnaryPrefixExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.optimizeTblRewriteDataSuffix_return optimizeTblRewriteDataSuffix() throws RecognitionException { return gAlterClauseParser.optimizeTblRewriteDataSuffix(); }\n+\tpublic HiveParser_FromClauseParser.subQuerySource_return subQuerySource() throws RecognitionException { return gFromClauseParser.subQuerySource(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressions_return expressions() throws RecognitionException { return gIdentifiersParser.expressions(); }\n+\tpublic HiveParser_CreateDDLParser.createDataConnectorStatement_return createDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.createDataConnectorStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.timeQualifiers_return timeQualifiers() throws RecognitionException { return gIdentifiersParser.timeQualifiers(); }\n+\tpublic HiveParser_ResourcePlanParser.alterPoolStatement_return alterPoolStatement() throws RecognitionException { return gResourcePlanParser.alterPoolStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceStarExpression_return precedenceStarExpression() throws RecognitionException { return gIdentifiersParser.precedenceStarExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementPartitionKeyType_return alterStatementPartitionKeyType() throws RecognitionException { return gAlterClauseParser.alterStatementPartitionKeyType(); }\n \n \tpublic HiveParser_FromClauseParser.partitionedTableFunction_return partitionedTableFunction() throws RecognitionException { return gFromClauseParser.partitionedTableFunction(); }\n \n-\tpublic HiveParser_FromClauseParser.whereClause_return whereClause() throws RecognitionException { return gFromClauseParser.whereClause(); }\n-\n-\tpublic HiveParser_SelectClauseParser.window_specification_return window_specification(CommonTree nullTreatment) throws RecognitionException { return gSelectClauseParser.window_specification(nullTreatment); }\n-\n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixProperties_return alterStatementSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixProperties(); }\n-\n-\tpublic HiveParser_IdentifiersParser.groupingSetExpression_return groupingSetExpression() throws RecognitionException { return gIdentifiersParser.groupingSetExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementChangeColPosition_return alterStatementChangeColPosition() throws RecognitionException { return gAlterClauseParser.alterStatementChangeColPosition(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionOrDefault_return expressionOrDefault() throws RecognitionException { return gIdentifiersParser.expressionOrDefault(); }\n+\tpublic HiveParser_FromClauseParser.fromSource_return fromSource() throws RecognitionException { return gFromClauseParser.fromSource(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpAssignList_return rpAssignList() throws RecognitionException { return gResourcePlanParser.rpAssignList(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrExpression_return precedenceBitwiseOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.extractExpression_return extractExpression() throws RecognitionException { return gIdentifiersParser.extractExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarOperator_return precedenceSimilarOperator() throws RecognitionException { return gIdentifiersParser.precedenceSimilarOperator(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_range_expression_return window_range_expression() throws RecognitionException { return gSelectClauseParser.window_range_expression(); }\n+\tpublic HiveParser_IdentifiersParser.function_return function() throws RecognitionException { return gIdentifiersParser.function(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropPartitions_return alterStatementSuffixDropPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropPartitions(table); }\n+\tpublic HiveParser_ResourcePlanParser.createMappingStatement_return createMappingStatement() throws RecognitionException { return gResourcePlanParser.createMappingStatement(); }\n \n-\tpublic HiveParser_ResourcePlanParser.poolAssign_return poolAssign() throws RecognitionException { return gResourcePlanParser.poolAssign(); }\n+\tpublic HiveParser_IdentifiersParser.partitionVal_return partitionVal() throws RecognitionException { return gIdentifiersParser.partitionVal(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropBranch_return alterStatementSuffixDropBranch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropBranch(); }\n+\tpublic HiveParser_ResourcePlanParser.poolAssignList_return poolAssignList() throws RecognitionException { return gResourcePlanParser.poolAssignList(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixArchive_return alterStatementSuffixArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixArchive(); }\n+\tpublic HiveParser_AlterClauseParser.retentionOfSnapshots_return retentionOfSnapshots() throws RecognitionException { return gAlterClauseParser.retentionOfSnapshots(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrOperator_return precedenceBitwiseOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrOperator(); }\n+\tpublic HiveParser_FromClauseParser.joinSource_return joinSource() throws RecognitionException { return gFromClauseParser.joinSource(); }\n \n-\tpublic HiveParser_FromClauseParser.tableName_return tableName() throws RecognitionException { return gFromClauseParser.tableName(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRename_return alterStatementSuffixRename(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRename(table); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorStatementSuffix_return alterDataConnectorStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDataConnectorStatementSuffix(); }\n+\tpublic HiveParser_IdentifiersParser.expressionsInParenthesis_return expressionsInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsInParenthesis(isStruct, forceStruct); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSerdeProperties_return alterStatementSuffixSerdeProperties(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSerdeProperties(partition); }\n+\tpublic HiveParser_IdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName_return sql11ReservedKeywordsUsedAsFunctionName() throws RecognitionException { return gIdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName(); }\n \n-\tpublic HiveParser_IdentifiersParser.orderByClause_return orderByClause() throws RecognitionException { return gIdentifiersParser.orderByClause(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixProperties_return alterDatabaseSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixProperties(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceConcatenateExpression_return precedenceConcatenateExpression() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateExpression(); }\n+\tpublic HiveParser_IdentifiersParser.timestampLiteral_return timestampLiteral() throws RecognitionException { return gIdentifiersParser.timestampLiteral(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffix_return alterTblPartitionStatementSuffix(boolean partition) throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffix(partition); }\n+\tpublic HiveParser_FromClauseParser.viewName_return viewName() throws RecognitionException { return gFromClauseParser.viewName(); }\n \n-\tpublic HiveParser_IdentifiersParser.floorDateQualifiers_return floorDateQualifiers() throws RecognitionException { return gIdentifiersParser.floorDateQualifiers(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceOrExpression_return precedenceOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceOrExpression(); }\n \n \tpublic final boolean synpred18_HiveParser() {\n \t\tstate.backtracking++;\n \t\tint start = input.mark();\n \t\ttry {\n \t\t\tsynpred18_HiveParser_fragment(); // can never throw exception\n \t\t} catch (RecognitionException re) {\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_AlterClauseParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_AlterClauseParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 AlterClauseParser.g 2024-09-26 14:10:39\n+// $ANTLR 3.5.2 AlterClauseParser.g 2025-01-31 11:23:52\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1229,15 +1229,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_alterViewStatementSuffix_in_alterStatement99);\n \t\t\t\t\talterViewStatementSuffix9=alterViewStatementSuffix();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_alterViewStatementSuffix.add(alterViewStatementSuffix9.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: alterViewStatementSuffix, tableName\n+\t\t\t\t\t// elements: tableName, alterViewStatementSuffix\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -2040,15 +2040,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec38.getTree());\n \t\t\t\t\tpushFollow(FOLLOW_alterTblPartitionStatementSuffix_in_alterTableStatementSuffix336);\n \t\t\t\t\talterTblPartitionStatementSuffix39=alterTblPartitionStatementSuffix(true);\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_alterTblPartitionStatementSuffix.add(alterTblPartitionStatementSuffix39.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: partitionSpec, alterTblPartitionStatementSuffix\n+\t\t\t\t\t// elements: alterTblPartitionStatementSuffix, partitionSpec\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3551,15 +3551,15 @@\n \n \t\t\tpushFollow(FOLLOW_dbProperties_in_alterDatabaseSuffixProperties908);\n \t\t\tdbProperties83=gHiveParser.dbProperties();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_dbProperties.add(dbProperties83.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: name, dbProperties\n+\t\t\t// elements: dbProperties, name\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3903,15 +3903,15 @@\n \t\t\t\t\tKW_MANAGEDLOCATION90=(Token)match(input,KW_MANAGEDLOCATION,FOLLOW_KW_MANAGEDLOCATION_in_alterDatabaseSuffixSetLocation1040); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_MANAGEDLOCATION.add(KW_MANAGEDLOCATION90);\n \n \t\t\t\t\tnewLocation=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_alterDatabaseSuffixSetLocation1044); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(newLocation);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: newLocation, dbName\n+\t\t\t\t\t// elements: dbName, newLocation\n \t\t\t\t\t// token labels: newLocation\n \t\t\t\t\t// rule labels: dbName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -4006,15 +4006,15 @@\n \t\t\tKW_MANAGEDLOCATION92=(Token)match(input,KW_MANAGEDLOCATION,FOLLOW_KW_MANAGEDLOCATION_in_alterDatabaseSuffixSetManagedLocation1093); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_KW_MANAGEDLOCATION.add(KW_MANAGEDLOCATION92);\n \n \t\t\tnewLocation=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_alterDatabaseSuffixSetManagedLocation1097); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(newLocation);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: dbName, newLocation\n+\t\t\t// elements: newLocation, dbName\n \t\t\t// token labels: newLocation\n \t\t\t// rule labels: dbName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4276,15 +4276,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_restrictOrCascade.add(restrictOrCascade100.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: columnNameTypeList, restrictOrCascade, columnNameTypeList, restrictOrCascade\n+\t\t\t// elements: columnNameTypeList, restrictOrCascade, restrictOrCascade, columnNameTypeList\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4485,15 +4485,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_alterConstraintWithName.add(alterConstraintWithName102.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: alterConstraintWithName, alterForeignKeyWithName\n+\t\t\t// elements: alterForeignKeyWithName, alterConstraintWithName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4915,15 +4915,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_restrictOrCascade.add(restrictOrCascade114.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: alterColumnConstraint, newName, colType, comment, alterStatementChangeColPosition, restrictOrCascade, oldName\n+\t\t\t// elements: alterStatementChangeColPosition, restrictOrCascade, comment, newName, alterColumnConstraint, oldName, colType\n \t\t\t// token labels: comment\n \t\t\t// rule labels: newName, oldName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5103,15 +5103,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colName, tableProperties, colName, tableProperties, comment, comment\n+\t\t\t// elements: comment, colName, tableProperties, tableProperties, colName, comment\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5507,15 +5507,15 @@\n \t\t\t\t\tEarlyExitException eee = new EarlyExitException(24, input);\n \t\t\t\t\tthrow eee;\n \t\t\t\t}\n \t\t\t\tcnt24++;\n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: ifNotExists, alterStatementSuffixAddPartitionsElement, ifNotExists, alterStatementSuffixAddPartitionsElement\n+\t\t\t// elements: alterStatementSuffixAddPartitionsElement, ifNotExists, alterStatementSuffixAddPartitionsElement, ifNotExists\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -6237,15 +6237,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replicationClause.add(replicationClause147.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: KW_PURGE, ifExists, partitionSelectorSpec, replicationClause, replicationClause, ifExists, partitionSelectorSpec\n+\t\t\t// elements: partitionSelectorSpec, KW_PURGE, replicationClause, partitionSelectorSpec, replicationClause, ifExists, ifExists\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -6475,15 +6475,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_tableProperties_in_alterStatementSuffixProperties2204);\n \t\t\t\t\ttableProperties154=gHiveParser.tableProperties();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableProperties.add(tableProperties154.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ifExists, tableProperties\n+\t\t\t\t\t// elements: tableProperties, ifExists\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -7809,15 +7809,15 @@\n \t\t\tEQUAL193=(Token)match(input,EQUAL,FOLLOW_EQUAL_in_skewedLocationMap2818); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_EQUAL.add(EQUAL193);\n \n \t\t\tvalue=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_skewedLocationMap2822); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(value);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: key, value\n+\t\t\t// elements: value, key\n \t\t\t// token labels: value\n \t\t\t// rule labels: key, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -8230,15 +8230,15 @@\n \n \t\t\tpushFollow(FOLLOW_tableName_in_alterStatementSuffixExchangePartition2998);\n \t\t\texchangename=gHiveParser.tableName();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_tableName.add(exchangename.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: exchangename, partitionSpec\n+\t\t\t// elements: partitionSpec, exchangename\n \t\t\t// token labels: \n \t\t\t// rule labels: exchangename, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -8483,15 +8483,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableProperties, colName, comment\n+\t\t\t// elements: colName, comment, tableProperties\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -9058,15 +9058,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_orderByClause.add(orderByClause229.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: compactPool, blocking, tableProperties, compactType, tableImplBuckets, orderByClause\n+\t\t\t// elements: compactPool, blocking, orderByClause, tableImplBuckets, tableProperties, compactType\n \t\t\t// token labels: compactType\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -9802,15 +9802,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: expireParam, KW_EXPIRE_SNAPSHOTS\n+\t\t\t\t\t// elements: KW_EXPIRE_SNAPSHOTS, expireParam\n \t\t\t\t\t// token labels: expireParam\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9926,15 +9926,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: sourceBranch, targetBranch, KW_FAST_FORWARD\n+\t\t\t\t\t// elements: KW_FAST_FORWARD, targetBranch, sourceBranch\n \t\t\t\t\t// token labels: targetBranch, sourceBranch\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9977,15 +9977,15 @@\n \t\t\t\t\tKW_CHERRY_PICK256=(Token)match(input,KW_CHERRY_PICK,FOLLOW_KW_CHERRY_PICK_in_alterStatementSuffixExecute3728); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_CHERRY_PICK.add(KW_CHERRY_PICK256);\n \n \t\t\t\t\tsnapshotId=(Token)match(input,Number,FOLLOW_Number_in_alterStatementSuffixExecute3732); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_Number.add(snapshotId);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: snapshotId, KW_CHERRY_PICK\n+\t\t\t\t\t// elements: KW_CHERRY_PICK, snapshotId\n \t\t\t\t\t// token labels: snapshotId\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10040,15 +10040,15 @@\n \t\t\t\t\t{\n \t\t\t\t\ttoTimestamp=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_alterStatementSuffixExecute3772); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(toTimestamp);\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: fromTimestamp, toTimestamp, KW_EXPIRE_SNAPSHOTS\n+\t\t\t\t\t// elements: fromTimestamp, KW_EXPIRE_SNAPSHOTS, toTimestamp\n \t\t\t\t\t// token labels: fromTimestamp, toTimestamp\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10295,15 +10295,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_alterStatementSuffixDropBranch3907);\n \t\t\tbranchName=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_identifier.add(branchName.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: ifExists, branchName\n+\t\t\t// elements: branchName, ifExists\n \t\t\t// token labels: \n \t\t\t// rule labels: branchName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -10460,15 +10460,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_retentionOfSnapshots.add(retentionOfSnapshots279.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: refRetain, retentionOfSnapshots, snapshotIdOfRef, branchName\n+\t\t\t// elements: snapshotIdOfRef, refRetain, retentionOfSnapshots, branchName\n \t\t\t// token labels: \n \t\t\t// rule labels: branchName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -10852,15 +10852,15 @@\n \n \t\t\tpushFollow(FOLLOW_timeUnitQualifiers_in_refRetain4146);\n \t\t\ttimeUnit=gHiveParser.timeUnitQualifiers();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_timeUnitQualifiers.add(timeUnit.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: timeUnit, maxRefAge\n+\t\t\t// elements: maxRefAge, timeUnit\n \t\t\t// token labels: maxRefAge\n \t\t\t// rule labels: retval, timeUnit\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -10990,15 +10990,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: maxSnapshotAge, timeUnit, minSnapshotsToKeep\n+\t\t\t// elements: timeUnit, minSnapshotsToKeep, maxSnapshotAge\n \t\t\t// token labels: maxSnapshotAge, minSnapshotsToKeep\n \t\t\t// rule labels: retval, timeUnit\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -11012,20 +11012,20 @@\n \t\t\t{\n \t\t\t\t// AlterClauseParser.g:546:8: ^( TOK_WITH_SNAPSHOT_RETENTION $minSnapshotsToKeep ( $maxSnapshotAge $timeUnit)? )\n \t\t\t\t{\n \t\t\t\tASTNode root_1 = (ASTNode)adaptor.nil();\n \t\t\t\troot_1 = (ASTNode)adaptor.becomeRoot((ASTNode)adaptor.create(TOK_WITH_SNAPSHOT_RETENTION, \"TOK_WITH_SNAPSHOT_RETENTION\"), root_1);\n \t\t\t\tadaptor.addChild(root_1, stream_minSnapshotsToKeep.nextNode());\n \t\t\t\t// AlterClauseParser.g:546:58: ( $maxSnapshotAge $timeUnit)?\n-\t\t\t\tif ( stream_maxSnapshotAge.hasNext()||stream_timeUnit.hasNext() ) {\n+\t\t\t\tif ( stream_timeUnit.hasNext()||stream_maxSnapshotAge.hasNext() ) {\n \t\t\t\t\tadaptor.addChild(root_1, stream_maxSnapshotAge.nextNode());\n \t\t\t\t\tadaptor.addChild(root_1, stream_timeUnit.nextTree());\n \t\t\t\t}\n-\t\t\t\tstream_maxSnapshotAge.reset();\n \t\t\t\tstream_timeUnit.reset();\n+\t\t\t\tstream_maxSnapshotAge.reset();\n \n \t\t\t\tadaptor.addChild(root_0, root_1);\n \t\t\t\t}\n \n \t\t\t}\n \n \n@@ -11258,15 +11258,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_refRetain.add(refRetain303.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: refRetain, snapshotIdOfRef, tagName\n+\t\t\t// elements: refRetain, tagName, snapshotIdOfRef\n \t\t\t// token labels: \n \t\t\t// rule labels: tagName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -11455,15 +11455,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: inFmt, serdeCls, outFmt, inDriver, outDriver\n+\t\t\t\t\t// elements: outFmt, inDriver, serdeCls, inFmt, outDriver\n \t\t\t\t\t// token labels: inFmt, inDriver, serdeCls, outDriver, outFmt\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11811,15 +11811,15 @@\n \n \t\t\tpushFollow(FOLLOW_dcProperties_in_alterDataConnectorSuffixProperties4528);\n \t\t\tdcProperties314=gHiveParser.dcProperties();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_dcProperties.add(dcProperties314.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: dcProperties, name\n+\t\t\t// elements: name, dcProperties\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -11912,15 +11912,15 @@\n \n \t\t\tpushFollow(FOLLOW_principalName_in_alterDataConnectorSuffixSetOwner4578);\n \t\t\tprincipalName317=gHiveParser.principalName();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_principalName.add(principalName317.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: dcName, principalName\n+\t\t\t// elements: principalName, dcName\n \t\t\t// token labels: \n \t\t\t// rule labels: dcName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -12012,15 +12012,15 @@\n \t\t\tKW_URL319=(Token)match(input,KW_URL,FOLLOW_KW_URL_in_alterDataConnectorSuffixSetUrl4626); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_KW_URL.add(KW_URL319);\n \n \t\t\tnewUri=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_alterDataConnectorSuffixSetUrl4630); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(newUri);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: newUri, dcName\n+\t\t\t// elements: dcName, newUri\n \t\t\t// token labels: newUri\n \t\t\t// rule labels: dcName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_CreateDDLParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_CreateDDLParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 CreateDDLParser.g 2024-09-26 14:10:42\n+// $ANTLR 3.5.2 CreateDDLParser.g 2025-01-31 11:23:54\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1135,15 +1135,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(format.getTree());\n \t\t\t\t\turi=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_likeTableOrFile84); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(uri);\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: uri, format\n+\t\t\t\t\t// elements: format, uri\n \t\t\t\t\t// token labels: uri\n \t\t\t\t\t// rule labels: format, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -1808,15 +1808,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableFileFormat, likeTableOrFile, selectStatementWithCTE, tableBuckets, tableSkewed, trans, tableRowFormat, tablePropertiesPrefixed, createTablePartitionSpec, columnNameTypeOrConstraintList, ifNotExists, tableComment, tableLocation, name, temp, ext\n+\t\t\t\t\t// elements: tableLocation, trans, ifNotExists, tableComment, temp, createTablePartitionSpec, tablePropertiesPrefixed, likeTableOrFile, tableRowFormat, columnNameTypeOrConstraintList, name, ext, tableFileFormat, tableSkewed, selectStatementWithCTE, tableBuckets\n \t\t\t\t\t// token labels: ext, temp, trans\n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -2294,15 +2294,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tablePropertiesPrefixed, tableFileFormat, tableSkewed, tableLocation, likeTableOrFile, mgd, name, createTablePartitionSpec, columnNameTypeOrConstraintList, ifNotExists, tableRowFormat, tableComment, tableBuckets, selectStatementWithCTE\n+\t\t\t\t\t// elements: tableFileFormat, likeTableOrFile, createTablePartitionSpec, selectStatementWithCTE, tableRowFormat, tableBuckets, ifNotExists, tableComment, mgd, tableLocation, name, columnNameTypeOrConstraintList, tableSkewed, tablePropertiesPrefixed\n \t\t\t\t\t// token labels: mgd\n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -2555,15 +2555,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_dcProperties.add(dcprops.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: ifNotExists, name, dcprops, dataConnectorComment, dataConnectorUrl, dataConnectorType\n+\t\t\t// elements: ifNotExists, dataConnectorType, dataConnectorComment, dcprops, name, dataConnectorUrl\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval, dcprops\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3046,15 +3046,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_dropDataConnectorStatement1229);\n \t\t\tidentifier66=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_identifier.add(identifier66.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: identifier, ifExists\n+\t\t\t// elements: ifExists, identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_FromClauseParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_FromClauseParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 FromClauseParser.g 2024-09-26 14:10:40\n+// $ANTLR 3.5.2 FromClauseParser.g 2025-01-31 11:23:52\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -3394,15 +3394,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: identifier, function, tableAlias\n+\t\t\t\t\t// elements: function, tableAlias, identifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3530,15 +3530,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: function, tableAlias, identifier\n+\t\t\t\t\t// elements: identifier, tableAlias, function\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3697,15 +3697,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: valuesClause, tableAlias, identifier\n+\t\t\t\t\t// elements: tableAlias, identifier, valuesClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -4002,24 +4002,24 @@\n \n \t\t\t}\n \n \t\t\tRPAREN114=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_tableBucketSample1292); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN114);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: expr, numerator, denominator\n-\t\t\t// token labels: numerator, denominator\n+\t\t\t// elements: denominator, numerator, expr\n+\t\t\t// token labels: denominator, numerator\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: expr\n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n-\t\t\tRewriteRuleTokenStream stream_numerator=new RewriteRuleTokenStream(adaptor,\"token numerator\",numerator);\n \t\t\tRewriteRuleTokenStream stream_denominator=new RewriteRuleTokenStream(adaptor,\"token denominator\",denominator);\n+\t\t\tRewriteRuleTokenStream stream_numerator=new RewriteRuleTokenStream(adaptor,\"token numerator\",numerator);\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_expr=new RewriteRuleSubtreeStream(adaptor,\"token expr\",list_expr);\n \t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t// 186:149: -> ^( TOK_TABLEBUCKETSAMPLE $numerator $denominator ( $expr)* )\n \t\t\t{\n \t\t\t\t// FromClauseParser.g:186:152: ^( TOK_TABLEBUCKETSAMPLE $numerator $denominator ( $expr)* )\n \t\t\t\t{\n@@ -4598,15 +4598,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(alias.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tabname, ts, props, alias, asOf\n+\t\t\t// elements: asOf, alias, ts, tabname, props\n \t\t\t// token labels: \n \t\t\t// rule labels: tabname, asOf, alias, retval, ts, props\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5288,15 +5288,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_viewName1859);\n \t\t\tview=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_identifier.add(view.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: view, db\n+\t\t\t// elements: db, view\n \t\t\t// token labels: \n \t\t\t// rule labels: view, db, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5562,15 +5562,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_orderByClause.add(orderByClause143.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: partitionByClause, orderByClause\n+\t\t\t\t\t// elements: orderByClause, partitionByClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -6431,15 +6431,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(alias.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: alias, ptfsrc, name, spec, expression\n+\t\t\t// elements: alias, name, expression, spec, ptfsrc\n \t\t\t// token labels: \n \t\t\t// rule labels: ptfsrc, name, alias, spec, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -7254,15 +7254,15 @@\n \n \t\t\t}\n \n \t\t\tRPAREN189=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_virtualTableSource2677); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN189);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: valuesClause, identifier\n+\t\t\t// elements: identifier, valuesClause\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_IdentifiersParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_IdentifiersParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 IdentifiersParser.g 2024-09-26 14:10:42\n+// $ANTLR 3.5.2 IdentifiersParser.g 2025-01-31 11:23:54\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -3966,15 +3966,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_selectExpression.add(str.getTree());\n \t\t\t}\n \n \t\t\tRPAREN79=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_trimFunction1438); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN79);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: trim_characters, trim_characters, str, str, str, trim_characters\n+\t\t\t// elements: str, trim_characters, str, trim_characters, trim_characters, str\n \t\t\t// token labels: \n \t\t\t// rule labels: str, trim_characters, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4785,15 +4785,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: functionName, functionName, ws, ws, ordBy, selectExpression, selectExpression, functionName, ws, selectExpression, functionName\n+\t\t\t\t\t// elements: selectExpression, selectExpression, functionName, ws, ws, functionName, selectExpression, ws, ordBy, functionName, functionName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: ordBy, ws, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -5271,15 +5271,15 @@\n \n \t\t\t}\n \n \t\t\tRPAREN107=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_castExpression2132); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN107);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: StringLiteral, expression, expression, toType, expression, StringLiteral\n+\t\t\t// elements: toType, expression, StringLiteral, expression, expression, StringLiteral\n \t\t\t// token labels: \n \t\t\t// rule labels: toType, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5642,15 +5642,15 @@\n \n \t\t\t}\n \n \t\t\tKW_END124=(Token)match(input,KW_END,FOLLOW_KW_END_in_whenExpression2357); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_KW_END.add(KW_END124);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, KW_WHEN\n+\t\t\t// elements: KW_WHEN, expression\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -6253,15 +6253,15 @@\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_expression.add(expression141.getTree());\n \t\t\tRPAREN142=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_extractExpression2666); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN142);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: timeUnit, expression\n+\t\t\t// elements: expression, timeUnit\n \t\t\t// token labels: \n \t\t\t// rule labels: retval, timeUnit\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -10338,15 +10338,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_subQueryExpression_in_precedenceSimilarExpression4385);\n \t\t\t\t\tsubQueryExpression253=subQueryExpression();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_subQueryExpression.add(subQueryExpression253.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_EXISTS, subQueryExpression\n+\t\t\t\t\t// elements: subQueryExpression, KW_EXISTS\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10993,15 +10993,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_quantifierType.add(quantifierType268.getTree());\n \t\t\tpushFollow(FOLLOW_subQueryExpression_in_precedenceSimilarExpressionQuantifierPredicate4718);\n \t\t\tsubQueryExpression269=subQueryExpression();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_subQueryExpression.add(subQueryExpression269.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: subQueryExpression, subQuerySelectorOperator, quantifierType\n+\t\t\t// elements: quantifierType, subQuerySelectorOperator, subQueryExpression\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -11438,15 +11438,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_precedenceRegexpOperator.add(precedenceRegexpOperator274.getTree());\n \t\t\t\t\tpushFollow(FOLLOW_precedenceBitwiseOrExpression_in_precedenceSimilarExpressionPartNot4900);\n \t\t\t\t\tnotExpr=precedenceBitwiseOrExpression();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_precedenceBitwiseOrExpression.add(notExpr.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: precedenceRegexpOperator, notExpr\n+\t\t\t\t\t// elements: notExpr, precedenceRegexpOperator\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: notExpr, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -12392,15 +12392,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_isCondition.add(isCondition300.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: precedenceEqualExpression, isCondition, precedenceEqualExpression\n+\t\t\t// elements: precedenceEqualExpression, precedenceEqualExpression, isCondition\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -13602,15 +13602,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_partitionSelectorOperator.add(partitionSelectorOperator333.getTree());\n \t\t\tpushFollow(FOLLOW_constant_in_partitionSelectorVal5709);\n \t\t\tconstant334=constant();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_constant.add(constant334.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: partitionSelectorOperator, constant, identifier\n+\t\t\t// elements: partitionSelectorOperator, identifier, constant\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_PrepareStatementParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_PrepareStatementParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 PrepareStatementParser.g 2024-09-26 14:10:42\n+// $ANTLR 3.5.2 PrepareStatementParser.g 2025-01-31 11:23:54\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1057,15 +1057,15 @@\n \n \t\t\tpushFollow(FOLLOW_queryStatementExpression_in_prepareStatement72);\n \t\t\tqueryStatementExpression4=gHiveParser.queryStatementExpression();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_queryStatementExpression.add(queryStatementExpression4.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: queryStatementExpression, identifier\n+\t\t\t// elements: identifier, queryStatementExpression\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_ResourcePlanParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_ResourcePlanParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 ResourcePlanParser.g 2024-09-26 14:10:42\n+// $ANTLR 3.5.2 ResourcePlanParser.g 2025-01-31 11:23:54\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -2848,15 +2848,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_rpAssignList_in_alterResourcePlanStatement675);\n \t\t\t\t\trpAssignList50=rpAssignList();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_rpAssignList.add(rpAssignList50.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, rpAssignList\n+\t\t\t\t\t// elements: rpAssignList, name\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -2894,15 +2894,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_rpUnassignList_in_alterResourcePlanStatement702);\n \t\t\t\t\trpUnassignList52=rpUnassignList();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_rpUnassignList.add(rpUnassignList52.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, rpUnassignList\n+\t\t\t\t\t// elements: rpUnassignList, name\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -2943,15 +2943,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_identifier_in_alterResourcePlanStatement733);\n \t\t\t\t\tnewName=gHiveParser.identifier();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_identifier.add(newName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: newName, name\n+\t\t\t\t\t// elements: name, newName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: newName, name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_newName=new RewriteRuleSubtreeStream(adaptor,\"rule newName\",newName!=null?newName.getTree():null);\n@@ -3062,15 +3062,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, enable, activate\n+\t\t\t\t\t// elements: enable, activate, name\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -3554,15 +3554,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_dropResourcePlanStatement998);\n \t\t\tname=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_identifier.add(name.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: ifExists, name\n+\t\t\t// elements: name, ifExists\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -4425,15 +4425,15 @@\n \n \t\t\tpushFollow(FOLLOW_triggerActionExpression_in_createTriggerStatement1362);\n \t\t\ttriggerActionExpression106=triggerActionExpression();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_triggerActionExpression.add(triggerActionExpression106.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: triggerExpression, triggerName, triggerActionExpression, rpName\n+\t\t\t// elements: triggerActionExpression, rpName, triggerName, triggerExpression\n \t\t\t// token labels: \n \t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -4693,15 +4693,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_triggerActionExpression_in_alterTriggerStatement1440);\n \t\t\t\t\ttriggerActionExpression113=triggerActionExpression();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_triggerActionExpression.add(triggerActionExpression113.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: triggerExpression, triggerActionExpression, triggerName, rpName\n+\t\t\t\t\t// elements: triggerExpression, rpName, triggerName, triggerActionExpression\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -4902,15 +4902,15 @@\n \t\t\t\t\tKW_FROM124=(Token)match(input,KW_FROM,FOLLOW_KW_FROM_in_alterTriggerStatement1581);  \n \t\t\t\t\tstream_KW_FROM.add(KW_FROM124);\n \n \t\t\t\t\tKW_UNMANAGED125=(Token)match(input,KW_UNMANAGED,FOLLOW_KW_UNMANAGED_in_alterTriggerStatement1583);  \n \t\t\t\t\tstream_KW_UNMANAGED.add(KW_UNMANAGED125);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: triggerName, rpName\n+\t\t\t\t\t// elements: rpName, triggerName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -5014,15 +5014,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_dropTriggerStatement1644);\n \t\t\ttriggerName=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_identifier.add(triggerName.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: rpName, triggerName\n+\t\t\t// elements: triggerName, rpName\n \t\t\t// token labels: \n \t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -5535,15 +5535,15 @@\n \n \t\t\tpushFollow(FOLLOW_poolAssignList_in_createPoolStatement1889);\n \t\t\tpoolAssignList145=poolAssignList();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_poolAssignList.add(poolAssignList145.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: rpName, poolAssignList, poolPath\n+\t\t\t// elements: poolPath, poolAssignList, rpName\n \t\t\t// token labels: \n \t\t\t// rule labels: rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -5706,15 +5706,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_poolAssignList_in_alterPoolStatement1958);\n \t\t\t\t\tpoolAssignList151=poolAssignList();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_poolAssignList.add(poolAssignList151.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: rpName, poolPath, poolAssignList\n+\t\t\t\t\t// elements: rpName, poolAssignList, poolPath\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -5751,15 +5751,15 @@\n \t\t\t\t\tKW_UNSET152=(Token)match(input,KW_UNSET,FOLLOW_KW_UNSET_in_alterPoolStatement1985);  \n \t\t\t\t\tstream_KW_UNSET.add(KW_UNSET152);\n \n \t\t\t\t\tKW_SCHEDULING_POLICY153=(Token)match(input,KW_SCHEDULING_POLICY,FOLLOW_KW_SCHEDULING_POLICY_in_alterPoolStatement1987);  \n \t\t\t\t\tstream_KW_SCHEDULING_POLICY.add(KW_SCHEDULING_POLICY153);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: rpName, poolPath\n+\t\t\t\t\t// elements: poolPath, rpName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -5972,15 +5972,15 @@\n \n \t\t\tpushFollow(FOLLOW_poolPath_in_dropPoolStatement2118);\n \t\t\tpoolPath161=poolPath();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_poolPath.add(poolPath161.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: rpName, poolPath\n+\t\t\t// elements: poolPath, rpName\n \t\t\t// token labels: \n \t\t\t// rule labels: rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -6224,15 +6224,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: name, order, path, unmanaged, rpName, mappingType\n+\t\t\t// elements: mappingType, rpName, path, unmanaged, name, order\n \t\t\t// token labels: mappingType, name, order\n \t\t\t// rule labels: path, rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleTokenStream stream_mappingType=new RewriteRuleTokenStream(adaptor,\"token mappingType\",mappingType);\n@@ -6499,15 +6499,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: mappingType, name, rpName, order, path, unmanaged\n+\t\t\t// elements: mappingType, unmanaged, name, order, rpName, path\n \t\t\t// token labels: mappingType, name, order\n \t\t\t// rule labels: path, rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleTokenStream stream_mappingType=new RewriteRuleTokenStream(adaptor,\"token mappingType\",mappingType);\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_SelectClauseParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_SelectClauseParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 SelectClauseParser.g 2024-09-26 14:10:39\n+// $ANTLR 3.5.2 SelectClauseParser.g 2025-01-31 11:23:52\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1176,15 +1176,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: QUERY_HINT, QUERY_HINT, selectList, selectTrfmClause, selectList, QUERY_HINT\n+\t\t\t\t\t// elements: QUERY_HINT, selectTrfmClause, QUERY_HINT, selectList, selectList, QUERY_HINT\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -1701,15 +1701,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_rowFormat.add(outSerde.getTree());\n \t\t\tpushFollow(FOLLOW_recordReader_in_selectTrfmClause357);\n \t\t\toutRec=gHiveParser.recordReader();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_recordReader.add(outRec.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: inRec, outSerde, selectExpressionList, outRec, aliasList, StringLiteral, columnNameTypeList, inSerde\n+\t\t\t// elements: StringLiteral, columnNameTypeList, inSerde, outRec, selectExpressionList, aliasList, inRec, outSerde\n \t\t\t// token labels: \n \t\t\t// rule labels: inRec, outRec, inSerde, outSerde, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -2321,15 +2321,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_rowFormat.add(outSerde.getTree());\n \t\t\tpushFollow(FOLLOW_recordReader_in_trfmClause622);\n \t\t\toutRec=gHiveParser.recordReader();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_recordReader.add(outRec.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: selectExpressionList, outSerde, inRec, columnNameTypeList, StringLiteral, aliasList, inSerde, outRec\n+\t\t\t// elements: selectExpressionList, columnNameTypeList, inSerde, outSerde, outRec, StringLiteral, inRec, aliasList\n \t\t\t// token labels: \n \t\t\t// rule labels: inRec, outRec, inSerde, outSerde, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3896,15 +3896,15 @@\n \t\t\t\t\tNumber77=(Token)match(input,Number,FOLLOW_Number_in_window_frame_start_boundary1097); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_Number.add(Number77);\n \n \t\t\t\t\tKW_PRECEDING78=(Token)match(input,KW_PRECEDING,FOLLOW_KW_PRECEDING_in_window_frame_start_boundary1099); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_PRECEDING.add(KW_PRECEDING78);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: Number, KW_PRECEDING\n+\t\t\t\t\t// elements: KW_PRECEDING, Number\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n"}]}
