{"diffoscope-json-version": 1, "source1": "/input1", "source2": "/input2", "unified_diff": null, "details": [{"source1": "zipinfo -v {}", "source2": "zipinfo -v {}", "unified_diff": "@@ -1376,15 +1376,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         ed3d1833\n+  32-bit CRC value (hex):                         76792329\n   compressed size:                                47644 bytes\n   uncompressed size:                              47644 bytes\n   length of filename:                             47 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1448,15 +1448,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         7c58f3ca\n+  32-bit CRC value (hex):                         085d21dc\n   compressed size:                                30210 bytes\n   uncompressed size:                              30210 bytes\n   length of filename:                             46 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1556,15 +1556,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         2bc76d8d\n+  32-bit CRC value (hex):                         da38c555\n   compressed size:                                29518 bytes\n   uncompressed size:                              29518 bytes\n   length of filename:                             54 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1592,15 +1592,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         e46be64a\n+  32-bit CRC value (hex):                         af6b4e4c\n   compressed size:                                386217 bytes\n   uncompressed size:                              386217 bytes\n   length of filename:                             70 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1628,15 +1628,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         1be2ae6d\n+  32-bit CRC value (hex):                         e5142d89\n   compressed size:                                387628 bytes\n   uncompressed size:                              387628 bytes\n   length of filename:                             62 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1700,15 +1700,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         7b681f7f\n+  32-bit CRC value (hex):                         21ca8c10\n   compressed size:                                2068462 bytes\n   uncompressed size:                              2068462 bytes\n   length of filename:                             47 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1736,15 +1736,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         3256b3b5\n+  32-bit CRC value (hex):                         b36960eb\n   compressed size:                                554655 bytes\n   uncompressed size:                              554655 bytes\n   length of filename:                             65 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1772,15 +1772,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         43774f29\n+  32-bit CRC value (hex):                         8b366b52\n   compressed size:                                150310 bytes\n   uncompressed size:                              150310 bytes\n   length of filename:                             63 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1808,15 +1808,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         0b38168e\n+  32-bit CRC value (hex):                         a6d3e9cd\n   compressed size:                                669381 bytes\n   uncompressed size:                              669381 bytes\n   length of filename:                             64 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1844,15 +1844,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         aafbdfc0\n+  32-bit CRC value (hex):                         e89351d4\n   compressed size:                                1418080 bytes\n   uncompressed size:                              1418080 bytes\n   length of filename:                             65 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1880,15 +1880,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         d0d86270\n+  32-bit CRC value (hex):                         a4e68ddf\n   compressed size:                                56565 bytes\n   uncompressed size:                              56565 bytes\n   length of filename:                             70 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1916,15 +1916,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         5c97168c\n+  32-bit CRC value (hex):                         d547a345\n   compressed size:                                287498 bytes\n   uncompressed size:                              287498 bytes\n   length of filename:                             66 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n@@ -1952,15 +1952,15 @@\n   minimum software version required to extract:   2.0\n   compression method:                             none (stored)\n   file security status:                           not encrypted\n   extended local header:                          yes\n   file last modified on (DOS date/time):          2098 Jan 1 00:00:00\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 local\n   file last modified on (UT extra field modtime): 1970 Jan 1 00:00:00 UTC\n-  32-bit CRC value (hex):                         d52e95fd\n+  32-bit CRC value (hex):                         0fd66499\n   compressed size:                                288697 bytes\n   uncompressed size:                              288697 bytes\n   length of filename:                             66 characters\n   length of extra field:                          9 bytes\n   length of file comment:                         0 characters\n   disk number on which file begins:               disk 1\n   apparent file type:                             binary\n"}, {"source1": "zipdetails --redact --scan --utc {}", "source2": "zipdetails --redact --scan --utc {}", "unified_diff": "@@ -1013,15 +1013,15 @@\n 0476E1 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 0476E3   Length              0005 (5)\n 0476E5   Flags               01 (1) 'Modification'\n 0476E6   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 0476EA PAYLOAD\n \n 053106 DATA DESCRIPTOR       08074B50 (134695760)\n-05310A CRC                   ED3D1833 (3980204083)\n+05310A CRC                   76792329 (1987650345)\n 05310E Compressed Size       0000BA1C (47644)\n 053112 Uncompressed Size     0000BA1C (47644)\n \n 053116 LOCAL HEADER #40      04034B50 (67324752)\n 05311A Extract Zip Spec      14 (20) '2.0'\n 05311B Extract OS            00 (0) 'MS-DOS'\n 05311C General Purpose Flag  0008 (8)\n@@ -1069,15 +1069,15 @@\n 0536C7 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 0536C9   Length              0005 (5)\n 0536CB   Flags               01 (1) 'Modification'\n 0536CC   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 0536D0 PAYLOAD\n \n 05ACD2 DATA DESCRIPTOR       08074B50 (134695760)\n-05ACD6 CRC                   7C58F3CA (2086204362)\n+05ACD6 CRC                   085D21DC (140321244)\n 05ACDA Compressed Size       00007602 (30210)\n 05ACDE Uncompressed Size     00007602 (30210)\n \n 05ACE2 LOCAL HEADER #42      04034B50 (67324752)\n 05ACE6 Extract Zip Spec      14 (20) '2.0'\n 05ACE7 Extract OS            00 (0) 'MS-DOS'\n 05ACE8 General Purpose Flag  0008 (8)\n@@ -1153,15 +1153,15 @@\n 05E2A9 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 05E2AB   Length              0005 (5)\n 05E2AD   Flags               01 (1) 'Modification'\n 05E2AE   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 05E2B2 PAYLOAD\n \n 065600 DATA DESCRIPTOR       08074B50 (134695760)\n-065604 CRC                   2BC76D8D (734489997)\n+065604 CRC                   DA38C555 (3661153621)\n 065608 Compressed Size       0000734E (29518)\n 06560C Uncompressed Size     0000734E (29518)\n \n 065610 LOCAL HEADER #45      04034B50 (67324752)\n 065614 Extract Zip Spec      14 (20) '2.0'\n 065615 Extract OS            00 (0) 'MS-DOS'\n 065616 General Purpose Flag  0008 (8)\n@@ -1181,15 +1181,15 @@\n 065674 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 065676   Length              0005 (5)\n 065678   Flags               01 (1) 'Modification'\n 065679   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 06567D PAYLOAD\n \n 0C3B26 DATA DESCRIPTOR       08074B50 (134695760)\n-0C3B2A CRC                   E46BE64A (3832276554)\n+0C3B2A CRC                   AF6B4E4C (2943045196)\n 0C3B2E Compressed Size       0005E4A9 (386217)\n 0C3B32 Uncompressed Size     0005E4A9 (386217)\n \n 0C3B36 LOCAL HEADER #46      04034B50 (67324752)\n 0C3B3A Extract Zip Spec      14 (20) '2.0'\n 0C3B3B Extract OS            00 (0) 'MS-DOS'\n 0C3B3C General Purpose Flag  0008 (8)\n@@ -1209,15 +1209,15 @@\n 0C3B92 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 0C3B94   Length              0005 (5)\n 0C3B96   Flags               01 (1) 'Modification'\n 0C3B97   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 0C3B9B PAYLOAD\n \n 1225C7 DATA DESCRIPTOR       08074B50 (134695760)\n-1225CB CRC                   1BE2AE6D (467840621)\n+1225CB CRC                   E5142D89 (3843304841)\n 1225CF Compressed Size       0005EA2C (387628)\n 1225D3 Uncompressed Size     0005EA2C (387628)\n \n 1225D7 LOCAL HEADER #47      04034B50 (67324752)\n 1225DB Extract Zip Spec      14 (20) '2.0'\n 1225DC Extract OS            00 (0) 'MS-DOS'\n 1225DD General Purpose Flag  0008 (8)\n@@ -1265,15 +1265,15 @@\n 138925 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 138927   Length              0005 (5)\n 138929   Flags               01 (1) 'Modification'\n 13892A   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 13892E PAYLOAD\n \n 33191C DATA DESCRIPTOR       08074B50 (134695760)\n-331920 CRC                   7B681F7F (2070421375)\n+331920 CRC                   21CA8C10 (566922256)\n 331924 Compressed Size       001F8FEE (2068462)\n 331928 Uncompressed Size     001F8FEE (2068462)\n \n 33192C LOCAL HEADER #49      04034B50 (67324752)\n 331930 Extract Zip Spec      14 (20) '2.0'\n 331931 Extract OS            00 (0) 'MS-DOS'\n 331932 General Purpose Flag  0008 (8)\n@@ -1293,15 +1293,15 @@\n 33198B Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 33198D   Length              0005 (5)\n 33198F   Flags               01 (1) 'Modification'\n 331990   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 331994 PAYLOAD\n \n 3B9033 DATA DESCRIPTOR       08074B50 (134695760)\n-3B9037 CRC                   3256B3B5 (844542901)\n+3B9037 CRC                   B36960EB (3010027755)\n 3B903B Compressed Size       0008769F (554655)\n 3B903F Uncompressed Size     0008769F (554655)\n \n 3B9043 LOCAL HEADER #50      04034B50 (67324752)\n 3B9047 Extract Zip Spec      14 (20) '2.0'\n 3B9048 Extract OS            00 (0) 'MS-DOS'\n 3B9049 General Purpose Flag  0008 (8)\n@@ -1321,15 +1321,15 @@\n 3B90A0 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 3B90A2   Length              0005 (5)\n 3B90A4   Flags               01 (1) 'Modification'\n 3B90A5   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 3B90A9 PAYLOAD\n \n 3DDBCF DATA DESCRIPTOR       08074B50 (134695760)\n-3DDBD3 CRC                   43774F29 (1131892521)\n+3DDBD3 CRC                   8B366B52 (2335599442)\n 3DDBD7 Compressed Size       00024B26 (150310)\n 3DDBDB Uncompressed Size     00024B26 (150310)\n \n 3DDBDF LOCAL HEADER #51      04034B50 (67324752)\n 3DDBE3 Extract Zip Spec      14 (20) '2.0'\n 3DDBE4 Extract OS            00 (0) 'MS-DOS'\n 3DDBE5 General Purpose Flag  0008 (8)\n@@ -1349,15 +1349,15 @@\n 3DDC3D Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 3DDC3F   Length              0005 (5)\n 3DDC41   Flags               01 (1) 'Modification'\n 3DDC42   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 3DDC46 PAYLOAD\n \n 48130B DATA DESCRIPTOR       08074B50 (134695760)\n-48130F CRC                   0B38168E (188225166)\n+48130F CRC                   A6D3E9CD (2798905805)\n 481313 Compressed Size       000A36C5 (669381)\n 481317 Uncompressed Size     000A36C5 (669381)\n \n 48131B LOCAL HEADER #52      04034B50 (67324752)\n 48131F Extract Zip Spec      14 (20) '2.0'\n 481320 Extract OS            00 (0) 'MS-DOS'\n 481321 General Purpose Flag  0008 (8)\n@@ -1377,15 +1377,15 @@\n 48137A Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 48137C   Length              0005 (5)\n 48137E   Flags               01 (1) 'Modification'\n 48137F   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 481383 PAYLOAD\n \n 5DB6E3 DATA DESCRIPTOR       08074B50 (134695760)\n-5DB6E7 CRC                   AAFBDFC0 (2868633536)\n+5DB6E7 CRC                   E89351D4 (3901968852)\n 5DB6EB Compressed Size       0015A360 (1418080)\n 5DB6EF Uncompressed Size     0015A360 (1418080)\n \n 5DB6F3 LOCAL HEADER #53      04034B50 (67324752)\n 5DB6F7 Extract Zip Spec      14 (20) '2.0'\n 5DB6F8 Extract OS            00 (0) 'MS-DOS'\n 5DB6F9 General Purpose Flag  0008 (8)\n@@ -1405,15 +1405,15 @@\n 5DB757 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 5DB759   Length              0005 (5)\n 5DB75B   Flags               01 (1) 'Modification'\n 5DB75C   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 5DB760 PAYLOAD\n \n 5E9455 DATA DESCRIPTOR       08074B50 (134695760)\n-5E9459 CRC                   D0D86270 (3503841904)\n+5E9459 CRC                   A4E68DDF (2766573023)\n 5E945D Compressed Size       0000DCF5 (56565)\n 5E9461 Uncompressed Size     0000DCF5 (56565)\n \n 5E9465 LOCAL HEADER #54      04034B50 (67324752)\n 5E9469 Extract Zip Spec      14 (20) '2.0'\n 5E946A Extract OS            00 (0) 'MS-DOS'\n 5E946B General Purpose Flag  0008 (8)\n@@ -1433,15 +1433,15 @@\n 5E94C5 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 5E94C7   Length              0005 (5)\n 5E94C9   Flags               01 (1) 'Modification'\n 5E94CA   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 5E94CE PAYLOAD\n \n 62F7D8 DATA DESCRIPTOR       08074B50 (134695760)\n-62F7DC CRC                   5C97168C (1553405580)\n+62F7DC CRC                   D547A345 (3578241861)\n 62F7E0 Compressed Size       0004630A (287498)\n 62F7E4 Uncompressed Size     0004630A (287498)\n \n 62F7E8 LOCAL HEADER #55      04034B50 (67324752)\n 62F7EC Extract Zip Spec      14 (20) '2.0'\n 62F7ED Extract OS            00 (0) 'MS-DOS'\n 62F7EE General Purpose Flag  0008 (8)\n@@ -1461,15 +1461,15 @@\n 62F848 Extra ID #1           5455 (21589) 'Extended Timestamp [UT]'\n 62F84A   Length              0005 (5)\n 62F84C   Flags               01 (1) 'Modification'\n 62F84D   Modification Time   00000000 (0) 'Thu Jan  1 00:00:00 1970'\n 62F851 PAYLOAD\n \n 67600A DATA DESCRIPTOR       08074B50 (134695760)\n-67600E CRC                   D52E95FD (3576600061)\n+67600E CRC                   0FD66499 (265708697)\n 676012 Compressed Size       000467B9 (288697)\n 676016 Uncompressed Size     000467B9 (288697)\n \n 67601A LOCAL HEADER #56      04034B50 (67324752)\n 67601E Extract Zip Spec      14 (20) '2.0'\n 67601F Extract OS            00 (0) 'MS-DOS'\n 676020 General Purpose Flag  0008 (8)\n@@ -2884,15 +2884,15 @@\n 687A26 Created OS            00 (0) 'MS-DOS'\n 687A27 Extract Zip Spec      14 (20) '2.0'\n 687A28 Extract OS            00 (0) 'MS-DOS'\n 687A29 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687A2B Compression Method    0000 (0) 'Stored'\n 687A2D Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687A31 CRC                   ED3D1833 (3980204083)\n+687A31 CRC                   76792329 (1987650345)\n 687A35 Compressed Size       0000BA1C (47644)\n 687A39 Uncompressed Size     0000BA1C (47644)\n 687A3D Filename Length       002F (47)\n 687A3F Extra Length          0009 (9)\n 687A41 Comment Length        0000 (0)\n 687A43 Disk Start            0000 (0)\n 687A45 Int File Attributes   0000 (0)\n@@ -2944,15 +2944,15 @@\n 687AEE Created OS            00 (0) 'MS-DOS'\n 687AEF Extract Zip Spec      14 (20) '2.0'\n 687AF0 Extract OS            00 (0) 'MS-DOS'\n 687AF1 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687AF3 Compression Method    0000 (0) 'Stored'\n 687AF5 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687AF9 CRC                   7C58F3CA (2086204362)\n+687AF9 CRC                   085D21DC (140321244)\n 687AFD Compressed Size       00007602 (30210)\n 687B01 Uncompressed Size     00007602 (30210)\n 687B05 Filename Length       002E (46)\n 687B07 Extra Length          0009 (9)\n 687B09 Comment Length        0000 (0)\n 687B0B Disk Start            0000 (0)\n 687B0D Int File Attributes   0000 (0)\n@@ -3034,15 +3034,15 @@\n 687C25 Created OS            00 (0) 'MS-DOS'\n 687C26 Extract Zip Spec      14 (20) '2.0'\n 687C27 Extract OS            00 (0) 'MS-DOS'\n 687C28 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687C2A Compression Method    0000 (0) 'Stored'\n 687C2C Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687C30 CRC                   2BC76D8D (734489997)\n+687C30 CRC                   DA38C555 (3661153621)\n 687C34 Compressed Size       0000734E (29518)\n 687C38 Uncompressed Size     0000734E (29518)\n 687C3C Filename Length       0036 (54)\n 687C3E Extra Length          0009 (9)\n 687C40 Comment Length        0000 (0)\n 687C42 Disk Start            0000 (0)\n 687C44 Int File Attributes   0000 (0)\n@@ -3064,15 +3064,15 @@\n 687C92 Created OS            00 (0) 'MS-DOS'\n 687C93 Extract Zip Spec      14 (20) '2.0'\n 687C94 Extract OS            00 (0) 'MS-DOS'\n 687C95 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687C97 Compression Method    0000 (0) 'Stored'\n 687C99 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687C9D CRC                   E46BE64A (3832276554)\n+687C9D CRC                   AF6B4E4C (2943045196)\n 687CA1 Compressed Size       0005E4A9 (386217)\n 687CA5 Uncompressed Size     0005E4A9 (386217)\n 687CA9 Filename Length       0046 (70)\n 687CAB Extra Length          0009 (9)\n 687CAD Comment Length        0000 (0)\n 687CAF Disk Start            0000 (0)\n 687CB1 Int File Attributes   0000 (0)\n@@ -3094,15 +3094,15 @@\n 687D0F Created OS            00 (0) 'MS-DOS'\n 687D10 Extract Zip Spec      14 (20) '2.0'\n 687D11 Extract OS            00 (0) 'MS-DOS'\n 687D12 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687D14 Compression Method    0000 (0) 'Stored'\n 687D16 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687D1A CRC                   1BE2AE6D (467840621)\n+687D1A CRC                   E5142D89 (3843304841)\n 687D1E Compressed Size       0005EA2C (387628)\n 687D22 Uncompressed Size     0005EA2C (387628)\n 687D26 Filename Length       003E (62)\n 687D28 Extra Length          0009 (9)\n 687D2A Comment Length        0000 (0)\n 687D2C Disk Start            0000 (0)\n 687D2E Int File Attributes   0000 (0)\n@@ -3154,15 +3154,15 @@\n 687DE7 Created OS            00 (0) 'MS-DOS'\n 687DE8 Extract Zip Spec      14 (20) '2.0'\n 687DE9 Extract OS            00 (0) 'MS-DOS'\n 687DEA General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687DEC Compression Method    0000 (0) 'Stored'\n 687DEE Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687DF2 CRC                   7B681F7F (2070421375)\n+687DF2 CRC                   21CA8C10 (566922256)\n 687DF6 Compressed Size       001F8FEE (2068462)\n 687DFA Uncompressed Size     001F8FEE (2068462)\n 687DFE Filename Length       002F (47)\n 687E00 Extra Length          0009 (9)\n 687E02 Comment Length        0000 (0)\n 687E04 Disk Start            0000 (0)\n 687E06 Int File Attributes   0000 (0)\n@@ -3184,15 +3184,15 @@\n 687E4D Created OS            00 (0) 'MS-DOS'\n 687E4E Extract Zip Spec      14 (20) '2.0'\n 687E4F Extract OS            00 (0) 'MS-DOS'\n 687E50 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687E52 Compression Method    0000 (0) 'Stored'\n 687E54 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687E58 CRC                   3256B3B5 (844542901)\n+687E58 CRC                   B36960EB (3010027755)\n 687E5C Compressed Size       0008769F (554655)\n 687E60 Uncompressed Size     0008769F (554655)\n 687E64 Filename Length       0041 (65)\n 687E66 Extra Length          0009 (9)\n 687E68 Comment Length        0000 (0)\n 687E6A Disk Start            0000 (0)\n 687E6C Int File Attributes   0000 (0)\n@@ -3214,15 +3214,15 @@\n 687EC5 Created OS            00 (0) 'MS-DOS'\n 687EC6 Extract Zip Spec      14 (20) '2.0'\n 687EC7 Extract OS            00 (0) 'MS-DOS'\n 687EC8 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687ECA Compression Method    0000 (0) 'Stored'\n 687ECC Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687ED0 CRC                   43774F29 (1131892521)\n+687ED0 CRC                   8B366B52 (2335599442)\n 687ED4 Compressed Size       00024B26 (150310)\n 687ED8 Uncompressed Size     00024B26 (150310)\n 687EDC Filename Length       003F (63)\n 687EDE Extra Length          0009 (9)\n 687EE0 Comment Length        0000 (0)\n 687EE2 Disk Start            0000 (0)\n 687EE4 Int File Attributes   0000 (0)\n@@ -3244,15 +3244,15 @@\n 687F3B Created OS            00 (0) 'MS-DOS'\n 687F3C Extract Zip Spec      14 (20) '2.0'\n 687F3D Extract OS            00 (0) 'MS-DOS'\n 687F3E General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687F40 Compression Method    0000 (0) 'Stored'\n 687F42 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687F46 CRC                   0B38168E (188225166)\n+687F46 CRC                   A6D3E9CD (2798905805)\n 687F4A Compressed Size       000A36C5 (669381)\n 687F4E Uncompressed Size     000A36C5 (669381)\n 687F52 Filename Length       0040 (64)\n 687F54 Extra Length          0009 (9)\n 687F56 Comment Length        0000 (0)\n 687F58 Disk Start            0000 (0)\n 687F5A Int File Attributes   0000 (0)\n@@ -3274,15 +3274,15 @@\n 687FB2 Created OS            00 (0) 'MS-DOS'\n 687FB3 Extract Zip Spec      14 (20) '2.0'\n 687FB4 Extract OS            00 (0) 'MS-DOS'\n 687FB5 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 687FB7 Compression Method    0000 (0) 'Stored'\n 687FB9 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-687FBD CRC                   AAFBDFC0 (2868633536)\n+687FBD CRC                   E89351D4 (3901968852)\n 687FC1 Compressed Size       0015A360 (1418080)\n 687FC5 Uncompressed Size     0015A360 (1418080)\n 687FC9 Filename Length       0041 (65)\n 687FCB Extra Length          0009 (9)\n 687FCD Comment Length        0000 (0)\n 687FCF Disk Start            0000 (0)\n 687FD1 Int File Attributes   0000 (0)\n@@ -3304,15 +3304,15 @@\n 68802A Created OS            00 (0) 'MS-DOS'\n 68802B Extract Zip Spec      14 (20) '2.0'\n 68802C Extract OS            00 (0) 'MS-DOS'\n 68802D General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 68802F Compression Method    0000 (0) 'Stored'\n 688031 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-688035 CRC                   D0D86270 (3503841904)\n+688035 CRC                   A4E68DDF (2766573023)\n 688039 Compressed Size       0000DCF5 (56565)\n 68803D Uncompressed Size     0000DCF5 (56565)\n 688041 Filename Length       0046 (70)\n 688043 Extra Length          0009 (9)\n 688045 Comment Length        0000 (0)\n 688047 Disk Start            0000 (0)\n 688049 Int File Attributes   0000 (0)\n@@ -3334,15 +3334,15 @@\n 6880A7 Created OS            00 (0) 'MS-DOS'\n 6880A8 Extract Zip Spec      14 (20) '2.0'\n 6880A9 Extract OS            00 (0) 'MS-DOS'\n 6880AA General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 6880AC Compression Method    0000 (0) 'Stored'\n 6880AE Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-6880B2 CRC                   5C97168C (1553405580)\n+6880B2 CRC                   D547A345 (3578241861)\n 6880B6 Compressed Size       0004630A (287498)\n 6880BA Uncompressed Size     0004630A (287498)\n 6880BE Filename Length       0042 (66)\n 6880C0 Extra Length          0009 (9)\n 6880C2 Comment Length        0000 (0)\n 6880C4 Disk Start            0000 (0)\n 6880C6 Int File Attributes   0000 (0)\n@@ -3364,15 +3364,15 @@\n 688120 Created OS            00 (0) 'MS-DOS'\n 688121 Extract Zip Spec      14 (20) '2.0'\n 688122 Extract OS            00 (0) 'MS-DOS'\n 688123 General Purpose Flag  0008 (8)\n        [Bit  3]              1 'Streamed'\n 688125 Compression Method    0000 (0) 'Stored'\n 688127 Modification Time     EC210000 (3961585664) 'Wed Jan  1 00:00:00 2098'\n-68812B CRC                   D52E95FD (3576600061)\n+68812B CRC                   0FD66499 (265708697)\n 68812F Compressed Size       000467B9 (288697)\n 688133 Uncompressed Size     000467B9 (288697)\n 688137 Filename Length       0042 (66)\n 688139 Extra Length          0009 (9)\n 68813B Comment Length        0000 (0)\n 68813D Disk Start            0000 (0)\n 68813F Int File Attributes   0000 (0)\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HintParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HintParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HintParser.g 2024-03-25 12:44:51\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HintParser.g 2025-01-31 11:33:02\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.hadoop.conf.Configuration;\n import org.apache.hadoop.hive.conf.HiveConf;\n \n \n@@ -839,15 +839,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: hintName, hintArgs\n+\t\t\t// elements: hintArgs, hintName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexer.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexer.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexer.g 2024-03-25 12:44:47\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexer.g 2025-01-31 11:32:58\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexerStandard.g 2024-03-25 12:44:51\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveLexerStandard.g 2025-01-31 11:33:01\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard_HiveLexerParent.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexerStandard_HiveLexerParent.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 HiveLexerParent.g 2024-03-25 12:44:51\n+// $ANTLR 3.5.2 HiveLexerParent.g 2025-01-31 11:33:01\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveLexer_HiveLexerParent.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveLexer_HiveLexerParent.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 HiveLexerParent.g 2024-03-25 12:44:48\n+// $ANTLR 3.5.2 HiveLexerParent.g 2025-01-31 11:32:58\n \n package org.apache.hadoop.hive.ql.parse;\n \n import org.apache.commons.lang3.StringUtils;\n \n \n import org.antlr.runtime.*;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveParser.g 2024-03-25 16:12:01\n+// $ANTLR 3.5.2 org/apache/hadoop/hive/ql/parse/HiveParser.g 2025-01-31 11:32:59\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1758,15 +1758,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_execStatement_in_explainStatement1599);\n \t\t\t\t\texecStatement7=execStatement();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_execStatement.add(execStatement7.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: explainOption, execStatement\n+\t\t\t\t\t// elements: execStatement, explainOption\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3110,15 +3110,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_inputFileFormat.add(inputFileFormat50.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: islocal, tab, path, inputFileFormat, isoverwrite\n+\t\t\t// elements: path, islocal, inputFileFormat, tab, isoverwrite\n \t\t\t// token labels: islocal, path, isoverwrite\n \t\t\t// rule labels: tab, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3263,15 +3263,15 @@\n \n \t\t\t}\n \n \t\t\tRPAREN54=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_replicationClause2181); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN54);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: replId, isMetadataOnly\n+\t\t\t// elements: isMetadataOnly, replId\n \t\t\t// token labels: replId, isMetadataOnly\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3408,15 +3408,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replicationClause.add(replicationClause58.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tab, replicationClause, path\n+\t\t\t// elements: replicationClause, path, tab\n \t\t\t// token labels: path\n \t\t\t// rule labels: tab, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3590,15 +3590,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableLocation.add(tableLocation62.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: ext, path, tableLocation, tab\n+\t\t\t// elements: path, tableLocation, tab, ext\n \t\t\t// token labels: ext, path\n \t\t\t// rule labels: tab, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3764,15 +3764,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replConfigs.add(replConf.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: dbPolicy, oldDbPolicy, replConf\n+\t\t\t// elements: dbPolicy, replConf, oldDbPolicy\n \t\t\t// token labels: \n \t\t\t// rule labels: oldDbPolicy, dbPolicy, retval, replConf\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4055,15 +4055,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replConfigs.add(replConf.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: sourceDbPolicy, dbName, replConf\n+\t\t\t// elements: replConf, sourceDbPolicy, dbName\n \t\t\t// token labels: \n \t\t\t// rule labels: dbName, sourceDbPolicy, retval, replConf\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4407,15 +4407,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: replTablesIncludeList, replTablesExcludeList\n+\t\t\t// elements: replTablesExcludeList, replTablesIncludeList\n \t\t\t// token labels: replTablesExcludeList, replTablesIncludeList\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4541,15 +4541,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replConfigs.add(replConf.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: replConf, dbName\n+\t\t\t// elements: dbName, replConf\n \t\t\t// token labels: \n \t\t\t// rule labels: dbName, retval, replConf\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -6313,15 +6313,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_dbProperties.add(dbprops.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ifNotExists, name, databaseComment, dbLocation, dbManagedLocation, dbprops\n+\t\t\t\t\t// elements: dbLocation, name, ifNotExists, databaseComment, dbManagedLocation, dbprops\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, dbprops, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -6497,15 +6497,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_dbProperties.add(dbprops.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, dbprops, dbConnectorName, databaseComment, ifNotExists\n+\t\t\t\t\t// elements: ifNotExists, databaseComment, dbprops, name, dbConnectorName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, dbprops, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -7257,15 +7257,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_restrictOrCascade.add(restrictOrCascade178.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: identifier, restrictOrCascade, ifExists\n+\t\t\t// elements: identifier, ifExists, restrictOrCascade\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -7696,15 +7696,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replicationClause.add(replicationClause193.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, ifExists, replicationClause, KW_PURGE\n+\t\t\t// elements: replicationClause, ifExists, tableName, KW_PURGE\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -7811,15 +7811,15 @@\n \t\t\tKW_SERDE195=(Token)match(input,KW_SERDE,FOLLOW_KW_SERDE_in_inputFileFormat4388); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_KW_SERDE.add(KW_SERDE195);\n \n \t\t\tserdeCls=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_inputFileFormat4392); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(serdeCls);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: inFmt, serdeCls\n+\t\t\t// elements: serdeCls, inFmt\n \t\t\t// token labels: inFmt, serdeCls\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -8792,15 +8792,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_extColumnName.add(extColumnName209.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: extColumnName, tableName, partitionSpec\n+\t\t\t// elements: extColumnName, partitionSpec, tableName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -9517,15 +9517,15 @@\n \t\t\t\t\tname=descFuncNames();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_descFuncNames.add(name.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, KW_EXTENDED\n+\t\t\t\t\t// elements: KW_EXTENDED, name\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9854,15 +9854,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: parttype, KW_COLUMNS, statsColumnName, noscan\n+\t\t\t\t\t// elements: KW_COLUMNS, parttype, statsColumnName, noscan\n \t\t\t\t\t// token labels: noscan\n \t\t\t\t\t// rule labels: statsColumnName, parttype, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10519,15 +10519,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showTablesFilterExpr.add(filter.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: filter, db_name, isExtended\n+\t\t\t\t\t// elements: db_name, filter, isExtended\n \t\t\t\t\t// token labels: isExtended\n \t\t\t\t\t// rule labels: filter, db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10820,15 +10820,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showStmtIdentifier.add(showStmtIdentifier250.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: showStmtIdentifier, db_name\n+\t\t\t\t\t// elements: db_name, showStmtIdentifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11025,15 +11025,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showStmtIdentifier.add(showStmtIdentifier261.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableName, db_name, showStmtIdentifier, KW_SORTED\n+\t\t\t\t\t// elements: db_name, KW_SORTED, tableName, showStmtIdentifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11107,15 +11107,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_showFunctionIdentifier.add(showFunctionIdentifier265.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_LIKE, showFunctionIdentifier\n+\t\t\t\t\t// elements: showFunctionIdentifier, KW_LIKE\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11242,15 +11242,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause271.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: orderByClause, whereClause, limitClause, partitionSpec, tabName\n+\t\t\t\t\t// elements: orderByClause, limitClause, tabName, partitionSpec, whereClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: tabName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11547,15 +11547,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec284.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: partitionSpec, db_name, showStmtIdentifier\n+\t\t\t\t\t// elements: db_name, showStmtIdentifier, partitionSpec\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: db_name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -11628,15 +11628,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableName, prptyName\n+\t\t\t\t\t// elements: prptyName, tableName\n \t\t\t\t\t// token labels: prptyName\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -12371,15 +12371,15 @@\n \t\t\t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause303.getTree());\n \t\t\t\t\t\t\t\t\t}\n \t\t\t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t\t\t}\n \n \t\t\t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t\t\t// elements: dbName, compactionType, limitClause, compactionPool, compactionStatus, orderByClause\n+\t\t\t\t\t\t\t// elements: compactionPool, orderByClause, limitClause, dbName, compactionStatus, compactionType\n \t\t\t\t\t\t\t// token labels: \n \t\t\t\t\t\t\t// rule labels: dbName, retval\n \t\t\t\t\t\t\t// token list labels: \n \t\t\t\t\t\t\t// rule list labels: \n \t\t\t\t\t\t\t// wildcard labels: \n \t\t\t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\t\t\tretval.tree = root_0;\n@@ -12823,15 +12823,15 @@\n \t\t\t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause308.getTree());\n \t\t\t\t\t\t\t\t\t}\n \t\t\t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t\t\t}\n \n \t\t\t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t\t\t// elements: limitClause, compactionPool, parttype, compactionType, compactionStatus, orderByClause\n+\t\t\t\t\t\t\t// elements: parttype, compactionPool, compactionType, limitClause, orderByClause, compactionStatus\n \t\t\t\t\t\t\t// token labels: \n \t\t\t\t\t\t\t// rule labels: parttype, retval\n \t\t\t\t\t\t\t// token list labels: \n \t\t\t\t\t\t\t// rule list labels: \n \t\t\t\t\t\t\t// wildcard labels: \n \t\t\t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\t\t\tretval.tree = root_0;\n@@ -13475,15 +13475,15 @@\n \t\t\t\t\tEQUAL322=(Token)match(input,EQUAL,FOLLOW_EQUAL_in_showTablesFilterExpr6036); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_EQUAL.add(EQUAL322);\n \n \t\t\t\t\tStringLiteral323=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_showTablesFilterExpr6038); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(StringLiteral323);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: identifier, StringLiteral\n+\t\t\t\t\t// elements: StringLiteral, identifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -13980,15 +13980,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec340.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, partitionSpec\n+\t\t\t// elements: partitionSpec, tableName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14454,15 +14454,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_withGrantOption.add(withGrantOption352.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: privilegeObject, withGrantOption, principalSpecification, privList\n+\t\t\t// elements: privilegeObject, withGrantOption, privList, principalSpecification\n \t\t\t// token labels: \n \t\t\t// rule labels: privList, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14612,15 +14612,15 @@\n \n \t\t\tpushFollow(FOLLOW_principalSpecification_in_revokePrivileges6509);\n \t\t\tprincipalSpecification358=principalSpecification();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_principalSpecification.add(principalSpecification358.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: grantOptionFor, principalSpecification, privilegeObject, privilegeList\n+\t\t\t// elements: principalSpecification, privilegeObject, privilegeList, grantOptionFor\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14803,15 +14803,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_withAdminOption.add(withAdminOption366.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: principalSpecification, withAdminOption, identifier\n+\t\t\t// elements: principalSpecification, identifier, withAdminOption\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -14999,15 +14999,15 @@\n \n \t\t\tpushFollow(FOLLOW_principalSpecification_in_revokeRole6639);\n \t\t\tprincipalSpecification374=principalSpecification();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_principalSpecification.add(principalSpecification374.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: adminOptionFor, identifier, principalSpecification\n+\t\t\t// elements: principalSpecification, identifier, adminOptionFor\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -15634,15 +15634,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_privilegeIncludeColObject.add(privilegeIncludeColObject391.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: principalName, privilegeIncludeColObject\n+\t\t\t// elements: privilegeIncludeColObject, principalName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -17184,15 +17184,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec414.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableName, partitionSpec, cols\n+\t\t\t\t\t// elements: partitionSpec, tableName, cols\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: cols, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -18924,15 +18924,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: opt, tableName, repair, partitionSelectorSpec\n+\t\t\t// elements: partitionSelectorSpec, tableName, repair, opt\n \t\t\t// token labels: repair, opt\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -19496,15 +19496,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_resourceList.add(rList.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: rList, functionIdentifier, rList, functionIdentifier, StringLiteral, StringLiteral\n+\t\t\t// elements: rList, functionIdentifier, StringLiteral, rList, functionIdentifier, StringLiteral\n \t\t\t// token labels: \n \t\t\t// rule labels: rList, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -19946,15 +19946,15 @@\n \n \t\t\tpushFollow(FOLLOW_expression_in_createMacroStatement8373);\n \t\t\texpression485=expression();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_expression.add(expression485.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: Identifier, columnNameTypeList, expression\n+\t\t\t// elements: columnNameTypeList, expression, Identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -20331,15 +20331,15 @@\n \n \t\t\tpushFollow(FOLLOW_selectStatementWithCTE_in_createViewStatement8538);\n \t\t\tselectStatementWithCTE502=selectStatementWithCTE();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_selectStatementWithCTE.add(selectStatementWithCTE502.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: orReplace, viewPartition, columnNameCommentList, tableComment, tablePropertiesPrefixed, ifNotExists, name, selectStatementWithCTE\n+\t\t\t// elements: selectStatementWithCTE, tablePropertiesPrefixed, ifNotExists, name, orReplace, viewPartition, columnNameCommentList, tableComment\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -21506,15 +21506,15 @@\n \n \t\t\tpushFollow(FOLLOW_selectStatementWithCTE_in_createMaterializedViewStatement9066);\n \t\t\tselectStatementWithCTE547=selectStatementWithCTE();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_selectStatementWithCTE.add(selectStatementWithCTE547.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: name, rewriteDisabled, viewPartition, viewOrganization, selectStatementWithCTE, tableRowFormat, ifNotExists, tableComment, tableLocation, tableFileFormat, tablePropertiesPrefixed\n+\t\t\t// elements: tableComment, tableRowFormat, tablePropertiesPrefixed, name, rewriteDisabled, viewOrganization, viewPartition, ifNotExists, tableFileFormat, tableLocation, selectStatementWithCTE\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -21846,15 +21846,15 @@\n \n \t\t\tpushFollow(FOLLOW_definedAsSpec_in_createScheduledQueryStatement9331);\n \t\t\tdefinedAsSpec559=definedAsSpec();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_definedAsSpec.add(definedAsSpec559.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: enableSpecification, executedAsSpec, name, scheduleSpec, definedAsSpec\n+\t\t\t// elements: definedAsSpec, name, enableSpecification, executedAsSpec, scheduleSpec\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -22490,15 +22490,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: value, offsetTs, qualifier\n+\t\t\t\t\t// elements: qualifier, offsetTs, value\n \t\t\t\t\t// token labels: offsetTs, value\n \t\t\t\t\t// rule labels: qualifier, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -23236,15 +23236,15 @@\n \n \t\t\t\t\t}\n \n \t\t\t\t\tRPAREN589=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_createTablePartitionSpec10035); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN589);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: opt1, opt2\n+\t\t\t\t\t// elements: opt2, opt1\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: opt1, opt2, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -24576,15 +24576,15 @@\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_columnName.add(columnName630.getTree());\n \t\t\t\t\tRPAREN631=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_partitionTransformType10536); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN631);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: value, columnName\n+\t\t\t\t\t// elements: columnName, value\n \t\t\t\t\t// token labels: value\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -24751,15 +24751,15 @@\n \t\t\tnum=(Token)match(input,Number,FOLLOW_Number_in_tableBuckets10626); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_Number.add(num);\n \n \t\t\tKW_BUCKETS641=(Token)match(input,KW_BUCKETS,FOLLOW_KW_BUCKETS_in_tableBuckets10628); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_KW_BUCKETS.add(KW_BUCKETS641);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: num, sortCols, bucketCols\n+\t\t\t// elements: num, bucketCols, sortCols\n \t\t\t// token labels: num\n \t\t\t// rule labels: bucketCols, sortCols, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -25024,15 +25024,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_storedAsDirs.add(storedAsDirs652.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: skewedCols, skewedValues, storedAsDirs\n+\t\t\t// elements: skewedValues, skewedCols, storedAsDirs\n \t\t\t// token labels: \n \t\t\t// rule labels: skewedCols, skewedValues, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -25653,15 +25653,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableProperties.add(serdeprops.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: serdeprops, name\n+\t\t\t// elements: name, serdeprops\n \t\t\t// token labels: name\n \t\t\t// rule labels: serdeprops, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -26546,15 +26546,15 @@\n \t\t\tEQUAL685=(Token)match(input,EQUAL,FOLLOW_EQUAL_in_keyValueProperty11357); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_EQUAL.add(EQUAL685);\n \n \t\t\tvalue=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_keyValueProperty11361); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(value);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: key, value\n+\t\t\t// elements: value, key\n \t\t\t// token labels: value, key\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -26760,15 +26760,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: fldIdnt, fldEscape\n+\t\t\t// elements: fldEscape, fldIdnt\n \t\t\t// token labels: fldIdnt, fldEscape\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -27446,15 +27446,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: inDriver, inFmt, outDriver, outFmt\n+\t\t\t\t\t// elements: inFmt, inDriver, outFmt, outDriver\n \t\t\t\t\t// token labels: inFmt, inDriver, outDriver, outFmt\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -27557,15 +27557,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(fileformat.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: storageHandler, fileformat, serdeprops\n+\t\t\t\t\t// elements: storageHandler, serdeprops, fileformat\n \t\t\t\t\t// token labels: storageHandler\n \t\t\t\t\t// rule labels: serdeprops, fileformat, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -27675,15 +27675,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(fileformat.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: genericSpec, serdeprops, fileformat\n+\t\t\t\t\t// elements: genericSpec, fileformat, serdeprops\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: serdeprops, genericSpec, fileformat, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -29779,15 +29779,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate765.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintOptsCreate, constraintName, constraintOptsCreate\n+\t\t\t// elements: constraintOptsCreate, constraintOptsCreate, constraintName\n \t\t\t// token labels: \n \t\t\t// rule labels: constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -29926,15 +29926,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsAlter.add(constraintOptsAlter768.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintOptsAlter, constraintName\n+\t\t\t// elements: constraintName, constraintOptsAlter\n \t\t\t// token labels: \n \t\t\t// rule labels: constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -30117,15 +30117,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_tableConstraintType.add(tableConstraintType771.getTree());\n \t\t\tpushFollow(FOLLOW_columnParenthesesList_in_pkUkConstraint12829);\n \t\t\tpkCols=columnParenthesesList();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_columnParenthesesList.add(pkCols.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableConstraintType, pkCols\n+\t\t\t// elements: pkCols, tableConstraintType\n \t\t\t// token labels: \n \t\t\t// rule labels: pkCols, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -30379,24 +30379,24 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate780.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tabName, parCols, fkCols, constraintOptsCreate, constraintOptsCreate, constraintName, parCols, fkCols, tabName\n+\t\t\t// elements: parCols, fkCols, constraintName, tabName, constraintOptsCreate, parCols, constraintOptsCreate, fkCols, tabName\n \t\t\t// token labels: \n-\t\t\t// rule labels: tabName, parCols, fkCols, constraintName, retval\n+\t\t\t// rule labels: parCols, tabName, fkCols, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n-\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_parCols=new RewriteRuleSubtreeStream(adaptor,\"rule parCols\",parCols!=null?parCols.getTree():null);\n+\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_fkCols=new RewriteRuleSubtreeStream(adaptor,\"rule fkCols\",fkCols!=null?fkCols.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_constraintName=new RewriteRuleSubtreeStream(adaptor,\"rule constraintName\",constraintName!=null?constraintName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t// 2222:5: -> {$constraintName.tree != null}? ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) $fkCols $tabName $parCols ( constraintOptsCreate )? )\n \t\t\tif ((constraintName!=null?((ASTNode)constraintName.getTree()):null) != null) {\n@@ -30566,15 +30566,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsAlter.add(constraintOptsAlter785.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintName, fkCols, parCols, constraintOptsAlter, tabName\n+\t\t\t// elements: fkCols, parCols, constraintName, tabName, constraintOptsAlter\n \t\t\t// token labels: \n \t\t\t// rule labels: parCols, tabName, fkCols, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -31513,15 +31513,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_nullOrdering.add(nullSpec.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: identifier, nullSpec, identifier, identifier, identifier, nullSpec, identifier, identifier, nullSpec\n+\t\t\t// elements: identifier, identifier, identifier, identifier, nullSpec, nullSpec, nullSpec, identifier, identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: nullSpec, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -31845,15 +31845,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colName, comment\n+\t\t\t// elements: comment, colName\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -32116,15 +32116,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_nullOrdering.add(nullSpec.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, expression, expression, orderSpec, nullSpec, expression, nullSpec, orderSpec, expression, orderSpec, expression, orderSpec, expression, expression, orderSpec\n+\t\t\t// elements: nullSpec, expression, expression, nullSpec, orderSpec, orderSpec, expression, orderSpec, orderSpec, expression, expression, orderSpec, expression, expression, expression\n \t\t\t// token labels: \n \t\t\t// rule labels: orderSpec, nullSpec, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -32373,15 +32373,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colName, colType, comment, colName, colType\n+\t\t\t// elements: colType, colName, comment, colType, colName\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -33073,15 +33073,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: comment, colType, columnConstraint, colName\n+\t\t\t// elements: colName, comment, colType, columnConstraint\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -33425,24 +33425,24 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate827.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintName, tabName, colName, tabName, constraintOptsCreate, constraintOptsCreate, colName\n+\t\t\t// elements: colName, constraintOptsCreate, constraintName, tabName, colName, tabName, constraintOptsCreate\n \t\t\t// token labels: \n-\t\t\t// rule labels: tabName, colName, constraintName, retval\n+\t\t\t// rule labels: colName, tabName, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n-\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_colName=new RewriteRuleSubtreeStream(adaptor,\"rule colName\",colName!=null?colName.getTree():null);\n+\t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_constraintName=new RewriteRuleSubtreeStream(adaptor,\"rule constraintName\",constraintName!=null?constraintName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t// 2397:5: -> {$constraintName.tree != null}? ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) ^( TOK_TABCOLNAME ) $tabName ^( TOK_TABCOLNAME $colName) ( constraintOptsCreate )? )\n \t\t\tif ((constraintName!=null?((ASTNode)constraintName.getTree()):null) != null) {\n \t\t\t\t// org/apache/hadoop/hive/ql/parse/HiveParser.g:2398:13: ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) ^( TOK_TABCOLNAME ) $tabName ^( TOK_TABCOLNAME $colName) ( constraintOptsCreate )? )\n@@ -33623,15 +33623,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsCreate.add(constraintOptsCreate830.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: constraintOptsCreate, constraintName, constraintOptsCreate\n+\t\t\t// elements: constraintOptsCreate, constraintOptsCreate, constraintName\n \t\t\t// token labels: \n \t\t\t// rule labels: constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -33986,24 +33986,24 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_constraintOptsAlter.add(constraintOptsAlter837.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colName, tabName, tabName, constraintName, constraintOptsAlter, colName, constraintOptsAlter\n+\t\t\t// elements: constraintName, constraintOptsAlter, tabName, constraintOptsAlter, tabName, colName, colName\n \t\t\t// token labels: \n-\t\t\t// rule labels: colName, tabName, constraintName, retval\n+\t\t\t// rule labels: tabName, colName, constraintName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n-\t\t\tRewriteRuleSubtreeStream stream_colName=new RewriteRuleSubtreeStream(adaptor,\"rule colName\",colName!=null?colName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_tabName=new RewriteRuleSubtreeStream(adaptor,\"rule tabName\",tabName!=null?tabName.getTree():null);\n+\t\t\tRewriteRuleSubtreeStream stream_colName=new RewriteRuleSubtreeStream(adaptor,\"rule colName\",colName!=null?colName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_constraintName=new RewriteRuleSubtreeStream(adaptor,\"rule constraintName\",constraintName!=null?constraintName.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t// 2422:5: -> {$constraintName.tree != null}? ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) ^( TOK_TABCOLNAME ) $tabName ^( TOK_TABCOLNAME $colName) ( constraintOptsAlter )? )\n \t\t\tif ((constraintName!=null?((ASTNode)constraintName.getTree()):null) != null) {\n \t\t\t\t// org/apache/hadoop/hive/ql/parse/HiveParser.g:2423:13: ^( TOK_FOREIGN_KEY ^( TOK_CONSTRAINT_NAME $constraintName) ^( TOK_TABCOLNAME ) $tabName ^( TOK_TABCOLNAME $colName) ( constraintOptsAlter )? )\n@@ -35272,15 +35272,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colName, colType, comment, colType, colName\n+\t\t\t// elements: colType, colName, colType, colName, comment\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -38672,27 +38672,27 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_qualifyClause.add(q.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: h, g, q, win, f, s, w\n+\t\t\t\t\t// elements: win, w, f, g, s, h, q\n \t\t\t\t\t// token labels: \n-\t\t\t\t\t// rule labels: q, s, f, g, w, h, win, retval\n+\t\t\t\t\t// rule labels: q, s, f, w, g, h, win, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_q=new RewriteRuleSubtreeStream(adaptor,\"rule q\",q!=null?q.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_s=new RewriteRuleSubtreeStream(adaptor,\"rule s\",s!=null?s.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_f=new RewriteRuleSubtreeStream(adaptor,\"rule f\",f!=null?f.getTree():null);\n-\t\t\t\t\tRewriteRuleSubtreeStream stream_g=new RewriteRuleSubtreeStream(adaptor,\"rule g\",g!=null?g.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_w=new RewriteRuleSubtreeStream(adaptor,\"rule w\",w!=null?w.getTree():null);\n+\t\t\t\t\tRewriteRuleSubtreeStream stream_g=new RewriteRuleSubtreeStream(adaptor,\"rule g\",g!=null?g.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_h=new RewriteRuleSubtreeStream(adaptor,\"rule h\",h!=null?h.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_win=new RewriteRuleSubtreeStream(adaptor,\"rule win\",win!=null?win.getTree():null);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t\t\t// 2640:4: -> ^( TOK_QUERY ( $f)? ^( TOK_INSERT ^( TOK_DESTINATION ^( TOK_DIR TOK_TMP_FILE ) ) $s ( $w)? ( $g)? ( $h)? ( $win)? ( $q)? ) )\n \t\t\t\t\t{\n@@ -38989,15 +38989,15 @@\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((c!=null?((ASTNode)c.getTree()):null));\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((d!=null?((ASTNode)d.getTree()):null));\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((sort!=null?((ASTNode)sort.getTree()):null));\n \t\t\t   (a!=null?((ASTNode)a.getTree()):null).getFirstChildWithType(TOK_INSERT).addChild((l!=null?((ASTNode)l.getTree()):null));\n \t\t\t   }\n \t\t\t   }\n \t\t\t// AST REWRITE\n-\t\t\t// elements: sort, l, c, d, o\n+\t\t\t// elements: c, d, o, l, sort\n \t\t\t// token labels: \n \t\t\t// rule labels: c, d, sort, l, retval, o\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -39186,15 +39186,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_setOperator.add(u.getTree());\n \t\t\t\t\tpushFollow(FOLLOW_atomSelectStatement_in_setOpSelectStatement17080);\n \t\t\t\t\tb=atomSelectStatement();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_atomSelectStatement.add(b.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: b, u, b, b, b, u\n+\t\t\t\t\t// elements: b, b, u, u, b, b\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: b, u, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -39904,15 +39904,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause965.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: lateralView, window_clause, orderByClause, groupByClause, qualifyClause, distributeByClause, sortByClause, whereClause, selectClause, insertClause, havingClause, clusterByClause, limitClause\n+\t\t\t\t\t// elements: whereClause, distributeByClause, clusterByClause, orderByClause, limitClause, window_clause, groupByClause, selectClause, sortByClause, lateralView, insertClause, havingClause, qualifyClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40229,15 +40229,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_limitClause.add(limitClause977.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: havingClause, lateralView, qualifyClause, whereClause, selectClause, orderByClause, clusterByClause, window_clause, groupByClause, distributeByClause, limitClause, sortByClause\n+\t\t\t\t\t// elements: havingClause, distributeByClause, sortByClause, lateralView, groupByClause, qualifyClause, selectClause, orderByClause, clusterByClause, window_clause, limitClause, whereClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40479,15 +40479,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_ifNotExists.add(ifNotExists981.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ifNotExists, destination\n+\t\t\t\t\t// elements: destination, ifNotExists\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40759,15 +40759,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_tableFileFormat.add(tableFileFormat991.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: local, tableRowFormat, StringLiteral, tableFileFormat\n+\t\t\t\t\t// elements: StringLiteral, tableFileFormat, local, tableRowFormat\n \t\t\t\t\t// token labels: local\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -40993,15 +40993,15 @@\n \n \t\t\t\t\tnum=(Token)match(input,Number,FOLLOW_Number_in_limitClause18297); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_Number.add(num);\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: num, offset\n+\t\t\t\t\t// elements: offset, num\n \t\t\t\t\t// token labels: offset, num\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -43035,15 +43035,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_expression.add(expression1056.getTree());\n \t\t\tpushFollow(FOLLOW_whenClauses_in_mergeStatement18982);\n \t\t\twhenClauses1057=whenClauses();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_whenClauses.add(whenClauses1057.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: identifier, expression, QUERY_HINT, tableName, joinSourcePart, whenClauses\n+\t\t\t// elements: joinSourcePart, expression, tableName, whenClauses, identifier, QUERY_HINT\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -43344,15 +43344,15 @@\n \n \t\t\tpushFollow(FOLLOW_valueRowConstructor_in_whenNotMatchedClause19090);\n \t\t\tvalueRowConstructor1069=valueRowConstructor();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_valueRowConstructor.add(valueRowConstructor1069.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: targetCols, expression, valueRowConstructor\n+\t\t\t// elements: valueRowConstructor, targetCols, expression\n \t\t\t// token labels: \n \t\t\t// rule labels: targetCols, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -43476,15 +43476,15 @@\n \n \t\t\tpushFollow(FOLLOW_updateOrDelete_in_whenMatchedAndClause19147);\n \t\t\tupdateOrDelete1075=updateOrDelete();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_updateOrDelete.add(updateOrDelete1075.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, updateOrDelete\n+\t\t\t// elements: updateOrDelete, expression\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -44586,551 +44586,551 @@\n \t\t}\n \t\t}\n \n \t}\n \t// $ANTLR end synpred24_HiveParser\n \n \t// Delegated rules\n-\tpublic HiveParser_AlterClauseParser.refRetain_return refRetain() throws RecognitionException { return gAlterClauseParser.refRetain(); }\n+\tpublic HiveParser_ResourcePlanParser.alterPoolStatement_return alterPoolStatement() throws RecognitionException { return gResourcePlanParser.alterPoolStatement(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetOwner_return alterDataConnectorSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetOwner(); }\n+\tpublic HiveParser_ResourcePlanParser.dropPoolStatement_return dropPoolStatement() throws RecognitionException { return gResourcePlanParser.dropPoolStatement(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectClause_return selectClause() throws RecognitionException { return gSelectClauseParser.selectClause(); }\n+\tpublic HiveParser_IdentifiersParser.orderByClause_return orderByClause() throws RecognitionException { return gIdentifiersParser.orderByClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalQualifiers_return intervalQualifiers() throws RecognitionException { return gIdentifiersParser.intervalQualifiers(); }\n+\tpublic HiveParser_FromClauseParser.valuesClause_return valuesClause() throws RecognitionException { return gFromClauseParser.valuesClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceStarOperator_return precedenceStarOperator() throws RecognitionException { return gIdentifiersParser.precedenceStarOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorStatementSuffix_return alterDataConnectorStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDataConnectorStatementSuffix(); }\n \n-\tpublic HiveParser_IdentifiersParser.functionIdentifier_return functionIdentifier() throws RecognitionException { return gIdentifiersParser.functionIdentifier(); }\n+\tpublic HiveParser_IdentifiersParser.groupby_expression_return groupby_expression() throws RecognitionException { return gIdentifiersParser.groupby_expression(); }\n \n-\tpublic HiveParser_IdentifiersParser.subQuerySelectorOperator_return subQuerySelectorOperator() throws RecognitionException { return gIdentifiersParser.subQuerySelectorOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUnArchive_return alterStatementSuffixUnArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUnArchive(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionsInParenthesis_return expressionsInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsInParenthesis(isStruct, forceStruct); }\n+\tpublic HiveParser_FromClauseParser.whereClause_return whereClause() throws RecognitionException { return gFromClauseParser.whereClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.null_treatment_return null_treatment() throws RecognitionException { return gIdentifiersParser.null_treatment(); }\n+\tpublic HiveParser_AlterClauseParser.alterTableStatementSuffix_return alterTableStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterTableStatementSuffix(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropBranch_return alterStatementSuffixDropBranch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropBranch(); }\n+\tpublic HiveParser_IdentifiersParser.functionIdentifier_return functionIdentifier() throws RecognitionException { return gIdentifiersParser.functionIdentifier(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStats_return alterStatementSuffixUpdateStats(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStats(partition); }\n+\tpublic HiveParser_IdentifiersParser.booleanValueTok_return booleanValueTok() throws RecognitionException { return gIdentifiersParser.booleanValueTok(); }\n \n-\tpublic HiveParser_FromClauseParser.tableAllColumns_return tableAllColumns() throws RecognitionException { return gFromClauseParser.tableAllColumns(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetLocation_return alterDatabaseSuffixSetLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetLocation(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedencePlusOperator_return precedencePlusOperator() throws RecognitionException { return gIdentifiersParser.precedencePlusOperator(); }\n+\tpublic HiveParser_CreateDDLParser.likeTableOrFile_return likeTableOrFile() throws RecognitionException { return gCreateDDLParser.likeTableOrFile(); }\n \n-\tpublic HiveParser_IdentifiersParser.caseExpression_return caseExpression() throws RecognitionException { return gIdentifiersParser.caseExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceNotOperator_return precedenceNotOperator() throws RecognitionException { return gIdentifiersParser.precedenceNotOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetOwner_return alterDatabaseSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetOwner(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCompact_return alterStatementSuffixCompact() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCompact(); }\n \n-\tpublic HiveParser_FromClauseParser.tableOrColumn_return tableOrColumn() throws RecognitionException { return gFromClauseParser.tableOrColumn(); }\n+\tpublic HiveParser_IdentifiersParser.firstExpressionsWithAlias_return firstExpressionsWithAlias() throws RecognitionException { return gIdentifiersParser.firstExpressionsWithAlias(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAndOperator_return precedenceAndOperator() throws RecognitionException { return gIdentifiersParser.precedenceAndOperator(); }\n+\tpublic HiveParser_FromClauseParser.valuesTableConstructor_return valuesTableConstructor() throws RecognitionException { return gFromClauseParser.valuesTableConstructor(); }\n \n-\tpublic HiveParser_IdentifiersParser.constant_return constant() throws RecognitionException { return gIdentifiersParser.constant(); }\n+\tpublic HiveParser_ResourcePlanParser.globalWmStatement_return globalWmStatement() throws RecognitionException { return gResourcePlanParser.globalWmStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressions_return expressions() throws RecognitionException { return gIdentifiersParser.expressions(); }\n+\tpublic HiveParser_AlterClauseParser.skewedLocations_return skewedLocations() throws RecognitionException { return gAlterClauseParser.skewedLocations(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerExpressionStandalone_return triggerExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerExpressionStandalone(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpression_return precedenceSimilarExpression() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.orderByClause_return orderByClause() throws RecognitionException { return gIdentifiersParser.orderByClause(); }\n+\tpublic HiveParser_ResourcePlanParser.poolAssign_return poolAssign() throws RecognitionException { return gResourcePlanParser.poolAssign(); }\n \n-\tpublic HiveParser_CreateDDLParser.dcProperties_return dcProperties() throws RecognitionException { return gCreateDDLParser.dcProperties(); }\n+\tpublic HiveParser_ResourcePlanParser.resourcePlanDdlStatements_return resourcePlanDdlStatements() throws RecognitionException { return gResourcePlanParser.resourcePlanDdlStatements(); }\n \n-\tpublic HiveParser_AlterClauseParser.optimizeTblRewriteDataSuffix_return optimizeTblRewriteDataSuffix() throws RecognitionException { return gAlterClauseParser.optimizeTblRewriteDataSuffix(); }\n+\tpublic HiveParser_ResourcePlanParser.replaceResourcePlanStatement_return replaceResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.replaceResourcePlanStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.valuesClause_return valuesClause() throws RecognitionException { return gFromClauseParser.valuesClause(); }\n+\tpublic HiveParser_AlterClauseParser.snapshotIdOfRef_return snapshotIdOfRef() throws RecognitionException { return gAlterClauseParser.snapshotIdOfRef(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupByClause_return groupByClause() throws RecognitionException { return gIdentifiersParser.groupByClause(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSelectorSpec_return partitionSelectorSpec() throws RecognitionException { return gIdentifiersParser.partitionSelectorSpec(); }\n \n-\tpublic HiveParser_ResourcePlanParser.unmanaged_return unmanaged() throws RecognitionException { return gResourcePlanParser.unmanaged(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixProperties_return alterDatabaseSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixProperties(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinToken_return uniqueJoinToken() throws RecognitionException { return gFromClauseParser.uniqueJoinToken(); }\n+\tpublic HiveParser_IdentifiersParser.floorExpression_return floorExpression() throws RecognitionException { return gIdentifiersParser.floorExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRename_return alterStatementSuffixRename(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRename(table); }\n+\tpublic HiveParser_SelectClauseParser.selectTrfmClause_return selectTrfmClause() throws RecognitionException { return gSelectClauseParser.selectTrfmClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixConvert_return alterStatementSuffixConvert() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixConvert(); }\n+\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRewrite_return alterMaterializedViewSuffixRewrite(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRewrite(tableNameTree); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerAtomExpression_return triggerAtomExpression() throws RecognitionException { return gResourcePlanParser.triggerAtomExpression(); }\n+\tpublic HiveParser_IdentifiersParser.charSetStringLiteral_return charSetStringLiteral() throws RecognitionException { return gIdentifiersParser.charSetStringLiteral(); }\n \n-\tpublic HiveParser_FromClauseParser.partitionTableFunctionSource_return partitionTableFunctionSource() throws RecognitionException { return gFromClauseParser.partitionTableFunctionSource(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceStarExpression_return precedenceStarExpression() throws RecognitionException { return gIdentifiersParser.precedenceStarExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.createPoolStatement_return createPoolStatement() throws RecognitionException { return gResourcePlanParser.createPoolStatement(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAndExpression_return precedenceAndExpression() throws RecognitionException { return gIdentifiersParser.precedenceAndExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementPartitionKeyType_return alterStatementPartitionKeyType() throws RecognitionException { return gAlterClauseParser.alterStatementPartitionKeyType(); }\n+\tpublic HiveParser_AlterClauseParser.optimizeTableStatementSuffix_return optimizeTableStatementSuffix() throws RecognitionException { return gAlterClauseParser.optimizeTableStatementSuffix(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixMergeFiles_return alterStatementSuffixMergeFiles(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixMergeFiles(partition); }\n+\tpublic HiveParser_IdentifiersParser.partitionSelectorOperator_return partitionSelectorOperator() throws RecognitionException { return gIdentifiersParser.partitionSelectorOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrExpression_return precedenceBitwiseOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrExpression(); }\n+\tpublic HiveParser_IdentifiersParser.subQueryExpression_return subQueryExpression() throws RecognitionException { return gIdentifiersParser.subQueryExpression(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectList_return selectList() throws RecognitionException { return gSelectClauseParser.selectList(); }\n+\tpublic HiveParser_FromClauseParser.joinToken_return joinToken() throws RecognitionException { return gFromClauseParser.joinToken(); }\n \n-\tpublic HiveParser_IdentifiersParser.castExpression_return castExpression() throws RecognitionException { return gIdentifiersParser.castExpression(); }\n+\tpublic HiveParser_SelectClauseParser.window_frame_start_boundary_return window_frame_start_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_start_boundary(); }\n \n-\tpublic HiveParser_IdentifiersParser.floorDateQualifiers_return floorDateQualifiers() throws RecognitionException { return gIdentifiersParser.floorDateQualifiers(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixStatsPart_return alterStatementSuffixStatsPart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixStatsPart(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionVal_return partitionVal() throws RecognitionException { return gIdentifiersParser.partitionVal(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionAtom_return precedenceSimilarExpressionAtom(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionAtom(t); }\n \n-\tpublic HiveParser_IdentifiersParser.trimFunction_return trimFunction() throws RecognitionException { return gIdentifiersParser.trimFunction(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropPartitions_return alterStatementSuffixDropPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropPartitions(table); }\n \n-\tpublic HiveParser_AlterClauseParser.retentionOfSnapshots_return retentionOfSnapshots() throws RecognitionException { return gAlterClauseParser.retentionOfSnapshots(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceUnarySuffixExpression_return precedenceUnarySuffixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnarySuffixExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.sortByClause_return sortByClause() throws RecognitionException { return gIdentifiersParser.sortByClause(); }\n+\tpublic HiveParser_AlterClauseParser.tablePartitionPrefix_return tablePartitionPrefix() throws RecognitionException { return gAlterClauseParser.tablePartitionPrefix(); }\n \n-\tpublic HiveParser_CreateDDLParser.createDataConnectorStatement_return createDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.createDataConnectorStatement(); }\n+\tpublic HiveParser_SelectClauseParser.window_range_expression_return window_range_expression() throws RecognitionException { return gSelectClauseParser.window_range_expression(); }\n \n-\tpublic HiveParser_AlterClauseParser.skewedLocationsList_return skewedLocationsList() throws RecognitionException { return gAlterClauseParser.skewedLocationsList(); }\n+\tpublic HiveParser_IdentifiersParser.havingCondition_return havingCondition() throws RecognitionException { return gIdentifiersParser.havingCondition(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionAtom_return precedenceSimilarExpressionAtom(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionAtom(t); }\n+\tpublic HiveParser_FromClauseParser.tableOrColumn_return tableOrColumn() throws RecognitionException { return gFromClauseParser.tableOrColumn(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenameCol_return alterStatementSuffixRenameCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenameCol(); }\n+\tpublic HiveParser_IdentifiersParser.sortByClause_return sortByClause() throws RecognitionException { return gIdentifiersParser.sortByClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterViewSuffixProperties_return alterViewSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterViewSuffixProperties(); }\n+\tpublic HiveParser_IdentifiersParser.extractExpression_return extractExpression() throws RecognitionException { return gIdentifiersParser.extractExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropTag_return alterStatementSuffixDropTag() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropTag(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitionsElement_return alterStatementSuffixAddPartitionsElement() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitionsElement(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceEqualOperator_return precedenceEqualOperator() throws RecognitionException { return gIdentifiersParser.precedenceEqualOperator(); }\n+\tpublic HiveParser_IdentifiersParser.expressionPart_return expressionPart(CommonTree firstExprTree, boolean isStruct) throws RecognitionException { return gIdentifiersParser.expressionPart(firstExprTree, isStruct); }\n \n-\tpublic HiveParser_ResourcePlanParser.withReplace_return withReplace() throws RecognitionException { return gResourcePlanParser.withReplace(); }\n+\tpublic HiveParser_IdentifiersParser.timestampLiteral_return timestampLiteral() throws RecognitionException { return gIdentifiersParser.timestampLiteral(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpression_return precedenceSimilarExpression() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRename_return alterStatementSuffixRename(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRename(table); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceConcatenateOperator_return precedenceConcatenateOperator() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropBranch_return alterStatementSuffixDropBranch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropBranch(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSelectorOperator_return partitionSelectorOperator() throws RecognitionException { return gIdentifiersParser.partitionSelectorOperator(); }\n+\tpublic HiveParser_FromClauseParser.viewName_return viewName() throws RecognitionException { return gFromClauseParser.viewName(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropPoolStatement_return dropPoolStatement() throws RecognitionException { return gResourcePlanParser.dropPoolStatement(); }\n+\tpublic HiveParser_SelectClauseParser.window_defn_return window_defn() throws RecognitionException { return gSelectClauseParser.window_defn(); }\n \n-\tpublic HiveParser_ResourcePlanParser.createMappingStatement_return createMappingStatement() throws RecognitionException { return gResourcePlanParser.createMappingStatement(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerActionExpression_return triggerActionExpression() throws RecognitionException { return gResourcePlanParser.triggerActionExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.timestampLocalTZLiteral_return timestampLocalTZLiteral() throws RecognitionException { return gIdentifiersParser.timestampLocalTZLiteral(); }\n+\tpublic HiveParser_FromClauseParser.atomjoinSource_return atomjoinSource() throws RecognitionException { return gFromClauseParser.atomjoinSource(); }\n \n-\tpublic HiveParser_IdentifiersParser.principalIdentifier_return principalIdentifier() throws RecognitionException { return gIdentifiersParser.principalIdentifier(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixProperties_return alterStatementSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixProperties(); }\n \n-\tpublic HiveParser_FromClauseParser.tableName_return tableName() throws RecognitionException { return gFromClauseParser.tableName(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceRegexpOperator_return precedenceRegexpOperator() throws RecognitionException { return gIdentifiersParser.precedenceRegexpOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.quantifierType_return quantifierType() throws RecognitionException { return gIdentifiersParser.quantifierType(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceConcatenateExpression_return precedenceConcatenateExpression() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorOperator_return precedenceBitwiseXorOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorOperator(); }\n+\tpublic HiveParser_IdentifiersParser.isCondition_return isCondition() throws RecognitionException { return gIdentifiersParser.isCondition(); }\n \n-\tpublic HiveParser_IdentifiersParser.subQueryExpression_return subQueryExpression() throws RecognitionException { return gIdentifiersParser.subQueryExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenameCol_return alterStatementSuffixRenameCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenameCol(); }\n \n-\tpublic HiveParser_IdentifiersParser.extractExpression_return extractExpression() throws RecognitionException { return gIdentifiersParser.extractExpression(); }\n+\tpublic HiveParser_FromClauseParser.joinSource_return joinSource() throws RecognitionException { return gFromClauseParser.joinSource(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixProperties_return alterStatementSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixProperties(); }\n+\tpublic HiveParser_IdentifiersParser.qualifyClause_return qualifyClause() throws RecognitionException { return gIdentifiersParser.qualifyClause(); }\n \n-\tpublic HiveParser_ResourcePlanParser.alterMappingStatement_return alterMappingStatement() throws RecognitionException { return gResourcePlanParser.alterMappingStatement(); }\n+\tpublic HiveParser_IdentifiersParser.booleanValue_return booleanValue() throws RecognitionException { return gIdentifiersParser.booleanValue(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceEqualExpression_return precedenceEqualExpression() throws RecognitionException { return gIdentifiersParser.precedenceEqualExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRebuild_return alterMaterializedViewSuffixRebuild(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRebuild(tableNameTree); }\n \n-\tpublic HiveParser_IdentifiersParser.columnRefOrderInParenthesis_return columnRefOrderInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderInParenthesis(); }\n+\tpublic HiveParser_ResourcePlanParser.dropResourcePlanStatement_return dropResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.dropResourcePlanStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionIn_return precedenceSimilarExpressionIn(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionIn(t); }\n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetUrl_return alterDataConnectorSuffixSetUrl() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetUrl(); }\n \n-\tpublic HiveParser_FromClauseParser.defaultValue_return defaultValue() throws RecognitionException { return gFromClauseParser.defaultValue(); }\n+\tpublic HiveParser_FromClauseParser.virtualTableSource_return virtualTableSource() throws RecognitionException { return gFromClauseParser.virtualTableSource(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerActionExpression_return triggerActionExpression() throws RecognitionException { return gResourcePlanParser.triggerActionExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrOperator_return precedenceBitwiseOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorExpression_return precedenceBitwiseXorExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceNotExpression_return precedenceNotExpression() throws RecognitionException { return gIdentifiersParser.precedenceNotExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.createResourcePlanStatement_return createResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.createResourcePlanStatement(); }\n+\tpublic HiveParser_IdentifiersParser.constant_return constant() throws RecognitionException { return gIdentifiersParser.constant(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetPartSpec_return alterStatementSuffixSetPartSpec() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetPartSpec(); }\n+\tpublic HiveParser_IdentifiersParser.dateLiteral_return dateLiteral() throws RecognitionException { return gIdentifiersParser.dateLiteral(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterMaterializedViewStatementSuffix_return alterMaterializedViewStatementSuffix(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewStatementSuffix(tableNameTree); }\n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetOwner_return alterDataConnectorSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetOwner(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAmpersandExpression_return precedenceAmpersandExpression() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandExpression(); }\n+\tpublic HiveParser_ResourcePlanParser.createTriggerStatement_return createTriggerStatement() throws RecognitionException { return gResourcePlanParser.createTriggerStatement(); }\n \n-\tpublic HiveParser_CreateDDLParser.dataConnectorUrl_return dataConnectorUrl() throws RecognitionException { return gCreateDDLParser.dataConnectorUrl(); }\n+\tpublic HiveParser_SelectClauseParser.window_value_expression_return window_value_expression() throws RecognitionException { return gSelectClauseParser.window_value_expression(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_frame_start_boundary_return window_frame_start_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_start_boundary(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetOwner_return alterDatabaseSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetOwner(); }\n \n-\tpublic HiveParser_IdentifiersParser.function_return function() throws RecognitionException { return gIdentifiersParser.function(); }\n+\tpublic HiveParser_IdentifiersParser.groupingExpressionSingle_return groupingExpressionSingle() throws RecognitionException { return gIdentifiersParser.groupingExpressionSingle(); }\n \n-\tpublic HiveParser_FromClauseParser.fromSource_return fromSource() throws RecognitionException { return gFromClauseParser.fromSource(); }\n+\tpublic HiveParser_IdentifiersParser.expression_return expression() throws RecognitionException { return gIdentifiersParser.expression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixClusterbySortby_return alterStatementSuffixClusterbySortby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixClusterbySortby(); }\n+\tpublic HiveParser_IdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName_return sql11ReservedKeywordsUsedAsFunctionName() throws RecognitionException { return gIdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropMappingStatement_return dropMappingStatement() throws RecognitionException { return gResourcePlanParser.dropMappingStatement(); }\n+\tpublic HiveParser_IdentifiersParser.groupByClause_return groupByClause() throws RecognitionException { return gIdentifiersParser.groupByClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceDistinctOperator_return precedenceDistinctOperator() throws RecognitionException { return gIdentifiersParser.precedenceDistinctOperator(); }\n+\tpublic HiveParser_IdentifiersParser.intervalValue_return intervalValue() throws RecognitionException { return gIdentifiersParser.intervalValue(); }\n \n-\tpublic HiveParser_FromClauseParser.searchCondition_return searchCondition() throws RecognitionException { return gFromClauseParser.searchCondition(); }\n+\tpublic HiveParser_IdentifiersParser.expressionsInParenthesis_return expressionsInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsInParenthesis(isStruct, forceStruct); }\n \n-\tpublic HiveParser_ResourcePlanParser.alterResourcePlanStatement_return alterResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.alterResourcePlanStatement(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddCol_return alterStatementSuffixAddCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddCol(); }\n \n-\tpublic HiveParser_IdentifiersParser.timestampLiteral_return timestampLiteral() throws RecognitionException { return gIdentifiersParser.timestampLiteral(); }\n+\tpublic HiveParser_IdentifiersParser.columnRefOrderInParenthesis_return columnRefOrderInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderInParenthesis(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerLiteral_return triggerLiteral() throws RecognitionException { return gResourcePlanParser.triggerLiteral(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceOrOperator_return precedenceOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceOrOperator(); }\n \n-\tpublic HiveParser_ResourcePlanParser.replaceResourcePlanStatement_return replaceResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.replaceResourcePlanStatement(); }\n+\tpublic HiveParser_ResourcePlanParser.enable_return enable() throws RecognitionException { return gResourcePlanParser.enable(); }\n \n-\tpublic HiveParser_FromClauseParser.fromClause_return fromClause() throws RecognitionException { return gFromClauseParser.fromClause(); }\n+\tpublic HiveParser_AlterClauseParser.skewedLocationsList_return skewedLocationsList() throws RecognitionException { return gAlterClauseParser.skewedLocationsList(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceNotOperator_return precedenceNotOperator() throws RecognitionException { return gIdentifiersParser.precedenceNotOperator(); }\n+\tpublic HiveParser_IdentifiersParser.rollupStandard_return rollupStandard() throws RecognitionException { return gIdentifiersParser.rollupStandard(); }\n \n-\tpublic HiveParser_FromClauseParser.partitioningSpec_return partitioningSpec() throws RecognitionException { return gFromClauseParser.partitioningSpec(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceConcatenateOperator_return precedenceConcatenateOperator() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateOperator(); }\n \n-\tpublic HiveParser_CreateDDLParser.dataConnectorType_return dataConnectorType() throws RecognitionException { return gCreateDDLParser.dataConnectorType(); }\n+\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffix_return alterTblPartitionStatementSuffix(boolean partition) throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffix(partition); }\n \n-\tpublic HiveParser_SelectClauseParser.selectExpression_return selectExpression() throws RecognitionException { return gSelectClauseParser.selectExpression(); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinToken_return uniqueJoinToken() throws RecognitionException { return gFromClauseParser.uniqueJoinToken(); }\n \n-\tpublic HiveParser_ResourcePlanParser.activate_return activate() throws RecognitionException { return gResourcePlanParser.activate(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionMain_return precedenceSimilarExpressionMain() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionMain(); }\n \n-\tpublic HiveParser_FromClauseParser.expressionList_return expressionList() throws RecognitionException { return gFromClauseParser.expressionList(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorOperator_return precedenceBitwiseXorOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.parameterIdx_return parameterIdx() throws RecognitionException { return gIdentifiersParser.parameterIdx(); }\n+\tpublic HiveParser_AlterClauseParser.fileFormat_return fileFormat() throws RecognitionException { return gAlterClauseParser.fileFormat(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPartNot_return precedenceSimilarExpressionPartNot(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPartNot(t); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenamePart_return alterStatementSuffixRenamePart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenamePart(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAndExpression_return precedenceAndExpression() throws RecognitionException { return gIdentifiersParser.precedenceAndExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixFileFormat_return alterStatementSuffixFileFormat(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixFileFormat(partition); }\n \n-\tpublic HiveParser_SelectClauseParser.window_clause_return window_clause() throws RecognitionException { return gSelectClauseParser.window_clause(); }\n+\tpublic HiveParser_ResourcePlanParser.alterMappingStatement_return alterMappingStatement() throws RecognitionException { return gResourcePlanParser.alterMappingStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.joinToken_return joinToken() throws RecognitionException { return gFromClauseParser.joinToken(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSerdeProperties_return alterStatementSuffixSerdeProperties(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSerdeProperties(partition); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixArchive_return alterStatementSuffixArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixArchive(); }\n+\tpublic HiveParser_FromClauseParser.splitSample_return splitSample() throws RecognitionException { return gFromClauseParser.splitSample(); }\n \n-\tpublic HiveParser_IdentifiersParser.prepareStmtParam_return prepareStmtParam() throws RecognitionException { return gIdentifiersParser.prepareStmtParam(); }\n+\tpublic HiveParser_IdentifiersParser.intervalLiteral_return intervalLiteral() throws RecognitionException { return gIdentifiersParser.intervalLiteral(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffix_return alterTblPartitionStatementSuffix(boolean partition) throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffix(partition); }\n+\tpublic HiveParser_AlterClauseParser.skewedLocationMap_return skewedLocationMap() throws RecognitionException { return gAlterClauseParser.skewedLocationMap(); }\n \n-\tpublic HiveParser_IdentifiersParser.tableOrPartition_return tableOrPartition() throws RecognitionException { return gIdentifiersParser.tableOrPartition(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixExchangePartition_return alterStatementSuffixExchangePartition() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExchangePartition(); }\n \n-\tpublic HiveParser_FromClauseParser.joinSource_return joinSource() throws RecognitionException { return gFromClauseParser.joinSource(); }\n+\tpublic HiveParser_IdentifiersParser.timeQualifiers_return timeQualifiers() throws RecognitionException { return gIdentifiersParser.timeQualifiers(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionPart_return expressionPart(CommonTree firstExprTree, boolean isStruct) throws RecognitionException { return gIdentifiersParser.expressionPart(firstExprTree, isStruct); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementPartitionKeyType_return alterStatementPartitionKeyType() throws RecognitionException { return gAlterClauseParser.alterStatementPartitionKeyType(); }\n \n-\tpublic HiveParser_AlterClauseParser.skewedLocationMap_return skewedLocationMap() throws RecognitionException { return gAlterClauseParser.skewedLocationMap(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddConstraint_return alterStatementSuffixAddConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddConstraint(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupingSetExpression_return groupingSetExpression() throws RecognitionException { return gIdentifiersParser.groupingSetExpression(); }\n+\tpublic HiveParser_ResourcePlanParser.createResourcePlanStatement_return createResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.createResourcePlanStatement(); }\n \n \tpublic HiveParser_IdentifiersParser.timeUnitQualifiers_return timeUnitQualifiers() throws RecognitionException { return gIdentifiersParser.timeUnitQualifiers(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseStatementSuffix_return alterDatabaseStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDatabaseStatementSuffix(); }\n-\n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixProperties_return alterDataConnectorSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixProperties(); }\n+\tpublic HiveParser_SelectClauseParser.window_clause_return window_clause() throws RecognitionException { return gSelectClauseParser.window_clause(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceNotExpression_return precedenceNotExpression() throws RecognitionException { return gIdentifiersParser.precedenceNotExpression(); }\n+\tpublic HiveParser_FromClauseParser.tableSource_return tableSource() throws RecognitionException { return gFromClauseParser.tableSource(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionWithAlias_return expressionWithAlias() throws RecognitionException { return gIdentifiersParser.expressionWithAlias(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetPartSpec_return alterStatementSuffixSetPartSpec() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetPartSpec(); }\n \n-\tpublic HiveParser_ResourcePlanParser.alterPoolStatement_return alterPoolStatement() throws RecognitionException { return gResourcePlanParser.alterPoolStatement(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceUnaryPrefixExpression_return precedenceUnaryPrefixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnaryPrefixExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.rollupOldSyntax_return rollupOldSyntax() throws RecognitionException { return gIdentifiersParser.rollupOldSyntax(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAndOperator_return precedenceAndOperator() throws RecognitionException { return gIdentifiersParser.precedenceAndOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrOperator_return precedenceBitwiseOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrOperator(); }\n+\tpublic HiveParser_CreateDDLParser.dcProperties_return dcProperties() throws RecognitionException { return gCreateDDLParser.dcProperties(); }\n \n-\tpublic HiveParser_FromClauseParser.whereClause_return whereClause() throws RecognitionException { return gFromClauseParser.whereClause(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionQuantifierPredicate_return precedenceSimilarExpressionQuantifierPredicate(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionQuantifierPredicate(t); }\n \n-\tpublic HiveParser_IdentifiersParser.groupByEmpty_return groupByEmpty() throws RecognitionException { return gIdentifiersParser.groupByEmpty(); }\n+\tpublic HiveParser_FromClauseParser.tableAllColumns_return tableAllColumns() throws RecognitionException { return gFromClauseParser.tableAllColumns(); }\n \n-\tpublic HiveParser_AlterClauseParser.tablePartitionPrefix_return tablePartitionPrefix() throws RecognitionException { return gAlterClauseParser.tablePartitionPrefix(); }\n+\tpublic HiveParser_FromClauseParser.joinSourcePart_return joinSourcePart() throws RecognitionException { return gFromClauseParser.joinSourcePart(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_defn_return window_defn() throws RecognitionException { return gSelectClauseParser.window_defn(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionIn_return precedenceSimilarExpressionIn(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionIn(t); }\n \n-\tpublic HiveParser_AlterClauseParser.optimizeTableStatementSuffix_return optimizeTableStatementSuffix() throws RecognitionException { return gAlterClauseParser.optimizeTableStatementSuffix(); }\n+\tpublic HiveParser_IdentifiersParser.functionName_return functionName() throws RecognitionException { return gIdentifiersParser.functionName(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupby_expression_return groupby_expression() throws RecognitionException { return gIdentifiersParser.groupby_expression(); }\n+\tpublic HiveParser_SelectClauseParser.window_frame_boundary_return window_frame_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_boundary(); }\n \n-\tpublic HiveParser_CreateDDLParser.createTableStatement_return createTableStatement() throws RecognitionException { return gCreateDDLParser.createTableStatement(); }\n+\tpublic HiveParser_IdentifiersParser.function_return function() throws RecognitionException { return gIdentifiersParser.function(); }\n \n-\tpublic HiveParser_ResourcePlanParser.poolAssign_return poolAssign() throws RecognitionException { return gResourcePlanParser.poolAssign(); }\n+\tpublic HiveParser_ResourcePlanParser.poolPath_return poolPath() throws RecognitionException { return gResourcePlanParser.poolPath(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_value_expression_return window_value_expression() throws RecognitionException { return gSelectClauseParser.window_value_expression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAmpersandOperator_return precedenceAmpersandOperator() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupingSetExpressionMultiple_return groupingSetExpressionMultiple() throws RecognitionException { return gIdentifiersParser.groupingSetExpressionMultiple(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixArchive_return alterStatementSuffixArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixArchive(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceUnarySuffixExpression_return precedenceUnarySuffixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnarySuffixExpression(); }\n+\tpublic HiveParser_SelectClauseParser.selectExpressionList_return selectExpressionList() throws RecognitionException { return gSelectClauseParser.selectExpressionList(); }\n \n-\tpublic HiveParser_AlterClauseParser.partitionLocation_return partitionLocation() throws RecognitionException { return gAlterClauseParser.partitionLocation(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropConstraint_return alterStatementSuffixDropConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropConstraint(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceStarExpression_return precedenceStarExpression() throws RecognitionException { return gIdentifiersParser.precedenceStarExpression(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceUnaryOperator_return precedenceUnaryOperator() throws RecognitionException { return gIdentifiersParser.precedenceUnaryOperator(); }\n \n-\tpublic HiveParser_ResourcePlanParser.comparisionOperator_return comparisionOperator() throws RecognitionException { return gResourcePlanParser.comparisionOperator(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerOrExpression_return triggerOrExpression() throws RecognitionException { return gResourcePlanParser.triggerOrExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitions_return alterStatementSuffixAddPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitions(table); }\n+\tpublic HiveParser_FromClauseParser.aliasList_return aliasList() throws RecognitionException { return gFromClauseParser.aliasList(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddCol_return alterStatementSuffixAddCol() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddCol(); }\n+\tpublic HiveParser_ResourcePlanParser.dropMappingStatement_return dropMappingStatement() throws RecognitionException { return gResourcePlanParser.dropMappingStatement(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetOwner_return alterStatementSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetOwner(); }\n+\tpublic HiveParser_IdentifiersParser.clusterByClause_return clusterByClause() throws RecognitionException { return gIdentifiersParser.clusterByClause(); }\n \n-\tpublic HiveParser_ResourcePlanParser.disable_return disable() throws RecognitionException { return gResourcePlanParser.disable(); }\n+\tpublic HiveParser_SelectClauseParser.window_specification_return window_specification(CommonTree nullTreatment) throws RecognitionException { return gSelectClauseParser.window_specification(nullTreatment); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSerdeProperties_return alterStatementSuffixSerdeProperties(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSerdeProperties(partition); }\n+\tpublic HiveParser_ResourcePlanParser.triggerExpression_return triggerExpression() throws RecognitionException { return gResourcePlanParser.triggerExpression(); }\n \n-\tpublic HiveParser_FromClauseParser.valueRowConstructor_return valueRowConstructor() throws RecognitionException { return gFromClauseParser.valueRowConstructor(); }\n+\tpublic HiveParser_IdentifiersParser.partitionByClause_return partitionByClause() throws RecognitionException { return gIdentifiersParser.partitionByClause(); }\n \n-\tpublic HiveParser_ResourcePlanParser.globalWmStatement_return globalWmStatement() throws RecognitionException { return gResourcePlanParser.globalWmStatement(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSpec_return partitionSpec() throws RecognitionException { return gIdentifiersParser.partitionSpec(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceConcatenateExpression_return precedenceConcatenateExpression() throws RecognitionException { return gIdentifiersParser.precedenceConcatenateExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixLocation_return alterStatementSuffixLocation(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixLocation(partition); }\n \n-\tpublic HiveParser_IdentifiersParser.stringLiteralSequence_return stringLiteralSequence() throws RecognitionException { return gIdentifiersParser.stringLiteralSequence(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixMergeFiles_return alterStatementSuffixMergeFiles(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixMergeFiles(partition); }\n \n-\tpublic HiveParser_AlterClauseParser.alterViewStatementSuffix_return alterViewStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterViewStatementSuffix(); }\n+\tpublic HiveParser_IdentifiersParser.atomExpression_return atomExpression() throws RecognitionException { return gIdentifiersParser.atomExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceUnaryPrefixExpression_return precedenceUnaryPrefixExpression() throws RecognitionException { return gIdentifiersParser.precedenceUnaryPrefixExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation_return alterTblPartitionStatementSuffixSkewedLocation() throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation(); }\n \n-\tpublic HiveParser_FromClauseParser.asOfClause_return asOfClause() throws RecognitionException { return gFromClauseParser.asOfClause(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerAndExpression_return triggerAndExpression() throws RecognitionException { return gResourcePlanParser.triggerAndExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropTriggerStatement_return dropTriggerStatement() throws RecognitionException { return gResourcePlanParser.dropTriggerStatement(); }\n+\tpublic HiveParser_IdentifiersParser.sysFuncNames_return sysFuncNames() throws RecognitionException { return gIdentifiersParser.sysFuncNames(); }\n \n-\tpublic HiveParser_FromClauseParser.viewName_return viewName() throws RecognitionException { return gFromClauseParser.viewName(); }\n+\tpublic HiveParser_FromClauseParser.expressionList_return expressionList() throws RecognitionException { return gFromClauseParser.expressionList(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetLocation_return alterDatabaseSuffixSetLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetLocation(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixBucketNum_return alterStatementSuffixBucketNum(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixBucketNum(partition); }\n \n-\tpublic HiveParser_FromClauseParser.splitSample_return splitSample() throws RecognitionException { return gFromClauseParser.splitSample(); }\n+\tpublic HiveParser_FromClauseParser.valueRowConstructor_return valueRowConstructor() throws RecognitionException { return gFromClauseParser.valueRowConstructor(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalValue_return intervalValue() throws RecognitionException { return gIdentifiersParser.intervalValue(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerAtomExpression_return triggerAtomExpression() throws RecognitionException { return gResourcePlanParser.triggerAtomExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionOrDefault_return expressionOrDefault() throws RecognitionException { return gIdentifiersParser.expressionOrDefault(); }\n+\tpublic HiveParser_IdentifiersParser.expressions_return expressions() throws RecognitionException { return gIdentifiersParser.expressions(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCompact_return alterStatementSuffixCompact() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCompact(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetManagedLocation_return alterDatabaseSuffixSetManagedLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetManagedLocation(); }\n \n-\tpublic HiveParser_FromClauseParser.tableAlias_return tableAlias() throws RecognitionException { return gFromClauseParser.tableAlias(); }\n+\tpublic HiveParser_FromClauseParser.fromSource_return fromSource() throws RecognitionException { return gFromClauseParser.fromSource(); }\n \n-\tpublic HiveParser_PrepareStatementParser.prepareStatement_return prepareStatement() throws RecognitionException { return gPrepareStatementParser.prepareStatement(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerLiteral_return triggerLiteral() throws RecognitionException { return gResourcePlanParser.triggerLiteral(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceOrOperator_return precedenceOrOperator() throws RecognitionException { return gIdentifiersParser.precedenceOrOperator(); }\n+\tpublic HiveParser_IdentifiersParser.floorDateQualifiers_return floorDateQualifiers() throws RecognitionException { return gIdentifiersParser.floorDateQualifiers(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_range_expression_return window_range_expression() throws RecognitionException { return gSelectClauseParser.window_range_expression(); }\n+\tpublic HiveParser_CreateDDLParser.dataConnectorType_return dataConnectorType() throws RecognitionException { return gCreateDDLParser.dataConnectorType(); }\n \n-\tpublic HiveParser_IdentifiersParser.atomExpression_return atomExpression() throws RecognitionException { return gIdentifiersParser.atomExpression(); }\n+\tpublic HiveParser_ResourcePlanParser.createMappingStatement_return createMappingStatement() throws RecognitionException { return gResourcePlanParser.createMappingStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPart_return precedenceSimilarExpressionPart(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPart(t); }\n+\tpublic HiveParser_IdentifiersParser.identifier_return identifier() throws RecognitionException { return gIdentifiersParser.identifier(); }\n \n-\tpublic HiveParser_IdentifiersParser.charSetStringLiteral_return charSetStringLiteral() throws RecognitionException { return gIdentifiersParser.charSetStringLiteral(); }\n+\tpublic HiveParser_IdentifiersParser.timestampLocalTZLiteral_return timestampLocalTZLiteral() throws RecognitionException { return gIdentifiersParser.timestampLocalTZLiteral(); }\n \n-\tpublic HiveParser_IdentifiersParser.expression_return expression() throws RecognitionException { return gIdentifiersParser.expression(); }\n+\tpublic HiveParser_SelectClauseParser.trfmClause_return trfmClause() throws RecognitionException { return gSelectClauseParser.trfmClause(); }\n \n-\tpublic HiveParser_ResourcePlanParser.poolAssignList_return poolAssignList() throws RecognitionException { return gResourcePlanParser.poolAssignList(); }\n+\tpublic HiveParser_IdentifiersParser.trimFunction_return trimFunction() throws RecognitionException { return gIdentifiersParser.trimFunction(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalLiteral_return intervalLiteral() throws RecognitionException { return gIdentifiersParser.intervalLiteral(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCreateBranch_return alterStatementSuffixCreateBranch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCreateBranch(); }\n \n-\tpublic HiveParser_IdentifiersParser.clusterByClause_return clusterByClause() throws RecognitionException { return gIdentifiersParser.clusterByClause(); }\n+\tpublic HiveParser_SelectClauseParser.selectList_return selectList() throws RecognitionException { return gSelectClauseParser.selectList(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionByClause_return partitionByClause() throws RecognitionException { return gIdentifiersParser.partitionByClause(); }\n+\tpublic HiveParser_AlterClauseParser.alterDatabaseStatementSuffix_return alterDatabaseStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDatabaseStatementSuffix(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementChangeColPosition_return alterStatementChangeColPosition() throws RecognitionException { return gAlterClauseParser.alterStatementChangeColPosition(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarOperator_return precedenceSimilarOperator() throws RecognitionException { return gIdentifiersParser.precedenceSimilarOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCreateTag_return alterStatementSuffixCreateTag() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCreateTag(); }\n+\tpublic HiveParser_FromClauseParser.tableName_return tableName() throws RecognitionException { return gFromClauseParser.tableName(); }\n \n-\tpublic HiveParser_FromClauseParser.aliasList_return aliasList() throws RecognitionException { return gFromClauseParser.aliasList(); }\n+\tpublic HiveParser_SelectClauseParser.window_frame_return window_frame() throws RecognitionException { return gSelectClauseParser.window_frame(); }\n \n-\tpublic HiveParser_FromClauseParser.virtualTableSource_return virtualTableSource() throws RecognitionException { return gFromClauseParser.virtualTableSource(); }\n+\tpublic HiveParser_IdentifiersParser.precedencePlusExpression_return precedencePlusExpression() throws RecognitionException { return gIdentifiersParser.precedencePlusExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixFileFormat_return alterStatementSuffixFileFormat(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixFileFormat(partition); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinTableSource_return uniqueJoinTableSource() throws RecognitionException { return gFromClauseParser.uniqueJoinTableSource(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRebuild_return alterMaterializedViewSuffixRebuild(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRebuild(tableNameTree); }\n+\tpublic HiveParser_FromClauseParser.fromClause_return fromClause() throws RecognitionException { return gFromClauseParser.fromClause(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSpec_return partitionSpec() throws RecognitionException { return gIdentifiersParser.partitionSpec(); }\n+\tpublic HiveParser_ResourcePlanParser.alterResourcePlanStatement_return alterResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.alterResourcePlanStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSelectorSpec_return partitionSelectorSpec() throws RecognitionException { return gIdentifiersParser.partitionSelectorSpec(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixConvert_return alterStatementSuffixConvert() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixConvert(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterTableStatementSuffix_return alterTableStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterTableStatementSuffix(); }\n+\tpublic HiveParser_AlterClauseParser.alterMaterializedViewStatementSuffix_return alterMaterializedViewStatementSuffix(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewStatementSuffix(tableNameTree); }\n \n-\tpublic HiveParser_FromClauseParser.tableBucketSample_return tableBucketSample() throws RecognitionException { return gFromClauseParser.tableBucketSample(); }\n+\tpublic HiveParser_AlterClauseParser.optimizeTblRewriteDataSuffix_return optimizeTblRewriteDataSuffix() throws RecognitionException { return gAlterClauseParser.optimizeTblRewriteDataSuffix(); }\n \n-\tpublic HiveParser_IdentifiersParser.rollupStandard_return rollupStandard() throws RecognitionException { return gIdentifiersParser.rollupStandard(); }\n+\tpublic HiveParser_IdentifiersParser.groupingSetExpression_return groupingSetExpression() throws RecognitionException { return gIdentifiersParser.groupingSetExpression(); }\n \n-\tpublic HiveParser_CreateDDLParser.likeTableOrFile_return likeTableOrFile() throws RecognitionException { return gCreateDDLParser.likeTableOrFile(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerExpressionStandalone_return triggerExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerExpressionStandalone(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitionsElement_return alterStatementSuffixAddPartitionsElement() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitionsElement(); }\n+\tpublic HiveParser_IdentifiersParser.intervalQualifiers_return intervalQualifiers() throws RecognitionException { return gIdentifiersParser.intervalQualifiers(); }\n \n-\tpublic HiveParser_FromClauseParser.tableSource_return tableSource() throws RecognitionException { return gFromClauseParser.tableSource(); }\n+\tpublic HiveParser_FromClauseParser.tableSample_return tableSample() throws RecognitionException { return gFromClauseParser.tableSample(); }\n \n-\tpublic HiveParser_CreateDDLParser.dataConnectorComment_return dataConnectorComment() throws RecognitionException { return gCreateDDLParser.dataConnectorComment(); }\n+\tpublic HiveParser_FromClauseParser.tableBucketSample_return tableBucketSample() throws RecognitionException { return gFromClauseParser.tableBucketSample(); }\n \n-\tpublic HiveParser_ResourcePlanParser.dropResourcePlanStatement_return dropResourcePlanStatement() throws RecognitionException { return gResourcePlanParser.dropResourcePlanStatement(); }\n+\tpublic HiveParser_AlterClauseParser.retentionOfSnapshots_return retentionOfSnapshots() throws RecognitionException { return gAlterClauseParser.retentionOfSnapshots(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpUnassignList_return rpUnassignList() throws RecognitionException { return gResourcePlanParser.rpUnassignList(); }\n+\tpublic HiveParser_IdentifiersParser.nonReserved_return nonReserved() throws RecognitionException { return gIdentifiersParser.nonReserved(); }\n \n-\tpublic HiveParser_FromClauseParser.partitionedTableFunction_return partitionedTableFunction() throws RecognitionException { return gFromClauseParser.partitionedTableFunction(); }\n+\tpublic HiveParser_FromClauseParser.tableAlias_return tableAlias() throws RecognitionException { return gFromClauseParser.tableAlias(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarOperator_return precedenceSimilarOperator() throws RecognitionException { return gIdentifiersParser.precedenceSimilarOperator(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatement_return alterStatement() throws RecognitionException { return gAlterClauseParser.alterStatement(); }\n \n-\tpublic HiveParser_IdentifiersParser.sysFuncNames_return sysFuncNames() throws RecognitionException { return gIdentifiersParser.sysFuncNames(); }\n+\tpublic HiveParser_IdentifiersParser.descFuncNames_return descFuncNames() throws RecognitionException { return gIdentifiersParser.descFuncNames(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixBucketNum_return alterStatementSuffixBucketNum(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixBucketNum(partition); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSetOwner_return alterStatementSuffixSetOwner() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSetOwner(); }\n \n-\tpublic HiveParser_AlterClauseParser.snapshotIdOfRef_return snapshotIdOfRef() throws RecognitionException { return gAlterClauseParser.snapshotIdOfRef(); }\n+\tpublic HiveParser_ResourcePlanParser.rpAssignList_return rpAssignList() throws RecognitionException { return gResourcePlanParser.rpAssignList(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerAndExpression_return triggerAndExpression() throws RecognitionException { return gResourcePlanParser.triggerAndExpression(); }\n+\tpublic HiveParser_FromClauseParser.searchCondition_return searchCondition() throws RecognitionException { return gFromClauseParser.searchCondition(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixExchangePartition_return alterStatementSuffixExchangePartition() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExchangePartition(); }\n+\tpublic HiveParser_ResourcePlanParser.rpAssign_return rpAssign() throws RecognitionException { return gResourcePlanParser.rpAssign(); }\n \n-\tpublic HiveParser_CreateDDLParser.dropDataConnectorStatement_return dropDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.dropDataConnectorStatement(); }\n+\tpublic HiveParser_IdentifiersParser.havingClause_return havingClause() throws RecognitionException { return gIdentifiersParser.havingClause(); }\n \n-\tpublic HiveParser_PrepareStatementParser.executeParamList_return executeParamList() throws RecognitionException { return gPrepareStatementParser.executeParamList(); }\n+\tpublic HiveParser_ResourcePlanParser.unmanaged_return unmanaged() throws RecognitionException { return gResourcePlanParser.unmanaged(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStatsCol_return alterStatementSuffixUpdateStatsCol(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStatsCol(partition); }\n+\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixProperties_return alterDataConnectorSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixProperties(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceFieldExpression_return precedenceFieldExpression() throws RecognitionException { return gIdentifiersParser.precedenceFieldExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateColumns_return alterStatementSuffixUpdateColumns() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateColumns(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceOrExpression_return precedenceOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceOrExpression(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropTag_return alterStatementSuffixDropTag() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropTag(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinTableSource_return uniqueJoinTableSource() throws RecognitionException { return gFromClauseParser.uniqueJoinTableSource(); }\n+\tpublic HiveParser_FromClauseParser.partitioningSpec_return partitioningSpec() throws RecognitionException { return gFromClauseParser.partitioningSpec(); }\n \n-\tpublic HiveParser_IdentifiersParser.identifier_return identifier() throws RecognitionException { return gIdentifiersParser.identifier(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseOrExpression_return precedenceBitwiseOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseOrExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceRegexpOperator_return precedenceRegexpOperator() throws RecognitionException { return gIdentifiersParser.precedenceRegexpOperator(); }\n+\tpublic HiveParser_IdentifiersParser.quantifierType_return quantifierType() throws RecognitionException { return gIdentifiersParser.quantifierType(); }\n \n-\tpublic HiveParser_ResourcePlanParser.resourcePlanDdlStatements_return resourcePlanDdlStatements() throws RecognitionException { return gResourcePlanParser.resourcePlanDdlStatements(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceEqualOperator_return precedenceEqualOperator() throws RecognitionException { return gIdentifiersParser.precedenceEqualOperator(); }\n \n \tpublic HiveParser_AlterClauseParser.alterStatementSuffixExecute_return alterStatementSuffixExecute() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixExecute(); }\n \n-\tpublic HiveParser_AlterClauseParser.skewedLocations_return skewedLocations() throws RecognitionException { return gAlterClauseParser.skewedLocations(); }\n+\tpublic HiveParser_FromClauseParser.firstValueRowConstructor_return firstValueRowConstructor() throws RecognitionException { return gFromClauseParser.firstValueRowConstructor(); }\n \n-\tpublic HiveParser_IdentifiersParser.qualifyClause_return qualifyClause() throws RecognitionException { return gIdentifiersParser.qualifyClause(); }\n+\tpublic HiveParser_IdentifiersParser.expressionOrDefault_return expressionOrDefault() throws RecognitionException { return gIdentifiersParser.expressionOrDefault(); }\n \n-\tpublic HiveParser_ResourcePlanParser.poolPath_return poolPath() throws RecognitionException { return gResourcePlanParser.poolPath(); }\n+\tpublic HiveParser_ResourcePlanParser.poolAssignList_return poolAssignList() throws RecognitionException { return gResourcePlanParser.poolAssignList(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceUnaryOperator_return precedenceUnaryOperator() throws RecognitionException { return gIdentifiersParser.precedenceUnaryOperator(); }\n+\tpublic HiveParser_FromClauseParser.partitionTableFunctionSource_return partitionTableFunctionSource() throws RecognitionException { return gFromClauseParser.partitionTableFunctionSource(); }\n \n-\tpublic HiveParser_IdentifiersParser.havingCondition_return havingCondition() throws RecognitionException { return gIdentifiersParser.havingCondition(); }\n+\tpublic HiveParser_IdentifiersParser.caseExpression_return caseExpression() throws RecognitionException { return gIdentifiersParser.caseExpression(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_frame_return window_frame() throws RecognitionException { return gSelectClauseParser.window_frame(); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinSource_return uniqueJoinSource() throws RecognitionException { return gFromClauseParser.uniqueJoinSource(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectExpressionList_return selectExpressionList() throws RecognitionException { return gSelectClauseParser.selectExpressionList(); }\n+\tpublic HiveParser_IdentifiersParser.parameterIdx_return parameterIdx() throws RecognitionException { return gIdentifiersParser.parameterIdx(); }\n \n-\tpublic HiveParser_IdentifiersParser.isCondition_return isCondition() throws RecognitionException { return gIdentifiersParser.isCondition(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStatsCol_return alterStatementSuffixUpdateStatsCol(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStatsCol(partition); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateColumns_return alterStatementSuffixUpdateColumns() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateColumns(); }\n+\tpublic HiveParser_ResourcePlanParser.comparisionOperator_return comparisionOperator() throws RecognitionException { return gResourcePlanParser.comparisionOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.distributeByClause_return distributeByClause() throws RecognitionException { return gIdentifiersParser.distributeByClause(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixTouch_return alterStatementSuffixTouch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixTouch(); }\n \n-\tpublic HiveParser_IdentifiersParser.descFuncNames_return descFuncNames() throws RecognitionException { return gIdentifiersParser.descFuncNames(); }\n+\tpublic HiveParser_AlterClauseParser.refRetain_return refRetain() throws RecognitionException { return gAlterClauseParser.refRetain(); }\n \n-\tpublic HiveParser_ResourcePlanParser.enable_return enable() throws RecognitionException { return gResourcePlanParser.enable(); }\n+\tpublic HiveParser_FromClauseParser.subQuerySource_return subQuerySource() throws RecognitionException { return gFromClauseParser.subQuerySource(); }\n \n-\tpublic HiveParser_IdentifiersParser.groupingExpressionSingle_return groupingExpressionSingle() throws RecognitionException { return gIdentifiersParser.groupingExpressionSingle(); }\n+\tpublic HiveParser_FromClauseParser.lateralView_return lateralView() throws RecognitionException { return gFromClauseParser.lateralView(); }\n \n-\tpublic HiveParser_IdentifiersParser.whenExpression_return whenExpression() throws RecognitionException { return gIdentifiersParser.whenExpression(); }\n+\tpublic HiveParser_ResourcePlanParser.activate_return activate() throws RecognitionException { return gResourcePlanParser.activate(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_frame_boundary_return window_frame_boundary() throws RecognitionException { return gSelectClauseParser.window_frame_boundary(); }\n+\tpublic HiveParser_IdentifiersParser.prepareStmtParam_return prepareStmtParam() throws RecognitionException { return gIdentifiersParser.prepareStmtParam(); }\n \n-\tpublic HiveParser_IdentifiersParser.functionName_return functionName() throws RecognitionException { return gIdentifiersParser.functionName(); }\n+\tpublic HiveParser_ResourcePlanParser.withReplace_return withReplace() throws RecognitionException { return gResourcePlanParser.withReplace(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinExpr_return uniqueJoinExpr() throws RecognitionException { return gFromClauseParser.uniqueJoinExpr(); }\n+\tpublic HiveParser_AlterClauseParser.compactPool_return compactPool() throws RecognitionException { return gAlterClauseParser.compactPool(); }\n \n-\tpublic HiveParser_IdentifiersParser.partitionSelectorVal_return partitionSelectorVal() throws RecognitionException { return gIdentifiersParser.partitionSelectorVal(); }\n+\tpublic HiveParser_IdentifiersParser.intervalExpression_return intervalExpression() throws RecognitionException { return gIdentifiersParser.intervalExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedencePlusExpression_return precedencePlusExpression() throws RecognitionException { return gIdentifiersParser.precedencePlusExpression(); }\n+\tpublic HiveParser_IdentifiersParser.partitionVal_return partitionVal() throws RecognitionException { return gIdentifiersParser.partitionVal(); }\n \n-\tpublic HiveParser_IdentifiersParser.booleanValue_return booleanValue() throws RecognitionException { return gIdentifiersParser.booleanValue(); }\n+\tpublic HiveParser_FromClauseParser.defaultValue_return defaultValue() throws RecognitionException { return gFromClauseParser.defaultValue(); }\n \n-\tpublic HiveParser_AlterClauseParser.blocking_return blocking() throws RecognitionException { return gAlterClauseParser.blocking(); }\n+\tpublic HiveParser_IdentifiersParser.tableOrPartition_return tableOrPartition() throws RecognitionException { return gIdentifiersParser.tableOrPartition(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropPartitions_return alterStatementSuffixDropPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropPartitions(table); }\n+\tpublic HiveParser_IdentifiersParser.precedenceFieldExpression_return precedenceFieldExpression() throws RecognitionException { return gIdentifiersParser.precedenceFieldExpression(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerActionExpressionStandalone_return triggerActionExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerActionExpressionStandalone(); }\n+\tpublic HiveParser_AlterClauseParser.alterViewStatementSuffix_return alterViewStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterViewStatementSuffix(); }\n \n-\tpublic HiveParser_IdentifiersParser.booleanValueTok_return booleanValueTok() throws RecognitionException { return gIdentifiersParser.booleanValueTok(); }\n+\tpublic HiveParser_AlterClauseParser.blocking_return blocking() throws RecognitionException { return gAlterClauseParser.blocking(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixDropConstraint_return alterStatementSuffixDropConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixDropConstraint(); }\n+\tpublic HiveParser_CreateDDLParser.createTableStatement_return createTableStatement() throws RecognitionException { return gCreateDDLParser.createTableStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.atomjoinSource_return atomjoinSource() throws RecognitionException { return gFromClauseParser.atomjoinSource(); }\n+\tpublic HiveParser_IdentifiersParser.partitionSelectorVal_return partitionSelectorVal() throws RecognitionException { return gIdentifiersParser.partitionSelectorVal(); }\n \n-\tpublic HiveParser_SelectClauseParser.trfmClause_return trfmClause() throws RecognitionException { return gSelectClauseParser.trfmClause(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPartNot_return precedenceSimilarExpressionPartNot(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPartNot(t); }\n \n-\tpublic HiveParser_FromClauseParser.lateralView_return lateralView() throws RecognitionException { return gFromClauseParser.lateralView(); }\n+\tpublic HiveParser_IdentifiersParser.null_treatment_return null_treatment() throws RecognitionException { return gIdentifiersParser.null_treatment(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionMain_return precedenceSimilarExpressionMain() throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionMain(); }\n+\tpublic HiveParser_IdentifiersParser.rollupOldSyntax_return rollupOldSyntax() throws RecognitionException { return gIdentifiersParser.rollupOldSyntax(); }\n \n-\tpublic HiveParser_IdentifiersParser.intervalExpression_return intervalExpression() throws RecognitionException { return gIdentifiersParser.intervalExpression(); }\n+\tpublic HiveParser_IdentifiersParser.principalIdentifier_return principalIdentifier() throws RecognitionException { return gIdentifiersParser.principalIdentifier(); }\n \n-\tpublic HiveParser_FromClauseParser.valuesTableConstructor_return valuesTableConstructor() throws RecognitionException { return gFromClauseParser.valuesTableConstructor(); }\n+\tpublic HiveParser_CreateDDLParser.dataConnectorComment_return dataConnectorComment() throws RecognitionException { return gCreateDDLParser.dataConnectorComment(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorSuffixSetUrl_return alterDataConnectorSuffixSetUrl() throws RecognitionException { return gAlterClauseParser.alterDataConnectorSuffixSetUrl(); }\n+\tpublic HiveParser_SelectClauseParser.selectExpression_return selectExpression() throws RecognitionException { return gSelectClauseParser.selectExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.floorExpression_return floorExpression() throws RecognitionException { return gIdentifiersParser.floorExpression(); }\n+\tpublic HiveParser_IdentifiersParser.distributeByClause_return distributeByClause() throws RecognitionException { return gIdentifiersParser.distributeByClause(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatement_return alterStatement() throws RecognitionException { return gAlterClauseParser.alterStatement(); }\n+\tpublic HiveParser_ResourcePlanParser.triggerActionExpressionStandalone_return triggerActionExpressionStandalone() throws RecognitionException { return gResourcePlanParser.triggerActionExpressionStandalone(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDataConnectorStatementSuffix_return alterDataConnectorStatementSuffix() throws RecognitionException { return gAlterClauseParser.alterDataConnectorStatementSuffix(); }\n+\tpublic HiveParser_AlterClauseParser.partitionLocation_return partitionLocation() throws RecognitionException { return gAlterClauseParser.partitionLocation(); }\n \n-\tpublic HiveParser_IdentifiersParser.havingClause_return havingClause() throws RecognitionException { return gIdentifiersParser.havingClause(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCreateTag_return alterStatementSuffixCreateTag() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCreateTag(); }\n \n-\tpublic HiveParser_AlterClauseParser.fileFormat_return fileFormat() throws RecognitionException { return gAlterClauseParser.fileFormat(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceStarOperator_return precedenceStarOperator() throws RecognitionException { return gIdentifiersParser.precedenceStarOperator(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerOrExpression_return triggerOrExpression() throws RecognitionException { return gResourcePlanParser.triggerOrExpression(); }\n+\tpublic HiveParser_SelectClauseParser.selectItem_return selectItem() throws RecognitionException { return gSelectClauseParser.selectItem(); }\n \n-\tpublic HiveParser_ResourcePlanParser.createTriggerStatement_return createTriggerStatement() throws RecognitionException { return gResourcePlanParser.createTriggerStatement(); }\n+\tpublic HiveParser_PrepareStatementParser.prepareStatement_return prepareStatement() throws RecognitionException { return gPrepareStatementParser.prepareStatement(); }\n \n-\tpublic HiveParser_ResourcePlanParser.alterTriggerStatement_return alterTriggerStatement() throws RecognitionException { return gResourcePlanParser.alterTriggerStatement(); }\n+\tpublic HiveParser_IdentifiersParser.precedencePlusOperator_return precedencePlusOperator() throws RecognitionException { return gIdentifiersParser.precedencePlusOperator(); }\n \n-\tpublic HiveParser_IdentifiersParser.expressionsNotInParenthesis_return expressionsNotInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsNotInParenthesis(isStruct, forceStruct); }\n+\tpublic HiveParser_ResourcePlanParser.disable_return disable() throws RecognitionException { return gResourcePlanParser.disable(); }\n \n-\tpublic HiveParser_IdentifiersParser.nonReserved_return nonReserved() throws RecognitionException { return gIdentifiersParser.nonReserved(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddPartitions_return alterStatementSuffixAddPartitions(boolean table) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddPartitions(table); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSkewedby_return alterStatementSuffixSkewedby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSkewedby(); }\n+\tpublic HiveParser_FromClauseParser.uniqueJoinExpr_return uniqueJoinExpr() throws RecognitionException { return gFromClauseParser.uniqueJoinExpr(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterMaterializedViewSuffixRewrite_return alterMaterializedViewSuffixRewrite(CommonTree tableNameTree) throws RecognitionException { return gAlterClauseParser.alterMaterializedViewSuffixRewrite(tableNameTree); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUpdateStats_return alterStatementSuffixUpdateStats(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUpdateStats(partition); }\n+\n+\tpublic HiveParser_SelectClauseParser.selectClause_return selectClause() throws RecognitionException { return gSelectClauseParser.selectClause(); }\n \n \tpublic HiveParser_PrepareStatementParser.executeStatement_return executeStatement() throws RecognitionException { return gPrepareStatementParser.executeStatement(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpAssign_return rpAssign() throws RecognitionException { return gResourcePlanParser.rpAssign(); }\n+\tpublic HiveParser_IdentifiersParser.subQuerySelectorOperator_return subQuerySelectorOperator() throws RecognitionException { return gIdentifiersParser.subQuerySelectorOperator(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpAssignList_return rpAssignList() throws RecognitionException { return gResourcePlanParser.rpAssignList(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixClusterbySortby_return alterStatementSuffixClusterbySortby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixClusterbySortby(); }\n \n-\tpublic HiveParser_IdentifiersParser.columnRefOrderNotInParenthesis_return columnRefOrderNotInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderNotInParenthesis(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceEqualExpression_return precedenceEqualExpression() throws RecognitionException { return gIdentifiersParser.precedenceEqualExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionQuantifierPredicate_return precedenceSimilarExpressionQuantifierPredicate(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionQuantifierPredicate(t); }\n+\tpublic HiveParser_FromClauseParser.asOfClause_return asOfClause() throws RecognitionException { return gFromClauseParser.asOfClause(); }\n \n-\tpublic HiveParser_SelectClauseParser.window_specification_return window_specification(CommonTree nullTreatment) throws RecognitionException { return gSelectClauseParser.window_specification(nullTreatment); }\n+\tpublic HiveParser_IdentifiersParser.precedenceBitwiseXorExpression_return precedenceBitwiseXorExpression() throws RecognitionException { return gIdentifiersParser.precedenceBitwiseXorExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixRenamePart_return alterStatementSuffixRenamePart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixRenamePart(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementChangeColPosition_return alterStatementChangeColPosition() throws RecognitionException { return gAlterClauseParser.alterStatementChangeColPosition(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixUnArchive_return alterStatementSuffixUnArchive() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixUnArchive(); }\n+\tpublic HiveParser_IdentifiersParser.expressionWithAlias_return expressionWithAlias() throws RecognitionException { return gIdentifiersParser.expressionWithAlias(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixAddConstraint_return alterStatementSuffixAddConstraint() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixAddConstraint(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceSimilarExpressionPart_return precedenceSimilarExpressionPart(CommonTree t) throws RecognitionException { return gIdentifiersParser.precedenceSimilarExpressionPart(t); }\n \n-\tpublic HiveParser_IdentifiersParser.dateLiteral_return dateLiteral() throws RecognitionException { return gIdentifiersParser.dateLiteral(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceAmpersandExpression_return precedenceAmpersandExpression() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixLocation_return alterStatementSuffixLocation(boolean partition) throws RecognitionException { return gAlterClauseParser.alterStatementSuffixLocation(partition); }\n+\tpublic HiveParser_ResourcePlanParser.rpUnassign_return rpUnassign() throws RecognitionException { return gResourcePlanParser.rpUnassign(); }\n \n-\tpublic HiveParser_IdentifiersParser.firstExpressionsWithAlias_return firstExpressionsWithAlias() throws RecognitionException { return gIdentifiersParser.firstExpressionsWithAlias(); }\n+\tpublic HiveParser_AlterClauseParser.alterStatementSuffixSkewedby_return alterStatementSuffixSkewedby() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixSkewedby(); }\n \n-\tpublic HiveParser_ResourcePlanParser.triggerExpression_return triggerExpression() throws RecognitionException { return gResourcePlanParser.triggerExpression(); }\n+\tpublic HiveParser_CreateDDLParser.dataConnectorUrl_return dataConnectorUrl() throws RecognitionException { return gCreateDDLParser.dataConnectorUrl(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectTrfmClause_return selectTrfmClause() throws RecognitionException { return gSelectClauseParser.selectTrfmClause(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceOrExpression_return precedenceOrExpression() throws RecognitionException { return gIdentifiersParser.precedenceOrExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.precedenceAmpersandOperator_return precedenceAmpersandOperator() throws RecognitionException { return gIdentifiersParser.precedenceAmpersandOperator(); }\n+\tpublic HiveParser_FromClauseParser.valuesSource_return valuesSource() throws RecognitionException { return gFromClauseParser.valuesSource(); }\n \n-\tpublic HiveParser_FromClauseParser.firstValueRowConstructor_return firstValueRowConstructor() throws RecognitionException { return gFromClauseParser.firstValueRowConstructor(); }\n+\tpublic HiveParser_IdentifiersParser.expressionsNotInParenthesis_return expressionsNotInParenthesis(boolean isStruct, boolean forceStruct) throws RecognitionException { return gIdentifiersParser.expressionsNotInParenthesis(isStruct, forceStruct); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixSetManagedLocation_return alterDatabaseSuffixSetManagedLocation() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixSetManagedLocation(); }\n+\tpublic HiveParser_IdentifiersParser.stringLiteralSequence_return stringLiteralSequence() throws RecognitionException { return gIdentifiersParser.stringLiteralSequence(); }\n \n-\tpublic HiveParser_SelectClauseParser.selectItem_return selectItem() throws RecognitionException { return gSelectClauseParser.selectItem(); }\n+\tpublic HiveParser_IdentifiersParser.precedenceDistinctOperator_return precedenceDistinctOperator() throws RecognitionException { return gIdentifiersParser.precedenceDistinctOperator(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixCreateBranch_return alterStatementSuffixCreateBranch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixCreateBranch(); }\n+\tpublic HiveParser_PrepareStatementParser.executeParamList_return executeParamList() throws RecognitionException { return gPrepareStatementParser.executeParamList(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixTouch_return alterStatementSuffixTouch() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixTouch(); }\n+\tpublic HiveParser_IdentifiersParser.groupingSetExpressionMultiple_return groupingSetExpressionMultiple() throws RecognitionException { return gIdentifiersParser.groupingSetExpressionMultiple(); }\n \n-\tpublic HiveParser_ResourcePlanParser.rpUnassign_return rpUnassign() throws RecognitionException { return gResourcePlanParser.rpUnassign(); }\n+\tpublic HiveParser_IdentifiersParser.castExpression_return castExpression() throws RecognitionException { return gIdentifiersParser.castExpression(); }\n \n-\tpublic HiveParser_IdentifiersParser.timeQualifiers_return timeQualifiers() throws RecognitionException { return gIdentifiersParser.timeQualifiers(); }\n+\tpublic HiveParser_ResourcePlanParser.dropTriggerStatement_return dropTriggerStatement() throws RecognitionException { return gResourcePlanParser.dropTriggerStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.subQuerySource_return subQuerySource() throws RecognitionException { return gFromClauseParser.subQuerySource(); }\n+\tpublic HiveParser_IdentifiersParser.whenExpression_return whenExpression() throws RecognitionException { return gIdentifiersParser.whenExpression(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation_return alterTblPartitionStatementSuffixSkewedLocation() throws RecognitionException { return gAlterClauseParser.alterTblPartitionStatementSuffixSkewedLocation(); }\n+\tpublic HiveParser_ResourcePlanParser.rpUnassignList_return rpUnassignList() throws RecognitionException { return gResourcePlanParser.rpUnassignList(); }\n \n-\tpublic HiveParser_FromClauseParser.valuesSource_return valuesSource() throws RecognitionException { return gFromClauseParser.valuesSource(); }\n+\tpublic HiveParser_FromClauseParser.partitionedTableFunction_return partitionedTableFunction() throws RecognitionException { return gFromClauseParser.partitionedTableFunction(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterDatabaseSuffixProperties_return alterDatabaseSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterDatabaseSuffixProperties(); }\n+\tpublic HiveParser_IdentifiersParser.columnRefOrderNotInParenthesis_return columnRefOrderNotInParenthesis() throws RecognitionException { return gIdentifiersParser.columnRefOrderNotInParenthesis(); }\n \n-\tpublic HiveParser_AlterClauseParser.compactPool_return compactPool() throws RecognitionException { return gAlterClauseParser.compactPool(); }\n+\tpublic HiveParser_CreateDDLParser.createDataConnectorStatement_return createDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.createDataConnectorStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.tableSample_return tableSample() throws RecognitionException { return gFromClauseParser.tableSample(); }\n+\tpublic HiveParser_ResourcePlanParser.alterTriggerStatement_return alterTriggerStatement() throws RecognitionException { return gResourcePlanParser.alterTriggerStatement(); }\n \n-\tpublic HiveParser_AlterClauseParser.alterStatementSuffixStatsPart_return alterStatementSuffixStatsPart() throws RecognitionException { return gAlterClauseParser.alterStatementSuffixStatsPart(); }\n+\tpublic HiveParser_CreateDDLParser.dropDataConnectorStatement_return dropDataConnectorStatement() throws RecognitionException { return gCreateDDLParser.dropDataConnectorStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.joinSourcePart_return joinSourcePart() throws RecognitionException { return gFromClauseParser.joinSourcePart(); }\n+\tpublic HiveParser_IdentifiersParser.groupByEmpty_return groupByEmpty() throws RecognitionException { return gIdentifiersParser.groupByEmpty(); }\n \n-\tpublic HiveParser_IdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName_return sql11ReservedKeywordsUsedAsFunctionName() throws RecognitionException { return gIdentifiersParser.sql11ReservedKeywordsUsedAsFunctionName(); }\n+\tpublic HiveParser_ResourcePlanParser.createPoolStatement_return createPoolStatement() throws RecognitionException { return gResourcePlanParser.createPoolStatement(); }\n \n-\tpublic HiveParser_FromClauseParser.uniqueJoinSource_return uniqueJoinSource() throws RecognitionException { return gFromClauseParser.uniqueJoinSource(); }\n+\tpublic HiveParser_AlterClauseParser.alterViewSuffixProperties_return alterViewSuffixProperties() throws RecognitionException { return gAlterClauseParser.alterViewSuffixProperties(); }\n \n \tpublic final boolean synpred18_HiveParser() {\n \t\tstate.backtracking++;\n \t\tint start = input.mark();\n \t\ttry {\n \t\t\tsynpred18_HiveParser_fragment(); // can never throw exception\n \t\t} catch (RecognitionException re) {\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_AlterClauseParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_AlterClauseParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 AlterClauseParser.g 2024-03-25 16:12:01\n+// $ANTLR 3.5.2 AlterClauseParser.g 2025-01-31 11:32:59\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1432,15 +1432,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableName.add(tableName23.getTree());\n \t\t\t\t\tpushFollow(FOLLOW_optimizeTableStatementSuffix_in_alterStatement182);\n \t\t\t\t\toptimizeTableStatementSuffix24=optimizeTableStatementSuffix();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_optimizeTableStatementSuffix.add(optimizeTableStatementSuffix24.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: optimizeTableStatementSuffix, tableName\n+\t\t\t\t\t// elements: tableName, optimizeTableStatementSuffix\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -2040,15 +2040,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec38.getTree());\n \t\t\t\t\tpushFollow(FOLLOW_alterTblPartitionStatementSuffix_in_alterTableStatementSuffix336);\n \t\t\t\t\talterTblPartitionStatementSuffix39=alterTblPartitionStatementSuffix(true);\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_alterTblPartitionStatementSuffix.add(alterTblPartitionStatementSuffix39.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: alterTblPartitionStatementSuffix, partitionSpec\n+\t\t\t\t\t// elements: partitionSpec, alterTblPartitionStatementSuffix\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3652,15 +3652,15 @@\n \n \t\t\tpushFollow(FOLLOW_principalName_in_alterDatabaseSuffixSetOwner958);\n \t\t\tprincipalName86=gHiveParser.principalName();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_principalName.add(principalName86.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: dbName, principalName\n+\t\t\t// elements: principalName, dbName\n \t\t\t// token labels: \n \t\t\t// rule labels: dbName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4276,15 +4276,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_restrictOrCascade.add(restrictOrCascade100.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: restrictOrCascade, restrictOrCascade, columnNameTypeList, columnNameTypeList\n+\t\t\t// elements: columnNameTypeList, restrictOrCascade, restrictOrCascade, columnNameTypeList\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4915,15 +4915,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_restrictOrCascade.add(restrictOrCascade114.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: alterStatementChangeColPosition, oldName, comment, restrictOrCascade, colType, newName, alterColumnConstraint\n+\t\t\t// elements: restrictOrCascade, newName, colType, comment, alterColumnConstraint, alterStatementChangeColPosition, oldName\n \t\t\t// token labels: comment\n \t\t\t// rule labels: newName, oldName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5103,15 +5103,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: colName, comment, tableProperties, tableProperties, colName, comment\n+\t\t\t// elements: tableProperties, comment, colName, comment, tableProperties, colName\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5507,15 +5507,15 @@\n \t\t\t\t\tEarlyExitException eee = new EarlyExitException(24, input);\n \t\t\t\t\tthrow eee;\n \t\t\t\t}\n \t\t\t\tcnt24++;\n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: ifNotExists, alterStatementSuffixAddPartitionsElement, ifNotExists, alterStatementSuffixAddPartitionsElement\n+\t\t\t// elements: alterStatementSuffixAddPartitionsElement, ifNotExists, alterStatementSuffixAddPartitionsElement, ifNotExists\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -6237,15 +6237,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_replicationClause.add(replicationClause147.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: replicationClause, ifExists, ifExists, partitionSelectorSpec, replicationClause, partitionSelectorSpec, KW_PURGE\n+\t\t\t// elements: partitionSelectorSpec, ifExists, replicationClause, KW_PURGE, ifExists, replicationClause, partitionSelectorSpec\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -6665,15 +6665,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_tableProperties_in_alterViewSuffixProperties2277);\n \t\t\t\t\ttableProperties161=gHiveParser.tableProperties();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_tableProperties.add(tableProperties161.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ifExists, tableProperties\n+\t\t\t\t\t// elements: tableProperties, ifExists\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -6847,15 +6847,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_tableProperties.add(tableProperties166.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: serdeName, serdeName, tableProperties, tableProperties\n+\t\t\t\t\t// elements: serdeName, tableProperties, serdeName, tableProperties\n \t\t\t\t\t// token labels: serdeName\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -7088,15 +7088,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_partitionSpec.add(partitionSpec174.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: tableName, partitionSpec\n+\t\t\t// elements: partitionSpec, tableName\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -8230,15 +8230,15 @@\n \n \t\t\tpushFollow(FOLLOW_tableName_in_alterStatementSuffixExchangePartition2998);\n \t\t\texchangename=gHiveParser.tableName();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_tableName.add(exchangename.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: partitionSpec, exchangename\n+\t\t\t// elements: exchangename, partitionSpec\n \t\t\t// token labels: \n \t\t\t// rule labels: exchangename, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -8483,15 +8483,15 @@\n \n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: comment, tableProperties, colName\n+\t\t\t// elements: tableProperties, comment, colName\n \t\t\t// token labels: comment\n \t\t\t// rule labels: colName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -9058,15 +9058,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_orderByClause.add(orderByClause229.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: compactType, tableImplBuckets, compactPool, blocking, orderByClause, tableProperties\n+\t\t\t// elements: blocking, tableProperties, orderByClause, compactType, tableImplBuckets, compactPool\n \t\t\t// token labels: compactType\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -9732,15 +9732,15 @@\n \n \t\t\t\t\t}\n \n \t\t\t\t\tRPAREN244=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_alterStatementSuffixExecute3589); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN244);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_ROLLBACK, rollbackParam\n+\t\t\t\t\t// elements: rollbackParam, KW_ROLLBACK\n \t\t\t\t\t// token labels: rollbackParam\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9802,15 +9802,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: expireParam, KW_EXPIRE_SNAPSHOTS\n+\t\t\t\t\t// elements: KW_EXPIRE_SNAPSHOTS, expireParam\n \t\t\t\t\t// token labels: expireParam\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9864,15 +9864,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_expression.add(snapshotParam.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\tRPAREN252=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_alterStatementSuffixExecute3663); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN252);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: snapshotParam, KW_SET_CURRENT_SNAPSHOT\n+\t\t\t\t\t// elements: KW_SET_CURRENT_SNAPSHOT, snapshotParam\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: snapshotParam, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -9926,15 +9926,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: targetBranch, sourceBranch, KW_FAST_FORWARD\n+\t\t\t\t\t// elements: sourceBranch, KW_FAST_FORWARD, targetBranch\n \t\t\t\t\t// token labels: targetBranch, sourceBranch\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10040,15 +10040,15 @@\n \t\t\t\t\t{\n \t\t\t\t\ttoTimestamp=(Token)match(input,StringLiteral,FOLLOW_StringLiteral_in_alterStatementSuffixExecute3772); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_StringLiteral.add(toTimestamp);\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: fromTimestamp, KW_EXPIRE_SNAPSHOTS, toTimestamp\n+\t\t\t\t\t// elements: toTimestamp, KW_EXPIRE_SNAPSHOTS, fromTimestamp\n \t\t\t\t\t// token labels: fromTimestamp, toTimestamp\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10092,15 +10092,15 @@\n \t\t\t\t\tKW_LAST264=(Token)match(input,KW_LAST,FOLLOW_KW_LAST_in_alterStatementSuffixExecute3805); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_LAST.add(KW_LAST264);\n \n \t\t\t\t\tnumToRetain=(Token)match(input,Number,FOLLOW_Number_in_alterStatementSuffixExecute3809); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_Number.add(numToRetain);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: numToRetain, KW_EXPIRE_SNAPSHOTS, KW_RETAIN\n+\t\t\t\t\t// elements: numToRetain, KW_RETAIN, KW_EXPIRE_SNAPSHOTS\n \t\t\t\t\t// token labels: numToRetain\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10460,15 +10460,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_retentionOfSnapshots.add(retentionOfSnapshots279.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: retentionOfSnapshots, refRetain, branchName, snapshotIdOfRef\n+\t\t\t// elements: refRetain, branchName, snapshotIdOfRef, retentionOfSnapshots\n \t\t\t// token labels: \n \t\t\t// rule labels: branchName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -10990,15 +10990,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: minSnapshotsToKeep, timeUnit, maxSnapshotAge\n+\t\t\t// elements: minSnapshotsToKeep, maxSnapshotAge, timeUnit\n \t\t\t// token labels: maxSnapshotAge, minSnapshotsToKeep\n \t\t\t// rule labels: retval, timeUnit\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -11012,20 +11012,20 @@\n \t\t\t{\n \t\t\t\t// AlterClauseParser.g:546:8: ^( TOK_WITH_SNAPSHOT_RETENTION $minSnapshotsToKeep ( $maxSnapshotAge $timeUnit)? )\n \t\t\t\t{\n \t\t\t\tASTNode root_1 = (ASTNode)adaptor.nil();\n \t\t\t\troot_1 = (ASTNode)adaptor.becomeRoot((ASTNode)adaptor.create(TOK_WITH_SNAPSHOT_RETENTION, \"TOK_WITH_SNAPSHOT_RETENTION\"), root_1);\n \t\t\t\tadaptor.addChild(root_1, stream_minSnapshotsToKeep.nextNode());\n \t\t\t\t// AlterClauseParser.g:546:58: ( $maxSnapshotAge $timeUnit)?\n-\t\t\t\tif ( stream_timeUnit.hasNext()||stream_maxSnapshotAge.hasNext() ) {\n+\t\t\t\tif ( stream_maxSnapshotAge.hasNext()||stream_timeUnit.hasNext() ) {\n \t\t\t\t\tadaptor.addChild(root_1, stream_maxSnapshotAge.nextNode());\n \t\t\t\t\tadaptor.addChild(root_1, stream_timeUnit.nextTree());\n \t\t\t\t}\n-\t\t\t\tstream_timeUnit.reset();\n \t\t\t\tstream_maxSnapshotAge.reset();\n+\t\t\t\tstream_timeUnit.reset();\n \n \t\t\t\tadaptor.addChild(root_0, root_1);\n \t\t\t\t}\n \n \t\t\t}\n \n \n@@ -11115,15 +11115,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_alterStatementSuffixDropTag4270);\n \t\t\ttagName=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_identifier.add(tagName.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: ifExists, tagName\n+\t\t\t// elements: tagName, ifExists\n \t\t\t// token labels: \n \t\t\t// rule labels: tagName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -11258,15 +11258,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_refRetain.add(refRetain303.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: refRetain, tagName, snapshotIdOfRef\n+\t\t\t// elements: snapshotIdOfRef, tagName, refRetain\n \t\t\t// token labels: \n \t\t\t// rule labels: tagName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -11455,26 +11455,26 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: serdeCls, outFmt, inDriver, inFmt, outDriver\n-\t\t\t\t\t// token labels: inFmt, inDriver, serdeCls, outDriver, outFmt\n+\t\t\t\t\t// elements: outDriver, outFmt, inDriver, inFmt, serdeCls\n+\t\t\t\t\t// token labels: inFmt, inDriver, outDriver, serdeCls, outFmt\n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleTokenStream stream_inFmt=new RewriteRuleTokenStream(adaptor,\"token inFmt\",inFmt);\n \t\t\t\t\tRewriteRuleTokenStream stream_inDriver=new RewriteRuleTokenStream(adaptor,\"token inDriver\",inDriver);\n-\t\t\t\t\tRewriteRuleTokenStream stream_serdeCls=new RewriteRuleTokenStream(adaptor,\"token serdeCls\",serdeCls);\n \t\t\t\t\tRewriteRuleTokenStream stream_outDriver=new RewriteRuleTokenStream(adaptor,\"token outDriver\",outDriver);\n+\t\t\t\t\tRewriteRuleTokenStream stream_serdeCls=new RewriteRuleTokenStream(adaptor,\"token serdeCls\",serdeCls);\n \t\t\t\t\tRewriteRuleTokenStream stream_outFmt=new RewriteRuleTokenStream(adaptor,\"token outFmt\",outFmt);\n \t\t\t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n \n \t\t\t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t\t\t// 567:7: -> ^( TOK_TABLEFILEFORMAT $inFmt $outFmt $serdeCls ( $inDriver)? ( $outDriver)? )\n \t\t\t\t\t{\n \t\t\t\t\t\t// AlterClauseParser.g:567:10: ^( TOK_TABLEFILEFORMAT $inFmt $outFmt $serdeCls ( $inDriver)? ( $outDriver)? )\n@@ -11811,15 +11811,15 @@\n \n \t\t\tpushFollow(FOLLOW_dcProperties_in_alterDataConnectorSuffixProperties4528);\n \t\t\tdcProperties314=gHiveParser.dcProperties();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_dcProperties.add(dcProperties314.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: name, dcProperties\n+\t\t\t// elements: dcProperties, name\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -11912,15 +11912,15 @@\n \n \t\t\tpushFollow(FOLLOW_principalName_in_alterDataConnectorSuffixSetOwner4578);\n \t\t\tprincipalName317=gHiveParser.principalName();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_principalName.add(principalName317.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: principalName, dcName\n+\t\t\t// elements: dcName, principalName\n \t\t\t// token labels: \n \t\t\t// rule labels: dcName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_CreateDDLParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_CreateDDLParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 CreateDDLParser.g 2024-03-25 16:12:04\n+// $ANTLR 3.5.2 CreateDDLParser.g 2025-01-31 11:33:01\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1808,15 +1808,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableComment, columnNameTypeOrConstraintList, temp, tableFileFormat, tableBuckets, tableSkewed, ext, ifNotExists, tableRowFormat, tablePropertiesPrefixed, name, likeTableOrFile, tableLocation, selectStatementWithCTE, trans, createTablePartitionSpec\n+\t\t\t\t\t// elements: tableFileFormat, name, ext, tableBuckets, selectStatementWithCTE, ifNotExists, tableRowFormat, likeTableOrFile, tableLocation, trans, columnNameTypeOrConstraintList, createTablePartitionSpec, tablePropertiesPrefixed, tableComment, temp, tableSkewed\n \t\t\t\t\t// token labels: ext, temp, trans\n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -2294,15 +2294,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ifNotExists, tablePropertiesPrefixed, likeTableOrFile, tableComment, columnNameTypeOrConstraintList, tableRowFormat, tableFileFormat, selectStatementWithCTE, createTablePartitionSpec, name, mgd, tableSkewed, tableBuckets, tableLocation\n+\t\t\t\t\t// elements: tablePropertiesPrefixed, tableSkewed, tableLocation, likeTableOrFile, tableBuckets, selectStatementWithCTE, ifNotExists, tableComment, name, columnNameTypeOrConstraintList, tableFileFormat, mgd, tableRowFormat, createTablePartitionSpec\n \t\t\t\t\t// token labels: mgd\n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -2555,15 +2555,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_dcProperties.add(dcprops.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: dcprops, dataConnectorUrl, dataConnectorType, name, dataConnectorComment, ifNotExists\n+\t\t\t// elements: dataConnectorUrl, dataConnectorComment, name, ifNotExists, dataConnectorType, dcprops\n \t\t\t// token labels: \n \t\t\t// rule labels: name, retval, dcprops\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3046,15 +3046,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_dropDataConnectorStatement1229);\n \t\t\tidentifier66=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_identifier.add(identifier66.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: identifier, ifExists\n+\t\t\t// elements: ifExists, identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_FromClauseParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_FromClauseParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 FromClauseParser.g 2024-03-25 16:12:02\n+// $ANTLR 3.5.2 FromClauseParser.g 2025-01-31 11:33:00\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -3394,15 +3394,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: function, tableAlias, identifier\n+\t\t\t\t\t// elements: function, identifier, tableAlias\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3530,15 +3530,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: tableAlias, identifier, function\n+\t\t\t\t\t// elements: tableAlias, function, identifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3697,15 +3697,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: valuesClause, identifier, tableAlias\n+\t\t\t\t\t// elements: valuesClause, tableAlias, identifier\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -4002,15 +4002,15 @@\n \n \t\t\t}\n \n \t\t\tRPAREN114=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_tableBucketSample1292); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN114);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: denominator, numerator, expr\n+\t\t\t// elements: denominator, expr, numerator\n \t\t\t// token labels: denominator, numerator\n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: expr\n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4598,28 +4598,28 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(alias.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: props, tabname, alias, ts, asOf\n+\t\t\t// elements: tabname, alias, ts, props, asOf\n \t\t\t// token labels: \n-\t\t\t// rule labels: tabname, asOf, alias, retval, props, ts\n+\t\t\t// rule labels: tabname, asOf, alias, retval, ts, props\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_tabname=new RewriteRuleSubtreeStream(adaptor,\"rule tabname\",tabname!=null?tabname.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_asOf=new RewriteRuleSubtreeStream(adaptor,\"rule asOf\",asOf!=null?asOf.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_alias=new RewriteRuleSubtreeStream(adaptor,\"rule alias\",alias!=null?alias.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n-\t\t\tRewriteRuleSubtreeStream stream_props=new RewriteRuleSubtreeStream(adaptor,\"rule props\",props!=null?props.getTree():null);\n \t\t\tRewriteRuleSubtreeStream stream_ts=new RewriteRuleSubtreeStream(adaptor,\"rule ts\",ts!=null?ts.getTree():null);\n+\t\t\tRewriteRuleSubtreeStream stream_props=new RewriteRuleSubtreeStream(adaptor,\"rule props\",props!=null?props.getTree():null);\n \n \t\t\troot_0 = (ASTNode)adaptor.nil();\n \t\t\t// 213:5: -> ^( TOK_TABREF $tabname ( $props)? ( $ts)? ( $alias)? ( $asOf)? )\n \t\t\t{\n \t\t\t\t// FromClauseParser.g:213:8: ^( TOK_TABREF $tabname ( $props)? ( $ts)? ( $alias)? ( $asOf)? )\n \t\t\t\t{\n \t\t\t\tASTNode root_1 = (ASTNode)adaptor.nil();\n@@ -4985,15 +4985,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(alias.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: ts, alias, tabname\n+\t\t\t// elements: tabname, ts, alias\n \t\t\t// token labels: \n \t\t\t// rule labels: tabname, alias, retval, ts\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5120,15 +5120,15 @@\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\tif ( state.backtracking==0 ) {tables.add(new ImmutablePair<>((db!=null?input.toString(db.start,db.stop):null), (tab!=null?input.toString(tab.start,tab.stop):null)));}\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: db, meta, tab\n+\t\t\t\t\t// elements: db, tab, meta\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: tab, meta, db, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -5416,15 +5416,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_subQuerySource1916);\n \t\t\tidentifier141=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_identifier.add(identifier141.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: identifier, queryStatementExpression\n+\t\t\t// elements: queryStatementExpression, identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5562,15 +5562,15 @@\n \t\t\t\t\t\t\tif ( state.backtracking==0 ) stream_orderByClause.add(orderByClause143.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: partitionByClause, orderByClause\n+\t\t\t\t\t// elements: orderByClause, partitionByClause\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -6431,15 +6431,15 @@\n \t\t\t\t\tif ( state.backtracking==0 ) stream_identifier.add(alias.getTree());\n \t\t\t\t\t}\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: name, ptfsrc, expression, alias, spec\n+\t\t\t// elements: alias, spec, expression, name, ptfsrc\n \t\t\t// token labels: \n \t\t\t// rule labels: ptfsrc, name, alias, spec, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_IdentifiersParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_IdentifiersParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 IdentifiersParser.g 2024-03-25 16:12:03\n+// $ANTLR 3.5.2 IdentifiersParser.g 2025-01-31 11:33:01\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -2733,15 +2733,15 @@\n \t\t\t\tdefault :\n \t\t\t\t\tbreak loop15;\n \t\t\t\t}\n \t\t\t}\n \n \t\t\tif ( state.backtracking==0 ) { incAliasCounter(); }\n \t\t\t// AST REWRITE\n-\t\t\t// elements: expressionWithAlias, expressionWithAlias, identifier\n+\t\t\t// elements: identifier, expressionWithAlias, expressionWithAlias\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3966,15 +3966,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_selectExpression.add(str.getTree());\n \t\t\t}\n \n \t\t\tRPAREN79=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_trimFunction1438); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN79);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: str, trim_characters, str, trim_characters, str, trim_characters\n+\t\t\t// elements: trim_characters, str, str, str, trim_characters, trim_characters\n \t\t\t// token labels: \n \t\t\t// rule labels: str, trim_characters, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -4785,15 +4785,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: selectExpression, functionName, ws, functionName, ws, ws, functionName, functionName, selectExpression, selectExpression, ordBy\n+\t\t\t\t\t// elements: selectExpression, functionName, functionName, ordBy, ws, ws, selectExpression, selectExpression, functionName, functionName, ws\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: ordBy, ws, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -5271,15 +5271,15 @@\n \n \t\t\t}\n \n \t\t\tRPAREN107=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_castExpression2132); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN107);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, toType, expression, StringLiteral, expression, StringLiteral\n+\t\t\t// elements: toType, StringLiteral, expression, expression, expression, StringLiteral\n \t\t\t// token labels: \n \t\t\t// rule labels: toType, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -5778,15 +5778,15 @@\n \n \t\t\t}\n \n \t\t\tRPAREN129=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_floorExpression2444); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN129);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, floorUnit, expression\n+\t\t\t// elements: floorUnit, expression, expression\n \t\t\t// token labels: \n \t\t\t// rule labels: retval, floorUnit\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -6253,15 +6253,15 @@\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_expression.add(expression141.getTree());\n \t\t\tRPAREN142=(Token)match(input,RPAREN,FOLLOW_RPAREN_in_extractExpression2666); if (state.failed) return retval; \n \t\t\tif ( state.backtracking==0 ) stream_RPAREN.add(RPAREN142);\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: expression, timeUnit\n+\t\t\t// elements: timeUnit, expression\n \t\t\t// token labels: \n \t\t\t// rule labels: retval, timeUnit\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -10338,15 +10338,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_subQueryExpression_in_precedenceSimilarExpression4385);\n \t\t\t\t\tsubQueryExpression253=subQueryExpression();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_subQueryExpression.add(subQueryExpression253.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_EXISTS, subQueryExpression\n+\t\t\t\t\t// elements: subQueryExpression, KW_EXISTS\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10541,15 +10541,15 @@\n \t\t\t\t\tequalExpr=precedenceBitwiseOrExpression();\n \t\t\t\t\tstate._fsp--;\n \t\t\t\t\tif (state.failed) return retval;\n \t\t\t\t\tif ( state.backtracking==0 ) stream_precedenceBitwiseOrExpression.add(equalExpr.getTree());\n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: precedenceSimilarOperator, equalExpr\n+\t\t\t\t\t// elements: equalExpr, precedenceSimilarOperator\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval, equalExpr\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -10993,15 +10993,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_quantifierType.add(quantifierType268.getTree());\n \t\t\tpushFollow(FOLLOW_subQueryExpression_in_precedenceSimilarExpressionQuantifierPredicate4718);\n \t\t\tsubQueryExpression269=subQueryExpression();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_subQueryExpression.add(subQueryExpression269.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: quantifierType, subQuerySelectorOperator, subQueryExpression\n+\t\t\t// elements: subQueryExpression, subQuerySelectorOperator, quantifierType\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_PrepareStatementParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_PrepareStatementParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 PrepareStatementParser.g 2024-03-25 16:12:04\n+// $ANTLR 3.5.2 PrepareStatementParser.g 2025-01-31 11:33:01\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1057,15 +1057,15 @@\n \n \t\t\tpushFollow(FOLLOW_queryStatementExpression_in_prepareStatement72);\n \t\t\tqueryStatementExpression4=gHiveParser.queryStatementExpression();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_queryStatementExpression.add(queryStatementExpression4.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: identifier, queryStatementExpression\n+\t\t\t// elements: queryStatementExpression, identifier\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_retval=new RewriteRuleSubtreeStream(adaptor,\"rule retval\",retval!=null?retval.getTree():null);\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_ResourcePlanParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_ResourcePlanParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 ResourcePlanParser.g 2024-03-25 16:12:04\n+// $ANTLR 3.5.2 ResourcePlanParser.g 2025-01-31 11:33:01\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -2091,15 +2091,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_identifier_in_createResourcePlanStatement435);\n \t\t\t\t\tlikeName=gHiveParser.identifier();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_identifier.add(likeName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: name, likeName, ifNotExists\n+\t\t\t\t\t// elements: name, ifNotExists, likeName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: likeName, name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_likeName=new RewriteRuleSubtreeStream(adaptor,\"rule likeName\",likeName!=null?likeName.getTree():null);\n@@ -2171,15 +2171,15 @@\n \t\t\t\t\t\t\tstream_rpAssignList.add(rpAssignList36.getTree());\n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: ifNotExists, rpAssignList, name\n+\t\t\t\t\t// elements: name, ifNotExists, rpAssignList\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -2943,15 +2943,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_identifier_in_alterResourcePlanStatement733);\n \t\t\t\t\tnewName=gHiveParser.identifier();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_identifier.add(newName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: newName, name\n+\t\t\t\t\t// elements: name, newName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: newName, name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_newName=new RewriteRuleSubtreeStream(adaptor,\"rule newName\",newName!=null?newName.getTree():null);\n@@ -3062,15 +3062,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: enable, name, activate\n+\t\t\t\t\t// elements: activate, name, enable\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: name, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_name=new RewriteRuleSubtreeStream(adaptor,\"rule name\",name!=null?name.getTree():null);\n@@ -4425,15 +4425,15 @@\n \n \t\t\tpushFollow(FOLLOW_triggerActionExpression_in_createTriggerStatement1362);\n \t\t\ttriggerActionExpression106=triggerActionExpression();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_triggerActionExpression.add(triggerActionExpression106.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: triggerName, triggerExpression, rpName, triggerActionExpression\n+\t\t\t// elements: triggerActionExpression, triggerName, triggerExpression, rpName\n \t\t\t// token labels: \n \t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -4693,15 +4693,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_triggerActionExpression_in_alterTriggerStatement1440);\n \t\t\t\t\ttriggerActionExpression113=triggerActionExpression();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_triggerActionExpression.add(triggerActionExpression113.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: triggerExpression, triggerName, rpName, triggerActionExpression\n+\t\t\t\t\t// elements: triggerActionExpression, triggerName, rpName, triggerExpression\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -4748,15 +4748,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_poolPath_in_alterTriggerStatement1486);\n \t\t\t\t\tpoolName=poolPath();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_poolPath.add(poolName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: rpName, triggerName, poolName\n+\t\t\t\t\t// elements: triggerName, poolName, rpName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval, poolName\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -4803,15 +4803,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_poolPath_in_alterTriggerStatement1521);\n \t\t\t\t\tpoolName=poolPath();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_poolPath.add(poolName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: rpName, triggerName, poolName\n+\t\t\t\t\t// elements: poolName, rpName, triggerName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval, poolName\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -4853,15 +4853,15 @@\n \t\t\t\t\tKW_TO121=(Token)match(input,KW_TO,FOLLOW_KW_TO_in_alterTriggerStatement1551);  \n \t\t\t\t\tstream_KW_TO.add(KW_TO121);\n \n \t\t\t\t\tKW_UNMANAGED122=(Token)match(input,KW_UNMANAGED,FOLLOW_KW_UNMANAGED_in_alterTriggerStatement1553);  \n \t\t\t\t\tstream_KW_UNMANAGED.add(KW_UNMANAGED122);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: triggerName, rpName\n+\t\t\t\t\t// elements: rpName, triggerName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -4902,15 +4902,15 @@\n \t\t\t\t\tKW_FROM124=(Token)match(input,KW_FROM,FOLLOW_KW_FROM_in_alterTriggerStatement1581);  \n \t\t\t\t\tstream_KW_FROM.add(KW_FROM124);\n \n \t\t\t\t\tKW_UNMANAGED125=(Token)match(input,KW_UNMANAGED,FOLLOW_KW_UNMANAGED_in_alterTriggerStatement1583);  \n \t\t\t\t\tstream_KW_UNMANAGED.add(KW_UNMANAGED125);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: triggerName, rpName\n+\t\t\t\t\t// elements: rpName, triggerName\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -5535,15 +5535,15 @@\n \n \t\t\tpushFollow(FOLLOW_poolAssignList_in_createPoolStatement1889);\n \t\t\tpoolAssignList145=poolAssignList();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_poolAssignList.add(poolAssignList145.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: poolAssignList, rpName, poolPath\n+\t\t\t// elements: rpName, poolPath, poolAssignList\n \t\t\t// token labels: \n \t\t\t// rule labels: rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -5706,15 +5706,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_poolAssignList_in_alterPoolStatement1958);\n \t\t\t\t\tpoolAssignList151=poolAssignList();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_poolAssignList.add(poolAssignList151.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: rpName, poolAssignList, poolPath\n+\t\t\t\t\t// elements: poolPath, rpName, poolAssignList\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -5808,15 +5808,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_identifier_in_alterPoolStatement2024);\n \t\t\t\t\ttriggerName=gHiveParser.identifier();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_identifier.add(triggerName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: triggerName, poolPath, rpName\n+\t\t\t\t\t// elements: rpName, triggerName, poolPath\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -5859,15 +5859,15 @@\n \n \t\t\t\t\tpushFollow(FOLLOW_identifier_in_alterPoolStatement2058);\n \t\t\t\t\ttriggerName=gHiveParser.identifier();\n \t\t\t\t\tstate._fsp--;\n \n \t\t\t\t\tstream_identifier.add(triggerName.getTree());\n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: poolPath, rpName, triggerName\n+\t\t\t\t\t// elements: rpName, triggerName, poolPath\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: triggerName, rpName, retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tretval.tree = root_0;\n \t\t\t\t\tRewriteRuleSubtreeStream stream_triggerName=new RewriteRuleSubtreeStream(adaptor,\"rule triggerName\",triggerName!=null?triggerName.getTree():null);\n@@ -5972,15 +5972,15 @@\n \n \t\t\tpushFollow(FOLLOW_poolPath_in_dropPoolStatement2118);\n \t\t\tpoolPath161=poolPath();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_poolPath.add(poolPath161.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: poolPath, rpName\n+\t\t\t// elements: rpName, poolPath\n \t\t\t// token labels: \n \t\t\t// rule labels: rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleSubtreeStream stream_rpName=new RewriteRuleSubtreeStream(adaptor,\"rule rpName\",rpName!=null?rpName.getTree():null);\n@@ -6224,15 +6224,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: order, unmanaged, mappingType, name, rpName, path\n+\t\t\t// elements: rpName, unmanaged, path, order, mappingType, name\n \t\t\t// token labels: mappingType, name, order\n \t\t\t// rule labels: path, rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleTokenStream stream_mappingType=new RewriteRuleTokenStream(adaptor,\"token mappingType\",mappingType);\n@@ -6499,15 +6499,15 @@\n \t\t\t\t\tbreak;\n \n \t\t\t}\n \n \t\t\t}\n \n \t\t\t// AST REWRITE\n-\t\t\t// elements: unmanaged, order, mappingType, path, rpName, name\n+\t\t\t// elements: rpName, name, order, mappingType, path, unmanaged\n \t\t\t// token labels: mappingType, name, order\n \t\t\t// rule labels: path, rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleTokenStream stream_mappingType=new RewriteRuleTokenStream(adaptor,\"token mappingType\",mappingType);\n@@ -6681,15 +6681,15 @@\n \n \t\t\tpushFollow(FOLLOW_identifier_in_dropMappingStatement2469);\n \t\t\trpName=gHiveParser.identifier();\n \t\t\tstate._fsp--;\n \n \t\t\tstream_identifier.add(rpName.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: rpName, name, mappingType\n+\t\t\t// elements: rpName, mappingType, name\n \t\t\t// token labels: mappingType, name\n \t\t\t// rule labels: rpName, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tretval.tree = root_0;\n \t\t\tRewriteRuleTokenStream stream_mappingType=new RewriteRuleTokenStream(adaptor,\"token mappingType\",mappingType);\n"}, {"source1": "org/apache/hadoop/hive/ql/parse/HiveParser_SelectClauseParser.java", "source2": "org/apache/hadoop/hive/ql/parse/HiveParser_SelectClauseParser.java", "unified_diff": "@@ -1,8 +1,8 @@\n-// $ANTLR 3.5.2 SelectClauseParser.g 2024-03-25 16:12:01\n+// $ANTLR 3.5.2 SelectClauseParser.g 2025-01-31 11:32:59\n \n package org.apache.hadoop.hive.ql.parse;\n \n import java.util.Arrays;\n import java.util.ArrayList;\n import java.util.Collection;\n import java.util.HashMap;\n@@ -1176,15 +1176,15 @@\n \n \t\t\t\t\t\t\t}\n \t\t\t\t\t\t\tbreak;\n \n \t\t\t\t\t}\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: selectList, selectTrfmClause, QUERY_HINT, QUERY_HINT, selectList, QUERY_HINT\n+\t\t\t\t\t// elements: selectList, QUERY_HINT, QUERY_HINT, QUERY_HINT, selectTrfmClause, selectList\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -1701,15 +1701,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_rowFormat.add(outSerde.getTree());\n \t\t\tpushFollow(FOLLOW_recordReader_in_selectTrfmClause357);\n \t\t\toutRec=gHiveParser.recordReader();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_recordReader.add(outRec.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: aliasList, inRec, inSerde, outSerde, selectExpressionList, StringLiteral, columnNameTypeList, outRec\n+\t\t\t// elements: StringLiteral, aliasList, inRec, outSerde, selectExpressionList, inSerde, columnNameTypeList, outRec\n \t\t\t// token labels: \n \t\t\t// rule labels: inRec, outRec, inSerde, outSerde, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -2321,15 +2321,15 @@\n \t\t\tif ( state.backtracking==0 ) stream_rowFormat.add(outSerde.getTree());\n \t\t\tpushFollow(FOLLOW_recordReader_in_trfmClause622);\n \t\t\toutRec=gHiveParser.recordReader();\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_recordReader.add(outRec.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: outRec, StringLiteral, selectExpressionList, inRec, inSerde, columnNameTypeList, outSerde, aliasList\n+\t\t\t// elements: columnNameTypeList, selectExpressionList, outSerde, StringLiteral, aliasList, inRec, outRec, inSerde\n \t\t\t// token labels: \n \t\t\t// rule labels: inRec, outRec, inSerde, outSerde, retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -2757,15 +2757,15 @@\n \n \t\t\tpushFollow(FOLLOW_window_specification_in_window_defn825);\n \t\t\twindow_specification56=window_specification(null);\n \t\t\tstate._fsp--;\n \t\t\tif (state.failed) return retval;\n \t\t\tif ( state.backtracking==0 ) stream_window_specification.add(window_specification56.getTree());\n \t\t\t// AST REWRITE\n-\t\t\t// elements: window_specification, identifier\n+\t\t\t// elements: identifier, window_specification\n \t\t\t// token labels: \n \t\t\t// rule labels: retval\n \t\t\t// token list labels: \n \t\t\t// rule list labels: \n \t\t\t// wildcard labels: \n \t\t\tif ( state.backtracking==0 ) {\n \t\t\tretval.tree = root_0;\n@@ -3819,15 +3819,15 @@\n \t\t\t\t\tKW_UNBOUNDED73=(Token)match(input,KW_UNBOUNDED,FOLLOW_KW_UNBOUNDED_in_window_frame_start_boundary1064); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_UNBOUNDED.add(KW_UNBOUNDED73);\n \n \t\t\t\t\tKW_PRECEDING74=(Token)match(input,KW_PRECEDING,FOLLOW_KW_PRECEDING_in_window_frame_start_boundary1066); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_PRECEDING.add(KW_PRECEDING74);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_UNBOUNDED, KW_PRECEDING\n+\t\t\t\t\t// elements: KW_PRECEDING, KW_UNBOUNDED\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n@@ -3896,15 +3896,15 @@\n \t\t\t\t\tNumber77=(Token)match(input,Number,FOLLOW_Number_in_window_frame_start_boundary1097); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_Number.add(Number77);\n \n \t\t\t\t\tKW_PRECEDING78=(Token)match(input,KW_PRECEDING,FOLLOW_KW_PRECEDING_in_window_frame_start_boundary1099); if (state.failed) return retval; \n \t\t\t\t\tif ( state.backtracking==0 ) stream_KW_PRECEDING.add(KW_PRECEDING78);\n \n \t\t\t\t\t// AST REWRITE\n-\t\t\t\t\t// elements: KW_PRECEDING, Number\n+\t\t\t\t\t// elements: Number, KW_PRECEDING\n \t\t\t\t\t// token labels: \n \t\t\t\t\t// rule labels: retval\n \t\t\t\t\t// token list labels: \n \t\t\t\t\t// rule list labels: \n \t\t\t\t\t// wildcard labels: \n \t\t\t\t\tif ( state.backtracking==0 ) {\n \t\t\t\t\tretval.tree = root_0;\n"}]}
