diff options
author | Patrick J Volkerding <volkerdi@slackware.com> | 2020-11-06 21:25:18 +0000 |
---|---|---|
committer | Eric Hameleers <alien@slackware.com> | 2020-11-07 08:59:53 +0100 |
commit | aae320a395fbbe0be771ead4963da6d61e50c5ae (patch) | |
tree | b28a177d33110ca0705d049a3bc6fac06034799e | |
parent | 1e2b70eb9ad995e99d86e49d96104c090a860709 (diff) | |
download | current-aae320a395fbbe0be771ead4963da6d61e50c5ae.tar.gz |
Fri Nov 6 21:25:18 UTC 202020201106212518
d/rust-1.47.0-x86_64-2.txz: Rebuilt.
Switch back to Rust 1.47.0 now that patches are available to fix Firefox
and Thunderbird.
xap/mozilla-thunderbird-78.4.1-x86_64-1.txz: Upgraded.
This is a bugfix release.
For more information, see:
https://www.mozilla.org/en-US/thunderbird/78.4.1/releasenotes/
-rw-r--r-- | ChangeLog.rss | 21 | ||||
-rw-r--r-- | ChangeLog.txt | 9 | ||||
-rw-r--r-- | FILELIST.TXT | 80 | ||||
-rwxr-xr-x | recompress.sh | 2 | ||||
-rwxr-xr-x | source/d/rust/rust.SlackBuild | 8 | ||||
-rw-r--r-- | source/d/rust/rust.url | 14 | ||||
-rwxr-xr-x | source/xap/mozilla-firefox/mozilla-firefox.SlackBuild | 3 | ||||
-rw-r--r-- | source/xap/mozilla-firefox/rust_1.47.0.patch | 30905 | ||||
-rwxr-xr-x | source/xap/mozilla-thunderbird/mozilla-thunderbird.SlackBuild | 3 | ||||
-rw-r--r-- | source/xap/mozilla-thunderbird/rust_1.47.0.patch | 30905 | ||||
-rwxr-xr-x | source/xap/seamonkey/seamonkey.SlackBuild | 4 |
11 files changed, 61900 insertions, 54 deletions
diff --git a/ChangeLog.rss b/ChangeLog.rss index 25e4c91c..107b7964 100644 --- a/ChangeLog.rss +++ b/ChangeLog.rss @@ -11,10 +11,27 @@ <description>Tracking Slackware development in git.</description> <language>en-us</language> <id xmlns="http://www.w3.org/2005/Atom">urn:uuid:c964f45e-6732-11e8-bbe5-107b4450212f</id> - <pubDate>Thu, 5 Nov 2020 20:40:51 GMT</pubDate> - <lastBuildDate>Fri, 6 Nov 2020 07:59:48 GMT</lastBuildDate> + <pubDate>Fri, 6 Nov 2020 21:25:18 GMT</pubDate> + <lastBuildDate>Sat, 7 Nov 2020 07:59:46 GMT</lastBuildDate> <generator>maintain_current_git.sh v 1.12</generator> <item> + <title>Fri, 6 Nov 2020 21:25:18 GMT</title> + <pubDate>Fri, 6 Nov 2020 21:25:18 GMT</pubDate> + <link>https://git.slackware.nl/current/tag/?h=20201106212518</link> + <guid isPermaLink="false">20201106212518</guid> + <description> + <![CDATA[<pre> +d/rust-1.47.0-x86_64-2.txz: Rebuilt. + Switch back to Rust 1.47.0 now that patches are available to fix Firefox + and Thunderbird. +xap/mozilla-thunderbird-78.4.1-x86_64-1.txz: Upgraded. + This is a bugfix release. + For more information, see: + https://www.mozilla.org/en-US/thunderbird/78.4.1/releasenotes/ + </pre>]]> + </description> + </item> + <item> <title>Thu, 5 Nov 2020 20:40:51 GMT</title> <pubDate>Thu, 5 Nov 2020 20:40:51 GMT</pubDate> <link>https://git.slackware.nl/current/tag/?h=20201105204051</link> diff --git a/ChangeLog.txt b/ChangeLog.txt index a28834d0..d64e2015 100644 --- a/ChangeLog.txt +++ b/ChangeLog.txt @@ -1,3 +1,12 @@ +Fri Nov 6 21:25:18 UTC 2020 +d/rust-1.47.0-x86_64-2.txz: Rebuilt. + Switch back to Rust 1.47.0 now that patches are available to fix Firefox + and Thunderbird. +xap/mozilla-thunderbird-78.4.1-x86_64-1.txz: Upgraded. + This is a bugfix release. + For more information, see: + https://www.mozilla.org/en-US/thunderbird/78.4.1/releasenotes/ ++--------------------------+ Thu Nov 5 20:40:51 UTC 2020 a/hwdata-0.341-noarch-1.txz: Upgraded. a/kernel-generic-5.4.75-x86_64-1.txz: Upgraded. diff --git a/FILELIST.TXT b/FILELIST.TXT index f5579cfd..06402b1a 100644 --- a/FILELIST.TXT +++ b/FILELIST.TXT @@ -1,20 +1,20 @@ -Thu Nov 5 20:47:14 UTC 2020 +Fri Nov 6 21:54:02 UTC 2020 Here is the file list for this directory. If you are using a mirror site and find missing or extra files in the disk subdirectories, please have the archive administrator refresh the mirror. -drwxr-xr-x 12 root root 4096 2020-11-05 20:40 . +drwxr-xr-x 12 root root 4096 2020-11-06 21:28 . -rw-r--r-- 1 root root 10064 2016-06-30 18:39 ./ANNOUNCE.14_2 -rw-r--r-- 1 root root 14975 2020-02-08 21:06 ./CHANGES_AND_HINTS.TXT --rw-r--r-- 1 root root 1350030 2020-11-04 19:37 ./CHECKSUMS.md5 --rw-r--r-- 1 root root 163 2020-11-04 19:37 ./CHECKSUMS.md5.asc +-rw-r--r-- 1 root root 1350290 2020-11-06 21:28 ./CHECKSUMS.md5 +-rw-r--r-- 1 root root 163 2020-11-06 21:28 ./CHECKSUMS.md5.asc -rw-r--r-- 1 root root 17976 1994-06-10 02:28 ./COPYING -rw-r--r-- 1 root root 35147 2007-06-30 04:21 ./COPYING3 -rw-r--r-- 1 root root 19573 2016-06-23 20:08 ./COPYRIGHT.TXT -rw-r--r-- 1 root root 616 2006-10-02 04:37 ./CRYPTO_NOTICE.TXT --rw-r--r-- 1 root root 1018366 2020-11-05 20:40 ./ChangeLog.txt +-rw-r--r-- 1 root root 1018736 2020-11-06 21:25 ./ChangeLog.txt drwxr-xr-x 3 root root 4096 2013-03-20 22:17 ./EFI drwxr-xr-x 2 root root 4096 2020-11-05 20:39 ./EFI/BOOT -rw-r--r-- 1 root root 1417216 2019-07-05 18:54 ./EFI/BOOT/bootx64.efi @@ -25,9 +25,9 @@ drwxr-xr-x 2 root root 4096 2020-11-05 20:39 ./EFI/BOOT -rwxr-xr-x 1 root root 2504 2019-07-05 18:54 ./EFI/BOOT/make-grub.sh -rw-r--r-- 1 root root 10722 2013-09-21 19:02 ./EFI/BOOT/osdetect.cfg -rw-r--r-- 1 root root 1273 2013-08-12 21:08 ./EFI/BOOT/tools.cfg --rw-r--r-- 1 root root 1731047 2020-11-04 19:37 ./FILELIST.TXT +-rw-r--r-- 1 root root 1731356 2020-11-06 21:28 ./FILELIST.TXT -rw-r--r-- 1 root root 1572 2012-08-29 18:27 ./GPG-KEY --rw-r--r-- 1 root root 766991 2020-11-05 20:46 ./PACKAGES.TXT +-rw-r--r-- 1 root root 766991 2020-11-06 21:28 ./PACKAGES.TXT -rw-r--r-- 1 root root 8564 2016-06-28 21:33 ./README.TXT -rw-r--r-- 1 root root 3629 2020-11-05 20:27 ./README.initrd -rw-r--r-- 1 root root 34412 2017-12-01 17:44 ./README_CRYPT.TXT @@ -791,11 +791,11 @@ drwxr-xr-x 2 root root 4096 2012-09-20 18:06 ./patches -rw-r--r-- 1 root root 575 2012-09-20 18:06 ./patches/FILE_LIST -rw-r--r-- 1 root root 14 2012-09-20 18:06 ./patches/MANIFEST.bz2 -rw-r--r-- 1 root root 224 2012-09-20 18:06 ./patches/PACKAGES.TXT -drwxr-xr-x 18 root root 4096 2020-11-05 20:46 ./slackware64 --rw-r--r-- 1 root root 302002 2020-11-05 20:46 ./slackware64/CHECKSUMS.md5 --rw-r--r-- 1 root root 163 2020-11-05 20:46 ./slackware64/CHECKSUMS.md5.asc --rw-r--r-- 1 root root 375825 2020-11-05 20:45 ./slackware64/FILE_LIST --rw-r--r-- 1 root root 3865286 2020-11-05 20:46 ./slackware64/MANIFEST.bz2 +drwxr-xr-x 18 root root 4096 2020-11-06 21:28 ./slackware64 +-rw-r--r-- 1 root root 302002 2020-11-06 21:28 ./slackware64/CHECKSUMS.md5 +-rw-r--r-- 1 root root 163 2020-11-06 21:28 ./slackware64/CHECKSUMS.md5.asc +-rw-r--r-- 1 root root 375825 2020-11-06 21:27 ./slackware64/FILE_LIST +-rw-r--r-- 1 root root 3877602 2020-11-06 21:27 ./slackware64/MANIFEST.bz2 lrwxrwxrwx 1 root root 15 2009-08-23 23:34 ./slackware64/PACKAGES.TXT -> ../PACKAGES.TXT drwxr-xr-x 2 root root 20480 2020-11-05 20:45 ./slackware64/a -rw-r--r-- 1 root root 327 2018-06-24 18:44 ./slackware64/a/aaa_base-14.2-x86_64-5.txt @@ -1433,7 +1433,7 @@ drwxr-xr-x 2 root root 20480 2020-11-04 19:35 ./slackware64/ap -rw-r--r-- 1 root root 506 2020-02-16 18:00 ./slackware64/ap/zsh-5.8-x86_64-1.txt -rw-r--r-- 1 root root 3053312 2020-02-16 18:00 ./slackware64/ap/zsh-5.8-x86_64-1.txz -rw-r--r-- 1 root root 163 2020-02-16 18:00 ./slackware64/ap/zsh-5.8-x86_64-1.txz.asc -drwxr-xr-x 2 root root 16384 2020-11-05 20:45 ./slackware64/d +drwxr-xr-x 2 root root 16384 2020-11-06 19:21 ./slackware64/d -rw-r--r-- 1 root root 360 2020-10-24 22:39 ./slackware64/d/Cython-0.29.21-x86_64-2.txt -rw-r--r-- 1 root root 2380688 2020-10-24 22:39 ./slackware64/d/Cython-0.29.21-x86_64-2.txz -rw-r--r-- 1 root root 163 2020-10-24 22:39 ./slackware64/d/Cython-0.29.21-x86_64-2.txz.asc @@ -1615,9 +1615,9 @@ drwxr-xr-x 2 root root 16384 2020-11-05 20:45 ./slackware64/d -rw-r--r-- 1 root root 385 2020-10-02 19:53 ./slackware64/d/ruby-2.7.2-x86_64-1.txt -rw-r--r-- 1 root root 6105584 2020-10-02 19:53 ./slackware64/d/ruby-2.7.2-x86_64-1.txz -rw-r--r-- 1 root root 163 2020-10-02 19:53 ./slackware64/d/ruby-2.7.2-x86_64-1.txz.asc --rw-r--r-- 1 root root 426 2020-10-16 19:40 ./slackware64/d/rust-1.46.0-x86_64-2.txt --rw-r--r-- 1 root root 57264220 2020-10-16 19:40 ./slackware64/d/rust-1.46.0-x86_64-2.txz --rw-r--r-- 1 root root 163 2020-10-16 19:40 ./slackware64/d/rust-1.46.0-x86_64-2.txz.asc +-rw-r--r-- 1 root root 426 2020-10-13 19:53 ./slackware64/d/rust-1.47.0-x86_64-2.txt +-rw-r--r-- 1 root root 65410800 2020-10-13 19:53 ./slackware64/d/rust-1.47.0-x86_64-2.txz +-rw-r--r-- 1 root root 163 2020-10-13 19:53 ./slackware64/d/rust-1.47.0-x86_64-2.txz.asc -rw-r--r-- 1 root root 456 2020-10-24 22:53 ./slackware64/d/scons-4.0.1-x86_64-2.txt -rw-r--r-- 1 root root 1752500 2020-10-24 22:53 ./slackware64/d/scons-4.0.1-x86_64-2.txz -rw-r--r-- 1 root root 163 2020-10-24 22:53 ./slackware64/d/scons-4.0.1-x86_64-2.txz.asc @@ -4922,7 +4922,7 @@ drwxr-xr-x 2 root root 61440 2020-11-05 20:45 ./slackware64/x -rw-r--r-- 1 root root 213 2018-04-13 06:10 ./slackware64/x/xwud-1.0.5-x86_64-2.txt -rw-r--r-- 1 root root 25288 2018-04-13 06:10 ./slackware64/x/xwud-1.0.5-x86_64-2.txz -rw-r--r-- 1 root root 163 2018-04-13 06:10 ./slackware64/x/xwud-1.0.5-x86_64-2.txz.asc -drwxr-xr-x 2 root root 12288 2020-10-29 21:58 ./slackware64/xap +drwxr-xr-x 2 root root 12288 2020-11-06 21:27 ./slackware64/xap -rw-r--r-- 1 root root 625 2020-04-05 20:17 ./slackware64/xap/MPlayer-20200103-x86_64-2.txt -rw-r--r-- 1 root root 2733304 2020-04-05 20:17 ./slackware64/xap/MPlayer-20200103-x86_64-2.txz -rw-r--r-- 1 root root 163 2020-04-05 20:17 ./slackware64/xap/MPlayer-20200103-x86_64-2.txz.asc @@ -4999,9 +4999,9 @@ drwxr-xr-x 2 root root 12288 2020-10-29 21:58 ./slackware64/xap -rw-r--r-- 1 root root 570 2020-10-19 18:07 ./slackware64/xap/mozilla-firefox-78.4.0esr-x86_64-1.txt -rw-r--r-- 1 root root 55346328 2020-10-19 18:07 ./slackware64/xap/mozilla-firefox-78.4.0esr-x86_64-1.txz -rw-r--r-- 1 root root 163 2020-10-19 18:07 ./slackware64/xap/mozilla-firefox-78.4.0esr-x86_64-1.txz.asc --rw-r--r-- 1 root root 663 2020-10-21 19:20 ./slackware64/xap/mozilla-thunderbird-78.4.0-x86_64-1.txt --rw-r--r-- 1 root root 56893792 2020-10-21 19:20 ./slackware64/xap/mozilla-thunderbird-78.4.0-x86_64-1.txz --rw-r--r-- 1 root root 163 2020-10-21 19:20 ./slackware64/xap/mozilla-thunderbird-78.4.0-x86_64-1.txz.asc +-rw-r--r-- 1 root root 663 2020-11-06 19:18 ./slackware64/xap/mozilla-thunderbird-78.4.1-x86_64-1.txt +-rw-r--r-- 1 root root 56951784 2020-11-06 19:18 ./slackware64/xap/mozilla-thunderbird-78.4.1-x86_64-1.txz +-rw-r--r-- 1 root root 163 2020-11-06 19:18 ./slackware64/xap/mozilla-thunderbird-78.4.1-x86_64-1.txz.asc -rw-r--r-- 1 root root 359 2020-06-23 18:24 ./slackware64/xap/network-manager-applet-1.18.0-x86_64-1.txt -rw-r--r-- 1 root root 905824 2020-06-23 18:24 ./slackware64/xap/network-manager-applet-1.18.0-x86_64-1.txz -rw-r--r-- 1 root root 163 2020-06-23 18:24 ./slackware64/xap/network-manager-applet-1.18.0-x86_64-1.txz.asc @@ -5168,11 +5168,11 @@ drwxr-xr-x 2 root root 4096 2019-02-17 23:51 ./slackware64/y -rw-r--r-- 1 root root 1147 2018-03-01 07:55 ./slackware64/y/maketag -rw-r--r-- 1 root root 1147 2018-03-01 07:55 ./slackware64/y/maketag.ez -rw-r--r-- 1 root root 14 2018-03-01 07:55 ./slackware64/y/tagfile -drwxr-xr-x 19 root root 4096 2020-11-05 20:46 ./source --rw-r--r-- 1 root root 503030 2020-11-05 20:46 ./source/CHECKSUMS.md5 --rw-r--r-- 1 root root 163 2020-11-05 20:46 ./source/CHECKSUMS.md5.asc --rw-r--r-- 1 root root 708529 2020-11-05 20:46 ./source/FILE_LIST --rw-r--r-- 1 root root 21899335 2020-11-05 20:46 ./source/MANIFEST.bz2 +drwxr-xr-x 19 root root 4096 2020-11-06 21:54 ./source +-rw-r--r-- 1 root root 503188 2020-11-06 21:54 ./source/CHECKSUMS.md5 +-rw-r--r-- 1 root root 163 2020-11-06 21:54 ./source/CHECKSUMS.md5.asc +-rw-r--r-- 1 root root 708714 2020-11-06 21:53 ./source/FILE_LIST +-rw-r--r-- 1 root root 21872411 2020-11-06 21:53 ./source/MANIFEST.bz2 -rw-r--r-- 1 root root 1314 2006-10-02 04:40 ./source/README.TXT drwxr-xr-x 119 root root 4096 2020-10-28 19:36 ./source/a -rw-r--r-- 1 root root 1034 2019-05-04 17:56 ./source/a/FTBFSlog @@ -6837,7 +6837,7 @@ drwxr-xr-x 2 root root 4096 2020-02-16 17:57 ./source/ap/zsh -rwxr-xr-x 1 root root 4352 2020-02-16 17:58 ./source/ap/zsh/zsh.SlackBuild -rw-r--r-- 1 root root 23 2018-09-14 17:50 ./source/ap/zsh/zsh.url -rwxr-xr-x 1 root root 11975 2020-02-19 20:57 ./source/buildlist-from-changelog.sh -drwxr-xr-x 60 root root 4096 2020-10-24 18:46 ./source/d +drwxr-xr-x 60 root root 4096 2020-11-06 19:20 ./source/d drwxr-xr-x 2 root root 4096 2020-10-24 18:47 ./source/d/Cython -rw-r--r-- 1 root root 1602538 2020-07-08 21:56 ./source/d/Cython/Cython-0.29.21.tar.lz -rwxr-xr-x 1 root root 3031 2020-10-24 18:47 ./source/d/Cython/Cython.SlackBuild @@ -7192,11 +7192,11 @@ drwxr-xr-x 2 root root 4096 2020-10-02 19:51 ./source/d/ruby -rw-r--r-- 1 root root 11694896 2020-10-02 09:28 ./source/d/ruby/ruby-2.7.2.tar.lz -rwxr-xr-x 1 root root 4807 2019-12-25 18:16 ./source/d/ruby/ruby.SlackBuild -rw-r--r-- 1 root root 837 2019-03-13 16:43 ./source/d/ruby/slack-desc -drwxr-xr-x 2 root root 4096 2020-10-16 20:06 ./source/d/rust +drwxr-xr-x 2 root root 4096 2020-10-08 18:05 ./source/d/rust -rw-r--r-- 1 root root 300 2017-10-16 20:00 ./source/d/rust/link_libffi.diff.gz --rwxr-xr-x 1 root root 9453 2020-10-16 18:54 ./source/d/rust/rust.SlackBuild --rw-r--r-- 1 root root 1457 2020-08-27 17:59 ./source/d/rust/rust.url --rw-r--r-- 1 root root 93811458 2020-08-27 15:49 ./source/d/rust/rustc-1.46.0-src.tar.lz +-rwxr-xr-x 1 root root 9453 2020-10-13 19:10 ./source/d/rust/rust.SlackBuild +-rw-r--r-- 1 root root 1457 2020-10-08 18:04 ./source/d/rust/rust.url +-rw-r--r-- 1 root root 98424391 2020-10-08 13:24 ./source/d/rust/rustc-1.47.0-src.tar.lz -rw-r--r-- 1 root root 910 2018-02-27 06:49 ./source/d/rust/slack-desc drwxr-xr-x 2 root root 4096 2020-10-24 20:35 ./source/d/scons -rw-r--r-- 1 root root 1274421 2020-07-17 02:00 ./source/d/scons/scons-4.0.1.tar.lz @@ -13161,7 +13161,7 @@ drwxr-xr-x 2 root root 4096 2020-03-07 19:29 ./source/xap/libnma -rw-r--r-- 1 root root 1366584 2020-03-06 19:49 ./source/xap/libnma/libnma-1.8.28.tar.xz -rwxr-xr-x 1 root root 3570 2020-03-07 19:31 ./source/xap/libnma/libnma.SlackBuild -rw-r--r-- 1 root root 792 2020-03-07 19:38 ./source/xap/libnma/slack-desc -drwxr-xr-x 3 root root 4096 2020-10-19 17:46 ./source/xap/mozilla-firefox +drwxr-xr-x 3 root root 4096 2020-11-06 21:53 ./source/xap/mozilla-firefox drwxr-xr-x 5 root root 4096 2019-08-27 19:34 ./source/xap/mozilla-firefox/build-deps -rwxr-xr-x 1 root root 1919 2019-07-09 19:35 ./source/xap/mozilla-firefox/build-deps.sh drwxr-xr-x 2 root root 4096 2016-07-03 18:05 ./source/xap/mozilla-firefox/build-deps/autoconf @@ -13182,11 +13182,12 @@ drwxr-xr-x 2 root root 4096 2020-07-03 19:10 ./source/xap/mozilla-firefox/ -rw-r--r-- 1 root root 327 2008-06-17 17:19 ./source/xap/mozilla-firefox/firefox.moz_plugin_path.diff.gz -rw-r--r-- 1 root root 462 2009-07-01 06:05 ./source/xap/mozilla-firefox/mimeTypes.rdf.gz -rw-r--r-- 1 root root 680 2009-07-01 13:28 ./source/xap/mozilla-firefox/mozilla-firefox-mimeTypes-fix.diff.gz --rwxr-xr-x 1 root root 15166 2020-07-27 16:28 ./source/xap/mozilla-firefox/mozilla-firefox.SlackBuild +-rwxr-xr-x 1 root root 15263 2020-11-06 19:25 ./source/xap/mozilla-firefox/mozilla-firefox.SlackBuild -rw-r--r-- 1 root root 2748 2017-12-04 21:30 ./source/xap/mozilla-firefox/mozilla-firefox.desktop +-rw-r--r-- 1 root root 173003 2020-10-20 13:21 ./source/xap/mozilla-firefox/rust_1.47.0.patch.gz -rw-r--r-- 1 root root 1033 2020-07-07 18:08 ./source/xap/mozilla-firefox/slack-desc -rw-r--r-- 1 root root 456 2019-07-08 18:54 ./source/xap/mozilla-firefox/unbreakdocs.diff.gz -drwxr-xr-x 5 root root 4096 2020-10-21 19:24 ./source/xap/mozilla-thunderbird +drwxr-xr-x 5 root root 4096 2020-11-06 19:26 ./source/xap/mozilla-thunderbird drwxr-xr-x 2 root root 4096 2016-07-03 18:05 ./source/xap/mozilla-thunderbird/autoconf -rw-r--r-- 1 root root 5869 2016-07-03 18:04 ./source/xap/mozilla-thunderbird/autoconf/autoconf-2.13-consolidated_fixes-1.patch.gz -rw-r--r-- 1 root root 300116 1999-01-15 21:03 ./source/xap/mozilla-thunderbird/autoconf/autoconf-2.13.tar.xz @@ -13210,12 +13211,13 @@ lrwxrwxrwx 1 root root 7 2019-08-28 16:42 ./source/xap/mozilla-thunderb lrwxrwxrwx 1 root root 7 2019-08-28 16:40 ./source/xap/mozilla-thunderbird/gold/ld -> ld.gold -rwxr-xr-x 1 root root 281 2019-08-28 19:15 ./source/xap/mozilla-thunderbird/gold/ld.gold -rw-r--r-- 1 root root 312 2012-03-28 18:01 ./source/xap/mozilla-thunderbird/mozilla-firefox.xpcom_arm.patch.gz --rwxr-xr-x 1 root root 13036 2020-08-25 16:18 ./source/xap/mozilla-thunderbird/mozilla-thunderbird.SlackBuild +-rwxr-xr-x 1 root root 13133 2020-11-06 18:57 ./source/xap/mozilla-thunderbird/mozilla-thunderbird.SlackBuild -rw-r--r-- 1 root root 3378 2005-03-08 05:13 ./source/xap/mozilla-thunderbird/mozilla-thunderbird.desktop +-rw-r--r-- 1 root root 173003 2020-10-20 13:21 ./source/xap/mozilla-thunderbird/rust_1.47.0.patch.gz -rw-r--r-- 1 root root 1130 2018-02-27 06:47 ./source/xap/mozilla-thunderbird/slack-desc -rw-r--r-- 1 root root 330 2019-08-27 16:35 ./source/xap/mozilla-thunderbird/tb.ui.scrollToClick.diff.gz --rw-r--r-- 1 root root 361176060 2020-10-21 14:38 ./source/xap/mozilla-thunderbird/thunderbird-78.4.0.source.tar.xz --rw-r--r-- 1 root root 833 2020-10-21 14:38 ./source/xap/mozilla-thunderbird/thunderbird-78.4.0.source.tar.xz.asc +-rw-r--r-- 1 root root 352778636 2020-11-06 13:00 ./source/xap/mozilla-thunderbird/thunderbird-78.4.1.source.tar.xz +-rw-r--r-- 1 root root 833 2020-11-06 13:00 ./source/xap/mozilla-thunderbird/thunderbird-78.4.1.source.tar.xz.asc -rw-r--r-- 1 root root 456 2019-07-08 18:54 ./source/xap/mozilla-thunderbird/unbreakdocs.diff.gz drwxr-xr-x 2 root root 4096 2020-06-23 18:18 ./source/xap/network-manager-applet -rw-r--r-- 1 root root 207 2012-08-10 23:59 ./source/xap/network-manager-applet/doinst.sh.gz @@ -13260,7 +13262,7 @@ drwxr-xr-x 2 root root 4096 2020-08-24 19:02 ./source/xap/sane -rwxr-xr-x 1 root root 6139 2020-08-24 19:02 ./source/xap/sane/sane.SlackBuild -rw-r--r-- 1 root root 40 2020-05-17 19:09 ./source/xap/sane/sane.url -rw-r--r-- 1 root root 811 2020-05-17 19:09 ./source/xap/sane/slack-desc -drwxr-xr-x 3 root root 4096 2020-09-22 17:59 ./source/xap/seamonkey +drwxr-xr-x 3 root root 4096 2020-11-06 20:41 ./source/xap/seamonkey drwxr-xr-x 2 root root 4096 2016-07-03 18:05 ./source/xap/seamonkey/autoconf -rw-r--r-- 1 root root 5869 2016-07-03 18:04 ./source/xap/seamonkey/autoconf/autoconf-2.13-consolidated_fixes-1.patch.gz -rw-r--r-- 1 root root 300116 1999-01-15 21:03 ./source/xap/seamonkey/autoconf/autoconf-2.13.tar.xz @@ -13270,7 +13272,7 @@ drwxr-xr-x 2 root root 4096 2016-07-03 18:05 ./source/xap/seamonkey/autoco -rw-r--r-- 1 root root 12796 2006-02-09 03:17 ./source/xap/seamonkey/seamonkey-icon.png -rw-r--r-- 1 root root 185 2006-02-09 03:40 ./source/xap/seamonkey/seamonkey-mail-icon.png -rw-r--r-- 1 root root 207 2006-02-09 03:23 ./source/xap/seamonkey/seamonkey-mail.desktop --rwxr-xr-x 1 root root 8571 2020-09-22 17:59 ./source/xap/seamonkey/seamonkey.SlackBuild +-rwxr-xr-x 1 root root 8565 2020-11-06 20:24 ./source/xap/seamonkey/seamonkey.SlackBuild -rw-r--r-- 1 root root 315 2013-08-13 03:23 ./source/xap/seamonkey/seamonkey.desktop -rw-r--r-- 1 root root 849 2020-03-01 04:26 ./source/xap/seamonkey/slack-desc -rw-r--r-- 1 root root 341 2017-07-21 16:21 ./source/xap/seamonkey/sm.ui.scrollToClick.diff.gz @@ -17077,7 +17079,7 @@ drwxr-xr-x 2 root root 4096 2020-10-31 19:30 ./testing/source/vtown/kde/sr -rw-r--r-- 1 root root 488 2020-10-06 04:30 ./testing/source/vtown/kde/src/applications/kdesdk-kioslaves-20.08.2.tar.xz.sig -rw-r--r-- 1 root root 17700 2020-10-06 04:30 ./testing/source/vtown/kde/src/applications/kdesdk-thumbnailers-20.08.2.tar.xz -rw-r--r-- 1 root root 488 2020-10-06 04:30 ./testing/source/vtown/kde/src/applications/kdesdk-thumbnailers-20.08.2.tar.xz.sig --rw-r--r-- 1 root root 467300 2020-10-06 04:30 ./testing/source/vtown/kde/src/applications/kdf-20.08.2.tar.xz +-rw-r--r-- 1 root root 467300 2020-10-06 04:30 ./testing/source/vtown/kde/src/applications/kdf-20.08.2.tar.xz -rw-r--r-- 1 root root 488 2020-10-06 04:30 ./testing/source/vtown/kde/src/applications/kdf-20.08.2.tar.xz.sig -rw-r--r-- 1 root root 107664 2020-10-06 04:30 ./testing/source/vtown/kde/src/applications/kdialog-20.08.2.tar.xz -rw-r--r-- 1 root root 488 2020-10-06 04:30 ./testing/source/vtown/kde/src/applications/kdialog-20.08.2.tar.xz.sig diff --git a/recompress.sh b/recompress.sh index e896e61a..9657d639 100755 --- a/recompress.sh +++ b/recompress.sh @@ -1097,6 +1097,7 @@ gzip ./source/xap/mozilla-thunderbird/build-deps/autoconf/autoconf-2.13-consolid gzip ./source/xap/mozilla-thunderbird/tb.ui.scrollToClick.diff gzip ./source/xap/mozilla-thunderbird/mozilla-firefox.xpcom_arm.patch gzip ./source/xap/mozilla-thunderbird/unbreakdocs.diff +gzip ./source/xap/mozilla-thunderbird/rust_1.47.0.patch gzip ./source/xap/mozilla-thunderbird/gkrust.a.no.networking.check.diff gzip ./source/xap/xmms/xmms.alsa.default.diff gzip ./source/xap/xmms/xmms.gtk.doublesize.diff @@ -1211,6 +1212,7 @@ gzip ./source/xap/mozilla-firefox/build-deps/autoconf/autoconf-2.13-consolidated gzip ./source/xap/mozilla-firefox/mimeTypes.rdf gzip ./source/xap/mozilla-firefox/mozilla-firefox-mimeTypes-fix.diff gzip ./source/xap/mozilla-firefox/unbreakdocs.diff +gzip ./source/xap/mozilla-firefox/rust_1.47.0.patch gzip ./source/xap/electricsheep/electricsheep.mplayer.diff gzip ./source/installer/sources/initrd/etc/keymaps.tar gzip ./source/x/liberation-fonts-ttf/doinst.sh diff --git a/source/d/rust/rust.SlackBuild b/source/d/rust/rust.SlackBuild index 8f3508a5..6e8e6af1 100755 --- a/source/d/rust/rust.SlackBuild +++ b/source/d/rust/rust.SlackBuild @@ -26,7 +26,7 @@ cd $(dirname $0) ; CWD=$(pwd) PKGNAM=rust SRCNAM="${PKGNAM}c" -VERSION=${VERSION:-1.46.0} +VERSION=${VERSION:-1.47.0} BUILD=${BUILD:-2} # Set this to YES to build with the system LLVM, or NO to use the bundled LLVM. @@ -34,9 +34,9 @@ BUILD=${BUILD:-2} SYSTEM_LLVM=${SYSTEM_LLVM:-YES} # Bootstrap variables (might not be kept updated for latest Rust): -RSTAGE0_VERSION=${RSTAGE0_VERSION:-1.45.2} -RSTAGE0_DIR=${RSTAGE0_DIR:-2020-08-03} -CSTAGE0_VERSION=${CSTAGE0_VERSION:-0.46.1} +RSTAGE0_VERSION=${RSTAGE0_VERSION:-1.46.0} +RSTAGE0_DIR=${RSTAGE0_DIR:-2020-08-27} +CSTAGE0_VERSION=${CSTAGE0_VERSION:-0.47.0} CSTAGE0_DIR=${CSTAGE0_DIR:-$RSTAGE0_DIR} # Automatically determine the architecture we're building on: diff --git a/source/d/rust/rust.url b/source/d/rust/rust.url index e88857ae..c9982da0 100644 --- a/source/d/rust/rust.url +++ b/source/d/rust/rust.url @@ -1,5 +1,5 @@ # Source code (repacked to .tar.lz): -VERSION=1.46.0 +VERSION=1.47.0 rm -f rustc-${VERSION}-src.tar.* lftpget https://static.rust-lang.org/dist/rustc-${VERSION}-src.tar.gz lftpget https://static.rust-lang.org/dist/rustc-${VERSION}-src.tar.gz.asc @@ -21,12 +21,12 @@ fi exit 0 # i686 bootstrap: -lftpget https://static.rust-lang.org/dist/2020-08-03/cargo-0.46.1-i686-unknown-linux-gnu.tar.gz -lftpget https://static.rust-lang.org/dist/2020-08-03/rust-std-1.45.2-i686-unknown-linux-gnu.tar.gz -lftpget https://static.rust-lang.org/dist/2020-08-03/rustc-1.45.2-i686-unknown-linux-gnu.tar.gz +lftpget https://static.rust-lang.org/dist/2020-08-27/cargo-0.47.0-i686-unknown-linux-gnu.tar.gz +lftpget https://static.rust-lang.org/dist/2020-08-27/rust-std-1.46.0-i686-unknown-linux-gnu.tar.gz +lftpget https://static.rust-lang.org/dist/2020-08-27/rustc-1.46.0-i686-unknown-linux-gnu.tar.gz # x86_64 bootstrap: -lftpget https://static.rust-lang.org/dist/2020-08-03/cargo-0.46.1-x86_64-unknown-linux-gnu.tar.gz -lftpget https://static.rust-lang.org/dist/2020-08-03/rust-std-1.45.2-x86_64-unknown-linux-gnu.tar.gz -lftpget https://static.rust-lang.org/dist/2020-08-03/rustc-1.45.2-x86_64-unknown-linux-gnu.tar.gz +lftpget https://static.rust-lang.org/dist/2020-08-27/cargo-0.47.0-x86_64-unknown-linux-gnu.tar.gz +lftpget https://static.rust-lang.org/dist/2020-08-27/rust-std-1.46.0-x86_64-unknown-linux-gnu.tar.gz +lftpget https://static.rust-lang.org/dist/2020-08-27/rustc-1.46.0-x86_64-unknown-linux-gnu.tar.gz diff --git a/source/xap/mozilla-firefox/mozilla-firefox.SlackBuild b/source/xap/mozilla-firefox/mozilla-firefox.SlackBuild index c8e20e3c..6f3dd999 100755 --- a/source/xap/mozilla-firefox/mozilla-firefox.SlackBuild +++ b/source/xap/mozilla-firefox/mozilla-firefox.SlackBuild @@ -198,6 +198,9 @@ zcat $CWD/ff.ui.scrollToClick.diff.gz | patch -p1 --verbose || exit 1 # Fix building with latest Rust: zcat $CWD/unbreakdocs.diff.gz | patch -p1 --verbose || exit 1 +# Fix building with rust-1.47.0: +zcat $CWD/rust_1.47.0.patch.gz | patch -p1 --verbose || exit 1 + # Fetch localization, if requested # https://bugzilla.mozilla.org/show_bug.cgi?id=1256955 if [ ! -z $MOZLOCALIZE ]; then diff --git a/source/xap/mozilla-firefox/rust_1.47.0.patch b/source/xap/mozilla-firefox/rust_1.47.0.patch new file mode 100644 index 00000000..9471b7bd --- /dev/null +++ b/source/xap/mozilla-firefox/rust_1.47.0.patch @@ -0,0 +1,30905 @@ +From 83fc2e3616ef15056be74f056a15e892038809b9 Mon Sep 17 00:00:00 2001 +From: Thomas Deutschmann <whissi@gentoo.org> +Date: Sat, 10 Oct 2020 16:10:20 +0200 +Subject: [PATCH 38/38] bmo#1663715: Update syn and proc-macro2 so that Firefox + can build on Rust nightly again + +Link: https://bugzilla.mozilla.org/show_bug.cgi?id=1663715#c7 +Signed-off-by: Thomas Deutschmann <whissi@gentoo.org> +--- + Cargo.lock | 8 +- + .../rust/lucet-wasi/.cargo-checksum.json | 2 +- + .../rust/packed_simd/.cargo-checksum.json | 2 +- + .../rust/proc-macro2/.cargo-checksum.json | 2 +- + third_party/rust/proc-macro2/Cargo.toml | 15 +- + third_party/rust/proc-macro2/README.md | 2 +- + third_party/rust/proc-macro2/build.rs | 20 + + third_party/rust/proc-macro2/src/detection.rs | 67 + + third_party/rust/proc-macro2/src/fallback.rs | 1010 ++---- + third_party/rust/proc-macro2/src/lib.rs | 225 +- + third_party/rust/proc-macro2/src/marker.rs | 18 + + third_party/rust/proc-macro2/src/parse.rs | 849 +++++ + third_party/rust/proc-macro2/src/strnom.rs | 391 --- + third_party/rust/proc-macro2/src/wrapper.rs | 258 +- + .../rust/proc-macro2/tests/comments.rs | 103 + + third_party/rust/proc-macro2/tests/marker.rs | 33 + + third_party/rust/proc-macro2/tests/test.rs | 240 +- + .../rust/proc-macro2/tests/test_fmt.rs | 26 + + .../spirv-cross-internal/.cargo-checksum.json | 2 +- + third_party/rust/syn/.cargo-checksum.json | 2 +- + third_party/rust/syn/Cargo.toml | 35 +- + third_party/rust/syn/README.md | 16 +- + third_party/rust/syn/benches/file.rs | 7 + + third_party/rust/syn/benches/rust.rs | 45 +- + third_party/rust/syn/build.rs | 38 +- + third_party/rust/syn/src/attr.rs | 126 +- + third_party/rust/syn/src/buffer.rs | 56 +- + third_party/rust/syn/src/custom_keyword.rs | 12 +- + .../rust/syn/src/custom_punctuation.rs | 50 +- + third_party/rust/syn/src/data.rs | 96 +- + third_party/rust/syn/src/derive.rs | 10 +- + third_party/rust/syn/src/discouraged.rs | 27 +- + third_party/rust/syn/src/error.rs | 33 +- + third_party/rust/syn/src/expr.rs | 826 +++-- + third_party/rust/syn/src/ext.rs | 12 +- + third_party/rust/syn/src/file.rs | 4 +- + third_party/rust/syn/src/gen/clone.rs | 2051 ++++++++++++ + third_party/rust/syn/src/gen/debug.rs | 2857 +++++++++++++++++ + third_party/rust/syn/src/gen/eq.rs | 1930 +++++++++++ + third_party/rust/syn/src/gen/fold.rs | 287 +- + third_party/rust/syn/src/gen/hash.rs | 2691 ++++++++++++++++ + third_party/rust/syn/src/gen/visit.rs | 19 +- + third_party/rust/syn/src/gen/visit_mut.rs | 19 +- + third_party/rust/syn/src/generics.rs | 255 +- + third_party/rust/syn/src/item.rs | 1515 +++++---- + third_party/rust/syn/src/keyword.rs | 0 + third_party/rust/syn/src/lib.rs | 109 +- + third_party/rust/syn/src/lifetime.rs | 13 +- + third_party/rust/syn/src/lit.rs | 581 ++-- + third_party/rust/syn/src/mac.rs | 55 +- + third_party/rust/syn/src/macros.rs | 61 +- + third_party/rust/syn/src/op.rs | 6 +- + third_party/rust/syn/src/parse.rs | 211 +- + third_party/rust/syn/src/parse_macro_input.rs | 32 +- + third_party/rust/syn/src/parse_quote.rs | 15 +- + third_party/rust/syn/src/pat.rs | 313 +- + third_party/rust/syn/src/path.rs | 33 +- + third_party/rust/syn/src/punctuated.rs | 123 +- + third_party/rust/syn/src/reserved.rs | 42 + + third_party/rust/syn/src/spanned.rs | 4 +- + third_party/rust/syn/src/stmt.rs | 141 +- + third_party/rust/syn/src/token.rs | 99 +- + third_party/rust/syn/src/tt.rs | 6 +- + third_party/rust/syn/src/ty.rs | 364 ++- + third_party/rust/syn/src/verbatim.rs | 15 + + third_party/rust/syn/src/whitespace.rs | 65 + + third_party/rust/syn/tests/clone.sh | 16 - + third_party/rust/syn/tests/common/eq.rs | 247 +- + third_party/rust/syn/tests/common/mod.rs | 13 + + third_party/rust/syn/tests/common/parse.rs | 24 +- + third_party/rust/syn/tests/debug/gen.rs | 50 +- + third_party/rust/syn/tests/debug/mod.rs | 17 +- + third_party/rust/syn/tests/features/error.rs | 1 - + third_party/rust/syn/tests/features/mod.rs | 22 - + third_party/rust/syn/tests/macros/mod.rs | 8 +- + third_party/rust/syn/tests/repo/mod.rs | 137 +- + third_party/rust/syn/tests/repo/progress.rs | 37 + + third_party/rust/syn/tests/test_asyncness.rs | 38 +- + third_party/rust/syn/tests/test_attribute.rs | 452 +-- + .../rust/syn/tests/test_derive_input.rs | 1321 ++++---- + third_party/rust/syn/tests/test_expr.rs | 314 +- + third_party/rust/syn/tests/test_generics.rs | 371 ++- + third_party/rust/syn/tests/test_grouping.rs | 53 +- + third_party/rust/syn/tests/test_ident.rs | 5 - + third_party/rust/syn/tests/test_item.rs | 45 + + third_party/rust/syn/tests/test_iterators.rs | 7 +- + third_party/rust/syn/tests/test_lit.rs | 75 +- + third_party/rust/syn/tests/test_meta.rs | 498 ++- + .../rust/syn/tests/test_parse_buffer.rs | 41 +- + .../rust/syn/tests/test_parse_stream.rs | 12 + + third_party/rust/syn/tests/test_pat.rs | 27 +- + third_party/rust/syn/tests/test_path.rs | 52 + + third_party/rust/syn/tests/test_precedence.rs | 196 +- + third_party/rust/syn/tests/test_receiver.rs | 127 + + third_party/rust/syn/tests/test_round_trip.rs | 41 +- + third_party/rust/syn/tests/test_shebang.rs | 59 + + .../rust/syn/tests/test_should_parse.rs | 4 - + third_party/rust/syn/tests/test_size.rs | 2 - + third_party/rust/syn/tests/test_stmt.rs | 44 + + .../rust/syn/tests/test_token_trees.rs | 12 +- + third_party/rust/syn/tests/test_ty.rs | 53 + + third_party/rust/syn/tests/test_visibility.rs | 145 + + third_party/rust/syn/tests/zzz_stable.rs | 4 +- + 103 files changed, 17319 insertions(+), 5831 deletions(-) + create mode 100644 third_party/rust/proc-macro2/src/detection.rs + create mode 100644 third_party/rust/proc-macro2/src/marker.rs + create mode 100644 third_party/rust/proc-macro2/src/parse.rs + delete mode 100644 third_party/rust/proc-macro2/src/strnom.rs + create mode 100644 third_party/rust/proc-macro2/tests/comments.rs + create mode 100644 third_party/rust/proc-macro2/tests/test_fmt.rs + create mode 100644 third_party/rust/syn/src/gen/clone.rs + create mode 100644 third_party/rust/syn/src/gen/debug.rs + create mode 100644 third_party/rust/syn/src/gen/eq.rs + create mode 100644 third_party/rust/syn/src/gen/hash.rs + delete mode 100644 third_party/rust/syn/src/keyword.rs + create mode 100644 third_party/rust/syn/src/reserved.rs + create mode 100644 third_party/rust/syn/src/verbatim.rs + create mode 100644 third_party/rust/syn/src/whitespace.rs + delete mode 100755 third_party/rust/syn/tests/clone.sh + delete mode 100644 third_party/rust/syn/tests/features/error.rs + delete mode 100644 third_party/rust/syn/tests/features/mod.rs + create mode 100644 third_party/rust/syn/tests/repo/progress.rs + create mode 100644 third_party/rust/syn/tests/test_item.rs + create mode 100644 third_party/rust/syn/tests/test_parse_stream.rs + create mode 100644 third_party/rust/syn/tests/test_path.rs + create mode 100644 third_party/rust/syn/tests/test_receiver.rs + create mode 100644 third_party/rust/syn/tests/test_shebang.rs + create mode 100644 third_party/rust/syn/tests/test_stmt.rs + create mode 100644 third_party/rust/syn/tests/test_ty.rs + create mode 100644 third_party/rust/syn/tests/test_visibility.rs + +diff --git a/Cargo.lock b/Cargo.lock +index 19117e8368..d5fe0f6457 100644 +--- a/Cargo.lock ++++ b/Cargo.lock +@@ -3717,9 +3717,9 @@ dependencies = [ + + [[package]] + name = "proc-macro2" +-version = "1.0.5" ++version = "1.0.24" + source = "registry+https://github.com/rust-lang/crates.io-index" +-checksum = "90cf5f418035b98e655e9cdb225047638296b862b42411c4e45bb88d700f7fc0" ++checksum = "1e0704ee1a7e00d7bb417d0770ea303c1bccbabf0ef1667dae92b5967f5f8a71" + dependencies = [ + "unicode-xid", + ] +@@ -4647,9 +4647,9 @@ dependencies = [ + + [[package]] + name = "syn" +-version = "1.0.5" ++version = "1.0.40" + source = "registry+https://github.com/rust-lang/crates.io-index" +-checksum = "66850e97125af79138385e9b88339cbcd037e3f28ceab8c5ad98e64f0f1f80bf" ++checksum = "963f7d3cc59b59b9325165add223142bbf1df27655d07789f109896d353d8350" + dependencies = [ + "proc-macro2", + "quote", +diff --git a/third_party/rust/lucet-wasi/.cargo-checksum.json b/third_party/rust/lucet-wasi/.cargo-checksum.json +index 229fc9978c..2c8c0a3c22 100644 +--- a/third_party/rust/lucet-wasi/.cargo-checksum.json ++++ b/third_party/rust/lucet-wasi/.cargo-checksum.json +@@ -1 +1 @@ +-{"files":{"Cargo.toml":"fea1408a1c1b1e84b06044a0b12cb26c8fd3253ca124debb6cd3e4faab48fcbd","LICENSE":"268872b9816f90fd8e85db5a28d33f8150ebb8dd016653fb39ef1f94f2686bc5","LICENSE.cloudabi-utils":"86a34251f0aab76b7dc3daf8d252afbdf481ea94aa5b46d020205178b7e2eac1","LICENSE.wasmtime":"a6c48161a09acc75a0e25503bab66a731eb5fba5392ed4bb4743e4ba5085327a","bindings.json":"fad8362f845e6f7a2af1d7547cee5730894e7b96804f338181fc070ffdcfae1e","build.rs":"593173ad03963afcbef43f1843be6f04cde1df3eae0298ca20bf881019dd350a","examples/.gitignore":"44575cf5b28512d75644bf54a517dcef304ff809fd511747621b4d64f19aac66","examples/Makefile":"d2d2ceeb1bc4435189ea9a2710b6f5f5331ce6aa73ae8a4f4edcca215058a9b4","examples/README.md":"f2a5be6cc88d511c9f4d3bfefdd42dcb2ace813bc23f6a4430b6b543f7373341","examples/hello.c":"9cbc0d3173e02309e15372835fa849d196b2a202d81806fea60378e1878d0c53","examples/pseudoquine.c":"8fd696f8e1b8fb86188564a05f4776875ead31d785a12e3aa4af9d9c1b46d5b5","include/lucet_wasi.h":"497f712c64f753ebdf73ab05b0b340d50094f9c59bb8637ccbf99d895cd20966","src/bindings.rs":"edbeb51d1a93fd31039ee1f1dc7c1b6c0bded2cf5dad10039e8b7da81a4d4a12","src/c_api.rs":"a9c73070a88a461882a28e3e2adfd773f569b964c7ffabde39a3cef907782f71","src/ctx.rs":"578f87c35cce12eaebec95d03e31954c3e6cd0afa214a0fec068f03814eb0cc7","src/fdentry.rs":"94a8480fa587e5586327dfd6b66d8a6a3ef1f8091ba8deb335bf45642f4f98e6","src/host.rs":"6f05f8fea2afed827abfc3c008a5854a8023d91d066580ecfb49e5c8036ef3a3","src/hostcalls/fs.rs":"4726e6f695f7d1d4e371ec52c57f4e36b0ba0d2302fc008b21a301f5fd7a5c97","src/hostcalls/fs_helpers.rs":"474bce0a1f15fa23b0b0b8aa83143d993dd2cbd7cdfc38c118d452d04e80caea","src/hostcalls/misc.rs":"83d087891d92af08cfa2d2e0c5f41cc47cb8219460f6dbcc8666b418dfef206e","src/hostcalls/mod.rs":"4c5d3f65c69503e11e647770879026c37c0e5e01a99b7116c8fb9411b4797187","src/hostcalls/timers.rs":"e65d6a491256b5d6051b6816f6c5049ba3cdc6142651bac81f34d659c1c2a104","src/lib.rs":"5554e1a3f0cd3756173ece6435a0d01b2f520b3401cd5fc33180a04fb9f69bbe","src/memory.rs":"0a09026b15d27f99d74e560cd94795f645cba414a8491bc961987fab9d9da69b","src/wasi_host.rs":"cacbdac28304a837b11e5ad400ae9de3ee79c0284be335e64606ecdfe426ad6e","src/wasm32.rs":"13a5dc6e59784662f1e55eccb457cbbae241a96f70cfa72c41d55858ca05b980","tests/guests/cant_dotdot.c":"609b8cece9443e375a0b38a7e43651b179f66ee9c686edba6696fe1bcd45b111","tests/guests/clock_getres.c":"f5e41c0a2b05a8d7cdb5b4da6c8b6778b858004c1e9d115503c45a1d976be33b","tests/guests/duplicate_import.wat":"4bd8d7a5c1d1597dbe7648300e94e3fab84d7ab068d56cfb656aa1a208026cee","tests/guests/exitcode.c":"b7c5cec3ead0ed82326c568287a1f3398e71ae7e447ce49a3c4c7114c82495af","tests/guests/follow_symlink.c":"de3143ad2bbbfe834c0c32b54c9fcf144ca4eba5cdcf7588929e5f47225ab616","tests/guests/fs.c":"0dca5232ff5da1b7745e3b44bca39333c01a20ba4eae1a6a0a1c492c71ca1efa","tests/guests/getentropy.c":"5d80bcc68dcf3ba91576969055099d61635ae713c057b3cb36afb122a5f26347","tests/guests/getrusage.c":"8114c103b85eb564d9ab43684958bc1939de3794d314b7c121762f3a2f0434a6","tests/guests/gettimeofday.c":"4a57f376b06f4228017b82695448a0bd213fb91455f5301d689cd87fcff01f06","tests/guests/notdir.c":"bd8f8b24360b7cf8d5dced9d9ba4c15843fcbbae89fecc13e3a457c33a275e28","tests/guests/poll.c":"aefaa9b58ce9906dc379e0bd25fa68dfbf8cdffb48cd5ecde1d67708b83b366d","tests/guests/preopen_populates.c":"f186e4eb4aab6a1d9ec7bc5c49eaea6d9d162e0159dfe8f953bb48ade9b58d43","tests/guests/read_file.c":"1aab9393f005f05b69592826d7c4d384a115d5bca42c66f10a901811b4b1dcac","tests/guests/read_file_twice.c":"04a3dad7a43b93e36efd4e2c822c11b3f129429ec799af304d82b358686c578a","tests/guests/stat.c":"02756933ea7d4337b4fa04344b32968851b02f9d0bd5ea1cb0e2f022e8c65ab0","tests/guests/stdin.c":"66efc4b54f68d1138046f1afefae15f7d4555b2904b4a988818e61e67fe8fefb","tests/guests/symlink_escape.c":"686e047b5c986e29c854bcd93996d027dcdc8721219fa9fa532efc98d2798f5c","tests/guests/symlink_loop.c":"2bbddf3a5edfc6e5f3c0fa82cee4ac92b18804810509e263abd17f5240cd37e5","tests/guests/write_file.c":"9e9b14552c2445cfa6d0aa26b334081a59e6e3428dbb17ceca005a9ba59d3220","tests/test_helpers/mod.rs":"bc18194317611fe1be5c439a7a9e0de75399555c3b6de4275af149fb180456c8","tests/tests.rs":"173a7e0f086f6ed46474686cc3413ee68bbd2ff67004f7790e963a1392c7c46e"},"package":null} +\ No newline at end of file ++{"files":{"Cargo.toml":"fea1408a1c1b1e84b06044a0b12cb26c8fd3253ca124debb6cd3e4faab48fcbd","LICENSE":"268872b9816f90fd8e85db5a28d33f8150ebb8dd016653fb39ef1f94f2686bc5","LICENSE.cloudabi-utils":"86a34251f0aab76b7dc3daf8d252afbdf481ea94aa5b46d020205178b7e2eac1","LICENSE.wasmtime":"a6c48161a09acc75a0e25503bab66a731eb5fba5392ed4bb4743e4ba5085327a","bindings.json":"fad8362f845e6f7a2af1d7547cee5730894e7b96804f338181fc070ffdcfae1e","build.rs":"593173ad03963afcbef43f1843be6f04cde1df3eae0298ca20bf881019dd350a","examples/Makefile":"d2d2ceeb1bc4435189ea9a2710b6f5f5331ce6aa73ae8a4f4edcca215058a9b4","examples/README.md":"f2a5be6cc88d511c9f4d3bfefdd42dcb2ace813bc23f6a4430b6b543f7373341","examples/hello.c":"9cbc0d3173e02309e15372835fa849d196b2a202d81806fea60378e1878d0c53","examples/pseudoquine.c":"8fd696f8e1b8fb86188564a05f4776875ead31d785a12e3aa4af9d9c1b46d5b5","include/lucet_wasi.h":"497f712c64f753ebdf73ab05b0b340d50094f9c59bb8637ccbf99d895cd20966","src/bindings.rs":"edbeb51d1a93fd31039ee1f1dc7c1b6c0bded2cf5dad10039e8b7da81a4d4a12","src/c_api.rs":"a9c73070a88a461882a28e3e2adfd773f569b964c7ffabde39a3cef907782f71","src/ctx.rs":"578f87c35cce12eaebec95d03e31954c3e6cd0afa214a0fec068f03814eb0cc7","src/fdentry.rs":"94a8480fa587e5586327dfd6b66d8a6a3ef1f8091ba8deb335bf45642f4f98e6","src/host.rs":"6f05f8fea2afed827abfc3c008a5854a8023d91d066580ecfb49e5c8036ef3a3","src/hostcalls/fs.rs":"4726e6f695f7d1d4e371ec52c57f4e36b0ba0d2302fc008b21a301f5fd7a5c97","src/hostcalls/fs_helpers.rs":"474bce0a1f15fa23b0b0b8aa83143d993dd2cbd7cdfc38c118d452d04e80caea","src/hostcalls/misc.rs":"83d087891d92af08cfa2d2e0c5f41cc47cb8219460f6dbcc8666b418dfef206e","src/hostcalls/mod.rs":"4c5d3f65c69503e11e647770879026c37c0e5e01a99b7116c8fb9411b4797187","src/hostcalls/timers.rs":"e65d6a491256b5d6051b6816f6c5049ba3cdc6142651bac81f34d659c1c2a104","src/lib.rs":"5554e1a3f0cd3756173ece6435a0d01b2f520b3401cd5fc33180a04fb9f69bbe","src/memory.rs":"0a09026b15d27f99d74e560cd94795f645cba414a8491bc961987fab9d9da69b","src/wasi_host.rs":"cacbdac28304a837b11e5ad400ae9de3ee79c0284be335e64606ecdfe426ad6e","src/wasm32.rs":"13a5dc6e59784662f1e55eccb457cbbae241a96f70cfa72c41d55858ca05b980","tests/guests/cant_dotdot.c":"609b8cece9443e375a0b38a7e43651b179f66ee9c686edba6696fe1bcd45b111","tests/guests/clock_getres.c":"f5e41c0a2b05a8d7cdb5b4da6c8b6778b858004c1e9d115503c45a1d976be33b","tests/guests/duplicate_import.wat":"4bd8d7a5c1d1597dbe7648300e94e3fab84d7ab068d56cfb656aa1a208026cee","tests/guests/exitcode.c":"b7c5cec3ead0ed82326c568287a1f3398e71ae7e447ce49a3c4c7114c82495af","tests/guests/follow_symlink.c":"de3143ad2bbbfe834c0c32b54c9fcf144ca4eba5cdcf7588929e5f47225ab616","tests/guests/fs.c":"0dca5232ff5da1b7745e3b44bca39333c01a20ba4eae1a6a0a1c492c71ca1efa","tests/guests/getentropy.c":"5d80bcc68dcf3ba91576969055099d61635ae713c057b3cb36afb122a5f26347","tests/guests/getrusage.c":"8114c103b85eb564d9ab43684958bc1939de3794d314b7c121762f3a2f0434a6","tests/guests/gettimeofday.c":"4a57f376b06f4228017b82695448a0bd213fb91455f5301d689cd87fcff01f06","tests/guests/notdir.c":"bd8f8b24360b7cf8d5dced9d9ba4c15843fcbbae89fecc13e3a457c33a275e28","tests/guests/poll.c":"aefaa9b58ce9906dc379e0bd25fa68dfbf8cdffb48cd5ecde1d67708b83b366d","tests/guests/preopen_populates.c":"f186e4eb4aab6a1d9ec7bc5c49eaea6d9d162e0159dfe8f953bb48ade9b58d43","tests/guests/read_file.c":"1aab9393f005f05b69592826d7c4d384a115d5bca42c66f10a901811b4b1dcac","tests/guests/read_file_twice.c":"04a3dad7a43b93e36efd4e2c822c11b3f129429ec799af304d82b358686c578a","tests/guests/stat.c":"02756933ea7d4337b4fa04344b32968851b02f9d0bd5ea1cb0e2f022e8c65ab0","tests/guests/stdin.c":"66efc4b54f68d1138046f1afefae15f7d4555b2904b4a988818e61e67fe8fefb","tests/guests/symlink_escape.c":"686e047b5c986e29c854bcd93996d027dcdc8721219fa9fa532efc98d2798f5c","tests/guests/symlink_loop.c":"2bbddf3a5edfc6e5f3c0fa82cee4ac92b18804810509e263abd17f5240cd37e5","tests/guests/write_file.c":"9e9b14552c2445cfa6d0aa26b334081a59e6e3428dbb17ceca005a9ba59d3220","tests/test_helpers/mod.rs":"bc18194317611fe1be5c439a7a9e0de75399555c3b6de4275af149fb180456c8","tests/tests.rs":"173a7e0f086f6ed46474686cc3413ee68bbd2ff67004f7790e963a1392c7c46e"},"package":null} +\ No newline at end of file +diff --git a/third_party/rust/packed_simd/.cargo-checksum.json b/third_party/rust/packed_simd/.cargo-checksum.json +index 01afcc1efd..c727a10006 100644 +--- a/third_party/rust/packed_simd/.cargo-checksum.json ++++ b/third_party/rust/packed_simd/.cargo-checksum.json +@@ -1 +1 @@ +-{"files":{".appveyor.yml":"f1ed01850e0d725f9498f52a1a63ddf40702ad6e0bf5b2d7c4c04d76e96794a3",".travis.yml":"e9258d9a54fdaf4cbc12405fe5993ac4497eb2b29021691dbc91b19cb9b52227","Cargo.toml":"089941ba3c89ea111cbea3cc3abdcdcf2b9d0ae0db268d7269ee38226db950e5","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"6485b8ed310d3f0340bf1ad1f47645069ce4069dcc6bb46c7d5c6faf41de1fdb","bors.toml":"dee881dc69b9b7834e4eba5d95c3ed5a416d4628815a167d6a22d4cb4fb064b8","build.rs":"f3baefc5e5bb9b250e762a1466371b922fd7ee4243c217b2d014307603c2f57a","ci/all.sh":"a23d14e10cb26a0eb719e389c30eb955fa53cddcd436890646df09af640bd2eb","ci/android-install-ndk.sh":"0f1746108cc30bf9b9ba45bcde7b19fc1a8bdf5b0258035b4eb8dc69b75efac4","ci/android-install-sdk.sh":"3490432022c5c8f5a115c084f7a9aca1626f96c0c87ffb62019228c4346b47e4","ci/android-sysimage.sh":"ebf4e5daa1f0fe1b2092b79f0f3f161c4c4275cb744e52352c4d81ab451e4c5a","ci/benchmark.sh":"b61d19ef6b90deba8fb79dee74c8b062d94844676293da346da87bb78a9a49a4","ci/deploy_and_run_on_ios_simulator.rs":"ec8ecf82d92072676aa47f0d1a3d021b60a7ae3531153ef12d2ff4541fc294dc","ci/docker/aarch64-linux-android/Dockerfile":"ace2e7d33c87bc0f6d3962a4a3408c04557646f7f51ab99cfbf574906796b016","ci/docker/aarch64-unknown-linux-gnu/Dockerfile":"1ecdac757101d951794fb2ab0deaa278199cf25f2e08a15c7d40ff31a8556184","ci/docker/arm-linux-androideabi/Dockerfile":"370e55d3330a413a3ccf677b3afb3e0ef9018a5fab263faa97ae8ac017fc2286","ci/docker/arm-unknown-linux-gnueabi/Dockerfile":"e25d88f6c0c94aada3d2e3f08243f755feb7e869dc5dc505b3799719cb1af591","ci/docker/arm-unknown-linux-gnueabihf/Dockerfile":"f126f4c7bae8c11ab8b16df06ad997863f0838825a9c08c9899a3eedb6d570bd","ci/docker/armv7-unknown-linux-gnueabihf/Dockerfile":"b647545c158ee480a4c581dbdc1f57833aef056c8d498acc04b573e842bf803c","ci/docker/i586-unknown-linux-gnu/Dockerfile":"0d492759017307ccf74dc2aa4a8cf6623daf3dc728c708dc2b18fa7940800cba","ci/docker/i686-unknown-linux-gnu/Dockerfile":"0d492759017307ccf74dc2aa4a8cf6623daf3dc728c708dc2b18fa7940800cba","ci/docker/mips-unknown-linux-gnu/Dockerfile":"323776469bb7b160385f3621d66e3ee14c75242f8180f916e65af048a29d4ea0","ci/docker/mips64-unknown-linux-gnuabi64/Dockerfile":"c647f6948a9a43b0be695cbed4eac752120d0faf28e5e69c718cb10406921dab","ci/docker/mips64el-unknown-linux-gnuabi64/Dockerfile":"77bfd00cc8639509be381b394f077e39b45a00158ad61b4e1656714c714665d1","ci/docker/mipsel-unknown-linux-musl/Dockerfile":"ec5bea6c98a3b626731fdb95f9ff2d1182639c76e8fb16d3271d0fc884901524","ci/docker/powerpc-unknown-linux-gnu/Dockerfile":"4f2b662de66e83d1354f650b7077692309637f786c2ea5516c31b5c2ee10af2d","ci/docker/powerpc64-unknown-linux-gnu/Dockerfile":"a9595402b772bc365982e22a0096a8988825d90b09b5faa97ab192e76072f71d","ci/docker/powerpc64le-unknown-linux-gnu/Dockerfile":"df3c381c157439695ae8cd10ab71664702c061e3b4ab22906a5ad6c2680acfed","ci/docker/s390x-unknown-linux-gnu/Dockerfile":"93fb44df3d7fd31ead158570667c97b5076a05c3d968af4a84bc13819a8f2db8","ci/docker/sparc64-unknown-linux-gnu/Dockerfile":"da1c39a3ff1fe22e41395fa7c8934e90b4c1788e551b9aec6e38bfd94effc437","ci/docker/thumbv7neon-linux-androideabi/Dockerfile":"c2decd5591bd7a09378901bef629cd944acf052eb55e4f35b79eb9cb4d62246a","ci/docker/thumbv7neon-unknown-linux-gnueabihf/Dockerfile":"75c0c56161c7382b439de74c00de1c0e3dc9d59560cd6720976a751034b78714","ci/docker/wasm32-unknown-unknown/Dockerfile":"3e5f294bc1e004aa599086c2af49d6f3e7459fa250f5fbdd60cf67d53db78758","ci/docker/x86_64-linux-android/Dockerfile":"685040273cf350d5509e580ac451555efa19790c8723ca2af066adadc6880ad2","ci/docker/x86_64-unknown-linux-gnu-emulated/Dockerfile":"44b6203d9290bfdc53d81219f0937e1110847a23dd982ec8c4de388354f01536","ci/docker/x86_64-unknown-linux-gnu/Dockerfile":"d253c86803b22da428fa9cc671a05f18d3318eca7733b8dccb4f7be1ddf524c5","ci/dox.sh":"5b61711be47a4e3dde0ddd15ba73d256ea95fd75af3897732c24db1dc7e66366","ci/linux-s390x.sh":"d6b732d7795b4ba131326aff893bca6228a7d2eb0e9402f135705413dbbe0dce","ci/linux-sparc64.sh":"c92966838b1ab7ad3b7a344833ee726aba6b647cf5952e56f0ad1ba420b13325","ci/lld-shim.rs":"3d7f71ec23a49e2b67f694a0168786f9a954dda15f5a138815d966643fd3fcc3","ci/max_line_width.sh":"0a1518bba4c9ecaa55694cb2e9930d0e19c265baabf73143f17f9cf285aaa5bb","ci/run-docker.sh":"92e036390ad9b0d16f109579df1b5ced2e72e9afea40c7d011400ebd3a2a90de","ci/run.sh":"63259e22a96ba539f53c06b1b39f53e3a78a71171652e7afc170836110ccd913","ci/run_examples.sh":"d1a23c6c35374a0678ba5114b9b8fefd8be0a79e774872a8bf0898d1baca18d0","ci/runtest-android.rs":"145a8e9799a5223975061fe7e586ade5669ee4877a7d7a4cf6b4ab48e8e36c7c","ci/setup_benchmarks.sh":"73fb981a8fdb1dcd54409d3c0fbbfb8f77a3ceabf8626a6b9bf9d21d6bc8ce72","ci/test-runner-linux":"c8aa6025cff5306f4f31d0c61dc5f9d4dd5a1d189ab613ef8d4c367c694d9ccd","contributing.md":"2cc8c9c560ae17867e69b06d09b758dbf7bc39eb774ada50a743724b10acc0a2","perf-guide/.gitignore":"fe82c7da551079d832cf74200b0b359b4df9828cb4a0416fa7384f07a2ae6a13","perf-guide/book.toml":"115a98284126c6b180178b44713314cc494f08a71662ee2ce15cf67f17a51064","perf-guide/src/SUMMARY.md":"3e03bffc991fdc2050f3d51842d72d9d21ea6abab56a3baf3b2d5973a78b89e1","perf-guide/src/ascii.css":"29afb08833b2fe2250f0412e1fa1161a2432a0820a14953c87124407417c741a","perf-guide/src/bound_checks.md":"5e4991ff58a183ef0cd9fdc1feb4cd12d083b44bdf87393bbb0927808ef3ce7d","perf-guide/src/float-math/approx.md":"8c09032fa2d795a0c5db1775826c850d28eb2627846d0965c60ee72de63735ad","perf-guide/src/float-math/fma.md":"311076ba4b741d604a82e74b83a8d7e8c318fcbd7f64c4392d1cf5af95c60243","perf-guide/src/float-math/fp.md":"04153e775ab6e4f0d7837bcc515230d327b04edfa34c84ce9c9e10ebaeef2be8","perf-guide/src/float-math/svml.md":"0798873b8eedaeda5fed62dc91645b57c20775a02d3cd74d8bd06958f1516506","perf-guide/src/introduction.md":"9f5a19e9e6751f25d2daad39891a0cc600974527ec4c8305843f9618910671bd","perf-guide/src/prof/linux.md":"447731eb5de7d69166728fdbc5ecb0c0c9db678ea493b45a592d67dd002184c0","perf-guide/src/prof/mca.md":"f56d54f3d20e7aa4d32052186e8237b03d65971eb5d112802b442570ff11d344","perf-guide/src/prof/profiling.md":"8a650c0fd6ede0964789bb6577557eeef1d8226a896788602ce61528e260e43c","perf-guide/src/target-feature/attribute.md":"615f88dca0a707b6c416fa605435dd6e1fb5361cc639429cbf68cd87624bd78b","perf-guide/src/target-feature/features.md":"17077760ff24c006b606dd21889c53d87228f4311f3ba3a574f9afdeacd86165","perf-guide/src/target-feature/inlining.md":"7ed1d7068d8173a00d84c16cfe5871cd68b9f04f8d0cca2d01ebc84957ebf2f6","perf-guide/src/target-feature/practice.md":"c4b371842e0086df178488fec97f20def8f0c62ee588bcd25fd948b9b1fa227e","perf-guide/src/target-feature/runtime.md":"835425f5ee597fb3e51d36e725a81ebee29f4561231d19563cd4da81dbb1cfcb","perf-guide/src/target-feature/rustflags.md":"ab49712e9293a65d74d540ba4784fcb57ff1119ec05a575d895c071f1a620f64","perf-guide/src/vert-hor-ops.md":"c6211c0ee91e60552ec592d89d9d957eedc21dee3cbd89e1ad6765ea06a27471","readme.md":"585a8f0e16877fb9abb00cd17a175fcb9d7857840c6c61209f1827ffab095070","rustfmt.toml":"de6101d0670bad65fb3b337d56957d2a024e017e5ab146ec784d77312daaf8ff","src/api.rs":"331a3a4abb19cee2df5f2df4ad7c3e88b45e62cf23fdacfc9bbaa633dc5cf788","src/api/bit_manip.rs":"e68290ee679cc5abc9c73afbe635c1035f8cbfe849e5c751a1680e459244c39e","src/api/cast.rs":"03b94a3d316ac7b7be7068810044911e965e889a0ace7bae762749ca74a92747","src/api/cast/macros.rs":"b0a14d0c83ad2ebb7a275180f6d9e3f2bc312ba57a7d3d6c39fad4e0f20f9408","src/api/cast/v128.rs":"63e28c6a3edf1a7a635f51b8d3c6adbb1d46f884d92a196b3d4a6e743d809416","src/api/cast/v16.rs":"2a584eeb57fd47baad6f3533764301b04aaaac23702b7a8db12598ac02899262","src/api/cast/v256.rs":"b91c15ed8d1536ecd97b4eb79ff9d5aba0552cd9b6f0ea6435b05f2273e23b3a","src/api/cast/v32.rs":"62ec89fcce7fa7f28497ee5770adc8f81d2d3a6b2925b02f7dc06504c40e8f38","src/api/cast/v512.rs":"d855cb943ae7106e9599ef38e30a3afb1c6bd5433178baca54cb128fd9a7d143","src/api/cast/v64.rs":"fe0f7dfaf4fc0c0c1a78c96fcfcdfdc2a1e2845843b11aa797a0c6fb52a8f774","src/api/cmp.rs":"357c3a2a09c6d4611c32dd7fa95be2fae933d513e229026ec9b44451a77b884e","src/api/cmp/eq.rs":"60f70f355bae4cb5b17db53204cacc3890f70670611c17df638d4c04f7cc8075","src/api/cmp/ord.rs":"589f7234761c294fa5df8f525bc4acd5a47cdb602207d524a0d4e19804cd9695","src/api/cmp/partial_eq.rs":"3ed23d2a930b0f9750c3a5309da766b03dc4f9c4d375b42ad3c50fe732693d15","src/api/cmp/partial_ord.rs":"e16b11805c94048acd058c93994b5bc74bb187f8d7e3b86a87df60e1601467f9","src/api/cmp/vertical.rs":"de3d62f38eba817299aa16f1e1939954c9a447e316509397465c2830852ba053","src/api/default.rs":"b61f92fc0e33a2633b3375eb405beba480da071cde03df4d437d8a6058afcd97","src/api/fmt.rs":"67fb804bb86b6cd77cf8cd492b5733ce437071b66fe3297278b8a6552c325dda","src/api/fmt/binary.rs":"35cb5c266197d6224d598fb3d286e5fe48ef0c01ed356c2ff6fe9ba946f96a92","src/api/fmt/debug.rs":"aa18eea443bf353fea3db8b1a025132bbcaf91e747ecfa43b8d9fce9af395a0c","src/api/fmt/lower_hex.rs":"69d5be366631af309f214e8031c8c20267fcc27a695eac6f45c6bc1df72a67e6","src/api/fmt/octal.rs":"9eb11ba3d990213f3c7f1ec25edba7ce997cb1320e16d308c83498ba6b9bfbd9","src/api/fmt/upper_hex.rs":"a4637d085b7bb20e759ce58e08435b510a563ba3dd468af2b03560fdc5511562","src/api/from.rs":"2e599d8329cb05eaf06224cc441355c4b7b51254fc19256619333be8c149d444","src/api/from/from_array.rs":"4151593c7bba7455821fffa5b59867005a77c95d32f1f0cc3fd87294000157d9","src/api/from/from_vector.rs":"9764371aa9e6005aace74dea14f59e5611a095b7cf42707940924749282c52f0","src/api/hash.rs":"562cfa3f1d8eb9a733c035a3665a599c2f1e341ee820d8fbdd102a4398a441bc","src/api/into_bits.rs":"82297f0697d67b5a015e904e7e6e7b2a7066ba825bc54b94b4ff3e22d7a1eefb","src/api/into_bits/arch_specific.rs":"1f925390b0ce7132587d95f2419c6e2ad3e1a9d17eb1d9c120a1c1c4bdf4277e","src/api/into_bits/macros.rs":"d762406de25aedff88d460dec7a80dc8e825a2a419d53218ce007efa6a1d3e04","src/api/into_bits/v128.rs":"ecdc5893664c71d7ab1ff3697c3fbe490d20d8748b9b76881d05e7625e40d74c","src/api/into_bits/v16.rs":"5459ec7dad1ad7bd30dc7e48374580b993abf23701d9c3cb22203fa0a9aabb6d","src/api/into_bits/v256.rs":"90ea351da0380ead1bf0f63b620afd40d01d638d09f7e7be31840bd2c1d9c663","src/api/into_bits/v32.rs":"ee1dc5a430050e16f51154b5fe85b1536f5feddf2ea23dd1d3859b67c4afc6fc","src/api/into_bits/v512.rs":"f72098ed1c9a23944f3d01abaf5e0f2d0e81d35a06fdadd2183e896d41b59867","src/api/into_bits/v64.rs":"6394462facdfe7827349c742b7801f1291e75a720dfb8c0b52100df46f371c98","src/api/math.rs":"8b2a2fc651917a850539f993aa0b9e5bf4da67b11685285b8de8cdca311719ec","src/api/math/float.rs":"61d2794d68262a1090ae473bd30793b5f65cf732f32a6694a3af2ce5d9225616","src/api/math/float/abs.rs":"5b6b2701e2e11135b7ce58a05052ea8120e10e4702c95d046b9d21b827b26bf8","src/api/math/float/consts.rs":"78acba000d3fa527111300b6327c1932de9c4c1e02d4174e1a5615c01463d38c","src/api/math/float/cos.rs":"4c2dd7173728ef189314f1576c9486e03be21b7da98843b2f9011282a7979e31","src/api/math/float/exp.rs":"7c6d5f1e304f498a01cfa23b92380c815d7da0ad94eae3483783bc377d287eef","src/api/math/float/ln.rs":"54c7583f3df793b39ff57534fade27b41bb992439e5dc178252f5ca3190a3e54","src/api/math/float/mul_add.rs":"62cac77660d20159276d4c9ef066eb90c81cbddb808e8e157182c607625ad2eb","src/api/math/float/mul_adde.rs":"bae056ee9f3a70df39ec3c3b2f6437c65303888a7b843ef1a5bcf1f5aca0e602","src/api/math/float/powf.rs":"9ddb938984b36d39d82a82f862f80df8f7fb013f1d222d45698d41d88472f568","src/api/math/float/recpre.rs":"589225794ff1dbf31158dff660e6d4509ecc8befbb57c633900dea5ac0b840d6","src/api/math/float/rsqrte.rs":"a32abdcc318d7ccc8448231f54d75b884b7cbeb03a7d595713ab6243036f4dbf","src/api/math/float/sin.rs":"cbd3622b7df74f19691743001c8cf747a201f8977ad90542fee915f37dcd1e49","src/api/math/float/sqrt.rs":"0c66d5d63fb08e4d99c6b82a8828e41173aff1ac9fa1a2764a11fac217ccf2ac","src/api/math/float/sqrte.rs":"731e1c9f321b662accdd27dacb3aac2e8043b7aecb2f2161dde733bd9f025362","src/api/minimal.rs":"1f22bcc528555444e76de569ec0ae2029b9ae9d04805efeafa93369c8098036b","src/api/minimal/iuf.rs":"c501a6696950cf5e521765f178de548af64fdfb6e10d026616d09fab93ca2d17","src/api/minimal/mask.rs":"42e415f536c5193d0218f5a754b34b87fd7c971bff068009f958712166ff056d","src/api/minimal/ptr.rs":"a9ee482d1dd1c956fb8f3f179e6e620b1de4e9d713961461d4c6923a4ef2e67c","src/api/ops.rs":"3e273b277a0f3019d42c3c59ca94a5afd4885d5ae6d2182e5089bbeec9de42ee","src/api/ops/scalar_arithmetic.rs":"d2d5ad897a59dd0787544f927e0e7ca4072c3e58b0f4a2324083312b0d5a21d7","src/api/ops/scalar_bitwise.rs":"482204e459ca6be79568e1c9f70adbe2d2151412ddf122fb2161be8ebb51c40c","src/api/ops/scalar_mask_bitwise.rs":"c250f52042e37b22d57256c80d4604104cfd2fbe2a2e127c676267270ca5d350","src/api/ops/scalar_shifts.rs":"987f8fdebeedc16e3d77c1b732e7826ef70633c541d16dfa290845d5c6289150","src/api/ops/vector_arithmetic.rs":"ddca15d09ddeef502c2ed66117a62300ca65d87e959e8b622d767bdf1c307910","src/api/ops/vector_bitwise.rs":"b3968f7005b649edcc22a54e2379b14d5ee19045f2e784029805781ae043b5ee","src/api/ops/vector_float_min_max.rs":"f5155dce75219f4ba11275b1f295d2fdcddd49d174a6f1fb2ace7ea42813ce41","src/api/ops/vector_int_min_max.rs":"a378789c6ff9b32a51fbd0a97ffd36ed102cd1fe6a067d2b02017c1df342def6","src/api/ops/vector_mask_bitwise.rs":"5052d18517d765415d40327e6e8e55a312daaca0a5e2aec959bfa54b1675f9c8","src/api/ops/vector_neg.rs":"5c62f6b0221983cdbd23cd0a3af3672e6ba1255f0dfe8b19aae6fbd6503e231b","src/api/ops/vector_rotates.rs":"03cbe8a400fd7c688e4ee771a990a6754f2031b1a59b19ae81158b21471167e5","src/api/ops/vector_shifts.rs":"9bf69d0087268f61009e39aea52e03a90f378910206b6a28e8393178b6a5d0e0","src/api/ptr.rs":"8a793251bed6130dcfb2f1519ceaa18b751bbb15875928d0fb6deb5a5e07523a","src/api/ptr/gather_scatter.rs":"9ddd960365e050674b25b2fd3116e24d94669b4375d74e71c03e3f1469576066","src/api/reductions.rs":"ae5baca81352ecd44526d6c30c0a1feeda475ec73ddd3c3ec6b14e944e5448ee","src/api/reductions/bitwise.rs":"8bf910ae226188bd15fc7e125f058cd2566b6186fcd0cd8fd020f352c39ce139","src/api/reductions/float_arithmetic.rs":"e58c8c87806a95df2b2b5b48ac5991036df024096d9d7c171a480fe9282896a4","src/api/reductions/integer_arithmetic.rs":"47471da1c5f859489680bb5d34ced3d3aa20081c16053a3af121a4496fcb57bf","src/api/reductions/mask.rs":"db83327a950e33a317f37fd33ca4e20c347fb415975ec024f3e23da8509425af","src/api/reductions/min_max.rs":"f27be3aa28e1c1f46de7890198db6e12f00c207085e89ef2de7e57ee443cdb98","src/api/select.rs":"a98e2ccf9fc6bdeed32d337c8675bc96c2fbe2cc34fbf149ad6047fb8e749774","src/api/shuffle.rs":"da58200790868c09659819322a489929a5b6e56c596ed07e6a44293ea02e7d09","src/api/shuffle1_dyn.rs":"bfea5a91905b31444e9ef7ca6eddb7a9606b7e22d3f71bb842eb2795a0346620","src/api/slice.rs":"ee87484e8af329547b9a5d4f2a69e8bed6ea10bbd96270d706083843d4eea2ac","src/api/slice/from_slice.rs":"4d4fe8a329c885fcb4fbcbedf99efb15a95296fe6b3f595056cc37037450d5ac","src/api/slice/write_to_slice.rs":"f5b23b2c4b91cfb26b713a9013a6c0da7f45eaefb79ba06dcbc27f3f23bda679","src/api/swap_bytes.rs":"4a6792a2e49a77475e1b237592b4b2804dbddb79c474331acd0dd71b36934259","src/codegen.rs":"c6eebc3d3665420aa6a2f317977e3c41a4f43e0550ac630cdbe8e4bbed5e2031","src/codegen/bit_manip.rs":"5559e095105a80003e0de35af1d19b0c65c9ab04eb743c7e01c5442d882eb34e","src/codegen/llvm.rs":"d1299c189abb17a6133f047574cffc7a6db4c1be37cb7d4785491cb5e8f8cf54","src/codegen/math.rs":"35f96e37a78fcf0cdb02146b7f27a45108fe06a37fc2a54d8851ce131a326178","src/codegen/math/float.rs":"dd86c0449e576c83b719700962ac017c332987fac08d91f2b7a2b1b883598170","src/codegen/math/float/abs.rs":"f56e2b4b8055ea861c1f5cbc6b6e1d8e7e5af163b62c13574ddee4e09513bfbc","src/codegen/math/float/cos.rs":"ef3b511a24d23045b310315e80348a9b7fedb576fc2de52d74290616a0abeb2a","src/codegen/math/float/cos_pi.rs":"4e7631a5d73dac21531e09ef1802d1180f8997509c2c8fa9f67f322194263a97","src/codegen/math/float/exp.rs":"61b691598c41b5622f24e4320c1bdd08701e612a516438bdddcc728fc3405c8c","src/codegen/math/float/ln.rs":"46b718b1ba8c9d99e1ad40f53d20dfde08a3063ca7bd2a9fdd6698e060da687e","src/codegen/math/float/macros.rs":"dd42135fff13f9aca4fd3a1a4e14c7e6c31aadc6d817d63b0d2fb9e62e062744","src/codegen/math/float/mul_add.rs":"a37bf764345d4b1714f97e83897b7cf0855fc2811704bcbc0012db91825339e1","src/codegen/math/float/mul_adde.rs":"c75702bfcb361de45964a93caf959a695ef2376bd069227600b8c6872665c755","src/codegen/math/float/powf.rs":"642346e982bc4c39203de0864d2149c4179cd7b21cf67a2951687932b4675872","src/codegen/math/float/sin.rs":"9d68164c90cdca6a85155040cdac42e27342ebe0b925273ef1593df721af4258","src/codegen/math/float/sin_cos_pi.rs":"9be02ad48585a1e8d99129382fbffbaed47852f15459256a708850b6b7a75405","src/codegen/math/float/sin_pi.rs":"9890347905b4d4a3c7341c3eb06406e46e60582bcf6960688bd727e5dadc6c57","src/codegen/math/float/sqrt.rs":"e3c60dcfb0c6d2fc62adabcc931b2d4040b83cab294dea36443fb4b89eb79e34","src/codegen/math/float/sqrte.rs":"f0f4ef9eb475ae41bcc7ec6a95ad744ba6b36925faa8b2c2814004396d196b63","src/codegen/pointer_sized_int.rs":"a70697169c28218b56fd2e8d5353f2e00671d1150d0c8cef77d613bdfacd84cb","src/codegen/reductions.rs":"645e2514746d01387ddd07f0aa4ffd8430cc9ab428d4fb13773ea319fa25dd95","src/codegen/reductions/mask.rs":"8f1afe6aabf096a3278e1fc3a30f736e04aa8b9ce96373cee22162d18cfe2702","src/codegen/reductions/mask/aarch64.rs":"cba6e17603d39795dcfe8339b6b7d8714c3e162a1f0a635979f037aa24fe4206","src/codegen/reductions/mask/arm.rs":"9447904818aa2c7c25d0963eead452a639a11ca7dbd6d21eedbfcaade07a0f33","src/codegen/reductions/mask/fallback.rs":"7a0ef9f7fd03ae318b495b95e121350cd61caffc5cc6ee17fabf130d5d933453","src/codegen/reductions/mask/fallback_impl.rs":"76547f396e55ef403327c77c314cf8db8c7a5c9b9819bfb925abeacf130249e5","src/codegen/reductions/mask/x86.rs":"14bd2c482071f2355beebcf7b7ecf950ff2dfcdb08c3ca50993092434a9de717","src/codegen/reductions/mask/x86/avx.rs":"b4913d87844c522903641cbbf10db4551addb1ce5e9e78278e21612fa65c733b","src/codegen/reductions/mask/x86/avx2.rs":"677aed3f056285285daa3adff8bc65e739630b4424defa6d9665e160f027507e","src/codegen/reductions/mask/x86/sse.rs":"226610b4ff88c676d5187114dd57b4a8800de6ce40884675e9198445b1ed0306","src/codegen/reductions/mask/x86/sse2.rs":"bc38e6c31cb4b3d62147eba6cac264e519e2a48e0f7ce9010cfa9ef0cf0ec9fd","src/codegen/shuffle.rs":"0abca97e92cdce49a58a39cc447eb09dc7d7715ef256c8dbd2181a186e61bb64","src/codegen/shuffle1_dyn.rs":"04523e9338133bdedb012dd076c2c564b79ce5593b0fc56d0fb6910e04190a81","src/codegen/swap_bytes.rs":"1d6cdc716eadddc92b4fd506b2445a821caa8dc00860447de09d7ebd69c2087f","src/codegen/v128.rs":"94226b31ec403d18d9d2fe06713f147c9c79e9b5f9105089088266313f843185","src/codegen/v16.rs":"ddec4ffb66b6f7aaffb9a1780c5ddba82557abd74f45073d335047e04cf74924","src/codegen/v256.rs":"6b63917f0444118d6b1595bff2045e59b97c4d24012bd575f69f1f0efc5a0241","src/codegen/v32.rs":"3477b3c5540aed86e61e2f5807dd31db947413cec9181c587d93ed6ec74f0eba","src/codegen/v512.rs":"5854f99d3aabc4cd42b28a20d9ce447756dc2ba024a409a69b6a8ae1f1842fc5","src/codegen/v64.rs":"e9e89caebfe63d10c0cbca61e4dfdba3b7e02ee0989170f80beed23237ddd950","src/codegen/vPtr.rs":"96d609a9eece4dcbbcc01ba0b8744d7f5958be12774176a2945bc676f4e6b5cb","src/codegen/vSize.rs":"eeee9858749aa82142b27bc120d1989bb74a6b82e1e4efbbeaccc9634dc9acfc","src/lib.rs":"1b5d419ff05ee0370d671810423ccc254708cc8d415c1dbac2a7a36be4bf63a8","src/masks.rs":"870f429967b2d7d5133f4d28d6c753fc5cef0570b27b29d4e966a066d22d2d0e","src/sealed.rs":"ff7f0324276408ae8249941cfa32c90b8835a54d750896b683efea857af19db2","src/testing.rs":"1d3a7862ef625e235a5734ad7204e68d350f902c0695182b1f08a0552432416e","src/testing/macros.rs":"6378856d7a40ba5ec5c7c0dad6327d79f0c77266921c24296d10aed6c68e9b98","src/testing/utils.rs":"d6fd5a5017f1f85d9d99585754f8f6ad06fc3d683b34083543e67a7cc6c1772c","src/v128.rs":"18fe263c4aa28cd06461c7070b0269f69f4a2e75749b8f142a83dfdfe4d22bf5","src/v16.rs":"e5c663c9fb3547eaeac78a5f7db9969f4d8b5ec96112bf2954602fff11f0aebd","src/v256.rs":"68732cd688ad12a56d8b4f8ddf279f77bdfe1be2943c7dc0c1b4f1a76798aa0f","src/v32.rs":"785b22a1ccb4a41bb53dfeb0670f624c0ce42e6cdf62d1747e3283777a1c70bd","src/v512.rs":"d1337bfe07f06a8f37f8e8fa7d4315b9307476ee435ad80dd5269eaed564fbfa","src/v64.rs":"3077468d65125b8f085e9454c8b2463a4d5225697464ba6a1300f8799528fd4b","src/vPtr.rs":"c9a53f41f466e17b6648a4ce390fd8f4d3a848d440eb8a9a803a11608d76eb05","src/vSize.rs":"5c46d3e8c3ee5863d9b6e37e681f871386e0efc254d6d84ba711edb529ce7b3c","tests/endianness.rs":"541a144be017e3dd7da7c8ea49d907dc02538245e8c5f3deb5bd43da92c929e1"},"package":null} +\ No newline at end of file ++{"files":{".appveyor.yml":"f1ed01850e0d725f9498f52a1a63ddf40702ad6e0bf5b2d7c4c04d76e96794a3",".travis.yml":"e9258d9a54fdaf4cbc12405fe5993ac4497eb2b29021691dbc91b19cb9b52227","Cargo.toml":"089941ba3c89ea111cbea3cc3abdcdcf2b9d0ae0db268d7269ee38226db950e5","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"6485b8ed310d3f0340bf1ad1f47645069ce4069dcc6bb46c7d5c6faf41de1fdb","bors.toml":"dee881dc69b9b7834e4eba5d95c3ed5a416d4628815a167d6a22d4cb4fb064b8","build.rs":"f3baefc5e5bb9b250e762a1466371b922fd7ee4243c217b2d014307603c2f57a","ci/all.sh":"a23d14e10cb26a0eb719e389c30eb955fa53cddcd436890646df09af640bd2eb","ci/android-install-ndk.sh":"0f1746108cc30bf9b9ba45bcde7b19fc1a8bdf5b0258035b4eb8dc69b75efac4","ci/android-install-sdk.sh":"3490432022c5c8f5a115c084f7a9aca1626f96c0c87ffb62019228c4346b47e4","ci/android-sysimage.sh":"ebf4e5daa1f0fe1b2092b79f0f3f161c4c4275cb744e52352c4d81ab451e4c5a","ci/benchmark.sh":"b61d19ef6b90deba8fb79dee74c8b062d94844676293da346da87bb78a9a49a4","ci/deploy_and_run_on_ios_simulator.rs":"ec8ecf82d92072676aa47f0d1a3d021b60a7ae3531153ef12d2ff4541fc294dc","ci/docker/aarch64-linux-android/Dockerfile":"ace2e7d33c87bc0f6d3962a4a3408c04557646f7f51ab99cfbf574906796b016","ci/docker/aarch64-unknown-linux-gnu/Dockerfile":"1ecdac757101d951794fb2ab0deaa278199cf25f2e08a15c7d40ff31a8556184","ci/docker/arm-linux-androideabi/Dockerfile":"370e55d3330a413a3ccf677b3afb3e0ef9018a5fab263faa97ae8ac017fc2286","ci/docker/arm-unknown-linux-gnueabi/Dockerfile":"e25d88f6c0c94aada3d2e3f08243f755feb7e869dc5dc505b3799719cb1af591","ci/docker/arm-unknown-linux-gnueabihf/Dockerfile":"f126f4c7bae8c11ab8b16df06ad997863f0838825a9c08c9899a3eedb6d570bd","ci/docker/armv7-unknown-linux-gnueabihf/Dockerfile":"b647545c158ee480a4c581dbdc1f57833aef056c8d498acc04b573e842bf803c","ci/docker/i586-unknown-linux-gnu/Dockerfile":"0d492759017307ccf74dc2aa4a8cf6623daf3dc728c708dc2b18fa7940800cba","ci/docker/i686-unknown-linux-gnu/Dockerfile":"0d492759017307ccf74dc2aa4a8cf6623daf3dc728c708dc2b18fa7940800cba","ci/docker/mips-unknown-linux-gnu/Dockerfile":"323776469bb7b160385f3621d66e3ee14c75242f8180f916e65af048a29d4ea0","ci/docker/mips64-unknown-linux-gnuabi64/Dockerfile":"c647f6948a9a43b0be695cbed4eac752120d0faf28e5e69c718cb10406921dab","ci/docker/mips64el-unknown-linux-gnuabi64/Dockerfile":"77bfd00cc8639509be381b394f077e39b45a00158ad61b4e1656714c714665d1","ci/docker/mipsel-unknown-linux-musl/Dockerfile":"ec5bea6c98a3b626731fdb95f9ff2d1182639c76e8fb16d3271d0fc884901524","ci/docker/powerpc-unknown-linux-gnu/Dockerfile":"4f2b662de66e83d1354f650b7077692309637f786c2ea5516c31b5c2ee10af2d","ci/docker/powerpc64-unknown-linux-gnu/Dockerfile":"a9595402b772bc365982e22a0096a8988825d90b09b5faa97ab192e76072f71d","ci/docker/powerpc64le-unknown-linux-gnu/Dockerfile":"df3c381c157439695ae8cd10ab71664702c061e3b4ab22906a5ad6c2680acfed","ci/docker/s390x-unknown-linux-gnu/Dockerfile":"93fb44df3d7fd31ead158570667c97b5076a05c3d968af4a84bc13819a8f2db8","ci/docker/sparc64-unknown-linux-gnu/Dockerfile":"da1c39a3ff1fe22e41395fa7c8934e90b4c1788e551b9aec6e38bfd94effc437","ci/docker/thumbv7neon-linux-androideabi/Dockerfile":"c2decd5591bd7a09378901bef629cd944acf052eb55e4f35b79eb9cb4d62246a","ci/docker/thumbv7neon-unknown-linux-gnueabihf/Dockerfile":"75c0c56161c7382b439de74c00de1c0e3dc9d59560cd6720976a751034b78714","ci/docker/wasm32-unknown-unknown/Dockerfile":"3e5f294bc1e004aa599086c2af49d6f3e7459fa250f5fbdd60cf67d53db78758","ci/docker/x86_64-linux-android/Dockerfile":"685040273cf350d5509e580ac451555efa19790c8723ca2af066adadc6880ad2","ci/docker/x86_64-unknown-linux-gnu-emulated/Dockerfile":"44b6203d9290bfdc53d81219f0937e1110847a23dd982ec8c4de388354f01536","ci/docker/x86_64-unknown-linux-gnu/Dockerfile":"d253c86803b22da428fa9cc671a05f18d3318eca7733b8dccb4f7be1ddf524c5","ci/dox.sh":"5b61711be47a4e3dde0ddd15ba73d256ea95fd75af3897732c24db1dc7e66366","ci/linux-s390x.sh":"d6b732d7795b4ba131326aff893bca6228a7d2eb0e9402f135705413dbbe0dce","ci/linux-sparc64.sh":"c92966838b1ab7ad3b7a344833ee726aba6b647cf5952e56f0ad1ba420b13325","ci/lld-shim.rs":"3d7f71ec23a49e2b67f694a0168786f9a954dda15f5a138815d966643fd3fcc3","ci/max_line_width.sh":"0a1518bba4c9ecaa55694cb2e9930d0e19c265baabf73143f17f9cf285aaa5bb","ci/run-docker.sh":"92e036390ad9b0d16f109579df1b5ced2e72e9afea40c7d011400ebd3a2a90de","ci/run.sh":"63259e22a96ba539f53c06b1b39f53e3a78a71171652e7afc170836110ccd913","ci/run_examples.sh":"d1a23c6c35374a0678ba5114b9b8fefd8be0a79e774872a8bf0898d1baca18d0","ci/runtest-android.rs":"145a8e9799a5223975061fe7e586ade5669ee4877a7d7a4cf6b4ab48e8e36c7c","ci/setup_benchmarks.sh":"73fb981a8fdb1dcd54409d3c0fbbfb8f77a3ceabf8626a6b9bf9d21d6bc8ce72","ci/test-runner-linux":"c8aa6025cff5306f4f31d0c61dc5f9d4dd5a1d189ab613ef8d4c367c694d9ccd","contributing.md":"2cc8c9c560ae17867e69b06d09b758dbf7bc39eb774ada50a743724b10acc0a2","perf-guide/book.toml":"115a98284126c6b180178b44713314cc494f08a71662ee2ce15cf67f17a51064","perf-guide/src/SUMMARY.md":"3e03bffc991fdc2050f3d51842d72d9d21ea6abab56a3baf3b2d5973a78b89e1","perf-guide/src/ascii.css":"29afb08833b2fe2250f0412e1fa1161a2432a0820a14953c87124407417c741a","perf-guide/src/bound_checks.md":"5e4991ff58a183ef0cd9fdc1feb4cd12d083b44bdf87393bbb0927808ef3ce7d","perf-guide/src/float-math/approx.md":"8c09032fa2d795a0c5db1775826c850d28eb2627846d0965c60ee72de63735ad","perf-guide/src/float-math/fma.md":"311076ba4b741d604a82e74b83a8d7e8c318fcbd7f64c4392d1cf5af95c60243","perf-guide/src/float-math/fp.md":"04153e775ab6e4f0d7837bcc515230d327b04edfa34c84ce9c9e10ebaeef2be8","perf-guide/src/float-math/svml.md":"0798873b8eedaeda5fed62dc91645b57c20775a02d3cd74d8bd06958f1516506","perf-guide/src/introduction.md":"9f5a19e9e6751f25d2daad39891a0cc600974527ec4c8305843f9618910671bd","perf-guide/src/prof/linux.md":"447731eb5de7d69166728fdbc5ecb0c0c9db678ea493b45a592d67dd002184c0","perf-guide/src/prof/mca.md":"f56d54f3d20e7aa4d32052186e8237b03d65971eb5d112802b442570ff11d344","perf-guide/src/prof/profiling.md":"8a650c0fd6ede0964789bb6577557eeef1d8226a896788602ce61528e260e43c","perf-guide/src/target-feature/attribute.md":"615f88dca0a707b6c416fa605435dd6e1fb5361cc639429cbf68cd87624bd78b","perf-guide/src/target-feature/features.md":"17077760ff24c006b606dd21889c53d87228f4311f3ba3a574f9afdeacd86165","perf-guide/src/target-feature/inlining.md":"7ed1d7068d8173a00d84c16cfe5871cd68b9f04f8d0cca2d01ebc84957ebf2f6","perf-guide/src/target-feature/practice.md":"c4b371842e0086df178488fec97f20def8f0c62ee588bcd25fd948b9b1fa227e","perf-guide/src/target-feature/runtime.md":"835425f5ee597fb3e51d36e725a81ebee29f4561231d19563cd4da81dbb1cfcb","perf-guide/src/target-feature/rustflags.md":"ab49712e9293a65d74d540ba4784fcb57ff1119ec05a575d895c071f1a620f64","perf-guide/src/vert-hor-ops.md":"c6211c0ee91e60552ec592d89d9d957eedc21dee3cbd89e1ad6765ea06a27471","readme.md":"585a8f0e16877fb9abb00cd17a175fcb9d7857840c6c61209f1827ffab095070","rustfmt.toml":"de6101d0670bad65fb3b337d56957d2a024e017e5ab146ec784d77312daaf8ff","src/api.rs":"331a3a4abb19cee2df5f2df4ad7c3e88b45e62cf23fdacfc9bbaa633dc5cf788","src/api/bit_manip.rs":"e68290ee679cc5abc9c73afbe635c1035f8cbfe849e5c751a1680e459244c39e","src/api/cast.rs":"03b94a3d316ac7b7be7068810044911e965e889a0ace7bae762749ca74a92747","src/api/cast/macros.rs":"b0a14d0c83ad2ebb7a275180f6d9e3f2bc312ba57a7d3d6c39fad4e0f20f9408","src/api/cast/v128.rs":"63e28c6a3edf1a7a635f51b8d3c6adbb1d46f884d92a196b3d4a6e743d809416","src/api/cast/v16.rs":"2a584eeb57fd47baad6f3533764301b04aaaac23702b7a8db12598ac02899262","src/api/cast/v256.rs":"b91c15ed8d1536ecd97b4eb79ff9d5aba0552cd9b6f0ea6435b05f2273e23b3a","src/api/cast/v32.rs":"62ec89fcce7fa7f28497ee5770adc8f81d2d3a6b2925b02f7dc06504c40e8f38","src/api/cast/v512.rs":"d855cb943ae7106e9599ef38e30a3afb1c6bd5433178baca54cb128fd9a7d143","src/api/cast/v64.rs":"fe0f7dfaf4fc0c0c1a78c96fcfcdfdc2a1e2845843b11aa797a0c6fb52a8f774","src/api/cmp.rs":"357c3a2a09c6d4611c32dd7fa95be2fae933d513e229026ec9b44451a77b884e","src/api/cmp/eq.rs":"60f70f355bae4cb5b17db53204cacc3890f70670611c17df638d4c04f7cc8075","src/api/cmp/ord.rs":"589f7234761c294fa5df8f525bc4acd5a47cdb602207d524a0d4e19804cd9695","src/api/cmp/partial_eq.rs":"3ed23d2a930b0f9750c3a5309da766b03dc4f9c4d375b42ad3c50fe732693d15","src/api/cmp/partial_ord.rs":"e16b11805c94048acd058c93994b5bc74bb187f8d7e3b86a87df60e1601467f9","src/api/cmp/vertical.rs":"de3d62f38eba817299aa16f1e1939954c9a447e316509397465c2830852ba053","src/api/default.rs":"b61f92fc0e33a2633b3375eb405beba480da071cde03df4d437d8a6058afcd97","src/api/fmt.rs":"67fb804bb86b6cd77cf8cd492b5733ce437071b66fe3297278b8a6552c325dda","src/api/fmt/binary.rs":"35cb5c266197d6224d598fb3d286e5fe48ef0c01ed356c2ff6fe9ba946f96a92","src/api/fmt/debug.rs":"aa18eea443bf353fea3db8b1a025132bbcaf91e747ecfa43b8d9fce9af395a0c","src/api/fmt/lower_hex.rs":"69d5be366631af309f214e8031c8c20267fcc27a695eac6f45c6bc1df72a67e6","src/api/fmt/octal.rs":"9eb11ba3d990213f3c7f1ec25edba7ce997cb1320e16d308c83498ba6b9bfbd9","src/api/fmt/upper_hex.rs":"a4637d085b7bb20e759ce58e08435b510a563ba3dd468af2b03560fdc5511562","src/api/from.rs":"2e599d8329cb05eaf06224cc441355c4b7b51254fc19256619333be8c149d444","src/api/from/from_array.rs":"4151593c7bba7455821fffa5b59867005a77c95d32f1f0cc3fd87294000157d9","src/api/from/from_vector.rs":"9764371aa9e6005aace74dea14f59e5611a095b7cf42707940924749282c52f0","src/api/hash.rs":"562cfa3f1d8eb9a733c035a3665a599c2f1e341ee820d8fbdd102a4398a441bc","src/api/into_bits.rs":"82297f0697d67b5a015e904e7e6e7b2a7066ba825bc54b94b4ff3e22d7a1eefb","src/api/into_bits/arch_specific.rs":"1f925390b0ce7132587d95f2419c6e2ad3e1a9d17eb1d9c120a1c1c4bdf4277e","src/api/into_bits/macros.rs":"d762406de25aedff88d460dec7a80dc8e825a2a419d53218ce007efa6a1d3e04","src/api/into_bits/v128.rs":"ecdc5893664c71d7ab1ff3697c3fbe490d20d8748b9b76881d05e7625e40d74c","src/api/into_bits/v16.rs":"5459ec7dad1ad7bd30dc7e48374580b993abf23701d9c3cb22203fa0a9aabb6d","src/api/into_bits/v256.rs":"90ea351da0380ead1bf0f63b620afd40d01d638d09f7e7be31840bd2c1d9c663","src/api/into_bits/v32.rs":"ee1dc5a430050e16f51154b5fe85b1536f5feddf2ea23dd1d3859b67c4afc6fc","src/api/into_bits/v512.rs":"f72098ed1c9a23944f3d01abaf5e0f2d0e81d35a06fdadd2183e896d41b59867","src/api/into_bits/v64.rs":"6394462facdfe7827349c742b7801f1291e75a720dfb8c0b52100df46f371c98","src/api/math.rs":"8b2a2fc651917a850539f993aa0b9e5bf4da67b11685285b8de8cdca311719ec","src/api/math/float.rs":"61d2794d68262a1090ae473bd30793b5f65cf732f32a6694a3af2ce5d9225616","src/api/math/float/abs.rs":"5b6b2701e2e11135b7ce58a05052ea8120e10e4702c95d046b9d21b827b26bf8","src/api/math/float/consts.rs":"78acba000d3fa527111300b6327c1932de9c4c1e02d4174e1a5615c01463d38c","src/api/math/float/cos.rs":"4c2dd7173728ef189314f1576c9486e03be21b7da98843b2f9011282a7979e31","src/api/math/float/exp.rs":"7c6d5f1e304f498a01cfa23b92380c815d7da0ad94eae3483783bc377d287eef","src/api/math/float/ln.rs":"54c7583f3df793b39ff57534fade27b41bb992439e5dc178252f5ca3190a3e54","src/api/math/float/mul_add.rs":"62cac77660d20159276d4c9ef066eb90c81cbddb808e8e157182c607625ad2eb","src/api/math/float/mul_adde.rs":"bae056ee9f3a70df39ec3c3b2f6437c65303888a7b843ef1a5bcf1f5aca0e602","src/api/math/float/powf.rs":"9ddb938984b36d39d82a82f862f80df8f7fb013f1d222d45698d41d88472f568","src/api/math/float/recpre.rs":"589225794ff1dbf31158dff660e6d4509ecc8befbb57c633900dea5ac0b840d6","src/api/math/float/rsqrte.rs":"a32abdcc318d7ccc8448231f54d75b884b7cbeb03a7d595713ab6243036f4dbf","src/api/math/float/sin.rs":"cbd3622b7df74f19691743001c8cf747a201f8977ad90542fee915f37dcd1e49","src/api/math/float/sqrt.rs":"0c66d5d63fb08e4d99c6b82a8828e41173aff1ac9fa1a2764a11fac217ccf2ac","src/api/math/float/sqrte.rs":"731e1c9f321b662accdd27dacb3aac2e8043b7aecb2f2161dde733bd9f025362","src/api/minimal.rs":"1f22bcc528555444e76de569ec0ae2029b9ae9d04805efeafa93369c8098036b","src/api/minimal/iuf.rs":"c501a6696950cf5e521765f178de548af64fdfb6e10d026616d09fab93ca2d17","src/api/minimal/mask.rs":"42e415f536c5193d0218f5a754b34b87fd7c971bff068009f958712166ff056d","src/api/minimal/ptr.rs":"a9ee482d1dd1c956fb8f3f179e6e620b1de4e9d713961461d4c6923a4ef2e67c","src/api/ops.rs":"3e273b277a0f3019d42c3c59ca94a5afd4885d5ae6d2182e5089bbeec9de42ee","src/api/ops/scalar_arithmetic.rs":"d2d5ad897a59dd0787544f927e0e7ca4072c3e58b0f4a2324083312b0d5a21d7","src/api/ops/scalar_bitwise.rs":"482204e459ca6be79568e1c9f70adbe2d2151412ddf122fb2161be8ebb51c40c","src/api/ops/scalar_mask_bitwise.rs":"c250f52042e37b22d57256c80d4604104cfd2fbe2a2e127c676267270ca5d350","src/api/ops/scalar_shifts.rs":"987f8fdebeedc16e3d77c1b732e7826ef70633c541d16dfa290845d5c6289150","src/api/ops/vector_arithmetic.rs":"ddca15d09ddeef502c2ed66117a62300ca65d87e959e8b622d767bdf1c307910","src/api/ops/vector_bitwise.rs":"b3968f7005b649edcc22a54e2379b14d5ee19045f2e784029805781ae043b5ee","src/api/ops/vector_float_min_max.rs":"f5155dce75219f4ba11275b1f295d2fdcddd49d174a6f1fb2ace7ea42813ce41","src/api/ops/vector_int_min_max.rs":"a378789c6ff9b32a51fbd0a97ffd36ed102cd1fe6a067d2b02017c1df342def6","src/api/ops/vector_mask_bitwise.rs":"5052d18517d765415d40327e6e8e55a312daaca0a5e2aec959bfa54b1675f9c8","src/api/ops/vector_neg.rs":"5c62f6b0221983cdbd23cd0a3af3672e6ba1255f0dfe8b19aae6fbd6503e231b","src/api/ops/vector_rotates.rs":"03cbe8a400fd7c688e4ee771a990a6754f2031b1a59b19ae81158b21471167e5","src/api/ops/vector_shifts.rs":"9bf69d0087268f61009e39aea52e03a90f378910206b6a28e8393178b6a5d0e0","src/api/ptr.rs":"8a793251bed6130dcfb2f1519ceaa18b751bbb15875928d0fb6deb5a5e07523a","src/api/ptr/gather_scatter.rs":"9ddd960365e050674b25b2fd3116e24d94669b4375d74e71c03e3f1469576066","src/api/reductions.rs":"ae5baca81352ecd44526d6c30c0a1feeda475ec73ddd3c3ec6b14e944e5448ee","src/api/reductions/bitwise.rs":"8bf910ae226188bd15fc7e125f058cd2566b6186fcd0cd8fd020f352c39ce139","src/api/reductions/float_arithmetic.rs":"e58c8c87806a95df2b2b5b48ac5991036df024096d9d7c171a480fe9282896a4","src/api/reductions/integer_arithmetic.rs":"47471da1c5f859489680bb5d34ced3d3aa20081c16053a3af121a4496fcb57bf","src/api/reductions/mask.rs":"db83327a950e33a317f37fd33ca4e20c347fb415975ec024f3e23da8509425af","src/api/reductions/min_max.rs":"f27be3aa28e1c1f46de7890198db6e12f00c207085e89ef2de7e57ee443cdb98","src/api/select.rs":"a98e2ccf9fc6bdeed32d337c8675bc96c2fbe2cc34fbf149ad6047fb8e749774","src/api/shuffle.rs":"da58200790868c09659819322a489929a5b6e56c596ed07e6a44293ea02e7d09","src/api/shuffle1_dyn.rs":"bfea5a91905b31444e9ef7ca6eddb7a9606b7e22d3f71bb842eb2795a0346620","src/api/slice.rs":"ee87484e8af329547b9a5d4f2a69e8bed6ea10bbd96270d706083843d4eea2ac","src/api/slice/from_slice.rs":"4d4fe8a329c885fcb4fbcbedf99efb15a95296fe6b3f595056cc37037450d5ac","src/api/slice/write_to_slice.rs":"f5b23b2c4b91cfb26b713a9013a6c0da7f45eaefb79ba06dcbc27f3f23bda679","src/api/swap_bytes.rs":"4a6792a2e49a77475e1b237592b4b2804dbddb79c474331acd0dd71b36934259","src/codegen.rs":"c6eebc3d3665420aa6a2f317977e3c41a4f43e0550ac630cdbe8e4bbed5e2031","src/codegen/bit_manip.rs":"5559e095105a80003e0de35af1d19b0c65c9ab04eb743c7e01c5442d882eb34e","src/codegen/llvm.rs":"d1299c189abb17a6133f047574cffc7a6db4c1be37cb7d4785491cb5e8f8cf54","src/codegen/math.rs":"35f96e37a78fcf0cdb02146b7f27a45108fe06a37fc2a54d8851ce131a326178","src/codegen/math/float.rs":"dd86c0449e576c83b719700962ac017c332987fac08d91f2b7a2b1b883598170","src/codegen/math/float/abs.rs":"f56e2b4b8055ea861c1f5cbc6b6e1d8e7e5af163b62c13574ddee4e09513bfbc","src/codegen/math/float/cos.rs":"ef3b511a24d23045b310315e80348a9b7fedb576fc2de52d74290616a0abeb2a","src/codegen/math/float/cos_pi.rs":"4e7631a5d73dac21531e09ef1802d1180f8997509c2c8fa9f67f322194263a97","src/codegen/math/float/exp.rs":"61b691598c41b5622f24e4320c1bdd08701e612a516438bdddcc728fc3405c8c","src/codegen/math/float/ln.rs":"46b718b1ba8c9d99e1ad40f53d20dfde08a3063ca7bd2a9fdd6698e060da687e","src/codegen/math/float/macros.rs":"dd42135fff13f9aca4fd3a1a4e14c7e6c31aadc6d817d63b0d2fb9e62e062744","src/codegen/math/float/mul_add.rs":"a37bf764345d4b1714f97e83897b7cf0855fc2811704bcbc0012db91825339e1","src/codegen/math/float/mul_adde.rs":"c75702bfcb361de45964a93caf959a695ef2376bd069227600b8c6872665c755","src/codegen/math/float/powf.rs":"642346e982bc4c39203de0864d2149c4179cd7b21cf67a2951687932b4675872","src/codegen/math/float/sin.rs":"9d68164c90cdca6a85155040cdac42e27342ebe0b925273ef1593df721af4258","src/codegen/math/float/sin_cos_pi.rs":"9be02ad48585a1e8d99129382fbffbaed47852f15459256a708850b6b7a75405","src/codegen/math/float/sin_pi.rs":"9890347905b4d4a3c7341c3eb06406e46e60582bcf6960688bd727e5dadc6c57","src/codegen/math/float/sqrt.rs":"e3c60dcfb0c6d2fc62adabcc931b2d4040b83cab294dea36443fb4b89eb79e34","src/codegen/math/float/sqrte.rs":"f0f4ef9eb475ae41bcc7ec6a95ad744ba6b36925faa8b2c2814004396d196b63","src/codegen/pointer_sized_int.rs":"a70697169c28218b56fd2e8d5353f2e00671d1150d0c8cef77d613bdfacd84cb","src/codegen/reductions.rs":"645e2514746d01387ddd07f0aa4ffd8430cc9ab428d4fb13773ea319fa25dd95","src/codegen/reductions/mask.rs":"8f1afe6aabf096a3278e1fc3a30f736e04aa8b9ce96373cee22162d18cfe2702","src/codegen/reductions/mask/aarch64.rs":"cba6e17603d39795dcfe8339b6b7d8714c3e162a1f0a635979f037aa24fe4206","src/codegen/reductions/mask/arm.rs":"9447904818aa2c7c25d0963eead452a639a11ca7dbd6d21eedbfcaade07a0f33","src/codegen/reductions/mask/fallback.rs":"7a0ef9f7fd03ae318b495b95e121350cd61caffc5cc6ee17fabf130d5d933453","src/codegen/reductions/mask/fallback_impl.rs":"76547f396e55ef403327c77c314cf8db8c7a5c9b9819bfb925abeacf130249e5","src/codegen/reductions/mask/x86.rs":"14bd2c482071f2355beebcf7b7ecf950ff2dfcdb08c3ca50993092434a9de717","src/codegen/reductions/mask/x86/avx.rs":"b4913d87844c522903641cbbf10db4551addb1ce5e9e78278e21612fa65c733b","src/codegen/reductions/mask/x86/avx2.rs":"677aed3f056285285daa3adff8bc65e739630b4424defa6d9665e160f027507e","src/codegen/reductions/mask/x86/sse.rs":"226610b4ff88c676d5187114dd57b4a8800de6ce40884675e9198445b1ed0306","src/codegen/reductions/mask/x86/sse2.rs":"bc38e6c31cb4b3d62147eba6cac264e519e2a48e0f7ce9010cfa9ef0cf0ec9fd","src/codegen/shuffle.rs":"0abca97e92cdce49a58a39cc447eb09dc7d7715ef256c8dbd2181a186e61bb64","src/codegen/shuffle1_dyn.rs":"04523e9338133bdedb012dd076c2c564b79ce5593b0fc56d0fb6910e04190a81","src/codegen/swap_bytes.rs":"1d6cdc716eadddc92b4fd506b2445a821caa8dc00860447de09d7ebd69c2087f","src/codegen/v128.rs":"94226b31ec403d18d9d2fe06713f147c9c79e9b5f9105089088266313f843185","src/codegen/v16.rs":"ddec4ffb66b6f7aaffb9a1780c5ddba82557abd74f45073d335047e04cf74924","src/codegen/v256.rs":"6b63917f0444118d6b1595bff2045e59b97c4d24012bd575f69f1f0efc5a0241","src/codegen/v32.rs":"3477b3c5540aed86e61e2f5807dd31db947413cec9181c587d93ed6ec74f0eba","src/codegen/v512.rs":"5854f99d3aabc4cd42b28a20d9ce447756dc2ba024a409a69b6a8ae1f1842fc5","src/codegen/v64.rs":"e9e89caebfe63d10c0cbca61e4dfdba3b7e02ee0989170f80beed23237ddd950","src/codegen/vPtr.rs":"96d609a9eece4dcbbcc01ba0b8744d7f5958be12774176a2945bc676f4e6b5cb","src/codegen/vSize.rs":"eeee9858749aa82142b27bc120d1989bb74a6b82e1e4efbbeaccc9634dc9acfc","src/lib.rs":"1b5d419ff05ee0370d671810423ccc254708cc8d415c1dbac2a7a36be4bf63a8","src/masks.rs":"870f429967b2d7d5133f4d28d6c753fc5cef0570b27b29d4e966a066d22d2d0e","src/sealed.rs":"ff7f0324276408ae8249941cfa32c90b8835a54d750896b683efea857af19db2","src/testing.rs":"1d3a7862ef625e235a5734ad7204e68d350f902c0695182b1f08a0552432416e","src/testing/macros.rs":"6378856d7a40ba5ec5c7c0dad6327d79f0c77266921c24296d10aed6c68e9b98","src/testing/utils.rs":"d6fd5a5017f1f85d9d99585754f8f6ad06fc3d683b34083543e67a7cc6c1772c","src/v128.rs":"18fe263c4aa28cd06461c7070b0269f69f4a2e75749b8f142a83dfdfe4d22bf5","src/v16.rs":"e5c663c9fb3547eaeac78a5f7db9969f4d8b5ec96112bf2954602fff11f0aebd","src/v256.rs":"68732cd688ad12a56d8b4f8ddf279f77bdfe1be2943c7dc0c1b4f1a76798aa0f","src/v32.rs":"785b22a1ccb4a41bb53dfeb0670f624c0ce42e6cdf62d1747e3283777a1c70bd","src/v512.rs":"d1337bfe07f06a8f37f8e8fa7d4315b9307476ee435ad80dd5269eaed564fbfa","src/v64.rs":"3077468d65125b8f085e9454c8b2463a4d5225697464ba6a1300f8799528fd4b","src/vPtr.rs":"c9a53f41f466e17b6648a4ce390fd8f4d3a848d440eb8a9a803a11608d76eb05","src/vSize.rs":"5c46d3e8c3ee5863d9b6e37e681f871386e0efc254d6d84ba711edb529ce7b3c","tests/endianness.rs":"541a144be017e3dd7da7c8ea49d907dc02538245e8c5f3deb5bd43da92c929e1"},"package":null} +\ No newline at end of file +diff --git a/third_party/rust/proc-macro2/.cargo-checksum.json b/third_party/rust/proc-macro2/.cargo-checksum.json +index eeef4120af..e7849f2896 100644 +--- a/third_party/rust/proc-macro2/.cargo-checksum.json ++++ b/third_party/rust/proc-macro2/.cargo-checksum.json +@@ -1 +1 @@ +-{"files":{"Cargo.toml":"e2c1fc6ed317eeef8462fcd192f6b6389e1d84f0d7afeac78f12c23903deddf8","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"378f5840b258e2779c39418f3f2d7b2ba96f1c7917dd6be0713f88305dbda397","README.md":"362a2156f7645528061b6e8487a2eb0f32f1693012ed82ee57afa05c039bba0d","build.rs":"0cc6e2cb919ddbff59cf1d810283939f97a59f0037540c0f2ee3453237635ff8","src/fallback.rs":"5c6379a90735e27abcc40253b223158c6b1e5784f3850bc423335363e87ef038","src/lib.rs":"ae5251296ad3fcd8b600919a993fec0afd8b56da3e11fef6bc7265b273129936","src/strnom.rs":"37f7791f73f123817ad5403af1d4e2a0714be27401729a2d451bc80b1f26bac9","src/wrapper.rs":"81372e910604217a625aa71c47d43e65f4e008456eae93ac39325c9abf10701a","tests/features.rs":"a86deb8644992a4eb64d9fd493eff16f9cf9c5cb6ade3a634ce0c990cf87d559","tests/marker.rs":"c2652e3ae1dfcb94d2e6313b29712c5dcbd0fe62026913e67bb7cebd7560aade","tests/test.rs":"8c427be9cba1fa8d4a16647e53e3545e5863e29e2c0b311c93c9dd1399abf6a1"},"package":"90cf5f418035b98e655e9cdb225047638296b862b42411c4e45bb88d700f7fc0"} +\ No newline at end of file ++{"files":{"Cargo.toml":"302d447d62c8d091d6241cf62bdad607c0d4ed8ff9f43d9b254c9d99c253ee8e","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"378f5840b258e2779c39418f3f2d7b2ba96f1c7917dd6be0713f88305dbda397","README.md":"e1f9d4fc22cff2c049f166a403b41458632a94357890d31cf0e3ad83807fb430","build.rs":"a71283fbc495095eebbbf46753df3fe2c19505c745b508dea157f65796b64dd7","src/detection.rs":"9d25d896889e65330858f2d6f6223c1b98cd1dad189813ad4161ff189fbda2b8","src/fallback.rs":"b114e013695260f6066395c8712cea112ec2a386010397a80f15a60f8b986444","src/lib.rs":"7f528764a958587f007f0c2a330a6a414bae2c8e73d5ed9fb64ff1b42b1805b1","src/marker.rs":"87fce2d0357f5b7998b6d9dfb064f4a0cbc9dabb19e33d4b514a446243ebe2e8","src/parse.rs":"1d2253eacbd40eb3a2a933be2adcee356af922bdb48cc89ff266252a41fd98a1","src/wrapper.rs":"f52646ce1705c1f6265516f30d4c43297b5f529dd31fb91f4c806be89d5a4122","tests/comments.rs":"ea6cbe6f4c8852e6a0612893c7d4f2c144a2e6a134a6c3db641a320cbfc3c800","tests/features.rs":"a86deb8644992a4eb64d9fd493eff16f9cf9c5cb6ade3a634ce0c990cf87d559","tests/marker.rs":"652db9f25c69ffc65baa60cdca8f195aa2e254d4de0a9ddc85de4dc2470544b6","tests/test.rs":"5f30a704eeb2b9198b57f416d622da72d25cb9bf8d8b12e6d0e90aa2cb0e43fc","tests/test_fmt.rs":"745dfdc41d09c5308c221395eb43f2041f0a1413d2927a813bc2ad4554438fe2"},"package":"1e0704ee1a7e00d7bb417d0770ea303c1bccbabf0ef1667dae92b5967f5f8a71"} +\ No newline at end of file +diff --git a/third_party/rust/proc-macro2/Cargo.toml b/third_party/rust/proc-macro2/Cargo.toml +index 95d653633d..22150c516a 100644 +--- a/third_party/rust/proc-macro2/Cargo.toml ++++ b/third_party/rust/proc-macro2/Cargo.toml +@@ -13,21 +13,22 @@ + [package] + edition = "2018" + name = "proc-macro2" +-version = "1.0.5" +-authors = ["Alex Crichton <alex@alexcrichton.com>"] +-description = "A stable implementation of the upcoming new `proc_macro` API. Comes with an\noption, off by default, to also reimplement itself in terms of the upstream\nunstable API.\n" +-homepage = "https://github.com/alexcrichton/proc-macro2" ++version = "1.0.24" ++authors = ["Alex Crichton <alex@alexcrichton.com>", "David Tolnay <dtolnay@gmail.com>"] ++description = "A substitute implementation of the compiler's `proc_macro` API to decouple\ntoken-based libraries from the procedural macro use case.\n" + documentation = "https://docs.rs/proc-macro2" + readme = "README.md" + keywords = ["macros"] ++categories = ["development-tools::procedural-macro-helpers"] + license = "MIT OR Apache-2.0" + repository = "https://github.com/alexcrichton/proc-macro2" + [package.metadata.docs.rs] + rustc-args = ["--cfg", "procmacro2_semver_exempt"] + rustdoc-args = ["--cfg", "procmacro2_semver_exempt"] ++targets = ["x86_64-unknown-linux-gnu"] + +-[lib] +-name = "proc_macro2" ++[package.metadata.playground] ++features = ["span-locations"] + [dependencies.unicode-xid] + version = "0.2" + [dev-dependencies.quote] +@@ -39,5 +40,3 @@ default = ["proc-macro"] + nightly = [] + proc-macro = [] + span-locations = [] +-[badges.travis-ci] +-repository = "alexcrichton/proc-macro2" +diff --git a/third_party/rust/proc-macro2/README.md b/third_party/rust/proc-macro2/README.md +index 19b0c3b5f8..3d05e871a7 100644 +--- a/third_party/rust/proc-macro2/README.md ++++ b/third_party/rust/proc-macro2/README.md +@@ -1,6 +1,6 @@ + # proc-macro2 + +-[![Build Status](https://api.travis-ci.com/alexcrichton/proc-macro2.svg?branch=master)](https://travis-ci.com/alexcrichton/proc-macro2) ++[![Build Status](https://img.shields.io/github/workflow/status/alexcrichton/proc-macro2/build%20and%20test)](https://github.com/alexcrichton/proc-macro2/actions) + [![Latest Version](https://img.shields.io/crates/v/proc-macro2.svg)](https://crates.io/crates/proc-macro2) + [![Rust Documentation](https://img.shields.io/badge/api-rustdoc-blue.svg)](https://docs.rs/proc-macro2) + +diff --git a/third_party/rust/proc-macro2/build.rs b/third_party/rust/proc-macro2/build.rs +index deb9b92719..b247d874f6 100644 +--- a/third_party/rust/proc-macro2/build.rs ++++ b/third_party/rust/proc-macro2/build.rs +@@ -14,6 +14,10 @@ + // procmacro2_semver_exempt surface area is implemented by using the + // nightly-only proc_macro API. + // ++// "hygiene" ++// Enable Span::mixed_site() and non-dummy behavior of Span::resolved_at ++// and Span::located_at. Enabled on Rust 1.45+. ++// + // "proc_macro_span" + // Enable non-dummy behavior of Span::start and Span::end methods which + // requires an unstable compiler feature. Enabled when building with +@@ -57,6 +61,22 @@ fn main() { + println!("cargo:rustc-cfg=span_locations"); + } + ++ if version.minor < 32 { ++ println!("cargo:rustc-cfg=no_libprocmacro_unwind_safe"); ++ } ++ ++ if version.minor < 39 { ++ println!("cargo:rustc-cfg=no_bind_by_move_pattern_guard"); ++ } ++ ++ if version.minor >= 44 { ++ println!("cargo:rustc-cfg=lexerror_display"); ++ } ++ ++ if version.minor >= 45 { ++ println!("cargo:rustc-cfg=hygiene"); ++ } ++ + let target = env::var("TARGET").unwrap(); + if !enable_use_proc_macro(&target) { + return; +diff --git a/third_party/rust/proc-macro2/src/detection.rs b/third_party/rust/proc-macro2/src/detection.rs +new file mode 100644 +index 0000000000..c597bc99c6 +--- /dev/null ++++ b/third_party/rust/proc-macro2/src/detection.rs +@@ -0,0 +1,67 @@ ++use std::panic::{self, PanicInfo}; ++use std::sync::atomic::*; ++use std::sync::Once; ++ ++static WORKS: AtomicUsize = AtomicUsize::new(0); ++static INIT: Once = Once::new(); ++ ++pub(crate) fn inside_proc_macro() -> bool { ++ match WORKS.load(Ordering::SeqCst) { ++ 1 => return false, ++ 2 => return true, ++ _ => {} ++ } ++ ++ INIT.call_once(initialize); ++ inside_proc_macro() ++} ++ ++pub(crate) fn force_fallback() { ++ WORKS.store(1, Ordering::SeqCst); ++} ++ ++pub(crate) fn unforce_fallback() { ++ initialize(); ++} ++ ++// Swap in a null panic hook to avoid printing "thread panicked" to stderr, ++// then use catch_unwind to determine whether the compiler's proc_macro is ++// working. When proc-macro2 is used from outside of a procedural macro all ++// of the proc_macro crate's APIs currently panic. ++// ++// The Once is to prevent the possibility of this ordering: ++// ++// thread 1 calls take_hook, gets the user's original hook ++// thread 1 calls set_hook with the null hook ++// thread 2 calls take_hook, thinks null hook is the original hook ++// thread 2 calls set_hook with the null hook ++// thread 1 calls set_hook with the actual original hook ++// thread 2 calls set_hook with what it thinks is the original hook ++// ++// in which the user's hook has been lost. ++// ++// There is still a race condition where a panic in a different thread can ++// happen during the interval that the user's original panic hook is ++// unregistered such that their hook is incorrectly not called. This is ++// sufficiently unlikely and less bad than printing panic messages to stderr ++// on correct use of this crate. Maybe there is a libstd feature request ++// here. For now, if a user needs to guarantee that this failure mode does ++// not occur, they need to call e.g. `proc_macro2::Span::call_site()` from ++// the main thread before launching any other threads. ++fn initialize() { ++ type PanicHook = dyn Fn(&PanicInfo) + Sync + Send + 'static; ++ ++ let null_hook: Box<PanicHook> = Box::new(|_panic_info| { /* ignore */ }); ++ let sanity_check = &*null_hook as *const PanicHook; ++ let original_hook = panic::take_hook(); ++ panic::set_hook(null_hook); ++ ++ let works = panic::catch_unwind(proc_macro::Span::call_site).is_ok(); ++ WORKS.store(works as usize + 1, Ordering::SeqCst); ++ ++ let hopefully_null_hook = panic::take_hook(); ++ panic::set_hook(original_hook); ++ if sanity_check != &*hopefully_null_hook { ++ panic!("observed race condition in proc_macro2::inside_proc_macro"); ++ } ++} +diff --git a/third_party/rust/proc-macro2/src/fallback.rs b/third_party/rust/proc-macro2/src/fallback.rs +index fe582b3b5f..8900c5ff0f 100644 +--- a/third_party/rust/proc-macro2/src/fallback.rs ++++ b/third_party/rust/proc-macro2/src/fallback.rs +@@ -1,27 +1,41 @@ ++use crate::parse::{token_stream, Cursor}; ++use crate::{Delimiter, Spacing, TokenTree}; + #[cfg(span_locations)] + use std::cell::RefCell; + #[cfg(span_locations)] + use std::cmp; +-use std::fmt; +-use std::iter; ++use std::fmt::{self, Debug, Display}; ++use std::iter::FromIterator; ++use std::mem; + use std::ops::RangeBounds; + #[cfg(procmacro2_semver_exempt)] + use std::path::Path; + use std::path::PathBuf; + use std::str::FromStr; + use std::vec; +- +-use crate::strnom::{block_comment, skip_whitespace, whitespace, word_break, Cursor, PResult}; +-use crate::{Delimiter, Punct, Spacing, TokenTree}; + use unicode_xid::UnicodeXID; + ++/// Force use of proc-macro2's fallback implementation of the API for now, even ++/// if the compiler's implementation is available. ++pub fn force() { ++ #[cfg(wrap_proc_macro)] ++ crate::detection::force_fallback(); ++} ++ ++/// Resume using the compiler's implementation of the proc macro API if it is ++/// available. ++pub fn unforce() { ++ #[cfg(wrap_proc_macro)] ++ crate::detection::unforce_fallback(); ++} ++ + #[derive(Clone)] +-pub struct TokenStream { +- inner: Vec<TokenTree>, ++pub(crate) struct TokenStream { ++ pub(crate) inner: Vec<TokenTree>, + } + + #[derive(Debug)] +-pub struct LexError; ++pub(crate) struct LexError; + + impl TokenStream { + pub fn new() -> TokenStream { +@@ -31,6 +45,72 @@ impl TokenStream { + pub fn is_empty(&self) -> bool { + self.inner.len() == 0 + } ++ ++ fn take_inner(&mut self) -> Vec<TokenTree> { ++ mem::replace(&mut self.inner, Vec::new()) ++ } ++ ++ fn push_token(&mut self, token: TokenTree) { ++ // https://github.com/alexcrichton/proc-macro2/issues/235 ++ match token { ++ #[cfg(not(no_bind_by_move_pattern_guard))] ++ TokenTree::Literal(crate::Literal { ++ #[cfg(wrap_proc_macro)] ++ inner: crate::imp::Literal::Fallback(literal), ++ #[cfg(not(wrap_proc_macro))] ++ inner: literal, ++ .. ++ }) if literal.text.starts_with('-') => { ++ push_negative_literal(self, literal); ++ } ++ #[cfg(no_bind_by_move_pattern_guard)] ++ TokenTree::Literal(crate::Literal { ++ #[cfg(wrap_proc_macro)] ++ inner: crate::imp::Literal::Fallback(literal), ++ #[cfg(not(wrap_proc_macro))] ++ inner: literal, ++ .. ++ }) => { ++ if literal.text.starts_with('-') { ++ push_negative_literal(self, literal); ++ } else { ++ self.inner ++ .push(TokenTree::Literal(crate::Literal::_new_stable(literal))); ++ } ++ } ++ _ => self.inner.push(token), ++ } ++ ++ #[cold] ++ fn push_negative_literal(stream: &mut TokenStream, mut literal: Literal) { ++ literal.text.remove(0); ++ let mut punct = crate::Punct::new('-', Spacing::Alone); ++ punct.set_span(crate::Span::_new_stable(literal.span)); ++ stream.inner.push(TokenTree::Punct(punct)); ++ stream ++ .inner ++ .push(TokenTree::Literal(crate::Literal::_new_stable(literal))); ++ } ++ } ++} ++ ++// Nonrecursive to prevent stack overflow. ++impl Drop for TokenStream { ++ fn drop(&mut self) { ++ while let Some(token) = self.inner.pop() { ++ let group = match token { ++ TokenTree::Group(group) => group.inner, ++ _ => continue, ++ }; ++ #[cfg(wrap_proc_macro)] ++ let group = match group { ++ crate::imp::Group::Fallback(group) => group, ++ _ => continue, ++ }; ++ let mut group = group; ++ self.inner.extend(group.stream.take_inner()); ++ } ++ } + } + + #[cfg(span_locations)] +@@ -59,20 +139,22 @@ impl FromStr for TokenStream { + // Create a dummy file & add it to the source map + let cursor = get_cursor(src); + +- match token_stream(cursor) { +- Ok((input, output)) => { +- if skip_whitespace(input).len() != 0 { +- Err(LexError) +- } else { +- Ok(output) +- } +- } +- Err(LexError) => Err(LexError), ++ let (rest, tokens) = token_stream(cursor)?; ++ if rest.is_empty() { ++ Ok(tokens) ++ } else { ++ Err(LexError) + } + } + } + +-impl fmt::Display for TokenStream { ++impl Display for LexError { ++ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { ++ f.write_str("cannot parse string into token stream") ++ } ++} ++ ++impl Display for TokenStream { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + let mut joint = false; + for (i, tt) in self.inner.iter().enumerate() { +@@ -80,37 +162,22 @@ impl fmt::Display for TokenStream { + write!(f, " ")?; + } + joint = false; +- match *tt { +- TokenTree::Group(ref tt) => { +- let (start, end) = match tt.delimiter() { +- Delimiter::Parenthesis => ("(", ")"), +- Delimiter::Brace => ("{", "}"), +- Delimiter::Bracket => ("[", "]"), +- Delimiter::None => ("", ""), +- }; +- if tt.stream().into_iter().next().is_none() { +- write!(f, "{} {}", start, end)? +- } else { +- write!(f, "{} {} {}", start, tt.stream(), end)? +- } +- } +- TokenTree::Ident(ref tt) => write!(f, "{}", tt)?, +- TokenTree::Punct(ref tt) => { +- write!(f, "{}", tt.as_char())?; +- match tt.spacing() { +- Spacing::Alone => {} +- Spacing::Joint => joint = true, +- } ++ match tt { ++ TokenTree::Group(tt) => Display::fmt(tt, f), ++ TokenTree::Ident(tt) => Display::fmt(tt, f), ++ TokenTree::Punct(tt) => { ++ joint = tt.spacing() == Spacing::Joint; ++ Display::fmt(tt, f) + } +- TokenTree::Literal(ref tt) => write!(f, "{}", tt)?, +- } ++ TokenTree::Literal(tt) => Display::fmt(tt, f), ++ }? + } + + Ok(()) + } + } + +-impl fmt::Debug for TokenStream { ++impl Debug for TokenStream { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + f.write_str("TokenStream ")?; + f.debug_list().entries(self.clone()).finish() +@@ -139,28 +206,26 @@ impl From<TokenStream> for proc_macro::TokenStream { + + impl From<TokenTree> for TokenStream { + fn from(tree: TokenTree) -> TokenStream { +- TokenStream { inner: vec![tree] } ++ let mut stream = TokenStream::new(); ++ stream.push_token(tree); ++ stream + } + } + +-impl iter::FromIterator<TokenTree> for TokenStream { +- fn from_iter<I: IntoIterator<Item = TokenTree>>(streams: I) -> Self { +- let mut v = Vec::new(); +- +- for token in streams.into_iter() { +- v.push(token); +- } +- +- TokenStream { inner: v } ++impl FromIterator<TokenTree> for TokenStream { ++ fn from_iter<I: IntoIterator<Item = TokenTree>>(tokens: I) -> Self { ++ let mut stream = TokenStream::new(); ++ stream.extend(tokens); ++ stream + } + } + +-impl iter::FromIterator<TokenStream> for TokenStream { ++impl FromIterator<TokenStream> for TokenStream { + fn from_iter<I: IntoIterator<Item = TokenStream>>(streams: I) -> Self { + let mut v = Vec::new(); + +- for stream in streams.into_iter() { +- v.extend(stream.inner); ++ for mut stream in streams { ++ v.extend(stream.take_inner()); + } + + TokenStream { inner: v } +@@ -168,31 +233,30 @@ impl iter::FromIterator<TokenStream> for TokenStream { + } + + impl Extend<TokenTree> for TokenStream { +- fn extend<I: IntoIterator<Item = TokenTree>>(&mut self, streams: I) { +- self.inner.extend(streams); ++ fn extend<I: IntoIterator<Item = TokenTree>>(&mut self, tokens: I) { ++ tokens.into_iter().for_each(|token| self.push_token(token)); + } + } + + impl Extend<TokenStream> for TokenStream { + fn extend<I: IntoIterator<Item = TokenStream>>(&mut self, streams: I) { +- self.inner +- .extend(streams.into_iter().flat_map(|stream| stream)); ++ self.inner.extend(streams.into_iter().flatten()); + } + } + +-pub type TokenTreeIter = vec::IntoIter<TokenTree>; ++pub(crate) type TokenTreeIter = vec::IntoIter<TokenTree>; + + impl IntoIterator for TokenStream { + type Item = TokenTree; + type IntoIter = TokenTreeIter; + +- fn into_iter(self) -> TokenTreeIter { +- self.inner.into_iter() ++ fn into_iter(mut self) -> TokenTreeIter { ++ self.take_inner().into_iter() + } + } + + #[derive(Clone, PartialEq, Eq)] +-pub struct SourceFile { ++pub(crate) struct SourceFile { + path: PathBuf, + } + +@@ -208,7 +272,7 @@ impl SourceFile { + } + } + +-impl fmt::Debug for SourceFile { ++impl Debug for SourceFile { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + f.debug_struct("SourceFile") + .field("path", &self.path()) +@@ -218,7 +282,7 @@ impl fmt::Debug for SourceFile { + } + + #[derive(Clone, Copy, Debug, PartialEq, Eq)] +-pub struct LineColumn { ++pub(crate) struct LineColumn { + pub line: usize, + pub column: usize, + } +@@ -228,23 +292,11 @@ thread_local! { + static SOURCE_MAP: RefCell<SourceMap> = RefCell::new(SourceMap { + // NOTE: We start with a single dummy file which all call_site() and + // def_site() spans reference. +- files: vec![{ ++ files: vec![FileInfo { + #[cfg(procmacro2_semver_exempt)] +- { +- FileInfo { +- name: "<unspecified>".to_owned(), +- span: Span { lo: 0, hi: 0 }, +- lines: vec![0], +- } +- } +- +- #[cfg(not(procmacro2_semver_exempt))] +- { +- FileInfo { +- span: Span { lo: 0, hi: 0 }, +- lines: vec![0], +- } +- } ++ name: "<unspecified>".to_owned(), ++ span: Span { lo: 0, hi: 0 }, ++ lines: vec![0], + }], + }); + } +@@ -282,16 +334,21 @@ impl FileInfo { + } + } + +-/// Computesthe offsets of each line in the given source string. ++/// Computes the offsets of each line in the given source string ++/// and the total number of characters + #[cfg(span_locations)] +-fn lines_offsets(s: &str) -> Vec<usize> { ++fn lines_offsets(s: &str) -> (usize, Vec<usize>) { + let mut lines = vec![0]; +- let mut prev = 0; +- while let Some(len) = s[prev..].find('\n') { +- prev += len + 1; +- lines.push(prev); ++ let mut total = 0; ++ ++ for ch in s.chars() { ++ total += 1; ++ if ch == '\n' { ++ lines.push(total); ++ } + } +- lines ++ ++ (total, lines) + } + + #[cfg(span_locations)] +@@ -310,23 +367,22 @@ impl SourceMap { + } + + fn add_file(&mut self, name: &str, src: &str) -> Span { +- let lines = lines_offsets(src); ++ let (len, lines) = lines_offsets(src); + let lo = self.next_start_pos(); + // XXX(nika): Shouild we bother doing a checked cast or checked add here? + let span = Span { + lo, +- hi: lo + (src.len() as u32), ++ hi: lo + (len as u32), + }; + +- #[cfg(procmacro2_semver_exempt)] + self.files.push(FileInfo { ++ #[cfg(procmacro2_semver_exempt)] + name: name.to_owned(), + span, + lines, + }); + + #[cfg(not(procmacro2_semver_exempt))] +- self.files.push(FileInfo { span, lines }); + let _ = name; + + span +@@ -343,11 +399,11 @@ impl SourceMap { + } + + #[derive(Clone, Copy, PartialEq, Eq)] +-pub struct Span { ++pub(crate) struct Span { + #[cfg(span_locations)] +- lo: u32, ++ pub(crate) lo: u32, + #[cfg(span_locations)] +- hi: u32, ++ pub(crate) hi: u32, + } + + impl Span { +@@ -361,12 +417,16 @@ impl Span { + Span { lo: 0, hi: 0 } + } + ++ #[cfg(hygiene)] ++ pub fn mixed_site() -> Span { ++ Span::call_site() ++ } ++ + #[cfg(procmacro2_semver_exempt)] + pub fn def_site() -> Span { + Span::call_site() + } + +- #[cfg(procmacro2_semver_exempt)] + pub fn resolved_at(&self, _other: Span) -> Span { + // Stable spans consist only of line/column information, so + // `resolved_at` and `located_at` only select which span the +@@ -374,7 +434,6 @@ impl Span { + *self + } + +- #[cfg(procmacro2_semver_exempt)] + pub fn located_at(&self, other: Span) -> Span { + other + } +@@ -427,26 +486,59 @@ impl Span { + }) + }) + } ++ ++ #[cfg(not(span_locations))] ++ fn first_byte(self) -> Self { ++ self ++ } ++ ++ #[cfg(span_locations)] ++ fn first_byte(self) -> Self { ++ Span { ++ lo: self.lo, ++ hi: cmp::min(self.lo.saturating_add(1), self.hi), ++ } ++ } ++ ++ #[cfg(not(span_locations))] ++ fn last_byte(self) -> Self { ++ self ++ } ++ ++ #[cfg(span_locations)] ++ fn last_byte(self) -> Self { ++ Span { ++ lo: cmp::max(self.hi.saturating_sub(1), self.lo), ++ hi: self.hi, ++ } ++ } + } + +-impl fmt::Debug for Span { ++impl Debug for Span { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- #[cfg(procmacro2_semver_exempt)] ++ #[cfg(span_locations)] + return write!(f, "bytes({}..{})", self.lo, self.hi); + +- #[cfg(not(procmacro2_semver_exempt))] ++ #[cfg(not(span_locations))] + write!(f, "Span") + } + } + +-pub fn debug_span_field_if_nontrivial(debug: &mut fmt::DebugStruct, span: Span) { +- if cfg!(procmacro2_semver_exempt) { ++pub(crate) fn debug_span_field_if_nontrivial(debug: &mut fmt::DebugStruct, span: Span) { ++ #[cfg(span_locations)] ++ { ++ if span.lo == 0 && span.hi == 0 { ++ return; ++ } ++ } ++ ++ if cfg!(span_locations) { + debug.field("span", &span); + } + } + + #[derive(Clone)] +-pub struct Group { ++pub(crate) struct Group { + delimiter: Delimiter, + stream: TokenStream, + span: Span, +@@ -474,11 +566,11 @@ impl Group { + } + + pub fn span_open(&self) -> Span { +- self.span ++ self.span.first_byte() + } + + pub fn span_close(&self) -> Span { +- self.span ++ self.span.last_byte() + } + + pub fn set_span(&mut self, span: Span) { +@@ -486,36 +578,45 @@ impl Group { + } + } + +-impl fmt::Display for Group { ++impl Display for Group { ++ // We attempt to match libproc_macro's formatting. ++ // Empty parens: () ++ // Nonempty parens: (...) ++ // Empty brackets: [] ++ // Nonempty brackets: [...] ++ // Empty braces: { } ++ // Nonempty braces: { ... } + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- let (left, right) = match self.delimiter { ++ let (open, close) = match self.delimiter { + Delimiter::Parenthesis => ("(", ")"), +- Delimiter::Brace => ("{", "}"), ++ Delimiter::Brace => ("{ ", "}"), + Delimiter::Bracket => ("[", "]"), + Delimiter::None => ("", ""), + }; + +- f.write_str(left)?; +- self.stream.fmt(f)?; +- f.write_str(right)?; ++ f.write_str(open)?; ++ Display::fmt(&self.stream, f)?; ++ if self.delimiter == Delimiter::Brace && !self.stream.inner.is_empty() { ++ f.write_str(" ")?; ++ } ++ f.write_str(close)?; + + Ok(()) + } + } + +-impl fmt::Debug for Group { ++impl Debug for Group { + fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result { + let mut debug = fmt.debug_struct("Group"); + debug.field("delimiter", &self.delimiter); + debug.field("stream", &self.stream); +- #[cfg(procmacro2_semver_exempt)] +- debug.field("span", &self.span); ++ debug_span_field_if_nontrivial(&mut debug, self.span); + debug.finish() + } + } + + #[derive(Clone)] +-pub struct Ident { ++pub(crate) struct Ident { + sym: String, + span: Span, + raw: bool, +@@ -549,16 +650,14 @@ impl Ident { + } + } + +-#[inline] +-fn is_ident_start(c: char) -> bool { ++pub(crate) fn is_ident_start(c: char) -> bool { + ('a' <= c && c <= 'z') + || ('A' <= c && c <= 'Z') + || c == '_' + || (c > '\x7f' && UnicodeXID::is_xid_start(c)) + } + +-#[inline] +-fn is_ident_continue(c: char) -> bool { ++pub(crate) fn is_ident_continue(c: char) -> bool { + ('a' <= c && c <= 'z') + || ('A' <= c && c <= 'Z') + || c == '_' +@@ -615,18 +714,18 @@ where + } + } + +-impl fmt::Display for Ident { ++impl Display for Ident { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + if self.raw { +- "r#".fmt(f)?; ++ f.write_str("r#")?; + } +- self.sym.fmt(f) ++ Display::fmt(&self.sym, f) + } + } + +-impl fmt::Debug for Ident { ++impl Debug for Ident { + // Ident(proc_macro), Ident(r#union) +- #[cfg(not(procmacro2_semver_exempt))] ++ #[cfg(not(span_locations))] + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + let mut debug = f.debug_tuple("Ident"); + debug.field(&format_args!("{}", self)); +@@ -637,17 +736,17 @@ impl fmt::Debug for Ident { + // sym: proc_macro, + // span: bytes(128..138) + // } +- #[cfg(procmacro2_semver_exempt)] ++ #[cfg(span_locations)] + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + let mut debug = f.debug_struct("Ident"); + debug.field("sym", &format_args!("{}", self)); +- debug.field("span", &self.span); ++ debug_span_field_if_nontrivial(&mut debug, self.span); + debug.finish() + } + } + + #[derive(Clone)] +-pub struct Literal { ++pub(crate) struct Literal { + text: String, + span: Span, + } +@@ -669,7 +768,7 @@ macro_rules! unsuffixed_numbers { + } + + impl Literal { +- fn _new(text: String) -> Literal { ++ pub(crate) fn _new(text: String) -> Literal { + Literal { + text, + span: Span::call_site(), +@@ -711,7 +810,7 @@ impl Literal { + + pub fn f32_unsuffixed(f: f32) -> Literal { + let mut s = f.to_string(); +- if !s.contains(".") { ++ if !s.contains('.') { + s.push_str(".0"); + } + Literal::_new(s) +@@ -719,7 +818,7 @@ impl Literal { + + pub fn f64_unsuffixed(f: f64) -> Literal { + let mut s = f.to_string(); +- if !s.contains(".") { ++ if !s.contains('.') { + s.push_str(".0"); + } + Literal::_new(s) +@@ -730,10 +829,10 @@ impl Literal { + text.push('"'); + for c in t.chars() { + if c == '\'' { +- // escape_default turns this into "\'" which is unnecessary. ++ // escape_debug turns this into "\'" which is unnecessary. + text.push(c); + } else { +- text.extend(c.escape_default()); ++ text.extend(c.escape_debug()); + } + } + text.push('"'); +@@ -744,10 +843,10 @@ impl Literal { + let mut text = String::new(); + text.push('\''); + if t == '"' { +- // escape_default turns this into '\"' which is unnecessary. ++ // escape_debug turns this into '\"' which is unnecessary. + text.push(t); + } else { +- text.extend(t.escape_default()); ++ text.extend(t.escape_debug()); + } + text.push('\''); + Literal::_new(text) +@@ -756,6 +855,7 @@ impl Literal { + pub fn byte_string(bytes: &[u8]) -> Literal { + let mut escaped = "b\"".to_string(); + for b in bytes { ++ #[allow(clippy::match_overlapping_arm)] + match *b { + b'\0' => escaped.push_str(r"\0"), + b'\t' => escaped.push_str(r"\t"), +@@ -784,651 +884,17 @@ impl Literal { + } + } + +-impl fmt::Display for Literal { ++impl Display for Literal { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.text.fmt(f) ++ Display::fmt(&self.text, f) + } + } + +-impl fmt::Debug for Literal { ++impl Debug for Literal { + fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result { + let mut debug = fmt.debug_struct("Literal"); + debug.field("lit", &format_args!("{}", self.text)); +- #[cfg(procmacro2_semver_exempt)] +- debug.field("span", &self.span); ++ debug_span_field_if_nontrivial(&mut debug, self.span); + debug.finish() + } + } +- +-fn token_stream(mut input: Cursor) -> PResult<TokenStream> { +- let mut trees = Vec::new(); +- loop { +- let input_no_ws = skip_whitespace(input); +- if input_no_ws.rest.len() == 0 { +- break; +- } +- if let Ok((a, tokens)) = doc_comment(input_no_ws) { +- input = a; +- trees.extend(tokens); +- continue; +- } +- +- let (a, tt) = match token_tree(input_no_ws) { +- Ok(p) => p, +- Err(_) => break, +- }; +- trees.push(tt); +- input = a; +- } +- Ok((input, TokenStream { inner: trees })) +-} +- +-#[cfg(not(span_locations))] +-fn spanned<'a, T>( +- input: Cursor<'a>, +- f: fn(Cursor<'a>) -> PResult<'a, T>, +-) -> PResult<'a, (T, crate::Span)> { +- let (a, b) = f(skip_whitespace(input))?; +- Ok((a, ((b, crate::Span::_new_stable(Span::call_site()))))) +-} +- +-#[cfg(span_locations)] +-fn spanned<'a, T>( +- input: Cursor<'a>, +- f: fn(Cursor<'a>) -> PResult<'a, T>, +-) -> PResult<'a, (T, crate::Span)> { +- let input = skip_whitespace(input); +- let lo = input.off; +- let (a, b) = f(input)?; +- let hi = a.off; +- let span = crate::Span::_new_stable(Span { lo, hi }); +- Ok((a, (b, span))) +-} +- +-fn token_tree(input: Cursor) -> PResult<TokenTree> { +- let (rest, (mut tt, span)) = spanned(input, token_kind)?; +- tt.set_span(span); +- Ok((rest, tt)) +-} +- +-named!(token_kind -> TokenTree, alt!( +- map!(group, |g| TokenTree::Group(crate::Group::_new_stable(g))) +- | +- map!(literal, |l| TokenTree::Literal(crate::Literal::_new_stable(l))) // must be before symbol +- | +- map!(op, TokenTree::Punct) +- | +- symbol_leading_ws +-)); +- +-named!(group -> Group, alt!( +- delimited!( +- punct!("("), +- token_stream, +- punct!(")") +- ) => { |ts| Group::new(Delimiter::Parenthesis, ts) } +- | +- delimited!( +- punct!("["), +- token_stream, +- punct!("]") +- ) => { |ts| Group::new(Delimiter::Bracket, ts) } +- | +- delimited!( +- punct!("{"), +- token_stream, +- punct!("}") +- ) => { |ts| Group::new(Delimiter::Brace, ts) } +-)); +- +-fn symbol_leading_ws(input: Cursor) -> PResult<TokenTree> { +- symbol(skip_whitespace(input)) +-} +- +-fn symbol(input: Cursor) -> PResult<TokenTree> { +- let raw = input.starts_with("r#"); +- let rest = input.advance((raw as usize) << 1); +- +- let (rest, sym) = symbol_not_raw(rest)?; +- +- if !raw { +- let ident = crate::Ident::new(sym, crate::Span::call_site()); +- return Ok((rest, ident.into())); +- } +- +- if sym == "_" { +- return Err(LexError); +- } +- +- let ident = crate::Ident::_new_raw(sym, crate::Span::call_site()); +- Ok((rest, ident.into())) +-} +- +-fn symbol_not_raw(input: Cursor) -> PResult<&str> { +- let mut chars = input.char_indices(); +- +- match chars.next() { +- Some((_, ch)) if is_ident_start(ch) => {} +- _ => return Err(LexError), +- } +- +- let mut end = input.len(); +- for (i, ch) in chars { +- if !is_ident_continue(ch) { +- end = i; +- break; +- } +- } +- +- Ok((input.advance(end), &input.rest[..end])) +-} +- +-fn literal(input: Cursor) -> PResult<Literal> { +- let input_no_ws = skip_whitespace(input); +- +- match literal_nocapture(input_no_ws) { +- Ok((a, ())) => { +- let start = input.len() - input_no_ws.len(); +- let len = input_no_ws.len() - a.len(); +- let end = start + len; +- Ok((a, Literal::_new(input.rest[start..end].to_string()))) +- } +- Err(LexError) => Err(LexError), +- } +-} +- +-named!(literal_nocapture -> (), alt!( +- string +- | +- byte_string +- | +- byte +- | +- character +- | +- float +- | +- int +-)); +- +-named!(string -> (), alt!( +- quoted_string +- | +- preceded!( +- punct!("r"), +- raw_string +- ) => { |_| () } +-)); +- +-named!(quoted_string -> (), do_parse!( +- punct!("\"") >> +- cooked_string >> +- tag!("\"") >> +- option!(symbol_not_raw) >> +- (()) +-)); +- +-fn cooked_string(input: Cursor) -> PResult<()> { +- let mut chars = input.char_indices().peekable(); +- while let Some((byte_offset, ch)) = chars.next() { +- match ch { +- '"' => { +- return Ok((input.advance(byte_offset), ())); +- } +- '\r' => { +- if let Some((_, '\n')) = chars.next() { +- // ... +- } else { +- break; +- } +- } +- '\\' => match chars.next() { +- Some((_, 'x')) => { +- if !backslash_x_char(&mut chars) { +- break; +- } +- } +- Some((_, 'n')) | Some((_, 'r')) | Some((_, 't')) | Some((_, '\\')) +- | Some((_, '\'')) | Some((_, '"')) | Some((_, '0')) => {} +- Some((_, 'u')) => { +- if !backslash_u(&mut chars) { +- break; +- } +- } +- Some((_, '\n')) | Some((_, '\r')) => { +- while let Some(&(_, ch)) = chars.peek() { +- if ch.is_whitespace() { +- chars.next(); +- } else { +- break; +- } +- } +- } +- _ => break, +- }, +- _ch => {} +- } +- } +- Err(LexError) +-} +- +-named!(byte_string -> (), alt!( +- delimited!( +- punct!("b\""), +- cooked_byte_string, +- tag!("\"") +- ) => { |_| () } +- | +- preceded!( +- punct!("br"), +- raw_string +- ) => { |_| () } +-)); +- +-fn cooked_byte_string(mut input: Cursor) -> PResult<()> { +- let mut bytes = input.bytes().enumerate(); +- 'outer: while let Some((offset, b)) = bytes.next() { +- match b { +- b'"' => { +- return Ok((input.advance(offset), ())); +- } +- b'\r' => { +- if let Some((_, b'\n')) = bytes.next() { +- // ... +- } else { +- break; +- } +- } +- b'\\' => match bytes.next() { +- Some((_, b'x')) => { +- if !backslash_x_byte(&mut bytes) { +- break; +- } +- } +- Some((_, b'n')) | Some((_, b'r')) | Some((_, b't')) | Some((_, b'\\')) +- | Some((_, b'0')) | Some((_, b'\'')) | Some((_, b'"')) => {} +- Some((newline, b'\n')) | Some((newline, b'\r')) => { +- let rest = input.advance(newline + 1); +- for (offset, ch) in rest.char_indices() { +- if !ch.is_whitespace() { +- input = rest.advance(offset); +- bytes = input.bytes().enumerate(); +- continue 'outer; +- } +- } +- break; +- } +- _ => break, +- }, +- b if b < 0x80 => {} +- _ => break, +- } +- } +- Err(LexError) +-} +- +-fn raw_string(input: Cursor) -> PResult<()> { +- let mut chars = input.char_indices(); +- let mut n = 0; +- while let Some((byte_offset, ch)) = chars.next() { +- match ch { +- '"' => { +- n = byte_offset; +- break; +- } +- '#' => {} +- _ => return Err(LexError), +- } +- } +- for (byte_offset, ch) in chars { +- match ch { +- '"' if input.advance(byte_offset + 1).starts_with(&input.rest[..n]) => { +- let rest = input.advance(byte_offset + 1 + n); +- return Ok((rest, ())); +- } +- '\r' => {} +- _ => {} +- } +- } +- Err(LexError) +-} +- +-named!(byte -> (), do_parse!( +- punct!("b") >> +- tag!("'") >> +- cooked_byte >> +- tag!("'") >> +- (()) +-)); +- +-fn cooked_byte(input: Cursor) -> PResult<()> { +- let mut bytes = input.bytes().enumerate(); +- let ok = match bytes.next().map(|(_, b)| b) { +- Some(b'\\') => match bytes.next().map(|(_, b)| b) { +- Some(b'x') => backslash_x_byte(&mut bytes), +- Some(b'n') | Some(b'r') | Some(b't') | Some(b'\\') | Some(b'0') | Some(b'\'') +- | Some(b'"') => true, +- _ => false, +- }, +- b => b.is_some(), +- }; +- if ok { +- match bytes.next() { +- Some((offset, _)) => { +- if input.chars().as_str().is_char_boundary(offset) { +- Ok((input.advance(offset), ())) +- } else { +- Err(LexError) +- } +- } +- None => Ok((input.advance(input.len()), ())), +- } +- } else { +- Err(LexError) +- } +-} +- +-named!(character -> (), do_parse!( +- punct!("'") >> +- cooked_char >> +- tag!("'") >> +- (()) +-)); +- +-fn cooked_char(input: Cursor) -> PResult<()> { +- let mut chars = input.char_indices(); +- let ok = match chars.next().map(|(_, ch)| ch) { +- Some('\\') => match chars.next().map(|(_, ch)| ch) { +- Some('x') => backslash_x_char(&mut chars), +- Some('u') => backslash_u(&mut chars), +- Some('n') | Some('r') | Some('t') | Some('\\') | Some('0') | Some('\'') | Some('"') => { +- true +- } +- _ => false, +- }, +- ch => ch.is_some(), +- }; +- if ok { +- match chars.next() { +- Some((idx, _)) => Ok((input.advance(idx), ())), +- None => Ok((input.advance(input.len()), ())), +- } +- } else { +- Err(LexError) +- } +-} +- +-macro_rules! next_ch { +- ($chars:ident @ $pat:pat $(| $rest:pat)*) => { +- match $chars.next() { +- Some((_, ch)) => match ch { +- $pat $(| $rest)* => ch, +- _ => return false, +- }, +- None => return false +- } +- }; +-} +- +-fn backslash_x_char<I>(chars: &mut I) -> bool +-where +- I: Iterator<Item = (usize, char)>, +-{ +- next_ch!(chars @ '0'..='7'); +- next_ch!(chars @ '0'..='9' | 'a'..='f' | 'A'..='F'); +- true +-} +- +-fn backslash_x_byte<I>(chars: &mut I) -> bool +-where +- I: Iterator<Item = (usize, u8)>, +-{ +- next_ch!(chars @ b'0'..=b'9' | b'a'..=b'f' | b'A'..=b'F'); +- next_ch!(chars @ b'0'..=b'9' | b'a'..=b'f' | b'A'..=b'F'); +- true +-} +- +-fn backslash_u<I>(chars: &mut I) -> bool +-where +- I: Iterator<Item = (usize, char)>, +-{ +- next_ch!(chars @ '{'); +- next_ch!(chars @ '0'..='9' | 'a'..='f' | 'A'..='F'); +- loop { +- let c = next_ch!(chars @ '0'..='9' | 'a'..='f' | 'A'..='F' | '_' | '}'); +- if c == '}' { +- return true; +- } +- } +-} +- +-fn float(input: Cursor) -> PResult<()> { +- let (mut rest, ()) = float_digits(input)?; +- if let Some(ch) = rest.chars().next() { +- if is_ident_start(ch) { +- rest = symbol_not_raw(rest)?.0; +- } +- } +- word_break(rest) +-} +- +-fn float_digits(input: Cursor) -> PResult<()> { +- let mut chars = input.chars().peekable(); +- match chars.next() { +- Some(ch) if ch >= '0' && ch <= '9' => {} +- _ => return Err(LexError), +- } +- +- let mut len = 1; +- let mut has_dot = false; +- let mut has_exp = false; +- while let Some(&ch) = chars.peek() { +- match ch { +- '0'..='9' | '_' => { +- chars.next(); +- len += 1; +- } +- '.' => { +- if has_dot { +- break; +- } +- chars.next(); +- if chars +- .peek() +- .map(|&ch| ch == '.' || is_ident_start(ch)) +- .unwrap_or(false) +- { +- return Err(LexError); +- } +- len += 1; +- has_dot = true; +- } +- 'e' | 'E' => { +- chars.next(); +- len += 1; +- has_exp = true; +- break; +- } +- _ => break, +- } +- } +- +- let rest = input.advance(len); +- if !(has_dot || has_exp || rest.starts_with("f32") || rest.starts_with("f64")) { +- return Err(LexError); +- } +- +- if has_exp { +- let mut has_exp_value = false; +- while let Some(&ch) = chars.peek() { +- match ch { +- '+' | '-' => { +- if has_exp_value { +- break; +- } +- chars.next(); +- len += 1; +- } +- '0'..='9' => { +- chars.next(); +- len += 1; +- has_exp_value = true; +- } +- '_' => { +- chars.next(); +- len += 1; +- } +- _ => break, +- } +- } +- if !has_exp_value { +- return Err(LexError); +- } +- } +- +- Ok((input.advance(len), ())) +-} +- +-fn int(input: Cursor) -> PResult<()> { +- let (mut rest, ()) = digits(input)?; +- if let Some(ch) = rest.chars().next() { +- if is_ident_start(ch) { +- rest = symbol_not_raw(rest)?.0; +- } +- } +- word_break(rest) +-} +- +-fn digits(mut input: Cursor) -> PResult<()> { +- let base = if input.starts_with("0x") { +- input = input.advance(2); +- 16 +- } else if input.starts_with("0o") { +- input = input.advance(2); +- 8 +- } else if input.starts_with("0b") { +- input = input.advance(2); +- 2 +- } else { +- 10 +- }; +- +- let mut len = 0; +- let mut empty = true; +- for b in input.bytes() { +- let digit = match b { +- b'0'..=b'9' => (b - b'0') as u64, +- b'a'..=b'f' => 10 + (b - b'a') as u64, +- b'A'..=b'F' => 10 + (b - b'A') as u64, +- b'_' => { +- if empty && base == 10 { +- return Err(LexError); +- } +- len += 1; +- continue; +- } +- _ => break, +- }; +- if digit >= base { +- return Err(LexError); +- } +- len += 1; +- empty = false; +- } +- if empty { +- Err(LexError) +- } else { +- Ok((input.advance(len), ())) +- } +-} +- +-fn op(input: Cursor) -> PResult<Punct> { +- let input = skip_whitespace(input); +- match op_char(input) { +- Ok((rest, '\'')) => { +- symbol(rest)?; +- Ok((rest, Punct::new('\'', Spacing::Joint))) +- } +- Ok((rest, ch)) => { +- let kind = match op_char(rest) { +- Ok(_) => Spacing::Joint, +- Err(LexError) => Spacing::Alone, +- }; +- Ok((rest, Punct::new(ch, kind))) +- } +- Err(LexError) => Err(LexError), +- } +-} +- +-fn op_char(input: Cursor) -> PResult<char> { +- if input.starts_with("//") || input.starts_with("/*") { +- // Do not accept `/` of a comment as an op. +- return Err(LexError); +- } +- +- let mut chars = input.chars(); +- let first = match chars.next() { +- Some(ch) => ch, +- None => { +- return Err(LexError); +- } +- }; +- let recognized = "~!@#$%^&*-=+|;:,<.>/?'"; +- if recognized.contains(first) { +- Ok((input.advance(first.len_utf8()), first)) +- } else { +- Err(LexError) +- } +-} +- +-fn doc_comment(input: Cursor) -> PResult<Vec<TokenTree>> { +- let mut trees = Vec::new(); +- let (rest, ((comment, inner), span)) = spanned(input, doc_comment_contents)?; +- trees.push(TokenTree::Punct(Punct::new('#', Spacing::Alone))); +- if inner { +- trees.push(Punct::new('!', Spacing::Alone).into()); +- } +- let mut stream = vec![ +- TokenTree::Ident(crate::Ident::new("doc", span)), +- TokenTree::Punct(Punct::new('=', Spacing::Alone)), +- TokenTree::Literal(crate::Literal::string(comment)), +- ]; +- for tt in stream.iter_mut() { +- tt.set_span(span); +- } +- let group = Group::new(Delimiter::Bracket, stream.into_iter().collect()); +- trees.push(crate::Group::_new_stable(group).into()); +- for tt in trees.iter_mut() { +- tt.set_span(span); +- } +- Ok((rest, trees)) +-} +- +-named!(doc_comment_contents -> (&str, bool), alt!( +- do_parse!( +- punct!("//!") >> +- s: take_until_newline_or_eof!() >> +- ((s, true)) +- ) +- | +- do_parse!( +- option!(whitespace) >> +- peek!(tag!("/*!")) >> +- s: block_comment >> +- ((s, true)) +- ) +- | +- do_parse!( +- punct!("///") >> +- not!(tag!("/")) >> +- s: take_until_newline_or_eof!() >> +- ((s, false)) +- ) +- | +- do_parse!( +- option!(whitespace) >> +- peek!(tuple!(tag!("/**"), not!(tag!("*")))) >> +- s: block_comment >> +- ((s, false)) +- ) +-)); +diff --git a/third_party/rust/proc-macro2/src/lib.rs b/third_party/rust/proc-macro2/src/lib.rs +index a08be3e815..c20fb50d4a 100644 +--- a/third_party/rust/proc-macro2/src/lib.rs ++++ b/third_party/rust/proc-macro2/src/lib.rs +@@ -78,27 +78,24 @@ + //! a different thread. + + // Proc-macro2 types in rustdoc of other crates get linked to here. +-#![doc(html_root_url = "https://docs.rs/proc-macro2/1.0.5")] ++#![doc(html_root_url = "https://docs.rs/proc-macro2/1.0.24")] + #![cfg_attr(any(proc_macro_span, super_unstable), feature(proc_macro_span))] + #![cfg_attr(super_unstable, feature(proc_macro_raw_ident, proc_macro_def_site))] ++#![allow(clippy::needless_doctest_main)] + + #[cfg(use_proc_macro)] + extern crate proc_macro; + +-use std::cmp::Ordering; +-use std::fmt; +-use std::hash::{Hash, Hasher}; +-use std::iter::FromIterator; +-use std::marker; +-use std::ops::RangeBounds; +-#[cfg(procmacro2_semver_exempt)] +-use std::path::PathBuf; +-use std::rc::Rc; +-use std::str::FromStr; ++mod marker; ++mod parse; ++ ++#[cfg(wrap_proc_macro)] ++mod detection; + +-#[macro_use] +-mod strnom; +-mod fallback; ++// Public for proc_macro2::fallback::force() and unforce(), but those are quite ++// a niche use case so we omit it from rustdoc. ++#[doc(hidden)] ++pub mod fallback; + + #[cfg(not(wrap_proc_macro))] + use crate::fallback as imp; +@@ -106,6 +103,17 @@ use crate::fallback as imp; + #[cfg(wrap_proc_macro)] + mod imp; + ++use crate::marker::Marker; ++use std::cmp::Ordering; ++use std::error::Error; ++use std::fmt::{self, Debug, Display}; ++use std::hash::{Hash, Hasher}; ++use std::iter::FromIterator; ++use std::ops::RangeBounds; ++#[cfg(procmacro2_semver_exempt)] ++use std::path::PathBuf; ++use std::str::FromStr; ++ + /// An abstract stream of tokens, or more concretely a sequence of token trees. + /// + /// This type provides interfaces for iterating over token trees and for +@@ -116,27 +124,27 @@ mod imp; + #[derive(Clone)] + pub struct TokenStream { + inner: imp::TokenStream, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + /// Error returned from `TokenStream::from_str`. + pub struct LexError { + inner: imp::LexError, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + impl TokenStream { + fn _new(inner: imp::TokenStream) -> TokenStream { + TokenStream { + inner, +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + + fn _new_stable(inner: fallback::TokenStream) -> TokenStream { + TokenStream { + inner: inner.into(), +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + +@@ -173,7 +181,7 @@ impl FromStr for TokenStream { + fn from_str(src: &str) -> Result<TokenStream, LexError> { + let e = src.parse().map_err(|e| LexError { + inner: e, +- _marker: marker::PhantomData, ++ _marker: Marker, + })?; + Ok(TokenStream::_new(e)) + } +@@ -228,25 +236,33 @@ impl FromIterator<TokenStream> for TokenStream { + /// convertible back into the same token stream (modulo spans), except for + /// possibly `TokenTree::Group`s with `Delimiter::None` delimiters and negative + /// numeric literals. +-impl fmt::Display for TokenStream { ++impl Display for TokenStream { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Display::fmt(&self.inner, f) + } + } + + /// Prints token in a form convenient for debugging. +-impl fmt::Debug for TokenStream { ++impl Debug for TokenStream { ++ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { ++ Debug::fmt(&self.inner, f) ++ } ++} ++ ++impl Debug for LexError { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +-impl fmt::Debug for LexError { ++impl Display for LexError { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Display::fmt(&self.inner, f) + } + } + ++impl Error for LexError {} ++ + /// The source file of a given `Span`. + /// + /// This type is semver exempt and not exposed by default. +@@ -254,7 +270,7 @@ impl fmt::Debug for LexError { + #[derive(Clone, PartialEq, Eq)] + pub struct SourceFile { + inner: imp::SourceFile, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + #[cfg(procmacro2_semver_exempt)] +@@ -262,7 +278,7 @@ impl SourceFile { + fn _new(inner: imp::SourceFile) -> Self { + SourceFile { + inner, +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + +@@ -291,9 +307,9 @@ impl SourceFile { + } + + #[cfg(procmacro2_semver_exempt)] +-impl fmt::Debug for SourceFile { ++impl Debug for SourceFile { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +@@ -311,25 +327,41 @@ pub struct LineColumn { + pub column: usize, + } + ++#[cfg(span_locations)] ++impl Ord for LineColumn { ++ fn cmp(&self, other: &Self) -> Ordering { ++ self.line ++ .cmp(&other.line) ++ .then(self.column.cmp(&other.column)) ++ } ++} ++ ++#[cfg(span_locations)] ++impl PartialOrd for LineColumn { ++ fn partial_cmp(&self, other: &Self) -> Option<Ordering> { ++ Some(self.cmp(other)) ++ } ++} ++ + /// A region of source code, along with macro expansion information. + #[derive(Copy, Clone)] + pub struct Span { + inner: imp::Span, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + impl Span { + fn _new(inner: imp::Span) -> Span { + Span { + inner, +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + + fn _new_stable(inner: fallback::Span) -> Span { + Span { + inner: inner.into(), +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + +@@ -342,6 +374,16 @@ impl Span { + Span::_new(imp::Span::call_site()) + } + ++ /// The span located at the invocation of the procedural macro, but with ++ /// local variables, labels, and `$crate` resolved at the definition site ++ /// of the macro. This is the same hygiene behavior as `macro_rules`. ++ /// ++ /// This function requires Rust 1.45 or later. ++ #[cfg(hygiene)] ++ pub fn mixed_site() -> Span { ++ Span::_new(imp::Span::mixed_site()) ++ } ++ + /// A span that resolves at the macro definition site. + /// + /// This method is semver exempt and not exposed by default. +@@ -352,18 +394,12 @@ impl Span { + + /// Creates a new span with the same line/column information as `self` but + /// that resolves symbols as though it were at `other`. +- /// +- /// This method is semver exempt and not exposed by default. +- #[cfg(procmacro2_semver_exempt)] + pub fn resolved_at(&self, other: Span) -> Span { + Span::_new(self.inner.resolved_at(other.inner)) + } + + /// Creates a new span with the same name resolution behavior as `self` but + /// with the line/column information of `other`. +- /// +- /// This method is semver exempt and not exposed by default. +- #[cfg(procmacro2_semver_exempt)] + pub fn located_at(&self, other: Span) -> Span { + Span::_new(self.inner.located_at(other.inner)) + } +@@ -439,9 +475,9 @@ impl Span { + } + + /// Prints a span in a form convenient for debugging. +-impl fmt::Debug for Span { ++impl Debug for Span { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +@@ -462,11 +498,11 @@ impl TokenTree { + /// Returns the span of this tree, delegating to the `span` method of + /// the contained token or a delimited stream. + pub fn span(&self) -> Span { +- match *self { +- TokenTree::Group(ref t) => t.span(), +- TokenTree::Ident(ref t) => t.span(), +- TokenTree::Punct(ref t) => t.span(), +- TokenTree::Literal(ref t) => t.span(), ++ match self { ++ TokenTree::Group(t) => t.span(), ++ TokenTree::Ident(t) => t.span(), ++ TokenTree::Punct(t) => t.span(), ++ TokenTree::Literal(t) => t.span(), + } + } + +@@ -476,11 +512,11 @@ impl TokenTree { + /// the span of each of the internal tokens, this will simply delegate to + /// the `set_span` method of each variant. + pub fn set_span(&mut self, span: Span) { +- match *self { +- TokenTree::Group(ref mut t) => t.set_span(span), +- TokenTree::Ident(ref mut t) => t.set_span(span), +- TokenTree::Punct(ref mut t) => t.set_span(span), +- TokenTree::Literal(ref mut t) => t.set_span(span), ++ match self { ++ TokenTree::Group(t) => t.set_span(span), ++ TokenTree::Ident(t) => t.set_span(span), ++ TokenTree::Punct(t) => t.set_span(span), ++ TokenTree::Literal(t) => t.set_span(span), + } + } + } +@@ -513,32 +549,32 @@ impl From<Literal> for TokenTree { + /// convertible back into the same token tree (modulo spans), except for + /// possibly `TokenTree::Group`s with `Delimiter::None` delimiters and negative + /// numeric literals. +-impl fmt::Display for TokenTree { ++impl Display for TokenTree { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- match *self { +- TokenTree::Group(ref t) => t.fmt(f), +- TokenTree::Ident(ref t) => t.fmt(f), +- TokenTree::Punct(ref t) => t.fmt(f), +- TokenTree::Literal(ref t) => t.fmt(f), ++ match self { ++ TokenTree::Group(t) => Display::fmt(t, f), ++ TokenTree::Ident(t) => Display::fmt(t, f), ++ TokenTree::Punct(t) => Display::fmt(t, f), ++ TokenTree::Literal(t) => Display::fmt(t, f), + } + } + } + + /// Prints token tree in a form convenient for debugging. +-impl fmt::Debug for TokenTree { ++impl Debug for TokenTree { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + // Each of these has the name in the struct type in the derived debug, + // so don't bother with an extra layer of indirection +- match *self { +- TokenTree::Group(ref t) => t.fmt(f), +- TokenTree::Ident(ref t) => { ++ match self { ++ TokenTree::Group(t) => Debug::fmt(t, f), ++ TokenTree::Ident(t) => { + let mut debug = f.debug_struct("Ident"); + debug.field("sym", &format_args!("{}", t)); + imp::debug_span_field_if_nontrivial(&mut debug, t.span().inner); + debug.finish() + } +- TokenTree::Punct(ref t) => t.fmt(f), +- TokenTree::Literal(ref t) => t.fmt(f), ++ TokenTree::Punct(t) => Debug::fmt(t, f), ++ TokenTree::Literal(t) => Debug::fmt(t, f), + } + } + } +@@ -651,15 +687,15 @@ impl Group { + /// Prints the group as a string that should be losslessly convertible back + /// into the same group (modulo spans), except for possibly `TokenTree::Group`s + /// with `Delimiter::None` delimiters. +-impl fmt::Display for Group { ++impl Display for Group { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { +- fmt::Display::fmt(&self.inner, formatter) ++ Display::fmt(&self.inner, formatter) + } + } + +-impl fmt::Debug for Group { ++impl Debug for Group { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { +- fmt::Debug::fmt(&self.inner, formatter) ++ Debug::fmt(&self.inner, formatter) + } + } + +@@ -669,7 +705,7 @@ impl fmt::Debug for Group { + /// `Punct` with different forms of `Spacing` returned. + #[derive(Clone)] + pub struct Punct { +- op: char, ++ ch: char, + spacing: Spacing, + span: Span, + } +@@ -695,9 +731,9 @@ impl Punct { + /// + /// The returned `Punct` will have the default span of `Span::call_site()` + /// which can be further configured with the `set_span` method below. +- pub fn new(op: char, spacing: Spacing) -> Punct { ++ pub fn new(ch: char, spacing: Spacing) -> Punct { + Punct { +- op, ++ ch, + spacing, + span: Span::call_site(), + } +@@ -705,7 +741,7 @@ impl Punct { + + /// Returns the value of this punctuation character as `char`. + pub fn as_char(&self) -> char { +- self.op ++ self.ch + } + + /// Returns the spacing of this punctuation character, indicating whether +@@ -730,16 +766,16 @@ impl Punct { + + /// Prints the punctuation character as a string that should be losslessly + /// convertible back into the same character. +-impl fmt::Display for Punct { ++impl Display for Punct { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.op.fmt(f) ++ Display::fmt(&self.ch, f) + } + } + +-impl fmt::Debug for Punct { ++impl Debug for Punct { + fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result { + let mut debug = fmt.debug_struct("Punct"); +- debug.field("op", &self.op); ++ debug.field("char", &self.ch); + debug.field("spacing", &self.spacing); + imp::debug_span_field_if_nontrivial(&mut debug, self.span.inner); + debug.finish() +@@ -813,14 +849,14 @@ impl fmt::Debug for Punct { + #[derive(Clone)] + pub struct Ident { + inner: imp::Ident, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + impl Ident { + fn _new(inner: imp::Ident) -> Ident { + Ident { + inner, +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + +@@ -920,15 +956,15 @@ impl Hash for Ident { + + /// Prints the identifier as a string that should be losslessly convertible back + /// into the same identifier. +-impl fmt::Display for Ident { ++impl Display for Ident { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Display::fmt(&self.inner, f) + } + } + +-impl fmt::Debug for Ident { ++impl Debug for Ident { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +@@ -941,7 +977,7 @@ impl fmt::Debug for Ident { + #[derive(Clone)] + pub struct Literal { + inner: imp::Literal, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + macro_rules! suffixed_int_literals { +@@ -988,14 +1024,14 @@ impl Literal { + fn _new(inner: imp::Literal) -> Literal { + Literal { + inner, +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + + fn _new_stable(inner: fallback::Literal) -> Literal { + Literal { + inner: inner.into(), +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + +@@ -1140,26 +1176,25 @@ impl Literal { + } + } + +-impl fmt::Debug for Literal { ++impl Debug for Literal { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +-impl fmt::Display for Literal { ++impl Display for Literal { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Display::fmt(&self.inner, f) + } + } + + /// Public implementation details for the `TokenStream` type, such as iterators. + pub mod token_stream { +- use std::fmt; +- use std::marker; +- use std::rc::Rc; ++ use crate::marker::Marker; ++ use crate::{imp, TokenTree}; ++ use std::fmt::{self, Debug}; + + pub use crate::TokenStream; +- use crate::{imp, TokenTree}; + + /// An iterator over `TokenStream`'s `TokenTree`s. + /// +@@ -1168,7 +1203,7 @@ pub mod token_stream { + #[derive(Clone)] + pub struct IntoIter { + inner: imp::TokenTreeIter, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + impl Iterator for IntoIter { +@@ -1179,9 +1214,9 @@ pub mod token_stream { + } + } + +- impl fmt::Debug for IntoIter { ++ impl Debug for IntoIter { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +@@ -1192,7 +1227,7 @@ pub mod token_stream { + fn into_iter(self) -> IntoIter { + IntoIter { + inner: self.inner.into_iter(), +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + } +diff --git a/third_party/rust/proc-macro2/src/marker.rs b/third_party/rust/proc-macro2/src/marker.rs +new file mode 100644 +index 0000000000..58729baf4a +--- /dev/null ++++ b/third_party/rust/proc-macro2/src/marker.rs +@@ -0,0 +1,18 @@ ++use std::marker::PhantomData; ++use std::panic::{RefUnwindSafe, UnwindSafe}; ++use std::rc::Rc; ++ ++// Zero sized marker with the correct set of autotrait impls we want all proc ++// macro types to have. ++pub(crate) type Marker = PhantomData<ProcMacroAutoTraits>; ++ ++pub(crate) use self::value::*; ++ ++mod value { ++ pub(crate) use std::marker::PhantomData as Marker; ++} ++ ++pub(crate) struct ProcMacroAutoTraits(Rc<()>); ++ ++impl UnwindSafe for ProcMacroAutoTraits {} ++impl RefUnwindSafe for ProcMacroAutoTraits {} +diff --git a/third_party/rust/proc-macro2/src/parse.rs b/third_party/rust/proc-macro2/src/parse.rs +new file mode 100644 +index 0000000000..365fe0484d +--- /dev/null ++++ b/third_party/rust/proc-macro2/src/parse.rs +@@ -0,0 +1,849 @@ ++use crate::fallback::{ ++ is_ident_continue, is_ident_start, Group, LexError, Literal, Span, TokenStream, ++}; ++use crate::{Delimiter, Punct, Spacing, TokenTree}; ++use std::char; ++use std::str::{Bytes, CharIndices, Chars}; ++ ++#[derive(Copy, Clone, Eq, PartialEq)] ++pub(crate) struct Cursor<'a> { ++ pub rest: &'a str, ++ #[cfg(span_locations)] ++ pub off: u32, ++} ++ ++impl<'a> Cursor<'a> { ++ fn advance(&self, bytes: usize) -> Cursor<'a> { ++ let (_front, rest) = self.rest.split_at(bytes); ++ Cursor { ++ rest, ++ #[cfg(span_locations)] ++ off: self.off + _front.chars().count() as u32, ++ } ++ } ++ ++ fn starts_with(&self, s: &str) -> bool { ++ self.rest.starts_with(s) ++ } ++ ++ pub(crate) fn is_empty(&self) -> bool { ++ self.rest.is_empty() ++ } ++ ++ fn len(&self) -> usize { ++ self.rest.len() ++ } ++ ++ fn as_bytes(&self) -> &'a [u8] { ++ self.rest.as_bytes() ++ } ++ ++ fn bytes(&self) -> Bytes<'a> { ++ self.rest.bytes() ++ } ++ ++ fn chars(&self) -> Chars<'a> { ++ self.rest.chars() ++ } ++ ++ fn char_indices(&self) -> CharIndices<'a> { ++ self.rest.char_indices() ++ } ++ ++ fn parse(&self, tag: &str) -> Result<Cursor<'a>, LexError> { ++ if self.starts_with(tag) { ++ Ok(self.advance(tag.len())) ++ } else { ++ Err(LexError) ++ } ++ } ++} ++ ++type PResult<'a, O> = Result<(Cursor<'a>, O), LexError>; ++ ++fn skip_whitespace(input: Cursor) -> Cursor { ++ let mut s = input; ++ ++ while !s.is_empty() { ++ let byte = s.as_bytes()[0]; ++ if byte == b'/' { ++ if s.starts_with("//") ++ && (!s.starts_with("///") || s.starts_with("////")) ++ && !s.starts_with("//!") ++ { ++ let (cursor, _) = take_until_newline_or_eof(s); ++ s = cursor; ++ continue; ++ } else if s.starts_with("/**/") { ++ s = s.advance(4); ++ continue; ++ } else if s.starts_with("/*") ++ && (!s.starts_with("/**") || s.starts_with("/***")) ++ && !s.starts_with("/*!") ++ { ++ match block_comment(s) { ++ Ok((rest, _)) => { ++ s = rest; ++ continue; ++ } ++ Err(LexError) => return s, ++ } ++ } ++ } ++ match byte { ++ b' ' | 0x09..=0x0d => { ++ s = s.advance(1); ++ continue; ++ } ++ b if b <= 0x7f => {} ++ _ => { ++ let ch = s.chars().next().unwrap(); ++ if is_whitespace(ch) { ++ s = s.advance(ch.len_utf8()); ++ continue; ++ } ++ } ++ } ++ return s; ++ } ++ s ++} ++ ++fn block_comment(input: Cursor) -> PResult<&str> { ++ if !input.starts_with("/*") { ++ return Err(LexError); ++ } ++ ++ let mut depth = 0; ++ let bytes = input.as_bytes(); ++ let mut i = 0; ++ let upper = bytes.len() - 1; ++ ++ while i < upper { ++ if bytes[i] == b'/' && bytes[i + 1] == b'*' { ++ depth += 1; ++ i += 1; // eat '*' ++ } else if bytes[i] == b'*' && bytes[i + 1] == b'/' { ++ depth -= 1; ++ if depth == 0 { ++ return Ok((input.advance(i + 2), &input.rest[..i + 2])); ++ } ++ i += 1; // eat '/' ++ } ++ i += 1; ++ } ++ ++ Err(LexError) ++} ++ ++fn is_whitespace(ch: char) -> bool { ++ // Rust treats left-to-right mark and right-to-left mark as whitespace ++ ch.is_whitespace() || ch == '\u{200e}' || ch == '\u{200f}' ++} ++ ++fn word_break(input: Cursor) -> Result<Cursor, LexError> { ++ match input.chars().next() { ++ Some(ch) if is_ident_continue(ch) => Err(LexError), ++ Some(_) | None => Ok(input), ++ } ++} ++ ++pub(crate) fn token_stream(mut input: Cursor) -> PResult<TokenStream> { ++ let mut trees = Vec::new(); ++ let mut stack = Vec::new(); ++ ++ loop { ++ input = skip_whitespace(input); ++ ++ if let Ok((rest, tt)) = doc_comment(input) { ++ trees.extend(tt); ++ input = rest; ++ continue; ++ } ++ ++ #[cfg(span_locations)] ++ let lo = input.off; ++ ++ let first = match input.bytes().next() { ++ Some(first) => first, ++ None => break, ++ }; ++ ++ if let Some(open_delimiter) = match first { ++ b'(' => Some(Delimiter::Parenthesis), ++ b'[' => Some(Delimiter::Bracket), ++ b'{' => Some(Delimiter::Brace), ++ _ => None, ++ } { ++ input = input.advance(1); ++ let frame = (open_delimiter, trees); ++ #[cfg(span_locations)] ++ let frame = (lo, frame); ++ stack.push(frame); ++ trees = Vec::new(); ++ } else if let Some(close_delimiter) = match first { ++ b')' => Some(Delimiter::Parenthesis), ++ b']' => Some(Delimiter::Bracket), ++ b'}' => Some(Delimiter::Brace), ++ _ => None, ++ } { ++ input = input.advance(1); ++ let frame = stack.pop().ok_or(LexError)?; ++ #[cfg(span_locations)] ++ let (lo, frame) = frame; ++ let (open_delimiter, outer) = frame; ++ if open_delimiter != close_delimiter { ++ return Err(LexError); ++ } ++ let mut g = Group::new(open_delimiter, TokenStream { inner: trees }); ++ g.set_span(Span { ++ #[cfg(span_locations)] ++ lo, ++ #[cfg(span_locations)] ++ hi: input.off, ++ }); ++ trees = outer; ++ trees.push(TokenTree::Group(crate::Group::_new_stable(g))); ++ } else { ++ let (rest, mut tt) = leaf_token(input)?; ++ tt.set_span(crate::Span::_new_stable(Span { ++ #[cfg(span_locations)] ++ lo, ++ #[cfg(span_locations)] ++ hi: rest.off, ++ })); ++ trees.push(tt); ++ input = rest; ++ } ++ } ++ ++ if stack.is_empty() { ++ Ok((input, TokenStream { inner: trees })) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn leaf_token(input: Cursor) -> PResult<TokenTree> { ++ if let Ok((input, l)) = literal(input) { ++ // must be parsed before ident ++ Ok((input, TokenTree::Literal(crate::Literal::_new_stable(l)))) ++ } else if let Ok((input, p)) = punct(input) { ++ Ok((input, TokenTree::Punct(p))) ++ } else if let Ok((input, i)) = ident(input) { ++ Ok((input, TokenTree::Ident(i))) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn ident(input: Cursor) -> PResult<crate::Ident> { ++ if ["r\"", "r#\"", "r##", "b\"", "b\'", "br\"", "br#"] ++ .iter() ++ .any(|prefix| input.starts_with(prefix)) ++ { ++ Err(LexError) ++ } else { ++ ident_any(input) ++ } ++} ++ ++fn ident_any(input: Cursor) -> PResult<crate::Ident> { ++ let raw = input.starts_with("r#"); ++ let rest = input.advance((raw as usize) << 1); ++ ++ let (rest, sym) = ident_not_raw(rest)?; ++ ++ if !raw { ++ let ident = crate::Ident::new(sym, crate::Span::call_site()); ++ return Ok((rest, ident)); ++ } ++ ++ if sym == "_" { ++ return Err(LexError); ++ } ++ ++ let ident = crate::Ident::_new_raw(sym, crate::Span::call_site()); ++ Ok((rest, ident)) ++} ++ ++fn ident_not_raw(input: Cursor) -> PResult<&str> { ++ let mut chars = input.char_indices(); ++ ++ match chars.next() { ++ Some((_, ch)) if is_ident_start(ch) => {} ++ _ => return Err(LexError), ++ } ++ ++ let mut end = input.len(); ++ for (i, ch) in chars { ++ if !is_ident_continue(ch) { ++ end = i; ++ break; ++ } ++ } ++ ++ Ok((input.advance(end), &input.rest[..end])) ++} ++ ++fn literal(input: Cursor) -> PResult<Literal> { ++ match literal_nocapture(input) { ++ Ok(a) => { ++ let end = input.len() - a.len(); ++ Ok((a, Literal::_new(input.rest[..end].to_string()))) ++ } ++ Err(LexError) => Err(LexError), ++ } ++} ++ ++fn literal_nocapture(input: Cursor) -> Result<Cursor, LexError> { ++ if let Ok(ok) = string(input) { ++ Ok(ok) ++ } else if let Ok(ok) = byte_string(input) { ++ Ok(ok) ++ } else if let Ok(ok) = byte(input) { ++ Ok(ok) ++ } else if let Ok(ok) = character(input) { ++ Ok(ok) ++ } else if let Ok(ok) = float(input) { ++ Ok(ok) ++ } else if let Ok(ok) = int(input) { ++ Ok(ok) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn literal_suffix(input: Cursor) -> Cursor { ++ match ident_not_raw(input) { ++ Ok((input, _)) => input, ++ Err(LexError) => input, ++ } ++} ++ ++fn string(input: Cursor) -> Result<Cursor, LexError> { ++ if let Ok(input) = input.parse("\"") { ++ cooked_string(input) ++ } else if let Ok(input) = input.parse("r") { ++ raw_string(input) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn cooked_string(input: Cursor) -> Result<Cursor, LexError> { ++ let mut chars = input.char_indices().peekable(); ++ ++ while let Some((i, ch)) = chars.next() { ++ match ch { ++ '"' => { ++ let input = input.advance(i + 1); ++ return Ok(literal_suffix(input)); ++ } ++ '\r' => match chars.next() { ++ Some((_, '\n')) => {} ++ _ => break, ++ }, ++ '\\' => match chars.next() { ++ Some((_, 'x')) => { ++ if !backslash_x_char(&mut chars) { ++ break; ++ } ++ } ++ Some((_, 'n')) | Some((_, 'r')) | Some((_, 't')) | Some((_, '\\')) ++ | Some((_, '\'')) | Some((_, '"')) | Some((_, '0')) => {} ++ Some((_, 'u')) => { ++ if !backslash_u(&mut chars) { ++ break; ++ } ++ } ++ Some((_, ch @ '\n')) | Some((_, ch @ '\r')) => { ++ let mut last = ch; ++ loop { ++ if last == '\r' && chars.next().map_or(true, |(_, ch)| ch != '\n') { ++ return Err(LexError); ++ } ++ match chars.peek() { ++ Some((_, ch)) if ch.is_whitespace() => { ++ last = *ch; ++ chars.next(); ++ } ++ _ => break, ++ } ++ } ++ } ++ _ => break, ++ }, ++ _ch => {} ++ } ++ } ++ Err(LexError) ++} ++ ++fn byte_string(input: Cursor) -> Result<Cursor, LexError> { ++ if let Ok(input) = input.parse("b\"") { ++ cooked_byte_string(input) ++ } else if let Ok(input) = input.parse("br") { ++ raw_string(input) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn cooked_byte_string(mut input: Cursor) -> Result<Cursor, LexError> { ++ let mut bytes = input.bytes().enumerate(); ++ while let Some((offset, b)) = bytes.next() { ++ match b { ++ b'"' => { ++ let input = input.advance(offset + 1); ++ return Ok(literal_suffix(input)); ++ } ++ b'\r' => match bytes.next() { ++ Some((_, b'\n')) => {} ++ _ => break, ++ }, ++ b'\\' => match bytes.next() { ++ Some((_, b'x')) => { ++ if !backslash_x_byte(&mut bytes) { ++ break; ++ } ++ } ++ Some((_, b'n')) | Some((_, b'r')) | Some((_, b't')) | Some((_, b'\\')) ++ | Some((_, b'0')) | Some((_, b'\'')) | Some((_, b'"')) => {} ++ Some((newline, b @ b'\n')) | Some((newline, b @ b'\r')) => { ++ let mut last = b as char; ++ let rest = input.advance(newline + 1); ++ let mut chars = rest.char_indices(); ++ loop { ++ if last == '\r' && chars.next().map_or(true, |(_, ch)| ch != '\n') { ++ return Err(LexError); ++ } ++ match chars.next() { ++ Some((_, ch)) if ch.is_whitespace() => last = ch, ++ Some((offset, _)) => { ++ input = rest.advance(offset); ++ bytes = input.bytes().enumerate(); ++ break; ++ } ++ None => return Err(LexError), ++ } ++ } ++ } ++ _ => break, ++ }, ++ b if b < 0x80 => {} ++ _ => break, ++ } ++ } ++ Err(LexError) ++} ++ ++fn raw_string(input: Cursor) -> Result<Cursor, LexError> { ++ let mut chars = input.char_indices(); ++ let mut n = 0; ++ while let Some((i, ch)) = chars.next() { ++ match ch { ++ '"' => { ++ n = i; ++ break; ++ } ++ '#' => {} ++ _ => return Err(LexError), ++ } ++ } ++ while let Some((i, ch)) = chars.next() { ++ match ch { ++ '"' if input.rest[i + 1..].starts_with(&input.rest[..n]) => { ++ let rest = input.advance(i + 1 + n); ++ return Ok(literal_suffix(rest)); ++ } ++ '\r' => match chars.next() { ++ Some((_, '\n')) => {} ++ _ => break, ++ }, ++ _ => {} ++ } ++ } ++ Err(LexError) ++} ++ ++fn byte(input: Cursor) -> Result<Cursor, LexError> { ++ let input = input.parse("b'")?; ++ let mut bytes = input.bytes().enumerate(); ++ let ok = match bytes.next().map(|(_, b)| b) { ++ Some(b'\\') => match bytes.next().map(|(_, b)| b) { ++ Some(b'x') => backslash_x_byte(&mut bytes), ++ Some(b'n') | Some(b'r') | Some(b't') | Some(b'\\') | Some(b'0') | Some(b'\'') ++ | Some(b'"') => true, ++ _ => false, ++ }, ++ b => b.is_some(), ++ }; ++ if !ok { ++ return Err(LexError); ++ } ++ let (offset, _) = bytes.next().ok_or(LexError)?; ++ if !input.chars().as_str().is_char_boundary(offset) { ++ return Err(LexError); ++ } ++ let input = input.advance(offset).parse("'")?; ++ Ok(literal_suffix(input)) ++} ++ ++fn character(input: Cursor) -> Result<Cursor, LexError> { ++ let input = input.parse("'")?; ++ let mut chars = input.char_indices(); ++ let ok = match chars.next().map(|(_, ch)| ch) { ++ Some('\\') => match chars.next().map(|(_, ch)| ch) { ++ Some('x') => backslash_x_char(&mut chars), ++ Some('u') => backslash_u(&mut chars), ++ Some('n') | Some('r') | Some('t') | Some('\\') | Some('0') | Some('\'') | Some('"') => { ++ true ++ } ++ _ => false, ++ }, ++ ch => ch.is_some(), ++ }; ++ if !ok { ++ return Err(LexError); ++ } ++ let (idx, _) = chars.next().ok_or(LexError)?; ++ let input = input.advance(idx).parse("'")?; ++ Ok(literal_suffix(input)) ++} ++ ++macro_rules! next_ch { ++ ($chars:ident @ $pat:pat $(| $rest:pat)*) => { ++ match $chars.next() { ++ Some((_, ch)) => match ch { ++ $pat $(| $rest)* => ch, ++ _ => return false, ++ }, ++ None => return false, ++ } ++ }; ++} ++ ++fn backslash_x_char<I>(chars: &mut I) -> bool ++where ++ I: Iterator<Item = (usize, char)>, ++{ ++ next_ch!(chars @ '0'..='7'); ++ next_ch!(chars @ '0'..='9' | 'a'..='f' | 'A'..='F'); ++ true ++} ++ ++fn backslash_x_byte<I>(chars: &mut I) -> bool ++where ++ I: Iterator<Item = (usize, u8)>, ++{ ++ next_ch!(chars @ b'0'..=b'9' | b'a'..=b'f' | b'A'..=b'F'); ++ next_ch!(chars @ b'0'..=b'9' | b'a'..=b'f' | b'A'..=b'F'); ++ true ++} ++ ++fn backslash_u<I>(chars: &mut I) -> bool ++where ++ I: Iterator<Item = (usize, char)>, ++{ ++ next_ch!(chars @ '{'); ++ let mut value = 0; ++ let mut len = 0; ++ while let Some((_, ch)) = chars.next() { ++ let digit = match ch { ++ '0'..='9' => ch as u8 - b'0', ++ 'a'..='f' => 10 + ch as u8 - b'a', ++ 'A'..='F' => 10 + ch as u8 - b'A', ++ '_' if len > 0 => continue, ++ '}' if len > 0 => return char::from_u32(value).is_some(), ++ _ => return false, ++ }; ++ if len == 6 { ++ return false; ++ } ++ value *= 0x10; ++ value += u32::from(digit); ++ len += 1; ++ } ++ false ++} ++ ++fn float(input: Cursor) -> Result<Cursor, LexError> { ++ let mut rest = float_digits(input)?; ++ if let Some(ch) = rest.chars().next() { ++ if is_ident_start(ch) { ++ rest = ident_not_raw(rest)?.0; ++ } ++ } ++ word_break(rest) ++} ++ ++fn float_digits(input: Cursor) -> Result<Cursor, LexError> { ++ let mut chars = input.chars().peekable(); ++ match chars.next() { ++ Some(ch) if ch >= '0' && ch <= '9' => {} ++ _ => return Err(LexError), ++ } ++ ++ let mut len = 1; ++ let mut has_dot = false; ++ let mut has_exp = false; ++ while let Some(&ch) = chars.peek() { ++ match ch { ++ '0'..='9' | '_' => { ++ chars.next(); ++ len += 1; ++ } ++ '.' => { ++ if has_dot { ++ break; ++ } ++ chars.next(); ++ if chars ++ .peek() ++ .map(|&ch| ch == '.' || is_ident_start(ch)) ++ .unwrap_or(false) ++ { ++ return Err(LexError); ++ } ++ len += 1; ++ has_dot = true; ++ } ++ 'e' | 'E' => { ++ chars.next(); ++ len += 1; ++ has_exp = true; ++ break; ++ } ++ _ => break, ++ } ++ } ++ ++ if !(has_dot || has_exp) { ++ return Err(LexError); ++ } ++ ++ if has_exp { ++ let token_before_exp = if has_dot { ++ Ok(input.advance(len - 1)) ++ } else { ++ Err(LexError) ++ }; ++ let mut has_sign = false; ++ let mut has_exp_value = false; ++ while let Some(&ch) = chars.peek() { ++ match ch { ++ '+' | '-' => { ++ if has_exp_value { ++ break; ++ } ++ if has_sign { ++ return token_before_exp; ++ } ++ chars.next(); ++ len += 1; ++ has_sign = true; ++ } ++ '0'..='9' => { ++ chars.next(); ++ len += 1; ++ has_exp_value = true; ++ } ++ '_' => { ++ chars.next(); ++ len += 1; ++ } ++ _ => break, ++ } ++ } ++ if !has_exp_value { ++ return token_before_exp; ++ } ++ } ++ ++ Ok(input.advance(len)) ++} ++ ++fn int(input: Cursor) -> Result<Cursor, LexError> { ++ let mut rest = digits(input)?; ++ if let Some(ch) = rest.chars().next() { ++ if is_ident_start(ch) { ++ rest = ident_not_raw(rest)?.0; ++ } ++ } ++ word_break(rest) ++} ++ ++fn digits(mut input: Cursor) -> Result<Cursor, LexError> { ++ let base = if input.starts_with("0x") { ++ input = input.advance(2); ++ 16 ++ } else if input.starts_with("0o") { ++ input = input.advance(2); ++ 8 ++ } else if input.starts_with("0b") { ++ input = input.advance(2); ++ 2 ++ } else { ++ 10 ++ }; ++ ++ let mut len = 0; ++ let mut empty = true; ++ for b in input.bytes() { ++ match b { ++ b'0'..=b'9' => { ++ let digit = (b - b'0') as u64; ++ if digit >= base { ++ return Err(LexError); ++ } ++ } ++ b'a'..=b'f' => { ++ let digit = 10 + (b - b'a') as u64; ++ if digit >= base { ++ break; ++ } ++ } ++ b'A'..=b'F' => { ++ let digit = 10 + (b - b'A') as u64; ++ if digit >= base { ++ break; ++ } ++ } ++ b'_' => { ++ if empty && base == 10 { ++ return Err(LexError); ++ } ++ len += 1; ++ continue; ++ } ++ _ => break, ++ }; ++ len += 1; ++ empty = false; ++ } ++ if empty { ++ Err(LexError) ++ } else { ++ Ok(input.advance(len)) ++ } ++} ++ ++fn punct(input: Cursor) -> PResult<Punct> { ++ match punct_char(input) { ++ Ok((rest, '\'')) => { ++ if ident_any(rest)?.0.starts_with("'") { ++ Err(LexError) ++ } else { ++ Ok((rest, Punct::new('\'', Spacing::Joint))) ++ } ++ } ++ Ok((rest, ch)) => { ++ let kind = match punct_char(rest) { ++ Ok(_) => Spacing::Joint, ++ Err(LexError) => Spacing::Alone, ++ }; ++ Ok((rest, Punct::new(ch, kind))) ++ } ++ Err(LexError) => Err(LexError), ++ } ++} ++ ++fn punct_char(input: Cursor) -> PResult<char> { ++ if input.starts_with("//") || input.starts_with("/*") { ++ // Do not accept `/` of a comment as a punct. ++ return Err(LexError); ++ } ++ ++ let mut chars = input.chars(); ++ let first = match chars.next() { ++ Some(ch) => ch, ++ None => { ++ return Err(LexError); ++ } ++ }; ++ let recognized = "~!@#$%^&*-=+|;:,<.>/?'"; ++ if recognized.contains(first) { ++ Ok((input.advance(first.len_utf8()), first)) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn doc_comment(input: Cursor) -> PResult<Vec<TokenTree>> { ++ #[cfg(span_locations)] ++ let lo = input.off; ++ let (rest, (comment, inner)) = doc_comment_contents(input)?; ++ let span = crate::Span::_new_stable(Span { ++ #[cfg(span_locations)] ++ lo, ++ #[cfg(span_locations)] ++ hi: rest.off, ++ }); ++ ++ let mut scan_for_bare_cr = comment; ++ while let Some(cr) = scan_for_bare_cr.find('\r') { ++ let rest = &scan_for_bare_cr[cr + 1..]; ++ if !rest.starts_with('\n') { ++ return Err(LexError); ++ } ++ scan_for_bare_cr = rest; ++ } ++ ++ let mut trees = Vec::new(); ++ trees.push(TokenTree::Punct(Punct::new('#', Spacing::Alone))); ++ if inner { ++ trees.push(Punct::new('!', Spacing::Alone).into()); ++ } ++ let mut stream = vec![ ++ TokenTree::Ident(crate::Ident::new("doc", span)), ++ TokenTree::Punct(Punct::new('=', Spacing::Alone)), ++ TokenTree::Literal(crate::Literal::string(comment)), ++ ]; ++ for tt in stream.iter_mut() { ++ tt.set_span(span); ++ } ++ let group = Group::new(Delimiter::Bracket, stream.into_iter().collect()); ++ trees.push(crate::Group::_new_stable(group).into()); ++ for tt in trees.iter_mut() { ++ tt.set_span(span); ++ } ++ Ok((rest, trees)) ++} ++ ++fn doc_comment_contents(input: Cursor) -> PResult<(&str, bool)> { ++ if input.starts_with("//!") { ++ let input = input.advance(3); ++ let (input, s) = take_until_newline_or_eof(input); ++ Ok((input, (s, true))) ++ } else if input.starts_with("/*!") { ++ let (input, s) = block_comment(input)?; ++ Ok((input, (&s[3..s.len() - 2], true))) ++ } else if input.starts_with("///") { ++ let input = input.advance(3); ++ if input.starts_with("/") { ++ return Err(LexError); ++ } ++ let (input, s) = take_until_newline_or_eof(input); ++ Ok((input, (s, false))) ++ } else if input.starts_with("/**") && !input.rest[3..].starts_with('*') { ++ let (input, s) = block_comment(input)?; ++ Ok((input, (&s[3..s.len() - 2], false))) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn take_until_newline_or_eof(input: Cursor) -> (Cursor, &str) { ++ let chars = input.char_indices(); ++ ++ for (i, ch) in chars { ++ if ch == '\n' { ++ return (input.advance(i), &input.rest[..i]); ++ } else if ch == '\r' && input.rest[i + 1..].starts_with('\n') { ++ return (input.advance(i + 1), &input.rest[..i]); ++ } ++ } ++ ++ (input.advance(input.len()), input.rest) ++} +diff --git a/third_party/rust/proc-macro2/src/strnom.rs b/third_party/rust/proc-macro2/src/strnom.rs +deleted file mode 100644 +index eb7d0b8a8e..0000000000 +--- a/third_party/rust/proc-macro2/src/strnom.rs ++++ /dev/null +@@ -1,391 +0,0 @@ +-//! Adapted from [`nom`](https://github.com/Geal/nom). +- +-use crate::fallback::LexError; +-use std::str::{Bytes, CharIndices, Chars}; +-use unicode_xid::UnicodeXID; +- +-#[derive(Copy, Clone, Eq, PartialEq)] +-pub struct Cursor<'a> { +- pub rest: &'a str, +- #[cfg(span_locations)] +- pub off: u32, +-} +- +-impl<'a> Cursor<'a> { +- #[cfg(not(span_locations))] +- pub fn advance(&self, amt: usize) -> Cursor<'a> { +- Cursor { +- rest: &self.rest[amt..], +- } +- } +- #[cfg(span_locations)] +- pub fn advance(&self, amt: usize) -> Cursor<'a> { +- Cursor { +- rest: &self.rest[amt..], +- off: self.off + (amt as u32), +- } +- } +- +- pub fn find(&self, p: char) -> Option<usize> { +- self.rest.find(p) +- } +- +- pub fn starts_with(&self, s: &str) -> bool { +- self.rest.starts_with(s) +- } +- +- pub fn is_empty(&self) -> bool { +- self.rest.is_empty() +- } +- +- pub fn len(&self) -> usize { +- self.rest.len() +- } +- +- pub fn as_bytes(&self) -> &'a [u8] { +- self.rest.as_bytes() +- } +- +- pub fn bytes(&self) -> Bytes<'a> { +- self.rest.bytes() +- } +- +- pub fn chars(&self) -> Chars<'a> { +- self.rest.chars() +- } +- +- pub fn char_indices(&self) -> CharIndices<'a> { +- self.rest.char_indices() +- } +-} +- +-pub type PResult<'a, O> = Result<(Cursor<'a>, O), LexError>; +- +-pub fn whitespace(input: Cursor) -> PResult<()> { +- if input.is_empty() { +- return Err(LexError); +- } +- +- let bytes = input.as_bytes(); +- let mut i = 0; +- while i < bytes.len() { +- let s = input.advance(i); +- if bytes[i] == b'/' { +- if s.starts_with("//") +- && (!s.starts_with("///") || s.starts_with("////")) +- && !s.starts_with("//!") +- { +- if let Some(len) = s.find('\n') { +- i += len + 1; +- continue; +- } +- break; +- } else if s.starts_with("/**/") { +- i += 4; +- continue; +- } else if s.starts_with("/*") +- && (!s.starts_with("/**") || s.starts_with("/***")) +- && !s.starts_with("/*!") +- { +- let (_, com) = block_comment(s)?; +- i += com.len(); +- continue; +- } +- } +- match bytes[i] { +- b' ' | 0x09..=0x0d => { +- i += 1; +- continue; +- } +- b if b <= 0x7f => {} +- _ => { +- let ch = s.chars().next().unwrap(); +- if is_whitespace(ch) { +- i += ch.len_utf8(); +- continue; +- } +- } +- } +- return if i > 0 { Ok((s, ())) } else { Err(LexError) }; +- } +- Ok((input.advance(input.len()), ())) +-} +- +-pub fn block_comment(input: Cursor) -> PResult<&str> { +- if !input.starts_with("/*") { +- return Err(LexError); +- } +- +- let mut depth = 0; +- let bytes = input.as_bytes(); +- let mut i = 0; +- let upper = bytes.len() - 1; +- while i < upper { +- if bytes[i] == b'/' && bytes[i + 1] == b'*' { +- depth += 1; +- i += 1; // eat '*' +- } else if bytes[i] == b'*' && bytes[i + 1] == b'/' { +- depth -= 1; +- if depth == 0 { +- return Ok((input.advance(i + 2), &input.rest[..i + 2])); +- } +- i += 1; // eat '/' +- } +- i += 1; +- } +- Err(LexError) +-} +- +-pub fn skip_whitespace(input: Cursor) -> Cursor { +- match whitespace(input) { +- Ok((rest, _)) => rest, +- Err(LexError) => input, +- } +-} +- +-fn is_whitespace(ch: char) -> bool { +- // Rust treats left-to-right mark and right-to-left mark as whitespace +- ch.is_whitespace() || ch == '\u{200e}' || ch == '\u{200f}' +-} +- +-pub fn word_break(input: Cursor) -> PResult<()> { +- match input.chars().next() { +- Some(ch) if UnicodeXID::is_xid_continue(ch) => Err(LexError), +- Some(_) | None => Ok((input, ())), +- } +-} +- +-macro_rules! named { +- ($name:ident -> $o:ty, $submac:ident!( $($args:tt)* )) => { +- fn $name<'a>(i: Cursor<'a>) -> $crate::strnom::PResult<'a, $o> { +- $submac!(i, $($args)*) +- } +- }; +-} +- +-macro_rules! alt { +- ($i:expr, $e:ident | $($rest:tt)*) => { +- alt!($i, call!($e) | $($rest)*) +- }; +- +- ($i:expr, $subrule:ident!( $($args:tt)*) | $($rest:tt)*) => { +- match $subrule!($i, $($args)*) { +- res @ Ok(_) => res, +- _ => alt!($i, $($rest)*) +- } +- }; +- +- ($i:expr, $subrule:ident!( $($args:tt)* ) => { $gen:expr } | $($rest:tt)+) => { +- match $subrule!($i, $($args)*) { +- Ok((i, o)) => Ok((i, $gen(o))), +- Err(LexError) => alt!($i, $($rest)*) +- } +- }; +- +- ($i:expr, $e:ident => { $gen:expr } | $($rest:tt)*) => { +- alt!($i, call!($e) => { $gen } | $($rest)*) +- }; +- +- ($i:expr, $e:ident => { $gen:expr }) => { +- alt!($i, call!($e) => { $gen }) +- }; +- +- ($i:expr, $subrule:ident!( $($args:tt)* ) => { $gen:expr }) => { +- match $subrule!($i, $($args)*) { +- Ok((i, o)) => Ok((i, $gen(o))), +- Err(LexError) => Err(LexError), +- } +- }; +- +- ($i:expr, $e:ident) => { +- alt!($i, call!($e)) +- }; +- +- ($i:expr, $subrule:ident!( $($args:tt)*)) => { +- $subrule!($i, $($args)*) +- }; +-} +- +-macro_rules! do_parse { +- ($i:expr, ( $($rest:expr),* )) => { +- Ok(($i, ( $($rest),* ))) +- }; +- +- ($i:expr, $e:ident >> $($rest:tt)*) => { +- do_parse!($i, call!($e) >> $($rest)*) +- }; +- +- ($i:expr, $submac:ident!( $($args:tt)* ) >> $($rest:tt)*) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, _)) => do_parse!(i, $($rest)*), +- } +- }; +- +- ($i:expr, $field:ident : $e:ident >> $($rest:tt)*) => { +- do_parse!($i, $field: call!($e) >> $($rest)*) +- }; +- +- ($i:expr, $field:ident : $submac:ident!( $($args:tt)* ) >> $($rest:tt)*) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, o)) => { +- let $field = o; +- do_parse!(i, $($rest)*) +- }, +- } +- }; +-} +- +-macro_rules! peek { +- ($i:expr, $submac:ident!( $($args:tt)* )) => { +- match $submac!($i, $($args)*) { +- Ok((_, o)) => Ok(($i, o)), +- Err(LexError) => Err(LexError), +- } +- }; +-} +- +-macro_rules! call { +- ($i:expr, $fun:expr $(, $args:expr)*) => { +- $fun($i $(, $args)*) +- }; +-} +- +-macro_rules! option { +- ($i:expr, $f:expr) => { +- match $f($i) { +- Ok((i, o)) => Ok((i, Some(o))), +- Err(LexError) => Ok(($i, None)), +- } +- }; +-} +- +-macro_rules! take_until_newline_or_eof { +- ($i:expr,) => {{ +- if $i.len() == 0 { +- Ok(($i, "")) +- } else { +- match $i.find('\n') { +- Some(i) => Ok(($i.advance(i), &$i.rest[..i])), +- None => Ok(($i.advance($i.len()), &$i.rest[..$i.len()])), +- } +- } +- }}; +-} +- +-macro_rules! tuple { +- ($i:expr, $($rest:tt)*) => { +- tuple_parser!($i, (), $($rest)*) +- }; +-} +- +-/// Do not use directly. Use `tuple!`. +-macro_rules! tuple_parser { +- ($i:expr, ($($parsed:tt),*), $e:ident, $($rest:tt)*) => { +- tuple_parser!($i, ($($parsed),*), call!($e), $($rest)*) +- }; +- +- ($i:expr, (), $submac:ident!( $($args:tt)* ), $($rest:tt)*) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, o)) => tuple_parser!(i, (o), $($rest)*), +- } +- }; +- +- ($i:expr, ($($parsed:tt)*), $submac:ident!( $($args:tt)* ), $($rest:tt)*) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, o)) => tuple_parser!(i, ($($parsed)* , o), $($rest)*), +- } +- }; +- +- ($i:expr, ($($parsed:tt),*), $e:ident) => { +- tuple_parser!($i, ($($parsed),*), call!($e)) +- }; +- +- ($i:expr, (), $submac:ident!( $($args:tt)* )) => { +- $submac!($i, $($args)*) +- }; +- +- ($i:expr, ($($parsed:expr),*), $submac:ident!( $($args:tt)* )) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, o)) => Ok((i, ($($parsed),*, o))) +- } +- }; +- +- ($i:expr, ($($parsed:expr),*)) => { +- Ok(($i, ($($parsed),*))) +- }; +-} +- +-macro_rules! not { +- ($i:expr, $submac:ident!( $($args:tt)* )) => { +- match $submac!($i, $($args)*) { +- Ok((_, _)) => Err(LexError), +- Err(LexError) => Ok(($i, ())), +- } +- }; +-} +- +-macro_rules! tag { +- ($i:expr, $tag:expr) => { +- if $i.starts_with($tag) { +- Ok(($i.advance($tag.len()), &$i.rest[..$tag.len()])) +- } else { +- Err(LexError) +- } +- }; +-} +- +-macro_rules! punct { +- ($i:expr, $punct:expr) => { +- $crate::strnom::punct($i, $punct) +- }; +-} +- +-/// Do not use directly. Use `punct!`. +-pub fn punct<'a>(input: Cursor<'a>, token: &'static str) -> PResult<'a, &'a str> { +- let input = skip_whitespace(input); +- if input.starts_with(token) { +- Ok((input.advance(token.len()), token)) +- } else { +- Err(LexError) +- } +-} +- +-macro_rules! preceded { +- ($i:expr, $submac:ident!( $($args:tt)* ), $submac2:ident!( $($args2:tt)* )) => { +- match tuple!($i, $submac!($($args)*), $submac2!($($args2)*)) { +- Ok((remaining, (_, o))) => Ok((remaining, o)), +- Err(LexError) => Err(LexError), +- } +- }; +- +- ($i:expr, $submac:ident!( $($args:tt)* ), $g:expr) => { +- preceded!($i, $submac!($($args)*), call!($g)) +- }; +-} +- +-macro_rules! delimited { +- ($i:expr, $submac:ident!( $($args:tt)* ), $($rest:tt)+) => { +- match tuple_parser!($i, (), $submac!($($args)*), $($rest)*) { +- Err(LexError) => Err(LexError), +- Ok((i1, (_, o, _))) => Ok((i1, o)) +- } +- }; +-} +- +-macro_rules! map { +- ($i:expr, $submac:ident!( $($args:tt)* ), $g:expr) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, o)) => Ok((i, call!(o, $g))) +- } +- }; +- +- ($i:expr, $f:expr, $g:expr) => { +- map!($i, call!($f), $g) +- }; +-} +diff --git a/third_party/rust/proc-macro2/src/wrapper.rs b/third_party/rust/proc-macro2/src/wrapper.rs +index 552b9381cf..3df044af17 100644 +--- a/third_party/rust/proc-macro2/src/wrapper.rs ++++ b/third_party/rust/proc-macro2/src/wrapper.rs +@@ -1,15 +1,15 @@ +-use std::fmt; +-use std::iter; ++use crate::detection::inside_proc_macro; ++use crate::{fallback, Delimiter, Punct, Spacing, TokenTree}; ++use std::fmt::{self, Debug, Display}; ++use std::iter::FromIterator; + use std::ops::RangeBounds; +-use std::panic::{self, PanicInfo}; ++use std::panic; + #[cfg(super_unstable)] + use std::path::PathBuf; + use std::str::FromStr; + +-use crate::{fallback, Delimiter, Punct, Spacing, TokenTree}; +- + #[derive(Clone)] +-pub enum TokenStream { ++pub(crate) enum TokenStream { + Compiler(DeferredTokenStream), + Fallback(fallback::TokenStream), + } +@@ -19,73 +19,16 @@ pub enum TokenStream { + // we hold on to the appended tokens and do proc_macro::TokenStream::extend as + // late as possible to batch together consecutive uses of the Extend impl. + #[derive(Clone)] +-pub struct DeferredTokenStream { ++pub(crate) struct DeferredTokenStream { + stream: proc_macro::TokenStream, + extra: Vec<proc_macro::TokenTree>, + } + +-pub enum LexError { ++pub(crate) enum LexError { + Compiler(proc_macro::LexError), + Fallback(fallback::LexError), + } + +-fn nightly_works() -> bool { +- use std::sync::atomic::*; +- use std::sync::Once; +- +- static WORKS: AtomicUsize = AtomicUsize::new(0); +- static INIT: Once = Once::new(); +- +- match WORKS.load(Ordering::SeqCst) { +- 1 => return false, +- 2 => return true, +- _ => {} +- } +- +- // Swap in a null panic hook to avoid printing "thread panicked" to stderr, +- // then use catch_unwind to determine whether the compiler's proc_macro is +- // working. When proc-macro2 is used from outside of a procedural macro all +- // of the proc_macro crate's APIs currently panic. +- // +- // The Once is to prevent the possibility of this ordering: +- // +- // thread 1 calls take_hook, gets the user's original hook +- // thread 1 calls set_hook with the null hook +- // thread 2 calls take_hook, thinks null hook is the original hook +- // thread 2 calls set_hook with the null hook +- // thread 1 calls set_hook with the actual original hook +- // thread 2 calls set_hook with what it thinks is the original hook +- // +- // in which the user's hook has been lost. +- // +- // There is still a race condition where a panic in a different thread can +- // happen during the interval that the user's original panic hook is +- // unregistered such that their hook is incorrectly not called. This is +- // sufficiently unlikely and less bad than printing panic messages to stderr +- // on correct use of this crate. Maybe there is a libstd feature request +- // here. For now, if a user needs to guarantee that this failure mode does +- // not occur, they need to call e.g. `proc_macro2::Span::call_site()` from +- // the main thread before launching any other threads. +- INIT.call_once(|| { +- type PanicHook = dyn Fn(&PanicInfo) + Sync + Send + 'static; +- +- let null_hook: Box<PanicHook> = Box::new(|_panic_info| { /* ignore */ }); +- let sanity_check = &*null_hook as *const PanicHook; +- let original_hook = panic::take_hook(); +- panic::set_hook(null_hook); +- +- let works = panic::catch_unwind(|| proc_macro::Span::call_site()).is_ok(); +- WORKS.store(works as usize + 1, Ordering::SeqCst); +- +- let hopefully_null_hook = panic::take_hook(); +- panic::set_hook(original_hook); +- if sanity_check != &*hopefully_null_hook { +- panic!("observed race condition in proc_macro2::nightly_works"); +- } +- }); +- nightly_works() +-} +- + fn mismatch() -> ! { + panic!("stable/nightly mismatch") + } +@@ -103,7 +46,12 @@ impl DeferredTokenStream { + } + + fn evaluate_now(&mut self) { +- self.stream.extend(self.extra.drain(..)); ++ // If-check provides a fast short circuit for the common case of `extra` ++ // being empty, which saves a round trip over the proc macro bridge. ++ // Improves macro expansion time in winrt by 6% in debug mode. ++ if !self.extra.is_empty() { ++ self.stream.extend(self.extra.drain(..)); ++ } + } + + fn into_token_stream(mut self) -> proc_macro::TokenStream { +@@ -114,7 +62,7 @@ impl DeferredTokenStream { + + impl TokenStream { + pub fn new() -> TokenStream { +- if nightly_works() { ++ if inside_proc_macro() { + TokenStream::Compiler(DeferredTokenStream::new(proc_macro::TokenStream::new())) + } else { + TokenStream::Fallback(fallback::TokenStream::new()) +@@ -147,9 +95,9 @@ impl FromStr for TokenStream { + type Err = LexError; + + fn from_str(src: &str) -> Result<TokenStream, LexError> { +- if nightly_works() { ++ if inside_proc_macro() { + Ok(TokenStream::Compiler(DeferredTokenStream::new( +- src.parse()?, ++ proc_macro_parse(src)?, + ))) + } else { + Ok(TokenStream::Fallback(src.parse()?)) +@@ -157,11 +105,17 @@ impl FromStr for TokenStream { + } + } + +-impl fmt::Display for TokenStream { ++// Work around https://github.com/rust-lang/rust/issues/58736. ++fn proc_macro_parse(src: &str) -> Result<proc_macro::TokenStream, LexError> { ++ panic::catch_unwind(|| src.parse().map_err(LexError::Compiler)) ++ .unwrap_or(Err(LexError::Fallback(fallback::LexError))) ++} ++ ++impl Display for TokenStream { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- TokenStream::Compiler(tts) => tts.clone().into_token_stream().fmt(f), +- TokenStream::Fallback(tts) => tts.fmt(f), ++ TokenStream::Compiler(tts) => Display::fmt(&tts.clone().into_token_stream(), f), ++ TokenStream::Fallback(tts) => Display::fmt(tts, f), + } + } + } +@@ -187,7 +141,7 @@ impl From<fallback::TokenStream> for TokenStream { + } + } + +-// Assumes nightly_works(). ++// Assumes inside_proc_macro(). + fn into_compiler_token(token: TokenTree) -> proc_macro::TokenTree { + match token { + TokenTree::Group(tt) => tt.inner.unwrap_nightly().into(), +@@ -196,9 +150,9 @@ fn into_compiler_token(token: TokenTree) -> proc_macro::TokenTree { + Spacing::Joint => proc_macro::Spacing::Joint, + Spacing::Alone => proc_macro::Spacing::Alone, + }; +- let mut op = proc_macro::Punct::new(tt.as_char(), spacing); +- op.set_span(tt.span().inner.unwrap_nightly()); +- op.into() ++ let mut punct = proc_macro::Punct::new(tt.as_char(), spacing); ++ punct.set_span(tt.span().inner.unwrap_nightly()); ++ punct.into() + } + TokenTree::Ident(tt) => tt.inner.unwrap_nightly().into(), + TokenTree::Literal(tt) => tt.inner.unwrap_nightly().into(), +@@ -207,7 +161,7 @@ fn into_compiler_token(token: TokenTree) -> proc_macro::TokenTree { + + impl From<TokenTree> for TokenStream { + fn from(token: TokenTree) -> TokenStream { +- if nightly_works() { ++ if inside_proc_macro() { + TokenStream::Compiler(DeferredTokenStream::new(into_compiler_token(token).into())) + } else { + TokenStream::Fallback(token.into()) +@@ -215,9 +169,9 @@ impl From<TokenTree> for TokenStream { + } + } + +-impl iter::FromIterator<TokenTree> for TokenStream { ++impl FromIterator<TokenTree> for TokenStream { + fn from_iter<I: IntoIterator<Item = TokenTree>>(trees: I) -> Self { +- if nightly_works() { ++ if inside_proc_macro() { + TokenStream::Compiler(DeferredTokenStream::new( + trees.into_iter().map(into_compiler_token).collect(), + )) +@@ -227,7 +181,7 @@ impl iter::FromIterator<TokenTree> for TokenStream { + } + } + +-impl iter::FromIterator<TokenStream> for TokenStream { ++impl FromIterator<TokenStream> for TokenStream { + fn from_iter<I: IntoIterator<Item = TokenStream>>(streams: I) -> Self { + let mut streams = streams.into_iter(); + match streams.next() { +@@ -252,14 +206,15 @@ impl iter::FromIterator<TokenStream> for TokenStream { + } + + impl Extend<TokenTree> for TokenStream { +- fn extend<I: IntoIterator<Item = TokenTree>>(&mut self, streams: I) { ++ fn extend<I: IntoIterator<Item = TokenTree>>(&mut self, stream: I) { + match self { + TokenStream::Compiler(tts) => { + // Here is the reason for DeferredTokenStream. +- tts.extra +- .extend(streams.into_iter().map(into_compiler_token)); ++ for token in stream { ++ tts.extra.push(into_compiler_token(token)); ++ } + } +- TokenStream::Fallback(tts) => tts.extend(streams), ++ TokenStream::Fallback(tts) => tts.extend(stream), + } + } + } +@@ -270,20 +225,20 @@ impl Extend<TokenStream> for TokenStream { + TokenStream::Compiler(tts) => { + tts.evaluate_now(); + tts.stream +- .extend(streams.into_iter().map(|stream| stream.unwrap_nightly())); ++ .extend(streams.into_iter().map(TokenStream::unwrap_nightly)); + } + TokenStream::Fallback(tts) => { +- tts.extend(streams.into_iter().map(|stream| stream.unwrap_stable())); ++ tts.extend(streams.into_iter().map(TokenStream::unwrap_stable)); + } + } + } + } + +-impl fmt::Debug for TokenStream { ++impl Debug for TokenStream { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- TokenStream::Compiler(tts) => tts.clone().into_token_stream().fmt(f), +- TokenStream::Fallback(tts) => tts.fmt(f), ++ TokenStream::Compiler(tts) => Debug::fmt(&tts.clone().into_token_stream(), f), ++ TokenStream::Fallback(tts) => Debug::fmt(tts, f), + } + } + } +@@ -300,17 +255,29 @@ impl From<fallback::LexError> for LexError { + } + } + +-impl fmt::Debug for LexError { ++impl Debug for LexError { ++ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ LexError::Compiler(e) => Debug::fmt(e, f), ++ LexError::Fallback(e) => Debug::fmt(e, f), ++ } ++ } ++} ++ ++impl Display for LexError { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- LexError::Compiler(e) => e.fmt(f), +- LexError::Fallback(e) => e.fmt(f), ++ #[cfg(lexerror_display)] ++ LexError::Compiler(e) => Display::fmt(e, f), ++ #[cfg(not(lexerror_display))] ++ LexError::Compiler(_e) => Display::fmt(&fallback::LexError, f), ++ LexError::Fallback(e) => Display::fmt(e, f), + } + } + } + + #[derive(Clone)] +-pub enum TokenTreeIter { ++pub(crate) enum TokenTreeIter { + Compiler(proc_macro::token_stream::IntoIter), + Fallback(fallback::TokenTreeIter), + } +@@ -361,7 +328,7 @@ impl Iterator for TokenTreeIter { + } + } + +-impl fmt::Debug for TokenTreeIter { ++impl Debug for TokenTreeIter { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + f.debug_struct("TokenTreeIter").finish() + } +@@ -369,7 +336,7 @@ impl fmt::Debug for TokenTreeIter { + + #[derive(Clone, PartialEq, Eq)] + #[cfg(super_unstable)] +-pub enum SourceFile { ++pub(crate) enum SourceFile { + Compiler(proc_macro::SourceFile), + Fallback(fallback::SourceFile), + } +@@ -397,58 +364,77 @@ impl SourceFile { + } + + #[cfg(super_unstable)] +-impl fmt::Debug for SourceFile { ++impl Debug for SourceFile { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- SourceFile::Compiler(a) => a.fmt(f), +- SourceFile::Fallback(a) => a.fmt(f), ++ SourceFile::Compiler(a) => Debug::fmt(a, f), ++ SourceFile::Fallback(a) => Debug::fmt(a, f), + } + } + } + + #[cfg(any(super_unstable, feature = "span-locations"))] +-pub struct LineColumn { ++pub(crate) struct LineColumn { + pub line: usize, + pub column: usize, + } + + #[derive(Copy, Clone)] +-pub enum Span { ++pub(crate) enum Span { + Compiler(proc_macro::Span), + Fallback(fallback::Span), + } + + impl Span { + pub fn call_site() -> Span { +- if nightly_works() { ++ if inside_proc_macro() { + Span::Compiler(proc_macro::Span::call_site()) + } else { + Span::Fallback(fallback::Span::call_site()) + } + } + ++ #[cfg(hygiene)] ++ pub fn mixed_site() -> Span { ++ if inside_proc_macro() { ++ Span::Compiler(proc_macro::Span::mixed_site()) ++ } else { ++ Span::Fallback(fallback::Span::mixed_site()) ++ } ++ } ++ + #[cfg(super_unstable)] + pub fn def_site() -> Span { +- if nightly_works() { ++ if inside_proc_macro() { + Span::Compiler(proc_macro::Span::def_site()) + } else { + Span::Fallback(fallback::Span::def_site()) + } + } + +- #[cfg(super_unstable)] + pub fn resolved_at(&self, other: Span) -> Span { + match (self, other) { ++ #[cfg(hygiene)] + (Span::Compiler(a), Span::Compiler(b)) => Span::Compiler(a.resolved_at(b)), ++ ++ // Name resolution affects semantics, but location is only cosmetic ++ #[cfg(not(hygiene))] ++ (Span::Compiler(_), Span::Compiler(_)) => other, ++ + (Span::Fallback(a), Span::Fallback(b)) => Span::Fallback(a.resolved_at(b)), + _ => mismatch(), + } + } + +- #[cfg(super_unstable)] + pub fn located_at(&self, other: Span) -> Span { + match (self, other) { ++ #[cfg(hygiene)] + (Span::Compiler(a), Span::Compiler(b)) => Span::Compiler(a.located_at(b)), ++ ++ // Name resolution affects semantics, but location is only cosmetic ++ #[cfg(not(hygiene))] ++ (Span::Compiler(_), Span::Compiler(_)) => *self, ++ + (Span::Fallback(a), Span::Fallback(b)) => Span::Fallback(a.located_at(b)), + _ => mismatch(), + } +@@ -542,16 +528,16 @@ impl From<fallback::Span> for Span { + } + } + +-impl fmt::Debug for Span { ++impl Debug for Span { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- Span::Compiler(s) => s.fmt(f), +- Span::Fallback(s) => s.fmt(f), ++ Span::Compiler(s) => Debug::fmt(s, f), ++ Span::Fallback(s) => Debug::fmt(s, f), + } + } + } + +-pub fn debug_span_field_if_nontrivial(debug: &mut fmt::DebugStruct, span: Span) { ++pub(crate) fn debug_span_field_if_nontrivial(debug: &mut fmt::DebugStruct, span: Span) { + match span { + Span::Compiler(s) => { + debug.field("span", &s); +@@ -561,7 +547,7 @@ pub fn debug_span_field_if_nontrivial(debug: &mut fmt::DebugStruct, span: Span) + } + + #[derive(Clone)] +-pub enum Group { ++pub(crate) enum Group { + Compiler(proc_macro::Group), + Fallback(fallback::Group), + } +@@ -652,26 +638,26 @@ impl From<fallback::Group> for Group { + } + } + +-impl fmt::Display for Group { ++impl Display for Group { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + match self { +- Group::Compiler(group) => group.fmt(formatter), +- Group::Fallback(group) => group.fmt(formatter), ++ Group::Compiler(group) => Display::fmt(group, formatter), ++ Group::Fallback(group) => Display::fmt(group, formatter), + } + } + } + +-impl fmt::Debug for Group { ++impl Debug for Group { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + match self { +- Group::Compiler(group) => group.fmt(formatter), +- Group::Fallback(group) => group.fmt(formatter), ++ Group::Compiler(group) => Debug::fmt(group, formatter), ++ Group::Fallback(group) => Debug::fmt(group, formatter), + } + } + } + + #[derive(Clone)] +-pub enum Ident { ++pub(crate) enum Ident { + Compiler(proc_macro::Ident), + Fallback(fallback::Ident), + } +@@ -747,26 +733,26 @@ where + } + } + +-impl fmt::Display for Ident { ++impl Display for Ident { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- Ident::Compiler(t) => t.fmt(f), +- Ident::Fallback(t) => t.fmt(f), ++ Ident::Compiler(t) => Display::fmt(t, f), ++ Ident::Fallback(t) => Display::fmt(t, f), + } + } + } + +-impl fmt::Debug for Ident { ++impl Debug for Ident { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- Ident::Compiler(t) => t.fmt(f), +- Ident::Fallback(t) => t.fmt(f), ++ Ident::Compiler(t) => Debug::fmt(t, f), ++ Ident::Fallback(t) => Debug::fmt(t, f), + } + } + } + + #[derive(Clone)] +-pub enum Literal { ++pub(crate) enum Literal { + Compiler(proc_macro::Literal), + Fallback(fallback::Literal), + } +@@ -774,7 +760,7 @@ pub enum Literal { + macro_rules! suffixed_numbers { + ($($name:ident => $kind:ident,)*) => ($( + pub fn $name(n: $kind) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::$name(n)) + } else { + Literal::Fallback(fallback::Literal::$name(n)) +@@ -786,7 +772,7 @@ macro_rules! suffixed_numbers { + macro_rules! unsuffixed_integers { + ($($name:ident => $kind:ident,)*) => ($( + pub fn $name(n: $kind) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::$name(n)) + } else { + Literal::Fallback(fallback::Literal::$name(n)) +@@ -830,7 +816,7 @@ impl Literal { + } + + pub fn f32_unsuffixed(f: f32) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::f32_unsuffixed(f)) + } else { + Literal::Fallback(fallback::Literal::f32_unsuffixed(f)) +@@ -838,7 +824,7 @@ impl Literal { + } + + pub fn f64_unsuffixed(f: f64) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::f64_unsuffixed(f)) + } else { + Literal::Fallback(fallback::Literal::f64_unsuffixed(f)) +@@ -846,7 +832,7 @@ impl Literal { + } + + pub fn string(t: &str) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::string(t)) + } else { + Literal::Fallback(fallback::Literal::string(t)) +@@ -854,7 +840,7 @@ impl Literal { + } + + pub fn character(t: char) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::character(t)) + } else { + Literal::Fallback(fallback::Literal::character(t)) +@@ -862,7 +848,7 @@ impl Literal { + } + + pub fn byte_string(bytes: &[u8]) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::byte_string(bytes)) + } else { + Literal::Fallback(fallback::Literal::byte_string(bytes)) +@@ -908,20 +894,20 @@ impl From<fallback::Literal> for Literal { + } + } + +-impl fmt::Display for Literal { ++impl Display for Literal { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- Literal::Compiler(t) => t.fmt(f), +- Literal::Fallback(t) => t.fmt(f), ++ Literal::Compiler(t) => Display::fmt(t, f), ++ Literal::Fallback(t) => Display::fmt(t, f), + } + } + } + +-impl fmt::Debug for Literal { ++impl Debug for Literal { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- Literal::Compiler(t) => t.fmt(f), +- Literal::Fallback(t) => t.fmt(f), ++ Literal::Compiler(t) => Debug::fmt(t, f), ++ Literal::Fallback(t) => Debug::fmt(t, f), + } + } + } +diff --git a/third_party/rust/proc-macro2/tests/comments.rs b/third_party/rust/proc-macro2/tests/comments.rs +new file mode 100644 +index 0000000000..708cccb880 +--- /dev/null ++++ b/third_party/rust/proc-macro2/tests/comments.rs +@@ -0,0 +1,103 @@ ++use proc_macro2::{Delimiter, Literal, Spacing, TokenStream, TokenTree}; ++ ++// #[doc = "..."] -> "..." ++fn lit_of_outer_doc_comment(tokens: TokenStream) -> Literal { ++ lit_of_doc_comment(tokens, false) ++} ++ ++// #![doc = "..."] -> "..." ++fn lit_of_inner_doc_comment(tokens: TokenStream) -> Literal { ++ lit_of_doc_comment(tokens, true) ++} ++ ++fn lit_of_doc_comment(tokens: TokenStream, inner: bool) -> Literal { ++ let mut iter = tokens.clone().into_iter(); ++ match iter.next().unwrap() { ++ TokenTree::Punct(punct) => { ++ assert_eq!(punct.as_char(), '#'); ++ assert_eq!(punct.spacing(), Spacing::Alone); ++ } ++ _ => panic!("wrong token {:?}", tokens), ++ } ++ if inner { ++ match iter.next().unwrap() { ++ TokenTree::Punct(punct) => { ++ assert_eq!(punct.as_char(), '!'); ++ assert_eq!(punct.spacing(), Spacing::Alone); ++ } ++ _ => panic!("wrong token {:?}", tokens), ++ } ++ } ++ iter = match iter.next().unwrap() { ++ TokenTree::Group(group) => { ++ assert_eq!(group.delimiter(), Delimiter::Bracket); ++ assert!(iter.next().is_none(), "unexpected token {:?}", tokens); ++ group.stream().into_iter() ++ } ++ _ => panic!("wrong token {:?}", tokens), ++ }; ++ match iter.next().unwrap() { ++ TokenTree::Ident(ident) => assert_eq!(ident.to_string(), "doc"), ++ _ => panic!("wrong token {:?}", tokens), ++ } ++ match iter.next().unwrap() { ++ TokenTree::Punct(punct) => { ++ assert_eq!(punct.as_char(), '='); ++ assert_eq!(punct.spacing(), Spacing::Alone); ++ } ++ _ => panic!("wrong token {:?}", tokens), ++ } ++ match iter.next().unwrap() { ++ TokenTree::Literal(literal) => { ++ assert!(iter.next().is_none(), "unexpected token {:?}", tokens); ++ literal ++ } ++ _ => panic!("wrong token {:?}", tokens), ++ } ++} ++ ++#[test] ++fn closed_immediately() { ++ let stream = "/**/".parse::<TokenStream>().unwrap(); ++ let tokens = stream.into_iter().collect::<Vec<_>>(); ++ assert!(tokens.is_empty(), "not empty -- {:?}", tokens); ++} ++ ++#[test] ++fn incomplete() { ++ assert!("/*/".parse::<TokenStream>().is_err()); ++} ++ ++#[test] ++fn lit() { ++ let stream = "/// doc".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_outer_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\" doc\""); ++ ++ let stream = "//! doc".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_inner_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\" doc\""); ++ ++ let stream = "/** doc */".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_outer_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\" doc \""); ++ ++ let stream = "/*! doc */".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_inner_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\" doc \""); ++} ++ ++#[test] ++fn carriage_return() { ++ let stream = "///\r\n".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_outer_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\"\""); ++ ++ let stream = "/**\r\n*/".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_outer_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\"\\r\\n\""); ++ ++ "///\r".parse::<TokenStream>().unwrap_err(); ++ "///\r \n".parse::<TokenStream>().unwrap_err(); ++ "/**\r \n*/".parse::<TokenStream>().unwrap_err(); ++} +diff --git a/third_party/rust/proc-macro2/tests/marker.rs b/third_party/rust/proc-macro2/tests/marker.rs +index 7af2539c1a..70e57677cd 100644 +--- a/third_party/rust/proc-macro2/tests/marker.rs ++++ b/third_party/rust/proc-macro2/tests/marker.rs +@@ -57,3 +57,36 @@ mod semver_exempt { + + assert_impl!(SourceFile is not Send or Sync); + } ++ ++#[cfg(not(no_libprocmacro_unwind_safe))] ++mod unwind_safe { ++ use super::*; ++ use std::panic::{RefUnwindSafe, UnwindSafe}; ++ ++ macro_rules! assert_unwind_safe { ++ ($($types:ident)*) => { ++ $( ++ assert_impl!($types is UnwindSafe and RefUnwindSafe); ++ )* ++ }; ++ } ++ ++ assert_unwind_safe! { ++ Delimiter ++ Group ++ Ident ++ LexError ++ Literal ++ Punct ++ Spacing ++ Span ++ TokenStream ++ TokenTree ++ } ++ ++ #[cfg(procmacro2_semver_exempt)] ++ assert_unwind_safe! { ++ LineColumn ++ SourceFile ++ } ++} +diff --git a/third_party/rust/proc-macro2/tests/test.rs b/third_party/rust/proc-macro2/tests/test.rs +index 7528388138..1e9f633944 100644 +--- a/third_party/rust/proc-macro2/tests/test.rs ++++ b/third_party/rust/proc-macro2/tests/test.rs +@@ -1,7 +1,6 @@ ++use proc_macro2::{Ident, Literal, Punct, Spacing, Span, TokenStream, TokenTree}; + use std::str::{self, FromStr}; + +-use proc_macro2::{Ident, Literal, Spacing, Span, TokenStream, TokenTree}; +- + #[test] + fn idents() { + assert_eq!( +@@ -84,6 +83,11 @@ fn literal_string() { + assert_eq!(Literal::string("didn't").to_string(), "\"didn't\""); + } + ++#[test] ++fn literal_raw_string() { ++ "r\"\r\n\"".parse::<TokenStream>().unwrap(); ++} ++ + #[test] + fn literal_character() { + assert_eq!(Literal::character('x').to_string(), "'x'"); +@@ -110,6 +114,37 @@ fn literal_suffix() { + assert_eq!(token_count("1._0"), 3); + assert_eq!(token_count("1._m"), 3); + assert_eq!(token_count("\"\"s"), 1); ++ assert_eq!(token_count("r\"\"r"), 1); ++ assert_eq!(token_count("b\"\"b"), 1); ++ assert_eq!(token_count("br\"\"br"), 1); ++ assert_eq!(token_count("r#\"\"#r"), 1); ++ assert_eq!(token_count("'c'c"), 1); ++ assert_eq!(token_count("b'b'b"), 1); ++ assert_eq!(token_count("0E"), 1); ++ assert_eq!(token_count("0o0A"), 1); ++ assert_eq!(token_count("0E--0"), 4); ++ assert_eq!(token_count("0.0ECMA"), 1); ++} ++ ++#[test] ++fn literal_iter_negative() { ++ let negative_literal = Literal::i32_suffixed(-3); ++ let tokens = TokenStream::from(TokenTree::Literal(negative_literal)); ++ let mut iter = tokens.into_iter(); ++ match iter.next().unwrap() { ++ TokenTree::Punct(punct) => { ++ assert_eq!(punct.as_char(), '-'); ++ assert_eq!(punct.spacing(), Spacing::Alone); ++ } ++ unexpected => panic!("unexpected token {:?}", unexpected), ++ } ++ match iter.next().unwrap() { ++ TokenTree::Literal(literal) => { ++ assert_eq!(literal.to_string(), "3i32"); ++ } ++ unexpected => panic!("unexpected token {:?}", unexpected), ++ } ++ assert!(iter.next().is_none()); + } + + #[test] +@@ -161,41 +196,21 @@ fn fail() { + fail("' static"); + fail("r#1"); + fail("r#_"); ++ fail("\"\\u{0000000}\""); // overlong unicode escape (rust allows at most 6 hex digits) ++ fail("\"\\u{999999}\""); // outside of valid range of char ++ fail("\"\\u{_0}\""); // leading underscore ++ fail("\"\\u{}\""); // empty ++ fail("b\"\r\""); // bare carriage return in byte string ++ fail("r\"\r\""); // bare carriage return in raw string ++ fail("\"\\\r \""); // backslash carriage return ++ fail("'aa'aa"); ++ fail("br##\"\"#"); ++ fail("\"\\\n\u{85}\r\""); + } + + #[cfg(span_locations)] + #[test] + fn span_test() { +- use proc_macro2::TokenTree; +- +- fn check_spans(p: &str, mut lines: &[(usize, usize, usize, usize)]) { +- let ts = p.parse::<TokenStream>().unwrap(); +- check_spans_internal(ts, &mut lines); +- } +- +- fn check_spans_internal(ts: TokenStream, lines: &mut &[(usize, usize, usize, usize)]) { +- for i in ts { +- if let Some((&(sline, scol, eline, ecol), rest)) = lines.split_first() { +- *lines = rest; +- +- let start = i.span().start(); +- assert_eq!(start.line, sline, "sline did not match for {}", i); +- assert_eq!(start.column, scol, "scol did not match for {}", i); +- +- let end = i.span().end(); +- assert_eq!(end.line, eline, "eline did not match for {}", i); +- assert_eq!(end.column, ecol, "ecol did not match for {}", i); +- +- match i { +- TokenTree::Group(ref g) => { +- check_spans_internal(g.stream().clone(), lines); +- } +- _ => {} +- } +- } +- } +- } +- + check_spans( + "\ + /// This is a document comment +@@ -274,53 +289,11 @@ fn span_join() { + #[test] + fn no_panic() { + let s = str::from_utf8(b"b\'\xc2\x86 \x00\x00\x00^\"").unwrap(); +- assert!(s.parse::<proc_macro2::TokenStream>().is_err()); ++ assert!(s.parse::<TokenStream>().is_err()); + } + + #[test] +-fn tricky_doc_comment() { +- let stream = "/**/".parse::<proc_macro2::TokenStream>().unwrap(); +- let tokens = stream.into_iter().collect::<Vec<_>>(); +- assert!(tokens.is_empty(), "not empty -- {:?}", tokens); +- +- let stream = "/// doc".parse::<proc_macro2::TokenStream>().unwrap(); +- let tokens = stream.into_iter().collect::<Vec<_>>(); +- assert!(tokens.len() == 2, "not length 2 -- {:?}", tokens); +- match tokens[0] { +- proc_macro2::TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '#'), +- _ => panic!("wrong token {:?}", tokens[0]), +- } +- let mut tokens = match tokens[1] { +- proc_macro2::TokenTree::Group(ref tt) => { +- assert_eq!(tt.delimiter(), proc_macro2::Delimiter::Bracket); +- tt.stream().into_iter() +- } +- _ => panic!("wrong token {:?}", tokens[0]), +- }; +- +- match tokens.next().unwrap() { +- proc_macro2::TokenTree::Ident(ref tt) => assert_eq!(tt.to_string(), "doc"), +- t => panic!("wrong token {:?}", t), +- } +- match tokens.next().unwrap() { +- proc_macro2::TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '='), +- t => panic!("wrong token {:?}", t), +- } +- match tokens.next().unwrap() { +- proc_macro2::TokenTree::Literal(ref tt) => { +- assert_eq!(tt.to_string(), "\" doc\""); +- } +- t => panic!("wrong token {:?}", t), +- } +- assert!(tokens.next().is_none()); +- +- let stream = "//! doc".parse::<proc_macro2::TokenStream>().unwrap(); +- let tokens = stream.into_iter().collect::<Vec<_>>(); +- assert!(tokens.len() == 3, "not length 3 -- {:?}", tokens); +-} +- +-#[test] +-fn op_before_comment() { ++fn punct_before_comment() { + let mut tts = TokenStream::from_str("~// comment").unwrap().into_iter(); + match tts.next().unwrap() { + TokenTree::Punct(tt) => { +@@ -331,6 +304,22 @@ fn op_before_comment() { + } + } + ++#[test] ++fn joint_last_token() { ++ // This test verifies that we match the behavior of libproc_macro *not* in ++ // the range nightly-2020-09-06 through nightly-2020-09-10, in which this ++ // behavior was temporarily broken. ++ // See https://github.com/rust-lang/rust/issues/76399 ++ ++ let joint_punct = Punct::new(':', Spacing::Joint); ++ let stream = TokenStream::from(TokenTree::Punct(joint_punct)); ++ let punct = match stream.into_iter().next().unwrap() { ++ TokenTree::Punct(punct) => punct, ++ _ => unreachable!(), ++ }; ++ assert_eq!(punct.spacing(), Spacing::Joint); ++} ++ + #[test] + fn raw_identifier() { + let mut tts = TokenStream::from_str("r#dyn").unwrap().into_iter(); +@@ -345,11 +334,11 @@ fn raw_identifier() { + fn test_debug_ident() { + let ident = Ident::new("proc_macro", Span::call_site()); + +- #[cfg(not(procmacro2_semver_exempt))] ++ #[cfg(not(span_locations))] + let expected = "Ident(proc_macro)"; + +- #[cfg(procmacro2_semver_exempt)] +- let expected = "Ident { sym: proc_macro, span: bytes(0..0) }"; ++ #[cfg(span_locations)] ++ let expected = "Ident { sym: proc_macro }"; + + assert_eq!(expected, format!("{:?}", ident)); + } +@@ -358,7 +347,7 @@ fn test_debug_ident() { + fn test_debug_tokenstream() { + let tts = TokenStream::from_str("[a + 1]").unwrap(); + +- #[cfg(not(procmacro2_semver_exempt))] ++ #[cfg(not(span_locations))] + let expected = "\ + TokenStream [ + Group { +@@ -368,7 +357,7 @@ TokenStream [ + sym: a, + }, + Punct { +- op: '+', ++ char: '+', + spacing: Alone, + }, + Literal { +@@ -379,7 +368,7 @@ TokenStream [ + ]\ + "; + +- #[cfg(not(procmacro2_semver_exempt))] ++ #[cfg(not(span_locations))] + let expected_before_trailing_commas = "\ + TokenStream [ + Group { +@@ -389,7 +378,7 @@ TokenStream [ + sym: a + }, + Punct { +- op: '+', ++ char: '+', + spacing: Alone + }, + Literal { +@@ -400,7 +389,7 @@ TokenStream [ + ]\ + "; + +- #[cfg(procmacro2_semver_exempt)] ++ #[cfg(span_locations)] + let expected = "\ + TokenStream [ + Group { +@@ -411,7 +400,7 @@ TokenStream [ + span: bytes(2..3), + }, + Punct { +- op: '+', ++ char: '+', + spacing: Alone, + span: bytes(4..5), + }, +@@ -425,7 +414,7 @@ TokenStream [ + ]\ + "; + +- #[cfg(procmacro2_semver_exempt)] ++ #[cfg(span_locations)] + let expected_before_trailing_commas = "\ + TokenStream [ + Group { +@@ -436,7 +425,7 @@ TokenStream [ + span: bytes(2..3) + }, + Punct { +- op: '+', ++ char: '+', + spacing: Alone, + span: bytes(4..5) + }, +@@ -464,3 +453,80 @@ fn default_tokenstream_is_empty() { + + assert!(default_token_stream.is_empty()); + } ++ ++#[test] ++fn tuple_indexing() { ++ // This behavior may change depending on https://github.com/rust-lang/rust/pull/71322 ++ let mut tokens = "tuple.0.0".parse::<TokenStream>().unwrap().into_iter(); ++ assert_eq!("tuple", tokens.next().unwrap().to_string()); ++ assert_eq!(".", tokens.next().unwrap().to_string()); ++ assert_eq!("0.0", tokens.next().unwrap().to_string()); ++ assert!(tokens.next().is_none()); ++} ++ ++#[cfg(span_locations)] ++#[test] ++fn non_ascii_tokens() { ++ check_spans("// abc", &[]); ++ check_spans("// ábc", &[]); ++ check_spans("// abc x", &[]); ++ check_spans("// ábc x", &[]); ++ check_spans("/* abc */ x", &[(1, 10, 1, 11)]); ++ check_spans("/* ábc */ x", &[(1, 10, 1, 11)]); ++ check_spans("/* ab\nc */ x", &[(2, 5, 2, 6)]); ++ check_spans("/* áb\nc */ x", &[(2, 5, 2, 6)]); ++ check_spans("/*** abc */ x", &[(1, 12, 1, 13)]); ++ check_spans("/*** ábc */ x", &[(1, 12, 1, 13)]); ++ check_spans(r#""abc""#, &[(1, 0, 1, 5)]); ++ check_spans(r#""ábc""#, &[(1, 0, 1, 5)]); ++ check_spans(r###"r#"abc"#"###, &[(1, 0, 1, 8)]); ++ check_spans(r###"r#"ábc"#"###, &[(1, 0, 1, 8)]); ++ check_spans("r#\"a\nc\"#", &[(1, 0, 2, 3)]); ++ check_spans("r#\"á\nc\"#", &[(1, 0, 2, 3)]); ++ check_spans("'a'", &[(1, 0, 1, 3)]); ++ check_spans("'á'", &[(1, 0, 1, 3)]); ++ check_spans("//! abc", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]); ++ check_spans("//! ábc", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]); ++ check_spans("//! abc\n", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]); ++ check_spans("//! ábc\n", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]); ++ check_spans("/*! abc */", &[(1, 0, 1, 10), (1, 0, 1, 10), (1, 0, 1, 10)]); ++ check_spans("/*! ábc */", &[(1, 0, 1, 10), (1, 0, 1, 10), (1, 0, 1, 10)]); ++ check_spans("/*! a\nc */", &[(1, 0, 2, 4), (1, 0, 2, 4), (1, 0, 2, 4)]); ++ check_spans("/*! á\nc */", &[(1, 0, 2, 4), (1, 0, 2, 4), (1, 0, 2, 4)]); ++ check_spans("abc", &[(1, 0, 1, 3)]); ++ check_spans("ábc", &[(1, 0, 1, 3)]); ++ check_spans("ábć", &[(1, 0, 1, 3)]); ++ check_spans("abc// foo", &[(1, 0, 1, 3)]); ++ check_spans("ábc// foo", &[(1, 0, 1, 3)]); ++ check_spans("ábć// foo", &[(1, 0, 1, 3)]); ++ check_spans("b\"a\\\n c\"", &[(1, 0, 2, 3)]); ++ check_spans("b\"a\\\n\u{00a0}c\"", &[(1, 0, 2, 3)]); ++} ++ ++#[cfg(span_locations)] ++fn check_spans(p: &str, mut lines: &[(usize, usize, usize, usize)]) { ++ let ts = p.parse::<TokenStream>().unwrap(); ++ check_spans_internal(ts, &mut lines); ++ assert!(lines.is_empty(), "leftover ranges: {:?}", lines); ++} ++ ++#[cfg(span_locations)] ++fn check_spans_internal(ts: TokenStream, lines: &mut &[(usize, usize, usize, usize)]) { ++ for i in ts { ++ if let Some((&(sline, scol, eline, ecol), rest)) = lines.split_first() { ++ *lines = rest; ++ ++ let start = i.span().start(); ++ assert_eq!(start.line, sline, "sline did not match for {}", i); ++ assert_eq!(start.column, scol, "scol did not match for {}", i); ++ ++ let end = i.span().end(); ++ assert_eq!(end.line, eline, "eline did not match for {}", i); ++ assert_eq!(end.column, ecol, "ecol did not match for {}", i); ++ ++ if let TokenTree::Group(g) = i { ++ check_spans_internal(g.stream().clone(), lines); ++ } ++ } ++ } ++} +diff --git a/third_party/rust/proc-macro2/tests/test_fmt.rs b/third_party/rust/proc-macro2/tests/test_fmt.rs +new file mode 100644 +index 0000000000..99a0aee5c8 +--- /dev/null ++++ b/third_party/rust/proc-macro2/tests/test_fmt.rs +@@ -0,0 +1,26 @@ ++use proc_macro2::{Delimiter, Group, Ident, Span, TokenStream, TokenTree}; ++use std::iter::{self, FromIterator}; ++ ++#[test] ++fn test_fmt_group() { ++ let ident = Ident::new("x", Span::call_site()); ++ let inner = TokenStream::from_iter(iter::once(TokenTree::Ident(ident))); ++ let parens_empty = Group::new(Delimiter::Parenthesis, TokenStream::new()); ++ let parens_nonempty = Group::new(Delimiter::Parenthesis, inner.clone()); ++ let brackets_empty = Group::new(Delimiter::Bracket, TokenStream::new()); ++ let brackets_nonempty = Group::new(Delimiter::Bracket, inner.clone()); ++ let braces_empty = Group::new(Delimiter::Brace, TokenStream::new()); ++ let braces_nonempty = Group::new(Delimiter::Brace, inner.clone()); ++ let none_empty = Group::new(Delimiter::None, TokenStream::new()); ++ let none_nonempty = Group::new(Delimiter::None, inner.clone()); ++ ++ // Matches libproc_macro. ++ assert_eq!("()", parens_empty.to_string()); ++ assert_eq!("(x)", parens_nonempty.to_string()); ++ assert_eq!("[]", brackets_empty.to_string()); ++ assert_eq!("[x]", brackets_nonempty.to_string()); ++ assert_eq!("{ }", braces_empty.to_string()); ++ assert_eq!("{ x }", braces_nonempty.to_string()); ++ assert_eq!("", none_empty.to_string()); ++ assert_eq!("x", none_nonempty.to_string()); ++} +diff --git a/third_party/rust/spirv-cross-internal/.cargo-checksum.json b/third_party/rust/spirv-cross-internal/.cargo-checksum.json +index 3c732d6d0e..014aa640e1 100644 +--- a/third_party/rust/spirv-cross-internal/.cargo-checksum.json ++++ b/third_party/rust/spirv-cross-internal/.cargo-checksum.json +@@ -1 +1 @@ +-{"files":{"Cargo.toml":"32644850d6dd8a8496fcf2dd0327dfc18a41f87bbd5f0922e174a905babde20d","build.rs":"68d1e7d99b4d890e21606aa7e7904d4f9b8caed6c0951bb8f5028745a7103d1e","src/bindings_native.rs":"6aa4e3a3bd73a638b3282405c0eaf2565ff8f364919c509f94f79e326679b03e","src/bindings_wasm.rs":"331d19691566beb163fc459d630e9830167cc86ec8cacde6be53fb0a7ca2b1be","src/bindings_wasm_functions.rs":"2515dd91e98e769fe282e7dc4f60820f4be7365c784a9f3bd17310d18a136da8","src/compiler.rs":"a53c7e0136ea3deddbfaf158295daca36623a0e244426b7c04d2856922d91c73","src/emscripten.rs":"3169890001970610013026468739910afca0d85e00d7e34beadfdd31bbcbeeb7","src/glsl.rs":"7ca1f6a11db8e7926a813fb91298dac1c0e2f90fe5426cc985cde7d8f36c05c9","src/hlsl.rs":"3734ad89a4f51647ebeb761753cb324a48ebca724970201e7598a3763d094d73","src/lib.rs":"cc41cbbe48f3e96791ba5338c66fa1fe0e533eaed6bbdced3f008d5e9fe6c6ce","src/msl.rs":"5505d1d626d8e7903ffa9d137e7392a1670fa7b574eacc4bbd33673937b36ea7","src/ptr_util.rs":"280404beede469b2c9ae40536323515a9213dac5d30014fac870a23b37672442","src/spirv.rs":"8a2cbe6b554bb6312f2aede0469338ace21e79720b5128c653e7a68c0e161277","src/vendor/SPIRV-Cross/.clang-format":"9ec4314e20afecad827a2dbd4832256be8464e88aab4a53fab45173ed129b2ed","src/vendor/SPIRV-Cross/.gitignore":"7f23cc92ddb5e1f584447e98d3e8ab6543fc182f1543f0f6ec29856f9250cdd6","src/vendor/SPIRV-Cross/.travis.yml":"abcc1b2f622b65feafd3e37a2b5e867fce3cf7211cae9fb2bf92a6de79100be4","src/vendor/SPIRV-Cross/CMakeLists.txt":"0972ab271e1eecdfe41ae0a85bf14ce7fe142b7609f8d2965b2c94f92c63b6aa","src/vendor/SPIRV-Cross/CODE_OF_CONDUCT.md":"a25e58cd66a9543e1500db9a5c3c027e874893ad1a264530bf26eb10918b5d80","src/vendor/SPIRV-Cross/GLSL.std.450.h":"20f32378793c5f416bc0704f44345c2a14c99cba3f411e3beaf1bcea372d58ba","src/vendor/SPIRV-Cross/LICENSE":"cfc7749b96f63bd31c3c42b5c471bf756814053e847c10f3eb003417bc523d30","src/vendor/SPIRV-Cross/Makefile":"e2255d32e47d447b100ce3cd0753d0758dc26460e0440d14cc157d4caf9b62f1","src/vendor/SPIRV-Cross/README.md":"b944fc5b1de709089fc3cae0dba09186ce2e43b64de6dcae1423ba1c139ee395","src/vendor/SPIRV-Cross/appveyor.yml":"0f18b8ae5fadc027a20f69307fc32e56a97edfbdcb55ac392852308f88545a04","src/vendor/SPIRV-Cross/include/spirv_cross/barrier.hpp":"bb796625e89f75e239e92f9a61597d421ffe5fb1902d200691ebe95cf856a1f8","src/vendor/SPIRV-Cross/include/spirv_cross/external_interface.h":"cdceda962d87133e44989510edc944e99052d713869b406a8b6b2d54e3d02dd7","src/vendor/SPIRV-Cross/include/spirv_cross/image.hpp":"681d0964b144c5009424196a8bc832cb81cfe5df5b91c2f3e1bfb625765a0c50","src/vendor/SPIRV-Cross/include/spirv_cross/internal_interface.hpp":"ab8851e5708b944a9bf340ce17297d94bef4876d30c833ea83d44b16f60726f6","src/vendor/SPIRV-Cross/include/spirv_cross/sampler.hpp":"b0b1bd4796b4315e826985c224ea5fa7d5bc823fcd6091f7527a5e18d9ae42fb","src/vendor/SPIRV-Cross/include/spirv_cross/thread_group.hpp":"70d9e0400f62de71d3775972eadc196ddb218254fa8155e8e33daf8d99957cc0","src/vendor/SPIRV-Cross/main.cpp":"ab2835f6dff9299f562a89755893f98a4bd946d4460abca885ba2172ebce851c","src/vendor/SPIRV-Cross/pkg-config/spirv-cross-c-shared.pc.in":"75a27e069ee07362bc6b7412b2c4e1270778dd639a8b96b8afbb113d21de79e7","src/vendor/SPIRV-Cross/spirv.h":"2c19cd1d262a3757e92276fbaa8e1a706b2328a92f47bda88f849b5cebebb630","src/vendor/SPIRV-Cross/spirv.hpp":"eb7c39a017f0a33e9d9bae7556c678c2840af4cc0beb98b2aeff9d651e370128","src/vendor/SPIRV-Cross/spirv_cfg.cpp":"09b756068d95a87a65a186831484fe21a973d3eb19675225b23864b9d37fe3d0","src/vendor/SPIRV-Cross/spirv_cfg.hpp":"ad98c4e82d6da3ed450dfc7939814115ae6b29ba2a25b0148cfc8c88774055b5","src/vendor/SPIRV-Cross/spirv_common.hpp":"5d0658a6fe9ac2f628754d2439b4d26ec8df7e50e4b196567cd4c7e7b8ce7574","src/vendor/SPIRV-Cross/spirv_cpp.cpp":"d067269ec6fc56860b4ed39edddf0a127f5e2d091512a85a79f02d5f3beccd19","src/vendor/SPIRV-Cross/spirv_cpp.hpp":"751ce82c70a399781f18aca54ea7fbb98083b98706f2a6be1cd1aa90627ef597","src/vendor/SPIRV-Cross/spirv_cross.cpp":"225b4cf4c7f8868019680d0a75a3da7983db428783ee1a3dcbb8c2765a71e79a","src/vendor/SPIRV-Cross/spirv_cross.hpp":"7eab2d7ae4b25a3c4fe85631e664727e036ae23a776aeb8c17db23f08020258a","src/vendor/SPIRV-Cross/spirv_cross_c.cpp":"80f438f4700f0b0b02547771f82be0610ba5e24ba3edb56d4331d59a206a4c2e","src/vendor/SPIRV-Cross/spirv_cross_c.h":"8b0ad63ed9fe52bc62f305afa27b5e33d9bd689a8b3f9935353742faf88a1e91","src/vendor/SPIRV-Cross/spirv_cross_containers.hpp":"409f60a5cb1dba1d0c96ef61761d8643402fc5ab67cecce12f6af3d135cfa338","src/vendor/SPIRV-Cross/spirv_cross_error_handling.hpp":"45868796346eef895cc05825582e60d26c34ec35d1a2b97f5e172f05fc565b19","src/vendor/SPIRV-Cross/spirv_cross_parsed_ir.cpp":"7c7e08db4dfc1084fd9cd266a6f829d35a6b2856198b16307a44f4d1c0743106","src/vendor/SPIRV-Cross/spirv_cross_parsed_ir.hpp":"d15446565eaa430bf2b14b596b3e579a72095a386b1eb256afc977cdf42ec5b9","src/vendor/SPIRV-Cross/spirv_cross_util.cpp":"f2426dccad2fff44053f6b1b9558719471a1ba0445c317f2c0311bc0d8b4636d","src/vendor/SPIRV-Cross/spirv_cross_util.hpp":"63ca5b2d6c80b5ad39473d6841676e48d5db3d2cbcb252fe6a6db0ef2077f0f5","src/vendor/SPIRV-Cross/spirv_glsl.cpp":"c1704189895442d593819472d3f852979107ac1798375d5a44d8f97be17fa3b5","src/vendor/SPIRV-Cross/spirv_glsl.hpp":"fb5e00ff589e3c5a25c5a15e78cf4809b753ebe04c84a008dc977330b8cbbc27","src/vendor/SPIRV-Cross/spirv_hlsl.cpp":"549ff42d0773a3f9e555d1aebd4c0084709e11528082f47c5f2b8b07de029f7c","src/vendor/SPIRV-Cross/spirv_hlsl.hpp":"18f303e7a384acf0e14677b579b100aee838813f54735330f0d1387f2493b363","src/vendor/SPIRV-Cross/spirv_msl.cpp":"7266afeb827e50782cb35362788d27fa9bd641992380008604ea1a4e48ec9c37","src/vendor/SPIRV-Cross/spirv_msl.hpp":"76b609e0b8d2f421c51e98d6357fcf44666b0d510935e083e4e02f20a80051e1","src/vendor/SPIRV-Cross/spirv_parser.cpp":"4def4db8c77967dcd2d6d93330343dc6ebb309e8cd696f52ee568a7c3ee97036","src/vendor/SPIRV-Cross/spirv_parser.hpp":"fc5f92184d1b4986f78088bee8ed7ddb9b342a34f9a3e55a4a594187292b66b8","src/vendor/SPIRV-Cross/spirv_reflect.cpp":"995a4e4315f67d4a2084e2b5d982de9baf72c7e265dde71d03ae0b8a29a24b92","src/vendor/SPIRV-Cross/spirv_reflect.hpp":"a874b40be0b210c69718e87a90e9add24653957656a995d83656f0918cfd10e4","src/wrapper.cpp":"7800cec5564dc1c7c13059cb1b1cac8d9be9248bbc887d5176600980c0d18142","src/wrapper.hpp":"b9a4f846bf57ee7a10fbb81f8f49df96ad11b1e5749c3c510925bb9b029ac08a","tests/common/mod.rs":"2843bf104c7938d93065f7b5688c9f063ad9e5720c407c737aedc5f2dee5a80f","tests/glsl_tests.rs":"ddd85107370dbfcde9d004d6717e229a7278d29ecba375828d8cb79e86625f66","tests/hlsl_tests.rs":"2058cb590c5ab85a636c93c18d0f2a79eb5ea5e647b7f10a2c83fbf394964dd6","tests/msl_tests.rs":"211d3b9cb43455a4c55bd619c05acdd21953358580c50ae75cac3f06eb26c5dd","tests/shaders/array.vert":"d0dab7ddea131e069961054f40a164602448aa78720b7ff480e141d1f7b0b2d6","tests/shaders/array.vert.spv":"8e44421590ade1716be66ad39f60fb1ce58eedeab8f0806335a7369687b308b1","tests/shaders/multiple_entry_points.cl":"2abbe57f2387f07f7f5f4cd375e47900be9c13bdc79aa0ed98a43a798cb0df81","tests/shaders/multiple_entry_points.cl.spv":"bdd34ce6765dbeab637631c3cbdf251532870d9fd6cd6c54883c0c872058ab3b","tests/shaders/rasterize_disabled.vert":"da6de172549830216933c44edf18b13113d7ca87462e3d09ad50dfc9c9836821","tests/shaders/rasterize_disabled.vert.spv":"2ba809eb500ed6e5a067389ccc056551e796e7019517593d110fb62c9dca2056","tests/shaders/sampler.frag":"4c568e65176afe596dd8ef279485e992607e94d612786214ae1c6702d0322e1f","tests/shaders/sampler.frag.spv":"bd7bd1973a82dcfdf5755361fa4dd420fdf1c32c5de0a6f0896a8d5971f98684","tests/shaders/simple.vert":"ea143c97dff5ef03728b96b2dd893bdc59d56651581ecf9fe50f10807b0efdd0","tests/shaders/simple.vert.spv":"a2b5094ffd76288e0d08c37ce0351e28f20bb6d80ddd73fc44a71c1c7cbbf7db","tests/shaders/specialization.comp":"ce32fa1615737209f2e465ea347d79030ddcb33a88c38447e7cde7dffc920163","tests/shaders/specialization.comp.spv":"848604e37b870b8999692b266677be2ce0df6ce38093a0d81e6bc43d0bdf8a3f","tests/shaders/struct.frag":"d8840bb1961d6f14609b00ee54406c1e3ea31cecd8231b77cfb73d28b71910c0","tests/shaders/struct.frag.spv":"774aa886374eb95abf9bb7d0045ee77d97e26529e9ec96b90991a515fdbca4be","tests/shaders/struct.vert":"9299cda83ddb5b5c3d95ab0d057e4df2af137dfd92d6c4d3e96295b7d42e29a1","tests/shaders/struct.vert.spv":"4a82bdee72616ac058bc60d4255efa3e78199a2b8597570c013bebbee7107fb7","tests/shaders/two_ubo.vert":"be109b2c65e5e9e1bb0dab968d7f651232f6b1c46935a3928f980bf7a40f2d62","tests/shaders/two_ubo.vert.spv":"efd14e2d0a782d61dd944711f30b9e7fcb14af17593c1fe4e11cf2b7e232bcc2","tests/shaders/workgroup.comp":"478044b5392e0d1fb042253d71ea6bf7b8a014af4a6ee35d8db4c5343ac69739","tests/shaders/workgroup.comp.spv":"72f636fe3d1d6d0c5963f71bf4349c7e40d544331b33b6b64fb5b65784e6abee","tests/spirv_tests.rs":"6de5c893f631d550fc86853537bddb824ffb534a0a7f130406be3b9ed166f198"},"package":null} +\ No newline at end of file ++{"files":{"Cargo.toml":"32644850d6dd8a8496fcf2dd0327dfc18a41f87bbd5f0922e174a905babde20d","build.rs":"68d1e7d99b4d890e21606aa7e7904d4f9b8caed6c0951bb8f5028745a7103d1e","src/bindings_native.rs":"6aa4e3a3bd73a638b3282405c0eaf2565ff8f364919c509f94f79e326679b03e","src/bindings_wasm.rs":"331d19691566beb163fc459d630e9830167cc86ec8cacde6be53fb0a7ca2b1be","src/bindings_wasm_functions.rs":"2515dd91e98e769fe282e7dc4f60820f4be7365c784a9f3bd17310d18a136da8","src/compiler.rs":"a53c7e0136ea3deddbfaf158295daca36623a0e244426b7c04d2856922d91c73","src/emscripten.rs":"3169890001970610013026468739910afca0d85e00d7e34beadfdd31bbcbeeb7","src/glsl.rs":"7ca1f6a11db8e7926a813fb91298dac1c0e2f90fe5426cc985cde7d8f36c05c9","src/hlsl.rs":"3734ad89a4f51647ebeb761753cb324a48ebca724970201e7598a3763d094d73","src/lib.rs":"cc41cbbe48f3e96791ba5338c66fa1fe0e533eaed6bbdced3f008d5e9fe6c6ce","src/msl.rs":"5505d1d626d8e7903ffa9d137e7392a1670fa7b574eacc4bbd33673937b36ea7","src/ptr_util.rs":"280404beede469b2c9ae40536323515a9213dac5d30014fac870a23b37672442","src/spirv.rs":"8a2cbe6b554bb6312f2aede0469338ace21e79720b5128c653e7a68c0e161277","src/vendor/SPIRV-Cross/.clang-format":"9ec4314e20afecad827a2dbd4832256be8464e88aab4a53fab45173ed129b2ed","src/vendor/SPIRV-Cross/.travis.yml":"abcc1b2f622b65feafd3e37a2b5e867fce3cf7211cae9fb2bf92a6de79100be4","src/vendor/SPIRV-Cross/CMakeLists.txt":"0972ab271e1eecdfe41ae0a85bf14ce7fe142b7609f8d2965b2c94f92c63b6aa","src/vendor/SPIRV-Cross/CODE_OF_CONDUCT.md":"a25e58cd66a9543e1500db9a5c3c027e874893ad1a264530bf26eb10918b5d80","src/vendor/SPIRV-Cross/GLSL.std.450.h":"20f32378793c5f416bc0704f44345c2a14c99cba3f411e3beaf1bcea372d58ba","src/vendor/SPIRV-Cross/LICENSE":"cfc7749b96f63bd31c3c42b5c471bf756814053e847c10f3eb003417bc523d30","src/vendor/SPIRV-Cross/Makefile":"e2255d32e47d447b100ce3cd0753d0758dc26460e0440d14cc157d4caf9b62f1","src/vendor/SPIRV-Cross/README.md":"b944fc5b1de709089fc3cae0dba09186ce2e43b64de6dcae1423ba1c139ee395","src/vendor/SPIRV-Cross/appveyor.yml":"0f18b8ae5fadc027a20f69307fc32e56a97edfbdcb55ac392852308f88545a04","src/vendor/SPIRV-Cross/include/spirv_cross/barrier.hpp":"bb796625e89f75e239e92f9a61597d421ffe5fb1902d200691ebe95cf856a1f8","src/vendor/SPIRV-Cross/include/spirv_cross/external_interface.h":"cdceda962d87133e44989510edc944e99052d713869b406a8b6b2d54e3d02dd7","src/vendor/SPIRV-Cross/include/spirv_cross/image.hpp":"681d0964b144c5009424196a8bc832cb81cfe5df5b91c2f3e1bfb625765a0c50","src/vendor/SPIRV-Cross/include/spirv_cross/internal_interface.hpp":"ab8851e5708b944a9bf340ce17297d94bef4876d30c833ea83d44b16f60726f6","src/vendor/SPIRV-Cross/include/spirv_cross/sampler.hpp":"b0b1bd4796b4315e826985c224ea5fa7d5bc823fcd6091f7527a5e18d9ae42fb","src/vendor/SPIRV-Cross/include/spirv_cross/thread_group.hpp":"70d9e0400f62de71d3775972eadc196ddb218254fa8155e8e33daf8d99957cc0","src/vendor/SPIRV-Cross/main.cpp":"ab2835f6dff9299f562a89755893f98a4bd946d4460abca885ba2172ebce851c","src/vendor/SPIRV-Cross/pkg-config/spirv-cross-c-shared.pc.in":"75a27e069ee07362bc6b7412b2c4e1270778dd639a8b96b8afbb113d21de79e7","src/vendor/SPIRV-Cross/spirv.h":"2c19cd1d262a3757e92276fbaa8e1a706b2328a92f47bda88f849b5cebebb630","src/vendor/SPIRV-Cross/spirv.hpp":"eb7c39a017f0a33e9d9bae7556c678c2840af4cc0beb98b2aeff9d651e370128","src/vendor/SPIRV-Cross/spirv_cfg.cpp":"09b756068d95a87a65a186831484fe21a973d3eb19675225b23864b9d37fe3d0","src/vendor/SPIRV-Cross/spirv_cfg.hpp":"ad98c4e82d6da3ed450dfc7939814115ae6b29ba2a25b0148cfc8c88774055b5","src/vendor/SPIRV-Cross/spirv_common.hpp":"5d0658a6fe9ac2f628754d2439b4d26ec8df7e50e4b196567cd4c7e7b8ce7574","src/vendor/SPIRV-Cross/spirv_cpp.cpp":"d067269ec6fc56860b4ed39edddf0a127f5e2d091512a85a79f02d5f3beccd19","src/vendor/SPIRV-Cross/spirv_cpp.hpp":"751ce82c70a399781f18aca54ea7fbb98083b98706f2a6be1cd1aa90627ef597","src/vendor/SPIRV-Cross/spirv_cross.cpp":"225b4cf4c7f8868019680d0a75a3da7983db428783ee1a3dcbb8c2765a71e79a","src/vendor/SPIRV-Cross/spirv_cross.hpp":"7eab2d7ae4b25a3c4fe85631e664727e036ae23a776aeb8c17db23f08020258a","src/vendor/SPIRV-Cross/spirv_cross_c.cpp":"80f438f4700f0b0b02547771f82be0610ba5e24ba3edb56d4331d59a206a4c2e","src/vendor/SPIRV-Cross/spirv_cross_c.h":"8b0ad63ed9fe52bc62f305afa27b5e33d9bd689a8b3f9935353742faf88a1e91","src/vendor/SPIRV-Cross/spirv_cross_containers.hpp":"409f60a5cb1dba1d0c96ef61761d8643402fc5ab67cecce12f6af3d135cfa338","src/vendor/SPIRV-Cross/spirv_cross_error_handling.hpp":"45868796346eef895cc05825582e60d26c34ec35d1a2b97f5e172f05fc565b19","src/vendor/SPIRV-Cross/spirv_cross_parsed_ir.cpp":"7c7e08db4dfc1084fd9cd266a6f829d35a6b2856198b16307a44f4d1c0743106","src/vendor/SPIRV-Cross/spirv_cross_parsed_ir.hpp":"d15446565eaa430bf2b14b596b3e579a72095a386b1eb256afc977cdf42ec5b9","src/vendor/SPIRV-Cross/spirv_cross_util.cpp":"f2426dccad2fff44053f6b1b9558719471a1ba0445c317f2c0311bc0d8b4636d","src/vendor/SPIRV-Cross/spirv_cross_util.hpp":"63ca5b2d6c80b5ad39473d6841676e48d5db3d2cbcb252fe6a6db0ef2077f0f5","src/vendor/SPIRV-Cross/spirv_glsl.cpp":"c1704189895442d593819472d3f852979107ac1798375d5a44d8f97be17fa3b5","src/vendor/SPIRV-Cross/spirv_glsl.hpp":"fb5e00ff589e3c5a25c5a15e78cf4809b753ebe04c84a008dc977330b8cbbc27","src/vendor/SPIRV-Cross/spirv_hlsl.cpp":"549ff42d0773a3f9e555d1aebd4c0084709e11528082f47c5f2b8b07de029f7c","src/vendor/SPIRV-Cross/spirv_hlsl.hpp":"18f303e7a384acf0e14677b579b100aee838813f54735330f0d1387f2493b363","src/vendor/SPIRV-Cross/spirv_msl.cpp":"7266afeb827e50782cb35362788d27fa9bd641992380008604ea1a4e48ec9c37","src/vendor/SPIRV-Cross/spirv_msl.hpp":"76b609e0b8d2f421c51e98d6357fcf44666b0d510935e083e4e02f20a80051e1","src/vendor/SPIRV-Cross/spirv_parser.cpp":"4def4db8c77967dcd2d6d93330343dc6ebb309e8cd696f52ee568a7c3ee97036","src/vendor/SPIRV-Cross/spirv_parser.hpp":"fc5f92184d1b4986f78088bee8ed7ddb9b342a34f9a3e55a4a594187292b66b8","src/vendor/SPIRV-Cross/spirv_reflect.cpp":"995a4e4315f67d4a2084e2b5d982de9baf72c7e265dde71d03ae0b8a29a24b92","src/vendor/SPIRV-Cross/spirv_reflect.hpp":"a874b40be0b210c69718e87a90e9add24653957656a995d83656f0918cfd10e4","src/wrapper.cpp":"7800cec5564dc1c7c13059cb1b1cac8d9be9248bbc887d5176600980c0d18142","src/wrapper.hpp":"b9a4f846bf57ee7a10fbb81f8f49df96ad11b1e5749c3c510925bb9b029ac08a","tests/common/mod.rs":"2843bf104c7938d93065f7b5688c9f063ad9e5720c407c737aedc5f2dee5a80f","tests/glsl_tests.rs":"ddd85107370dbfcde9d004d6717e229a7278d29ecba375828d8cb79e86625f66","tests/hlsl_tests.rs":"2058cb590c5ab85a636c93c18d0f2a79eb5ea5e647b7f10a2c83fbf394964dd6","tests/msl_tests.rs":"211d3b9cb43455a4c55bd619c05acdd21953358580c50ae75cac3f06eb26c5dd","tests/shaders/array.vert":"d0dab7ddea131e069961054f40a164602448aa78720b7ff480e141d1f7b0b2d6","tests/shaders/array.vert.spv":"8e44421590ade1716be66ad39f60fb1ce58eedeab8f0806335a7369687b308b1","tests/shaders/multiple_entry_points.cl":"2abbe57f2387f07f7f5f4cd375e47900be9c13bdc79aa0ed98a43a798cb0df81","tests/shaders/multiple_entry_points.cl.spv":"bdd34ce6765dbeab637631c3cbdf251532870d9fd6cd6c54883c0c872058ab3b","tests/shaders/rasterize_disabled.vert":"da6de172549830216933c44edf18b13113d7ca87462e3d09ad50dfc9c9836821","tests/shaders/rasterize_disabled.vert.spv":"2ba809eb500ed6e5a067389ccc056551e796e7019517593d110fb62c9dca2056","tests/shaders/sampler.frag":"4c568e65176afe596dd8ef279485e992607e94d612786214ae1c6702d0322e1f","tests/shaders/sampler.frag.spv":"bd7bd1973a82dcfdf5755361fa4dd420fdf1c32c5de0a6f0896a8d5971f98684","tests/shaders/simple.vert":"ea143c97dff5ef03728b96b2dd893bdc59d56651581ecf9fe50f10807b0efdd0","tests/shaders/simple.vert.spv":"a2b5094ffd76288e0d08c37ce0351e28f20bb6d80ddd73fc44a71c1c7cbbf7db","tests/shaders/specialization.comp":"ce32fa1615737209f2e465ea347d79030ddcb33a88c38447e7cde7dffc920163","tests/shaders/specialization.comp.spv":"848604e37b870b8999692b266677be2ce0df6ce38093a0d81e6bc43d0bdf8a3f","tests/shaders/struct.frag":"d8840bb1961d6f14609b00ee54406c1e3ea31cecd8231b77cfb73d28b71910c0","tests/shaders/struct.frag.spv":"774aa886374eb95abf9bb7d0045ee77d97e26529e9ec96b90991a515fdbca4be","tests/shaders/struct.vert":"9299cda83ddb5b5c3d95ab0d057e4df2af137dfd92d6c4d3e96295b7d42e29a1","tests/shaders/struct.vert.spv":"4a82bdee72616ac058bc60d4255efa3e78199a2b8597570c013bebbee7107fb7","tests/shaders/two_ubo.vert":"be109b2c65e5e9e1bb0dab968d7f651232f6b1c46935a3928f980bf7a40f2d62","tests/shaders/two_ubo.vert.spv":"efd14e2d0a782d61dd944711f30b9e7fcb14af17593c1fe4e11cf2b7e232bcc2","tests/shaders/workgroup.comp":"478044b5392e0d1fb042253d71ea6bf7b8a014af4a6ee35d8db4c5343ac69739","tests/shaders/workgroup.comp.spv":"72f636fe3d1d6d0c5963f71bf4349c7e40d544331b33b6b64fb5b65784e6abee","tests/spirv_tests.rs":"6de5c893f631d550fc86853537bddb824ffb534a0a7f130406be3b9ed166f198"},"package":null} +\ No newline at end of file +diff --git a/third_party/rust/syn/.cargo-checksum.json b/third_party/rust/syn/.cargo-checksum.json +index 77939d8fc6..704f2ed200 100644 +--- a/third_party/rust/syn/.cargo-checksum.json ++++ b/third_party/rust/syn/.cargo-checksum.json +@@ -1 +1 @@ +-{"files":{"Cargo.toml":"484d29864d333a361652fa4e24e1dcfab9efa47705ffd8c106d802eb03b78da7","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"23f18e03dc49df91622fe2a76176497404e46ced8a715d9d2b67a7446571cca3","README.md":"ca605417b6db8c995458f8407afaad6c177aedcc2274004283600f5638fa1b0c","benches/file.rs":"b45211cc4a0296a77aac2b4de16dbc6b5cb66adfb5afac00a77bccea87f43968","benches/rust.rs":"9cc0f62e944f1583d05c43a395a1556731501cf5976ef67a081f4f6387f883ba","build.rs":"7423ab199728d55c7d64c44b7c6729cfd93bd8273366a77707353003e27565d7","src/attr.rs":"cf81add298f0e75c35a9980a59bc3c2fd3fe933635830d1591374eeb2487c225","src/await.rs":"18f0b2ecb319991f891e300011424985e3cf33d166ea9f29f22d575fc8c83a76","src/bigint.rs":"efc7f64959980653d73fe4f8bc2a3a2904dc05f45b02c6dc15cd316fa3d7c338","src/buffer.rs":"2a432c11a3da67a21d46c2272bf9ce60a0bb20893b5750027bbd8ca3e843ab35","src/custom_keyword.rs":"589e46ec1be9a04d6de12c0b8cadf87cc1c05606ed46ddea62e9869cbca4a191","src/custom_punctuation.rs":"2ba2e294e15a0fce7ede3686c42b2891797079a724dd1193b66e7d305624c891","src/data.rs":"cc9b250d084e444782d3ff5e63c1ba387cbde8f7f2e977eab9846d920b4b8c3f","src/derive.rs":"c18878f14be5d5ab11fd7dda2d2ff1ff75c9662daf11eed033de62e4d0670a89","src/discouraged.rs":"50e10915695c4d14f64a78e20ecbef90a2cd53a7c26ee3426a2524a8ee5c9cbf","src/error.rs":"2c17a402f83ed5ae4ad96e753216771bef620235c2ff1ccc23f4bbafc7266fe1","src/export.rs":"dcae67456278c0339acfbcbb4737b8d37cfba5a150ae789f31f4be79abf7e726","src/expr.rs":"871d8eeb43cef02ef88de3bea7477b79b4eabc096a0899dde0e5750edf482f49","src/ext.rs":"b97ed549490b9248b5b5df31b3d5b08ba8791e23e6c5d3a1157a0363eb683ff3","src/file.rs":"3cc2bf5c709238d515a557f721f231c8c725b196400de051f945b549299d38a7","src/gen/fold.rs":"10b3ae33d0ce410d6bbe8b93be9d5f9e856c7dc8212133cc46b703f97d548190","src/gen/visit.rs":"e0f5798552d186024696b7bfc7219d4ff53b0e45f735a83e77cbb6b6578c5fa4","src/gen/visit_mut.rs":"9f7dda83907969971dba84d545aaa563b0728e54db97ffab5050fdf43a79c731","src/gen_helper.rs":"ea6c66388365971db6a2fc86cbb208f7eacde77e245bc8623f27a3642a3d7741","src/generics.rs":"d845d7a828863123a5187fd0fe59c9dae3636f63bad302bd035792eed3dcb1ba","src/group.rs":"119b62d8481b4b1c327639bed40e114bf1969765250b68186628247fd4144b3b","src/ident.rs":"503156ce51a7ef0420892e8dbf2ecf8fe51f42a84d52cc2c05654e1a83020cbf","src/item.rs":"213f2f58c65ee1aa222f111bc9b1be681f8fb069caed04ca56586839979318d0","src/keyword.rs":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855","src/lib.rs":"24778e9f15e8025e75aca114c712716ada586b471adb3b3b69278f4d39b8a21b","src/lifetime.rs":"905359708f772ec858954badde69ee016d29e6eeba1dd205b268445b1aff6f3a","src/lit.rs":"5bb0bddb94cbd256e50e92dc091a0baa09f1be40a77058b897507f3b17191e5d","src/lookahead.rs":"5cce8b4cb345a85c24a452ea6d78eadb76f01ca0a789cbf5ce35108334904173","src/mac.rs":"6b468244cc07e3f2f10419f833d9e2ed23edbcd6dc34cf21c5947633699db964","src/macros.rs":"0d8c3bab47539aa2d00bec64e92c901ea2c9c0af74c868051c0905b82650f970","src/op.rs":"93cd44770bb110deadf807a01d9a666efe644b6e3010f4b51cae77ee7438cfbb","src/parse.rs":"5017123c249ebc65866af113a0ad671814b9873f47568180e6539a305eb0317d","src/parse_macro_input.rs":"f799aadb7216c2d333b579f48ed2fedfe07b5e96f004b25b569649ffbaa958d2","src/parse_quote.rs":"81575bf60b18b0d8624d7025a5bcc8dcd6633ad70c454dee2a06e4c391700b6c","src/pat.rs":"db0f2263b9813de1f4e3e3e0396fe0080b1e11c8090c6b4fb6fca3cfbe22bc96","src/path.rs":"32e685ac7fd2d4b9989802de8f326a8d47fa710f86ec3e45fd9d3ff8fdfe97ef","src/print.rs":"da6529c1d9d21aaf6c835f66b4e67eacb7cf91a10eb5e9a2143b49bf99b3b5e1","src/punctuated.rs":"384e7b317b26f24118eb4b0c39e949ee9f4f3e700a4c80e462342c83b2cc3282","src/sealed.rs":"896a495a5340eec898527f18bd4ddca408ea03ea0ee3af30074ff48deace778d","src/span.rs":"748c51c6feb223c26d3b1701f5bb98aee823666c775c98106cfa24fe29d8cec1","src/spanned.rs":"adddb6acae14a0fa340df302b932c31e34b259706ce56fd82ab597ec424500e1","src/stmt.rs":"fbccf2b4da7980fe6ea8d99457d291577c0f225b370c1dd97da41abf2a18fcf7","src/thread.rs":"815eca6bd64f4eef7c447f0809e84108f5428ff50225224b373efd8fbb696874","src/token.rs":"761d8d1793560eb2b631c36ddfdbb14ac65178405f095453aa0e75e8816bdbb9","src/tt.rs":"1e32ae216d14d895ff621bf32bc611f138aa00108b0090be2cbaa0affebe8e2a","src/ty.rs":"ce052e0079b65b66bea4e9502d2ff2c90ad4b867904bf7eb892eb60aa3ef219a","tests/clone.sh":"961243d42540d9992090efbbd5478b7aa395671db64a6c28cdadb6edc610ebdf","tests/common/eq.rs":"a42d339808fd32dd4bfd440c254add8c56d5e2cde3a6bf0c88621b618ce5eaa7","tests/common/mod.rs":"20a3300016351fa210a193fbb0db059ef5125fa7909585ded64790004d4977ed","tests/common/parse.rs":"17ba6d1e74aaa3f8096c6d379d803221f12d95cca69927be047d6ddf8367647f","tests/debug/gen.rs":"57bd5cf585e0b86ad00f29f09ff3db3390c4a756d503514a9b28407500dcea3c","tests/debug/mod.rs":"462d6fe34ee75c3ca1207d4db2ff3bdee5b430b9f9ca632e5671d1588d3f76b3","tests/features/error.rs":"e0581a2869cbd237c2bc18a0a85272296e1577bb5f7317a67fa85e28e04eea6f","tests/features/mod.rs":"66a2605ec54ede29208da350f2bed536dfa763b58408d64d3fca3b13de64b64f","tests/macros/mod.rs":"3f2d758c0ba76b93f54b0c1fc22ad50edff8ef42629ba4d47ac7d7f823da8359","tests/repo/mod.rs":"e851a68972c9194a9a8d7b68538b16ed79ae81cba55e1a2ce210d1b759fb1a21","tests/test_asyncness.rs":"b6c46118b036e6807d24eb0e1779244b4fca23dac0d8031e9843b3edec484ce8","tests/test_attribute.rs":"2d8f18a98c989d3f7adaaeb1aeebd4f8413365ace63feecb37cb3f9db9db4d8f","tests/test_derive_input.rs":"477d80f914c54b526f8ff229788dc0e7798d118f6dcfa348f4c99755edb347b9","tests/test_expr.rs":"f35ca80566849a36e6ba6403d9663519eff37e4224360c468fedff8b561a643e","tests/test_generics.rs":"83a5dc07f5c5701c12625399262f7120b66f01a742523f3eda28da2cf2c87eb3","tests/test_grouping.rs":"aadd75215addd9e5a8fa2f9472117d4cb80f1e8b84e07f4c0845675c9014164f","tests/test_ident.rs":"236c239dd66f543f084f44ff747d7bc3962cf11a019a279777fe972f6e17aa4c","tests/test_iterators.rs":"718938da14778dcba06324d36a99d9317c9d45d81a34c6a44c47e1fa38085e9f","tests/test_lit.rs":"7dff2661a5ac586d6ed2fe27501cb8ff62f4cf3f6c91f596bff6057c67ad7857","tests/test_meta.rs":"8444dee084882243b107dfc8a6aac27f9382f9774162d1ac8ed8ec30d60c048e","tests/test_parse_buffer.rs":"b244bb4bc41ff06d21f239e60a3d663fdec5aa4af33f2a354afef36d34f0aefc","tests/test_pat.rs":"41776b878efae9b8e340f21ffe6296e921cf309f618482efd98609c33e32c28b","tests/test_precedence.rs":"71f3ea52cda8b40166bb7416fb98774e6a653542497b521f8e183e283dcf579d","tests/test_round_trip.rs":"e0de37f45fa223b488d25a41beab185eb92abb7bf765a9f13fe5d870ff31f5f1","tests/test_should_parse.rs":"4da4e25ee2baa7e75135c375042a7f958de136c5698dab03f99ff7a774dcd463","tests/test_size.rs":"970150b9d49ef91ab4c8f8c6a59b83f9a68a02acb779f0280733a5efaec6487a","tests/test_token_trees.rs":"a07ea657bf03b9c667c821b2db2af49b176ca737e3e01217a73cca78b7f11380","tests/zzz_stable.rs":"961d4940a926db4ca523d834b060c62de988e6a8e01c9f5efaa7bb4c86745b47"},"package":"66850e97125af79138385e9b88339cbcd037e3f28ceab8c5ad98e64f0f1f80bf"} +\ No newline at end of file ++{"files":{"Cargo.toml":"28ddb678a5ccac4423435384c8b7116f804e896eabc5aae9d5c2bc666aaebbb4","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"23f18e03dc49df91622fe2a76176497404e46ced8a715d9d2b67a7446571cca3","README.md":"03f3b53cf858536a0883aa5b5882ee61dcd0f1e71c0930c9106fcfa1d6aad2df","benches/file.rs":"b4724fc7c0f48b8f488e2632a1064f6c0bf16ded3969680fc3f4a2369536269b","benches/rust.rs":"ea6291ef2d2a83d94a3312fe179d48259f8ec0b04c961993ddd181d0a4ab740e","build.rs":"aeca2312f05aec658eaa66980a0ef3d578837db107a55702b39419ea0422eb4a","src/attr.rs":"7d79482634d6544eb4a4825405407b53660d0f5f8b929f7e1671e005b9d92038","src/await.rs":"18f0b2ecb319991f891e300011424985e3cf33d166ea9f29f22d575fc8c83a76","src/bigint.rs":"efc7f64959980653d73fe4f8bc2a3a2904dc05f45b02c6dc15cd316fa3d7c338","src/buffer.rs":"cf2a4b3bdc247b80c85ff5625a1dfb7a5f517fd835f6e1518a7b924990e4c293","src/custom_keyword.rs":"9627467063e41776315a6a14b2aaea3875592d8e0ebd2dc6df1fc2f12c06f146","src/custom_punctuation.rs":"b00e7bee96eb473507527e39db65e74e71592dc06421d2cfe45ed899c17d4847","src/data.rs":"7aec9a745cd53ec95688afa353f6efb9576e7fc0143757b51d28bc3d900b1d2a","src/derive.rs":"fa71866df6e383673dd3329f455a9f953585b83f9739050be3bf1f8c6d526b96","src/discouraged.rs":"a1f3d85e20dedf50b1b7b4571d970a3a6e9b2de4afde7dd0c986fe240df2ba46","src/error.rs":"c3005b50e3132026250c5356d0d391bf96db8087f0f5f744de98e360d8a20a3e","src/export.rs":"dcae67456278c0339acfbcbb4737b8d37cfba5a150ae789f31f4be79abf7e726","src/expr.rs":"54455fd20041996653ca5379b03cdf3c2fc1b3dd2e1149b5bc6b1dd492545d55","src/ext.rs":"870086d9021e6a6fcefa2f00cd91b55c4b74dcee8f0f6a07e76d96fb44707d61","src/file.rs":"75167ebc77e7870122078eabde1b872c337142d4b0962c20cedffcaaa2a5b7c6","src/gen/clone.rs":"0845c1bf8624c3f235cd247b4eb748e7e16b4c240097cb0ff16751f688c079ae","src/gen/debug.rs":"d24fe37f4ce1dd74f2dc54136e893782d3c4d0908323c036c97599551a56960c","src/gen/eq.rs":"1e6ef09b17ca7f36861ef23ce2a6991b231ed5f087f046469b5f23da40f5b419","src/gen/fold.rs":"3f59e59ed8ad2ab5dd347bfbe41bbc785c2aabd8ae902087a584a6daed597182","src/gen/hash.rs":"e5b2a52587173076777233a9e57e2b3c8e0dd6d6f41d16fa7c9fde68b05c2bfc","src/gen/visit.rs":"23008c170d4dd3975232876a0a654921d9b6af57372cb9fcc133ca740588d666","src/gen/visit_mut.rs":"42886c3ee02ded72d9c3eec006e20431eaee0c6b90ddefc1a36ec7bf50c6a24a","src/gen_helper.rs":"ea6c66388365971db6a2fc86cbb208f7eacde77e245bc8623f27a3642a3d7741","src/generics.rs":"d1c175284ca21e777ef0414c28383929b170ccb00aaf7a929eb18d3b05e18da8","src/group.rs":"119b62d8481b4b1c327639bed40e114bf1969765250b68186628247fd4144b3b","src/ident.rs":"503156ce51a7ef0420892e8dbf2ecf8fe51f42a84d52cc2c05654e1a83020cbf","src/item.rs":"c9ad9881e8cda8ee3f157f0c7602fc53d08a7e3288b9afc388c393689eac5aea","src/lib.rs":"558ad13779233b27bebc4b2fc8025eb1c7e57b32130dc1dd911391e27b427500","src/lifetime.rs":"f390fe06692fc51fbf3eb490bb9f795da70e4452f51c5b0df3bbaa899084ddf1","src/lit.rs":"9fab84e38756b092fbb055dcdf01e31d42d916c49e3eaae8c9019043b0ee4301","src/lookahead.rs":"5cce8b4cb345a85c24a452ea6d78eadb76f01ca0a789cbf5ce35108334904173","src/mac.rs":"e5cecea397fd01a44958162781d8d94343fe2a1b9b9754a5666c3d2ab4d7ef64","src/macros.rs":"2ce05b553f14da4ee550bb681cb0733b7186ad94719cd36f96d53e15fd02cf2b","src/op.rs":"449514e146deab0ab020bc6f764544c294dbc780941c9802bf60cf1b2839d550","src/parse.rs":"bde888c98ee259f2a73489a693515ed4875432b0d79486ac83aea19f441992a3","src/parse_macro_input.rs":"653a020f023cac0eccbc1fcc34aa7bf80567b43e5475deab4ad3e487a5363201","src/parse_quote.rs":"642f21e5fa54df4b7c373fb158289ee1005d49e1a49b1d194df5438faee71c46","src/pat.rs":"1473b258162cc822f1ee0c0869f521053ed345a140c39ed83b9b4dfb6f9f2aca","src/path.rs":"f119f0c2af12fabd360eac9a2312e0f6e6c28c633c9671bde6ef0bece7c5ba3c","src/print.rs":"da6529c1d9d21aaf6c835f66b4e67eacb7cf91a10eb5e9a2143b49bf99b3b5e1","src/punctuated.rs":"212f5a601d6c2eb8b8fa679be1167b455b595bee964d2775b0101ebb16c3eaa5","src/reserved.rs":"3625eb2a64589a4992ab79a1674e9679f465bea613ab139a671df5337e88cee6","src/sealed.rs":"896a495a5340eec898527f18bd4ddca408ea03ea0ee3af30074ff48deace778d","src/span.rs":"748c51c6feb223c26d3b1701f5bb98aee823666c775c98106cfa24fe29d8cec1","src/spanned.rs":"7d77714d585e6f42397091ffb3a799fd7b20c05c5442c737683c429ea7d409a5","src/stmt.rs":"3917fbc897f80efe838267833c55650ff8d636cb49a6d1084e28eff65d0e3ccd","src/thread.rs":"815eca6bd64f4eef7c447f0809e84108f5428ff50225224b373efd8fbb696874","src/token.rs":"a1ca6298bf6592cb80cbab1db4eac2fa4e3fa56729bb807bfb0f08ab0f229ca5","src/tt.rs":"1cc9e200624288322f800f32e3d6e2e53da946467bb312dd40a52c02cdcc4730","src/ty.rs":"cb167cbb16240c59a31b44adec175172caaf75ffef9a0bb168584b51bf105795","src/verbatim.rs":"802a97df997432f18cac6e6200ff6ea29fb2474986005e0fcdbc2b65197f87f7","src/whitespace.rs":"e63dd0aa3d34029f17766a8b09c1a6e4479e36c552c8b7023d710a399333aace","tests/common/eq.rs":"4b190a3833bdfd20a4cb1e3dff25a698751dec71d6f30249cf09426e061a4fb1","tests/common/mod.rs":"25ef6d7daa09bad3198a0e9e91b2812425f92db7c585c1e34a03a84d7362ccd8","tests/common/parse.rs":"8b7ba32f4988c30758c108536c4877dc5a039a237bf9b0687220ef2295797bbd","tests/debug/gen.rs":"d6e2abf2a7bb58a7895a60c2f094a98a4f85c9189d02011d0dcef6ef053f26e3","tests/debug/mod.rs":"868763d0ef1609a3ad5e05e9f1bfa0f813e91e7e9a36653414a188bb2fdaa425","tests/macros/mod.rs":"c0eafa4e3845fc08f6efe6021bac37822c0ac325eb7b51194a5f35236f648d92","tests/repo/mod.rs":"9e316b88d57ae213e81950c35e45443078ec90e702798353bc3528cb8a2810b6","tests/repo/progress.rs":"c08d0314a7f3ecf760d471f27da3cd2a500aeb9f1c8331bffb2aa648f9fabf3f","tests/test_asyncness.rs":"cff01db49d28ab23b0b258bc6c0a5cc4071be4fe7248eef344a5d79d2fb649b7","tests/test_attribute.rs":"0ffd99384e1a52ae17d9fed5c4053e411e8f9018decef07ffa621d1faa7329d8","tests/test_derive_input.rs":"610444351e3bf99366976bbf1da109c334a70ac9500caef366bcf9b68819829f","tests/test_expr.rs":"0ee83f6f6de950018c043efcc3e85776b4227dae3068309998a8d9709f2fc66c","tests/test_generics.rs":"9d713f90a79d6145efc89fb6f946029ca03486c632219950889da39940152ba0","tests/test_grouping.rs":"46c27baec4daaaf1e891892f0b0515ea8a44619071c7d0cc9192580916f1569f","tests/test_ident.rs":"9eb53d1e21edf23e7c9e14dc74dcc2b2538e9221e19dbcc0a44e3acc2e90f3f6","tests/test_item.rs":"461ed0c8648afffcea3217f52c9a88298182b4d39d73a11803b1281d99c98c25","tests/test_iterators.rs":"53ed6078d37550bd6765d2411e3660be401aef8a31a407350cc064a7d08c7c33","tests/test_lit.rs":"2a46c5f2f2ad1dcbb7e9b0cd11b55861c5ff818c2c4c51351d07e2daa7c74674","tests/test_meta.rs":"1fc98af3279cadc3d8db3c7e8d4d7f9e9dbd4d17548cf6a2f6f4536ed65367f6","tests/test_parse_buffer.rs":"8bbe2d24ca8a3788f72c6908fc96c26d546f11c69687bf8d72727f851d5e2d27","tests/test_parse_stream.rs":"2f449a2c41a3dee6fd14bee24e1666a453cb808eda17332fd91afd127fcdd2a6","tests/test_pat.rs":"2cb331fe404496d51e7cc7e283ae13c519a2265ca82e1c88e113296f860c2cba","tests/test_path.rs":"fcd5591e639fc787acc9763d828a811c8114525c9341282eefda8f331e082a51","tests/test_precedence.rs":"8d03656741b01e577d7501ce24332d1a4febec3e31a043e47c61062b8c527ed2","tests/test_receiver.rs":"084eca59984b9a18651da52f2c4407355da3de1335916a12477652999e2d01cc","tests/test_round_trip.rs":"ba01bf4ec04cd2d6f9e4800c343563925ae960c5f16752dc0797fda4451b6cc2","tests/test_shebang.rs":"f5772cadad5b56e3112cb16308b779f92bce1c3a48091fc9933deb2276a69331","tests/test_should_parse.rs":"1d3535698a446e2755bfc360676bdb161841a1f454cdef6e7556c6d06a95c89d","tests/test_size.rs":"5fae772bab66809d6708232f35cfb4a287882486763b0f763feec2ad79fbb68b","tests/test_stmt.rs":"17e4355843ee2982b51faba2721a18966f8c2b9422e16b052a123b8ee8b80752","tests/test_token_trees.rs":"43e56a701817e3c3bfd0cae54a457dd7a38ccb3ca19da41e2b995fdf20e6ed18","tests/test_ty.rs":"5b7c0bfc4963d41920dd0b39fdea419e34f00409ba86ad4211d6c3c7e8bbe1c0","tests/test_visibility.rs":"3f958e2b3b5908005e756a80eea326a91eac97cc4ab60599bebde8d4b942d65c","tests/zzz_stable.rs":"2a862e59cb446235ed99aec0e6ada8e16d3ecc30229b29d825b7c0bbc2602989"},"package":"963f7d3cc59b59b9325165add223142bbf1df27655d07789f109896d353d8350"} +\ No newline at end of file +diff --git a/third_party/rust/syn/Cargo.toml b/third_party/rust/syn/Cargo.toml +index 7a5c962f06..20277fc461 100644 +--- a/third_party/rust/syn/Cargo.toml ++++ b/third_party/rust/syn/Cargo.toml +@@ -13,7 +13,7 @@ + [package] + edition = "2018" + name = "syn" +-version = "1.0.5" ++version = "1.0.40" + authors = ["David Tolnay <dtolnay@gmail.com>"] + include = ["/benches/**", "/build.rs", "/Cargo.toml", "/LICENSE-APACHE", "/LICENSE-MIT", "/README.md", "/src/**", "/tests/**"] + description = "Parser for Rust source code" +@@ -24,25 +24,21 @@ license = "MIT OR Apache-2.0" + repository = "https://github.com/dtolnay/syn" + [package.metadata.docs.rs] + all-features = true ++targets = ["x86_64-unknown-linux-gnu"] + + [package.metadata.playground] +-all-features = true +- +-[lib] +-name = "syn" ++features = ["full", "visit", "visit-mut", "fold", "extra-traits"] + + [[bench]] + name = "rust" + harness = false + required-features = ["full", "parsing"] +-edition = "2018" + + [[bench]] + name = "file" + required-features = ["full", "parsing"] +-edition = "2018" + [dependencies.proc-macro2] +-version = "1.0" ++version = "1.0.13" + default-features = false + + [dependencies.quote] +@@ -52,18 +48,34 @@ default-features = false + + [dependencies.unicode-xid] + version = "0.2" ++[dev-dependencies.anyhow] ++version = "1.0" ++ ++[dev-dependencies.flate2] ++version = "1.0" ++ + [dev-dependencies.insta] +-version = "0.9" ++version = "0.16" + + [dev-dependencies.rayon] + version = "1.0" + + [dev-dependencies.ref-cast] +-version = "0.2" ++version = "1.0" + + [dev-dependencies.regex] + version = "1.0" + ++[dev-dependencies.reqwest] ++version = "0.10" ++features = ["blocking"] ++ ++[dev-dependencies.syn-test-suite] ++version = "0" ++ ++[dev-dependencies.tar] ++version = "0.4" ++ + [dev-dependencies.termcolor] + version = "1.0" + +@@ -80,7 +92,6 @@ full = [] + parsing = [] + printing = ["quote"] + proc-macro = ["proc-macro2/proc-macro", "quote/proc-macro"] ++test = ["syn-test-suite/all-features"] + visit = [] + visit-mut = [] +-[badges.travis-ci] +-repository = "dtolnay/syn" +diff --git a/third_party/rust/syn/README.md b/third_party/rust/syn/README.md +index 29a7f32a46..12b5f45b3d 100644 +--- a/third_party/rust/syn/README.md ++++ b/third_party/rust/syn/README.md +@@ -1,10 +1,10 @@ + Parser for Rust source code + =========================== + +-[![Build Status](https://api.travis-ci.org/dtolnay/syn.svg?branch=master)](https://travis-ci.org/dtolnay/syn) +-[![Latest Version](https://img.shields.io/crates/v/syn.svg)](https://crates.io/crates/syn) +-[![Rust Documentation](https://img.shields.io/badge/api-rustdoc-blue.svg)](https://docs.rs/syn/1.0/syn/) +-[![Rustc Version 1.31+](https://img.shields.io/badge/rustc-1.31+-lightgray.svg)](https://blog.rust-lang.org/2018/12/06/Rust-1.31-and-rust-2018.html) ++[<img alt="github" src="https://img.shields.io/badge/github-dtolnay/syn-8da0cb?style=for-the-badge&labelColor=555555&logo=github" height="20">](https://github.com/dtolnay/syn) ++[<img alt="crates.io" src="https://img.shields.io/crates/v/syn.svg?style=for-the-badge&color=fc8d62&logo=rust" height="20">](https://crates.io/crates/syn) ++[<img alt="docs.rs" src="https://img.shields.io/badge/docs.rs-syn-66c2a5?style=for-the-badge&labelColor=555555&logoColor=white&logo=data:image/svg+xml;base64,PHN2ZyByb2xlPSJpbWciIHhtbG5zPSJodHRwOi8vd3d3LnczLm9yZy8yMDAwL3N2ZyIgdmlld0JveD0iMCAwIDUxMiA1MTIiPjxwYXRoIGZpbGw9IiNmNWY1ZjUiIGQ9Ik00ODguNiAyNTAuMkwzOTIgMjE0VjEwNS41YzAtMTUtOS4zLTI4LjQtMjMuNC0zMy43bC0xMDAtMzcuNWMtOC4xLTMuMS0xNy4xLTMuMS0yNS4zIDBsLTEwMCAzNy41Yy0xNC4xIDUuMy0yMy40IDE4LjctMjMuNCAzMy43VjIxNGwtOTYuNiAzNi4yQzkuMyAyNTUuNSAwIDI2OC45IDAgMjgzLjlWMzk0YzAgMTMuNiA3LjcgMjYuMSAxOS45IDMyLjJsMTAwIDUwYzEwLjEgNS4xIDIyLjEgNS4xIDMyLjIgMGwxMDMuOS01MiAxMDMuOSA1MmMxMC4xIDUuMSAyMi4xIDUuMSAzMi4yIDBsMTAwLTUwYzEyLjItNi4xIDE5LjktMTguNiAxOS45LTMyLjJWMjgzLjljMC0xNS05LjMtMjguNC0yMy40LTMzLjd6TTM1OCAyMTQuOGwtODUgMzEuOXYtNjguMmw4NS0zN3Y3My4zek0xNTQgMTA0LjFsMTAyLTM4LjIgMTAyIDM4LjJ2LjZsLTEwMiA0MS40LTEwMi00MS40di0uNnptODQgMjkxLjFsLTg1IDQyLjV2LTc5LjFsODUtMzguOHY3NS40em0wLTExMmwtMTAyIDQxLjQtMTAyLTQxLjR2LS42bDEwMi0zOC4yIDEwMiAzOC4ydi42em0yNDAgMTEybC04NSA0Mi41di03OS4xbDg1LTM4Ljh2NzUuNHptMC0xMTJsLTEwMiA0MS40LTEwMi00MS40di0uNmwxMDItMzguMiAxMDIgMzguMnYuNnoiPjwvcGF0aD48L3N2Zz4K" height="20">](https://docs.rs/syn) ++[<img alt="build status" src="https://img.shields.io/github/workflow/status/dtolnay/syn/CI/master?style=for-the-badge" height="20">](https://github.com/dtolnay/syn/actions?query=branch%3Amaster) + + Syn is a parsing library for parsing a stream of Rust tokens into a syntax tree + of Rust source code. +@@ -46,10 +46,6 @@ contains some APIs that may be useful more generally. + [`syn::DeriveInput`]: https://docs.rs/syn/1.0/syn/struct.DeriveInput.html + [parser functions]: https://docs.rs/syn/1.0/syn/parse/index.html + +-If you get stuck with anything involving procedural macros in Rust I am happy to +-provide help even if the issue is not related to Syn. Please file a ticket in +-this repo. +- + *Version requirement: Syn supports rustc 1.31 and up.* + + [*Release notes*](https://github.com/dtolnay/syn/releases) +@@ -88,8 +84,6 @@ proc-macro = true + ``` + + ```rust +-extern crate proc_macro; +- + use proc_macro::TokenStream; + use quote::quote; + use syn::{parse_macro_input, DeriveInput}; +@@ -271,7 +265,7 @@ points, which are required by the language to use `proc_macro::TokenStream`. + The proc-macro2 crate will automatically detect and use the compiler's data + structures when a procedural macro is active. + +-[proc-macro2]: https://docs.rs/proc-macro2/1.0.0/proc_macro2/ ++[proc-macro2]: https://docs.rs/proc-macro2/1.0/proc_macro2/ + + <br> + +diff --git a/third_party/rust/syn/benches/file.rs b/third_party/rust/syn/benches/file.rs +index 08ecd90960..58ab8df297 100644 +--- a/third_party/rust/syn/benches/file.rs ++++ b/third_party/rust/syn/benches/file.rs +@@ -1,9 +1,16 @@ + // $ cargo bench --features full --bench file + + #![feature(rustc_private, test)] ++#![recursion_limit = "1024"] + + extern crate test; + ++#[macro_use] ++#[path = "../tests/macros/mod.rs"] ++mod macros; ++ ++#[path = "../tests/common/mod.rs"] ++mod common; + #[path = "../tests/repo/mod.rs"] + pub mod repo; + +diff --git a/third_party/rust/syn/benches/rust.rs b/third_party/rust/syn/benches/rust.rs +index e3d9cd29ba..50e1a7f601 100644 +--- a/third_party/rust/syn/benches/rust.rs ++++ b/third_party/rust/syn/benches/rust.rs +@@ -4,7 +4,14 @@ + // $ RUSTFLAGS='--cfg syn_only' cargo build --release --features full --bench rust + + #![cfg_attr(not(syn_only), feature(rustc_private))] ++#![recursion_limit = "1024"] + ++#[macro_use] ++#[path = "../tests/macros/mod.rs"] ++mod macros; ++ ++#[path = "../tests/common/mod.rs"] ++mod common; + #[path = "../tests/repo/mod.rs"] + mod repo; + +@@ -28,31 +35,35 @@ mod syn_parse { + } + + #[cfg(not(syn_only))] +-mod libsyntax_parse { ++mod librustc_parse { + extern crate rustc_data_structures; +- extern crate syntax; +- extern crate syntax_pos; ++ extern crate rustc_errors; ++ extern crate rustc_parse; ++ extern crate rustc_session; ++ extern crate rustc_span; + + use rustc_data_structures::sync::Lrc; +- use syntax::edition::Edition; +- use syntax::errors::{emitter::Emitter, DiagnosticBuilder, Handler}; +- use syntax::parse::ParseSess; +- use syntax::source_map::{FilePathMapping, SourceMap}; +- use syntax_pos::FileName; ++ use rustc_errors::{emitter::Emitter, Diagnostic, Handler}; ++ use rustc_session::parse::ParseSess; ++ use rustc_span::source_map::{FilePathMapping, SourceMap}; ++ use rustc_span::{edition::Edition, FileName}; + + pub fn bench(content: &str) -> Result<(), ()> { + struct SilentEmitter; + + impl Emitter for SilentEmitter { +- fn emit_diagnostic(&mut self, _db: &DiagnosticBuilder) {} ++ fn emit_diagnostic(&mut self, _diag: &Diagnostic) {} ++ fn source_map(&self) -> Option<&Lrc<SourceMap>> { ++ None ++ } + } + +- syntax::with_globals(Edition::Edition2018, || { ++ rustc_span::with_session_globals(Edition::Edition2018, || { + let cm = Lrc::new(SourceMap::new(FilePathMapping::empty())); + let emitter = Box::new(SilentEmitter); + let handler = Handler::with_emitter(false, None, emitter); + let sess = ParseSess::with_span_handler(handler, cm); +- if let Err(mut diagnostic) = syntax::parse::parse_crate_from_source_str( ++ if let Err(mut diagnostic) = rustc_parse::parse_crate_from_source_str( + FileName::Custom("bench".to_owned()), + content.to_owned(), + &sess, +@@ -104,11 +115,11 @@ fn main() { + repo::clone_rust(); + + macro_rules! testcases { +- ($($(#[$cfg:meta])* $name:path,)*) => { ++ ($($(#[$cfg:meta])* $name:ident,)*) => { + vec![ + $( + $(#[$cfg])* +- (stringify!($name), $name as fn(&str) -> Result<(), ()>), ++ (stringify!($name), $name::bench as fn(&str) -> Result<(), ()>), + )* + ] + }; +@@ -128,12 +139,12 @@ fn main() { + + for (name, f) in testcases!( + #[cfg(not(syn_only))] +- read_from_disk::bench, ++ read_from_disk, + #[cfg(not(syn_only))] +- tokenstream_parse::bench, +- syn_parse::bench, ++ tokenstream_parse, ++ syn_parse, + #[cfg(not(syn_only))] +- libsyntax_parse::bench, ++ librustc_parse, + ) { + eprint!("{:20}", format!("{}:", name)); + let elapsed = exec(f); +diff --git a/third_party/rust/syn/build.rs b/third_party/rust/syn/build.rs +index c0f9ed3406..cf7681c3f9 100644 +--- a/third_party/rust/syn/build.rs ++++ b/third_party/rust/syn/build.rs +@@ -1,6 +1,6 @@ + use std::env; + use std::process::Command; +-use std::str::{self, FromStr}; ++use std::str; + + // The rustc-cfg strings below are *not* public API. Please let us know by + // opening a GitHub issue if your build environment requires some way to enable +@@ -26,38 +26,14 @@ struct Compiler { + } + + fn rustc_version() -> Option<Compiler> { +- let rustc = match env::var_os("RUSTC") { +- Some(rustc) => rustc, +- None => return None, +- }; +- +- let output = match Command::new(rustc).arg("--version").output() { +- Ok(output) => output, +- Err(_) => return None, +- }; +- +- let version = match str::from_utf8(&output.stdout) { +- Ok(version) => version, +- Err(_) => return None, +- }; +- ++ let rustc = env::var_os("RUSTC")?; ++ let output = Command::new(rustc).arg("--version").output().ok()?; ++ let version = str::from_utf8(&output.stdout).ok()?; + let mut pieces = version.split('.'); + if pieces.next() != Some("rustc 1") { + return None; + } +- +- let next = match pieces.next() { +- Some(next) => next, +- None => return None, +- }; +- +- let minor = match u32::from_str(next) { +- Ok(minor) => minor, +- Err(_) => return None, +- }; +- +- Some(Compiler { +- minor: minor, +- nightly: version.contains("nightly"), +- }) ++ let minor = pieces.next()?.parse().ok()?; ++ let nightly = version.contains("nightly"); ++ Some(Compiler { minor, nightly }) + } +diff --git a/third_party/rust/syn/src/attr.rs b/third_party/rust/syn/src/attr.rs +index 34009deabc..fa4f1cb2a3 100644 +--- a/third_party/rust/syn/src/attr.rs ++++ b/third_party/rust/syn/src/attr.rs +@@ -9,15 +9,11 @@ use proc_macro2::TokenStream; + use crate::parse::{Parse, ParseBuffer, ParseStream, Parser, Result}; + #[cfg(feature = "parsing")] + use crate::punctuated::Pair; +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; +-#[cfg(feature = "extra-traits")] +-use std::hash::{Hash, Hasher}; + + ast_struct! { + /// An attribute like `#[repr(transparent)]`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// <br> +@@ -111,7 +107,46 @@ ast_struct! { + /// + /// [`parse_meta()`]: Attribute::parse_meta + /// [`parse_args()`]: Attribute::parse_args +- pub struct Attribute #manual_extra_traits { ++ /// ++ /// <p><br></p> ++ /// ++ /// # Doc comments ++ /// ++ /// The compiler transforms doc comments, such as `/// comment` and `/*! ++ /// comment */`, into attributes before macros are expanded. Each comment is ++ /// expanded into an attribute of the form `#[doc = r"comment"]`. ++ /// ++ /// As an example, the following `mod` items are expanded identically: ++ /// ++ /// ``` ++ /// # use syn::{ItemMod, parse_quote}; ++ /// let doc: ItemMod = parse_quote! { ++ /// /// Single line doc comments ++ /// /// We write so many! ++ /// /** ++ /// * Multi-line comments... ++ /// * May span many lines ++ /// */ ++ /// mod example { ++ /// //! Of course, they can be inner too ++ /// /*! And fit in a single line */ ++ /// } ++ /// }; ++ /// let attr: ItemMod = parse_quote! { ++ /// #[doc = r" Single line doc comments"] ++ /// #[doc = r" We write so many!"] ++ /// #[doc = r" ++ /// * Multi-line comments... ++ /// * May span many lines ++ /// "] ++ /// mod example { ++ /// #![doc = r" Of course, they can be inner too"] ++ /// #![doc = r" And fit in a single line "] ++ /// } ++ /// }; ++ /// assert_eq!(doc, attr); ++ /// ``` ++ pub struct Attribute { + pub pound_token: Token![#], + pub style: AttrStyle, + pub bracket_token: token::Bracket, +@@ -120,39 +155,11 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Attribute {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Attribute { +- fn eq(&self, other: &Self) -> bool { +- self.style == other.style +- && self.pound_token == other.pound_token +- && self.bracket_token == other.bracket_token +- && self.path == other.path +- && TokenStreamHelper(&self.tokens) == TokenStreamHelper(&other.tokens) +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Attribute { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- self.style.hash(state); +- self.pound_token.hash(state); +- self.bracket_token.hash(state); +- self.path.hash(state); +- TokenStreamHelper(&self.tokens).hash(state); +- } +-} +- + impl Attribute { + /// Parses the content of the attribute, consisting of the path and tokens, + /// as a [`Meta`] if possible. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_meta(&self) -> Result<Meta> { +@@ -199,7 +206,7 @@ impl Attribute { + /// ^^^^^^^^^ what gets parsed + /// ``` + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_args<T: Parse>(&self) -> Result<T> { +@@ -208,7 +215,7 @@ impl Attribute { + + /// Parse the arguments to the attribute using the given parser. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_args_with<F: Parser>(&self, parser: F) -> Result<F::Output> { +@@ -221,7 +228,7 @@ impl Attribute { + + /// Parses zero or more outer attributes from the stream. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_outer(input: ParseStream) -> Result<Vec<Self>> { +@@ -234,7 +241,7 @@ impl Attribute { + + /// Parses zero or more inner attributes from the stream. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_inner(input: ParseStream) -> Result<Vec<Self>> { +@@ -247,7 +254,7 @@ impl Attribute { + } + + #[cfg(feature = "parsing")] +-fn error_expected_args(attr: &Attribute) -> Error { ++fn expected_parentheses(attr: &Attribute) -> String { + let style = match attr.style { + AttrStyle::Outer => "#", + AttrStyle::Inner(_) => "#!", +@@ -261,19 +268,23 @@ fn error_expected_args(attr: &Attribute) -> Error { + path += &segment.ident.to_string(); + } + +- let msg = format!("expected attribute arguments: {}[{}(...)]", style, path); +- +- #[cfg(feature = "printing")] +- return Error::new_spanned(attr, msg); +- +- #[cfg(not(feature = "printing"))] +- return Error::new(attr.bracket_token.span, msg); ++ format!("{}[{}(...)]", style, path) + } + + #[cfg(feature = "parsing")] + fn enter_args<'a>(attr: &Attribute, input: ParseStream<'a>) -> Result<ParseBuffer<'a>> { + if input.is_empty() { +- return Err(error_expected_args(attr)); ++ let expected = expected_parentheses(attr); ++ let msg = format!("expected attribute arguments in parentheses: {}", expected); ++ return Err(crate::error::new2( ++ attr.pound_token.span, ++ attr.bracket_token.span, ++ msg, ++ )); ++ } else if input.peek(Token![=]) { ++ let expected = expected_parentheses(attr); ++ let msg = format!("expected parentheses: {}", expected); ++ return Err(input.error(msg)); + }; + + let content; +@@ -298,7 +309,7 @@ ast_enum! { + /// Distinguishes between attributes that decorate an item and attributes + /// that are contained within an item. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Outer attributes +@@ -312,7 +323,6 @@ ast_enum! { + /// - `#![feature(proc_macro)]` + /// - `//! # Example` + /// - `/*! Please file an issue */` +- #[cfg_attr(feature = "clone-impls", derive(Copy))] + pub enum AttrStyle { + Outer, + Inner(Token![!]), +@@ -322,7 +332,7 @@ ast_enum! { + ast_enum_of_structs! { + /// Content of a compile-time structured attribute. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// ## Path +@@ -360,7 +370,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A structured list within an attribute, like `derive(Copy, Clone)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct MetaList { + pub path: Path, +@@ -372,7 +382,7 @@ ast_struct! { + ast_struct! { + /// A name-value pair within an attribute, like `feature = "nightly"`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct MetaNameValue { + pub path: Path, +@@ -398,7 +408,7 @@ impl Meta { + ast_enum_of_structs! { + /// Element of a compile-time attribute list. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum NestedMeta { + /// A structured meta item, like the `Copy` in `#[derive(Copy)]` which +@@ -429,8 +439,8 @@ ast_enum_of_structs! { + /// as type `AttributeArgs`. + /// + /// ``` +-/// extern crate proc_macro; +-/// ++/// # extern crate proc_macro; ++/// # + /// use proc_macro::TokenStream; + /// use syn::{parse_macro_input, AttributeArgs, ItemFn}; + /// +@@ -464,7 +474,7 @@ where + fn is_outer(attr: &&Attribute) -> bool { + match attr.style { + AttrStyle::Outer => true, +- _ => false, ++ AttrStyle::Inner(_) => false, + } + } + self.into_iter().filter(is_outer) +@@ -474,7 +484,7 @@ where + fn is_inner(attr: &&Attribute) -> bool { + match attr.style { + AttrStyle::Inner(_) => true, +- _ => false, ++ AttrStyle::Outer => false, + } + } + self.into_iter().filter(is_inner) +diff --git a/third_party/rust/syn/src/buffer.rs b/third_party/rust/syn/src/buffer.rs +index 551a5ac816..a461cc49ea 100644 +--- a/third_party/rust/syn/src/buffer.rs ++++ b/third_party/rust/syn/src/buffer.rs +@@ -1,7 +1,7 @@ + //! A stably addressed token buffer supporting efficient traversal based on a + //! cheaply copyable cursor. + //! +-//! *This module is available if Syn is built with the `"parsing"` feature.* ++//! *This module is available only if Syn is built with the `"parsing"` feature.* + + // This module is heavily commented as it contains most of the unsafe code in + // Syn, and caution should be used when editing it. The public-facing interface +@@ -36,7 +36,7 @@ enum Entry { + /// `TokenStream` which requires a deep copy in order to traverse more than + /// once. + /// +-/// *This type is available if Syn is built with the `"parsing"` feature.* ++/// *This type is available only if Syn is built with the `"parsing"` feature.* + pub struct TokenBuffer { + // NOTE: Do not derive clone on this - there are raw pointers inside which + // will be messed up. Moving the `TokenBuffer` itself is safe as the actual +@@ -98,7 +98,7 @@ impl TokenBuffer { + /// Creates a `TokenBuffer` containing all the tokens from the input + /// `TokenStream`. + /// +- /// *This method is available if Syn is built with both the `"parsing"` and ++ /// *This method is available only if Syn is built with both the `"parsing"` and + /// `"proc-macro"` features.* + #[cfg(all( + not(all(target_arch = "wasm32", any(target_os = "unknown", target_os = "wasi"))), +@@ -133,8 +133,7 @@ impl TokenBuffer { + /// Two cursors are equal if they have the same location in the same input + /// stream, and have the same scope. + /// +-/// *This type is available if Syn is built with the `"parsing"` feature.* +-#[derive(Copy, Clone, Eq, PartialEq)] ++/// *This type is available only if Syn is built with the `"parsing"` feature.* + pub struct Cursor<'a> { + // The current entry which the `Cursor` is pointing at. + ptr: *const Entry, +@@ -201,13 +200,13 @@ impl<'a> Cursor<'a> { + Cursor::create(self.ptr.offset(1), self.scope) + } + +- /// If the cursor is looking at a `None`-delimited group, move it to look at +- /// the first token inside instead. If the group is empty, this will move ++ /// While the cursor is looking at a `None`-delimited group, move it to look ++ /// at the first token inside instead. If the group is empty, this will move + /// the cursor past the `None`-delimited group. + /// + /// WARNING: This mutates its argument. + fn ignore_none(&mut self) { +- if let Entry::Group(group, buf) = self.entry() { ++ while let Entry::Group(group, buf) = self.entry() { + if group.delimiter() == Delimiter::None { + // NOTE: We call `Cursor::create` here to make sure that + // situations where we should immediately exit the span after +@@ -215,13 +214,14 @@ impl<'a> Cursor<'a> { + unsafe { + *self = Cursor::create(&buf.data[0], self.scope); + } ++ } else { ++ break; + } + } + } + + /// Checks whether the cursor is currently pointing at the end of its valid + /// scope. +- #[inline] + pub fn eof(self) -> bool { + // We're at eof if we're at the end of our scope. + self.ptr == self.scope +@@ -342,6 +342,44 @@ impl<'a> Cursor<'a> { + Entry::End(..) => Span::call_site(), + } + } ++ ++ /// Skip over the next token without cloning it. Returns `None` if this ++ /// cursor points to eof. ++ /// ++ /// This method treats `'lifetimes` as a single token. ++ pub(crate) fn skip(self) -> Option<Cursor<'a>> { ++ match self.entry() { ++ Entry::End(..) => None, ++ ++ // Treat lifetimes as a single tt for the purposes of 'skip'. ++ Entry::Punct(op) if op.as_char() == '\'' && op.spacing() == Spacing::Joint => { ++ let next = unsafe { self.bump() }; ++ match next.entry() { ++ Entry::Ident(_) => Some(unsafe { next.bump() }), ++ _ => Some(next), ++ } ++ } ++ _ => Some(unsafe { self.bump() }), ++ } ++ } ++} ++ ++impl<'a> Copy for Cursor<'a> {} ++ ++impl<'a> Clone for Cursor<'a> { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++ ++impl<'a> Eq for Cursor<'a> {} ++ ++impl<'a> PartialEq for Cursor<'a> { ++ fn eq(&self, other: &Self) -> bool { ++ let Cursor { ptr, scope, marker } = self; ++ let _ = marker; ++ *ptr == other.ptr && *scope == other.scope ++ } + } + + pub(crate) fn same_scope(a: Cursor, b: Cursor) -> bool { +diff --git a/third_party/rust/syn/src/custom_keyword.rs b/third_party/rust/syn/src/custom_keyword.rs +index 200e8478ef..a33044a564 100644 +--- a/third_party/rust/syn/src/custom_keyword.rs ++++ b/third_party/rust/syn/src/custom_keyword.rs +@@ -86,7 +86,7 @@ + /// } + /// } + /// ``` +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! custom_keyword { + ($ident:ident) => { + #[allow(non_camel_case_types)] +@@ -95,7 +95,7 @@ macro_rules! custom_keyword { + } + + #[doc(hidden)] +- #[allow(non_snake_case)] ++ #[allow(dead_code, non_snake_case)] + pub fn $ident<__S: $crate::export::IntoSpans<[$crate::export::Span; 1]>>( + span: __S, + ) -> $ident { +@@ -112,10 +112,10 @@ macro_rules! custom_keyword { + } + } + +- impl_parse_for_custom_keyword!($ident); +- impl_to_tokens_for_custom_keyword!($ident); +- impl_clone_for_custom_keyword!($ident); +- impl_extra_traits_for_custom_keyword!($ident); ++ $crate::impl_parse_for_custom_keyword!($ident); ++ $crate::impl_to_tokens_for_custom_keyword!($ident); ++ $crate::impl_clone_for_custom_keyword!($ident); ++ $crate::impl_extra_traits_for_custom_keyword!($ident); + }; + } + +diff --git a/third_party/rust/syn/src/custom_punctuation.rs b/third_party/rust/syn/src/custom_punctuation.rs +index 29fa448bd8..70dff42851 100644 +--- a/third_party/rust/syn/src/custom_punctuation.rs ++++ b/third_party/rust/syn/src/custom_punctuation.rs +@@ -74,19 +74,19 @@ + /// let _: PathSegments = syn::parse_str(input).unwrap(); + /// } + /// ``` +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! custom_punctuation { + ($ident:ident, $($tt:tt)+) => { + pub struct $ident { +- pub spans: custom_punctuation_repr!($($tt)+), ++ pub spans: $crate::custom_punctuation_repr!($($tt)+), + } + + #[doc(hidden)] +- #[allow(non_snake_case)] +- pub fn $ident<__S: $crate::export::IntoSpans<custom_punctuation_repr!($($tt)+)>>( ++ #[allow(dead_code, non_snake_case)] ++ pub fn $ident<__S: $crate::export::IntoSpans<$crate::custom_punctuation_repr!($($tt)+)>>( + spans: __S, + ) -> $ident { +- let _validate_len = 0 $(+ custom_punctuation_len!(strict, $tt))*; ++ let _validate_len = 0 $(+ $crate::custom_punctuation_len!(strict, $tt))*; + $ident { + spans: $crate::export::IntoSpans::into_spans(spans) + } +@@ -98,33 +98,33 @@ macro_rules! custom_punctuation { + } + } + +- impl_parse_for_custom_punctuation!($ident, $($tt)+); +- impl_to_tokens_for_custom_punctuation!($ident, $($tt)+); +- impl_clone_for_custom_punctuation!($ident, $($tt)+); +- impl_extra_traits_for_custom_punctuation!($ident, $($tt)+); ++ $crate::impl_parse_for_custom_punctuation!($ident, $($tt)+); ++ $crate::impl_to_tokens_for_custom_punctuation!($ident, $($tt)+); ++ $crate::impl_clone_for_custom_punctuation!($ident, $($tt)+); ++ $crate::impl_extra_traits_for_custom_punctuation!($ident, $($tt)+); + }; + } + + // Not public API. + #[cfg(feature = "parsing")] + #[doc(hidden)] +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! impl_parse_for_custom_punctuation { + ($ident:ident, $($tt:tt)+) => { + impl $crate::token::CustomToken for $ident { + fn peek(cursor: $crate::buffer::Cursor) -> bool { +- $crate::token::parsing::peek_punct(cursor, stringify_punct!($($tt)+)) ++ $crate::token::parsing::peek_punct(cursor, $crate::stringify_punct!($($tt)+)) + } + + fn display() -> &'static $crate::export::str { +- custom_punctuation_concat!("`", stringify_punct!($($tt)+), "`") ++ concat!("`", $crate::stringify_punct!($($tt)+), "`") + } + } + + impl $crate::parse::Parse for $ident { + fn parse(input: $crate::parse::ParseStream) -> $crate::parse::Result<$ident> { +- let spans: custom_punctuation_repr!($($tt)+) = +- $crate::token::parsing::punct(input, stringify_punct!($($tt)+))?; ++ let spans: $crate::custom_punctuation_repr!($($tt)+) = ++ $crate::token::parsing::punct(input, $crate::stringify_punct!($($tt)+))?; + Ok($ident(spans)) + } + } +@@ -142,12 +142,12 @@ macro_rules! impl_parse_for_custom_punctuation { + // Not public API. + #[cfg(feature = "printing")] + #[doc(hidden)] +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! impl_to_tokens_for_custom_punctuation { + ($ident:ident, $($tt:tt)+) => { + impl $crate::export::ToTokens for $ident { + fn to_tokens(&self, tokens: &mut $crate::export::TokenStream2) { +- $crate::token::printing::punct(stringify_punct!($($tt)+), &self.spans, tokens) ++ $crate::token::printing::punct($crate::stringify_punct!($($tt)+), &self.spans, tokens) + } + } + }; +@@ -221,16 +221,16 @@ macro_rules! impl_extra_traits_for_custom_punctuation { + + // Not public API. + #[doc(hidden)] +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! custom_punctuation_repr { + ($($tt:tt)+) => { +- [$crate::export::Span; 0 $(+ custom_punctuation_len!(lenient, $tt))+] ++ [$crate::export::Span; 0 $(+ $crate::custom_punctuation_len!(lenient, $tt))+] + }; + } + + // Not public API. + #[doc(hidden)] +-#[macro_export(local_inner_macros)] ++#[macro_export] + #[rustfmt::skip] + macro_rules! custom_punctuation_len { + ($mode:ident, +) => { 1 }; +@@ -279,7 +279,7 @@ macro_rules! custom_punctuation_len { + ($mode:ident, -=) => { 2 }; + ($mode:ident, ~) => { 1 }; + (lenient, $tt:tt) => { 0 }; +- (strict, $tt:tt) => {{ custom_punctuation_unexpected!($tt); 0 }}; ++ (strict, $tt:tt) => {{ $crate::custom_punctuation_unexpected!($tt); 0 }}; + } + + // Not public API. +@@ -297,13 +297,3 @@ macro_rules! stringify_punct { + concat!($(stringify!($tt)),+) + }; + } +- +-// Not public API. +-// Without this, local_inner_macros breaks when looking for concat! +-#[doc(hidden)] +-#[macro_export] +-macro_rules! custom_punctuation_concat { +- ($($tt:tt)*) => { +- concat!($($tt)*) +- }; +-} +diff --git a/third_party/rust/syn/src/data.rs b/third_party/rust/syn/src/data.rs +index be43679874..b217b8ca6f 100644 +--- a/third_party/rust/syn/src/data.rs ++++ b/third_party/rust/syn/src/data.rs +@@ -4,7 +4,7 @@ use crate::punctuated::Punctuated; + ast_struct! { + /// An enum variant. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Variant { + /// Attributes tagged on the variant. +@@ -24,7 +24,7 @@ ast_struct! { + ast_enum_of_structs! { + /// Data stored within an enum variant or struct. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Syntax tree enum +@@ -52,7 +52,7 @@ ast_struct! { + /// Named fields of a struct or struct variant such as `Point { x: f64, + /// y: f64 }`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct FieldsNamed { + pub brace_token: token::Brace, +@@ -63,7 +63,7 @@ ast_struct! { + ast_struct! { + /// Unnamed fields of a tuple struct or tuple variant such as `Some(T)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct FieldsUnnamed { + pub paren_token: token::Paren, +@@ -93,6 +93,24 @@ impl Fields { + Fields::Unnamed(f) => f.unnamed.iter_mut(), + } + } ++ ++ /// Returns the number of fields. ++ pub fn len(&self) -> usize { ++ match self { ++ Fields::Unit => 0, ++ Fields::Named(f) => f.named.len(), ++ Fields::Unnamed(f) => f.unnamed.len(), ++ } ++ } ++ ++ /// Returns `true` if there are zero fields. ++ pub fn is_empty(&self) -> bool { ++ match self { ++ Fields::Unit => true, ++ Fields::Named(f) => f.named.is_empty(), ++ Fields::Unnamed(f) => f.unnamed.is_empty(), ++ } ++ } + } + + impl IntoIterator for Fields { +@@ -129,7 +147,7 @@ impl<'a> IntoIterator for &'a mut Fields { + ast_struct! { + /// A field of a struct or enum variant. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Field { + /// Attributes tagged on the field. +@@ -154,7 +172,7 @@ ast_enum_of_structs! { + /// The visibility level of an item: inherited or `pub` or + /// `pub(restricted)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Syntax tree enum +@@ -184,7 +202,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A public visibility level: `pub`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct VisPublic { + pub pub_token: Token![pub], +@@ -194,7 +212,7 @@ ast_struct! { + ast_struct! { + /// A crate-level visibility: `crate`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct VisCrate { + pub crate_token: Token![crate], +@@ -205,7 +223,7 @@ ast_struct! { + /// A visibility level restricted to some path: `pub(self)` or + /// `pub(super)` or `pub(crate)` or `pub(in some::module)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct VisRestricted { + pub pub_token: Token![pub], +@@ -220,12 +238,15 @@ pub mod parsing { + use super::*; + + use crate::ext::IdentExt; ++ use crate::parse::discouraged::Speculative; + use crate::parse::{Parse, ParseStream, Result}; + + impl Parse for Variant { + fn parse(input: ParseStream) -> Result<Self> { ++ let attrs = input.call(Attribute::parse_outer)?; ++ let _visibility: Visibility = input.parse()?; + Ok(Variant { +- attrs: input.call(Attribute::parse_outer)?, ++ attrs, + ident: input.parse()?, + fields: { + if input.peek(token::Brace) { +@@ -295,6 +316,17 @@ pub mod parsing { + + impl Parse for Visibility { + fn parse(input: ParseStream) -> Result<Self> { ++ // Recognize an empty None-delimited group, as produced by a $:vis ++ // matcher that matched no tokens. ++ if input.peek(token::Group) { ++ let ahead = input.fork(); ++ let group = crate::group::parse_group(&ahead)?; ++ if group.content.is_empty() { ++ input.advance_to(&ahead); ++ return Ok(Visibility::Inherited); ++ } ++ } ++ + if input.peek(Token![pub]) { + Self::parse_pub(input) + } else if input.peek(Token![crate]) { +@@ -310,27 +342,39 @@ pub mod parsing { + let pub_token = input.parse::<Token![pub]>()?; + + if input.peek(token::Paren) { +- // TODO: optimize using advance_to + let ahead = input.fork(); +- let mut content; +- parenthesized!(content in ahead); + ++ let content; ++ let paren_token = parenthesized!(content in ahead); + if content.peek(Token![crate]) + || content.peek(Token![self]) + || content.peek(Token![super]) + { +- return Ok(Visibility::Restricted(VisRestricted { +- pub_token, +- paren_token: parenthesized!(content in input), +- in_token: None, +- path: Box::new(Path::from(content.call(Ident::parse_any)?)), +- })); ++ let path = content.call(Ident::parse_any)?; ++ ++ // Ensure there are no additional tokens within `content`. ++ // Without explicitly checking, we may misinterpret a tuple ++ // field as a restricted visibility, causing a parse error. ++ // e.g. `pub (crate::A, crate::B)` (Issue #720). ++ if content.is_empty() { ++ input.advance_to(&ahead); ++ return Ok(Visibility::Restricted(VisRestricted { ++ pub_token, ++ paren_token, ++ in_token: None, ++ path: Box::new(Path::from(path)), ++ })); ++ } + } else if content.peek(Token![in]) { ++ let in_token: Token![in] = content.parse()?; ++ let path = content.call(Path::parse_mod_style)?; ++ ++ input.advance_to(&ahead); + return Ok(Visibility::Restricted(VisRestricted { + pub_token, +- paren_token: parenthesized!(content in input), +- in_token: Some(content.parse()?), +- path: Box::new(content.call(Path::parse_mod_style)?), ++ paren_token, ++ in_token: Some(in_token), ++ path: Box::new(path), + })); + } + } +@@ -347,6 +391,14 @@ pub mod parsing { + })) + } + } ++ ++ #[cfg(feature = "full")] ++ pub(crate) fn is_some(&self) -> bool { ++ match self { ++ Visibility::Inherited => false, ++ _ => true, ++ } ++ } + } + } + +diff --git a/third_party/rust/syn/src/derive.rs b/third_party/rust/syn/src/derive.rs +index 8cb9cf7b6d..3fa9d89a93 100644 +--- a/third_party/rust/syn/src/derive.rs ++++ b/third_party/rust/syn/src/derive.rs +@@ -4,7 +4,7 @@ use crate::punctuated::Punctuated; + ast_struct! { + /// Data structure sent to a `proc_macro_derive` macro. + /// +- /// *This type is available if Syn is built with the `"derive"` feature.* ++ /// *This type is available only if Syn is built with the `"derive"` feature.* + pub struct DeriveInput { + /// Attributes tagged on the whole struct or enum. + pub attrs: Vec<Attribute>, +@@ -26,7 +26,7 @@ ast_struct! { + ast_enum_of_structs! { + /// The storage of a struct, enum or union data structure. + /// +- /// *This type is available if Syn is built with the `"derive"` feature.* ++ /// *This type is available only if Syn is built with the `"derive"` feature.* + /// + /// # Syntax tree enum + /// +@@ -53,7 +53,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A struct input to a `proc_macro_derive` macro. + /// +- /// *This type is available if Syn is built with the `"derive"` ++ /// *This type is available only if Syn is built with the `"derive"` + /// feature.* + pub struct DataStruct { + pub struct_token: Token![struct], +@@ -65,7 +65,7 @@ ast_struct! { + ast_struct! { + /// An enum input to a `proc_macro_derive` macro. + /// +- /// *This type is available if Syn is built with the `"derive"` ++ /// *This type is available only if Syn is built with the `"derive"` + /// feature.* + pub struct DataEnum { + pub enum_token: Token![enum], +@@ -77,7 +77,7 @@ ast_struct! { + ast_struct! { + /// An untagged union input to a `proc_macro_derive` macro. + /// +- /// *This type is available if Syn is built with the `"derive"` ++ /// *This type is available only if Syn is built with the `"derive"` + /// feature.* + pub struct DataUnion { + pub union_token: Token![union], +diff --git a/third_party/rust/syn/src/discouraged.rs b/third_party/rust/syn/src/discouraged.rs +index 4d9ff93728..76c9fce6f8 100644 +--- a/third_party/rust/syn/src/discouraged.rs ++++ b/third_party/rust/syn/src/discouraged.rs +@@ -16,7 +16,7 @@ pub trait Speculative { + /// syntax of the form `A* B*` for arbitrary syntax `A` and `B`. The problem + /// is that when the fork fails to parse an `A`, it's impossible to tell + /// whether that was because of a syntax error and the user meant to provide +- /// an `A`, or that the `A`s are finished and its time to start parsing ++ /// an `A`, or that the `A`s are finished and it's time to start parsing + /// `B`s. Use with care. + /// + /// Also note that if `A` is a subset of `B`, `A* B*` can be parsed by +@@ -72,7 +72,6 @@ pub trait Speculative { + /// || input.peek(Token![self]) + /// || input.peek(Token![Self]) + /// || input.peek(Token![crate]) +- /// || input.peek(Token![extern]) + /// { + /// let ident = input.call(Ident::parse_any)?; + /// return Ok(PathSegment::from(ident)); +@@ -164,6 +163,30 @@ impl<'a> Speculative for ParseBuffer<'a> { + panic!("Fork was not derived from the advancing parse stream"); + } + ++ let (self_unexp, self_sp) = inner_unexpected(self); ++ let (fork_unexp, fork_sp) = inner_unexpected(fork); ++ if !Rc::ptr_eq(&self_unexp, &fork_unexp) { ++ match (fork_sp, self_sp) { ++ // Unexpected set on the fork, but not on `self`, copy it over. ++ (Some(span), None) => { ++ self_unexp.set(Unexpected::Some(span)); ++ } ++ // Unexpected unset. Use chain to propagate errors from fork. ++ (None, None) => { ++ fork_unexp.set(Unexpected::Chain(self_unexp)); ++ ++ // Ensure toplevel 'unexpected' tokens from the fork don't ++ // bubble up the chain by replacing the root `unexpected` ++ // pointer, only 'unexpected' tokens from existing group ++ // parsers should bubble. ++ fork.unexpected ++ .set(Some(Rc::new(Cell::new(Unexpected::None)))); ++ } ++ // Unexpected has been set on `self`. No changes needed. ++ (_, Some(_)) => {} ++ } ++ } ++ + // See comment on `cell` in the struct definition. + self.cell + .set(unsafe { mem::transmute::<Cursor, Cursor<'static>>(fork.cursor()) }) +diff --git a/third_party/rust/syn/src/error.rs b/third_party/rust/syn/src/error.rs +index 146d652299..dba34f9254 100644 +--- a/third_party/rust/syn/src/error.rs ++++ b/third_party/rust/syn/src/error.rs +@@ -1,4 +1,3 @@ +-use std; + use std::fmt::{self, Debug, Display}; + use std::iter::FromIterator; + use std::slice; +@@ -32,8 +31,8 @@ pub type Result<T> = std::result::Result<T, Error>; + /// conversion to `compile_error!` automatically. + /// + /// ``` +-/// extern crate proc_macro; +-/// ++/// # extern crate proc_macro; ++/// # + /// use proc_macro::TokenStream; + /// use syn::{parse_macro_input, AttributeArgs, ItemFn}; + /// +@@ -82,7 +81,6 @@ pub type Result<T> = std::result::Result<T, Error>; + /// # } + /// # } + /// ``` +-#[derive(Clone)] + pub struct Error { + messages: Vec<ErrorMessage>, + } +@@ -250,6 +248,17 @@ pub fn new_at<T: Display>(scope: Span, cursor: Cursor, message: T) -> Error { + } + } + ++#[cfg(all(feature = "parsing", any(feature = "full", feature = "derive")))] ++pub fn new2<T: Display>(start: Span, end: Span, message: T) -> Error { ++ Error { ++ messages: vec![ErrorMessage { ++ start_span: ThreadBound::new(start), ++ end_span: ThreadBound::new(end), ++ message: message.to_string(), ++ }], ++ } ++} ++ + impl Debug for Error { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + if self.messages.len() == 1 { +@@ -278,6 +287,14 @@ impl Display for Error { + } + } + ++impl Clone for Error { ++ fn clone(&self) -> Self { ++ Error { ++ messages: self.messages.clone(), ++ } ++ } ++} ++ + impl Clone for ErrorMessage { + fn clone(&self) -> Self { + let start = self +@@ -355,3 +372,11 @@ impl<'a> Iterator for Iter<'a> { + }) + } + } ++ ++impl Extend<Error> for Error { ++ fn extend<T: IntoIterator<Item = Error>>(&mut self, iter: T) { ++ for err in iter { ++ self.combine(err); ++ } ++ } ++} +diff --git a/third_party/rust/syn/src/expr.rs b/third_party/rust/syn/src/expr.rs +index 2874a463aa..2fe0e0b5d8 100644 +--- a/third_party/rust/syn/src/expr.rs ++++ b/third_party/rust/syn/src/expr.rs +@@ -1,18 +1,21 @@ + use super::*; + use crate::punctuated::Punctuated; +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; ++#[cfg(feature = "full")] ++use crate::reserved::Reserved; + use proc_macro2::{Span, TokenStream}; +-#[cfg(feature = "extra-traits")] ++#[cfg(feature = "printing")] ++use quote::IdentFragment; ++#[cfg(feature = "printing")] ++use std::fmt::{self, Display}; + use std::hash::{Hash, Hasher}; +-#[cfg(all(feature = "parsing", feature = "full"))] ++#[cfg(feature = "parsing")] + use std::mem; + + ast_enum_of_structs! { + /// A Rust expression. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` +- /// feature.* ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` ++ /// feature, but most of the variants are not available unless "full" is enabled.* + /// + /// # Syntax tree enums + /// +@@ -83,7 +86,7 @@ ast_enum_of_structs! { + /// A sign that you may not be choosing the right variable names is if you + /// see names getting repeated in your code, like accessing + /// `receiver.receiver` or `pat.pat` or `cond.cond`. +- pub enum Expr #manual_extra_traits { ++ pub enum Expr { + /// A slice literal expression: `[a, b, c, d]`. + Array(ExprArray), + +@@ -228,7 +231,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A slice literal expression: `[a, b, c, d]`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprArray #full { + pub attrs: Vec<Attribute>, + pub bracket_token: token::Bracket, +@@ -239,7 +242,7 @@ ast_struct! { + ast_struct! { + /// An assignment expression: `a = compute()`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprAssign #full { + pub attrs: Vec<Attribute>, + pub left: Box<Expr>, +@@ -251,7 +254,7 @@ ast_struct! { + ast_struct! { + /// A compound assignment expression: `counter += 1`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprAssignOp #full { + pub attrs: Vec<Attribute>, + pub left: Box<Expr>, +@@ -263,7 +266,7 @@ ast_struct! { + ast_struct! { + /// An async block: `async { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprAsync #full { + pub attrs: Vec<Attribute>, + pub async_token: Token![async], +@@ -275,7 +278,7 @@ ast_struct! { + ast_struct! { + /// An await expression: `fut.await`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprAwait #full { + pub attrs: Vec<Attribute>, + pub base: Box<Expr>, +@@ -287,7 +290,7 @@ ast_struct! { + ast_struct! { + /// A binary operation: `a + b`, `a * b`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprBinary { + pub attrs: Vec<Attribute>, +@@ -300,7 +303,7 @@ ast_struct! { + ast_struct! { + /// A blocked scope: `{ ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprBlock #full { + pub attrs: Vec<Attribute>, + pub label: Option<Label>, +@@ -311,7 +314,7 @@ ast_struct! { + ast_struct! { + /// A box expression: `box f`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprBox #full { + pub attrs: Vec<Attribute>, + pub box_token: Token![box], +@@ -323,7 +326,7 @@ ast_struct! { + /// A `break`, with an optional label to break and an optional + /// expression. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprBreak #full { + pub attrs: Vec<Attribute>, + pub break_token: Token![break], +@@ -335,7 +338,7 @@ ast_struct! { + ast_struct! { + /// A function call expression: `invoke(a, b)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprCall { + pub attrs: Vec<Attribute>, +@@ -348,7 +351,7 @@ ast_struct! { + ast_struct! { + /// A cast expression: `foo as f64`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprCast { + pub attrs: Vec<Attribute>, +@@ -361,7 +364,7 @@ ast_struct! { + ast_struct! { + /// A closure expression: `|a, b| a + b`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprClosure #full { + pub attrs: Vec<Attribute>, + pub asyncness: Option<Token![async]>, +@@ -378,7 +381,7 @@ ast_struct! { + ast_struct! { + /// A `continue`, with an optional label. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprContinue #full { + pub attrs: Vec<Attribute>, + pub continue_token: Token![continue], +@@ -390,7 +393,7 @@ ast_struct! { + /// Access of a named struct field (`obj.k`) or unnamed tuple struct + /// field (`obj.0`). + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprField { + pub attrs: Vec<Attribute>, + pub base: Box<Expr>, +@@ -402,7 +405,7 @@ ast_struct! { + ast_struct! { + /// A for loop: `for pat in expr { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprForLoop #full { + pub attrs: Vec<Attribute>, + pub label: Option<Label>, +@@ -421,7 +424,7 @@ ast_struct! { + /// of expressions and is related to `None`-delimited spans in a + /// `TokenStream`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprGroup #full { + pub attrs: Vec<Attribute>, + pub group_token: token::Group, +@@ -436,7 +439,7 @@ ast_struct! { + /// The `else` branch expression may only be an `If` or `Block` + /// expression, not any of the other types of expression. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprIf #full { + pub attrs: Vec<Attribute>, + pub if_token: Token![if], +@@ -449,7 +452,7 @@ ast_struct! { + ast_struct! { + /// A square bracketed indexing expression: `vector[2]`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprIndex { + pub attrs: Vec<Attribute>, +@@ -462,7 +465,7 @@ ast_struct! { + ast_struct! { + /// A `let` guard: `let Some(x) = opt`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprLet #full { + pub attrs: Vec<Attribute>, + pub let_token: Token![let], +@@ -475,7 +478,7 @@ ast_struct! { + ast_struct! { + /// A literal in place of an expression: `1`, `"foo"`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprLit { + pub attrs: Vec<Attribute>, +@@ -486,7 +489,7 @@ ast_struct! { + ast_struct! { + /// Conditionless loop: `loop { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprLoop #full { + pub attrs: Vec<Attribute>, + pub label: Option<Label>, +@@ -498,7 +501,7 @@ ast_struct! { + ast_struct! { + /// A macro invocation expression: `format!("{}", q)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprMacro #full { + pub attrs: Vec<Attribute>, + pub mac: Macro, +@@ -508,7 +511,7 @@ ast_struct! { + ast_struct! { + /// A `match` expression: `match n { Some(n) => {}, None => {} }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprMatch #full { + pub attrs: Vec<Attribute>, + pub match_token: Token![match], +@@ -521,7 +524,7 @@ ast_struct! { + ast_struct! { + /// A method call expression: `x.foo::<T>(a, b)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprMethodCall #full { + pub attrs: Vec<Attribute>, + pub receiver: Box<Expr>, +@@ -536,7 +539,7 @@ ast_struct! { + ast_struct! { + /// A parenthesized expression: `(a + b)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprParen { + pub attrs: Vec<Attribute>, + pub paren_token: token::Paren, +@@ -550,7 +553,7 @@ ast_struct! { + /// + /// A plain identifier like `x` is a path of length 1. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprPath { + pub attrs: Vec<Attribute>, +@@ -562,7 +565,7 @@ ast_struct! { + ast_struct! { + /// A range expression: `1..2`, `1..`, `..2`, `1..=2`, `..=2`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprRange #full { + pub attrs: Vec<Attribute>, + pub from: Option<Box<Expr>>, +@@ -574,7 +577,7 @@ ast_struct! { + ast_struct! { + /// A referencing operation: `&a` or `&mut a`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprReference #full { + pub attrs: Vec<Attribute>, + pub and_token: Token![&], +@@ -587,7 +590,7 @@ ast_struct! { + ast_struct! { + /// An array literal constructed from one repeated element: `[0u8; N]`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprRepeat #full { + pub attrs: Vec<Attribute>, + pub bracket_token: token::Bracket, +@@ -600,7 +603,7 @@ ast_struct! { + ast_struct! { + /// A `return`, with an optional value to be returned. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprReturn #full { + pub attrs: Vec<Attribute>, + pub return_token: Token![return], +@@ -614,7 +617,7 @@ ast_struct! { + /// The `rest` provides the value of the remaining fields as in `S { a: + /// 1, b: 1, ..rest }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprStruct #full { + pub attrs: Vec<Attribute>, + pub path: Path, +@@ -628,7 +631,7 @@ ast_struct! { + ast_struct! { + /// A try-expression: `expr?`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprTry #full { + pub attrs: Vec<Attribute>, + pub expr: Box<Expr>, +@@ -639,7 +642,7 @@ ast_struct! { + ast_struct! { + /// A try block: `try { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprTryBlock #full { + pub attrs: Vec<Attribute>, + pub try_token: Token![try], +@@ -650,7 +653,7 @@ ast_struct! { + ast_struct! { + /// A tuple expression: `(a, b, c, d)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprTuple #full { + pub attrs: Vec<Attribute>, + pub paren_token: token::Paren, +@@ -661,7 +664,7 @@ ast_struct! { + ast_struct! { + /// A type ascription expression: `foo: f64`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprType #full { + pub attrs: Vec<Attribute>, + pub expr: Box<Expr>, +@@ -673,7 +676,7 @@ ast_struct! { + ast_struct! { + /// A unary operation: `!x`, `*x`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprUnary { + pub attrs: Vec<Attribute>, +@@ -685,7 +688,7 @@ ast_struct! { + ast_struct! { + /// An unsafe block: `unsafe { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprUnsafe #full { + pub attrs: Vec<Attribute>, + pub unsafe_token: Token![unsafe], +@@ -696,7 +699,7 @@ ast_struct! { + ast_struct! { + /// A while loop: `while expr { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprWhile #full { + pub attrs: Vec<Attribute>, + pub label: Option<Label>, +@@ -709,7 +712,7 @@ ast_struct! { + ast_struct! { + /// A yield expression: `yield expr`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprYield #full { + pub attrs: Vec<Attribute>, + pub yield_token: Token![yield], +@@ -717,232 +720,6 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Expr {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Expr { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (Expr::Array(this), Expr::Array(other)) => this == other, +- (Expr::Assign(this), Expr::Assign(other)) => this == other, +- (Expr::AssignOp(this), Expr::AssignOp(other)) => this == other, +- (Expr::Async(this), Expr::Async(other)) => this == other, +- (Expr::Await(this), Expr::Await(other)) => this == other, +- (Expr::Binary(this), Expr::Binary(other)) => this == other, +- (Expr::Block(this), Expr::Block(other)) => this == other, +- (Expr::Box(this), Expr::Box(other)) => this == other, +- (Expr::Break(this), Expr::Break(other)) => this == other, +- (Expr::Call(this), Expr::Call(other)) => this == other, +- (Expr::Cast(this), Expr::Cast(other)) => this == other, +- (Expr::Closure(this), Expr::Closure(other)) => this == other, +- (Expr::Continue(this), Expr::Continue(other)) => this == other, +- (Expr::Field(this), Expr::Field(other)) => this == other, +- (Expr::ForLoop(this), Expr::ForLoop(other)) => this == other, +- (Expr::Group(this), Expr::Group(other)) => this == other, +- (Expr::If(this), Expr::If(other)) => this == other, +- (Expr::Index(this), Expr::Index(other)) => this == other, +- (Expr::Let(this), Expr::Let(other)) => this == other, +- (Expr::Lit(this), Expr::Lit(other)) => this == other, +- (Expr::Loop(this), Expr::Loop(other)) => this == other, +- (Expr::Macro(this), Expr::Macro(other)) => this == other, +- (Expr::Match(this), Expr::Match(other)) => this == other, +- (Expr::MethodCall(this), Expr::MethodCall(other)) => this == other, +- (Expr::Paren(this), Expr::Paren(other)) => this == other, +- (Expr::Path(this), Expr::Path(other)) => this == other, +- (Expr::Range(this), Expr::Range(other)) => this == other, +- (Expr::Reference(this), Expr::Reference(other)) => this == other, +- (Expr::Repeat(this), Expr::Repeat(other)) => this == other, +- (Expr::Return(this), Expr::Return(other)) => this == other, +- (Expr::Struct(this), Expr::Struct(other)) => this == other, +- (Expr::Try(this), Expr::Try(other)) => this == other, +- (Expr::TryBlock(this), Expr::TryBlock(other)) => this == other, +- (Expr::Tuple(this), Expr::Tuple(other)) => this == other, +- (Expr::Type(this), Expr::Type(other)) => this == other, +- (Expr::Unary(this), Expr::Unary(other)) => this == other, +- (Expr::Unsafe(this), Expr::Unsafe(other)) => this == other, +- (Expr::Verbatim(this), Expr::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- (Expr::While(this), Expr::While(other)) => this == other, +- (Expr::Yield(this), Expr::Yield(other)) => this == other, +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Expr { +- fn hash<H>(&self, hash: &mut H) +- where +- H: Hasher, +- { +- match self { +- Expr::Array(expr) => { +- hash.write_u8(0); +- expr.hash(hash); +- } +- Expr::Assign(expr) => { +- hash.write_u8(1); +- expr.hash(hash); +- } +- Expr::AssignOp(expr) => { +- hash.write_u8(2); +- expr.hash(hash); +- } +- Expr::Async(expr) => { +- hash.write_u8(3); +- expr.hash(hash); +- } +- Expr::Await(expr) => { +- hash.write_u8(4); +- expr.hash(hash); +- } +- Expr::Binary(expr) => { +- hash.write_u8(5); +- expr.hash(hash); +- } +- Expr::Block(expr) => { +- hash.write_u8(6); +- expr.hash(hash); +- } +- Expr::Box(expr) => { +- hash.write_u8(7); +- expr.hash(hash); +- } +- Expr::Break(expr) => { +- hash.write_u8(8); +- expr.hash(hash); +- } +- Expr::Call(expr) => { +- hash.write_u8(9); +- expr.hash(hash); +- } +- Expr::Cast(expr) => { +- hash.write_u8(10); +- expr.hash(hash); +- } +- Expr::Closure(expr) => { +- hash.write_u8(11); +- expr.hash(hash); +- } +- Expr::Continue(expr) => { +- hash.write_u8(12); +- expr.hash(hash); +- } +- Expr::Field(expr) => { +- hash.write_u8(13); +- expr.hash(hash); +- } +- Expr::ForLoop(expr) => { +- hash.write_u8(14); +- expr.hash(hash); +- } +- Expr::Group(expr) => { +- hash.write_u8(15); +- expr.hash(hash); +- } +- Expr::If(expr) => { +- hash.write_u8(16); +- expr.hash(hash); +- } +- Expr::Index(expr) => { +- hash.write_u8(17); +- expr.hash(hash); +- } +- Expr::Let(expr) => { +- hash.write_u8(18); +- expr.hash(hash); +- } +- Expr::Lit(expr) => { +- hash.write_u8(19); +- expr.hash(hash); +- } +- Expr::Loop(expr) => { +- hash.write_u8(20); +- expr.hash(hash); +- } +- Expr::Macro(expr) => { +- hash.write_u8(21); +- expr.hash(hash); +- } +- Expr::Match(expr) => { +- hash.write_u8(22); +- expr.hash(hash); +- } +- Expr::MethodCall(expr) => { +- hash.write_u8(23); +- expr.hash(hash); +- } +- Expr::Paren(expr) => { +- hash.write_u8(24); +- expr.hash(hash); +- } +- Expr::Path(expr) => { +- hash.write_u8(25); +- expr.hash(hash); +- } +- Expr::Range(expr) => { +- hash.write_u8(26); +- expr.hash(hash); +- } +- Expr::Reference(expr) => { +- hash.write_u8(27); +- expr.hash(hash); +- } +- Expr::Repeat(expr) => { +- hash.write_u8(28); +- expr.hash(hash); +- } +- Expr::Return(expr) => { +- hash.write_u8(29); +- expr.hash(hash); +- } +- Expr::Struct(expr) => { +- hash.write_u8(30); +- expr.hash(hash); +- } +- Expr::Try(expr) => { +- hash.write_u8(31); +- expr.hash(hash); +- } +- Expr::TryBlock(expr) => { +- hash.write_u8(32); +- expr.hash(hash); +- } +- Expr::Tuple(expr) => { +- hash.write_u8(33); +- expr.hash(hash); +- } +- Expr::Type(expr) => { +- hash.write_u8(34); +- expr.hash(hash); +- } +- Expr::Unary(expr) => { +- hash.write_u8(35); +- expr.hash(hash); +- } +- Expr::Unsafe(expr) => { +- hash.write_u8(36); +- expr.hash(hash); +- } +- Expr::Verbatim(expr) => { +- hash.write_u8(37); +- TokenStreamHelper(expr).hash(hash); +- } +- Expr::While(expr) => { +- hash.write_u8(38); +- expr.hash(hash); +- } +- Expr::Yield(expr) => { +- hash.write_u8(39); +- expr.hash(hash); +- } +- Expr::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + impl Expr { + #[cfg(all(feature = "parsing", feature = "full"))] + pub(crate) fn replace_attrs(&mut self, new: Vec<Attribute>) -> Vec<Attribute> { +@@ -996,7 +773,7 @@ ast_enum! { + /// A struct or tuple struct field accessed in a struct literal or field + /// expression. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum Member { + /// A named field like `self.x`. +@@ -1006,12 +783,50 @@ ast_enum! { + } + } + ++impl Eq for Member {} ++ ++impl PartialEq for Member { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Member::Named(this), Member::Named(other)) => this == other, ++ (Member::Unnamed(this), Member::Unnamed(other)) => this == other, ++ _ => false, ++ } ++ } ++} ++ ++impl Hash for Member { ++ fn hash<H: Hasher>(&self, state: &mut H) { ++ match self { ++ Member::Named(m) => m.hash(state), ++ Member::Unnamed(m) => m.hash(state), ++ } ++ } ++} ++ ++#[cfg(feature = "printing")] ++impl IdentFragment for Member { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Member::Named(m) => Display::fmt(m, formatter), ++ Member::Unnamed(m) => Display::fmt(&m.index, formatter), ++ } ++ } ++ ++ fn span(&self) -> Option<Span> { ++ match self { ++ Member::Named(m) => Some(m.span()), ++ Member::Unnamed(m) => Some(m.span), ++ } ++ } ++} ++ + ast_struct! { + /// The index of an unnamed tuple struct field. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- pub struct Index #manual_extra_traits { ++ pub struct Index { + pub index: u32, + pub span: Span, + } +@@ -1027,28 +842,28 @@ impl From<usize> for Index { + } + } + +-#[cfg(feature = "extra-traits")] + impl Eq for Index {} + +-#[cfg(feature = "extra-traits")] + impl PartialEq for Index { + fn eq(&self, other: &Self) -> bool { + self.index == other.index + } + } + +-#[cfg(feature = "extra-traits")] + impl Hash for Index { + fn hash<H: Hasher>(&self, state: &mut H) { + self.index.hash(state); + } + } + +-#[cfg(feature = "full")] +-ast_struct! { +- #[derive(Default)] +- pub struct Reserved { +- private: (), ++#[cfg(feature = "printing")] ++impl IdentFragment for Index { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ Display::fmt(&self.index, formatter) ++ } ++ ++ fn span(&self) -> Option<Span> { ++ Some(self.span) + } + } + +@@ -1057,7 +872,7 @@ ast_struct! { + /// The `::<>` explicit type parameters passed to a method call: + /// `parse::<u64>()`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct MethodTurbofish { + pub colon2_token: Token![::], + pub lt_token: Token![<], +@@ -1070,7 +885,7 @@ ast_struct! { + ast_enum! { + /// An individual generic argument to a method, like `T`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub enum GenericMethodArgument { + /// A type argument. + Type(Type), +@@ -1086,7 +901,7 @@ ast_enum! { + ast_struct! { + /// A field-value pair in a struct literal. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct FieldValue { + /// Attributes tagged on the field. + pub attrs: Vec<Attribute>, +@@ -1107,7 +922,7 @@ ast_struct! { + ast_struct! { + /// A lifetime labeling a `for`, `while`, or `loop`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Label { + pub name: Lifetime, + pub colon_token: Token![:], +@@ -1134,7 +949,7 @@ ast_struct! { + /// # } + /// ``` + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Arm { + pub attrs: Vec<Attribute>, + pub pat: Pat, +@@ -1149,8 +964,7 @@ ast_struct! { + ast_enum! { + /// Limit types of a range, inclusive or exclusive. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* +- #[cfg_attr(feature = "clone-impls", derive(Copy))] ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub enum RangeLimits { + /// Inclusive at the beginning, exclusive at the end. + HalfOpen(Token![..]), +@@ -1162,7 +976,7 @@ ast_enum! { + #[cfg(any(feature = "parsing", feature = "printing"))] + #[cfg(feature = "full")] + pub(crate) fn requires_terminator(expr: &Expr) -> bool { +- // see https://github.com/rust-lang/rust/blob/eb8f2586e/src/libsyntax/parse/classify.rs#L17-L37 ++ // see https://github.com/rust-lang/rust/blob/2679c38fc/src/librustc_ast/util/classify.rs#L7-L25 + match *expr { + Expr::Unsafe(..) + | Expr::Block(..) +@@ -1183,16 +997,17 @@ pub(crate) mod parsing { + + use crate::parse::{Parse, ParseStream, Result}; + use crate::path; ++ use std::cmp::Ordering; ++ ++ crate::custom_keyword!(raw); + + // When we're parsing expressions which occur before blocks, like in an if + // statement's condition, we cannot parse a struct literal. + // + // Struct literals are ambiguous in certain positions + // https://github.com/rust-lang/rfcs/pull/92 +- #[derive(Copy, Clone)] + pub struct AllowStruct(bool); + +- #[derive(Copy, Clone, PartialEq, PartialOrd)] + enum Precedence { + Any, + Assign, +@@ -1246,9 +1061,121 @@ pub(crate) mod parsing { + } + } + +- #[cfg(feature = "full")] +- fn expr_no_struct(input: ParseStream) -> Result<Expr> { +- ambiguous_expr(input, AllowStruct(false)) ++ impl Expr { ++ /// An alternative to the primary `Expr::parse` parser (from the ++ /// [`Parse`] trait) for ambiguous syntactic positions in which a ++ /// trailing brace should not be taken as part of the expression. ++ /// ++ /// Rust grammar has an ambiguity where braces sometimes turn a path ++ /// expression into a struct initialization and sometimes do not. In the ++ /// following code, the expression `S {}` is one expression. Presumably ++ /// there is an empty struct `struct S {}` defined somewhere which it is ++ /// instantiating. ++ /// ++ /// ``` ++ /// # struct S; ++ /// # impl std::ops::Deref for S { ++ /// # type Target = bool; ++ /// # fn deref(&self) -> &Self::Target { ++ /// # &true ++ /// # } ++ /// # } ++ /// let _ = *S {}; ++ /// ++ /// // parsed by rustc as: `*(S {})` ++ /// ``` ++ /// ++ /// We would want to parse the above using `Expr::parse` after the `=` ++ /// token. ++ /// ++ /// But in the following, `S {}` is *not* a struct init expression. ++ /// ++ /// ``` ++ /// # const S: &bool = &true; ++ /// if *S {} {} ++ /// ++ /// // parsed by rustc as: ++ /// // ++ /// // if (*S) { ++ /// // /* empty block */ ++ /// // } ++ /// // { ++ /// // /* another empty block */ ++ /// // } ++ /// ``` ++ /// ++ /// For that reason we would want to parse if-conditions using ++ /// `Expr::parse_without_eager_brace` after the `if` token. Same for ++ /// similar syntactic positions such as the condition expr after a ++ /// `while` token or the expr at the top of a `match`. ++ /// ++ /// The Rust grammar's choices around which way this ambiguity is ++ /// resolved at various syntactic positions is fairly arbitrary. Really ++ /// either parse behavior could work in most positions, and language ++ /// designers just decide each case based on which is more likely to be ++ /// what the programmer had in mind most of the time. ++ /// ++ /// ``` ++ /// # struct S; ++ /// # fn doc() -> S { ++ /// if return S {} {} ++ /// # unreachable!() ++ /// # } ++ /// ++ /// // parsed by rustc as: ++ /// // ++ /// // if (return (S {})) { ++ /// // } ++ /// // ++ /// // but could equally well have been this other arbitrary choice: ++ /// // ++ /// // if (return S) { ++ /// // } ++ /// // {} ++ /// ``` ++ /// ++ /// Note the grammar ambiguity on trailing braces is distinct from ++ /// precedence and is not captured by assigning a precedence level to ++ /// the braced struct init expr in relation to other operators. This can ++ /// be illustrated by `return 0..S {}` vs `match 0..S {}`. The former ++ /// parses as `return (0..(S {}))` implying tighter precedence for ++ /// struct init than `..`, while the latter parses as `match (0..S) {}` ++ /// implying tighter precedence for `..` than struct init, a ++ /// contradiction. ++ #[cfg(feature = "full")] ++ pub fn parse_without_eager_brace(input: ParseStream) -> Result<Expr> { ++ ambiguous_expr(input, AllowStruct(false)) ++ } ++ } ++ ++ impl Copy for AllowStruct {} ++ ++ impl Clone for AllowStruct { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } ++ ++ impl Copy for Precedence {} ++ ++ impl Clone for Precedence { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } ++ ++ impl PartialEq for Precedence { ++ fn eq(&self, other: &Self) -> bool { ++ *self as u8 == *other as u8 ++ } ++ } ++ ++ impl PartialOrd for Precedence { ++ fn partial_cmp(&self, other: &Self) -> Option<Ordering> { ++ let this = *self as u8; ++ let other = *other as u8; ++ Some(this.cmp(&other)) ++ } + } + + #[cfg(feature = "full")] +@@ -1430,56 +1357,84 @@ pub(crate) mod parsing { + parse_expr(input, lhs, allow_struct, Precedence::Any) + } + ++ #[cfg(feature = "full")] ++ fn expr_attrs(input: ParseStream) -> Result<Vec<Attribute>> { ++ let mut attrs = Vec::new(); ++ loop { ++ if input.peek(token::Group) { ++ let ahead = input.fork(); ++ let group = crate::group::parse_group(&ahead)?; ++ if !group.content.peek(Token![#]) || group.content.peek2(Token![!]) { ++ break; ++ } ++ let attr = group.content.call(attr::parsing::single_parse_outer)?; ++ if !group.content.is_empty() { ++ break; ++ } ++ attrs.push(attr); ++ } else if input.peek(Token![#]) { ++ attrs.push(input.call(attr::parsing::single_parse_outer)?); ++ } else { ++ break; ++ } ++ } ++ Ok(attrs) ++ } ++ + // <UnOp> <trailer> + // & <trailer> + // &mut <trailer> + // box <trailer> + #[cfg(feature = "full")] + fn unary_expr(input: ParseStream, allow_struct: AllowStruct) -> Result<Expr> { +- // TODO: optimize using advance_to +- let ahead = input.fork(); +- ahead.call(Attribute::parse_outer)?; +- if ahead.peek(Token![&]) +- || ahead.peek(Token![box]) +- || ahead.peek(Token![*]) +- || ahead.peek(Token![!]) +- || ahead.peek(Token![-]) +- { +- let attrs = input.call(Attribute::parse_outer)?; +- if input.peek(Token![&]) { ++ let begin = input.fork(); ++ let attrs = input.call(expr_attrs)?; ++ if input.peek(Token![&]) { ++ let and_token: Token![&] = input.parse()?; ++ let raw: Option<raw> = ++ if input.peek(raw) && (input.peek2(Token![mut]) || input.peek2(Token![const])) { ++ Some(input.parse()?) ++ } else { ++ None ++ }; ++ let mutability: Option<Token![mut]> = input.parse()?; ++ if raw.is_some() && mutability.is_none() { ++ input.parse::<Token![const]>()?; ++ } ++ let expr = Box::new(unary_expr(input, allow_struct)?); ++ if raw.is_some() { ++ Ok(Expr::Verbatim(verbatim::between(begin, input))) ++ } else { + Ok(Expr::Reference(ExprReference { + attrs, +- and_token: input.parse()?, ++ and_token, + raw: Reserved::default(), +- mutability: input.parse()?, +- expr: Box::new(unary_expr(input, allow_struct)?), +- })) +- } else if input.peek(Token![box]) { +- Ok(Expr::Box(ExprBox { +- attrs, +- box_token: input.parse()?, +- expr: Box::new(unary_expr(input, allow_struct)?), +- })) +- } else { +- Ok(Expr::Unary(ExprUnary { +- attrs, +- op: input.parse()?, +- expr: Box::new(unary_expr(input, allow_struct)?), ++ mutability, ++ expr, + })) + } ++ } else if input.peek(Token![box]) { ++ Ok(Expr::Box(ExprBox { ++ attrs, ++ box_token: input.parse()?, ++ expr: Box::new(unary_expr(input, allow_struct)?), ++ })) ++ } else if input.peek(Token![*]) || input.peek(Token![!]) || input.peek(Token![-]) { ++ Ok(Expr::Unary(ExprUnary { ++ attrs, ++ op: input.parse()?, ++ expr: Box::new(unary_expr(input, allow_struct)?), ++ })) + } else { +- trailer_expr(input, allow_struct) ++ trailer_expr(attrs, input, allow_struct) + } + } + + #[cfg(not(feature = "full"))] + fn unary_expr(input: ParseStream, allow_struct: AllowStruct) -> Result<Expr> { +- // TODO: optimize using advance_to +- let ahead = input.fork(); +- ahead.call(Attribute::parse_outer)?; +- if ahead.peek(Token![*]) || ahead.peek(Token![!]) || ahead.peek(Token![-]) { ++ if input.peek(Token![*]) || input.peek(Token![!]) || input.peek(Token![-]) { + Ok(Expr::Unary(ExprUnary { +- attrs: input.call(Attribute::parse_outer)?, ++ attrs: Vec::new(), + op: input.parse()?, + expr: Box::new(unary_expr(input, allow_struct)?), + })) +@@ -1495,13 +1450,11 @@ pub(crate) mod parsing { + // <atom> [ <expr> ] ... + // <atom> ? ... + #[cfg(feature = "full")] +- fn trailer_expr(input: ParseStream, allow_struct: AllowStruct) -> Result<Expr> { +- if input.peek(token::Group) { +- return input.call(expr_group).map(Expr::Group); +- } +- +- let outer_attrs = input.call(Attribute::parse_outer)?; +- ++ fn trailer_expr( ++ outer_attrs: Vec<Attribute>, ++ input: ParseStream, ++ allow_struct: AllowStruct, ++ ) -> Result<Expr> { + let atom = atom_expr(input, allow_struct)?; + let mut e = trailer_helper(input, atom)?; + +@@ -1523,18 +1476,26 @@ pub(crate) mod parsing { + args: content.parse_terminated(Expr::parse)?, + }); + } else if input.peek(Token![.]) && !input.peek(Token![..]) { +- let dot_token: Token![.] = input.parse()?; ++ let mut dot_token: Token![.] = input.parse()?; + +- if input.peek(token::Await) { ++ let await_token: Option<token::Await> = input.parse()?; ++ if let Some(await_token) = await_token { + e = Expr::Await(ExprAwait { + attrs: Vec::new(), + base: Box::new(e), + dot_token, +- await_token: input.parse()?, ++ await_token, + }); + continue; + } + ++ let float_token: Option<LitFloat> = input.parse()?; ++ if let Some(float_token) = float_token { ++ if multi_index(&mut e, &mut dot_token, float_token)? { ++ continue; ++ } ++ } ++ + let member: Member = input.parse()?; + let turbofish = if member.is_named() && input.peek(Token![::]) { + Some(MethodTurbofish { +@@ -1620,10 +1581,17 @@ pub(crate) mod parsing { + }); + } else if input.peek(Token![.]) && !input.peek(Token![..]) && !input.peek2(token::Await) + { ++ let mut dot_token: Token![.] = input.parse()?; ++ let float_token: Option<LitFloat> = input.parse()?; ++ if let Some(float_token) = float_token { ++ if multi_index(&mut e, &mut dot_token, float_token)? { ++ continue; ++ } ++ } + e = Expr::Field(ExprField { + attrs: Vec::new(), + base: Box::new(e), +- dot_token: input.parse()?, ++ dot_token, + member: input.parse()?, + }); + } else if input.peek(token::Bracket) { +@@ -1646,7 +1614,11 @@ pub(crate) mod parsing { + // interactions, as they are fully contained. + #[cfg(feature = "full")] + fn atom_expr(input: ParseStream, allow_struct: AllowStruct) -> Result<Expr> { +- if input.peek(token::Group) { ++ if input.peek(token::Group) ++ && !input.peek2(Token![::]) ++ && !input.peek2(Token![!]) ++ && !input.peek2(token::Brace) ++ { + input.call(expr_group).map(Expr::Group) + } else if input.peek(Lit) { + input.parse().map(Expr::Lit) +@@ -1668,7 +1640,6 @@ pub(crate) mod parsing { + || input.peek(Token![self]) + || input.peek(Token![Self]) + || input.peek(Token![super]) +- || input.peek(Token![extern]) + || input.peek(Token![crate]) + { + path_or_macro_or_struct(input, allow_struct) +@@ -1740,7 +1711,6 @@ pub(crate) mod parsing { + || input.peek(Token![self]) + || input.peek(Token![Self]) + || input.peek(Token![super]) +- || input.peek(Token![extern]) + || input.peek(Token![crate]) + { + input.parse().map(Expr::Path) +@@ -1878,7 +1848,7 @@ pub(crate) mod parsing { + + #[cfg(feature = "full")] + pub(crate) fn expr_early(input: ParseStream) -> Result<Expr> { +- let mut attrs = input.call(Attribute::parse_outer)?; ++ let mut attrs = input.call(expr_attrs)?; + let mut expr = if input.peek(Token![if]) { + Expr::If(input.parse()?) + } else if input.peek(Token![while]) { +@@ -1905,7 +1875,7 @@ pub(crate) mod parsing { + return parse_expr(input, expr, allow_struct, Precedence::Any); + }; + +- if input.peek(Token![.]) || input.peek(Token![?]) { ++ if input.peek(Token![.]) && !input.peek(Token![..]) || input.peek(Token![?]) { + expr = trailer_helper(input, expr)?; + + attrs.extend(expr.replace_attrs(Vec::new())); +@@ -1951,7 +1921,16 @@ pub(crate) mod parsing { + + #[cfg(feature = "full")] + fn generic_method_argument(input: ParseStream) -> Result<GenericMethodArgument> { +- // TODO parse const generics as well ++ if input.peek(Lit) { ++ let lit = input.parse()?; ++ return Ok(GenericMethodArgument::Const(Expr::Lit(lit))); ++ } ++ ++ if input.peek(token::Brace) { ++ let block = input.call(expr::parsing::expr_block)?; ++ return Ok(GenericMethodArgument::Const(Expr::Block(block))); ++ } ++ + input.parse().map(GenericMethodArgument::Type) + } + +@@ -1960,44 +1939,20 @@ pub(crate) mod parsing { + Ok(ExprLet { + attrs: Vec::new(), + let_token: input.parse()?, +- pat: { +- let leading_vert: Option<Token![|]> = input.parse()?; +- let pat: Pat = input.parse()?; +- if leading_vert.is_some() +- || input.peek(Token![|]) && !input.peek(Token![||]) && !input.peek(Token![|=]) +- { +- let mut cases = Punctuated::new(); +- cases.push_value(pat); +- while input.peek(Token![|]) +- && !input.peek(Token![||]) +- && !input.peek(Token![|=]) +- { +- let punct = input.parse()?; +- cases.push_punct(punct); +- let pat: Pat = input.parse()?; +- cases.push_value(pat); +- } +- Pat::Or(PatOr { +- attrs: Vec::new(), +- leading_vert, +- cases, +- }) +- } else { +- pat +- } +- }, ++ pat: pat::parsing::multi_pat_with_leading_vert(input)?, + eq_token: input.parse()?, +- expr: Box::new(input.call(expr_no_struct)?), ++ expr: Box::new(input.call(Expr::parse_without_eager_brace)?), + }) + } + + #[cfg(feature = "full")] + impl Parse for ExprIf { + fn parse(input: ParseStream) -> Result<Self> { ++ let attrs = input.call(Attribute::parse_outer)?; + Ok(ExprIf { +- attrs: Vec::new(), ++ attrs, + if_token: input.parse()?, +- cond: Box::new(input.call(expr_no_struct)?), ++ cond: Box::new(input.call(Expr::parse_without_eager_brace)?), + then_branch: input.parse()?, + else_branch: { + if input.peek(Token![else]) { +@@ -2033,29 +1988,14 @@ pub(crate) mod parsing { + #[cfg(feature = "full")] + impl Parse for ExprForLoop { + fn parse(input: ParseStream) -> Result<Self> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; + let label: Option<Label> = input.parse()?; + let for_token: Token![for] = input.parse()?; + +- let leading_vert: Option<Token![|]> = input.parse()?; +- let mut pat: Pat = input.parse()?; +- if leading_vert.is_some() || input.peek(Token![|]) { +- let mut cases = Punctuated::new(); +- cases.push_value(pat); +- while input.peek(Token![|]) { +- let punct = input.parse()?; +- cases.push_punct(punct); +- let pat: Pat = input.parse()?; +- cases.push_value(pat); +- } +- pat = Pat::Or(PatOr { +- attrs: Vec::new(), +- leading_vert, +- cases, +- }); +- } ++ let pat = pat::parsing::multi_pat_with_leading_vert(input)?; + + let in_token: Token![in] = input.parse()?; +- let expr: Expr = input.call(expr_no_struct)?; ++ let expr: Expr = input.call(Expr::parse_without_eager_brace)?; + + let content; + let brace_token = braced!(content in input); +@@ -2063,7 +2003,7 @@ pub(crate) mod parsing { + let stmts = content.call(Block::parse_within)?; + + Ok(ExprForLoop { +- attrs: inner_attrs, ++ attrs: private::attrs(outer_attrs, inner_attrs), + label, + for_token, + pat, +@@ -2077,6 +2017,7 @@ pub(crate) mod parsing { + #[cfg(feature = "full")] + impl Parse for ExprLoop { + fn parse(input: ParseStream) -> Result<Self> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; + let label: Option<Label> = input.parse()?; + let loop_token: Token![loop] = input.parse()?; + +@@ -2086,7 +2027,7 @@ pub(crate) mod parsing { + let stmts = content.call(Block::parse_within)?; + + Ok(ExprLoop { +- attrs: inner_attrs, ++ attrs: private::attrs(outer_attrs, inner_attrs), + label, + loop_token, + body: Block { brace_token, stmts }, +@@ -2097,8 +2038,9 @@ pub(crate) mod parsing { + #[cfg(feature = "full")] + impl Parse for ExprMatch { + fn parse(input: ParseStream) -> Result<Self> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; + let match_token: Token![match] = input.parse()?; +- let expr = expr_no_struct(input)?; ++ let expr = Expr::parse_without_eager_brace(input)?; + + let content; + let brace_token = braced!(content in input); +@@ -2110,7 +2052,7 @@ pub(crate) mod parsing { + } + + Ok(ExprMatch { +- attrs: inner_attrs, ++ attrs: private::attrs(outer_attrs, inner_attrs), + match_token, + expr: Box::new(expr), + brace_token, +@@ -2305,9 +2247,10 @@ pub(crate) mod parsing { + #[cfg(feature = "full")] + impl Parse for ExprWhile { + fn parse(input: ParseStream) -> Result<Self> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; + let label: Option<Label> = input.parse()?; + let while_token: Token![while] = input.parse()?; +- let cond = expr_no_struct(input)?; ++ let cond = Expr::parse_without_eager_brace(input)?; + + let content; + let brace_token = braced!(content in input); +@@ -2315,7 +2258,7 @@ pub(crate) mod parsing { + let stmts = content.call(Block::parse_within)?; + + Ok(ExprWhile { +- attrs: inner_attrs, ++ attrs: private::attrs(outer_attrs, inner_attrs), + label, + while_token, + cond: Box::new(cond), +@@ -2399,6 +2342,7 @@ pub(crate) mod parsing { + #[cfg(feature = "full")] + impl Parse for FieldValue { + fn parse(input: ParseStream) -> Result<Self> { ++ let attrs = input.call(Attribute::parse_outer)?; + let member: Member = input.parse()?; + let (colon_token, value) = if input.peek(Token![:]) || !member.is_named() { + let colon_token: Token![:] = input.parse()?; +@@ -2416,7 +2360,7 @@ pub(crate) mod parsing { + }; + + Ok(FieldValue { +- attrs: Vec::new(), ++ attrs, + member, + colon_token, + expr: value, +@@ -2433,46 +2377,36 @@ pub(crate) mod parsing { + let content; + let brace_token = braced!(content in input); + let inner_attrs = content.call(Attribute::parse_inner)?; ++ let attrs = private::attrs(outer_attrs, inner_attrs); + + let mut fields = Punctuated::new(); +- loop { +- let attrs = content.call(Attribute::parse_outer)?; +- // TODO: optimize using advance_to +- if content.fork().parse::<Member>().is_err() { +- if attrs.is_empty() { +- break; +- } else { +- return Err(content.error("expected struct field")); +- } ++ while !content.is_empty() { ++ if content.peek(Token![..]) { ++ return Ok(ExprStruct { ++ attrs, ++ brace_token, ++ path, ++ fields, ++ dot2_token: Some(content.parse()?), ++ rest: Some(Box::new(content.parse()?)), ++ }); + } + +- fields.push(FieldValue { +- attrs, +- ..content.parse()? +- }); +- +- if !content.peek(Token![,]) { ++ fields.push(content.parse()?); ++ if content.is_empty() { + break; + } + let punct: Token![,] = content.parse()?; + fields.push_punct(punct); + } + +- let (dot2_token, rest) = if fields.empty_or_trailing() && content.peek(Token![..]) { +- let dot2_token: Token![..] = content.parse()?; +- let rest: Expr = content.parse()?; +- (Some(dot2_token), Some(Box::new(rest))) +- } else { +- (None, None) +- }; +- + Ok(ExprStruct { +- attrs: private::attrs(outer_attrs, inner_attrs), ++ attrs, + brace_token, + path, + fields, +- dot2_token, +- rest, ++ dot2_token: None, ++ rest: None, + }) + } + +@@ -2577,27 +2511,7 @@ pub(crate) mod parsing { + let requires_comma; + Ok(Arm { + attrs: input.call(Attribute::parse_outer)?, +- pat: { +- let leading_vert: Option<Token![|]> = input.parse()?; +- let pat: Pat = input.parse()?; +- if leading_vert.is_some() || input.peek(Token![|]) { +- let mut cases = Punctuated::new(); +- cases.push_value(pat); +- while input.peek(Token![|]) { +- let punct = input.parse()?; +- cases.push_punct(punct); +- let pat: Pat = input.parse()?; +- cases.push_value(pat); +- } +- Pat::Or(PatOr { +- attrs: Vec::new(), +- leading_vert, +- cases, +- }) +- } else { +- pat +- } +- }, ++ pat: pat::parsing::multi_pat_with_leading_vert(input)?, + guard: { + if input.peek(Token![if]) { + let if_token: Token![if] = input.parse()?; +@@ -2641,6 +2555,26 @@ pub(crate) mod parsing { + } + } + ++ fn multi_index(e: &mut Expr, dot_token: &mut Token![.], float: LitFloat) -> Result<bool> { ++ let mut float_repr = float.to_string(); ++ let trailing_dot = float_repr.ends_with('.'); ++ if trailing_dot { ++ float_repr.truncate(float_repr.len() - 1); ++ } ++ for part in float_repr.split('.') { ++ let index = crate::parse_str(part).map_err(|err| Error::new(float.span(), err))?; ++ let base = mem::replace(e, Expr::__Nonexhaustive); ++ *e = Expr::Field(ExprField { ++ attrs: Vec::new(), ++ base: Box::new(base), ++ dot_token: Token![.](dot_token.span), ++ member: Member::Unnamed(index), ++ }); ++ *dot_token = Token![.](float.span()); ++ } ++ Ok(!trailing_dot) ++ } ++ + #[cfg(feature = "full")] + impl Member { + fn is_named(&self) -> bool { +diff --git a/third_party/rust/syn/src/ext.rs b/third_party/rust/syn/src/ext.rs +index d09577a27a..4f9bc145d9 100644 +--- a/third_party/rust/syn/src/ext.rs ++++ b/third_party/rust/syn/src/ext.rs +@@ -1,6 +1,6 @@ + //! Extension traits to provide parsing methods on foreign types. + //! +-//! *This module is available if Syn is built with the `"parsing"` feature.* ++//! *This module is available only if Syn is built with the `"parsing"` feature.* + + use proc_macro2::Ident; + +@@ -16,7 +16,7 @@ use crate::token::CustomToken; + /// This trait is sealed and cannot be implemented for types outside of Syn. It + /// is implemented only for `proc_macro2::Ident`. + /// +-/// *This trait is available if Syn is built with the `"parsing"` feature.* ++/// *This trait is available only if Syn is built with the `"parsing"` feature.* + pub trait IdentExt: Sized + private::Sealed { + /// Parses any identifier including keywords. + /// +@@ -129,7 +129,13 @@ mod private { + + impl Sealed for Ident {} + +- #[derive(Copy, Clone)] + pub struct PeekFn; + pub struct IdentAny; ++ ++ impl Copy for PeekFn {} ++ impl Clone for PeekFn { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } + } +diff --git a/third_party/rust/syn/src/file.rs b/third_party/rust/syn/src/file.rs +index 88c02fe832..c8fab63cd9 100644 +--- a/third_party/rust/syn/src/file.rs ++++ b/third_party/rust/syn/src/file.rs +@@ -3,7 +3,7 @@ use super::*; + ast_struct! { + /// A complete file of Rust source code. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Example + /// +@@ -37,6 +37,8 @@ ast_struct! { + /// file.read_to_string(&mut src).expect("Unable to read file"); + /// + /// let syntax = syn::parse_file(&src).expect("Unable to parse file"); ++ /// ++ /// // Debug impl is available if Syn is built with "extra-traits" feature. + /// println!("{:#?}", syntax); + /// } + /// ``` +diff --git a/third_party/rust/syn/src/gen/clone.rs b/third_party/rust/syn/src/gen/clone.rs +new file mode 100644 +index 0000000000..bea3887013 +--- /dev/null ++++ b/third_party/rust/syn/src/gen/clone.rs +@@ -0,0 +1,2051 @@ ++// This file is @generated by syn-internal-codegen. ++// It is not intended for manual editing. ++ ++#![allow(clippy::clone_on_copy, clippy::expl_impl_clone_on_copy)] ++use crate::*; ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Abi { ++ fn clone(&self) -> Self { ++ Abi { ++ extern_token: self.extern_token.clone(), ++ name: self.name.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for AngleBracketedGenericArguments { ++ fn clone(&self) -> Self { ++ AngleBracketedGenericArguments { ++ colon2_token: self.colon2_token.clone(), ++ lt_token: self.lt_token.clone(), ++ args: self.args.clone(), ++ gt_token: self.gt_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Arm { ++ fn clone(&self) -> Self { ++ Arm { ++ attrs: self.attrs.clone(), ++ pat: self.pat.clone(), ++ guard: self.guard.clone(), ++ fat_arrow_token: self.fat_arrow_token.clone(), ++ body: self.body.clone(), ++ comma: self.comma.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Copy for AttrStyle {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for AttrStyle { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Attribute { ++ fn clone(&self) -> Self { ++ Attribute { ++ pound_token: self.pound_token.clone(), ++ style: self.style.clone(), ++ bracket_token: self.bracket_token.clone(), ++ path: self.path.clone(), ++ tokens: self.tokens.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for BareFnArg { ++ fn clone(&self) -> Self { ++ BareFnArg { ++ attrs: self.attrs.clone(), ++ name: self.name.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Copy for BinOp {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for BinOp { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Binding { ++ fn clone(&self) -> Self { ++ Binding { ++ ident: self.ident.clone(), ++ eq_token: self.eq_token.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Block { ++ fn clone(&self) -> Self { ++ Block { ++ brace_token: self.brace_token.clone(), ++ stmts: self.stmts.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for BoundLifetimes { ++ fn clone(&self) -> Self { ++ BoundLifetimes { ++ for_token: self.for_token.clone(), ++ lt_token: self.lt_token.clone(), ++ lifetimes: self.lifetimes.clone(), ++ gt_token: self.gt_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ConstParam { ++ fn clone(&self) -> Self { ++ ConstParam { ++ attrs: self.attrs.clone(), ++ const_token: self.const_token.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ eq_token: self.eq_token.clone(), ++ default: self.default.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Constraint { ++ fn clone(&self) -> Self { ++ Constraint { ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Clone for Data { ++ fn clone(&self) -> Self { ++ match self { ++ Data::Struct(v0) => Data::Struct(v0.clone()), ++ Data::Enum(v0) => Data::Enum(v0.clone()), ++ Data::Union(v0) => Data::Union(v0.clone()), ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Clone for DataEnum { ++ fn clone(&self) -> Self { ++ DataEnum { ++ enum_token: self.enum_token.clone(), ++ brace_token: self.brace_token.clone(), ++ variants: self.variants.clone(), ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Clone for DataStruct { ++ fn clone(&self) -> Self { ++ DataStruct { ++ struct_token: self.struct_token.clone(), ++ fields: self.fields.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Clone for DataUnion { ++ fn clone(&self) -> Self { ++ DataUnion { ++ union_token: self.union_token.clone(), ++ fields: self.fields.clone(), ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Clone for DeriveInput { ++ fn clone(&self) -> Self { ++ DeriveInput { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ data: self.data.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Expr { ++ fn clone(&self) -> Self { ++ match self { ++ #[cfg(feature = "full")] ++ Expr::Array(v0) => Expr::Array(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Assign(v0) => Expr::Assign(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::AssignOp(v0) => Expr::AssignOp(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Async(v0) => Expr::Async(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Await(v0) => Expr::Await(v0.clone()), ++ Expr::Binary(v0) => Expr::Binary(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Block(v0) => Expr::Block(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Box(v0) => Expr::Box(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Break(v0) => Expr::Break(v0.clone()), ++ Expr::Call(v0) => Expr::Call(v0.clone()), ++ Expr::Cast(v0) => Expr::Cast(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Closure(v0) => Expr::Closure(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Continue(v0) => Expr::Continue(v0.clone()), ++ Expr::Field(v0) => Expr::Field(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::ForLoop(v0) => Expr::ForLoop(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Group(v0) => Expr::Group(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::If(v0) => Expr::If(v0.clone()), ++ Expr::Index(v0) => Expr::Index(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Let(v0) => Expr::Let(v0.clone()), ++ Expr::Lit(v0) => Expr::Lit(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Loop(v0) => Expr::Loop(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Macro(v0) => Expr::Macro(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Match(v0) => Expr::Match(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::MethodCall(v0) => Expr::MethodCall(v0.clone()), ++ Expr::Paren(v0) => Expr::Paren(v0.clone()), ++ Expr::Path(v0) => Expr::Path(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Range(v0) => Expr::Range(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Reference(v0) => Expr::Reference(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Repeat(v0) => Expr::Repeat(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Return(v0) => Expr::Return(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Struct(v0) => Expr::Struct(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Try(v0) => Expr::Try(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::TryBlock(v0) => Expr::TryBlock(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Tuple(v0) => Expr::Tuple(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Type(v0) => Expr::Type(v0.clone()), ++ Expr::Unary(v0) => Expr::Unary(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Unsafe(v0) => Expr::Unsafe(v0.clone()), ++ Expr::Verbatim(v0) => Expr::Verbatim(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::While(v0) => Expr::While(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Yield(v0) => Expr::Yield(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprArray { ++ fn clone(&self) -> Self { ++ ExprArray { ++ attrs: self.attrs.clone(), ++ bracket_token: self.bracket_token.clone(), ++ elems: self.elems.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprAssign { ++ fn clone(&self) -> Self { ++ ExprAssign { ++ attrs: self.attrs.clone(), ++ left: self.left.clone(), ++ eq_token: self.eq_token.clone(), ++ right: self.right.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprAssignOp { ++ fn clone(&self) -> Self { ++ ExprAssignOp { ++ attrs: self.attrs.clone(), ++ left: self.left.clone(), ++ op: self.op.clone(), ++ right: self.right.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprAsync { ++ fn clone(&self) -> Self { ++ ExprAsync { ++ attrs: self.attrs.clone(), ++ async_token: self.async_token.clone(), ++ capture: self.capture.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprAwait { ++ fn clone(&self) -> Self { ++ ExprAwait { ++ attrs: self.attrs.clone(), ++ base: self.base.clone(), ++ dot_token: self.dot_token.clone(), ++ await_token: self.await_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprBinary { ++ fn clone(&self) -> Self { ++ ExprBinary { ++ attrs: self.attrs.clone(), ++ left: self.left.clone(), ++ op: self.op.clone(), ++ right: self.right.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprBlock { ++ fn clone(&self) -> Self { ++ ExprBlock { ++ attrs: self.attrs.clone(), ++ label: self.label.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprBox { ++ fn clone(&self) -> Self { ++ ExprBox { ++ attrs: self.attrs.clone(), ++ box_token: self.box_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprBreak { ++ fn clone(&self) -> Self { ++ ExprBreak { ++ attrs: self.attrs.clone(), ++ break_token: self.break_token.clone(), ++ label: self.label.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprCall { ++ fn clone(&self) -> Self { ++ ExprCall { ++ attrs: self.attrs.clone(), ++ func: self.func.clone(), ++ paren_token: self.paren_token.clone(), ++ args: self.args.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprCast { ++ fn clone(&self) -> Self { ++ ExprCast { ++ attrs: self.attrs.clone(), ++ expr: self.expr.clone(), ++ as_token: self.as_token.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprClosure { ++ fn clone(&self) -> Self { ++ ExprClosure { ++ attrs: self.attrs.clone(), ++ asyncness: self.asyncness.clone(), ++ movability: self.movability.clone(), ++ capture: self.capture.clone(), ++ or1_token: self.or1_token.clone(), ++ inputs: self.inputs.clone(), ++ or2_token: self.or2_token.clone(), ++ output: self.output.clone(), ++ body: self.body.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprContinue { ++ fn clone(&self) -> Self { ++ ExprContinue { ++ attrs: self.attrs.clone(), ++ continue_token: self.continue_token.clone(), ++ label: self.label.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprField { ++ fn clone(&self) -> Self { ++ ExprField { ++ attrs: self.attrs.clone(), ++ base: self.base.clone(), ++ dot_token: self.dot_token.clone(), ++ member: self.member.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprForLoop { ++ fn clone(&self) -> Self { ++ ExprForLoop { ++ attrs: self.attrs.clone(), ++ label: self.label.clone(), ++ for_token: self.for_token.clone(), ++ pat: self.pat.clone(), ++ in_token: self.in_token.clone(), ++ expr: self.expr.clone(), ++ body: self.body.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprGroup { ++ fn clone(&self) -> Self { ++ ExprGroup { ++ attrs: self.attrs.clone(), ++ group_token: self.group_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprIf { ++ fn clone(&self) -> Self { ++ ExprIf { ++ attrs: self.attrs.clone(), ++ if_token: self.if_token.clone(), ++ cond: self.cond.clone(), ++ then_branch: self.then_branch.clone(), ++ else_branch: self.else_branch.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprIndex { ++ fn clone(&self) -> Self { ++ ExprIndex { ++ attrs: self.attrs.clone(), ++ expr: self.expr.clone(), ++ bracket_token: self.bracket_token.clone(), ++ index: self.index.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprLet { ++ fn clone(&self) -> Self { ++ ExprLet { ++ attrs: self.attrs.clone(), ++ let_token: self.let_token.clone(), ++ pat: self.pat.clone(), ++ eq_token: self.eq_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprLit { ++ fn clone(&self) -> Self { ++ ExprLit { ++ attrs: self.attrs.clone(), ++ lit: self.lit.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprLoop { ++ fn clone(&self) -> Self { ++ ExprLoop { ++ attrs: self.attrs.clone(), ++ label: self.label.clone(), ++ loop_token: self.loop_token.clone(), ++ body: self.body.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprMacro { ++ fn clone(&self) -> Self { ++ ExprMacro { ++ attrs: self.attrs.clone(), ++ mac: self.mac.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprMatch { ++ fn clone(&self) -> Self { ++ ExprMatch { ++ attrs: self.attrs.clone(), ++ match_token: self.match_token.clone(), ++ expr: self.expr.clone(), ++ brace_token: self.brace_token.clone(), ++ arms: self.arms.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprMethodCall { ++ fn clone(&self) -> Self { ++ ExprMethodCall { ++ attrs: self.attrs.clone(), ++ receiver: self.receiver.clone(), ++ dot_token: self.dot_token.clone(), ++ method: self.method.clone(), ++ turbofish: self.turbofish.clone(), ++ paren_token: self.paren_token.clone(), ++ args: self.args.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprParen { ++ fn clone(&self) -> Self { ++ ExprParen { ++ attrs: self.attrs.clone(), ++ paren_token: self.paren_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprPath { ++ fn clone(&self) -> Self { ++ ExprPath { ++ attrs: self.attrs.clone(), ++ qself: self.qself.clone(), ++ path: self.path.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprRange { ++ fn clone(&self) -> Self { ++ ExprRange { ++ attrs: self.attrs.clone(), ++ from: self.from.clone(), ++ limits: self.limits.clone(), ++ to: self.to.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprReference { ++ fn clone(&self) -> Self { ++ ExprReference { ++ attrs: self.attrs.clone(), ++ and_token: self.and_token.clone(), ++ raw: self.raw.clone(), ++ mutability: self.mutability.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprRepeat { ++ fn clone(&self) -> Self { ++ ExprRepeat { ++ attrs: self.attrs.clone(), ++ bracket_token: self.bracket_token.clone(), ++ expr: self.expr.clone(), ++ semi_token: self.semi_token.clone(), ++ len: self.len.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprReturn { ++ fn clone(&self) -> Self { ++ ExprReturn { ++ attrs: self.attrs.clone(), ++ return_token: self.return_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprStruct { ++ fn clone(&self) -> Self { ++ ExprStruct { ++ attrs: self.attrs.clone(), ++ path: self.path.clone(), ++ brace_token: self.brace_token.clone(), ++ fields: self.fields.clone(), ++ dot2_token: self.dot2_token.clone(), ++ rest: self.rest.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprTry { ++ fn clone(&self) -> Self { ++ ExprTry { ++ attrs: self.attrs.clone(), ++ expr: self.expr.clone(), ++ question_token: self.question_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprTryBlock { ++ fn clone(&self) -> Self { ++ ExprTryBlock { ++ attrs: self.attrs.clone(), ++ try_token: self.try_token.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprTuple { ++ fn clone(&self) -> Self { ++ ExprTuple { ++ attrs: self.attrs.clone(), ++ paren_token: self.paren_token.clone(), ++ elems: self.elems.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprType { ++ fn clone(&self) -> Self { ++ ExprType { ++ attrs: self.attrs.clone(), ++ expr: self.expr.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprUnary { ++ fn clone(&self) -> Self { ++ ExprUnary { ++ attrs: self.attrs.clone(), ++ op: self.op.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprUnsafe { ++ fn clone(&self) -> Self { ++ ExprUnsafe { ++ attrs: self.attrs.clone(), ++ unsafe_token: self.unsafe_token.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprWhile { ++ fn clone(&self) -> Self { ++ ExprWhile { ++ attrs: self.attrs.clone(), ++ label: self.label.clone(), ++ while_token: self.while_token.clone(), ++ cond: self.cond.clone(), ++ body: self.body.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprYield { ++ fn clone(&self) -> Self { ++ ExprYield { ++ attrs: self.attrs.clone(), ++ yield_token: self.yield_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Field { ++ fn clone(&self) -> Self { ++ Field { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for FieldPat { ++ fn clone(&self) -> Self { ++ FieldPat { ++ attrs: self.attrs.clone(), ++ member: self.member.clone(), ++ colon_token: self.colon_token.clone(), ++ pat: self.pat.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for FieldValue { ++ fn clone(&self) -> Self { ++ FieldValue { ++ attrs: self.attrs.clone(), ++ member: self.member.clone(), ++ colon_token: self.colon_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Fields { ++ fn clone(&self) -> Self { ++ match self { ++ Fields::Named(v0) => Fields::Named(v0.clone()), ++ Fields::Unnamed(v0) => Fields::Unnamed(v0.clone()), ++ Fields::Unit => Fields::Unit, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for FieldsNamed { ++ fn clone(&self) -> Self { ++ FieldsNamed { ++ brace_token: self.brace_token.clone(), ++ named: self.named.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for FieldsUnnamed { ++ fn clone(&self) -> Self { ++ FieldsUnnamed { ++ paren_token: self.paren_token.clone(), ++ unnamed: self.unnamed.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for File { ++ fn clone(&self) -> Self { ++ File { ++ shebang: self.shebang.clone(), ++ attrs: self.attrs.clone(), ++ items: self.items.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for FnArg { ++ fn clone(&self) -> Self { ++ match self { ++ FnArg::Receiver(v0) => FnArg::Receiver(v0.clone()), ++ FnArg::Typed(v0) => FnArg::Typed(v0.clone()), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ForeignItem { ++ fn clone(&self) -> Self { ++ match self { ++ ForeignItem::Fn(v0) => ForeignItem::Fn(v0.clone()), ++ ForeignItem::Static(v0) => ForeignItem::Static(v0.clone()), ++ ForeignItem::Type(v0) => ForeignItem::Type(v0.clone()), ++ ForeignItem::Macro(v0) => ForeignItem::Macro(v0.clone()), ++ ForeignItem::Verbatim(v0) => ForeignItem::Verbatim(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ForeignItemFn { ++ fn clone(&self) -> Self { ++ ForeignItemFn { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ sig: self.sig.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ForeignItemMacro { ++ fn clone(&self) -> Self { ++ ForeignItemMacro { ++ attrs: self.attrs.clone(), ++ mac: self.mac.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ForeignItemStatic { ++ fn clone(&self) -> Self { ++ ForeignItemStatic { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ static_token: self.static_token.clone(), ++ mutability: self.mutability.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ForeignItemType { ++ fn clone(&self) -> Self { ++ ForeignItemType { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ type_token: self.type_token.clone(), ++ ident: self.ident.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for GenericArgument { ++ fn clone(&self) -> Self { ++ match self { ++ GenericArgument::Lifetime(v0) => GenericArgument::Lifetime(v0.clone()), ++ GenericArgument::Type(v0) => GenericArgument::Type(v0.clone()), ++ GenericArgument::Binding(v0) => GenericArgument::Binding(v0.clone()), ++ GenericArgument::Constraint(v0) => GenericArgument::Constraint(v0.clone()), ++ GenericArgument::Const(v0) => GenericArgument::Const(v0.clone()), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for GenericMethodArgument { ++ fn clone(&self) -> Self { ++ match self { ++ GenericMethodArgument::Type(v0) => GenericMethodArgument::Type(v0.clone()), ++ GenericMethodArgument::Const(v0) => GenericMethodArgument::Const(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for GenericParam { ++ fn clone(&self) -> Self { ++ match self { ++ GenericParam::Type(v0) => GenericParam::Type(v0.clone()), ++ GenericParam::Lifetime(v0) => GenericParam::Lifetime(v0.clone()), ++ GenericParam::Const(v0) => GenericParam::Const(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Generics { ++ fn clone(&self) -> Self { ++ Generics { ++ lt_token: self.lt_token.clone(), ++ params: self.params.clone(), ++ gt_token: self.gt_token.clone(), ++ where_clause: self.where_clause.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ImplItem { ++ fn clone(&self) -> Self { ++ match self { ++ ImplItem::Const(v0) => ImplItem::Const(v0.clone()), ++ ImplItem::Method(v0) => ImplItem::Method(v0.clone()), ++ ImplItem::Type(v0) => ImplItem::Type(v0.clone()), ++ ImplItem::Macro(v0) => ImplItem::Macro(v0.clone()), ++ ImplItem::Verbatim(v0) => ImplItem::Verbatim(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ImplItemConst { ++ fn clone(&self) -> Self { ++ ImplItemConst { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ defaultness: self.defaultness.clone(), ++ const_token: self.const_token.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ eq_token: self.eq_token.clone(), ++ expr: self.expr.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ImplItemMacro { ++ fn clone(&self) -> Self { ++ ImplItemMacro { ++ attrs: self.attrs.clone(), ++ mac: self.mac.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ImplItemMethod { ++ fn clone(&self) -> Self { ++ ImplItemMethod { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ defaultness: self.defaultness.clone(), ++ sig: self.sig.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ImplItemType { ++ fn clone(&self) -> Self { ++ ImplItemType { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ defaultness: self.defaultness.clone(), ++ type_token: self.type_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ eq_token: self.eq_token.clone(), ++ ty: self.ty.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Index { ++ fn clone(&self) -> Self { ++ Index { ++ index: self.index.clone(), ++ span: self.span.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Item { ++ fn clone(&self) -> Self { ++ match self { ++ Item::Const(v0) => Item::Const(v0.clone()), ++ Item::Enum(v0) => Item::Enum(v0.clone()), ++ Item::ExternCrate(v0) => Item::ExternCrate(v0.clone()), ++ Item::Fn(v0) => Item::Fn(v0.clone()), ++ Item::ForeignMod(v0) => Item::ForeignMod(v0.clone()), ++ Item::Impl(v0) => Item::Impl(v0.clone()), ++ Item::Macro(v0) => Item::Macro(v0.clone()), ++ Item::Macro2(v0) => Item::Macro2(v0.clone()), ++ Item::Mod(v0) => Item::Mod(v0.clone()), ++ Item::Static(v0) => Item::Static(v0.clone()), ++ Item::Struct(v0) => Item::Struct(v0.clone()), ++ Item::Trait(v0) => Item::Trait(v0.clone()), ++ Item::TraitAlias(v0) => Item::TraitAlias(v0.clone()), ++ Item::Type(v0) => Item::Type(v0.clone()), ++ Item::Union(v0) => Item::Union(v0.clone()), ++ Item::Use(v0) => Item::Use(v0.clone()), ++ Item::Verbatim(v0) => Item::Verbatim(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemConst { ++ fn clone(&self) -> Self { ++ ItemConst { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ const_token: self.const_token.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ eq_token: self.eq_token.clone(), ++ expr: self.expr.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemEnum { ++ fn clone(&self) -> Self { ++ ItemEnum { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ enum_token: self.enum_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ brace_token: self.brace_token.clone(), ++ variants: self.variants.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemExternCrate { ++ fn clone(&self) -> Self { ++ ItemExternCrate { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ extern_token: self.extern_token.clone(), ++ crate_token: self.crate_token.clone(), ++ ident: self.ident.clone(), ++ rename: self.rename.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemFn { ++ fn clone(&self) -> Self { ++ ItemFn { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ sig: self.sig.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemForeignMod { ++ fn clone(&self) -> Self { ++ ItemForeignMod { ++ attrs: self.attrs.clone(), ++ abi: self.abi.clone(), ++ brace_token: self.brace_token.clone(), ++ items: self.items.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemImpl { ++ fn clone(&self) -> Self { ++ ItemImpl { ++ attrs: self.attrs.clone(), ++ defaultness: self.defaultness.clone(), ++ unsafety: self.unsafety.clone(), ++ impl_token: self.impl_token.clone(), ++ generics: self.generics.clone(), ++ trait_: self.trait_.clone(), ++ self_ty: self.self_ty.clone(), ++ brace_token: self.brace_token.clone(), ++ items: self.items.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemMacro { ++ fn clone(&self) -> Self { ++ ItemMacro { ++ attrs: self.attrs.clone(), ++ ident: self.ident.clone(), ++ mac: self.mac.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemMacro2 { ++ fn clone(&self) -> Self { ++ ItemMacro2 { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ macro_token: self.macro_token.clone(), ++ ident: self.ident.clone(), ++ rules: self.rules.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemMod { ++ fn clone(&self) -> Self { ++ ItemMod { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ mod_token: self.mod_token.clone(), ++ ident: self.ident.clone(), ++ content: self.content.clone(), ++ semi: self.semi.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemStatic { ++ fn clone(&self) -> Self { ++ ItemStatic { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ static_token: self.static_token.clone(), ++ mutability: self.mutability.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ eq_token: self.eq_token.clone(), ++ expr: self.expr.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemStruct { ++ fn clone(&self) -> Self { ++ ItemStruct { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ struct_token: self.struct_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ fields: self.fields.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemTrait { ++ fn clone(&self) -> Self { ++ ItemTrait { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ unsafety: self.unsafety.clone(), ++ auto_token: self.auto_token.clone(), ++ trait_token: self.trait_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ colon_token: self.colon_token.clone(), ++ supertraits: self.supertraits.clone(), ++ brace_token: self.brace_token.clone(), ++ items: self.items.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemTraitAlias { ++ fn clone(&self) -> Self { ++ ItemTraitAlias { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ trait_token: self.trait_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ eq_token: self.eq_token.clone(), ++ bounds: self.bounds.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemType { ++ fn clone(&self) -> Self { ++ ItemType { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ type_token: self.type_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ eq_token: self.eq_token.clone(), ++ ty: self.ty.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemUnion { ++ fn clone(&self) -> Self { ++ ItemUnion { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ union_token: self.union_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ fields: self.fields.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemUse { ++ fn clone(&self) -> Self { ++ ItemUse { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ use_token: self.use_token.clone(), ++ leading_colon: self.leading_colon.clone(), ++ tree: self.tree.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Label { ++ fn clone(&self) -> Self { ++ Label { ++ name: self.name.clone(), ++ colon_token: self.colon_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for LifetimeDef { ++ fn clone(&self) -> Self { ++ LifetimeDef { ++ attrs: self.attrs.clone(), ++ lifetime: self.lifetime.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++impl Clone for Lit { ++ fn clone(&self) -> Self { ++ match self { ++ Lit::Str(v0) => Lit::Str(v0.clone()), ++ Lit::ByteStr(v0) => Lit::ByteStr(v0.clone()), ++ Lit::Byte(v0) => Lit::Byte(v0.clone()), ++ Lit::Char(v0) => Lit::Char(v0.clone()), ++ Lit::Int(v0) => Lit::Int(v0.clone()), ++ Lit::Float(v0) => Lit::Float(v0.clone()), ++ Lit::Bool(v0) => Lit::Bool(v0.clone()), ++ Lit::Verbatim(v0) => Lit::Verbatim(v0.clone()), ++ } ++ } ++} ++impl Clone for LitBool { ++ fn clone(&self) -> Self { ++ LitBool { ++ value: self.value.clone(), ++ span: self.span.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Local { ++ fn clone(&self) -> Self { ++ Local { ++ attrs: self.attrs.clone(), ++ let_token: self.let_token.clone(), ++ pat: self.pat.clone(), ++ init: self.init.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Macro { ++ fn clone(&self) -> Self { ++ Macro { ++ path: self.path.clone(), ++ bang_token: self.bang_token.clone(), ++ delimiter: self.delimiter.clone(), ++ tokens: self.tokens.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for MacroDelimiter { ++ fn clone(&self) -> Self { ++ match self { ++ MacroDelimiter::Paren(v0) => MacroDelimiter::Paren(v0.clone()), ++ MacroDelimiter::Brace(v0) => MacroDelimiter::Brace(v0.clone()), ++ MacroDelimiter::Bracket(v0) => MacroDelimiter::Bracket(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Member { ++ fn clone(&self) -> Self { ++ match self { ++ Member::Named(v0) => Member::Named(v0.clone()), ++ Member::Unnamed(v0) => Member::Unnamed(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Meta { ++ fn clone(&self) -> Self { ++ match self { ++ Meta::Path(v0) => Meta::Path(v0.clone()), ++ Meta::List(v0) => Meta::List(v0.clone()), ++ Meta::NameValue(v0) => Meta::NameValue(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for MetaList { ++ fn clone(&self) -> Self { ++ MetaList { ++ path: self.path.clone(), ++ paren_token: self.paren_token.clone(), ++ nested: self.nested.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for MetaNameValue { ++ fn clone(&self) -> Self { ++ MetaNameValue { ++ path: self.path.clone(), ++ eq_token: self.eq_token.clone(), ++ lit: self.lit.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for MethodTurbofish { ++ fn clone(&self) -> Self { ++ MethodTurbofish { ++ colon2_token: self.colon2_token.clone(), ++ lt_token: self.lt_token.clone(), ++ args: self.args.clone(), ++ gt_token: self.gt_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for NestedMeta { ++ fn clone(&self) -> Self { ++ match self { ++ NestedMeta::Meta(v0) => NestedMeta::Meta(v0.clone()), ++ NestedMeta::Lit(v0) => NestedMeta::Lit(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ParenthesizedGenericArguments { ++ fn clone(&self) -> Self { ++ ParenthesizedGenericArguments { ++ paren_token: self.paren_token.clone(), ++ inputs: self.inputs.clone(), ++ output: self.output.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Pat { ++ fn clone(&self) -> Self { ++ match self { ++ Pat::Box(v0) => Pat::Box(v0.clone()), ++ Pat::Ident(v0) => Pat::Ident(v0.clone()), ++ Pat::Lit(v0) => Pat::Lit(v0.clone()), ++ Pat::Macro(v0) => Pat::Macro(v0.clone()), ++ Pat::Or(v0) => Pat::Or(v0.clone()), ++ Pat::Path(v0) => Pat::Path(v0.clone()), ++ Pat::Range(v0) => Pat::Range(v0.clone()), ++ Pat::Reference(v0) => Pat::Reference(v0.clone()), ++ Pat::Rest(v0) => Pat::Rest(v0.clone()), ++ Pat::Slice(v0) => Pat::Slice(v0.clone()), ++ Pat::Struct(v0) => Pat::Struct(v0.clone()), ++ Pat::Tuple(v0) => Pat::Tuple(v0.clone()), ++ Pat::TupleStruct(v0) => Pat::TupleStruct(v0.clone()), ++ Pat::Type(v0) => Pat::Type(v0.clone()), ++ Pat::Verbatim(v0) => Pat::Verbatim(v0.clone()), ++ Pat::Wild(v0) => Pat::Wild(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatBox { ++ fn clone(&self) -> Self { ++ PatBox { ++ attrs: self.attrs.clone(), ++ box_token: self.box_token.clone(), ++ pat: self.pat.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatIdent { ++ fn clone(&self) -> Self { ++ PatIdent { ++ attrs: self.attrs.clone(), ++ by_ref: self.by_ref.clone(), ++ mutability: self.mutability.clone(), ++ ident: self.ident.clone(), ++ subpat: self.subpat.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatLit { ++ fn clone(&self) -> Self { ++ PatLit { ++ attrs: self.attrs.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatMacro { ++ fn clone(&self) -> Self { ++ PatMacro { ++ attrs: self.attrs.clone(), ++ mac: self.mac.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatOr { ++ fn clone(&self) -> Self { ++ PatOr { ++ attrs: self.attrs.clone(), ++ leading_vert: self.leading_vert.clone(), ++ cases: self.cases.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatPath { ++ fn clone(&self) -> Self { ++ PatPath { ++ attrs: self.attrs.clone(), ++ qself: self.qself.clone(), ++ path: self.path.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatRange { ++ fn clone(&self) -> Self { ++ PatRange { ++ attrs: self.attrs.clone(), ++ lo: self.lo.clone(), ++ limits: self.limits.clone(), ++ hi: self.hi.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatReference { ++ fn clone(&self) -> Self { ++ PatReference { ++ attrs: self.attrs.clone(), ++ and_token: self.and_token.clone(), ++ mutability: self.mutability.clone(), ++ pat: self.pat.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatRest { ++ fn clone(&self) -> Self { ++ PatRest { ++ attrs: self.attrs.clone(), ++ dot2_token: self.dot2_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatSlice { ++ fn clone(&self) -> Self { ++ PatSlice { ++ attrs: self.attrs.clone(), ++ bracket_token: self.bracket_token.clone(), ++ elems: self.elems.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatStruct { ++ fn clone(&self) -> Self { ++ PatStruct { ++ attrs: self.attrs.clone(), ++ path: self.path.clone(), ++ brace_token: self.brace_token.clone(), ++ fields: self.fields.clone(), ++ dot2_token: self.dot2_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatTuple { ++ fn clone(&self) -> Self { ++ PatTuple { ++ attrs: self.attrs.clone(), ++ paren_token: self.paren_token.clone(), ++ elems: self.elems.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatTupleStruct { ++ fn clone(&self) -> Self { ++ PatTupleStruct { ++ attrs: self.attrs.clone(), ++ path: self.path.clone(), ++ pat: self.pat.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatType { ++ fn clone(&self) -> Self { ++ PatType { ++ attrs: self.attrs.clone(), ++ pat: self.pat.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatWild { ++ fn clone(&self) -> Self { ++ PatWild { ++ attrs: self.attrs.clone(), ++ underscore_token: self.underscore_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Path { ++ fn clone(&self) -> Self { ++ Path { ++ leading_colon: self.leading_colon.clone(), ++ segments: self.segments.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for PathArguments { ++ fn clone(&self) -> Self { ++ match self { ++ PathArguments::None => PathArguments::None, ++ PathArguments::AngleBracketed(v0) => PathArguments::AngleBracketed(v0.clone()), ++ PathArguments::Parenthesized(v0) => PathArguments::Parenthesized(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for PathSegment { ++ fn clone(&self) -> Self { ++ PathSegment { ++ ident: self.ident.clone(), ++ arguments: self.arguments.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for PredicateEq { ++ fn clone(&self) -> Self { ++ PredicateEq { ++ lhs_ty: self.lhs_ty.clone(), ++ eq_token: self.eq_token.clone(), ++ rhs_ty: self.rhs_ty.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for PredicateLifetime { ++ fn clone(&self) -> Self { ++ PredicateLifetime { ++ lifetime: self.lifetime.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for PredicateType { ++ fn clone(&self) -> Self { ++ PredicateType { ++ lifetimes: self.lifetimes.clone(), ++ bounded_ty: self.bounded_ty.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for QSelf { ++ fn clone(&self) -> Self { ++ QSelf { ++ lt_token: self.lt_token.clone(), ++ ty: self.ty.clone(), ++ position: self.position.clone(), ++ as_token: self.as_token.clone(), ++ gt_token: self.gt_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Copy for RangeLimits {} ++#[cfg(feature = "full")] ++impl Clone for RangeLimits { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Receiver { ++ fn clone(&self) -> Self { ++ Receiver { ++ attrs: self.attrs.clone(), ++ reference: self.reference.clone(), ++ mutability: self.mutability.clone(), ++ self_token: self.self_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ReturnType { ++ fn clone(&self) -> Self { ++ match self { ++ ReturnType::Default => ReturnType::Default, ++ ReturnType::Type(v0, v1) => ReturnType::Type(v0.clone(), v1.clone()), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Signature { ++ fn clone(&self) -> Self { ++ Signature { ++ constness: self.constness.clone(), ++ asyncness: self.asyncness.clone(), ++ unsafety: self.unsafety.clone(), ++ abi: self.abi.clone(), ++ fn_token: self.fn_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ paren_token: self.paren_token.clone(), ++ inputs: self.inputs.clone(), ++ variadic: self.variadic.clone(), ++ output: self.output.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Stmt { ++ fn clone(&self) -> Self { ++ match self { ++ Stmt::Local(v0) => Stmt::Local(v0.clone()), ++ Stmt::Item(v0) => Stmt::Item(v0.clone()), ++ Stmt::Expr(v0) => Stmt::Expr(v0.clone()), ++ Stmt::Semi(v0, v1) => Stmt::Semi(v0.clone(), v1.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TraitBound { ++ fn clone(&self) -> Self { ++ TraitBound { ++ paren_token: self.paren_token.clone(), ++ modifier: self.modifier.clone(), ++ lifetimes: self.lifetimes.clone(), ++ path: self.path.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Copy for TraitBoundModifier {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TraitBoundModifier { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for TraitItem { ++ fn clone(&self) -> Self { ++ match self { ++ TraitItem::Const(v0) => TraitItem::Const(v0.clone()), ++ TraitItem::Method(v0) => TraitItem::Method(v0.clone()), ++ TraitItem::Type(v0) => TraitItem::Type(v0.clone()), ++ TraitItem::Macro(v0) => TraitItem::Macro(v0.clone()), ++ TraitItem::Verbatim(v0) => TraitItem::Verbatim(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for TraitItemConst { ++ fn clone(&self) -> Self { ++ TraitItemConst { ++ attrs: self.attrs.clone(), ++ const_token: self.const_token.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ default: self.default.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for TraitItemMacro { ++ fn clone(&self) -> Self { ++ TraitItemMacro { ++ attrs: self.attrs.clone(), ++ mac: self.mac.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for TraitItemMethod { ++ fn clone(&self) -> Self { ++ TraitItemMethod { ++ attrs: self.attrs.clone(), ++ sig: self.sig.clone(), ++ default: self.default.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for TraitItemType { ++ fn clone(&self) -> Self { ++ TraitItemType { ++ attrs: self.attrs.clone(), ++ type_token: self.type_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ default: self.default.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Type { ++ fn clone(&self) -> Self { ++ match self { ++ Type::Array(v0) => Type::Array(v0.clone()), ++ Type::BareFn(v0) => Type::BareFn(v0.clone()), ++ Type::Group(v0) => Type::Group(v0.clone()), ++ Type::ImplTrait(v0) => Type::ImplTrait(v0.clone()), ++ Type::Infer(v0) => Type::Infer(v0.clone()), ++ Type::Macro(v0) => Type::Macro(v0.clone()), ++ Type::Never(v0) => Type::Never(v0.clone()), ++ Type::Paren(v0) => Type::Paren(v0.clone()), ++ Type::Path(v0) => Type::Path(v0.clone()), ++ Type::Ptr(v0) => Type::Ptr(v0.clone()), ++ Type::Reference(v0) => Type::Reference(v0.clone()), ++ Type::Slice(v0) => Type::Slice(v0.clone()), ++ Type::TraitObject(v0) => Type::TraitObject(v0.clone()), ++ Type::Tuple(v0) => Type::Tuple(v0.clone()), ++ Type::Verbatim(v0) => Type::Verbatim(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeArray { ++ fn clone(&self) -> Self { ++ TypeArray { ++ bracket_token: self.bracket_token.clone(), ++ elem: self.elem.clone(), ++ semi_token: self.semi_token.clone(), ++ len: self.len.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeBareFn { ++ fn clone(&self) -> Self { ++ TypeBareFn { ++ lifetimes: self.lifetimes.clone(), ++ unsafety: self.unsafety.clone(), ++ abi: self.abi.clone(), ++ fn_token: self.fn_token.clone(), ++ paren_token: self.paren_token.clone(), ++ inputs: self.inputs.clone(), ++ variadic: self.variadic.clone(), ++ output: self.output.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeGroup { ++ fn clone(&self) -> Self { ++ TypeGroup { ++ group_token: self.group_token.clone(), ++ elem: self.elem.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeImplTrait { ++ fn clone(&self) -> Self { ++ TypeImplTrait { ++ impl_token: self.impl_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeInfer { ++ fn clone(&self) -> Self { ++ TypeInfer { ++ underscore_token: self.underscore_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeMacro { ++ fn clone(&self) -> Self { ++ TypeMacro { ++ mac: self.mac.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeNever { ++ fn clone(&self) -> Self { ++ TypeNever { ++ bang_token: self.bang_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeParam { ++ fn clone(&self) -> Self { ++ TypeParam { ++ attrs: self.attrs.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ eq_token: self.eq_token.clone(), ++ default: self.default.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeParamBound { ++ fn clone(&self) -> Self { ++ match self { ++ TypeParamBound::Trait(v0) => TypeParamBound::Trait(v0.clone()), ++ TypeParamBound::Lifetime(v0) => TypeParamBound::Lifetime(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeParen { ++ fn clone(&self) -> Self { ++ TypeParen { ++ paren_token: self.paren_token.clone(), ++ elem: self.elem.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypePath { ++ fn clone(&self) -> Self { ++ TypePath { ++ qself: self.qself.clone(), ++ path: self.path.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypePtr { ++ fn clone(&self) -> Self { ++ TypePtr { ++ star_token: self.star_token.clone(), ++ const_token: self.const_token.clone(), ++ mutability: self.mutability.clone(), ++ elem: self.elem.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeReference { ++ fn clone(&self) -> Self { ++ TypeReference { ++ and_token: self.and_token.clone(), ++ lifetime: self.lifetime.clone(), ++ mutability: self.mutability.clone(), ++ elem: self.elem.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeSlice { ++ fn clone(&self) -> Self { ++ TypeSlice { ++ bracket_token: self.bracket_token.clone(), ++ elem: self.elem.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeTraitObject { ++ fn clone(&self) -> Self { ++ TypeTraitObject { ++ dyn_token: self.dyn_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeTuple { ++ fn clone(&self) -> Self { ++ TypeTuple { ++ paren_token: self.paren_token.clone(), ++ elems: self.elems.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Copy for UnOp {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for UnOp { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UseGlob { ++ fn clone(&self) -> Self { ++ UseGlob { ++ star_token: self.star_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UseGroup { ++ fn clone(&self) -> Self { ++ UseGroup { ++ brace_token: self.brace_token.clone(), ++ items: self.items.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UseName { ++ fn clone(&self) -> Self { ++ UseName { ++ ident: self.ident.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UsePath { ++ fn clone(&self) -> Self { ++ UsePath { ++ ident: self.ident.clone(), ++ colon2_token: self.colon2_token.clone(), ++ tree: self.tree.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UseRename { ++ fn clone(&self) -> Self { ++ UseRename { ++ ident: self.ident.clone(), ++ as_token: self.as_token.clone(), ++ rename: self.rename.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UseTree { ++ fn clone(&self) -> Self { ++ match self { ++ UseTree::Path(v0) => UseTree::Path(v0.clone()), ++ UseTree::Name(v0) => UseTree::Name(v0.clone()), ++ UseTree::Rename(v0) => UseTree::Rename(v0.clone()), ++ UseTree::Glob(v0) => UseTree::Glob(v0.clone()), ++ UseTree::Group(v0) => UseTree::Group(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Variadic { ++ fn clone(&self) -> Self { ++ Variadic { ++ attrs: self.attrs.clone(), ++ dots: self.dots.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Variant { ++ fn clone(&self) -> Self { ++ Variant { ++ attrs: self.attrs.clone(), ++ ident: self.ident.clone(), ++ fields: self.fields.clone(), ++ discriminant: self.discriminant.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for VisCrate { ++ fn clone(&self) -> Self { ++ VisCrate { ++ crate_token: self.crate_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for VisPublic { ++ fn clone(&self) -> Self { ++ VisPublic { ++ pub_token: self.pub_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for VisRestricted { ++ fn clone(&self) -> Self { ++ VisRestricted { ++ pub_token: self.pub_token.clone(), ++ paren_token: self.paren_token.clone(), ++ in_token: self.in_token.clone(), ++ path: self.path.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Visibility { ++ fn clone(&self) -> Self { ++ match self { ++ Visibility::Public(v0) => Visibility::Public(v0.clone()), ++ Visibility::Crate(v0) => Visibility::Crate(v0.clone()), ++ Visibility::Restricted(v0) => Visibility::Restricted(v0.clone()), ++ Visibility::Inherited => Visibility::Inherited, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for WhereClause { ++ fn clone(&self) -> Self { ++ WhereClause { ++ where_token: self.where_token.clone(), ++ predicates: self.predicates.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for WherePredicate { ++ fn clone(&self) -> Self { ++ match self { ++ WherePredicate::Type(v0) => WherePredicate::Type(v0.clone()), ++ WherePredicate::Lifetime(v0) => WherePredicate::Lifetime(v0.clone()), ++ WherePredicate::Eq(v0) => WherePredicate::Eq(v0.clone()), ++ } ++ } ++} +diff --git a/third_party/rust/syn/src/gen/debug.rs b/third_party/rust/syn/src/gen/debug.rs +new file mode 100644 +index 0000000000..72baab05f4 +--- /dev/null ++++ b/third_party/rust/syn/src/gen/debug.rs +@@ -0,0 +1,2857 @@ ++// This file is @generated by syn-internal-codegen. ++// It is not intended for manual editing. ++ ++use crate::*; ++use std::fmt::{self, Debug}; ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Abi { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Abi"); ++ formatter.field("extern_token", &self.extern_token); ++ formatter.field("name", &self.name); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for AngleBracketedGenericArguments { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("AngleBracketedGenericArguments"); ++ formatter.field("colon2_token", &self.colon2_token); ++ formatter.field("lt_token", &self.lt_token); ++ formatter.field("args", &self.args); ++ formatter.field("gt_token", &self.gt_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Arm { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Arm"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("pat", &self.pat); ++ formatter.field("guard", &self.guard); ++ formatter.field("fat_arrow_token", &self.fat_arrow_token); ++ formatter.field("body", &self.body); ++ formatter.field("comma", &self.comma); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for AttrStyle { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ AttrStyle::Outer => formatter.write_str("Outer"), ++ AttrStyle::Inner(v0) => { ++ let mut formatter = formatter.debug_tuple("Inner"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Attribute { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Attribute"); ++ formatter.field("pound_token", &self.pound_token); ++ formatter.field("style", &self.style); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("path", &self.path); ++ formatter.field("tokens", &self.tokens); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for BareFnArg { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("BareFnArg"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("name", &self.name); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for BinOp { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ BinOp::Add(v0) => { ++ let mut formatter = formatter.debug_tuple("Add"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Sub(v0) => { ++ let mut formatter = formatter.debug_tuple("Sub"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Mul(v0) => { ++ let mut formatter = formatter.debug_tuple("Mul"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Div(v0) => { ++ let mut formatter = formatter.debug_tuple("Div"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Rem(v0) => { ++ let mut formatter = formatter.debug_tuple("Rem"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::And(v0) => { ++ let mut formatter = formatter.debug_tuple("And"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Or(v0) => { ++ let mut formatter = formatter.debug_tuple("Or"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitXor(v0) => { ++ let mut formatter = formatter.debug_tuple("BitXor"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitAnd(v0) => { ++ let mut formatter = formatter.debug_tuple("BitAnd"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitOr(v0) => { ++ let mut formatter = formatter.debug_tuple("BitOr"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Shl(v0) => { ++ let mut formatter = formatter.debug_tuple("Shl"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Shr(v0) => { ++ let mut formatter = formatter.debug_tuple("Shr"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Eq(v0) => { ++ let mut formatter = formatter.debug_tuple("Eq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Lt(v0) => { ++ let mut formatter = formatter.debug_tuple("Lt"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Le(v0) => { ++ let mut formatter = formatter.debug_tuple("Le"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Ne(v0) => { ++ let mut formatter = formatter.debug_tuple("Ne"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Ge(v0) => { ++ let mut formatter = formatter.debug_tuple("Ge"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Gt(v0) => { ++ let mut formatter = formatter.debug_tuple("Gt"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::AddEq(v0) => { ++ let mut formatter = formatter.debug_tuple("AddEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::SubEq(v0) => { ++ let mut formatter = formatter.debug_tuple("SubEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::MulEq(v0) => { ++ let mut formatter = formatter.debug_tuple("MulEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::DivEq(v0) => { ++ let mut formatter = formatter.debug_tuple("DivEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::RemEq(v0) => { ++ let mut formatter = formatter.debug_tuple("RemEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitXorEq(v0) => { ++ let mut formatter = formatter.debug_tuple("BitXorEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitAndEq(v0) => { ++ let mut formatter = formatter.debug_tuple("BitAndEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitOrEq(v0) => { ++ let mut formatter = formatter.debug_tuple("BitOrEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::ShlEq(v0) => { ++ let mut formatter = formatter.debug_tuple("ShlEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::ShrEq(v0) => { ++ let mut formatter = formatter.debug_tuple("ShrEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Binding { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Binding"); ++ formatter.field("ident", &self.ident); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Block { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Block"); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("stmts", &self.stmts); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for BoundLifetimes { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("BoundLifetimes"); ++ formatter.field("for_token", &self.for_token); ++ formatter.field("lt_token", &self.lt_token); ++ formatter.field("lifetimes", &self.lifetimes); ++ formatter.field("gt_token", &self.gt_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ConstParam { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ConstParam"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("const_token", &self.const_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("default", &self.default); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Constraint { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Constraint"); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "derive")] ++impl Debug for Data { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Data::Struct(v0) => { ++ let mut formatter = formatter.debug_tuple("Struct"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Data::Enum(v0) => { ++ let mut formatter = formatter.debug_tuple("Enum"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Data::Union(v0) => { ++ let mut formatter = formatter.debug_tuple("Union"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Debug for DataEnum { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("DataEnum"); ++ formatter.field("enum_token", &self.enum_token); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("variants", &self.variants); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "derive")] ++impl Debug for DataStruct { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("DataStruct"); ++ formatter.field("struct_token", &self.struct_token); ++ formatter.field("fields", &self.fields); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "derive")] ++impl Debug for DataUnion { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("DataUnion"); ++ formatter.field("union_token", &self.union_token); ++ formatter.field("fields", &self.fields); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "derive")] ++impl Debug for DeriveInput { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("DeriveInput"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("data", &self.data); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Expr { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ #[cfg(feature = "full")] ++ Expr::Array(v0) => { ++ let mut formatter = formatter.debug_tuple("Array"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Assign(v0) => { ++ let mut formatter = formatter.debug_tuple("Assign"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::AssignOp(v0) => { ++ let mut formatter = formatter.debug_tuple("AssignOp"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Async(v0) => { ++ let mut formatter = formatter.debug_tuple("Async"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Await(v0) => { ++ let mut formatter = formatter.debug_tuple("Await"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Binary(v0) => { ++ let mut formatter = formatter.debug_tuple("Binary"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Block(v0) => { ++ let mut formatter = formatter.debug_tuple("Block"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Box(v0) => { ++ let mut formatter = formatter.debug_tuple("Box"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Break(v0) => { ++ let mut formatter = formatter.debug_tuple("Break"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Call(v0) => { ++ let mut formatter = formatter.debug_tuple("Call"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Cast(v0) => { ++ let mut formatter = formatter.debug_tuple("Cast"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Closure(v0) => { ++ let mut formatter = formatter.debug_tuple("Closure"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Continue(v0) => { ++ let mut formatter = formatter.debug_tuple("Continue"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Field(v0) => { ++ let mut formatter = formatter.debug_tuple("Field"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::ForLoop(v0) => { ++ let mut formatter = formatter.debug_tuple("ForLoop"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Group(v0) => { ++ let mut formatter = formatter.debug_tuple("Group"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::If(v0) => { ++ let mut formatter = formatter.debug_tuple("If"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Index(v0) => { ++ let mut formatter = formatter.debug_tuple("Index"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Let(v0) => { ++ let mut formatter = formatter.debug_tuple("Let"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Lit(v0) => { ++ let mut formatter = formatter.debug_tuple("Lit"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Loop(v0) => { ++ let mut formatter = formatter.debug_tuple("Loop"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Match(v0) => { ++ let mut formatter = formatter.debug_tuple("Match"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::MethodCall(v0) => { ++ let mut formatter = formatter.debug_tuple("MethodCall"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Paren(v0) => { ++ let mut formatter = formatter.debug_tuple("Paren"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Path(v0) => { ++ let mut formatter = formatter.debug_tuple("Path"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Range(v0) => { ++ let mut formatter = formatter.debug_tuple("Range"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Reference(v0) => { ++ let mut formatter = formatter.debug_tuple("Reference"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Repeat(v0) => { ++ let mut formatter = formatter.debug_tuple("Repeat"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Return(v0) => { ++ let mut formatter = formatter.debug_tuple("Return"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Struct(v0) => { ++ let mut formatter = formatter.debug_tuple("Struct"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Try(v0) => { ++ let mut formatter = formatter.debug_tuple("Try"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::TryBlock(v0) => { ++ let mut formatter = formatter.debug_tuple("TryBlock"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Tuple(v0) => { ++ let mut formatter = formatter.debug_tuple("Tuple"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Unary(v0) => { ++ let mut formatter = formatter.debug_tuple("Unary"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Unsafe(v0) => { ++ let mut formatter = formatter.debug_tuple("Unsafe"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::While(v0) => { ++ let mut formatter = formatter.debug_tuple("While"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Yield(v0) => { ++ let mut formatter = formatter.debug_tuple("Yield"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprArray { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprArray"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("elems", &self.elems); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprAssign { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprAssign"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("left", &self.left); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("right", &self.right); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprAssignOp { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprAssignOp"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("left", &self.left); ++ formatter.field("op", &self.op); ++ formatter.field("right", &self.right); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprAsync { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprAsync"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("async_token", &self.async_token); ++ formatter.field("capture", &self.capture); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprAwait { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprAwait"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("base", &self.base); ++ formatter.field("dot_token", &self.dot_token); ++ formatter.field("await_token", &self.await_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprBinary { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprBinary"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("left", &self.left); ++ formatter.field("op", &self.op); ++ formatter.field("right", &self.right); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprBlock { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprBlock"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("label", &self.label); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprBox { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprBox"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("box_token", &self.box_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprBreak { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprBreak"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("break_token", &self.break_token); ++ formatter.field("label", &self.label); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprCall { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprCall"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("func", &self.func); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("args", &self.args); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprCast { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprCast"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("expr", &self.expr); ++ formatter.field("as_token", &self.as_token); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprClosure { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprClosure"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("asyncness", &self.asyncness); ++ formatter.field("movability", &self.movability); ++ formatter.field("capture", &self.capture); ++ formatter.field("or1_token", &self.or1_token); ++ formatter.field("inputs", &self.inputs); ++ formatter.field("or2_token", &self.or2_token); ++ formatter.field("output", &self.output); ++ formatter.field("body", &self.body); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprContinue { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprContinue"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("continue_token", &self.continue_token); ++ formatter.field("label", &self.label); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprField { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprField"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("base", &self.base); ++ formatter.field("dot_token", &self.dot_token); ++ formatter.field("member", &self.member); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprForLoop { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprForLoop"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("label", &self.label); ++ formatter.field("for_token", &self.for_token); ++ formatter.field("pat", &self.pat); ++ formatter.field("in_token", &self.in_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("body", &self.body); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprGroup { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprGroup"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("group_token", &self.group_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprIf { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprIf"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("if_token", &self.if_token); ++ formatter.field("cond", &self.cond); ++ formatter.field("then_branch", &self.then_branch); ++ formatter.field("else_branch", &self.else_branch); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprIndex { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprIndex"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("expr", &self.expr); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("index", &self.index); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprLet { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprLet"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("let_token", &self.let_token); ++ formatter.field("pat", &self.pat); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprLit { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprLit"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("lit", &self.lit); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprLoop { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprLoop"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("label", &self.label); ++ formatter.field("loop_token", &self.loop_token); ++ formatter.field("body", &self.body); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("mac", &self.mac); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprMatch { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprMatch"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("match_token", &self.match_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("arms", &self.arms); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprMethodCall { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprMethodCall"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("receiver", &self.receiver); ++ formatter.field("dot_token", &self.dot_token); ++ formatter.field("method", &self.method); ++ formatter.field("turbofish", &self.turbofish); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("args", &self.args); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprParen { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprParen"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprPath { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprPath"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("qself", &self.qself); ++ formatter.field("path", &self.path); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprRange { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprRange"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("from", &self.from); ++ formatter.field("limits", &self.limits); ++ formatter.field("to", &self.to); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprReference { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprReference"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("and_token", &self.and_token); ++ formatter.field("raw", &self.raw); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprRepeat { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprRepeat"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.field("len", &self.len); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprReturn { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprReturn"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("return_token", &self.return_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprStruct { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprStruct"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("path", &self.path); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("fields", &self.fields); ++ formatter.field("dot2_token", &self.dot2_token); ++ formatter.field("rest", &self.rest); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprTry { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprTry"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("expr", &self.expr); ++ formatter.field("question_token", &self.question_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprTryBlock { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprTryBlock"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("try_token", &self.try_token); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprTuple { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprTuple"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("elems", &self.elems); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("expr", &self.expr); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprUnary { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprUnary"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("op", &self.op); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprUnsafe { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprUnsafe"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("unsafe_token", &self.unsafe_token); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprWhile { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprWhile"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("label", &self.label); ++ formatter.field("while_token", &self.while_token); ++ formatter.field("cond", &self.cond); ++ formatter.field("body", &self.body); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprYield { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprYield"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("yield_token", &self.yield_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Field { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Field"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for FieldPat { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("FieldPat"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("member", &self.member); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("pat", &self.pat); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for FieldValue { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("FieldValue"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("member", &self.member); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Fields { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Fields::Named(v0) => { ++ let mut formatter = formatter.debug_tuple("Named"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Fields::Unnamed(v0) => { ++ let mut formatter = formatter.debug_tuple("Unnamed"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Fields::Unit => formatter.write_str("Unit"), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for FieldsNamed { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("FieldsNamed"); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("named", &self.named); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for FieldsUnnamed { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("FieldsUnnamed"); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("unnamed", &self.unnamed); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for File { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("File"); ++ formatter.field("shebang", &self.shebang); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("items", &self.items); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for FnArg { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ FnArg::Receiver(v0) => { ++ let mut formatter = formatter.debug_tuple("Receiver"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ FnArg::Typed(v0) => { ++ let mut formatter = formatter.debug_tuple("Typed"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ForeignItem { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ ForeignItem::Fn(v0) => { ++ let mut formatter = formatter.debug_tuple("Fn"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ForeignItem::Static(v0) => { ++ let mut formatter = formatter.debug_tuple("Static"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ForeignItem::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ForeignItem::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ForeignItem::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ForeignItemFn { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ForeignItemFn"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("sig", &self.sig); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ForeignItemMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ForeignItemMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("mac", &self.mac); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ForeignItemStatic { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ForeignItemStatic"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("static_token", &self.static_token); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ForeignItemType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ForeignItemType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("type_token", &self.type_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for GenericArgument { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ GenericArgument::Lifetime(v0) => { ++ let mut formatter = formatter.debug_tuple("Lifetime"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericArgument::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericArgument::Binding(v0) => { ++ let mut formatter = formatter.debug_tuple("Binding"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericArgument::Constraint(v0) => { ++ let mut formatter = formatter.debug_tuple("Constraint"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericArgument::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for GenericMethodArgument { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ GenericMethodArgument::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericMethodArgument::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for GenericParam { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ GenericParam::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericParam::Lifetime(v0) => { ++ let mut formatter = formatter.debug_tuple("Lifetime"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericParam::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Generics { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Generics"); ++ formatter.field("lt_token", &self.lt_token); ++ formatter.field("params", &self.params); ++ formatter.field("gt_token", &self.gt_token); ++ formatter.field("where_clause", &self.where_clause); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ImplItem { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ ImplItem::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ImplItem::Method(v0) => { ++ let mut formatter = formatter.debug_tuple("Method"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ImplItem::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ImplItem::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ImplItem::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ImplItemConst { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ImplItemConst"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("defaultness", &self.defaultness); ++ formatter.field("const_token", &self.const_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ImplItemMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ImplItemMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("mac", &self.mac); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ImplItemMethod { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ImplItemMethod"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("defaultness", &self.defaultness); ++ formatter.field("sig", &self.sig); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ImplItemType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ImplItemType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("defaultness", &self.defaultness); ++ formatter.field("type_token", &self.type_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Index { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Index"); ++ formatter.field("index", &self.index); ++ formatter.field("span", &self.span); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Item { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Item::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Enum(v0) => { ++ let mut formatter = formatter.debug_tuple("Enum"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::ExternCrate(v0) => { ++ let mut formatter = formatter.debug_tuple("ExternCrate"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Fn(v0) => { ++ let mut formatter = formatter.debug_tuple("Fn"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::ForeignMod(v0) => { ++ let mut formatter = formatter.debug_tuple("ForeignMod"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Impl(v0) => { ++ let mut formatter = formatter.debug_tuple("Impl"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Macro2(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro2"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Mod(v0) => { ++ let mut formatter = formatter.debug_tuple("Mod"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Static(v0) => { ++ let mut formatter = formatter.debug_tuple("Static"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Struct(v0) => { ++ let mut formatter = formatter.debug_tuple("Struct"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Trait(v0) => { ++ let mut formatter = formatter.debug_tuple("Trait"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::TraitAlias(v0) => { ++ let mut formatter = formatter.debug_tuple("TraitAlias"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Union(v0) => { ++ let mut formatter = formatter.debug_tuple("Union"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Use(v0) => { ++ let mut formatter = formatter.debug_tuple("Use"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemConst { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemConst"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("const_token", &self.const_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemEnum { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemEnum"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("enum_token", &self.enum_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("variants", &self.variants); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemExternCrate { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemExternCrate"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("extern_token", &self.extern_token); ++ formatter.field("crate_token", &self.crate_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("rename", &self.rename); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemFn { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemFn"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("sig", &self.sig); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemForeignMod { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemForeignMod"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("abi", &self.abi); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("items", &self.items); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemImpl { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemImpl"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("defaultness", &self.defaultness); ++ formatter.field("unsafety", &self.unsafety); ++ formatter.field("impl_token", &self.impl_token); ++ formatter.field("generics", &self.generics); ++ formatter.field("trait_", &self.trait_); ++ formatter.field("self_ty", &self.self_ty); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("items", &self.items); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("ident", &self.ident); ++ formatter.field("mac", &self.mac); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemMacro2 { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemMacro2"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("macro_token", &self.macro_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("rules", &self.rules); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemMod { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemMod"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("mod_token", &self.mod_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("content", &self.content); ++ formatter.field("semi", &self.semi); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemStatic { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemStatic"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("static_token", &self.static_token); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemStruct { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemStruct"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("struct_token", &self.struct_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("fields", &self.fields); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemTrait { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemTrait"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("unsafety", &self.unsafety); ++ formatter.field("auto_token", &self.auto_token); ++ formatter.field("trait_token", &self.trait_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("supertraits", &self.supertraits); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("items", &self.items); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemTraitAlias { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemTraitAlias"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("trait_token", &self.trait_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("type_token", &self.type_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemUnion { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemUnion"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("union_token", &self.union_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("fields", &self.fields); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemUse { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemUse"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("use_token", &self.use_token); ++ formatter.field("leading_colon", &self.leading_colon); ++ formatter.field("tree", &self.tree); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Label { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Label"); ++ formatter.field("name", &self.name); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.finish() ++ } ++} ++impl Debug for Lifetime { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Lifetime"); ++ formatter.field("apostrophe", &self.apostrophe); ++ formatter.field("ident", &self.ident); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for LifetimeDef { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("LifetimeDef"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("lifetime", &self.lifetime); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++impl Debug for Lit { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Lit::Str(v0) => { ++ let mut formatter = formatter.debug_tuple("Str"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::ByteStr(v0) => { ++ let mut formatter = formatter.debug_tuple("ByteStr"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Byte(v0) => { ++ let mut formatter = formatter.debug_tuple("Byte"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Char(v0) => { ++ let mut formatter = formatter.debug_tuple("Char"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Int(v0) => { ++ let mut formatter = formatter.debug_tuple("Int"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Float(v0) => { ++ let mut formatter = formatter.debug_tuple("Float"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Bool(v0) => { ++ let mut formatter = formatter.debug_tuple("Bool"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Local { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Local"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("let_token", &self.let_token); ++ formatter.field("pat", &self.pat); ++ formatter.field("init", &self.init); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Macro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Macro"); ++ formatter.field("path", &self.path); ++ formatter.field("bang_token", &self.bang_token); ++ formatter.field("delimiter", &self.delimiter); ++ formatter.field("tokens", &self.tokens); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for MacroDelimiter { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ MacroDelimiter::Paren(v0) => { ++ let mut formatter = formatter.debug_tuple("Paren"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ MacroDelimiter::Brace(v0) => { ++ let mut formatter = formatter.debug_tuple("Brace"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ MacroDelimiter::Bracket(v0) => { ++ let mut formatter = formatter.debug_tuple("Bracket"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Member { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Member::Named(v0) => { ++ let mut formatter = formatter.debug_tuple("Named"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Member::Unnamed(v0) => { ++ let mut formatter = formatter.debug_tuple("Unnamed"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Meta { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Meta::Path(v0) => { ++ let mut formatter = formatter.debug_tuple("Path"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Meta::List(v0) => { ++ let mut formatter = formatter.debug_tuple("List"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Meta::NameValue(v0) => { ++ let mut formatter = formatter.debug_tuple("NameValue"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for MetaList { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("MetaList"); ++ formatter.field("path", &self.path); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("nested", &self.nested); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for MetaNameValue { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("MetaNameValue"); ++ formatter.field("path", &self.path); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("lit", &self.lit); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for MethodTurbofish { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("MethodTurbofish"); ++ formatter.field("colon2_token", &self.colon2_token); ++ formatter.field("lt_token", &self.lt_token); ++ formatter.field("args", &self.args); ++ formatter.field("gt_token", &self.gt_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for NestedMeta { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ NestedMeta::Meta(v0) => { ++ let mut formatter = formatter.debug_tuple("Meta"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ NestedMeta::Lit(v0) => { ++ let mut formatter = formatter.debug_tuple("Lit"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ParenthesizedGenericArguments { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ParenthesizedGenericArguments"); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("inputs", &self.inputs); ++ formatter.field("output", &self.output); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Pat { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Pat::Box(v0) => { ++ let mut formatter = formatter.debug_tuple("Box"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Ident(v0) => { ++ let mut formatter = formatter.debug_tuple("Ident"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Lit(v0) => { ++ let mut formatter = formatter.debug_tuple("Lit"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Or(v0) => { ++ let mut formatter = formatter.debug_tuple("Or"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Path(v0) => { ++ let mut formatter = formatter.debug_tuple("Path"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Range(v0) => { ++ let mut formatter = formatter.debug_tuple("Range"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Reference(v0) => { ++ let mut formatter = formatter.debug_tuple("Reference"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Rest(v0) => { ++ let mut formatter = formatter.debug_tuple("Rest"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Slice(v0) => { ++ let mut formatter = formatter.debug_tuple("Slice"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Struct(v0) => { ++ let mut formatter = formatter.debug_tuple("Struct"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Tuple(v0) => { ++ let mut formatter = formatter.debug_tuple("Tuple"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::TupleStruct(v0) => { ++ let mut formatter = formatter.debug_tuple("TupleStruct"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Wild(v0) => { ++ let mut formatter = formatter.debug_tuple("Wild"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatBox { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatBox"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("box_token", &self.box_token); ++ formatter.field("pat", &self.pat); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatIdent { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatIdent"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("by_ref", &self.by_ref); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("ident", &self.ident); ++ formatter.field("subpat", &self.subpat); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatLit { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatLit"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("mac", &self.mac); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatOr { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatOr"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("leading_vert", &self.leading_vert); ++ formatter.field("cases", &self.cases); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatPath { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatPath"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("qself", &self.qself); ++ formatter.field("path", &self.path); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatRange { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatRange"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("lo", &self.lo); ++ formatter.field("limits", &self.limits); ++ formatter.field("hi", &self.hi); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatReference { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatReference"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("and_token", &self.and_token); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("pat", &self.pat); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatRest { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatRest"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("dot2_token", &self.dot2_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatSlice { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatSlice"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("elems", &self.elems); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatStruct { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatStruct"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("path", &self.path); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("fields", &self.fields); ++ formatter.field("dot2_token", &self.dot2_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatTuple { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatTuple"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("elems", &self.elems); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatTupleStruct { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatTupleStruct"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("path", &self.path); ++ formatter.field("pat", &self.pat); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("pat", &self.pat); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatWild { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatWild"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("underscore_token", &self.underscore_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Path { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Path"); ++ formatter.field("leading_colon", &self.leading_colon); ++ formatter.field("segments", &self.segments); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for PathArguments { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ PathArguments::None => formatter.write_str("None"), ++ PathArguments::AngleBracketed(v0) => { ++ let mut formatter = formatter.debug_tuple("AngleBracketed"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ PathArguments::Parenthesized(v0) => { ++ let mut formatter = formatter.debug_tuple("Parenthesized"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for PathSegment { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PathSegment"); ++ formatter.field("ident", &self.ident); ++ formatter.field("arguments", &self.arguments); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for PredicateEq { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PredicateEq"); ++ formatter.field("lhs_ty", &self.lhs_ty); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("rhs_ty", &self.rhs_ty); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for PredicateLifetime { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PredicateLifetime"); ++ formatter.field("lifetime", &self.lifetime); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for PredicateType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PredicateType"); ++ formatter.field("lifetimes", &self.lifetimes); ++ formatter.field("bounded_ty", &self.bounded_ty); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for QSelf { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("QSelf"); ++ formatter.field("lt_token", &self.lt_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("position", &self.position); ++ formatter.field("as_token", &self.as_token); ++ formatter.field("gt_token", &self.gt_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for RangeLimits { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ RangeLimits::HalfOpen(v0) => { ++ let mut formatter = formatter.debug_tuple("HalfOpen"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ RangeLimits::Closed(v0) => { ++ let mut formatter = formatter.debug_tuple("Closed"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Receiver { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Receiver"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("reference", &self.reference); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("self_token", &self.self_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ReturnType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ ReturnType::Default => formatter.write_str("Default"), ++ ReturnType::Type(v0, v1) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.field(v1); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Signature { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Signature"); ++ formatter.field("constness", &self.constness); ++ formatter.field("asyncness", &self.asyncness); ++ formatter.field("unsafety", &self.unsafety); ++ formatter.field("abi", &self.abi); ++ formatter.field("fn_token", &self.fn_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("inputs", &self.inputs); ++ formatter.field("variadic", &self.variadic); ++ formatter.field("output", &self.output); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Stmt { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Stmt::Local(v0) => { ++ let mut formatter = formatter.debug_tuple("Local"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Stmt::Item(v0) => { ++ let mut formatter = formatter.debug_tuple("Item"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Stmt::Expr(v0) => { ++ let mut formatter = formatter.debug_tuple("Expr"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Stmt::Semi(v0, v1) => { ++ let mut formatter = formatter.debug_tuple("Semi"); ++ formatter.field(v0); ++ formatter.field(v1); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TraitBound { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TraitBound"); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("modifier", &self.modifier); ++ formatter.field("lifetimes", &self.lifetimes); ++ formatter.field("path", &self.path); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TraitBoundModifier { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ TraitBoundModifier::None => formatter.write_str("None"), ++ TraitBoundModifier::Maybe(v0) => { ++ let mut formatter = formatter.debug_tuple("Maybe"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for TraitItem { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ TraitItem::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ TraitItem::Method(v0) => { ++ let mut formatter = formatter.debug_tuple("Method"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ TraitItem::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ TraitItem::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ TraitItem::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for TraitItemConst { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TraitItemConst"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("const_token", &self.const_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("default", &self.default); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for TraitItemMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TraitItemMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("mac", &self.mac); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for TraitItemMethod { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TraitItemMethod"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("sig", &self.sig); ++ formatter.field("default", &self.default); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for TraitItemType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TraitItemType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("type_token", &self.type_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.field("default", &self.default); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Type { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Type::Array(v0) => { ++ let mut formatter = formatter.debug_tuple("Array"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::BareFn(v0) => { ++ let mut formatter = formatter.debug_tuple("BareFn"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Group(v0) => { ++ let mut formatter = formatter.debug_tuple("Group"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::ImplTrait(v0) => { ++ let mut formatter = formatter.debug_tuple("ImplTrait"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Infer(v0) => { ++ let mut formatter = formatter.debug_tuple("Infer"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Never(v0) => { ++ let mut formatter = formatter.debug_tuple("Never"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Paren(v0) => { ++ let mut formatter = formatter.debug_tuple("Paren"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Path(v0) => { ++ let mut formatter = formatter.debug_tuple("Path"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Ptr(v0) => { ++ let mut formatter = formatter.debug_tuple("Ptr"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Reference(v0) => { ++ let mut formatter = formatter.debug_tuple("Reference"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Slice(v0) => { ++ let mut formatter = formatter.debug_tuple("Slice"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::TraitObject(v0) => { ++ let mut formatter = formatter.debug_tuple("TraitObject"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Tuple(v0) => { ++ let mut formatter = formatter.debug_tuple("Tuple"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeArray { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeArray"); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("elem", &self.elem); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.field("len", &self.len); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeBareFn { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeBareFn"); ++ formatter.field("lifetimes", &self.lifetimes); ++ formatter.field("unsafety", &self.unsafety); ++ formatter.field("abi", &self.abi); ++ formatter.field("fn_token", &self.fn_token); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("inputs", &self.inputs); ++ formatter.field("variadic", &self.variadic); ++ formatter.field("output", &self.output); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeGroup { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeGroup"); ++ formatter.field("group_token", &self.group_token); ++ formatter.field("elem", &self.elem); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeImplTrait { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeImplTrait"); ++ formatter.field("impl_token", &self.impl_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeInfer { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeInfer"); ++ formatter.field("underscore_token", &self.underscore_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeMacro"); ++ formatter.field("mac", &self.mac); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeNever { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeNever"); ++ formatter.field("bang_token", &self.bang_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeParam { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeParam"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("default", &self.default); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeParamBound { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ TypeParamBound::Trait(v0) => { ++ let mut formatter = formatter.debug_tuple("Trait"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ TypeParamBound::Lifetime(v0) => { ++ let mut formatter = formatter.debug_tuple("Lifetime"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeParen { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeParen"); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("elem", &self.elem); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypePath { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypePath"); ++ formatter.field("qself", &self.qself); ++ formatter.field("path", &self.path); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypePtr { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypePtr"); ++ formatter.field("star_token", &self.star_token); ++ formatter.field("const_token", &self.const_token); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("elem", &self.elem); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeReference { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeReference"); ++ formatter.field("and_token", &self.and_token); ++ formatter.field("lifetime", &self.lifetime); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("elem", &self.elem); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeSlice { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeSlice"); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("elem", &self.elem); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeTraitObject { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeTraitObject"); ++ formatter.field("dyn_token", &self.dyn_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeTuple { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeTuple"); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("elems", &self.elems); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for UnOp { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ UnOp::Deref(v0) => { ++ let mut formatter = formatter.debug_tuple("Deref"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UnOp::Not(v0) => { ++ let mut formatter = formatter.debug_tuple("Not"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UnOp::Neg(v0) => { ++ let mut formatter = formatter.debug_tuple("Neg"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UseGlob { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("UseGlob"); ++ formatter.field("star_token", &self.star_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UseGroup { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("UseGroup"); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("items", &self.items); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UseName { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("UseName"); ++ formatter.field("ident", &self.ident); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UsePath { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("UsePath"); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon2_token", &self.colon2_token); ++ formatter.field("tree", &self.tree); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UseRename { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("UseRename"); ++ formatter.field("ident", &self.ident); ++ formatter.field("as_token", &self.as_token); ++ formatter.field("rename", &self.rename); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UseTree { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ UseTree::Path(v0) => { ++ let mut formatter = formatter.debug_tuple("Path"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UseTree::Name(v0) => { ++ let mut formatter = formatter.debug_tuple("Name"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UseTree::Rename(v0) => { ++ let mut formatter = formatter.debug_tuple("Rename"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UseTree::Glob(v0) => { ++ let mut formatter = formatter.debug_tuple("Glob"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UseTree::Group(v0) => { ++ let mut formatter = formatter.debug_tuple("Group"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Variadic { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Variadic"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("dots", &self.dots); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Variant { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Variant"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("ident", &self.ident); ++ formatter.field("fields", &self.fields); ++ formatter.field("discriminant", &self.discriminant); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for VisCrate { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("VisCrate"); ++ formatter.field("crate_token", &self.crate_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for VisPublic { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("VisPublic"); ++ formatter.field("pub_token", &self.pub_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for VisRestricted { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("VisRestricted"); ++ formatter.field("pub_token", &self.pub_token); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("in_token", &self.in_token); ++ formatter.field("path", &self.path); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Visibility { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Visibility::Public(v0) => { ++ let mut formatter = formatter.debug_tuple("Public"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Visibility::Crate(v0) => { ++ let mut formatter = formatter.debug_tuple("Crate"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Visibility::Restricted(v0) => { ++ let mut formatter = formatter.debug_tuple("Restricted"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Visibility::Inherited => formatter.write_str("Inherited"), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for WhereClause { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("WhereClause"); ++ formatter.field("where_token", &self.where_token); ++ formatter.field("predicates", &self.predicates); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for WherePredicate { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ WherePredicate::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ WherePredicate::Lifetime(v0) => { ++ let mut formatter = formatter.debug_tuple("Lifetime"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ WherePredicate::Eq(v0) => { ++ let mut formatter = formatter.debug_tuple("Eq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} +diff --git a/third_party/rust/syn/src/gen/eq.rs b/third_party/rust/syn/src/gen/eq.rs +new file mode 100644 +index 0000000000..15b2bcbbde +--- /dev/null ++++ b/third_party/rust/syn/src/gen/eq.rs +@@ -0,0 +1,1930 @@ ++// This file is @generated by syn-internal-codegen. ++// It is not intended for manual editing. ++ ++#[cfg(any(feature = "derive", feature = "full"))] ++use crate::tt::TokenStreamHelper; ++use crate::*; ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Abi {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Abi { ++ fn eq(&self, other: &Self) -> bool { ++ self.name == other.name ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for AngleBracketedGenericArguments {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for AngleBracketedGenericArguments { ++ fn eq(&self, other: &Self) -> bool { ++ self.colon2_token == other.colon2_token && self.args == other.args ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Arm {} ++#[cfg(feature = "full")] ++impl PartialEq for Arm { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.pat == other.pat ++ && self.guard == other.guard ++ && self.body == other.body ++ && self.comma == other.comma ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for AttrStyle {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for AttrStyle { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (AttrStyle::Outer, AttrStyle::Outer) => true, ++ (AttrStyle::Inner(_), AttrStyle::Inner(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Attribute {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Attribute { ++ fn eq(&self, other: &Self) -> bool { ++ self.style == other.style ++ && self.path == other.path ++ && TokenStreamHelper(&self.tokens) == TokenStreamHelper(&other.tokens) ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for BareFnArg {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for BareFnArg { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.name == other.name && self.ty == other.ty ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for BinOp {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for BinOp { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (BinOp::Add(_), BinOp::Add(_)) => true, ++ (BinOp::Sub(_), BinOp::Sub(_)) => true, ++ (BinOp::Mul(_), BinOp::Mul(_)) => true, ++ (BinOp::Div(_), BinOp::Div(_)) => true, ++ (BinOp::Rem(_), BinOp::Rem(_)) => true, ++ (BinOp::And(_), BinOp::And(_)) => true, ++ (BinOp::Or(_), BinOp::Or(_)) => true, ++ (BinOp::BitXor(_), BinOp::BitXor(_)) => true, ++ (BinOp::BitAnd(_), BinOp::BitAnd(_)) => true, ++ (BinOp::BitOr(_), BinOp::BitOr(_)) => true, ++ (BinOp::Shl(_), BinOp::Shl(_)) => true, ++ (BinOp::Shr(_), BinOp::Shr(_)) => true, ++ (BinOp::Eq(_), BinOp::Eq(_)) => true, ++ (BinOp::Lt(_), BinOp::Lt(_)) => true, ++ (BinOp::Le(_), BinOp::Le(_)) => true, ++ (BinOp::Ne(_), BinOp::Ne(_)) => true, ++ (BinOp::Ge(_), BinOp::Ge(_)) => true, ++ (BinOp::Gt(_), BinOp::Gt(_)) => true, ++ (BinOp::AddEq(_), BinOp::AddEq(_)) => true, ++ (BinOp::SubEq(_), BinOp::SubEq(_)) => true, ++ (BinOp::MulEq(_), BinOp::MulEq(_)) => true, ++ (BinOp::DivEq(_), BinOp::DivEq(_)) => true, ++ (BinOp::RemEq(_), BinOp::RemEq(_)) => true, ++ (BinOp::BitXorEq(_), BinOp::BitXorEq(_)) => true, ++ (BinOp::BitAndEq(_), BinOp::BitAndEq(_)) => true, ++ (BinOp::BitOrEq(_), BinOp::BitOrEq(_)) => true, ++ (BinOp::ShlEq(_), BinOp::ShlEq(_)) => true, ++ (BinOp::ShrEq(_), BinOp::ShrEq(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Binding {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Binding { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Block {} ++#[cfg(feature = "full")] ++impl PartialEq for Block { ++ fn eq(&self, other: &Self) -> bool { ++ self.stmts == other.stmts ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for BoundLifetimes {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for BoundLifetimes { ++ fn eq(&self, other: &Self) -> bool { ++ self.lifetimes == other.lifetimes ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ConstParam {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ConstParam { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.ty == other.ty ++ && self.eq_token == other.eq_token ++ && self.default == other.default ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Constraint {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Constraint { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident && self.bounds == other.bounds ++ } ++} ++#[cfg(feature = "derive")] ++impl Eq for Data {} ++#[cfg(feature = "derive")] ++impl PartialEq for Data { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Data::Struct(self0), Data::Struct(other0)) => self0 == other0, ++ (Data::Enum(self0), Data::Enum(other0)) => self0 == other0, ++ (Data::Union(self0), Data::Union(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Eq for DataEnum {} ++#[cfg(feature = "derive")] ++impl PartialEq for DataEnum { ++ fn eq(&self, other: &Self) -> bool { ++ self.variants == other.variants ++ } ++} ++#[cfg(feature = "derive")] ++impl Eq for DataStruct {} ++#[cfg(feature = "derive")] ++impl PartialEq for DataStruct { ++ fn eq(&self, other: &Self) -> bool { ++ self.fields == other.fields && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "derive")] ++impl Eq for DataUnion {} ++#[cfg(feature = "derive")] ++impl PartialEq for DataUnion { ++ fn eq(&self, other: &Self) -> bool { ++ self.fields == other.fields ++ } ++} ++#[cfg(feature = "derive")] ++impl Eq for DeriveInput {} ++#[cfg(feature = "derive")] ++impl PartialEq for DeriveInput { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.data == other.data ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Expr {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Expr { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ #[cfg(feature = "full")] ++ (Expr::Array(self0), Expr::Array(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Assign(self0), Expr::Assign(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::AssignOp(self0), Expr::AssignOp(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Async(self0), Expr::Async(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Await(self0), Expr::Await(other0)) => self0 == other0, ++ (Expr::Binary(self0), Expr::Binary(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Block(self0), Expr::Block(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Box(self0), Expr::Box(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Break(self0), Expr::Break(other0)) => self0 == other0, ++ (Expr::Call(self0), Expr::Call(other0)) => self0 == other0, ++ (Expr::Cast(self0), Expr::Cast(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Closure(self0), Expr::Closure(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Continue(self0), Expr::Continue(other0)) => self0 == other0, ++ (Expr::Field(self0), Expr::Field(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::ForLoop(self0), Expr::ForLoop(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Group(self0), Expr::Group(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::If(self0), Expr::If(other0)) => self0 == other0, ++ (Expr::Index(self0), Expr::Index(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Let(self0), Expr::Let(other0)) => self0 == other0, ++ (Expr::Lit(self0), Expr::Lit(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Loop(self0), Expr::Loop(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Macro(self0), Expr::Macro(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Match(self0), Expr::Match(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::MethodCall(self0), Expr::MethodCall(other0)) => self0 == other0, ++ (Expr::Paren(self0), Expr::Paren(other0)) => self0 == other0, ++ (Expr::Path(self0), Expr::Path(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Range(self0), Expr::Range(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Reference(self0), Expr::Reference(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Repeat(self0), Expr::Repeat(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Return(self0), Expr::Return(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Struct(self0), Expr::Struct(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Try(self0), Expr::Try(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::TryBlock(self0), Expr::TryBlock(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Tuple(self0), Expr::Tuple(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Type(self0), Expr::Type(other0)) => self0 == other0, ++ (Expr::Unary(self0), Expr::Unary(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Unsafe(self0), Expr::Unsafe(other0)) => self0 == other0, ++ (Expr::Verbatim(self0), Expr::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ #[cfg(feature = "full")] ++ (Expr::While(self0), Expr::While(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Yield(self0), Expr::Yield(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprArray {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprArray { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.elems == other.elems ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprAssign {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprAssign { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.left == other.left && self.right == other.right ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprAssignOp {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprAssignOp { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.left == other.left ++ && self.op == other.op ++ && self.right == other.right ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprAsync {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprAsync { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.capture == other.capture && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprAwait {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprAwait { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.base == other.base ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprBinary {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprBinary { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.left == other.left ++ && self.op == other.op ++ && self.right == other.right ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprBlock {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprBlock { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.label == other.label && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprBox {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprBox { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprBreak {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprBreak { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.label == other.label && self.expr == other.expr ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprCall {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprCall { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.func == other.func && self.args == other.args ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprCast {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprCast { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprClosure {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprClosure { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.asyncness == other.asyncness ++ && self.movability == other.movability ++ && self.capture == other.capture ++ && self.inputs == other.inputs ++ && self.output == other.output ++ && self.body == other.body ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprContinue {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprContinue { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.label == other.label ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprField {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprField { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.base == other.base && self.member == other.member ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprForLoop {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprForLoop { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.label == other.label ++ && self.pat == other.pat ++ && self.expr == other.expr ++ && self.body == other.body ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprGroup {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprGroup { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprIf {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprIf { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.cond == other.cond ++ && self.then_branch == other.then_branch ++ && self.else_branch == other.else_branch ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprIndex {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprIndex { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr && self.index == other.index ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprLet {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprLet { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.pat == other.pat && self.expr == other.expr ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprLit {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprLit { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.lit == other.lit ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprLoop {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprLoop { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.label == other.label && self.body == other.body ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mac == other.mac ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprMatch {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprMatch { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr && self.arms == other.arms ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprMethodCall {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprMethodCall { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.receiver == other.receiver ++ && self.method == other.method ++ && self.turbofish == other.turbofish ++ && self.args == other.args ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprParen {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprParen { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprPath {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprPath { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.qself == other.qself && self.path == other.path ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprRange {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprRange { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.from == other.from ++ && self.limits == other.limits ++ && self.to == other.to ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprReference {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprReference { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mutability == other.mutability && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprRepeat {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprRepeat { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr && self.len == other.len ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprReturn {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprReturn { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprStruct {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprStruct { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.path == other.path ++ && self.fields == other.fields ++ && self.dot2_token == other.dot2_token ++ && self.rest == other.rest ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprTry {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprTry { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprTryBlock {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprTryBlock { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprTuple {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprTuple { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.elems == other.elems ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprType {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr && self.ty == other.ty ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprUnary {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprUnary { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.op == other.op && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprUnsafe {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprUnsafe { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprWhile {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprWhile { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.label == other.label ++ && self.cond == other.cond ++ && self.body == other.body ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprYield {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprYield { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Field {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Field { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.colon_token == other.colon_token ++ && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for FieldPat {} ++#[cfg(feature = "full")] ++impl PartialEq for FieldPat { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.member == other.member ++ && self.colon_token == other.colon_token ++ && self.pat == other.pat ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for FieldValue {} ++#[cfg(feature = "full")] ++impl PartialEq for FieldValue { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.member == other.member ++ && self.colon_token == other.colon_token ++ && self.expr == other.expr ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Fields {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Fields { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Fields::Named(self0), Fields::Named(other0)) => self0 == other0, ++ (Fields::Unnamed(self0), Fields::Unnamed(other0)) => self0 == other0, ++ (Fields::Unit, Fields::Unit) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for FieldsNamed {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for FieldsNamed { ++ fn eq(&self, other: &Self) -> bool { ++ self.named == other.named ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for FieldsUnnamed {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for FieldsUnnamed { ++ fn eq(&self, other: &Self) -> bool { ++ self.unnamed == other.unnamed ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for File {} ++#[cfg(feature = "full")] ++impl PartialEq for File { ++ fn eq(&self, other: &Self) -> bool { ++ self.shebang == other.shebang && self.attrs == other.attrs && self.items == other.items ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for FnArg {} ++#[cfg(feature = "full")] ++impl PartialEq for FnArg { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (FnArg::Receiver(self0), FnArg::Receiver(other0)) => self0 == other0, ++ (FnArg::Typed(self0), FnArg::Typed(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ForeignItem {} ++#[cfg(feature = "full")] ++impl PartialEq for ForeignItem { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (ForeignItem::Fn(self0), ForeignItem::Fn(other0)) => self0 == other0, ++ (ForeignItem::Static(self0), ForeignItem::Static(other0)) => self0 == other0, ++ (ForeignItem::Type(self0), ForeignItem::Type(other0)) => self0 == other0, ++ (ForeignItem::Macro(self0), ForeignItem::Macro(other0)) => self0 == other0, ++ (ForeignItem::Verbatim(self0), ForeignItem::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ForeignItemFn {} ++#[cfg(feature = "full")] ++impl PartialEq for ForeignItemFn { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.vis == other.vis && self.sig == other.sig ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ForeignItemMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for ForeignItemMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mac == other.mac && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ForeignItemStatic {} ++#[cfg(feature = "full")] ++impl PartialEq for ForeignItemStatic { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.mutability == other.mutability ++ && self.ident == other.ident ++ && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ForeignItemType {} ++#[cfg(feature = "full")] ++impl PartialEq for ForeignItemType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.vis == other.vis && self.ident == other.ident ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for GenericArgument {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for GenericArgument { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (GenericArgument::Lifetime(self0), GenericArgument::Lifetime(other0)) => { ++ self0 == other0 ++ } ++ (GenericArgument::Type(self0), GenericArgument::Type(other0)) => self0 == other0, ++ (GenericArgument::Binding(self0), GenericArgument::Binding(other0)) => self0 == other0, ++ (GenericArgument::Constraint(self0), GenericArgument::Constraint(other0)) => { ++ self0 == other0 ++ } ++ (GenericArgument::Const(self0), GenericArgument::Const(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for GenericMethodArgument {} ++#[cfg(feature = "full")] ++impl PartialEq for GenericMethodArgument { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (GenericMethodArgument::Type(self0), GenericMethodArgument::Type(other0)) => { ++ self0 == other0 ++ } ++ (GenericMethodArgument::Const(self0), GenericMethodArgument::Const(other0)) => { ++ self0 == other0 ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for GenericParam {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for GenericParam { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (GenericParam::Type(self0), GenericParam::Type(other0)) => self0 == other0, ++ (GenericParam::Lifetime(self0), GenericParam::Lifetime(other0)) => self0 == other0, ++ (GenericParam::Const(self0), GenericParam::Const(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Generics {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Generics { ++ fn eq(&self, other: &Self) -> bool { ++ self.lt_token == other.lt_token ++ && self.params == other.params ++ && self.gt_token == other.gt_token ++ && self.where_clause == other.where_clause ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ImplItem {} ++#[cfg(feature = "full")] ++impl PartialEq for ImplItem { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (ImplItem::Const(self0), ImplItem::Const(other0)) => self0 == other0, ++ (ImplItem::Method(self0), ImplItem::Method(other0)) => self0 == other0, ++ (ImplItem::Type(self0), ImplItem::Type(other0)) => self0 == other0, ++ (ImplItem::Macro(self0), ImplItem::Macro(other0)) => self0 == other0, ++ (ImplItem::Verbatim(self0), ImplItem::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ImplItemConst {} ++#[cfg(feature = "full")] ++impl PartialEq for ImplItemConst { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.defaultness == other.defaultness ++ && self.ident == other.ident ++ && self.ty == other.ty ++ && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ImplItemMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for ImplItemMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mac == other.mac && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ImplItemMethod {} ++#[cfg(feature = "full")] ++impl PartialEq for ImplItemMethod { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.defaultness == other.defaultness ++ && self.sig == other.sig ++ && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ImplItemType {} ++#[cfg(feature = "full")] ++impl PartialEq for ImplItemType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.defaultness == other.defaultness ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Item {} ++#[cfg(feature = "full")] ++impl PartialEq for Item { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Item::Const(self0), Item::Const(other0)) => self0 == other0, ++ (Item::Enum(self0), Item::Enum(other0)) => self0 == other0, ++ (Item::ExternCrate(self0), Item::ExternCrate(other0)) => self0 == other0, ++ (Item::Fn(self0), Item::Fn(other0)) => self0 == other0, ++ (Item::ForeignMod(self0), Item::ForeignMod(other0)) => self0 == other0, ++ (Item::Impl(self0), Item::Impl(other0)) => self0 == other0, ++ (Item::Macro(self0), Item::Macro(other0)) => self0 == other0, ++ (Item::Macro2(self0), Item::Macro2(other0)) => self0 == other0, ++ (Item::Mod(self0), Item::Mod(other0)) => self0 == other0, ++ (Item::Static(self0), Item::Static(other0)) => self0 == other0, ++ (Item::Struct(self0), Item::Struct(other0)) => self0 == other0, ++ (Item::Trait(self0), Item::Trait(other0)) => self0 == other0, ++ (Item::TraitAlias(self0), Item::TraitAlias(other0)) => self0 == other0, ++ (Item::Type(self0), Item::Type(other0)) => self0 == other0, ++ (Item::Union(self0), Item::Union(other0)) => self0 == other0, ++ (Item::Use(self0), Item::Use(other0)) => self0 == other0, ++ (Item::Verbatim(self0), Item::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemConst {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemConst { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.ty == other.ty ++ && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemEnum {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemEnum { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.variants == other.variants ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemExternCrate {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemExternCrate { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.rename == other.rename ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemFn {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemFn { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.sig == other.sig ++ && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemForeignMod {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemForeignMod { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.abi == other.abi && self.items == other.items ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemImpl {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemImpl { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.defaultness == other.defaultness ++ && self.unsafety == other.unsafety ++ && self.generics == other.generics ++ && self.trait_ == other.trait_ ++ && self.self_ty == other.self_ty ++ && self.items == other.items ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.mac == other.mac ++ && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemMacro2 {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemMacro2 { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && TokenStreamHelper(&self.rules) == TokenStreamHelper(&other.rules) ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemMod {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemMod { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.content == other.content ++ && self.semi == other.semi ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemStatic {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemStatic { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.mutability == other.mutability ++ && self.ident == other.ident ++ && self.ty == other.ty ++ && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemStruct {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemStruct { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.fields == other.fields ++ && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemTrait {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemTrait { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.unsafety == other.unsafety ++ && self.auto_token == other.auto_token ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.colon_token == other.colon_token ++ && self.supertraits == other.supertraits ++ && self.items == other.items ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemTraitAlias {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemTraitAlias { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.bounds == other.bounds ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemType {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemUnion {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemUnion { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.fields == other.fields ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemUse {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemUse { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.leading_colon == other.leading_colon ++ && self.tree == other.tree ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Label {} ++#[cfg(feature = "full")] ++impl PartialEq for Label { ++ fn eq(&self, other: &Self) -> bool { ++ self.name == other.name ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for LifetimeDef {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for LifetimeDef { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.lifetime == other.lifetime ++ && self.colon_token == other.colon_token ++ && self.bounds == other.bounds ++ } ++} ++impl Eq for Lit {} ++impl PartialEq for Lit { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Lit::Str(self0), Lit::Str(other0)) => self0 == other0, ++ (Lit::ByteStr(self0), Lit::ByteStr(other0)) => self0 == other0, ++ (Lit::Byte(self0), Lit::Byte(other0)) => self0 == other0, ++ (Lit::Char(self0), Lit::Char(other0)) => self0 == other0, ++ (Lit::Int(self0), Lit::Int(other0)) => self0 == other0, ++ (Lit::Float(self0), Lit::Float(other0)) => self0 == other0, ++ (Lit::Bool(self0), Lit::Bool(other0)) => self0 == other0, ++ (Lit::Verbatim(self0), Lit::Verbatim(other0)) => { ++ self0.to_string() == other0.to_string() ++ } ++ _ => false, ++ } ++ } ++} ++impl Eq for LitBool {} ++impl PartialEq for LitBool { ++ fn eq(&self, other: &Self) -> bool { ++ self.value == other.value ++ } ++} ++impl Eq for LitByte {} ++impl Eq for LitByteStr {} ++impl Eq for LitChar {} ++impl Eq for LitFloat {} ++impl Eq for LitInt {} ++impl Eq for LitStr {} ++#[cfg(feature = "full")] ++impl Eq for Local {} ++#[cfg(feature = "full")] ++impl PartialEq for Local { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.pat == other.pat && self.init == other.init ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Macro {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Macro { ++ fn eq(&self, other: &Self) -> bool { ++ self.path == other.path ++ && self.delimiter == other.delimiter ++ && TokenStreamHelper(&self.tokens) == TokenStreamHelper(&other.tokens) ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for MacroDelimiter {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for MacroDelimiter { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (MacroDelimiter::Paren(_), MacroDelimiter::Paren(_)) => true, ++ (MacroDelimiter::Brace(_), MacroDelimiter::Brace(_)) => true, ++ (MacroDelimiter::Bracket(_), MacroDelimiter::Bracket(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Meta {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Meta { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Meta::Path(self0), Meta::Path(other0)) => self0 == other0, ++ (Meta::List(self0), Meta::List(other0)) => self0 == other0, ++ (Meta::NameValue(self0), Meta::NameValue(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for MetaList {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for MetaList { ++ fn eq(&self, other: &Self) -> bool { ++ self.path == other.path && self.nested == other.nested ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for MetaNameValue {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for MetaNameValue { ++ fn eq(&self, other: &Self) -> bool { ++ self.path == other.path && self.lit == other.lit ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for MethodTurbofish {} ++#[cfg(feature = "full")] ++impl PartialEq for MethodTurbofish { ++ fn eq(&self, other: &Self) -> bool { ++ self.args == other.args ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for NestedMeta {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for NestedMeta { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (NestedMeta::Meta(self0), NestedMeta::Meta(other0)) => self0 == other0, ++ (NestedMeta::Lit(self0), NestedMeta::Lit(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ParenthesizedGenericArguments {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ParenthesizedGenericArguments { ++ fn eq(&self, other: &Self) -> bool { ++ self.inputs == other.inputs && self.output == other.output ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Pat {} ++#[cfg(feature = "full")] ++impl PartialEq for Pat { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Pat::Box(self0), Pat::Box(other0)) => self0 == other0, ++ (Pat::Ident(self0), Pat::Ident(other0)) => self0 == other0, ++ (Pat::Lit(self0), Pat::Lit(other0)) => self0 == other0, ++ (Pat::Macro(self0), Pat::Macro(other0)) => self0 == other0, ++ (Pat::Or(self0), Pat::Or(other0)) => self0 == other0, ++ (Pat::Path(self0), Pat::Path(other0)) => self0 == other0, ++ (Pat::Range(self0), Pat::Range(other0)) => self0 == other0, ++ (Pat::Reference(self0), Pat::Reference(other0)) => self0 == other0, ++ (Pat::Rest(self0), Pat::Rest(other0)) => self0 == other0, ++ (Pat::Slice(self0), Pat::Slice(other0)) => self0 == other0, ++ (Pat::Struct(self0), Pat::Struct(other0)) => self0 == other0, ++ (Pat::Tuple(self0), Pat::Tuple(other0)) => self0 == other0, ++ (Pat::TupleStruct(self0), Pat::TupleStruct(other0)) => self0 == other0, ++ (Pat::Type(self0), Pat::Type(other0)) => self0 == other0, ++ (Pat::Verbatim(self0), Pat::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ (Pat::Wild(self0), Pat::Wild(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatBox {} ++#[cfg(feature = "full")] ++impl PartialEq for PatBox { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.pat == other.pat ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatIdent {} ++#[cfg(feature = "full")] ++impl PartialEq for PatIdent { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.by_ref == other.by_ref ++ && self.mutability == other.mutability ++ && self.ident == other.ident ++ && self.subpat == other.subpat ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatLit {} ++#[cfg(feature = "full")] ++impl PartialEq for PatLit { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for PatMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mac == other.mac ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatOr {} ++#[cfg(feature = "full")] ++impl PartialEq for PatOr { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.leading_vert == other.leading_vert ++ && self.cases == other.cases ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatPath {} ++#[cfg(feature = "full")] ++impl PartialEq for PatPath { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.qself == other.qself && self.path == other.path ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatRange {} ++#[cfg(feature = "full")] ++impl PartialEq for PatRange { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.lo == other.lo ++ && self.limits == other.limits ++ && self.hi == other.hi ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatReference {} ++#[cfg(feature = "full")] ++impl PartialEq for PatReference { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mutability == other.mutability && self.pat == other.pat ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatRest {} ++#[cfg(feature = "full")] ++impl PartialEq for PatRest { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatSlice {} ++#[cfg(feature = "full")] ++impl PartialEq for PatSlice { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.elems == other.elems ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatStruct {} ++#[cfg(feature = "full")] ++impl PartialEq for PatStruct { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.path == other.path ++ && self.fields == other.fields ++ && self.dot2_token == other.dot2_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatTuple {} ++#[cfg(feature = "full")] ++impl PartialEq for PatTuple { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.elems == other.elems ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatTupleStruct {} ++#[cfg(feature = "full")] ++impl PartialEq for PatTupleStruct { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.path == other.path && self.pat == other.pat ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatType {} ++#[cfg(feature = "full")] ++impl PartialEq for PatType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.pat == other.pat && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatWild {} ++#[cfg(feature = "full")] ++impl PartialEq for PatWild { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Path {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Path { ++ fn eq(&self, other: &Self) -> bool { ++ self.leading_colon == other.leading_colon && self.segments == other.segments ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for PathArguments {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for PathArguments { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (PathArguments::None, PathArguments::None) => true, ++ (PathArguments::AngleBracketed(self0), PathArguments::AngleBracketed(other0)) => { ++ self0 == other0 ++ } ++ (PathArguments::Parenthesized(self0), PathArguments::Parenthesized(other0)) => { ++ self0 == other0 ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for PathSegment {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for PathSegment { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident && self.arguments == other.arguments ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for PredicateEq {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for PredicateEq { ++ fn eq(&self, other: &Self) -> bool { ++ self.lhs_ty == other.lhs_ty && self.rhs_ty == other.rhs_ty ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for PredicateLifetime {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for PredicateLifetime { ++ fn eq(&self, other: &Self) -> bool { ++ self.lifetime == other.lifetime && self.bounds == other.bounds ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for PredicateType {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for PredicateType { ++ fn eq(&self, other: &Self) -> bool { ++ self.lifetimes == other.lifetimes ++ && self.bounded_ty == other.bounded_ty ++ && self.bounds == other.bounds ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for QSelf {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for QSelf { ++ fn eq(&self, other: &Self) -> bool { ++ self.ty == other.ty && self.position == other.position && self.as_token == other.as_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for RangeLimits {} ++#[cfg(feature = "full")] ++impl PartialEq for RangeLimits { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (RangeLimits::HalfOpen(_), RangeLimits::HalfOpen(_)) => true, ++ (RangeLimits::Closed(_), RangeLimits::Closed(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Receiver {} ++#[cfg(feature = "full")] ++impl PartialEq for Receiver { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.reference == other.reference ++ && self.mutability == other.mutability ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ReturnType {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ReturnType { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (ReturnType::Default, ReturnType::Default) => true, ++ (ReturnType::Type(_, self1), ReturnType::Type(_, other1)) => self1 == other1, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Signature {} ++#[cfg(feature = "full")] ++impl PartialEq for Signature { ++ fn eq(&self, other: &Self) -> bool { ++ self.constness == other.constness ++ && self.asyncness == other.asyncness ++ && self.unsafety == other.unsafety ++ && self.abi == other.abi ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.inputs == other.inputs ++ && self.variadic == other.variadic ++ && self.output == other.output ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Stmt {} ++#[cfg(feature = "full")] ++impl PartialEq for Stmt { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Stmt::Local(self0), Stmt::Local(other0)) => self0 == other0, ++ (Stmt::Item(self0), Stmt::Item(other0)) => self0 == other0, ++ (Stmt::Expr(self0), Stmt::Expr(other0)) => self0 == other0, ++ (Stmt::Semi(self0, _), Stmt::Semi(other0, _)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TraitBound {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TraitBound { ++ fn eq(&self, other: &Self) -> bool { ++ self.paren_token == other.paren_token ++ && self.modifier == other.modifier ++ && self.lifetimes == other.lifetimes ++ && self.path == other.path ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TraitBoundModifier {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TraitBoundModifier { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (TraitBoundModifier::None, TraitBoundModifier::None) => true, ++ (TraitBoundModifier::Maybe(_), TraitBoundModifier::Maybe(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for TraitItem {} ++#[cfg(feature = "full")] ++impl PartialEq for TraitItem { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (TraitItem::Const(self0), TraitItem::Const(other0)) => self0 == other0, ++ (TraitItem::Method(self0), TraitItem::Method(other0)) => self0 == other0, ++ (TraitItem::Type(self0), TraitItem::Type(other0)) => self0 == other0, ++ (TraitItem::Macro(self0), TraitItem::Macro(other0)) => self0 == other0, ++ (TraitItem::Verbatim(self0), TraitItem::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for TraitItemConst {} ++#[cfg(feature = "full")] ++impl PartialEq for TraitItemConst { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.ty == other.ty ++ && self.default == other.default ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for TraitItemMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for TraitItemMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mac == other.mac && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for TraitItemMethod {} ++#[cfg(feature = "full")] ++impl PartialEq for TraitItemMethod { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.sig == other.sig ++ && self.default == other.default ++ && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for TraitItemType {} ++#[cfg(feature = "full")] ++impl PartialEq for TraitItemType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.colon_token == other.colon_token ++ && self.bounds == other.bounds ++ && self.default == other.default ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Type {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Type { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Type::Array(self0), Type::Array(other0)) => self0 == other0, ++ (Type::BareFn(self0), Type::BareFn(other0)) => self0 == other0, ++ (Type::Group(self0), Type::Group(other0)) => self0 == other0, ++ (Type::ImplTrait(self0), Type::ImplTrait(other0)) => self0 == other0, ++ (Type::Infer(self0), Type::Infer(other0)) => self0 == other0, ++ (Type::Macro(self0), Type::Macro(other0)) => self0 == other0, ++ (Type::Never(self0), Type::Never(other0)) => self0 == other0, ++ (Type::Paren(self0), Type::Paren(other0)) => self0 == other0, ++ (Type::Path(self0), Type::Path(other0)) => self0 == other0, ++ (Type::Ptr(self0), Type::Ptr(other0)) => self0 == other0, ++ (Type::Reference(self0), Type::Reference(other0)) => self0 == other0, ++ (Type::Slice(self0), Type::Slice(other0)) => self0 == other0, ++ (Type::TraitObject(self0), Type::TraitObject(other0)) => self0 == other0, ++ (Type::Tuple(self0), Type::Tuple(other0)) => self0 == other0, ++ (Type::Verbatim(self0), Type::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeArray {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeArray { ++ fn eq(&self, other: &Self) -> bool { ++ self.elem == other.elem && self.len == other.len ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeBareFn {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeBareFn { ++ fn eq(&self, other: &Self) -> bool { ++ self.lifetimes == other.lifetimes ++ && self.unsafety == other.unsafety ++ && self.abi == other.abi ++ && self.inputs == other.inputs ++ && self.variadic == other.variadic ++ && self.output == other.output ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeGroup {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeGroup { ++ fn eq(&self, other: &Self) -> bool { ++ self.elem == other.elem ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeImplTrait {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeImplTrait { ++ fn eq(&self, other: &Self) -> bool { ++ self.bounds == other.bounds ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeInfer {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeInfer { ++ fn eq(&self, _other: &Self) -> bool { ++ true ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeMacro {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.mac == other.mac ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeNever {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeNever { ++ fn eq(&self, _other: &Self) -> bool { ++ true ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeParam {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeParam { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.colon_token == other.colon_token ++ && self.bounds == other.bounds ++ && self.eq_token == other.eq_token ++ && self.default == other.default ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeParamBound {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeParamBound { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (TypeParamBound::Trait(self0), TypeParamBound::Trait(other0)) => self0 == other0, ++ (TypeParamBound::Lifetime(self0), TypeParamBound::Lifetime(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeParen {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeParen { ++ fn eq(&self, other: &Self) -> bool { ++ self.elem == other.elem ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypePath {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypePath { ++ fn eq(&self, other: &Self) -> bool { ++ self.qself == other.qself && self.path == other.path ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypePtr {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypePtr { ++ fn eq(&self, other: &Self) -> bool { ++ self.const_token == other.const_token ++ && self.mutability == other.mutability ++ && self.elem == other.elem ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeReference {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeReference { ++ fn eq(&self, other: &Self) -> bool { ++ self.lifetime == other.lifetime ++ && self.mutability == other.mutability ++ && self.elem == other.elem ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeSlice {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeSlice { ++ fn eq(&self, other: &Self) -> bool { ++ self.elem == other.elem ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeTraitObject {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeTraitObject { ++ fn eq(&self, other: &Self) -> bool { ++ self.dyn_token == other.dyn_token && self.bounds == other.bounds ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeTuple {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeTuple { ++ fn eq(&self, other: &Self) -> bool { ++ self.elems == other.elems ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for UnOp {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for UnOp { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (UnOp::Deref(_), UnOp::Deref(_)) => true, ++ (UnOp::Not(_), UnOp::Not(_)) => true, ++ (UnOp::Neg(_), UnOp::Neg(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UseGlob {} ++#[cfg(feature = "full")] ++impl PartialEq for UseGlob { ++ fn eq(&self, _other: &Self) -> bool { ++ true ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UseGroup {} ++#[cfg(feature = "full")] ++impl PartialEq for UseGroup { ++ fn eq(&self, other: &Self) -> bool { ++ self.items == other.items ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UseName {} ++#[cfg(feature = "full")] ++impl PartialEq for UseName { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UsePath {} ++#[cfg(feature = "full")] ++impl PartialEq for UsePath { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident && self.tree == other.tree ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UseRename {} ++#[cfg(feature = "full")] ++impl PartialEq for UseRename { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident && self.rename == other.rename ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UseTree {} ++#[cfg(feature = "full")] ++impl PartialEq for UseTree { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (UseTree::Path(self0), UseTree::Path(other0)) => self0 == other0, ++ (UseTree::Name(self0), UseTree::Name(other0)) => self0 == other0, ++ (UseTree::Rename(self0), UseTree::Rename(other0)) => self0 == other0, ++ (UseTree::Glob(self0), UseTree::Glob(other0)) => self0 == other0, ++ (UseTree::Group(self0), UseTree::Group(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Variadic {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Variadic { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Variant {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Variant { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.fields == other.fields ++ && self.discriminant == other.discriminant ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for VisCrate {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for VisCrate { ++ fn eq(&self, _other: &Self) -> bool { ++ true ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for VisPublic {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for VisPublic { ++ fn eq(&self, _other: &Self) -> bool { ++ true ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for VisRestricted {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for VisRestricted { ++ fn eq(&self, other: &Self) -> bool { ++ self.in_token == other.in_token && self.path == other.path ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Visibility {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Visibility { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Visibility::Public(self0), Visibility::Public(other0)) => self0 == other0, ++ (Visibility::Crate(self0), Visibility::Crate(other0)) => self0 == other0, ++ (Visibility::Restricted(self0), Visibility::Restricted(other0)) => self0 == other0, ++ (Visibility::Inherited, Visibility::Inherited) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for WhereClause {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for WhereClause { ++ fn eq(&self, other: &Self) -> bool { ++ self.predicates == other.predicates ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for WherePredicate {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for WherePredicate { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (WherePredicate::Type(self0), WherePredicate::Type(other0)) => self0 == other0, ++ (WherePredicate::Lifetime(self0), WherePredicate::Lifetime(other0)) => self0 == other0, ++ (WherePredicate::Eq(self0), WherePredicate::Eq(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} +diff --git a/third_party/rust/syn/src/gen/fold.rs b/third_party/rust/syn/src/gen/fold.rs +index f51218b78c..d9dd32a420 100644 +--- a/third_party/rust/syn/src/gen/fold.rs ++++ b/third_party/rust/syn/src/gen/fold.rs +@@ -2,6 +2,7 @@ + // It is not intended for manual editing. + + #![allow(unreachable_code, unused_variables)] ++#![allow(clippy::match_wildcard_for_single_variants)] + #[cfg(any(feature = "full", feature = "derive"))] + use crate::gen::helper::fold::*; + #[cfg(any(feature = "full", feature = "derive"))] +@@ -26,7 +27,7 @@ macro_rules! full { + /// + /// [module documentation]: self + /// +-/// *This trait is available if Syn is built with the `"fold"` feature.* ++/// *This trait is available only if Syn is built with the `"fold"` feature.* + pub trait Fold { + #[cfg(any(feature = "derive", feature = "full"))] + fn fold_abi(&mut self, i: Abi) -> Abi { +@@ -433,35 +434,27 @@ pub trait Fold { + fn fold_lifetime_def(&mut self, i: LifetimeDef) -> LifetimeDef { + fold_lifetime_def(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit(&mut self, i: Lit) -> Lit { + fold_lit(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_bool(&mut self, i: LitBool) -> LitBool { + fold_lit_bool(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_byte(&mut self, i: LitByte) -> LitByte { + fold_lit_byte(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_byte_str(&mut self, i: LitByteStr) -> LitByteStr { + fold_lit_byte_str(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_char(&mut self, i: LitChar) -> LitChar { + fold_lit_char(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_float(&mut self, i: LitFloat) -> LitFloat { + fold_lit_float(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_int(&mut self, i: LitInt) -> LitInt { + fold_lit_int(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_str(&mut self, i: LitStr) -> LitStr { + fold_lit_str(self, i) + } +@@ -799,10 +792,10 @@ where + F: Fold + ?Sized, + { + AngleBracketedGenericArguments { +- colon2_token: (node.colon2_token).map(|it| Token ! [ :: ](tokens_helper(f, &it.spans))), +- lt_token: Token ! [ < ](tokens_helper(f, &node.lt_token.spans)), ++ colon2_token: (node.colon2_token).map(|it| Token ! [::](tokens_helper(f, &it.spans))), ++ lt_token: Token ! [<](tokens_helper(f, &node.lt_token.spans)), + args: FoldHelper::lift(node.args, |it| f.fold_generic_argument(it)), +- gt_token: Token ! [ > ](tokens_helper(f, &node.gt_token.spans)), ++ gt_token: Token ! [>](tokens_helper(f, &node.gt_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -819,9 +812,9 @@ where + Box::new(f.fold_expr(*(it).1)), + ) + }), +- fat_arrow_token: Token ! [ => ](tokens_helper(f, &node.fat_arrow_token.spans)), ++ fat_arrow_token: Token ! [=>](tokens_helper(f, &node.fat_arrow_token.spans)), + body: Box::new(f.fold_expr(*node.body)), +- comma: (node.comma).map(|it| Token ! [ , ](tokens_helper(f, &it.spans))), ++ comma: (node.comma).map(|it| Token ! [,](tokens_helper(f, &it.spans))), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -842,7 +835,7 @@ where + F: Fold + ?Sized, + { + Attribute { +- pound_token: Token ! [ # ](tokens_helper(f, &node.pound_token.spans)), ++ pound_token: Token ! [#](tokens_helper(f, &node.pound_token.spans)), + style: f.fold_attr_style(node.style), + bracket_token: Bracket(tokens_helper(f, &node.bracket_token.span)), + path: f.fold_path(node.path), +@@ -859,7 +852,7 @@ where + name: (node.name).map(|it| { + ( + f.fold_ident((it).0), +- Token ! [ : ](tokens_helper(f, &(it).1.spans)), ++ Token ! [:](tokens_helper(f, &(it).1.spans)), + ) + }), + ty: f.fold_type(node.ty), +@@ -871,59 +864,47 @@ where + F: Fold + ?Sized, + { + match node { +- BinOp::Add(_binding_0) => BinOp::Add(Token ! [ + ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Sub(_binding_0) => BinOp::Sub(Token ! [ - ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Mul(_binding_0) => BinOp::Mul(Token ! [ * ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Div(_binding_0) => BinOp::Div(Token ! [ / ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Rem(_binding_0) => BinOp::Rem(Token ! [ % ](tokens_helper(f, &_binding_0.spans))), +- BinOp::And(_binding_0) => BinOp::And(Token ! [ && ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Or(_binding_0) => BinOp::Or(Token ! [ || ](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Add(_binding_0) => BinOp::Add(Token ! [+](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Sub(_binding_0) => BinOp::Sub(Token ! [-](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Mul(_binding_0) => BinOp::Mul(Token ! [*](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Div(_binding_0) => BinOp::Div(Token ! [/](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Rem(_binding_0) => BinOp::Rem(Token ! [%](tokens_helper(f, &_binding_0.spans))), ++ BinOp::And(_binding_0) => BinOp::And(Token ! [&&](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Or(_binding_0) => BinOp::Or(Token ! [||](tokens_helper(f, &_binding_0.spans))), + BinOp::BitXor(_binding_0) => { +- BinOp::BitXor(Token ! [ ^ ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::BitXor(Token ! [^](tokens_helper(f, &_binding_0.spans))) + } + BinOp::BitAnd(_binding_0) => { +- BinOp::BitAnd(Token ! [ & ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::BitOr(_binding_0) => { +- BinOp::BitOr(Token ! [ | ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::Shl(_binding_0) => BinOp::Shl(Token ! [ << ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Shr(_binding_0) => BinOp::Shr(Token ! [ >> ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Eq(_binding_0) => BinOp::Eq(Token ! [ == ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Lt(_binding_0) => BinOp::Lt(Token ! [ < ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Le(_binding_0) => BinOp::Le(Token ! [ <= ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Ne(_binding_0) => BinOp::Ne(Token ! [ != ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Ge(_binding_0) => BinOp::Ge(Token ! [ >= ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Gt(_binding_0) => BinOp::Gt(Token ! [ > ](tokens_helper(f, &_binding_0.spans))), +- BinOp::AddEq(_binding_0) => { +- BinOp::AddEq(Token ! [ += ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::SubEq(_binding_0) => { +- BinOp::SubEq(Token ! [ -= ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::MulEq(_binding_0) => { +- BinOp::MulEq(Token ! [ *= ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::DivEq(_binding_0) => { +- BinOp::DivEq(Token ! [ /= ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::RemEq(_binding_0) => { +- BinOp::RemEq(Token ! [ %= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::BitAnd(Token ! [&](tokens_helper(f, &_binding_0.spans))) + } ++ BinOp::BitOr(_binding_0) => BinOp::BitOr(Token ! [|](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Shl(_binding_0) => BinOp::Shl(Token ! [<<](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Shr(_binding_0) => BinOp::Shr(Token ! [>>](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Eq(_binding_0) => BinOp::Eq(Token ! [==](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Lt(_binding_0) => BinOp::Lt(Token ! [<](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Le(_binding_0) => BinOp::Le(Token ! [<=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Ne(_binding_0) => BinOp::Ne(Token ! [!=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Ge(_binding_0) => BinOp::Ge(Token ! [>=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Gt(_binding_0) => BinOp::Gt(Token ! [>](tokens_helper(f, &_binding_0.spans))), ++ BinOp::AddEq(_binding_0) => BinOp::AddEq(Token ! [+=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::SubEq(_binding_0) => BinOp::SubEq(Token ! [-=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::MulEq(_binding_0) => BinOp::MulEq(Token ! [*=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::DivEq(_binding_0) => BinOp::DivEq(Token ! [/=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::RemEq(_binding_0) => BinOp::RemEq(Token ! [%=](tokens_helper(f, &_binding_0.spans))), + BinOp::BitXorEq(_binding_0) => { +- BinOp::BitXorEq(Token ! [ ^= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::BitXorEq(Token ! [^=](tokens_helper(f, &_binding_0.spans))) + } + BinOp::BitAndEq(_binding_0) => { +- BinOp::BitAndEq(Token ! [ &= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::BitAndEq(Token ! [&=](tokens_helper(f, &_binding_0.spans))) + } + BinOp::BitOrEq(_binding_0) => { +- BinOp::BitOrEq(Token ! [ |= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::BitOrEq(Token ! [|=](tokens_helper(f, &_binding_0.spans))) + } + BinOp::ShlEq(_binding_0) => { +- BinOp::ShlEq(Token ! [ <<= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::ShlEq(Token ! [<<=](tokens_helper(f, &_binding_0.spans))) + } + BinOp::ShrEq(_binding_0) => { +- BinOp::ShrEq(Token ! [ >>= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::ShrEq(Token ! [>>=](tokens_helper(f, &_binding_0.spans))) + } + } + } +@@ -934,7 +915,7 @@ where + { + Binding { + ident: f.fold_ident(node.ident), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + ty: f.fold_type(node.ty), + } + } +@@ -955,9 +936,9 @@ where + { + BoundLifetimes { + for_token: Token![for](tokens_helper(f, &node.for_token.span)), +- lt_token: Token ! [ < ](tokens_helper(f, &node.lt_token.spans)), ++ lt_token: Token ! [<](tokens_helper(f, &node.lt_token.spans)), + lifetimes: FoldHelper::lift(node.lifetimes, |it| f.fold_lifetime_def(it)), +- gt_token: Token ! [ > ](tokens_helper(f, &node.gt_token.spans)), ++ gt_token: Token ! [>](tokens_helper(f, &node.gt_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -969,9 +950,9 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + const_token: Token![const](tokens_helper(f, &node.const_token.span)), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: f.fold_type(node.ty), +- eq_token: (node.eq_token).map(|it| Token ! [ = ](tokens_helper(f, &it.spans))), ++ eq_token: (node.eq_token).map(|it| Token ! [=](tokens_helper(f, &it.spans))), + default: (node.default).map(|it| f.fold_expr(it)), + } + } +@@ -982,7 +963,7 @@ where + { + Constraint { + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_type_param_bound(it)), + } + } +@@ -1016,7 +997,7 @@ where + DataStruct { + struct_token: Token![struct](tokens_helper(f, &node.struct_token.span)), + fields: f.fold_fields(node.fields), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "derive")] +@@ -1112,7 +1093,7 @@ where + ExprAssign { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + left: Box::new(f.fold_expr(*node.left)), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + right: Box::new(f.fold_expr(*node.right)), + } + } +@@ -1148,7 +1129,7 @@ where + ExprAwait { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + base: Box::new(f.fold_expr(*node.base)), +- dot_token: Token ! [ . ](tokens_helper(f, &node.dot_token.spans)), ++ dot_token: Token ! [.](tokens_helper(f, &node.dot_token.spans)), + await_token: crate::token::Await(tokens_helper(f, &node.await_token.span)), + } + } +@@ -1232,9 +1213,9 @@ where + asyncness: (node.asyncness).map(|it| Token![async](tokens_helper(f, &it.span))), + movability: (node.movability).map(|it| Token![static](tokens_helper(f, &it.span))), + capture: (node.capture).map(|it| Token![move](tokens_helper(f, &it.span))), +- or1_token: Token ! [ | ](tokens_helper(f, &node.or1_token.spans)), ++ or1_token: Token ! [|](tokens_helper(f, &node.or1_token.spans)), + inputs: FoldHelper::lift(node.inputs, |it| f.fold_pat(it)), +- or2_token: Token ! [ | ](tokens_helper(f, &node.or2_token.spans)), ++ or2_token: Token ! [|](tokens_helper(f, &node.or2_token.spans)), + output: f.fold_return_type(node.output), + body: Box::new(f.fold_expr(*node.body)), + } +@@ -1258,7 +1239,7 @@ where + ExprField { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + base: Box::new(f.fold_expr(*node.base)), +- dot_token: Token ! [ . ](tokens_helper(f, &node.dot_token.spans)), ++ dot_token: Token ! [.](tokens_helper(f, &node.dot_token.spans)), + member: f.fold_member(node.member), + } + } +@@ -1327,7 +1308,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + let_token: Token![let](tokens_helper(f, &node.let_token.span)), + pat: f.fold_pat(node.pat), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + expr: Box::new(f.fold_expr(*node.expr)), + } + } +@@ -1384,7 +1365,7 @@ where + ExprMethodCall { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + receiver: Box::new(f.fold_expr(*node.receiver)), +- dot_token: Token ! [ . ](tokens_helper(f, &node.dot_token.spans)), ++ dot_token: Token ! [.](tokens_helper(f, &node.dot_token.spans)), + method: f.fold_ident(node.method), + turbofish: (node.turbofish).map(|it| f.fold_method_turbofish(it)), + paren_token: Paren(tokens_helper(f, &node.paren_token.span)), +@@ -1432,7 +1413,7 @@ where + { + ExprReference { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), +- and_token: Token ! [ & ](tokens_helper(f, &node.and_token.spans)), ++ and_token: Token ! [&](tokens_helper(f, &node.and_token.spans)), + raw: node.raw, + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + expr: Box::new(f.fold_expr(*node.expr)), +@@ -1447,7 +1428,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + bracket_token: Bracket(tokens_helper(f, &node.bracket_token.span)), + expr: Box::new(f.fold_expr(*node.expr)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + len: Box::new(f.fold_expr(*node.len)), + } + } +@@ -1484,7 +1465,7 @@ where + ExprTry { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + expr: Box::new(f.fold_expr(*node.expr)), +- question_token: Token ! [ ? ](tokens_helper(f, &node.question_token.spans)), ++ question_token: Token ! [?](tokens_helper(f, &node.question_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -1517,7 +1498,7 @@ where + ExprType { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + expr: Box::new(f.fold_expr(*node.expr)), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), + } + } +@@ -1576,7 +1557,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + vis: f.fold_visibility(node.vis), + ident: (node.ident).map(|it| f.fold_ident(it)), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + ty: f.fold_type(node.ty), + } + } +@@ -1588,7 +1569,7 @@ where + FieldPat { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + member: f.fold_member(node.member), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + pat: Box::new(f.fold_pat(*node.pat)), + } + } +@@ -1600,7 +1581,7 @@ where + FieldValue { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + member: f.fold_member(node.member), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + expr: f.fold_expr(node.expr), + } + } +@@ -1681,7 +1662,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + vis: f.fold_visibility(node.vis), + sig: f.fold_signature(node.sig), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -1692,7 +1673,7 @@ where + ForeignItemMacro { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + mac: f.fold_macro(node.mac), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -1706,9 +1687,9 @@ where + static_token: Token![static](tokens_helper(f, &node.static_token.span)), + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -1721,7 +1702,7 @@ where + vis: f.fold_visibility(node.vis), + type_token: Token![type](tokens_helper(f, &node.type_token.span)), + ident: f.fold_ident(node.ident), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -1779,9 +1760,9 @@ where + F: Fold + ?Sized, + { + Generics { +- lt_token: (node.lt_token).map(|it| Token ! [ < ](tokens_helper(f, &it.spans))), ++ lt_token: (node.lt_token).map(|it| Token ! [<](tokens_helper(f, &it.spans))), + params: FoldHelper::lift(node.params, |it| f.fold_generic_param(it)), +- gt_token: (node.gt_token).map(|it| Token ! [ > ](tokens_helper(f, &it.spans))), ++ gt_token: (node.gt_token).map(|it| Token ! [>](tokens_helper(f, &it.spans))), + where_clause: (node.where_clause).map(|it| f.fold_where_clause(it)), + } + } +@@ -1819,11 +1800,11 @@ where + defaultness: (node.defaultness).map(|it| Token![default](tokens_helper(f, &it.span))), + const_token: Token![const](tokens_helper(f, &node.const_token.span)), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: f.fold_type(node.ty), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + expr: f.fold_expr(node.expr), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -1834,7 +1815,7 @@ where + ImplItemMacro { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + mac: f.fold_macro(node.mac), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -1862,9 +1843,9 @@ where + type_token: Token![type](tokens_helper(f, &node.type_token.span)), + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + ty: f.fold_type(node.ty), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -1913,11 +1894,11 @@ where + vis: f.fold_visibility(node.vis), + const_token: Token![const](tokens_helper(f, &node.const_token.span)), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + expr: Box::new(f.fold_expr(*node.expr)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -1952,7 +1933,7 @@ where + f.fold_ident((it).1), + ) + }), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2011,7 +1992,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + ident: (node.ident).map(|it| f.fold_ident(it)), + mac: f.fold_macro(node.mac), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -2043,7 +2024,7 @@ where + FoldHelper::lift((it).1, |it| f.fold_item(it)), + ) + }), +- semi: (node.semi).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi: (node.semi).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -2057,11 +2038,11 @@ where + static_token: Token![static](tokens_helper(f, &node.static_token.span)), + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + expr: Box::new(f.fold_expr(*node.expr)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2076,7 +2057,7 @@ where + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), + fields: f.fold_fields(node.fields), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -2092,7 +2073,7 @@ where + trait_token: Token![trait](tokens_helper(f, &node.trait_token.span)), + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + supertraits: FoldHelper::lift(node.supertraits, |it| f.fold_type_param_bound(it)), + brace_token: Brace(tokens_helper(f, &node.brace_token.span)), + items: FoldHelper::lift(node.items, |it| f.fold_trait_item(it)), +@@ -2109,9 +2090,9 @@ where + trait_token: Token![trait](tokens_helper(f, &node.trait_token.span)), + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_type_param_bound(it)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2125,9 +2106,9 @@ where + type_token: Token![type](tokens_helper(f, &node.type_token.span)), + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2153,9 +2134,9 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + vis: f.fold_visibility(node.vis), + use_token: Token![use](tokens_helper(f, &node.use_token.span)), +- leading_colon: (node.leading_colon).map(|it| Token ! [ :: ](tokens_helper(f, &it.spans))), ++ leading_colon: (node.leading_colon).map(|it| Token ! [::](tokens_helper(f, &it.spans))), + tree: f.fold_use_tree(node.tree), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2165,7 +2146,7 @@ where + { + Label { + name: f.fold_lifetime(node.name), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + } + } + pub fn fold_lifetime<F>(f: &mut F, node: Lifetime) -> Lifetime +@@ -2185,11 +2166,10 @@ where + LifetimeDef { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + lifetime: f.fold_lifetime(node.lifetime), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_lifetime(it)), + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit<F>(f: &mut F, node: Lit) -> Lit + where + F: Fold + ?Sized, +@@ -2205,7 +2185,6 @@ where + Lit::Verbatim(_binding_0) => Lit::Verbatim(_binding_0), + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_bool<F>(f: &mut F, node: LitBool) -> LitBool + where + F: Fold + ?Sized, +@@ -2215,7 +2194,6 @@ where + span: f.fold_span(node.span), + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_byte<F>(f: &mut F, node: LitByte) -> LitByte + where + F: Fold + ?Sized, +@@ -2225,7 +2203,6 @@ where + node.set_span(span); + node + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_byte_str<F>(f: &mut F, node: LitByteStr) -> LitByteStr + where + F: Fold + ?Sized, +@@ -2235,7 +2212,6 @@ where + node.set_span(span); + node + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_char<F>(f: &mut F, node: LitChar) -> LitChar + where + F: Fold + ?Sized, +@@ -2245,7 +2221,6 @@ where + node.set_span(span); + node + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_float<F>(f: &mut F, node: LitFloat) -> LitFloat + where + F: Fold + ?Sized, +@@ -2255,7 +2230,6 @@ where + node.set_span(span); + node + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_int<F>(f: &mut F, node: LitInt) -> LitInt + where + F: Fold + ?Sized, +@@ -2265,7 +2239,6 @@ where + node.set_span(span); + node + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_str<F>(f: &mut F, node: LitStr) -> LitStr + where + F: Fold + ?Sized, +@@ -2286,11 +2259,11 @@ where + pat: f.fold_pat(node.pat), + init: (node.init).map(|it| { + ( +- Token ! [ = ](tokens_helper(f, &(it).0.spans)), ++ Token ! [=](tokens_helper(f, &(it).0.spans)), + Box::new(f.fold_expr(*(it).1)), + ) + }), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -2361,7 +2334,7 @@ where + { + MetaNameValue { + path: f.fold_path(node.path), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + lit: f.fold_lit(node.lit), + } + } +@@ -2371,10 +2344,10 @@ where + F: Fold + ?Sized, + { + MethodTurbofish { +- colon2_token: Token ! [ :: ](tokens_helper(f, &node.colon2_token.spans)), +- lt_token: Token ! [ < ](tokens_helper(f, &node.lt_token.spans)), ++ colon2_token: Token ! [::](tokens_helper(f, &node.colon2_token.spans)), ++ lt_token: Token ! [<](tokens_helper(f, &node.lt_token.spans)), + args: FoldHelper::lift(node.args, |it| f.fold_generic_method_argument(it)), +- gt_token: Token ! [ > ](tokens_helper(f, &node.gt_token.spans)), ++ gt_token: Token ! [>](tokens_helper(f, &node.gt_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -2449,7 +2422,7 @@ where + ident: f.fold_ident(node.ident), + subpat: (node.subpat).map(|it| { + ( +- Token ! [ @ ](tokens_helper(f, &(it).0.spans)), ++ Token ! [@](tokens_helper(f, &(it).0.spans)), + Box::new(f.fold_pat(*(it).1)), + ) + }), +@@ -2482,7 +2455,7 @@ where + { + PatOr { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), +- leading_vert: (node.leading_vert).map(|it| Token ! [ | ](tokens_helper(f, &it.spans))), ++ leading_vert: (node.leading_vert).map(|it| Token ! [|](tokens_helper(f, &it.spans))), + cases: FoldHelper::lift(node.cases, |it| f.fold_pat(it)), + } + } +@@ -2516,7 +2489,7 @@ where + { + PatReference { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), +- and_token: Token ! [ & ](tokens_helper(f, &node.and_token.spans)), ++ and_token: Token ! [&](tokens_helper(f, &node.and_token.spans)), + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + pat: Box::new(f.fold_pat(*node.pat)), + } +@@ -2585,7 +2558,7 @@ where + PatType { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + pat: Box::new(f.fold_pat(*node.pat)), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), + } + } +@@ -2605,7 +2578,7 @@ where + F: Fold + ?Sized, + { + Path { +- leading_colon: (node.leading_colon).map(|it| Token ! [ :: ](tokens_helper(f, &it.spans))), ++ leading_colon: (node.leading_colon).map(|it| Token ! [::](tokens_helper(f, &it.spans))), + segments: FoldHelper::lift(node.segments, |it| f.fold_path_segment(it)), + } + } +@@ -2641,7 +2614,7 @@ where + { + PredicateEq { + lhs_ty: f.fold_type(node.lhs_ty), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + rhs_ty: f.fold_type(node.rhs_ty), + } + } +@@ -2652,7 +2625,7 @@ where + { + PredicateLifetime { + lifetime: f.fold_lifetime(node.lifetime), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_lifetime(it)), + } + } +@@ -2664,7 +2637,7 @@ where + PredicateType { + lifetimes: (node.lifetimes).map(|it| f.fold_bound_lifetimes(it)), + bounded_ty: f.fold_type(node.bounded_ty), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_type_param_bound(it)), + } + } +@@ -2674,11 +2647,11 @@ where + F: Fold + ?Sized, + { + QSelf { +- lt_token: Token ! [ < ](tokens_helper(f, &node.lt_token.spans)), ++ lt_token: Token ! [<](tokens_helper(f, &node.lt_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), + position: node.position, + as_token: (node.as_token).map(|it| Token![as](tokens_helper(f, &it.span))), +- gt_token: Token ! [ > ](tokens_helper(f, &node.gt_token.spans)), ++ gt_token: Token ! [>](tokens_helper(f, &node.gt_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2691,7 +2664,7 @@ where + RangeLimits::HalfOpen(Token![..](tokens_helper(f, &_binding_0.spans))) + } + RangeLimits::Closed(_binding_0) => { +- RangeLimits::Closed(Token ! [ ..= ](tokens_helper(f, &_binding_0.spans))) ++ RangeLimits::Closed(Token ! [..=](tokens_helper(f, &_binding_0.spans))) + } + } + } +@@ -2704,7 +2677,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + reference: (node.reference).map(|it| { + ( +- Token ! [ & ](tokens_helper(f, &(it).0.spans)), ++ Token ! [&](tokens_helper(f, &(it).0.spans)), + ((it).1).map(|it| f.fold_lifetime(it)), + ) + }), +@@ -2720,7 +2693,7 @@ where + match node { + ReturnType::Default => ReturnType::Default, + ReturnType::Type(_binding_0, _binding_1) => ReturnType::Type( +- Token ! [ -> ](tokens_helper(f, &_binding_0.spans)), ++ Token ! [->](tokens_helper(f, &_binding_0.spans)), + Box::new(f.fold_type(*_binding_1)), + ), + } +@@ -2761,7 +2734,7 @@ where + Stmt::Expr(_binding_0) => Stmt::Expr(f.fold_expr(_binding_0)), + Stmt::Semi(_binding_0, _binding_1) => Stmt::Semi( + f.fold_expr(_binding_0), +- Token ! [ ; ](tokens_helper(f, &_binding_1.spans)), ++ Token ! [;](tokens_helper(f, &_binding_1.spans)), + ), + } + } +@@ -2785,7 +2758,7 @@ where + match node { + TraitBoundModifier::None => TraitBoundModifier::None, + TraitBoundModifier::Maybe(_binding_0) => { +- TraitBoundModifier::Maybe(Token ! [ ? ](tokens_helper(f, &_binding_0.spans))) ++ TraitBoundModifier::Maybe(Token ! [?](tokens_helper(f, &_binding_0.spans))) + } + } + } +@@ -2812,15 +2785,15 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + const_token: Token![const](tokens_helper(f, &node.const_token.span)), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: f.fold_type(node.ty), + default: (node.default).map(|it| { + ( +- Token ! [ = ](tokens_helper(f, &(it).0.spans)), ++ Token ! [=](tokens_helper(f, &(it).0.spans)), + f.fold_expr((it).1), + ) + }), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2831,7 +2804,7 @@ where + TraitItemMacro { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + mac: f.fold_macro(node.mac), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -2843,7 +2816,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + sig: f.fold_signature(node.sig), + default: (node.default).map(|it| f.fold_block(it)), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -2856,15 +2829,15 @@ where + type_token: Token![type](tokens_helper(f, &node.type_token.span)), + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_type_param_bound(it)), + default: (node.default).map(|it| { + ( +- Token ! [ = ](tokens_helper(f, &(it).0.spans)), ++ Token ! [=](tokens_helper(f, &(it).0.spans)), + f.fold_type((it).1), + ) + }), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -2899,7 +2872,7 @@ where + TypeArray { + bracket_token: Bracket(tokens_helper(f, &node.bracket_token.span)), + elem: Box::new(f.fold_type(*node.elem)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + len: f.fold_expr(node.len), + } + } +@@ -2974,9 +2947,9 @@ where + TypeParam { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + ident: f.fold_ident(node.ident), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_type_param_bound(it)), +- eq_token: (node.eq_token).map(|it| Token ! [ = ](tokens_helper(f, &it.spans))), ++ eq_token: (node.eq_token).map(|it| Token ! [=](tokens_helper(f, &it.spans))), + default: (node.default).map(|it| f.fold_type(it)), + } + } +@@ -3018,7 +2991,7 @@ where + F: Fold + ?Sized, + { + TypePtr { +- star_token: Token ! [ * ](tokens_helper(f, &node.star_token.spans)), ++ star_token: Token ! [*](tokens_helper(f, &node.star_token.spans)), + const_token: (node.const_token).map(|it| Token![const](tokens_helper(f, &it.span))), + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + elem: Box::new(f.fold_type(*node.elem)), +@@ -3030,7 +3003,7 @@ where + F: Fold + ?Sized, + { + TypeReference { +- and_token: Token ! [ & ](tokens_helper(f, &node.and_token.spans)), ++ and_token: Token ! [&](tokens_helper(f, &node.and_token.spans)), + lifetime: (node.lifetime).map(|it| f.fold_lifetime(it)), + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + elem: Box::new(f.fold_type(*node.elem)), +@@ -3072,9 +3045,9 @@ where + F: Fold + ?Sized, + { + match node { +- UnOp::Deref(_binding_0) => UnOp::Deref(Token ! [ * ](tokens_helper(f, &_binding_0.spans))), ++ UnOp::Deref(_binding_0) => UnOp::Deref(Token ! [*](tokens_helper(f, &_binding_0.spans))), + UnOp::Not(_binding_0) => UnOp::Not(Token![!](tokens_helper(f, &_binding_0.spans))), +- UnOp::Neg(_binding_0) => UnOp::Neg(Token ! [ - ](tokens_helper(f, &_binding_0.spans))), ++ UnOp::Neg(_binding_0) => UnOp::Neg(Token ! [-](tokens_helper(f, &_binding_0.spans))), + } + } + #[cfg(feature = "full")] +@@ -3083,7 +3056,7 @@ where + F: Fold + ?Sized, + { + UseGlob { +- star_token: Token ! [ * ](tokens_helper(f, &node.star_token.spans)), ++ star_token: Token ! [*](tokens_helper(f, &node.star_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -3112,7 +3085,7 @@ where + { + UsePath { + ident: f.fold_ident(node.ident), +- colon2_token: Token ! [ :: ](tokens_helper(f, &node.colon2_token.spans)), ++ colon2_token: Token ! [::](tokens_helper(f, &node.colon2_token.spans)), + tree: Box::new(f.fold_use_tree(*node.tree)), + } + } +@@ -3147,7 +3120,7 @@ where + { + Variadic { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), +- dots: Token ! [ ... ](tokens_helper(f, &node.dots.spans)), ++ dots: Token ! [...](tokens_helper(f, &node.dots.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -3161,7 +3134,7 @@ where + fields: f.fold_fields(node.fields), + discriminant: (node.discriminant).map(|it| { + ( +- Token ! [ = ](tokens_helper(f, &(it).0.spans)), ++ Token ! [=](tokens_helper(f, &(it).0.spans)), + f.fold_expr((it).1), + ) + }), +diff --git a/third_party/rust/syn/src/gen/hash.rs b/third_party/rust/syn/src/gen/hash.rs +new file mode 100644 +index 0000000000..9e9e84a7af +--- /dev/null ++++ b/third_party/rust/syn/src/gen/hash.rs +@@ -0,0 +1,2691 @@ ++// This file is @generated by syn-internal-codegen. ++// It is not intended for manual editing. ++ ++#[cfg(any(feature = "derive", feature = "full"))] ++use crate::tt::TokenStreamHelper; ++use crate::*; ++use std::hash::{Hash, Hasher}; ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Abi { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.name.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for AngleBracketedGenericArguments { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.colon2_token.hash(state); ++ self.args.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Arm { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.pat.hash(state); ++ self.guard.hash(state); ++ self.body.hash(state); ++ self.comma.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for AttrStyle { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ AttrStyle::Outer => { ++ state.write_u8(0u8); ++ } ++ AttrStyle::Inner(_) => { ++ state.write_u8(1u8); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Attribute { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.style.hash(state); ++ self.path.hash(state); ++ TokenStreamHelper(&self.tokens).hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for BareFnArg { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.name.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for BinOp { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ BinOp::Add(_) => { ++ state.write_u8(0u8); ++ } ++ BinOp::Sub(_) => { ++ state.write_u8(1u8); ++ } ++ BinOp::Mul(_) => { ++ state.write_u8(2u8); ++ } ++ BinOp::Div(_) => { ++ state.write_u8(3u8); ++ } ++ BinOp::Rem(_) => { ++ state.write_u8(4u8); ++ } ++ BinOp::And(_) => { ++ state.write_u8(5u8); ++ } ++ BinOp::Or(_) => { ++ state.write_u8(6u8); ++ } ++ BinOp::BitXor(_) => { ++ state.write_u8(7u8); ++ } ++ BinOp::BitAnd(_) => { ++ state.write_u8(8u8); ++ } ++ BinOp::BitOr(_) => { ++ state.write_u8(9u8); ++ } ++ BinOp::Shl(_) => { ++ state.write_u8(10u8); ++ } ++ BinOp::Shr(_) => { ++ state.write_u8(11u8); ++ } ++ BinOp::Eq(_) => { ++ state.write_u8(12u8); ++ } ++ BinOp::Lt(_) => { ++ state.write_u8(13u8); ++ } ++ BinOp::Le(_) => { ++ state.write_u8(14u8); ++ } ++ BinOp::Ne(_) => { ++ state.write_u8(15u8); ++ } ++ BinOp::Ge(_) => { ++ state.write_u8(16u8); ++ } ++ BinOp::Gt(_) => { ++ state.write_u8(17u8); ++ } ++ BinOp::AddEq(_) => { ++ state.write_u8(18u8); ++ } ++ BinOp::SubEq(_) => { ++ state.write_u8(19u8); ++ } ++ BinOp::MulEq(_) => { ++ state.write_u8(20u8); ++ } ++ BinOp::DivEq(_) => { ++ state.write_u8(21u8); ++ } ++ BinOp::RemEq(_) => { ++ state.write_u8(22u8); ++ } ++ BinOp::BitXorEq(_) => { ++ state.write_u8(23u8); ++ } ++ BinOp::BitAndEq(_) => { ++ state.write_u8(24u8); ++ } ++ BinOp::BitOrEq(_) => { ++ state.write_u8(25u8); ++ } ++ BinOp::ShlEq(_) => { ++ state.write_u8(26u8); ++ } ++ BinOp::ShrEq(_) => { ++ state.write_u8(27u8); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Binding { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Block { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.stmts.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for BoundLifetimes { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lifetimes.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ConstParam { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ self.eq_token.hash(state); ++ self.default.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Constraint { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ self.bounds.hash(state); ++ } ++} ++#[cfg(feature = "derive")] ++impl Hash for Data { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Data::Struct(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Data::Enum(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Data::Union(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Hash for DataEnum { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.variants.hash(state); ++ } ++} ++#[cfg(feature = "derive")] ++impl Hash for DataStruct { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.fields.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "derive")] ++impl Hash for DataUnion { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.fields.hash(state); ++ } ++} ++#[cfg(feature = "derive")] ++impl Hash for DeriveInput { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.data.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Expr { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ #[cfg(feature = "full")] ++ Expr::Array(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Assign(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::AssignOp(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Async(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Await(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ Expr::Binary(v0) => { ++ state.write_u8(5u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Block(v0) => { ++ state.write_u8(6u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Box(v0) => { ++ state.write_u8(7u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Break(v0) => { ++ state.write_u8(8u8); ++ v0.hash(state); ++ } ++ Expr::Call(v0) => { ++ state.write_u8(9u8); ++ v0.hash(state); ++ } ++ Expr::Cast(v0) => { ++ state.write_u8(10u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Closure(v0) => { ++ state.write_u8(11u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Continue(v0) => { ++ state.write_u8(12u8); ++ v0.hash(state); ++ } ++ Expr::Field(v0) => { ++ state.write_u8(13u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::ForLoop(v0) => { ++ state.write_u8(14u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Group(v0) => { ++ state.write_u8(15u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::If(v0) => { ++ state.write_u8(16u8); ++ v0.hash(state); ++ } ++ Expr::Index(v0) => { ++ state.write_u8(17u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Let(v0) => { ++ state.write_u8(18u8); ++ v0.hash(state); ++ } ++ Expr::Lit(v0) => { ++ state.write_u8(19u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Loop(v0) => { ++ state.write_u8(20u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Macro(v0) => { ++ state.write_u8(21u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Match(v0) => { ++ state.write_u8(22u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::MethodCall(v0) => { ++ state.write_u8(23u8); ++ v0.hash(state); ++ } ++ Expr::Paren(v0) => { ++ state.write_u8(24u8); ++ v0.hash(state); ++ } ++ Expr::Path(v0) => { ++ state.write_u8(25u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Range(v0) => { ++ state.write_u8(26u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Reference(v0) => { ++ state.write_u8(27u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Repeat(v0) => { ++ state.write_u8(28u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Return(v0) => { ++ state.write_u8(29u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Struct(v0) => { ++ state.write_u8(30u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Try(v0) => { ++ state.write_u8(31u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::TryBlock(v0) => { ++ state.write_u8(32u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Tuple(v0) => { ++ state.write_u8(33u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Type(v0) => { ++ state.write_u8(34u8); ++ v0.hash(state); ++ } ++ Expr::Unary(v0) => { ++ state.write_u8(35u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Unsafe(v0) => { ++ state.write_u8(36u8); ++ v0.hash(state); ++ } ++ Expr::Verbatim(v0) => { ++ state.write_u8(37u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::While(v0) => { ++ state.write_u8(38u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Yield(v0) => { ++ state.write_u8(39u8); ++ v0.hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprArray { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.elems.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprAssign { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.left.hash(state); ++ self.right.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprAssignOp { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.left.hash(state); ++ self.op.hash(state); ++ self.right.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprAsync { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.capture.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprAwait { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.base.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprBinary { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.left.hash(state); ++ self.op.hash(state); ++ self.right.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprBlock { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprBox { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprBreak { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprCall { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.func.hash(state); ++ self.args.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprCast { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprClosure { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.asyncness.hash(state); ++ self.movability.hash(state); ++ self.capture.hash(state); ++ self.inputs.hash(state); ++ self.output.hash(state); ++ self.body.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprContinue { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprField { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.base.hash(state); ++ self.member.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprForLoop { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ self.pat.hash(state); ++ self.expr.hash(state); ++ self.body.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprGroup { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprIf { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.cond.hash(state); ++ self.then_branch.hash(state); ++ self.else_branch.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprIndex { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ self.index.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprLet { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.pat.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprLit { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.lit.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprLoop { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ self.body.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mac.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprMatch { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ self.arms.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprMethodCall { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.receiver.hash(state); ++ self.method.hash(state); ++ self.turbofish.hash(state); ++ self.args.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprParen { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprPath { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.qself.hash(state); ++ self.path.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprRange { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.from.hash(state); ++ self.limits.hash(state); ++ self.to.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprReference { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mutability.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprRepeat { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ self.len.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprReturn { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprStruct { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.path.hash(state); ++ self.fields.hash(state); ++ self.dot2_token.hash(state); ++ self.rest.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprTry { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprTryBlock { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprTuple { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.elems.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprUnary { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.op.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprUnsafe { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprWhile { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ self.cond.hash(state); ++ self.body.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprYield { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Field { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.colon_token.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for FieldPat { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.member.hash(state); ++ self.colon_token.hash(state); ++ self.pat.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for FieldValue { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.member.hash(state); ++ self.colon_token.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Fields { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Fields::Named(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Fields::Unnamed(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Fields::Unit => { ++ state.write_u8(2u8); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for FieldsNamed { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.named.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for FieldsUnnamed { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.unnamed.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for File { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.shebang.hash(state); ++ self.attrs.hash(state); ++ self.items.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for FnArg { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ FnArg::Receiver(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ FnArg::Typed(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ForeignItem { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ ForeignItem::Fn(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ ForeignItem::Static(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ ForeignItem::Type(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ ForeignItem::Macro(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ ForeignItem::Verbatim(v0) => { ++ state.write_u8(4u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ForeignItemFn { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.sig.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ForeignItemMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mac.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ForeignItemStatic { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.mutability.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ForeignItemType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for GenericArgument { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ GenericArgument::Lifetime(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ GenericArgument::Type(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ GenericArgument::Binding(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ GenericArgument::Constraint(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ GenericArgument::Const(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for GenericMethodArgument { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ GenericMethodArgument::Type(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ GenericMethodArgument::Const(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for GenericParam { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ GenericParam::Type(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ GenericParam::Lifetime(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ GenericParam::Const(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Generics { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lt_token.hash(state); ++ self.params.hash(state); ++ self.gt_token.hash(state); ++ self.where_clause.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ImplItem { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ ImplItem::Const(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ ImplItem::Method(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ ImplItem::Type(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ ImplItem::Macro(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ ImplItem::Verbatim(v0) => { ++ state.write_u8(4u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ImplItemConst { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.defaultness.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ImplItemMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mac.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ImplItemMethod { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.defaultness.hash(state); ++ self.sig.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ImplItemType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.defaultness.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Item { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Item::Const(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Item::Enum(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Item::ExternCrate(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Item::Fn(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ Item::ForeignMod(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ Item::Impl(v0) => { ++ state.write_u8(5u8); ++ v0.hash(state); ++ } ++ Item::Macro(v0) => { ++ state.write_u8(6u8); ++ v0.hash(state); ++ } ++ Item::Macro2(v0) => { ++ state.write_u8(7u8); ++ v0.hash(state); ++ } ++ Item::Mod(v0) => { ++ state.write_u8(8u8); ++ v0.hash(state); ++ } ++ Item::Static(v0) => { ++ state.write_u8(9u8); ++ v0.hash(state); ++ } ++ Item::Struct(v0) => { ++ state.write_u8(10u8); ++ v0.hash(state); ++ } ++ Item::Trait(v0) => { ++ state.write_u8(11u8); ++ v0.hash(state); ++ } ++ Item::TraitAlias(v0) => { ++ state.write_u8(12u8); ++ v0.hash(state); ++ } ++ Item::Type(v0) => { ++ state.write_u8(13u8); ++ v0.hash(state); ++ } ++ Item::Union(v0) => { ++ state.write_u8(14u8); ++ v0.hash(state); ++ } ++ Item::Use(v0) => { ++ state.write_u8(15u8); ++ v0.hash(state); ++ } ++ Item::Verbatim(v0) => { ++ state.write_u8(16u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemConst { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemEnum { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.variants.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemExternCrate { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.rename.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemFn { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.sig.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemForeignMod { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.abi.hash(state); ++ self.items.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemImpl { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.defaultness.hash(state); ++ self.unsafety.hash(state); ++ self.generics.hash(state); ++ self.trait_.hash(state); ++ self.self_ty.hash(state); ++ self.items.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.mac.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemMacro2 { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ TokenStreamHelper(&self.rules).hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemMod { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.content.hash(state); ++ self.semi.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemStatic { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.mutability.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemStruct { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.fields.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemTrait { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.unsafety.hash(state); ++ self.auto_token.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.colon_token.hash(state); ++ self.supertraits.hash(state); ++ self.items.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemTraitAlias { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.bounds.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemUnion { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.fields.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemUse { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.leading_colon.hash(state); ++ self.tree.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Label { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.name.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for LifetimeDef { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.lifetime.hash(state); ++ self.colon_token.hash(state); ++ self.bounds.hash(state); ++ } ++} ++impl Hash for Lit { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Lit::Str(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Lit::ByteStr(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Lit::Byte(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Lit::Char(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ Lit::Int(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ Lit::Float(v0) => { ++ state.write_u8(5u8); ++ v0.hash(state); ++ } ++ Lit::Bool(v0) => { ++ state.write_u8(6u8); ++ v0.hash(state); ++ } ++ Lit::Verbatim(v0) => { ++ state.write_u8(7u8); ++ v0.to_string().hash(state); ++ } ++ } ++ } ++} ++impl Hash for LitBool { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.value.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Local { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.pat.hash(state); ++ self.init.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Macro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.path.hash(state); ++ self.delimiter.hash(state); ++ TokenStreamHelper(&self.tokens).hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for MacroDelimiter { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ MacroDelimiter::Paren(_) => { ++ state.write_u8(0u8); ++ } ++ MacroDelimiter::Brace(_) => { ++ state.write_u8(1u8); ++ } ++ MacroDelimiter::Bracket(_) => { ++ state.write_u8(2u8); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Meta { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Meta::Path(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Meta::List(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Meta::NameValue(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for MetaList { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.path.hash(state); ++ self.nested.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for MetaNameValue { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.path.hash(state); ++ self.lit.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for MethodTurbofish { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.args.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for NestedMeta { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ NestedMeta::Meta(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ NestedMeta::Lit(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ParenthesizedGenericArguments { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.inputs.hash(state); ++ self.output.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Pat { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Pat::Box(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Pat::Ident(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Pat::Lit(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Pat::Macro(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ Pat::Or(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ Pat::Path(v0) => { ++ state.write_u8(5u8); ++ v0.hash(state); ++ } ++ Pat::Range(v0) => { ++ state.write_u8(6u8); ++ v0.hash(state); ++ } ++ Pat::Reference(v0) => { ++ state.write_u8(7u8); ++ v0.hash(state); ++ } ++ Pat::Rest(v0) => { ++ state.write_u8(8u8); ++ v0.hash(state); ++ } ++ Pat::Slice(v0) => { ++ state.write_u8(9u8); ++ v0.hash(state); ++ } ++ Pat::Struct(v0) => { ++ state.write_u8(10u8); ++ v0.hash(state); ++ } ++ Pat::Tuple(v0) => { ++ state.write_u8(11u8); ++ v0.hash(state); ++ } ++ Pat::TupleStruct(v0) => { ++ state.write_u8(12u8); ++ v0.hash(state); ++ } ++ Pat::Type(v0) => { ++ state.write_u8(13u8); ++ v0.hash(state); ++ } ++ Pat::Verbatim(v0) => { ++ state.write_u8(14u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ Pat::Wild(v0) => { ++ state.write_u8(15u8); ++ v0.hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatBox { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.pat.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatIdent { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.by_ref.hash(state); ++ self.mutability.hash(state); ++ self.ident.hash(state); ++ self.subpat.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatLit { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mac.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatOr { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.leading_vert.hash(state); ++ self.cases.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatPath { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.qself.hash(state); ++ self.path.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatRange { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.lo.hash(state); ++ self.limits.hash(state); ++ self.hi.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatReference { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mutability.hash(state); ++ self.pat.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatRest { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatSlice { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.elems.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatStruct { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.path.hash(state); ++ self.fields.hash(state); ++ self.dot2_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatTuple { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.elems.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatTupleStruct { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.path.hash(state); ++ self.pat.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.pat.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatWild { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Path { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.leading_colon.hash(state); ++ self.segments.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for PathArguments { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ PathArguments::None => { ++ state.write_u8(0u8); ++ } ++ PathArguments::AngleBracketed(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ PathArguments::Parenthesized(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for PathSegment { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ self.arguments.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for PredicateEq { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lhs_ty.hash(state); ++ self.rhs_ty.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for PredicateLifetime { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lifetime.hash(state); ++ self.bounds.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for PredicateType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lifetimes.hash(state); ++ self.bounded_ty.hash(state); ++ self.bounds.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for QSelf { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ty.hash(state); ++ self.position.hash(state); ++ self.as_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for RangeLimits { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ RangeLimits::HalfOpen(_) => { ++ state.write_u8(0u8); ++ } ++ RangeLimits::Closed(_) => { ++ state.write_u8(1u8); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Receiver { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.reference.hash(state); ++ self.mutability.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ReturnType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ ReturnType::Default => { ++ state.write_u8(0u8); ++ } ++ ReturnType::Type(_, v1) => { ++ state.write_u8(1u8); ++ v1.hash(state); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Signature { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.constness.hash(state); ++ self.asyncness.hash(state); ++ self.unsafety.hash(state); ++ self.abi.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.inputs.hash(state); ++ self.variadic.hash(state); ++ self.output.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Stmt { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Stmt::Local(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Stmt::Item(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Stmt::Expr(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Stmt::Semi(v0, _) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TraitBound { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.paren_token.hash(state); ++ self.modifier.hash(state); ++ self.lifetimes.hash(state); ++ self.path.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TraitBoundModifier { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ TraitBoundModifier::None => { ++ state.write_u8(0u8); ++ } ++ TraitBoundModifier::Maybe(_) => { ++ state.write_u8(1u8); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for TraitItem { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ TraitItem::Const(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ TraitItem::Method(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ TraitItem::Type(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ TraitItem::Macro(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ TraitItem::Verbatim(v0) => { ++ state.write_u8(4u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for TraitItemConst { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ self.default.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for TraitItemMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mac.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for TraitItemMethod { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.sig.hash(state); ++ self.default.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for TraitItemType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.colon_token.hash(state); ++ self.bounds.hash(state); ++ self.default.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Type { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Type::Array(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Type::BareFn(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Type::Group(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Type::ImplTrait(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ Type::Infer(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ Type::Macro(v0) => { ++ state.write_u8(5u8); ++ v0.hash(state); ++ } ++ Type::Never(v0) => { ++ state.write_u8(6u8); ++ v0.hash(state); ++ } ++ Type::Paren(v0) => { ++ state.write_u8(7u8); ++ v0.hash(state); ++ } ++ Type::Path(v0) => { ++ state.write_u8(8u8); ++ v0.hash(state); ++ } ++ Type::Ptr(v0) => { ++ state.write_u8(9u8); ++ v0.hash(state); ++ } ++ Type::Reference(v0) => { ++ state.write_u8(10u8); ++ v0.hash(state); ++ } ++ Type::Slice(v0) => { ++ state.write_u8(11u8); ++ v0.hash(state); ++ } ++ Type::TraitObject(v0) => { ++ state.write_u8(12u8); ++ v0.hash(state); ++ } ++ Type::Tuple(v0) => { ++ state.write_u8(13u8); ++ v0.hash(state); ++ } ++ Type::Verbatim(v0) => { ++ state.write_u8(14u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeArray { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.elem.hash(state); ++ self.len.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeBareFn { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lifetimes.hash(state); ++ self.unsafety.hash(state); ++ self.abi.hash(state); ++ self.inputs.hash(state); ++ self.variadic.hash(state); ++ self.output.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeGroup { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.elem.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeImplTrait { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.bounds.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeInfer { ++ fn hash<H>(&self, _state: &mut H) ++ where ++ H: Hasher, ++ { ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.mac.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeNever { ++ fn hash<H>(&self, _state: &mut H) ++ where ++ H: Hasher, ++ { ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeParam { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.colon_token.hash(state); ++ self.bounds.hash(state); ++ self.eq_token.hash(state); ++ self.default.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeParamBound { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ TypeParamBound::Trait(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ TypeParamBound::Lifetime(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeParen { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.elem.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypePath { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.qself.hash(state); ++ self.path.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypePtr { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.const_token.hash(state); ++ self.mutability.hash(state); ++ self.elem.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeReference { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lifetime.hash(state); ++ self.mutability.hash(state); ++ self.elem.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeSlice { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.elem.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeTraitObject { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.dyn_token.hash(state); ++ self.bounds.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeTuple { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.elems.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for UnOp { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ UnOp::Deref(_) => { ++ state.write_u8(0u8); ++ } ++ UnOp::Not(_) => { ++ state.write_u8(1u8); ++ } ++ UnOp::Neg(_) => { ++ state.write_u8(2u8); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UseGlob { ++ fn hash<H>(&self, _state: &mut H) ++ where ++ H: Hasher, ++ { ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UseGroup { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.items.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UseName { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UsePath { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ self.tree.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UseRename { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ self.rename.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UseTree { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ UseTree::Path(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ UseTree::Name(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ UseTree::Rename(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ UseTree::Glob(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ UseTree::Group(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Variadic { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Variant { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.fields.hash(state); ++ self.discriminant.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for VisCrate { ++ fn hash<H>(&self, _state: &mut H) ++ where ++ H: Hasher, ++ { ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for VisPublic { ++ fn hash<H>(&self, _state: &mut H) ++ where ++ H: Hasher, ++ { ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for VisRestricted { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.in_token.hash(state); ++ self.path.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Visibility { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Visibility::Public(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Visibility::Crate(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Visibility::Restricted(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Visibility::Inherited => { ++ state.write_u8(3u8); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for WhereClause { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.predicates.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for WherePredicate { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ WherePredicate::Type(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ WherePredicate::Lifetime(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ WherePredicate::Eq(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ } ++ } ++} +diff --git a/third_party/rust/syn/src/gen/visit.rs b/third_party/rust/syn/src/gen/visit.rs +index b667f530c3..24d34b7480 100644 +--- a/third_party/rust/syn/src/gen/visit.rs ++++ b/third_party/rust/syn/src/gen/visit.rs +@@ -20,7 +20,6 @@ macro_rules! full { + unreachable!() + }; + } +-#[cfg(any(feature = "full", feature = "derive"))] + macro_rules! skip { + ($($tt:tt)*) => {}; + } +@@ -30,7 +29,7 @@ macro_rules! skip { + /// + /// [module documentation]: self + /// +-/// *This trait is available if Syn is built with the `"visit"` feature.* ++/// *This trait is available only if Syn is built with the `"visit"` feature.* + pub trait Visit<'ast> { + #[cfg(any(feature = "derive", feature = "full"))] + fn visit_abi(&mut self, i: &'ast Abi) { +@@ -434,35 +433,27 @@ pub trait Visit<'ast> { + fn visit_lifetime_def(&mut self, i: &'ast LifetimeDef) { + visit_lifetime_def(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit(&mut self, i: &'ast Lit) { + visit_lit(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_bool(&mut self, i: &'ast LitBool) { + visit_lit_bool(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_byte(&mut self, i: &'ast LitByte) { + visit_lit_byte(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_byte_str(&mut self, i: &'ast LitByteStr) { + visit_lit_byte_str(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_char(&mut self, i: &'ast LitChar) { + visit_lit_char(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_float(&mut self, i: &'ast LitFloat) { + visit_lit_float(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_int(&mut self, i: &'ast LitInt) { + visit_lit_int(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_str(&mut self, i: &'ast LitStr) { + visit_lit_str(self, i) + } +@@ -2537,7 +2528,6 @@ where + } + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit<'ast, V>(v: &mut V, node: &'ast Lit) + where + V: Visit<'ast> + ?Sized, +@@ -2569,7 +2559,6 @@ where + } + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_bool<'ast, V>(v: &mut V, node: &'ast LitBool) + where + V: Visit<'ast> + ?Sized, +@@ -2577,37 +2566,31 @@ where + skip!(node.value); + v.visit_span(&node.span); + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_byte<'ast, V>(v: &mut V, node: &'ast LitByte) + where + V: Visit<'ast> + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_byte_str<'ast, V>(v: &mut V, node: &'ast LitByteStr) + where + V: Visit<'ast> + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_char<'ast, V>(v: &mut V, node: &'ast LitChar) + where + V: Visit<'ast> + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_float<'ast, V>(v: &mut V, node: &'ast LitFloat) + where + V: Visit<'ast> + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_int<'ast, V>(v: &mut V, node: &'ast LitInt) + where + V: Visit<'ast> + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_str<'ast, V>(v: &mut V, node: &'ast LitStr) + where + V: Visit<'ast> + ?Sized, +diff --git a/third_party/rust/syn/src/gen/visit_mut.rs b/third_party/rust/syn/src/gen/visit_mut.rs +index 5cddb827c6..5ce11f0b2e 100644 +--- a/third_party/rust/syn/src/gen/visit_mut.rs ++++ b/third_party/rust/syn/src/gen/visit_mut.rs +@@ -20,7 +20,6 @@ macro_rules! full { + unreachable!() + }; + } +-#[cfg(any(feature = "full", feature = "derive"))] + macro_rules! skip { + ($($tt:tt)*) => {}; + } +@@ -31,7 +30,7 @@ macro_rules! skip { + /// + /// [module documentation]: self + /// +-/// *This trait is available if Syn is built with the `"visit-mut"` feature.* ++/// *This trait is available only if Syn is built with the `"visit-mut"` feature.* + pub trait VisitMut { + #[cfg(any(feature = "derive", feature = "full"))] + fn visit_abi_mut(&mut self, i: &mut Abi) { +@@ -438,35 +437,27 @@ pub trait VisitMut { + fn visit_lifetime_def_mut(&mut self, i: &mut LifetimeDef) { + visit_lifetime_def_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_mut(&mut self, i: &mut Lit) { + visit_lit_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_bool_mut(&mut self, i: &mut LitBool) { + visit_lit_bool_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_byte_mut(&mut self, i: &mut LitByte) { + visit_lit_byte_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_byte_str_mut(&mut self, i: &mut LitByteStr) { + visit_lit_byte_str_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_char_mut(&mut self, i: &mut LitChar) { + visit_lit_char_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_float_mut(&mut self, i: &mut LitFloat) { + visit_lit_float_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_int_mut(&mut self, i: &mut LitInt) { + visit_lit_int_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_str_mut(&mut self, i: &mut LitStr) { + visit_lit_str_mut(self, i) + } +@@ -2543,7 +2534,6 @@ where + } + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_mut<V>(v: &mut V, node: &mut Lit) + where + V: VisitMut + ?Sized, +@@ -2575,7 +2565,6 @@ where + } + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_bool_mut<V>(v: &mut V, node: &mut LitBool) + where + V: VisitMut + ?Sized, +@@ -2583,37 +2572,31 @@ where + skip!(node.value); + v.visit_span_mut(&mut node.span); + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_byte_mut<V>(v: &mut V, node: &mut LitByte) + where + V: VisitMut + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_byte_str_mut<V>(v: &mut V, node: &mut LitByteStr) + where + V: VisitMut + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_char_mut<V>(v: &mut V, node: &mut LitChar) + where + V: VisitMut + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_float_mut<V>(v: &mut V, node: &mut LitFloat) + where + V: VisitMut + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_int_mut<V>(v: &mut V, node: &mut LitInt) + where + V: VisitMut + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_str_mut<V>(v: &mut V, node: &mut LitStr) + where + V: VisitMut + ?Sized, +diff --git a/third_party/rust/syn/src/generics.rs b/third_party/rust/syn/src/generics.rs +index 95ab2e404a..05e8ef5cdf 100644 +--- a/third_party/rust/syn/src/generics.rs ++++ b/third_party/rust/syn/src/generics.rs +@@ -1,13 +1,16 @@ + use super::*; + use crate::punctuated::{Iter, IterMut, Punctuated}; ++#[cfg(all(feature = "printing", feature = "extra-traits"))] ++use std::fmt::{self, Debug}; ++#[cfg(all(feature = "printing", feature = "extra-traits"))] ++use std::hash::{Hash, Hasher}; + + ast_struct! { + /// Lifetimes and type parameters attached to a declaration of a function, + /// enum, trait, etc. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- #[derive(Default)] + pub struct Generics { + pub lt_token: Option<Token![<]>, + pub params: Punctuated<GenericParam, Token![,]>, +@@ -20,7 +23,7 @@ ast_enum_of_structs! { + /// A generic type parameter, lifetime, or const generic: `T: Into<String>`, + /// `'a: 'b`, `const LEN: usize`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Syntax tree enum +@@ -28,9 +31,6 @@ ast_enum_of_structs! { + /// This type is a [syntax tree enum]. + /// + /// [syntax tree enum]: enum.Expr.html#syntax-tree-enums +- // +- // TODO: change syntax-tree-enum link to an intra rustdoc link, currently +- // blocked on https://github.com/rust-lang/rust/issues/62833 + pub enum GenericParam { + /// A generic type parameter: `T: Into<String>`. + Type(TypeParam), +@@ -46,7 +46,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A generic type parameter: `T: Into<String>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeParam { + pub attrs: Vec<Attribute>, +@@ -61,7 +61,7 @@ ast_struct! { + ast_struct! { + /// A lifetime definition: `'a: 'b + 'c + 'd`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct LifetimeDef { + pub attrs: Vec<Attribute>, +@@ -74,7 +74,7 @@ ast_struct! { + ast_struct! { + /// A const generic parameter: `const LENGTH: usize`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ConstParam { + pub attrs: Vec<Attribute>, +@@ -87,6 +87,17 @@ ast_struct! { + } + } + ++impl Default for Generics { ++ fn default() -> Self { ++ Generics { ++ lt_token: None, ++ params: Punctuated::new(), ++ gt_token: None, ++ where_clause: None, ++ } ++ } ++} ++ + impl Generics { + /// Returns an + /// <code +@@ -280,29 +291,23 @@ impl<'a> Iterator for ConstParamsMut<'a> { + + /// Returned by `Generics::split_for_impl`. + /// +-/// *This type is available if Syn is built with the `"derive"` or `"full"` ++/// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature and the `"printing"` feature.* + #[cfg(feature = "printing")] +-#[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + pub struct ImplGenerics<'a>(&'a Generics); + + /// Returned by `Generics::split_for_impl`. + /// +-/// *This type is available if Syn is built with the `"derive"` or `"full"` ++/// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature and the `"printing"` feature.* + #[cfg(feature = "printing")] +-#[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + pub struct TypeGenerics<'a>(&'a Generics); + + /// Returned by `TypeGenerics::as_turbofish`. + /// +-/// *This type is available if Syn is built with the `"derive"` or `"full"` ++/// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature and the `"printing"` feature.* + #[cfg(feature = "printing")] +-#[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + pub struct Turbofish<'a>(&'a Generics); + + #[cfg(feature = "printing")] +@@ -314,9 +319,8 @@ impl Generics { + /// # use proc_macro2::{Span, Ident}; + /// # use quote::quote; + /// # +- /// # fn main() { +- /// # let generics: syn::Generics = Default::default(); +- /// # let name = Ident::new("MyType", Span::call_site()); ++ /// # let generics: syn::Generics = Default::default(); ++ /// # let name = Ident::new("MyType", Span::call_site()); + /// # + /// let (impl_generics, ty_generics, where_clause) = generics.split_for_impl(); + /// quote! { +@@ -324,11 +328,10 @@ impl Generics { + /// // ... + /// } + /// } +- /// # ; +- /// # } ++ /// # ; + /// ``` + /// +- /// *This method is available if Syn is built with the `"derive"` or ++ /// *This method is available only if Syn is built with the `"derive"` or + /// `"full"` feature and the `"printing"` feature.* + pub fn split_for_impl(&self) -> (ImplGenerics, TypeGenerics, Option<&WhereClause>) { + ( +@@ -339,11 +342,57 @@ impl Generics { + } + } + ++#[cfg(feature = "printing")] ++macro_rules! generics_wrapper_impls { ++ ($ty:ident) => { ++ #[cfg(feature = "clone-impls")] ++ impl<'a> Clone for $ty<'a> { ++ fn clone(&self) -> Self { ++ $ty(self.0) ++ } ++ } ++ ++ #[cfg(feature = "extra-traits")] ++ impl<'a> Debug for $ty<'a> { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ formatter ++ .debug_tuple(stringify!($ty)) ++ .field(self.0) ++ .finish() ++ } ++ } ++ ++ #[cfg(feature = "extra-traits")] ++ impl<'a> Eq for $ty<'a> {} ++ ++ #[cfg(feature = "extra-traits")] ++ impl<'a> PartialEq for $ty<'a> { ++ fn eq(&self, other: &Self) -> bool { ++ self.0 == other.0 ++ } ++ } ++ ++ #[cfg(feature = "extra-traits")] ++ impl<'a> Hash for $ty<'a> { ++ fn hash<H: Hasher>(&self, state: &mut H) { ++ self.0.hash(state); ++ } ++ } ++ }; ++} ++ ++#[cfg(feature = "printing")] ++generics_wrapper_impls!(ImplGenerics); ++#[cfg(feature = "printing")] ++generics_wrapper_impls!(TypeGenerics); ++#[cfg(feature = "printing")] ++generics_wrapper_impls!(Turbofish); ++ + #[cfg(feature = "printing")] + impl<'a> TypeGenerics<'a> { + /// Turn a type's generics like `<X, Y>` into a turbofish like `::<X, Y>`. + /// +- /// *This method is available if Syn is built with the `"derive"` or ++ /// *This method is available only if Syn is built with the `"derive"` or + /// `"full"` feature and the `"printing"` feature.* + pub fn as_turbofish(&self) -> Turbofish { + Turbofish(self.0) +@@ -353,9 +402,8 @@ impl<'a> TypeGenerics<'a> { + ast_struct! { + /// A set of bound lifetimes: `for<'a, 'b, 'c>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- #[derive(Default)] + pub struct BoundLifetimes { + pub for_token: Token![for], + pub lt_token: Token![<], +@@ -364,6 +412,17 @@ ast_struct! { + } + } + ++impl Default for BoundLifetimes { ++ fn default() -> Self { ++ BoundLifetimes { ++ for_token: Default::default(), ++ lt_token: Default::default(), ++ lifetimes: Punctuated::new(), ++ gt_token: Default::default(), ++ } ++ } ++} ++ + impl LifetimeDef { + pub fn new(lifetime: Lifetime) -> Self { + LifetimeDef { +@@ -391,7 +450,7 @@ impl From<Ident> for TypeParam { + ast_enum_of_structs! { + /// A trait or lifetime used as a bound on a type parameter. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum TypeParamBound { + Trait(TraitBound), +@@ -402,7 +461,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A trait used as a bound on a type parameter. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct TraitBound { + pub paren_token: Option<token::Paren>, +@@ -418,9 +477,8 @@ ast_enum! { + /// A modifier on a trait bound, currently only used for the `?` in + /// `?Sized`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- #[cfg_attr(feature = "clone-impls", derive(Copy))] + pub enum TraitBoundModifier { + None, + Maybe(Token![?]), +@@ -431,7 +489,7 @@ ast_struct! { + /// A `where` clause in a definition: `where T: Deserialize<'de>, D: + /// 'static`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct WhereClause { + pub where_token: Token![where], +@@ -442,7 +500,7 @@ ast_struct! { + ast_enum_of_structs! { + /// A single predicate in a `where` clause: `T: Deserialize<'de>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Syntax tree enum +@@ -450,9 +508,6 @@ ast_enum_of_structs! { + /// This type is a [syntax tree enum]. + /// + /// [syntax tree enum]: enum.Expr.html#syntax-tree-enums +- // +- // TODO: change syntax-tree-enum link to an intra rustdoc link, currently +- // blocked on https://github.com/rust-lang/rust/issues/62833 + pub enum WherePredicate { + /// A type predicate in a `where` clause: `for<'c> Foo<'c>: Trait<'c>`. + Type(PredicateType), +@@ -468,7 +523,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A type predicate in a `where` clause: `for<'c> Foo<'c>: Trait<'c>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct PredicateType { + /// Any lifetimes from a `for` binding +@@ -484,7 +539,7 @@ ast_struct! { + ast_struct! { + /// A lifetime predicate in a `where` clause: `'a: 'b + 'c`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct PredicateLifetime { + pub lifetime: Lifetime, +@@ -496,7 +551,7 @@ ast_struct! { + ast_struct! { + /// An equality predicate in a `where` clause (unsupported). + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct PredicateEq { + pub lhs_ty: Type, +@@ -521,7 +576,6 @@ pub mod parsing { + + let mut params = Punctuated::new(); + let mut allow_lifetime_param = true; +- let mut allow_type_param = true; + loop { + if input.peek(Token![>]) { + break; +@@ -534,7 +588,7 @@ pub mod parsing { + attrs, + ..input.parse()? + })); +- } else if allow_type_param && lookahead.peek(Ident) { ++ } else if lookahead.peek(Ident) { + allow_lifetime_param = false; + params.push_value(GenericParam::Type(TypeParam { + attrs, +@@ -542,7 +596,6 @@ pub mod parsing { + })); + } else if lookahead.peek(Token![const]) { + allow_lifetime_param = false; +- allow_type_param = false; + params.push_value(GenericParam::Const(ConstParam { + attrs, + ..input.parse()? +@@ -665,57 +718,53 @@ pub mod parsing { + + impl Parse for TypeParam { + fn parse(input: ParseStream) -> Result<Self> { +- let has_colon; +- let has_default; +- Ok(TypeParam { +- attrs: input.call(Attribute::parse_outer)?, +- ident: input.parse()?, +- colon_token: { +- if input.peek(Token![:]) { +- has_colon = true; +- Some(input.parse()?) +- } else { +- has_colon = false; +- None +- } +- }, +- bounds: { +- let mut bounds = Punctuated::new(); +- if has_colon { +- loop { +- if input.peek(Token![,]) +- || input.peek(Token![>]) +- || input.peek(Token![=]) +- { +- break; +- } +- let value = input.parse()?; +- bounds.push_value(value); +- if !input.peek(Token![+]) { +- break; +- } +- let punct = input.parse()?; +- bounds.push_punct(punct); +- } ++ let attrs = input.call(Attribute::parse_outer)?; ++ let ident: Ident = input.parse()?; ++ let colon_token: Option<Token![:]> = input.parse()?; ++ ++ let begin_bound = input.fork(); ++ let mut is_maybe_const = false; ++ let mut bounds = Punctuated::new(); ++ if colon_token.is_some() { ++ loop { ++ if input.peek(Token![,]) || input.peek(Token![>]) || input.peek(Token![=]) { ++ break; + } +- bounds +- }, +- eq_token: { +- if input.peek(Token![=]) { +- has_default = true; +- Some(input.parse()?) +- } else { +- has_default = false; +- None ++ if input.peek(Token![?]) && input.peek2(Token![const]) { ++ input.parse::<Token![?]>()?; ++ input.parse::<Token![const]>()?; ++ is_maybe_const = true; + } +- }, +- default: { +- if has_default { +- Some(input.parse()?) +- } else { +- None ++ let value: TypeParamBound = input.parse()?; ++ bounds.push_value(value); ++ if !input.peek(Token![+]) { ++ break; + } +- }, ++ let punct: Token![+] = input.parse()?; ++ bounds.push_punct(punct); ++ } ++ } ++ ++ let mut eq_token: Option<Token![=]> = input.parse()?; ++ let mut default = if eq_token.is_some() { ++ Some(input.parse::<Type>()?) ++ } else { ++ None ++ }; ++ ++ if is_maybe_const { ++ bounds.clear(); ++ eq_token = None; ++ default = Some(Type::Verbatim(verbatim::between(begin_bound, input))); ++ } ++ ++ Ok(TypeParam { ++ attrs, ++ ident, ++ colon_token, ++ bounds, ++ eq_token, ++ default, + }) + } + } +@@ -898,6 +947,8 @@ mod printing { + use super::*; + + use proc_macro2::TokenStream; ++ #[cfg(feature = "full")] ++ use proc_macro2::TokenTree; + use quote::{ToTokens, TokenStreamExt}; + + use crate::attr::FilterAttrs; +@@ -1080,9 +1131,25 @@ mod printing { + TokensOrDefault(&self.colon_token).to_tokens(tokens); + self.bounds.to_tokens(tokens); + } +- if self.default.is_some() { ++ if let Some(default) = &self.default { ++ #[cfg(feature = "full")] ++ { ++ if self.eq_token.is_none() { ++ if let Type::Verbatim(default) = default { ++ let mut iter = default.clone().into_iter(); ++ match (iter.next(), iter.next()) { ++ (Some(TokenTree::Punct(ref q)), Some(TokenTree::Ident(ref c))) ++ if q.as_char() == '?' && c == "const" => ++ { ++ return default.to_tokens(tokens); ++ } ++ _ => {} ++ } ++ } ++ } ++ } + TokensOrDefault(&self.eq_token).to_tokens(tokens); +- self.default.to_tokens(tokens); ++ default.to_tokens(tokens); + } + } + } +@@ -1117,9 +1184,9 @@ mod printing { + self.ident.to_tokens(tokens); + self.colon_token.to_tokens(tokens); + self.ty.to_tokens(tokens); +- if self.default.is_some() { ++ if let Some(default) = &self.default { + TokensOrDefault(&self.eq_token).to_tokens(tokens); +- self.default.to_tokens(tokens); ++ default.to_tokens(tokens); + } + } + } +diff --git a/third_party/rust/syn/src/item.rs b/third_party/rust/syn/src/item.rs +index ff4485ace9..0d8f7d3ddc 100644 +--- a/third_party/rust/syn/src/item.rs ++++ b/third_party/rust/syn/src/item.rs +@@ -1,17 +1,15 @@ + use super::*; +-use crate::derive::{Data, DeriveInput}; ++use crate::derive::{Data, DataEnum, DataStruct, DataUnion, DeriveInput}; + use crate::punctuated::Punctuated; + use proc_macro2::TokenStream; + +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; +-#[cfg(feature = "extra-traits")] +-use std::hash::{Hash, Hasher}; ++#[cfg(feature = "parsing")] ++use std::mem; + + ast_enum_of_structs! { + /// Things that can appear directly inside of a module or scope. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -21,7 +19,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum Item #manual_extra_traits { ++ pub enum Item { + /// A constant item: `const MAX: u16 = 65535`. + Const(ItemConst), + +@@ -83,7 +81,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A constant item: `const MAX: u16 = 65535`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemConst { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -100,7 +98,7 @@ ast_struct! { + ast_struct! { + /// An enum definition: `enum Foo<A, B> { A(A), B(B) }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemEnum { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -115,7 +113,7 @@ ast_struct! { + ast_struct! { + /// An `extern crate` item: `extern crate serde`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemExternCrate { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -131,7 +129,7 @@ ast_struct! { + /// A free-standing function: `fn process(n: usize) -> Result<()> { ... + /// }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemFn { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -143,7 +141,7 @@ ast_struct! { + ast_struct! { + /// A block of foreign items: `extern "C" { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemForeignMod { + pub attrs: Vec<Attribute>, + pub abi: Abi, +@@ -156,7 +154,7 @@ ast_struct! { + /// An impl block providing trait or associated items: `impl<A> Trait + /// for Data<A> { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemImpl { + pub attrs: Vec<Attribute>, + pub defaultness: Option<Token![default]>, +@@ -175,7 +173,7 @@ ast_struct! { + ast_struct! { + /// A macro invocation, which includes `macro_rules!` definitions. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemMacro { + pub attrs: Vec<Attribute>, + /// The `example` in `macro_rules! example { ... }`. +@@ -188,8 +186,8 @@ ast_struct! { + ast_struct! { + /// A 2.0-style declarative macro introduced by the `macro` keyword. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* +- pub struct ItemMacro2 #manual_extra_traits { ++ /// *This type is available only if Syn is built with the `"full"` feature.* ++ pub struct ItemMacro2 { + pub attrs: Vec<Attribute>, + pub vis: Visibility, + pub macro_token: Token![macro], +@@ -201,7 +199,7 @@ ast_struct! { + ast_struct! { + /// A module or module declaration: `mod m` or `mod m { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemMod { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -215,7 +213,7 @@ ast_struct! { + ast_struct! { + /// A static item: `static BIKE: Shed = Shed(42)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemStatic { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -233,7 +231,7 @@ ast_struct! { + ast_struct! { + /// A struct definition: `struct Foo<A> { x: A }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemStruct { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -248,7 +246,7 @@ ast_struct! { + ast_struct! { + /// A trait definition: `pub trait Iterator { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemTrait { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -267,7 +265,7 @@ ast_struct! { + ast_struct! { + /// A trait alias: `pub trait SharableIterator = Iterator + Sync`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemTraitAlias { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -283,7 +281,7 @@ ast_struct! { + ast_struct! { + /// A type alias: `type Result<T> = std::result::Result<T, MyError>`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemType { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -299,7 +297,7 @@ ast_struct! { + ast_struct! { + /// A union definition: `union Foo<A, B> { x: A, y: B }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemUnion { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -313,7 +311,7 @@ ast_struct! { + ast_struct! { + /// A use declaration: `use std::collections::HashMap`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemUse { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -324,145 +322,32 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Item {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Item { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (Item::Const(this), Item::Const(other)) => this == other, +- (Item::Enum(this), Item::Enum(other)) => this == other, +- (Item::ExternCrate(this), Item::ExternCrate(other)) => this == other, +- (Item::Fn(this), Item::Fn(other)) => this == other, +- (Item::ForeignMod(this), Item::ForeignMod(other)) => this == other, +- (Item::Impl(this), Item::Impl(other)) => this == other, +- (Item::Macro(this), Item::Macro(other)) => this == other, +- (Item::Macro2(this), Item::Macro2(other)) => this == other, +- (Item::Mod(this), Item::Mod(other)) => this == other, +- (Item::Static(this), Item::Static(other)) => this == other, +- (Item::Struct(this), Item::Struct(other)) => this == other, +- (Item::Trait(this), Item::Trait(other)) => this == other, +- (Item::TraitAlias(this), Item::TraitAlias(other)) => this == other, +- (Item::Type(this), Item::Type(other)) => this == other, +- (Item::Union(this), Item::Union(other)) => this == other, +- (Item::Use(this), Item::Use(other)) => this == other, +- (Item::Verbatim(this), Item::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Item { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { ++impl Item { ++ #[cfg(feature = "parsing")] ++ pub(crate) fn replace_attrs(&mut self, new: Vec<Attribute>) -> Vec<Attribute> { + match self { +- Item::Const(item) => { +- state.write_u8(0); +- item.hash(state); +- } +- Item::Enum(item) => { +- state.write_u8(1); +- item.hash(state); +- } +- Item::ExternCrate(item) => { +- state.write_u8(2); +- item.hash(state); +- } +- Item::Fn(item) => { +- state.write_u8(3); +- item.hash(state); +- } +- Item::ForeignMod(item) => { +- state.write_u8(4); +- item.hash(state); +- } +- Item::Impl(item) => { +- state.write_u8(5); +- item.hash(state); +- } +- Item::Macro(item) => { +- state.write_u8(6); +- item.hash(state); +- } +- Item::Macro2(item) => { +- state.write_u8(7); +- item.hash(state); +- } +- Item::Mod(item) => { +- state.write_u8(8); +- item.hash(state); +- } +- Item::Static(item) => { +- state.write_u8(9); +- item.hash(state); +- } +- Item::Struct(item) => { +- state.write_u8(10); +- item.hash(state); +- } +- Item::Trait(item) => { +- state.write_u8(11); +- item.hash(state); +- } +- Item::TraitAlias(item) => { +- state.write_u8(12); +- item.hash(state); +- } +- Item::Type(item) => { +- state.write_u8(13); +- item.hash(state); +- } +- Item::Union(item) => { +- state.write_u8(14); +- item.hash(state); +- } +- Item::Use(item) => { +- state.write_u8(15); +- item.hash(state); +- } +- Item::Verbatim(item) => { +- state.write_u8(16); +- TokenStreamHelper(item).hash(state); +- } ++ Item::ExternCrate(ItemExternCrate { attrs, .. }) ++ | Item::Use(ItemUse { attrs, .. }) ++ | Item::Static(ItemStatic { attrs, .. }) ++ | Item::Const(ItemConst { attrs, .. }) ++ | Item::Fn(ItemFn { attrs, .. }) ++ | Item::Mod(ItemMod { attrs, .. }) ++ | Item::ForeignMod(ItemForeignMod { attrs, .. }) ++ | Item::Type(ItemType { attrs, .. }) ++ | Item::Struct(ItemStruct { attrs, .. }) ++ | Item::Enum(ItemEnum { attrs, .. }) ++ | Item::Union(ItemUnion { attrs, .. }) ++ | Item::Trait(ItemTrait { attrs, .. }) ++ | Item::TraitAlias(ItemTraitAlias { attrs, .. }) ++ | Item::Impl(ItemImpl { attrs, .. }) ++ | Item::Macro(ItemMacro { attrs, .. }) ++ | Item::Macro2(ItemMacro2 { attrs, .. }) => mem::replace(attrs, new), ++ Item::Verbatim(_) => Vec::new(), + Item::__Nonexhaustive => unreachable!(), + } + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for ItemMacro2 {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for ItemMacro2 { +- fn eq(&self, other: &Self) -> bool { +- self.attrs == other.attrs +- && self.vis == other.vis +- && self.macro_token == other.macro_token +- && self.ident == other.ident +- && TokenStreamHelper(&self.rules) == TokenStreamHelper(&other.rules) +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for ItemMacro2 { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- self.attrs.hash(state); +- self.vis.hash(state); +- self.macro_token.hash(state); +- self.ident.hash(state); +- TokenStreamHelper(&self.rules).hash(state); +- } +-} +- + impl From<DeriveInput> for Item { + fn from(input: DeriveInput) -> Item { + match input.data { +@@ -496,10 +381,57 @@ impl From<DeriveInput> for Item { + } + } + ++impl From<ItemStruct> for DeriveInput { ++ fn from(input: ItemStruct) -> DeriveInput { ++ DeriveInput { ++ attrs: input.attrs, ++ vis: input.vis, ++ ident: input.ident, ++ generics: input.generics, ++ data: Data::Struct(DataStruct { ++ struct_token: input.struct_token, ++ fields: input.fields, ++ semi_token: input.semi_token, ++ }), ++ } ++ } ++} ++ ++impl From<ItemEnum> for DeriveInput { ++ fn from(input: ItemEnum) -> DeriveInput { ++ DeriveInput { ++ attrs: input.attrs, ++ vis: input.vis, ++ ident: input.ident, ++ generics: input.generics, ++ data: Data::Enum(DataEnum { ++ enum_token: input.enum_token, ++ brace_token: input.brace_token, ++ variants: input.variants, ++ }), ++ } ++ } ++} ++ ++impl From<ItemUnion> for DeriveInput { ++ fn from(input: ItemUnion) -> DeriveInput { ++ DeriveInput { ++ attrs: input.attrs, ++ vis: input.vis, ++ ident: input.ident, ++ generics: input.generics, ++ data: Data::Union(DataUnion { ++ union_token: input.union_token, ++ fields: input.fields, ++ }), ++ } ++ } ++} ++ + ast_enum_of_structs! { + /// A suffix of an import tree in a `use` item: `Type as Renamed` or `*`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -530,7 +462,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A path prefix of imports in a `use` item: `std::...`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct UsePath { + pub ident: Ident, + pub colon2_token: Token![::], +@@ -541,7 +473,7 @@ ast_struct! { + ast_struct! { + /// An identifier imported by a `use` item: `HashMap`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct UseName { + pub ident: Ident, + } +@@ -550,7 +482,7 @@ ast_struct! { + ast_struct! { + /// An renamed identifier imported by a `use` item: `HashMap as Map`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct UseRename { + pub ident: Ident, + pub as_token: Token![as], +@@ -561,7 +493,7 @@ ast_struct! { + ast_struct! { + /// A glob import in a `use` item: `*`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct UseGlob { + pub star_token: Token![*], + } +@@ -570,7 +502,7 @@ ast_struct! { + ast_struct! { + /// A braced group of imports in a `use` item: `{A, B, C}`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct UseGroup { + pub brace_token: token::Brace, + pub items: Punctuated<UseTree, Token![,]>, +@@ -580,7 +512,7 @@ ast_struct! { + ast_enum_of_structs! { + /// An item within an `extern` block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -590,7 +522,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum ForeignItem #manual_extra_traits { ++ pub enum ForeignItem { + /// A foreign function in an `extern` block. + Fn(ForeignItemFn), + +@@ -614,7 +546,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A foreign function in an `extern` block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ForeignItemFn { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -626,7 +558,7 @@ ast_struct! { + ast_struct! { + /// A foreign static item in an `extern` block: `static ext: u8`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ForeignItemStatic { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -642,7 +574,7 @@ ast_struct! { + ast_struct! { + /// A foreign type in an `extern` block: `type void`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ForeignItemType { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -655,7 +587,7 @@ ast_struct! { + ast_struct! { + /// A macro invocation within an extern block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ForeignItemMacro { + pub attrs: Vec<Attribute>, + pub mac: Macro, +@@ -663,61 +595,10 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for ForeignItem {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for ForeignItem { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (ForeignItem::Fn(this), ForeignItem::Fn(other)) => this == other, +- (ForeignItem::Static(this), ForeignItem::Static(other)) => this == other, +- (ForeignItem::Type(this), ForeignItem::Type(other)) => this == other, +- (ForeignItem::Macro(this), ForeignItem::Macro(other)) => this == other, +- (ForeignItem::Verbatim(this), ForeignItem::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for ForeignItem { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- match self { +- ForeignItem::Fn(item) => { +- state.write_u8(0); +- item.hash(state); +- } +- ForeignItem::Static(item) => { +- state.write_u8(1); +- item.hash(state); +- } +- ForeignItem::Type(item) => { +- state.write_u8(2); +- item.hash(state); +- } +- ForeignItem::Macro(item) => { +- state.write_u8(3); +- item.hash(state); +- } +- ForeignItem::Verbatim(item) => { +- state.write_u8(4); +- TokenStreamHelper(item).hash(state); +- } +- ForeignItem::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + ast_enum_of_structs! { + /// An item declaration within the definition of a trait. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -727,7 +608,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum TraitItem #manual_extra_traits { ++ pub enum TraitItem { + /// An associated constant within the definition of a trait. + Const(TraitItemConst), + +@@ -751,7 +632,7 @@ ast_enum_of_structs! { + ast_struct! { + /// An associated constant within the definition of a trait. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct TraitItemConst { + pub attrs: Vec<Attribute>, + pub const_token: Token![const], +@@ -766,7 +647,7 @@ ast_struct! { + ast_struct! { + /// A trait method within the definition of a trait. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct TraitItemMethod { + pub attrs: Vec<Attribute>, + pub sig: Signature, +@@ -778,7 +659,7 @@ ast_struct! { + ast_struct! { + /// An associated type within the definition of a trait. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct TraitItemType { + pub attrs: Vec<Attribute>, + pub type_token: Token![type], +@@ -794,7 +675,7 @@ ast_struct! { + ast_struct! { + /// A macro invocation within the definition of a trait. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct TraitItemMacro { + pub attrs: Vec<Attribute>, + pub mac: Macro, +@@ -802,61 +683,10 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for TraitItem {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for TraitItem { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (TraitItem::Const(this), TraitItem::Const(other)) => this == other, +- (TraitItem::Method(this), TraitItem::Method(other)) => this == other, +- (TraitItem::Type(this), TraitItem::Type(other)) => this == other, +- (TraitItem::Macro(this), TraitItem::Macro(other)) => this == other, +- (TraitItem::Verbatim(this), TraitItem::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for TraitItem { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- match self { +- TraitItem::Const(item) => { +- state.write_u8(0); +- item.hash(state); +- } +- TraitItem::Method(item) => { +- state.write_u8(1); +- item.hash(state); +- } +- TraitItem::Type(item) => { +- state.write_u8(2); +- item.hash(state); +- } +- TraitItem::Macro(item) => { +- state.write_u8(3); +- item.hash(state); +- } +- TraitItem::Verbatim(item) => { +- state.write_u8(4); +- TokenStreamHelper(item).hash(state); +- } +- TraitItem::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + ast_enum_of_structs! { + /// An item within an impl block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -866,7 +696,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum ImplItem #manual_extra_traits { ++ pub enum ImplItem { + /// An associated constant within an impl block. + Const(ImplItemConst), + +@@ -890,7 +720,7 @@ ast_enum_of_structs! { + ast_struct! { + /// An associated constant within an impl block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ImplItemConst { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -908,7 +738,7 @@ ast_struct! { + ast_struct! { + /// A method within an impl block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ImplItemMethod { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -921,7 +751,7 @@ ast_struct! { + ast_struct! { + /// An associated type within an impl block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ImplItemType { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -938,7 +768,7 @@ ast_struct! { + ast_struct! { + /// A macro invocation within an impl block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ImplItemMacro { + pub attrs: Vec<Attribute>, + pub mac: Macro, +@@ -946,62 +776,11 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for ImplItem {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for ImplItem { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (ImplItem::Const(this), ImplItem::Const(other)) => this == other, +- (ImplItem::Method(this), ImplItem::Method(other)) => this == other, +- (ImplItem::Type(this), ImplItem::Type(other)) => this == other, +- (ImplItem::Macro(this), ImplItem::Macro(other)) => this == other, +- (ImplItem::Verbatim(this), ImplItem::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for ImplItem { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- match self { +- ImplItem::Const(item) => { +- state.write_u8(0); +- item.hash(state); +- } +- ImplItem::Method(item) => { +- state.write_u8(1); +- item.hash(state); +- } +- ImplItem::Type(item) => { +- state.write_u8(2); +- item.hash(state); +- } +- ImplItem::Macro(item) => { +- state.write_u8(3); +- item.hash(state); +- } +- ImplItem::Verbatim(item) => { +- state.write_u8(4); +- TokenStreamHelper(item).hash(state); +- } +- ImplItem::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + ast_struct! { + /// A function signature in a trait or implementation: `unsafe fn + /// initialize(&self)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Signature { + pub constness: Option<Token![const]>, + pub asyncness: Option<Token![async]>, +@@ -1017,13 +796,34 @@ ast_struct! { + } + } + ++impl Signature { ++ /// A method's `self` receiver, such as `&self` or `self: Box<Self>`. ++ pub fn receiver(&self) -> Option<&FnArg> { ++ let arg = self.inputs.first()?; ++ match arg { ++ FnArg::Receiver(_) => Some(arg), ++ FnArg::Typed(PatType { pat, .. }) => { ++ if let Pat::Ident(PatIdent { ident, .. }) = &**pat { ++ if ident == "self" { ++ return Some(arg); ++ } ++ } ++ None ++ } ++ } ++ } ++} ++ + ast_enum_of_structs! { + /// An argument in a function signature: the `n: usize` in `fn f(n: usize)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub enum FnArg { + /// The `self` argument of an associated method, whether taken by value + /// or by reference. ++ /// ++ /// Note that `self` receivers with a specified type, such as `self: ++ /// Box<Self>`, are parsed as a `FnArg::Typed`. + Receiver(Receiver), + + /// A function argument accepted by pattern and type. +@@ -1035,7 +835,10 @@ ast_struct! { + /// The `self` argument of an associated method, whether taken by value + /// or by reference. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// Note that `self` receivers with a specified type, such as `self: ++ /// Box<Self>`, are parsed as a `FnArg::Typed`. ++ /// ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Receiver { + pub attrs: Vec<Attribute>, + pub reference: Option<(Token![&], Option<Lifetime>)>, +@@ -1056,7 +859,8 @@ pub mod parsing { + + use crate::ext::IdentExt; + use crate::parse::discouraged::Speculative; +- use crate::parse::{Parse, ParseStream, Result}; ++ use crate::parse::{Parse, ParseBuffer, ParseStream, Result}; ++ use crate::token::Brace; + use proc_macro2::{Delimiter, Group, Punct, Spacing, TokenTree}; + use std::iter::{self, FromIterator}; + +@@ -1064,18 +868,26 @@ pub mod parsing { + + impl Parse for Item { + fn parse(input: ParseStream) -> Result<Self> { ++ let begin = input.fork(); + let mut attrs = input.call(Attribute::parse_outer)?; + let ahead = input.fork(); + let vis: Visibility = ahead.parse()?; + + let lookahead = ahead.lookahead1(); +- let mut item = if lookahead.peek(Token![extern]) { ++ let mut item = if lookahead.peek(Token![fn]) || peek_signature(&ahead) { ++ let vis: Visibility = input.parse()?; ++ let sig = parse_signature(input)?; ++ if input.peek(Token![;]) { ++ input.parse::<Token![;]>()?; ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } else { ++ parse_rest_of_fn(input, Vec::new(), vis, sig).map(Item::Fn) ++ } ++ } else if lookahead.peek(Token![extern]) { + ahead.parse::<Token![extern]>()?; + let lookahead = ahead.lookahead1(); + if lookahead.peek(Token![crate]) { + input.parse().map(Item::ExternCrate) +- } else if lookahead.peek(Token![fn]) { +- input.parse().map(Item::Fn) + } else if lookahead.peek(token::Brace) { + input.parse().map(Item::ForeignMod) + } else if lookahead.peek(LitStr) { +@@ -1083,8 +895,6 @@ pub mod parsing { + let lookahead = ahead.lookahead1(); + if lookahead.peek(token::Brace) { + input.parse().map(Item::ForeignMod) +- } else if lookahead.peek(Token![fn]) { +- input.parse().map(Item::Fn) + } else { + Err(lookahead.error()) + } +@@ -1094,18 +904,61 @@ pub mod parsing { + } else if lookahead.peek(Token![use]) { + input.parse().map(Item::Use) + } else if lookahead.peek(Token![static]) { +- input.parse().map(Item::Static) ++ let vis = input.parse()?; ++ let static_token = input.parse()?; ++ let mutability = input.parse()?; ++ let ident = input.parse()?; ++ let colon_token = input.parse()?; ++ let ty = input.parse()?; ++ if input.peek(Token![;]) { ++ input.parse::<Token![;]>()?; ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(Item::Static(ItemStatic { ++ attrs: Vec::new(), ++ vis, ++ static_token, ++ mutability, ++ ident, ++ colon_token, ++ ty, ++ eq_token: input.parse()?, ++ expr: input.parse()?, ++ semi_token: input.parse()?, ++ })) ++ } + } else if lookahead.peek(Token![const]) { + ahead.parse::<Token![const]>()?; + let lookahead = ahead.lookahead1(); + if lookahead.peek(Ident) || lookahead.peek(Token![_]) { +- input.parse().map(Item::Const) +- } else if lookahead.peek(Token![unsafe]) +- || lookahead.peek(Token![async]) +- || lookahead.peek(Token![extern]) +- || lookahead.peek(Token![fn]) +- { +- input.parse().map(Item::Fn) ++ let vis = input.parse()?; ++ let const_token = input.parse()?; ++ let ident = { ++ let lookahead = input.lookahead1(); ++ if lookahead.peek(Ident) || lookahead.peek(Token![_]) { ++ input.call(Ident::parse_any)? ++ } else { ++ return Err(lookahead.error()); ++ } ++ }; ++ let colon_token = input.parse()?; ++ let ty = input.parse()?; ++ if input.peek(Token![;]) { ++ input.parse::<Token![;]>()?; ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(Item::Const(ItemConst { ++ attrs: Vec::new(), ++ vis, ++ const_token, ++ ident, ++ colon_token, ++ ty, ++ eq_token: input.parse()?, ++ expr: input.parse()?, ++ semi_token: input.parse()?, ++ })) ++ } + } else { + Err(lookahead.error()) + } +@@ -1117,21 +970,19 @@ pub mod parsing { + { + input.parse().map(Item::Trait) + } else if lookahead.peek(Token![impl]) { +- input.parse().map(Item::Impl) +- } else if lookahead.peek(Token![async]) +- || lookahead.peek(Token![extern]) +- || lookahead.peek(Token![fn]) +- { +- input.parse().map(Item::Fn) ++ let allow_const_impl = true; ++ if let Some(item) = parse_impl(input, allow_const_impl)? { ++ Ok(Item::Impl(item)) ++ } else { ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } + } else { + Err(lookahead.error()) + } +- } else if lookahead.peek(Token![async]) || lookahead.peek(Token![fn]) { +- input.parse().map(Item::Fn) + } else if lookahead.peek(Token![mod]) { + input.parse().map(Item::Mod) + } else if lookahead.peek(Token![type]) { +- input.parse().map(Item::Type) ++ parse_item_type(begin, input) + } else if lookahead.peek(existential) { + input.call(item_existential).map(Item::Verbatim) + } else if lookahead.peek(Token![struct]) { +@@ -1147,14 +998,18 @@ pub mod parsing { + } else if lookahead.peek(Token![impl]) + || lookahead.peek(Token![default]) && !ahead.peek2(Token![!]) + { +- input.parse().map(Item::Impl) ++ let allow_const_impl = true; ++ if let Some(item) = parse_impl(input, allow_const_impl)? { ++ Ok(Item::Impl(item)) ++ } else { ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } + } else if lookahead.peek(Token![macro]) { + input.parse().map(Item::Macro2) + } else if vis.is_inherited() + && (lookahead.peek(Ident) + || lookahead.peek(Token![self]) + || lookahead.peek(Token![super]) +- || lookahead.peek(Token![extern]) + || lookahead.peek(Token![crate]) + || lookahead.peek(Token![::])) + { +@@ -1163,32 +1018,64 @@ pub mod parsing { + Err(lookahead.error()) + }?; + +- { +- let item_attrs = match &mut item { +- Item::ExternCrate(item) => &mut item.attrs, +- Item::Use(item) => &mut item.attrs, +- Item::Static(item) => &mut item.attrs, +- Item::Const(item) => &mut item.attrs, +- Item::Fn(item) => &mut item.attrs, +- Item::Mod(item) => &mut item.attrs, +- Item::ForeignMod(item) => &mut item.attrs, +- Item::Type(item) => &mut item.attrs, +- Item::Struct(item) => &mut item.attrs, +- Item::Enum(item) => &mut item.attrs, +- Item::Union(item) => &mut item.attrs, +- Item::Trait(item) => &mut item.attrs, +- Item::TraitAlias(item) => &mut item.attrs, +- Item::Impl(item) => &mut item.attrs, +- Item::Macro(item) => &mut item.attrs, +- Item::Macro2(item) => &mut item.attrs, +- Item::Verbatim(_) => return Ok(item), +- Item::__Nonexhaustive => unreachable!(), +- }; +- attrs.extend(item_attrs.drain(..)); +- *item_attrs = attrs; ++ attrs.extend(item.replace_attrs(Vec::new())); ++ item.replace_attrs(attrs); ++ Ok(item) ++ } ++ } ++ ++ struct FlexibleItemType { ++ vis: Visibility, ++ defaultness: Option<Token![default]>, ++ type_token: Token![type], ++ ident: Ident, ++ generics: Generics, ++ colon_token: Option<Token![:]>, ++ bounds: Punctuated<TypeParamBound, Token![+]>, ++ ty: Option<(Token![=], Type)>, ++ semi_token: Token![;], ++ } ++ ++ impl Parse for FlexibleItemType { ++ fn parse(input: ParseStream) -> Result<Self> { ++ let vis: Visibility = input.parse()?; ++ let defaultness: Option<Token![default]> = input.parse()?; ++ let type_token: Token![type] = input.parse()?; ++ let ident: Ident = input.parse()?; ++ let mut generics: Generics = input.parse()?; ++ let colon_token: Option<Token![:]> = input.parse()?; ++ let mut bounds = Punctuated::new(); ++ if colon_token.is_some() { ++ loop { ++ bounds.push_value(input.parse::<TypeParamBound>()?); ++ if input.peek(Token![where]) || input.peek(Token![=]) || input.peek(Token![;]) { ++ break; ++ } ++ bounds.push_punct(input.parse::<Token![+]>()?); ++ if input.peek(Token![where]) || input.peek(Token![=]) || input.peek(Token![;]) { ++ break; ++ } ++ } + } ++ generics.where_clause = input.parse()?; ++ let ty = if let Some(eq_token) = input.parse()? { ++ Some((eq_token, input.parse::<Type>()?)) ++ } else { ++ None ++ }; ++ let semi_token: Token![;] = input.parse()?; + +- Ok(item) ++ Ok(FlexibleItemType { ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds, ++ ty, ++ semi_token, ++ }) + } + } + +@@ -1310,7 +1197,6 @@ pub mod parsing { + || lookahead.peek(Token![self]) + || lookahead.peek(Token![super]) + || lookahead.peek(Token![crate]) +- || lookahead.peek(Token![extern]) + { + let ident = input.call(Ident::parse_any)?; + if input.peek(Token![::]) { +@@ -1392,69 +1278,126 @@ pub mod parsing { + } + } + +- impl Parse for ItemFn { +- fn parse(input: ParseStream) -> Result<Self> { +- let outer_attrs = input.call(Attribute::parse_outer)?; +- let vis: Visibility = input.parse()?; +- let constness: Option<Token![const]> = input.parse()?; +- let asyncness: Option<Token![async]> = input.parse()?; +- let unsafety: Option<Token![unsafe]> = input.parse()?; +- let abi: Option<Abi> = input.parse()?; +- let fn_token: Token![fn] = input.parse()?; +- let ident: Ident = input.parse()?; +- let generics: Generics = input.parse()?; ++ fn pop_variadic(args: &mut Punctuated<FnArg, Token![,]>) -> Option<Variadic> { ++ let trailing_punct = args.trailing_punct(); + +- let content; +- let paren_token = parenthesized!(content in input); +- let inputs = content.parse_terminated(FnArg::parse)?; +- let variadic = inputs.last().as_ref().and_then(get_variadic); +- +- fn get_variadic(input: &&FnArg) -> Option<Variadic> { +- if let FnArg::Typed(PatType { ty, .. }) = input { +- if let Type::Verbatim(tokens) = &**ty { +- if let Ok(dots) = parse2(tokens.clone()) { +- return Some(Variadic { +- attrs: Vec::new(), +- dots, +- }); +- } +- } +- } +- None ++ let last = match args.last_mut()? { ++ FnArg::Typed(last) => last, ++ _ => return None, ++ }; ++ ++ let ty = match last.ty.as_ref() { ++ Type::Verbatim(ty) => ty, ++ _ => return None, ++ }; ++ ++ let mut variadic = Variadic { ++ attrs: Vec::new(), ++ dots: parse2(ty.clone()).ok()?, ++ }; ++ ++ if let Pat::Verbatim(pat) = last.pat.as_ref() { ++ if pat.to_string() == "..." && !trailing_punct { ++ variadic.attrs = mem::replace(&mut last.attrs, Vec::new()); ++ args.pop(); + } ++ } + +- let output: ReturnType = input.parse()?; +- let where_clause: Option<WhereClause> = input.parse()?; ++ Some(variadic) ++ } + +- let content; +- let brace_token = braced!(content in input); +- let inner_attrs = content.call(Attribute::parse_inner)?; +- let stmts = content.call(Block::parse_within)?; ++ fn variadic_to_tokens(dots: &Token![...]) -> TokenStream { ++ TokenStream::from_iter(vec![ ++ TokenTree::Punct({ ++ let mut dot = Punct::new('.', Spacing::Joint); ++ dot.set_span(dots.spans[0]); ++ dot ++ }), ++ TokenTree::Punct({ ++ let mut dot = Punct::new('.', Spacing::Joint); ++ dot.set_span(dots.spans[1]); ++ dot ++ }), ++ TokenTree::Punct({ ++ let mut dot = Punct::new('.', Spacing::Alone); ++ dot.set_span(dots.spans[2]); ++ dot ++ }), ++ ]) ++ } + +- Ok(ItemFn { +- attrs: private::attrs(outer_attrs, inner_attrs), +- vis, +- sig: Signature { +- constness, +- asyncness, +- unsafety, +- abi, +- fn_token, +- ident, +- paren_token, +- inputs, +- output, +- variadic, +- generics: Generics { +- where_clause, +- ..generics +- }, +- }, +- block: Box::new(Block { brace_token, stmts }), +- }) ++ fn peek_signature(input: ParseStream) -> bool { ++ let fork = input.fork(); ++ fork.parse::<Option<Token![const]>>().is_ok() ++ && fork.parse::<Option<Token![async]>>().is_ok() ++ && fork.parse::<Option<Token![unsafe]>>().is_ok() ++ && fork.parse::<Option<Abi>>().is_ok() ++ && fork.peek(Token![fn]) ++ } ++ ++ fn parse_signature(input: ParseStream) -> Result<Signature> { ++ let constness: Option<Token![const]> = input.parse()?; ++ let asyncness: Option<Token![async]> = input.parse()?; ++ let unsafety: Option<Token![unsafe]> = input.parse()?; ++ let abi: Option<Abi> = input.parse()?; ++ let fn_token: Token![fn] = input.parse()?; ++ let ident: Ident = input.parse()?; ++ let generics: Generics = input.parse()?; ++ ++ let content; ++ let paren_token = parenthesized!(content in input); ++ let mut inputs = parse_fn_args(&content)?; ++ let variadic = pop_variadic(&mut inputs); ++ ++ let output: ReturnType = input.parse()?; ++ let where_clause: Option<WhereClause> = input.parse()?; ++ ++ Ok(Signature { ++ constness, ++ asyncness, ++ unsafety, ++ abi, ++ fn_token, ++ ident, ++ paren_token, ++ inputs, ++ output, ++ variadic, ++ generics: Generics { ++ where_clause, ++ ..generics ++ }, ++ }) ++ } ++ ++ impl Parse for ItemFn { ++ fn parse(input: ParseStream) -> Result<Self> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; ++ let vis: Visibility = input.parse()?; ++ let sig = parse_signature(input)?; ++ parse_rest_of_fn(input, outer_attrs, vis, sig) + } + } + ++ fn parse_rest_of_fn( ++ input: ParseStream, ++ outer_attrs: Vec<Attribute>, ++ vis: Visibility, ++ sig: Signature, ++ ) -> Result<ItemFn> { ++ let content; ++ let brace_token = braced!(content in input); ++ let inner_attrs = content.call(Attribute::parse_inner)?; ++ let stmts = content.call(Block::parse_within)?; ++ ++ Ok(ItemFn { ++ attrs: private::attrs(outer_attrs, inner_attrs), ++ vis, ++ sig, ++ block: Box::new(Block { brace_token, stmts }), ++ }) ++ } ++ + impl Parse for FnArg { + fn parse(input: ParseStream) -> Result<Self> { + let attrs = input.call(Attribute::parse_outer)?; +@@ -1491,26 +1434,79 @@ pub mod parsing { + } + } + ++ fn parse_fn_args(input: ParseStream) -> Result<Punctuated<FnArg, Token![,]>> { ++ let mut args = Punctuated::new(); ++ let mut has_receiver = false; ++ ++ while !input.is_empty() { ++ let attrs = input.call(Attribute::parse_outer)?; ++ ++ let arg = if let Some(dots) = input.parse::<Option<Token![...]>>()? { ++ FnArg::Typed(PatType { ++ attrs, ++ pat: Box::new(Pat::Verbatim(variadic_to_tokens(&dots))), ++ colon_token: Token![:](dots.spans[0]), ++ ty: Box::new(Type::Verbatim(variadic_to_tokens(&dots))), ++ }) ++ } else { ++ let mut arg: FnArg = input.parse()?; ++ match &mut arg { ++ FnArg::Receiver(receiver) if has_receiver => { ++ return Err(Error::new( ++ receiver.self_token.span, ++ "unexpected second method receiver", ++ )); ++ } ++ FnArg::Receiver(receiver) if !args.is_empty() => { ++ return Err(Error::new( ++ receiver.self_token.span, ++ "unexpected method receiver", ++ )); ++ } ++ FnArg::Receiver(receiver) => { ++ has_receiver = true; ++ receiver.attrs = attrs; ++ } ++ FnArg::Typed(arg) => arg.attrs = attrs, ++ } ++ arg ++ }; ++ args.push_value(arg); ++ ++ if input.is_empty() { ++ break; ++ } ++ ++ let comma: Token![,] = input.parse()?; ++ args.push_punct(comma); ++ } ++ ++ Ok(args) ++ } ++ + fn fn_arg_typed(input: ParseStream) -> Result<PatType> { ++ // Hack to parse pre-2018 syntax in ++ // test/ui/rfc-2565-param-attrs/param-attrs-pretty.rs ++ // because the rest of the test case is valuable. ++ if input.peek(Ident) && input.peek2(Token![<]) { ++ let span = input.fork().parse::<Ident>()?.span(); ++ return Ok(PatType { ++ attrs: Vec::new(), ++ pat: Box::new(Pat::Wild(PatWild { ++ attrs: Vec::new(), ++ underscore_token: Token![_](span), ++ })), ++ colon_token: Token![:](span), ++ ty: input.parse()?, ++ }); ++ } ++ + Ok(PatType { + attrs: Vec::new(), +- pat: input.parse()?, ++ pat: Box::new(pat::parsing::multi_pat(input)?), + colon_token: input.parse()?, + ty: Box::new(match input.parse::<Option<Token![...]>>()? { +- Some(dot3) => { +- let args = vec![ +- TokenTree::Punct(Punct::new('.', Spacing::Joint)), +- TokenTree::Punct(Punct::new('.', Spacing::Joint)), +- TokenTree::Punct(Punct::new('.', Spacing::Alone)), +- ]; +- let tokens = TokenStream::from_iter(args.into_iter().zip(&dot3.spans).map( +- |(mut arg, span)| { +- arg.set_span(*span); +- arg +- }, +- )); +- Type::Verbatim(tokens) +- } ++ Some(dot3) => Type::Verbatim(variadic_to_tokens(&dot3)), + None => input.parse()?, + }), + }) +@@ -1581,22 +1577,60 @@ pub mod parsing { + + impl Parse for ForeignItem { + fn parse(input: ParseStream) -> Result<Self> { ++ let begin = input.fork(); + let mut attrs = input.call(Attribute::parse_outer)?; + let ahead = input.fork(); + let vis: Visibility = ahead.parse()?; + + let lookahead = ahead.lookahead1(); +- let mut item = if lookahead.peek(Token![fn]) { +- input.parse().map(ForeignItem::Fn) ++ let mut item = if lookahead.peek(Token![fn]) || peek_signature(&ahead) { ++ let vis: Visibility = input.parse()?; ++ let sig = parse_signature(input)?; ++ if input.peek(token::Brace) { ++ let content; ++ braced!(content in input); ++ content.call(Attribute::parse_inner)?; ++ content.call(Block::parse_within)?; ++ ++ Ok(ForeignItem::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(ForeignItem::Fn(ForeignItemFn { ++ attrs: Vec::new(), ++ vis, ++ sig, ++ semi_token: input.parse()?, ++ })) ++ } + } else if lookahead.peek(Token![static]) { +- input.parse().map(ForeignItem::Static) ++ let vis = input.parse()?; ++ let static_token = input.parse()?; ++ let mutability = input.parse()?; ++ let ident = input.parse()?; ++ let colon_token = input.parse()?; ++ let ty = input.parse()?; ++ if input.peek(Token![=]) { ++ input.parse::<Token![=]>()?; ++ input.parse::<Expr>()?; ++ input.parse::<Token![;]>()?; ++ Ok(ForeignItem::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(ForeignItem::Static(ForeignItemStatic { ++ attrs: Vec::new(), ++ vis, ++ static_token, ++ mutability, ++ ident, ++ colon_token, ++ ty, ++ semi_token: input.parse()?, ++ })) ++ } + } else if lookahead.peek(Token![type]) { +- input.parse().map(ForeignItem::Type) ++ parse_foreign_item_type(begin, input) + } else if vis.is_inherited() + && (lookahead.peek(Ident) + || lookahead.peek(Token![self]) + || lookahead.peek(Token![super]) +- || lookahead.peek(Token![extern]) + || lookahead.peek(Token![crate]) + || lookahead.peek(Token![::])) + { +@@ -1605,17 +1639,16 @@ pub mod parsing { + Err(lookahead.error()) + }?; + +- { +- let item_attrs = match &mut item { +- ForeignItem::Fn(item) => &mut item.attrs, +- ForeignItem::Static(item) => &mut item.attrs, +- ForeignItem::Type(item) => &mut item.attrs, +- ForeignItem::Macro(item) => &mut item.attrs, +- ForeignItem::Verbatim(_) | ForeignItem::__Nonexhaustive => unreachable!(), +- }; +- attrs.extend(item_attrs.drain(..)); +- *item_attrs = attrs; +- } ++ let item_attrs = match &mut item { ++ ForeignItem::Fn(item) => &mut item.attrs, ++ ForeignItem::Static(item) => &mut item.attrs, ++ ForeignItem::Type(item) => &mut item.attrs, ++ ForeignItem::Macro(item) => &mut item.attrs, ++ ForeignItem::Verbatim(_) => return Ok(item), ++ ForeignItem::__Nonexhaustive => unreachable!(), ++ }; ++ attrs.extend(item_attrs.drain(..)); ++ *item_attrs = attrs; + + Ok(item) + } +@@ -1625,55 +1658,12 @@ pub mod parsing { + fn parse(input: ParseStream) -> Result<Self> { + let attrs = input.call(Attribute::parse_outer)?; + let vis: Visibility = input.parse()?; +- let fn_token: Token![fn] = input.parse()?; +- let ident: Ident = input.parse()?; +- let generics: Generics = input.parse()?; +- +- let content; +- let paren_token = parenthesized!(content in input); +- let mut inputs = Punctuated::new(); +- let mut variadic = None; +- while !content.is_empty() { +- let attrs = content.call(Attribute::parse_outer)?; +- +- if let Some(dots) = content.parse()? { +- variadic = Some(Variadic { attrs, dots }); +- break; +- } +- +- let mut arg = content.call(fn_arg_typed)?; +- arg.attrs = attrs; +- inputs.push_value(FnArg::Typed(arg)); +- if content.is_empty() { +- break; +- } +- +- inputs.push_punct(content.parse()?); +- } +- +- let output: ReturnType = input.parse()?; +- let where_clause: Option<WhereClause> = input.parse()?; ++ let sig = parse_signature(input)?; + let semi_token: Token![;] = input.parse()?; +- + Ok(ForeignItemFn { + attrs, + vis, +- sig: Signature { +- constness: None, +- asyncness: None, +- unsafety: None, +- abi: None, +- fn_token, +- ident, +- paren_token, +- inputs, +- output, +- variadic, +- generics: Generics { +- where_clause, +- ..generics +- }, +- }, ++ sig, + semi_token, + }) + } +@@ -1706,6 +1696,37 @@ pub mod parsing { + } + } + ++ fn parse_foreign_item_type(begin: ParseBuffer, input: ParseStream) -> Result<ForeignItem> { ++ let FlexibleItemType { ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds: _, ++ ty, ++ semi_token, ++ } = input.parse()?; ++ ++ if defaultness.is_some() ++ || generics.lt_token.is_some() ++ || generics.where_clause.is_some() ++ || colon_token.is_some() ++ || ty.is_some() ++ { ++ Ok(ForeignItem::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(ForeignItem::Type(ForeignItemType { ++ attrs: Vec::new(), ++ vis, ++ type_token, ++ ident, ++ semi_token, ++ })) ++ } ++ } ++ + impl Parse for ForeignItemMacro { + fn parse(input: ParseStream) -> Result<Self> { + let attrs = input.call(Attribute::parse_outer)?; +@@ -1742,6 +1763,36 @@ pub mod parsing { + } + } + ++ fn parse_item_type(begin: ParseBuffer, input: ParseStream) -> Result<Item> { ++ let FlexibleItemType { ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds: _, ++ ty, ++ semi_token, ++ } = input.parse()?; ++ ++ if defaultness.is_some() || colon_token.is_some() || ty.is_none() { ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } else { ++ let (eq_token, ty) = ty.unwrap(); ++ Ok(Item::Type(ItemType { ++ attrs: Vec::new(), ++ vis, ++ type_token, ++ ident, ++ generics, ++ eq_token, ++ ty: Box::new(ty), ++ semi_token, ++ })) ++ } ++ } ++ + #[cfg(not(feature = "printing"))] + fn item_existential(input: ParseStream) -> Result<TokenStream> { + Err(input.error("existential type is not supported")) +@@ -1887,7 +1938,7 @@ pub mod parsing { + + impl Parse for ItemTrait { + fn parse(input: ParseStream) -> Result<Self> { +- let attrs = input.call(Attribute::parse_outer)?; ++ let outer_attrs = input.call(Attribute::parse_outer)?; + let vis: Visibility = input.parse()?; + let unsafety: Option<Token![unsafe]> = input.parse()?; + let auto_token: Option<Token![auto]> = input.parse()?; +@@ -1896,7 +1947,7 @@ pub mod parsing { + let generics: Generics = input.parse()?; + parse_rest_of_trait( + input, +- attrs, ++ outer_attrs, + vis, + unsafety, + auto_token, +@@ -1909,7 +1960,7 @@ pub mod parsing { + + fn parse_rest_of_trait( + input: ParseStream, +- attrs: Vec<Attribute>, ++ outer_attrs: Vec<Attribute>, + vis: Visibility, + unsafety: Option<Token![unsafe]>, + auto_token: Option<Token![auto]>, +@@ -1937,13 +1988,14 @@ pub mod parsing { + + let content; + let brace_token = braced!(content in input); ++ let inner_attrs = content.call(Attribute::parse_inner)?; + let mut items = Vec::new(); + while !content.is_empty() { + items.push(content.parse()?); + } + + Ok(ItemTrait { +- attrs, ++ attrs: private::attrs(outer_attrs, inner_attrs), + vis, + unsafety, + auto_token, +@@ -2014,14 +2066,19 @@ pub mod parsing { + + impl Parse for TraitItem { + fn parse(input: ParseStream) -> Result<Self> { ++ let begin = input.fork(); + let mut attrs = input.call(Attribute::parse_outer)?; ++ let vis: Visibility = input.parse()?; ++ let defaultness: Option<Token![default]> = input.parse()?; + let ahead = input.fork(); + + let lookahead = ahead.lookahead1(); +- let mut item = if lookahead.peek(Token![const]) { ++ let mut item = if lookahead.peek(Token![fn]) || peek_signature(&ahead) { ++ input.parse().map(TraitItem::Method) ++ } else if lookahead.peek(Token![const]) { + ahead.parse::<Token![const]>()?; + let lookahead = ahead.lookahead1(); +- if lookahead.peek(Ident) { ++ if lookahead.peek(Ident) || lookahead.peek(Token![_]) { + input.parse().map(TraitItem::Const) + } else if lookahead.peek(Token![async]) + || lookahead.peek(Token![unsafe]) +@@ -2032,18 +2089,11 @@ pub mod parsing { + } else { + Err(lookahead.error()) + } +- } else if lookahead.peek(Token![async]) +- || lookahead.peek(Token![unsafe]) +- || lookahead.peek(Token![extern]) +- || lookahead.peek(Token![fn]) +- { +- input.parse().map(TraitItem::Method) + } else if lookahead.peek(Token![type]) { +- input.parse().map(TraitItem::Type) ++ parse_trait_item_type(begin.fork(), input) + } else if lookahead.peek(Ident) + || lookahead.peek(Token![self]) + || lookahead.peek(Token![super]) +- || lookahead.peek(Token![extern]) + || lookahead.peek(Token![crate]) + || lookahead.peek(Token![::]) + { +@@ -2052,18 +2102,20 @@ pub mod parsing { + Err(lookahead.error()) + }?; + +- { +- let item_attrs = match &mut item { +- TraitItem::Const(item) => &mut item.attrs, +- TraitItem::Method(item) => &mut item.attrs, +- TraitItem::Type(item) => &mut item.attrs, +- TraitItem::Macro(item) => &mut item.attrs, +- TraitItem::Verbatim(_) | TraitItem::__Nonexhaustive => unreachable!(), +- }; +- attrs.extend(item_attrs.drain(..)); +- *item_attrs = attrs; ++ match (vis, defaultness) { ++ (Visibility::Inherited, None) => {} ++ _ => return Ok(TraitItem::Verbatim(verbatim::between(begin, input))), + } + ++ let item_attrs = match &mut item { ++ TraitItem::Const(item) => &mut item.attrs, ++ TraitItem::Method(item) => &mut item.attrs, ++ TraitItem::Type(item) => &mut item.attrs, ++ TraitItem::Macro(item) => &mut item.attrs, ++ TraitItem::Verbatim(_) | TraitItem::__Nonexhaustive => unreachable!(), ++ }; ++ attrs.extend(item_attrs.drain(..)); ++ *item_attrs = attrs; + Ok(item) + } + } +@@ -2073,7 +2125,14 @@ pub mod parsing { + Ok(TraitItemConst { + attrs: input.call(Attribute::parse_outer)?, + const_token: input.parse()?, +- ident: input.parse()?, ++ ident: { ++ let lookahead = input.lookahead1(); ++ if lookahead.peek(Ident) || lookahead.peek(Token![_]) { ++ input.call(Ident::parse_any)? ++ } else { ++ return Err(lookahead.error()); ++ } ++ }, + colon_token: input.parse()?, + ty: input.parse()?, + default: { +@@ -2093,20 +2152,7 @@ pub mod parsing { + impl Parse for TraitItemMethod { + fn parse(input: ParseStream) -> Result<Self> { + let outer_attrs = input.call(Attribute::parse_outer)?; +- let constness: Option<Token![const]> = input.parse()?; +- let asyncness: Option<Token![async]> = input.parse()?; +- let unsafety: Option<Token![unsafe]> = input.parse()?; +- let abi: Option<Abi> = input.parse()?; +- let fn_token: Token![fn] = input.parse()?; +- let ident: Ident = input.parse()?; +- let generics: Generics = input.parse()?; +- +- let content; +- let paren_token = parenthesized!(content in input); +- let inputs = content.parse_terminated(FnArg::parse)?; +- +- let output: ReturnType = input.parse()?; +- let where_clause: Option<WhereClause> = input.parse()?; ++ let sig = parse_signature(input)?; + + let lookahead = input.lookahead1(); + let (brace_token, inner_attrs, stmts, semi_token) = if lookahead.peek(token::Brace) { +@@ -2124,22 +2170,7 @@ pub mod parsing { + + Ok(TraitItemMethod { + attrs: private::attrs(outer_attrs, inner_attrs), +- sig: Signature { +- constness, +- asyncness, +- unsafety, +- abi, +- fn_token, +- ident, +- paren_token, +- inputs, +- output, +- variadic: None, +- generics: Generics { +- where_clause, +- ..generics +- }, +- }, ++ sig, + default: brace_token.map(|brace_token| Block { brace_token, stmts }), + semi_token, + }) +@@ -2188,6 +2219,35 @@ pub mod parsing { + } + } + ++ fn parse_trait_item_type(begin: ParseBuffer, input: ParseStream) -> Result<TraitItem> { ++ let FlexibleItemType { ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds, ++ ty, ++ semi_token, ++ } = input.parse()?; ++ ++ if defaultness.is_some() || vis.is_some() { ++ Ok(TraitItem::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(TraitItem::Type(TraitItemType { ++ attrs: Vec::new(), ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds, ++ default: ty, ++ semi_token, ++ })) ++ } ++ } ++ + impl Parse for TraitItemMacro { + fn parse(input: ParseStream) -> Result<Self> { + let attrs = input.call(Attribute::parse_outer)?; +@@ -2207,52 +2267,67 @@ pub mod parsing { + + impl Parse for ItemImpl { + fn parse(input: ParseStream) -> Result<Self> { +- let outer_attrs = input.call(Attribute::parse_outer)?; +- let defaultness: Option<Token![default]> = input.parse()?; +- let unsafety: Option<Token![unsafe]> = input.parse()?; +- let impl_token: Token![impl] = input.parse()?; +- +- let has_generics = input.peek(Token![<]) +- && (input.peek2(Token![>]) +- || input.peek2(Token![#]) +- || (input.peek2(Ident) || input.peek2(Lifetime)) +- && (input.peek3(Token![:]) +- || input.peek3(Token![,]) +- || input.peek3(Token![>]))); +- let generics: Generics = if has_generics { +- input.parse()? +- } else { +- Generics::default() +- }; +- +- let trait_ = { +- // TODO: optimize using advance_to +- let ahead = input.fork(); +- if ahead.parse::<Option<Token![!]>>().is_ok() +- && ahead.parse::<Path>().is_ok() +- && ahead.parse::<Token![for]>().is_ok() +- { +- let polarity: Option<Token![!]> = input.parse()?; +- let path: Path = input.parse()?; +- let for_token: Token![for] = input.parse()?; +- Some((polarity, path, for_token)) +- } else { +- None +- } +- }; +- let self_ty: Type = input.parse()?; +- let where_clause: Option<WhereClause> = input.parse()?; ++ let allow_const_impl = false; ++ parse_impl(input, allow_const_impl).map(Option::unwrap) ++ } ++ } ++ ++ fn parse_impl(input: ParseStream, allow_const_impl: bool) -> Result<Option<ItemImpl>> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; ++ let defaultness: Option<Token![default]> = input.parse()?; ++ let unsafety: Option<Token![unsafe]> = input.parse()?; ++ let impl_token: Token![impl] = input.parse()?; ++ ++ let has_generics = input.peek(Token![<]) ++ && (input.peek2(Token![>]) ++ || input.peek2(Token![#]) ++ || (input.peek2(Ident) || input.peek2(Lifetime)) ++ && (input.peek3(Token![:]) ++ || input.peek3(Token![,]) ++ || input.peek3(Token![>])) ++ || input.peek2(Token![const])); ++ let generics: Generics = if has_generics { ++ input.parse()? ++ } else { ++ Generics::default() ++ }; + +- let content; +- let brace_token = braced!(content in input); +- let inner_attrs = content.call(Attribute::parse_inner)?; ++ let is_const_impl = allow_const_impl ++ && (input.peek(Token![const]) || input.peek(Token![?]) && input.peek2(Token![const])); ++ if is_const_impl { ++ input.parse::<Option<Token![?]>>()?; ++ input.parse::<Token![const]>()?; ++ } + +- let mut items = Vec::new(); +- while !content.is_empty() { +- items.push(content.parse()?); ++ let trait_ = (|| -> Option<_> { ++ let ahead = input.fork(); ++ let polarity: Option<Token![!]> = ahead.parse().ok()?; ++ let mut path: Path = ahead.parse().ok()?; ++ if path.segments.last().unwrap().arguments.is_empty() && ahead.peek(token::Paren) { ++ let parenthesized = PathArguments::Parenthesized(ahead.parse().ok()?); ++ path.segments.last_mut().unwrap().arguments = parenthesized; + } ++ let for_token: Token![for] = ahead.parse().ok()?; ++ input.advance_to(&ahead); ++ Some((polarity, path, for_token)) ++ })(); ++ ++ let self_ty: Type = input.parse()?; ++ let where_clause: Option<WhereClause> = input.parse()?; ++ ++ let content; ++ let brace_token = braced!(content in input); ++ let inner_attrs = content.call(Attribute::parse_inner)?; ++ ++ let mut items = Vec::new(); ++ while !content.is_empty() { ++ items.push(content.parse()?); ++ } + +- Ok(ItemImpl { ++ if is_const_impl { ++ Ok(None) ++ } else { ++ Ok(Some(ItemImpl { + attrs: private::attrs(outer_attrs, inner_attrs), + defaultness, + unsafety, +@@ -2265,12 +2340,13 @@ pub mod parsing { + self_ty: Box::new(self_ty), + brace_token, + items, +- }) ++ })) + } + } + + impl Parse for ImplItem { + fn parse(input: ParseStream) -> Result<Self> { ++ let begin = input.fork(); + let mut attrs = input.call(Attribute::parse_outer)?; + let ahead = input.fork(); + let vis: Visibility = ahead.parse()?; +@@ -2284,28 +2360,38 @@ pub mod parsing { + None + }; + +- let mut item = if lookahead.peek(Token![const]) { +- ahead.parse::<Token![const]>()?; ++ let mut item = if lookahead.peek(Token![fn]) || peek_signature(&ahead) { ++ input.parse().map(ImplItem::Method) ++ } else if lookahead.peek(Token![const]) { ++ let const_token: Token![const] = ahead.parse()?; + let lookahead = ahead.lookahead1(); +- if lookahead.peek(Ident) { +- input.parse().map(ImplItem::Const) +- } else if lookahead.peek(Token![unsafe]) +- || lookahead.peek(Token![async]) +- || lookahead.peek(Token![extern]) +- || lookahead.peek(Token![fn]) +- { +- input.parse().map(ImplItem::Method) ++ if lookahead.peek(Ident) || lookahead.peek(Token![_]) { ++ input.advance_to(&ahead); ++ let ident: Ident = input.call(Ident::parse_any)?; ++ let colon_token: Token![:] = input.parse()?; ++ let ty: Type = input.parse()?; ++ if let Some(eq_token) = input.parse()? { ++ return Ok(ImplItem::Const(ImplItemConst { ++ attrs, ++ vis, ++ defaultness, ++ const_token, ++ ident, ++ colon_token, ++ ty, ++ eq_token, ++ expr: input.parse()?, ++ semi_token: input.parse()?, ++ })); ++ } else { ++ input.parse::<Token![;]>()?; ++ return Ok(ImplItem::Verbatim(verbatim::between(begin, input))); ++ } + } else { + Err(lookahead.error()) + } +- } else if lookahead.peek(Token![unsafe]) +- || lookahead.peek(Token![async]) +- || lookahead.peek(Token![extern]) +- || lookahead.peek(Token![fn]) +- { +- input.parse().map(ImplItem::Method) + } else if lookahead.peek(Token![type]) { +- input.parse().map(ImplItem::Type) ++ parse_impl_item_type(begin, input) + } else if vis.is_inherited() && defaultness.is_none() && lookahead.peek(existential) { + input.call(item_existential).map(ImplItem::Verbatim) + } else if vis.is_inherited() +@@ -2313,7 +2399,6 @@ pub mod parsing { + && (lookahead.peek(Ident) + || lookahead.peek(Token![self]) + || lookahead.peek(Token![super]) +- || lookahead.peek(Token![extern]) + || lookahead.peek(Token![crate]) + || lookahead.peek(Token![::])) + { +@@ -2346,7 +2431,14 @@ pub mod parsing { + vis: input.parse()?, + defaultness: input.parse()?, + const_token: input.parse()?, +- ident: input.parse()?, ++ ident: { ++ let lookahead = input.lookahead1(); ++ if lookahead.peek(Ident) || lookahead.peek(Token![_]) { ++ input.call(Ident::parse_any)? ++ } else { ++ return Err(lookahead.error()); ++ } ++ }, + colon_token: input.parse()?, + ty: input.parse()?, + eq_token: input.parse()?, +@@ -2358,50 +2450,39 @@ pub mod parsing { + + impl Parse for ImplItemMethod { + fn parse(input: ParseStream) -> Result<Self> { +- let outer_attrs = input.call(Attribute::parse_outer)?; ++ let mut attrs = input.call(Attribute::parse_outer)?; + let vis: Visibility = input.parse()?; + let defaultness: Option<Token![default]> = input.parse()?; +- let constness: Option<Token![const]> = input.parse()?; +- let asyncness: Option<Token![async]> = input.parse()?; +- let unsafety: Option<Token![unsafe]> = input.parse()?; +- let abi: Option<Abi> = input.parse()?; +- let fn_token: Token![fn] = input.parse()?; +- let ident: Ident = input.parse()?; +- let generics: Generics = input.parse()?; +- +- let content; +- let paren_token = parenthesized!(content in input); +- let inputs = content.parse_terminated(FnArg::parse)?; +- +- let output: ReturnType = input.parse()?; +- let where_clause: Option<WhereClause> = input.parse()?; +- +- let content; +- let brace_token = braced!(content in input); +- let inner_attrs = content.call(Attribute::parse_inner)?; +- let stmts = content.call(Block::parse_within)?; ++ let sig = parse_signature(input)?; ++ ++ let block = if let Some(semi) = input.parse::<Option<Token![;]>>()? { ++ // Accept methods without a body in an impl block because ++ // rustc's *parser* does not reject them (the compilation error ++ // is emitted later than parsing) and it can be useful for macro ++ // DSLs. ++ let mut punct = Punct::new(';', Spacing::Alone); ++ punct.set_span(semi.span); ++ let tokens = TokenStream::from_iter(vec![TokenTree::Punct(punct)]); ++ Block { ++ brace_token: Brace::default(), ++ stmts: vec![Stmt::Item(Item::Verbatim(tokens))], ++ } ++ } else { ++ let content; ++ let brace_token = braced!(content in input); ++ attrs.extend(content.call(Attribute::parse_inner)?); ++ Block { ++ brace_token, ++ stmts: content.call(Block::parse_within)?, ++ } ++ }; + + Ok(ImplItemMethod { +- attrs: private::attrs(outer_attrs, inner_attrs), ++ attrs, + vis, + defaultness, +- sig: Signature { +- constness, +- asyncness, +- unsafety, +- abi, +- fn_token, +- ident, +- paren_token, +- inputs, +- output, +- variadic: None, +- generics: Generics { +- where_clause, +- ..generics +- }, +- }, +- block: Block { brace_token, stmts }, ++ sig, ++ block, + }) + } + } +@@ -2426,6 +2507,37 @@ pub mod parsing { + } + } + ++ fn parse_impl_item_type(begin: ParseBuffer, input: ParseStream) -> Result<ImplItem> { ++ let FlexibleItemType { ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds: _, ++ ty, ++ semi_token, ++ } = input.parse()?; ++ ++ if colon_token.is_some() || ty.is_none() { ++ Ok(ImplItem::Verbatim(verbatim::between(begin, input))) ++ } else { ++ let (eq_token, ty) = ty.unwrap(); ++ Ok(ImplItem::Type(ImplItemType { ++ attrs: Vec::new(), ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ eq_token, ++ ty, ++ semi_token, ++ })) ++ } ++ } ++ + impl Parse for ImplItemMacro { + fn parse(input: ParseStream) -> Result<Self> { + let attrs = input.call(Attribute::parse_outer)?; +@@ -2471,6 +2583,7 @@ mod printing { + + use crate::attr::FilterAttrs; + use crate::print::TokensOrDefault; ++ use crate::punctuated::Pair; + + impl ToTokens for ItemExternCrate { + fn to_tokens(&self, tokens: &mut TokenStream) { +@@ -2835,6 +2948,14 @@ mod printing { + self.vis.to_tokens(tokens); + self.defaultness.to_tokens(tokens); + self.sig.to_tokens(tokens); ++ if self.block.stmts.len() == 1 { ++ if let Stmt::Item(Item::Verbatim(verbatim)) = &self.block.stmts[0] { ++ if verbatim.to_string() == ";" { ++ verbatim.to_tokens(tokens); ++ return; ++ } ++ } ++ } + self.block.brace_token.surround(tokens, |tokens| { + tokens.append_all(self.attrs.inner()); + tokens.append_all(&self.block.stmts); +@@ -2905,6 +3026,33 @@ mod printing { + } + } + ++ fn maybe_variadic_to_tokens(arg: &FnArg, tokens: &mut TokenStream) -> bool { ++ let arg = match arg { ++ FnArg::Typed(arg) => arg, ++ FnArg::Receiver(receiver) => { ++ receiver.to_tokens(tokens); ++ return false; ++ } ++ }; ++ ++ match arg.ty.as_ref() { ++ Type::Verbatim(ty) if ty.to_string() == "..." => { ++ match arg.pat.as_ref() { ++ Pat::Verbatim(pat) if pat.to_string() == "..." => { ++ tokens.append_all(arg.attrs.outer()); ++ pat.to_tokens(tokens); ++ } ++ _ => arg.to_tokens(tokens), ++ } ++ true ++ } ++ _ => { ++ arg.to_tokens(tokens); ++ false ++ } ++ } ++ } ++ + impl ToTokens for Signature { + fn to_tokens(&self, tokens: &mut TokenStream) { + self.constness.to_tokens(tokens); +@@ -2915,11 +3063,24 @@ mod printing { + self.ident.to_tokens(tokens); + self.generics.to_tokens(tokens); + self.paren_token.surround(tokens, |tokens| { +- self.inputs.to_tokens(tokens); +- if self.variadic.is_some() && !self.inputs.empty_or_trailing() { +- <Token![,]>::default().to_tokens(tokens); ++ let mut last_is_variadic = false; ++ for input in self.inputs.pairs() { ++ match input { ++ Pair::Punctuated(input, comma) => { ++ maybe_variadic_to_tokens(input, tokens); ++ comma.to_tokens(tokens); ++ } ++ Pair::End(input) => { ++ last_is_variadic = maybe_variadic_to_tokens(input, tokens); ++ } ++ } ++ } ++ if self.variadic.is_some() && !last_is_variadic { ++ if !self.inputs.empty_or_trailing() { ++ <Token![,]>::default().to_tokens(tokens); ++ } ++ self.variadic.to_tokens(tokens); + } +- self.variadic.to_tokens(tokens); + }); + self.output.to_tokens(tokens); + self.generics.where_clause.to_tokens(tokens); +diff --git a/third_party/rust/syn/src/keyword.rs b/third_party/rust/syn/src/keyword.rs +deleted file mode 100644 +index e69de29bb2..0000000000 +diff --git a/third_party/rust/syn/src/lib.rs b/third_party/rust/syn/src/lib.rs +index c8ada7e638..3da506731e 100644 +--- a/third_party/rust/syn/src/lib.rs ++++ b/third_party/rust/syn/src/lib.rs +@@ -1,3 +1,11 @@ ++//! [![github]](https://github.com/dtolnay/syn) [![crates-io]](https://crates.io/crates/syn) [![docs-rs]](https://docs.rs/syn) ++//! ++//! [github]: https://img.shields.io/badge/github-8da0cb?style=for-the-badge&labelColor=555555&logo=github ++//! [crates-io]: https://img.shields.io/badge/crates.io-fc8d62?style=for-the-badge&labelColor=555555&logo=rust ++//! [docs-rs]: https://img.shields.io/badge/docs.rs-66c2a5?style=for-the-badge&labelColor=555555&logoColor=white&logo=data:image/svg+xml;base64,PHN2ZyByb2xlPSJpbWciIHhtbG5zPSJodHRwOi8vd3d3LnczLm9yZy8yMDAwL3N2ZyIgdmlld0JveD0iMCAwIDUxMiA1MTIiPjxwYXRoIGZpbGw9IiNmNWY1ZjUiIGQ9Ik00ODguNiAyNTAuMkwzOTIgMjE0VjEwNS41YzAtMTUtOS4zLTI4LjQtMjMuNC0zMy43bC0xMDAtMzcuNWMtOC4xLTMuMS0xNy4xLTMuMS0yNS4zIDBsLTEwMCAzNy41Yy0xNC4xIDUuMy0yMy40IDE4LjctMjMuNCAzMy43VjIxNGwtOTYuNiAzNi4yQzkuMyAyNTUuNSAwIDI2OC45IDAgMjgzLjlWMzk0YzAgMTMuNiA3LjcgMjYuMSAxOS45IDMyLjJsMTAwIDUwYzEwLjEgNS4xIDIyLjEgNS4xIDMyLjIgMGwxMDMuOS01MiAxMDMuOSA1MmMxMC4xIDUuMSAyMi4xIDUuMSAzMi4yIDBsMTAwLTUwYzEyLjItNi4xIDE5LjktMTguNiAxOS45LTMyLjJWMjgzLjljMC0xNS05LjMtMjguNC0yMy40LTMzLjd6TTM1OCAyMTQuOGwtODUgMzEuOXYtNjguMmw4NS0zN3Y3My4zek0xNTQgMTA0LjFsMTAyLTM4LjIgMTAyIDM4LjJ2LjZsLTEwMiA0MS40LTEwMi00MS40di0uNnptODQgMjkxLjFsLTg1IDQyLjV2LTc5LjFsODUtMzguOHY3NS40em0wLTExMmwtMTAyIDQxLjQtMTAyLTQxLjR2LS42bDEwMi0zOC4yIDEwMiAzOC4ydi42em0yNDAgMTEybC04NSA0Mi41di03OS4xbDg1LTM4Ljh2NzUuNHptMC0xMTJsLTEwMiA0MS40LTEwMi00MS40di0uNmwxMDItMzguMiAxMDIgMzguMnYuNnoiPjwvcGF0aD48L3N2Zz4K ++//! ++//! <br> ++//! + //! Syn is a parsing library for parsing a stream of Rust tokens into a syntax + //! tree of Rust source code. + //! +@@ -62,8 +70,8 @@ + //! ``` + //! + //! ``` +-//! extern crate proc_macro; +-//! ++//! # extern crate proc_macro; ++//! # + //! use proc_macro::TokenStream; + //! use quote::quote; + //! use syn::{parse_macro_input, DeriveInput}; +@@ -242,35 +250,48 @@ + //! dynamic library libproc_macro from rustc toolchain. + + // Syn types in rustdoc of other crates get linked to here. +-#![doc(html_root_url = "https://docs.rs/syn/1.0.5")] ++#![doc(html_root_url = "https://docs.rs/syn/1.0.40")] + #![deny(clippy::all, clippy::pedantic)] + // Ignored clippy lints. + #![allow( +- clippy::block_in_if_condition_stmt, ++ clippy::blocks_in_if_conditions, + clippy::cognitive_complexity, + clippy::doc_markdown, + clippy::eval_order_dependence, + clippy::inherent_to_string, + clippy::large_enum_variant, ++ clippy::manual_non_exhaustive, ++ clippy::match_like_matches_macro, ++ clippy::match_on_vec_items, ++ clippy::needless_doctest_main, + clippy::needless_pass_by_value, + clippy::never_loop, + clippy::suspicious_op_assign_impl, + clippy::too_many_arguments, +- clippy::trivially_copy_pass_by_ref ++ clippy::trivially_copy_pass_by_ref, ++ clippy::unnecessary_unwrap + )] + // Ignored clippy_pedantic lints. + #![allow( + clippy::cast_possible_truncation, ++ clippy::default_trait_access, + clippy::empty_enum, ++ clippy::expl_impl_clone_on_copy, + clippy::if_not_else, + clippy::items_after_statements, ++ clippy::match_same_arms, ++ clippy::missing_errors_doc, + clippy::module_name_repetitions, ++ clippy::must_use_candidate, ++ clippy::option_if_let_else, + clippy::shadow_unrelated, + clippy::similar_names, + clippy::single_match_else, ++ clippy::too_many_lines, + clippy::unseparated_literal_suffix, + clippy::use_self, +- clippy::used_underscore_binding ++ clippy::used_underscore_binding, ++ clippy::wildcard_imports + )] + + #[cfg(all( +@@ -284,7 +305,6 @@ extern crate unicode_xid; + #[cfg(feature = "printing")] + extern crate quote; + +-#[cfg(any(feature = "full", feature = "derive"))] + #[macro_use] + mod macros; + +@@ -307,7 +327,6 @@ pub use crate::attr::{ + AttrStyle, Attribute, AttributeArgs, Meta, MetaList, MetaNameValue, NestedMeta, + }; + +-#[cfg(any(feature = "full", feature = "derive"))] + mod bigint; + + #[cfg(any(feature = "full", feature = "derive"))] +@@ -364,9 +383,7 @@ pub use crate::file::File; + mod lifetime; + pub use crate::lifetime::Lifetime; + +-#[cfg(any(feature = "full", feature = "derive"))] + mod lit; +-#[cfg(any(feature = "full", feature = "derive"))] + pub use crate::lit::{ + Lit, LitBool, LitByte, LitByteStr, LitChar, LitFloat, LitInt, LitStr, StrStyle, + }; +@@ -441,6 +458,9 @@ pub mod parse_macro_input; + #[cfg(all(feature = "parsing", feature = "printing"))] + pub mod spanned; + ++#[cfg(all(feature = "parsing", feature = "full"))] ++mod whitespace; ++ + mod gen { + /// Syntax tree traversal to walk a shared borrow of a syntax tree. + /// +@@ -482,7 +502,7 @@ mod gen { + /// /* ... */ + /// ``` + /// +- /// *This module is available if Syn is built with the `"visit"` feature.* ++ /// *This module is available only if Syn is built with the `"visit"` feature.* + /// + /// <br> + /// +@@ -603,7 +623,7 @@ mod gen { + /// /* ... */ + /// ``` + /// +- /// *This module is available if Syn is built with the `"visit-mut"` ++ /// *This module is available only if Syn is built with the `"visit-mut"` + /// feature.* + /// + /// <br> +@@ -702,7 +722,7 @@ mod gen { + /// /* ... */ + /// ``` + /// +- /// *This module is available if Syn is built with the `"fold"` feature.* ++ /// *This module is available only if Syn is built with the `"fold"` feature.* + /// + /// <br> + /// +@@ -744,6 +764,22 @@ mod gen { + #[rustfmt::skip] + pub mod fold; + ++ #[cfg(feature = "clone-impls")] ++ #[rustfmt::skip] ++ mod clone; ++ ++ #[cfg(feature = "extra-traits")] ++ #[rustfmt::skip] ++ mod eq; ++ ++ #[cfg(feature = "extra-traits")] ++ #[rustfmt::skip] ++ mod hash; ++ ++ #[cfg(feature = "extra-traits")] ++ #[rustfmt::skip] ++ mod debug; ++ + #[cfg(any(feature = "full", feature = "derive"))] + #[path = "../gen_helper.rs"] + mod helper; +@@ -757,6 +793,8 @@ pub mod export; + mod custom_keyword; + mod custom_punctuation; + mod sealed; ++mod span; ++mod thread; + + #[cfg(feature = "parsing")] + mod lookahead; +@@ -764,13 +802,15 @@ mod lookahead; + #[cfg(feature = "parsing")] + pub mod parse; + +-mod span; ++#[cfg(feature = "full")] ++mod reserved; ++ ++#[cfg(all(any(feature = "full", feature = "derive"), feature = "parsing"))] ++mod verbatim; + + #[cfg(all(any(feature = "full", feature = "derive"), feature = "printing"))] + mod print; + +-mod thread; +- + //////////////////////////////////////////////////////////////////////////////// + + #[allow(dead_code, non_camel_case_types)] +@@ -800,14 +840,14 @@ pub use crate::error::{Error, Result}; + /// + /// [`syn::parse2`]: parse2 + /// +-/// *This function is available if Syn is built with both the `"parsing"` and ++/// *This function is available only if Syn is built with both the `"parsing"` and + /// `"proc-macro"` features.* + /// + /// # Examples + /// + /// ``` +-/// extern crate proc_macro; +-/// ++/// # extern crate proc_macro; ++/// # + /// use proc_macro::TokenStream; + /// use quote::quote; + /// use syn::DeriveInput; +@@ -847,7 +887,7 @@ pub fn parse<T: parse::Parse>(tokens: proc_macro::TokenStream) -> Result<T> { + /// + /// [`syn::parse`]: parse() + /// +-/// *This function is available if Syn is built with the `"parsing"` feature.* ++/// *This function is available only if Syn is built with the `"parsing"` feature.* + #[cfg(feature = "parsing")] + pub fn parse2<T: parse::Parse>(tokens: proc_macro2::TokenStream) -> Result<T> { + parse::Parser::parse2(T::parse, tokens) +@@ -855,7 +895,7 @@ pub fn parse2<T: parse::Parse>(tokens: proc_macro2::TokenStream) -> Result<T> { + + /// Parse a string of Rust code into the chosen syntax tree node. + /// +-/// *This function is available if Syn is built with the `"parsing"` feature.* ++/// *This function is available only if Syn is built with the `"parsing"` feature.* + /// + /// # Hygiene + /// +@@ -874,9 +914,7 @@ pub fn parse2<T: parse::Parse>(tokens: proc_macro2::TokenStream) -> Result<T> { + /// Ok(()) + /// } + /// # +-/// # fn main() { +-/// # run().unwrap(); +-/// # } ++/// # run().unwrap(); + /// ``` + #[cfg(feature = "parsing")] + pub fn parse_str<T: parse::Parse>(s: &str) -> Result<T> { +@@ -894,7 +932,7 @@ pub fn parse_str<T: parse::Parse>(s: &str) -> Result<T> { + /// + /// If present, either of these would be an error using `from_str`. + /// +-/// *This function is available if Syn is built with the `"parsing"` and ++/// *This function is available only if Syn is built with the `"parsing"` and + /// `"full"` features.* + /// + /// # Examples +@@ -918,9 +956,7 @@ pub fn parse_str<T: parse::Parse>(s: &str) -> Result<T> { + /// Ok(()) + /// } + /// # +-/// # fn main() { +-/// # run().unwrap(); +-/// # } ++/// # run().unwrap(); + /// ``` + #[cfg(all(feature = "parsing", feature = "full"))] + pub fn parse_file(mut content: &str) -> Result<File> { +@@ -931,13 +967,16 @@ pub fn parse_file(mut content: &str) -> Result<File> { + } + + let mut shebang = None; +- if content.starts_with("#!") && !content.starts_with("#![") { +- if let Some(idx) = content.find('\n') { +- shebang = Some(content[..idx].to_string()); +- content = &content[idx..]; +- } else { +- shebang = Some(content.to_string()); +- content = ""; ++ if content.starts_with("#!") { ++ let rest = whitespace::skip(&content[2..]); ++ if !rest.starts_with('[') { ++ if let Some(idx) = content.find('\n') { ++ shebang = Some(content[..idx].to_string()); ++ content = &content[idx..]; ++ } else { ++ shebang = Some(content.to_string()); ++ content = ""; ++ } + } + } + +diff --git a/third_party/rust/syn/src/lifetime.rs b/third_party/rust/syn/src/lifetime.rs +index d51c48e827..959cc5f9c6 100644 +--- a/third_party/rust/syn/src/lifetime.rs ++++ b/third_party/rust/syn/src/lifetime.rs +@@ -18,10 +18,8 @@ use crate::lookahead; + /// - All following characters must be Unicode code points with the XID_Continue + /// property. + /// +-/// *This type is available if Syn is built with the `"derive"` or `"full"` ++/// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +-#[cfg_attr(feature = "extra-traits", derive(Debug))] +-#[derive(Clone)] + pub struct Lifetime { + pub apostrophe: Span, + pub ident: Ident, +@@ -72,6 +70,15 @@ impl Display for Lifetime { + } + } + ++impl Clone for Lifetime { ++ fn clone(&self) -> Self { ++ Lifetime { ++ apostrophe: self.apostrophe, ++ ident: self.ident.clone(), ++ } ++ } ++} ++ + impl PartialEq for Lifetime { + fn eq(&self, other: &Lifetime) -> bool { + self.ident.eq(&other.ident) +diff --git a/third_party/rust/syn/src/lit.rs b/third_party/rust/syn/src/lit.rs +index f2209a2980..ee77e75bec 100644 +--- a/third_party/rust/syn/src/lit.rs ++++ b/third_party/rust/syn/src/lit.rs +@@ -22,9 +22,6 @@ use crate::{Error, Result}; + ast_enum_of_structs! { + /// A Rust literal such as a string or integer or boolean. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` +- /// feature.* +- /// + /// # Syntax tree enum + /// + /// This type is a [syntax tree enum]. +@@ -33,7 +30,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum Lit #manual_extra_traits { ++ pub enum Lit { + /// A UTF-8 string literal: `"foo"`. + Str(LitStr), + +@@ -64,61 +61,44 @@ ast_enum_of_structs! { + + ast_struct! { + /// A UTF-8 string literal: `"foo"`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitStr #manual_extra_traits_debug { +- repr: Box<LitStrRepr>, ++ pub struct LitStr { ++ repr: Box<LitRepr>, + } + } + +-#[cfg_attr(feature = "clone-impls", derive(Clone))] +-struct LitStrRepr { +- token: Literal, +- suffix: Box<str>, +-} +- + ast_struct! { + /// A byte string literal: `b"foo"`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitByteStr #manual_extra_traits_debug { +- token: Literal, ++ pub struct LitByteStr { ++ repr: Box<LitRepr>, + } + } + + ast_struct! { + /// A byte literal: `b'f'`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitByte #manual_extra_traits_debug { +- token: Literal, ++ pub struct LitByte { ++ repr: Box<LitRepr>, + } + } + + ast_struct! { + /// A character literal: `'a'`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitChar #manual_extra_traits_debug { +- token: Literal, ++ pub struct LitChar { ++ repr: Box<LitRepr>, + } + } + ++struct LitRepr { ++ token: Literal, ++ suffix: Box<str>, ++} ++ + ast_struct! { + /// An integer literal: `1` or `1u16`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitInt #manual_extra_traits_debug { ++ pub struct LitInt { + repr: Box<LitIntRepr>, + } + } + +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + struct LitIntRepr { + token: Literal, + digits: Box<str>, +@@ -129,15 +109,11 @@ ast_struct! { + /// A floating point literal: `1f64` or `1.0e10f64`. + /// + /// Must be finite. May not be infinte or NaN. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitFloat #manual_extra_traits_debug { ++ pub struct LitFloat { + repr: Box<LitFloatRepr>, + } + } + +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + struct LitFloatRepr { + token: Literal, + digits: Box<str>, +@@ -146,92 +122,27 @@ struct LitFloatRepr { + + ast_struct! { + /// A boolean literal: `true` or `false`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitBool #manual_extra_traits_debug { ++ pub struct LitBool { + pub value: bool, + pub span: Span, + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Lit {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Lit { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (Lit::Str(this), Lit::Str(other)) => this == other, +- (Lit::ByteStr(this), Lit::ByteStr(other)) => this == other, +- (Lit::Byte(this), Lit::Byte(other)) => this == other, +- (Lit::Char(this), Lit::Char(other)) => this == other, +- (Lit::Int(this), Lit::Int(other)) => this == other, +- (Lit::Float(this), Lit::Float(other)) => this == other, +- (Lit::Bool(this), Lit::Bool(other)) => this == other, +- (Lit::Verbatim(this), Lit::Verbatim(other)) => this.to_string() == other.to_string(), +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Lit { +- fn hash<H>(&self, hash: &mut H) +- where +- H: Hasher, +- { +- match self { +- Lit::Str(lit) => { +- hash.write_u8(0); +- lit.hash(hash); +- } +- Lit::ByteStr(lit) => { +- hash.write_u8(1); +- lit.hash(hash); +- } +- Lit::Byte(lit) => { +- hash.write_u8(2); +- lit.hash(hash); +- } +- Lit::Char(lit) => { +- hash.write_u8(3); +- lit.hash(hash); +- } +- Lit::Int(lit) => { +- hash.write_u8(4); +- lit.hash(hash); +- } +- Lit::Float(lit) => { +- hash.write_u8(5); +- lit.hash(hash); +- } +- Lit::Bool(lit) => { +- hash.write_u8(6); +- lit.hash(hash); +- } +- Lit::Verbatim(lit) => { +- hash.write_u8(7); +- lit.to_string().hash(hash); +- } +- } +- } +-} +- + impl LitStr { + pub fn new(value: &str, span: Span) -> Self { +- let mut lit = Literal::string(value); +- lit.set_span(span); ++ let mut token = Literal::string(value); ++ token.set_span(span); + LitStr { +- repr: Box::new(LitStrRepr { +- token: lit, ++ repr: Box::new(LitRepr { ++ token, + suffix: Box::<str>::default(), + }), + } + } + + pub fn value(&self) -> String { +- let (value, _) = value::parse_lit_str(&self.repr.token.to_string()); ++ let repr = self.repr.token.to_string(); ++ let (value, _suffix) = value::parse_lit_str(&repr); + String::from(value) + } + +@@ -311,7 +222,7 @@ impl LitStr { + fn respan_token_tree(mut token: TokenTree, span: Span) -> TokenTree { + match &mut token { + TokenTree::Group(g) => { +- let stream = respan_token_stream(g.stream().clone(), span); ++ let stream = respan_token_stream(g.stream(), span); + *g = Group::new(g.delimiter(), stream); + g.set_span(span); + } +@@ -345,19 +256,30 @@ impl LitByteStr { + pub fn new(value: &[u8], span: Span) -> Self { + let mut token = Literal::byte_string(value); + token.set_span(span); +- LitByteStr { token } ++ LitByteStr { ++ repr: Box::new(LitRepr { ++ token, ++ suffix: Box::<str>::default(), ++ }), ++ } + } + + pub fn value(&self) -> Vec<u8> { +- value::parse_lit_byte_str(&self.token.to_string()) ++ let repr = self.repr.token.to_string(); ++ let (value, _suffix) = value::parse_lit_byte_str(&repr); ++ value + } + + pub fn span(&self) -> Span { +- self.token.span() ++ self.repr.token.span() + } + + pub fn set_span(&mut self, span: Span) { +- self.token.set_span(span) ++ self.repr.token.set_span(span) ++ } ++ ++ pub fn suffix(&self) -> &str { ++ &self.repr.suffix + } + } + +@@ -365,19 +287,30 @@ impl LitByte { + pub fn new(value: u8, span: Span) -> Self { + let mut token = Literal::u8_suffixed(value); + token.set_span(span); +- LitByte { token } ++ LitByte { ++ repr: Box::new(LitRepr { ++ token, ++ suffix: Box::<str>::default(), ++ }), ++ } + } + + pub fn value(&self) -> u8 { +- value::parse_lit_byte(&self.token.to_string()) ++ let repr = self.repr.token.to_string(); ++ let (value, _suffix) = value::parse_lit_byte(&repr); ++ value + } + + pub fn span(&self) -> Span { +- self.token.span() ++ self.repr.token.span() + } + + pub fn set_span(&mut self, span: Span) { +- self.token.set_span(span) ++ self.repr.token.set_span(span) ++ } ++ ++ pub fn suffix(&self) -> &str { ++ &self.repr.suffix + } + } + +@@ -385,36 +318,52 @@ impl LitChar { + pub fn new(value: char, span: Span) -> Self { + let mut token = Literal::character(value); + token.set_span(span); +- LitChar { token } ++ LitChar { ++ repr: Box::new(LitRepr { ++ token, ++ suffix: Box::<str>::default(), ++ }), ++ } + } + + pub fn value(&self) -> char { +- value::parse_lit_char(&self.token.to_string()) ++ let repr = self.repr.token.to_string(); ++ let (value, _suffix) = value::parse_lit_char(&repr); ++ value + } + + pub fn span(&self) -> Span { +- self.token.span() ++ self.repr.token.span() + } + + pub fn set_span(&mut self, span: Span) { +- self.token.set_span(span) ++ self.repr.token.set_span(span) ++ } ++ ++ pub fn suffix(&self) -> &str { ++ &self.repr.suffix + } + } + + impl LitInt { + pub fn new(repr: &str, span: Span) -> Self { +- if let Some((digits, suffix)) = value::parse_lit_int(repr) { +- let mut token = value::to_literal(repr); +- token.set_span(span); +- LitInt { +- repr: Box::new(LitIntRepr { +- token, +- digits, +- suffix, +- }), +- } +- } else { +- panic!("Not an integer literal: `{}`", repr); ++ let (digits, suffix) = match value::parse_lit_int(repr) { ++ Some(parse) => parse, ++ None => panic!("Not an integer literal: `{}`", repr), ++ }; ++ ++ let mut token = match value::to_literal(repr, &digits, &suffix) { ++ Some(token) => token, ++ None => panic!("Unsupported integer literal: `{}`", repr), ++ }; ++ ++ token.set_span(span); ++ LitInt { ++ repr: Box::new(LitIntRepr { ++ token, ++ digits, ++ suffix, ++ }), + } + } + +@@ -492,18 +441,23 @@ impl Display for LitInt { + + impl LitFloat { + pub fn new(repr: &str, span: Span) -> Self { +- if let Some((digits, suffix)) = value::parse_lit_float(repr) { +- let mut token = value::to_literal(repr); +- token.set_span(span); +- LitFloat { +- repr: Box::new(LitFloatRepr { +- token, +- digits, +- suffix, +- }), +- } +- } else { +- panic!("Not a float literal: `{}`", repr); ++ let (digits, suffix) = match value::parse_lit_float(repr) { ++ Some(parse) => parse, ++ None => panic!("Not a float literal: `{}`", repr), ++ }; ++ ++ let mut token = match value::to_literal(repr, &digits, &suffix) { ++ Some(token) => token, ++ None => panic!("Unsupported float literal: `{}`", repr), ++ }; ++ ++ token.set_span(span); ++ LitFloat { ++ repr: Box::new(LitFloatRepr { ++ token, ++ digits, ++ suffix, ++ }), + } + } + +@@ -575,7 +529,7 @@ mod debug_impls { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + formatter + .debug_struct("LitByteStr") +- .field("token", &format_args!("{}", self.token)) ++ .field("token", &format_args!("{}", self.repr.token)) + .finish() + } + } +@@ -584,7 +538,7 @@ mod debug_impls { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + formatter + .debug_struct("LitByte") +- .field("token", &format_args!("{}", self.token)) ++ .field("token", &format_args!("{}", self.repr.token)) + .finish() + } + } +@@ -593,7 +547,7 @@ mod debug_impls { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + formatter + .debug_struct("LitChar") +- .field("token", &format_args!("{}", self.token)) ++ .field("token", &format_args!("{}", self.repr.token)) + .finish() + } + } +@@ -626,15 +580,53 @@ mod debug_impls { + } + } + ++#[cfg(feature = "clone-impls")] ++impl Clone for LitRepr { ++ fn clone(&self) -> Self { ++ LitRepr { ++ token: self.token.clone(), ++ suffix: self.suffix.clone(), ++ } ++ } ++} ++ ++#[cfg(feature = "clone-impls")] ++impl Clone for LitIntRepr { ++ fn clone(&self) -> Self { ++ LitIntRepr { ++ token: self.token.clone(), ++ digits: self.digits.clone(), ++ suffix: self.suffix.clone(), ++ } ++ } ++} ++ ++#[cfg(feature = "clone-impls")] ++impl Clone for LitFloatRepr { ++ fn clone(&self) -> Self { ++ LitFloatRepr { ++ token: self.token.clone(), ++ digits: self.digits.clone(), ++ suffix: self.suffix.clone(), ++ } ++ } ++} ++ + macro_rules! lit_extra_traits { +- ($ty:ident, $($field:ident).+) => { +- #[cfg(feature = "extra-traits")] +- impl Eq for $ty {} ++ ($ty:ident) => { ++ #[cfg(feature = "clone-impls")] ++ impl Clone for $ty { ++ fn clone(&self) -> Self { ++ $ty { ++ repr: self.repr.clone(), ++ } ++ } ++ } + + #[cfg(feature = "extra-traits")] + impl PartialEq for $ty { + fn eq(&self, other: &Self) -> bool { +- self.$($field).+.to_string() == other.$($field).+.to_string() ++ self.repr.token.to_string() == other.repr.token.to_string() + } + } + +@@ -644,7 +636,7 @@ macro_rules! lit_extra_traits { + where + H: Hasher, + { +- self.$($field).+.to_string().hash(state); ++ self.repr.token.to_string().hash(state); + } + } + +@@ -657,20 +649,23 @@ macro_rules! lit_extra_traits { + }; + } + +-lit_extra_traits!(LitStr, repr.token); +-lit_extra_traits!(LitByteStr, token); +-lit_extra_traits!(LitByte, token); +-lit_extra_traits!(LitChar, token); +-lit_extra_traits!(LitInt, repr.token); +-lit_extra_traits!(LitFloat, repr.token); +-lit_extra_traits!(LitBool, value); ++lit_extra_traits!(LitStr); ++lit_extra_traits!(LitByteStr); ++lit_extra_traits!(LitByte); ++lit_extra_traits!(LitChar); ++lit_extra_traits!(LitInt); ++lit_extra_traits!(LitFloat); ++ ++#[cfg(feature = "parsing")] ++#[doc(hidden)] ++#[allow(non_snake_case)] ++pub fn LitBool(marker: lookahead::TokenMarker) -> LitBool { ++ match marker {} ++} + + ast_enum! { + /// The style of a string literal, either plain quoted or a raw string like + /// `r##"data"##`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` +- /// feature.* + pub enum StrStyle #no_visit { + /// An ordinary string like `"data"`. + Cooked, +@@ -691,7 +686,9 @@ pub fn Lit(marker: lookahead::TokenMarker) -> Lit { + #[cfg(feature = "parsing")] + pub mod parsing { + use super::*; ++ use crate::buffer::Cursor; + use crate::parse::{Parse, ParseStream, Result}; ++ use proc_macro2::Punct; + + impl Parse for Lit { + fn parse(input: ParseStream) -> Result<Self> { +@@ -699,25 +696,73 @@ pub mod parsing { + if let Some((lit, rest)) = cursor.literal() { + return Ok((Lit::new(lit), rest)); + } +- while let Some((ident, rest)) = cursor.ident() { +- let value = if ident == "true" { +- true +- } else if ident == "false" { +- false +- } else { +- break; +- }; +- let lit_bool = LitBool { +- value, +- span: ident.span(), +- }; +- return Ok((Lit::Bool(lit_bool), rest)); ++ ++ if let Some((ident, rest)) = cursor.ident() { ++ let value = ident == "true"; ++ if value || ident == "false" { ++ let lit_bool = LitBool { ++ value, ++ span: ident.span(), ++ }; ++ return Ok((Lit::Bool(lit_bool), rest)); ++ } + } ++ ++ if let Some((punct, rest)) = cursor.punct() { ++ if punct.as_char() == '-' { ++ if let Some((lit, rest)) = parse_negative_lit(punct, rest) { ++ return Ok((lit, rest)); ++ } ++ } ++ } ++ + Err(cursor.error("expected literal")) + }) + } + } + ++ fn parse_negative_lit(neg: Punct, cursor: Cursor) -> Option<(Lit, Cursor)> { ++ let (lit, rest) = cursor.literal()?; ++ ++ let mut span = neg.span(); ++ span = span.join(lit.span()).unwrap_or(span); ++ ++ let mut repr = lit.to_string(); ++ repr.insert(0, '-'); ++ ++ if !(repr.ends_with("f32") || repr.ends_with("f64")) { ++ if let Some((digits, suffix)) = value::parse_lit_int(&repr) { ++ if let Some(mut token) = value::to_literal(&repr, &digits, &suffix) { ++ token.set_span(span); ++ return Some(( ++ Lit::Int(LitInt { ++ repr: Box::new(LitIntRepr { ++ token, ++ digits, ++ suffix, ++ }), ++ }), ++ rest, ++ )); ++ } ++ } ++ } ++ ++ let (digits, suffix) = value::parse_lit_float(&repr)?; ++ let mut token = value::to_literal(&repr, &digits, &suffix)?; ++ token.set_span(span); ++ Some(( ++ Lit::Float(LitFloat { ++ repr: Box::new(LitFloatRepr { ++ token, ++ digits, ++ suffix, ++ }), ++ }), ++ rest, ++ )) ++ } ++ + impl Parse for LitStr { + fn parse(input: ParseStream) -> Result<Self> { + let head = input.fork(); +@@ -803,19 +848,19 @@ mod printing { + + impl ToTokens for LitByteStr { + fn to_tokens(&self, tokens: &mut TokenStream) { +- self.token.to_tokens(tokens); ++ self.repr.token.to_tokens(tokens); + } + } + + impl ToTokens for LitByte { + fn to_tokens(&self, tokens: &mut TokenStream) { +- self.token.to_tokens(tokens); ++ self.repr.token.to_tokens(tokens); + } + } + + impl ToTokens for LitChar { + fn to_tokens(&self, tokens: &mut TokenStream) { +- self.token.to_tokens(tokens); ++ self.repr.token.to_tokens(tokens); + } + } + +@@ -855,20 +900,29 @@ mod value { + b'"' | b'r' => { + let (_, suffix) = parse_lit_str(&repr); + return Lit::Str(LitStr { +- repr: Box::new(LitStrRepr { token, suffix }), ++ repr: Box::new(LitRepr { token, suffix }), + }); + } + b'b' => match byte(&repr, 1) { + b'"' | b'r' => { +- return Lit::ByteStr(LitByteStr { token }); ++ let (_, suffix) = parse_lit_byte_str(&repr); ++ return Lit::ByteStr(LitByteStr { ++ repr: Box::new(LitRepr { token, suffix }), ++ }); + } + b'\'' => { +- return Lit::Byte(LitByte { token }); ++ let (_, suffix) = parse_lit_byte(&repr); ++ return Lit::Byte(LitByte { ++ repr: Box::new(LitRepr { token, suffix }), ++ }); + } + _ => {} + }, + b'\'' => { +- return Lit::Char(LitChar { token }); ++ let (_, suffix) = parse_lit_char(&repr); ++ return Lit::Char(LitChar { ++ repr: Box::new(LitRepr { token, suffix }), ++ }); + } + b'0'..=b'9' | b'-' => { + if !(repr.ends_with("f32") || repr.ends_with("f64")) { +@@ -905,6 +959,44 @@ mod value { + + panic!("Unrecognized literal: `{}`", repr); + } ++ ++ pub fn suffix(&self) -> &str { ++ match self { ++ Lit::Str(lit) => lit.suffix(), ++ Lit::ByteStr(lit) => lit.suffix(), ++ Lit::Byte(lit) => lit.suffix(), ++ Lit::Char(lit) => lit.suffix(), ++ Lit::Int(lit) => lit.suffix(), ++ Lit::Float(lit) => lit.suffix(), ++ Lit::Bool(_) | Lit::Verbatim(_) => "", ++ } ++ } ++ ++ pub fn span(&self) -> Span { ++ match self { ++ Lit::Str(lit) => lit.span(), ++ Lit::ByteStr(lit) => lit.span(), ++ Lit::Byte(lit) => lit.span(), ++ Lit::Char(lit) => lit.span(), ++ Lit::Int(lit) => lit.span(), ++ Lit::Float(lit) => lit.span(), ++ Lit::Bool(lit) => lit.span, ++ Lit::Verbatim(lit) => lit.span(), ++ } ++ } ++ ++ pub fn set_span(&mut self, span: Span) { ++ match self { ++ Lit::Str(lit) => lit.set_span(span), ++ Lit::ByteStr(lit) => lit.set_span(span), ++ Lit::Byte(lit) => lit.set_span(span), ++ Lit::Char(lit) => lit.set_span(span), ++ Lit::Int(lit) => lit.set_span(span), ++ Lit::Float(lit) => lit.set_span(span), ++ Lit::Bool(lit) => lit.span = span, ++ Lit::Verbatim(lit) => lit.set_span(span), ++ } ++ } + } + + /// Get the byte at offset idx, or a default of `b'\0'` if we're looking +@@ -1004,19 +1096,18 @@ mod value { + pounds += 1; + } + assert_eq!(byte(s, pounds), b'"'); +- assert_eq!(byte(s, s.len() - pounds - 1), b'"'); +- for end in s[s.len() - pounds..].bytes() { ++ let close = s.rfind('"').unwrap(); ++ for end in s[close + 1..close + 1 + pounds].bytes() { + assert_eq!(end, b'#'); + } + +- let content = s[pounds + 1..s.len() - pounds - 1] +- .to_owned() +- .into_boxed_str(); +- let suffix = Box::<str>::default(); // todo ++ let content = s[pounds + 1..close].to_owned().into_boxed_str(); ++ let suffix = s[close + 1 + pounds..].to_owned().into_boxed_str(); + (content, suffix) + } + +- pub fn parse_lit_byte_str(s: &str) -> Vec<u8> { ++ // Returns (content, suffix). ++ pub fn parse_lit_byte_str(s: &str) -> (Vec<u8>, Box<str>) { + assert_eq!(byte(s, 0), b'b'); + match byte(s, 1) { + b'"' => parse_lit_byte_str_cooked(s), +@@ -1028,25 +1119,25 @@ mod value { + // Clippy false positive + // https://github.com/rust-lang-nursery/rust-clippy/issues/2329 + #[allow(clippy::needless_continue)] +- fn parse_lit_byte_str_cooked(mut s: &str) -> Vec<u8> { ++ fn parse_lit_byte_str_cooked(mut s: &str) -> (Vec<u8>, Box<str>) { + assert_eq!(byte(s, 0), b'b'); + assert_eq!(byte(s, 1), b'"'); + s = &s[2..]; + + // We're going to want to have slices which don't respect codepoint boundaries. +- let mut s = s.as_bytes(); ++ let mut v = s.as_bytes(); + + let mut out = Vec::new(); + 'outer: loop { +- let byte = match byte(s, 0) { ++ let byte = match byte(v, 0) { + b'"' => break, + b'\\' => { +- let b = byte(s, 1); +- s = &s[2..]; ++ let b = byte(v, 1); ++ v = &v[2..]; + match b { + b'x' => { +- let (b, rest) = backslash_x(s); +- s = rest; ++ let (b, rest) = backslash_x(v); ++ v = rest; + b + } + b'n' => b'\n', +@@ -1057,10 +1148,10 @@ mod value { + b'\'' => b'\'', + b'"' => b'"', + b'\r' | b'\n' => loop { +- let byte = byte(s, 0); ++ let byte = byte(v, 0); + let ch = char::from_u32(u32::from(byte)).unwrap(); + if ch.is_whitespace() { +- s = &s[1..]; ++ v = &v[1..]; + } else { + continue 'outer; + } +@@ -1069,42 +1160,45 @@ mod value { + } + } + b'\r' => { +- assert_eq!(byte(s, 1), b'\n', "Bare CR not allowed in string"); +- s = &s[2..]; ++ assert_eq!(byte(v, 1), b'\n', "Bare CR not allowed in string"); ++ v = &v[2..]; + b'\n' + } + b => { +- s = &s[1..]; ++ v = &v[1..]; + b + } + }; + out.push(byte); + } + +- assert_eq!(s, b"\""); +- out ++ assert_eq!(byte(v, 0), b'"'); ++ let suffix = s[s.len() - v.len() + 1..].to_owned().into_boxed_str(); ++ (out, suffix) + } + +- fn parse_lit_byte_str_raw(s: &str) -> Vec<u8> { ++ fn parse_lit_byte_str_raw(s: &str) -> (Vec<u8>, Box<str>) { + assert_eq!(byte(s, 0), b'b'); +- String::from(parse_lit_str_raw(&s[1..]).0).into_bytes() ++ let (value, suffix) = parse_lit_str_raw(&s[1..]); ++ (String::from(value).into_bytes(), suffix) + } + +- pub fn parse_lit_byte(s: &str) -> u8 { ++ // Returns (value, suffix). ++ pub fn parse_lit_byte(s: &str) -> (u8, Box<str>) { + assert_eq!(byte(s, 0), b'b'); + assert_eq!(byte(s, 1), b'\''); + + // We're going to want to have slices which don't respect codepoint boundaries. +- let mut s = s[2..].as_bytes(); ++ let mut v = s[2..].as_bytes(); + +- let b = match byte(s, 0) { ++ let b = match byte(v, 0) { + b'\\' => { +- let b = byte(s, 1); +- s = &s[2..]; ++ let b = byte(v, 1); ++ v = &v[2..]; + match b { + b'x' => { +- let (b, rest) = backslash_x(s); +- s = rest; ++ let (b, rest) = backslash_x(v); ++ v = rest; + b + } + b'n' => b'\n', +@@ -1118,16 +1212,18 @@ mod value { + } + } + b => { +- s = &s[1..]; ++ v = &v[1..]; + b + } + }; + +- assert_eq!(byte(s, 0), b'\''); +- b ++ assert_eq!(byte(v, 0), b'\''); ++ let suffix = s[s.len() - v.len() + 1..].to_owned().into_boxed_str(); ++ (b, suffix) + } + +- pub fn parse_lit_char(mut s: &str) -> char { ++ // Returns (value, suffix). ++ pub fn parse_lit_char(mut s: &str) -> (char, Box<str>) { + assert_eq!(byte(s, 0), b'\''); + s = &s[1..]; + +@@ -1163,8 +1259,9 @@ mod value { + ch + } + }; +- assert_eq!(s, "\'", "Expected end of char literal"); +- ch ++ assert_eq!(byte(s, 0), b'\''); ++ let suffix = s[1..].to_owned().into_boxed_str(); ++ (ch, suffix) + } + + fn backslash_x<S>(s: &S) -> (u8, &S) +@@ -1334,7 +1431,11 @@ mod value { + } + b'e' | b'E' => { + if has_e { +- return None; ++ if has_exponent { ++ break; ++ } else { ++ return None; ++ } + } + has_e = true; + bytes[write] = b'e'; +@@ -1372,11 +1473,33 @@ mod value { + } + } + +- pub fn to_literal(s: &str) -> Literal { +- let stream = s.parse::<TokenStream>().unwrap(); +- match stream.into_iter().next().unwrap() { +- TokenTree::Literal(l) => l, +- _ => unreachable!(), ++ pub fn to_literal(repr: &str, digits: &str, suffix: &str) -> Option<Literal> { ++ if repr.starts_with('-') { ++ if suffix == "f64" { ++ digits.parse().ok().map(Literal::f64_suffixed) ++ } else if suffix == "f32" { ++ digits.parse().ok().map(Literal::f32_suffixed) ++ } else if suffix == "i64" { ++ digits.parse().ok().map(Literal::i64_suffixed) ++ } else if suffix == "i32" { ++ digits.parse().ok().map(Literal::i32_suffixed) ++ } else if suffix == "i16" { ++ digits.parse().ok().map(Literal::i16_suffixed) ++ } else if suffix == "i8" { ++ digits.parse().ok().map(Literal::i8_suffixed) ++ } else if !suffix.is_empty() { ++ None ++ } else if digits.contains('.') { ++ digits.parse().ok().map(Literal::f64_unsuffixed) ++ } else { ++ digits.parse().ok().map(Literal::i64_unsuffixed) ++ } ++ } else { ++ let stream = repr.parse::<TokenStream>().unwrap(); ++ match stream.into_iter().next().unwrap() { ++ TokenTree::Literal(l) => Some(l), ++ _ => unreachable!(), ++ } + } + } + } +diff --git a/third_party/rust/syn/src/mac.rs b/third_party/rust/syn/src/mac.rs +index 6c3dcae92a..de288a34e1 100644 +--- a/third_party/rust/syn/src/mac.rs ++++ b/third_party/rust/syn/src/mac.rs +@@ -2,21 +2,17 @@ use super::*; + use crate::token::{Brace, Bracket, Paren}; + use proc_macro2::TokenStream; + #[cfg(feature = "parsing")] +-use proc_macro2::{Delimiter, Span, TokenTree}; ++use proc_macro2::{Delimiter, Group, Span, TokenTree}; + + #[cfg(feature = "parsing")] + use crate::parse::{Parse, ParseStream, Parser, Result}; +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; +-#[cfg(feature = "extra-traits")] +-use std::hash::{Hash, Hasher}; + + ast_struct! { + /// A macro invocation: `println!("{}", mac)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- pub struct Macro #manual_extra_traits { ++ pub struct Macro { + pub path: Path, + pub bang_token: Token![!], + pub delimiter: MacroDelimiter, +@@ -27,7 +23,7 @@ ast_struct! { + ast_enum! { + /// A grouping token that surrounds a macro body: `m!(...)` or `m!{...}` or `m![...]`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum MacroDelimiter { + Paren(Paren), +@@ -36,39 +32,20 @@ ast_enum! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Macro {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Macro { +- fn eq(&self, other: &Self) -> bool { +- self.path == other.path +- && self.bang_token == other.bang_token +- && self.delimiter == other.delimiter +- && TokenStreamHelper(&self.tokens) == TokenStreamHelper(&other.tokens) +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Macro { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- self.path.hash(state); +- self.bang_token.hash(state); +- self.delimiter.hash(state); +- TokenStreamHelper(&self.tokens).hash(state); +- } +-} +- + #[cfg(feature = "parsing")] +-fn delimiter_span(delimiter: &MacroDelimiter) -> Span { +- match delimiter { ++fn delimiter_span_close(macro_delimiter: &MacroDelimiter) -> Span { ++ let delimiter = match macro_delimiter { ++ MacroDelimiter::Paren(_) => Delimiter::Parenthesis, ++ MacroDelimiter::Brace(_) => Delimiter::Brace, ++ MacroDelimiter::Bracket(_) => Delimiter::Bracket, ++ }; ++ let mut group = Group::new(delimiter, TokenStream::new()); ++ group.set_span(match macro_delimiter { + MacroDelimiter::Paren(token) => token.span, + MacroDelimiter::Brace(token) => token.span, + MacroDelimiter::Bracket(token) => token.span, +- } ++ }); ++ group.span_close() + } + + impl Macro { +@@ -163,9 +140,7 @@ impl Macro { + /// given parser. + #[cfg(feature = "parsing")] + pub fn parse_body_with<F: Parser>(&self, parser: F) -> Result<F::Output> { +- // TODO: see if we can get a group.span_close() span in here as the +- // scope, rather than the span of the whole group. +- let scope = delimiter_span(&self.delimiter); ++ let scope = delimiter_span_close(&self.delimiter); + crate::parse::parse_scoped(parser, scope, self.tokens.clone()) + } + } +diff --git a/third_party/rust/syn/src/macros.rs b/third_party/rust/syn/src/macros.rs +index 9cac5c15df..8060224381 100644 +--- a/third_party/rust/syn/src/macros.rs ++++ b/third_party/rust/syn/src/macros.rs +@@ -4,15 +4,11 @@ macro_rules! ast_struct { + struct $name:ident #full $($rest:tt)* + ) => { + #[cfg(feature = "full")] +- #[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] + $($attrs_pub)* struct $name $($rest)* + + #[cfg(not(feature = "full"))] +- #[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] + $($attrs_pub)* struct $name { +- _noconstruct: (), ++ _noconstruct: ::std::marker::PhantomData<::proc_macro2::Span>, + } + + #[cfg(all(not(feature = "full"), feature = "printing"))] +@@ -23,29 +19,10 @@ macro_rules! ast_struct { + } + }; + +- ( +- [$($attrs_pub:tt)*] +- struct $name:ident #manual_extra_traits $($rest:tt)* +- ) => { +- #[cfg_attr(feature = "extra-traits", derive(Debug))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] +- $($attrs_pub)* struct $name $($rest)* +- }; +- +- ( +- [$($attrs_pub:tt)*] +- struct $name:ident #manual_extra_traits_debug $($rest:tt)* +- ) => { +- #[cfg_attr(feature = "clone-impls", derive(Clone))] +- $($attrs_pub)* struct $name $($rest)* +- }; +- + ( + [$($attrs_pub:tt)*] + struct $name:ident $($rest:tt)* + ) => { +- #[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] + $($attrs_pub)* struct $name $($rest)* + }; + +@@ -63,21 +40,10 @@ macro_rules! ast_enum { + ast_enum!([$($attrs_pub)*] enum $name $($rest)*); + ); + +- ( +- [$($attrs_pub:tt)*] +- enum $name:ident #manual_extra_traits $($rest:tt)* +- ) => ( +- #[cfg_attr(feature = "extra-traits", derive(Debug))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] +- $($attrs_pub)* enum $name $($rest)* +- ); +- + ( + [$($attrs_pub:tt)*] + enum $name:ident $($rest:tt)* + ) => ( +- #[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] + $($attrs_pub)* enum $name $($rest)* + ); + +@@ -120,15 +86,9 @@ macro_rules! ast_enum_of_structs_impl { + check_keyword_matches!(pub $pub); + check_keyword_matches!(enum $enum); + +- $( +- $( +- impl From<$member> for $name { +- fn from(e: $member) -> $name { +- $name::$variant(e) +- } +- } +- )* +- )* ++ $($( ++ ast_enum_from_struct!($name::$variant, $member); ++ )*)* + + #[cfg(feature = "printing")] + generate_to_tokens! { +@@ -140,6 +100,19 @@ macro_rules! ast_enum_of_structs_impl { + }; + } + ++macro_rules! ast_enum_from_struct { ++ // No From<TokenStream> for verbatim variants. ++ ($name:ident::Verbatim, $member:ident) => {}; ++ ++ ($name:ident::$variant:ident, $member:ident) => { ++ impl From<$member> for $name { ++ fn from(e: $member) -> $name { ++ $name::$variant(e) ++ } ++ } ++ }; ++} ++ + #[cfg(feature = "printing")] + macro_rules! generate_to_tokens { + (do_not_generate_to_tokens $($foo:tt)*) => (); +diff --git a/third_party/rust/syn/src/op.rs b/third_party/rust/syn/src/op.rs +index 49fb853c79..d254673b40 100644 +--- a/third_party/rust/syn/src/op.rs ++++ b/third_party/rust/syn/src/op.rs +@@ -1,9 +1,8 @@ + ast_enum! { + /// A binary operator: `+`, `+=`, `&`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- #[cfg_attr(feature = "clone-impls", derive(Copy))] + pub enum BinOp { + /// The `+` operator (addition) + Add(Token![+]), +@@ -67,9 +66,8 @@ ast_enum! { + ast_enum! { + /// A unary operator: `*`, `!`, `-`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- #[cfg_attr(feature = "clone-impls", derive(Copy))] + pub enum UnOp { + /// The `*` operator for dereferencing + Deref(Token![*]), +diff --git a/third_party/rust/syn/src/parse.rs b/third_party/rust/syn/src/parse.rs +index 7c7b194308..abb4c4c14f 100644 +--- a/third_party/rust/syn/src/parse.rs ++++ b/third_party/rust/syn/src/parse.rs +@@ -26,8 +26,8 @@ + //! [`parse_macro_input!`]: ../macro.parse_macro_input.html + //! + //! ``` +-//! extern crate proc_macro; +-//! ++//! # extern crate proc_macro; ++//! # + //! use proc_macro::TokenStream; + //! use syn::{braced, parse_macro_input, token, Field, Ident, Result, Token}; + //! use syn::parse::{Parse, ParseStream}; +@@ -109,9 +109,7 @@ + //! # Ok(()) + //! # } + //! # +-//! # fn main() { +-//! # run_parser().unwrap(); +-//! # } ++//! # run_parser().unwrap(); + //! ``` + //! + //! The [`parse_quote!`] macro also uses this approach. +@@ -155,8 +153,8 @@ + //! [`Parser`]: trait.Parser.html + //! + //! ``` +-//! extern crate proc_macro; +-//! ++//! # extern crate proc_macro; ++//! # + //! use proc_macro::TokenStream; + //! use syn::parse::Parser; + //! use syn::punctuated::Punctuated; +@@ -186,7 +184,7 @@ + //! + //! --- + //! +-//! *This module is available if Syn is built with the `"parsing"` feature.* ++//! *This module is available only if Syn is built with the `"parsing"` feature.* + + #[path = "discouraged.rs"] + pub mod discouraged; +@@ -217,6 +215,11 @@ pub use crate::lookahead::{Lookahead1, Peek}; + + /// Parsing interface implemented by all types that can be parsed in a default + /// way from a token stream. ++/// ++/// Refer to the [module documentation] for details about implementing and using ++/// the `Parse` trait. ++/// ++/// [module documentation]: self + pub trait Parse: Sized { + fn parse(input: ParseStream) -> Result<Self>; + } +@@ -263,13 +266,16 @@ pub struct ParseBuffer<'a> { + // the cell. + cell: Cell<Cursor<'static>>, + marker: PhantomData<Cursor<'a>>, +- unexpected: Rc<Cell<Option<Span>>>, ++ unexpected: Cell<Option<Rc<Cell<Unexpected>>>>, + } + + impl<'a> Drop for ParseBuffer<'a> { + fn drop(&mut self) { +- if !self.is_empty() && self.unexpected.get().is_none() { +- self.unexpected.set(Some(self.cursor().span())); ++ if let Some(unexpected_span) = span_of_unexpected_ignoring_nones(self.cursor()) { ++ let (inner, old_span) = inner_unexpected(self); ++ if old_span.is_none() { ++ inner.set(Unexpected::Some(unexpected_span)); ++ } + } + } + } +@@ -324,15 +330,12 @@ impl<'a> Debug for ParseBuffer<'a> { + /// # input.parse() + /// # } + /// # +-/// # fn main() { +-/// # use syn::parse::Parser; +-/// # let remainder = remainder_after_skipping_past_next_at +-/// # .parse_str("a @ b c") +-/// # .unwrap(); +-/// # assert_eq!(remainder.to_string(), "b c"); +-/// # } ++/// # use syn::parse::Parser; ++/// # let remainder = remainder_after_skipping_past_next_at ++/// # .parse_str("a @ b c") ++/// # .unwrap(); ++/// # assert_eq!(remainder.to_string(), "b c"); + /// ``` +-#[derive(Copy, Clone)] + pub struct StepCursor<'c, 'a> { + scope: Span, + // This field is covariant in 'c. +@@ -356,6 +359,14 @@ impl<'c, 'a> Deref for StepCursor<'c, 'a> { + } + } + ++impl<'c, 'a> Copy for StepCursor<'c, 'a> {} ++ ++impl<'c, 'a> Clone for StepCursor<'c, 'a> { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++ + impl<'c, 'a> StepCursor<'c, 'a> { + /// Triggers an error at the current position of the parse stream. + /// +@@ -375,36 +386,81 @@ pub(crate) fn advance_step_cursor<'c, 'a>(proof: StepCursor<'c, 'a>, to: Cursor< + unsafe { mem::transmute::<Cursor<'c>, Cursor<'a>>(to) } + } + +-fn skip(input: ParseStream) -> bool { +- input +- .step(|cursor| { +- if let Some((_lifetime, rest)) = cursor.lifetime() { +- Ok((true, rest)) +- } else if let Some((_token, rest)) = cursor.token_tree() { +- Ok((true, rest)) +- } else { +- Ok((false, *cursor)) +- } +- }) +- .unwrap() +-} +- + pub(crate) fn new_parse_buffer( + scope: Span, + cursor: Cursor, +- unexpected: Rc<Cell<Option<Span>>>, ++ unexpected: Rc<Cell<Unexpected>>, + ) -> ParseBuffer { + ParseBuffer { + scope, + // See comment on `cell` in the struct definition. + cell: Cell::new(unsafe { mem::transmute::<Cursor, Cursor<'static>>(cursor) }), + marker: PhantomData, +- unexpected, ++ unexpected: Cell::new(Some(unexpected)), ++ } ++} ++ ++pub(crate) enum Unexpected { ++ None, ++ Some(Span), ++ Chain(Rc<Cell<Unexpected>>), ++} ++ ++impl Default for Unexpected { ++ fn default() -> Self { ++ Unexpected::None + } + } + +-pub(crate) fn get_unexpected(buffer: &ParseBuffer) -> Rc<Cell<Option<Span>>> { +- buffer.unexpected.clone() ++impl Clone for Unexpected { ++ fn clone(&self) -> Self { ++ match self { ++ Unexpected::None => Unexpected::None, ++ Unexpected::Some(span) => Unexpected::Some(*span), ++ Unexpected::Chain(next) => Unexpected::Chain(next.clone()), ++ } ++ } ++} ++ ++// We call this on Cell<Unexpected> and Cell<Option<T>> where temporarily ++// swapping in a None is cheap. ++fn cell_clone<T: Default + Clone>(cell: &Cell<T>) -> T { ++ let prev = cell.take(); ++ let ret = prev.clone(); ++ cell.set(prev); ++ ret ++} ++ ++fn inner_unexpected(buffer: &ParseBuffer) -> (Rc<Cell<Unexpected>>, Option<Span>) { ++ let mut unexpected = get_unexpected(buffer); ++ loop { ++ match cell_clone(&unexpected) { ++ Unexpected::None => return (unexpected, None), ++ Unexpected::Some(span) => return (unexpected, Some(span)), ++ Unexpected::Chain(next) => unexpected = next, ++ } ++ } ++} ++ ++pub(crate) fn get_unexpected(buffer: &ParseBuffer) -> Rc<Cell<Unexpected>> { ++ cell_clone(&buffer.unexpected).unwrap() ++} ++ ++fn span_of_unexpected_ignoring_nones(mut cursor: Cursor) -> Option<Span> { ++ if cursor.eof() { ++ return None; ++ } ++ while let Some((inner, _span, rest)) = cursor.group(Delimiter::None) { ++ if let Some(unexpected) = span_of_unexpected_ignoring_nones(inner) { ++ return Some(unexpected); ++ } ++ cursor = rest; ++ } ++ if cursor.eof() { ++ None ++ } else { ++ Some(cursor.span()) ++ } + } + + impl<'a> ParseBuffer<'a> { +@@ -566,14 +622,17 @@ impl<'a> ParseBuffer<'a> { + /// } + /// ``` + pub fn peek2<T: Peek>(&self, token: T) -> bool { +- let ahead = self.fork(); +- skip(&ahead) && ahead.peek(token) ++ let _ = token; ++ self.cursor().skip().map_or(false, T::Token::peek) + } + + /// Looks at the third-next token in the parse stream. + pub fn peek3<T: Peek>(&self, token: T) -> bool { +- let ahead = self.fork(); +- skip(&ahead) && skip(&ahead) && ahead.peek(token) ++ let _ = token; ++ self.cursor() ++ .skip() ++ .and_then(Cursor::skip) ++ .map_or(false, T::Token::peek) + } + + /// Parses zero or more occurrences of `T` separated by punctuation of type +@@ -615,12 +674,10 @@ impl<'a> ParseBuffer<'a> { + /// } + /// } + /// # +- /// # fn main() { +- /// # let input = quote! { +- /// # struct S(A, B); +- /// # }; +- /// # syn::parse2::<TupleStruct>(input).unwrap(); +- /// # } ++ /// # let input = quote! { ++ /// # struct S(A, B); ++ /// # }; ++ /// # syn::parse2::<TupleStruct>(input).unwrap(); + /// ``` + pub fn parse_terminated<T, P: Parse>( + &self, +@@ -847,8 +904,8 @@ impl<'a> ParseBuffer<'a> { + cell: self.cell.clone(), + marker: PhantomData, + // Not the parent's unexpected. Nothing cares whether the clone +- // parses all the way. +- unexpected: Rc::new(Cell::new(None)), ++ // parses all the way unless we `advance_to`. ++ unexpected: Cell::new(Some(Rc::new(Cell::new(Unexpected::None)))), + } + } + +@@ -923,13 +980,11 @@ impl<'a> ParseBuffer<'a> { + /// # input.parse() + /// # } + /// # +- /// # fn main() { +- /// # use syn::parse::Parser; +- /// # let remainder = remainder_after_skipping_past_next_at +- /// # .parse_str("a @ b c") +- /// # .unwrap(); +- /// # assert_eq!(remainder.to_string(), "b c"); +- /// # } ++ /// # use syn::parse::Parser; ++ /// # let remainder = remainder_after_skipping_past_next_at ++ /// # .parse_str("a @ b c") ++ /// # .unwrap(); ++ /// # assert_eq!(remainder.to_string(), "b c"); + /// ``` + pub fn step<F, R>(&self, function: F) -> Result<R> + where +@@ -961,6 +1016,18 @@ impl<'a> ParseBuffer<'a> { + Ok(node) + } + ++ /// Returns the `Span` of the next token in the parse stream, or ++ /// `Span::call_site()` if this parse stream has completely exhausted its ++ /// input `TokenStream`. ++ pub fn span(&self) -> Span { ++ let cursor = self.cursor(); ++ if cursor.eof() { ++ self.scope ++ } else { ++ crate::buffer::open_span_of_group(cursor) ++ } ++ } ++ + /// Provides low-level access to the token representation underlying this + /// parse stream. + /// +@@ -971,7 +1038,7 @@ impl<'a> ParseBuffer<'a> { + } + + fn check_unexpected(&self) -> Result<()> { +- match self.unexpected.get() { ++ match inner_unexpected(self).1 { + Some(span) => Err(Error::new(span, "unexpected token")), + None => Ok(()), + } +@@ -1048,7 +1115,7 @@ impl Parse for Literal { + /// + /// [module documentation]: self + /// +-/// *This trait is available if Syn is built with the `"parsing"` feature.* ++/// *This trait is available only if Syn is built with the `"parsing"` feature.* + pub trait Parser: Sized { + type Output; + +@@ -1063,7 +1130,7 @@ pub trait Parser: Sized { + /// This function will check that the input is fully parsed. If there are + /// any unparsed tokens at the end of the stream, an error is returned. + /// +- /// *This method is available if Syn is built with both the `"parsing"` and ++ /// *This method is available only if Syn is built with both the `"parsing"` and + /// `"proc-macro"` features.* + #[cfg(all( + not(all(target_arch = "wasm32", any(target_os = "unknown", target_os = "wasi"))), +@@ -1088,6 +1155,7 @@ pub trait Parser: Sized { + + // Not public API. + #[doc(hidden)] ++ #[cfg(any(feature = "full", feature = "derive"))] + fn __parse_scoped(self, scope: Span, tokens: TokenStream) -> Result<Self::Output> { + let _ = scope; + self.parse2(tokens) +@@ -1095,6 +1163,7 @@ pub trait Parser: Sized { + + // Not public API. + #[doc(hidden)] ++ #[cfg(any(feature = "full", feature = "derive"))] + fn __parse_stream(self, input: ParseStream) -> Result<Self::Output> { + input.parse().and_then(|tokens| self.parse2(tokens)) + } +@@ -1103,7 +1172,7 @@ pub trait Parser: Sized { + fn tokens_to_parse_buffer(tokens: &TokenBuffer) -> ParseBuffer { + let scope = Span::call_site(); + let cursor = tokens.begin(); +- let unexpected = Rc::new(Cell::new(None)); ++ let unexpected = Rc::new(Cell::new(Unexpected::None)); + new_parse_buffer(scope, cursor, unexpected) + } + +@@ -1118,38 +1187,42 @@ where + let state = tokens_to_parse_buffer(&buf); + let node = self(&state)?; + state.check_unexpected()?; +- if state.is_empty() { +- Ok(node) ++ if let Some(unexpected_span) = span_of_unexpected_ignoring_nones(state.cursor()) { ++ Err(Error::new(unexpected_span, "unexpected token")) + } else { +- Err(state.error("unexpected token")) ++ Ok(node) + } + } + + #[doc(hidden)] ++ #[cfg(any(feature = "full", feature = "derive"))] + fn __parse_scoped(self, scope: Span, tokens: TokenStream) -> Result<Self::Output> { + let buf = TokenBuffer::new2(tokens); + let cursor = buf.begin(); +- let unexpected = Rc::new(Cell::new(None)); ++ let unexpected = Rc::new(Cell::new(Unexpected::None)); + let state = new_parse_buffer(scope, cursor, unexpected); + let node = self(&state)?; + state.check_unexpected()?; +- if state.is_empty() { +- Ok(node) ++ if let Some(unexpected_span) = span_of_unexpected_ignoring_nones(state.cursor()) { ++ Err(Error::new(unexpected_span, "unexpected token")) + } else { +- Err(state.error("unexpected token")) ++ Ok(node) + } + } + + #[doc(hidden)] ++ #[cfg(any(feature = "full", feature = "derive"))] + fn __parse_stream(self, input: ParseStream) -> Result<Self::Output> { + self(input) + } + } + ++#[cfg(any(feature = "full", feature = "derive"))] + pub(crate) fn parse_scoped<F: Parser>(f: F, scope: Span, tokens: TokenStream) -> Result<F::Output> { + f.__parse_scoped(scope, tokens) + } + ++#[cfg(any(feature = "full", feature = "derive"))] + pub(crate) fn parse_stream<F: Parser>(f: F, input: ParseStream) -> Result<F::Output> { + f.__parse_stream(input) + } +@@ -1160,8 +1233,8 @@ pub(crate) fn parse_stream<F: Parser>(f: F, input: ParseStream) -> Result<F::Out + /// provided any attribute args. + /// + /// ``` +-/// extern crate proc_macro; +-/// ++/// # extern crate proc_macro; ++/// # + /// use proc_macro::TokenStream; + /// use syn::parse_macro_input; + /// use syn::parse::Nothing; +diff --git a/third_party/rust/syn/src/parse_macro_input.rs b/third_party/rust/syn/src/parse_macro_input.rs +index d6e0725c17..c8fc1cea37 100644 +--- a/third_party/rust/syn/src/parse_macro_input.rs ++++ b/third_party/rust/syn/src/parse_macro_input.rs +@@ -16,8 +16,8 @@ + /// #\[proc_macro_attribute\] attribute. + /// + /// ``` +-/// extern crate proc_macro; +-/// ++/// # extern crate proc_macro; ++/// # + /// use proc_macro::TokenStream; + /// use syn::{parse_macro_input, Result}; + /// use syn::parse::{Parse, ParseStream}; +@@ -43,7 +43,31 @@ + /// # "".parse().unwrap() + /// } + /// ``` +-#[macro_export(local_inner_macros)] ++/// ++/// <br> ++/// ++/// # Expansion ++/// ++/// `parse_macro_input!($variable as $Type)` expands to something like: ++/// ++/// ```no_run ++/// # extern crate proc_macro; ++/// # ++/// # macro_rules! doc_test { ++/// # ($variable:ident as $Type:ty) => { ++/// match syn::parse::<$Type>($variable) { ++/// Ok(syntax_tree) => syntax_tree, ++/// Err(err) => return proc_macro::TokenStream::from(err.to_compile_error()), ++/// } ++/// # }; ++/// # } ++/// # ++/// # fn test(input: proc_macro::TokenStream) -> proc_macro::TokenStream { ++/// # let _ = doc_test!(input as syn::Ident); ++/// # proc_macro::TokenStream::new() ++/// # } ++/// ``` ++#[macro_export] + macro_rules! parse_macro_input { + ($tokenstream:ident as $ty:ty) => { + match $crate::parse_macro_input::parse::<$ty>($tokenstream) { +@@ -54,7 +78,7 @@ macro_rules! parse_macro_input { + } + }; + ($tokenstream:ident) => { +- parse_macro_input!($tokenstream as _) ++ $crate::parse_macro_input!($tokenstream as _) + }; + } + +diff --git a/third_party/rust/syn/src/parse_quote.rs b/third_party/rust/syn/src/parse_quote.rs +index 18a47b95c7..66aa818cd0 100644 +--- a/third_party/rust/syn/src/parse_quote.rs ++++ b/third_party/rust/syn/src/parse_quote.rs +@@ -24,7 +24,7 @@ + /// } + /// ``` + /// +-/// *This macro is available if Syn is built with the `"parsing"` feature, ++/// *This macro is available only if Syn is built with the `"parsing"` feature, + /// although interpolation of syntax tree nodes into the quoted tokens is only + /// supported if Syn is built with the `"printing"` feature as well.* + /// +@@ -56,8 +56,10 @@ + /// or inner like `#![...]` + /// - [`Punctuated<T, P>`] — parses zero or more `T` separated by punctuation + /// `P` with optional trailing punctuation ++/// - [`Vec<Stmt>`] — parses the same as `Block::parse_within` + /// + /// [`Punctuated<T, P>`]: punctuated::Punctuated ++/// [`Vec<Stmt>`]: Block::parse_within + /// + /// # Panics + /// +@@ -67,7 +69,7 @@ + // + // TODO: allow Punctuated to be inferred as intra doc link, currently blocked on + // https://github.com/rust-lang/rust/issues/62834 +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! parse_quote { + ($($tt:tt)*) => { + $crate::parse_quote::parse( +@@ -112,6 +114,8 @@ impl<T: Parse> ParseQuote for T { + use crate::punctuated::Punctuated; + #[cfg(any(feature = "full", feature = "derive"))] + use crate::{attr, Attribute}; ++#[cfg(feature = "full")] ++use crate::{Block, Stmt}; + + #[cfg(any(feature = "full", feature = "derive"))] + impl ParseQuote for Attribute { +@@ -129,3 +133,10 @@ impl<T: Parse, P: Parse> ParseQuote for Punctuated<T, P> { + Self::parse_terminated(input) + } + } ++ ++#[cfg(feature = "full")] ++impl ParseQuote for Vec<Stmt> { ++ fn parse(input: ParseStream) -> Result<Self> { ++ Block::parse_within(input) ++ } ++} +diff --git a/third_party/rust/syn/src/pat.rs b/third_party/rust/syn/src/pat.rs +index 9371e05493..e9576a2361 100644 +--- a/third_party/rust/syn/src/pat.rs ++++ b/third_party/rust/syn/src/pat.rs +@@ -1,16 +1,12 @@ + use super::*; + use crate::punctuated::Punctuated; +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; + use proc_macro2::TokenStream; +-#[cfg(feature = "extra-traits")] +-use std::hash::{Hash, Hasher}; + + ast_enum_of_structs! { + /// A pattern in a local binding, function signature, match expression, or + /// various other places. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -20,7 +16,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum Pat #manual_extra_traits { ++ pub enum Pat { + /// A box pattern: `box v`. + Box(PatBox), + +@@ -86,7 +82,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A box pattern: `box v`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatBox { + pub attrs: Vec<Attribute>, + pub box_token: Token![box], +@@ -97,7 +93,10 @@ ast_struct! { + ast_struct! { + /// A pattern that binds a new variable: `ref mut binding @ SUBPATTERN`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// It may also be a unit struct or struct variant (e.g. `None`), or a ++ /// constant; these cannot be distinguished syntactically. ++ /// ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatIdent { + pub attrs: Vec<Attribute>, + pub by_ref: Option<Token![ref]>, +@@ -113,7 +112,7 @@ ast_struct! { + /// This holds an `Expr` rather than a `Lit` because negative numbers + /// are represented as an `Expr::Unary`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatLit { + pub attrs: Vec<Attribute>, + pub expr: Box<Expr>, +@@ -123,7 +122,7 @@ ast_struct! { + ast_struct! { + /// A macro in pattern position. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatMacro { + pub attrs: Vec<Attribute>, + pub mac: Macro, +@@ -133,7 +132,7 @@ ast_struct! { + ast_struct! { + /// A pattern that matches any one of a set of cases. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatOr { + pub attrs: Vec<Attribute>, + pub leading_vert: Option<Token![|]>, +@@ -150,7 +149,7 @@ ast_struct! { + /// `<A>::B::C` and `<A as Trait>::B::C` can only legally refer to + /// associated constants. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatPath { + pub attrs: Vec<Attribute>, + pub qself: Option<QSelf>, +@@ -161,7 +160,7 @@ ast_struct! { + ast_struct! { + /// A range pattern: `1..=2`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatRange { + pub attrs: Vec<Attribute>, + pub lo: Box<Expr>, +@@ -173,7 +172,7 @@ ast_struct! { + ast_struct! { + /// A reference pattern: `&mut var`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatReference { + pub attrs: Vec<Attribute>, + pub and_token: Token![&], +@@ -185,7 +184,7 @@ ast_struct! { + ast_struct! { + /// The dots in a tuple or slice pattern: `[0, 1, ..]` + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatRest { + pub attrs: Vec<Attribute>, + pub dot2_token: Token![..], +@@ -195,7 +194,7 @@ ast_struct! { + ast_struct! { + /// A dynamically sized slice pattern: `[a, b, ref i @ .., y, z]`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatSlice { + pub attrs: Vec<Attribute>, + pub bracket_token: token::Bracket, +@@ -206,7 +205,7 @@ ast_struct! { + ast_struct! { + /// A struct or struct variant pattern: `Variant { x, y, .. }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatStruct { + pub attrs: Vec<Attribute>, + pub path: Path, +@@ -219,7 +218,7 @@ ast_struct! { + ast_struct! { + /// A tuple pattern: `(a, b)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatTuple { + pub attrs: Vec<Attribute>, + pub paren_token: token::Paren, +@@ -230,7 +229,7 @@ ast_struct! { + ast_struct! { + /// A tuple struct or tuple variant pattern: `Variant(x, y, .., z)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatTupleStruct { + pub attrs: Vec<Attribute>, + pub path: Path, +@@ -241,7 +240,7 @@ ast_struct! { + ast_struct! { + /// A type ascription pattern: `foo: f64`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatType { + pub attrs: Vec<Attribute>, + pub pat: Box<Pat>, +@@ -253,7 +252,7 @@ ast_struct! { + ast_struct! { + /// A pattern that matches any value: `_`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatWild { + pub attrs: Vec<Attribute>, + pub underscore_token: Token![_], +@@ -266,7 +265,7 @@ ast_struct! { + /// Patterns like the fields of Foo `{ x, ref y, ref mut z }` are treated + /// the same as `x: x, y: ref y, z: ref mut z` but there is no colon token. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct FieldPat { + pub attrs: Vec<Attribute>, + pub member: Member, +@@ -275,122 +274,17 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Pat {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Pat { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (Pat::Box(this), Pat::Box(other)) => this == other, +- (Pat::Ident(this), Pat::Ident(other)) => this == other, +- (Pat::Lit(this), Pat::Lit(other)) => this == other, +- (Pat::Macro(this), Pat::Macro(other)) => this == other, +- (Pat::Or(this), Pat::Or(other)) => this == other, +- (Pat::Path(this), Pat::Path(other)) => this == other, +- (Pat::Range(this), Pat::Range(other)) => this == other, +- (Pat::Reference(this), Pat::Reference(other)) => this == other, +- (Pat::Rest(this), Pat::Rest(other)) => this == other, +- (Pat::Slice(this), Pat::Slice(other)) => this == other, +- (Pat::Struct(this), Pat::Struct(other)) => this == other, +- (Pat::Tuple(this), Pat::Tuple(other)) => this == other, +- (Pat::TupleStruct(this), Pat::TupleStruct(other)) => this == other, +- (Pat::Type(this), Pat::Type(other)) => this == other, +- (Pat::Verbatim(this), Pat::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- (Pat::Wild(this), Pat::Wild(other)) => this == other, +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Pat { +- fn hash<H>(&self, hash: &mut H) +- where +- H: Hasher, +- { +- match self { +- Pat::Box(pat) => { +- hash.write_u8(0); +- pat.hash(hash); +- } +- Pat::Ident(pat) => { +- hash.write_u8(1); +- pat.hash(hash); +- } +- Pat::Lit(pat) => { +- hash.write_u8(2); +- pat.hash(hash); +- } +- Pat::Macro(pat) => { +- hash.write_u8(3); +- pat.hash(hash); +- } +- Pat::Or(pat) => { +- hash.write_u8(4); +- pat.hash(hash); +- } +- Pat::Path(pat) => { +- hash.write_u8(5); +- pat.hash(hash); +- } +- Pat::Range(pat) => { +- hash.write_u8(6); +- pat.hash(hash); +- } +- Pat::Reference(pat) => { +- hash.write_u8(7); +- pat.hash(hash); +- } +- Pat::Rest(pat) => { +- hash.write_u8(8); +- pat.hash(hash); +- } +- Pat::Slice(pat) => { +- hash.write_u8(9); +- pat.hash(hash); +- } +- Pat::Struct(pat) => { +- hash.write_u8(10); +- pat.hash(hash); +- } +- Pat::Tuple(pat) => { +- hash.write_u8(11); +- pat.hash(hash); +- } +- Pat::TupleStruct(pat) => { +- hash.write_u8(12); +- pat.hash(hash); +- } +- Pat::Type(pat) => { +- hash.write_u8(13); +- pat.hash(hash); +- } +- Pat::Verbatim(pat) => { +- hash.write_u8(14); +- TokenStreamHelper(pat).hash(hash); +- } +- Pat::Wild(pat) => { +- hash.write_u8(15); +- pat.hash(hash); +- } +- Pat::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + #[cfg(feature = "parsing")] +-mod parsing { ++pub mod parsing { + use super::*; + + use crate::ext::IdentExt; +- use crate::parse::{Parse, ParseStream, Result}; ++ use crate::parse::{Parse, ParseBuffer, ParseStream, Result}; + use crate::path; + + impl Parse for Pat { + fn parse(input: ParseStream) -> Result<Self> { ++ let begin = input.fork(); + let lookahead = input.lookahead1(); + if lookahead.peek(Ident) + && ({ +@@ -411,7 +305,6 @@ mod parsing { + || lookahead.peek(Token![<]) + || input.peek(Token![Self]) + || input.peek(Token![super]) +- || input.peek(Token![extern]) + || input.peek(Token![crate]) + { + pat_path_or_macro_or_struct_or_range(input) +@@ -434,7 +327,7 @@ mod parsing { + } else if lookahead.peek(token::Bracket) { + input.call(pat_slice).map(Pat::Slice) + } else if lookahead.peek(Token![..]) && !input.peek(Token![...]) { +- input.call(pat_rest).map(Pat::Rest) ++ pat_range_half_open(input, begin) + } else { + Err(lookahead.error()) + } +@@ -442,10 +335,11 @@ mod parsing { + } + + fn pat_path_or_macro_or_struct_or_range(input: ParseStream) -> Result<Pat> { ++ let begin = input.fork(); + let (qself, path) = path::parsing::qpath(input, true)?; + + if input.peek(Token![..]) { +- return pat_range(input, qself, path).map(Pat::Range); ++ return pat_range(input, begin, qself, path); + } + + if qself.is_some() { +@@ -487,7 +381,7 @@ mod parsing { + } else if input.peek(token::Paren) { + pat_tuple_struct(input, path).map(Pat::TupleStruct) + } else if input.peek(Token![..]) { +- pat_range(input, qself, path).map(Pat::Range) ++ pat_range(input, begin, qself, path) + } else { + Ok(Pat::Path(PatPath { + attrs: Vec::new(), +@@ -546,7 +440,7 @@ mod parsing { + while !content.is_empty() && !content.peek(Token![..]) { + let value = content.call(field_pat)?; + fields.push_value(value); +- if !content.peek(Token![,]) { ++ if content.is_empty() { + break; + } + let punct: Token![,] = content.parse()?; +@@ -578,6 +472,7 @@ mod parsing { + } + + fn field_pat(input: ParseStream) -> Result<FieldPat> { ++ let attrs = input.call(Attribute::parse_outer)?; + let boxed: Option<Token![box]> = input.parse()?; + let by_ref: Option<Token![ref]> = input.parse()?; + let mutability: Option<Token![mut]> = input.parse()?; +@@ -587,10 +482,10 @@ mod parsing { + || member.is_unnamed() + { + return Ok(FieldPat { +- attrs: Vec::new(), ++ attrs, + member, + colon_token: input.parse()?, +- pat: input.parse()?, ++ pat: Box::new(multi_pat(input)?), + }); + } + +@@ -610,30 +505,57 @@ mod parsing { + if let Some(boxed) = boxed { + pat = Pat::Box(PatBox { + attrs: Vec::new(), +- pat: Box::new(pat), + box_token: boxed, ++ pat: Box::new(pat), + }); + } + + Ok(FieldPat { ++ attrs, + member: Member::Named(ident), +- pat: Box::new(pat), +- attrs: Vec::new(), + colon_token: None, ++ pat: Box::new(pat), + }) + } + +- fn pat_range(input: ParseStream, qself: Option<QSelf>, path: Path) -> Result<PatRange> { +- Ok(PatRange { +- attrs: Vec::new(), +- lo: Box::new(Expr::Path(ExprPath { ++ fn pat_range( ++ input: ParseStream, ++ begin: ParseBuffer, ++ qself: Option<QSelf>, ++ path: Path, ++ ) -> Result<Pat> { ++ let limits: RangeLimits = input.parse()?; ++ let hi = input.call(pat_lit_expr)?; ++ if let Some(hi) = hi { ++ Ok(Pat::Range(PatRange { + attrs: Vec::new(), +- qself, +- path, +- })), +- limits: input.parse()?, +- hi: input.call(pat_lit_expr)?, +- }) ++ lo: Box::new(Expr::Path(ExprPath { ++ attrs: Vec::new(), ++ qself, ++ path, ++ })), ++ limits, ++ hi, ++ })) ++ } else { ++ Ok(Pat::Verbatim(verbatim::between(begin, input))) ++ } ++ } ++ ++ fn pat_range_half_open(input: ParseStream, begin: ParseBuffer) -> Result<Pat> { ++ let limits: RangeLimits = input.parse()?; ++ let hi = input.call(pat_lit_expr)?; ++ if hi.is_some() { ++ Ok(Pat::Verbatim(verbatim::between(begin, input))) ++ } else { ++ match limits { ++ RangeLimits::HalfOpen(dot2_token) => Ok(Pat::Rest(PatRest { ++ attrs: Vec::new(), ++ dot2_token, ++ })), ++ RangeLimits::Closed(_) => Err(input.error("expected range upper bound")), ++ } ++ } + } + + fn pat_tuple(input: ParseStream) -> Result<PatTuple> { +@@ -642,7 +564,7 @@ mod parsing { + + let mut elems = Punctuated::new(); + while !content.is_empty() { +- let value: Pat = content.parse()?; ++ let value = multi_pat(&content)?; + elems.push_value(value); + if content.is_empty() { + break; +@@ -668,14 +590,21 @@ mod parsing { + } + + fn pat_lit_or_range(input: ParseStream) -> Result<Pat> { +- let lo = input.call(pat_lit_expr)?; ++ let begin = input.fork(); ++ let lo = input.call(pat_lit_expr)?.unwrap(); + if input.peek(Token![..]) { +- Ok(Pat::Range(PatRange { +- attrs: Vec::new(), +- lo, +- limits: input.parse()?, +- hi: input.call(pat_lit_expr)?, +- })) ++ let limits: RangeLimits = input.parse()?; ++ let hi = input.call(pat_lit_expr)?; ++ if let Some(hi) = hi { ++ Ok(Pat::Range(PatRange { ++ attrs: Vec::new(), ++ lo, ++ limits, ++ hi, ++ })) ++ } else { ++ Ok(Pat::Verbatim(verbatim::between(begin, input))) ++ } + } else { + Ok(Pat::Lit(PatLit { + attrs: Vec::new(), +@@ -684,7 +613,17 @@ mod parsing { + } + } + +- fn pat_lit_expr(input: ParseStream) -> Result<Box<Expr>> { ++ fn pat_lit_expr(input: ParseStream) -> Result<Option<Box<Expr>>> { ++ if input.is_empty() ++ || input.peek(Token![|]) ++ || input.peek(Token![=>]) ++ || input.peek(Token![:]) && !input.peek(Token![::]) ++ || input.peek(Token![,]) ++ || input.peek(Token![;]) ++ { ++ return Ok(None); ++ } ++ + let neg: Option<Token![-]> = input.parse()?; + + let lookahead = input.lookahead1(); +@@ -696,7 +635,6 @@ mod parsing { + || lookahead.peek(Token![self]) + || lookahead.peek(Token![Self]) + || lookahead.peek(Token![super]) +- || lookahead.peek(Token![extern]) + || lookahead.peek(Token![crate]) + { + Expr::Path(input.parse()?) +@@ -704,7 +642,7 @@ mod parsing { + return Err(lookahead.error()); + }; + +- Ok(Box::new(if let Some(neg) = neg { ++ Ok(Some(Box::new(if let Some(neg) = neg { + Expr::Unary(ExprUnary { + attrs: Vec::new(), + op: UnOp::Neg(neg), +@@ -712,7 +650,7 @@ mod parsing { + }) + } else { + expr +- })) ++ }))) + } + + fn pat_slice(input: ParseStream) -> Result<PatSlice> { +@@ -721,7 +659,7 @@ mod parsing { + + let mut elems = Punctuated::new(); + while !content.is_empty() { +- let value: Pat = content.parse()?; ++ let value = multi_pat(&content)?; + elems.push_value(value); + if content.is_empty() { + break; +@@ -737,11 +675,35 @@ mod parsing { + }) + } + +- fn pat_rest(input: ParseStream) -> Result<PatRest> { +- Ok(PatRest { +- attrs: Vec::new(), +- dot2_token: input.parse()?, +- }) ++ pub fn multi_pat(input: ParseStream) -> Result<Pat> { ++ multi_pat_impl(input, None) ++ } ++ ++ pub fn multi_pat_with_leading_vert(input: ParseStream) -> Result<Pat> { ++ let leading_vert: Option<Token![|]> = input.parse()?; ++ multi_pat_impl(input, leading_vert) ++ } ++ ++ fn multi_pat_impl(input: ParseStream, leading_vert: Option<Token![|]>) -> Result<Pat> { ++ let mut pat: Pat = input.parse()?; ++ if leading_vert.is_some() ++ || input.peek(Token![|]) && !input.peek(Token![||]) && !input.peek(Token![|=]) ++ { ++ let mut cases = Punctuated::new(); ++ cases.push_value(pat); ++ while input.peek(Token![|]) && !input.peek(Token![||]) && !input.peek(Token![|=]) { ++ let punct = input.parse()?; ++ cases.push_punct(punct); ++ let pat: Pat = input.parse()?; ++ cases.push_value(pat); ++ } ++ pat = Pat::Or(PatOr { ++ attrs: Vec::new(), ++ leading_vert, ++ cases, ++ }); ++ } ++ Ok(pat) + } + } + +@@ -756,12 +718,14 @@ mod printing { + + impl ToTokens for PatWild { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.underscore_token.to_tokens(tokens); + } + } + + impl ToTokens for PatIdent { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.by_ref.to_tokens(tokens); + self.mutability.to_tokens(tokens); + self.ident.to_tokens(tokens); +@@ -774,6 +738,7 @@ mod printing { + + impl ToTokens for PatStruct { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.path.to_tokens(tokens); + self.brace_token.surround(tokens, |tokens| { + self.fields.to_tokens(tokens); +@@ -788,6 +753,7 @@ mod printing { + + impl ToTokens for PatTupleStruct { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.path.to_tokens(tokens); + self.pat.to_tokens(tokens); + } +@@ -804,12 +770,14 @@ mod printing { + + impl ToTokens for PatPath { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + private::print_path(tokens, &self.qself, &self.path); + } + } + + impl ToTokens for PatTuple { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.paren_token.surround(tokens, |tokens| { + self.elems.to_tokens(tokens); + }); +@@ -818,6 +786,7 @@ mod printing { + + impl ToTokens for PatBox { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.box_token.to_tokens(tokens); + self.pat.to_tokens(tokens); + } +@@ -825,6 +794,7 @@ mod printing { + + impl ToTokens for PatReference { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.and_token.to_tokens(tokens); + self.mutability.to_tokens(tokens); + self.pat.to_tokens(tokens); +@@ -833,18 +803,21 @@ mod printing { + + impl ToTokens for PatRest { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.dot2_token.to_tokens(tokens); + } + } + + impl ToTokens for PatLit { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.expr.to_tokens(tokens); + } + } + + impl ToTokens for PatRange { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.lo.to_tokens(tokens); + match &self.limits { + RangeLimits::HalfOpen(t) => t.to_tokens(tokens), +@@ -856,6 +829,7 @@ mod printing { + + impl ToTokens for PatSlice { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.bracket_token.surround(tokens, |tokens| { + self.elems.to_tokens(tokens); + }); +@@ -864,12 +838,14 @@ mod printing { + + impl ToTokens for PatMacro { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.mac.to_tokens(tokens); + } + } + + impl ToTokens for PatOr { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.leading_vert.to_tokens(tokens); + self.cases.to_tokens(tokens); + } +@@ -877,6 +853,7 @@ mod printing { + + impl ToTokens for FieldPat { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + if let Some(colon_token) = &self.colon_token { + self.member.to_tokens(tokens); + colon_token.to_tokens(tokens); +diff --git a/third_party/rust/syn/src/path.rs b/third_party/rust/syn/src/path.rs +index 8dda43ee67..15c0fcc664 100644 +--- a/third_party/rust/syn/src/path.rs ++++ b/third_party/rust/syn/src/path.rs +@@ -2,9 +2,9 @@ use super::*; + use crate::punctuated::Punctuated; + + ast_struct! { +- /// A path at which a named item is exported: `std::collections::HashMap`. ++ /// A path at which a named item is exported (e.g. `std::collections::HashMap`). + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Path { + pub leading_colon: Option<Token![::]>, +@@ -29,7 +29,7 @@ where + ast_struct! { + /// A segment of a path together with any path arguments on that segment. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct PathSegment { + pub ident: Ident, +@@ -52,7 +52,7 @@ where + ast_enum! { + /// Angle bracketed or parenthesized arguments of a path segment. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// ## Angle bracketed +@@ -98,7 +98,7 @@ impl PathArguments { + ast_enum! { + /// An individual generic argument, like `'a`, `T`, or `Item = T`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum GenericArgument { + /// A lifetime argument. +@@ -122,7 +122,7 @@ ast_struct! { + /// Angle bracketed arguments of a path segment: the `<K, V>` in `HashMap<K, + /// V>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct AngleBracketedGenericArguments { + pub colon2_token: Option<Token![::]>, +@@ -135,7 +135,7 @@ ast_struct! { + ast_struct! { + /// A binding (equality constraint) on an associated type: `Item = u8`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Binding { + pub ident: Ident, +@@ -147,7 +147,7 @@ ast_struct! { + ast_struct! { + /// An associated type bound: `Iterator<Item: Display>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Constraint { + pub ident: Ident, +@@ -160,7 +160,7 @@ ast_struct! { + /// Arguments of a function path segment: the `(A, B) -> C` in `Fn(A,B) -> + /// C`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct ParenthesizedGenericArguments { + pub paren_token: token::Paren, +@@ -189,7 +189,7 @@ ast_struct! { + /// ty position = 0 + /// ``` + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct QSelf { + pub lt_token: Token![<], +@@ -291,11 +291,7 @@ pub mod parsing { + + impl PathSegment { + fn parse_helper(input: ParseStream, expr_style: bool) -> Result<Self> { +- if input.peek(Token![super]) +- || input.peek(Token![self]) +- || input.peek(Token![crate]) +- || input.peek(Token![extern]) +- { ++ if input.peek(Token![super]) || input.peek(Token![self]) || input.peek(Token![crate]) { + let ident = input.call(Ident::parse_any)?; + return Ok(PathSegment::from(ident)); + } +@@ -358,7 +354,7 @@ pub mod parsing { + impl Path { + /// Parse a `Path` containing no path arguments on any of its segments. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + /// + /// # Example +@@ -400,7 +396,6 @@ pub mod parsing { + && !input.peek(Token![self]) + && !input.peek(Token![Self]) + && !input.peek(Token![crate]) +- && !input.peek(Token![extern]) + { + break; + } +@@ -433,7 +428,7 @@ pub mod parsing { + /// path arguments, and + /// - the ident of the first path segment is equal to the given one. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + /// + /// # Example +@@ -472,7 +467,7 @@ pub mod parsing { + /// - the first path segment has no angle bracketed or parenthesized + /// path arguments. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + pub fn get_ident(&self) -> Option<&Ident> { + if self.leading_colon.is_none() +diff --git a/third_party/rust/syn/src/punctuated.rs b/third_party/rust/syn/src/punctuated.rs +index 38c7bf4e82..46c82a65b1 100644 +--- a/third_party/rust/syn/src/punctuated.rs ++++ b/third_party/rust/syn/src/punctuated.rs +@@ -22,6 +22,8 @@ + + #[cfg(feature = "extra-traits")] + use std::fmt::{self, Debug}; ++#[cfg(feature = "extra-traits")] ++use std::hash::{Hash, Hasher}; + #[cfg(any(feature = "full", feature = "derive"))] + use std::iter; + use std::iter::FromIterator; +@@ -41,8 +43,6 @@ use crate::token::Token; + /// Refer to the [module documentation] for details about punctuated sequences. + /// + /// [module documentation]: self +-#[cfg_attr(feature = "extra-traits", derive(Eq, PartialEq, Hash))] +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + pub struct Punctuated<T, P> { + inner: Vec<(T, P)>, + last: Option<Box<T>>, +@@ -76,22 +76,19 @@ impl<T, P> Punctuated<T, P> { + self.iter().next() + } + ++ /// Mutably borrows the first element in this sequence. ++ pub fn first_mut(&mut self) -> Option<&mut T> { ++ self.iter_mut().next() ++ } ++ + /// Borrows the last element in this sequence. + pub fn last(&self) -> Option<&T> { +- if self.last.is_some() { +- self.last.as_ref().map(Box::as_ref) +- } else { +- self.inner.last().map(|pair| &pair.0) +- } ++ self.iter().next_back() + } + + /// Mutably borrows the last element in this sequence. + pub fn last_mut(&mut self) -> Option<&mut T> { +- if self.last.is_some() { +- self.last.as_mut().map(Box::as_mut) +- } else { +- self.inner.last_mut().map(|pair| &mut pair.0) +- } ++ self.iter_mut().next_back() + } + + /// Returns an iterator over borrowed syntax tree nodes of type `&T`. +@@ -230,13 +227,19 @@ impl<T, P> Punctuated<T, P> { + } + } + ++ /// Clears the sequence of all values and punctuation, making it empty. ++ pub fn clear(&mut self) { ++ self.inner.clear(); ++ self.last = None; ++ } ++ + /// Parses zero or more occurrences of `T` separated by punctuation of type + /// `P`, with optional trailing punctuation. + /// + /// Parsing continues until the end of this parse stream. The entire content + /// of this parse stream must consist of `T` and `P`. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_terminated(input: ParseStream) -> Result<Self> +@@ -256,7 +259,7 @@ impl<T, P> Punctuated<T, P> { + /// + /// [`parse_terminated`]: Punctuated::parse_terminated + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_terminated_with( +@@ -292,7 +295,7 @@ impl<T, P> Punctuated<T, P> { + /// is not followed by a `P`, even if there are remaining tokens in the + /// stream. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_separated_nonempty(input: ParseStream) -> Result<Self> +@@ -312,7 +315,7 @@ impl<T, P> Punctuated<T, P> { + /// + /// [`parse_separated_nonempty`]: Punctuated::parse_separated_nonempty + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_separated_nonempty_with( +@@ -338,6 +341,53 @@ impl<T, P> Punctuated<T, P> { + } + } + ++#[cfg(feature = "clone-impls")] ++impl<T, P> Clone for Punctuated<T, P> ++where ++ T: Clone, ++ P: Clone, ++{ ++ fn clone(&self) -> Self { ++ Punctuated { ++ inner: self.inner.clone(), ++ last: self.last.clone(), ++ } ++ } ++} ++ ++#[cfg(feature = "extra-traits")] ++impl<T, P> Eq for Punctuated<T, P> ++where ++ T: Eq, ++ P: Eq, ++{ ++} ++ ++#[cfg(feature = "extra-traits")] ++impl<T, P> PartialEq for Punctuated<T, P> ++where ++ T: PartialEq, ++ P: PartialEq, ++{ ++ fn eq(&self, other: &Self) -> bool { ++ let Punctuated { inner, last } = self; ++ *inner == other.inner && *last == other.last ++ } ++} ++ ++#[cfg(feature = "extra-traits")] ++impl<T, P> Hash for Punctuated<T, P> ++where ++ T: Hash, ++ P: Hash, ++{ ++ fn hash<H: Hasher>(&self, state: &mut H) { ++ let Punctuated { inner, last } = self; ++ inner.hash(state); ++ last.hash(state); ++ } ++} ++ + #[cfg(feature = "extra-traits")] + impl<T: Debug, P: Debug> Debug for Punctuated<T, P> { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +@@ -536,7 +586,6 @@ impl<'a, T, P> ExactSizeIterator for PairsMut<'a, T, P> { + /// Refer to the [module documentation] for details about punctuated sequences. + /// + /// [module documentation]: self +-#[derive(Clone)] + pub struct IntoPairs<T, P> { + inner: vec::IntoIter<(T, P)>, + last: option::IntoIter<T>, +@@ -572,12 +621,24 @@ impl<T, P> ExactSizeIterator for IntoPairs<T, P> { + } + } + ++impl<T, P> Clone for IntoPairs<T, P> ++where ++ T: Clone, ++ P: Clone, ++{ ++ fn clone(&self) -> Self { ++ IntoPairs { ++ inner: self.inner.clone(), ++ last: self.last.clone(), ++ } ++ } ++} ++ + /// An iterator over owned values of type `T`. + /// + /// Refer to the [module documentation] for details about punctuated sequences. + /// + /// [module documentation]: self +-#[derive(Clone)] + pub struct IntoIter<T> { + inner: vec::IntoIter<T>, + } +@@ -606,6 +667,17 @@ impl<T> ExactSizeIterator for IntoIter<T> { + } + } + ++impl<T> Clone for IntoIter<T> ++where ++ T: Clone, ++{ ++ fn clone(&self) -> Self { ++ IntoIter { ++ inner: self.inner.clone(), ++ } ++ } ++} ++ + /// An iterator over borrowed values of type `&T`. + /// + /// Refer to the [module documentation] for details about punctuated sequences. +@@ -799,7 +871,6 @@ impl<'a, T: 'a, I: 'a> IterMutTrait<'a, T> for I where + /// Refer to the [module documentation] for details about punctuated sequences. + /// + /// [module documentation]: self +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + pub enum Pair<T, P> { + Punctuated(T, P), + End(T), +@@ -856,6 +927,20 @@ impl<T, P> Pair<T, P> { + } + } + ++#[cfg(feature = "clone-impls")] ++impl<T, P> Clone for Pair<T, P> ++where ++ T: Clone, ++ P: Clone, ++{ ++ fn clone(&self) -> Self { ++ match self { ++ Pair::Punctuated(t, p) => Pair::Punctuated(t.clone(), p.clone()), ++ Pair::End(t) => Pair::End(t.clone()), ++ } ++ } ++} ++ + impl<T, P> Index<usize> for Punctuated<T, P> { + type Output = T; + +diff --git a/third_party/rust/syn/src/reserved.rs b/third_party/rust/syn/src/reserved.rs +new file mode 100644 +index 0000000000..ccfb8b5ad0 +--- /dev/null ++++ b/third_party/rust/syn/src/reserved.rs +@@ -0,0 +1,42 @@ ++// Type for a syntax tree node that is reserved for future use. ++// ++// For example ExprReference contains a field `raw` of type Reserved. If `&raw ++// place` syntax becomes a thing as per https://github.com/rust-lang/rfcs/pull/2582, ++// we can backward compatibly change `raw`'s type to Option<Token![raw]> without ++// the possibility of breaking any code. ++ ++use proc_macro2::Span; ++use std::marker::PhantomData; ++ ++#[cfg(feature = "extra-traits")] ++use std::fmt::{self, Debug}; ++ ++ast_struct! { ++ pub struct Reserved { ++ _private: PhantomData<Span>, ++ } ++} ++ ++impl Default for Reserved { ++ fn default() -> Self { ++ Reserved { ++ _private: PhantomData, ++ } ++ } ++} ++ ++#[cfg(feature = "clone-impls")] ++impl Clone for Reserved { ++ fn clone(&self) -> Self { ++ Reserved { ++ _private: self._private, ++ } ++ } ++} ++ ++#[cfg(feature = "extra-traits")] ++impl Debug for Reserved { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ formatter.debug_struct("Reserved").finish() ++ } ++} +diff --git a/third_party/rust/syn/src/spanned.rs b/third_party/rust/syn/src/spanned.rs +index 71ffe26b81..01591cedcb 100644 +--- a/third_party/rust/syn/src/spanned.rs ++++ b/third_party/rust/syn/src/spanned.rs +@@ -1,7 +1,7 @@ + //! A trait that can provide the `Span` of the complete contents of a syntax + //! tree node. + //! +-//! *This module is available if Syn is built with both the `"parsing"` and ++//! *This module is available only if Syn is built with both the `"parsing"` and + //! `"printing"` features.* + //! + //! <br> +@@ -97,7 +97,7 @@ use quote::spanned::Spanned as ToTokens; + /// + /// [module documentation]: self + /// +-/// *This trait is available if Syn is built with both the `"parsing"` and ++/// *This trait is available only if Syn is built with both the `"parsing"` and + /// `"printing"` features.* + pub trait Spanned { + /// Returns a `Span` covering the complete contents of this syntax tree +diff --git a/third_party/rust/syn/src/stmt.rs b/third_party/rust/syn/src/stmt.rs +index e4277fdbaa..b06e843d75 100644 +--- a/third_party/rust/syn/src/stmt.rs ++++ b/third_party/rust/syn/src/stmt.rs +@@ -3,7 +3,7 @@ use super::*; + ast_struct! { + /// A braced block containing Rust statements. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Block { + pub brace_token: token::Brace, + /// Statements in a block +@@ -14,7 +14,7 @@ ast_struct! { + ast_enum! { + /// A statement, usually ending in a semicolon. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub enum Stmt { + /// A local (let) binding. + Local(Local), +@@ -33,7 +33,7 @@ ast_enum! { + ast_struct! { + /// A local `let` binding: `let x: u64 = s.parse()?`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Local { + pub attrs: Vec<Attribute>, + pub let_token: Token![let], +@@ -47,14 +47,15 @@ ast_struct! { + pub mod parsing { + use super::*; + ++ use crate::parse::discouraged::Speculative; + use crate::parse::{Parse, ParseStream, Result}; +- use crate::punctuated::Punctuated; ++ use proc_macro2::TokenStream; + + impl Block { + /// Parse the body of a block as zero or more statements, possibly + /// including one trailing expression. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + /// + /// # Example +@@ -106,8 +107,8 @@ pub mod parsing { + pub fn parse_within(input: ParseStream) -> Result<Vec<Stmt>> { + let mut stmts = Vec::new(); + loop { +- while input.peek(Token![;]) { +- input.parse::<Token![;]>()?; ++ while let Some(semi) = input.parse::<Option<Token![;]>>()? { ++ stmts.push(Stmt::Semi(Expr::Verbatim(TokenStream::new()), semi)); + } + if input.is_empty() { + break; +@@ -146,55 +147,55 @@ pub mod parsing { + } + + fn parse_stmt(input: ParseStream, allow_nosemi: bool) -> Result<Stmt> { +- // TODO: optimize using advance_to ++ let mut attrs = input.call(Attribute::parse_outer)?; ++ ++ // brace-style macros; paren and bracket macros get parsed as ++ // expression statements. + let ahead = input.fork(); +- ahead.call(Attribute::parse_outer)?; ++ if let Ok(path) = ahead.call(Path::parse_mod_style) { ++ if ahead.peek(Token![!]) && (ahead.peek2(token::Brace) || ahead.peek2(Ident)) { ++ input.advance_to(&ahead); ++ return stmt_mac(input, attrs, path); ++ } ++ } + +- if { +- let ahead = ahead.fork(); +- // Only parse braces here; paren and bracket will get parsed as +- // expression statements +- ahead.call(Path::parse_mod_style).is_ok() +- && ahead.parse::<Token![!]>().is_ok() +- && (ahead.peek(token::Brace) || ahead.peek(Ident)) +- } { +- stmt_mac(input) +- } else if ahead.peek(Token![let]) { +- stmt_local(input).map(Stmt::Local) +- } else if ahead.peek(Token![pub]) +- || ahead.peek(Token![crate]) && !ahead.peek2(Token![::]) +- || ahead.peek(Token![extern]) && !ahead.peek2(Token![::]) +- || ahead.peek(Token![use]) +- || ahead.peek(Token![static]) && (ahead.peek2(Token![mut]) || ahead.peek2(Ident)) +- || ahead.peek(Token![const]) +- || ahead.peek(Token![unsafe]) && !ahead.peek2(token::Brace) +- || ahead.peek(Token![async]) +- && (ahead.peek2(Token![unsafe]) +- || ahead.peek2(Token![extern]) +- || ahead.peek2(Token![fn])) +- || ahead.peek(Token![fn]) +- || ahead.peek(Token![mod]) +- || ahead.peek(Token![type]) +- || ahead.peek(item::parsing::existential) && ahead.peek2(Token![type]) +- || ahead.peek(Token![struct]) +- || ahead.peek(Token![enum]) +- || ahead.peek(Token![union]) && ahead.peek2(Ident) +- || ahead.peek(Token![auto]) && ahead.peek2(Token![trait]) +- || ahead.peek(Token![trait]) +- || ahead.peek(Token![default]) +- && (ahead.peek2(Token![unsafe]) || ahead.peek2(Token![impl])) +- || ahead.peek(Token![impl]) +- || ahead.peek(Token![macro]) ++ if input.peek(Token![let]) { ++ stmt_local(input, attrs).map(Stmt::Local) ++ } else if input.peek(Token![pub]) ++ || input.peek(Token![crate]) && !input.peek2(Token![::]) ++ || input.peek(Token![extern]) ++ || input.peek(Token![use]) ++ || input.peek(Token![static]) && (input.peek2(Token![mut]) || input.peek2(Ident)) ++ || input.peek(Token![const]) ++ || input.peek(Token![unsafe]) && !input.peek2(token::Brace) ++ || input.peek(Token![async]) ++ && (input.peek2(Token![unsafe]) ++ || input.peek2(Token![extern]) ++ || input.peek2(Token![fn])) ++ || input.peek(Token![fn]) ++ || input.peek(Token![mod]) ++ || input.peek(Token![type]) ++ || input.peek(item::parsing::existential) && input.peek2(Token![type]) ++ || input.peek(Token![struct]) ++ || input.peek(Token![enum]) ++ || input.peek(Token![union]) && input.peek2(Ident) ++ || input.peek(Token![auto]) && input.peek2(Token![trait]) ++ || input.peek(Token![trait]) ++ || input.peek(Token![default]) ++ && (input.peek2(Token![unsafe]) || input.peek2(Token![impl])) ++ || input.peek(Token![impl]) ++ || input.peek(Token![macro]) + { +- input.parse().map(Stmt::Item) ++ let mut item: Item = input.parse()?; ++ attrs.extend(item.replace_attrs(Vec::new())); ++ item.replace_attrs(attrs); ++ Ok(Stmt::Item(item)) + } else { +- stmt_expr(input, allow_nosemi) ++ stmt_expr(input, allow_nosemi, attrs) + } + } + +- fn stmt_mac(input: ParseStream) -> Result<Stmt> { +- let attrs = input.call(Attribute::parse_outer)?; +- let path = input.call(Path::parse_mod_style)?; ++ fn stmt_mac(input: ParseStream, attrs: Vec<Attribute>, path: Path) -> Result<Stmt> { + let bang_token: Token![!] = input.parse()?; + let ident: Option<Ident> = input.parse()?; + let (delimiter, tokens) = mac::parse_delimiter(input)?; +@@ -213,33 +214,12 @@ pub mod parsing { + }))) + } + +- fn stmt_local(input: ParseStream) -> Result<Local> { ++ fn stmt_local(input: ParseStream, attrs: Vec<Attribute>) -> Result<Local> { + Ok(Local { +- attrs: input.call(Attribute::parse_outer)?, ++ attrs, + let_token: input.parse()?, + pat: { +- let leading_vert: Option<Token![|]> = input.parse()?; +- let mut pat: Pat = input.parse()?; +- if leading_vert.is_some() +- || input.peek(Token![|]) && !input.peek(Token![||]) && !input.peek(Token![|=]) +- { +- let mut cases = Punctuated::new(); +- cases.push_value(pat); +- while input.peek(Token![|]) +- && !input.peek(Token![||]) +- && !input.peek(Token![|=]) +- { +- let punct = input.parse()?; +- cases.push_punct(punct); +- let pat: Pat = input.parse()?; +- cases.push_value(pat); +- } +- pat = Pat::Or(PatOr { +- attrs: Vec::new(), +- leading_vert, +- cases, +- }); +- } ++ let mut pat: Pat = pat::parsing::multi_pat_with_leading_vert(input)?; + if input.peek(Token![:]) { + let colon_token: Token![:] = input.parse()?; + let ty: Type = input.parse()?; +@@ -265,12 +245,19 @@ pub mod parsing { + }) + } + +- fn stmt_expr(input: ParseStream, allow_nosemi: bool) -> Result<Stmt> { +- let mut attrs = input.call(Attribute::parse_outer)?; ++ fn stmt_expr( ++ input: ParseStream, ++ allow_nosemi: bool, ++ mut attrs: Vec<Attribute>, ++ ) -> Result<Stmt> { + let mut e = expr::parsing::expr_early(input)?; + +- attrs.extend(e.replace_attrs(Vec::new())); +- e.replace_attrs(attrs); ++ let mut attr_target = &mut e; ++ while let Expr::Binary(e) = attr_target { ++ attr_target = &mut e.left; ++ } ++ attrs.extend(attr_target.replace_attrs(Vec::new())); ++ attr_target.replace_attrs(attrs); + + if input.peek(Token![;]) { + return Ok(Stmt::Semi(e, input.parse()?)); +diff --git a/third_party/rust/syn/src/token.rs b/third_party/rust/syn/src/token.rs +index 0b8c18192f..8539378c5e 100644 +--- a/third_party/rust/syn/src/token.rs ++++ b/third_party/rust/syn/src/token.rs +@@ -88,7 +88,6 @@ + //! [Printing]: https://docs.rs/quote/1.0/quote/trait.ToTokens.html + //! [`Span`]: https://docs.rs/proc-macro2/1.0/proc_macro2/struct.Span.html + +-use std; + #[cfg(feature = "extra-traits")] + use std::cmp; + #[cfg(feature = "extra-traits")] +@@ -97,13 +96,13 @@ use std::fmt::{self, Debug}; + use std::hash::{Hash, Hasher}; + use std::ops::{Deref, DerefMut}; + +-#[cfg(feature = "parsing")] +-use proc_macro2::Delimiter; + #[cfg(any(feature = "parsing", feature = "printing"))] + use proc_macro2::Ident; + use proc_macro2::Span; + #[cfg(feature = "printing")] + use proc_macro2::TokenStream; ++#[cfg(feature = "parsing")] ++use proc_macro2::{Delimiter, Literal, Punct, TokenTree}; + #[cfg(feature = "printing")] + use quote::{ToTokens, TokenStreamExt}; + +@@ -112,10 +111,8 @@ use self::private::WithSpan; + use crate::buffer::Cursor; + #[cfg(feature = "parsing")] + use crate::error::Result; +-#[cfg(any(feature = "full", feature = "derive"))] + #[cfg(feature = "parsing")] + use crate::lifetime::Lifetime; +-#[cfg(any(feature = "full", feature = "derive"))] + #[cfg(feature = "parsing")] + use crate::lit::{Lit, LitBool, LitByte, LitByteStr, LitChar, LitFloat, LitInt, LitStr}; + #[cfg(feature = "parsing")] +@@ -155,21 +152,20 @@ mod private { + #[cfg(feature = "parsing")] + impl private::Sealed for Ident {} + +-#[cfg(any(feature = "full", feature = "derive"))] + #[cfg(feature = "parsing")] + fn peek_impl(cursor: Cursor, peek: fn(ParseStream) -> bool) -> bool { ++ use crate::parse::Unexpected; + use std::cell::Cell; + use std::rc::Rc; + + let scope = Span::call_site(); +- let unexpected = Rc::new(Cell::new(None)); ++ let unexpected = Rc::new(Cell::new(Unexpected::None)); + let buffer = crate::parse::new_parse_buffer(scope, cursor, unexpected); + peek(&buffer) + } + +-#[cfg(any(feature = "full", feature = "derive"))] + macro_rules! impl_token { +- ($name:ident $display:expr) => { ++ ($display:tt $name:ty) => { + #[cfg(feature = "parsing")] + impl Token for $name { + fn peek(cursor: Cursor) -> bool { +@@ -189,24 +185,38 @@ macro_rules! impl_token { + }; + } + +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(Lifetime "lifetime"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(Lit "literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitStr "string literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitByteStr "byte string literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitByte "byte literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitChar "character literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitInt "integer literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitFloat "floating point literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitBool "boolean literal"); ++impl_token!("lifetime" Lifetime); ++impl_token!("literal" Lit); ++impl_token!("string literal" LitStr); ++impl_token!("byte string literal" LitByteStr); ++impl_token!("byte literal" LitByte); ++impl_token!("character literal" LitChar); ++impl_token!("integer literal" LitInt); ++impl_token!("floating point literal" LitFloat); ++impl_token!("boolean literal" LitBool); ++impl_token!("group token" proc_macro2::Group); ++ ++macro_rules! impl_low_level_token { ++ ($display:tt $ty:ident $get:ident) => { ++ #[cfg(feature = "parsing")] ++ impl Token for $ty { ++ fn peek(cursor: Cursor) -> bool { ++ cursor.$get().is_some() ++ } ++ ++ fn display() -> &'static str { ++ $display ++ } ++ } ++ ++ #[cfg(feature = "parsing")] ++ impl private::Sealed for $ty {} ++ }; ++} ++ ++impl_low_level_token!("punctuation token" Punct punct); ++impl_low_level_token!("literal" Literal literal); ++impl_low_level_token!("token" TokenTree token_tree); + + // Not public API. + #[doc(hidden)] +@@ -233,7 +243,6 @@ impl<T: CustomToken> Token for T { + macro_rules! define_keywords { + ($($token:tt pub struct $name:ident #[$doc:meta])*) => { + $( +- #[cfg_attr(feature = "clone-impls", derive(Copy, Clone))] + #[$doc] + /// + /// Don't try to remember the name of this type — use the +@@ -260,6 +269,16 @@ macro_rules! define_keywords { + } + } + ++ #[cfg(feature = "clone-impls")] ++ impl Copy for $name {} ++ ++ #[cfg(feature = "clone-impls")] ++ impl Clone for $name { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } ++ + #[cfg(feature = "extra-traits")] + impl Debug for $name { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +@@ -338,7 +357,6 @@ macro_rules! impl_deref_if_len_is_1 { + macro_rules! define_punctuation_structs { + ($($token:tt pub struct $name:ident/$len:tt #[$doc:meta])*) => { + $( +- #[cfg_attr(feature = "clone-impls", derive(Copy, Clone))] + #[repr(C)] + #[$doc] + /// +@@ -366,6 +384,16 @@ macro_rules! define_punctuation_structs { + } + } + ++ #[cfg(feature = "clone-impls")] ++ impl Copy for $name {} ++ ++ #[cfg(feature = "clone-impls")] ++ impl Clone for $name { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } ++ + #[cfg(feature = "extra-traits")] + impl Debug for $name { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +@@ -436,7 +464,6 @@ macro_rules! define_punctuation { + macro_rules! define_delimiters { + ($($token:tt pub struct $name:ident #[$doc:meta])*) => { + $( +- #[cfg_attr(feature = "clone-impls", derive(Copy, Clone))] + #[$doc] + pub struct $name { + pub span: Span, +@@ -458,6 +485,16 @@ macro_rules! define_delimiters { + } + } + ++ #[cfg(feature = "clone-impls")] ++ impl Copy for $name {} ++ ++ #[cfg(feature = "clone-impls")] ++ impl Clone for $name { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } ++ + #[cfg(feature = "extra-traits")] + impl Debug for $name { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +@@ -855,7 +892,7 @@ pub mod parsing { + } + + pub fn punct<S: FromSpans>(input: ParseStream, token: &str) -> Result<S> { +- let mut spans = [input.cursor().span(); 3]; ++ let mut spans = [input.span(); 3]; + punct_helper(input, token, &mut spans)?; + Ok(S::from_spans(&spans)) + } +diff --git a/third_party/rust/syn/src/tt.rs b/third_party/rust/syn/src/tt.rs +index f860eebb4f..8dba0627cd 100644 +--- a/third_party/rust/syn/src/tt.rs ++++ b/third_party/rust/syn/src/tt.rs +@@ -18,8 +18,8 @@ impl<'a> PartialEq for TokenTreeHelper<'a> { + _ => return false, + } + +- let s1 = g1.stream().clone().into_iter(); +- let mut s2 = g2.stream().clone().into_iter(); ++ let s1 = g1.stream().into_iter(); ++ let mut s2 = g2.stream().into_iter(); + + for item1 in s1 { + let item2 = match s2.next() { +@@ -60,7 +60,7 @@ impl<'a> Hash for TokenTreeHelper<'a> { + Delimiter::None => 3u8.hash(h), + } + +- for item in g.stream().clone() { ++ for item in g.stream() { + TokenTreeHelper(&item).hash(h); + } + 0xffu8.hash(h); // terminator w/ a variant we don't normally hash +diff --git a/third_party/rust/syn/src/ty.rs b/third_party/rust/syn/src/ty.rs +index 4ee59bda2a..fd7c97eab7 100644 +--- a/third_party/rust/syn/src/ty.rs ++++ b/third_party/rust/syn/src/ty.rs +@@ -1,15 +1,11 @@ + use super::*; + use crate::punctuated::Punctuated; +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; + use proc_macro2::TokenStream; +-#[cfg(feature = "extra-traits")] +-use std::hash::{Hash, Hasher}; + + ast_enum_of_structs! { + /// The possible types that a Rust value could have. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Syntax tree enum +@@ -20,7 +16,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum Type #manual_extra_traits { ++ pub enum Type { + /// A fixed size array type: `[T; n]`. + Array(TypeArray), + +@@ -77,7 +73,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A fixed size array type: `[T; n]`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeArray { + pub bracket_token: token::Bracket, +@@ -90,7 +86,7 @@ ast_struct! { + ast_struct! { + /// A bare function type: `fn(usize) -> bool`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeBareFn { + pub lifetimes: Option<BoundLifetimes>, +@@ -107,7 +103,7 @@ ast_struct! { + ast_struct! { + /// A type contained within invisible delimiters. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeGroup { + pub group_token: token::Group, +@@ -119,7 +115,7 @@ ast_struct! { + /// An `impl Bound1 + Bound2 + Bound3` type where `Bound` is a trait or + /// a lifetime. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeImplTrait { + pub impl_token: Token![impl], +@@ -130,7 +126,7 @@ ast_struct! { + ast_struct! { + /// Indication that a type should be inferred by the compiler: `_`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeInfer { + pub underscore_token: Token![_], +@@ -140,7 +136,7 @@ ast_struct! { + ast_struct! { + /// A macro in the type position. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeMacro { + pub mac: Macro, +@@ -150,7 +146,7 @@ ast_struct! { + ast_struct! { + /// The never type: `!`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeNever { + pub bang_token: Token![!], +@@ -160,7 +156,7 @@ ast_struct! { + ast_struct! { + /// A parenthesized type equivalent to the inner type. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeParen { + pub paren_token: token::Paren, +@@ -172,7 +168,7 @@ ast_struct! { + /// A path like `std::slice::Iter`, optionally qualified with a + /// self-type as in `<Vec<T> as SomeTrait>::Associated`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypePath { + pub qself: Option<QSelf>, +@@ -183,7 +179,7 @@ ast_struct! { + ast_struct! { + /// A raw pointer type: `*const T` or `*mut T`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypePtr { + pub star_token: Token![*], +@@ -196,7 +192,7 @@ ast_struct! { + ast_struct! { + /// A reference type: `&'a T` or `&'a mut T`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeReference { + pub and_token: Token![&], +@@ -209,7 +205,7 @@ ast_struct! { + ast_struct! { + /// A dynamically sized slice type: `[T]`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeSlice { + pub bracket_token: token::Bracket, +@@ -221,7 +217,7 @@ ast_struct! { + /// A trait object type `Bound1 + Bound2 + Bound3` where `Bound` is a + /// trait or a lifetime. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeTraitObject { + pub dyn_token: Option<Token![dyn]>, +@@ -232,7 +228,7 @@ ast_struct! { + ast_struct! { + /// A tuple type: `(A, B, C, String)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeTuple { + pub paren_token: token::Paren, +@@ -240,111 +236,10 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Type {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Type { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (Type::Array(this), Type::Array(other)) => this == other, +- (Type::BareFn(this), Type::BareFn(other)) => this == other, +- (Type::Group(this), Type::Group(other)) => this == other, +- (Type::ImplTrait(this), Type::ImplTrait(other)) => this == other, +- (Type::Infer(this), Type::Infer(other)) => this == other, +- (Type::Macro(this), Type::Macro(other)) => this == other, +- (Type::Never(this), Type::Never(other)) => this == other, +- (Type::Paren(this), Type::Paren(other)) => this == other, +- (Type::Path(this), Type::Path(other)) => this == other, +- (Type::Ptr(this), Type::Ptr(other)) => this == other, +- (Type::Reference(this), Type::Reference(other)) => this == other, +- (Type::Slice(this), Type::Slice(other)) => this == other, +- (Type::TraitObject(this), Type::TraitObject(other)) => this == other, +- (Type::Tuple(this), Type::Tuple(other)) => this == other, +- (Type::Verbatim(this), Type::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Type { +- fn hash<H>(&self, hash: &mut H) +- where +- H: Hasher, +- { +- match self { +- Type::Array(ty) => { +- hash.write_u8(0); +- ty.hash(hash); +- } +- Type::BareFn(ty) => { +- hash.write_u8(1); +- ty.hash(hash); +- } +- Type::Group(ty) => { +- hash.write_u8(2); +- ty.hash(hash); +- } +- Type::ImplTrait(ty) => { +- hash.write_u8(3); +- ty.hash(hash); +- } +- Type::Infer(ty) => { +- hash.write_u8(4); +- ty.hash(hash); +- } +- Type::Macro(ty) => { +- hash.write_u8(5); +- ty.hash(hash); +- } +- Type::Never(ty) => { +- hash.write_u8(6); +- ty.hash(hash); +- } +- Type::Paren(ty) => { +- hash.write_u8(7); +- ty.hash(hash); +- } +- Type::Path(ty) => { +- hash.write_u8(8); +- ty.hash(hash); +- } +- Type::Ptr(ty) => { +- hash.write_u8(9); +- ty.hash(hash); +- } +- Type::Reference(ty) => { +- hash.write_u8(10); +- ty.hash(hash); +- } +- Type::Slice(ty) => { +- hash.write_u8(11); +- ty.hash(hash); +- } +- Type::TraitObject(ty) => { +- hash.write_u8(12); +- ty.hash(hash); +- } +- Type::Tuple(ty) => { +- hash.write_u8(13); +- ty.hash(hash); +- } +- Type::Verbatim(ty) => { +- hash.write_u8(14); +- TokenStreamHelper(ty).hash(hash); +- } +- Type::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + ast_struct! { + /// The binary interface of a function: `extern "C"`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Abi { + pub extern_token: Token![extern], +@@ -355,7 +250,7 @@ ast_struct! { + ast_struct! { + /// An argument in a function type: the `usize` in `fn(usize) -> bool`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct BareFnArg { + pub attrs: Vec<Attribute>, +@@ -377,7 +272,7 @@ ast_struct! { + /// } + /// ``` + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Variadic { + pub attrs: Vec<Attribute>, +@@ -388,7 +283,7 @@ ast_struct! { + ast_enum! { + /// Return type of a function signature. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum ReturnType { + /// Return type is not specified. +@@ -407,10 +302,13 @@ pub mod parsing { + use crate::ext::IdentExt; + use crate::parse::{Parse, ParseStream, Result}; + use crate::path; ++ use proc_macro2::{Punct, Spacing, TokenTree}; ++ use std::iter::FromIterator; + + impl Parse for Type { + fn parse(input: ParseStream) -> Result<Self> { +- ambig_ty(input, true) ++ let allow_plus = true; ++ ambig_ty(input, allow_plus) + } + } + +@@ -421,15 +319,17 @@ pub mod parsing { + /// + /// This parser does not allow a `+`, while the default parser does. + pub fn without_plus(input: ParseStream) -> Result<Self> { +- ambig_ty(input, false) ++ let allow_plus = false; ++ ambig_ty(input, allow_plus) + } + } + + fn ambig_ty(input: ParseStream, allow_plus: bool) -> Result<Type> { +- if input.peek(token::Group) { ++ if input.peek(token::Group) && !input.peek2(Token![::]) && !input.peek2(Token![<]) { + return input.parse().map(Type::Group); + } + ++ let begin = input.fork(); + let mut lifetimes = None::<BoundLifetimes>; + let mut lookahead = input.lookahead1(); + if lookahead.peek(Token![for]) { +@@ -524,7 +424,7 @@ pub mod parsing { + ..trait_bound + }) + } +- other => other, ++ other @ TypeParamBound::Lifetime(_) => other, + } + } + _ => break, +@@ -549,17 +449,20 @@ pub mod parsing { + })) + } else if lookahead.peek(Token![fn]) + || lookahead.peek(Token![unsafe]) +- || lookahead.peek(Token![extern]) && !input.peek2(Token![::]) ++ || lookahead.peek(Token![extern]) + { +- let mut bare_fn: TypeBareFn = input.parse()?; +- bare_fn.lifetimes = lifetimes; +- Ok(Type::BareFn(bare_fn)) ++ let allow_mut_self = true; ++ if let Some(mut bare_fn) = parse_bare_fn(input, allow_mut_self)? { ++ bare_fn.lifetimes = lifetimes; ++ Ok(Type::BareFn(bare_fn)) ++ } else { ++ Ok(Type::Verbatim(verbatim::between(begin, input))) ++ } + } else if lookahead.peek(Ident) + || input.peek(Token![super]) + || input.peek(Token![self]) + || input.peek(Token![Self]) + || input.peek(Token![crate]) +- || input.peek(Token![extern]) + || lookahead.peek(Token![::]) + || lookahead.peek(Token![<]) + { +@@ -722,38 +625,58 @@ pub mod parsing { + + impl Parse for TypeBareFn { + fn parse(input: ParseStream) -> Result<Self> { +- let args; +- let allow_variadic; +- Ok(TypeBareFn { +- lifetimes: input.parse()?, +- unsafety: input.parse()?, +- abi: input.parse()?, +- fn_token: input.parse()?, +- paren_token: parenthesized!(args in input), +- inputs: { +- let mut inputs = Punctuated::new(); +- while !args.is_empty() && !args.peek(Token![...]) { +- inputs.push_value(args.parse()?); +- if args.is_empty() { +- break; +- } +- inputs.push_punct(args.parse()?); +- } +- allow_variadic = inputs.empty_or_trailing(); +- inputs +- }, +- variadic: { +- if allow_variadic && args.peek(Token![...]) { +- Some(Variadic { +- attrs: Vec::new(), ++ let allow_mut_self = false; ++ parse_bare_fn(input, allow_mut_self).map(Option::unwrap) ++ } ++ } ++ ++ fn parse_bare_fn(input: ParseStream, allow_mut_self: bool) -> Result<Option<TypeBareFn>> { ++ let args; ++ let mut variadic = None; ++ let mut has_mut_self = false; ++ ++ let bare_fn = TypeBareFn { ++ lifetimes: input.parse()?, ++ unsafety: input.parse()?, ++ abi: input.parse()?, ++ fn_token: input.parse()?, ++ paren_token: parenthesized!(args in input), ++ inputs: { ++ let mut inputs = Punctuated::new(); ++ ++ while !args.is_empty() { ++ let attrs = args.call(Attribute::parse_outer)?; ++ ++ if inputs.empty_or_trailing() && args.peek(Token![...]) { ++ variadic = Some(Variadic { ++ attrs, + dots: args.parse()?, +- }) ++ }); ++ break; ++ } ++ ++ if let Some(arg) = parse_bare_fn_arg(&args, allow_mut_self)? { ++ inputs.push_value(BareFnArg { attrs, ..arg }); + } else { +- None ++ has_mut_self = true; + } +- }, +- output: input.call(ReturnType::without_plus)?, +- }) ++ if args.is_empty() { ++ break; ++ } ++ ++ inputs.push_punct(args.parse()?); ++ } ++ ++ inputs ++ }, ++ variadic, ++ output: input.call(ReturnType::without_plus)?, ++ }; ++ ++ if has_mut_self { ++ Ok(None) ++ } else { ++ Ok(Some(bare_fn)) + } + } + +@@ -776,9 +699,27 @@ pub mod parsing { + impl Parse for TypeTuple { + fn parse(input: ParseStream) -> Result<Self> { + let content; ++ let paren_token = parenthesized!(content in input); ++ ++ if content.is_empty() { ++ return Ok(TypeTuple { ++ paren_token, ++ elems: Punctuated::new(), ++ }); ++ } ++ ++ let first: Type = content.parse()?; + Ok(TypeTuple { +- paren_token: parenthesized!(content in input), +- elems: content.parse_terminated(Type::parse)?, ++ paren_token, ++ elems: { ++ let mut elems = Punctuated::new(); ++ elems.push_value(first); ++ elems.push_punct(content.parse()?); ++ let rest: Punctuated<Type, Token![,]> = ++ content.parse_terminated(Parse::parse)?; ++ elems.extend(rest); ++ elems ++ }, + }) + } + } +@@ -807,9 +748,11 @@ pub mod parsing { + + impl ReturnType { + pub fn without_plus(input: ParseStream) -> Result<Self> { +- Self::parse(input, false) ++ let allow_plus = false; ++ Self::parse(input, allow_plus) + } + ++ #[doc(hidden)] + pub fn parse(input: ParseStream, allow_plus: bool) -> Result<Self> { + if input.peek(Token![->]) { + let arrow = input.parse()?; +@@ -844,10 +787,12 @@ pub mod parsing { + + impl TypeTraitObject { + pub fn without_plus(input: ParseStream) -> Result<Self> { +- Self::parse(input, false) ++ let allow_plus = false; ++ Self::parse(input, allow_plus) + } + + // Only allow multiple trait references if allow_plus is true. ++ #[doc(hidden)] + pub fn parse(input: ParseStream, allow_plus: bool) -> Result<Self> { + Ok(TypeTraitObject { + dyn_token: input.parse()?, +@@ -910,7 +855,8 @@ pub mod parsing { + + impl Parse for TypeParen { + fn parse(input: ParseStream) -> Result<Self> { +- Self::parse(input, false) ++ let allow_plus = false; ++ Self::parse(input, allow_plus) + } + } + +@@ -926,22 +872,72 @@ pub mod parsing { + + impl Parse for BareFnArg { + fn parse(input: ParseStream) -> Result<Self> { +- Ok(BareFnArg { +- attrs: input.call(Attribute::parse_outer)?, +- name: { +- if (input.peek(Ident) || input.peek(Token![_])) +- && input.peek2(Token![:]) +- && !input.peek2(Token![::]) +- { +- let name = input.call(Ident::parse_any)?; +- let colon: Token![:] = input.parse()?; +- Some((name, colon)) +- } else { +- None +- } +- }, +- ty: input.parse()?, +- }) ++ let allow_mut_self = false; ++ parse_bare_fn_arg(input, allow_mut_self).map(Option::unwrap) ++ } ++ } ++ ++ fn parse_bare_fn_arg( ++ input: ParseStream, ++ mut allow_mut_self: bool, ++ ) -> Result<Option<BareFnArg>> { ++ let mut has_mut_self = false; ++ let arg = BareFnArg { ++ attrs: input.call(Attribute::parse_outer)?, ++ name: { ++ if (input.peek(Ident) || input.peek(Token![_]) || input.peek(Token![self])) ++ && input.peek2(Token![:]) ++ && !input.peek2(Token![::]) ++ { ++ let name = input.call(Ident::parse_any)?; ++ let colon: Token![:] = input.parse()?; ++ Some((name, colon)) ++ } else if allow_mut_self ++ && input.peek(Token![mut]) ++ && input.peek2(Token![self]) ++ && input.peek3(Token![:]) ++ && !input.peek3(Token![::]) ++ { ++ has_mut_self = true; ++ allow_mut_self = false; ++ input.parse::<Token![mut]>()?; ++ input.parse::<Token![self]>()?; ++ input.parse::<Token![:]>()?; ++ None ++ } else { ++ None ++ } ++ }, ++ ty: if !has_mut_self && input.peek(Token![...]) { ++ let dot3 = input.parse::<Token![...]>()?; ++ let args = vec![ ++ TokenTree::Punct(Punct::new('.', Spacing::Joint)), ++ TokenTree::Punct(Punct::new('.', Spacing::Joint)), ++ TokenTree::Punct(Punct::new('.', Spacing::Alone)), ++ ]; ++ let tokens = TokenStream::from_iter(args.into_iter().zip(&dot3.spans).map( ++ |(mut arg, span)| { ++ arg.set_span(*span); ++ arg ++ }, ++ )); ++ Type::Verbatim(tokens) ++ } else if allow_mut_self && input.peek(Token![mut]) && input.peek2(Token![self]) { ++ has_mut_self = true; ++ input.parse::<Token![mut]>()?; ++ Type::Path(TypePath { ++ qself: None, ++ path: input.parse::<Token![self]>()?.into(), ++ }) ++ } else { ++ input.parse()? ++ }, ++ }; ++ ++ if has_mut_self { ++ Ok(None) ++ } else { ++ Ok(Some(arg)) + } + } + +diff --git a/third_party/rust/syn/src/verbatim.rs b/third_party/rust/syn/src/verbatim.rs +new file mode 100644 +index 0000000000..0686352f7a +--- /dev/null ++++ b/third_party/rust/syn/src/verbatim.rs +@@ -0,0 +1,15 @@ ++use crate::parse::{ParseBuffer, ParseStream}; ++use proc_macro2::TokenStream; ++use std::iter; ++ ++pub fn between<'a>(begin: ParseBuffer<'a>, end: ParseStream<'a>) -> TokenStream { ++ let end = end.cursor(); ++ let mut cursor = begin.cursor(); ++ let mut tokens = TokenStream::new(); ++ while cursor != end { ++ let (tt, next) = cursor.token_tree().unwrap(); ++ tokens.extend(iter::once(tt)); ++ cursor = next; ++ } ++ tokens ++} +diff --git a/third_party/rust/syn/src/whitespace.rs b/third_party/rust/syn/src/whitespace.rs +new file mode 100644 +index 0000000000..7be082e1a2 +--- /dev/null ++++ b/third_party/rust/syn/src/whitespace.rs +@@ -0,0 +1,65 @@ ++pub fn skip(mut s: &str) -> &str { ++ 'skip: while !s.is_empty() { ++ let byte = s.as_bytes()[0]; ++ if byte == b'/' { ++ if s.starts_with("//") ++ && (!s.starts_with("///") || s.starts_with("////")) ++ && !s.starts_with("//!") ++ { ++ if let Some(i) = s.find('\n') { ++ s = &s[i + 1..]; ++ continue; ++ } else { ++ return ""; ++ } ++ } else if s.starts_with("/**/") { ++ s = &s[4..]; ++ continue; ++ } else if s.starts_with("/*") ++ && (!s.starts_with("/**") || s.starts_with("/***")) ++ && !s.starts_with("/*!") ++ { ++ let mut depth = 0; ++ let bytes = s.as_bytes(); ++ let mut i = 0; ++ let upper = bytes.len() - 1; ++ while i < upper { ++ if bytes[i] == b'/' && bytes[i + 1] == b'*' { ++ depth += 1; ++ i += 1; // eat '*' ++ } else if bytes[i] == b'*' && bytes[i + 1] == b'/' { ++ depth -= 1; ++ if depth == 0 { ++ s = &s[i + 2..]; ++ continue 'skip; ++ } ++ i += 1; // eat '/' ++ } ++ i += 1; ++ } ++ return s; ++ } ++ } ++ match byte { ++ b' ' | 0x09..=0x0d => { ++ s = &s[1..]; ++ continue; ++ } ++ b if b <= 0x7f => {} ++ _ => { ++ let ch = s.chars().next().unwrap(); ++ if is_whitespace(ch) { ++ s = &s[ch.len_utf8()..]; ++ continue; ++ } ++ } ++ } ++ return s; ++ } ++ s ++} ++ ++fn is_whitespace(ch: char) -> bool { ++ // Rust treats left-to-right mark and right-to-left mark as whitespace ++ ch.is_whitespace() || ch == '\u{200e}' || ch == '\u{200f}' ++} +diff --git a/third_party/rust/syn/tests/clone.sh b/third_party/rust/syn/tests/clone.sh +deleted file mode 100755 +index 8e0863cba6..0000000000 +--- a/third_party/rust/syn/tests/clone.sh ++++ /dev/null +@@ -1,16 +0,0 @@ +-#!/bin/bash +- +-REV=4560cb830fce63fcffdc4558f4281aaac6a3a1ba +- +-set -euo pipefail +-cd "$(dirname "${BASH_SOURCE[0]}")" +-mkdir -p rust +-touch rust/COMMIT +- +-if [ "$(cat rust/COMMIT)" != "$REV" ]; then +- rm -rf rust +- mkdir rust +- curl -L "https://github.com/rust-lang/rust/archive/${REV}.tar.gz" \ +- | tar xz --directory rust --strip-components 1 +- echo "$REV" > rust/COMMIT +-fi +diff --git a/third_party/rust/syn/tests/common/eq.rs b/third_party/rust/syn/tests/common/eq.rs +index 13a6c36ae5..7589a07573 100644 +--- a/third_party/rust/syn/tests/common/eq.rs ++++ b/third_party/rust/syn/tests/common/eq.rs +@@ -1,36 +1,35 @@ ++extern crate rustc_ast; + extern crate rustc_data_structures; ++extern crate rustc_span; + extern crate rustc_target; +-extern crate syntax; +-extern crate syntax_pos; + + use std::mem; + +-use self::rustc_data_structures::sync::Lrc; +-use self::rustc_data_structures::thin_vec::ThinVec; +-use self::rustc_target::abi::FloatTy; +-use self::rustc_target::spec::abi::Abi; +-use self::syntax::ast::{ +- AngleBracketedArgs, AnonConst, Arg, Arm, AsmDialect, AssocTyConstraint, AssocTyConstraintKind, +- AttrId, AttrStyle, Attribute, BareFnTy, BinOpKind, BindingMode, Block, BlockCheckMode, +- CaptureBy, Constness, Crate, CrateSugar, Defaultness, EnumDef, Expr, ExprKind, Field, FieldPat, +- FnDecl, FnHeader, ForeignItem, ForeignItemKind, ForeignMod, FunctionRetTy, GenericArg, +- GenericArgs, GenericBound, GenericParam, GenericParamKind, Generics, GlobalAsm, Ident, +- ImplItem, ImplItemKind, ImplPolarity, InlineAsm, InlineAsmOutput, IntTy, IsAsync, IsAuto, Item, +- ItemKind, Label, Lifetime, Lit, LitIntType, LitKind, Local, Mac, MacDelimiter, MacStmtStyle, +- MacroDef, MethodSig, Mod, Movability, MutTy, Mutability, NodeId, ParenthesizedArgs, Pat, +- PatKind, Path, PathSegment, PolyTraitRef, QSelf, RangeEnd, RangeLimits, RangeSyntax, Stmt, +- StmtKind, StrStyle, StructField, TraitBoundModifier, TraitItem, TraitItemKind, +- TraitObjectSyntax, TraitRef, Ty, TyKind, UintTy, UnOp, UnsafeSource, Unsafety, UseTree, +- UseTreeKind, Variant, VariantData, VisibilityKind, WhereBoundPredicate, WhereClause, +- WhereEqPredicate, WherePredicate, WhereRegionPredicate, ++use rustc_ast::ast::{ ++ AngleBracketedArg, AngleBracketedArgs, AnonConst, Arm, AssocItemKind, AssocTyConstraint, ++ AssocTyConstraintKind, Async, AttrId, AttrItem, AttrKind, AttrStyle, Attribute, BareFnTy, ++ BinOpKind, BindingMode, Block, BlockCheckMode, BorrowKind, CaptureBy, Const, Crate, CrateSugar, ++ Defaultness, EnumDef, Expr, ExprKind, Extern, Field, FieldPat, FloatTy, FnDecl, FnHeader, ++ FnRetTy, FnSig, ForeignItemKind, ForeignMod, GenericArg, GenericArgs, GenericBound, ++ GenericParam, GenericParamKind, Generics, GlobalAsm, ImplPolarity, InlineAsm, InlineAsmOperand, ++ InlineAsmOptions, InlineAsmRegOrRegClass, InlineAsmTemplatePiece, IntTy, IsAuto, Item, ++ ItemKind, Label, Lifetime, Lit, LitFloatType, LitIntType, LitKind, LlvmAsmDialect, ++ LlvmInlineAsm, LlvmInlineAsmOutput, Local, MacArgs, MacCall, MacCallStmt, MacDelimiter, ++ MacStmtStyle, MacroDef, Mod, Movability, MutTy, Mutability, NodeId, Param, ParenthesizedArgs, ++ Pat, PatKind, Path, PathSegment, PolyTraitRef, QSelf, RangeEnd, RangeLimits, RangeSyntax, Stmt, ++ StmtKind, StrLit, StrStyle, StructField, TraitBoundModifier, TraitObjectSyntax, TraitRef, Ty, ++ TyKind, UintTy, UnOp, Unsafe, UnsafeSource, UseTree, UseTreeKind, Variant, VariantData, ++ VisibilityKind, WhereBoundPredicate, WhereClause, WhereEqPredicate, WherePredicate, ++ WhereRegionPredicate, + }; +-use self::syntax::parse::lexer::comments; +-use self::syntax::parse::token::{self, DelimToken, Token, TokenKind}; +-use self::syntax::ptr::P; +-use self::syntax::source_map::Spanned; +-use self::syntax::symbol::{sym, Symbol}; +-use self::syntax::tokenstream::{DelimSpan, TokenStream, TokenTree}; +-use self::syntax_pos::{Span, SyntaxContext, DUMMY_SP}; ++use rustc_ast::ptr::P; ++use rustc_ast::token::{self, CommentKind, DelimToken, Token, TokenKind}; ++use rustc_ast::tokenstream::{DelimSpan, TokenStream, TokenTree}; ++use rustc_data_structures::sync::Lrc; ++use rustc_data_structures::thin_vec::ThinVec; ++use rustc_span::source_map::Spanned; ++use rustc_span::symbol::Ident; ++use rustc_span::{Span, Symbol, SyntaxContext}; + + pub trait SpanlessEq { + fn eq(&self, other: &Self) -> bool; +@@ -86,14 +85,6 @@ impl<A: SpanlessEq, B: SpanlessEq> SpanlessEq for (A, B) { + } + } + +-impl<A: SpanlessEq, B: SpanlessEq, C: SpanlessEq> SpanlessEq for (A, B, C) { +- fn eq(&self, other: &Self) -> bool { +- SpanlessEq::eq(&self.0, &other.0) +- && SpanlessEq::eq(&self.1, &other.1) +- && SpanlessEq::eq(&self.2, &other.2) +- } +-} +- + macro_rules! spanless_eq_true { + ($name:ident) => { + impl SpanlessEq for $name { +@@ -126,17 +117,19 @@ spanless_eq_partial_eq!(u16); + spanless_eq_partial_eq!(u128); + spanless_eq_partial_eq!(usize); + spanless_eq_partial_eq!(char); ++spanless_eq_partial_eq!(String); + spanless_eq_partial_eq!(Symbol); +-spanless_eq_partial_eq!(Abi); ++spanless_eq_partial_eq!(CommentKind); + spanless_eq_partial_eq!(DelimToken); ++spanless_eq_partial_eq!(InlineAsmOptions); + + macro_rules! spanless_eq_struct { + { +- $name:ident; ++ $name:ident $(<$param:ident>)?; + $([$field:ident $other:ident])* + $(![$ignore:ident])* + } => { +- impl SpanlessEq for $name { ++ impl $(<$param: SpanlessEq>)* SpanlessEq for $name $(<$param>)* { + fn eq(&self, other: &Self) -> bool { + let $name { $($field,)* $($ignore: _,)* } = self; + let $name { $($field: $other,)* $($ignore: _,)* } = other; +@@ -146,14 +139,14 @@ macro_rules! spanless_eq_struct { + }; + + { +- $name:ident; ++ $name:ident $(<$param:ident>)?; + $([$field:ident $other:ident])* + $next:ident + $($rest:ident)* + $(!$ignore:ident)* + } => { + spanless_eq_struct! { +- $name; ++ $name $(<$param>)*; + $([$field $other])* + [$next other] + $($rest)* +@@ -162,14 +155,14 @@ macro_rules! spanless_eq_struct { + }; + + { +- $name:ident; ++ $name:ident $(<$param:ident>)?; + $([$field:ident $other:ident])* + $(![$ignore:ident])* + !$next:ident + $(!$rest:ident)* + } => { + spanless_eq_struct! { +- $name; ++ $name $(<$param>)*; + $([$field $other])* + $(![$ignore])* + ![$next] +@@ -263,119 +256,131 @@ macro_rules! spanless_eq_enum { + }; + } + +-spanless_eq_struct!(AngleBracketedArgs; span args constraints); ++spanless_eq_struct!(AngleBracketedArgs; span args); + spanless_eq_struct!(AnonConst; id value); +-spanless_eq_struct!(Arg; attrs ty pat id span); +-spanless_eq_struct!(Arm; attrs pats guard body span id); ++spanless_eq_struct!(Arm; attrs pat guard body span id is_placeholder); + spanless_eq_struct!(AssocTyConstraint; id ident kind span); +-spanless_eq_struct!(Attribute; id style path tokens span !is_sugared_doc); +-spanless_eq_struct!(BareFnTy; unsafety abi generic_params decl); ++spanless_eq_struct!(AttrItem; path args); ++spanless_eq_struct!(Attribute; kind id style span); ++spanless_eq_struct!(BareFnTy; unsafety ext generic_params decl); + spanless_eq_struct!(Block; stmts id rules span); +-spanless_eq_struct!(Crate; module attrs span); ++spanless_eq_struct!(Crate; module attrs span proc_macros); + spanless_eq_struct!(EnumDef; variants); +-spanless_eq_struct!(Expr; id node span attrs); +-spanless_eq_struct!(Field; ident expr span is_shorthand attrs id); +-spanless_eq_struct!(FieldPat; ident pat is_shorthand attrs id span); +-spanless_eq_struct!(FnDecl; inputs output c_variadic); +-spanless_eq_struct!(FnHeader; constness asyncness unsafety abi); +-spanless_eq_struct!(ForeignItem; ident attrs node id span vis); ++spanless_eq_struct!(Expr; id kind span attrs !tokens); ++spanless_eq_struct!(Field; attrs id span ident expr is_shorthand is_placeholder); ++spanless_eq_struct!(FieldPat; ident pat is_shorthand attrs id span is_placeholder); ++spanless_eq_struct!(FnDecl; inputs output); ++spanless_eq_struct!(FnHeader; constness asyncness unsafety ext); ++spanless_eq_struct!(FnSig; header decl span); + spanless_eq_struct!(ForeignMod; abi items); +-spanless_eq_struct!(GenericParam; id ident attrs bounds kind); ++spanless_eq_struct!(GenericParam; id ident attrs bounds is_placeholder kind); + spanless_eq_struct!(Generics; params where_clause span); + spanless_eq_struct!(GlobalAsm; asm); +-spanless_eq_struct!(ImplItem; id ident vis defaultness attrs generics node span !tokens); +-spanless_eq_struct!(InlineAsm; asm asm_str_style outputs inputs clobbers volatile alignstack dialect); +-spanless_eq_struct!(InlineAsmOutput; constraint expr is_rw is_indirect); +-spanless_eq_struct!(Item; ident attrs id node vis span !tokens); ++spanless_eq_struct!(InlineAsm; template operands options line_spans); ++spanless_eq_struct!(Item<K>; attrs id span vis ident kind !tokens); + spanless_eq_struct!(Label; ident); + spanless_eq_struct!(Lifetime; id ident); +-spanless_eq_struct!(Lit; token node span); ++spanless_eq_struct!(Lit; token kind span); ++spanless_eq_struct!(LlvmInlineAsm; asm asm_str_style outputs inputs clobbers volatile alignstack dialect); ++spanless_eq_struct!(LlvmInlineAsmOutput; constraint expr is_rw is_indirect); + spanless_eq_struct!(Local; pat ty init id span attrs); +-spanless_eq_struct!(Mac; path delim tts span prior_type_ascription); +-spanless_eq_struct!(MacroDef; tokens legacy); +-spanless_eq_struct!(MethodSig; header decl); ++spanless_eq_struct!(MacCall; path args prior_type_ascription); ++spanless_eq_struct!(MacCallStmt; mac style attrs); ++spanless_eq_struct!(MacroDef; body macro_rules); + spanless_eq_struct!(Mod; inner items inline); + spanless_eq_struct!(MutTy; ty mutbl); ++spanless_eq_struct!(Param; attrs ty pat id span is_placeholder); + spanless_eq_struct!(ParenthesizedArgs; span inputs output); +-spanless_eq_struct!(Pat; id node span); ++spanless_eq_struct!(Pat; id kind span tokens); + spanless_eq_struct!(Path; span segments); + spanless_eq_struct!(PathSegment; ident id args); + spanless_eq_struct!(PolyTraitRef; bound_generic_params trait_ref span); + spanless_eq_struct!(QSelf; ty path_span position); +-spanless_eq_struct!(Stmt; id node span); +-spanless_eq_struct!(StructField; span ident vis id ty attrs); ++spanless_eq_struct!(Stmt; id kind span); ++spanless_eq_struct!(StrLit; style symbol suffix span symbol_unescaped); ++spanless_eq_struct!(StructField; attrs id span vis ident ty is_placeholder); + spanless_eq_struct!(Token; kind span); +-spanless_eq_struct!(TraitItem; id ident attrs generics node span !tokens); + spanless_eq_struct!(TraitRef; path ref_id); +-spanless_eq_struct!(Ty; id node span); ++spanless_eq_struct!(Ty; id kind span); + spanless_eq_struct!(UseTree; prefix kind span); +-spanless_eq_struct!(Variant; ident attrs id data disr_expr span); ++spanless_eq_struct!(Variant; attrs id span vis ident data disr_expr is_placeholder); + spanless_eq_struct!(WhereBoundPredicate; span bound_generic_params bounded_ty bounds); +-spanless_eq_struct!(WhereClause; predicates span); ++spanless_eq_struct!(WhereClause; has_where_token predicates span); + spanless_eq_struct!(WhereEqPredicate; id span lhs_ty rhs_ty); + spanless_eq_struct!(WhereRegionPredicate; span lifetime bounds); +-spanless_eq_enum!(AsmDialect; Att Intel); ++spanless_eq_enum!(AngleBracketedArg; Arg(0) Constraint(0)); ++spanless_eq_enum!(AssocItemKind; Const(0 1 2) Fn(0 1 2 3) TyAlias(0 1 2 3) MacCall(0)); + spanless_eq_enum!(AssocTyConstraintKind; Equality(ty) Bound(bounds)); ++spanless_eq_enum!(Async; Yes(span closure_id return_impl_trait_id) No); ++spanless_eq_enum!(AttrKind; Normal(0) DocComment(0 1)); + spanless_eq_enum!(AttrStyle; Outer Inner); + spanless_eq_enum!(BinOpKind; Add Sub Mul Div Rem And Or BitXor BitAnd BitOr Shl Shr Eq Lt Le Ne Ge Gt); + spanless_eq_enum!(BindingMode; ByRef(0) ByValue(0)); + spanless_eq_enum!(BlockCheckMode; Default Unsafe(0)); ++spanless_eq_enum!(BorrowKind; Ref Raw); + spanless_eq_enum!(CaptureBy; Value Ref); +-spanless_eq_enum!(Constness; Const NotConst); ++spanless_eq_enum!(Const; Yes(0) No); + spanless_eq_enum!(CrateSugar; PubCrate JustCrate); +-spanless_eq_enum!(Defaultness; Default Final); ++spanless_eq_enum!(Defaultness; Default(0) Final); ++spanless_eq_enum!(Extern; None Implicit Explicit(0)); + spanless_eq_enum!(FloatTy; F32 F64); +-spanless_eq_enum!(ForeignItemKind; Fn(0 1) Static(0 1) Ty Macro(0)); +-spanless_eq_enum!(FunctionRetTy; Default(0) Ty(0)); ++spanless_eq_enum!(FnRetTy; Default(0) Ty(0)); ++spanless_eq_enum!(ForeignItemKind; Static(0 1 2) Fn(0 1 2 3) TyAlias(0 1 2 3) MacCall(0)); + spanless_eq_enum!(GenericArg; Lifetime(0) Type(0) Const(0)); + spanless_eq_enum!(GenericArgs; AngleBracketed(0) Parenthesized(0)); + spanless_eq_enum!(GenericBound; Trait(0 1) Outlives(0)); +-spanless_eq_enum!(GenericParamKind; Lifetime Type(default) Const(ty)); +-spanless_eq_enum!(ImplItemKind; Const(0 1) Method(0 1) TyAlias(0) OpaqueTy(0) Macro(0)); +-spanless_eq_enum!(ImplPolarity; Positive Negative); ++spanless_eq_enum!(GenericParamKind; Lifetime Type(default) Const(ty kw_span)); ++spanless_eq_enum!(ImplPolarity; Positive Negative(0)); ++spanless_eq_enum!(InlineAsmRegOrRegClass; Reg(0) RegClass(0)); ++spanless_eq_enum!(InlineAsmTemplatePiece; String(0) Placeholder(operand_idx modifier span)); + spanless_eq_enum!(IntTy; Isize I8 I16 I32 I64 I128); +-spanless_eq_enum!(IsAsync; Async(closure_id return_impl_trait_id) NotAsync); + spanless_eq_enum!(IsAuto; Yes No); ++spanless_eq_enum!(LitFloatType; Suffixed(0) Unsuffixed); + spanless_eq_enum!(LitIntType; Signed(0) Unsigned(0) Unsuffixed); ++spanless_eq_enum!(LlvmAsmDialect; Att Intel); ++spanless_eq_enum!(MacArgs; Empty Delimited(0 1 2) Eq(0 1)); + spanless_eq_enum!(MacDelimiter; Parenthesis Bracket Brace); + spanless_eq_enum!(MacStmtStyle; Semicolon Braces NoBraces); + spanless_eq_enum!(Movability; Static Movable); +-spanless_eq_enum!(Mutability; Mutable Immutable); ++spanless_eq_enum!(Mutability; Mut Not); + spanless_eq_enum!(RangeEnd; Included(0) Excluded); + spanless_eq_enum!(RangeLimits; HalfOpen Closed); +-spanless_eq_enum!(StmtKind; Local(0) Item(0) Expr(0) Semi(0) Mac(0)); ++spanless_eq_enum!(StmtKind; Local(0) Item(0) Expr(0) Semi(0) Empty MacCall(0)); + spanless_eq_enum!(StrStyle; Cooked Raw(0)); + spanless_eq_enum!(TokenTree; Token(0) Delimited(0 1 2)); +-spanless_eq_enum!(TraitBoundModifier; None Maybe); +-spanless_eq_enum!(TraitItemKind; Const(0 1) Method(0 1) Type(0 1) Macro(0)); ++spanless_eq_enum!(TraitBoundModifier; None Maybe MaybeConst MaybeConstMaybe); + spanless_eq_enum!(TraitObjectSyntax; Dyn None); + spanless_eq_enum!(UintTy; Usize U8 U16 U32 U64 U128); + spanless_eq_enum!(UnOp; Deref Not Neg); ++spanless_eq_enum!(Unsafe; Yes(0) No); + spanless_eq_enum!(UnsafeSource; CompilerGenerated UserProvided); +-spanless_eq_enum!(Unsafety; Unsafe Normal); + spanless_eq_enum!(UseTreeKind; Simple(0 1 2) Nested(0) Glob); + spanless_eq_enum!(VariantData; Struct(0 1) Tuple(0 1) Unit(0)); + spanless_eq_enum!(VisibilityKind; Public Crate(0) Restricted(path id) Inherited); + spanless_eq_enum!(WherePredicate; BoundPredicate(0) RegionPredicate(0) EqPredicate(0)); +-spanless_eq_enum!(ExprKind; Box(0) Array(0) Call(0 1) MethodCall(0 1) Tup(0) ++spanless_eq_enum!(ExprKind; Box(0) Array(0) Call(0 1) MethodCall(0 1 2) Tup(0) + Binary(0 1 2) Unary(0 1) Lit(0) Cast(0 1) Type(0 1) Let(0 1) If(0 1 2) + While(0 1 2) ForLoop(0 1 2 3) Loop(0 1) Match(0 1) Closure(0 1 2 3 4 5) +- Block(0 1) Async(0 1 2) Await(0) TryBlock(0) Assign(0 1) AssignOp(0 1 2) +- Field(0 1) Index(0 1) Range(0 1 2) Path(0 1) AddrOf(0 1) Break(0 1) +- Continue(0) Ret(0) InlineAsm(0) Mac(0) Struct(0 1 2) Repeat(0 1) Paren(0) +- Try(0) Yield(0) Err); +-spanless_eq_enum!(ItemKind; ExternCrate(0) Use(0) Static(0 1 2) Const(0 1) +- Fn(0 1 2 3) Mod(0) ForeignMod(0) GlobalAsm(0) TyAlias(0 1) OpaqueTy(0 1) +- Enum(0 1) Struct(0 1) Union(0 1) Trait(0 1 2 3 4) TraitAlias(0 1) +- Impl(0 1 2 3 4 5 6) Mac(0) MacroDef(0)); ++ Block(0 1) Async(0 1 2) Await(0) TryBlock(0) Assign(0 1 2) AssignOp(0 1 2) ++ Field(0 1) Index(0 1) Range(0 1 2) Path(0 1) AddrOf(0 1 2) Break(0 1) ++ Continue(0) Ret(0) InlineAsm(0) LlvmInlineAsm(0) MacCall(0) Struct(0 1 2) ++ Repeat(0 1) Paren(0) Try(0) Yield(0) Err); ++spanless_eq_enum!(InlineAsmOperand; In(reg expr) Out(reg late expr) ++ InOut(reg late expr) SplitInOut(reg late in_expr out_expr) Const(expr) ++ Sym(expr)); ++spanless_eq_enum!(ItemKind; ExternCrate(0) Use(0) Static(0 1 2) Const(0 1 2) ++ Fn(0 1 2 3) Mod(0) ForeignMod(0) GlobalAsm(0) TyAlias(0 1 2 3) Enum(0 1) ++ Struct(0 1) Union(0 1) Trait(0 1 2 3 4) TraitAlias(0 1) ++ Impl(unsafety polarity defaultness constness generics of_trait self_ty items) ++ MacCall(0) MacroDef(0)); + spanless_eq_enum!(LitKind; Str(0 1) ByteStr(0) Byte(0) Char(0) Int(0 1) +- Float(0 1) FloatUnsuffixed(0) Bool(0) Err(0)); ++ Float(0 1) Bool(0) Err(0)); + spanless_eq_enum!(PatKind; Wild Ident(0 1 2) Struct(0 1 2) TupleStruct(0 1) + Or(0) Path(0 1) Tuple(0) Box(0) Ref(0 1) Lit(0) Range(0 1 2) Slice(0) Rest +- Paren(0) Mac(0)); ++ Paren(0) MacCall(0)); + spanless_eq_enum!(TyKind; Slice(0) Array(0 1) Ptr(0) Rptr(0 1) BareFn(0) Never + Tup(0) Path(0 1) TraitObject(0 1) ImplTrait(0 1) Paren(0) Typeof(0) Infer +- ImplicitSelf Mac(0) Err CVarArgs); ++ ImplicitSelf MacCall(0) Err CVarArgs); + + impl SpanlessEq for Ident { + fn eq(&self, other: &Self) -> bool { +@@ -414,44 +419,20 @@ impl SpanlessEq for TokenKind { + + impl SpanlessEq for TokenStream { + fn eq(&self, other: &Self) -> bool { +- SpanlessEq::eq(&expand_tts(self), &expand_tts(other)) +- } +-} +- +-fn expand_tts(tts: &TokenStream) -> Vec<TokenTree> { +- let mut tokens = Vec::new(); +- for tt in tts.clone().into_trees() { +- let c = match tt { +- TokenTree::Token(Token { +- kind: TokenKind::DocComment(c), +- .. +- }) => c, +- _ => { +- tokens.push(tt); +- continue; ++ let mut this = self.clone().into_trees(); ++ let mut other = other.clone().into_trees(); ++ loop { ++ let this = match this.next() { ++ None => return other.next().is_none(), ++ Some(val) => val, ++ }; ++ let other = match other.next() { ++ None => return false, ++ Some(val) => val, ++ }; ++ if !SpanlessEq::eq(&this, &other) { ++ return false; + } +- }; +- let contents = comments::strip_doc_comment_decoration(&c.as_str()); +- let style = comments::doc_comment_style(&c.as_str()); +- tokens.push(TokenTree::token(TokenKind::Pound, DUMMY_SP)); +- if style == AttrStyle::Inner { +- tokens.push(TokenTree::token(TokenKind::Not, DUMMY_SP)); + } +- let lit = token::Lit { +- kind: token::LitKind::Str, +- symbol: Symbol::intern(&contents), +- suffix: None, +- }; +- let tts = vec![ +- TokenTree::token(TokenKind::Ident(sym::doc, false), DUMMY_SP), +- TokenTree::token(TokenKind::Eq, DUMMY_SP), +- TokenTree::token(TokenKind::Literal(lit), DUMMY_SP), +- ]; +- tokens.push(TokenTree::Delimited( +- DelimSpan::dummy(), +- DelimToken::Bracket, +- tts.into_iter().collect::<TokenStream>().into(), +- )); + } +- tokens + } +diff --git a/third_party/rust/syn/tests/common/mod.rs b/third_party/rust/syn/tests/common/mod.rs +index 8b784beed7..a1cc80a16f 100644 +--- a/third_party/rust/syn/tests/common/mod.rs ++++ b/third_party/rust/syn/tests/common/mod.rs +@@ -1,5 +1,6 @@ + #![allow(dead_code)] + ++use rayon::ThreadPoolBuilder; + use std::env; + + pub mod eq; +@@ -12,3 +13,15 @@ pub fn abort_after() -> usize { + Err(_) => usize::max_value(), + } + } ++ ++/// Configure Rayon threadpool. ++pub fn rayon_init() { ++ let stack_size = match env::var("RUST_MIN_STACK") { ++ Ok(s) => s.parse().expect("failed to parse RUST_MIN_STACK"), ++ Err(_) => 20 * 1024 * 1024, ++ }; ++ ThreadPoolBuilder::new() ++ .stack_size(stack_size) ++ .build_global() ++ .unwrap(); ++} +diff --git a/third_party/rust/syn/tests/common/parse.rs b/third_party/rust/syn/tests/common/parse.rs +index 41d192f6fb..192828fedd 100644 +--- a/third_party/rust/syn/tests/common/parse.rs ++++ b/third_party/rust/syn/tests/common/parse.rs +@@ -1,20 +1,20 @@ +-extern crate proc_macro2; +-extern crate syn; +-extern crate syntax; +-extern crate syntax_pos; ++extern crate rustc_ast; ++extern crate rustc_expand; ++extern crate rustc_parse as parse; ++extern crate rustc_session; ++extern crate rustc_span; + +-use self::syntax::ast; +-use self::syntax::parse::{self, ParseSess}; +-use self::syntax::ptr::P; +-use self::syntax::source_map::FilePathMapping; +-use self::syntax_pos::FileName; ++use rustc_ast::ast; ++use rustc_ast::ptr::P; ++use rustc_session::parse::ParseSess; ++use rustc_span::source_map::FilePathMapping; ++use rustc_span::FileName; + + use std::panic; + +-pub fn libsyntax_expr(input: &str) -> Option<P<ast::Expr>> { ++pub fn librustc_expr(input: &str) -> Option<P<ast::Expr>> { + match panic::catch_unwind(|| { + let sess = ParseSess::new(FilePathMapping::empty()); +- sess.span_diagnostic.set_continue_after_error(false); + let e = parse::new_parser_from_source_str( + &sess, + FileName::Custom("test_precedence".to_string()), +@@ -32,7 +32,7 @@ pub fn libsyntax_expr(input: &str) -> Option<P<ast::Expr>> { + Ok(Some(e)) => Some(e), + Ok(None) => None, + Err(_) => { +- errorf!("libsyntax panicked\n"); ++ errorf!("librustc panicked\n"); + None + } + } +diff --git a/third_party/rust/syn/tests/debug/gen.rs b/third_party/rust/syn/tests/debug/gen.rs +index 8450c09ecf..85a1a39079 100644 +--- a/third_party/rust/syn/tests/debug/gen.rs ++++ b/third_party/rust/syn/tests/debug/gen.rs +@@ -2,7 +2,7 @@ + // It is not intended for manual editing. + + use super::{Lite, RefCast}; +-use std::fmt::{self, Debug}; ++use std::fmt::{self, Debug, Display}; + impl Debug for Lite<syn::Abi> { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + let _val = &self.value; +@@ -1039,9 +1039,9 @@ impl Debug for Lite<syn::Expr> { + } + syn::Expr::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + syn::Expr::While(_val) => { +@@ -2116,9 +2116,9 @@ impl Debug for Lite<syn::ForeignItem> { + } + syn::ForeignItem::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + _ => unreachable!(), +@@ -2432,9 +2432,9 @@ impl Debug for Lite<syn::ImplItem> { + } + syn::ImplItem::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + _ => unreachable!(), +@@ -2940,9 +2940,9 @@ impl Debug for Lite<syn::Item> { + } + syn::Item::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + _ => unreachable!(), +@@ -3437,9 +3437,9 @@ impl Debug for Lite<syn::Lit> { + } + syn::Lit::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + } +@@ -3878,9 +3878,9 @@ impl Debug for Lite<syn::Pat> { + } + syn::Pat::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + syn::Pat::Wild(_val) => { +@@ -4674,9 +4674,9 @@ impl Debug for Lite<syn::TraitItem> { + } + syn::TraitItem::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + _ => unreachable!(), +@@ -5040,9 +5040,9 @@ impl Debug for Lite<syn::Type> { + } + syn::Type::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + _ => unreachable!(), +diff --git a/third_party/rust/syn/tests/debug/mod.rs b/third_party/rust/syn/tests/debug/mod.rs +index c1180532ec..cefebacef7 100644 +--- a/third_party/rust/syn/tests/debug/mod.rs ++++ b/third_party/rust/syn/tests/debug/mod.rs +@@ -1,10 +1,7 @@ +-extern crate proc_macro2; +-extern crate ref_cast; +- + mod gen; + +-use self::proc_macro2::{Ident, Literal, TokenStream}; +-use self::ref_cast::RefCast; ++use proc_macro2::{Ident, Literal, TokenStream}; ++use ref_cast::RefCast; + use std::fmt::{self, Debug}; + use std::ops::Deref; + use syn::punctuated::Punctuated; +@@ -66,7 +63,15 @@ impl Debug for Lite<Literal> { + + impl Debug for Lite<TokenStream> { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { +- write!(formatter, "`{}`", self.value) ++ let string = self.value.to_string(); ++ if string.len() <= 80 { ++ write!(formatter, "TokenStream(`{}`)", self.value) ++ } else { ++ formatter ++ .debug_tuple("TokenStream") ++ .field(&format_args!("`{}`", string)) ++ .finish() ++ } + } + } + +diff --git a/third_party/rust/syn/tests/features/error.rs b/third_party/rust/syn/tests/features/error.rs +deleted file mode 100644 +index 10ac88965d..0000000000 +--- a/third_party/rust/syn/tests/features/error.rs ++++ /dev/null +@@ -1 +0,0 @@ +-"Hello! You want: cargo test --release --all-features" +diff --git a/third_party/rust/syn/tests/features/mod.rs b/third_party/rust/syn/tests/features/mod.rs +deleted file mode 100644 +index 83fbe13e7e..0000000000 +--- a/third_party/rust/syn/tests/features/mod.rs ++++ /dev/null +@@ -1,22 +0,0 @@ +-#[allow(unused_macros)] +-macro_rules! hide_from_rustfmt { +- ($mod:item) => { +- $mod +- }; +-} +- +-#[cfg(not(all( +- feature = "derive", +- feature = "full", +- feature = "parsing", +- feature = "printing", +- feature = "visit", +- feature = "visit-mut", +- feature = "fold", +- feature = "clone-impls", +- feature = "extra-traits", +- feature = "proc-macro", +-)))] +-hide_from_rustfmt! { +- mod error; +-} +diff --git a/third_party/rust/syn/tests/macros/mod.rs b/third_party/rust/syn/tests/macros/mod.rs +index c72fd01058..3994615fc4 100644 +--- a/third_party/rust/syn/tests/macros/mod.rs ++++ b/third_party/rust/syn/tests/macros/mod.rs +@@ -1,5 +1,3 @@ +-extern crate proc_macro2; +- + #[path = "../debug/mod.rs"] + pub mod debug; + +@@ -42,18 +40,18 @@ macro_rules! snapshot_impl { + (($expr:ident) as $t:ty, @$snapshot:literal) => { + let $expr = crate::macros::Tokens::parse::<$t>($expr).unwrap(); + let debug = crate::macros::debug::Lite(&$expr); +- insta::assert_debug_snapshot_matches!(debug, @$snapshot); ++ insta::assert_debug_snapshot!(debug, @$snapshot); + }; + (($($expr:tt)*) as $t:ty, @$snapshot:literal) => {{ + let syntax_tree = crate::macros::Tokens::parse::<$t>($($expr)*).unwrap(); + let debug = crate::macros::debug::Lite(&syntax_tree); +- insta::assert_debug_snapshot_matches!(debug, @$snapshot); ++ insta::assert_debug_snapshot!(debug, @$snapshot); + syntax_tree + }}; + (($($expr:tt)*) , @$snapshot:literal) => {{ + let syntax_tree = $($expr)*; + let debug = crate::macros::debug::Lite(&syntax_tree); +- insta::assert_debug_snapshot_matches!(debug, @$snapshot); ++ insta::assert_debug_snapshot!(debug, @$snapshot); + syntax_tree + }}; + (($($expr:tt)*) $next:tt $($rest:tt)*) => { +diff --git a/third_party/rust/syn/tests/repo/mod.rs b/third_party/rust/syn/tests/repo/mod.rs +index c22cb03758..1d3e1f0e74 100644 +--- a/third_party/rust/syn/tests/repo/mod.rs ++++ b/third_party/rust/syn/tests/repo/mod.rs +@@ -1,8 +1,37 @@ +-extern crate walkdir; ++mod progress; + +-use std::process::Command; ++use self::progress::Progress; ++use anyhow::Result; ++use flate2::read::GzDecoder; ++use std::fs; ++use std::path::Path; ++use tar::Archive; ++use walkdir::DirEntry; + +-use self::walkdir::DirEntry; ++const REVISION: &str = "792c645ca7d11a8d254df307d019c5bf01445c37"; ++ ++#[rustfmt::skip] ++static EXCLUDE: &[&str] = &[ ++ // Compile-fail expr parameter in const generic position: f::<1 + 2>() ++ "test/ui/const-generics/const-expression-parameter.rs", ++ ++ // Deprecated anonymous parameter syntax in traits ++ "test/ui/issues/issue-13105.rs", ++ "test/ui/issues/issue-13775.rs", ++ "test/ui/issues/issue-34074.rs", ++ "test/ui/proc-macro/trait-fn-args-2015.rs", ++ ++ // Not actually test cases ++ "test/rustdoc-ui/test-compile-fail2.rs", ++ "test/rustdoc-ui/test-compile-fail3.rs", ++ "test/ui/include-single-expr-helper.rs", ++ "test/ui/include-single-expr-helper-1.rs", ++ "test/ui/issues/auxiliary/issue-21146-inc.rs", ++ "test/ui/json-bom-plus-crlf-multifile-aux.rs", ++ "test/ui/lint/expansion-time-include.rs", ++ "test/ui/macros/auxiliary/macro-comma-support.rs", ++ "test/ui/macros/auxiliary/macro-include-items-expr.rs", ++]; + + pub fn base_dir_filter(entry: &DirEntry) -> bool { + let path = entry.path(); +@@ -12,49 +41,95 @@ pub fn base_dir_filter(entry: &DirEntry) -> bool { + if path.extension().map(|e| e != "rs").unwrap_or(true) { + return false; + } +- let path_string = path.to_string_lossy(); +- let path_string = if cfg!(windows) { +- path_string.replace('\\', "/").into() ++ ++ let mut path_string = path.to_string_lossy(); ++ if cfg!(windows) { ++ path_string = path_string.replace('\\', "/").into(); ++ } ++ let path = if let Some(path) = path_string.strip_prefix("tests/rust/src/") { ++ path ++ } else if let Some(path) = path_string.strip_prefix("tests/rust/library/") { ++ path + } else { +- path_string ++ panic!("unexpected path in Rust dist: {}", path_string); + }; ++ + // TODO assert that parsing fails on the parse-fail cases +- if path_string.starts_with("tests/rust/src/test/parse-fail") +- || path_string.starts_with("tests/rust/src/test/compile-fail") +- || path_string.starts_with("tests/rust/src/test/rustfix") ++ if path.starts_with("test/parse-fail") ++ || path.starts_with("test/compile-fail") ++ || path.starts_with("test/rustfix") + { + return false; + } + +- if path_string.starts_with("tests/rust/src/test/ui") { +- let stderr_path = path.with_extension("stderr"); ++ if path.starts_with("test/ui") { ++ let stderr_path = entry.path().with_extension("stderr"); + if stderr_path.exists() { + // Expected to fail in some way + return false; + } + } + +- match path_string.as_ref() { +- // Deprecated placement syntax +- "tests/rust/src/test/ui/obsolete-in-place/bad.rs" | +- // Deprecated anonymous parameter syntax in traits +- "tests/rust/src/test/ui/error-codes/e0119/auxiliary/issue-23563-a.rs" | +- "tests/rust/src/test/ui/issues/issue-13105.rs" | +- "tests/rust/src/test/ui/issues/issue-13775.rs" | +- "tests/rust/src/test/ui/issues/issue-34074.rs" | +- // Deprecated await macro syntax +- "tests/rust/src/test/ui/async-await/await-macro.rs" | +- // 2015-style dyn that libsyntax rejects +- "tests/rust/src/test/ui/dyn-keyword/dyn-2015-no-warnings-without-lints.rs" | +- // not actually test cases +- "tests/rust/src/test/ui/macros/auxiliary/macro-comma-support.rs" | +- "tests/rust/src/test/ui/macros/auxiliary/macro-include-items-expr.rs" | +- "tests/rust/src/test/ui/issues/auxiliary/issue-21146-inc.rs" => false, +- _ => true, ++ !EXCLUDE.contains(&path) ++} ++ ++#[allow(dead_code)] ++pub fn edition(path: &Path) -> &'static str { ++ if path.ends_with("dyn-2015-no-warnings-without-lints.rs") { ++ "2015" ++ } else { ++ "2018" + } + } + + pub fn clone_rust() { +- let result = Command::new("tests/clone.sh").status().unwrap(); +- assert!(result.success()); ++ let needs_clone = match fs::read_to_string("tests/rust/COMMIT") { ++ Err(_) => true, ++ Ok(contents) => contents.trim() != REVISION, ++ }; ++ if needs_clone { ++ download_and_unpack().unwrap(); ++ } ++ let mut missing = String::new(); ++ let test_src = Path::new("tests/rust/src"); ++ for exclude in EXCLUDE { ++ if !test_src.join(exclude).exists() { ++ missing += "\ntests/rust/src/"; ++ missing += exclude; ++ } ++ } ++ if !missing.is_empty() { ++ panic!("excluded test file does not exist:{}\n", missing); ++ } ++} ++ ++fn download_and_unpack() -> Result<()> { ++ let url = format!( ++ "https://github.com/rust-lang/rust/archive/{}.tar.gz", ++ REVISION ++ ); ++ let response = reqwest::blocking::get(&url)?.error_for_status()?; ++ let progress = Progress::new(response); ++ let decoder = GzDecoder::new(progress); ++ let mut archive = Archive::new(decoder); ++ let prefix = format!("rust-{}", REVISION); ++ ++ let tests_rust = Path::new("tests/rust"); ++ if tests_rust.exists() { ++ fs::remove_dir_all(tests_rust)?; ++ } ++ ++ for entry in archive.entries()? { ++ let mut entry = entry?; ++ let path = entry.path()?; ++ if path == Path::new("pax_global_header") { ++ continue; ++ } ++ let relative = path.strip_prefix(&prefix)?; ++ let out = tests_rust.join(relative); ++ entry.unpack(&out)?; ++ } ++ ++ fs::write("tests/rust/COMMIT", REVISION)?; ++ Ok(()) + } +diff --git a/third_party/rust/syn/tests/repo/progress.rs b/third_party/rust/syn/tests/repo/progress.rs +new file mode 100644 +index 0000000000..28c8a44b12 +--- /dev/null ++++ b/third_party/rust/syn/tests/repo/progress.rs +@@ -0,0 +1,37 @@ ++use std::io::{Read, Result}; ++use std::time::{Duration, Instant}; ++ ++pub struct Progress<R> { ++ bytes: usize, ++ tick: Instant, ++ stream: R, ++} ++ ++impl<R> Progress<R> { ++ pub fn new(stream: R) -> Self { ++ Progress { ++ bytes: 0, ++ tick: Instant::now() + Duration::from_millis(2000), ++ stream, ++ } ++ } ++} ++ ++impl<R: Read> Read for Progress<R> { ++ fn read(&mut self, buf: &mut [u8]) -> Result<usize> { ++ let num = self.stream.read(buf)?; ++ self.bytes += num; ++ let now = Instant::now(); ++ if now > self.tick { ++ self.tick = now + Duration::from_millis(500); ++ errorf!("downloading... {} bytes\n", self.bytes); ++ } ++ Ok(num) ++ } ++} ++ ++impl<R> Drop for Progress<R> { ++ fn drop(&mut self) { ++ errorf!("done ({} bytes)\n", self.bytes); ++ } ++} +diff --git a/third_party/rust/syn/tests/test_asyncness.rs b/third_party/rust/syn/tests/test_asyncness.rs +index f868fbcc20..0efef5976f 100644 +--- a/third_party/rust/syn/tests/test_asyncness.rs ++++ b/third_party/rust/syn/tests/test_asyncness.rs +@@ -1,7 +1,3 @@ +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -12,16 +8,16 @@ fn test_async_fn() { + let input = "async fn process() {}"; + + snapshot!(input as Item, @r###" +- ⋮Item::Fn { +- ⋮ vis: Inherited, +- ⋮ sig: Signature { +- ⋮ asyncness: Some, +- ⋮ ident: "process", +- ⋮ generics: Generics, +- ⋮ output: Default, +- ⋮ }, +- ⋮ block: Block, +- ⋮} ++ Item::Fn { ++ vis: Inherited, ++ sig: Signature { ++ asyncness: Some, ++ ident: "process", ++ generics: Generics, ++ output: Default, ++ }, ++ block: Block, ++ } + "###); + } + +@@ -30,12 +26,12 @@ fn test_async_closure() { + let input = "async || {}"; + + snapshot!(input as Expr, @r###" +- ⋮Expr::Closure { +- ⋮ asyncness: Some, +- ⋮ output: Default, +- ⋮ body: Expr::Block { +- ⋮ block: Block, +- ⋮ }, +- ⋮} ++ Expr::Closure { ++ asyncness: Some, ++ output: Default, ++ body: Expr::Block { ++ block: Block, ++ }, ++ } + "###); + } +diff --git a/third_party/rust/syn/tests/test_attribute.rs b/third_party/rust/syn/tests/test_attribute.rs +index aff6294fc3..c26bd090ec 100644 +--- a/third_party/rust/syn/tests/test_attribute.rs ++++ b/third_party/rust/syn/tests/test_attribute.rs +@@ -1,7 +1,3 @@ +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -13,14 +9,14 @@ fn test_meta_item_word() { + let meta = test("#[foo]"); + + snapshot!(meta, @r###" +- ⋮Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮}) ++ Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }) + "###); + } + +@@ -29,17 +25,17 @@ fn test_meta_item_name_value() { + let meta = test("#[foo = 5]"); + + snapshot!(meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + } + +@@ -48,37 +44,37 @@ fn test_meta_item_bool_value() { + let meta = test("#[foo = true]"); + + snapshot!(meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: Lit::Bool { +- ⋮ value: true, +- ⋮ }, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: Lit::Bool { ++ value: true, ++ }, ++ } + "###); + + let meta = test("#[foo = false]"); + + snapshot!(meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: Lit::Bool { +- ⋮ value: false, +- ⋮ }, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: Lit::Bool { ++ value: false, ++ }, ++ } + "###); + } + +@@ -87,19 +83,19 @@ fn test_meta_item_list_lit() { + let meta = test("#[foo(5)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Lit(5), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Lit(5), ++ ], ++ } + "###); + } + +@@ -108,26 +104,26 @@ fn test_meta_item_list_word() { + let meta = test("#[foo(bar)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "bar", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "bar", ++ arguments: None, ++ }, ++ ], ++ })), ++ ], ++ } + "###); + } + +@@ -136,29 +132,29 @@ fn test_meta_item_list_name_value() { + let meta = test("#[foo(bar = 5)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "bar", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮ }), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "bar", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ }), ++ ], ++ } + "###); + } + +@@ -167,31 +163,31 @@ fn test_meta_item_list_bool_value() { + let meta = test("#[foo(bar = true)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "bar", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: Lit::Bool { +- ⋮ value: true, +- ⋮ }, +- ⋮ }), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "bar", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: Lit::Bool { ++ value: true, ++ }, ++ }), ++ ], ++ } + "###); + } + +@@ -200,68 +196,68 @@ fn test_meta_item_multiple() { + let meta = test("#[foo(word, name = 5, list(name2 = 6), word2)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮ }), +- ⋮ Meta(Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "list", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 6, +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word", ++ arguments: None, ++ }, ++ ], ++ })), ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ }), ++ Meta(Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "list", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name2", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 6, ++ }), ++ ], ++ }), ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word2", ++ arguments: None, ++ }, ++ ], ++ })), ++ ], ++ } + "###); + } + +@@ -270,21 +266,63 @@ fn test_bool_lit() { + let meta = test("#[foo(true)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Lit(Lit::Bool { +- ⋮ value: true, +- ⋮ }), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Lit(Lit::Bool { ++ value: true, ++ }), ++ ], ++ } ++ "###); ++} ++ ++#[test] ++fn test_negative_lit() { ++ let meta = test("#[form(min = -1, max = 200)]"); ++ ++ snapshot!(meta, @r###" ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "form", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "min", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: -1, ++ }), ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "max", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 200, ++ }), ++ ], ++ } + "###); + } + +diff --git a/third_party/rust/syn/tests/test_derive_input.rs b/third_party/rust/syn/tests/test_derive_input.rs +index de68240166..bf1ebdb67d 100644 +--- a/third_party/rust/syn/tests/test_derive_input.rs ++++ b/third_party/rust/syn/tests/test_derive_input.rs +@@ -1,8 +1,3 @@ +-extern crate quote; +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -16,15 +11,15 @@ fn test_unit() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "Unit", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "Unit", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -39,105 +34,105 @@ fn test_struct() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "derive", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: `( Debug , Clone )`, +- ⋮ }, +- ⋮ ], +- ⋮ vis: Visibility::Public, +- ⋮ ident: "Item", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Fields::Named { +- ⋮ named: [ +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ident: Some("ident"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Ident", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ident: Some("attrs"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Vec", +- ⋮ arguments: PathArguments::AngleBracketed { +- ⋮ args: [ +- ⋮ Type(Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Attribute", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }), +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "derive", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(`(Debug , Clone)`), ++ }, ++ ], ++ vis: Visibility::Public, ++ ident: "Item", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Named { ++ named: [ ++ Field { ++ vis: Visibility::Public, ++ ident: Some("ident"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Ident", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ Field { ++ vis: Visibility::Public, ++ ident: Some("attrs"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Vec", ++ arguments: PathArguments::AngleBracketed { ++ args: [ ++ Type(Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Attribute", ++ arguments: None, ++ }, ++ ], ++ }, ++ }), ++ ], ++ }, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ }, ++ } + "###); + + snapshot!(input.attrs[0].parse_meta().unwrap(), @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "derive", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Debug", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Clone", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "derive", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "Debug", ++ arguments: None, ++ }, ++ ], ++ })), ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "Clone", ++ arguments: None, ++ }, ++ ], ++ })), ++ ], ++ } + "###); + } + +@@ -151,46 +146,46 @@ fn test_union() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "MaybeUninit", +- ⋮ generics: Generics { +- ⋮ lt_token: Some, +- ⋮ params: [ +- ⋮ Type(TypeParam { +- ⋮ ident: "T", +- ⋮ }), +- ⋮ ], +- ⋮ gt_token: Some, +- ⋮ }, +- ⋮ data: Data::Union { +- ⋮ fields: FieldsNamed { +- ⋮ named: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ident: Some("uninit"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Tuple, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ident: Some("value"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "T", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "MaybeUninit", ++ generics: Generics { ++ lt_token: Some, ++ params: [ ++ Type(TypeParam { ++ ident: "T", ++ }), ++ ], ++ gt_token: Some, ++ }, ++ data: Data::Union { ++ fields: FieldsNamed { ++ named: [ ++ Field { ++ vis: Inherited, ++ ident: Some("uninit"), ++ colon_token: Some, ++ ty: Type::Tuple, ++ }, ++ Field { ++ vis: Inherited, ++ ident: Some("value"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "T", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ }, ++ } + "###); + } + +@@ -212,118 +207,118 @@ fn test_enum() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "doc", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: `= r" See the std::result module documentation for details."`, +- ⋮ }, +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "must_use", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: ``, +- ⋮ }, +- ⋮ ], +- ⋮ vis: Visibility::Public, +- ⋮ ident: "Result", +- ⋮ generics: Generics { +- ⋮ lt_token: Some, +- ⋮ params: [ +- ⋮ Type(TypeParam { +- ⋮ ident: "T", +- ⋮ }), +- ⋮ Type(TypeParam { +- ⋮ ident: "E", +- ⋮ }), +- ⋮ ], +- ⋮ gt_token: Some, +- ⋮ }, +- ⋮ data: Data::Enum { +- ⋮ variants: [ +- ⋮ Variant { +- ⋮ ident: "Ok", +- ⋮ fields: Fields::Unnamed { +- ⋮ unnamed: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "T", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ Variant { +- ⋮ ident: "Err", +- ⋮ fields: Fields::Unnamed { +- ⋮ unnamed: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "E", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ Variant { +- ⋮ ident: "Surprise", +- ⋮ fields: Unit, +- ⋮ discriminant: Some(Expr::Lit { +- ⋮ lit: 0isize, +- ⋮ }), +- ⋮ }, +- ⋮ Variant { +- ⋮ ident: "ProcMacroHack", +- ⋮ fields: Unit, +- ⋮ discriminant: Some(Expr::Field { +- ⋮ base: Expr::Tuple { +- ⋮ elems: [ +- ⋮ Expr::Lit { +- ⋮ lit: 0, +- ⋮ }, +- ⋮ Expr::Lit { +- ⋮ lit: "data", +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ member: Unnamed(Index { +- ⋮ index: 0, +- ⋮ }), +- ⋮ }), +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮} ++ DeriveInput { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "doc", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(`= r" See the std::result module documentation for details."`), ++ }, ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "must_use", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ vis: Visibility::Public, ++ ident: "Result", ++ generics: Generics { ++ lt_token: Some, ++ params: [ ++ Type(TypeParam { ++ ident: "T", ++ }), ++ Type(TypeParam { ++ ident: "E", ++ }), ++ ], ++ gt_token: Some, ++ }, ++ data: Data::Enum { ++ variants: [ ++ Variant { ++ ident: "Ok", ++ fields: Fields::Unnamed { ++ unnamed: [ ++ Field { ++ vis: Inherited, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "T", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ }, ++ Variant { ++ ident: "Err", ++ fields: Fields::Unnamed { ++ unnamed: [ ++ Field { ++ vis: Inherited, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "E", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ }, ++ Variant { ++ ident: "Surprise", ++ fields: Unit, ++ discriminant: Some(Expr::Lit { ++ lit: 0isize, ++ }), ++ }, ++ Variant { ++ ident: "ProcMacroHack", ++ fields: Unit, ++ discriminant: Some(Expr::Field { ++ base: Expr::Tuple { ++ elems: [ ++ Expr::Lit { ++ lit: 0, ++ }, ++ Expr::Lit { ++ lit: "data", ++ }, ++ ], ++ }, ++ member: Unnamed(Index { ++ index: 0, ++ }), ++ }), ++ }, ++ ], ++ }, ++ } + "###); + + let meta_items: Vec<_> = input +@@ -333,27 +328,27 @@ fn test_enum() { + .collect(); + + snapshot!(meta_items, @r###" +- ⋮[ +- ⋮ Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "doc", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: " See the std::result module documentation for details.", +- ⋮ }, +- ⋮ Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "must_use", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }), +- ⋮] ++ [ ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "doc", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: " See the std::result module documentation for details.", ++ }, ++ Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "must_use", ++ arguments: None, ++ }, ++ ], ++ }), ++ ] + "###); + } + +@@ -366,34 +361,34 @@ fn test_attr_with_path() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ leading_colon: Some, +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "attr_args", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ PathSegment { +- ⋮ ident: "identity", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: `fn main ( ) { assert_eq ! ( foo ( ) , "Hello, world!" ) ; }`, +- ⋮ }, +- ⋮ ], +- ⋮ vis: Inherited, +- ⋮ ident: "Dummy", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ leading_colon: Some, ++ segments: [ ++ PathSegment { ++ ident: "attr_args", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "identity", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(`fn main () { assert_eq ! (foo () , "Hello, world!") ; }`), ++ }, ++ ], ++ vis: Inherited, ++ ident: "Dummy", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + + assert!(input.attrs[0].parse_meta().is_err()); +@@ -407,29 +402,29 @@ fn test_attr_with_non_mod_style_path() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "inert", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: `< T >`, +- ⋮ }, +- ⋮ ], +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "inert", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(`< T >`), ++ }, ++ ], ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + + assert!(input.attrs[0].parse_meta().is_err()); +@@ -443,48 +438,48 @@ fn test_attr_with_mod_style_path_with_self() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ PathSegment { +- ⋮ ident: "self", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: ``, +- ⋮ }, +- ⋮ ], +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "self", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + + snapshot!(input.attrs[0].parse_meta().unwrap(), @r###" +- ⋮Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ PathSegment { +- ⋮ ident: "self", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮}) ++ Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "self", ++ arguments: None, ++ }, ++ ], ++ }) + "###); + } + +@@ -496,55 +491,55 @@ fn test_pub_restricted() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Visibility::Restricted { +- ⋮ in_token: Some, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "m", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ident: "Z", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Fields::Unnamed { +- ⋮ unnamed: [ +- ⋮ Field { +- ⋮ vis: Visibility::Restricted { +- ⋮ in_token: Some, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "m", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ PathSegment { +- ⋮ ident: "n", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "u8", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Visibility::Restricted { ++ in_token: Some, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "m", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ident: "Z", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Unnamed { ++ unnamed: [ ++ Field { ++ vis: Visibility::Restricted { ++ in_token: Some, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "m", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "n", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "u8", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -555,15 +550,15 @@ fn test_vis_crate() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Visibility::Crate, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Visibility::Crate, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -574,24 +569,24 @@ fn test_pub_restricted_crate() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Visibility::Restricted { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "crate", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Visibility::Restricted { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "crate", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -602,24 +597,24 @@ fn test_pub_restricted_super() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Visibility::Restricted { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "super", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Visibility::Restricted { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "super", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -630,25 +625,25 @@ fn test_pub_restricted_in_super() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Visibility::Restricted { +- ⋮ in_token: Some, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "super", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Visibility::Restricted { ++ in_token: Some, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "super", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -659,15 +654,15 @@ fn test_fields_on_unit_struct() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + + let data = match input.data { +@@ -688,47 +683,47 @@ fn test_fields_on_named_struct() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Fields::Named { +- ⋮ named: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ident: Some("foo"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "i32", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ident: Some("bar"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "String", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Named { ++ named: [ ++ Field { ++ vis: Inherited, ++ ident: Some("foo"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "i32", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ Field { ++ vis: Visibility::Public, ++ ident: Some("bar"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "String", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ }, ++ } + "###); + + let data = match input.data { +@@ -737,38 +732,38 @@ fn test_fields_on_named_struct() { + }; + + snapshot!(data.fields.into_iter().collect::<Vec<_>>(), @r###" +- ⋮[ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ident: Some("foo"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "i32", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ident: Some("bar"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "String", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮] ++ [ ++ Field { ++ vis: Inherited, ++ ident: Some("foo"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "i32", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ Field { ++ vis: Visibility::Public, ++ ident: Some("bar"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "String", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ] + "###); + } + +@@ -779,44 +774,44 @@ fn test_fields_on_tuple_struct() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Fields::Unnamed { +- ⋮ unnamed: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "i32", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "String", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Unnamed { ++ unnamed: [ ++ Field { ++ vis: Inherited, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "i32", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ Field { ++ vis: Visibility::Public, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "String", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ semi_token: Some, ++ }, ++ } + "###); + + let data = match input.data { +@@ -825,34 +820,34 @@ fn test_fields_on_tuple_struct() { + }; + + snapshot!(data.fields.iter().collect::<Vec<_>>(), @r###" +- ⋮[ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "i32", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "String", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮] ++ [ ++ Field { ++ vis: Inherited, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "i32", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ Field { ++ vis: Visibility::Public, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "String", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ] + "###); + } + +@@ -864,34 +859,34 @@ fn test_ambiguous_crate() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Fields::Unnamed { +- ⋮ unnamed: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "crate", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ PathSegment { +- ⋮ ident: "X", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Unnamed { ++ unnamed: [ ++ Field { ++ vis: Inherited, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "crate", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "X", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ semi_token: Some, ++ }, ++ } + "###); + } +diff --git a/third_party/rust/syn/tests/test_expr.rs b/third_party/rust/syn/tests/test_expr.rs +index c8a11cec2c..b2b65a254f 100644 +--- a/third_party/rust/syn/tests/test_expr.rs ++++ b/third_party/rust/syn/tests/test_expr.rs +@@ -1,40 +1,302 @@ +-extern crate proc_macro2; +-extern crate syn; +- + #[macro_use] + mod macros; + +-use std::str::FromStr; +- +-use proc_macro2::TokenStream; ++use proc_macro2::{Delimiter, Group, Ident, Punct, Spacing, Span, TokenStream, TokenTree}; ++use quote::quote; ++use std::iter::FromIterator; + use syn::{Expr, ExprRange}; + + #[test] + fn test_expr_parse() { +- let code = "..100u32"; +- let tt = TokenStream::from_str(code).unwrap(); +- let expr: Expr = syn::parse2(tt.clone()).unwrap(); +- let expr_range: ExprRange = syn::parse2(tt).unwrap(); +- assert_eq!(expr, Expr::Range(expr_range)); ++ let tokens = quote!(..100u32); ++ snapshot!(tokens as Expr, @r###" ++ Expr::Range { ++ limits: HalfOpen, ++ to: Some(Expr::Lit { ++ lit: 100u32, ++ }), ++ } ++ "###); ++ ++ let tokens = quote!(..100u32); ++ snapshot!(tokens as ExprRange, @r###" ++ ExprRange { ++ limits: HalfOpen, ++ to: Some(Expr::Lit { ++ lit: 100u32, ++ }), ++ } ++ "###); + } + + #[test] + fn test_await() { + // Must not parse as Expr::Field. +- let expr = syn::parse_str::<Expr>("fut.await").unwrap(); +- +- snapshot!(expr, @r###" +- ⋮Expr::Await { +- ⋮ base: Expr::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "fut", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮} ++ let tokens = quote!(fut.await); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Await { ++ base: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "fut", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ } ++ "###); ++} ++ ++#[rustfmt::skip] ++#[test] ++fn test_tuple_multi_index() { ++ for &input in &[ ++ "tuple.0.0", ++ "tuple .0.0", ++ "tuple. 0.0", ++ "tuple.0 .0", ++ "tuple.0. 0", ++ "tuple . 0 . 0", ++ ] { ++ snapshot!(input as Expr, @r###" ++ Expr::Field { ++ base: Expr::Field { ++ base: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "tuple", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ member: Unnamed(Index { ++ index: 0, ++ }), ++ }, ++ member: Unnamed(Index { ++ index: 0, ++ }), ++ } ++ "###); ++ } ++ ++ for tokens in vec![ ++ quote!(tuple.0.0), ++ quote!(tuple .0.0), ++ quote!(tuple. 0.0), ++ quote!(tuple.0 .0), ++ quote!(tuple.0. 0), ++ quote!(tuple . 0 . 0), ++ ] { ++ snapshot!(tokens as Expr, @r###" ++ Expr::Field { ++ base: Expr::Field { ++ base: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "tuple", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ member: Unnamed(Index { ++ index: 0, ++ }), ++ }, ++ member: Unnamed(Index { ++ index: 0, ++ }), ++ } ++ "###); ++ } ++} ++ ++#[test] ++fn test_macro_variable_func() { ++ // mimics the token stream corresponding to `$fn()` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { f })), ++ TokenTree::Group(Group::new(Delimiter::Parenthesis, TokenStream::new())), ++ ]); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Call { ++ func: Expr::Group { ++ expr: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "f", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ } ++ "###); ++ ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Punct(Punct::new('#', Spacing::Alone)), ++ TokenTree::Group(Group::new(Delimiter::Bracket, quote! { outside })), ++ TokenTree::Group(Group::new(Delimiter::None, quote! { #[inside] f })), ++ TokenTree::Group(Group::new(Delimiter::Parenthesis, TokenStream::new())), ++ ]); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Call { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "outside", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ func: Expr::Group { ++ expr: Expr::Path { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "inside", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "f", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ } ++ "###); ++} ++ ++#[test] ++fn test_macro_variable_macro() { ++ // mimics the token stream corresponding to `$macro!()` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { m })), ++ TokenTree::Punct(Punct::new('!', Spacing::Alone)), ++ TokenTree::Group(Group::new(Delimiter::Parenthesis, TokenStream::new())), ++ ]); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Macro { ++ mac: Macro { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "m", ++ arguments: None, ++ }, ++ ], ++ }, ++ delimiter: Paren, ++ tokens: TokenStream(``), ++ }, ++ } ++ "###); ++} ++ ++#[test] ++fn test_macro_variable_struct() { ++ // mimics the token stream corresponding to `$struct {}` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { S })), ++ TokenTree::Group(Group::new(Delimiter::Brace, TokenStream::new())), ++ ]); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Struct { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "S", ++ arguments: None, ++ }, ++ ], ++ }, ++ } ++ "###); ++} ++ ++#[test] ++fn test_macro_variable_match_arm() { ++ // mimics the token stream corresponding to `match v { _ => $expr }` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Ident(Ident::new("match", Span::call_site())), ++ TokenTree::Ident(Ident::new("v", Span::call_site())), ++ TokenTree::Group(Group::new( ++ Delimiter::Brace, ++ TokenStream::from_iter(vec![ ++ TokenTree::Punct(Punct::new('_', Spacing::Alone)), ++ TokenTree::Punct(Punct::new('=', Spacing::Joint)), ++ TokenTree::Punct(Punct::new('>', Spacing::Alone)), ++ TokenTree::Group(Group::new(Delimiter::None, quote! { #[a] () })), ++ ]), ++ )), ++ ]); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Match { ++ expr: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "v", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ arms: [ ++ Arm { ++ pat: Pat::Wild, ++ body: Expr::Group { ++ expr: Expr::Tuple { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "a", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ } + "###); + } +diff --git a/third_party/rust/syn/tests/test_generics.rs b/third_party/rust/syn/tests/test_generics.rs +index 55c79e066b..b29434a147 100644 +--- a/third_party/rust/syn/tests/test_generics.rs ++++ b/third_party/rust/syn/tests/test_generics.rs +@@ -1,8 +1,3 @@ +-extern crate quote; +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -16,90 +11,90 @@ fn test_split_for_impl() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics { +- ⋮ lt_token: Some, +- ⋮ params: [ +- ⋮ Lifetime(LifetimeDef { +- ⋮ lifetime: Lifetime { +- ⋮ ident: "a", +- ⋮ }, +- ⋮ }), +- ⋮ Lifetime(LifetimeDef { +- ⋮ lifetime: Lifetime { +- ⋮ ident: "b", +- ⋮ }, +- ⋮ colon_token: Some, +- ⋮ bounds: [ +- ⋮ Lifetime { +- ⋮ ident: "a", +- ⋮ }, +- ⋮ ], +- ⋮ }), +- ⋮ Type(TypeParam { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "may_dangle", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: ``, +- ⋮ }, +- ⋮ ], +- ⋮ ident: "T", +- ⋮ colon_token: Some, +- ⋮ bounds: [ +- ⋮ Lifetime(Lifetime { +- ⋮ ident: "a", +- ⋮ }), +- ⋮ ], +- ⋮ eq_token: Some, +- ⋮ default: Some(Type::Tuple), +- ⋮ }), +- ⋮ ], +- ⋮ gt_token: Some, +- ⋮ where_clause: Some(WhereClause { +- ⋮ predicates: [ +- ⋮ Type(PredicateType { +- ⋮ bounded_ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "T", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ bounds: [ +- ⋮ Trait(TraitBound { +- ⋮ modifier: None, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Debug", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ }, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics { ++ lt_token: Some, ++ params: [ ++ Lifetime(LifetimeDef { ++ lifetime: Lifetime { ++ ident: "a", ++ }, ++ }), ++ Lifetime(LifetimeDef { ++ lifetime: Lifetime { ++ ident: "b", ++ }, ++ colon_token: Some, ++ bounds: [ ++ Lifetime { ++ ident: "a", ++ }, ++ ], ++ }), ++ Type(TypeParam { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "may_dangle", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ ident: "T", ++ colon_token: Some, ++ bounds: [ ++ Lifetime(Lifetime { ++ ident: "a", ++ }), ++ ], ++ eq_token: Some, ++ default: Some(Type::Tuple), ++ }), ++ ], ++ gt_token: Some, ++ where_clause: Some(WhereClause { ++ predicates: [ ++ Type(PredicateType { ++ bounded_ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "T", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ bounds: [ ++ Trait(TraitBound { ++ modifier: None, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Debug", ++ arguments: None, ++ }, ++ ], ++ }, ++ }), ++ ], ++ }), ++ ], ++ }), ++ }, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + + let generics = input.generics; +@@ -131,46 +126,46 @@ fn test_split_for_impl() { + fn test_ty_param_bound() { + let tokens = quote!('a); + snapshot!(tokens as TypeParamBound, @r###" +- ⋮Lifetime(Lifetime { +- ⋮ ident: "a", +- ⋮}) ++ Lifetime(Lifetime { ++ ident: "a", ++ }) + "###); + + let tokens = quote!('_); + snapshot!(tokens as TypeParamBound, @r###" +- ⋮Lifetime(Lifetime { +- ⋮ ident: "_", +- ⋮}) ++ Lifetime(Lifetime { ++ ident: "_", ++ }) + "###); + + let tokens = quote!(Debug); + snapshot!(tokens as TypeParamBound, @r###" +- ⋮Trait(TraitBound { +- ⋮ modifier: None, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Debug", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮}) ++ Trait(TraitBound { ++ modifier: None, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Debug", ++ arguments: None, ++ }, ++ ], ++ }, ++ }) + "###); + + let tokens = quote!(?Sized); + snapshot!(tokens as TypeParamBound, @r###" +- ⋮Trait(TraitBound { +- ⋮ modifier: Maybe, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Sized", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮}) ++ Trait(TraitBound { ++ modifier: Maybe, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Sized", ++ arguments: None, ++ }, ++ ], ++ }, ++ }) + "###); + } + +@@ -187,76 +182,76 @@ fn test_fn_precedence_in_where_clause() { + }; + + snapshot!(input as ItemFn, @r###" +- ⋮ItemFn { +- ⋮ vis: Inherited, +- ⋮ sig: Signature { +- ⋮ ident: "f", +- ⋮ generics: Generics { +- ⋮ lt_token: Some, +- ⋮ params: [ +- ⋮ Type(TypeParam { +- ⋮ ident: "G", +- ⋮ }), +- ⋮ ], +- ⋮ gt_token: Some, +- ⋮ where_clause: Some(WhereClause { +- ⋮ predicates: [ +- ⋮ Type(PredicateType { +- ⋮ bounded_ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "G", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ bounds: [ +- ⋮ Trait(TraitBound { +- ⋮ modifier: None, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "FnOnce", +- ⋮ arguments: PathArguments::Parenthesized { +- ⋮ output: Type( +- ⋮ Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "i32", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ), +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }), +- ⋮ Trait(TraitBound { +- ⋮ modifier: None, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Send", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ }, +- ⋮ output: Default, +- ⋮ }, +- ⋮ block: Block, +- ⋮} ++ ItemFn { ++ vis: Inherited, ++ sig: Signature { ++ ident: "f", ++ generics: Generics { ++ lt_token: Some, ++ params: [ ++ Type(TypeParam { ++ ident: "G", ++ }), ++ ], ++ gt_token: Some, ++ where_clause: Some(WhereClause { ++ predicates: [ ++ Type(PredicateType { ++ bounded_ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "G", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ bounds: [ ++ Trait(TraitBound { ++ modifier: None, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "FnOnce", ++ arguments: PathArguments::Parenthesized { ++ output: Type( ++ Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "i32", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ), ++ }, ++ }, ++ ], ++ }, ++ }), ++ Trait(TraitBound { ++ modifier: None, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Send", ++ arguments: None, ++ }, ++ ], ++ }, ++ }), ++ ], ++ }), ++ ], ++ }), ++ }, ++ output: Default, ++ }, ++ block: Block, ++ } + "###); + + let where_clause = input.sig.generics.where_clause.as_ref().unwrap(); +@@ -270,7 +265,7 @@ fn test_fn_precedence_in_where_clause() { + assert_eq!(predicate.bounds.len(), 2, "{:#?}", predicate.bounds); + + let first_bound = &predicate.bounds[0]; +- assert_eq!(quote!(#first_bound).to_string(), "FnOnce ( ) -> i32"); ++ assert_eq!(quote!(#first_bound).to_string(), "FnOnce () -> i32"); + + let second_bound = &predicate.bounds[1]; + assert_eq!(quote!(#second_bound).to_string(), "Send"); +diff --git a/third_party/rust/syn/tests/test_grouping.rs b/third_party/rust/syn/tests/test_grouping.rs +index 1558a47b4b..a0fe716390 100644 +--- a/third_party/rust/syn/tests/test_grouping.rs ++++ b/third_party/rust/syn/tests/test_grouping.rs +@@ -1,8 +1,3 @@ +-extern crate proc_macro2; +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -28,31 +23,31 @@ fn test_grouping() { + TokenTree::Literal(Literal::i32_suffixed(4)), + ]); + +- assert_eq!(tokens.to_string(), "1i32 + 2i32 + 3i32 * 4i32"); ++ assert_eq!(tokens.to_string(), "1i32 + 2i32 + 3i32 * 4i32"); + + snapshot!(tokens as Expr, @r###" +- ⋮Expr::Binary { +- ⋮ left: Expr::Lit { +- ⋮ lit: 1i32, +- ⋮ }, +- ⋮ op: Add, +- ⋮ right: Expr::Binary { +- ⋮ left: Expr::Group { +- ⋮ expr: Expr::Binary { +- ⋮ left: Expr::Lit { +- ⋮ lit: 2i32, +- ⋮ }, +- ⋮ op: Add, +- ⋮ right: Expr::Lit { +- ⋮ lit: 3i32, +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ op: Mul, +- ⋮ right: Expr::Lit { +- ⋮ lit: 4i32, +- ⋮ }, +- ⋮ }, +- ⋮} ++ Expr::Binary { ++ left: Expr::Lit { ++ lit: 1i32, ++ }, ++ op: Add, ++ right: Expr::Binary { ++ left: Expr::Group { ++ expr: Expr::Binary { ++ left: Expr::Lit { ++ lit: 2i32, ++ }, ++ op: Add, ++ right: Expr::Lit { ++ lit: 3i32, ++ }, ++ }, ++ }, ++ op: Mul, ++ right: Expr::Lit { ++ lit: 4i32, ++ }, ++ }, ++ } + "###); + } +diff --git a/third_party/rust/syn/tests/test_ident.rs b/third_party/rust/syn/tests/test_ident.rs +index bec00a70c9..ee01bfcc9f 100644 +--- a/third_party/rust/syn/tests/test_ident.rs ++++ b/third_party/rust/syn/tests/test_ident.rs +@@ -1,8 +1,3 @@ +-extern crate proc_macro2; +-extern crate syn; +- +-mod features; +- + use proc_macro2::{Ident, Span, TokenStream}; + use std::str::FromStr; + use syn::Result; +diff --git a/third_party/rust/syn/tests/test_item.rs b/third_party/rust/syn/tests/test_item.rs +new file mode 100644 +index 0000000000..74ac4baec6 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_item.rs +@@ -0,0 +1,45 @@ ++#[macro_use] ++mod macros; ++ ++use proc_macro2::{Delimiter, Group, Ident, Span, TokenStream, TokenTree}; ++use quote::quote; ++use std::iter::FromIterator; ++use syn::Item; ++ ++#[test] ++fn test_macro_variable_attr() { ++ // mimics the token stream corresponding to `$attr fn f() {}` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { #[test] })), ++ TokenTree::Ident(Ident::new("fn", Span::call_site())), ++ TokenTree::Ident(Ident::new("f", Span::call_site())), ++ TokenTree::Group(Group::new(Delimiter::Parenthesis, TokenStream::new())), ++ TokenTree::Group(Group::new(Delimiter::Brace, TokenStream::new())), ++ ]); ++ ++ snapshot!(tokens as Item, @r###" ++ Item::Fn { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "test", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ vis: Inherited, ++ sig: Signature { ++ ident: "f", ++ generics: Generics, ++ output: Default, ++ }, ++ block: Block, ++ } ++ "###); ++} +diff --git a/third_party/rust/syn/tests/test_iterators.rs b/third_party/rust/syn/tests/test_iterators.rs +index 1cf7157e6f..2c8359c157 100644 +--- a/third_party/rust/syn/tests/test_iterators.rs ++++ b/third_party/rust/syn/tests/test_iterators.rs +@@ -1,10 +1,5 @@ + use syn::punctuated::{Pair, Punctuated}; +- +-extern crate quote; +-#[macro_use] +-extern crate syn; +- +-mod features; ++use syn::Token; + + #[macro_use] + mod macros; +diff --git a/third_party/rust/syn/tests/test_lit.rs b/third_party/rust/syn/tests/test_lit.rs +index 1e8f49d19b..e995f2287f 100644 +--- a/third_party/rust/syn/tests/test_lit.rs ++++ b/third_party/rust/syn/tests/test_lit.rs +@@ -1,13 +1,11 @@ +-extern crate proc_macro2; +-extern crate quote; +-extern crate syn; ++#[macro_use] ++mod macros; + +-mod features; +- +-use proc_macro2::{TokenStream, TokenTree}; ++use proc_macro2::{Delimiter, Group, Literal, Span, TokenStream, TokenTree}; + use quote::ToTokens; ++use std::iter::FromIterator; + use std::str::FromStr; +-use syn::Lit; ++use syn::{Lit, LitFloat, LitInt}; + + fn lit(s: &str) -> Lit { + match TokenStream::from_str(s) +@@ -50,6 +48,9 @@ fn strings() { + "contains\nnewlinesescaped newlines", + ); + test_string("r\"raw\nstring\\\nhere\"", "raw\nstring\\\nhere"); ++ test_string("\"...\"q", "..."); ++ test_string("r\"...\"q", "..."); ++ test_string("r##\"...\"##q", "..."); + } + + #[test] +@@ -79,6 +80,9 @@ fn byte_strings() { + b"contains\nnewlinesescaped newlines", + ); + test_byte_string("br\"raw\nstring\\\nhere\"", b"raw\nstring\\\nhere"); ++ test_byte_string("b\"...\"q", b"..."); ++ test_byte_string("br\"...\"q", b"..."); ++ test_byte_string("br##\"...\"##q", b"..."); + } + + #[test] +@@ -100,6 +104,7 @@ fn bytes() { + test_byte("b'\\t'", b'\t'); + test_byte("b'\\''", b'\''); + test_byte("b'\"'", b'"'); ++ test_byte("b'a'q", b'a'); + } + + #[test] +@@ -125,6 +130,7 @@ fn chars() { + test_char("'\\''", '\''); + test_char("'\"'", '"'); + test_char("'\\u{1F415}'", '\u{1F415}'); ++ test_char("'a'q", 'a'); + } + + #[test] +@@ -185,4 +191,59 @@ fn floats() { + test_float("5.5e12", 5.5e12, ""); + test_float("1.0__3e-12", 1.03e-12, ""); + test_float("1.03e+12", 1.03e12, ""); ++ test_float("9e99e99", 9e99, "e99"); ++} ++ ++#[test] ++fn negative() { ++ let span = Span::call_site(); ++ assert_eq!("-1", LitInt::new("-1", span).to_string()); ++ assert_eq!("-1i8", LitInt::new("-1i8", span).to_string()); ++ assert_eq!("-1i16", LitInt::new("-1i16", span).to_string()); ++ assert_eq!("-1i32", LitInt::new("-1i32", span).to_string()); ++ assert_eq!("-1i64", LitInt::new("-1i64", span).to_string()); ++ assert_eq!("-1.5", LitFloat::new("-1.5", span).to_string()); ++ assert_eq!("-1.5f32", LitFloat::new("-1.5f32", span).to_string()); ++ assert_eq!("-1.5f64", LitFloat::new("-1.5f64", span).to_string()); ++} ++ ++#[test] ++fn suffix() { ++ fn get_suffix(token: &str) -> String { ++ let lit = syn::parse_str::<Lit>(token).unwrap(); ++ match lit { ++ Lit::Str(lit) => lit.suffix().to_owned(), ++ Lit::ByteStr(lit) => lit.suffix().to_owned(), ++ Lit::Byte(lit) => lit.suffix().to_owned(), ++ Lit::Char(lit) => lit.suffix().to_owned(), ++ Lit::Int(lit) => lit.suffix().to_owned(), ++ Lit::Float(lit) => lit.suffix().to_owned(), ++ _ => unimplemented!(), ++ } ++ } ++ ++ assert_eq!(get_suffix("\"\"s"), "s"); ++ assert_eq!(get_suffix("r\"\"r"), "r"); ++ assert_eq!(get_suffix("b\"\"b"), "b"); ++ assert_eq!(get_suffix("br\"\"br"), "br"); ++ assert_eq!(get_suffix("r#\"\"#r"), "r"); ++ assert_eq!(get_suffix("'c'c"), "c"); ++ assert_eq!(get_suffix("b'b'b"), "b"); ++ assert_eq!(get_suffix("1i32"), "i32"); ++ assert_eq!(get_suffix("1_i32"), "i32"); ++ assert_eq!(get_suffix("1.0f32"), "f32"); ++ assert_eq!(get_suffix("1.0_f32"), "f32"); ++} ++ ++#[test] ++fn test_deep_group_empty() { ++ let tokens = TokenStream::from_iter(vec![TokenTree::Group(Group::new( ++ Delimiter::None, ++ TokenStream::from_iter(vec![TokenTree::Group(Group::new( ++ Delimiter::None, ++ TokenStream::from_iter(vec![TokenTree::Literal(Literal::string("hi"))]), ++ ))]), ++ ))]); ++ ++ snapshot!(tokens as Lit, @r#""hi""# ); + } +diff --git a/third_party/rust/syn/tests/test_meta.rs b/third_party/rust/syn/tests/test_meta.rs +index 547472d6f4..d37dda948a 100644 +--- a/third_party/rust/syn/tests/test_meta.rs ++++ b/third_party/rust/syn/tests/test_meta.rs +@@ -1,7 +1,3 @@ +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -12,14 +8,14 @@ fn test_parse_meta_item_word() { + let input = "hello"; + + snapshot!(input as Meta, @r###" +- ⋮Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "hello", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮}) ++ Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "hello", ++ arguments: None, ++ }, ++ ], ++ }) + "###); + } + +@@ -29,31 +25,31 @@ fn test_parse_meta_name_value() { + let (inner, meta) = (input, input); + + snapshot!(inner as MetaNameValue, @r###" +- ⋮MetaNameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ MetaNameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + snapshot!(meta as Meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + assert_eq!(meta, inner.into()); +@@ -65,31 +61,31 @@ fn test_parse_meta_name_value_with_keyword() { + let (inner, meta) = (input, input); + + snapshot!(inner as MetaNameValue, @r###" +- ⋮MetaNameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "static", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ MetaNameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "static", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + snapshot!(meta as Meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "static", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "static", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + assert_eq!(meta, inner.into()); +@@ -101,31 +97,31 @@ fn test_parse_meta_name_value_with_bool() { + let (inner, meta) = (input, input); + + snapshot!(inner as MetaNameValue, @r###" +- ⋮MetaNameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "true", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ MetaNameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "true", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + snapshot!(meta as Meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "true", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "true", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + assert_eq!(meta, inner.into()); +@@ -137,35 +133,35 @@ fn test_parse_meta_item_list_lit() { + let (inner, meta) = (input, input); + + snapshot!(inner as MetaList, @r###" +- ⋮MetaList { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Lit(5), +- ⋮ ], +- ⋮} ++ MetaList { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Lit(5), ++ ], ++ } + "###); + + snapshot!(meta as Meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Lit(5), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Lit(5), ++ ], ++ } + "###); + + assert_eq!(meta, inner.into()); +@@ -177,133 +173,133 @@ fn test_parse_meta_item_multiple() { + let (inner, meta) = (input, input); + + snapshot!(inner as MetaList, @r###" +- ⋮MetaList { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮ }), +- ⋮ Meta(Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "list", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 6, +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ ], +- ⋮} ++ MetaList { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word", ++ arguments: None, ++ }, ++ ], ++ })), ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ }), ++ Meta(Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "list", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name2", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 6, ++ }), ++ ], ++ }), ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word2", ++ arguments: None, ++ }, ++ ], ++ })), ++ ], ++ } + "###); + + snapshot!(meta as Meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮ }), +- ⋮ Meta(Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "list", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 6, +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word", ++ arguments: None, ++ }, ++ ], ++ })), ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ }), ++ Meta(Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "list", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name2", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 6, ++ }), ++ ], ++ }), ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word2", ++ arguments: None, ++ }, ++ ], ++ })), ++ ], ++ } + "###); + + assert_eq!(meta, inner.into()); +@@ -316,28 +312,28 @@ fn test_parse_nested_meta() { + + let input = "list(name2 = 6)"; + snapshot!(input as NestedMeta, @r###" +- ⋮Meta(Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "list", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 6, +- ⋮ }), +- ⋮ ], +- ⋮}) ++ Meta(Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "list", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name2", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 6, ++ }), ++ ], ++ }) + "###); + } +diff --git a/third_party/rust/syn/tests/test_parse_buffer.rs b/third_party/rust/syn/tests/test_parse_buffer.rs +index f09495187f..57a3c7c38c 100644 +--- a/third_party/rust/syn/tests/test_parse_buffer.rs ++++ b/third_party/rust/syn/tests/test_parse_buffer.rs +@@ -1,7 +1,7 @@ +-#[macro_use] +-extern crate syn; +- ++use proc_macro2::{Delimiter, Group, Punct, Spacing, TokenStream, TokenTree}; ++use std::iter::FromIterator; + use syn::parse::{discouraged::Speculative, Parse, ParseStream, Parser, Result}; ++use syn::{parenthesized, Token}; + + #[test] + #[should_panic(expected = "Fork was not derived from the advancing parse stream")] +@@ -53,3 +53,38 @@ fn smuggled_speculative_cursor_into_brackets() { + + syn::parse_str::<BreakRules>("()").unwrap(); + } ++ ++#[test] ++fn trailing_empty_none_group() { ++ fn parse(input: ParseStream) -> Result<()> { ++ input.parse::<Token![+]>()?; ++ ++ let content; ++ parenthesized!(content in input); ++ content.parse::<Token![+]>()?; ++ ++ Ok(()) ++ } ++ ++ // `+ ( + <Ø Ø> ) <Ø <Ø Ø> Ø>` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Punct(Punct::new('+', Spacing::Alone)), ++ TokenTree::Group(Group::new( ++ Delimiter::Parenthesis, ++ TokenStream::from_iter(vec![ ++ TokenTree::Punct(Punct::new('+', Spacing::Alone)), ++ TokenTree::Group(Group::new(Delimiter::None, TokenStream::new())), ++ ]), ++ )), ++ TokenTree::Group(Group::new(Delimiter::None, TokenStream::new())), ++ TokenTree::Group(Group::new( ++ Delimiter::None, ++ TokenStream::from_iter(vec![TokenTree::Group(Group::new( ++ Delimiter::None, ++ TokenStream::new(), ++ ))]), ++ )), ++ ]); ++ ++ parse.parse2(tokens).unwrap(); ++} +diff --git a/third_party/rust/syn/tests/test_parse_stream.rs b/third_party/rust/syn/tests/test_parse_stream.rs +new file mode 100644 +index 0000000000..76bd065777 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_parse_stream.rs +@@ -0,0 +1,12 @@ ++use syn::ext::IdentExt; ++use syn::parse::ParseStream; ++use syn::{Ident, Token}; ++ ++#[test] ++fn test_peek() { ++ let _ = |input: ParseStream| { ++ let _ = input.peek(Ident); ++ let _ = input.peek(Ident::peek_any); ++ let _ = input.peek(Token![::]); ++ }; ++} +diff --git a/third_party/rust/syn/tests/test_pat.rs b/third_party/rust/syn/tests/test_pat.rs +index 1343aa646f..73388dd79d 100644 +--- a/third_party/rust/syn/tests/test_pat.rs ++++ b/third_party/rust/syn/tests/test_pat.rs +@@ -1,10 +1,5 @@ +-extern crate quote; +-extern crate syn; +- +-mod features; +- + use quote::quote; +-use syn::Pat; ++use syn::{Item, Pat, Stmt}; + + #[test] + fn test_pat_ident() { +@@ -21,3 +16,23 @@ fn test_pat_path() { + value => panic!("expected PatPath, got {:?}", value), + } + } ++ ++#[test] ++fn test_leading_vert() { ++ // https://github.com/rust-lang/rust/blob/1.43.0/src/test/ui/or-patterns/remove-leading-vert.rs ++ ++ syn::parse_str::<Item>("fn f() {}").unwrap(); ++ syn::parse_str::<Item>("fn fun1(| A: E) {}").unwrap_err(); ++ syn::parse_str::<Item>("fn fun2(|| A: E) {}").unwrap_err(); ++ ++ syn::parse_str::<Stmt>("let | () = ();").unwrap(); ++ syn::parse_str::<Stmt>("let (| A): E;").unwrap_err(); ++ syn::parse_str::<Stmt>("let (|| A): (E);").unwrap_err(); ++ syn::parse_str::<Stmt>("let (| A,): (E,);").unwrap_err(); ++ syn::parse_str::<Stmt>("let [| A]: [E; 1];").unwrap_err(); ++ syn::parse_str::<Stmt>("let [|| A]: [E; 1];").unwrap_err(); ++ syn::parse_str::<Stmt>("let TS(| A): TS;").unwrap_err(); ++ syn::parse_str::<Stmt>("let TS(|| A): TS;").unwrap_err(); ++ syn::parse_str::<Stmt>("let NS { f: | A }: NS;").unwrap_err(); ++ syn::parse_str::<Stmt>("let NS { f: || A }: NS;").unwrap_err(); ++} +diff --git a/third_party/rust/syn/tests/test_path.rs b/third_party/rust/syn/tests/test_path.rs +new file mode 100644 +index 0000000000..2ce12066f5 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_path.rs +@@ -0,0 +1,52 @@ ++#[macro_use] ++mod macros; ++ ++use proc_macro2::{Delimiter, Group, Ident, Punct, Spacing, Span, TokenStream, TokenTree}; ++use quote::quote; ++use std::iter::FromIterator; ++use syn::{Expr, Type}; ++ ++#[test] ++fn parse_interpolated_leading_component() { ++ // mimics the token stream corresponding to `$mod::rest` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { first })), ++ TokenTree::Punct(Punct::new(':', Spacing::Joint)), ++ TokenTree::Punct(Punct::new(':', Spacing::Alone)), ++ TokenTree::Ident(Ident::new("rest", Span::call_site())), ++ ]); ++ ++ snapshot!(tokens.clone() as Expr, @r###" ++ Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "first", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "rest", ++ arguments: None, ++ }, ++ ], ++ }, ++ } ++ "###); ++ ++ snapshot!(tokens as Type, @r###" ++ Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "first", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "rest", ++ arguments: None, ++ }, ++ ], ++ }, ++ } ++ "###); ++} +diff --git a/third_party/rust/syn/tests/test_precedence.rs b/third_party/rust/syn/tests/test_precedence.rs +index 53ee66e372..a586b3fe48 100644 +--- a/third_party/rust/syn/tests/test_precedence.rs ++++ b/third_party/rust/syn/tests/test_precedence.rs +@@ -4,35 +4,26 @@ + + //! The tests in this module do the following: + //! +-//! 1. Parse a given expression in both `syn` and `libsyntax`. ++//! 1. Parse a given expression in both `syn` and `librustc`. + //! 2. Fold over the expression adding brackets around each subexpression (with +-//! some complications - see the `syn_brackets` and `libsyntax_brackets` ++//! some complications - see the `syn_brackets` and `librustc_brackets` + //! methods). + //! 3. Serialize the `syn` expression back into a string, and re-parse it with +-//! `libsyntax`. ++//! `librustc`. + //! 4. Respan all of the expressions, replacing the spans with the default + //! spans. + //! 5. Compare the expressions with one another, if they are not equal fail. + +-extern crate quote; +-extern crate rayon; +-extern crate regex; ++extern crate rustc_ast; + extern crate rustc_data_structures; +-extern crate smallvec; +-extern crate syn; +-extern crate syntax; +-extern crate syntax_pos; +-extern crate walkdir; +- +-mod features; ++extern crate rustc_span; + + use quote::quote; + use rayon::iter::{IntoParallelIterator, ParallelIterator}; + use regex::Regex; +-use smallvec::smallvec; +-use syntax::ast; +-use syntax::ptr::P; +-use syntax_pos::edition::Edition; ++use rustc_ast::ast; ++use rustc_ast::ptr::P; ++use rustc_span::edition::Edition; + use walkdir::{DirEntry, WalkDir}; + + use std::fs::File; +@@ -73,7 +64,7 @@ fn test_simple_precedence() { + continue; + }; + +- let pf = match test_expressions(vec![expr]) { ++ let pf = match test_expressions(Edition::Edition2018, vec![expr]) { + (1, 0) => "passed", + (0, 1) => { + failed += 1; +@@ -91,8 +82,8 @@ fn test_simple_precedence() { + + /// Test expressions from rustc, like in `test_round_trip`. + #[test] +-#[cfg_attr(target_os = "windows", ignore = "requires nix .sh")] + fn test_rustc_precedence() { ++ common::rayon_init(); + repo::clone_rust(); + let abort_after = common::abort_after(); + if abort_after == 0 { +@@ -118,15 +109,6 @@ fn test_rustc_precedence() { + return; + } + +- // Our version of `libsyntax` can't parse this tests +- if path +- .to_str() +- .unwrap() +- .ends_with("optional_comma_in_match_arm.rs") +- { +- return; +- } +- + let mut file = File::open(path).unwrap(); + let mut content = String::new(); + file.read_to_string(&mut content).unwrap(); +@@ -134,8 +116,9 @@ fn test_rustc_precedence() { + + let (l_passed, l_failed) = match syn::parse_file(&content) { + Ok(file) => { ++ let edition = repo::edition(path).parse().unwrap(); + let exprs = collect_exprs(file); +- test_expressions(exprs) ++ test_expressions(edition, exprs) + } + Err(msg) => { + errorf!("syn failed to parse\n{:?}\n", msg); +@@ -169,36 +152,36 @@ fn test_rustc_precedence() { + } + } + +-fn test_expressions(exprs: Vec<syn::Expr>) -> (usize, usize) { ++fn test_expressions(edition: Edition, exprs: Vec<syn::Expr>) -> (usize, usize) { + let mut passed = 0; + let mut failed = 0; + +- syntax::with_globals(Edition::Edition2018, || { ++ rustc_span::with_session_globals(edition, || { + for expr in exprs { + let raw = quote!(#expr).to_string(); + +- let libsyntax_ast = if let Some(e) = libsyntax_parse_and_rewrite(&raw) { ++ let librustc_ast = if let Some(e) = librustc_parse_and_rewrite(&raw) { + e + } else { + failed += 1; +- errorf!("\nFAIL - libsyntax failed to parse raw\n"); ++ errorf!("\nFAIL - librustc failed to parse raw\n"); + continue; + }; + + let syn_expr = syn_brackets(expr); +- let syn_ast = if let Some(e) = parse::libsyntax_expr("e!(#syn_expr).to_string()) { ++ let syn_ast = if let Some(e) = parse::librustc_expr("e!(#syn_expr).to_string()) { + e + } else { + failed += 1; +- errorf!("\nFAIL - libsyntax failed to parse bracketed\n"); ++ errorf!("\nFAIL - librustc failed to parse bracketed\n"); + continue; + }; + +- if SpanlessEq::eq(&syn_ast, &libsyntax_ast) { ++ if SpanlessEq::eq(&syn_ast, &librustc_ast) { + passed += 1; + } else { + failed += 1; +- errorf!("\nFAIL\n{:?}\n!=\n{:?}\n", syn_ast, libsyntax_ast); ++ errorf!("\nFAIL\n{:?}\n!=\n{:?}\n", syn_ast, librustc_ast); + } + } + }); +@@ -206,54 +189,106 @@ fn test_expressions(exprs: Vec<syn::Expr>) -> (usize, usize) { + (passed, failed) + } + +-fn libsyntax_parse_and_rewrite(input: &str) -> Option<P<ast::Expr>> { +- parse::libsyntax_expr(input).and_then(libsyntax_brackets) ++fn librustc_parse_and_rewrite(input: &str) -> Option<P<ast::Expr>> { ++ parse::librustc_expr(input).and_then(librustc_brackets) + } + + /// Wrap every expression which is not already wrapped in parens with parens, to + /// reveal the precidence of the parsed expressions, and produce a stringified + /// form of the resulting expression. + /// +-/// This method operates on libsyntax objects. +-fn libsyntax_brackets(mut libsyntax_expr: P<ast::Expr>) -> Option<P<ast::Expr>> { ++/// This method operates on librustc objects. ++fn librustc_brackets(mut librustc_expr: P<ast::Expr>) -> Option<P<ast::Expr>> { ++ use rustc_ast::ast::{ ++ Block, BorrowKind, Expr, ExprKind, Field, GenericArg, MacCall, Pat, Stmt, StmtKind, Ty, ++ }; ++ use rustc_ast::mut_visit::{noop_visit_generic_arg, MutVisitor}; ++ use rustc_data_structures::map_in_place::MapInPlace; + use rustc_data_structures::thin_vec::ThinVec; +- use smallvec::SmallVec; ++ use rustc_span::DUMMY_SP; + use std::mem; +- use syntax::ast::{Expr, ExprKind, Field, Mac, Pat, Stmt, StmtKind, Ty}; +- use syntax::mut_visit::{noop_visit_expr, MutVisitor}; +- use syntax_pos::DUMMY_SP; + + struct BracketsVisitor { + failed: bool, + }; + ++ fn flat_map_field<T: MutVisitor>(mut f: Field, vis: &mut T) -> Vec<Field> { ++ if f.is_shorthand { ++ noop_visit_expr(&mut f.expr, vis); ++ } else { ++ vis.visit_expr(&mut f.expr); ++ } ++ vec![f] ++ } ++ ++ fn flat_map_stmt<T: MutVisitor>(stmt: Stmt, vis: &mut T) -> Vec<Stmt> { ++ let kind = match stmt.kind { ++ // Don't wrap toplevel expressions in statements. ++ StmtKind::Expr(mut e) => { ++ noop_visit_expr(&mut e, vis); ++ StmtKind::Expr(e) ++ } ++ StmtKind::Semi(mut e) => { ++ noop_visit_expr(&mut e, vis); ++ StmtKind::Semi(e) ++ } ++ s => s, ++ }; ++ ++ vec![Stmt { kind, ..stmt }] ++ } ++ ++ fn noop_visit_expr<T: MutVisitor>(e: &mut Expr, vis: &mut T) { ++ use rustc_ast::mut_visit::{noop_visit_expr, visit_opt, visit_thin_attrs}; ++ match &mut e.kind { ++ ExprKind::AddrOf(BorrowKind::Raw, ..) => {} ++ ExprKind::Struct(path, fields, expr) => { ++ vis.visit_path(path); ++ fields.flat_map_in_place(|field| flat_map_field(field, vis)); ++ visit_opt(expr, |expr| vis.visit_expr(expr)); ++ vis.visit_id(&mut e.id); ++ vis.visit_span(&mut e.span); ++ visit_thin_attrs(&mut e.attrs, vis); ++ } ++ _ => noop_visit_expr(e, vis), ++ } ++ } ++ + impl MutVisitor for BracketsVisitor { + fn visit_expr(&mut self, e: &mut P<Expr>) { + noop_visit_expr(e, self); +- match e.node { ++ match e.kind { + ExprKind::If(..) | ExprKind::Block(..) | ExprKind::Let(..) => {} + _ => { + let inner = mem::replace( + e, + P(Expr { + id: ast::DUMMY_NODE_ID, +- node: ExprKind::Err, ++ kind: ExprKind::Err, + span: DUMMY_SP, + attrs: ThinVec::new(), ++ tokens: None, + }), + ); +- e.node = ExprKind::Paren(inner); ++ e.kind = ExprKind::Paren(inner); + } + } + } + +- fn flat_map_field(&mut self, mut f: Field) -> SmallVec<[Field; 1]> { +- if f.is_shorthand { +- noop_visit_expr(&mut f.expr, self); +- } else { +- self.visit_expr(&mut f.expr); ++ fn visit_generic_arg(&mut self, arg: &mut GenericArg) { ++ match arg { ++ // Don't wrap const generic arg as that's invalid syntax. ++ GenericArg::Const(arg) => noop_visit_expr(&mut arg.value, self), ++ _ => noop_visit_generic_arg(arg, self), + } +- SmallVec::from([f]) ++ } ++ ++ fn visit_block(&mut self, block: &mut P<Block>) { ++ self.visit_id(&mut block.id); ++ block ++ .stmts ++ .flat_map_in_place(|stmt| flat_map_stmt(stmt, self)); ++ self.visit_span(&mut block.span); + } + + // We don't want to look at expressions that might appear in patterns or +@@ -267,25 +302,8 @@ fn libsyntax_brackets(mut libsyntax_expr: P<ast::Expr>) -> Option<P<ast::Expr>> + let _ = ty; + } + +- fn flat_map_stmt(&mut self, stmt: Stmt) -> SmallVec<[Stmt; 1]> { +- let node = match stmt.node { +- // Don't wrap toplevel expressions in statements. +- StmtKind::Expr(mut e) => { +- noop_visit_expr(&mut e, self); +- StmtKind::Expr(e) +- } +- StmtKind::Semi(mut e) => { +- noop_visit_expr(&mut e, self); +- StmtKind::Semi(e) +- } +- s => s, +- }; +- +- smallvec![Stmt { node, ..stmt }] +- } +- +- fn visit_mac(&mut self, mac: &mut Mac) { +- // By default when folding over macros, libsyntax panics. This is ++ fn visit_mac(&mut self, mac: &mut MacCall) { ++ // By default when folding over macros, librustc panics. This is + // because it's usually not what you want, you want to run after + // macro expansion. We do want to do that (syn doesn't do macro + // expansion), so we implement visit_mac to just return the macro +@@ -295,11 +313,11 @@ fn libsyntax_brackets(mut libsyntax_expr: P<ast::Expr>) -> Option<P<ast::Expr>> + } + + let mut folder = BracketsVisitor { failed: false }; +- folder.visit_expr(&mut libsyntax_expr); ++ folder.visit_expr(&mut librustc_expr); + if folder.failed { + None + } else { +- Some(libsyntax_expr) ++ Some(librustc_expr) + } + } + +@@ -318,14 +336,33 @@ fn syn_brackets(syn_expr: syn::Expr) -> syn::Expr { + Expr::If(..) | Expr::Unsafe(..) | Expr::Block(..) | Expr::Let(..) => { + fold_expr(self, expr) + } +- node => Expr::Paren(ExprParen { ++ _ => Expr::Paren(ExprParen { + attrs: Vec::new(), +- expr: Box::new(fold_expr(self, node)), ++ expr: Box::new(fold_expr(self, expr)), + paren_token: token::Paren::default(), + }), + } + } + ++ fn fold_generic_argument(&mut self, arg: GenericArgument) -> GenericArgument { ++ match arg { ++ // Don't wrap const generic arg as that's invalid syntax. ++ GenericArgument::Const(a) => GenericArgument::Const(fold_expr(self, a)), ++ _ => fold_generic_argument(self, arg), ++ } ++ } ++ ++ fn fold_generic_method_argument( ++ &mut self, ++ arg: GenericMethodArgument, ++ ) -> GenericMethodArgument { ++ match arg { ++ // Don't wrap const generic arg as that's invalid syntax. ++ GenericMethodArgument::Const(a) => GenericMethodArgument::Const(fold_expr(self, a)), ++ _ => fold_generic_method_argument(self, arg), ++ } ++ } ++ + fn fold_stmt(&mut self, stmt: Stmt) -> Stmt { + match stmt { + // Don't wrap toplevel expressions in statements. +@@ -360,7 +397,10 @@ fn collect_exprs(file: syn::File) -> Vec<syn::Expr> { + struct CollectExprs(Vec<Expr>); + impl Fold for CollectExprs { + fn fold_expr(&mut self, expr: Expr) -> Expr { +- self.0.push(expr); ++ match expr { ++ Expr::Verbatim(tokens) if tokens.is_empty() => {} ++ _ => self.0.push(expr), ++ } + + Expr::Tuple(ExprTuple { + attrs: vec![], +diff --git a/third_party/rust/syn/tests/test_receiver.rs b/third_party/rust/syn/tests/test_receiver.rs +new file mode 100644 +index 0000000000..923df96ba9 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_receiver.rs +@@ -0,0 +1,127 @@ ++use syn::{parse_quote, FnArg, Receiver, TraitItemMethod}; ++ ++#[test] ++fn test_by_value() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn by_value(self: Self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_by_mut_value() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn by_mut(mut self: Self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_by_ref() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn by_ref(self: &Self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_by_box() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn by_box(self: Box<Self>); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_by_pin() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn by_pin(self: Pin<Self>); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_explicit_type() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn explicit_type(self: Pin<MyType>); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_value_shorthand() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn value_shorthand(self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Receiver(Receiver { ++ reference: None, ++ mutability: None, ++ .. ++ })) => (), ++ value => panic!("expected FnArg::Receiver without ref/mut, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_mut_value_shorthand() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn mut_value_shorthand(mut self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Receiver(Receiver { ++ reference: None, ++ mutability: Some(_), ++ .. ++ })) => (), ++ value => panic!("expected FnArg::Receiver with mut, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_ref_shorthand() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn ref_shorthand(&self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Receiver(Receiver { ++ reference: Some(_), ++ mutability: None, ++ .. ++ })) => (), ++ value => panic!("expected FnArg::Receiver with ref, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_ref_mut_shorthand() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn ref_mut_shorthand(&mut self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Receiver(Receiver { ++ reference: Some(_), ++ mutability: Some(_), ++ .. ++ })) => (), ++ value => panic!("expected FnArg::Receiver with ref+mut, got {:?}", value), ++ } ++} +diff --git a/third_party/rust/syn/tests/test_round_trip.rs b/third_party/rust/syn/tests/test_round_trip.rs +index 2fc9cecd86..260dd0c3d9 100644 +--- a/third_party/rust/syn/tests/test_round_trip.rs ++++ b/third_party/rust/syn/tests/test_round_trip.rs +@@ -2,22 +2,20 @@ + #![recursion_limit = "1024"] + #![feature(rustc_private)] + +-extern crate quote; +-extern crate rayon; +-extern crate syn; +-extern crate syntax; +-extern crate syntax_pos; +-extern crate walkdir; +- +-mod features; ++extern crate rustc_ast; ++extern crate rustc_errors; ++extern crate rustc_expand; ++extern crate rustc_parse as parse; ++extern crate rustc_session; ++extern crate rustc_span; + + use quote::quote; + use rayon::iter::{IntoParallelIterator, ParallelIterator}; +-use syntax::ast; +-use syntax::parse::{self, PResult, ParseSess}; +-use syntax::source_map::FilePathMapping; +-use syntax_pos::edition::Edition; +-use syntax_pos::FileName; ++use rustc_ast::ast; ++use rustc_errors::PResult; ++use rustc_session::parse::ParseSess; ++use rustc_span::source_map::FilePathMapping; ++use rustc_span::FileName; + use walkdir::{DirEntry, WalkDir}; + + use std::fs::File; +@@ -38,8 +36,8 @@ mod repo; + use common::eq::SpanlessEq; + + #[test] +-#[cfg_attr(target_os = "windows", ignore = "requires nix .sh")] + fn test_round_trip() { ++ common::rayon_init(); + repo::clone_rust(); + let abort_after = common::abort_after(); + if abort_after == 0 { +@@ -78,11 +76,12 @@ fn test_round_trip() { + } + }; + let back = quote!(#krate).to_string(); ++ let edition = repo::edition(path).parse().unwrap(); + + let equal = panic::catch_unwind(|| { +- syntax::with_globals(Edition::Edition2018, || { ++ rustc_span::with_session_globals(edition, || { + let sess = ParseSess::new(FilePathMapping::empty()); +- let before = match libsyntax_parse(content, &sess) { ++ let before = match librustc_parse(content, &sess) { + Ok(before) => before, + Err(mut diagnostic) => { + diagnostic.cancel(); +@@ -93,7 +92,7 @@ fn test_round_trip() { + errorf!("=== {}: ignore\n", path.display()); + } else { + errorf!( +- "=== {}: ignore - libsyntax failed to parse original content: {}\n", ++ "=== {}: ignore - librustc failed to parse original content: {}\n", + path.display(), + diagnostic.message() + ); +@@ -101,10 +100,10 @@ fn test_round_trip() { + return true; + } + }; +- let after = match libsyntax_parse(back, &sess) { ++ let after = match librustc_parse(back, &sess) { + Ok(after) => after, + Err(mut diagnostic) => { +- errorf!("=== {}: libsyntax failed to parse", path.display()); ++ errorf!("=== {}: librustc failed to parse", path.display()); + diagnostic.emit(); + return false; + } +@@ -130,7 +129,7 @@ fn test_round_trip() { + }) + }); + match equal { +- Err(_) => errorf!("=== {}: ignoring libsyntax panic\n", path.display()), ++ Err(_) => errorf!("=== {}: ignoring librustc panic\n", path.display()), + Ok(true) => {} + Ok(false) => { + let prev_failed = failed.fetch_add(1, Ordering::SeqCst); +@@ -147,7 +146,7 @@ fn test_round_trip() { + } + } + +-fn libsyntax_parse(content: String, sess: &ParseSess) -> PResult<ast::Crate> { ++fn librustc_parse(content: String, sess: &ParseSess) -> PResult<ast::Crate> { + let name = FileName::Custom("test_round_trip".to_string()); + parse::parse_crate_from_source_str(name, content, sess) + } +diff --git a/third_party/rust/syn/tests/test_shebang.rs b/third_party/rust/syn/tests/test_shebang.rs +new file mode 100644 +index 0000000000..dc26b9aab3 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_shebang.rs +@@ -0,0 +1,59 @@ ++#[macro_use] ++mod macros; ++ ++#[test] ++fn test_basic() { ++ let content = "#!/usr/bin/env rustx\nfn main() {}"; ++ let file = syn::parse_file(content).unwrap(); ++ snapshot!(file, @r###" ++ File { ++ shebang: Some("#!/usr/bin/env rustx"), ++ items: [ ++ Item::Fn { ++ vis: Inherited, ++ sig: Signature { ++ ident: "main", ++ generics: Generics, ++ output: Default, ++ }, ++ block: Block, ++ }, ++ ], ++ } ++ "###); ++} ++ ++#[test] ++fn test_comment() { ++ let content = "#!//am/i/a/comment\n[allow(dead_code)] fn main() {}"; ++ let file = syn::parse_file(content).unwrap(); ++ snapshot!(file, @r###" ++ File { ++ attrs: [ ++ Attribute { ++ style: Inner, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "allow", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(`(dead_code)`), ++ }, ++ ], ++ items: [ ++ Item::Fn { ++ vis: Inherited, ++ sig: Signature { ++ ident: "main", ++ generics: Generics, ++ output: Default, ++ }, ++ block: Block, ++ }, ++ ], ++ } ++ "###); ++} +diff --git a/third_party/rust/syn/tests/test_should_parse.rs b/third_party/rust/syn/tests/test_should_parse.rs +index aadf42e3af..180d859916 100644 +--- a/third_party/rust/syn/tests/test_should_parse.rs ++++ b/third_party/rust/syn/tests/test_should_parse.rs +@@ -1,7 +1,3 @@ +-extern crate syn; +- +-mod features; +- + macro_rules! should_parse { + ($name:ident, { $($in:tt)* }) => { + #[test] +diff --git a/third_party/rust/syn/tests/test_size.rs b/third_party/rust/syn/tests/test_size.rs +index 386d4df889..01e8401158 100644 +--- a/third_party/rust/syn/tests/test_size.rs ++++ b/third_party/rust/syn/tests/test_size.rs +@@ -1,7 +1,5 @@ + #![cfg(target_pointer_width = "64")] + +-mod features; +- + use std::mem; + use syn::*; + +diff --git a/third_party/rust/syn/tests/test_stmt.rs b/third_party/rust/syn/tests/test_stmt.rs +new file mode 100644 +index 0000000000..d68b47fd2f +--- /dev/null ++++ b/third_party/rust/syn/tests/test_stmt.rs +@@ -0,0 +1,44 @@ ++#[macro_use] ++mod macros; ++ ++use syn::Stmt; ++ ++#[test] ++fn test_raw_operator() { ++ let stmt = syn::parse_str::<Stmt>("let _ = &raw const x;").unwrap(); ++ ++ snapshot!(stmt, @r###" ++ Local(Local { ++ pat: Pat::Wild, ++ init: Some(Verbatim(`& raw const x`)), ++ }) ++ "###); ++} ++ ++#[test] ++fn test_raw_variable() { ++ let stmt = syn::parse_str::<Stmt>("let _ = &raw;").unwrap(); ++ ++ snapshot!(stmt, @r###" ++ Local(Local { ++ pat: Pat::Wild, ++ init: Some(Expr::Reference { ++ expr: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "raw", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }), ++ }) ++ "###); ++} ++ ++#[test] ++fn test_raw_invalid() { ++ assert!(syn::parse_str::<Stmt>("let _ = &raw x;").is_err()); ++} +diff --git a/third_party/rust/syn/tests/test_token_trees.rs b/third_party/rust/syn/tests/test_token_trees.rs +index 70a9a72aab..5b00448af8 100644 +--- a/third_party/rust/syn/tests/test_token_trees.rs ++++ b/third_party/rust/syn/tests/test_token_trees.rs +@@ -1,9 +1,3 @@ +-extern crate proc_macro2; +-extern crate quote; +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -21,7 +15,11 @@ fn test_struct() { + } + "; + +- snapshot!(input as TokenStream, @"`# [ derive ( Debug , Clone ) ] pub struct Item { pub ident : Ident , pub attrs : Vec < Attribute >, }`"); ++ snapshot!(input as TokenStream, @r###" ++ TokenStream( ++ `# [derive (Debug , Clone)] pub struct Item { pub ident : Ident , pub attrs : Vec < Attribute >, }`, ++ ) ++ "###); + } + + #[test] +diff --git a/third_party/rust/syn/tests/test_ty.rs b/third_party/rust/syn/tests/test_ty.rs +new file mode 100644 +index 0000000000..9cbdcd6b99 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_ty.rs +@@ -0,0 +1,53 @@ ++#[macro_use] ++mod macros; ++ ++use proc_macro2::{Delimiter, Group, Ident, Punct, Spacing, Span, TokenStream, TokenTree}; ++use quote::quote; ++use std::iter::FromIterator; ++use syn::Type; ++ ++#[test] ++fn test_mut_self() { ++ syn::parse_str::<Type>("fn(mut self)").unwrap(); ++ syn::parse_str::<Type>("fn(mut self: ())").unwrap(); ++ syn::parse_str::<Type>("fn(mut self: ...)").unwrap_err(); ++ syn::parse_str::<Type>("fn(mut self: mut self)").unwrap_err(); ++ syn::parse_str::<Type>("fn(mut self::T)").unwrap_err(); ++} ++ ++#[test] ++fn test_macro_variable_type() { ++ // mimics the token stream corresponding to `$ty<T>` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { ty })), ++ TokenTree::Punct(Punct::new('<', Spacing::Alone)), ++ TokenTree::Ident(Ident::new("T", Span::call_site())), ++ TokenTree::Punct(Punct::new('>', Spacing::Alone)), ++ ]); ++ ++ snapshot!(tokens as Type, @r###" ++ Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "ty", ++ arguments: PathArguments::AngleBracketed { ++ args: [ ++ Type(Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "T", ++ arguments: None, ++ }, ++ ], ++ }, ++ }), ++ ], ++ }, ++ }, ++ ], ++ }, ++ } ++ "###); ++} +diff --git a/third_party/rust/syn/tests/test_visibility.rs b/third_party/rust/syn/tests/test_visibility.rs +new file mode 100644 +index 0000000000..c3d0ac7a5b +--- /dev/null ++++ b/third_party/rust/syn/tests/test_visibility.rs +@@ -0,0 +1,145 @@ ++#[macro_use] ++mod macros; ++ ++use proc_macro2::{Delimiter, Group, Ident, Punct, Spacing, Span, TokenStream, TokenTree}; ++use std::iter::FromIterator; ++use syn::parse::{Parse, ParseStream}; ++use syn::{DeriveInput, Result, Visibility}; ++ ++#[derive(Debug)] ++struct VisRest { ++ vis: Visibility, ++ rest: TokenStream, ++} ++ ++impl Parse for VisRest { ++ fn parse(input: ParseStream) -> Result<Self> { ++ Ok(VisRest { ++ vis: input.parse()?, ++ rest: input.parse()?, ++ }) ++ } ++} ++ ++macro_rules! assert_vis_parse { ++ ($input:expr, Ok($p:pat)) => { ++ assert_vis_parse!($input, Ok($p) + ""); ++ }; ++ ++ ($input:expr, Ok($p:pat) + $rest:expr) => { ++ let expected = $rest.parse::<TokenStream>().unwrap(); ++ let parse: VisRest = syn::parse_str($input).unwrap(); ++ ++ match parse.vis { ++ $p => {} ++ _ => panic!("Expected {}, got {:?}", stringify!($p), parse.vis), ++ } ++ ++ // NOTE: Round-trips through `to_string` to avoid potential whitespace ++ // diffs. ++ assert_eq!(parse.rest.to_string(), expected.to_string()); ++ }; ++ ++ ($input:expr, Err) => { ++ syn::parse2::<VisRest>($input.parse().unwrap()).unwrap_err(); ++ }; ++} ++ ++#[test] ++fn test_pub() { ++ assert_vis_parse!("pub", Ok(Visibility::Public(_))); ++} ++ ++#[test] ++fn test_crate() { ++ assert_vis_parse!("crate", Ok(Visibility::Crate(_))); ++} ++ ++#[test] ++fn test_inherited() { ++ assert_vis_parse!("", Ok(Visibility::Inherited)); ++} ++ ++#[test] ++fn test_in() { ++ assert_vis_parse!("pub(in foo::bar)", Ok(Visibility::Restricted(_))); ++} ++ ++#[test] ++fn test_pub_crate() { ++ assert_vis_parse!("pub(crate)", Ok(Visibility::Restricted(_))); ++} ++ ++#[test] ++fn test_pub_self() { ++ assert_vis_parse!("pub(self)", Ok(Visibility::Restricted(_))); ++} ++ ++#[test] ++fn test_pub_super() { ++ assert_vis_parse!("pub(super)", Ok(Visibility::Restricted(_))); ++} ++ ++#[test] ++fn test_missing_in() { ++ assert_vis_parse!("pub(foo::bar)", Ok(Visibility::Public(_)) + "(foo::bar)"); ++} ++ ++#[test] ++fn test_missing_in_path() { ++ assert_vis_parse!("pub(in)", Err); ++} ++ ++#[test] ++fn test_crate_path() { ++ assert_vis_parse!("pub(crate::A, crate::B)", Ok(Visibility::Public(_)) + "(crate::A, crate::B)"); ++} ++ ++#[test] ++fn test_junk_after_in() { ++ assert_vis_parse!("pub(in some::path @@garbage)", Err); ++} ++ ++#[test] ++fn test_empty_group_vis() { ++ // mimics `struct S { $vis $field: () }` where $vis is empty ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Ident(Ident::new("struct", Span::call_site())), ++ TokenTree::Ident(Ident::new("S", Span::call_site())), ++ TokenTree::Group(Group::new( ++ Delimiter::Brace, ++ TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, TokenStream::new())), ++ TokenTree::Group(Group::new( ++ Delimiter::None, ++ TokenStream::from_iter(vec![TokenTree::Ident(Ident::new( ++ "f", ++ Span::call_site(), ++ ))]), ++ )), ++ TokenTree::Punct(Punct::new(':', Spacing::Alone)), ++ TokenTree::Group(Group::new(Delimiter::Parenthesis, TokenStream::new())), ++ ]), ++ )), ++ ]); ++ ++ snapshot!(tokens as DeriveInput, @r###" ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Named { ++ named: [ ++ Field { ++ vis: Inherited, ++ ident: Some("f"), ++ colon_token: Some, ++ ty: Type::Tuple, ++ }, ++ ], ++ }, ++ }, ++ } ++ "###); ++} +diff --git a/third_party/rust/syn/tests/zzz_stable.rs b/third_party/rust/syn/tests/zzz_stable.rs +index a81b3df4d0..a1a670d9ed 100644 +--- a/third_party/rust/syn/tests/zzz_stable.rs ++++ b/third_party/rust/syn/tests/zzz_stable.rs +@@ -1,7 +1,5 @@ + #![cfg(syn_disable_nightly_tests)] + +-extern crate termcolor; +- + use std::io::{self, Write}; + use termcolor::{Color, ColorChoice, ColorSpec, StandardStream, WriteColor}; + +@@ -10,7 +8,7 @@ const MSG: &str = "\ + ‖ WARNING: + ‖ This is not a nightly compiler so not all tests were able to + ‖ run. Syn includes tests that compare Syn's parser against the +-‖ compiler's parser, which requires access to unstable libsyntax ++‖ compiler's parser, which requires access to unstable librustc + ‖ data structures and a nightly compiler. + ‖ + "; +-- +2.28.0 + diff --git a/source/xap/mozilla-thunderbird/mozilla-thunderbird.SlackBuild b/source/xap/mozilla-thunderbird/mozilla-thunderbird.SlackBuild index 48476c96..4804e26d 100755 --- a/source/xap/mozilla-thunderbird/mozilla-thunderbird.SlackBuild +++ b/source/xap/mozilla-thunderbird/mozilla-thunderbird.SlackBuild @@ -182,6 +182,9 @@ zcat $CWD/unbreakdocs.diff.gz | patch -p1 --verbose || exit 1 # Bypass a test that fails the build: zcat $CWD/gkrust.a.no.networking.check.diff.gz | patch -p1 --verbose || exit 1 +# Fix building with rust-1.47.0: +zcat $CWD/rust_1.47.0.patch.gz | patch -p1 --verbose || exit 1 + # Fetch localization, if requested: if [ ! -z $MOZLOCALIZE ]; then LOC_TAG="THUNDERBIRD_$( echo $VERSION | tr \. _ )_RELEASE" diff --git a/source/xap/mozilla-thunderbird/rust_1.47.0.patch b/source/xap/mozilla-thunderbird/rust_1.47.0.patch new file mode 100644 index 00000000..9471b7bd --- /dev/null +++ b/source/xap/mozilla-thunderbird/rust_1.47.0.patch @@ -0,0 +1,30905 @@ +From 83fc2e3616ef15056be74f056a15e892038809b9 Mon Sep 17 00:00:00 2001 +From: Thomas Deutschmann <whissi@gentoo.org> +Date: Sat, 10 Oct 2020 16:10:20 +0200 +Subject: [PATCH 38/38] bmo#1663715: Update syn and proc-macro2 so that Firefox + can build on Rust nightly again + +Link: https://bugzilla.mozilla.org/show_bug.cgi?id=1663715#c7 +Signed-off-by: Thomas Deutschmann <whissi@gentoo.org> +--- + Cargo.lock | 8 +- + .../rust/lucet-wasi/.cargo-checksum.json | 2 +- + .../rust/packed_simd/.cargo-checksum.json | 2 +- + .../rust/proc-macro2/.cargo-checksum.json | 2 +- + third_party/rust/proc-macro2/Cargo.toml | 15 +- + third_party/rust/proc-macro2/README.md | 2 +- + third_party/rust/proc-macro2/build.rs | 20 + + third_party/rust/proc-macro2/src/detection.rs | 67 + + third_party/rust/proc-macro2/src/fallback.rs | 1010 ++---- + third_party/rust/proc-macro2/src/lib.rs | 225 +- + third_party/rust/proc-macro2/src/marker.rs | 18 + + third_party/rust/proc-macro2/src/parse.rs | 849 +++++ + third_party/rust/proc-macro2/src/strnom.rs | 391 --- + third_party/rust/proc-macro2/src/wrapper.rs | 258 +- + .../rust/proc-macro2/tests/comments.rs | 103 + + third_party/rust/proc-macro2/tests/marker.rs | 33 + + third_party/rust/proc-macro2/tests/test.rs | 240 +- + .../rust/proc-macro2/tests/test_fmt.rs | 26 + + .../spirv-cross-internal/.cargo-checksum.json | 2 +- + third_party/rust/syn/.cargo-checksum.json | 2 +- + third_party/rust/syn/Cargo.toml | 35 +- + third_party/rust/syn/README.md | 16 +- + third_party/rust/syn/benches/file.rs | 7 + + third_party/rust/syn/benches/rust.rs | 45 +- + third_party/rust/syn/build.rs | 38 +- + third_party/rust/syn/src/attr.rs | 126 +- + third_party/rust/syn/src/buffer.rs | 56 +- + third_party/rust/syn/src/custom_keyword.rs | 12 +- + .../rust/syn/src/custom_punctuation.rs | 50 +- + third_party/rust/syn/src/data.rs | 96 +- + third_party/rust/syn/src/derive.rs | 10 +- + third_party/rust/syn/src/discouraged.rs | 27 +- + third_party/rust/syn/src/error.rs | 33 +- + third_party/rust/syn/src/expr.rs | 826 +++-- + third_party/rust/syn/src/ext.rs | 12 +- + third_party/rust/syn/src/file.rs | 4 +- + third_party/rust/syn/src/gen/clone.rs | 2051 ++++++++++++ + third_party/rust/syn/src/gen/debug.rs | 2857 +++++++++++++++++ + third_party/rust/syn/src/gen/eq.rs | 1930 +++++++++++ + third_party/rust/syn/src/gen/fold.rs | 287 +- + third_party/rust/syn/src/gen/hash.rs | 2691 ++++++++++++++++ + third_party/rust/syn/src/gen/visit.rs | 19 +- + third_party/rust/syn/src/gen/visit_mut.rs | 19 +- + third_party/rust/syn/src/generics.rs | 255 +- + third_party/rust/syn/src/item.rs | 1515 +++++---- + third_party/rust/syn/src/keyword.rs | 0 + third_party/rust/syn/src/lib.rs | 109 +- + third_party/rust/syn/src/lifetime.rs | 13 +- + third_party/rust/syn/src/lit.rs | 581 ++-- + third_party/rust/syn/src/mac.rs | 55 +- + third_party/rust/syn/src/macros.rs | 61 +- + third_party/rust/syn/src/op.rs | 6 +- + third_party/rust/syn/src/parse.rs | 211 +- + third_party/rust/syn/src/parse_macro_input.rs | 32 +- + third_party/rust/syn/src/parse_quote.rs | 15 +- + third_party/rust/syn/src/pat.rs | 313 +- + third_party/rust/syn/src/path.rs | 33 +- + third_party/rust/syn/src/punctuated.rs | 123 +- + third_party/rust/syn/src/reserved.rs | 42 + + third_party/rust/syn/src/spanned.rs | 4 +- + third_party/rust/syn/src/stmt.rs | 141 +- + third_party/rust/syn/src/token.rs | 99 +- + third_party/rust/syn/src/tt.rs | 6 +- + third_party/rust/syn/src/ty.rs | 364 ++- + third_party/rust/syn/src/verbatim.rs | 15 + + third_party/rust/syn/src/whitespace.rs | 65 + + third_party/rust/syn/tests/clone.sh | 16 - + third_party/rust/syn/tests/common/eq.rs | 247 +- + third_party/rust/syn/tests/common/mod.rs | 13 + + third_party/rust/syn/tests/common/parse.rs | 24 +- + third_party/rust/syn/tests/debug/gen.rs | 50 +- + third_party/rust/syn/tests/debug/mod.rs | 17 +- + third_party/rust/syn/tests/features/error.rs | 1 - + third_party/rust/syn/tests/features/mod.rs | 22 - + third_party/rust/syn/tests/macros/mod.rs | 8 +- + third_party/rust/syn/tests/repo/mod.rs | 137 +- + third_party/rust/syn/tests/repo/progress.rs | 37 + + third_party/rust/syn/tests/test_asyncness.rs | 38 +- + third_party/rust/syn/tests/test_attribute.rs | 452 +-- + .../rust/syn/tests/test_derive_input.rs | 1321 ++++---- + third_party/rust/syn/tests/test_expr.rs | 314 +- + third_party/rust/syn/tests/test_generics.rs | 371 ++- + third_party/rust/syn/tests/test_grouping.rs | 53 +- + third_party/rust/syn/tests/test_ident.rs | 5 - + third_party/rust/syn/tests/test_item.rs | 45 + + third_party/rust/syn/tests/test_iterators.rs | 7 +- + third_party/rust/syn/tests/test_lit.rs | 75 +- + third_party/rust/syn/tests/test_meta.rs | 498 ++- + .../rust/syn/tests/test_parse_buffer.rs | 41 +- + .../rust/syn/tests/test_parse_stream.rs | 12 + + third_party/rust/syn/tests/test_pat.rs | 27 +- + third_party/rust/syn/tests/test_path.rs | 52 + + third_party/rust/syn/tests/test_precedence.rs | 196 +- + third_party/rust/syn/tests/test_receiver.rs | 127 + + third_party/rust/syn/tests/test_round_trip.rs | 41 +- + third_party/rust/syn/tests/test_shebang.rs | 59 + + .../rust/syn/tests/test_should_parse.rs | 4 - + third_party/rust/syn/tests/test_size.rs | 2 - + third_party/rust/syn/tests/test_stmt.rs | 44 + + .../rust/syn/tests/test_token_trees.rs | 12 +- + third_party/rust/syn/tests/test_ty.rs | 53 + + third_party/rust/syn/tests/test_visibility.rs | 145 + + third_party/rust/syn/tests/zzz_stable.rs | 4 +- + 103 files changed, 17319 insertions(+), 5831 deletions(-) + create mode 100644 third_party/rust/proc-macro2/src/detection.rs + create mode 100644 third_party/rust/proc-macro2/src/marker.rs + create mode 100644 third_party/rust/proc-macro2/src/parse.rs + delete mode 100644 third_party/rust/proc-macro2/src/strnom.rs + create mode 100644 third_party/rust/proc-macro2/tests/comments.rs + create mode 100644 third_party/rust/proc-macro2/tests/test_fmt.rs + create mode 100644 third_party/rust/syn/src/gen/clone.rs + create mode 100644 third_party/rust/syn/src/gen/debug.rs + create mode 100644 third_party/rust/syn/src/gen/eq.rs + create mode 100644 third_party/rust/syn/src/gen/hash.rs + delete mode 100644 third_party/rust/syn/src/keyword.rs + create mode 100644 third_party/rust/syn/src/reserved.rs + create mode 100644 third_party/rust/syn/src/verbatim.rs + create mode 100644 third_party/rust/syn/src/whitespace.rs + delete mode 100755 third_party/rust/syn/tests/clone.sh + delete mode 100644 third_party/rust/syn/tests/features/error.rs + delete mode 100644 third_party/rust/syn/tests/features/mod.rs + create mode 100644 third_party/rust/syn/tests/repo/progress.rs + create mode 100644 third_party/rust/syn/tests/test_item.rs + create mode 100644 third_party/rust/syn/tests/test_parse_stream.rs + create mode 100644 third_party/rust/syn/tests/test_path.rs + create mode 100644 third_party/rust/syn/tests/test_receiver.rs + create mode 100644 third_party/rust/syn/tests/test_shebang.rs + create mode 100644 third_party/rust/syn/tests/test_stmt.rs + create mode 100644 third_party/rust/syn/tests/test_ty.rs + create mode 100644 third_party/rust/syn/tests/test_visibility.rs + +diff --git a/Cargo.lock b/Cargo.lock +index 19117e8368..d5fe0f6457 100644 +--- a/Cargo.lock ++++ b/Cargo.lock +@@ -3717,9 +3717,9 @@ dependencies = [ + + [[package]] + name = "proc-macro2" +-version = "1.0.5" ++version = "1.0.24" + source = "registry+https://github.com/rust-lang/crates.io-index" +-checksum = "90cf5f418035b98e655e9cdb225047638296b862b42411c4e45bb88d700f7fc0" ++checksum = "1e0704ee1a7e00d7bb417d0770ea303c1bccbabf0ef1667dae92b5967f5f8a71" + dependencies = [ + "unicode-xid", + ] +@@ -4647,9 +4647,9 @@ dependencies = [ + + [[package]] + name = "syn" +-version = "1.0.5" ++version = "1.0.40" + source = "registry+https://github.com/rust-lang/crates.io-index" +-checksum = "66850e97125af79138385e9b88339cbcd037e3f28ceab8c5ad98e64f0f1f80bf" ++checksum = "963f7d3cc59b59b9325165add223142bbf1df27655d07789f109896d353d8350" + dependencies = [ + "proc-macro2", + "quote", +diff --git a/third_party/rust/lucet-wasi/.cargo-checksum.json b/third_party/rust/lucet-wasi/.cargo-checksum.json +index 229fc9978c..2c8c0a3c22 100644 +--- a/third_party/rust/lucet-wasi/.cargo-checksum.json ++++ b/third_party/rust/lucet-wasi/.cargo-checksum.json +@@ -1 +1 @@ +-{"files":{"Cargo.toml":"fea1408a1c1b1e84b06044a0b12cb26c8fd3253ca124debb6cd3e4faab48fcbd","LICENSE":"268872b9816f90fd8e85db5a28d33f8150ebb8dd016653fb39ef1f94f2686bc5","LICENSE.cloudabi-utils":"86a34251f0aab76b7dc3daf8d252afbdf481ea94aa5b46d020205178b7e2eac1","LICENSE.wasmtime":"a6c48161a09acc75a0e25503bab66a731eb5fba5392ed4bb4743e4ba5085327a","bindings.json":"fad8362f845e6f7a2af1d7547cee5730894e7b96804f338181fc070ffdcfae1e","build.rs":"593173ad03963afcbef43f1843be6f04cde1df3eae0298ca20bf881019dd350a","examples/.gitignore":"44575cf5b28512d75644bf54a517dcef304ff809fd511747621b4d64f19aac66","examples/Makefile":"d2d2ceeb1bc4435189ea9a2710b6f5f5331ce6aa73ae8a4f4edcca215058a9b4","examples/README.md":"f2a5be6cc88d511c9f4d3bfefdd42dcb2ace813bc23f6a4430b6b543f7373341","examples/hello.c":"9cbc0d3173e02309e15372835fa849d196b2a202d81806fea60378e1878d0c53","examples/pseudoquine.c":"8fd696f8e1b8fb86188564a05f4776875ead31d785a12e3aa4af9d9c1b46d5b5","include/lucet_wasi.h":"497f712c64f753ebdf73ab05b0b340d50094f9c59bb8637ccbf99d895cd20966","src/bindings.rs":"edbeb51d1a93fd31039ee1f1dc7c1b6c0bded2cf5dad10039e8b7da81a4d4a12","src/c_api.rs":"a9c73070a88a461882a28e3e2adfd773f569b964c7ffabde39a3cef907782f71","src/ctx.rs":"578f87c35cce12eaebec95d03e31954c3e6cd0afa214a0fec068f03814eb0cc7","src/fdentry.rs":"94a8480fa587e5586327dfd6b66d8a6a3ef1f8091ba8deb335bf45642f4f98e6","src/host.rs":"6f05f8fea2afed827abfc3c008a5854a8023d91d066580ecfb49e5c8036ef3a3","src/hostcalls/fs.rs":"4726e6f695f7d1d4e371ec52c57f4e36b0ba0d2302fc008b21a301f5fd7a5c97","src/hostcalls/fs_helpers.rs":"474bce0a1f15fa23b0b0b8aa83143d993dd2cbd7cdfc38c118d452d04e80caea","src/hostcalls/misc.rs":"83d087891d92af08cfa2d2e0c5f41cc47cb8219460f6dbcc8666b418dfef206e","src/hostcalls/mod.rs":"4c5d3f65c69503e11e647770879026c37c0e5e01a99b7116c8fb9411b4797187","src/hostcalls/timers.rs":"e65d6a491256b5d6051b6816f6c5049ba3cdc6142651bac81f34d659c1c2a104","src/lib.rs":"5554e1a3f0cd3756173ece6435a0d01b2f520b3401cd5fc33180a04fb9f69bbe","src/memory.rs":"0a09026b15d27f99d74e560cd94795f645cba414a8491bc961987fab9d9da69b","src/wasi_host.rs":"cacbdac28304a837b11e5ad400ae9de3ee79c0284be335e64606ecdfe426ad6e","src/wasm32.rs":"13a5dc6e59784662f1e55eccb457cbbae241a96f70cfa72c41d55858ca05b980","tests/guests/cant_dotdot.c":"609b8cece9443e375a0b38a7e43651b179f66ee9c686edba6696fe1bcd45b111","tests/guests/clock_getres.c":"f5e41c0a2b05a8d7cdb5b4da6c8b6778b858004c1e9d115503c45a1d976be33b","tests/guests/duplicate_import.wat":"4bd8d7a5c1d1597dbe7648300e94e3fab84d7ab068d56cfb656aa1a208026cee","tests/guests/exitcode.c":"b7c5cec3ead0ed82326c568287a1f3398e71ae7e447ce49a3c4c7114c82495af","tests/guests/follow_symlink.c":"de3143ad2bbbfe834c0c32b54c9fcf144ca4eba5cdcf7588929e5f47225ab616","tests/guests/fs.c":"0dca5232ff5da1b7745e3b44bca39333c01a20ba4eae1a6a0a1c492c71ca1efa","tests/guests/getentropy.c":"5d80bcc68dcf3ba91576969055099d61635ae713c057b3cb36afb122a5f26347","tests/guests/getrusage.c":"8114c103b85eb564d9ab43684958bc1939de3794d314b7c121762f3a2f0434a6","tests/guests/gettimeofday.c":"4a57f376b06f4228017b82695448a0bd213fb91455f5301d689cd87fcff01f06","tests/guests/notdir.c":"bd8f8b24360b7cf8d5dced9d9ba4c15843fcbbae89fecc13e3a457c33a275e28","tests/guests/poll.c":"aefaa9b58ce9906dc379e0bd25fa68dfbf8cdffb48cd5ecde1d67708b83b366d","tests/guests/preopen_populates.c":"f186e4eb4aab6a1d9ec7bc5c49eaea6d9d162e0159dfe8f953bb48ade9b58d43","tests/guests/read_file.c":"1aab9393f005f05b69592826d7c4d384a115d5bca42c66f10a901811b4b1dcac","tests/guests/read_file_twice.c":"04a3dad7a43b93e36efd4e2c822c11b3f129429ec799af304d82b358686c578a","tests/guests/stat.c":"02756933ea7d4337b4fa04344b32968851b02f9d0bd5ea1cb0e2f022e8c65ab0","tests/guests/stdin.c":"66efc4b54f68d1138046f1afefae15f7d4555b2904b4a988818e61e67fe8fefb","tests/guests/symlink_escape.c":"686e047b5c986e29c854bcd93996d027dcdc8721219fa9fa532efc98d2798f5c","tests/guests/symlink_loop.c":"2bbddf3a5edfc6e5f3c0fa82cee4ac92b18804810509e263abd17f5240cd37e5","tests/guests/write_file.c":"9e9b14552c2445cfa6d0aa26b334081a59e6e3428dbb17ceca005a9ba59d3220","tests/test_helpers/mod.rs":"bc18194317611fe1be5c439a7a9e0de75399555c3b6de4275af149fb180456c8","tests/tests.rs":"173a7e0f086f6ed46474686cc3413ee68bbd2ff67004f7790e963a1392c7c46e"},"package":null} +\ No newline at end of file ++{"files":{"Cargo.toml":"fea1408a1c1b1e84b06044a0b12cb26c8fd3253ca124debb6cd3e4faab48fcbd","LICENSE":"268872b9816f90fd8e85db5a28d33f8150ebb8dd016653fb39ef1f94f2686bc5","LICENSE.cloudabi-utils":"86a34251f0aab76b7dc3daf8d252afbdf481ea94aa5b46d020205178b7e2eac1","LICENSE.wasmtime":"a6c48161a09acc75a0e25503bab66a731eb5fba5392ed4bb4743e4ba5085327a","bindings.json":"fad8362f845e6f7a2af1d7547cee5730894e7b96804f338181fc070ffdcfae1e","build.rs":"593173ad03963afcbef43f1843be6f04cde1df3eae0298ca20bf881019dd350a","examples/Makefile":"d2d2ceeb1bc4435189ea9a2710b6f5f5331ce6aa73ae8a4f4edcca215058a9b4","examples/README.md":"f2a5be6cc88d511c9f4d3bfefdd42dcb2ace813bc23f6a4430b6b543f7373341","examples/hello.c":"9cbc0d3173e02309e15372835fa849d196b2a202d81806fea60378e1878d0c53","examples/pseudoquine.c":"8fd696f8e1b8fb86188564a05f4776875ead31d785a12e3aa4af9d9c1b46d5b5","include/lucet_wasi.h":"497f712c64f753ebdf73ab05b0b340d50094f9c59bb8637ccbf99d895cd20966","src/bindings.rs":"edbeb51d1a93fd31039ee1f1dc7c1b6c0bded2cf5dad10039e8b7da81a4d4a12","src/c_api.rs":"a9c73070a88a461882a28e3e2adfd773f569b964c7ffabde39a3cef907782f71","src/ctx.rs":"578f87c35cce12eaebec95d03e31954c3e6cd0afa214a0fec068f03814eb0cc7","src/fdentry.rs":"94a8480fa587e5586327dfd6b66d8a6a3ef1f8091ba8deb335bf45642f4f98e6","src/host.rs":"6f05f8fea2afed827abfc3c008a5854a8023d91d066580ecfb49e5c8036ef3a3","src/hostcalls/fs.rs":"4726e6f695f7d1d4e371ec52c57f4e36b0ba0d2302fc008b21a301f5fd7a5c97","src/hostcalls/fs_helpers.rs":"474bce0a1f15fa23b0b0b8aa83143d993dd2cbd7cdfc38c118d452d04e80caea","src/hostcalls/misc.rs":"83d087891d92af08cfa2d2e0c5f41cc47cb8219460f6dbcc8666b418dfef206e","src/hostcalls/mod.rs":"4c5d3f65c69503e11e647770879026c37c0e5e01a99b7116c8fb9411b4797187","src/hostcalls/timers.rs":"e65d6a491256b5d6051b6816f6c5049ba3cdc6142651bac81f34d659c1c2a104","src/lib.rs":"5554e1a3f0cd3756173ece6435a0d01b2f520b3401cd5fc33180a04fb9f69bbe","src/memory.rs":"0a09026b15d27f99d74e560cd94795f645cba414a8491bc961987fab9d9da69b","src/wasi_host.rs":"cacbdac28304a837b11e5ad400ae9de3ee79c0284be335e64606ecdfe426ad6e","src/wasm32.rs":"13a5dc6e59784662f1e55eccb457cbbae241a96f70cfa72c41d55858ca05b980","tests/guests/cant_dotdot.c":"609b8cece9443e375a0b38a7e43651b179f66ee9c686edba6696fe1bcd45b111","tests/guests/clock_getres.c":"f5e41c0a2b05a8d7cdb5b4da6c8b6778b858004c1e9d115503c45a1d976be33b","tests/guests/duplicate_import.wat":"4bd8d7a5c1d1597dbe7648300e94e3fab84d7ab068d56cfb656aa1a208026cee","tests/guests/exitcode.c":"b7c5cec3ead0ed82326c568287a1f3398e71ae7e447ce49a3c4c7114c82495af","tests/guests/follow_symlink.c":"de3143ad2bbbfe834c0c32b54c9fcf144ca4eba5cdcf7588929e5f47225ab616","tests/guests/fs.c":"0dca5232ff5da1b7745e3b44bca39333c01a20ba4eae1a6a0a1c492c71ca1efa","tests/guests/getentropy.c":"5d80bcc68dcf3ba91576969055099d61635ae713c057b3cb36afb122a5f26347","tests/guests/getrusage.c":"8114c103b85eb564d9ab43684958bc1939de3794d314b7c121762f3a2f0434a6","tests/guests/gettimeofday.c":"4a57f376b06f4228017b82695448a0bd213fb91455f5301d689cd87fcff01f06","tests/guests/notdir.c":"bd8f8b24360b7cf8d5dced9d9ba4c15843fcbbae89fecc13e3a457c33a275e28","tests/guests/poll.c":"aefaa9b58ce9906dc379e0bd25fa68dfbf8cdffb48cd5ecde1d67708b83b366d","tests/guests/preopen_populates.c":"f186e4eb4aab6a1d9ec7bc5c49eaea6d9d162e0159dfe8f953bb48ade9b58d43","tests/guests/read_file.c":"1aab9393f005f05b69592826d7c4d384a115d5bca42c66f10a901811b4b1dcac","tests/guests/read_file_twice.c":"04a3dad7a43b93e36efd4e2c822c11b3f129429ec799af304d82b358686c578a","tests/guests/stat.c":"02756933ea7d4337b4fa04344b32968851b02f9d0bd5ea1cb0e2f022e8c65ab0","tests/guests/stdin.c":"66efc4b54f68d1138046f1afefae15f7d4555b2904b4a988818e61e67fe8fefb","tests/guests/symlink_escape.c":"686e047b5c986e29c854bcd93996d027dcdc8721219fa9fa532efc98d2798f5c","tests/guests/symlink_loop.c":"2bbddf3a5edfc6e5f3c0fa82cee4ac92b18804810509e263abd17f5240cd37e5","tests/guests/write_file.c":"9e9b14552c2445cfa6d0aa26b334081a59e6e3428dbb17ceca005a9ba59d3220","tests/test_helpers/mod.rs":"bc18194317611fe1be5c439a7a9e0de75399555c3b6de4275af149fb180456c8","tests/tests.rs":"173a7e0f086f6ed46474686cc3413ee68bbd2ff67004f7790e963a1392c7c46e"},"package":null} +\ No newline at end of file +diff --git a/third_party/rust/packed_simd/.cargo-checksum.json b/third_party/rust/packed_simd/.cargo-checksum.json +index 01afcc1efd..c727a10006 100644 +--- a/third_party/rust/packed_simd/.cargo-checksum.json ++++ b/third_party/rust/packed_simd/.cargo-checksum.json +@@ -1 +1 @@ +-{"files":{".appveyor.yml":"f1ed01850e0d725f9498f52a1a63ddf40702ad6e0bf5b2d7c4c04d76e96794a3",".travis.yml":"e9258d9a54fdaf4cbc12405fe5993ac4497eb2b29021691dbc91b19cb9b52227","Cargo.toml":"089941ba3c89ea111cbea3cc3abdcdcf2b9d0ae0db268d7269ee38226db950e5","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"6485b8ed310d3f0340bf1ad1f47645069ce4069dcc6bb46c7d5c6faf41de1fdb","bors.toml":"dee881dc69b9b7834e4eba5d95c3ed5a416d4628815a167d6a22d4cb4fb064b8","build.rs":"f3baefc5e5bb9b250e762a1466371b922fd7ee4243c217b2d014307603c2f57a","ci/all.sh":"a23d14e10cb26a0eb719e389c30eb955fa53cddcd436890646df09af640bd2eb","ci/android-install-ndk.sh":"0f1746108cc30bf9b9ba45bcde7b19fc1a8bdf5b0258035b4eb8dc69b75efac4","ci/android-install-sdk.sh":"3490432022c5c8f5a115c084f7a9aca1626f96c0c87ffb62019228c4346b47e4","ci/android-sysimage.sh":"ebf4e5daa1f0fe1b2092b79f0f3f161c4c4275cb744e52352c4d81ab451e4c5a","ci/benchmark.sh":"b61d19ef6b90deba8fb79dee74c8b062d94844676293da346da87bb78a9a49a4","ci/deploy_and_run_on_ios_simulator.rs":"ec8ecf82d92072676aa47f0d1a3d021b60a7ae3531153ef12d2ff4541fc294dc","ci/docker/aarch64-linux-android/Dockerfile":"ace2e7d33c87bc0f6d3962a4a3408c04557646f7f51ab99cfbf574906796b016","ci/docker/aarch64-unknown-linux-gnu/Dockerfile":"1ecdac757101d951794fb2ab0deaa278199cf25f2e08a15c7d40ff31a8556184","ci/docker/arm-linux-androideabi/Dockerfile":"370e55d3330a413a3ccf677b3afb3e0ef9018a5fab263faa97ae8ac017fc2286","ci/docker/arm-unknown-linux-gnueabi/Dockerfile":"e25d88f6c0c94aada3d2e3f08243f755feb7e869dc5dc505b3799719cb1af591","ci/docker/arm-unknown-linux-gnueabihf/Dockerfile":"f126f4c7bae8c11ab8b16df06ad997863f0838825a9c08c9899a3eedb6d570bd","ci/docker/armv7-unknown-linux-gnueabihf/Dockerfile":"b647545c158ee480a4c581dbdc1f57833aef056c8d498acc04b573e842bf803c","ci/docker/i586-unknown-linux-gnu/Dockerfile":"0d492759017307ccf74dc2aa4a8cf6623daf3dc728c708dc2b18fa7940800cba","ci/docker/i686-unknown-linux-gnu/Dockerfile":"0d492759017307ccf74dc2aa4a8cf6623daf3dc728c708dc2b18fa7940800cba","ci/docker/mips-unknown-linux-gnu/Dockerfile":"323776469bb7b160385f3621d66e3ee14c75242f8180f916e65af048a29d4ea0","ci/docker/mips64-unknown-linux-gnuabi64/Dockerfile":"c647f6948a9a43b0be695cbed4eac752120d0faf28e5e69c718cb10406921dab","ci/docker/mips64el-unknown-linux-gnuabi64/Dockerfile":"77bfd00cc8639509be381b394f077e39b45a00158ad61b4e1656714c714665d1","ci/docker/mipsel-unknown-linux-musl/Dockerfile":"ec5bea6c98a3b626731fdb95f9ff2d1182639c76e8fb16d3271d0fc884901524","ci/docker/powerpc-unknown-linux-gnu/Dockerfile":"4f2b662de66e83d1354f650b7077692309637f786c2ea5516c31b5c2ee10af2d","ci/docker/powerpc64-unknown-linux-gnu/Dockerfile":"a9595402b772bc365982e22a0096a8988825d90b09b5faa97ab192e76072f71d","ci/docker/powerpc64le-unknown-linux-gnu/Dockerfile":"df3c381c157439695ae8cd10ab71664702c061e3b4ab22906a5ad6c2680acfed","ci/docker/s390x-unknown-linux-gnu/Dockerfile":"93fb44df3d7fd31ead158570667c97b5076a05c3d968af4a84bc13819a8f2db8","ci/docker/sparc64-unknown-linux-gnu/Dockerfile":"da1c39a3ff1fe22e41395fa7c8934e90b4c1788e551b9aec6e38bfd94effc437","ci/docker/thumbv7neon-linux-androideabi/Dockerfile":"c2decd5591bd7a09378901bef629cd944acf052eb55e4f35b79eb9cb4d62246a","ci/docker/thumbv7neon-unknown-linux-gnueabihf/Dockerfile":"75c0c56161c7382b439de74c00de1c0e3dc9d59560cd6720976a751034b78714","ci/docker/wasm32-unknown-unknown/Dockerfile":"3e5f294bc1e004aa599086c2af49d6f3e7459fa250f5fbdd60cf67d53db78758","ci/docker/x86_64-linux-android/Dockerfile":"685040273cf350d5509e580ac451555efa19790c8723ca2af066adadc6880ad2","ci/docker/x86_64-unknown-linux-gnu-emulated/Dockerfile":"44b6203d9290bfdc53d81219f0937e1110847a23dd982ec8c4de388354f01536","ci/docker/x86_64-unknown-linux-gnu/Dockerfile":"d253c86803b22da428fa9cc671a05f18d3318eca7733b8dccb4f7be1ddf524c5","ci/dox.sh":"5b61711be47a4e3dde0ddd15ba73d256ea95fd75af3897732c24db1dc7e66366","ci/linux-s390x.sh":"d6b732d7795b4ba131326aff893bca6228a7d2eb0e9402f135705413dbbe0dce","ci/linux-sparc64.sh":"c92966838b1ab7ad3b7a344833ee726aba6b647cf5952e56f0ad1ba420b13325","ci/lld-shim.rs":"3d7f71ec23a49e2b67f694a0168786f9a954dda15f5a138815d966643fd3fcc3","ci/max_line_width.sh":"0a1518bba4c9ecaa55694cb2e9930d0e19c265baabf73143f17f9cf285aaa5bb","ci/run-docker.sh":"92e036390ad9b0d16f109579df1b5ced2e72e9afea40c7d011400ebd3a2a90de","ci/run.sh":"63259e22a96ba539f53c06b1b39f53e3a78a71171652e7afc170836110ccd913","ci/run_examples.sh":"d1a23c6c35374a0678ba5114b9b8fefd8be0a79e774872a8bf0898d1baca18d0","ci/runtest-android.rs":"145a8e9799a5223975061fe7e586ade5669ee4877a7d7a4cf6b4ab48e8e36c7c","ci/setup_benchmarks.sh":"73fb981a8fdb1dcd54409d3c0fbbfb8f77a3ceabf8626a6b9bf9d21d6bc8ce72","ci/test-runner-linux":"c8aa6025cff5306f4f31d0c61dc5f9d4dd5a1d189ab613ef8d4c367c694d9ccd","contributing.md":"2cc8c9c560ae17867e69b06d09b758dbf7bc39eb774ada50a743724b10acc0a2","perf-guide/.gitignore":"fe82c7da551079d832cf74200b0b359b4df9828cb4a0416fa7384f07a2ae6a13","perf-guide/book.toml":"115a98284126c6b180178b44713314cc494f08a71662ee2ce15cf67f17a51064","perf-guide/src/SUMMARY.md":"3e03bffc991fdc2050f3d51842d72d9d21ea6abab56a3baf3b2d5973a78b89e1","perf-guide/src/ascii.css":"29afb08833b2fe2250f0412e1fa1161a2432a0820a14953c87124407417c741a","perf-guide/src/bound_checks.md":"5e4991ff58a183ef0cd9fdc1feb4cd12d083b44bdf87393bbb0927808ef3ce7d","perf-guide/src/float-math/approx.md":"8c09032fa2d795a0c5db1775826c850d28eb2627846d0965c60ee72de63735ad","perf-guide/src/float-math/fma.md":"311076ba4b741d604a82e74b83a8d7e8c318fcbd7f64c4392d1cf5af95c60243","perf-guide/src/float-math/fp.md":"04153e775ab6e4f0d7837bcc515230d327b04edfa34c84ce9c9e10ebaeef2be8","perf-guide/src/float-math/svml.md":"0798873b8eedaeda5fed62dc91645b57c20775a02d3cd74d8bd06958f1516506","perf-guide/src/introduction.md":"9f5a19e9e6751f25d2daad39891a0cc600974527ec4c8305843f9618910671bd","perf-guide/src/prof/linux.md":"447731eb5de7d69166728fdbc5ecb0c0c9db678ea493b45a592d67dd002184c0","perf-guide/src/prof/mca.md":"f56d54f3d20e7aa4d32052186e8237b03d65971eb5d112802b442570ff11d344","perf-guide/src/prof/profiling.md":"8a650c0fd6ede0964789bb6577557eeef1d8226a896788602ce61528e260e43c","perf-guide/src/target-feature/attribute.md":"615f88dca0a707b6c416fa605435dd6e1fb5361cc639429cbf68cd87624bd78b","perf-guide/src/target-feature/features.md":"17077760ff24c006b606dd21889c53d87228f4311f3ba3a574f9afdeacd86165","perf-guide/src/target-feature/inlining.md":"7ed1d7068d8173a00d84c16cfe5871cd68b9f04f8d0cca2d01ebc84957ebf2f6","perf-guide/src/target-feature/practice.md":"c4b371842e0086df178488fec97f20def8f0c62ee588bcd25fd948b9b1fa227e","perf-guide/src/target-feature/runtime.md":"835425f5ee597fb3e51d36e725a81ebee29f4561231d19563cd4da81dbb1cfcb","perf-guide/src/target-feature/rustflags.md":"ab49712e9293a65d74d540ba4784fcb57ff1119ec05a575d895c071f1a620f64","perf-guide/src/vert-hor-ops.md":"c6211c0ee91e60552ec592d89d9d957eedc21dee3cbd89e1ad6765ea06a27471","readme.md":"585a8f0e16877fb9abb00cd17a175fcb9d7857840c6c61209f1827ffab095070","rustfmt.toml":"de6101d0670bad65fb3b337d56957d2a024e017e5ab146ec784d77312daaf8ff","src/api.rs":"331a3a4abb19cee2df5f2df4ad7c3e88b45e62cf23fdacfc9bbaa633dc5cf788","src/api/bit_manip.rs":"e68290ee679cc5abc9c73afbe635c1035f8cbfe849e5c751a1680e459244c39e","src/api/cast.rs":"03b94a3d316ac7b7be7068810044911e965e889a0ace7bae762749ca74a92747","src/api/cast/macros.rs":"b0a14d0c83ad2ebb7a275180f6d9e3f2bc312ba57a7d3d6c39fad4e0f20f9408","src/api/cast/v128.rs":"63e28c6a3edf1a7a635f51b8d3c6adbb1d46f884d92a196b3d4a6e743d809416","src/api/cast/v16.rs":"2a584eeb57fd47baad6f3533764301b04aaaac23702b7a8db12598ac02899262","src/api/cast/v256.rs":"b91c15ed8d1536ecd97b4eb79ff9d5aba0552cd9b6f0ea6435b05f2273e23b3a","src/api/cast/v32.rs":"62ec89fcce7fa7f28497ee5770adc8f81d2d3a6b2925b02f7dc06504c40e8f38","src/api/cast/v512.rs":"d855cb943ae7106e9599ef38e30a3afb1c6bd5433178baca54cb128fd9a7d143","src/api/cast/v64.rs":"fe0f7dfaf4fc0c0c1a78c96fcfcdfdc2a1e2845843b11aa797a0c6fb52a8f774","src/api/cmp.rs":"357c3a2a09c6d4611c32dd7fa95be2fae933d513e229026ec9b44451a77b884e","src/api/cmp/eq.rs":"60f70f355bae4cb5b17db53204cacc3890f70670611c17df638d4c04f7cc8075","src/api/cmp/ord.rs":"589f7234761c294fa5df8f525bc4acd5a47cdb602207d524a0d4e19804cd9695","src/api/cmp/partial_eq.rs":"3ed23d2a930b0f9750c3a5309da766b03dc4f9c4d375b42ad3c50fe732693d15","src/api/cmp/partial_ord.rs":"e16b11805c94048acd058c93994b5bc74bb187f8d7e3b86a87df60e1601467f9","src/api/cmp/vertical.rs":"de3d62f38eba817299aa16f1e1939954c9a447e316509397465c2830852ba053","src/api/default.rs":"b61f92fc0e33a2633b3375eb405beba480da071cde03df4d437d8a6058afcd97","src/api/fmt.rs":"67fb804bb86b6cd77cf8cd492b5733ce437071b66fe3297278b8a6552c325dda","src/api/fmt/binary.rs":"35cb5c266197d6224d598fb3d286e5fe48ef0c01ed356c2ff6fe9ba946f96a92","src/api/fmt/debug.rs":"aa18eea443bf353fea3db8b1a025132bbcaf91e747ecfa43b8d9fce9af395a0c","src/api/fmt/lower_hex.rs":"69d5be366631af309f214e8031c8c20267fcc27a695eac6f45c6bc1df72a67e6","src/api/fmt/octal.rs":"9eb11ba3d990213f3c7f1ec25edba7ce997cb1320e16d308c83498ba6b9bfbd9","src/api/fmt/upper_hex.rs":"a4637d085b7bb20e759ce58e08435b510a563ba3dd468af2b03560fdc5511562","src/api/from.rs":"2e599d8329cb05eaf06224cc441355c4b7b51254fc19256619333be8c149d444","src/api/from/from_array.rs":"4151593c7bba7455821fffa5b59867005a77c95d32f1f0cc3fd87294000157d9","src/api/from/from_vector.rs":"9764371aa9e6005aace74dea14f59e5611a095b7cf42707940924749282c52f0","src/api/hash.rs":"562cfa3f1d8eb9a733c035a3665a599c2f1e341ee820d8fbdd102a4398a441bc","src/api/into_bits.rs":"82297f0697d67b5a015e904e7e6e7b2a7066ba825bc54b94b4ff3e22d7a1eefb","src/api/into_bits/arch_specific.rs":"1f925390b0ce7132587d95f2419c6e2ad3e1a9d17eb1d9c120a1c1c4bdf4277e","src/api/into_bits/macros.rs":"d762406de25aedff88d460dec7a80dc8e825a2a419d53218ce007efa6a1d3e04","src/api/into_bits/v128.rs":"ecdc5893664c71d7ab1ff3697c3fbe490d20d8748b9b76881d05e7625e40d74c","src/api/into_bits/v16.rs":"5459ec7dad1ad7bd30dc7e48374580b993abf23701d9c3cb22203fa0a9aabb6d","src/api/into_bits/v256.rs":"90ea351da0380ead1bf0f63b620afd40d01d638d09f7e7be31840bd2c1d9c663","src/api/into_bits/v32.rs":"ee1dc5a430050e16f51154b5fe85b1536f5feddf2ea23dd1d3859b67c4afc6fc","src/api/into_bits/v512.rs":"f72098ed1c9a23944f3d01abaf5e0f2d0e81d35a06fdadd2183e896d41b59867","src/api/into_bits/v64.rs":"6394462facdfe7827349c742b7801f1291e75a720dfb8c0b52100df46f371c98","src/api/math.rs":"8b2a2fc651917a850539f993aa0b9e5bf4da67b11685285b8de8cdca311719ec","src/api/math/float.rs":"61d2794d68262a1090ae473bd30793b5f65cf732f32a6694a3af2ce5d9225616","src/api/math/float/abs.rs":"5b6b2701e2e11135b7ce58a05052ea8120e10e4702c95d046b9d21b827b26bf8","src/api/math/float/consts.rs":"78acba000d3fa527111300b6327c1932de9c4c1e02d4174e1a5615c01463d38c","src/api/math/float/cos.rs":"4c2dd7173728ef189314f1576c9486e03be21b7da98843b2f9011282a7979e31","src/api/math/float/exp.rs":"7c6d5f1e304f498a01cfa23b92380c815d7da0ad94eae3483783bc377d287eef","src/api/math/float/ln.rs":"54c7583f3df793b39ff57534fade27b41bb992439e5dc178252f5ca3190a3e54","src/api/math/float/mul_add.rs":"62cac77660d20159276d4c9ef066eb90c81cbddb808e8e157182c607625ad2eb","src/api/math/float/mul_adde.rs":"bae056ee9f3a70df39ec3c3b2f6437c65303888a7b843ef1a5bcf1f5aca0e602","src/api/math/float/powf.rs":"9ddb938984b36d39d82a82f862f80df8f7fb013f1d222d45698d41d88472f568","src/api/math/float/recpre.rs":"589225794ff1dbf31158dff660e6d4509ecc8befbb57c633900dea5ac0b840d6","src/api/math/float/rsqrte.rs":"a32abdcc318d7ccc8448231f54d75b884b7cbeb03a7d595713ab6243036f4dbf","src/api/math/float/sin.rs":"cbd3622b7df74f19691743001c8cf747a201f8977ad90542fee915f37dcd1e49","src/api/math/float/sqrt.rs":"0c66d5d63fb08e4d99c6b82a8828e41173aff1ac9fa1a2764a11fac217ccf2ac","src/api/math/float/sqrte.rs":"731e1c9f321b662accdd27dacb3aac2e8043b7aecb2f2161dde733bd9f025362","src/api/minimal.rs":"1f22bcc528555444e76de569ec0ae2029b9ae9d04805efeafa93369c8098036b","src/api/minimal/iuf.rs":"c501a6696950cf5e521765f178de548af64fdfb6e10d026616d09fab93ca2d17","src/api/minimal/mask.rs":"42e415f536c5193d0218f5a754b34b87fd7c971bff068009f958712166ff056d","src/api/minimal/ptr.rs":"a9ee482d1dd1c956fb8f3f179e6e620b1de4e9d713961461d4c6923a4ef2e67c","src/api/ops.rs":"3e273b277a0f3019d42c3c59ca94a5afd4885d5ae6d2182e5089bbeec9de42ee","src/api/ops/scalar_arithmetic.rs":"d2d5ad897a59dd0787544f927e0e7ca4072c3e58b0f4a2324083312b0d5a21d7","src/api/ops/scalar_bitwise.rs":"482204e459ca6be79568e1c9f70adbe2d2151412ddf122fb2161be8ebb51c40c","src/api/ops/scalar_mask_bitwise.rs":"c250f52042e37b22d57256c80d4604104cfd2fbe2a2e127c676267270ca5d350","src/api/ops/scalar_shifts.rs":"987f8fdebeedc16e3d77c1b732e7826ef70633c541d16dfa290845d5c6289150","src/api/ops/vector_arithmetic.rs":"ddca15d09ddeef502c2ed66117a62300ca65d87e959e8b622d767bdf1c307910","src/api/ops/vector_bitwise.rs":"b3968f7005b649edcc22a54e2379b14d5ee19045f2e784029805781ae043b5ee","src/api/ops/vector_float_min_max.rs":"f5155dce75219f4ba11275b1f295d2fdcddd49d174a6f1fb2ace7ea42813ce41","src/api/ops/vector_int_min_max.rs":"a378789c6ff9b32a51fbd0a97ffd36ed102cd1fe6a067d2b02017c1df342def6","src/api/ops/vector_mask_bitwise.rs":"5052d18517d765415d40327e6e8e55a312daaca0a5e2aec959bfa54b1675f9c8","src/api/ops/vector_neg.rs":"5c62f6b0221983cdbd23cd0a3af3672e6ba1255f0dfe8b19aae6fbd6503e231b","src/api/ops/vector_rotates.rs":"03cbe8a400fd7c688e4ee771a990a6754f2031b1a59b19ae81158b21471167e5","src/api/ops/vector_shifts.rs":"9bf69d0087268f61009e39aea52e03a90f378910206b6a28e8393178b6a5d0e0","src/api/ptr.rs":"8a793251bed6130dcfb2f1519ceaa18b751bbb15875928d0fb6deb5a5e07523a","src/api/ptr/gather_scatter.rs":"9ddd960365e050674b25b2fd3116e24d94669b4375d74e71c03e3f1469576066","src/api/reductions.rs":"ae5baca81352ecd44526d6c30c0a1feeda475ec73ddd3c3ec6b14e944e5448ee","src/api/reductions/bitwise.rs":"8bf910ae226188bd15fc7e125f058cd2566b6186fcd0cd8fd020f352c39ce139","src/api/reductions/float_arithmetic.rs":"e58c8c87806a95df2b2b5b48ac5991036df024096d9d7c171a480fe9282896a4","src/api/reductions/integer_arithmetic.rs":"47471da1c5f859489680bb5d34ced3d3aa20081c16053a3af121a4496fcb57bf","src/api/reductions/mask.rs":"db83327a950e33a317f37fd33ca4e20c347fb415975ec024f3e23da8509425af","src/api/reductions/min_max.rs":"f27be3aa28e1c1f46de7890198db6e12f00c207085e89ef2de7e57ee443cdb98","src/api/select.rs":"a98e2ccf9fc6bdeed32d337c8675bc96c2fbe2cc34fbf149ad6047fb8e749774","src/api/shuffle.rs":"da58200790868c09659819322a489929a5b6e56c596ed07e6a44293ea02e7d09","src/api/shuffle1_dyn.rs":"bfea5a91905b31444e9ef7ca6eddb7a9606b7e22d3f71bb842eb2795a0346620","src/api/slice.rs":"ee87484e8af329547b9a5d4f2a69e8bed6ea10bbd96270d706083843d4eea2ac","src/api/slice/from_slice.rs":"4d4fe8a329c885fcb4fbcbedf99efb15a95296fe6b3f595056cc37037450d5ac","src/api/slice/write_to_slice.rs":"f5b23b2c4b91cfb26b713a9013a6c0da7f45eaefb79ba06dcbc27f3f23bda679","src/api/swap_bytes.rs":"4a6792a2e49a77475e1b237592b4b2804dbddb79c474331acd0dd71b36934259","src/codegen.rs":"c6eebc3d3665420aa6a2f317977e3c41a4f43e0550ac630cdbe8e4bbed5e2031","src/codegen/bit_manip.rs":"5559e095105a80003e0de35af1d19b0c65c9ab04eb743c7e01c5442d882eb34e","src/codegen/llvm.rs":"d1299c189abb17a6133f047574cffc7a6db4c1be37cb7d4785491cb5e8f8cf54","src/codegen/math.rs":"35f96e37a78fcf0cdb02146b7f27a45108fe06a37fc2a54d8851ce131a326178","src/codegen/math/float.rs":"dd86c0449e576c83b719700962ac017c332987fac08d91f2b7a2b1b883598170","src/codegen/math/float/abs.rs":"f56e2b4b8055ea861c1f5cbc6b6e1d8e7e5af163b62c13574ddee4e09513bfbc","src/codegen/math/float/cos.rs":"ef3b511a24d23045b310315e80348a9b7fedb576fc2de52d74290616a0abeb2a","src/codegen/math/float/cos_pi.rs":"4e7631a5d73dac21531e09ef1802d1180f8997509c2c8fa9f67f322194263a97","src/codegen/math/float/exp.rs":"61b691598c41b5622f24e4320c1bdd08701e612a516438bdddcc728fc3405c8c","src/codegen/math/float/ln.rs":"46b718b1ba8c9d99e1ad40f53d20dfde08a3063ca7bd2a9fdd6698e060da687e","src/codegen/math/float/macros.rs":"dd42135fff13f9aca4fd3a1a4e14c7e6c31aadc6d817d63b0d2fb9e62e062744","src/codegen/math/float/mul_add.rs":"a37bf764345d4b1714f97e83897b7cf0855fc2811704bcbc0012db91825339e1","src/codegen/math/float/mul_adde.rs":"c75702bfcb361de45964a93caf959a695ef2376bd069227600b8c6872665c755","src/codegen/math/float/powf.rs":"642346e982bc4c39203de0864d2149c4179cd7b21cf67a2951687932b4675872","src/codegen/math/float/sin.rs":"9d68164c90cdca6a85155040cdac42e27342ebe0b925273ef1593df721af4258","src/codegen/math/float/sin_cos_pi.rs":"9be02ad48585a1e8d99129382fbffbaed47852f15459256a708850b6b7a75405","src/codegen/math/float/sin_pi.rs":"9890347905b4d4a3c7341c3eb06406e46e60582bcf6960688bd727e5dadc6c57","src/codegen/math/float/sqrt.rs":"e3c60dcfb0c6d2fc62adabcc931b2d4040b83cab294dea36443fb4b89eb79e34","src/codegen/math/float/sqrte.rs":"f0f4ef9eb475ae41bcc7ec6a95ad744ba6b36925faa8b2c2814004396d196b63","src/codegen/pointer_sized_int.rs":"a70697169c28218b56fd2e8d5353f2e00671d1150d0c8cef77d613bdfacd84cb","src/codegen/reductions.rs":"645e2514746d01387ddd07f0aa4ffd8430cc9ab428d4fb13773ea319fa25dd95","src/codegen/reductions/mask.rs":"8f1afe6aabf096a3278e1fc3a30f736e04aa8b9ce96373cee22162d18cfe2702","src/codegen/reductions/mask/aarch64.rs":"cba6e17603d39795dcfe8339b6b7d8714c3e162a1f0a635979f037aa24fe4206","src/codegen/reductions/mask/arm.rs":"9447904818aa2c7c25d0963eead452a639a11ca7dbd6d21eedbfcaade07a0f33","src/codegen/reductions/mask/fallback.rs":"7a0ef9f7fd03ae318b495b95e121350cd61caffc5cc6ee17fabf130d5d933453","src/codegen/reductions/mask/fallback_impl.rs":"76547f396e55ef403327c77c314cf8db8c7a5c9b9819bfb925abeacf130249e5","src/codegen/reductions/mask/x86.rs":"14bd2c482071f2355beebcf7b7ecf950ff2dfcdb08c3ca50993092434a9de717","src/codegen/reductions/mask/x86/avx.rs":"b4913d87844c522903641cbbf10db4551addb1ce5e9e78278e21612fa65c733b","src/codegen/reductions/mask/x86/avx2.rs":"677aed3f056285285daa3adff8bc65e739630b4424defa6d9665e160f027507e","src/codegen/reductions/mask/x86/sse.rs":"226610b4ff88c676d5187114dd57b4a8800de6ce40884675e9198445b1ed0306","src/codegen/reductions/mask/x86/sse2.rs":"bc38e6c31cb4b3d62147eba6cac264e519e2a48e0f7ce9010cfa9ef0cf0ec9fd","src/codegen/shuffle.rs":"0abca97e92cdce49a58a39cc447eb09dc7d7715ef256c8dbd2181a186e61bb64","src/codegen/shuffle1_dyn.rs":"04523e9338133bdedb012dd076c2c564b79ce5593b0fc56d0fb6910e04190a81","src/codegen/swap_bytes.rs":"1d6cdc716eadddc92b4fd506b2445a821caa8dc00860447de09d7ebd69c2087f","src/codegen/v128.rs":"94226b31ec403d18d9d2fe06713f147c9c79e9b5f9105089088266313f843185","src/codegen/v16.rs":"ddec4ffb66b6f7aaffb9a1780c5ddba82557abd74f45073d335047e04cf74924","src/codegen/v256.rs":"6b63917f0444118d6b1595bff2045e59b97c4d24012bd575f69f1f0efc5a0241","src/codegen/v32.rs":"3477b3c5540aed86e61e2f5807dd31db947413cec9181c587d93ed6ec74f0eba","src/codegen/v512.rs":"5854f99d3aabc4cd42b28a20d9ce447756dc2ba024a409a69b6a8ae1f1842fc5","src/codegen/v64.rs":"e9e89caebfe63d10c0cbca61e4dfdba3b7e02ee0989170f80beed23237ddd950","src/codegen/vPtr.rs":"96d609a9eece4dcbbcc01ba0b8744d7f5958be12774176a2945bc676f4e6b5cb","src/codegen/vSize.rs":"eeee9858749aa82142b27bc120d1989bb74a6b82e1e4efbbeaccc9634dc9acfc","src/lib.rs":"1b5d419ff05ee0370d671810423ccc254708cc8d415c1dbac2a7a36be4bf63a8","src/masks.rs":"870f429967b2d7d5133f4d28d6c753fc5cef0570b27b29d4e966a066d22d2d0e","src/sealed.rs":"ff7f0324276408ae8249941cfa32c90b8835a54d750896b683efea857af19db2","src/testing.rs":"1d3a7862ef625e235a5734ad7204e68d350f902c0695182b1f08a0552432416e","src/testing/macros.rs":"6378856d7a40ba5ec5c7c0dad6327d79f0c77266921c24296d10aed6c68e9b98","src/testing/utils.rs":"d6fd5a5017f1f85d9d99585754f8f6ad06fc3d683b34083543e67a7cc6c1772c","src/v128.rs":"18fe263c4aa28cd06461c7070b0269f69f4a2e75749b8f142a83dfdfe4d22bf5","src/v16.rs":"e5c663c9fb3547eaeac78a5f7db9969f4d8b5ec96112bf2954602fff11f0aebd","src/v256.rs":"68732cd688ad12a56d8b4f8ddf279f77bdfe1be2943c7dc0c1b4f1a76798aa0f","src/v32.rs":"785b22a1ccb4a41bb53dfeb0670f624c0ce42e6cdf62d1747e3283777a1c70bd","src/v512.rs":"d1337bfe07f06a8f37f8e8fa7d4315b9307476ee435ad80dd5269eaed564fbfa","src/v64.rs":"3077468d65125b8f085e9454c8b2463a4d5225697464ba6a1300f8799528fd4b","src/vPtr.rs":"c9a53f41f466e17b6648a4ce390fd8f4d3a848d440eb8a9a803a11608d76eb05","src/vSize.rs":"5c46d3e8c3ee5863d9b6e37e681f871386e0efc254d6d84ba711edb529ce7b3c","tests/endianness.rs":"541a144be017e3dd7da7c8ea49d907dc02538245e8c5f3deb5bd43da92c929e1"},"package":null} +\ No newline at end of file ++{"files":{".appveyor.yml":"f1ed01850e0d725f9498f52a1a63ddf40702ad6e0bf5b2d7c4c04d76e96794a3",".travis.yml":"e9258d9a54fdaf4cbc12405fe5993ac4497eb2b29021691dbc91b19cb9b52227","Cargo.toml":"089941ba3c89ea111cbea3cc3abdcdcf2b9d0ae0db268d7269ee38226db950e5","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"6485b8ed310d3f0340bf1ad1f47645069ce4069dcc6bb46c7d5c6faf41de1fdb","bors.toml":"dee881dc69b9b7834e4eba5d95c3ed5a416d4628815a167d6a22d4cb4fb064b8","build.rs":"f3baefc5e5bb9b250e762a1466371b922fd7ee4243c217b2d014307603c2f57a","ci/all.sh":"a23d14e10cb26a0eb719e389c30eb955fa53cddcd436890646df09af640bd2eb","ci/android-install-ndk.sh":"0f1746108cc30bf9b9ba45bcde7b19fc1a8bdf5b0258035b4eb8dc69b75efac4","ci/android-install-sdk.sh":"3490432022c5c8f5a115c084f7a9aca1626f96c0c87ffb62019228c4346b47e4","ci/android-sysimage.sh":"ebf4e5daa1f0fe1b2092b79f0f3f161c4c4275cb744e52352c4d81ab451e4c5a","ci/benchmark.sh":"b61d19ef6b90deba8fb79dee74c8b062d94844676293da346da87bb78a9a49a4","ci/deploy_and_run_on_ios_simulator.rs":"ec8ecf82d92072676aa47f0d1a3d021b60a7ae3531153ef12d2ff4541fc294dc","ci/docker/aarch64-linux-android/Dockerfile":"ace2e7d33c87bc0f6d3962a4a3408c04557646f7f51ab99cfbf574906796b016","ci/docker/aarch64-unknown-linux-gnu/Dockerfile":"1ecdac757101d951794fb2ab0deaa278199cf25f2e08a15c7d40ff31a8556184","ci/docker/arm-linux-androideabi/Dockerfile":"370e55d3330a413a3ccf677b3afb3e0ef9018a5fab263faa97ae8ac017fc2286","ci/docker/arm-unknown-linux-gnueabi/Dockerfile":"e25d88f6c0c94aada3d2e3f08243f755feb7e869dc5dc505b3799719cb1af591","ci/docker/arm-unknown-linux-gnueabihf/Dockerfile":"f126f4c7bae8c11ab8b16df06ad997863f0838825a9c08c9899a3eedb6d570bd","ci/docker/armv7-unknown-linux-gnueabihf/Dockerfile":"b647545c158ee480a4c581dbdc1f57833aef056c8d498acc04b573e842bf803c","ci/docker/i586-unknown-linux-gnu/Dockerfile":"0d492759017307ccf74dc2aa4a8cf6623daf3dc728c708dc2b18fa7940800cba","ci/docker/i686-unknown-linux-gnu/Dockerfile":"0d492759017307ccf74dc2aa4a8cf6623daf3dc728c708dc2b18fa7940800cba","ci/docker/mips-unknown-linux-gnu/Dockerfile":"323776469bb7b160385f3621d66e3ee14c75242f8180f916e65af048a29d4ea0","ci/docker/mips64-unknown-linux-gnuabi64/Dockerfile":"c647f6948a9a43b0be695cbed4eac752120d0faf28e5e69c718cb10406921dab","ci/docker/mips64el-unknown-linux-gnuabi64/Dockerfile":"77bfd00cc8639509be381b394f077e39b45a00158ad61b4e1656714c714665d1","ci/docker/mipsel-unknown-linux-musl/Dockerfile":"ec5bea6c98a3b626731fdb95f9ff2d1182639c76e8fb16d3271d0fc884901524","ci/docker/powerpc-unknown-linux-gnu/Dockerfile":"4f2b662de66e83d1354f650b7077692309637f786c2ea5516c31b5c2ee10af2d","ci/docker/powerpc64-unknown-linux-gnu/Dockerfile":"a9595402b772bc365982e22a0096a8988825d90b09b5faa97ab192e76072f71d","ci/docker/powerpc64le-unknown-linux-gnu/Dockerfile":"df3c381c157439695ae8cd10ab71664702c061e3b4ab22906a5ad6c2680acfed","ci/docker/s390x-unknown-linux-gnu/Dockerfile":"93fb44df3d7fd31ead158570667c97b5076a05c3d968af4a84bc13819a8f2db8","ci/docker/sparc64-unknown-linux-gnu/Dockerfile":"da1c39a3ff1fe22e41395fa7c8934e90b4c1788e551b9aec6e38bfd94effc437","ci/docker/thumbv7neon-linux-androideabi/Dockerfile":"c2decd5591bd7a09378901bef629cd944acf052eb55e4f35b79eb9cb4d62246a","ci/docker/thumbv7neon-unknown-linux-gnueabihf/Dockerfile":"75c0c56161c7382b439de74c00de1c0e3dc9d59560cd6720976a751034b78714","ci/docker/wasm32-unknown-unknown/Dockerfile":"3e5f294bc1e004aa599086c2af49d6f3e7459fa250f5fbdd60cf67d53db78758","ci/docker/x86_64-linux-android/Dockerfile":"685040273cf350d5509e580ac451555efa19790c8723ca2af066adadc6880ad2","ci/docker/x86_64-unknown-linux-gnu-emulated/Dockerfile":"44b6203d9290bfdc53d81219f0937e1110847a23dd982ec8c4de388354f01536","ci/docker/x86_64-unknown-linux-gnu/Dockerfile":"d253c86803b22da428fa9cc671a05f18d3318eca7733b8dccb4f7be1ddf524c5","ci/dox.sh":"5b61711be47a4e3dde0ddd15ba73d256ea95fd75af3897732c24db1dc7e66366","ci/linux-s390x.sh":"d6b732d7795b4ba131326aff893bca6228a7d2eb0e9402f135705413dbbe0dce","ci/linux-sparc64.sh":"c92966838b1ab7ad3b7a344833ee726aba6b647cf5952e56f0ad1ba420b13325","ci/lld-shim.rs":"3d7f71ec23a49e2b67f694a0168786f9a954dda15f5a138815d966643fd3fcc3","ci/max_line_width.sh":"0a1518bba4c9ecaa55694cb2e9930d0e19c265baabf73143f17f9cf285aaa5bb","ci/run-docker.sh":"92e036390ad9b0d16f109579df1b5ced2e72e9afea40c7d011400ebd3a2a90de","ci/run.sh":"63259e22a96ba539f53c06b1b39f53e3a78a71171652e7afc170836110ccd913","ci/run_examples.sh":"d1a23c6c35374a0678ba5114b9b8fefd8be0a79e774872a8bf0898d1baca18d0","ci/runtest-android.rs":"145a8e9799a5223975061fe7e586ade5669ee4877a7d7a4cf6b4ab48e8e36c7c","ci/setup_benchmarks.sh":"73fb981a8fdb1dcd54409d3c0fbbfb8f77a3ceabf8626a6b9bf9d21d6bc8ce72","ci/test-runner-linux":"c8aa6025cff5306f4f31d0c61dc5f9d4dd5a1d189ab613ef8d4c367c694d9ccd","contributing.md":"2cc8c9c560ae17867e69b06d09b758dbf7bc39eb774ada50a743724b10acc0a2","perf-guide/book.toml":"115a98284126c6b180178b44713314cc494f08a71662ee2ce15cf67f17a51064","perf-guide/src/SUMMARY.md":"3e03bffc991fdc2050f3d51842d72d9d21ea6abab56a3baf3b2d5973a78b89e1","perf-guide/src/ascii.css":"29afb08833b2fe2250f0412e1fa1161a2432a0820a14953c87124407417c741a","perf-guide/src/bound_checks.md":"5e4991ff58a183ef0cd9fdc1feb4cd12d083b44bdf87393bbb0927808ef3ce7d","perf-guide/src/float-math/approx.md":"8c09032fa2d795a0c5db1775826c850d28eb2627846d0965c60ee72de63735ad","perf-guide/src/float-math/fma.md":"311076ba4b741d604a82e74b83a8d7e8c318fcbd7f64c4392d1cf5af95c60243","perf-guide/src/float-math/fp.md":"04153e775ab6e4f0d7837bcc515230d327b04edfa34c84ce9c9e10ebaeef2be8","perf-guide/src/float-math/svml.md":"0798873b8eedaeda5fed62dc91645b57c20775a02d3cd74d8bd06958f1516506","perf-guide/src/introduction.md":"9f5a19e9e6751f25d2daad39891a0cc600974527ec4c8305843f9618910671bd","perf-guide/src/prof/linux.md":"447731eb5de7d69166728fdbc5ecb0c0c9db678ea493b45a592d67dd002184c0","perf-guide/src/prof/mca.md":"f56d54f3d20e7aa4d32052186e8237b03d65971eb5d112802b442570ff11d344","perf-guide/src/prof/profiling.md":"8a650c0fd6ede0964789bb6577557eeef1d8226a896788602ce61528e260e43c","perf-guide/src/target-feature/attribute.md":"615f88dca0a707b6c416fa605435dd6e1fb5361cc639429cbf68cd87624bd78b","perf-guide/src/target-feature/features.md":"17077760ff24c006b606dd21889c53d87228f4311f3ba3a574f9afdeacd86165","perf-guide/src/target-feature/inlining.md":"7ed1d7068d8173a00d84c16cfe5871cd68b9f04f8d0cca2d01ebc84957ebf2f6","perf-guide/src/target-feature/practice.md":"c4b371842e0086df178488fec97f20def8f0c62ee588bcd25fd948b9b1fa227e","perf-guide/src/target-feature/runtime.md":"835425f5ee597fb3e51d36e725a81ebee29f4561231d19563cd4da81dbb1cfcb","perf-guide/src/target-feature/rustflags.md":"ab49712e9293a65d74d540ba4784fcb57ff1119ec05a575d895c071f1a620f64","perf-guide/src/vert-hor-ops.md":"c6211c0ee91e60552ec592d89d9d957eedc21dee3cbd89e1ad6765ea06a27471","readme.md":"585a8f0e16877fb9abb00cd17a175fcb9d7857840c6c61209f1827ffab095070","rustfmt.toml":"de6101d0670bad65fb3b337d56957d2a024e017e5ab146ec784d77312daaf8ff","src/api.rs":"331a3a4abb19cee2df5f2df4ad7c3e88b45e62cf23fdacfc9bbaa633dc5cf788","src/api/bit_manip.rs":"e68290ee679cc5abc9c73afbe635c1035f8cbfe849e5c751a1680e459244c39e","src/api/cast.rs":"03b94a3d316ac7b7be7068810044911e965e889a0ace7bae762749ca74a92747","src/api/cast/macros.rs":"b0a14d0c83ad2ebb7a275180f6d9e3f2bc312ba57a7d3d6c39fad4e0f20f9408","src/api/cast/v128.rs":"63e28c6a3edf1a7a635f51b8d3c6adbb1d46f884d92a196b3d4a6e743d809416","src/api/cast/v16.rs":"2a584eeb57fd47baad6f3533764301b04aaaac23702b7a8db12598ac02899262","src/api/cast/v256.rs":"b91c15ed8d1536ecd97b4eb79ff9d5aba0552cd9b6f0ea6435b05f2273e23b3a","src/api/cast/v32.rs":"62ec89fcce7fa7f28497ee5770adc8f81d2d3a6b2925b02f7dc06504c40e8f38","src/api/cast/v512.rs":"d855cb943ae7106e9599ef38e30a3afb1c6bd5433178baca54cb128fd9a7d143","src/api/cast/v64.rs":"fe0f7dfaf4fc0c0c1a78c96fcfcdfdc2a1e2845843b11aa797a0c6fb52a8f774","src/api/cmp.rs":"357c3a2a09c6d4611c32dd7fa95be2fae933d513e229026ec9b44451a77b884e","src/api/cmp/eq.rs":"60f70f355bae4cb5b17db53204cacc3890f70670611c17df638d4c04f7cc8075","src/api/cmp/ord.rs":"589f7234761c294fa5df8f525bc4acd5a47cdb602207d524a0d4e19804cd9695","src/api/cmp/partial_eq.rs":"3ed23d2a930b0f9750c3a5309da766b03dc4f9c4d375b42ad3c50fe732693d15","src/api/cmp/partial_ord.rs":"e16b11805c94048acd058c93994b5bc74bb187f8d7e3b86a87df60e1601467f9","src/api/cmp/vertical.rs":"de3d62f38eba817299aa16f1e1939954c9a447e316509397465c2830852ba053","src/api/default.rs":"b61f92fc0e33a2633b3375eb405beba480da071cde03df4d437d8a6058afcd97","src/api/fmt.rs":"67fb804bb86b6cd77cf8cd492b5733ce437071b66fe3297278b8a6552c325dda","src/api/fmt/binary.rs":"35cb5c266197d6224d598fb3d286e5fe48ef0c01ed356c2ff6fe9ba946f96a92","src/api/fmt/debug.rs":"aa18eea443bf353fea3db8b1a025132bbcaf91e747ecfa43b8d9fce9af395a0c","src/api/fmt/lower_hex.rs":"69d5be366631af309f214e8031c8c20267fcc27a695eac6f45c6bc1df72a67e6","src/api/fmt/octal.rs":"9eb11ba3d990213f3c7f1ec25edba7ce997cb1320e16d308c83498ba6b9bfbd9","src/api/fmt/upper_hex.rs":"a4637d085b7bb20e759ce58e08435b510a563ba3dd468af2b03560fdc5511562","src/api/from.rs":"2e599d8329cb05eaf06224cc441355c4b7b51254fc19256619333be8c149d444","src/api/from/from_array.rs":"4151593c7bba7455821fffa5b59867005a77c95d32f1f0cc3fd87294000157d9","src/api/from/from_vector.rs":"9764371aa9e6005aace74dea14f59e5611a095b7cf42707940924749282c52f0","src/api/hash.rs":"562cfa3f1d8eb9a733c035a3665a599c2f1e341ee820d8fbdd102a4398a441bc","src/api/into_bits.rs":"82297f0697d67b5a015e904e7e6e7b2a7066ba825bc54b94b4ff3e22d7a1eefb","src/api/into_bits/arch_specific.rs":"1f925390b0ce7132587d95f2419c6e2ad3e1a9d17eb1d9c120a1c1c4bdf4277e","src/api/into_bits/macros.rs":"d762406de25aedff88d460dec7a80dc8e825a2a419d53218ce007efa6a1d3e04","src/api/into_bits/v128.rs":"ecdc5893664c71d7ab1ff3697c3fbe490d20d8748b9b76881d05e7625e40d74c","src/api/into_bits/v16.rs":"5459ec7dad1ad7bd30dc7e48374580b993abf23701d9c3cb22203fa0a9aabb6d","src/api/into_bits/v256.rs":"90ea351da0380ead1bf0f63b620afd40d01d638d09f7e7be31840bd2c1d9c663","src/api/into_bits/v32.rs":"ee1dc5a430050e16f51154b5fe85b1536f5feddf2ea23dd1d3859b67c4afc6fc","src/api/into_bits/v512.rs":"f72098ed1c9a23944f3d01abaf5e0f2d0e81d35a06fdadd2183e896d41b59867","src/api/into_bits/v64.rs":"6394462facdfe7827349c742b7801f1291e75a720dfb8c0b52100df46f371c98","src/api/math.rs":"8b2a2fc651917a850539f993aa0b9e5bf4da67b11685285b8de8cdca311719ec","src/api/math/float.rs":"61d2794d68262a1090ae473bd30793b5f65cf732f32a6694a3af2ce5d9225616","src/api/math/float/abs.rs":"5b6b2701e2e11135b7ce58a05052ea8120e10e4702c95d046b9d21b827b26bf8","src/api/math/float/consts.rs":"78acba000d3fa527111300b6327c1932de9c4c1e02d4174e1a5615c01463d38c","src/api/math/float/cos.rs":"4c2dd7173728ef189314f1576c9486e03be21b7da98843b2f9011282a7979e31","src/api/math/float/exp.rs":"7c6d5f1e304f498a01cfa23b92380c815d7da0ad94eae3483783bc377d287eef","src/api/math/float/ln.rs":"54c7583f3df793b39ff57534fade27b41bb992439e5dc178252f5ca3190a3e54","src/api/math/float/mul_add.rs":"62cac77660d20159276d4c9ef066eb90c81cbddb808e8e157182c607625ad2eb","src/api/math/float/mul_adde.rs":"bae056ee9f3a70df39ec3c3b2f6437c65303888a7b843ef1a5bcf1f5aca0e602","src/api/math/float/powf.rs":"9ddb938984b36d39d82a82f862f80df8f7fb013f1d222d45698d41d88472f568","src/api/math/float/recpre.rs":"589225794ff1dbf31158dff660e6d4509ecc8befbb57c633900dea5ac0b840d6","src/api/math/float/rsqrte.rs":"a32abdcc318d7ccc8448231f54d75b884b7cbeb03a7d595713ab6243036f4dbf","src/api/math/float/sin.rs":"cbd3622b7df74f19691743001c8cf747a201f8977ad90542fee915f37dcd1e49","src/api/math/float/sqrt.rs":"0c66d5d63fb08e4d99c6b82a8828e41173aff1ac9fa1a2764a11fac217ccf2ac","src/api/math/float/sqrte.rs":"731e1c9f321b662accdd27dacb3aac2e8043b7aecb2f2161dde733bd9f025362","src/api/minimal.rs":"1f22bcc528555444e76de569ec0ae2029b9ae9d04805efeafa93369c8098036b","src/api/minimal/iuf.rs":"c501a6696950cf5e521765f178de548af64fdfb6e10d026616d09fab93ca2d17","src/api/minimal/mask.rs":"42e415f536c5193d0218f5a754b34b87fd7c971bff068009f958712166ff056d","src/api/minimal/ptr.rs":"a9ee482d1dd1c956fb8f3f179e6e620b1de4e9d713961461d4c6923a4ef2e67c","src/api/ops.rs":"3e273b277a0f3019d42c3c59ca94a5afd4885d5ae6d2182e5089bbeec9de42ee","src/api/ops/scalar_arithmetic.rs":"d2d5ad897a59dd0787544f927e0e7ca4072c3e58b0f4a2324083312b0d5a21d7","src/api/ops/scalar_bitwise.rs":"482204e459ca6be79568e1c9f70adbe2d2151412ddf122fb2161be8ebb51c40c","src/api/ops/scalar_mask_bitwise.rs":"c250f52042e37b22d57256c80d4604104cfd2fbe2a2e127c676267270ca5d350","src/api/ops/scalar_shifts.rs":"987f8fdebeedc16e3d77c1b732e7826ef70633c541d16dfa290845d5c6289150","src/api/ops/vector_arithmetic.rs":"ddca15d09ddeef502c2ed66117a62300ca65d87e959e8b622d767bdf1c307910","src/api/ops/vector_bitwise.rs":"b3968f7005b649edcc22a54e2379b14d5ee19045f2e784029805781ae043b5ee","src/api/ops/vector_float_min_max.rs":"f5155dce75219f4ba11275b1f295d2fdcddd49d174a6f1fb2ace7ea42813ce41","src/api/ops/vector_int_min_max.rs":"a378789c6ff9b32a51fbd0a97ffd36ed102cd1fe6a067d2b02017c1df342def6","src/api/ops/vector_mask_bitwise.rs":"5052d18517d765415d40327e6e8e55a312daaca0a5e2aec959bfa54b1675f9c8","src/api/ops/vector_neg.rs":"5c62f6b0221983cdbd23cd0a3af3672e6ba1255f0dfe8b19aae6fbd6503e231b","src/api/ops/vector_rotates.rs":"03cbe8a400fd7c688e4ee771a990a6754f2031b1a59b19ae81158b21471167e5","src/api/ops/vector_shifts.rs":"9bf69d0087268f61009e39aea52e03a90f378910206b6a28e8393178b6a5d0e0","src/api/ptr.rs":"8a793251bed6130dcfb2f1519ceaa18b751bbb15875928d0fb6deb5a5e07523a","src/api/ptr/gather_scatter.rs":"9ddd960365e050674b25b2fd3116e24d94669b4375d74e71c03e3f1469576066","src/api/reductions.rs":"ae5baca81352ecd44526d6c30c0a1feeda475ec73ddd3c3ec6b14e944e5448ee","src/api/reductions/bitwise.rs":"8bf910ae226188bd15fc7e125f058cd2566b6186fcd0cd8fd020f352c39ce139","src/api/reductions/float_arithmetic.rs":"e58c8c87806a95df2b2b5b48ac5991036df024096d9d7c171a480fe9282896a4","src/api/reductions/integer_arithmetic.rs":"47471da1c5f859489680bb5d34ced3d3aa20081c16053a3af121a4496fcb57bf","src/api/reductions/mask.rs":"db83327a950e33a317f37fd33ca4e20c347fb415975ec024f3e23da8509425af","src/api/reductions/min_max.rs":"f27be3aa28e1c1f46de7890198db6e12f00c207085e89ef2de7e57ee443cdb98","src/api/select.rs":"a98e2ccf9fc6bdeed32d337c8675bc96c2fbe2cc34fbf149ad6047fb8e749774","src/api/shuffle.rs":"da58200790868c09659819322a489929a5b6e56c596ed07e6a44293ea02e7d09","src/api/shuffle1_dyn.rs":"bfea5a91905b31444e9ef7ca6eddb7a9606b7e22d3f71bb842eb2795a0346620","src/api/slice.rs":"ee87484e8af329547b9a5d4f2a69e8bed6ea10bbd96270d706083843d4eea2ac","src/api/slice/from_slice.rs":"4d4fe8a329c885fcb4fbcbedf99efb15a95296fe6b3f595056cc37037450d5ac","src/api/slice/write_to_slice.rs":"f5b23b2c4b91cfb26b713a9013a6c0da7f45eaefb79ba06dcbc27f3f23bda679","src/api/swap_bytes.rs":"4a6792a2e49a77475e1b237592b4b2804dbddb79c474331acd0dd71b36934259","src/codegen.rs":"c6eebc3d3665420aa6a2f317977e3c41a4f43e0550ac630cdbe8e4bbed5e2031","src/codegen/bit_manip.rs":"5559e095105a80003e0de35af1d19b0c65c9ab04eb743c7e01c5442d882eb34e","src/codegen/llvm.rs":"d1299c189abb17a6133f047574cffc7a6db4c1be37cb7d4785491cb5e8f8cf54","src/codegen/math.rs":"35f96e37a78fcf0cdb02146b7f27a45108fe06a37fc2a54d8851ce131a326178","src/codegen/math/float.rs":"dd86c0449e576c83b719700962ac017c332987fac08d91f2b7a2b1b883598170","src/codegen/math/float/abs.rs":"f56e2b4b8055ea861c1f5cbc6b6e1d8e7e5af163b62c13574ddee4e09513bfbc","src/codegen/math/float/cos.rs":"ef3b511a24d23045b310315e80348a9b7fedb576fc2de52d74290616a0abeb2a","src/codegen/math/float/cos_pi.rs":"4e7631a5d73dac21531e09ef1802d1180f8997509c2c8fa9f67f322194263a97","src/codegen/math/float/exp.rs":"61b691598c41b5622f24e4320c1bdd08701e612a516438bdddcc728fc3405c8c","src/codegen/math/float/ln.rs":"46b718b1ba8c9d99e1ad40f53d20dfde08a3063ca7bd2a9fdd6698e060da687e","src/codegen/math/float/macros.rs":"dd42135fff13f9aca4fd3a1a4e14c7e6c31aadc6d817d63b0d2fb9e62e062744","src/codegen/math/float/mul_add.rs":"a37bf764345d4b1714f97e83897b7cf0855fc2811704bcbc0012db91825339e1","src/codegen/math/float/mul_adde.rs":"c75702bfcb361de45964a93caf959a695ef2376bd069227600b8c6872665c755","src/codegen/math/float/powf.rs":"642346e982bc4c39203de0864d2149c4179cd7b21cf67a2951687932b4675872","src/codegen/math/float/sin.rs":"9d68164c90cdca6a85155040cdac42e27342ebe0b925273ef1593df721af4258","src/codegen/math/float/sin_cos_pi.rs":"9be02ad48585a1e8d99129382fbffbaed47852f15459256a708850b6b7a75405","src/codegen/math/float/sin_pi.rs":"9890347905b4d4a3c7341c3eb06406e46e60582bcf6960688bd727e5dadc6c57","src/codegen/math/float/sqrt.rs":"e3c60dcfb0c6d2fc62adabcc931b2d4040b83cab294dea36443fb4b89eb79e34","src/codegen/math/float/sqrte.rs":"f0f4ef9eb475ae41bcc7ec6a95ad744ba6b36925faa8b2c2814004396d196b63","src/codegen/pointer_sized_int.rs":"a70697169c28218b56fd2e8d5353f2e00671d1150d0c8cef77d613bdfacd84cb","src/codegen/reductions.rs":"645e2514746d01387ddd07f0aa4ffd8430cc9ab428d4fb13773ea319fa25dd95","src/codegen/reductions/mask.rs":"8f1afe6aabf096a3278e1fc3a30f736e04aa8b9ce96373cee22162d18cfe2702","src/codegen/reductions/mask/aarch64.rs":"cba6e17603d39795dcfe8339b6b7d8714c3e162a1f0a635979f037aa24fe4206","src/codegen/reductions/mask/arm.rs":"9447904818aa2c7c25d0963eead452a639a11ca7dbd6d21eedbfcaade07a0f33","src/codegen/reductions/mask/fallback.rs":"7a0ef9f7fd03ae318b495b95e121350cd61caffc5cc6ee17fabf130d5d933453","src/codegen/reductions/mask/fallback_impl.rs":"76547f396e55ef403327c77c314cf8db8c7a5c9b9819bfb925abeacf130249e5","src/codegen/reductions/mask/x86.rs":"14bd2c482071f2355beebcf7b7ecf950ff2dfcdb08c3ca50993092434a9de717","src/codegen/reductions/mask/x86/avx.rs":"b4913d87844c522903641cbbf10db4551addb1ce5e9e78278e21612fa65c733b","src/codegen/reductions/mask/x86/avx2.rs":"677aed3f056285285daa3adff8bc65e739630b4424defa6d9665e160f027507e","src/codegen/reductions/mask/x86/sse.rs":"226610b4ff88c676d5187114dd57b4a8800de6ce40884675e9198445b1ed0306","src/codegen/reductions/mask/x86/sse2.rs":"bc38e6c31cb4b3d62147eba6cac264e519e2a48e0f7ce9010cfa9ef0cf0ec9fd","src/codegen/shuffle.rs":"0abca97e92cdce49a58a39cc447eb09dc7d7715ef256c8dbd2181a186e61bb64","src/codegen/shuffle1_dyn.rs":"04523e9338133bdedb012dd076c2c564b79ce5593b0fc56d0fb6910e04190a81","src/codegen/swap_bytes.rs":"1d6cdc716eadddc92b4fd506b2445a821caa8dc00860447de09d7ebd69c2087f","src/codegen/v128.rs":"94226b31ec403d18d9d2fe06713f147c9c79e9b5f9105089088266313f843185","src/codegen/v16.rs":"ddec4ffb66b6f7aaffb9a1780c5ddba82557abd74f45073d335047e04cf74924","src/codegen/v256.rs":"6b63917f0444118d6b1595bff2045e59b97c4d24012bd575f69f1f0efc5a0241","src/codegen/v32.rs":"3477b3c5540aed86e61e2f5807dd31db947413cec9181c587d93ed6ec74f0eba","src/codegen/v512.rs":"5854f99d3aabc4cd42b28a20d9ce447756dc2ba024a409a69b6a8ae1f1842fc5","src/codegen/v64.rs":"e9e89caebfe63d10c0cbca61e4dfdba3b7e02ee0989170f80beed23237ddd950","src/codegen/vPtr.rs":"96d609a9eece4dcbbcc01ba0b8744d7f5958be12774176a2945bc676f4e6b5cb","src/codegen/vSize.rs":"eeee9858749aa82142b27bc120d1989bb74a6b82e1e4efbbeaccc9634dc9acfc","src/lib.rs":"1b5d419ff05ee0370d671810423ccc254708cc8d415c1dbac2a7a36be4bf63a8","src/masks.rs":"870f429967b2d7d5133f4d28d6c753fc5cef0570b27b29d4e966a066d22d2d0e","src/sealed.rs":"ff7f0324276408ae8249941cfa32c90b8835a54d750896b683efea857af19db2","src/testing.rs":"1d3a7862ef625e235a5734ad7204e68d350f902c0695182b1f08a0552432416e","src/testing/macros.rs":"6378856d7a40ba5ec5c7c0dad6327d79f0c77266921c24296d10aed6c68e9b98","src/testing/utils.rs":"d6fd5a5017f1f85d9d99585754f8f6ad06fc3d683b34083543e67a7cc6c1772c","src/v128.rs":"18fe263c4aa28cd06461c7070b0269f69f4a2e75749b8f142a83dfdfe4d22bf5","src/v16.rs":"e5c663c9fb3547eaeac78a5f7db9969f4d8b5ec96112bf2954602fff11f0aebd","src/v256.rs":"68732cd688ad12a56d8b4f8ddf279f77bdfe1be2943c7dc0c1b4f1a76798aa0f","src/v32.rs":"785b22a1ccb4a41bb53dfeb0670f624c0ce42e6cdf62d1747e3283777a1c70bd","src/v512.rs":"d1337bfe07f06a8f37f8e8fa7d4315b9307476ee435ad80dd5269eaed564fbfa","src/v64.rs":"3077468d65125b8f085e9454c8b2463a4d5225697464ba6a1300f8799528fd4b","src/vPtr.rs":"c9a53f41f466e17b6648a4ce390fd8f4d3a848d440eb8a9a803a11608d76eb05","src/vSize.rs":"5c46d3e8c3ee5863d9b6e37e681f871386e0efc254d6d84ba711edb529ce7b3c","tests/endianness.rs":"541a144be017e3dd7da7c8ea49d907dc02538245e8c5f3deb5bd43da92c929e1"},"package":null} +\ No newline at end of file +diff --git a/third_party/rust/proc-macro2/.cargo-checksum.json b/third_party/rust/proc-macro2/.cargo-checksum.json +index eeef4120af..e7849f2896 100644 +--- a/third_party/rust/proc-macro2/.cargo-checksum.json ++++ b/third_party/rust/proc-macro2/.cargo-checksum.json +@@ -1 +1 @@ +-{"files":{"Cargo.toml":"e2c1fc6ed317eeef8462fcd192f6b6389e1d84f0d7afeac78f12c23903deddf8","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"378f5840b258e2779c39418f3f2d7b2ba96f1c7917dd6be0713f88305dbda397","README.md":"362a2156f7645528061b6e8487a2eb0f32f1693012ed82ee57afa05c039bba0d","build.rs":"0cc6e2cb919ddbff59cf1d810283939f97a59f0037540c0f2ee3453237635ff8","src/fallback.rs":"5c6379a90735e27abcc40253b223158c6b1e5784f3850bc423335363e87ef038","src/lib.rs":"ae5251296ad3fcd8b600919a993fec0afd8b56da3e11fef6bc7265b273129936","src/strnom.rs":"37f7791f73f123817ad5403af1d4e2a0714be27401729a2d451bc80b1f26bac9","src/wrapper.rs":"81372e910604217a625aa71c47d43e65f4e008456eae93ac39325c9abf10701a","tests/features.rs":"a86deb8644992a4eb64d9fd493eff16f9cf9c5cb6ade3a634ce0c990cf87d559","tests/marker.rs":"c2652e3ae1dfcb94d2e6313b29712c5dcbd0fe62026913e67bb7cebd7560aade","tests/test.rs":"8c427be9cba1fa8d4a16647e53e3545e5863e29e2c0b311c93c9dd1399abf6a1"},"package":"90cf5f418035b98e655e9cdb225047638296b862b42411c4e45bb88d700f7fc0"} +\ No newline at end of file ++{"files":{"Cargo.toml":"302d447d62c8d091d6241cf62bdad607c0d4ed8ff9f43d9b254c9d99c253ee8e","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"378f5840b258e2779c39418f3f2d7b2ba96f1c7917dd6be0713f88305dbda397","README.md":"e1f9d4fc22cff2c049f166a403b41458632a94357890d31cf0e3ad83807fb430","build.rs":"a71283fbc495095eebbbf46753df3fe2c19505c745b508dea157f65796b64dd7","src/detection.rs":"9d25d896889e65330858f2d6f6223c1b98cd1dad189813ad4161ff189fbda2b8","src/fallback.rs":"b114e013695260f6066395c8712cea112ec2a386010397a80f15a60f8b986444","src/lib.rs":"7f528764a958587f007f0c2a330a6a414bae2c8e73d5ed9fb64ff1b42b1805b1","src/marker.rs":"87fce2d0357f5b7998b6d9dfb064f4a0cbc9dabb19e33d4b514a446243ebe2e8","src/parse.rs":"1d2253eacbd40eb3a2a933be2adcee356af922bdb48cc89ff266252a41fd98a1","src/wrapper.rs":"f52646ce1705c1f6265516f30d4c43297b5f529dd31fb91f4c806be89d5a4122","tests/comments.rs":"ea6cbe6f4c8852e6a0612893c7d4f2c144a2e6a134a6c3db641a320cbfc3c800","tests/features.rs":"a86deb8644992a4eb64d9fd493eff16f9cf9c5cb6ade3a634ce0c990cf87d559","tests/marker.rs":"652db9f25c69ffc65baa60cdca8f195aa2e254d4de0a9ddc85de4dc2470544b6","tests/test.rs":"5f30a704eeb2b9198b57f416d622da72d25cb9bf8d8b12e6d0e90aa2cb0e43fc","tests/test_fmt.rs":"745dfdc41d09c5308c221395eb43f2041f0a1413d2927a813bc2ad4554438fe2"},"package":"1e0704ee1a7e00d7bb417d0770ea303c1bccbabf0ef1667dae92b5967f5f8a71"} +\ No newline at end of file +diff --git a/third_party/rust/proc-macro2/Cargo.toml b/third_party/rust/proc-macro2/Cargo.toml +index 95d653633d..22150c516a 100644 +--- a/third_party/rust/proc-macro2/Cargo.toml ++++ b/third_party/rust/proc-macro2/Cargo.toml +@@ -13,21 +13,22 @@ + [package] + edition = "2018" + name = "proc-macro2" +-version = "1.0.5" +-authors = ["Alex Crichton <alex@alexcrichton.com>"] +-description = "A stable implementation of the upcoming new `proc_macro` API. Comes with an\noption, off by default, to also reimplement itself in terms of the upstream\nunstable API.\n" +-homepage = "https://github.com/alexcrichton/proc-macro2" ++version = "1.0.24" ++authors = ["Alex Crichton <alex@alexcrichton.com>", "David Tolnay <dtolnay@gmail.com>"] ++description = "A substitute implementation of the compiler's `proc_macro` API to decouple\ntoken-based libraries from the procedural macro use case.\n" + documentation = "https://docs.rs/proc-macro2" + readme = "README.md" + keywords = ["macros"] ++categories = ["development-tools::procedural-macro-helpers"] + license = "MIT OR Apache-2.0" + repository = "https://github.com/alexcrichton/proc-macro2" + [package.metadata.docs.rs] + rustc-args = ["--cfg", "procmacro2_semver_exempt"] + rustdoc-args = ["--cfg", "procmacro2_semver_exempt"] ++targets = ["x86_64-unknown-linux-gnu"] + +-[lib] +-name = "proc_macro2" ++[package.metadata.playground] ++features = ["span-locations"] + [dependencies.unicode-xid] + version = "0.2" + [dev-dependencies.quote] +@@ -39,5 +40,3 @@ default = ["proc-macro"] + nightly = [] + proc-macro = [] + span-locations = [] +-[badges.travis-ci] +-repository = "alexcrichton/proc-macro2" +diff --git a/third_party/rust/proc-macro2/README.md b/third_party/rust/proc-macro2/README.md +index 19b0c3b5f8..3d05e871a7 100644 +--- a/third_party/rust/proc-macro2/README.md ++++ b/third_party/rust/proc-macro2/README.md +@@ -1,6 +1,6 @@ + # proc-macro2 + +-[![Build Status](https://api.travis-ci.com/alexcrichton/proc-macro2.svg?branch=master)](https://travis-ci.com/alexcrichton/proc-macro2) ++[![Build Status](https://img.shields.io/github/workflow/status/alexcrichton/proc-macro2/build%20and%20test)](https://github.com/alexcrichton/proc-macro2/actions) + [![Latest Version](https://img.shields.io/crates/v/proc-macro2.svg)](https://crates.io/crates/proc-macro2) + [![Rust Documentation](https://img.shields.io/badge/api-rustdoc-blue.svg)](https://docs.rs/proc-macro2) + +diff --git a/third_party/rust/proc-macro2/build.rs b/third_party/rust/proc-macro2/build.rs +index deb9b92719..b247d874f6 100644 +--- a/third_party/rust/proc-macro2/build.rs ++++ b/third_party/rust/proc-macro2/build.rs +@@ -14,6 +14,10 @@ + // procmacro2_semver_exempt surface area is implemented by using the + // nightly-only proc_macro API. + // ++// "hygiene" ++// Enable Span::mixed_site() and non-dummy behavior of Span::resolved_at ++// and Span::located_at. Enabled on Rust 1.45+. ++// + // "proc_macro_span" + // Enable non-dummy behavior of Span::start and Span::end methods which + // requires an unstable compiler feature. Enabled when building with +@@ -57,6 +61,22 @@ fn main() { + println!("cargo:rustc-cfg=span_locations"); + } + ++ if version.minor < 32 { ++ println!("cargo:rustc-cfg=no_libprocmacro_unwind_safe"); ++ } ++ ++ if version.minor < 39 { ++ println!("cargo:rustc-cfg=no_bind_by_move_pattern_guard"); ++ } ++ ++ if version.minor >= 44 { ++ println!("cargo:rustc-cfg=lexerror_display"); ++ } ++ ++ if version.minor >= 45 { ++ println!("cargo:rustc-cfg=hygiene"); ++ } ++ + let target = env::var("TARGET").unwrap(); + if !enable_use_proc_macro(&target) { + return; +diff --git a/third_party/rust/proc-macro2/src/detection.rs b/third_party/rust/proc-macro2/src/detection.rs +new file mode 100644 +index 0000000000..c597bc99c6 +--- /dev/null ++++ b/third_party/rust/proc-macro2/src/detection.rs +@@ -0,0 +1,67 @@ ++use std::panic::{self, PanicInfo}; ++use std::sync::atomic::*; ++use std::sync::Once; ++ ++static WORKS: AtomicUsize = AtomicUsize::new(0); ++static INIT: Once = Once::new(); ++ ++pub(crate) fn inside_proc_macro() -> bool { ++ match WORKS.load(Ordering::SeqCst) { ++ 1 => return false, ++ 2 => return true, ++ _ => {} ++ } ++ ++ INIT.call_once(initialize); ++ inside_proc_macro() ++} ++ ++pub(crate) fn force_fallback() { ++ WORKS.store(1, Ordering::SeqCst); ++} ++ ++pub(crate) fn unforce_fallback() { ++ initialize(); ++} ++ ++// Swap in a null panic hook to avoid printing "thread panicked" to stderr, ++// then use catch_unwind to determine whether the compiler's proc_macro is ++// working. When proc-macro2 is used from outside of a procedural macro all ++// of the proc_macro crate's APIs currently panic. ++// ++// The Once is to prevent the possibility of this ordering: ++// ++// thread 1 calls take_hook, gets the user's original hook ++// thread 1 calls set_hook with the null hook ++// thread 2 calls take_hook, thinks null hook is the original hook ++// thread 2 calls set_hook with the null hook ++// thread 1 calls set_hook with the actual original hook ++// thread 2 calls set_hook with what it thinks is the original hook ++// ++// in which the user's hook has been lost. ++// ++// There is still a race condition where a panic in a different thread can ++// happen during the interval that the user's original panic hook is ++// unregistered such that their hook is incorrectly not called. This is ++// sufficiently unlikely and less bad than printing panic messages to stderr ++// on correct use of this crate. Maybe there is a libstd feature request ++// here. For now, if a user needs to guarantee that this failure mode does ++// not occur, they need to call e.g. `proc_macro2::Span::call_site()` from ++// the main thread before launching any other threads. ++fn initialize() { ++ type PanicHook = dyn Fn(&PanicInfo) + Sync + Send + 'static; ++ ++ let null_hook: Box<PanicHook> = Box::new(|_panic_info| { /* ignore */ }); ++ let sanity_check = &*null_hook as *const PanicHook; ++ let original_hook = panic::take_hook(); ++ panic::set_hook(null_hook); ++ ++ let works = panic::catch_unwind(proc_macro::Span::call_site).is_ok(); ++ WORKS.store(works as usize + 1, Ordering::SeqCst); ++ ++ let hopefully_null_hook = panic::take_hook(); ++ panic::set_hook(original_hook); ++ if sanity_check != &*hopefully_null_hook { ++ panic!("observed race condition in proc_macro2::inside_proc_macro"); ++ } ++} +diff --git a/third_party/rust/proc-macro2/src/fallback.rs b/third_party/rust/proc-macro2/src/fallback.rs +index fe582b3b5f..8900c5ff0f 100644 +--- a/third_party/rust/proc-macro2/src/fallback.rs ++++ b/third_party/rust/proc-macro2/src/fallback.rs +@@ -1,27 +1,41 @@ ++use crate::parse::{token_stream, Cursor}; ++use crate::{Delimiter, Spacing, TokenTree}; + #[cfg(span_locations)] + use std::cell::RefCell; + #[cfg(span_locations)] + use std::cmp; +-use std::fmt; +-use std::iter; ++use std::fmt::{self, Debug, Display}; ++use std::iter::FromIterator; ++use std::mem; + use std::ops::RangeBounds; + #[cfg(procmacro2_semver_exempt)] + use std::path::Path; + use std::path::PathBuf; + use std::str::FromStr; + use std::vec; +- +-use crate::strnom::{block_comment, skip_whitespace, whitespace, word_break, Cursor, PResult}; +-use crate::{Delimiter, Punct, Spacing, TokenTree}; + use unicode_xid::UnicodeXID; + ++/// Force use of proc-macro2's fallback implementation of the API for now, even ++/// if the compiler's implementation is available. ++pub fn force() { ++ #[cfg(wrap_proc_macro)] ++ crate::detection::force_fallback(); ++} ++ ++/// Resume using the compiler's implementation of the proc macro API if it is ++/// available. ++pub fn unforce() { ++ #[cfg(wrap_proc_macro)] ++ crate::detection::unforce_fallback(); ++} ++ + #[derive(Clone)] +-pub struct TokenStream { +- inner: Vec<TokenTree>, ++pub(crate) struct TokenStream { ++ pub(crate) inner: Vec<TokenTree>, + } + + #[derive(Debug)] +-pub struct LexError; ++pub(crate) struct LexError; + + impl TokenStream { + pub fn new() -> TokenStream { +@@ -31,6 +45,72 @@ impl TokenStream { + pub fn is_empty(&self) -> bool { + self.inner.len() == 0 + } ++ ++ fn take_inner(&mut self) -> Vec<TokenTree> { ++ mem::replace(&mut self.inner, Vec::new()) ++ } ++ ++ fn push_token(&mut self, token: TokenTree) { ++ // https://github.com/alexcrichton/proc-macro2/issues/235 ++ match token { ++ #[cfg(not(no_bind_by_move_pattern_guard))] ++ TokenTree::Literal(crate::Literal { ++ #[cfg(wrap_proc_macro)] ++ inner: crate::imp::Literal::Fallback(literal), ++ #[cfg(not(wrap_proc_macro))] ++ inner: literal, ++ .. ++ }) if literal.text.starts_with('-') => { ++ push_negative_literal(self, literal); ++ } ++ #[cfg(no_bind_by_move_pattern_guard)] ++ TokenTree::Literal(crate::Literal { ++ #[cfg(wrap_proc_macro)] ++ inner: crate::imp::Literal::Fallback(literal), ++ #[cfg(not(wrap_proc_macro))] ++ inner: literal, ++ .. ++ }) => { ++ if literal.text.starts_with('-') { ++ push_negative_literal(self, literal); ++ } else { ++ self.inner ++ .push(TokenTree::Literal(crate::Literal::_new_stable(literal))); ++ } ++ } ++ _ => self.inner.push(token), ++ } ++ ++ #[cold] ++ fn push_negative_literal(stream: &mut TokenStream, mut literal: Literal) { ++ literal.text.remove(0); ++ let mut punct = crate::Punct::new('-', Spacing::Alone); ++ punct.set_span(crate::Span::_new_stable(literal.span)); ++ stream.inner.push(TokenTree::Punct(punct)); ++ stream ++ .inner ++ .push(TokenTree::Literal(crate::Literal::_new_stable(literal))); ++ } ++ } ++} ++ ++// Nonrecursive to prevent stack overflow. ++impl Drop for TokenStream { ++ fn drop(&mut self) { ++ while let Some(token) = self.inner.pop() { ++ let group = match token { ++ TokenTree::Group(group) => group.inner, ++ _ => continue, ++ }; ++ #[cfg(wrap_proc_macro)] ++ let group = match group { ++ crate::imp::Group::Fallback(group) => group, ++ _ => continue, ++ }; ++ let mut group = group; ++ self.inner.extend(group.stream.take_inner()); ++ } ++ } + } + + #[cfg(span_locations)] +@@ -59,20 +139,22 @@ impl FromStr for TokenStream { + // Create a dummy file & add it to the source map + let cursor = get_cursor(src); + +- match token_stream(cursor) { +- Ok((input, output)) => { +- if skip_whitespace(input).len() != 0 { +- Err(LexError) +- } else { +- Ok(output) +- } +- } +- Err(LexError) => Err(LexError), ++ let (rest, tokens) = token_stream(cursor)?; ++ if rest.is_empty() { ++ Ok(tokens) ++ } else { ++ Err(LexError) + } + } + } + +-impl fmt::Display for TokenStream { ++impl Display for LexError { ++ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { ++ f.write_str("cannot parse string into token stream") ++ } ++} ++ ++impl Display for TokenStream { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + let mut joint = false; + for (i, tt) in self.inner.iter().enumerate() { +@@ -80,37 +162,22 @@ impl fmt::Display for TokenStream { + write!(f, " ")?; + } + joint = false; +- match *tt { +- TokenTree::Group(ref tt) => { +- let (start, end) = match tt.delimiter() { +- Delimiter::Parenthesis => ("(", ")"), +- Delimiter::Brace => ("{", "}"), +- Delimiter::Bracket => ("[", "]"), +- Delimiter::None => ("", ""), +- }; +- if tt.stream().into_iter().next().is_none() { +- write!(f, "{} {}", start, end)? +- } else { +- write!(f, "{} {} {}", start, tt.stream(), end)? +- } +- } +- TokenTree::Ident(ref tt) => write!(f, "{}", tt)?, +- TokenTree::Punct(ref tt) => { +- write!(f, "{}", tt.as_char())?; +- match tt.spacing() { +- Spacing::Alone => {} +- Spacing::Joint => joint = true, +- } ++ match tt { ++ TokenTree::Group(tt) => Display::fmt(tt, f), ++ TokenTree::Ident(tt) => Display::fmt(tt, f), ++ TokenTree::Punct(tt) => { ++ joint = tt.spacing() == Spacing::Joint; ++ Display::fmt(tt, f) + } +- TokenTree::Literal(ref tt) => write!(f, "{}", tt)?, +- } ++ TokenTree::Literal(tt) => Display::fmt(tt, f), ++ }? + } + + Ok(()) + } + } + +-impl fmt::Debug for TokenStream { ++impl Debug for TokenStream { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + f.write_str("TokenStream ")?; + f.debug_list().entries(self.clone()).finish() +@@ -139,28 +206,26 @@ impl From<TokenStream> for proc_macro::TokenStream { + + impl From<TokenTree> for TokenStream { + fn from(tree: TokenTree) -> TokenStream { +- TokenStream { inner: vec![tree] } ++ let mut stream = TokenStream::new(); ++ stream.push_token(tree); ++ stream + } + } + +-impl iter::FromIterator<TokenTree> for TokenStream { +- fn from_iter<I: IntoIterator<Item = TokenTree>>(streams: I) -> Self { +- let mut v = Vec::new(); +- +- for token in streams.into_iter() { +- v.push(token); +- } +- +- TokenStream { inner: v } ++impl FromIterator<TokenTree> for TokenStream { ++ fn from_iter<I: IntoIterator<Item = TokenTree>>(tokens: I) -> Self { ++ let mut stream = TokenStream::new(); ++ stream.extend(tokens); ++ stream + } + } + +-impl iter::FromIterator<TokenStream> for TokenStream { ++impl FromIterator<TokenStream> for TokenStream { + fn from_iter<I: IntoIterator<Item = TokenStream>>(streams: I) -> Self { + let mut v = Vec::new(); + +- for stream in streams.into_iter() { +- v.extend(stream.inner); ++ for mut stream in streams { ++ v.extend(stream.take_inner()); + } + + TokenStream { inner: v } +@@ -168,31 +233,30 @@ impl iter::FromIterator<TokenStream> for TokenStream { + } + + impl Extend<TokenTree> for TokenStream { +- fn extend<I: IntoIterator<Item = TokenTree>>(&mut self, streams: I) { +- self.inner.extend(streams); ++ fn extend<I: IntoIterator<Item = TokenTree>>(&mut self, tokens: I) { ++ tokens.into_iter().for_each(|token| self.push_token(token)); + } + } + + impl Extend<TokenStream> for TokenStream { + fn extend<I: IntoIterator<Item = TokenStream>>(&mut self, streams: I) { +- self.inner +- .extend(streams.into_iter().flat_map(|stream| stream)); ++ self.inner.extend(streams.into_iter().flatten()); + } + } + +-pub type TokenTreeIter = vec::IntoIter<TokenTree>; ++pub(crate) type TokenTreeIter = vec::IntoIter<TokenTree>; + + impl IntoIterator for TokenStream { + type Item = TokenTree; + type IntoIter = TokenTreeIter; + +- fn into_iter(self) -> TokenTreeIter { +- self.inner.into_iter() ++ fn into_iter(mut self) -> TokenTreeIter { ++ self.take_inner().into_iter() + } + } + + #[derive(Clone, PartialEq, Eq)] +-pub struct SourceFile { ++pub(crate) struct SourceFile { + path: PathBuf, + } + +@@ -208,7 +272,7 @@ impl SourceFile { + } + } + +-impl fmt::Debug for SourceFile { ++impl Debug for SourceFile { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + f.debug_struct("SourceFile") + .field("path", &self.path()) +@@ -218,7 +282,7 @@ impl fmt::Debug for SourceFile { + } + + #[derive(Clone, Copy, Debug, PartialEq, Eq)] +-pub struct LineColumn { ++pub(crate) struct LineColumn { + pub line: usize, + pub column: usize, + } +@@ -228,23 +292,11 @@ thread_local! { + static SOURCE_MAP: RefCell<SourceMap> = RefCell::new(SourceMap { + // NOTE: We start with a single dummy file which all call_site() and + // def_site() spans reference. +- files: vec![{ ++ files: vec![FileInfo { + #[cfg(procmacro2_semver_exempt)] +- { +- FileInfo { +- name: "<unspecified>".to_owned(), +- span: Span { lo: 0, hi: 0 }, +- lines: vec![0], +- } +- } +- +- #[cfg(not(procmacro2_semver_exempt))] +- { +- FileInfo { +- span: Span { lo: 0, hi: 0 }, +- lines: vec![0], +- } +- } ++ name: "<unspecified>".to_owned(), ++ span: Span { lo: 0, hi: 0 }, ++ lines: vec![0], + }], + }); + } +@@ -282,16 +334,21 @@ impl FileInfo { + } + } + +-/// Computesthe offsets of each line in the given source string. ++/// Computes the offsets of each line in the given source string ++/// and the total number of characters + #[cfg(span_locations)] +-fn lines_offsets(s: &str) -> Vec<usize> { ++fn lines_offsets(s: &str) -> (usize, Vec<usize>) { + let mut lines = vec![0]; +- let mut prev = 0; +- while let Some(len) = s[prev..].find('\n') { +- prev += len + 1; +- lines.push(prev); ++ let mut total = 0; ++ ++ for ch in s.chars() { ++ total += 1; ++ if ch == '\n' { ++ lines.push(total); ++ } + } +- lines ++ ++ (total, lines) + } + + #[cfg(span_locations)] +@@ -310,23 +367,22 @@ impl SourceMap { + } + + fn add_file(&mut self, name: &str, src: &str) -> Span { +- let lines = lines_offsets(src); ++ let (len, lines) = lines_offsets(src); + let lo = self.next_start_pos(); + // XXX(nika): Shouild we bother doing a checked cast or checked add here? + let span = Span { + lo, +- hi: lo + (src.len() as u32), ++ hi: lo + (len as u32), + }; + +- #[cfg(procmacro2_semver_exempt)] + self.files.push(FileInfo { ++ #[cfg(procmacro2_semver_exempt)] + name: name.to_owned(), + span, + lines, + }); + + #[cfg(not(procmacro2_semver_exempt))] +- self.files.push(FileInfo { span, lines }); + let _ = name; + + span +@@ -343,11 +399,11 @@ impl SourceMap { + } + + #[derive(Clone, Copy, PartialEq, Eq)] +-pub struct Span { ++pub(crate) struct Span { + #[cfg(span_locations)] +- lo: u32, ++ pub(crate) lo: u32, + #[cfg(span_locations)] +- hi: u32, ++ pub(crate) hi: u32, + } + + impl Span { +@@ -361,12 +417,16 @@ impl Span { + Span { lo: 0, hi: 0 } + } + ++ #[cfg(hygiene)] ++ pub fn mixed_site() -> Span { ++ Span::call_site() ++ } ++ + #[cfg(procmacro2_semver_exempt)] + pub fn def_site() -> Span { + Span::call_site() + } + +- #[cfg(procmacro2_semver_exempt)] + pub fn resolved_at(&self, _other: Span) -> Span { + // Stable spans consist only of line/column information, so + // `resolved_at` and `located_at` only select which span the +@@ -374,7 +434,6 @@ impl Span { + *self + } + +- #[cfg(procmacro2_semver_exempt)] + pub fn located_at(&self, other: Span) -> Span { + other + } +@@ -427,26 +486,59 @@ impl Span { + }) + }) + } ++ ++ #[cfg(not(span_locations))] ++ fn first_byte(self) -> Self { ++ self ++ } ++ ++ #[cfg(span_locations)] ++ fn first_byte(self) -> Self { ++ Span { ++ lo: self.lo, ++ hi: cmp::min(self.lo.saturating_add(1), self.hi), ++ } ++ } ++ ++ #[cfg(not(span_locations))] ++ fn last_byte(self) -> Self { ++ self ++ } ++ ++ #[cfg(span_locations)] ++ fn last_byte(self) -> Self { ++ Span { ++ lo: cmp::max(self.hi.saturating_sub(1), self.lo), ++ hi: self.hi, ++ } ++ } + } + +-impl fmt::Debug for Span { ++impl Debug for Span { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- #[cfg(procmacro2_semver_exempt)] ++ #[cfg(span_locations)] + return write!(f, "bytes({}..{})", self.lo, self.hi); + +- #[cfg(not(procmacro2_semver_exempt))] ++ #[cfg(not(span_locations))] + write!(f, "Span") + } + } + +-pub fn debug_span_field_if_nontrivial(debug: &mut fmt::DebugStruct, span: Span) { +- if cfg!(procmacro2_semver_exempt) { ++pub(crate) fn debug_span_field_if_nontrivial(debug: &mut fmt::DebugStruct, span: Span) { ++ #[cfg(span_locations)] ++ { ++ if span.lo == 0 && span.hi == 0 { ++ return; ++ } ++ } ++ ++ if cfg!(span_locations) { + debug.field("span", &span); + } + } + + #[derive(Clone)] +-pub struct Group { ++pub(crate) struct Group { + delimiter: Delimiter, + stream: TokenStream, + span: Span, +@@ -474,11 +566,11 @@ impl Group { + } + + pub fn span_open(&self) -> Span { +- self.span ++ self.span.first_byte() + } + + pub fn span_close(&self) -> Span { +- self.span ++ self.span.last_byte() + } + + pub fn set_span(&mut self, span: Span) { +@@ -486,36 +578,45 @@ impl Group { + } + } + +-impl fmt::Display for Group { ++impl Display for Group { ++ // We attempt to match libproc_macro's formatting. ++ // Empty parens: () ++ // Nonempty parens: (...) ++ // Empty brackets: [] ++ // Nonempty brackets: [...] ++ // Empty braces: { } ++ // Nonempty braces: { ... } + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- let (left, right) = match self.delimiter { ++ let (open, close) = match self.delimiter { + Delimiter::Parenthesis => ("(", ")"), +- Delimiter::Brace => ("{", "}"), ++ Delimiter::Brace => ("{ ", "}"), + Delimiter::Bracket => ("[", "]"), + Delimiter::None => ("", ""), + }; + +- f.write_str(left)?; +- self.stream.fmt(f)?; +- f.write_str(right)?; ++ f.write_str(open)?; ++ Display::fmt(&self.stream, f)?; ++ if self.delimiter == Delimiter::Brace && !self.stream.inner.is_empty() { ++ f.write_str(" ")?; ++ } ++ f.write_str(close)?; + + Ok(()) + } + } + +-impl fmt::Debug for Group { ++impl Debug for Group { + fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result { + let mut debug = fmt.debug_struct("Group"); + debug.field("delimiter", &self.delimiter); + debug.field("stream", &self.stream); +- #[cfg(procmacro2_semver_exempt)] +- debug.field("span", &self.span); ++ debug_span_field_if_nontrivial(&mut debug, self.span); + debug.finish() + } + } + + #[derive(Clone)] +-pub struct Ident { ++pub(crate) struct Ident { + sym: String, + span: Span, + raw: bool, +@@ -549,16 +650,14 @@ impl Ident { + } + } + +-#[inline] +-fn is_ident_start(c: char) -> bool { ++pub(crate) fn is_ident_start(c: char) -> bool { + ('a' <= c && c <= 'z') + || ('A' <= c && c <= 'Z') + || c == '_' + || (c > '\x7f' && UnicodeXID::is_xid_start(c)) + } + +-#[inline] +-fn is_ident_continue(c: char) -> bool { ++pub(crate) fn is_ident_continue(c: char) -> bool { + ('a' <= c && c <= 'z') + || ('A' <= c && c <= 'Z') + || c == '_' +@@ -615,18 +714,18 @@ where + } + } + +-impl fmt::Display for Ident { ++impl Display for Ident { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + if self.raw { +- "r#".fmt(f)?; ++ f.write_str("r#")?; + } +- self.sym.fmt(f) ++ Display::fmt(&self.sym, f) + } + } + +-impl fmt::Debug for Ident { ++impl Debug for Ident { + // Ident(proc_macro), Ident(r#union) +- #[cfg(not(procmacro2_semver_exempt))] ++ #[cfg(not(span_locations))] + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + let mut debug = f.debug_tuple("Ident"); + debug.field(&format_args!("{}", self)); +@@ -637,17 +736,17 @@ impl fmt::Debug for Ident { + // sym: proc_macro, + // span: bytes(128..138) + // } +- #[cfg(procmacro2_semver_exempt)] ++ #[cfg(span_locations)] + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + let mut debug = f.debug_struct("Ident"); + debug.field("sym", &format_args!("{}", self)); +- debug.field("span", &self.span); ++ debug_span_field_if_nontrivial(&mut debug, self.span); + debug.finish() + } + } + + #[derive(Clone)] +-pub struct Literal { ++pub(crate) struct Literal { + text: String, + span: Span, + } +@@ -669,7 +768,7 @@ macro_rules! unsuffixed_numbers { + } + + impl Literal { +- fn _new(text: String) -> Literal { ++ pub(crate) fn _new(text: String) -> Literal { + Literal { + text, + span: Span::call_site(), +@@ -711,7 +810,7 @@ impl Literal { + + pub fn f32_unsuffixed(f: f32) -> Literal { + let mut s = f.to_string(); +- if !s.contains(".") { ++ if !s.contains('.') { + s.push_str(".0"); + } + Literal::_new(s) +@@ -719,7 +818,7 @@ impl Literal { + + pub fn f64_unsuffixed(f: f64) -> Literal { + let mut s = f.to_string(); +- if !s.contains(".") { ++ if !s.contains('.') { + s.push_str(".0"); + } + Literal::_new(s) +@@ -730,10 +829,10 @@ impl Literal { + text.push('"'); + for c in t.chars() { + if c == '\'' { +- // escape_default turns this into "\'" which is unnecessary. ++ // escape_debug turns this into "\'" which is unnecessary. + text.push(c); + } else { +- text.extend(c.escape_default()); ++ text.extend(c.escape_debug()); + } + } + text.push('"'); +@@ -744,10 +843,10 @@ impl Literal { + let mut text = String::new(); + text.push('\''); + if t == '"' { +- // escape_default turns this into '\"' which is unnecessary. ++ // escape_debug turns this into '\"' which is unnecessary. + text.push(t); + } else { +- text.extend(t.escape_default()); ++ text.extend(t.escape_debug()); + } + text.push('\''); + Literal::_new(text) +@@ -756,6 +855,7 @@ impl Literal { + pub fn byte_string(bytes: &[u8]) -> Literal { + let mut escaped = "b\"".to_string(); + for b in bytes { ++ #[allow(clippy::match_overlapping_arm)] + match *b { + b'\0' => escaped.push_str(r"\0"), + b'\t' => escaped.push_str(r"\t"), +@@ -784,651 +884,17 @@ impl Literal { + } + } + +-impl fmt::Display for Literal { ++impl Display for Literal { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.text.fmt(f) ++ Display::fmt(&self.text, f) + } + } + +-impl fmt::Debug for Literal { ++impl Debug for Literal { + fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result { + let mut debug = fmt.debug_struct("Literal"); + debug.field("lit", &format_args!("{}", self.text)); +- #[cfg(procmacro2_semver_exempt)] +- debug.field("span", &self.span); ++ debug_span_field_if_nontrivial(&mut debug, self.span); + debug.finish() + } + } +- +-fn token_stream(mut input: Cursor) -> PResult<TokenStream> { +- let mut trees = Vec::new(); +- loop { +- let input_no_ws = skip_whitespace(input); +- if input_no_ws.rest.len() == 0 { +- break; +- } +- if let Ok((a, tokens)) = doc_comment(input_no_ws) { +- input = a; +- trees.extend(tokens); +- continue; +- } +- +- let (a, tt) = match token_tree(input_no_ws) { +- Ok(p) => p, +- Err(_) => break, +- }; +- trees.push(tt); +- input = a; +- } +- Ok((input, TokenStream { inner: trees })) +-} +- +-#[cfg(not(span_locations))] +-fn spanned<'a, T>( +- input: Cursor<'a>, +- f: fn(Cursor<'a>) -> PResult<'a, T>, +-) -> PResult<'a, (T, crate::Span)> { +- let (a, b) = f(skip_whitespace(input))?; +- Ok((a, ((b, crate::Span::_new_stable(Span::call_site()))))) +-} +- +-#[cfg(span_locations)] +-fn spanned<'a, T>( +- input: Cursor<'a>, +- f: fn(Cursor<'a>) -> PResult<'a, T>, +-) -> PResult<'a, (T, crate::Span)> { +- let input = skip_whitespace(input); +- let lo = input.off; +- let (a, b) = f(input)?; +- let hi = a.off; +- let span = crate::Span::_new_stable(Span { lo, hi }); +- Ok((a, (b, span))) +-} +- +-fn token_tree(input: Cursor) -> PResult<TokenTree> { +- let (rest, (mut tt, span)) = spanned(input, token_kind)?; +- tt.set_span(span); +- Ok((rest, tt)) +-} +- +-named!(token_kind -> TokenTree, alt!( +- map!(group, |g| TokenTree::Group(crate::Group::_new_stable(g))) +- | +- map!(literal, |l| TokenTree::Literal(crate::Literal::_new_stable(l))) // must be before symbol +- | +- map!(op, TokenTree::Punct) +- | +- symbol_leading_ws +-)); +- +-named!(group -> Group, alt!( +- delimited!( +- punct!("("), +- token_stream, +- punct!(")") +- ) => { |ts| Group::new(Delimiter::Parenthesis, ts) } +- | +- delimited!( +- punct!("["), +- token_stream, +- punct!("]") +- ) => { |ts| Group::new(Delimiter::Bracket, ts) } +- | +- delimited!( +- punct!("{"), +- token_stream, +- punct!("}") +- ) => { |ts| Group::new(Delimiter::Brace, ts) } +-)); +- +-fn symbol_leading_ws(input: Cursor) -> PResult<TokenTree> { +- symbol(skip_whitespace(input)) +-} +- +-fn symbol(input: Cursor) -> PResult<TokenTree> { +- let raw = input.starts_with("r#"); +- let rest = input.advance((raw as usize) << 1); +- +- let (rest, sym) = symbol_not_raw(rest)?; +- +- if !raw { +- let ident = crate::Ident::new(sym, crate::Span::call_site()); +- return Ok((rest, ident.into())); +- } +- +- if sym == "_" { +- return Err(LexError); +- } +- +- let ident = crate::Ident::_new_raw(sym, crate::Span::call_site()); +- Ok((rest, ident.into())) +-} +- +-fn symbol_not_raw(input: Cursor) -> PResult<&str> { +- let mut chars = input.char_indices(); +- +- match chars.next() { +- Some((_, ch)) if is_ident_start(ch) => {} +- _ => return Err(LexError), +- } +- +- let mut end = input.len(); +- for (i, ch) in chars { +- if !is_ident_continue(ch) { +- end = i; +- break; +- } +- } +- +- Ok((input.advance(end), &input.rest[..end])) +-} +- +-fn literal(input: Cursor) -> PResult<Literal> { +- let input_no_ws = skip_whitespace(input); +- +- match literal_nocapture(input_no_ws) { +- Ok((a, ())) => { +- let start = input.len() - input_no_ws.len(); +- let len = input_no_ws.len() - a.len(); +- let end = start + len; +- Ok((a, Literal::_new(input.rest[start..end].to_string()))) +- } +- Err(LexError) => Err(LexError), +- } +-} +- +-named!(literal_nocapture -> (), alt!( +- string +- | +- byte_string +- | +- byte +- | +- character +- | +- float +- | +- int +-)); +- +-named!(string -> (), alt!( +- quoted_string +- | +- preceded!( +- punct!("r"), +- raw_string +- ) => { |_| () } +-)); +- +-named!(quoted_string -> (), do_parse!( +- punct!("\"") >> +- cooked_string >> +- tag!("\"") >> +- option!(symbol_not_raw) >> +- (()) +-)); +- +-fn cooked_string(input: Cursor) -> PResult<()> { +- let mut chars = input.char_indices().peekable(); +- while let Some((byte_offset, ch)) = chars.next() { +- match ch { +- '"' => { +- return Ok((input.advance(byte_offset), ())); +- } +- '\r' => { +- if let Some((_, '\n')) = chars.next() { +- // ... +- } else { +- break; +- } +- } +- '\\' => match chars.next() { +- Some((_, 'x')) => { +- if !backslash_x_char(&mut chars) { +- break; +- } +- } +- Some((_, 'n')) | Some((_, 'r')) | Some((_, 't')) | Some((_, '\\')) +- | Some((_, '\'')) | Some((_, '"')) | Some((_, '0')) => {} +- Some((_, 'u')) => { +- if !backslash_u(&mut chars) { +- break; +- } +- } +- Some((_, '\n')) | Some((_, '\r')) => { +- while let Some(&(_, ch)) = chars.peek() { +- if ch.is_whitespace() { +- chars.next(); +- } else { +- break; +- } +- } +- } +- _ => break, +- }, +- _ch => {} +- } +- } +- Err(LexError) +-} +- +-named!(byte_string -> (), alt!( +- delimited!( +- punct!("b\""), +- cooked_byte_string, +- tag!("\"") +- ) => { |_| () } +- | +- preceded!( +- punct!("br"), +- raw_string +- ) => { |_| () } +-)); +- +-fn cooked_byte_string(mut input: Cursor) -> PResult<()> { +- let mut bytes = input.bytes().enumerate(); +- 'outer: while let Some((offset, b)) = bytes.next() { +- match b { +- b'"' => { +- return Ok((input.advance(offset), ())); +- } +- b'\r' => { +- if let Some((_, b'\n')) = bytes.next() { +- // ... +- } else { +- break; +- } +- } +- b'\\' => match bytes.next() { +- Some((_, b'x')) => { +- if !backslash_x_byte(&mut bytes) { +- break; +- } +- } +- Some((_, b'n')) | Some((_, b'r')) | Some((_, b't')) | Some((_, b'\\')) +- | Some((_, b'0')) | Some((_, b'\'')) | Some((_, b'"')) => {} +- Some((newline, b'\n')) | Some((newline, b'\r')) => { +- let rest = input.advance(newline + 1); +- for (offset, ch) in rest.char_indices() { +- if !ch.is_whitespace() { +- input = rest.advance(offset); +- bytes = input.bytes().enumerate(); +- continue 'outer; +- } +- } +- break; +- } +- _ => break, +- }, +- b if b < 0x80 => {} +- _ => break, +- } +- } +- Err(LexError) +-} +- +-fn raw_string(input: Cursor) -> PResult<()> { +- let mut chars = input.char_indices(); +- let mut n = 0; +- while let Some((byte_offset, ch)) = chars.next() { +- match ch { +- '"' => { +- n = byte_offset; +- break; +- } +- '#' => {} +- _ => return Err(LexError), +- } +- } +- for (byte_offset, ch) in chars { +- match ch { +- '"' if input.advance(byte_offset + 1).starts_with(&input.rest[..n]) => { +- let rest = input.advance(byte_offset + 1 + n); +- return Ok((rest, ())); +- } +- '\r' => {} +- _ => {} +- } +- } +- Err(LexError) +-} +- +-named!(byte -> (), do_parse!( +- punct!("b") >> +- tag!("'") >> +- cooked_byte >> +- tag!("'") >> +- (()) +-)); +- +-fn cooked_byte(input: Cursor) -> PResult<()> { +- let mut bytes = input.bytes().enumerate(); +- let ok = match bytes.next().map(|(_, b)| b) { +- Some(b'\\') => match bytes.next().map(|(_, b)| b) { +- Some(b'x') => backslash_x_byte(&mut bytes), +- Some(b'n') | Some(b'r') | Some(b't') | Some(b'\\') | Some(b'0') | Some(b'\'') +- | Some(b'"') => true, +- _ => false, +- }, +- b => b.is_some(), +- }; +- if ok { +- match bytes.next() { +- Some((offset, _)) => { +- if input.chars().as_str().is_char_boundary(offset) { +- Ok((input.advance(offset), ())) +- } else { +- Err(LexError) +- } +- } +- None => Ok((input.advance(input.len()), ())), +- } +- } else { +- Err(LexError) +- } +-} +- +-named!(character -> (), do_parse!( +- punct!("'") >> +- cooked_char >> +- tag!("'") >> +- (()) +-)); +- +-fn cooked_char(input: Cursor) -> PResult<()> { +- let mut chars = input.char_indices(); +- let ok = match chars.next().map(|(_, ch)| ch) { +- Some('\\') => match chars.next().map(|(_, ch)| ch) { +- Some('x') => backslash_x_char(&mut chars), +- Some('u') => backslash_u(&mut chars), +- Some('n') | Some('r') | Some('t') | Some('\\') | Some('0') | Some('\'') | Some('"') => { +- true +- } +- _ => false, +- }, +- ch => ch.is_some(), +- }; +- if ok { +- match chars.next() { +- Some((idx, _)) => Ok((input.advance(idx), ())), +- None => Ok((input.advance(input.len()), ())), +- } +- } else { +- Err(LexError) +- } +-} +- +-macro_rules! next_ch { +- ($chars:ident @ $pat:pat $(| $rest:pat)*) => { +- match $chars.next() { +- Some((_, ch)) => match ch { +- $pat $(| $rest)* => ch, +- _ => return false, +- }, +- None => return false +- } +- }; +-} +- +-fn backslash_x_char<I>(chars: &mut I) -> bool +-where +- I: Iterator<Item = (usize, char)>, +-{ +- next_ch!(chars @ '0'..='7'); +- next_ch!(chars @ '0'..='9' | 'a'..='f' | 'A'..='F'); +- true +-} +- +-fn backslash_x_byte<I>(chars: &mut I) -> bool +-where +- I: Iterator<Item = (usize, u8)>, +-{ +- next_ch!(chars @ b'0'..=b'9' | b'a'..=b'f' | b'A'..=b'F'); +- next_ch!(chars @ b'0'..=b'9' | b'a'..=b'f' | b'A'..=b'F'); +- true +-} +- +-fn backslash_u<I>(chars: &mut I) -> bool +-where +- I: Iterator<Item = (usize, char)>, +-{ +- next_ch!(chars @ '{'); +- next_ch!(chars @ '0'..='9' | 'a'..='f' | 'A'..='F'); +- loop { +- let c = next_ch!(chars @ '0'..='9' | 'a'..='f' | 'A'..='F' | '_' | '}'); +- if c == '}' { +- return true; +- } +- } +-} +- +-fn float(input: Cursor) -> PResult<()> { +- let (mut rest, ()) = float_digits(input)?; +- if let Some(ch) = rest.chars().next() { +- if is_ident_start(ch) { +- rest = symbol_not_raw(rest)?.0; +- } +- } +- word_break(rest) +-} +- +-fn float_digits(input: Cursor) -> PResult<()> { +- let mut chars = input.chars().peekable(); +- match chars.next() { +- Some(ch) if ch >= '0' && ch <= '9' => {} +- _ => return Err(LexError), +- } +- +- let mut len = 1; +- let mut has_dot = false; +- let mut has_exp = false; +- while let Some(&ch) = chars.peek() { +- match ch { +- '0'..='9' | '_' => { +- chars.next(); +- len += 1; +- } +- '.' => { +- if has_dot { +- break; +- } +- chars.next(); +- if chars +- .peek() +- .map(|&ch| ch == '.' || is_ident_start(ch)) +- .unwrap_or(false) +- { +- return Err(LexError); +- } +- len += 1; +- has_dot = true; +- } +- 'e' | 'E' => { +- chars.next(); +- len += 1; +- has_exp = true; +- break; +- } +- _ => break, +- } +- } +- +- let rest = input.advance(len); +- if !(has_dot || has_exp || rest.starts_with("f32") || rest.starts_with("f64")) { +- return Err(LexError); +- } +- +- if has_exp { +- let mut has_exp_value = false; +- while let Some(&ch) = chars.peek() { +- match ch { +- '+' | '-' => { +- if has_exp_value { +- break; +- } +- chars.next(); +- len += 1; +- } +- '0'..='9' => { +- chars.next(); +- len += 1; +- has_exp_value = true; +- } +- '_' => { +- chars.next(); +- len += 1; +- } +- _ => break, +- } +- } +- if !has_exp_value { +- return Err(LexError); +- } +- } +- +- Ok((input.advance(len), ())) +-} +- +-fn int(input: Cursor) -> PResult<()> { +- let (mut rest, ()) = digits(input)?; +- if let Some(ch) = rest.chars().next() { +- if is_ident_start(ch) { +- rest = symbol_not_raw(rest)?.0; +- } +- } +- word_break(rest) +-} +- +-fn digits(mut input: Cursor) -> PResult<()> { +- let base = if input.starts_with("0x") { +- input = input.advance(2); +- 16 +- } else if input.starts_with("0o") { +- input = input.advance(2); +- 8 +- } else if input.starts_with("0b") { +- input = input.advance(2); +- 2 +- } else { +- 10 +- }; +- +- let mut len = 0; +- let mut empty = true; +- for b in input.bytes() { +- let digit = match b { +- b'0'..=b'9' => (b - b'0') as u64, +- b'a'..=b'f' => 10 + (b - b'a') as u64, +- b'A'..=b'F' => 10 + (b - b'A') as u64, +- b'_' => { +- if empty && base == 10 { +- return Err(LexError); +- } +- len += 1; +- continue; +- } +- _ => break, +- }; +- if digit >= base { +- return Err(LexError); +- } +- len += 1; +- empty = false; +- } +- if empty { +- Err(LexError) +- } else { +- Ok((input.advance(len), ())) +- } +-} +- +-fn op(input: Cursor) -> PResult<Punct> { +- let input = skip_whitespace(input); +- match op_char(input) { +- Ok((rest, '\'')) => { +- symbol(rest)?; +- Ok((rest, Punct::new('\'', Spacing::Joint))) +- } +- Ok((rest, ch)) => { +- let kind = match op_char(rest) { +- Ok(_) => Spacing::Joint, +- Err(LexError) => Spacing::Alone, +- }; +- Ok((rest, Punct::new(ch, kind))) +- } +- Err(LexError) => Err(LexError), +- } +-} +- +-fn op_char(input: Cursor) -> PResult<char> { +- if input.starts_with("//") || input.starts_with("/*") { +- // Do not accept `/` of a comment as an op. +- return Err(LexError); +- } +- +- let mut chars = input.chars(); +- let first = match chars.next() { +- Some(ch) => ch, +- None => { +- return Err(LexError); +- } +- }; +- let recognized = "~!@#$%^&*-=+|;:,<.>/?'"; +- if recognized.contains(first) { +- Ok((input.advance(first.len_utf8()), first)) +- } else { +- Err(LexError) +- } +-} +- +-fn doc_comment(input: Cursor) -> PResult<Vec<TokenTree>> { +- let mut trees = Vec::new(); +- let (rest, ((comment, inner), span)) = spanned(input, doc_comment_contents)?; +- trees.push(TokenTree::Punct(Punct::new('#', Spacing::Alone))); +- if inner { +- trees.push(Punct::new('!', Spacing::Alone).into()); +- } +- let mut stream = vec![ +- TokenTree::Ident(crate::Ident::new("doc", span)), +- TokenTree::Punct(Punct::new('=', Spacing::Alone)), +- TokenTree::Literal(crate::Literal::string(comment)), +- ]; +- for tt in stream.iter_mut() { +- tt.set_span(span); +- } +- let group = Group::new(Delimiter::Bracket, stream.into_iter().collect()); +- trees.push(crate::Group::_new_stable(group).into()); +- for tt in trees.iter_mut() { +- tt.set_span(span); +- } +- Ok((rest, trees)) +-} +- +-named!(doc_comment_contents -> (&str, bool), alt!( +- do_parse!( +- punct!("//!") >> +- s: take_until_newline_or_eof!() >> +- ((s, true)) +- ) +- | +- do_parse!( +- option!(whitespace) >> +- peek!(tag!("/*!")) >> +- s: block_comment >> +- ((s, true)) +- ) +- | +- do_parse!( +- punct!("///") >> +- not!(tag!("/")) >> +- s: take_until_newline_or_eof!() >> +- ((s, false)) +- ) +- | +- do_parse!( +- option!(whitespace) >> +- peek!(tuple!(tag!("/**"), not!(tag!("*")))) >> +- s: block_comment >> +- ((s, false)) +- ) +-)); +diff --git a/third_party/rust/proc-macro2/src/lib.rs b/third_party/rust/proc-macro2/src/lib.rs +index a08be3e815..c20fb50d4a 100644 +--- a/third_party/rust/proc-macro2/src/lib.rs ++++ b/third_party/rust/proc-macro2/src/lib.rs +@@ -78,27 +78,24 @@ + //! a different thread. + + // Proc-macro2 types in rustdoc of other crates get linked to here. +-#![doc(html_root_url = "https://docs.rs/proc-macro2/1.0.5")] ++#![doc(html_root_url = "https://docs.rs/proc-macro2/1.0.24")] + #![cfg_attr(any(proc_macro_span, super_unstable), feature(proc_macro_span))] + #![cfg_attr(super_unstable, feature(proc_macro_raw_ident, proc_macro_def_site))] ++#![allow(clippy::needless_doctest_main)] + + #[cfg(use_proc_macro)] + extern crate proc_macro; + +-use std::cmp::Ordering; +-use std::fmt; +-use std::hash::{Hash, Hasher}; +-use std::iter::FromIterator; +-use std::marker; +-use std::ops::RangeBounds; +-#[cfg(procmacro2_semver_exempt)] +-use std::path::PathBuf; +-use std::rc::Rc; +-use std::str::FromStr; ++mod marker; ++mod parse; ++ ++#[cfg(wrap_proc_macro)] ++mod detection; + +-#[macro_use] +-mod strnom; +-mod fallback; ++// Public for proc_macro2::fallback::force() and unforce(), but those are quite ++// a niche use case so we omit it from rustdoc. ++#[doc(hidden)] ++pub mod fallback; + + #[cfg(not(wrap_proc_macro))] + use crate::fallback as imp; +@@ -106,6 +103,17 @@ use crate::fallback as imp; + #[cfg(wrap_proc_macro)] + mod imp; + ++use crate::marker::Marker; ++use std::cmp::Ordering; ++use std::error::Error; ++use std::fmt::{self, Debug, Display}; ++use std::hash::{Hash, Hasher}; ++use std::iter::FromIterator; ++use std::ops::RangeBounds; ++#[cfg(procmacro2_semver_exempt)] ++use std::path::PathBuf; ++use std::str::FromStr; ++ + /// An abstract stream of tokens, or more concretely a sequence of token trees. + /// + /// This type provides interfaces for iterating over token trees and for +@@ -116,27 +124,27 @@ mod imp; + #[derive(Clone)] + pub struct TokenStream { + inner: imp::TokenStream, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + /// Error returned from `TokenStream::from_str`. + pub struct LexError { + inner: imp::LexError, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + impl TokenStream { + fn _new(inner: imp::TokenStream) -> TokenStream { + TokenStream { + inner, +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + + fn _new_stable(inner: fallback::TokenStream) -> TokenStream { + TokenStream { + inner: inner.into(), +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + +@@ -173,7 +181,7 @@ impl FromStr for TokenStream { + fn from_str(src: &str) -> Result<TokenStream, LexError> { + let e = src.parse().map_err(|e| LexError { + inner: e, +- _marker: marker::PhantomData, ++ _marker: Marker, + })?; + Ok(TokenStream::_new(e)) + } +@@ -228,25 +236,33 @@ impl FromIterator<TokenStream> for TokenStream { + /// convertible back into the same token stream (modulo spans), except for + /// possibly `TokenTree::Group`s with `Delimiter::None` delimiters and negative + /// numeric literals. +-impl fmt::Display for TokenStream { ++impl Display for TokenStream { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Display::fmt(&self.inner, f) + } + } + + /// Prints token in a form convenient for debugging. +-impl fmt::Debug for TokenStream { ++impl Debug for TokenStream { ++ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { ++ Debug::fmt(&self.inner, f) ++ } ++} ++ ++impl Debug for LexError { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +-impl fmt::Debug for LexError { ++impl Display for LexError { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Display::fmt(&self.inner, f) + } + } + ++impl Error for LexError {} ++ + /// The source file of a given `Span`. + /// + /// This type is semver exempt and not exposed by default. +@@ -254,7 +270,7 @@ impl fmt::Debug for LexError { + #[derive(Clone, PartialEq, Eq)] + pub struct SourceFile { + inner: imp::SourceFile, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + #[cfg(procmacro2_semver_exempt)] +@@ -262,7 +278,7 @@ impl SourceFile { + fn _new(inner: imp::SourceFile) -> Self { + SourceFile { + inner, +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + +@@ -291,9 +307,9 @@ impl SourceFile { + } + + #[cfg(procmacro2_semver_exempt)] +-impl fmt::Debug for SourceFile { ++impl Debug for SourceFile { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +@@ -311,25 +327,41 @@ pub struct LineColumn { + pub column: usize, + } + ++#[cfg(span_locations)] ++impl Ord for LineColumn { ++ fn cmp(&self, other: &Self) -> Ordering { ++ self.line ++ .cmp(&other.line) ++ .then(self.column.cmp(&other.column)) ++ } ++} ++ ++#[cfg(span_locations)] ++impl PartialOrd for LineColumn { ++ fn partial_cmp(&self, other: &Self) -> Option<Ordering> { ++ Some(self.cmp(other)) ++ } ++} ++ + /// A region of source code, along with macro expansion information. + #[derive(Copy, Clone)] + pub struct Span { + inner: imp::Span, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + impl Span { + fn _new(inner: imp::Span) -> Span { + Span { + inner, +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + + fn _new_stable(inner: fallback::Span) -> Span { + Span { + inner: inner.into(), +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + +@@ -342,6 +374,16 @@ impl Span { + Span::_new(imp::Span::call_site()) + } + ++ /// The span located at the invocation of the procedural macro, but with ++ /// local variables, labels, and `$crate` resolved at the definition site ++ /// of the macro. This is the same hygiene behavior as `macro_rules`. ++ /// ++ /// This function requires Rust 1.45 or later. ++ #[cfg(hygiene)] ++ pub fn mixed_site() -> Span { ++ Span::_new(imp::Span::mixed_site()) ++ } ++ + /// A span that resolves at the macro definition site. + /// + /// This method is semver exempt and not exposed by default. +@@ -352,18 +394,12 @@ impl Span { + + /// Creates a new span with the same line/column information as `self` but + /// that resolves symbols as though it were at `other`. +- /// +- /// This method is semver exempt and not exposed by default. +- #[cfg(procmacro2_semver_exempt)] + pub fn resolved_at(&self, other: Span) -> Span { + Span::_new(self.inner.resolved_at(other.inner)) + } + + /// Creates a new span with the same name resolution behavior as `self` but + /// with the line/column information of `other`. +- /// +- /// This method is semver exempt and not exposed by default. +- #[cfg(procmacro2_semver_exempt)] + pub fn located_at(&self, other: Span) -> Span { + Span::_new(self.inner.located_at(other.inner)) + } +@@ -439,9 +475,9 @@ impl Span { + } + + /// Prints a span in a form convenient for debugging. +-impl fmt::Debug for Span { ++impl Debug for Span { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +@@ -462,11 +498,11 @@ impl TokenTree { + /// Returns the span of this tree, delegating to the `span` method of + /// the contained token or a delimited stream. + pub fn span(&self) -> Span { +- match *self { +- TokenTree::Group(ref t) => t.span(), +- TokenTree::Ident(ref t) => t.span(), +- TokenTree::Punct(ref t) => t.span(), +- TokenTree::Literal(ref t) => t.span(), ++ match self { ++ TokenTree::Group(t) => t.span(), ++ TokenTree::Ident(t) => t.span(), ++ TokenTree::Punct(t) => t.span(), ++ TokenTree::Literal(t) => t.span(), + } + } + +@@ -476,11 +512,11 @@ impl TokenTree { + /// the span of each of the internal tokens, this will simply delegate to + /// the `set_span` method of each variant. + pub fn set_span(&mut self, span: Span) { +- match *self { +- TokenTree::Group(ref mut t) => t.set_span(span), +- TokenTree::Ident(ref mut t) => t.set_span(span), +- TokenTree::Punct(ref mut t) => t.set_span(span), +- TokenTree::Literal(ref mut t) => t.set_span(span), ++ match self { ++ TokenTree::Group(t) => t.set_span(span), ++ TokenTree::Ident(t) => t.set_span(span), ++ TokenTree::Punct(t) => t.set_span(span), ++ TokenTree::Literal(t) => t.set_span(span), + } + } + } +@@ -513,32 +549,32 @@ impl From<Literal> for TokenTree { + /// convertible back into the same token tree (modulo spans), except for + /// possibly `TokenTree::Group`s with `Delimiter::None` delimiters and negative + /// numeric literals. +-impl fmt::Display for TokenTree { ++impl Display for TokenTree { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- match *self { +- TokenTree::Group(ref t) => t.fmt(f), +- TokenTree::Ident(ref t) => t.fmt(f), +- TokenTree::Punct(ref t) => t.fmt(f), +- TokenTree::Literal(ref t) => t.fmt(f), ++ match self { ++ TokenTree::Group(t) => Display::fmt(t, f), ++ TokenTree::Ident(t) => Display::fmt(t, f), ++ TokenTree::Punct(t) => Display::fmt(t, f), ++ TokenTree::Literal(t) => Display::fmt(t, f), + } + } + } + + /// Prints token tree in a form convenient for debugging. +-impl fmt::Debug for TokenTree { ++impl Debug for TokenTree { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + // Each of these has the name in the struct type in the derived debug, + // so don't bother with an extra layer of indirection +- match *self { +- TokenTree::Group(ref t) => t.fmt(f), +- TokenTree::Ident(ref t) => { ++ match self { ++ TokenTree::Group(t) => Debug::fmt(t, f), ++ TokenTree::Ident(t) => { + let mut debug = f.debug_struct("Ident"); + debug.field("sym", &format_args!("{}", t)); + imp::debug_span_field_if_nontrivial(&mut debug, t.span().inner); + debug.finish() + } +- TokenTree::Punct(ref t) => t.fmt(f), +- TokenTree::Literal(ref t) => t.fmt(f), ++ TokenTree::Punct(t) => Debug::fmt(t, f), ++ TokenTree::Literal(t) => Debug::fmt(t, f), + } + } + } +@@ -651,15 +687,15 @@ impl Group { + /// Prints the group as a string that should be losslessly convertible back + /// into the same group (modulo spans), except for possibly `TokenTree::Group`s + /// with `Delimiter::None` delimiters. +-impl fmt::Display for Group { ++impl Display for Group { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { +- fmt::Display::fmt(&self.inner, formatter) ++ Display::fmt(&self.inner, formatter) + } + } + +-impl fmt::Debug for Group { ++impl Debug for Group { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { +- fmt::Debug::fmt(&self.inner, formatter) ++ Debug::fmt(&self.inner, formatter) + } + } + +@@ -669,7 +705,7 @@ impl fmt::Debug for Group { + /// `Punct` with different forms of `Spacing` returned. + #[derive(Clone)] + pub struct Punct { +- op: char, ++ ch: char, + spacing: Spacing, + span: Span, + } +@@ -695,9 +731,9 @@ impl Punct { + /// + /// The returned `Punct` will have the default span of `Span::call_site()` + /// which can be further configured with the `set_span` method below. +- pub fn new(op: char, spacing: Spacing) -> Punct { ++ pub fn new(ch: char, spacing: Spacing) -> Punct { + Punct { +- op, ++ ch, + spacing, + span: Span::call_site(), + } +@@ -705,7 +741,7 @@ impl Punct { + + /// Returns the value of this punctuation character as `char`. + pub fn as_char(&self) -> char { +- self.op ++ self.ch + } + + /// Returns the spacing of this punctuation character, indicating whether +@@ -730,16 +766,16 @@ impl Punct { + + /// Prints the punctuation character as a string that should be losslessly + /// convertible back into the same character. +-impl fmt::Display for Punct { ++impl Display for Punct { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.op.fmt(f) ++ Display::fmt(&self.ch, f) + } + } + +-impl fmt::Debug for Punct { ++impl Debug for Punct { + fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result { + let mut debug = fmt.debug_struct("Punct"); +- debug.field("op", &self.op); ++ debug.field("char", &self.ch); + debug.field("spacing", &self.spacing); + imp::debug_span_field_if_nontrivial(&mut debug, self.span.inner); + debug.finish() +@@ -813,14 +849,14 @@ impl fmt::Debug for Punct { + #[derive(Clone)] + pub struct Ident { + inner: imp::Ident, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + impl Ident { + fn _new(inner: imp::Ident) -> Ident { + Ident { + inner, +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + +@@ -920,15 +956,15 @@ impl Hash for Ident { + + /// Prints the identifier as a string that should be losslessly convertible back + /// into the same identifier. +-impl fmt::Display for Ident { ++impl Display for Ident { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Display::fmt(&self.inner, f) + } + } + +-impl fmt::Debug for Ident { ++impl Debug for Ident { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +@@ -941,7 +977,7 @@ impl fmt::Debug for Ident { + #[derive(Clone)] + pub struct Literal { + inner: imp::Literal, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + macro_rules! suffixed_int_literals { +@@ -988,14 +1024,14 @@ impl Literal { + fn _new(inner: imp::Literal) -> Literal { + Literal { + inner, +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + + fn _new_stable(inner: fallback::Literal) -> Literal { + Literal { + inner: inner.into(), +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + +@@ -1140,26 +1176,25 @@ impl Literal { + } + } + +-impl fmt::Debug for Literal { ++impl Debug for Literal { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +-impl fmt::Display for Literal { ++impl Display for Literal { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Display::fmt(&self.inner, f) + } + } + + /// Public implementation details for the `TokenStream` type, such as iterators. + pub mod token_stream { +- use std::fmt; +- use std::marker; +- use std::rc::Rc; ++ use crate::marker::Marker; ++ use crate::{imp, TokenTree}; ++ use std::fmt::{self, Debug}; + + pub use crate::TokenStream; +- use crate::{imp, TokenTree}; + + /// An iterator over `TokenStream`'s `TokenTree`s. + /// +@@ -1168,7 +1203,7 @@ pub mod token_stream { + #[derive(Clone)] + pub struct IntoIter { + inner: imp::TokenTreeIter, +- _marker: marker::PhantomData<Rc<()>>, ++ _marker: Marker, + } + + impl Iterator for IntoIter { +@@ -1179,9 +1214,9 @@ pub mod token_stream { + } + } + +- impl fmt::Debug for IntoIter { ++ impl Debug for IntoIter { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +- self.inner.fmt(f) ++ Debug::fmt(&self.inner, f) + } + } + +@@ -1192,7 +1227,7 @@ pub mod token_stream { + fn into_iter(self) -> IntoIter { + IntoIter { + inner: self.inner.into_iter(), +- _marker: marker::PhantomData, ++ _marker: Marker, + } + } + } +diff --git a/third_party/rust/proc-macro2/src/marker.rs b/third_party/rust/proc-macro2/src/marker.rs +new file mode 100644 +index 0000000000..58729baf4a +--- /dev/null ++++ b/third_party/rust/proc-macro2/src/marker.rs +@@ -0,0 +1,18 @@ ++use std::marker::PhantomData; ++use std::panic::{RefUnwindSafe, UnwindSafe}; ++use std::rc::Rc; ++ ++// Zero sized marker with the correct set of autotrait impls we want all proc ++// macro types to have. ++pub(crate) type Marker = PhantomData<ProcMacroAutoTraits>; ++ ++pub(crate) use self::value::*; ++ ++mod value { ++ pub(crate) use std::marker::PhantomData as Marker; ++} ++ ++pub(crate) struct ProcMacroAutoTraits(Rc<()>); ++ ++impl UnwindSafe for ProcMacroAutoTraits {} ++impl RefUnwindSafe for ProcMacroAutoTraits {} +diff --git a/third_party/rust/proc-macro2/src/parse.rs b/third_party/rust/proc-macro2/src/parse.rs +new file mode 100644 +index 0000000000..365fe0484d +--- /dev/null ++++ b/third_party/rust/proc-macro2/src/parse.rs +@@ -0,0 +1,849 @@ ++use crate::fallback::{ ++ is_ident_continue, is_ident_start, Group, LexError, Literal, Span, TokenStream, ++}; ++use crate::{Delimiter, Punct, Spacing, TokenTree}; ++use std::char; ++use std::str::{Bytes, CharIndices, Chars}; ++ ++#[derive(Copy, Clone, Eq, PartialEq)] ++pub(crate) struct Cursor<'a> { ++ pub rest: &'a str, ++ #[cfg(span_locations)] ++ pub off: u32, ++} ++ ++impl<'a> Cursor<'a> { ++ fn advance(&self, bytes: usize) -> Cursor<'a> { ++ let (_front, rest) = self.rest.split_at(bytes); ++ Cursor { ++ rest, ++ #[cfg(span_locations)] ++ off: self.off + _front.chars().count() as u32, ++ } ++ } ++ ++ fn starts_with(&self, s: &str) -> bool { ++ self.rest.starts_with(s) ++ } ++ ++ pub(crate) fn is_empty(&self) -> bool { ++ self.rest.is_empty() ++ } ++ ++ fn len(&self) -> usize { ++ self.rest.len() ++ } ++ ++ fn as_bytes(&self) -> &'a [u8] { ++ self.rest.as_bytes() ++ } ++ ++ fn bytes(&self) -> Bytes<'a> { ++ self.rest.bytes() ++ } ++ ++ fn chars(&self) -> Chars<'a> { ++ self.rest.chars() ++ } ++ ++ fn char_indices(&self) -> CharIndices<'a> { ++ self.rest.char_indices() ++ } ++ ++ fn parse(&self, tag: &str) -> Result<Cursor<'a>, LexError> { ++ if self.starts_with(tag) { ++ Ok(self.advance(tag.len())) ++ } else { ++ Err(LexError) ++ } ++ } ++} ++ ++type PResult<'a, O> = Result<(Cursor<'a>, O), LexError>; ++ ++fn skip_whitespace(input: Cursor) -> Cursor { ++ let mut s = input; ++ ++ while !s.is_empty() { ++ let byte = s.as_bytes()[0]; ++ if byte == b'/' { ++ if s.starts_with("//") ++ && (!s.starts_with("///") || s.starts_with("////")) ++ && !s.starts_with("//!") ++ { ++ let (cursor, _) = take_until_newline_or_eof(s); ++ s = cursor; ++ continue; ++ } else if s.starts_with("/**/") { ++ s = s.advance(4); ++ continue; ++ } else if s.starts_with("/*") ++ && (!s.starts_with("/**") || s.starts_with("/***")) ++ && !s.starts_with("/*!") ++ { ++ match block_comment(s) { ++ Ok((rest, _)) => { ++ s = rest; ++ continue; ++ } ++ Err(LexError) => return s, ++ } ++ } ++ } ++ match byte { ++ b' ' | 0x09..=0x0d => { ++ s = s.advance(1); ++ continue; ++ } ++ b if b <= 0x7f => {} ++ _ => { ++ let ch = s.chars().next().unwrap(); ++ if is_whitespace(ch) { ++ s = s.advance(ch.len_utf8()); ++ continue; ++ } ++ } ++ } ++ return s; ++ } ++ s ++} ++ ++fn block_comment(input: Cursor) -> PResult<&str> { ++ if !input.starts_with("/*") { ++ return Err(LexError); ++ } ++ ++ let mut depth = 0; ++ let bytes = input.as_bytes(); ++ let mut i = 0; ++ let upper = bytes.len() - 1; ++ ++ while i < upper { ++ if bytes[i] == b'/' && bytes[i + 1] == b'*' { ++ depth += 1; ++ i += 1; // eat '*' ++ } else if bytes[i] == b'*' && bytes[i + 1] == b'/' { ++ depth -= 1; ++ if depth == 0 { ++ return Ok((input.advance(i + 2), &input.rest[..i + 2])); ++ } ++ i += 1; // eat '/' ++ } ++ i += 1; ++ } ++ ++ Err(LexError) ++} ++ ++fn is_whitespace(ch: char) -> bool { ++ // Rust treats left-to-right mark and right-to-left mark as whitespace ++ ch.is_whitespace() || ch == '\u{200e}' || ch == '\u{200f}' ++} ++ ++fn word_break(input: Cursor) -> Result<Cursor, LexError> { ++ match input.chars().next() { ++ Some(ch) if is_ident_continue(ch) => Err(LexError), ++ Some(_) | None => Ok(input), ++ } ++} ++ ++pub(crate) fn token_stream(mut input: Cursor) -> PResult<TokenStream> { ++ let mut trees = Vec::new(); ++ let mut stack = Vec::new(); ++ ++ loop { ++ input = skip_whitespace(input); ++ ++ if let Ok((rest, tt)) = doc_comment(input) { ++ trees.extend(tt); ++ input = rest; ++ continue; ++ } ++ ++ #[cfg(span_locations)] ++ let lo = input.off; ++ ++ let first = match input.bytes().next() { ++ Some(first) => first, ++ None => break, ++ }; ++ ++ if let Some(open_delimiter) = match first { ++ b'(' => Some(Delimiter::Parenthesis), ++ b'[' => Some(Delimiter::Bracket), ++ b'{' => Some(Delimiter::Brace), ++ _ => None, ++ } { ++ input = input.advance(1); ++ let frame = (open_delimiter, trees); ++ #[cfg(span_locations)] ++ let frame = (lo, frame); ++ stack.push(frame); ++ trees = Vec::new(); ++ } else if let Some(close_delimiter) = match first { ++ b')' => Some(Delimiter::Parenthesis), ++ b']' => Some(Delimiter::Bracket), ++ b'}' => Some(Delimiter::Brace), ++ _ => None, ++ } { ++ input = input.advance(1); ++ let frame = stack.pop().ok_or(LexError)?; ++ #[cfg(span_locations)] ++ let (lo, frame) = frame; ++ let (open_delimiter, outer) = frame; ++ if open_delimiter != close_delimiter { ++ return Err(LexError); ++ } ++ let mut g = Group::new(open_delimiter, TokenStream { inner: trees }); ++ g.set_span(Span { ++ #[cfg(span_locations)] ++ lo, ++ #[cfg(span_locations)] ++ hi: input.off, ++ }); ++ trees = outer; ++ trees.push(TokenTree::Group(crate::Group::_new_stable(g))); ++ } else { ++ let (rest, mut tt) = leaf_token(input)?; ++ tt.set_span(crate::Span::_new_stable(Span { ++ #[cfg(span_locations)] ++ lo, ++ #[cfg(span_locations)] ++ hi: rest.off, ++ })); ++ trees.push(tt); ++ input = rest; ++ } ++ } ++ ++ if stack.is_empty() { ++ Ok((input, TokenStream { inner: trees })) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn leaf_token(input: Cursor) -> PResult<TokenTree> { ++ if let Ok((input, l)) = literal(input) { ++ // must be parsed before ident ++ Ok((input, TokenTree::Literal(crate::Literal::_new_stable(l)))) ++ } else if let Ok((input, p)) = punct(input) { ++ Ok((input, TokenTree::Punct(p))) ++ } else if let Ok((input, i)) = ident(input) { ++ Ok((input, TokenTree::Ident(i))) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn ident(input: Cursor) -> PResult<crate::Ident> { ++ if ["r\"", "r#\"", "r##", "b\"", "b\'", "br\"", "br#"] ++ .iter() ++ .any(|prefix| input.starts_with(prefix)) ++ { ++ Err(LexError) ++ } else { ++ ident_any(input) ++ } ++} ++ ++fn ident_any(input: Cursor) -> PResult<crate::Ident> { ++ let raw = input.starts_with("r#"); ++ let rest = input.advance((raw as usize) << 1); ++ ++ let (rest, sym) = ident_not_raw(rest)?; ++ ++ if !raw { ++ let ident = crate::Ident::new(sym, crate::Span::call_site()); ++ return Ok((rest, ident)); ++ } ++ ++ if sym == "_" { ++ return Err(LexError); ++ } ++ ++ let ident = crate::Ident::_new_raw(sym, crate::Span::call_site()); ++ Ok((rest, ident)) ++} ++ ++fn ident_not_raw(input: Cursor) -> PResult<&str> { ++ let mut chars = input.char_indices(); ++ ++ match chars.next() { ++ Some((_, ch)) if is_ident_start(ch) => {} ++ _ => return Err(LexError), ++ } ++ ++ let mut end = input.len(); ++ for (i, ch) in chars { ++ if !is_ident_continue(ch) { ++ end = i; ++ break; ++ } ++ } ++ ++ Ok((input.advance(end), &input.rest[..end])) ++} ++ ++fn literal(input: Cursor) -> PResult<Literal> { ++ match literal_nocapture(input) { ++ Ok(a) => { ++ let end = input.len() - a.len(); ++ Ok((a, Literal::_new(input.rest[..end].to_string()))) ++ } ++ Err(LexError) => Err(LexError), ++ } ++} ++ ++fn literal_nocapture(input: Cursor) -> Result<Cursor, LexError> { ++ if let Ok(ok) = string(input) { ++ Ok(ok) ++ } else if let Ok(ok) = byte_string(input) { ++ Ok(ok) ++ } else if let Ok(ok) = byte(input) { ++ Ok(ok) ++ } else if let Ok(ok) = character(input) { ++ Ok(ok) ++ } else if let Ok(ok) = float(input) { ++ Ok(ok) ++ } else if let Ok(ok) = int(input) { ++ Ok(ok) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn literal_suffix(input: Cursor) -> Cursor { ++ match ident_not_raw(input) { ++ Ok((input, _)) => input, ++ Err(LexError) => input, ++ } ++} ++ ++fn string(input: Cursor) -> Result<Cursor, LexError> { ++ if let Ok(input) = input.parse("\"") { ++ cooked_string(input) ++ } else if let Ok(input) = input.parse("r") { ++ raw_string(input) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn cooked_string(input: Cursor) -> Result<Cursor, LexError> { ++ let mut chars = input.char_indices().peekable(); ++ ++ while let Some((i, ch)) = chars.next() { ++ match ch { ++ '"' => { ++ let input = input.advance(i + 1); ++ return Ok(literal_suffix(input)); ++ } ++ '\r' => match chars.next() { ++ Some((_, '\n')) => {} ++ _ => break, ++ }, ++ '\\' => match chars.next() { ++ Some((_, 'x')) => { ++ if !backslash_x_char(&mut chars) { ++ break; ++ } ++ } ++ Some((_, 'n')) | Some((_, 'r')) | Some((_, 't')) | Some((_, '\\')) ++ | Some((_, '\'')) | Some((_, '"')) | Some((_, '0')) => {} ++ Some((_, 'u')) => { ++ if !backslash_u(&mut chars) { ++ break; ++ } ++ } ++ Some((_, ch @ '\n')) | Some((_, ch @ '\r')) => { ++ let mut last = ch; ++ loop { ++ if last == '\r' && chars.next().map_or(true, |(_, ch)| ch != '\n') { ++ return Err(LexError); ++ } ++ match chars.peek() { ++ Some((_, ch)) if ch.is_whitespace() => { ++ last = *ch; ++ chars.next(); ++ } ++ _ => break, ++ } ++ } ++ } ++ _ => break, ++ }, ++ _ch => {} ++ } ++ } ++ Err(LexError) ++} ++ ++fn byte_string(input: Cursor) -> Result<Cursor, LexError> { ++ if let Ok(input) = input.parse("b\"") { ++ cooked_byte_string(input) ++ } else if let Ok(input) = input.parse("br") { ++ raw_string(input) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn cooked_byte_string(mut input: Cursor) -> Result<Cursor, LexError> { ++ let mut bytes = input.bytes().enumerate(); ++ while let Some((offset, b)) = bytes.next() { ++ match b { ++ b'"' => { ++ let input = input.advance(offset + 1); ++ return Ok(literal_suffix(input)); ++ } ++ b'\r' => match bytes.next() { ++ Some((_, b'\n')) => {} ++ _ => break, ++ }, ++ b'\\' => match bytes.next() { ++ Some((_, b'x')) => { ++ if !backslash_x_byte(&mut bytes) { ++ break; ++ } ++ } ++ Some((_, b'n')) | Some((_, b'r')) | Some((_, b't')) | Some((_, b'\\')) ++ | Some((_, b'0')) | Some((_, b'\'')) | Some((_, b'"')) => {} ++ Some((newline, b @ b'\n')) | Some((newline, b @ b'\r')) => { ++ let mut last = b as char; ++ let rest = input.advance(newline + 1); ++ let mut chars = rest.char_indices(); ++ loop { ++ if last == '\r' && chars.next().map_or(true, |(_, ch)| ch != '\n') { ++ return Err(LexError); ++ } ++ match chars.next() { ++ Some((_, ch)) if ch.is_whitespace() => last = ch, ++ Some((offset, _)) => { ++ input = rest.advance(offset); ++ bytes = input.bytes().enumerate(); ++ break; ++ } ++ None => return Err(LexError), ++ } ++ } ++ } ++ _ => break, ++ }, ++ b if b < 0x80 => {} ++ _ => break, ++ } ++ } ++ Err(LexError) ++} ++ ++fn raw_string(input: Cursor) -> Result<Cursor, LexError> { ++ let mut chars = input.char_indices(); ++ let mut n = 0; ++ while let Some((i, ch)) = chars.next() { ++ match ch { ++ '"' => { ++ n = i; ++ break; ++ } ++ '#' => {} ++ _ => return Err(LexError), ++ } ++ } ++ while let Some((i, ch)) = chars.next() { ++ match ch { ++ '"' if input.rest[i + 1..].starts_with(&input.rest[..n]) => { ++ let rest = input.advance(i + 1 + n); ++ return Ok(literal_suffix(rest)); ++ } ++ '\r' => match chars.next() { ++ Some((_, '\n')) => {} ++ _ => break, ++ }, ++ _ => {} ++ } ++ } ++ Err(LexError) ++} ++ ++fn byte(input: Cursor) -> Result<Cursor, LexError> { ++ let input = input.parse("b'")?; ++ let mut bytes = input.bytes().enumerate(); ++ let ok = match bytes.next().map(|(_, b)| b) { ++ Some(b'\\') => match bytes.next().map(|(_, b)| b) { ++ Some(b'x') => backslash_x_byte(&mut bytes), ++ Some(b'n') | Some(b'r') | Some(b't') | Some(b'\\') | Some(b'0') | Some(b'\'') ++ | Some(b'"') => true, ++ _ => false, ++ }, ++ b => b.is_some(), ++ }; ++ if !ok { ++ return Err(LexError); ++ } ++ let (offset, _) = bytes.next().ok_or(LexError)?; ++ if !input.chars().as_str().is_char_boundary(offset) { ++ return Err(LexError); ++ } ++ let input = input.advance(offset).parse("'")?; ++ Ok(literal_suffix(input)) ++} ++ ++fn character(input: Cursor) -> Result<Cursor, LexError> { ++ let input = input.parse("'")?; ++ let mut chars = input.char_indices(); ++ let ok = match chars.next().map(|(_, ch)| ch) { ++ Some('\\') => match chars.next().map(|(_, ch)| ch) { ++ Some('x') => backslash_x_char(&mut chars), ++ Some('u') => backslash_u(&mut chars), ++ Some('n') | Some('r') | Some('t') | Some('\\') | Some('0') | Some('\'') | Some('"') => { ++ true ++ } ++ _ => false, ++ }, ++ ch => ch.is_some(), ++ }; ++ if !ok { ++ return Err(LexError); ++ } ++ let (idx, _) = chars.next().ok_or(LexError)?; ++ let input = input.advance(idx).parse("'")?; ++ Ok(literal_suffix(input)) ++} ++ ++macro_rules! next_ch { ++ ($chars:ident @ $pat:pat $(| $rest:pat)*) => { ++ match $chars.next() { ++ Some((_, ch)) => match ch { ++ $pat $(| $rest)* => ch, ++ _ => return false, ++ }, ++ None => return false, ++ } ++ }; ++} ++ ++fn backslash_x_char<I>(chars: &mut I) -> bool ++where ++ I: Iterator<Item = (usize, char)>, ++{ ++ next_ch!(chars @ '0'..='7'); ++ next_ch!(chars @ '0'..='9' | 'a'..='f' | 'A'..='F'); ++ true ++} ++ ++fn backslash_x_byte<I>(chars: &mut I) -> bool ++where ++ I: Iterator<Item = (usize, u8)>, ++{ ++ next_ch!(chars @ b'0'..=b'9' | b'a'..=b'f' | b'A'..=b'F'); ++ next_ch!(chars @ b'0'..=b'9' | b'a'..=b'f' | b'A'..=b'F'); ++ true ++} ++ ++fn backslash_u<I>(chars: &mut I) -> bool ++where ++ I: Iterator<Item = (usize, char)>, ++{ ++ next_ch!(chars @ '{'); ++ let mut value = 0; ++ let mut len = 0; ++ while let Some((_, ch)) = chars.next() { ++ let digit = match ch { ++ '0'..='9' => ch as u8 - b'0', ++ 'a'..='f' => 10 + ch as u8 - b'a', ++ 'A'..='F' => 10 + ch as u8 - b'A', ++ '_' if len > 0 => continue, ++ '}' if len > 0 => return char::from_u32(value).is_some(), ++ _ => return false, ++ }; ++ if len == 6 { ++ return false; ++ } ++ value *= 0x10; ++ value += u32::from(digit); ++ len += 1; ++ } ++ false ++} ++ ++fn float(input: Cursor) -> Result<Cursor, LexError> { ++ let mut rest = float_digits(input)?; ++ if let Some(ch) = rest.chars().next() { ++ if is_ident_start(ch) { ++ rest = ident_not_raw(rest)?.0; ++ } ++ } ++ word_break(rest) ++} ++ ++fn float_digits(input: Cursor) -> Result<Cursor, LexError> { ++ let mut chars = input.chars().peekable(); ++ match chars.next() { ++ Some(ch) if ch >= '0' && ch <= '9' => {} ++ _ => return Err(LexError), ++ } ++ ++ let mut len = 1; ++ let mut has_dot = false; ++ let mut has_exp = false; ++ while let Some(&ch) = chars.peek() { ++ match ch { ++ '0'..='9' | '_' => { ++ chars.next(); ++ len += 1; ++ } ++ '.' => { ++ if has_dot { ++ break; ++ } ++ chars.next(); ++ if chars ++ .peek() ++ .map(|&ch| ch == '.' || is_ident_start(ch)) ++ .unwrap_or(false) ++ { ++ return Err(LexError); ++ } ++ len += 1; ++ has_dot = true; ++ } ++ 'e' | 'E' => { ++ chars.next(); ++ len += 1; ++ has_exp = true; ++ break; ++ } ++ _ => break, ++ } ++ } ++ ++ if !(has_dot || has_exp) { ++ return Err(LexError); ++ } ++ ++ if has_exp { ++ let token_before_exp = if has_dot { ++ Ok(input.advance(len - 1)) ++ } else { ++ Err(LexError) ++ }; ++ let mut has_sign = false; ++ let mut has_exp_value = false; ++ while let Some(&ch) = chars.peek() { ++ match ch { ++ '+' | '-' => { ++ if has_exp_value { ++ break; ++ } ++ if has_sign { ++ return token_before_exp; ++ } ++ chars.next(); ++ len += 1; ++ has_sign = true; ++ } ++ '0'..='9' => { ++ chars.next(); ++ len += 1; ++ has_exp_value = true; ++ } ++ '_' => { ++ chars.next(); ++ len += 1; ++ } ++ _ => break, ++ } ++ } ++ if !has_exp_value { ++ return token_before_exp; ++ } ++ } ++ ++ Ok(input.advance(len)) ++} ++ ++fn int(input: Cursor) -> Result<Cursor, LexError> { ++ let mut rest = digits(input)?; ++ if let Some(ch) = rest.chars().next() { ++ if is_ident_start(ch) { ++ rest = ident_not_raw(rest)?.0; ++ } ++ } ++ word_break(rest) ++} ++ ++fn digits(mut input: Cursor) -> Result<Cursor, LexError> { ++ let base = if input.starts_with("0x") { ++ input = input.advance(2); ++ 16 ++ } else if input.starts_with("0o") { ++ input = input.advance(2); ++ 8 ++ } else if input.starts_with("0b") { ++ input = input.advance(2); ++ 2 ++ } else { ++ 10 ++ }; ++ ++ let mut len = 0; ++ let mut empty = true; ++ for b in input.bytes() { ++ match b { ++ b'0'..=b'9' => { ++ let digit = (b - b'0') as u64; ++ if digit >= base { ++ return Err(LexError); ++ } ++ } ++ b'a'..=b'f' => { ++ let digit = 10 + (b - b'a') as u64; ++ if digit >= base { ++ break; ++ } ++ } ++ b'A'..=b'F' => { ++ let digit = 10 + (b - b'A') as u64; ++ if digit >= base { ++ break; ++ } ++ } ++ b'_' => { ++ if empty && base == 10 { ++ return Err(LexError); ++ } ++ len += 1; ++ continue; ++ } ++ _ => break, ++ }; ++ len += 1; ++ empty = false; ++ } ++ if empty { ++ Err(LexError) ++ } else { ++ Ok(input.advance(len)) ++ } ++} ++ ++fn punct(input: Cursor) -> PResult<Punct> { ++ match punct_char(input) { ++ Ok((rest, '\'')) => { ++ if ident_any(rest)?.0.starts_with("'") { ++ Err(LexError) ++ } else { ++ Ok((rest, Punct::new('\'', Spacing::Joint))) ++ } ++ } ++ Ok((rest, ch)) => { ++ let kind = match punct_char(rest) { ++ Ok(_) => Spacing::Joint, ++ Err(LexError) => Spacing::Alone, ++ }; ++ Ok((rest, Punct::new(ch, kind))) ++ } ++ Err(LexError) => Err(LexError), ++ } ++} ++ ++fn punct_char(input: Cursor) -> PResult<char> { ++ if input.starts_with("//") || input.starts_with("/*") { ++ // Do not accept `/` of a comment as a punct. ++ return Err(LexError); ++ } ++ ++ let mut chars = input.chars(); ++ let first = match chars.next() { ++ Some(ch) => ch, ++ None => { ++ return Err(LexError); ++ } ++ }; ++ let recognized = "~!@#$%^&*-=+|;:,<.>/?'"; ++ if recognized.contains(first) { ++ Ok((input.advance(first.len_utf8()), first)) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn doc_comment(input: Cursor) -> PResult<Vec<TokenTree>> { ++ #[cfg(span_locations)] ++ let lo = input.off; ++ let (rest, (comment, inner)) = doc_comment_contents(input)?; ++ let span = crate::Span::_new_stable(Span { ++ #[cfg(span_locations)] ++ lo, ++ #[cfg(span_locations)] ++ hi: rest.off, ++ }); ++ ++ let mut scan_for_bare_cr = comment; ++ while let Some(cr) = scan_for_bare_cr.find('\r') { ++ let rest = &scan_for_bare_cr[cr + 1..]; ++ if !rest.starts_with('\n') { ++ return Err(LexError); ++ } ++ scan_for_bare_cr = rest; ++ } ++ ++ let mut trees = Vec::new(); ++ trees.push(TokenTree::Punct(Punct::new('#', Spacing::Alone))); ++ if inner { ++ trees.push(Punct::new('!', Spacing::Alone).into()); ++ } ++ let mut stream = vec![ ++ TokenTree::Ident(crate::Ident::new("doc", span)), ++ TokenTree::Punct(Punct::new('=', Spacing::Alone)), ++ TokenTree::Literal(crate::Literal::string(comment)), ++ ]; ++ for tt in stream.iter_mut() { ++ tt.set_span(span); ++ } ++ let group = Group::new(Delimiter::Bracket, stream.into_iter().collect()); ++ trees.push(crate::Group::_new_stable(group).into()); ++ for tt in trees.iter_mut() { ++ tt.set_span(span); ++ } ++ Ok((rest, trees)) ++} ++ ++fn doc_comment_contents(input: Cursor) -> PResult<(&str, bool)> { ++ if input.starts_with("//!") { ++ let input = input.advance(3); ++ let (input, s) = take_until_newline_or_eof(input); ++ Ok((input, (s, true))) ++ } else if input.starts_with("/*!") { ++ let (input, s) = block_comment(input)?; ++ Ok((input, (&s[3..s.len() - 2], true))) ++ } else if input.starts_with("///") { ++ let input = input.advance(3); ++ if input.starts_with("/") { ++ return Err(LexError); ++ } ++ let (input, s) = take_until_newline_or_eof(input); ++ Ok((input, (s, false))) ++ } else if input.starts_with("/**") && !input.rest[3..].starts_with('*') { ++ let (input, s) = block_comment(input)?; ++ Ok((input, (&s[3..s.len() - 2], false))) ++ } else { ++ Err(LexError) ++ } ++} ++ ++fn take_until_newline_or_eof(input: Cursor) -> (Cursor, &str) { ++ let chars = input.char_indices(); ++ ++ for (i, ch) in chars { ++ if ch == '\n' { ++ return (input.advance(i), &input.rest[..i]); ++ } else if ch == '\r' && input.rest[i + 1..].starts_with('\n') { ++ return (input.advance(i + 1), &input.rest[..i]); ++ } ++ } ++ ++ (input.advance(input.len()), input.rest) ++} +diff --git a/third_party/rust/proc-macro2/src/strnom.rs b/third_party/rust/proc-macro2/src/strnom.rs +deleted file mode 100644 +index eb7d0b8a8e..0000000000 +--- a/third_party/rust/proc-macro2/src/strnom.rs ++++ /dev/null +@@ -1,391 +0,0 @@ +-//! Adapted from [`nom`](https://github.com/Geal/nom). +- +-use crate::fallback::LexError; +-use std::str::{Bytes, CharIndices, Chars}; +-use unicode_xid::UnicodeXID; +- +-#[derive(Copy, Clone, Eq, PartialEq)] +-pub struct Cursor<'a> { +- pub rest: &'a str, +- #[cfg(span_locations)] +- pub off: u32, +-} +- +-impl<'a> Cursor<'a> { +- #[cfg(not(span_locations))] +- pub fn advance(&self, amt: usize) -> Cursor<'a> { +- Cursor { +- rest: &self.rest[amt..], +- } +- } +- #[cfg(span_locations)] +- pub fn advance(&self, amt: usize) -> Cursor<'a> { +- Cursor { +- rest: &self.rest[amt..], +- off: self.off + (amt as u32), +- } +- } +- +- pub fn find(&self, p: char) -> Option<usize> { +- self.rest.find(p) +- } +- +- pub fn starts_with(&self, s: &str) -> bool { +- self.rest.starts_with(s) +- } +- +- pub fn is_empty(&self) -> bool { +- self.rest.is_empty() +- } +- +- pub fn len(&self) -> usize { +- self.rest.len() +- } +- +- pub fn as_bytes(&self) -> &'a [u8] { +- self.rest.as_bytes() +- } +- +- pub fn bytes(&self) -> Bytes<'a> { +- self.rest.bytes() +- } +- +- pub fn chars(&self) -> Chars<'a> { +- self.rest.chars() +- } +- +- pub fn char_indices(&self) -> CharIndices<'a> { +- self.rest.char_indices() +- } +-} +- +-pub type PResult<'a, O> = Result<(Cursor<'a>, O), LexError>; +- +-pub fn whitespace(input: Cursor) -> PResult<()> { +- if input.is_empty() { +- return Err(LexError); +- } +- +- let bytes = input.as_bytes(); +- let mut i = 0; +- while i < bytes.len() { +- let s = input.advance(i); +- if bytes[i] == b'/' { +- if s.starts_with("//") +- && (!s.starts_with("///") || s.starts_with("////")) +- && !s.starts_with("//!") +- { +- if let Some(len) = s.find('\n') { +- i += len + 1; +- continue; +- } +- break; +- } else if s.starts_with("/**/") { +- i += 4; +- continue; +- } else if s.starts_with("/*") +- && (!s.starts_with("/**") || s.starts_with("/***")) +- && !s.starts_with("/*!") +- { +- let (_, com) = block_comment(s)?; +- i += com.len(); +- continue; +- } +- } +- match bytes[i] { +- b' ' | 0x09..=0x0d => { +- i += 1; +- continue; +- } +- b if b <= 0x7f => {} +- _ => { +- let ch = s.chars().next().unwrap(); +- if is_whitespace(ch) { +- i += ch.len_utf8(); +- continue; +- } +- } +- } +- return if i > 0 { Ok((s, ())) } else { Err(LexError) }; +- } +- Ok((input.advance(input.len()), ())) +-} +- +-pub fn block_comment(input: Cursor) -> PResult<&str> { +- if !input.starts_with("/*") { +- return Err(LexError); +- } +- +- let mut depth = 0; +- let bytes = input.as_bytes(); +- let mut i = 0; +- let upper = bytes.len() - 1; +- while i < upper { +- if bytes[i] == b'/' && bytes[i + 1] == b'*' { +- depth += 1; +- i += 1; // eat '*' +- } else if bytes[i] == b'*' && bytes[i + 1] == b'/' { +- depth -= 1; +- if depth == 0 { +- return Ok((input.advance(i + 2), &input.rest[..i + 2])); +- } +- i += 1; // eat '/' +- } +- i += 1; +- } +- Err(LexError) +-} +- +-pub fn skip_whitespace(input: Cursor) -> Cursor { +- match whitespace(input) { +- Ok((rest, _)) => rest, +- Err(LexError) => input, +- } +-} +- +-fn is_whitespace(ch: char) -> bool { +- // Rust treats left-to-right mark and right-to-left mark as whitespace +- ch.is_whitespace() || ch == '\u{200e}' || ch == '\u{200f}' +-} +- +-pub fn word_break(input: Cursor) -> PResult<()> { +- match input.chars().next() { +- Some(ch) if UnicodeXID::is_xid_continue(ch) => Err(LexError), +- Some(_) | None => Ok((input, ())), +- } +-} +- +-macro_rules! named { +- ($name:ident -> $o:ty, $submac:ident!( $($args:tt)* )) => { +- fn $name<'a>(i: Cursor<'a>) -> $crate::strnom::PResult<'a, $o> { +- $submac!(i, $($args)*) +- } +- }; +-} +- +-macro_rules! alt { +- ($i:expr, $e:ident | $($rest:tt)*) => { +- alt!($i, call!($e) | $($rest)*) +- }; +- +- ($i:expr, $subrule:ident!( $($args:tt)*) | $($rest:tt)*) => { +- match $subrule!($i, $($args)*) { +- res @ Ok(_) => res, +- _ => alt!($i, $($rest)*) +- } +- }; +- +- ($i:expr, $subrule:ident!( $($args:tt)* ) => { $gen:expr } | $($rest:tt)+) => { +- match $subrule!($i, $($args)*) { +- Ok((i, o)) => Ok((i, $gen(o))), +- Err(LexError) => alt!($i, $($rest)*) +- } +- }; +- +- ($i:expr, $e:ident => { $gen:expr } | $($rest:tt)*) => { +- alt!($i, call!($e) => { $gen } | $($rest)*) +- }; +- +- ($i:expr, $e:ident => { $gen:expr }) => { +- alt!($i, call!($e) => { $gen }) +- }; +- +- ($i:expr, $subrule:ident!( $($args:tt)* ) => { $gen:expr }) => { +- match $subrule!($i, $($args)*) { +- Ok((i, o)) => Ok((i, $gen(o))), +- Err(LexError) => Err(LexError), +- } +- }; +- +- ($i:expr, $e:ident) => { +- alt!($i, call!($e)) +- }; +- +- ($i:expr, $subrule:ident!( $($args:tt)*)) => { +- $subrule!($i, $($args)*) +- }; +-} +- +-macro_rules! do_parse { +- ($i:expr, ( $($rest:expr),* )) => { +- Ok(($i, ( $($rest),* ))) +- }; +- +- ($i:expr, $e:ident >> $($rest:tt)*) => { +- do_parse!($i, call!($e) >> $($rest)*) +- }; +- +- ($i:expr, $submac:ident!( $($args:tt)* ) >> $($rest:tt)*) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, _)) => do_parse!(i, $($rest)*), +- } +- }; +- +- ($i:expr, $field:ident : $e:ident >> $($rest:tt)*) => { +- do_parse!($i, $field: call!($e) >> $($rest)*) +- }; +- +- ($i:expr, $field:ident : $submac:ident!( $($args:tt)* ) >> $($rest:tt)*) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, o)) => { +- let $field = o; +- do_parse!(i, $($rest)*) +- }, +- } +- }; +-} +- +-macro_rules! peek { +- ($i:expr, $submac:ident!( $($args:tt)* )) => { +- match $submac!($i, $($args)*) { +- Ok((_, o)) => Ok(($i, o)), +- Err(LexError) => Err(LexError), +- } +- }; +-} +- +-macro_rules! call { +- ($i:expr, $fun:expr $(, $args:expr)*) => { +- $fun($i $(, $args)*) +- }; +-} +- +-macro_rules! option { +- ($i:expr, $f:expr) => { +- match $f($i) { +- Ok((i, o)) => Ok((i, Some(o))), +- Err(LexError) => Ok(($i, None)), +- } +- }; +-} +- +-macro_rules! take_until_newline_or_eof { +- ($i:expr,) => {{ +- if $i.len() == 0 { +- Ok(($i, "")) +- } else { +- match $i.find('\n') { +- Some(i) => Ok(($i.advance(i), &$i.rest[..i])), +- None => Ok(($i.advance($i.len()), &$i.rest[..$i.len()])), +- } +- } +- }}; +-} +- +-macro_rules! tuple { +- ($i:expr, $($rest:tt)*) => { +- tuple_parser!($i, (), $($rest)*) +- }; +-} +- +-/// Do not use directly. Use `tuple!`. +-macro_rules! tuple_parser { +- ($i:expr, ($($parsed:tt),*), $e:ident, $($rest:tt)*) => { +- tuple_parser!($i, ($($parsed),*), call!($e), $($rest)*) +- }; +- +- ($i:expr, (), $submac:ident!( $($args:tt)* ), $($rest:tt)*) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, o)) => tuple_parser!(i, (o), $($rest)*), +- } +- }; +- +- ($i:expr, ($($parsed:tt)*), $submac:ident!( $($args:tt)* ), $($rest:tt)*) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, o)) => tuple_parser!(i, ($($parsed)* , o), $($rest)*), +- } +- }; +- +- ($i:expr, ($($parsed:tt),*), $e:ident) => { +- tuple_parser!($i, ($($parsed),*), call!($e)) +- }; +- +- ($i:expr, (), $submac:ident!( $($args:tt)* )) => { +- $submac!($i, $($args)*) +- }; +- +- ($i:expr, ($($parsed:expr),*), $submac:ident!( $($args:tt)* )) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, o)) => Ok((i, ($($parsed),*, o))) +- } +- }; +- +- ($i:expr, ($($parsed:expr),*)) => { +- Ok(($i, ($($parsed),*))) +- }; +-} +- +-macro_rules! not { +- ($i:expr, $submac:ident!( $($args:tt)* )) => { +- match $submac!($i, $($args)*) { +- Ok((_, _)) => Err(LexError), +- Err(LexError) => Ok(($i, ())), +- } +- }; +-} +- +-macro_rules! tag { +- ($i:expr, $tag:expr) => { +- if $i.starts_with($tag) { +- Ok(($i.advance($tag.len()), &$i.rest[..$tag.len()])) +- } else { +- Err(LexError) +- } +- }; +-} +- +-macro_rules! punct { +- ($i:expr, $punct:expr) => { +- $crate::strnom::punct($i, $punct) +- }; +-} +- +-/// Do not use directly. Use `punct!`. +-pub fn punct<'a>(input: Cursor<'a>, token: &'static str) -> PResult<'a, &'a str> { +- let input = skip_whitespace(input); +- if input.starts_with(token) { +- Ok((input.advance(token.len()), token)) +- } else { +- Err(LexError) +- } +-} +- +-macro_rules! preceded { +- ($i:expr, $submac:ident!( $($args:tt)* ), $submac2:ident!( $($args2:tt)* )) => { +- match tuple!($i, $submac!($($args)*), $submac2!($($args2)*)) { +- Ok((remaining, (_, o))) => Ok((remaining, o)), +- Err(LexError) => Err(LexError), +- } +- }; +- +- ($i:expr, $submac:ident!( $($args:tt)* ), $g:expr) => { +- preceded!($i, $submac!($($args)*), call!($g)) +- }; +-} +- +-macro_rules! delimited { +- ($i:expr, $submac:ident!( $($args:tt)* ), $($rest:tt)+) => { +- match tuple_parser!($i, (), $submac!($($args)*), $($rest)*) { +- Err(LexError) => Err(LexError), +- Ok((i1, (_, o, _))) => Ok((i1, o)) +- } +- }; +-} +- +-macro_rules! map { +- ($i:expr, $submac:ident!( $($args:tt)* ), $g:expr) => { +- match $submac!($i, $($args)*) { +- Err(LexError) => Err(LexError), +- Ok((i, o)) => Ok((i, call!(o, $g))) +- } +- }; +- +- ($i:expr, $f:expr, $g:expr) => { +- map!($i, call!($f), $g) +- }; +-} +diff --git a/third_party/rust/proc-macro2/src/wrapper.rs b/third_party/rust/proc-macro2/src/wrapper.rs +index 552b9381cf..3df044af17 100644 +--- a/third_party/rust/proc-macro2/src/wrapper.rs ++++ b/third_party/rust/proc-macro2/src/wrapper.rs +@@ -1,15 +1,15 @@ +-use std::fmt; +-use std::iter; ++use crate::detection::inside_proc_macro; ++use crate::{fallback, Delimiter, Punct, Spacing, TokenTree}; ++use std::fmt::{self, Debug, Display}; ++use std::iter::FromIterator; + use std::ops::RangeBounds; +-use std::panic::{self, PanicInfo}; ++use std::panic; + #[cfg(super_unstable)] + use std::path::PathBuf; + use std::str::FromStr; + +-use crate::{fallback, Delimiter, Punct, Spacing, TokenTree}; +- + #[derive(Clone)] +-pub enum TokenStream { ++pub(crate) enum TokenStream { + Compiler(DeferredTokenStream), + Fallback(fallback::TokenStream), + } +@@ -19,73 +19,16 @@ pub enum TokenStream { + // we hold on to the appended tokens and do proc_macro::TokenStream::extend as + // late as possible to batch together consecutive uses of the Extend impl. + #[derive(Clone)] +-pub struct DeferredTokenStream { ++pub(crate) struct DeferredTokenStream { + stream: proc_macro::TokenStream, + extra: Vec<proc_macro::TokenTree>, + } + +-pub enum LexError { ++pub(crate) enum LexError { + Compiler(proc_macro::LexError), + Fallback(fallback::LexError), + } + +-fn nightly_works() -> bool { +- use std::sync::atomic::*; +- use std::sync::Once; +- +- static WORKS: AtomicUsize = AtomicUsize::new(0); +- static INIT: Once = Once::new(); +- +- match WORKS.load(Ordering::SeqCst) { +- 1 => return false, +- 2 => return true, +- _ => {} +- } +- +- // Swap in a null panic hook to avoid printing "thread panicked" to stderr, +- // then use catch_unwind to determine whether the compiler's proc_macro is +- // working. When proc-macro2 is used from outside of a procedural macro all +- // of the proc_macro crate's APIs currently panic. +- // +- // The Once is to prevent the possibility of this ordering: +- // +- // thread 1 calls take_hook, gets the user's original hook +- // thread 1 calls set_hook with the null hook +- // thread 2 calls take_hook, thinks null hook is the original hook +- // thread 2 calls set_hook with the null hook +- // thread 1 calls set_hook with the actual original hook +- // thread 2 calls set_hook with what it thinks is the original hook +- // +- // in which the user's hook has been lost. +- // +- // There is still a race condition where a panic in a different thread can +- // happen during the interval that the user's original panic hook is +- // unregistered such that their hook is incorrectly not called. This is +- // sufficiently unlikely and less bad than printing panic messages to stderr +- // on correct use of this crate. Maybe there is a libstd feature request +- // here. For now, if a user needs to guarantee that this failure mode does +- // not occur, they need to call e.g. `proc_macro2::Span::call_site()` from +- // the main thread before launching any other threads. +- INIT.call_once(|| { +- type PanicHook = dyn Fn(&PanicInfo) + Sync + Send + 'static; +- +- let null_hook: Box<PanicHook> = Box::new(|_panic_info| { /* ignore */ }); +- let sanity_check = &*null_hook as *const PanicHook; +- let original_hook = panic::take_hook(); +- panic::set_hook(null_hook); +- +- let works = panic::catch_unwind(|| proc_macro::Span::call_site()).is_ok(); +- WORKS.store(works as usize + 1, Ordering::SeqCst); +- +- let hopefully_null_hook = panic::take_hook(); +- panic::set_hook(original_hook); +- if sanity_check != &*hopefully_null_hook { +- panic!("observed race condition in proc_macro2::nightly_works"); +- } +- }); +- nightly_works() +-} +- + fn mismatch() -> ! { + panic!("stable/nightly mismatch") + } +@@ -103,7 +46,12 @@ impl DeferredTokenStream { + } + + fn evaluate_now(&mut self) { +- self.stream.extend(self.extra.drain(..)); ++ // If-check provides a fast short circuit for the common case of `extra` ++ // being empty, which saves a round trip over the proc macro bridge. ++ // Improves macro expansion time in winrt by 6% in debug mode. ++ if !self.extra.is_empty() { ++ self.stream.extend(self.extra.drain(..)); ++ } + } + + fn into_token_stream(mut self) -> proc_macro::TokenStream { +@@ -114,7 +62,7 @@ impl DeferredTokenStream { + + impl TokenStream { + pub fn new() -> TokenStream { +- if nightly_works() { ++ if inside_proc_macro() { + TokenStream::Compiler(DeferredTokenStream::new(proc_macro::TokenStream::new())) + } else { + TokenStream::Fallback(fallback::TokenStream::new()) +@@ -147,9 +95,9 @@ impl FromStr for TokenStream { + type Err = LexError; + + fn from_str(src: &str) -> Result<TokenStream, LexError> { +- if nightly_works() { ++ if inside_proc_macro() { + Ok(TokenStream::Compiler(DeferredTokenStream::new( +- src.parse()?, ++ proc_macro_parse(src)?, + ))) + } else { + Ok(TokenStream::Fallback(src.parse()?)) +@@ -157,11 +105,17 @@ impl FromStr for TokenStream { + } + } + +-impl fmt::Display for TokenStream { ++// Work around https://github.com/rust-lang/rust/issues/58736. ++fn proc_macro_parse(src: &str) -> Result<proc_macro::TokenStream, LexError> { ++ panic::catch_unwind(|| src.parse().map_err(LexError::Compiler)) ++ .unwrap_or(Err(LexError::Fallback(fallback::LexError))) ++} ++ ++impl Display for TokenStream { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- TokenStream::Compiler(tts) => tts.clone().into_token_stream().fmt(f), +- TokenStream::Fallback(tts) => tts.fmt(f), ++ TokenStream::Compiler(tts) => Display::fmt(&tts.clone().into_token_stream(), f), ++ TokenStream::Fallback(tts) => Display::fmt(tts, f), + } + } + } +@@ -187,7 +141,7 @@ impl From<fallback::TokenStream> for TokenStream { + } + } + +-// Assumes nightly_works(). ++// Assumes inside_proc_macro(). + fn into_compiler_token(token: TokenTree) -> proc_macro::TokenTree { + match token { + TokenTree::Group(tt) => tt.inner.unwrap_nightly().into(), +@@ -196,9 +150,9 @@ fn into_compiler_token(token: TokenTree) -> proc_macro::TokenTree { + Spacing::Joint => proc_macro::Spacing::Joint, + Spacing::Alone => proc_macro::Spacing::Alone, + }; +- let mut op = proc_macro::Punct::new(tt.as_char(), spacing); +- op.set_span(tt.span().inner.unwrap_nightly()); +- op.into() ++ let mut punct = proc_macro::Punct::new(tt.as_char(), spacing); ++ punct.set_span(tt.span().inner.unwrap_nightly()); ++ punct.into() + } + TokenTree::Ident(tt) => tt.inner.unwrap_nightly().into(), + TokenTree::Literal(tt) => tt.inner.unwrap_nightly().into(), +@@ -207,7 +161,7 @@ fn into_compiler_token(token: TokenTree) -> proc_macro::TokenTree { + + impl From<TokenTree> for TokenStream { + fn from(token: TokenTree) -> TokenStream { +- if nightly_works() { ++ if inside_proc_macro() { + TokenStream::Compiler(DeferredTokenStream::new(into_compiler_token(token).into())) + } else { + TokenStream::Fallback(token.into()) +@@ -215,9 +169,9 @@ impl From<TokenTree> for TokenStream { + } + } + +-impl iter::FromIterator<TokenTree> for TokenStream { ++impl FromIterator<TokenTree> for TokenStream { + fn from_iter<I: IntoIterator<Item = TokenTree>>(trees: I) -> Self { +- if nightly_works() { ++ if inside_proc_macro() { + TokenStream::Compiler(DeferredTokenStream::new( + trees.into_iter().map(into_compiler_token).collect(), + )) +@@ -227,7 +181,7 @@ impl iter::FromIterator<TokenTree> for TokenStream { + } + } + +-impl iter::FromIterator<TokenStream> for TokenStream { ++impl FromIterator<TokenStream> for TokenStream { + fn from_iter<I: IntoIterator<Item = TokenStream>>(streams: I) -> Self { + let mut streams = streams.into_iter(); + match streams.next() { +@@ -252,14 +206,15 @@ impl iter::FromIterator<TokenStream> for TokenStream { + } + + impl Extend<TokenTree> for TokenStream { +- fn extend<I: IntoIterator<Item = TokenTree>>(&mut self, streams: I) { ++ fn extend<I: IntoIterator<Item = TokenTree>>(&mut self, stream: I) { + match self { + TokenStream::Compiler(tts) => { + // Here is the reason for DeferredTokenStream. +- tts.extra +- .extend(streams.into_iter().map(into_compiler_token)); ++ for token in stream { ++ tts.extra.push(into_compiler_token(token)); ++ } + } +- TokenStream::Fallback(tts) => tts.extend(streams), ++ TokenStream::Fallback(tts) => tts.extend(stream), + } + } + } +@@ -270,20 +225,20 @@ impl Extend<TokenStream> for TokenStream { + TokenStream::Compiler(tts) => { + tts.evaluate_now(); + tts.stream +- .extend(streams.into_iter().map(|stream| stream.unwrap_nightly())); ++ .extend(streams.into_iter().map(TokenStream::unwrap_nightly)); + } + TokenStream::Fallback(tts) => { +- tts.extend(streams.into_iter().map(|stream| stream.unwrap_stable())); ++ tts.extend(streams.into_iter().map(TokenStream::unwrap_stable)); + } + } + } + } + +-impl fmt::Debug for TokenStream { ++impl Debug for TokenStream { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- TokenStream::Compiler(tts) => tts.clone().into_token_stream().fmt(f), +- TokenStream::Fallback(tts) => tts.fmt(f), ++ TokenStream::Compiler(tts) => Debug::fmt(&tts.clone().into_token_stream(), f), ++ TokenStream::Fallback(tts) => Debug::fmt(tts, f), + } + } + } +@@ -300,17 +255,29 @@ impl From<fallback::LexError> for LexError { + } + } + +-impl fmt::Debug for LexError { ++impl Debug for LexError { ++ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ LexError::Compiler(e) => Debug::fmt(e, f), ++ LexError::Fallback(e) => Debug::fmt(e, f), ++ } ++ } ++} ++ ++impl Display for LexError { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- LexError::Compiler(e) => e.fmt(f), +- LexError::Fallback(e) => e.fmt(f), ++ #[cfg(lexerror_display)] ++ LexError::Compiler(e) => Display::fmt(e, f), ++ #[cfg(not(lexerror_display))] ++ LexError::Compiler(_e) => Display::fmt(&fallback::LexError, f), ++ LexError::Fallback(e) => Display::fmt(e, f), + } + } + } + + #[derive(Clone)] +-pub enum TokenTreeIter { ++pub(crate) enum TokenTreeIter { + Compiler(proc_macro::token_stream::IntoIter), + Fallback(fallback::TokenTreeIter), + } +@@ -361,7 +328,7 @@ impl Iterator for TokenTreeIter { + } + } + +-impl fmt::Debug for TokenTreeIter { ++impl Debug for TokenTreeIter { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + f.debug_struct("TokenTreeIter").finish() + } +@@ -369,7 +336,7 @@ impl fmt::Debug for TokenTreeIter { + + #[derive(Clone, PartialEq, Eq)] + #[cfg(super_unstable)] +-pub enum SourceFile { ++pub(crate) enum SourceFile { + Compiler(proc_macro::SourceFile), + Fallback(fallback::SourceFile), + } +@@ -397,58 +364,77 @@ impl SourceFile { + } + + #[cfg(super_unstable)] +-impl fmt::Debug for SourceFile { ++impl Debug for SourceFile { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- SourceFile::Compiler(a) => a.fmt(f), +- SourceFile::Fallback(a) => a.fmt(f), ++ SourceFile::Compiler(a) => Debug::fmt(a, f), ++ SourceFile::Fallback(a) => Debug::fmt(a, f), + } + } + } + + #[cfg(any(super_unstable, feature = "span-locations"))] +-pub struct LineColumn { ++pub(crate) struct LineColumn { + pub line: usize, + pub column: usize, + } + + #[derive(Copy, Clone)] +-pub enum Span { ++pub(crate) enum Span { + Compiler(proc_macro::Span), + Fallback(fallback::Span), + } + + impl Span { + pub fn call_site() -> Span { +- if nightly_works() { ++ if inside_proc_macro() { + Span::Compiler(proc_macro::Span::call_site()) + } else { + Span::Fallback(fallback::Span::call_site()) + } + } + ++ #[cfg(hygiene)] ++ pub fn mixed_site() -> Span { ++ if inside_proc_macro() { ++ Span::Compiler(proc_macro::Span::mixed_site()) ++ } else { ++ Span::Fallback(fallback::Span::mixed_site()) ++ } ++ } ++ + #[cfg(super_unstable)] + pub fn def_site() -> Span { +- if nightly_works() { ++ if inside_proc_macro() { + Span::Compiler(proc_macro::Span::def_site()) + } else { + Span::Fallback(fallback::Span::def_site()) + } + } + +- #[cfg(super_unstable)] + pub fn resolved_at(&self, other: Span) -> Span { + match (self, other) { ++ #[cfg(hygiene)] + (Span::Compiler(a), Span::Compiler(b)) => Span::Compiler(a.resolved_at(b)), ++ ++ // Name resolution affects semantics, but location is only cosmetic ++ #[cfg(not(hygiene))] ++ (Span::Compiler(_), Span::Compiler(_)) => other, ++ + (Span::Fallback(a), Span::Fallback(b)) => Span::Fallback(a.resolved_at(b)), + _ => mismatch(), + } + } + +- #[cfg(super_unstable)] + pub fn located_at(&self, other: Span) -> Span { + match (self, other) { ++ #[cfg(hygiene)] + (Span::Compiler(a), Span::Compiler(b)) => Span::Compiler(a.located_at(b)), ++ ++ // Name resolution affects semantics, but location is only cosmetic ++ #[cfg(not(hygiene))] ++ (Span::Compiler(_), Span::Compiler(_)) => *self, ++ + (Span::Fallback(a), Span::Fallback(b)) => Span::Fallback(a.located_at(b)), + _ => mismatch(), + } +@@ -542,16 +528,16 @@ impl From<fallback::Span> for Span { + } + } + +-impl fmt::Debug for Span { ++impl Debug for Span { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- Span::Compiler(s) => s.fmt(f), +- Span::Fallback(s) => s.fmt(f), ++ Span::Compiler(s) => Debug::fmt(s, f), ++ Span::Fallback(s) => Debug::fmt(s, f), + } + } + } + +-pub fn debug_span_field_if_nontrivial(debug: &mut fmt::DebugStruct, span: Span) { ++pub(crate) fn debug_span_field_if_nontrivial(debug: &mut fmt::DebugStruct, span: Span) { + match span { + Span::Compiler(s) => { + debug.field("span", &s); +@@ -561,7 +547,7 @@ pub fn debug_span_field_if_nontrivial(debug: &mut fmt::DebugStruct, span: Span) + } + + #[derive(Clone)] +-pub enum Group { ++pub(crate) enum Group { + Compiler(proc_macro::Group), + Fallback(fallback::Group), + } +@@ -652,26 +638,26 @@ impl From<fallback::Group> for Group { + } + } + +-impl fmt::Display for Group { ++impl Display for Group { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + match self { +- Group::Compiler(group) => group.fmt(formatter), +- Group::Fallback(group) => group.fmt(formatter), ++ Group::Compiler(group) => Display::fmt(group, formatter), ++ Group::Fallback(group) => Display::fmt(group, formatter), + } + } + } + +-impl fmt::Debug for Group { ++impl Debug for Group { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + match self { +- Group::Compiler(group) => group.fmt(formatter), +- Group::Fallback(group) => group.fmt(formatter), ++ Group::Compiler(group) => Debug::fmt(group, formatter), ++ Group::Fallback(group) => Debug::fmt(group, formatter), + } + } + } + + #[derive(Clone)] +-pub enum Ident { ++pub(crate) enum Ident { + Compiler(proc_macro::Ident), + Fallback(fallback::Ident), + } +@@ -747,26 +733,26 @@ where + } + } + +-impl fmt::Display for Ident { ++impl Display for Ident { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- Ident::Compiler(t) => t.fmt(f), +- Ident::Fallback(t) => t.fmt(f), ++ Ident::Compiler(t) => Display::fmt(t, f), ++ Ident::Fallback(t) => Display::fmt(t, f), + } + } + } + +-impl fmt::Debug for Ident { ++impl Debug for Ident { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- Ident::Compiler(t) => t.fmt(f), +- Ident::Fallback(t) => t.fmt(f), ++ Ident::Compiler(t) => Debug::fmt(t, f), ++ Ident::Fallback(t) => Debug::fmt(t, f), + } + } + } + + #[derive(Clone)] +-pub enum Literal { ++pub(crate) enum Literal { + Compiler(proc_macro::Literal), + Fallback(fallback::Literal), + } +@@ -774,7 +760,7 @@ pub enum Literal { + macro_rules! suffixed_numbers { + ($($name:ident => $kind:ident,)*) => ($( + pub fn $name(n: $kind) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::$name(n)) + } else { + Literal::Fallback(fallback::Literal::$name(n)) +@@ -786,7 +772,7 @@ macro_rules! suffixed_numbers { + macro_rules! unsuffixed_integers { + ($($name:ident => $kind:ident,)*) => ($( + pub fn $name(n: $kind) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::$name(n)) + } else { + Literal::Fallback(fallback::Literal::$name(n)) +@@ -830,7 +816,7 @@ impl Literal { + } + + pub fn f32_unsuffixed(f: f32) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::f32_unsuffixed(f)) + } else { + Literal::Fallback(fallback::Literal::f32_unsuffixed(f)) +@@ -838,7 +824,7 @@ impl Literal { + } + + pub fn f64_unsuffixed(f: f64) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::f64_unsuffixed(f)) + } else { + Literal::Fallback(fallback::Literal::f64_unsuffixed(f)) +@@ -846,7 +832,7 @@ impl Literal { + } + + pub fn string(t: &str) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::string(t)) + } else { + Literal::Fallback(fallback::Literal::string(t)) +@@ -854,7 +840,7 @@ impl Literal { + } + + pub fn character(t: char) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::character(t)) + } else { + Literal::Fallback(fallback::Literal::character(t)) +@@ -862,7 +848,7 @@ impl Literal { + } + + pub fn byte_string(bytes: &[u8]) -> Literal { +- if nightly_works() { ++ if inside_proc_macro() { + Literal::Compiler(proc_macro::Literal::byte_string(bytes)) + } else { + Literal::Fallback(fallback::Literal::byte_string(bytes)) +@@ -908,20 +894,20 @@ impl From<fallback::Literal> for Literal { + } + } + +-impl fmt::Display for Literal { ++impl Display for Literal { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- Literal::Compiler(t) => t.fmt(f), +- Literal::Fallback(t) => t.fmt(f), ++ Literal::Compiler(t) => Display::fmt(t, f), ++ Literal::Fallback(t) => Display::fmt(t, f), + } + } + } + +-impl fmt::Debug for Literal { ++impl Debug for Literal { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + match self { +- Literal::Compiler(t) => t.fmt(f), +- Literal::Fallback(t) => t.fmt(f), ++ Literal::Compiler(t) => Debug::fmt(t, f), ++ Literal::Fallback(t) => Debug::fmt(t, f), + } + } + } +diff --git a/third_party/rust/proc-macro2/tests/comments.rs b/third_party/rust/proc-macro2/tests/comments.rs +new file mode 100644 +index 0000000000..708cccb880 +--- /dev/null ++++ b/third_party/rust/proc-macro2/tests/comments.rs +@@ -0,0 +1,103 @@ ++use proc_macro2::{Delimiter, Literal, Spacing, TokenStream, TokenTree}; ++ ++// #[doc = "..."] -> "..." ++fn lit_of_outer_doc_comment(tokens: TokenStream) -> Literal { ++ lit_of_doc_comment(tokens, false) ++} ++ ++// #![doc = "..."] -> "..." ++fn lit_of_inner_doc_comment(tokens: TokenStream) -> Literal { ++ lit_of_doc_comment(tokens, true) ++} ++ ++fn lit_of_doc_comment(tokens: TokenStream, inner: bool) -> Literal { ++ let mut iter = tokens.clone().into_iter(); ++ match iter.next().unwrap() { ++ TokenTree::Punct(punct) => { ++ assert_eq!(punct.as_char(), '#'); ++ assert_eq!(punct.spacing(), Spacing::Alone); ++ } ++ _ => panic!("wrong token {:?}", tokens), ++ } ++ if inner { ++ match iter.next().unwrap() { ++ TokenTree::Punct(punct) => { ++ assert_eq!(punct.as_char(), '!'); ++ assert_eq!(punct.spacing(), Spacing::Alone); ++ } ++ _ => panic!("wrong token {:?}", tokens), ++ } ++ } ++ iter = match iter.next().unwrap() { ++ TokenTree::Group(group) => { ++ assert_eq!(group.delimiter(), Delimiter::Bracket); ++ assert!(iter.next().is_none(), "unexpected token {:?}", tokens); ++ group.stream().into_iter() ++ } ++ _ => panic!("wrong token {:?}", tokens), ++ }; ++ match iter.next().unwrap() { ++ TokenTree::Ident(ident) => assert_eq!(ident.to_string(), "doc"), ++ _ => panic!("wrong token {:?}", tokens), ++ } ++ match iter.next().unwrap() { ++ TokenTree::Punct(punct) => { ++ assert_eq!(punct.as_char(), '='); ++ assert_eq!(punct.spacing(), Spacing::Alone); ++ } ++ _ => panic!("wrong token {:?}", tokens), ++ } ++ match iter.next().unwrap() { ++ TokenTree::Literal(literal) => { ++ assert!(iter.next().is_none(), "unexpected token {:?}", tokens); ++ literal ++ } ++ _ => panic!("wrong token {:?}", tokens), ++ } ++} ++ ++#[test] ++fn closed_immediately() { ++ let stream = "/**/".parse::<TokenStream>().unwrap(); ++ let tokens = stream.into_iter().collect::<Vec<_>>(); ++ assert!(tokens.is_empty(), "not empty -- {:?}", tokens); ++} ++ ++#[test] ++fn incomplete() { ++ assert!("/*/".parse::<TokenStream>().is_err()); ++} ++ ++#[test] ++fn lit() { ++ let stream = "/// doc".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_outer_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\" doc\""); ++ ++ let stream = "//! doc".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_inner_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\" doc\""); ++ ++ let stream = "/** doc */".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_outer_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\" doc \""); ++ ++ let stream = "/*! doc */".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_inner_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\" doc \""); ++} ++ ++#[test] ++fn carriage_return() { ++ let stream = "///\r\n".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_outer_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\"\""); ++ ++ let stream = "/**\r\n*/".parse::<TokenStream>().unwrap(); ++ let lit = lit_of_outer_doc_comment(stream); ++ assert_eq!(lit.to_string(), "\"\\r\\n\""); ++ ++ "///\r".parse::<TokenStream>().unwrap_err(); ++ "///\r \n".parse::<TokenStream>().unwrap_err(); ++ "/**\r \n*/".parse::<TokenStream>().unwrap_err(); ++} +diff --git a/third_party/rust/proc-macro2/tests/marker.rs b/third_party/rust/proc-macro2/tests/marker.rs +index 7af2539c1a..70e57677cd 100644 +--- a/third_party/rust/proc-macro2/tests/marker.rs ++++ b/third_party/rust/proc-macro2/tests/marker.rs +@@ -57,3 +57,36 @@ mod semver_exempt { + + assert_impl!(SourceFile is not Send or Sync); + } ++ ++#[cfg(not(no_libprocmacro_unwind_safe))] ++mod unwind_safe { ++ use super::*; ++ use std::panic::{RefUnwindSafe, UnwindSafe}; ++ ++ macro_rules! assert_unwind_safe { ++ ($($types:ident)*) => { ++ $( ++ assert_impl!($types is UnwindSafe and RefUnwindSafe); ++ )* ++ }; ++ } ++ ++ assert_unwind_safe! { ++ Delimiter ++ Group ++ Ident ++ LexError ++ Literal ++ Punct ++ Spacing ++ Span ++ TokenStream ++ TokenTree ++ } ++ ++ #[cfg(procmacro2_semver_exempt)] ++ assert_unwind_safe! { ++ LineColumn ++ SourceFile ++ } ++} +diff --git a/third_party/rust/proc-macro2/tests/test.rs b/third_party/rust/proc-macro2/tests/test.rs +index 7528388138..1e9f633944 100644 +--- a/third_party/rust/proc-macro2/tests/test.rs ++++ b/third_party/rust/proc-macro2/tests/test.rs +@@ -1,7 +1,6 @@ ++use proc_macro2::{Ident, Literal, Punct, Spacing, Span, TokenStream, TokenTree}; + use std::str::{self, FromStr}; + +-use proc_macro2::{Ident, Literal, Spacing, Span, TokenStream, TokenTree}; +- + #[test] + fn idents() { + assert_eq!( +@@ -84,6 +83,11 @@ fn literal_string() { + assert_eq!(Literal::string("didn't").to_string(), "\"didn't\""); + } + ++#[test] ++fn literal_raw_string() { ++ "r\"\r\n\"".parse::<TokenStream>().unwrap(); ++} ++ + #[test] + fn literal_character() { + assert_eq!(Literal::character('x').to_string(), "'x'"); +@@ -110,6 +114,37 @@ fn literal_suffix() { + assert_eq!(token_count("1._0"), 3); + assert_eq!(token_count("1._m"), 3); + assert_eq!(token_count("\"\"s"), 1); ++ assert_eq!(token_count("r\"\"r"), 1); ++ assert_eq!(token_count("b\"\"b"), 1); ++ assert_eq!(token_count("br\"\"br"), 1); ++ assert_eq!(token_count("r#\"\"#r"), 1); ++ assert_eq!(token_count("'c'c"), 1); ++ assert_eq!(token_count("b'b'b"), 1); ++ assert_eq!(token_count("0E"), 1); ++ assert_eq!(token_count("0o0A"), 1); ++ assert_eq!(token_count("0E--0"), 4); ++ assert_eq!(token_count("0.0ECMA"), 1); ++} ++ ++#[test] ++fn literal_iter_negative() { ++ let negative_literal = Literal::i32_suffixed(-3); ++ let tokens = TokenStream::from(TokenTree::Literal(negative_literal)); ++ let mut iter = tokens.into_iter(); ++ match iter.next().unwrap() { ++ TokenTree::Punct(punct) => { ++ assert_eq!(punct.as_char(), '-'); ++ assert_eq!(punct.spacing(), Spacing::Alone); ++ } ++ unexpected => panic!("unexpected token {:?}", unexpected), ++ } ++ match iter.next().unwrap() { ++ TokenTree::Literal(literal) => { ++ assert_eq!(literal.to_string(), "3i32"); ++ } ++ unexpected => panic!("unexpected token {:?}", unexpected), ++ } ++ assert!(iter.next().is_none()); + } + + #[test] +@@ -161,41 +196,21 @@ fn fail() { + fail("' static"); + fail("r#1"); + fail("r#_"); ++ fail("\"\\u{0000000}\""); // overlong unicode escape (rust allows at most 6 hex digits) ++ fail("\"\\u{999999}\""); // outside of valid range of char ++ fail("\"\\u{_0}\""); // leading underscore ++ fail("\"\\u{}\""); // empty ++ fail("b\"\r\""); // bare carriage return in byte string ++ fail("r\"\r\""); // bare carriage return in raw string ++ fail("\"\\\r \""); // backslash carriage return ++ fail("'aa'aa"); ++ fail("br##\"\"#"); ++ fail("\"\\\n\u{85}\r\""); + } + + #[cfg(span_locations)] + #[test] + fn span_test() { +- use proc_macro2::TokenTree; +- +- fn check_spans(p: &str, mut lines: &[(usize, usize, usize, usize)]) { +- let ts = p.parse::<TokenStream>().unwrap(); +- check_spans_internal(ts, &mut lines); +- } +- +- fn check_spans_internal(ts: TokenStream, lines: &mut &[(usize, usize, usize, usize)]) { +- for i in ts { +- if let Some((&(sline, scol, eline, ecol), rest)) = lines.split_first() { +- *lines = rest; +- +- let start = i.span().start(); +- assert_eq!(start.line, sline, "sline did not match for {}", i); +- assert_eq!(start.column, scol, "scol did not match for {}", i); +- +- let end = i.span().end(); +- assert_eq!(end.line, eline, "eline did not match for {}", i); +- assert_eq!(end.column, ecol, "ecol did not match for {}", i); +- +- match i { +- TokenTree::Group(ref g) => { +- check_spans_internal(g.stream().clone(), lines); +- } +- _ => {} +- } +- } +- } +- } +- + check_spans( + "\ + /// This is a document comment +@@ -274,53 +289,11 @@ fn span_join() { + #[test] + fn no_panic() { + let s = str::from_utf8(b"b\'\xc2\x86 \x00\x00\x00^\"").unwrap(); +- assert!(s.parse::<proc_macro2::TokenStream>().is_err()); ++ assert!(s.parse::<TokenStream>().is_err()); + } + + #[test] +-fn tricky_doc_comment() { +- let stream = "/**/".parse::<proc_macro2::TokenStream>().unwrap(); +- let tokens = stream.into_iter().collect::<Vec<_>>(); +- assert!(tokens.is_empty(), "not empty -- {:?}", tokens); +- +- let stream = "/// doc".parse::<proc_macro2::TokenStream>().unwrap(); +- let tokens = stream.into_iter().collect::<Vec<_>>(); +- assert!(tokens.len() == 2, "not length 2 -- {:?}", tokens); +- match tokens[0] { +- proc_macro2::TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '#'), +- _ => panic!("wrong token {:?}", tokens[0]), +- } +- let mut tokens = match tokens[1] { +- proc_macro2::TokenTree::Group(ref tt) => { +- assert_eq!(tt.delimiter(), proc_macro2::Delimiter::Bracket); +- tt.stream().into_iter() +- } +- _ => panic!("wrong token {:?}", tokens[0]), +- }; +- +- match tokens.next().unwrap() { +- proc_macro2::TokenTree::Ident(ref tt) => assert_eq!(tt.to_string(), "doc"), +- t => panic!("wrong token {:?}", t), +- } +- match tokens.next().unwrap() { +- proc_macro2::TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '='), +- t => panic!("wrong token {:?}", t), +- } +- match tokens.next().unwrap() { +- proc_macro2::TokenTree::Literal(ref tt) => { +- assert_eq!(tt.to_string(), "\" doc\""); +- } +- t => panic!("wrong token {:?}", t), +- } +- assert!(tokens.next().is_none()); +- +- let stream = "//! doc".parse::<proc_macro2::TokenStream>().unwrap(); +- let tokens = stream.into_iter().collect::<Vec<_>>(); +- assert!(tokens.len() == 3, "not length 3 -- {:?}", tokens); +-} +- +-#[test] +-fn op_before_comment() { ++fn punct_before_comment() { + let mut tts = TokenStream::from_str("~// comment").unwrap().into_iter(); + match tts.next().unwrap() { + TokenTree::Punct(tt) => { +@@ -331,6 +304,22 @@ fn op_before_comment() { + } + } + ++#[test] ++fn joint_last_token() { ++ // This test verifies that we match the behavior of libproc_macro *not* in ++ // the range nightly-2020-09-06 through nightly-2020-09-10, in which this ++ // behavior was temporarily broken. ++ // See https://github.com/rust-lang/rust/issues/76399 ++ ++ let joint_punct = Punct::new(':', Spacing::Joint); ++ let stream = TokenStream::from(TokenTree::Punct(joint_punct)); ++ let punct = match stream.into_iter().next().unwrap() { ++ TokenTree::Punct(punct) => punct, ++ _ => unreachable!(), ++ }; ++ assert_eq!(punct.spacing(), Spacing::Joint); ++} ++ + #[test] + fn raw_identifier() { + let mut tts = TokenStream::from_str("r#dyn").unwrap().into_iter(); +@@ -345,11 +334,11 @@ fn raw_identifier() { + fn test_debug_ident() { + let ident = Ident::new("proc_macro", Span::call_site()); + +- #[cfg(not(procmacro2_semver_exempt))] ++ #[cfg(not(span_locations))] + let expected = "Ident(proc_macro)"; + +- #[cfg(procmacro2_semver_exempt)] +- let expected = "Ident { sym: proc_macro, span: bytes(0..0) }"; ++ #[cfg(span_locations)] ++ let expected = "Ident { sym: proc_macro }"; + + assert_eq!(expected, format!("{:?}", ident)); + } +@@ -358,7 +347,7 @@ fn test_debug_ident() { + fn test_debug_tokenstream() { + let tts = TokenStream::from_str("[a + 1]").unwrap(); + +- #[cfg(not(procmacro2_semver_exempt))] ++ #[cfg(not(span_locations))] + let expected = "\ + TokenStream [ + Group { +@@ -368,7 +357,7 @@ TokenStream [ + sym: a, + }, + Punct { +- op: '+', ++ char: '+', + spacing: Alone, + }, + Literal { +@@ -379,7 +368,7 @@ TokenStream [ + ]\ + "; + +- #[cfg(not(procmacro2_semver_exempt))] ++ #[cfg(not(span_locations))] + let expected_before_trailing_commas = "\ + TokenStream [ + Group { +@@ -389,7 +378,7 @@ TokenStream [ + sym: a + }, + Punct { +- op: '+', ++ char: '+', + spacing: Alone + }, + Literal { +@@ -400,7 +389,7 @@ TokenStream [ + ]\ + "; + +- #[cfg(procmacro2_semver_exempt)] ++ #[cfg(span_locations)] + let expected = "\ + TokenStream [ + Group { +@@ -411,7 +400,7 @@ TokenStream [ + span: bytes(2..3), + }, + Punct { +- op: '+', ++ char: '+', + spacing: Alone, + span: bytes(4..5), + }, +@@ -425,7 +414,7 @@ TokenStream [ + ]\ + "; + +- #[cfg(procmacro2_semver_exempt)] ++ #[cfg(span_locations)] + let expected_before_trailing_commas = "\ + TokenStream [ + Group { +@@ -436,7 +425,7 @@ TokenStream [ + span: bytes(2..3) + }, + Punct { +- op: '+', ++ char: '+', + spacing: Alone, + span: bytes(4..5) + }, +@@ -464,3 +453,80 @@ fn default_tokenstream_is_empty() { + + assert!(default_token_stream.is_empty()); + } ++ ++#[test] ++fn tuple_indexing() { ++ // This behavior may change depending on https://github.com/rust-lang/rust/pull/71322 ++ let mut tokens = "tuple.0.0".parse::<TokenStream>().unwrap().into_iter(); ++ assert_eq!("tuple", tokens.next().unwrap().to_string()); ++ assert_eq!(".", tokens.next().unwrap().to_string()); ++ assert_eq!("0.0", tokens.next().unwrap().to_string()); ++ assert!(tokens.next().is_none()); ++} ++ ++#[cfg(span_locations)] ++#[test] ++fn non_ascii_tokens() { ++ check_spans("// abc", &[]); ++ check_spans("// ábc", &[]); ++ check_spans("// abc x", &[]); ++ check_spans("// ábc x", &[]); ++ check_spans("/* abc */ x", &[(1, 10, 1, 11)]); ++ check_spans("/* ábc */ x", &[(1, 10, 1, 11)]); ++ check_spans("/* ab\nc */ x", &[(2, 5, 2, 6)]); ++ check_spans("/* áb\nc */ x", &[(2, 5, 2, 6)]); ++ check_spans("/*** abc */ x", &[(1, 12, 1, 13)]); ++ check_spans("/*** ábc */ x", &[(1, 12, 1, 13)]); ++ check_spans(r#""abc""#, &[(1, 0, 1, 5)]); ++ check_spans(r#""ábc""#, &[(1, 0, 1, 5)]); ++ check_spans(r###"r#"abc"#"###, &[(1, 0, 1, 8)]); ++ check_spans(r###"r#"ábc"#"###, &[(1, 0, 1, 8)]); ++ check_spans("r#\"a\nc\"#", &[(1, 0, 2, 3)]); ++ check_spans("r#\"á\nc\"#", &[(1, 0, 2, 3)]); ++ check_spans("'a'", &[(1, 0, 1, 3)]); ++ check_spans("'á'", &[(1, 0, 1, 3)]); ++ check_spans("//! abc", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]); ++ check_spans("//! ábc", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]); ++ check_spans("//! abc\n", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]); ++ check_spans("//! ábc\n", &[(1, 0, 1, 7), (1, 0, 1, 7), (1, 0, 1, 7)]); ++ check_spans("/*! abc */", &[(1, 0, 1, 10), (1, 0, 1, 10), (1, 0, 1, 10)]); ++ check_spans("/*! ábc */", &[(1, 0, 1, 10), (1, 0, 1, 10), (1, 0, 1, 10)]); ++ check_spans("/*! a\nc */", &[(1, 0, 2, 4), (1, 0, 2, 4), (1, 0, 2, 4)]); ++ check_spans("/*! á\nc */", &[(1, 0, 2, 4), (1, 0, 2, 4), (1, 0, 2, 4)]); ++ check_spans("abc", &[(1, 0, 1, 3)]); ++ check_spans("ábc", &[(1, 0, 1, 3)]); ++ check_spans("ábć", &[(1, 0, 1, 3)]); ++ check_spans("abc// foo", &[(1, 0, 1, 3)]); ++ check_spans("ábc// foo", &[(1, 0, 1, 3)]); ++ check_spans("ábć// foo", &[(1, 0, 1, 3)]); ++ check_spans("b\"a\\\n c\"", &[(1, 0, 2, 3)]); ++ check_spans("b\"a\\\n\u{00a0}c\"", &[(1, 0, 2, 3)]); ++} ++ ++#[cfg(span_locations)] ++fn check_spans(p: &str, mut lines: &[(usize, usize, usize, usize)]) { ++ let ts = p.parse::<TokenStream>().unwrap(); ++ check_spans_internal(ts, &mut lines); ++ assert!(lines.is_empty(), "leftover ranges: {:?}", lines); ++} ++ ++#[cfg(span_locations)] ++fn check_spans_internal(ts: TokenStream, lines: &mut &[(usize, usize, usize, usize)]) { ++ for i in ts { ++ if let Some((&(sline, scol, eline, ecol), rest)) = lines.split_first() { ++ *lines = rest; ++ ++ let start = i.span().start(); ++ assert_eq!(start.line, sline, "sline did not match for {}", i); ++ assert_eq!(start.column, scol, "scol did not match for {}", i); ++ ++ let end = i.span().end(); ++ assert_eq!(end.line, eline, "eline did not match for {}", i); ++ assert_eq!(end.column, ecol, "ecol did not match for {}", i); ++ ++ if let TokenTree::Group(g) = i { ++ check_spans_internal(g.stream().clone(), lines); ++ } ++ } ++ } ++} +diff --git a/third_party/rust/proc-macro2/tests/test_fmt.rs b/third_party/rust/proc-macro2/tests/test_fmt.rs +new file mode 100644 +index 0000000000..99a0aee5c8 +--- /dev/null ++++ b/third_party/rust/proc-macro2/tests/test_fmt.rs +@@ -0,0 +1,26 @@ ++use proc_macro2::{Delimiter, Group, Ident, Span, TokenStream, TokenTree}; ++use std::iter::{self, FromIterator}; ++ ++#[test] ++fn test_fmt_group() { ++ let ident = Ident::new("x", Span::call_site()); ++ let inner = TokenStream::from_iter(iter::once(TokenTree::Ident(ident))); ++ let parens_empty = Group::new(Delimiter::Parenthesis, TokenStream::new()); ++ let parens_nonempty = Group::new(Delimiter::Parenthesis, inner.clone()); ++ let brackets_empty = Group::new(Delimiter::Bracket, TokenStream::new()); ++ let brackets_nonempty = Group::new(Delimiter::Bracket, inner.clone()); ++ let braces_empty = Group::new(Delimiter::Brace, TokenStream::new()); ++ let braces_nonempty = Group::new(Delimiter::Brace, inner.clone()); ++ let none_empty = Group::new(Delimiter::None, TokenStream::new()); ++ let none_nonempty = Group::new(Delimiter::None, inner.clone()); ++ ++ // Matches libproc_macro. ++ assert_eq!("()", parens_empty.to_string()); ++ assert_eq!("(x)", parens_nonempty.to_string()); ++ assert_eq!("[]", brackets_empty.to_string()); ++ assert_eq!("[x]", brackets_nonempty.to_string()); ++ assert_eq!("{ }", braces_empty.to_string()); ++ assert_eq!("{ x }", braces_nonempty.to_string()); ++ assert_eq!("", none_empty.to_string()); ++ assert_eq!("x", none_nonempty.to_string()); ++} +diff --git a/third_party/rust/spirv-cross-internal/.cargo-checksum.json b/third_party/rust/spirv-cross-internal/.cargo-checksum.json +index 3c732d6d0e..014aa640e1 100644 +--- a/third_party/rust/spirv-cross-internal/.cargo-checksum.json ++++ b/third_party/rust/spirv-cross-internal/.cargo-checksum.json +@@ -1 +1 @@ +-{"files":{"Cargo.toml":"32644850d6dd8a8496fcf2dd0327dfc18a41f87bbd5f0922e174a905babde20d","build.rs":"68d1e7d99b4d890e21606aa7e7904d4f9b8caed6c0951bb8f5028745a7103d1e","src/bindings_native.rs":"6aa4e3a3bd73a638b3282405c0eaf2565ff8f364919c509f94f79e326679b03e","src/bindings_wasm.rs":"331d19691566beb163fc459d630e9830167cc86ec8cacde6be53fb0a7ca2b1be","src/bindings_wasm_functions.rs":"2515dd91e98e769fe282e7dc4f60820f4be7365c784a9f3bd17310d18a136da8","src/compiler.rs":"a53c7e0136ea3deddbfaf158295daca36623a0e244426b7c04d2856922d91c73","src/emscripten.rs":"3169890001970610013026468739910afca0d85e00d7e34beadfdd31bbcbeeb7","src/glsl.rs":"7ca1f6a11db8e7926a813fb91298dac1c0e2f90fe5426cc985cde7d8f36c05c9","src/hlsl.rs":"3734ad89a4f51647ebeb761753cb324a48ebca724970201e7598a3763d094d73","src/lib.rs":"cc41cbbe48f3e96791ba5338c66fa1fe0e533eaed6bbdced3f008d5e9fe6c6ce","src/msl.rs":"5505d1d626d8e7903ffa9d137e7392a1670fa7b574eacc4bbd33673937b36ea7","src/ptr_util.rs":"280404beede469b2c9ae40536323515a9213dac5d30014fac870a23b37672442","src/spirv.rs":"8a2cbe6b554bb6312f2aede0469338ace21e79720b5128c653e7a68c0e161277","src/vendor/SPIRV-Cross/.clang-format":"9ec4314e20afecad827a2dbd4832256be8464e88aab4a53fab45173ed129b2ed","src/vendor/SPIRV-Cross/.gitignore":"7f23cc92ddb5e1f584447e98d3e8ab6543fc182f1543f0f6ec29856f9250cdd6","src/vendor/SPIRV-Cross/.travis.yml":"abcc1b2f622b65feafd3e37a2b5e867fce3cf7211cae9fb2bf92a6de79100be4","src/vendor/SPIRV-Cross/CMakeLists.txt":"0972ab271e1eecdfe41ae0a85bf14ce7fe142b7609f8d2965b2c94f92c63b6aa","src/vendor/SPIRV-Cross/CODE_OF_CONDUCT.md":"a25e58cd66a9543e1500db9a5c3c027e874893ad1a264530bf26eb10918b5d80","src/vendor/SPIRV-Cross/GLSL.std.450.h":"20f32378793c5f416bc0704f44345c2a14c99cba3f411e3beaf1bcea372d58ba","src/vendor/SPIRV-Cross/LICENSE":"cfc7749b96f63bd31c3c42b5c471bf756814053e847c10f3eb003417bc523d30","src/vendor/SPIRV-Cross/Makefile":"e2255d32e47d447b100ce3cd0753d0758dc26460e0440d14cc157d4caf9b62f1","src/vendor/SPIRV-Cross/README.md":"b944fc5b1de709089fc3cae0dba09186ce2e43b64de6dcae1423ba1c139ee395","src/vendor/SPIRV-Cross/appveyor.yml":"0f18b8ae5fadc027a20f69307fc32e56a97edfbdcb55ac392852308f88545a04","src/vendor/SPIRV-Cross/include/spirv_cross/barrier.hpp":"bb796625e89f75e239e92f9a61597d421ffe5fb1902d200691ebe95cf856a1f8","src/vendor/SPIRV-Cross/include/spirv_cross/external_interface.h":"cdceda962d87133e44989510edc944e99052d713869b406a8b6b2d54e3d02dd7","src/vendor/SPIRV-Cross/include/spirv_cross/image.hpp":"681d0964b144c5009424196a8bc832cb81cfe5df5b91c2f3e1bfb625765a0c50","src/vendor/SPIRV-Cross/include/spirv_cross/internal_interface.hpp":"ab8851e5708b944a9bf340ce17297d94bef4876d30c833ea83d44b16f60726f6","src/vendor/SPIRV-Cross/include/spirv_cross/sampler.hpp":"b0b1bd4796b4315e826985c224ea5fa7d5bc823fcd6091f7527a5e18d9ae42fb","src/vendor/SPIRV-Cross/include/spirv_cross/thread_group.hpp":"70d9e0400f62de71d3775972eadc196ddb218254fa8155e8e33daf8d99957cc0","src/vendor/SPIRV-Cross/main.cpp":"ab2835f6dff9299f562a89755893f98a4bd946d4460abca885ba2172ebce851c","src/vendor/SPIRV-Cross/pkg-config/spirv-cross-c-shared.pc.in":"75a27e069ee07362bc6b7412b2c4e1270778dd639a8b96b8afbb113d21de79e7","src/vendor/SPIRV-Cross/spirv.h":"2c19cd1d262a3757e92276fbaa8e1a706b2328a92f47bda88f849b5cebebb630","src/vendor/SPIRV-Cross/spirv.hpp":"eb7c39a017f0a33e9d9bae7556c678c2840af4cc0beb98b2aeff9d651e370128","src/vendor/SPIRV-Cross/spirv_cfg.cpp":"09b756068d95a87a65a186831484fe21a973d3eb19675225b23864b9d37fe3d0","src/vendor/SPIRV-Cross/spirv_cfg.hpp":"ad98c4e82d6da3ed450dfc7939814115ae6b29ba2a25b0148cfc8c88774055b5","src/vendor/SPIRV-Cross/spirv_common.hpp":"5d0658a6fe9ac2f628754d2439b4d26ec8df7e50e4b196567cd4c7e7b8ce7574","src/vendor/SPIRV-Cross/spirv_cpp.cpp":"d067269ec6fc56860b4ed39edddf0a127f5e2d091512a85a79f02d5f3beccd19","src/vendor/SPIRV-Cross/spirv_cpp.hpp":"751ce82c70a399781f18aca54ea7fbb98083b98706f2a6be1cd1aa90627ef597","src/vendor/SPIRV-Cross/spirv_cross.cpp":"225b4cf4c7f8868019680d0a75a3da7983db428783ee1a3dcbb8c2765a71e79a","src/vendor/SPIRV-Cross/spirv_cross.hpp":"7eab2d7ae4b25a3c4fe85631e664727e036ae23a776aeb8c17db23f08020258a","src/vendor/SPIRV-Cross/spirv_cross_c.cpp":"80f438f4700f0b0b02547771f82be0610ba5e24ba3edb56d4331d59a206a4c2e","src/vendor/SPIRV-Cross/spirv_cross_c.h":"8b0ad63ed9fe52bc62f305afa27b5e33d9bd689a8b3f9935353742faf88a1e91","src/vendor/SPIRV-Cross/spirv_cross_containers.hpp":"409f60a5cb1dba1d0c96ef61761d8643402fc5ab67cecce12f6af3d135cfa338","src/vendor/SPIRV-Cross/spirv_cross_error_handling.hpp":"45868796346eef895cc05825582e60d26c34ec35d1a2b97f5e172f05fc565b19","src/vendor/SPIRV-Cross/spirv_cross_parsed_ir.cpp":"7c7e08db4dfc1084fd9cd266a6f829d35a6b2856198b16307a44f4d1c0743106","src/vendor/SPIRV-Cross/spirv_cross_parsed_ir.hpp":"d15446565eaa430bf2b14b596b3e579a72095a386b1eb256afc977cdf42ec5b9","src/vendor/SPIRV-Cross/spirv_cross_util.cpp":"f2426dccad2fff44053f6b1b9558719471a1ba0445c317f2c0311bc0d8b4636d","src/vendor/SPIRV-Cross/spirv_cross_util.hpp":"63ca5b2d6c80b5ad39473d6841676e48d5db3d2cbcb252fe6a6db0ef2077f0f5","src/vendor/SPIRV-Cross/spirv_glsl.cpp":"c1704189895442d593819472d3f852979107ac1798375d5a44d8f97be17fa3b5","src/vendor/SPIRV-Cross/spirv_glsl.hpp":"fb5e00ff589e3c5a25c5a15e78cf4809b753ebe04c84a008dc977330b8cbbc27","src/vendor/SPIRV-Cross/spirv_hlsl.cpp":"549ff42d0773a3f9e555d1aebd4c0084709e11528082f47c5f2b8b07de029f7c","src/vendor/SPIRV-Cross/spirv_hlsl.hpp":"18f303e7a384acf0e14677b579b100aee838813f54735330f0d1387f2493b363","src/vendor/SPIRV-Cross/spirv_msl.cpp":"7266afeb827e50782cb35362788d27fa9bd641992380008604ea1a4e48ec9c37","src/vendor/SPIRV-Cross/spirv_msl.hpp":"76b609e0b8d2f421c51e98d6357fcf44666b0d510935e083e4e02f20a80051e1","src/vendor/SPIRV-Cross/spirv_parser.cpp":"4def4db8c77967dcd2d6d93330343dc6ebb309e8cd696f52ee568a7c3ee97036","src/vendor/SPIRV-Cross/spirv_parser.hpp":"fc5f92184d1b4986f78088bee8ed7ddb9b342a34f9a3e55a4a594187292b66b8","src/vendor/SPIRV-Cross/spirv_reflect.cpp":"995a4e4315f67d4a2084e2b5d982de9baf72c7e265dde71d03ae0b8a29a24b92","src/vendor/SPIRV-Cross/spirv_reflect.hpp":"a874b40be0b210c69718e87a90e9add24653957656a995d83656f0918cfd10e4","src/wrapper.cpp":"7800cec5564dc1c7c13059cb1b1cac8d9be9248bbc887d5176600980c0d18142","src/wrapper.hpp":"b9a4f846bf57ee7a10fbb81f8f49df96ad11b1e5749c3c510925bb9b029ac08a","tests/common/mod.rs":"2843bf104c7938d93065f7b5688c9f063ad9e5720c407c737aedc5f2dee5a80f","tests/glsl_tests.rs":"ddd85107370dbfcde9d004d6717e229a7278d29ecba375828d8cb79e86625f66","tests/hlsl_tests.rs":"2058cb590c5ab85a636c93c18d0f2a79eb5ea5e647b7f10a2c83fbf394964dd6","tests/msl_tests.rs":"211d3b9cb43455a4c55bd619c05acdd21953358580c50ae75cac3f06eb26c5dd","tests/shaders/array.vert":"d0dab7ddea131e069961054f40a164602448aa78720b7ff480e141d1f7b0b2d6","tests/shaders/array.vert.spv":"8e44421590ade1716be66ad39f60fb1ce58eedeab8f0806335a7369687b308b1","tests/shaders/multiple_entry_points.cl":"2abbe57f2387f07f7f5f4cd375e47900be9c13bdc79aa0ed98a43a798cb0df81","tests/shaders/multiple_entry_points.cl.spv":"bdd34ce6765dbeab637631c3cbdf251532870d9fd6cd6c54883c0c872058ab3b","tests/shaders/rasterize_disabled.vert":"da6de172549830216933c44edf18b13113d7ca87462e3d09ad50dfc9c9836821","tests/shaders/rasterize_disabled.vert.spv":"2ba809eb500ed6e5a067389ccc056551e796e7019517593d110fb62c9dca2056","tests/shaders/sampler.frag":"4c568e65176afe596dd8ef279485e992607e94d612786214ae1c6702d0322e1f","tests/shaders/sampler.frag.spv":"bd7bd1973a82dcfdf5755361fa4dd420fdf1c32c5de0a6f0896a8d5971f98684","tests/shaders/simple.vert":"ea143c97dff5ef03728b96b2dd893bdc59d56651581ecf9fe50f10807b0efdd0","tests/shaders/simple.vert.spv":"a2b5094ffd76288e0d08c37ce0351e28f20bb6d80ddd73fc44a71c1c7cbbf7db","tests/shaders/specialization.comp":"ce32fa1615737209f2e465ea347d79030ddcb33a88c38447e7cde7dffc920163","tests/shaders/specialization.comp.spv":"848604e37b870b8999692b266677be2ce0df6ce38093a0d81e6bc43d0bdf8a3f","tests/shaders/struct.frag":"d8840bb1961d6f14609b00ee54406c1e3ea31cecd8231b77cfb73d28b71910c0","tests/shaders/struct.frag.spv":"774aa886374eb95abf9bb7d0045ee77d97e26529e9ec96b90991a515fdbca4be","tests/shaders/struct.vert":"9299cda83ddb5b5c3d95ab0d057e4df2af137dfd92d6c4d3e96295b7d42e29a1","tests/shaders/struct.vert.spv":"4a82bdee72616ac058bc60d4255efa3e78199a2b8597570c013bebbee7107fb7","tests/shaders/two_ubo.vert":"be109b2c65e5e9e1bb0dab968d7f651232f6b1c46935a3928f980bf7a40f2d62","tests/shaders/two_ubo.vert.spv":"efd14e2d0a782d61dd944711f30b9e7fcb14af17593c1fe4e11cf2b7e232bcc2","tests/shaders/workgroup.comp":"478044b5392e0d1fb042253d71ea6bf7b8a014af4a6ee35d8db4c5343ac69739","tests/shaders/workgroup.comp.spv":"72f636fe3d1d6d0c5963f71bf4349c7e40d544331b33b6b64fb5b65784e6abee","tests/spirv_tests.rs":"6de5c893f631d550fc86853537bddb824ffb534a0a7f130406be3b9ed166f198"},"package":null} +\ No newline at end of file ++{"files":{"Cargo.toml":"32644850d6dd8a8496fcf2dd0327dfc18a41f87bbd5f0922e174a905babde20d","build.rs":"68d1e7d99b4d890e21606aa7e7904d4f9b8caed6c0951bb8f5028745a7103d1e","src/bindings_native.rs":"6aa4e3a3bd73a638b3282405c0eaf2565ff8f364919c509f94f79e326679b03e","src/bindings_wasm.rs":"331d19691566beb163fc459d630e9830167cc86ec8cacde6be53fb0a7ca2b1be","src/bindings_wasm_functions.rs":"2515dd91e98e769fe282e7dc4f60820f4be7365c784a9f3bd17310d18a136da8","src/compiler.rs":"a53c7e0136ea3deddbfaf158295daca36623a0e244426b7c04d2856922d91c73","src/emscripten.rs":"3169890001970610013026468739910afca0d85e00d7e34beadfdd31bbcbeeb7","src/glsl.rs":"7ca1f6a11db8e7926a813fb91298dac1c0e2f90fe5426cc985cde7d8f36c05c9","src/hlsl.rs":"3734ad89a4f51647ebeb761753cb324a48ebca724970201e7598a3763d094d73","src/lib.rs":"cc41cbbe48f3e96791ba5338c66fa1fe0e533eaed6bbdced3f008d5e9fe6c6ce","src/msl.rs":"5505d1d626d8e7903ffa9d137e7392a1670fa7b574eacc4bbd33673937b36ea7","src/ptr_util.rs":"280404beede469b2c9ae40536323515a9213dac5d30014fac870a23b37672442","src/spirv.rs":"8a2cbe6b554bb6312f2aede0469338ace21e79720b5128c653e7a68c0e161277","src/vendor/SPIRV-Cross/.clang-format":"9ec4314e20afecad827a2dbd4832256be8464e88aab4a53fab45173ed129b2ed","src/vendor/SPIRV-Cross/.travis.yml":"abcc1b2f622b65feafd3e37a2b5e867fce3cf7211cae9fb2bf92a6de79100be4","src/vendor/SPIRV-Cross/CMakeLists.txt":"0972ab271e1eecdfe41ae0a85bf14ce7fe142b7609f8d2965b2c94f92c63b6aa","src/vendor/SPIRV-Cross/CODE_OF_CONDUCT.md":"a25e58cd66a9543e1500db9a5c3c027e874893ad1a264530bf26eb10918b5d80","src/vendor/SPIRV-Cross/GLSL.std.450.h":"20f32378793c5f416bc0704f44345c2a14c99cba3f411e3beaf1bcea372d58ba","src/vendor/SPIRV-Cross/LICENSE":"cfc7749b96f63bd31c3c42b5c471bf756814053e847c10f3eb003417bc523d30","src/vendor/SPIRV-Cross/Makefile":"e2255d32e47d447b100ce3cd0753d0758dc26460e0440d14cc157d4caf9b62f1","src/vendor/SPIRV-Cross/README.md":"b944fc5b1de709089fc3cae0dba09186ce2e43b64de6dcae1423ba1c139ee395","src/vendor/SPIRV-Cross/appveyor.yml":"0f18b8ae5fadc027a20f69307fc32e56a97edfbdcb55ac392852308f88545a04","src/vendor/SPIRV-Cross/include/spirv_cross/barrier.hpp":"bb796625e89f75e239e92f9a61597d421ffe5fb1902d200691ebe95cf856a1f8","src/vendor/SPIRV-Cross/include/spirv_cross/external_interface.h":"cdceda962d87133e44989510edc944e99052d713869b406a8b6b2d54e3d02dd7","src/vendor/SPIRV-Cross/include/spirv_cross/image.hpp":"681d0964b144c5009424196a8bc832cb81cfe5df5b91c2f3e1bfb625765a0c50","src/vendor/SPIRV-Cross/include/spirv_cross/internal_interface.hpp":"ab8851e5708b944a9bf340ce17297d94bef4876d30c833ea83d44b16f60726f6","src/vendor/SPIRV-Cross/include/spirv_cross/sampler.hpp":"b0b1bd4796b4315e826985c224ea5fa7d5bc823fcd6091f7527a5e18d9ae42fb","src/vendor/SPIRV-Cross/include/spirv_cross/thread_group.hpp":"70d9e0400f62de71d3775972eadc196ddb218254fa8155e8e33daf8d99957cc0","src/vendor/SPIRV-Cross/main.cpp":"ab2835f6dff9299f562a89755893f98a4bd946d4460abca885ba2172ebce851c","src/vendor/SPIRV-Cross/pkg-config/spirv-cross-c-shared.pc.in":"75a27e069ee07362bc6b7412b2c4e1270778dd639a8b96b8afbb113d21de79e7","src/vendor/SPIRV-Cross/spirv.h":"2c19cd1d262a3757e92276fbaa8e1a706b2328a92f47bda88f849b5cebebb630","src/vendor/SPIRV-Cross/spirv.hpp":"eb7c39a017f0a33e9d9bae7556c678c2840af4cc0beb98b2aeff9d651e370128","src/vendor/SPIRV-Cross/spirv_cfg.cpp":"09b756068d95a87a65a186831484fe21a973d3eb19675225b23864b9d37fe3d0","src/vendor/SPIRV-Cross/spirv_cfg.hpp":"ad98c4e82d6da3ed450dfc7939814115ae6b29ba2a25b0148cfc8c88774055b5","src/vendor/SPIRV-Cross/spirv_common.hpp":"5d0658a6fe9ac2f628754d2439b4d26ec8df7e50e4b196567cd4c7e7b8ce7574","src/vendor/SPIRV-Cross/spirv_cpp.cpp":"d067269ec6fc56860b4ed39edddf0a127f5e2d091512a85a79f02d5f3beccd19","src/vendor/SPIRV-Cross/spirv_cpp.hpp":"751ce82c70a399781f18aca54ea7fbb98083b98706f2a6be1cd1aa90627ef597","src/vendor/SPIRV-Cross/spirv_cross.cpp":"225b4cf4c7f8868019680d0a75a3da7983db428783ee1a3dcbb8c2765a71e79a","src/vendor/SPIRV-Cross/spirv_cross.hpp":"7eab2d7ae4b25a3c4fe85631e664727e036ae23a776aeb8c17db23f08020258a","src/vendor/SPIRV-Cross/spirv_cross_c.cpp":"80f438f4700f0b0b02547771f82be0610ba5e24ba3edb56d4331d59a206a4c2e","src/vendor/SPIRV-Cross/spirv_cross_c.h":"8b0ad63ed9fe52bc62f305afa27b5e33d9bd689a8b3f9935353742faf88a1e91","src/vendor/SPIRV-Cross/spirv_cross_containers.hpp":"409f60a5cb1dba1d0c96ef61761d8643402fc5ab67cecce12f6af3d135cfa338","src/vendor/SPIRV-Cross/spirv_cross_error_handling.hpp":"45868796346eef895cc05825582e60d26c34ec35d1a2b97f5e172f05fc565b19","src/vendor/SPIRV-Cross/spirv_cross_parsed_ir.cpp":"7c7e08db4dfc1084fd9cd266a6f829d35a6b2856198b16307a44f4d1c0743106","src/vendor/SPIRV-Cross/spirv_cross_parsed_ir.hpp":"d15446565eaa430bf2b14b596b3e579a72095a386b1eb256afc977cdf42ec5b9","src/vendor/SPIRV-Cross/spirv_cross_util.cpp":"f2426dccad2fff44053f6b1b9558719471a1ba0445c317f2c0311bc0d8b4636d","src/vendor/SPIRV-Cross/spirv_cross_util.hpp":"63ca5b2d6c80b5ad39473d6841676e48d5db3d2cbcb252fe6a6db0ef2077f0f5","src/vendor/SPIRV-Cross/spirv_glsl.cpp":"c1704189895442d593819472d3f852979107ac1798375d5a44d8f97be17fa3b5","src/vendor/SPIRV-Cross/spirv_glsl.hpp":"fb5e00ff589e3c5a25c5a15e78cf4809b753ebe04c84a008dc977330b8cbbc27","src/vendor/SPIRV-Cross/spirv_hlsl.cpp":"549ff42d0773a3f9e555d1aebd4c0084709e11528082f47c5f2b8b07de029f7c","src/vendor/SPIRV-Cross/spirv_hlsl.hpp":"18f303e7a384acf0e14677b579b100aee838813f54735330f0d1387f2493b363","src/vendor/SPIRV-Cross/spirv_msl.cpp":"7266afeb827e50782cb35362788d27fa9bd641992380008604ea1a4e48ec9c37","src/vendor/SPIRV-Cross/spirv_msl.hpp":"76b609e0b8d2f421c51e98d6357fcf44666b0d510935e083e4e02f20a80051e1","src/vendor/SPIRV-Cross/spirv_parser.cpp":"4def4db8c77967dcd2d6d93330343dc6ebb309e8cd696f52ee568a7c3ee97036","src/vendor/SPIRV-Cross/spirv_parser.hpp":"fc5f92184d1b4986f78088bee8ed7ddb9b342a34f9a3e55a4a594187292b66b8","src/vendor/SPIRV-Cross/spirv_reflect.cpp":"995a4e4315f67d4a2084e2b5d982de9baf72c7e265dde71d03ae0b8a29a24b92","src/vendor/SPIRV-Cross/spirv_reflect.hpp":"a874b40be0b210c69718e87a90e9add24653957656a995d83656f0918cfd10e4","src/wrapper.cpp":"7800cec5564dc1c7c13059cb1b1cac8d9be9248bbc887d5176600980c0d18142","src/wrapper.hpp":"b9a4f846bf57ee7a10fbb81f8f49df96ad11b1e5749c3c510925bb9b029ac08a","tests/common/mod.rs":"2843bf104c7938d93065f7b5688c9f063ad9e5720c407c737aedc5f2dee5a80f","tests/glsl_tests.rs":"ddd85107370dbfcde9d004d6717e229a7278d29ecba375828d8cb79e86625f66","tests/hlsl_tests.rs":"2058cb590c5ab85a636c93c18d0f2a79eb5ea5e647b7f10a2c83fbf394964dd6","tests/msl_tests.rs":"211d3b9cb43455a4c55bd619c05acdd21953358580c50ae75cac3f06eb26c5dd","tests/shaders/array.vert":"d0dab7ddea131e069961054f40a164602448aa78720b7ff480e141d1f7b0b2d6","tests/shaders/array.vert.spv":"8e44421590ade1716be66ad39f60fb1ce58eedeab8f0806335a7369687b308b1","tests/shaders/multiple_entry_points.cl":"2abbe57f2387f07f7f5f4cd375e47900be9c13bdc79aa0ed98a43a798cb0df81","tests/shaders/multiple_entry_points.cl.spv":"bdd34ce6765dbeab637631c3cbdf251532870d9fd6cd6c54883c0c872058ab3b","tests/shaders/rasterize_disabled.vert":"da6de172549830216933c44edf18b13113d7ca87462e3d09ad50dfc9c9836821","tests/shaders/rasterize_disabled.vert.spv":"2ba809eb500ed6e5a067389ccc056551e796e7019517593d110fb62c9dca2056","tests/shaders/sampler.frag":"4c568e65176afe596dd8ef279485e992607e94d612786214ae1c6702d0322e1f","tests/shaders/sampler.frag.spv":"bd7bd1973a82dcfdf5755361fa4dd420fdf1c32c5de0a6f0896a8d5971f98684","tests/shaders/simple.vert":"ea143c97dff5ef03728b96b2dd893bdc59d56651581ecf9fe50f10807b0efdd0","tests/shaders/simple.vert.spv":"a2b5094ffd76288e0d08c37ce0351e28f20bb6d80ddd73fc44a71c1c7cbbf7db","tests/shaders/specialization.comp":"ce32fa1615737209f2e465ea347d79030ddcb33a88c38447e7cde7dffc920163","tests/shaders/specialization.comp.spv":"848604e37b870b8999692b266677be2ce0df6ce38093a0d81e6bc43d0bdf8a3f","tests/shaders/struct.frag":"d8840bb1961d6f14609b00ee54406c1e3ea31cecd8231b77cfb73d28b71910c0","tests/shaders/struct.frag.spv":"774aa886374eb95abf9bb7d0045ee77d97e26529e9ec96b90991a515fdbca4be","tests/shaders/struct.vert":"9299cda83ddb5b5c3d95ab0d057e4df2af137dfd92d6c4d3e96295b7d42e29a1","tests/shaders/struct.vert.spv":"4a82bdee72616ac058bc60d4255efa3e78199a2b8597570c013bebbee7107fb7","tests/shaders/two_ubo.vert":"be109b2c65e5e9e1bb0dab968d7f651232f6b1c46935a3928f980bf7a40f2d62","tests/shaders/two_ubo.vert.spv":"efd14e2d0a782d61dd944711f30b9e7fcb14af17593c1fe4e11cf2b7e232bcc2","tests/shaders/workgroup.comp":"478044b5392e0d1fb042253d71ea6bf7b8a014af4a6ee35d8db4c5343ac69739","tests/shaders/workgroup.comp.spv":"72f636fe3d1d6d0c5963f71bf4349c7e40d544331b33b6b64fb5b65784e6abee","tests/spirv_tests.rs":"6de5c893f631d550fc86853537bddb824ffb534a0a7f130406be3b9ed166f198"},"package":null} +\ No newline at end of file +diff --git a/third_party/rust/syn/.cargo-checksum.json b/third_party/rust/syn/.cargo-checksum.json +index 77939d8fc6..704f2ed200 100644 +--- a/third_party/rust/syn/.cargo-checksum.json ++++ b/third_party/rust/syn/.cargo-checksum.json +@@ -1 +1 @@ +-{"files":{"Cargo.toml":"484d29864d333a361652fa4e24e1dcfab9efa47705ffd8c106d802eb03b78da7","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"23f18e03dc49df91622fe2a76176497404e46ced8a715d9d2b67a7446571cca3","README.md":"ca605417b6db8c995458f8407afaad6c177aedcc2274004283600f5638fa1b0c","benches/file.rs":"b45211cc4a0296a77aac2b4de16dbc6b5cb66adfb5afac00a77bccea87f43968","benches/rust.rs":"9cc0f62e944f1583d05c43a395a1556731501cf5976ef67a081f4f6387f883ba","build.rs":"7423ab199728d55c7d64c44b7c6729cfd93bd8273366a77707353003e27565d7","src/attr.rs":"cf81add298f0e75c35a9980a59bc3c2fd3fe933635830d1591374eeb2487c225","src/await.rs":"18f0b2ecb319991f891e300011424985e3cf33d166ea9f29f22d575fc8c83a76","src/bigint.rs":"efc7f64959980653d73fe4f8bc2a3a2904dc05f45b02c6dc15cd316fa3d7c338","src/buffer.rs":"2a432c11a3da67a21d46c2272bf9ce60a0bb20893b5750027bbd8ca3e843ab35","src/custom_keyword.rs":"589e46ec1be9a04d6de12c0b8cadf87cc1c05606ed46ddea62e9869cbca4a191","src/custom_punctuation.rs":"2ba2e294e15a0fce7ede3686c42b2891797079a724dd1193b66e7d305624c891","src/data.rs":"cc9b250d084e444782d3ff5e63c1ba387cbde8f7f2e977eab9846d920b4b8c3f","src/derive.rs":"c18878f14be5d5ab11fd7dda2d2ff1ff75c9662daf11eed033de62e4d0670a89","src/discouraged.rs":"50e10915695c4d14f64a78e20ecbef90a2cd53a7c26ee3426a2524a8ee5c9cbf","src/error.rs":"2c17a402f83ed5ae4ad96e753216771bef620235c2ff1ccc23f4bbafc7266fe1","src/export.rs":"dcae67456278c0339acfbcbb4737b8d37cfba5a150ae789f31f4be79abf7e726","src/expr.rs":"871d8eeb43cef02ef88de3bea7477b79b4eabc096a0899dde0e5750edf482f49","src/ext.rs":"b97ed549490b9248b5b5df31b3d5b08ba8791e23e6c5d3a1157a0363eb683ff3","src/file.rs":"3cc2bf5c709238d515a557f721f231c8c725b196400de051f945b549299d38a7","src/gen/fold.rs":"10b3ae33d0ce410d6bbe8b93be9d5f9e856c7dc8212133cc46b703f97d548190","src/gen/visit.rs":"e0f5798552d186024696b7bfc7219d4ff53b0e45f735a83e77cbb6b6578c5fa4","src/gen/visit_mut.rs":"9f7dda83907969971dba84d545aaa563b0728e54db97ffab5050fdf43a79c731","src/gen_helper.rs":"ea6c66388365971db6a2fc86cbb208f7eacde77e245bc8623f27a3642a3d7741","src/generics.rs":"d845d7a828863123a5187fd0fe59c9dae3636f63bad302bd035792eed3dcb1ba","src/group.rs":"119b62d8481b4b1c327639bed40e114bf1969765250b68186628247fd4144b3b","src/ident.rs":"503156ce51a7ef0420892e8dbf2ecf8fe51f42a84d52cc2c05654e1a83020cbf","src/item.rs":"213f2f58c65ee1aa222f111bc9b1be681f8fb069caed04ca56586839979318d0","src/keyword.rs":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855","src/lib.rs":"24778e9f15e8025e75aca114c712716ada586b471adb3b3b69278f4d39b8a21b","src/lifetime.rs":"905359708f772ec858954badde69ee016d29e6eeba1dd205b268445b1aff6f3a","src/lit.rs":"5bb0bddb94cbd256e50e92dc091a0baa09f1be40a77058b897507f3b17191e5d","src/lookahead.rs":"5cce8b4cb345a85c24a452ea6d78eadb76f01ca0a789cbf5ce35108334904173","src/mac.rs":"6b468244cc07e3f2f10419f833d9e2ed23edbcd6dc34cf21c5947633699db964","src/macros.rs":"0d8c3bab47539aa2d00bec64e92c901ea2c9c0af74c868051c0905b82650f970","src/op.rs":"93cd44770bb110deadf807a01d9a666efe644b6e3010f4b51cae77ee7438cfbb","src/parse.rs":"5017123c249ebc65866af113a0ad671814b9873f47568180e6539a305eb0317d","src/parse_macro_input.rs":"f799aadb7216c2d333b579f48ed2fedfe07b5e96f004b25b569649ffbaa958d2","src/parse_quote.rs":"81575bf60b18b0d8624d7025a5bcc8dcd6633ad70c454dee2a06e4c391700b6c","src/pat.rs":"db0f2263b9813de1f4e3e3e0396fe0080b1e11c8090c6b4fb6fca3cfbe22bc96","src/path.rs":"32e685ac7fd2d4b9989802de8f326a8d47fa710f86ec3e45fd9d3ff8fdfe97ef","src/print.rs":"da6529c1d9d21aaf6c835f66b4e67eacb7cf91a10eb5e9a2143b49bf99b3b5e1","src/punctuated.rs":"384e7b317b26f24118eb4b0c39e949ee9f4f3e700a4c80e462342c83b2cc3282","src/sealed.rs":"896a495a5340eec898527f18bd4ddca408ea03ea0ee3af30074ff48deace778d","src/span.rs":"748c51c6feb223c26d3b1701f5bb98aee823666c775c98106cfa24fe29d8cec1","src/spanned.rs":"adddb6acae14a0fa340df302b932c31e34b259706ce56fd82ab597ec424500e1","src/stmt.rs":"fbccf2b4da7980fe6ea8d99457d291577c0f225b370c1dd97da41abf2a18fcf7","src/thread.rs":"815eca6bd64f4eef7c447f0809e84108f5428ff50225224b373efd8fbb696874","src/token.rs":"761d8d1793560eb2b631c36ddfdbb14ac65178405f095453aa0e75e8816bdbb9","src/tt.rs":"1e32ae216d14d895ff621bf32bc611f138aa00108b0090be2cbaa0affebe8e2a","src/ty.rs":"ce052e0079b65b66bea4e9502d2ff2c90ad4b867904bf7eb892eb60aa3ef219a","tests/clone.sh":"961243d42540d9992090efbbd5478b7aa395671db64a6c28cdadb6edc610ebdf","tests/common/eq.rs":"a42d339808fd32dd4bfd440c254add8c56d5e2cde3a6bf0c88621b618ce5eaa7","tests/common/mod.rs":"20a3300016351fa210a193fbb0db059ef5125fa7909585ded64790004d4977ed","tests/common/parse.rs":"17ba6d1e74aaa3f8096c6d379d803221f12d95cca69927be047d6ddf8367647f","tests/debug/gen.rs":"57bd5cf585e0b86ad00f29f09ff3db3390c4a756d503514a9b28407500dcea3c","tests/debug/mod.rs":"462d6fe34ee75c3ca1207d4db2ff3bdee5b430b9f9ca632e5671d1588d3f76b3","tests/features/error.rs":"e0581a2869cbd237c2bc18a0a85272296e1577bb5f7317a67fa85e28e04eea6f","tests/features/mod.rs":"66a2605ec54ede29208da350f2bed536dfa763b58408d64d3fca3b13de64b64f","tests/macros/mod.rs":"3f2d758c0ba76b93f54b0c1fc22ad50edff8ef42629ba4d47ac7d7f823da8359","tests/repo/mod.rs":"e851a68972c9194a9a8d7b68538b16ed79ae81cba55e1a2ce210d1b759fb1a21","tests/test_asyncness.rs":"b6c46118b036e6807d24eb0e1779244b4fca23dac0d8031e9843b3edec484ce8","tests/test_attribute.rs":"2d8f18a98c989d3f7adaaeb1aeebd4f8413365ace63feecb37cb3f9db9db4d8f","tests/test_derive_input.rs":"477d80f914c54b526f8ff229788dc0e7798d118f6dcfa348f4c99755edb347b9","tests/test_expr.rs":"f35ca80566849a36e6ba6403d9663519eff37e4224360c468fedff8b561a643e","tests/test_generics.rs":"83a5dc07f5c5701c12625399262f7120b66f01a742523f3eda28da2cf2c87eb3","tests/test_grouping.rs":"aadd75215addd9e5a8fa2f9472117d4cb80f1e8b84e07f4c0845675c9014164f","tests/test_ident.rs":"236c239dd66f543f084f44ff747d7bc3962cf11a019a279777fe972f6e17aa4c","tests/test_iterators.rs":"718938da14778dcba06324d36a99d9317c9d45d81a34c6a44c47e1fa38085e9f","tests/test_lit.rs":"7dff2661a5ac586d6ed2fe27501cb8ff62f4cf3f6c91f596bff6057c67ad7857","tests/test_meta.rs":"8444dee084882243b107dfc8a6aac27f9382f9774162d1ac8ed8ec30d60c048e","tests/test_parse_buffer.rs":"b244bb4bc41ff06d21f239e60a3d663fdec5aa4af33f2a354afef36d34f0aefc","tests/test_pat.rs":"41776b878efae9b8e340f21ffe6296e921cf309f618482efd98609c33e32c28b","tests/test_precedence.rs":"71f3ea52cda8b40166bb7416fb98774e6a653542497b521f8e183e283dcf579d","tests/test_round_trip.rs":"e0de37f45fa223b488d25a41beab185eb92abb7bf765a9f13fe5d870ff31f5f1","tests/test_should_parse.rs":"4da4e25ee2baa7e75135c375042a7f958de136c5698dab03f99ff7a774dcd463","tests/test_size.rs":"970150b9d49ef91ab4c8f8c6a59b83f9a68a02acb779f0280733a5efaec6487a","tests/test_token_trees.rs":"a07ea657bf03b9c667c821b2db2af49b176ca737e3e01217a73cca78b7f11380","tests/zzz_stable.rs":"961d4940a926db4ca523d834b060c62de988e6a8e01c9f5efaa7bb4c86745b47"},"package":"66850e97125af79138385e9b88339cbcd037e3f28ceab8c5ad98e64f0f1f80bf"} +\ No newline at end of file ++{"files":{"Cargo.toml":"28ddb678a5ccac4423435384c8b7116f804e896eabc5aae9d5c2bc666aaebbb4","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"23f18e03dc49df91622fe2a76176497404e46ced8a715d9d2b67a7446571cca3","README.md":"03f3b53cf858536a0883aa5b5882ee61dcd0f1e71c0930c9106fcfa1d6aad2df","benches/file.rs":"b4724fc7c0f48b8f488e2632a1064f6c0bf16ded3969680fc3f4a2369536269b","benches/rust.rs":"ea6291ef2d2a83d94a3312fe179d48259f8ec0b04c961993ddd181d0a4ab740e","build.rs":"aeca2312f05aec658eaa66980a0ef3d578837db107a55702b39419ea0422eb4a","src/attr.rs":"7d79482634d6544eb4a4825405407b53660d0f5f8b929f7e1671e005b9d92038","src/await.rs":"18f0b2ecb319991f891e300011424985e3cf33d166ea9f29f22d575fc8c83a76","src/bigint.rs":"efc7f64959980653d73fe4f8bc2a3a2904dc05f45b02c6dc15cd316fa3d7c338","src/buffer.rs":"cf2a4b3bdc247b80c85ff5625a1dfb7a5f517fd835f6e1518a7b924990e4c293","src/custom_keyword.rs":"9627467063e41776315a6a14b2aaea3875592d8e0ebd2dc6df1fc2f12c06f146","src/custom_punctuation.rs":"b00e7bee96eb473507527e39db65e74e71592dc06421d2cfe45ed899c17d4847","src/data.rs":"7aec9a745cd53ec95688afa353f6efb9576e7fc0143757b51d28bc3d900b1d2a","src/derive.rs":"fa71866df6e383673dd3329f455a9f953585b83f9739050be3bf1f8c6d526b96","src/discouraged.rs":"a1f3d85e20dedf50b1b7b4571d970a3a6e9b2de4afde7dd0c986fe240df2ba46","src/error.rs":"c3005b50e3132026250c5356d0d391bf96db8087f0f5f744de98e360d8a20a3e","src/export.rs":"dcae67456278c0339acfbcbb4737b8d37cfba5a150ae789f31f4be79abf7e726","src/expr.rs":"54455fd20041996653ca5379b03cdf3c2fc1b3dd2e1149b5bc6b1dd492545d55","src/ext.rs":"870086d9021e6a6fcefa2f00cd91b55c4b74dcee8f0f6a07e76d96fb44707d61","src/file.rs":"75167ebc77e7870122078eabde1b872c337142d4b0962c20cedffcaaa2a5b7c6","src/gen/clone.rs":"0845c1bf8624c3f235cd247b4eb748e7e16b4c240097cb0ff16751f688c079ae","src/gen/debug.rs":"d24fe37f4ce1dd74f2dc54136e893782d3c4d0908323c036c97599551a56960c","src/gen/eq.rs":"1e6ef09b17ca7f36861ef23ce2a6991b231ed5f087f046469b5f23da40f5b419","src/gen/fold.rs":"3f59e59ed8ad2ab5dd347bfbe41bbc785c2aabd8ae902087a584a6daed597182","src/gen/hash.rs":"e5b2a52587173076777233a9e57e2b3c8e0dd6d6f41d16fa7c9fde68b05c2bfc","src/gen/visit.rs":"23008c170d4dd3975232876a0a654921d9b6af57372cb9fcc133ca740588d666","src/gen/visit_mut.rs":"42886c3ee02ded72d9c3eec006e20431eaee0c6b90ddefc1a36ec7bf50c6a24a","src/gen_helper.rs":"ea6c66388365971db6a2fc86cbb208f7eacde77e245bc8623f27a3642a3d7741","src/generics.rs":"d1c175284ca21e777ef0414c28383929b170ccb00aaf7a929eb18d3b05e18da8","src/group.rs":"119b62d8481b4b1c327639bed40e114bf1969765250b68186628247fd4144b3b","src/ident.rs":"503156ce51a7ef0420892e8dbf2ecf8fe51f42a84d52cc2c05654e1a83020cbf","src/item.rs":"c9ad9881e8cda8ee3f157f0c7602fc53d08a7e3288b9afc388c393689eac5aea","src/lib.rs":"558ad13779233b27bebc4b2fc8025eb1c7e57b32130dc1dd911391e27b427500","src/lifetime.rs":"f390fe06692fc51fbf3eb490bb9f795da70e4452f51c5b0df3bbaa899084ddf1","src/lit.rs":"9fab84e38756b092fbb055dcdf01e31d42d916c49e3eaae8c9019043b0ee4301","src/lookahead.rs":"5cce8b4cb345a85c24a452ea6d78eadb76f01ca0a789cbf5ce35108334904173","src/mac.rs":"e5cecea397fd01a44958162781d8d94343fe2a1b9b9754a5666c3d2ab4d7ef64","src/macros.rs":"2ce05b553f14da4ee550bb681cb0733b7186ad94719cd36f96d53e15fd02cf2b","src/op.rs":"449514e146deab0ab020bc6f764544c294dbc780941c9802bf60cf1b2839d550","src/parse.rs":"bde888c98ee259f2a73489a693515ed4875432b0d79486ac83aea19f441992a3","src/parse_macro_input.rs":"653a020f023cac0eccbc1fcc34aa7bf80567b43e5475deab4ad3e487a5363201","src/parse_quote.rs":"642f21e5fa54df4b7c373fb158289ee1005d49e1a49b1d194df5438faee71c46","src/pat.rs":"1473b258162cc822f1ee0c0869f521053ed345a140c39ed83b9b4dfb6f9f2aca","src/path.rs":"f119f0c2af12fabd360eac9a2312e0f6e6c28c633c9671bde6ef0bece7c5ba3c","src/print.rs":"da6529c1d9d21aaf6c835f66b4e67eacb7cf91a10eb5e9a2143b49bf99b3b5e1","src/punctuated.rs":"212f5a601d6c2eb8b8fa679be1167b455b595bee964d2775b0101ebb16c3eaa5","src/reserved.rs":"3625eb2a64589a4992ab79a1674e9679f465bea613ab139a671df5337e88cee6","src/sealed.rs":"896a495a5340eec898527f18bd4ddca408ea03ea0ee3af30074ff48deace778d","src/span.rs":"748c51c6feb223c26d3b1701f5bb98aee823666c775c98106cfa24fe29d8cec1","src/spanned.rs":"7d77714d585e6f42397091ffb3a799fd7b20c05c5442c737683c429ea7d409a5","src/stmt.rs":"3917fbc897f80efe838267833c55650ff8d636cb49a6d1084e28eff65d0e3ccd","src/thread.rs":"815eca6bd64f4eef7c447f0809e84108f5428ff50225224b373efd8fbb696874","src/token.rs":"a1ca6298bf6592cb80cbab1db4eac2fa4e3fa56729bb807bfb0f08ab0f229ca5","src/tt.rs":"1cc9e200624288322f800f32e3d6e2e53da946467bb312dd40a52c02cdcc4730","src/ty.rs":"cb167cbb16240c59a31b44adec175172caaf75ffef9a0bb168584b51bf105795","src/verbatim.rs":"802a97df997432f18cac6e6200ff6ea29fb2474986005e0fcdbc2b65197f87f7","src/whitespace.rs":"e63dd0aa3d34029f17766a8b09c1a6e4479e36c552c8b7023d710a399333aace","tests/common/eq.rs":"4b190a3833bdfd20a4cb1e3dff25a698751dec71d6f30249cf09426e061a4fb1","tests/common/mod.rs":"25ef6d7daa09bad3198a0e9e91b2812425f92db7c585c1e34a03a84d7362ccd8","tests/common/parse.rs":"8b7ba32f4988c30758c108536c4877dc5a039a237bf9b0687220ef2295797bbd","tests/debug/gen.rs":"d6e2abf2a7bb58a7895a60c2f094a98a4f85c9189d02011d0dcef6ef053f26e3","tests/debug/mod.rs":"868763d0ef1609a3ad5e05e9f1bfa0f813e91e7e9a36653414a188bb2fdaa425","tests/macros/mod.rs":"c0eafa4e3845fc08f6efe6021bac37822c0ac325eb7b51194a5f35236f648d92","tests/repo/mod.rs":"9e316b88d57ae213e81950c35e45443078ec90e702798353bc3528cb8a2810b6","tests/repo/progress.rs":"c08d0314a7f3ecf760d471f27da3cd2a500aeb9f1c8331bffb2aa648f9fabf3f","tests/test_asyncness.rs":"cff01db49d28ab23b0b258bc6c0a5cc4071be4fe7248eef344a5d79d2fb649b7","tests/test_attribute.rs":"0ffd99384e1a52ae17d9fed5c4053e411e8f9018decef07ffa621d1faa7329d8","tests/test_derive_input.rs":"610444351e3bf99366976bbf1da109c334a70ac9500caef366bcf9b68819829f","tests/test_expr.rs":"0ee83f6f6de950018c043efcc3e85776b4227dae3068309998a8d9709f2fc66c","tests/test_generics.rs":"9d713f90a79d6145efc89fb6f946029ca03486c632219950889da39940152ba0","tests/test_grouping.rs":"46c27baec4daaaf1e891892f0b0515ea8a44619071c7d0cc9192580916f1569f","tests/test_ident.rs":"9eb53d1e21edf23e7c9e14dc74dcc2b2538e9221e19dbcc0a44e3acc2e90f3f6","tests/test_item.rs":"461ed0c8648afffcea3217f52c9a88298182b4d39d73a11803b1281d99c98c25","tests/test_iterators.rs":"53ed6078d37550bd6765d2411e3660be401aef8a31a407350cc064a7d08c7c33","tests/test_lit.rs":"2a46c5f2f2ad1dcbb7e9b0cd11b55861c5ff818c2c4c51351d07e2daa7c74674","tests/test_meta.rs":"1fc98af3279cadc3d8db3c7e8d4d7f9e9dbd4d17548cf6a2f6f4536ed65367f6","tests/test_parse_buffer.rs":"8bbe2d24ca8a3788f72c6908fc96c26d546f11c69687bf8d72727f851d5e2d27","tests/test_parse_stream.rs":"2f449a2c41a3dee6fd14bee24e1666a453cb808eda17332fd91afd127fcdd2a6","tests/test_pat.rs":"2cb331fe404496d51e7cc7e283ae13c519a2265ca82e1c88e113296f860c2cba","tests/test_path.rs":"fcd5591e639fc787acc9763d828a811c8114525c9341282eefda8f331e082a51","tests/test_precedence.rs":"8d03656741b01e577d7501ce24332d1a4febec3e31a043e47c61062b8c527ed2","tests/test_receiver.rs":"084eca59984b9a18651da52f2c4407355da3de1335916a12477652999e2d01cc","tests/test_round_trip.rs":"ba01bf4ec04cd2d6f9e4800c343563925ae960c5f16752dc0797fda4451b6cc2","tests/test_shebang.rs":"f5772cadad5b56e3112cb16308b779f92bce1c3a48091fc9933deb2276a69331","tests/test_should_parse.rs":"1d3535698a446e2755bfc360676bdb161841a1f454cdef6e7556c6d06a95c89d","tests/test_size.rs":"5fae772bab66809d6708232f35cfb4a287882486763b0f763feec2ad79fbb68b","tests/test_stmt.rs":"17e4355843ee2982b51faba2721a18966f8c2b9422e16b052a123b8ee8b80752","tests/test_token_trees.rs":"43e56a701817e3c3bfd0cae54a457dd7a38ccb3ca19da41e2b995fdf20e6ed18","tests/test_ty.rs":"5b7c0bfc4963d41920dd0b39fdea419e34f00409ba86ad4211d6c3c7e8bbe1c0","tests/test_visibility.rs":"3f958e2b3b5908005e756a80eea326a91eac97cc4ab60599bebde8d4b942d65c","tests/zzz_stable.rs":"2a862e59cb446235ed99aec0e6ada8e16d3ecc30229b29d825b7c0bbc2602989"},"package":"963f7d3cc59b59b9325165add223142bbf1df27655d07789f109896d353d8350"} +\ No newline at end of file +diff --git a/third_party/rust/syn/Cargo.toml b/third_party/rust/syn/Cargo.toml +index 7a5c962f06..20277fc461 100644 +--- a/third_party/rust/syn/Cargo.toml ++++ b/third_party/rust/syn/Cargo.toml +@@ -13,7 +13,7 @@ + [package] + edition = "2018" + name = "syn" +-version = "1.0.5" ++version = "1.0.40" + authors = ["David Tolnay <dtolnay@gmail.com>"] + include = ["/benches/**", "/build.rs", "/Cargo.toml", "/LICENSE-APACHE", "/LICENSE-MIT", "/README.md", "/src/**", "/tests/**"] + description = "Parser for Rust source code" +@@ -24,25 +24,21 @@ license = "MIT OR Apache-2.0" + repository = "https://github.com/dtolnay/syn" + [package.metadata.docs.rs] + all-features = true ++targets = ["x86_64-unknown-linux-gnu"] + + [package.metadata.playground] +-all-features = true +- +-[lib] +-name = "syn" ++features = ["full", "visit", "visit-mut", "fold", "extra-traits"] + + [[bench]] + name = "rust" + harness = false + required-features = ["full", "parsing"] +-edition = "2018" + + [[bench]] + name = "file" + required-features = ["full", "parsing"] +-edition = "2018" + [dependencies.proc-macro2] +-version = "1.0" ++version = "1.0.13" + default-features = false + + [dependencies.quote] +@@ -52,18 +48,34 @@ default-features = false + + [dependencies.unicode-xid] + version = "0.2" ++[dev-dependencies.anyhow] ++version = "1.0" ++ ++[dev-dependencies.flate2] ++version = "1.0" ++ + [dev-dependencies.insta] +-version = "0.9" ++version = "0.16" + + [dev-dependencies.rayon] + version = "1.0" + + [dev-dependencies.ref-cast] +-version = "0.2" ++version = "1.0" + + [dev-dependencies.regex] + version = "1.0" + ++[dev-dependencies.reqwest] ++version = "0.10" ++features = ["blocking"] ++ ++[dev-dependencies.syn-test-suite] ++version = "0" ++ ++[dev-dependencies.tar] ++version = "0.4" ++ + [dev-dependencies.termcolor] + version = "1.0" + +@@ -80,7 +92,6 @@ full = [] + parsing = [] + printing = ["quote"] + proc-macro = ["proc-macro2/proc-macro", "quote/proc-macro"] ++test = ["syn-test-suite/all-features"] + visit = [] + visit-mut = [] +-[badges.travis-ci] +-repository = "dtolnay/syn" +diff --git a/third_party/rust/syn/README.md b/third_party/rust/syn/README.md +index 29a7f32a46..12b5f45b3d 100644 +--- a/third_party/rust/syn/README.md ++++ b/third_party/rust/syn/README.md +@@ -1,10 +1,10 @@ + Parser for Rust source code + =========================== + +-[![Build Status](https://api.travis-ci.org/dtolnay/syn.svg?branch=master)](https://travis-ci.org/dtolnay/syn) +-[![Latest Version](https://img.shields.io/crates/v/syn.svg)](https://crates.io/crates/syn) +-[![Rust Documentation](https://img.shields.io/badge/api-rustdoc-blue.svg)](https://docs.rs/syn/1.0/syn/) +-[![Rustc Version 1.31+](https://img.shields.io/badge/rustc-1.31+-lightgray.svg)](https://blog.rust-lang.org/2018/12/06/Rust-1.31-and-rust-2018.html) ++[<img alt="github" src="https://img.shields.io/badge/github-dtolnay/syn-8da0cb?style=for-the-badge&labelColor=555555&logo=github" height="20">](https://github.com/dtolnay/syn) ++[<img alt="crates.io" src="https://img.shields.io/crates/v/syn.svg?style=for-the-badge&color=fc8d62&logo=rust" height="20">](https://crates.io/crates/syn) ++[<img alt="docs.rs" src="https://img.shields.io/badge/docs.rs-syn-66c2a5?style=for-the-badge&labelColor=555555&logoColor=white&logo=data:image/svg+xml;base64,PHN2ZyByb2xlPSJpbWciIHhtbG5zPSJodHRwOi8vd3d3LnczLm9yZy8yMDAwL3N2ZyIgdmlld0JveD0iMCAwIDUxMiA1MTIiPjxwYXRoIGZpbGw9IiNmNWY1ZjUiIGQ9Ik00ODguNiAyNTAuMkwzOTIgMjE0VjEwNS41YzAtMTUtOS4zLTI4LjQtMjMuNC0zMy43bC0xMDAtMzcuNWMtOC4xLTMuMS0xNy4xLTMuMS0yNS4zIDBsLTEwMCAzNy41Yy0xNC4xIDUuMy0yMy40IDE4LjctMjMuNCAzMy43VjIxNGwtOTYuNiAzNi4yQzkuMyAyNTUuNSAwIDI2OC45IDAgMjgzLjlWMzk0YzAgMTMuNiA3LjcgMjYuMSAxOS45IDMyLjJsMTAwIDUwYzEwLjEgNS4xIDIyLjEgNS4xIDMyLjIgMGwxMDMuOS01MiAxMDMuOSA1MmMxMC4xIDUuMSAyMi4xIDUuMSAzMi4yIDBsMTAwLTUwYzEyLjItNi4xIDE5LjktMTguNiAxOS45LTMyLjJWMjgzLjljMC0xNS05LjMtMjguNC0yMy40LTMzLjd6TTM1OCAyMTQuOGwtODUgMzEuOXYtNjguMmw4NS0zN3Y3My4zek0xNTQgMTA0LjFsMTAyLTM4LjIgMTAyIDM4LjJ2LjZsLTEwMiA0MS40LTEwMi00MS40di0uNnptODQgMjkxLjFsLTg1IDQyLjV2LTc5LjFsODUtMzguOHY3NS40em0wLTExMmwtMTAyIDQxLjQtMTAyLTQxLjR2LS42bDEwMi0zOC4yIDEwMiAzOC4ydi42em0yNDAgMTEybC04NSA0Mi41di03OS4xbDg1LTM4Ljh2NzUuNHptMC0xMTJsLTEwMiA0MS40LTEwMi00MS40di0uNmwxMDItMzguMiAxMDIgMzguMnYuNnoiPjwvcGF0aD48L3N2Zz4K" height="20">](https://docs.rs/syn) ++[<img alt="build status" src="https://img.shields.io/github/workflow/status/dtolnay/syn/CI/master?style=for-the-badge" height="20">](https://github.com/dtolnay/syn/actions?query=branch%3Amaster) + + Syn is a parsing library for parsing a stream of Rust tokens into a syntax tree + of Rust source code. +@@ -46,10 +46,6 @@ contains some APIs that may be useful more generally. + [`syn::DeriveInput`]: https://docs.rs/syn/1.0/syn/struct.DeriveInput.html + [parser functions]: https://docs.rs/syn/1.0/syn/parse/index.html + +-If you get stuck with anything involving procedural macros in Rust I am happy to +-provide help even if the issue is not related to Syn. Please file a ticket in +-this repo. +- + *Version requirement: Syn supports rustc 1.31 and up.* + + [*Release notes*](https://github.com/dtolnay/syn/releases) +@@ -88,8 +84,6 @@ proc-macro = true + ``` + + ```rust +-extern crate proc_macro; +- + use proc_macro::TokenStream; + use quote::quote; + use syn::{parse_macro_input, DeriveInput}; +@@ -271,7 +265,7 @@ points, which are required by the language to use `proc_macro::TokenStream`. + The proc-macro2 crate will automatically detect and use the compiler's data + structures when a procedural macro is active. + +-[proc-macro2]: https://docs.rs/proc-macro2/1.0.0/proc_macro2/ ++[proc-macro2]: https://docs.rs/proc-macro2/1.0/proc_macro2/ + + <br> + +diff --git a/third_party/rust/syn/benches/file.rs b/third_party/rust/syn/benches/file.rs +index 08ecd90960..58ab8df297 100644 +--- a/third_party/rust/syn/benches/file.rs ++++ b/third_party/rust/syn/benches/file.rs +@@ -1,9 +1,16 @@ + // $ cargo bench --features full --bench file + + #![feature(rustc_private, test)] ++#![recursion_limit = "1024"] + + extern crate test; + ++#[macro_use] ++#[path = "../tests/macros/mod.rs"] ++mod macros; ++ ++#[path = "../tests/common/mod.rs"] ++mod common; + #[path = "../tests/repo/mod.rs"] + pub mod repo; + +diff --git a/third_party/rust/syn/benches/rust.rs b/third_party/rust/syn/benches/rust.rs +index e3d9cd29ba..50e1a7f601 100644 +--- a/third_party/rust/syn/benches/rust.rs ++++ b/third_party/rust/syn/benches/rust.rs +@@ -4,7 +4,14 @@ + // $ RUSTFLAGS='--cfg syn_only' cargo build --release --features full --bench rust + + #![cfg_attr(not(syn_only), feature(rustc_private))] ++#![recursion_limit = "1024"] + ++#[macro_use] ++#[path = "../tests/macros/mod.rs"] ++mod macros; ++ ++#[path = "../tests/common/mod.rs"] ++mod common; + #[path = "../tests/repo/mod.rs"] + mod repo; + +@@ -28,31 +35,35 @@ mod syn_parse { + } + + #[cfg(not(syn_only))] +-mod libsyntax_parse { ++mod librustc_parse { + extern crate rustc_data_structures; +- extern crate syntax; +- extern crate syntax_pos; ++ extern crate rustc_errors; ++ extern crate rustc_parse; ++ extern crate rustc_session; ++ extern crate rustc_span; + + use rustc_data_structures::sync::Lrc; +- use syntax::edition::Edition; +- use syntax::errors::{emitter::Emitter, DiagnosticBuilder, Handler}; +- use syntax::parse::ParseSess; +- use syntax::source_map::{FilePathMapping, SourceMap}; +- use syntax_pos::FileName; ++ use rustc_errors::{emitter::Emitter, Diagnostic, Handler}; ++ use rustc_session::parse::ParseSess; ++ use rustc_span::source_map::{FilePathMapping, SourceMap}; ++ use rustc_span::{edition::Edition, FileName}; + + pub fn bench(content: &str) -> Result<(), ()> { + struct SilentEmitter; + + impl Emitter for SilentEmitter { +- fn emit_diagnostic(&mut self, _db: &DiagnosticBuilder) {} ++ fn emit_diagnostic(&mut self, _diag: &Diagnostic) {} ++ fn source_map(&self) -> Option<&Lrc<SourceMap>> { ++ None ++ } + } + +- syntax::with_globals(Edition::Edition2018, || { ++ rustc_span::with_session_globals(Edition::Edition2018, || { + let cm = Lrc::new(SourceMap::new(FilePathMapping::empty())); + let emitter = Box::new(SilentEmitter); + let handler = Handler::with_emitter(false, None, emitter); + let sess = ParseSess::with_span_handler(handler, cm); +- if let Err(mut diagnostic) = syntax::parse::parse_crate_from_source_str( ++ if let Err(mut diagnostic) = rustc_parse::parse_crate_from_source_str( + FileName::Custom("bench".to_owned()), + content.to_owned(), + &sess, +@@ -104,11 +115,11 @@ fn main() { + repo::clone_rust(); + + macro_rules! testcases { +- ($($(#[$cfg:meta])* $name:path,)*) => { ++ ($($(#[$cfg:meta])* $name:ident,)*) => { + vec![ + $( + $(#[$cfg])* +- (stringify!($name), $name as fn(&str) -> Result<(), ()>), ++ (stringify!($name), $name::bench as fn(&str) -> Result<(), ()>), + )* + ] + }; +@@ -128,12 +139,12 @@ fn main() { + + for (name, f) in testcases!( + #[cfg(not(syn_only))] +- read_from_disk::bench, ++ read_from_disk, + #[cfg(not(syn_only))] +- tokenstream_parse::bench, +- syn_parse::bench, ++ tokenstream_parse, ++ syn_parse, + #[cfg(not(syn_only))] +- libsyntax_parse::bench, ++ librustc_parse, + ) { + eprint!("{:20}", format!("{}:", name)); + let elapsed = exec(f); +diff --git a/third_party/rust/syn/build.rs b/third_party/rust/syn/build.rs +index c0f9ed3406..cf7681c3f9 100644 +--- a/third_party/rust/syn/build.rs ++++ b/third_party/rust/syn/build.rs +@@ -1,6 +1,6 @@ + use std::env; + use std::process::Command; +-use std::str::{self, FromStr}; ++use std::str; + + // The rustc-cfg strings below are *not* public API. Please let us know by + // opening a GitHub issue if your build environment requires some way to enable +@@ -26,38 +26,14 @@ struct Compiler { + } + + fn rustc_version() -> Option<Compiler> { +- let rustc = match env::var_os("RUSTC") { +- Some(rustc) => rustc, +- None => return None, +- }; +- +- let output = match Command::new(rustc).arg("--version").output() { +- Ok(output) => output, +- Err(_) => return None, +- }; +- +- let version = match str::from_utf8(&output.stdout) { +- Ok(version) => version, +- Err(_) => return None, +- }; +- ++ let rustc = env::var_os("RUSTC")?; ++ let output = Command::new(rustc).arg("--version").output().ok()?; ++ let version = str::from_utf8(&output.stdout).ok()?; + let mut pieces = version.split('.'); + if pieces.next() != Some("rustc 1") { + return None; + } +- +- let next = match pieces.next() { +- Some(next) => next, +- None => return None, +- }; +- +- let minor = match u32::from_str(next) { +- Ok(minor) => minor, +- Err(_) => return None, +- }; +- +- Some(Compiler { +- minor: minor, +- nightly: version.contains("nightly"), +- }) ++ let minor = pieces.next()?.parse().ok()?; ++ let nightly = version.contains("nightly"); ++ Some(Compiler { minor, nightly }) + } +diff --git a/third_party/rust/syn/src/attr.rs b/third_party/rust/syn/src/attr.rs +index 34009deabc..fa4f1cb2a3 100644 +--- a/third_party/rust/syn/src/attr.rs ++++ b/third_party/rust/syn/src/attr.rs +@@ -9,15 +9,11 @@ use proc_macro2::TokenStream; + use crate::parse::{Parse, ParseBuffer, ParseStream, Parser, Result}; + #[cfg(feature = "parsing")] + use crate::punctuated::Pair; +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; +-#[cfg(feature = "extra-traits")] +-use std::hash::{Hash, Hasher}; + + ast_struct! { + /// An attribute like `#[repr(transparent)]`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// <br> +@@ -111,7 +107,46 @@ ast_struct! { + /// + /// [`parse_meta()`]: Attribute::parse_meta + /// [`parse_args()`]: Attribute::parse_args +- pub struct Attribute #manual_extra_traits { ++ /// ++ /// <p><br></p> ++ /// ++ /// # Doc comments ++ /// ++ /// The compiler transforms doc comments, such as `/// comment` and `/*! ++ /// comment */`, into attributes before macros are expanded. Each comment is ++ /// expanded into an attribute of the form `#[doc = r"comment"]`. ++ /// ++ /// As an example, the following `mod` items are expanded identically: ++ /// ++ /// ``` ++ /// # use syn::{ItemMod, parse_quote}; ++ /// let doc: ItemMod = parse_quote! { ++ /// /// Single line doc comments ++ /// /// We write so many! ++ /// /** ++ /// * Multi-line comments... ++ /// * May span many lines ++ /// */ ++ /// mod example { ++ /// //! Of course, they can be inner too ++ /// /*! And fit in a single line */ ++ /// } ++ /// }; ++ /// let attr: ItemMod = parse_quote! { ++ /// #[doc = r" Single line doc comments"] ++ /// #[doc = r" We write so many!"] ++ /// #[doc = r" ++ /// * Multi-line comments... ++ /// * May span many lines ++ /// "] ++ /// mod example { ++ /// #![doc = r" Of course, they can be inner too"] ++ /// #![doc = r" And fit in a single line "] ++ /// } ++ /// }; ++ /// assert_eq!(doc, attr); ++ /// ``` ++ pub struct Attribute { + pub pound_token: Token![#], + pub style: AttrStyle, + pub bracket_token: token::Bracket, +@@ -120,39 +155,11 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Attribute {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Attribute { +- fn eq(&self, other: &Self) -> bool { +- self.style == other.style +- && self.pound_token == other.pound_token +- && self.bracket_token == other.bracket_token +- && self.path == other.path +- && TokenStreamHelper(&self.tokens) == TokenStreamHelper(&other.tokens) +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Attribute { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- self.style.hash(state); +- self.pound_token.hash(state); +- self.bracket_token.hash(state); +- self.path.hash(state); +- TokenStreamHelper(&self.tokens).hash(state); +- } +-} +- + impl Attribute { + /// Parses the content of the attribute, consisting of the path and tokens, + /// as a [`Meta`] if possible. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_meta(&self) -> Result<Meta> { +@@ -199,7 +206,7 @@ impl Attribute { + /// ^^^^^^^^^ what gets parsed + /// ``` + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_args<T: Parse>(&self) -> Result<T> { +@@ -208,7 +215,7 @@ impl Attribute { + + /// Parse the arguments to the attribute using the given parser. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_args_with<F: Parser>(&self, parser: F) -> Result<F::Output> { +@@ -221,7 +228,7 @@ impl Attribute { + + /// Parses zero or more outer attributes from the stream. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_outer(input: ParseStream) -> Result<Vec<Self>> { +@@ -234,7 +241,7 @@ impl Attribute { + + /// Parses zero or more inner attributes from the stream. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_inner(input: ParseStream) -> Result<Vec<Self>> { +@@ -247,7 +254,7 @@ impl Attribute { + } + + #[cfg(feature = "parsing")] +-fn error_expected_args(attr: &Attribute) -> Error { ++fn expected_parentheses(attr: &Attribute) -> String { + let style = match attr.style { + AttrStyle::Outer => "#", + AttrStyle::Inner(_) => "#!", +@@ -261,19 +268,23 @@ fn error_expected_args(attr: &Attribute) -> Error { + path += &segment.ident.to_string(); + } + +- let msg = format!("expected attribute arguments: {}[{}(...)]", style, path); +- +- #[cfg(feature = "printing")] +- return Error::new_spanned(attr, msg); +- +- #[cfg(not(feature = "printing"))] +- return Error::new(attr.bracket_token.span, msg); ++ format!("{}[{}(...)]", style, path) + } + + #[cfg(feature = "parsing")] + fn enter_args<'a>(attr: &Attribute, input: ParseStream<'a>) -> Result<ParseBuffer<'a>> { + if input.is_empty() { +- return Err(error_expected_args(attr)); ++ let expected = expected_parentheses(attr); ++ let msg = format!("expected attribute arguments in parentheses: {}", expected); ++ return Err(crate::error::new2( ++ attr.pound_token.span, ++ attr.bracket_token.span, ++ msg, ++ )); ++ } else if input.peek(Token![=]) { ++ let expected = expected_parentheses(attr); ++ let msg = format!("expected parentheses: {}", expected); ++ return Err(input.error(msg)); + }; + + let content; +@@ -298,7 +309,7 @@ ast_enum! { + /// Distinguishes between attributes that decorate an item and attributes + /// that are contained within an item. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Outer attributes +@@ -312,7 +323,6 @@ ast_enum! { + /// - `#![feature(proc_macro)]` + /// - `//! # Example` + /// - `/*! Please file an issue */` +- #[cfg_attr(feature = "clone-impls", derive(Copy))] + pub enum AttrStyle { + Outer, + Inner(Token![!]), +@@ -322,7 +332,7 @@ ast_enum! { + ast_enum_of_structs! { + /// Content of a compile-time structured attribute. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// ## Path +@@ -360,7 +370,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A structured list within an attribute, like `derive(Copy, Clone)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct MetaList { + pub path: Path, +@@ -372,7 +382,7 @@ ast_struct! { + ast_struct! { + /// A name-value pair within an attribute, like `feature = "nightly"`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct MetaNameValue { + pub path: Path, +@@ -398,7 +408,7 @@ impl Meta { + ast_enum_of_structs! { + /// Element of a compile-time attribute list. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum NestedMeta { + /// A structured meta item, like the `Copy` in `#[derive(Copy)]` which +@@ -429,8 +439,8 @@ ast_enum_of_structs! { + /// as type `AttributeArgs`. + /// + /// ``` +-/// extern crate proc_macro; +-/// ++/// # extern crate proc_macro; ++/// # + /// use proc_macro::TokenStream; + /// use syn::{parse_macro_input, AttributeArgs, ItemFn}; + /// +@@ -464,7 +474,7 @@ where + fn is_outer(attr: &&Attribute) -> bool { + match attr.style { + AttrStyle::Outer => true, +- _ => false, ++ AttrStyle::Inner(_) => false, + } + } + self.into_iter().filter(is_outer) +@@ -474,7 +484,7 @@ where + fn is_inner(attr: &&Attribute) -> bool { + match attr.style { + AttrStyle::Inner(_) => true, +- _ => false, ++ AttrStyle::Outer => false, + } + } + self.into_iter().filter(is_inner) +diff --git a/third_party/rust/syn/src/buffer.rs b/third_party/rust/syn/src/buffer.rs +index 551a5ac816..a461cc49ea 100644 +--- a/third_party/rust/syn/src/buffer.rs ++++ b/third_party/rust/syn/src/buffer.rs +@@ -1,7 +1,7 @@ + //! A stably addressed token buffer supporting efficient traversal based on a + //! cheaply copyable cursor. + //! +-//! *This module is available if Syn is built with the `"parsing"` feature.* ++//! *This module is available only if Syn is built with the `"parsing"` feature.* + + // This module is heavily commented as it contains most of the unsafe code in + // Syn, and caution should be used when editing it. The public-facing interface +@@ -36,7 +36,7 @@ enum Entry { + /// `TokenStream` which requires a deep copy in order to traverse more than + /// once. + /// +-/// *This type is available if Syn is built with the `"parsing"` feature.* ++/// *This type is available only if Syn is built with the `"parsing"` feature.* + pub struct TokenBuffer { + // NOTE: Do not derive clone on this - there are raw pointers inside which + // will be messed up. Moving the `TokenBuffer` itself is safe as the actual +@@ -98,7 +98,7 @@ impl TokenBuffer { + /// Creates a `TokenBuffer` containing all the tokens from the input + /// `TokenStream`. + /// +- /// *This method is available if Syn is built with both the `"parsing"` and ++ /// *This method is available only if Syn is built with both the `"parsing"` and + /// `"proc-macro"` features.* + #[cfg(all( + not(all(target_arch = "wasm32", any(target_os = "unknown", target_os = "wasi"))), +@@ -133,8 +133,7 @@ impl TokenBuffer { + /// Two cursors are equal if they have the same location in the same input + /// stream, and have the same scope. + /// +-/// *This type is available if Syn is built with the `"parsing"` feature.* +-#[derive(Copy, Clone, Eq, PartialEq)] ++/// *This type is available only if Syn is built with the `"parsing"` feature.* + pub struct Cursor<'a> { + // The current entry which the `Cursor` is pointing at. + ptr: *const Entry, +@@ -201,13 +200,13 @@ impl<'a> Cursor<'a> { + Cursor::create(self.ptr.offset(1), self.scope) + } + +- /// If the cursor is looking at a `None`-delimited group, move it to look at +- /// the first token inside instead. If the group is empty, this will move ++ /// While the cursor is looking at a `None`-delimited group, move it to look ++ /// at the first token inside instead. If the group is empty, this will move + /// the cursor past the `None`-delimited group. + /// + /// WARNING: This mutates its argument. + fn ignore_none(&mut self) { +- if let Entry::Group(group, buf) = self.entry() { ++ while let Entry::Group(group, buf) = self.entry() { + if group.delimiter() == Delimiter::None { + // NOTE: We call `Cursor::create` here to make sure that + // situations where we should immediately exit the span after +@@ -215,13 +214,14 @@ impl<'a> Cursor<'a> { + unsafe { + *self = Cursor::create(&buf.data[0], self.scope); + } ++ } else { ++ break; + } + } + } + + /// Checks whether the cursor is currently pointing at the end of its valid + /// scope. +- #[inline] + pub fn eof(self) -> bool { + // We're at eof if we're at the end of our scope. + self.ptr == self.scope +@@ -342,6 +342,44 @@ impl<'a> Cursor<'a> { + Entry::End(..) => Span::call_site(), + } + } ++ ++ /// Skip over the next token without cloning it. Returns `None` if this ++ /// cursor points to eof. ++ /// ++ /// This method treats `'lifetimes` as a single token. ++ pub(crate) fn skip(self) -> Option<Cursor<'a>> { ++ match self.entry() { ++ Entry::End(..) => None, ++ ++ // Treat lifetimes as a single tt for the purposes of 'skip'. ++ Entry::Punct(op) if op.as_char() == '\'' && op.spacing() == Spacing::Joint => { ++ let next = unsafe { self.bump() }; ++ match next.entry() { ++ Entry::Ident(_) => Some(unsafe { next.bump() }), ++ _ => Some(next), ++ } ++ } ++ _ => Some(unsafe { self.bump() }), ++ } ++ } ++} ++ ++impl<'a> Copy for Cursor<'a> {} ++ ++impl<'a> Clone for Cursor<'a> { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++ ++impl<'a> Eq for Cursor<'a> {} ++ ++impl<'a> PartialEq for Cursor<'a> { ++ fn eq(&self, other: &Self) -> bool { ++ let Cursor { ptr, scope, marker } = self; ++ let _ = marker; ++ *ptr == other.ptr && *scope == other.scope ++ } + } + + pub(crate) fn same_scope(a: Cursor, b: Cursor) -> bool { +diff --git a/third_party/rust/syn/src/custom_keyword.rs b/third_party/rust/syn/src/custom_keyword.rs +index 200e8478ef..a33044a564 100644 +--- a/third_party/rust/syn/src/custom_keyword.rs ++++ b/third_party/rust/syn/src/custom_keyword.rs +@@ -86,7 +86,7 @@ + /// } + /// } + /// ``` +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! custom_keyword { + ($ident:ident) => { + #[allow(non_camel_case_types)] +@@ -95,7 +95,7 @@ macro_rules! custom_keyword { + } + + #[doc(hidden)] +- #[allow(non_snake_case)] ++ #[allow(dead_code, non_snake_case)] + pub fn $ident<__S: $crate::export::IntoSpans<[$crate::export::Span; 1]>>( + span: __S, + ) -> $ident { +@@ -112,10 +112,10 @@ macro_rules! custom_keyword { + } + } + +- impl_parse_for_custom_keyword!($ident); +- impl_to_tokens_for_custom_keyword!($ident); +- impl_clone_for_custom_keyword!($ident); +- impl_extra_traits_for_custom_keyword!($ident); ++ $crate::impl_parse_for_custom_keyword!($ident); ++ $crate::impl_to_tokens_for_custom_keyword!($ident); ++ $crate::impl_clone_for_custom_keyword!($ident); ++ $crate::impl_extra_traits_for_custom_keyword!($ident); + }; + } + +diff --git a/third_party/rust/syn/src/custom_punctuation.rs b/third_party/rust/syn/src/custom_punctuation.rs +index 29fa448bd8..70dff42851 100644 +--- a/third_party/rust/syn/src/custom_punctuation.rs ++++ b/third_party/rust/syn/src/custom_punctuation.rs +@@ -74,19 +74,19 @@ + /// let _: PathSegments = syn::parse_str(input).unwrap(); + /// } + /// ``` +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! custom_punctuation { + ($ident:ident, $($tt:tt)+) => { + pub struct $ident { +- pub spans: custom_punctuation_repr!($($tt)+), ++ pub spans: $crate::custom_punctuation_repr!($($tt)+), + } + + #[doc(hidden)] +- #[allow(non_snake_case)] +- pub fn $ident<__S: $crate::export::IntoSpans<custom_punctuation_repr!($($tt)+)>>( ++ #[allow(dead_code, non_snake_case)] ++ pub fn $ident<__S: $crate::export::IntoSpans<$crate::custom_punctuation_repr!($($tt)+)>>( + spans: __S, + ) -> $ident { +- let _validate_len = 0 $(+ custom_punctuation_len!(strict, $tt))*; ++ let _validate_len = 0 $(+ $crate::custom_punctuation_len!(strict, $tt))*; + $ident { + spans: $crate::export::IntoSpans::into_spans(spans) + } +@@ -98,33 +98,33 @@ macro_rules! custom_punctuation { + } + } + +- impl_parse_for_custom_punctuation!($ident, $($tt)+); +- impl_to_tokens_for_custom_punctuation!($ident, $($tt)+); +- impl_clone_for_custom_punctuation!($ident, $($tt)+); +- impl_extra_traits_for_custom_punctuation!($ident, $($tt)+); ++ $crate::impl_parse_for_custom_punctuation!($ident, $($tt)+); ++ $crate::impl_to_tokens_for_custom_punctuation!($ident, $($tt)+); ++ $crate::impl_clone_for_custom_punctuation!($ident, $($tt)+); ++ $crate::impl_extra_traits_for_custom_punctuation!($ident, $($tt)+); + }; + } + + // Not public API. + #[cfg(feature = "parsing")] + #[doc(hidden)] +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! impl_parse_for_custom_punctuation { + ($ident:ident, $($tt:tt)+) => { + impl $crate::token::CustomToken for $ident { + fn peek(cursor: $crate::buffer::Cursor) -> bool { +- $crate::token::parsing::peek_punct(cursor, stringify_punct!($($tt)+)) ++ $crate::token::parsing::peek_punct(cursor, $crate::stringify_punct!($($tt)+)) + } + + fn display() -> &'static $crate::export::str { +- custom_punctuation_concat!("`", stringify_punct!($($tt)+), "`") ++ concat!("`", $crate::stringify_punct!($($tt)+), "`") + } + } + + impl $crate::parse::Parse for $ident { + fn parse(input: $crate::parse::ParseStream) -> $crate::parse::Result<$ident> { +- let spans: custom_punctuation_repr!($($tt)+) = +- $crate::token::parsing::punct(input, stringify_punct!($($tt)+))?; ++ let spans: $crate::custom_punctuation_repr!($($tt)+) = ++ $crate::token::parsing::punct(input, $crate::stringify_punct!($($tt)+))?; + Ok($ident(spans)) + } + } +@@ -142,12 +142,12 @@ macro_rules! impl_parse_for_custom_punctuation { + // Not public API. + #[cfg(feature = "printing")] + #[doc(hidden)] +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! impl_to_tokens_for_custom_punctuation { + ($ident:ident, $($tt:tt)+) => { + impl $crate::export::ToTokens for $ident { + fn to_tokens(&self, tokens: &mut $crate::export::TokenStream2) { +- $crate::token::printing::punct(stringify_punct!($($tt)+), &self.spans, tokens) ++ $crate::token::printing::punct($crate::stringify_punct!($($tt)+), &self.spans, tokens) + } + } + }; +@@ -221,16 +221,16 @@ macro_rules! impl_extra_traits_for_custom_punctuation { + + // Not public API. + #[doc(hidden)] +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! custom_punctuation_repr { + ($($tt:tt)+) => { +- [$crate::export::Span; 0 $(+ custom_punctuation_len!(lenient, $tt))+] ++ [$crate::export::Span; 0 $(+ $crate::custom_punctuation_len!(lenient, $tt))+] + }; + } + + // Not public API. + #[doc(hidden)] +-#[macro_export(local_inner_macros)] ++#[macro_export] + #[rustfmt::skip] + macro_rules! custom_punctuation_len { + ($mode:ident, +) => { 1 }; +@@ -279,7 +279,7 @@ macro_rules! custom_punctuation_len { + ($mode:ident, -=) => { 2 }; + ($mode:ident, ~) => { 1 }; + (lenient, $tt:tt) => { 0 }; +- (strict, $tt:tt) => {{ custom_punctuation_unexpected!($tt); 0 }}; ++ (strict, $tt:tt) => {{ $crate::custom_punctuation_unexpected!($tt); 0 }}; + } + + // Not public API. +@@ -297,13 +297,3 @@ macro_rules! stringify_punct { + concat!($(stringify!($tt)),+) + }; + } +- +-// Not public API. +-// Without this, local_inner_macros breaks when looking for concat! +-#[doc(hidden)] +-#[macro_export] +-macro_rules! custom_punctuation_concat { +- ($($tt:tt)*) => { +- concat!($($tt)*) +- }; +-} +diff --git a/third_party/rust/syn/src/data.rs b/third_party/rust/syn/src/data.rs +index be43679874..b217b8ca6f 100644 +--- a/third_party/rust/syn/src/data.rs ++++ b/third_party/rust/syn/src/data.rs +@@ -4,7 +4,7 @@ use crate::punctuated::Punctuated; + ast_struct! { + /// An enum variant. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Variant { + /// Attributes tagged on the variant. +@@ -24,7 +24,7 @@ ast_struct! { + ast_enum_of_structs! { + /// Data stored within an enum variant or struct. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Syntax tree enum +@@ -52,7 +52,7 @@ ast_struct! { + /// Named fields of a struct or struct variant such as `Point { x: f64, + /// y: f64 }`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct FieldsNamed { + pub brace_token: token::Brace, +@@ -63,7 +63,7 @@ ast_struct! { + ast_struct! { + /// Unnamed fields of a tuple struct or tuple variant such as `Some(T)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct FieldsUnnamed { + pub paren_token: token::Paren, +@@ -93,6 +93,24 @@ impl Fields { + Fields::Unnamed(f) => f.unnamed.iter_mut(), + } + } ++ ++ /// Returns the number of fields. ++ pub fn len(&self) -> usize { ++ match self { ++ Fields::Unit => 0, ++ Fields::Named(f) => f.named.len(), ++ Fields::Unnamed(f) => f.unnamed.len(), ++ } ++ } ++ ++ /// Returns `true` if there are zero fields. ++ pub fn is_empty(&self) -> bool { ++ match self { ++ Fields::Unit => true, ++ Fields::Named(f) => f.named.is_empty(), ++ Fields::Unnamed(f) => f.unnamed.is_empty(), ++ } ++ } + } + + impl IntoIterator for Fields { +@@ -129,7 +147,7 @@ impl<'a> IntoIterator for &'a mut Fields { + ast_struct! { + /// A field of a struct or enum variant. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Field { + /// Attributes tagged on the field. +@@ -154,7 +172,7 @@ ast_enum_of_structs! { + /// The visibility level of an item: inherited or `pub` or + /// `pub(restricted)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Syntax tree enum +@@ -184,7 +202,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A public visibility level: `pub`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct VisPublic { + pub pub_token: Token![pub], +@@ -194,7 +212,7 @@ ast_struct! { + ast_struct! { + /// A crate-level visibility: `crate`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct VisCrate { + pub crate_token: Token![crate], +@@ -205,7 +223,7 @@ ast_struct! { + /// A visibility level restricted to some path: `pub(self)` or + /// `pub(super)` or `pub(crate)` or `pub(in some::module)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct VisRestricted { + pub pub_token: Token![pub], +@@ -220,12 +238,15 @@ pub mod parsing { + use super::*; + + use crate::ext::IdentExt; ++ use crate::parse::discouraged::Speculative; + use crate::parse::{Parse, ParseStream, Result}; + + impl Parse for Variant { + fn parse(input: ParseStream) -> Result<Self> { ++ let attrs = input.call(Attribute::parse_outer)?; ++ let _visibility: Visibility = input.parse()?; + Ok(Variant { +- attrs: input.call(Attribute::parse_outer)?, ++ attrs, + ident: input.parse()?, + fields: { + if input.peek(token::Brace) { +@@ -295,6 +316,17 @@ pub mod parsing { + + impl Parse for Visibility { + fn parse(input: ParseStream) -> Result<Self> { ++ // Recognize an empty None-delimited group, as produced by a $:vis ++ // matcher that matched no tokens. ++ if input.peek(token::Group) { ++ let ahead = input.fork(); ++ let group = crate::group::parse_group(&ahead)?; ++ if group.content.is_empty() { ++ input.advance_to(&ahead); ++ return Ok(Visibility::Inherited); ++ } ++ } ++ + if input.peek(Token![pub]) { + Self::parse_pub(input) + } else if input.peek(Token![crate]) { +@@ -310,27 +342,39 @@ pub mod parsing { + let pub_token = input.parse::<Token![pub]>()?; + + if input.peek(token::Paren) { +- // TODO: optimize using advance_to + let ahead = input.fork(); +- let mut content; +- parenthesized!(content in ahead); + ++ let content; ++ let paren_token = parenthesized!(content in ahead); + if content.peek(Token![crate]) + || content.peek(Token![self]) + || content.peek(Token![super]) + { +- return Ok(Visibility::Restricted(VisRestricted { +- pub_token, +- paren_token: parenthesized!(content in input), +- in_token: None, +- path: Box::new(Path::from(content.call(Ident::parse_any)?)), +- })); ++ let path = content.call(Ident::parse_any)?; ++ ++ // Ensure there are no additional tokens within `content`. ++ // Without explicitly checking, we may misinterpret a tuple ++ // field as a restricted visibility, causing a parse error. ++ // e.g. `pub (crate::A, crate::B)` (Issue #720). ++ if content.is_empty() { ++ input.advance_to(&ahead); ++ return Ok(Visibility::Restricted(VisRestricted { ++ pub_token, ++ paren_token, ++ in_token: None, ++ path: Box::new(Path::from(path)), ++ })); ++ } + } else if content.peek(Token![in]) { ++ let in_token: Token![in] = content.parse()?; ++ let path = content.call(Path::parse_mod_style)?; ++ ++ input.advance_to(&ahead); + return Ok(Visibility::Restricted(VisRestricted { + pub_token, +- paren_token: parenthesized!(content in input), +- in_token: Some(content.parse()?), +- path: Box::new(content.call(Path::parse_mod_style)?), ++ paren_token, ++ in_token: Some(in_token), ++ path: Box::new(path), + })); + } + } +@@ -347,6 +391,14 @@ pub mod parsing { + })) + } + } ++ ++ #[cfg(feature = "full")] ++ pub(crate) fn is_some(&self) -> bool { ++ match self { ++ Visibility::Inherited => false, ++ _ => true, ++ } ++ } + } + } + +diff --git a/third_party/rust/syn/src/derive.rs b/third_party/rust/syn/src/derive.rs +index 8cb9cf7b6d..3fa9d89a93 100644 +--- a/third_party/rust/syn/src/derive.rs ++++ b/third_party/rust/syn/src/derive.rs +@@ -4,7 +4,7 @@ use crate::punctuated::Punctuated; + ast_struct! { + /// Data structure sent to a `proc_macro_derive` macro. + /// +- /// *This type is available if Syn is built with the `"derive"` feature.* ++ /// *This type is available only if Syn is built with the `"derive"` feature.* + pub struct DeriveInput { + /// Attributes tagged on the whole struct or enum. + pub attrs: Vec<Attribute>, +@@ -26,7 +26,7 @@ ast_struct! { + ast_enum_of_structs! { + /// The storage of a struct, enum or union data structure. + /// +- /// *This type is available if Syn is built with the `"derive"` feature.* ++ /// *This type is available only if Syn is built with the `"derive"` feature.* + /// + /// # Syntax tree enum + /// +@@ -53,7 +53,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A struct input to a `proc_macro_derive` macro. + /// +- /// *This type is available if Syn is built with the `"derive"` ++ /// *This type is available only if Syn is built with the `"derive"` + /// feature.* + pub struct DataStruct { + pub struct_token: Token![struct], +@@ -65,7 +65,7 @@ ast_struct! { + ast_struct! { + /// An enum input to a `proc_macro_derive` macro. + /// +- /// *This type is available if Syn is built with the `"derive"` ++ /// *This type is available only if Syn is built with the `"derive"` + /// feature.* + pub struct DataEnum { + pub enum_token: Token![enum], +@@ -77,7 +77,7 @@ ast_struct! { + ast_struct! { + /// An untagged union input to a `proc_macro_derive` macro. + /// +- /// *This type is available if Syn is built with the `"derive"` ++ /// *This type is available only if Syn is built with the `"derive"` + /// feature.* + pub struct DataUnion { + pub union_token: Token![union], +diff --git a/third_party/rust/syn/src/discouraged.rs b/third_party/rust/syn/src/discouraged.rs +index 4d9ff93728..76c9fce6f8 100644 +--- a/third_party/rust/syn/src/discouraged.rs ++++ b/third_party/rust/syn/src/discouraged.rs +@@ -16,7 +16,7 @@ pub trait Speculative { + /// syntax of the form `A* B*` for arbitrary syntax `A` and `B`. The problem + /// is that when the fork fails to parse an `A`, it's impossible to tell + /// whether that was because of a syntax error and the user meant to provide +- /// an `A`, or that the `A`s are finished and its time to start parsing ++ /// an `A`, or that the `A`s are finished and it's time to start parsing + /// `B`s. Use with care. + /// + /// Also note that if `A` is a subset of `B`, `A* B*` can be parsed by +@@ -72,7 +72,6 @@ pub trait Speculative { + /// || input.peek(Token![self]) + /// || input.peek(Token![Self]) + /// || input.peek(Token![crate]) +- /// || input.peek(Token![extern]) + /// { + /// let ident = input.call(Ident::parse_any)?; + /// return Ok(PathSegment::from(ident)); +@@ -164,6 +163,30 @@ impl<'a> Speculative for ParseBuffer<'a> { + panic!("Fork was not derived from the advancing parse stream"); + } + ++ let (self_unexp, self_sp) = inner_unexpected(self); ++ let (fork_unexp, fork_sp) = inner_unexpected(fork); ++ if !Rc::ptr_eq(&self_unexp, &fork_unexp) { ++ match (fork_sp, self_sp) { ++ // Unexpected set on the fork, but not on `self`, copy it over. ++ (Some(span), None) => { ++ self_unexp.set(Unexpected::Some(span)); ++ } ++ // Unexpected unset. Use chain to propagate errors from fork. ++ (None, None) => { ++ fork_unexp.set(Unexpected::Chain(self_unexp)); ++ ++ // Ensure toplevel 'unexpected' tokens from the fork don't ++ // bubble up the chain by replacing the root `unexpected` ++ // pointer, only 'unexpected' tokens from existing group ++ // parsers should bubble. ++ fork.unexpected ++ .set(Some(Rc::new(Cell::new(Unexpected::None)))); ++ } ++ // Unexpected has been set on `self`. No changes needed. ++ (_, Some(_)) => {} ++ } ++ } ++ + // See comment on `cell` in the struct definition. + self.cell + .set(unsafe { mem::transmute::<Cursor, Cursor<'static>>(fork.cursor()) }) +diff --git a/third_party/rust/syn/src/error.rs b/third_party/rust/syn/src/error.rs +index 146d652299..dba34f9254 100644 +--- a/third_party/rust/syn/src/error.rs ++++ b/third_party/rust/syn/src/error.rs +@@ -1,4 +1,3 @@ +-use std; + use std::fmt::{self, Debug, Display}; + use std::iter::FromIterator; + use std::slice; +@@ -32,8 +31,8 @@ pub type Result<T> = std::result::Result<T, Error>; + /// conversion to `compile_error!` automatically. + /// + /// ``` +-/// extern crate proc_macro; +-/// ++/// # extern crate proc_macro; ++/// # + /// use proc_macro::TokenStream; + /// use syn::{parse_macro_input, AttributeArgs, ItemFn}; + /// +@@ -82,7 +81,6 @@ pub type Result<T> = std::result::Result<T, Error>; + /// # } + /// # } + /// ``` +-#[derive(Clone)] + pub struct Error { + messages: Vec<ErrorMessage>, + } +@@ -250,6 +248,17 @@ pub fn new_at<T: Display>(scope: Span, cursor: Cursor, message: T) -> Error { + } + } + ++#[cfg(all(feature = "parsing", any(feature = "full", feature = "derive")))] ++pub fn new2<T: Display>(start: Span, end: Span, message: T) -> Error { ++ Error { ++ messages: vec![ErrorMessage { ++ start_span: ThreadBound::new(start), ++ end_span: ThreadBound::new(end), ++ message: message.to_string(), ++ }], ++ } ++} ++ + impl Debug for Error { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + if self.messages.len() == 1 { +@@ -278,6 +287,14 @@ impl Display for Error { + } + } + ++impl Clone for Error { ++ fn clone(&self) -> Self { ++ Error { ++ messages: self.messages.clone(), ++ } ++ } ++} ++ + impl Clone for ErrorMessage { + fn clone(&self) -> Self { + let start = self +@@ -355,3 +372,11 @@ impl<'a> Iterator for Iter<'a> { + }) + } + } ++ ++impl Extend<Error> for Error { ++ fn extend<T: IntoIterator<Item = Error>>(&mut self, iter: T) { ++ for err in iter { ++ self.combine(err); ++ } ++ } ++} +diff --git a/third_party/rust/syn/src/expr.rs b/third_party/rust/syn/src/expr.rs +index 2874a463aa..2fe0e0b5d8 100644 +--- a/third_party/rust/syn/src/expr.rs ++++ b/third_party/rust/syn/src/expr.rs +@@ -1,18 +1,21 @@ + use super::*; + use crate::punctuated::Punctuated; +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; ++#[cfg(feature = "full")] ++use crate::reserved::Reserved; + use proc_macro2::{Span, TokenStream}; +-#[cfg(feature = "extra-traits")] ++#[cfg(feature = "printing")] ++use quote::IdentFragment; ++#[cfg(feature = "printing")] ++use std::fmt::{self, Display}; + use std::hash::{Hash, Hasher}; +-#[cfg(all(feature = "parsing", feature = "full"))] ++#[cfg(feature = "parsing")] + use std::mem; + + ast_enum_of_structs! { + /// A Rust expression. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` +- /// feature.* ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` ++ /// feature, but most of the variants are not available unless "full" is enabled.* + /// + /// # Syntax tree enums + /// +@@ -83,7 +86,7 @@ ast_enum_of_structs! { + /// A sign that you may not be choosing the right variable names is if you + /// see names getting repeated in your code, like accessing + /// `receiver.receiver` or `pat.pat` or `cond.cond`. +- pub enum Expr #manual_extra_traits { ++ pub enum Expr { + /// A slice literal expression: `[a, b, c, d]`. + Array(ExprArray), + +@@ -228,7 +231,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A slice literal expression: `[a, b, c, d]`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprArray #full { + pub attrs: Vec<Attribute>, + pub bracket_token: token::Bracket, +@@ -239,7 +242,7 @@ ast_struct! { + ast_struct! { + /// An assignment expression: `a = compute()`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprAssign #full { + pub attrs: Vec<Attribute>, + pub left: Box<Expr>, +@@ -251,7 +254,7 @@ ast_struct! { + ast_struct! { + /// A compound assignment expression: `counter += 1`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprAssignOp #full { + pub attrs: Vec<Attribute>, + pub left: Box<Expr>, +@@ -263,7 +266,7 @@ ast_struct! { + ast_struct! { + /// An async block: `async { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprAsync #full { + pub attrs: Vec<Attribute>, + pub async_token: Token![async], +@@ -275,7 +278,7 @@ ast_struct! { + ast_struct! { + /// An await expression: `fut.await`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprAwait #full { + pub attrs: Vec<Attribute>, + pub base: Box<Expr>, +@@ -287,7 +290,7 @@ ast_struct! { + ast_struct! { + /// A binary operation: `a + b`, `a * b`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprBinary { + pub attrs: Vec<Attribute>, +@@ -300,7 +303,7 @@ ast_struct! { + ast_struct! { + /// A blocked scope: `{ ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprBlock #full { + pub attrs: Vec<Attribute>, + pub label: Option<Label>, +@@ -311,7 +314,7 @@ ast_struct! { + ast_struct! { + /// A box expression: `box f`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprBox #full { + pub attrs: Vec<Attribute>, + pub box_token: Token![box], +@@ -323,7 +326,7 @@ ast_struct! { + /// A `break`, with an optional label to break and an optional + /// expression. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprBreak #full { + pub attrs: Vec<Attribute>, + pub break_token: Token![break], +@@ -335,7 +338,7 @@ ast_struct! { + ast_struct! { + /// A function call expression: `invoke(a, b)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprCall { + pub attrs: Vec<Attribute>, +@@ -348,7 +351,7 @@ ast_struct! { + ast_struct! { + /// A cast expression: `foo as f64`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprCast { + pub attrs: Vec<Attribute>, +@@ -361,7 +364,7 @@ ast_struct! { + ast_struct! { + /// A closure expression: `|a, b| a + b`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprClosure #full { + pub attrs: Vec<Attribute>, + pub asyncness: Option<Token![async]>, +@@ -378,7 +381,7 @@ ast_struct! { + ast_struct! { + /// A `continue`, with an optional label. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprContinue #full { + pub attrs: Vec<Attribute>, + pub continue_token: Token![continue], +@@ -390,7 +393,7 @@ ast_struct! { + /// Access of a named struct field (`obj.k`) or unnamed tuple struct + /// field (`obj.0`). + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprField { + pub attrs: Vec<Attribute>, + pub base: Box<Expr>, +@@ -402,7 +405,7 @@ ast_struct! { + ast_struct! { + /// A for loop: `for pat in expr { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprForLoop #full { + pub attrs: Vec<Attribute>, + pub label: Option<Label>, +@@ -421,7 +424,7 @@ ast_struct! { + /// of expressions and is related to `None`-delimited spans in a + /// `TokenStream`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprGroup #full { + pub attrs: Vec<Attribute>, + pub group_token: token::Group, +@@ -436,7 +439,7 @@ ast_struct! { + /// The `else` branch expression may only be an `If` or `Block` + /// expression, not any of the other types of expression. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprIf #full { + pub attrs: Vec<Attribute>, + pub if_token: Token![if], +@@ -449,7 +452,7 @@ ast_struct! { + ast_struct! { + /// A square bracketed indexing expression: `vector[2]`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprIndex { + pub attrs: Vec<Attribute>, +@@ -462,7 +465,7 @@ ast_struct! { + ast_struct! { + /// A `let` guard: `let Some(x) = opt`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprLet #full { + pub attrs: Vec<Attribute>, + pub let_token: Token![let], +@@ -475,7 +478,7 @@ ast_struct! { + ast_struct! { + /// A literal in place of an expression: `1`, `"foo"`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprLit { + pub attrs: Vec<Attribute>, +@@ -486,7 +489,7 @@ ast_struct! { + ast_struct! { + /// Conditionless loop: `loop { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprLoop #full { + pub attrs: Vec<Attribute>, + pub label: Option<Label>, +@@ -498,7 +501,7 @@ ast_struct! { + ast_struct! { + /// A macro invocation expression: `format!("{}", q)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprMacro #full { + pub attrs: Vec<Attribute>, + pub mac: Macro, +@@ -508,7 +511,7 @@ ast_struct! { + ast_struct! { + /// A `match` expression: `match n { Some(n) => {}, None => {} }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprMatch #full { + pub attrs: Vec<Attribute>, + pub match_token: Token![match], +@@ -521,7 +524,7 @@ ast_struct! { + ast_struct! { + /// A method call expression: `x.foo::<T>(a, b)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprMethodCall #full { + pub attrs: Vec<Attribute>, + pub receiver: Box<Expr>, +@@ -536,7 +539,7 @@ ast_struct! { + ast_struct! { + /// A parenthesized expression: `(a + b)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprParen { + pub attrs: Vec<Attribute>, + pub paren_token: token::Paren, +@@ -550,7 +553,7 @@ ast_struct! { + /// + /// A plain identifier like `x` is a path of length 1. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprPath { + pub attrs: Vec<Attribute>, +@@ -562,7 +565,7 @@ ast_struct! { + ast_struct! { + /// A range expression: `1..2`, `1..`, `..2`, `1..=2`, `..=2`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprRange #full { + pub attrs: Vec<Attribute>, + pub from: Option<Box<Expr>>, +@@ -574,7 +577,7 @@ ast_struct! { + ast_struct! { + /// A referencing operation: `&a` or `&mut a`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprReference #full { + pub attrs: Vec<Attribute>, + pub and_token: Token![&], +@@ -587,7 +590,7 @@ ast_struct! { + ast_struct! { + /// An array literal constructed from one repeated element: `[0u8; N]`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprRepeat #full { + pub attrs: Vec<Attribute>, + pub bracket_token: token::Bracket, +@@ -600,7 +603,7 @@ ast_struct! { + ast_struct! { + /// A `return`, with an optional value to be returned. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprReturn #full { + pub attrs: Vec<Attribute>, + pub return_token: Token![return], +@@ -614,7 +617,7 @@ ast_struct! { + /// The `rest` provides the value of the remaining fields as in `S { a: + /// 1, b: 1, ..rest }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprStruct #full { + pub attrs: Vec<Attribute>, + pub path: Path, +@@ -628,7 +631,7 @@ ast_struct! { + ast_struct! { + /// A try-expression: `expr?`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprTry #full { + pub attrs: Vec<Attribute>, + pub expr: Box<Expr>, +@@ -639,7 +642,7 @@ ast_struct! { + ast_struct! { + /// A try block: `try { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprTryBlock #full { + pub attrs: Vec<Attribute>, + pub try_token: Token![try], +@@ -650,7 +653,7 @@ ast_struct! { + ast_struct! { + /// A tuple expression: `(a, b, c, d)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprTuple #full { + pub attrs: Vec<Attribute>, + pub paren_token: token::Paren, +@@ -661,7 +664,7 @@ ast_struct! { + ast_struct! { + /// A type ascription expression: `foo: f64`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprType #full { + pub attrs: Vec<Attribute>, + pub expr: Box<Expr>, +@@ -673,7 +676,7 @@ ast_struct! { + ast_struct! { + /// A unary operation: `!x`, `*x`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ExprUnary { + pub attrs: Vec<Attribute>, +@@ -685,7 +688,7 @@ ast_struct! { + ast_struct! { + /// An unsafe block: `unsafe { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprUnsafe #full { + pub attrs: Vec<Attribute>, + pub unsafe_token: Token![unsafe], +@@ -696,7 +699,7 @@ ast_struct! { + ast_struct! { + /// A while loop: `while expr { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprWhile #full { + pub attrs: Vec<Attribute>, + pub label: Option<Label>, +@@ -709,7 +712,7 @@ ast_struct! { + ast_struct! { + /// A yield expression: `yield expr`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ExprYield #full { + pub attrs: Vec<Attribute>, + pub yield_token: Token![yield], +@@ -717,232 +720,6 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Expr {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Expr { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (Expr::Array(this), Expr::Array(other)) => this == other, +- (Expr::Assign(this), Expr::Assign(other)) => this == other, +- (Expr::AssignOp(this), Expr::AssignOp(other)) => this == other, +- (Expr::Async(this), Expr::Async(other)) => this == other, +- (Expr::Await(this), Expr::Await(other)) => this == other, +- (Expr::Binary(this), Expr::Binary(other)) => this == other, +- (Expr::Block(this), Expr::Block(other)) => this == other, +- (Expr::Box(this), Expr::Box(other)) => this == other, +- (Expr::Break(this), Expr::Break(other)) => this == other, +- (Expr::Call(this), Expr::Call(other)) => this == other, +- (Expr::Cast(this), Expr::Cast(other)) => this == other, +- (Expr::Closure(this), Expr::Closure(other)) => this == other, +- (Expr::Continue(this), Expr::Continue(other)) => this == other, +- (Expr::Field(this), Expr::Field(other)) => this == other, +- (Expr::ForLoop(this), Expr::ForLoop(other)) => this == other, +- (Expr::Group(this), Expr::Group(other)) => this == other, +- (Expr::If(this), Expr::If(other)) => this == other, +- (Expr::Index(this), Expr::Index(other)) => this == other, +- (Expr::Let(this), Expr::Let(other)) => this == other, +- (Expr::Lit(this), Expr::Lit(other)) => this == other, +- (Expr::Loop(this), Expr::Loop(other)) => this == other, +- (Expr::Macro(this), Expr::Macro(other)) => this == other, +- (Expr::Match(this), Expr::Match(other)) => this == other, +- (Expr::MethodCall(this), Expr::MethodCall(other)) => this == other, +- (Expr::Paren(this), Expr::Paren(other)) => this == other, +- (Expr::Path(this), Expr::Path(other)) => this == other, +- (Expr::Range(this), Expr::Range(other)) => this == other, +- (Expr::Reference(this), Expr::Reference(other)) => this == other, +- (Expr::Repeat(this), Expr::Repeat(other)) => this == other, +- (Expr::Return(this), Expr::Return(other)) => this == other, +- (Expr::Struct(this), Expr::Struct(other)) => this == other, +- (Expr::Try(this), Expr::Try(other)) => this == other, +- (Expr::TryBlock(this), Expr::TryBlock(other)) => this == other, +- (Expr::Tuple(this), Expr::Tuple(other)) => this == other, +- (Expr::Type(this), Expr::Type(other)) => this == other, +- (Expr::Unary(this), Expr::Unary(other)) => this == other, +- (Expr::Unsafe(this), Expr::Unsafe(other)) => this == other, +- (Expr::Verbatim(this), Expr::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- (Expr::While(this), Expr::While(other)) => this == other, +- (Expr::Yield(this), Expr::Yield(other)) => this == other, +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Expr { +- fn hash<H>(&self, hash: &mut H) +- where +- H: Hasher, +- { +- match self { +- Expr::Array(expr) => { +- hash.write_u8(0); +- expr.hash(hash); +- } +- Expr::Assign(expr) => { +- hash.write_u8(1); +- expr.hash(hash); +- } +- Expr::AssignOp(expr) => { +- hash.write_u8(2); +- expr.hash(hash); +- } +- Expr::Async(expr) => { +- hash.write_u8(3); +- expr.hash(hash); +- } +- Expr::Await(expr) => { +- hash.write_u8(4); +- expr.hash(hash); +- } +- Expr::Binary(expr) => { +- hash.write_u8(5); +- expr.hash(hash); +- } +- Expr::Block(expr) => { +- hash.write_u8(6); +- expr.hash(hash); +- } +- Expr::Box(expr) => { +- hash.write_u8(7); +- expr.hash(hash); +- } +- Expr::Break(expr) => { +- hash.write_u8(8); +- expr.hash(hash); +- } +- Expr::Call(expr) => { +- hash.write_u8(9); +- expr.hash(hash); +- } +- Expr::Cast(expr) => { +- hash.write_u8(10); +- expr.hash(hash); +- } +- Expr::Closure(expr) => { +- hash.write_u8(11); +- expr.hash(hash); +- } +- Expr::Continue(expr) => { +- hash.write_u8(12); +- expr.hash(hash); +- } +- Expr::Field(expr) => { +- hash.write_u8(13); +- expr.hash(hash); +- } +- Expr::ForLoop(expr) => { +- hash.write_u8(14); +- expr.hash(hash); +- } +- Expr::Group(expr) => { +- hash.write_u8(15); +- expr.hash(hash); +- } +- Expr::If(expr) => { +- hash.write_u8(16); +- expr.hash(hash); +- } +- Expr::Index(expr) => { +- hash.write_u8(17); +- expr.hash(hash); +- } +- Expr::Let(expr) => { +- hash.write_u8(18); +- expr.hash(hash); +- } +- Expr::Lit(expr) => { +- hash.write_u8(19); +- expr.hash(hash); +- } +- Expr::Loop(expr) => { +- hash.write_u8(20); +- expr.hash(hash); +- } +- Expr::Macro(expr) => { +- hash.write_u8(21); +- expr.hash(hash); +- } +- Expr::Match(expr) => { +- hash.write_u8(22); +- expr.hash(hash); +- } +- Expr::MethodCall(expr) => { +- hash.write_u8(23); +- expr.hash(hash); +- } +- Expr::Paren(expr) => { +- hash.write_u8(24); +- expr.hash(hash); +- } +- Expr::Path(expr) => { +- hash.write_u8(25); +- expr.hash(hash); +- } +- Expr::Range(expr) => { +- hash.write_u8(26); +- expr.hash(hash); +- } +- Expr::Reference(expr) => { +- hash.write_u8(27); +- expr.hash(hash); +- } +- Expr::Repeat(expr) => { +- hash.write_u8(28); +- expr.hash(hash); +- } +- Expr::Return(expr) => { +- hash.write_u8(29); +- expr.hash(hash); +- } +- Expr::Struct(expr) => { +- hash.write_u8(30); +- expr.hash(hash); +- } +- Expr::Try(expr) => { +- hash.write_u8(31); +- expr.hash(hash); +- } +- Expr::TryBlock(expr) => { +- hash.write_u8(32); +- expr.hash(hash); +- } +- Expr::Tuple(expr) => { +- hash.write_u8(33); +- expr.hash(hash); +- } +- Expr::Type(expr) => { +- hash.write_u8(34); +- expr.hash(hash); +- } +- Expr::Unary(expr) => { +- hash.write_u8(35); +- expr.hash(hash); +- } +- Expr::Unsafe(expr) => { +- hash.write_u8(36); +- expr.hash(hash); +- } +- Expr::Verbatim(expr) => { +- hash.write_u8(37); +- TokenStreamHelper(expr).hash(hash); +- } +- Expr::While(expr) => { +- hash.write_u8(38); +- expr.hash(hash); +- } +- Expr::Yield(expr) => { +- hash.write_u8(39); +- expr.hash(hash); +- } +- Expr::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + impl Expr { + #[cfg(all(feature = "parsing", feature = "full"))] + pub(crate) fn replace_attrs(&mut self, new: Vec<Attribute>) -> Vec<Attribute> { +@@ -996,7 +773,7 @@ ast_enum! { + /// A struct or tuple struct field accessed in a struct literal or field + /// expression. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum Member { + /// A named field like `self.x`. +@@ -1006,12 +783,50 @@ ast_enum! { + } + } + ++impl Eq for Member {} ++ ++impl PartialEq for Member { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Member::Named(this), Member::Named(other)) => this == other, ++ (Member::Unnamed(this), Member::Unnamed(other)) => this == other, ++ _ => false, ++ } ++ } ++} ++ ++impl Hash for Member { ++ fn hash<H: Hasher>(&self, state: &mut H) { ++ match self { ++ Member::Named(m) => m.hash(state), ++ Member::Unnamed(m) => m.hash(state), ++ } ++ } ++} ++ ++#[cfg(feature = "printing")] ++impl IdentFragment for Member { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Member::Named(m) => Display::fmt(m, formatter), ++ Member::Unnamed(m) => Display::fmt(&m.index, formatter), ++ } ++ } ++ ++ fn span(&self) -> Option<Span> { ++ match self { ++ Member::Named(m) => Some(m.span()), ++ Member::Unnamed(m) => Some(m.span), ++ } ++ } ++} ++ + ast_struct! { + /// The index of an unnamed tuple struct field. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- pub struct Index #manual_extra_traits { ++ pub struct Index { + pub index: u32, + pub span: Span, + } +@@ -1027,28 +842,28 @@ impl From<usize> for Index { + } + } + +-#[cfg(feature = "extra-traits")] + impl Eq for Index {} + +-#[cfg(feature = "extra-traits")] + impl PartialEq for Index { + fn eq(&self, other: &Self) -> bool { + self.index == other.index + } + } + +-#[cfg(feature = "extra-traits")] + impl Hash for Index { + fn hash<H: Hasher>(&self, state: &mut H) { + self.index.hash(state); + } + } + +-#[cfg(feature = "full")] +-ast_struct! { +- #[derive(Default)] +- pub struct Reserved { +- private: (), ++#[cfg(feature = "printing")] ++impl IdentFragment for Index { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ Display::fmt(&self.index, formatter) ++ } ++ ++ fn span(&self) -> Option<Span> { ++ Some(self.span) + } + } + +@@ -1057,7 +872,7 @@ ast_struct! { + /// The `::<>` explicit type parameters passed to a method call: + /// `parse::<u64>()`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct MethodTurbofish { + pub colon2_token: Token![::], + pub lt_token: Token![<], +@@ -1070,7 +885,7 @@ ast_struct! { + ast_enum! { + /// An individual generic argument to a method, like `T`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub enum GenericMethodArgument { + /// A type argument. + Type(Type), +@@ -1086,7 +901,7 @@ ast_enum! { + ast_struct! { + /// A field-value pair in a struct literal. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct FieldValue { + /// Attributes tagged on the field. + pub attrs: Vec<Attribute>, +@@ -1107,7 +922,7 @@ ast_struct! { + ast_struct! { + /// A lifetime labeling a `for`, `while`, or `loop`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Label { + pub name: Lifetime, + pub colon_token: Token![:], +@@ -1134,7 +949,7 @@ ast_struct! { + /// # } + /// ``` + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Arm { + pub attrs: Vec<Attribute>, + pub pat: Pat, +@@ -1149,8 +964,7 @@ ast_struct! { + ast_enum! { + /// Limit types of a range, inclusive or exclusive. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* +- #[cfg_attr(feature = "clone-impls", derive(Copy))] ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub enum RangeLimits { + /// Inclusive at the beginning, exclusive at the end. + HalfOpen(Token![..]), +@@ -1162,7 +976,7 @@ ast_enum! { + #[cfg(any(feature = "parsing", feature = "printing"))] + #[cfg(feature = "full")] + pub(crate) fn requires_terminator(expr: &Expr) -> bool { +- // see https://github.com/rust-lang/rust/blob/eb8f2586e/src/libsyntax/parse/classify.rs#L17-L37 ++ // see https://github.com/rust-lang/rust/blob/2679c38fc/src/librustc_ast/util/classify.rs#L7-L25 + match *expr { + Expr::Unsafe(..) + | Expr::Block(..) +@@ -1183,16 +997,17 @@ pub(crate) mod parsing { + + use crate::parse::{Parse, ParseStream, Result}; + use crate::path; ++ use std::cmp::Ordering; ++ ++ crate::custom_keyword!(raw); + + // When we're parsing expressions which occur before blocks, like in an if + // statement's condition, we cannot parse a struct literal. + // + // Struct literals are ambiguous in certain positions + // https://github.com/rust-lang/rfcs/pull/92 +- #[derive(Copy, Clone)] + pub struct AllowStruct(bool); + +- #[derive(Copy, Clone, PartialEq, PartialOrd)] + enum Precedence { + Any, + Assign, +@@ -1246,9 +1061,121 @@ pub(crate) mod parsing { + } + } + +- #[cfg(feature = "full")] +- fn expr_no_struct(input: ParseStream) -> Result<Expr> { +- ambiguous_expr(input, AllowStruct(false)) ++ impl Expr { ++ /// An alternative to the primary `Expr::parse` parser (from the ++ /// [`Parse`] trait) for ambiguous syntactic positions in which a ++ /// trailing brace should not be taken as part of the expression. ++ /// ++ /// Rust grammar has an ambiguity where braces sometimes turn a path ++ /// expression into a struct initialization and sometimes do not. In the ++ /// following code, the expression `S {}` is one expression. Presumably ++ /// there is an empty struct `struct S {}` defined somewhere which it is ++ /// instantiating. ++ /// ++ /// ``` ++ /// # struct S; ++ /// # impl std::ops::Deref for S { ++ /// # type Target = bool; ++ /// # fn deref(&self) -> &Self::Target { ++ /// # &true ++ /// # } ++ /// # } ++ /// let _ = *S {}; ++ /// ++ /// // parsed by rustc as: `*(S {})` ++ /// ``` ++ /// ++ /// We would want to parse the above using `Expr::parse` after the `=` ++ /// token. ++ /// ++ /// But in the following, `S {}` is *not* a struct init expression. ++ /// ++ /// ``` ++ /// # const S: &bool = &true; ++ /// if *S {} {} ++ /// ++ /// // parsed by rustc as: ++ /// // ++ /// // if (*S) { ++ /// // /* empty block */ ++ /// // } ++ /// // { ++ /// // /* another empty block */ ++ /// // } ++ /// ``` ++ /// ++ /// For that reason we would want to parse if-conditions using ++ /// `Expr::parse_without_eager_brace` after the `if` token. Same for ++ /// similar syntactic positions such as the condition expr after a ++ /// `while` token or the expr at the top of a `match`. ++ /// ++ /// The Rust grammar's choices around which way this ambiguity is ++ /// resolved at various syntactic positions is fairly arbitrary. Really ++ /// either parse behavior could work in most positions, and language ++ /// designers just decide each case based on which is more likely to be ++ /// what the programmer had in mind most of the time. ++ /// ++ /// ``` ++ /// # struct S; ++ /// # fn doc() -> S { ++ /// if return S {} {} ++ /// # unreachable!() ++ /// # } ++ /// ++ /// // parsed by rustc as: ++ /// // ++ /// // if (return (S {})) { ++ /// // } ++ /// // ++ /// // but could equally well have been this other arbitrary choice: ++ /// // ++ /// // if (return S) { ++ /// // } ++ /// // {} ++ /// ``` ++ /// ++ /// Note the grammar ambiguity on trailing braces is distinct from ++ /// precedence and is not captured by assigning a precedence level to ++ /// the braced struct init expr in relation to other operators. This can ++ /// be illustrated by `return 0..S {}` vs `match 0..S {}`. The former ++ /// parses as `return (0..(S {}))` implying tighter precedence for ++ /// struct init than `..`, while the latter parses as `match (0..S) {}` ++ /// implying tighter precedence for `..` than struct init, a ++ /// contradiction. ++ #[cfg(feature = "full")] ++ pub fn parse_without_eager_brace(input: ParseStream) -> Result<Expr> { ++ ambiguous_expr(input, AllowStruct(false)) ++ } ++ } ++ ++ impl Copy for AllowStruct {} ++ ++ impl Clone for AllowStruct { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } ++ ++ impl Copy for Precedence {} ++ ++ impl Clone for Precedence { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } ++ ++ impl PartialEq for Precedence { ++ fn eq(&self, other: &Self) -> bool { ++ *self as u8 == *other as u8 ++ } ++ } ++ ++ impl PartialOrd for Precedence { ++ fn partial_cmp(&self, other: &Self) -> Option<Ordering> { ++ let this = *self as u8; ++ let other = *other as u8; ++ Some(this.cmp(&other)) ++ } + } + + #[cfg(feature = "full")] +@@ -1430,56 +1357,84 @@ pub(crate) mod parsing { + parse_expr(input, lhs, allow_struct, Precedence::Any) + } + ++ #[cfg(feature = "full")] ++ fn expr_attrs(input: ParseStream) -> Result<Vec<Attribute>> { ++ let mut attrs = Vec::new(); ++ loop { ++ if input.peek(token::Group) { ++ let ahead = input.fork(); ++ let group = crate::group::parse_group(&ahead)?; ++ if !group.content.peek(Token![#]) || group.content.peek2(Token![!]) { ++ break; ++ } ++ let attr = group.content.call(attr::parsing::single_parse_outer)?; ++ if !group.content.is_empty() { ++ break; ++ } ++ attrs.push(attr); ++ } else if input.peek(Token![#]) { ++ attrs.push(input.call(attr::parsing::single_parse_outer)?); ++ } else { ++ break; ++ } ++ } ++ Ok(attrs) ++ } ++ + // <UnOp> <trailer> + // & <trailer> + // &mut <trailer> + // box <trailer> + #[cfg(feature = "full")] + fn unary_expr(input: ParseStream, allow_struct: AllowStruct) -> Result<Expr> { +- // TODO: optimize using advance_to +- let ahead = input.fork(); +- ahead.call(Attribute::parse_outer)?; +- if ahead.peek(Token![&]) +- || ahead.peek(Token![box]) +- || ahead.peek(Token![*]) +- || ahead.peek(Token![!]) +- || ahead.peek(Token![-]) +- { +- let attrs = input.call(Attribute::parse_outer)?; +- if input.peek(Token![&]) { ++ let begin = input.fork(); ++ let attrs = input.call(expr_attrs)?; ++ if input.peek(Token![&]) { ++ let and_token: Token![&] = input.parse()?; ++ let raw: Option<raw> = ++ if input.peek(raw) && (input.peek2(Token![mut]) || input.peek2(Token![const])) { ++ Some(input.parse()?) ++ } else { ++ None ++ }; ++ let mutability: Option<Token![mut]> = input.parse()?; ++ if raw.is_some() && mutability.is_none() { ++ input.parse::<Token![const]>()?; ++ } ++ let expr = Box::new(unary_expr(input, allow_struct)?); ++ if raw.is_some() { ++ Ok(Expr::Verbatim(verbatim::between(begin, input))) ++ } else { + Ok(Expr::Reference(ExprReference { + attrs, +- and_token: input.parse()?, ++ and_token, + raw: Reserved::default(), +- mutability: input.parse()?, +- expr: Box::new(unary_expr(input, allow_struct)?), +- })) +- } else if input.peek(Token![box]) { +- Ok(Expr::Box(ExprBox { +- attrs, +- box_token: input.parse()?, +- expr: Box::new(unary_expr(input, allow_struct)?), +- })) +- } else { +- Ok(Expr::Unary(ExprUnary { +- attrs, +- op: input.parse()?, +- expr: Box::new(unary_expr(input, allow_struct)?), ++ mutability, ++ expr, + })) + } ++ } else if input.peek(Token![box]) { ++ Ok(Expr::Box(ExprBox { ++ attrs, ++ box_token: input.parse()?, ++ expr: Box::new(unary_expr(input, allow_struct)?), ++ })) ++ } else if input.peek(Token![*]) || input.peek(Token![!]) || input.peek(Token![-]) { ++ Ok(Expr::Unary(ExprUnary { ++ attrs, ++ op: input.parse()?, ++ expr: Box::new(unary_expr(input, allow_struct)?), ++ })) + } else { +- trailer_expr(input, allow_struct) ++ trailer_expr(attrs, input, allow_struct) + } + } + + #[cfg(not(feature = "full"))] + fn unary_expr(input: ParseStream, allow_struct: AllowStruct) -> Result<Expr> { +- // TODO: optimize using advance_to +- let ahead = input.fork(); +- ahead.call(Attribute::parse_outer)?; +- if ahead.peek(Token![*]) || ahead.peek(Token![!]) || ahead.peek(Token![-]) { ++ if input.peek(Token![*]) || input.peek(Token![!]) || input.peek(Token![-]) { + Ok(Expr::Unary(ExprUnary { +- attrs: input.call(Attribute::parse_outer)?, ++ attrs: Vec::new(), + op: input.parse()?, + expr: Box::new(unary_expr(input, allow_struct)?), + })) +@@ -1495,13 +1450,11 @@ pub(crate) mod parsing { + // <atom> [ <expr> ] ... + // <atom> ? ... + #[cfg(feature = "full")] +- fn trailer_expr(input: ParseStream, allow_struct: AllowStruct) -> Result<Expr> { +- if input.peek(token::Group) { +- return input.call(expr_group).map(Expr::Group); +- } +- +- let outer_attrs = input.call(Attribute::parse_outer)?; +- ++ fn trailer_expr( ++ outer_attrs: Vec<Attribute>, ++ input: ParseStream, ++ allow_struct: AllowStruct, ++ ) -> Result<Expr> { + let atom = atom_expr(input, allow_struct)?; + let mut e = trailer_helper(input, atom)?; + +@@ -1523,18 +1476,26 @@ pub(crate) mod parsing { + args: content.parse_terminated(Expr::parse)?, + }); + } else if input.peek(Token![.]) && !input.peek(Token![..]) { +- let dot_token: Token![.] = input.parse()?; ++ let mut dot_token: Token![.] = input.parse()?; + +- if input.peek(token::Await) { ++ let await_token: Option<token::Await> = input.parse()?; ++ if let Some(await_token) = await_token { + e = Expr::Await(ExprAwait { + attrs: Vec::new(), + base: Box::new(e), + dot_token, +- await_token: input.parse()?, ++ await_token, + }); + continue; + } + ++ let float_token: Option<LitFloat> = input.parse()?; ++ if let Some(float_token) = float_token { ++ if multi_index(&mut e, &mut dot_token, float_token)? { ++ continue; ++ } ++ } ++ + let member: Member = input.parse()?; + let turbofish = if member.is_named() && input.peek(Token![::]) { + Some(MethodTurbofish { +@@ -1620,10 +1581,17 @@ pub(crate) mod parsing { + }); + } else if input.peek(Token![.]) && !input.peek(Token![..]) && !input.peek2(token::Await) + { ++ let mut dot_token: Token![.] = input.parse()?; ++ let float_token: Option<LitFloat> = input.parse()?; ++ if let Some(float_token) = float_token { ++ if multi_index(&mut e, &mut dot_token, float_token)? { ++ continue; ++ } ++ } + e = Expr::Field(ExprField { + attrs: Vec::new(), + base: Box::new(e), +- dot_token: input.parse()?, ++ dot_token, + member: input.parse()?, + }); + } else if input.peek(token::Bracket) { +@@ -1646,7 +1614,11 @@ pub(crate) mod parsing { + // interactions, as they are fully contained. + #[cfg(feature = "full")] + fn atom_expr(input: ParseStream, allow_struct: AllowStruct) -> Result<Expr> { +- if input.peek(token::Group) { ++ if input.peek(token::Group) ++ && !input.peek2(Token![::]) ++ && !input.peek2(Token![!]) ++ && !input.peek2(token::Brace) ++ { + input.call(expr_group).map(Expr::Group) + } else if input.peek(Lit) { + input.parse().map(Expr::Lit) +@@ -1668,7 +1640,6 @@ pub(crate) mod parsing { + || input.peek(Token![self]) + || input.peek(Token![Self]) + || input.peek(Token![super]) +- || input.peek(Token![extern]) + || input.peek(Token![crate]) + { + path_or_macro_or_struct(input, allow_struct) +@@ -1740,7 +1711,6 @@ pub(crate) mod parsing { + || input.peek(Token![self]) + || input.peek(Token![Self]) + || input.peek(Token![super]) +- || input.peek(Token![extern]) + || input.peek(Token![crate]) + { + input.parse().map(Expr::Path) +@@ -1878,7 +1848,7 @@ pub(crate) mod parsing { + + #[cfg(feature = "full")] + pub(crate) fn expr_early(input: ParseStream) -> Result<Expr> { +- let mut attrs = input.call(Attribute::parse_outer)?; ++ let mut attrs = input.call(expr_attrs)?; + let mut expr = if input.peek(Token![if]) { + Expr::If(input.parse()?) + } else if input.peek(Token![while]) { +@@ -1905,7 +1875,7 @@ pub(crate) mod parsing { + return parse_expr(input, expr, allow_struct, Precedence::Any); + }; + +- if input.peek(Token![.]) || input.peek(Token![?]) { ++ if input.peek(Token![.]) && !input.peek(Token![..]) || input.peek(Token![?]) { + expr = trailer_helper(input, expr)?; + + attrs.extend(expr.replace_attrs(Vec::new())); +@@ -1951,7 +1921,16 @@ pub(crate) mod parsing { + + #[cfg(feature = "full")] + fn generic_method_argument(input: ParseStream) -> Result<GenericMethodArgument> { +- // TODO parse const generics as well ++ if input.peek(Lit) { ++ let lit = input.parse()?; ++ return Ok(GenericMethodArgument::Const(Expr::Lit(lit))); ++ } ++ ++ if input.peek(token::Brace) { ++ let block = input.call(expr::parsing::expr_block)?; ++ return Ok(GenericMethodArgument::Const(Expr::Block(block))); ++ } ++ + input.parse().map(GenericMethodArgument::Type) + } + +@@ -1960,44 +1939,20 @@ pub(crate) mod parsing { + Ok(ExprLet { + attrs: Vec::new(), + let_token: input.parse()?, +- pat: { +- let leading_vert: Option<Token![|]> = input.parse()?; +- let pat: Pat = input.parse()?; +- if leading_vert.is_some() +- || input.peek(Token![|]) && !input.peek(Token![||]) && !input.peek(Token![|=]) +- { +- let mut cases = Punctuated::new(); +- cases.push_value(pat); +- while input.peek(Token![|]) +- && !input.peek(Token![||]) +- && !input.peek(Token![|=]) +- { +- let punct = input.parse()?; +- cases.push_punct(punct); +- let pat: Pat = input.parse()?; +- cases.push_value(pat); +- } +- Pat::Or(PatOr { +- attrs: Vec::new(), +- leading_vert, +- cases, +- }) +- } else { +- pat +- } +- }, ++ pat: pat::parsing::multi_pat_with_leading_vert(input)?, + eq_token: input.parse()?, +- expr: Box::new(input.call(expr_no_struct)?), ++ expr: Box::new(input.call(Expr::parse_without_eager_brace)?), + }) + } + + #[cfg(feature = "full")] + impl Parse for ExprIf { + fn parse(input: ParseStream) -> Result<Self> { ++ let attrs = input.call(Attribute::parse_outer)?; + Ok(ExprIf { +- attrs: Vec::new(), ++ attrs, + if_token: input.parse()?, +- cond: Box::new(input.call(expr_no_struct)?), ++ cond: Box::new(input.call(Expr::parse_without_eager_brace)?), + then_branch: input.parse()?, + else_branch: { + if input.peek(Token![else]) { +@@ -2033,29 +1988,14 @@ pub(crate) mod parsing { + #[cfg(feature = "full")] + impl Parse for ExprForLoop { + fn parse(input: ParseStream) -> Result<Self> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; + let label: Option<Label> = input.parse()?; + let for_token: Token![for] = input.parse()?; + +- let leading_vert: Option<Token![|]> = input.parse()?; +- let mut pat: Pat = input.parse()?; +- if leading_vert.is_some() || input.peek(Token![|]) { +- let mut cases = Punctuated::new(); +- cases.push_value(pat); +- while input.peek(Token![|]) { +- let punct = input.parse()?; +- cases.push_punct(punct); +- let pat: Pat = input.parse()?; +- cases.push_value(pat); +- } +- pat = Pat::Or(PatOr { +- attrs: Vec::new(), +- leading_vert, +- cases, +- }); +- } ++ let pat = pat::parsing::multi_pat_with_leading_vert(input)?; + + let in_token: Token![in] = input.parse()?; +- let expr: Expr = input.call(expr_no_struct)?; ++ let expr: Expr = input.call(Expr::parse_without_eager_brace)?; + + let content; + let brace_token = braced!(content in input); +@@ -2063,7 +2003,7 @@ pub(crate) mod parsing { + let stmts = content.call(Block::parse_within)?; + + Ok(ExprForLoop { +- attrs: inner_attrs, ++ attrs: private::attrs(outer_attrs, inner_attrs), + label, + for_token, + pat, +@@ -2077,6 +2017,7 @@ pub(crate) mod parsing { + #[cfg(feature = "full")] + impl Parse for ExprLoop { + fn parse(input: ParseStream) -> Result<Self> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; + let label: Option<Label> = input.parse()?; + let loop_token: Token![loop] = input.parse()?; + +@@ -2086,7 +2027,7 @@ pub(crate) mod parsing { + let stmts = content.call(Block::parse_within)?; + + Ok(ExprLoop { +- attrs: inner_attrs, ++ attrs: private::attrs(outer_attrs, inner_attrs), + label, + loop_token, + body: Block { brace_token, stmts }, +@@ -2097,8 +2038,9 @@ pub(crate) mod parsing { + #[cfg(feature = "full")] + impl Parse for ExprMatch { + fn parse(input: ParseStream) -> Result<Self> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; + let match_token: Token![match] = input.parse()?; +- let expr = expr_no_struct(input)?; ++ let expr = Expr::parse_without_eager_brace(input)?; + + let content; + let brace_token = braced!(content in input); +@@ -2110,7 +2052,7 @@ pub(crate) mod parsing { + } + + Ok(ExprMatch { +- attrs: inner_attrs, ++ attrs: private::attrs(outer_attrs, inner_attrs), + match_token, + expr: Box::new(expr), + brace_token, +@@ -2305,9 +2247,10 @@ pub(crate) mod parsing { + #[cfg(feature = "full")] + impl Parse for ExprWhile { + fn parse(input: ParseStream) -> Result<Self> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; + let label: Option<Label> = input.parse()?; + let while_token: Token![while] = input.parse()?; +- let cond = expr_no_struct(input)?; ++ let cond = Expr::parse_without_eager_brace(input)?; + + let content; + let brace_token = braced!(content in input); +@@ -2315,7 +2258,7 @@ pub(crate) mod parsing { + let stmts = content.call(Block::parse_within)?; + + Ok(ExprWhile { +- attrs: inner_attrs, ++ attrs: private::attrs(outer_attrs, inner_attrs), + label, + while_token, + cond: Box::new(cond), +@@ -2399,6 +2342,7 @@ pub(crate) mod parsing { + #[cfg(feature = "full")] + impl Parse for FieldValue { + fn parse(input: ParseStream) -> Result<Self> { ++ let attrs = input.call(Attribute::parse_outer)?; + let member: Member = input.parse()?; + let (colon_token, value) = if input.peek(Token![:]) || !member.is_named() { + let colon_token: Token![:] = input.parse()?; +@@ -2416,7 +2360,7 @@ pub(crate) mod parsing { + }; + + Ok(FieldValue { +- attrs: Vec::new(), ++ attrs, + member, + colon_token, + expr: value, +@@ -2433,46 +2377,36 @@ pub(crate) mod parsing { + let content; + let brace_token = braced!(content in input); + let inner_attrs = content.call(Attribute::parse_inner)?; ++ let attrs = private::attrs(outer_attrs, inner_attrs); + + let mut fields = Punctuated::new(); +- loop { +- let attrs = content.call(Attribute::parse_outer)?; +- // TODO: optimize using advance_to +- if content.fork().parse::<Member>().is_err() { +- if attrs.is_empty() { +- break; +- } else { +- return Err(content.error("expected struct field")); +- } ++ while !content.is_empty() { ++ if content.peek(Token![..]) { ++ return Ok(ExprStruct { ++ attrs, ++ brace_token, ++ path, ++ fields, ++ dot2_token: Some(content.parse()?), ++ rest: Some(Box::new(content.parse()?)), ++ }); + } + +- fields.push(FieldValue { +- attrs, +- ..content.parse()? +- }); +- +- if !content.peek(Token![,]) { ++ fields.push(content.parse()?); ++ if content.is_empty() { + break; + } + let punct: Token![,] = content.parse()?; + fields.push_punct(punct); + } + +- let (dot2_token, rest) = if fields.empty_or_trailing() && content.peek(Token![..]) { +- let dot2_token: Token![..] = content.parse()?; +- let rest: Expr = content.parse()?; +- (Some(dot2_token), Some(Box::new(rest))) +- } else { +- (None, None) +- }; +- + Ok(ExprStruct { +- attrs: private::attrs(outer_attrs, inner_attrs), ++ attrs, + brace_token, + path, + fields, +- dot2_token, +- rest, ++ dot2_token: None, ++ rest: None, + }) + } + +@@ -2577,27 +2511,7 @@ pub(crate) mod parsing { + let requires_comma; + Ok(Arm { + attrs: input.call(Attribute::parse_outer)?, +- pat: { +- let leading_vert: Option<Token![|]> = input.parse()?; +- let pat: Pat = input.parse()?; +- if leading_vert.is_some() || input.peek(Token![|]) { +- let mut cases = Punctuated::new(); +- cases.push_value(pat); +- while input.peek(Token![|]) { +- let punct = input.parse()?; +- cases.push_punct(punct); +- let pat: Pat = input.parse()?; +- cases.push_value(pat); +- } +- Pat::Or(PatOr { +- attrs: Vec::new(), +- leading_vert, +- cases, +- }) +- } else { +- pat +- } +- }, ++ pat: pat::parsing::multi_pat_with_leading_vert(input)?, + guard: { + if input.peek(Token![if]) { + let if_token: Token![if] = input.parse()?; +@@ -2641,6 +2555,26 @@ pub(crate) mod parsing { + } + } + ++ fn multi_index(e: &mut Expr, dot_token: &mut Token![.], float: LitFloat) -> Result<bool> { ++ let mut float_repr = float.to_string(); ++ let trailing_dot = float_repr.ends_with('.'); ++ if trailing_dot { ++ float_repr.truncate(float_repr.len() - 1); ++ } ++ for part in float_repr.split('.') { ++ let index = crate::parse_str(part).map_err(|err| Error::new(float.span(), err))?; ++ let base = mem::replace(e, Expr::__Nonexhaustive); ++ *e = Expr::Field(ExprField { ++ attrs: Vec::new(), ++ base: Box::new(base), ++ dot_token: Token![.](dot_token.span), ++ member: Member::Unnamed(index), ++ }); ++ *dot_token = Token![.](float.span()); ++ } ++ Ok(!trailing_dot) ++ } ++ + #[cfg(feature = "full")] + impl Member { + fn is_named(&self) -> bool { +diff --git a/third_party/rust/syn/src/ext.rs b/third_party/rust/syn/src/ext.rs +index d09577a27a..4f9bc145d9 100644 +--- a/third_party/rust/syn/src/ext.rs ++++ b/third_party/rust/syn/src/ext.rs +@@ -1,6 +1,6 @@ + //! Extension traits to provide parsing methods on foreign types. + //! +-//! *This module is available if Syn is built with the `"parsing"` feature.* ++//! *This module is available only if Syn is built with the `"parsing"` feature.* + + use proc_macro2::Ident; + +@@ -16,7 +16,7 @@ use crate::token::CustomToken; + /// This trait is sealed and cannot be implemented for types outside of Syn. It + /// is implemented only for `proc_macro2::Ident`. + /// +-/// *This trait is available if Syn is built with the `"parsing"` feature.* ++/// *This trait is available only if Syn is built with the `"parsing"` feature.* + pub trait IdentExt: Sized + private::Sealed { + /// Parses any identifier including keywords. + /// +@@ -129,7 +129,13 @@ mod private { + + impl Sealed for Ident {} + +- #[derive(Copy, Clone)] + pub struct PeekFn; + pub struct IdentAny; ++ ++ impl Copy for PeekFn {} ++ impl Clone for PeekFn { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } + } +diff --git a/third_party/rust/syn/src/file.rs b/third_party/rust/syn/src/file.rs +index 88c02fe832..c8fab63cd9 100644 +--- a/third_party/rust/syn/src/file.rs ++++ b/third_party/rust/syn/src/file.rs +@@ -3,7 +3,7 @@ use super::*; + ast_struct! { + /// A complete file of Rust source code. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Example + /// +@@ -37,6 +37,8 @@ ast_struct! { + /// file.read_to_string(&mut src).expect("Unable to read file"); + /// + /// let syntax = syn::parse_file(&src).expect("Unable to parse file"); ++ /// ++ /// // Debug impl is available if Syn is built with "extra-traits" feature. + /// println!("{:#?}", syntax); + /// } + /// ``` +diff --git a/third_party/rust/syn/src/gen/clone.rs b/third_party/rust/syn/src/gen/clone.rs +new file mode 100644 +index 0000000000..bea3887013 +--- /dev/null ++++ b/third_party/rust/syn/src/gen/clone.rs +@@ -0,0 +1,2051 @@ ++// This file is @generated by syn-internal-codegen. ++// It is not intended for manual editing. ++ ++#![allow(clippy::clone_on_copy, clippy::expl_impl_clone_on_copy)] ++use crate::*; ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Abi { ++ fn clone(&self) -> Self { ++ Abi { ++ extern_token: self.extern_token.clone(), ++ name: self.name.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for AngleBracketedGenericArguments { ++ fn clone(&self) -> Self { ++ AngleBracketedGenericArguments { ++ colon2_token: self.colon2_token.clone(), ++ lt_token: self.lt_token.clone(), ++ args: self.args.clone(), ++ gt_token: self.gt_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Arm { ++ fn clone(&self) -> Self { ++ Arm { ++ attrs: self.attrs.clone(), ++ pat: self.pat.clone(), ++ guard: self.guard.clone(), ++ fat_arrow_token: self.fat_arrow_token.clone(), ++ body: self.body.clone(), ++ comma: self.comma.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Copy for AttrStyle {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for AttrStyle { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Attribute { ++ fn clone(&self) -> Self { ++ Attribute { ++ pound_token: self.pound_token.clone(), ++ style: self.style.clone(), ++ bracket_token: self.bracket_token.clone(), ++ path: self.path.clone(), ++ tokens: self.tokens.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for BareFnArg { ++ fn clone(&self) -> Self { ++ BareFnArg { ++ attrs: self.attrs.clone(), ++ name: self.name.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Copy for BinOp {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for BinOp { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Binding { ++ fn clone(&self) -> Self { ++ Binding { ++ ident: self.ident.clone(), ++ eq_token: self.eq_token.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Block { ++ fn clone(&self) -> Self { ++ Block { ++ brace_token: self.brace_token.clone(), ++ stmts: self.stmts.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for BoundLifetimes { ++ fn clone(&self) -> Self { ++ BoundLifetimes { ++ for_token: self.for_token.clone(), ++ lt_token: self.lt_token.clone(), ++ lifetimes: self.lifetimes.clone(), ++ gt_token: self.gt_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ConstParam { ++ fn clone(&self) -> Self { ++ ConstParam { ++ attrs: self.attrs.clone(), ++ const_token: self.const_token.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ eq_token: self.eq_token.clone(), ++ default: self.default.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Constraint { ++ fn clone(&self) -> Self { ++ Constraint { ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Clone for Data { ++ fn clone(&self) -> Self { ++ match self { ++ Data::Struct(v0) => Data::Struct(v0.clone()), ++ Data::Enum(v0) => Data::Enum(v0.clone()), ++ Data::Union(v0) => Data::Union(v0.clone()), ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Clone for DataEnum { ++ fn clone(&self) -> Self { ++ DataEnum { ++ enum_token: self.enum_token.clone(), ++ brace_token: self.brace_token.clone(), ++ variants: self.variants.clone(), ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Clone for DataStruct { ++ fn clone(&self) -> Self { ++ DataStruct { ++ struct_token: self.struct_token.clone(), ++ fields: self.fields.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Clone for DataUnion { ++ fn clone(&self) -> Self { ++ DataUnion { ++ union_token: self.union_token.clone(), ++ fields: self.fields.clone(), ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Clone for DeriveInput { ++ fn clone(&self) -> Self { ++ DeriveInput { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ data: self.data.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Expr { ++ fn clone(&self) -> Self { ++ match self { ++ #[cfg(feature = "full")] ++ Expr::Array(v0) => Expr::Array(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Assign(v0) => Expr::Assign(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::AssignOp(v0) => Expr::AssignOp(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Async(v0) => Expr::Async(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Await(v0) => Expr::Await(v0.clone()), ++ Expr::Binary(v0) => Expr::Binary(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Block(v0) => Expr::Block(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Box(v0) => Expr::Box(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Break(v0) => Expr::Break(v0.clone()), ++ Expr::Call(v0) => Expr::Call(v0.clone()), ++ Expr::Cast(v0) => Expr::Cast(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Closure(v0) => Expr::Closure(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Continue(v0) => Expr::Continue(v0.clone()), ++ Expr::Field(v0) => Expr::Field(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::ForLoop(v0) => Expr::ForLoop(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Group(v0) => Expr::Group(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::If(v0) => Expr::If(v0.clone()), ++ Expr::Index(v0) => Expr::Index(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Let(v0) => Expr::Let(v0.clone()), ++ Expr::Lit(v0) => Expr::Lit(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Loop(v0) => Expr::Loop(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Macro(v0) => Expr::Macro(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Match(v0) => Expr::Match(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::MethodCall(v0) => Expr::MethodCall(v0.clone()), ++ Expr::Paren(v0) => Expr::Paren(v0.clone()), ++ Expr::Path(v0) => Expr::Path(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Range(v0) => Expr::Range(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Reference(v0) => Expr::Reference(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Repeat(v0) => Expr::Repeat(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Return(v0) => Expr::Return(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Struct(v0) => Expr::Struct(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Try(v0) => Expr::Try(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::TryBlock(v0) => Expr::TryBlock(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Tuple(v0) => Expr::Tuple(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Type(v0) => Expr::Type(v0.clone()), ++ Expr::Unary(v0) => Expr::Unary(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Unsafe(v0) => Expr::Unsafe(v0.clone()), ++ Expr::Verbatim(v0) => Expr::Verbatim(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::While(v0) => Expr::While(v0.clone()), ++ #[cfg(feature = "full")] ++ Expr::Yield(v0) => Expr::Yield(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprArray { ++ fn clone(&self) -> Self { ++ ExprArray { ++ attrs: self.attrs.clone(), ++ bracket_token: self.bracket_token.clone(), ++ elems: self.elems.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprAssign { ++ fn clone(&self) -> Self { ++ ExprAssign { ++ attrs: self.attrs.clone(), ++ left: self.left.clone(), ++ eq_token: self.eq_token.clone(), ++ right: self.right.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprAssignOp { ++ fn clone(&self) -> Self { ++ ExprAssignOp { ++ attrs: self.attrs.clone(), ++ left: self.left.clone(), ++ op: self.op.clone(), ++ right: self.right.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprAsync { ++ fn clone(&self) -> Self { ++ ExprAsync { ++ attrs: self.attrs.clone(), ++ async_token: self.async_token.clone(), ++ capture: self.capture.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprAwait { ++ fn clone(&self) -> Self { ++ ExprAwait { ++ attrs: self.attrs.clone(), ++ base: self.base.clone(), ++ dot_token: self.dot_token.clone(), ++ await_token: self.await_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprBinary { ++ fn clone(&self) -> Self { ++ ExprBinary { ++ attrs: self.attrs.clone(), ++ left: self.left.clone(), ++ op: self.op.clone(), ++ right: self.right.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprBlock { ++ fn clone(&self) -> Self { ++ ExprBlock { ++ attrs: self.attrs.clone(), ++ label: self.label.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprBox { ++ fn clone(&self) -> Self { ++ ExprBox { ++ attrs: self.attrs.clone(), ++ box_token: self.box_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprBreak { ++ fn clone(&self) -> Self { ++ ExprBreak { ++ attrs: self.attrs.clone(), ++ break_token: self.break_token.clone(), ++ label: self.label.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprCall { ++ fn clone(&self) -> Self { ++ ExprCall { ++ attrs: self.attrs.clone(), ++ func: self.func.clone(), ++ paren_token: self.paren_token.clone(), ++ args: self.args.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprCast { ++ fn clone(&self) -> Self { ++ ExprCast { ++ attrs: self.attrs.clone(), ++ expr: self.expr.clone(), ++ as_token: self.as_token.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprClosure { ++ fn clone(&self) -> Self { ++ ExprClosure { ++ attrs: self.attrs.clone(), ++ asyncness: self.asyncness.clone(), ++ movability: self.movability.clone(), ++ capture: self.capture.clone(), ++ or1_token: self.or1_token.clone(), ++ inputs: self.inputs.clone(), ++ or2_token: self.or2_token.clone(), ++ output: self.output.clone(), ++ body: self.body.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprContinue { ++ fn clone(&self) -> Self { ++ ExprContinue { ++ attrs: self.attrs.clone(), ++ continue_token: self.continue_token.clone(), ++ label: self.label.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprField { ++ fn clone(&self) -> Self { ++ ExprField { ++ attrs: self.attrs.clone(), ++ base: self.base.clone(), ++ dot_token: self.dot_token.clone(), ++ member: self.member.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprForLoop { ++ fn clone(&self) -> Self { ++ ExprForLoop { ++ attrs: self.attrs.clone(), ++ label: self.label.clone(), ++ for_token: self.for_token.clone(), ++ pat: self.pat.clone(), ++ in_token: self.in_token.clone(), ++ expr: self.expr.clone(), ++ body: self.body.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprGroup { ++ fn clone(&self) -> Self { ++ ExprGroup { ++ attrs: self.attrs.clone(), ++ group_token: self.group_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprIf { ++ fn clone(&self) -> Self { ++ ExprIf { ++ attrs: self.attrs.clone(), ++ if_token: self.if_token.clone(), ++ cond: self.cond.clone(), ++ then_branch: self.then_branch.clone(), ++ else_branch: self.else_branch.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprIndex { ++ fn clone(&self) -> Self { ++ ExprIndex { ++ attrs: self.attrs.clone(), ++ expr: self.expr.clone(), ++ bracket_token: self.bracket_token.clone(), ++ index: self.index.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprLet { ++ fn clone(&self) -> Self { ++ ExprLet { ++ attrs: self.attrs.clone(), ++ let_token: self.let_token.clone(), ++ pat: self.pat.clone(), ++ eq_token: self.eq_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprLit { ++ fn clone(&self) -> Self { ++ ExprLit { ++ attrs: self.attrs.clone(), ++ lit: self.lit.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprLoop { ++ fn clone(&self) -> Self { ++ ExprLoop { ++ attrs: self.attrs.clone(), ++ label: self.label.clone(), ++ loop_token: self.loop_token.clone(), ++ body: self.body.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprMacro { ++ fn clone(&self) -> Self { ++ ExprMacro { ++ attrs: self.attrs.clone(), ++ mac: self.mac.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprMatch { ++ fn clone(&self) -> Self { ++ ExprMatch { ++ attrs: self.attrs.clone(), ++ match_token: self.match_token.clone(), ++ expr: self.expr.clone(), ++ brace_token: self.brace_token.clone(), ++ arms: self.arms.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprMethodCall { ++ fn clone(&self) -> Self { ++ ExprMethodCall { ++ attrs: self.attrs.clone(), ++ receiver: self.receiver.clone(), ++ dot_token: self.dot_token.clone(), ++ method: self.method.clone(), ++ turbofish: self.turbofish.clone(), ++ paren_token: self.paren_token.clone(), ++ args: self.args.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprParen { ++ fn clone(&self) -> Self { ++ ExprParen { ++ attrs: self.attrs.clone(), ++ paren_token: self.paren_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprPath { ++ fn clone(&self) -> Self { ++ ExprPath { ++ attrs: self.attrs.clone(), ++ qself: self.qself.clone(), ++ path: self.path.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprRange { ++ fn clone(&self) -> Self { ++ ExprRange { ++ attrs: self.attrs.clone(), ++ from: self.from.clone(), ++ limits: self.limits.clone(), ++ to: self.to.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprReference { ++ fn clone(&self) -> Self { ++ ExprReference { ++ attrs: self.attrs.clone(), ++ and_token: self.and_token.clone(), ++ raw: self.raw.clone(), ++ mutability: self.mutability.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprRepeat { ++ fn clone(&self) -> Self { ++ ExprRepeat { ++ attrs: self.attrs.clone(), ++ bracket_token: self.bracket_token.clone(), ++ expr: self.expr.clone(), ++ semi_token: self.semi_token.clone(), ++ len: self.len.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprReturn { ++ fn clone(&self) -> Self { ++ ExprReturn { ++ attrs: self.attrs.clone(), ++ return_token: self.return_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprStruct { ++ fn clone(&self) -> Self { ++ ExprStruct { ++ attrs: self.attrs.clone(), ++ path: self.path.clone(), ++ brace_token: self.brace_token.clone(), ++ fields: self.fields.clone(), ++ dot2_token: self.dot2_token.clone(), ++ rest: self.rest.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprTry { ++ fn clone(&self) -> Self { ++ ExprTry { ++ attrs: self.attrs.clone(), ++ expr: self.expr.clone(), ++ question_token: self.question_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprTryBlock { ++ fn clone(&self) -> Self { ++ ExprTryBlock { ++ attrs: self.attrs.clone(), ++ try_token: self.try_token.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprTuple { ++ fn clone(&self) -> Self { ++ ExprTuple { ++ attrs: self.attrs.clone(), ++ paren_token: self.paren_token.clone(), ++ elems: self.elems.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprType { ++ fn clone(&self) -> Self { ++ ExprType { ++ attrs: self.attrs.clone(), ++ expr: self.expr.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ExprUnary { ++ fn clone(&self) -> Self { ++ ExprUnary { ++ attrs: self.attrs.clone(), ++ op: self.op.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprUnsafe { ++ fn clone(&self) -> Self { ++ ExprUnsafe { ++ attrs: self.attrs.clone(), ++ unsafe_token: self.unsafe_token.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprWhile { ++ fn clone(&self) -> Self { ++ ExprWhile { ++ attrs: self.attrs.clone(), ++ label: self.label.clone(), ++ while_token: self.while_token.clone(), ++ cond: self.cond.clone(), ++ body: self.body.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ExprYield { ++ fn clone(&self) -> Self { ++ ExprYield { ++ attrs: self.attrs.clone(), ++ yield_token: self.yield_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Field { ++ fn clone(&self) -> Self { ++ Field { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for FieldPat { ++ fn clone(&self) -> Self { ++ FieldPat { ++ attrs: self.attrs.clone(), ++ member: self.member.clone(), ++ colon_token: self.colon_token.clone(), ++ pat: self.pat.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for FieldValue { ++ fn clone(&self) -> Self { ++ FieldValue { ++ attrs: self.attrs.clone(), ++ member: self.member.clone(), ++ colon_token: self.colon_token.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Fields { ++ fn clone(&self) -> Self { ++ match self { ++ Fields::Named(v0) => Fields::Named(v0.clone()), ++ Fields::Unnamed(v0) => Fields::Unnamed(v0.clone()), ++ Fields::Unit => Fields::Unit, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for FieldsNamed { ++ fn clone(&self) -> Self { ++ FieldsNamed { ++ brace_token: self.brace_token.clone(), ++ named: self.named.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for FieldsUnnamed { ++ fn clone(&self) -> Self { ++ FieldsUnnamed { ++ paren_token: self.paren_token.clone(), ++ unnamed: self.unnamed.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for File { ++ fn clone(&self) -> Self { ++ File { ++ shebang: self.shebang.clone(), ++ attrs: self.attrs.clone(), ++ items: self.items.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for FnArg { ++ fn clone(&self) -> Self { ++ match self { ++ FnArg::Receiver(v0) => FnArg::Receiver(v0.clone()), ++ FnArg::Typed(v0) => FnArg::Typed(v0.clone()), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ForeignItem { ++ fn clone(&self) -> Self { ++ match self { ++ ForeignItem::Fn(v0) => ForeignItem::Fn(v0.clone()), ++ ForeignItem::Static(v0) => ForeignItem::Static(v0.clone()), ++ ForeignItem::Type(v0) => ForeignItem::Type(v0.clone()), ++ ForeignItem::Macro(v0) => ForeignItem::Macro(v0.clone()), ++ ForeignItem::Verbatim(v0) => ForeignItem::Verbatim(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ForeignItemFn { ++ fn clone(&self) -> Self { ++ ForeignItemFn { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ sig: self.sig.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ForeignItemMacro { ++ fn clone(&self) -> Self { ++ ForeignItemMacro { ++ attrs: self.attrs.clone(), ++ mac: self.mac.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ForeignItemStatic { ++ fn clone(&self) -> Self { ++ ForeignItemStatic { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ static_token: self.static_token.clone(), ++ mutability: self.mutability.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ForeignItemType { ++ fn clone(&self) -> Self { ++ ForeignItemType { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ type_token: self.type_token.clone(), ++ ident: self.ident.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for GenericArgument { ++ fn clone(&self) -> Self { ++ match self { ++ GenericArgument::Lifetime(v0) => GenericArgument::Lifetime(v0.clone()), ++ GenericArgument::Type(v0) => GenericArgument::Type(v0.clone()), ++ GenericArgument::Binding(v0) => GenericArgument::Binding(v0.clone()), ++ GenericArgument::Constraint(v0) => GenericArgument::Constraint(v0.clone()), ++ GenericArgument::Const(v0) => GenericArgument::Const(v0.clone()), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for GenericMethodArgument { ++ fn clone(&self) -> Self { ++ match self { ++ GenericMethodArgument::Type(v0) => GenericMethodArgument::Type(v0.clone()), ++ GenericMethodArgument::Const(v0) => GenericMethodArgument::Const(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for GenericParam { ++ fn clone(&self) -> Self { ++ match self { ++ GenericParam::Type(v0) => GenericParam::Type(v0.clone()), ++ GenericParam::Lifetime(v0) => GenericParam::Lifetime(v0.clone()), ++ GenericParam::Const(v0) => GenericParam::Const(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Generics { ++ fn clone(&self) -> Self { ++ Generics { ++ lt_token: self.lt_token.clone(), ++ params: self.params.clone(), ++ gt_token: self.gt_token.clone(), ++ where_clause: self.where_clause.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ImplItem { ++ fn clone(&self) -> Self { ++ match self { ++ ImplItem::Const(v0) => ImplItem::Const(v0.clone()), ++ ImplItem::Method(v0) => ImplItem::Method(v0.clone()), ++ ImplItem::Type(v0) => ImplItem::Type(v0.clone()), ++ ImplItem::Macro(v0) => ImplItem::Macro(v0.clone()), ++ ImplItem::Verbatim(v0) => ImplItem::Verbatim(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ImplItemConst { ++ fn clone(&self) -> Self { ++ ImplItemConst { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ defaultness: self.defaultness.clone(), ++ const_token: self.const_token.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ eq_token: self.eq_token.clone(), ++ expr: self.expr.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ImplItemMacro { ++ fn clone(&self) -> Self { ++ ImplItemMacro { ++ attrs: self.attrs.clone(), ++ mac: self.mac.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ImplItemMethod { ++ fn clone(&self) -> Self { ++ ImplItemMethod { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ defaultness: self.defaultness.clone(), ++ sig: self.sig.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ImplItemType { ++ fn clone(&self) -> Self { ++ ImplItemType { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ defaultness: self.defaultness.clone(), ++ type_token: self.type_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ eq_token: self.eq_token.clone(), ++ ty: self.ty.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Index { ++ fn clone(&self) -> Self { ++ Index { ++ index: self.index.clone(), ++ span: self.span.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Item { ++ fn clone(&self) -> Self { ++ match self { ++ Item::Const(v0) => Item::Const(v0.clone()), ++ Item::Enum(v0) => Item::Enum(v0.clone()), ++ Item::ExternCrate(v0) => Item::ExternCrate(v0.clone()), ++ Item::Fn(v0) => Item::Fn(v0.clone()), ++ Item::ForeignMod(v0) => Item::ForeignMod(v0.clone()), ++ Item::Impl(v0) => Item::Impl(v0.clone()), ++ Item::Macro(v0) => Item::Macro(v0.clone()), ++ Item::Macro2(v0) => Item::Macro2(v0.clone()), ++ Item::Mod(v0) => Item::Mod(v0.clone()), ++ Item::Static(v0) => Item::Static(v0.clone()), ++ Item::Struct(v0) => Item::Struct(v0.clone()), ++ Item::Trait(v0) => Item::Trait(v0.clone()), ++ Item::TraitAlias(v0) => Item::TraitAlias(v0.clone()), ++ Item::Type(v0) => Item::Type(v0.clone()), ++ Item::Union(v0) => Item::Union(v0.clone()), ++ Item::Use(v0) => Item::Use(v0.clone()), ++ Item::Verbatim(v0) => Item::Verbatim(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemConst { ++ fn clone(&self) -> Self { ++ ItemConst { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ const_token: self.const_token.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ eq_token: self.eq_token.clone(), ++ expr: self.expr.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemEnum { ++ fn clone(&self) -> Self { ++ ItemEnum { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ enum_token: self.enum_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ brace_token: self.brace_token.clone(), ++ variants: self.variants.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemExternCrate { ++ fn clone(&self) -> Self { ++ ItemExternCrate { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ extern_token: self.extern_token.clone(), ++ crate_token: self.crate_token.clone(), ++ ident: self.ident.clone(), ++ rename: self.rename.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemFn { ++ fn clone(&self) -> Self { ++ ItemFn { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ sig: self.sig.clone(), ++ block: self.block.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemForeignMod { ++ fn clone(&self) -> Self { ++ ItemForeignMod { ++ attrs: self.attrs.clone(), ++ abi: self.abi.clone(), ++ brace_token: self.brace_token.clone(), ++ items: self.items.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemImpl { ++ fn clone(&self) -> Self { ++ ItemImpl { ++ attrs: self.attrs.clone(), ++ defaultness: self.defaultness.clone(), ++ unsafety: self.unsafety.clone(), ++ impl_token: self.impl_token.clone(), ++ generics: self.generics.clone(), ++ trait_: self.trait_.clone(), ++ self_ty: self.self_ty.clone(), ++ brace_token: self.brace_token.clone(), ++ items: self.items.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemMacro { ++ fn clone(&self) -> Self { ++ ItemMacro { ++ attrs: self.attrs.clone(), ++ ident: self.ident.clone(), ++ mac: self.mac.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemMacro2 { ++ fn clone(&self) -> Self { ++ ItemMacro2 { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ macro_token: self.macro_token.clone(), ++ ident: self.ident.clone(), ++ rules: self.rules.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemMod { ++ fn clone(&self) -> Self { ++ ItemMod { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ mod_token: self.mod_token.clone(), ++ ident: self.ident.clone(), ++ content: self.content.clone(), ++ semi: self.semi.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemStatic { ++ fn clone(&self) -> Self { ++ ItemStatic { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ static_token: self.static_token.clone(), ++ mutability: self.mutability.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ eq_token: self.eq_token.clone(), ++ expr: self.expr.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemStruct { ++ fn clone(&self) -> Self { ++ ItemStruct { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ struct_token: self.struct_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ fields: self.fields.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemTrait { ++ fn clone(&self) -> Self { ++ ItemTrait { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ unsafety: self.unsafety.clone(), ++ auto_token: self.auto_token.clone(), ++ trait_token: self.trait_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ colon_token: self.colon_token.clone(), ++ supertraits: self.supertraits.clone(), ++ brace_token: self.brace_token.clone(), ++ items: self.items.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemTraitAlias { ++ fn clone(&self) -> Self { ++ ItemTraitAlias { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ trait_token: self.trait_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ eq_token: self.eq_token.clone(), ++ bounds: self.bounds.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemType { ++ fn clone(&self) -> Self { ++ ItemType { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ type_token: self.type_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ eq_token: self.eq_token.clone(), ++ ty: self.ty.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemUnion { ++ fn clone(&self) -> Self { ++ ItemUnion { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ union_token: self.union_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ fields: self.fields.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for ItemUse { ++ fn clone(&self) -> Self { ++ ItemUse { ++ attrs: self.attrs.clone(), ++ vis: self.vis.clone(), ++ use_token: self.use_token.clone(), ++ leading_colon: self.leading_colon.clone(), ++ tree: self.tree.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Label { ++ fn clone(&self) -> Self { ++ Label { ++ name: self.name.clone(), ++ colon_token: self.colon_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for LifetimeDef { ++ fn clone(&self) -> Self { ++ LifetimeDef { ++ attrs: self.attrs.clone(), ++ lifetime: self.lifetime.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++impl Clone for Lit { ++ fn clone(&self) -> Self { ++ match self { ++ Lit::Str(v0) => Lit::Str(v0.clone()), ++ Lit::ByteStr(v0) => Lit::ByteStr(v0.clone()), ++ Lit::Byte(v0) => Lit::Byte(v0.clone()), ++ Lit::Char(v0) => Lit::Char(v0.clone()), ++ Lit::Int(v0) => Lit::Int(v0.clone()), ++ Lit::Float(v0) => Lit::Float(v0.clone()), ++ Lit::Bool(v0) => Lit::Bool(v0.clone()), ++ Lit::Verbatim(v0) => Lit::Verbatim(v0.clone()), ++ } ++ } ++} ++impl Clone for LitBool { ++ fn clone(&self) -> Self { ++ LitBool { ++ value: self.value.clone(), ++ span: self.span.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Local { ++ fn clone(&self) -> Self { ++ Local { ++ attrs: self.attrs.clone(), ++ let_token: self.let_token.clone(), ++ pat: self.pat.clone(), ++ init: self.init.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Macro { ++ fn clone(&self) -> Self { ++ Macro { ++ path: self.path.clone(), ++ bang_token: self.bang_token.clone(), ++ delimiter: self.delimiter.clone(), ++ tokens: self.tokens.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for MacroDelimiter { ++ fn clone(&self) -> Self { ++ match self { ++ MacroDelimiter::Paren(v0) => MacroDelimiter::Paren(v0.clone()), ++ MacroDelimiter::Brace(v0) => MacroDelimiter::Brace(v0.clone()), ++ MacroDelimiter::Bracket(v0) => MacroDelimiter::Bracket(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Member { ++ fn clone(&self) -> Self { ++ match self { ++ Member::Named(v0) => Member::Named(v0.clone()), ++ Member::Unnamed(v0) => Member::Unnamed(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Meta { ++ fn clone(&self) -> Self { ++ match self { ++ Meta::Path(v0) => Meta::Path(v0.clone()), ++ Meta::List(v0) => Meta::List(v0.clone()), ++ Meta::NameValue(v0) => Meta::NameValue(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for MetaList { ++ fn clone(&self) -> Self { ++ MetaList { ++ path: self.path.clone(), ++ paren_token: self.paren_token.clone(), ++ nested: self.nested.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for MetaNameValue { ++ fn clone(&self) -> Self { ++ MetaNameValue { ++ path: self.path.clone(), ++ eq_token: self.eq_token.clone(), ++ lit: self.lit.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for MethodTurbofish { ++ fn clone(&self) -> Self { ++ MethodTurbofish { ++ colon2_token: self.colon2_token.clone(), ++ lt_token: self.lt_token.clone(), ++ args: self.args.clone(), ++ gt_token: self.gt_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for NestedMeta { ++ fn clone(&self) -> Self { ++ match self { ++ NestedMeta::Meta(v0) => NestedMeta::Meta(v0.clone()), ++ NestedMeta::Lit(v0) => NestedMeta::Lit(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ParenthesizedGenericArguments { ++ fn clone(&self) -> Self { ++ ParenthesizedGenericArguments { ++ paren_token: self.paren_token.clone(), ++ inputs: self.inputs.clone(), ++ output: self.output.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Pat { ++ fn clone(&self) -> Self { ++ match self { ++ Pat::Box(v0) => Pat::Box(v0.clone()), ++ Pat::Ident(v0) => Pat::Ident(v0.clone()), ++ Pat::Lit(v0) => Pat::Lit(v0.clone()), ++ Pat::Macro(v0) => Pat::Macro(v0.clone()), ++ Pat::Or(v0) => Pat::Or(v0.clone()), ++ Pat::Path(v0) => Pat::Path(v0.clone()), ++ Pat::Range(v0) => Pat::Range(v0.clone()), ++ Pat::Reference(v0) => Pat::Reference(v0.clone()), ++ Pat::Rest(v0) => Pat::Rest(v0.clone()), ++ Pat::Slice(v0) => Pat::Slice(v0.clone()), ++ Pat::Struct(v0) => Pat::Struct(v0.clone()), ++ Pat::Tuple(v0) => Pat::Tuple(v0.clone()), ++ Pat::TupleStruct(v0) => Pat::TupleStruct(v0.clone()), ++ Pat::Type(v0) => Pat::Type(v0.clone()), ++ Pat::Verbatim(v0) => Pat::Verbatim(v0.clone()), ++ Pat::Wild(v0) => Pat::Wild(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatBox { ++ fn clone(&self) -> Self { ++ PatBox { ++ attrs: self.attrs.clone(), ++ box_token: self.box_token.clone(), ++ pat: self.pat.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatIdent { ++ fn clone(&self) -> Self { ++ PatIdent { ++ attrs: self.attrs.clone(), ++ by_ref: self.by_ref.clone(), ++ mutability: self.mutability.clone(), ++ ident: self.ident.clone(), ++ subpat: self.subpat.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatLit { ++ fn clone(&self) -> Self { ++ PatLit { ++ attrs: self.attrs.clone(), ++ expr: self.expr.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatMacro { ++ fn clone(&self) -> Self { ++ PatMacro { ++ attrs: self.attrs.clone(), ++ mac: self.mac.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatOr { ++ fn clone(&self) -> Self { ++ PatOr { ++ attrs: self.attrs.clone(), ++ leading_vert: self.leading_vert.clone(), ++ cases: self.cases.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatPath { ++ fn clone(&self) -> Self { ++ PatPath { ++ attrs: self.attrs.clone(), ++ qself: self.qself.clone(), ++ path: self.path.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatRange { ++ fn clone(&self) -> Self { ++ PatRange { ++ attrs: self.attrs.clone(), ++ lo: self.lo.clone(), ++ limits: self.limits.clone(), ++ hi: self.hi.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatReference { ++ fn clone(&self) -> Self { ++ PatReference { ++ attrs: self.attrs.clone(), ++ and_token: self.and_token.clone(), ++ mutability: self.mutability.clone(), ++ pat: self.pat.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatRest { ++ fn clone(&self) -> Self { ++ PatRest { ++ attrs: self.attrs.clone(), ++ dot2_token: self.dot2_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatSlice { ++ fn clone(&self) -> Self { ++ PatSlice { ++ attrs: self.attrs.clone(), ++ bracket_token: self.bracket_token.clone(), ++ elems: self.elems.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatStruct { ++ fn clone(&self) -> Self { ++ PatStruct { ++ attrs: self.attrs.clone(), ++ path: self.path.clone(), ++ brace_token: self.brace_token.clone(), ++ fields: self.fields.clone(), ++ dot2_token: self.dot2_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatTuple { ++ fn clone(&self) -> Self { ++ PatTuple { ++ attrs: self.attrs.clone(), ++ paren_token: self.paren_token.clone(), ++ elems: self.elems.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatTupleStruct { ++ fn clone(&self) -> Self { ++ PatTupleStruct { ++ attrs: self.attrs.clone(), ++ path: self.path.clone(), ++ pat: self.pat.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatType { ++ fn clone(&self) -> Self { ++ PatType { ++ attrs: self.attrs.clone(), ++ pat: self.pat.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for PatWild { ++ fn clone(&self) -> Self { ++ PatWild { ++ attrs: self.attrs.clone(), ++ underscore_token: self.underscore_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Path { ++ fn clone(&self) -> Self { ++ Path { ++ leading_colon: self.leading_colon.clone(), ++ segments: self.segments.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for PathArguments { ++ fn clone(&self) -> Self { ++ match self { ++ PathArguments::None => PathArguments::None, ++ PathArguments::AngleBracketed(v0) => PathArguments::AngleBracketed(v0.clone()), ++ PathArguments::Parenthesized(v0) => PathArguments::Parenthesized(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for PathSegment { ++ fn clone(&self) -> Self { ++ PathSegment { ++ ident: self.ident.clone(), ++ arguments: self.arguments.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for PredicateEq { ++ fn clone(&self) -> Self { ++ PredicateEq { ++ lhs_ty: self.lhs_ty.clone(), ++ eq_token: self.eq_token.clone(), ++ rhs_ty: self.rhs_ty.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for PredicateLifetime { ++ fn clone(&self) -> Self { ++ PredicateLifetime { ++ lifetime: self.lifetime.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for PredicateType { ++ fn clone(&self) -> Self { ++ PredicateType { ++ lifetimes: self.lifetimes.clone(), ++ bounded_ty: self.bounded_ty.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for QSelf { ++ fn clone(&self) -> Self { ++ QSelf { ++ lt_token: self.lt_token.clone(), ++ ty: self.ty.clone(), ++ position: self.position.clone(), ++ as_token: self.as_token.clone(), ++ gt_token: self.gt_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Copy for RangeLimits {} ++#[cfg(feature = "full")] ++impl Clone for RangeLimits { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Receiver { ++ fn clone(&self) -> Self { ++ Receiver { ++ attrs: self.attrs.clone(), ++ reference: self.reference.clone(), ++ mutability: self.mutability.clone(), ++ self_token: self.self_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for ReturnType { ++ fn clone(&self) -> Self { ++ match self { ++ ReturnType::Default => ReturnType::Default, ++ ReturnType::Type(v0, v1) => ReturnType::Type(v0.clone(), v1.clone()), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Signature { ++ fn clone(&self) -> Self { ++ Signature { ++ constness: self.constness.clone(), ++ asyncness: self.asyncness.clone(), ++ unsafety: self.unsafety.clone(), ++ abi: self.abi.clone(), ++ fn_token: self.fn_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ paren_token: self.paren_token.clone(), ++ inputs: self.inputs.clone(), ++ variadic: self.variadic.clone(), ++ output: self.output.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for Stmt { ++ fn clone(&self) -> Self { ++ match self { ++ Stmt::Local(v0) => Stmt::Local(v0.clone()), ++ Stmt::Item(v0) => Stmt::Item(v0.clone()), ++ Stmt::Expr(v0) => Stmt::Expr(v0.clone()), ++ Stmt::Semi(v0, v1) => Stmt::Semi(v0.clone(), v1.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TraitBound { ++ fn clone(&self) -> Self { ++ TraitBound { ++ paren_token: self.paren_token.clone(), ++ modifier: self.modifier.clone(), ++ lifetimes: self.lifetimes.clone(), ++ path: self.path.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Copy for TraitBoundModifier {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TraitBoundModifier { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for TraitItem { ++ fn clone(&self) -> Self { ++ match self { ++ TraitItem::Const(v0) => TraitItem::Const(v0.clone()), ++ TraitItem::Method(v0) => TraitItem::Method(v0.clone()), ++ TraitItem::Type(v0) => TraitItem::Type(v0.clone()), ++ TraitItem::Macro(v0) => TraitItem::Macro(v0.clone()), ++ TraitItem::Verbatim(v0) => TraitItem::Verbatim(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for TraitItemConst { ++ fn clone(&self) -> Self { ++ TraitItemConst { ++ attrs: self.attrs.clone(), ++ const_token: self.const_token.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ ty: self.ty.clone(), ++ default: self.default.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for TraitItemMacro { ++ fn clone(&self) -> Self { ++ TraitItemMacro { ++ attrs: self.attrs.clone(), ++ mac: self.mac.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for TraitItemMethod { ++ fn clone(&self) -> Self { ++ TraitItemMethod { ++ attrs: self.attrs.clone(), ++ sig: self.sig.clone(), ++ default: self.default.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for TraitItemType { ++ fn clone(&self) -> Self { ++ TraitItemType { ++ attrs: self.attrs.clone(), ++ type_token: self.type_token.clone(), ++ ident: self.ident.clone(), ++ generics: self.generics.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ default: self.default.clone(), ++ semi_token: self.semi_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Type { ++ fn clone(&self) -> Self { ++ match self { ++ Type::Array(v0) => Type::Array(v0.clone()), ++ Type::BareFn(v0) => Type::BareFn(v0.clone()), ++ Type::Group(v0) => Type::Group(v0.clone()), ++ Type::ImplTrait(v0) => Type::ImplTrait(v0.clone()), ++ Type::Infer(v0) => Type::Infer(v0.clone()), ++ Type::Macro(v0) => Type::Macro(v0.clone()), ++ Type::Never(v0) => Type::Never(v0.clone()), ++ Type::Paren(v0) => Type::Paren(v0.clone()), ++ Type::Path(v0) => Type::Path(v0.clone()), ++ Type::Ptr(v0) => Type::Ptr(v0.clone()), ++ Type::Reference(v0) => Type::Reference(v0.clone()), ++ Type::Slice(v0) => Type::Slice(v0.clone()), ++ Type::TraitObject(v0) => Type::TraitObject(v0.clone()), ++ Type::Tuple(v0) => Type::Tuple(v0.clone()), ++ Type::Verbatim(v0) => Type::Verbatim(v0.clone()), ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeArray { ++ fn clone(&self) -> Self { ++ TypeArray { ++ bracket_token: self.bracket_token.clone(), ++ elem: self.elem.clone(), ++ semi_token: self.semi_token.clone(), ++ len: self.len.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeBareFn { ++ fn clone(&self) -> Self { ++ TypeBareFn { ++ lifetimes: self.lifetimes.clone(), ++ unsafety: self.unsafety.clone(), ++ abi: self.abi.clone(), ++ fn_token: self.fn_token.clone(), ++ paren_token: self.paren_token.clone(), ++ inputs: self.inputs.clone(), ++ variadic: self.variadic.clone(), ++ output: self.output.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeGroup { ++ fn clone(&self) -> Self { ++ TypeGroup { ++ group_token: self.group_token.clone(), ++ elem: self.elem.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeImplTrait { ++ fn clone(&self) -> Self { ++ TypeImplTrait { ++ impl_token: self.impl_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeInfer { ++ fn clone(&self) -> Self { ++ TypeInfer { ++ underscore_token: self.underscore_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeMacro { ++ fn clone(&self) -> Self { ++ TypeMacro { ++ mac: self.mac.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeNever { ++ fn clone(&self) -> Self { ++ TypeNever { ++ bang_token: self.bang_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeParam { ++ fn clone(&self) -> Self { ++ TypeParam { ++ attrs: self.attrs.clone(), ++ ident: self.ident.clone(), ++ colon_token: self.colon_token.clone(), ++ bounds: self.bounds.clone(), ++ eq_token: self.eq_token.clone(), ++ default: self.default.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeParamBound { ++ fn clone(&self) -> Self { ++ match self { ++ TypeParamBound::Trait(v0) => TypeParamBound::Trait(v0.clone()), ++ TypeParamBound::Lifetime(v0) => TypeParamBound::Lifetime(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeParen { ++ fn clone(&self) -> Self { ++ TypeParen { ++ paren_token: self.paren_token.clone(), ++ elem: self.elem.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypePath { ++ fn clone(&self) -> Self { ++ TypePath { ++ qself: self.qself.clone(), ++ path: self.path.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypePtr { ++ fn clone(&self) -> Self { ++ TypePtr { ++ star_token: self.star_token.clone(), ++ const_token: self.const_token.clone(), ++ mutability: self.mutability.clone(), ++ elem: self.elem.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeReference { ++ fn clone(&self) -> Self { ++ TypeReference { ++ and_token: self.and_token.clone(), ++ lifetime: self.lifetime.clone(), ++ mutability: self.mutability.clone(), ++ elem: self.elem.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeSlice { ++ fn clone(&self) -> Self { ++ TypeSlice { ++ bracket_token: self.bracket_token.clone(), ++ elem: self.elem.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeTraitObject { ++ fn clone(&self) -> Self { ++ TypeTraitObject { ++ dyn_token: self.dyn_token.clone(), ++ bounds: self.bounds.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for TypeTuple { ++ fn clone(&self) -> Self { ++ TypeTuple { ++ paren_token: self.paren_token.clone(), ++ elems: self.elems.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Copy for UnOp {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for UnOp { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UseGlob { ++ fn clone(&self) -> Self { ++ UseGlob { ++ star_token: self.star_token.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UseGroup { ++ fn clone(&self) -> Self { ++ UseGroup { ++ brace_token: self.brace_token.clone(), ++ items: self.items.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UseName { ++ fn clone(&self) -> Self { ++ UseName { ++ ident: self.ident.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UsePath { ++ fn clone(&self) -> Self { ++ UsePath { ++ ident: self.ident.clone(), ++ colon2_token: self.colon2_token.clone(), ++ tree: self.tree.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UseRename { ++ fn clone(&self) -> Self { ++ UseRename { ++ ident: self.ident.clone(), ++ as_token: self.as_token.clone(), ++ rename: self.rename.clone(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Clone for UseTree { ++ fn clone(&self) -> Self { ++ match self { ++ UseTree::Path(v0) => UseTree::Path(v0.clone()), ++ UseTree::Name(v0) => UseTree::Name(v0.clone()), ++ UseTree::Rename(v0) => UseTree::Rename(v0.clone()), ++ UseTree::Glob(v0) => UseTree::Glob(v0.clone()), ++ UseTree::Group(v0) => UseTree::Group(v0.clone()), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Variadic { ++ fn clone(&self) -> Self { ++ Variadic { ++ attrs: self.attrs.clone(), ++ dots: self.dots.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Variant { ++ fn clone(&self) -> Self { ++ Variant { ++ attrs: self.attrs.clone(), ++ ident: self.ident.clone(), ++ fields: self.fields.clone(), ++ discriminant: self.discriminant.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for VisCrate { ++ fn clone(&self) -> Self { ++ VisCrate { ++ crate_token: self.crate_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for VisPublic { ++ fn clone(&self) -> Self { ++ VisPublic { ++ pub_token: self.pub_token.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for VisRestricted { ++ fn clone(&self) -> Self { ++ VisRestricted { ++ pub_token: self.pub_token.clone(), ++ paren_token: self.paren_token.clone(), ++ in_token: self.in_token.clone(), ++ path: self.path.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for Visibility { ++ fn clone(&self) -> Self { ++ match self { ++ Visibility::Public(v0) => Visibility::Public(v0.clone()), ++ Visibility::Crate(v0) => Visibility::Crate(v0.clone()), ++ Visibility::Restricted(v0) => Visibility::Restricted(v0.clone()), ++ Visibility::Inherited => Visibility::Inherited, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for WhereClause { ++ fn clone(&self) -> Self { ++ WhereClause { ++ where_token: self.where_token.clone(), ++ predicates: self.predicates.clone(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Clone for WherePredicate { ++ fn clone(&self) -> Self { ++ match self { ++ WherePredicate::Type(v0) => WherePredicate::Type(v0.clone()), ++ WherePredicate::Lifetime(v0) => WherePredicate::Lifetime(v0.clone()), ++ WherePredicate::Eq(v0) => WherePredicate::Eq(v0.clone()), ++ } ++ } ++} +diff --git a/third_party/rust/syn/src/gen/debug.rs b/third_party/rust/syn/src/gen/debug.rs +new file mode 100644 +index 0000000000..72baab05f4 +--- /dev/null ++++ b/third_party/rust/syn/src/gen/debug.rs +@@ -0,0 +1,2857 @@ ++// This file is @generated by syn-internal-codegen. ++// It is not intended for manual editing. ++ ++use crate::*; ++use std::fmt::{self, Debug}; ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Abi { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Abi"); ++ formatter.field("extern_token", &self.extern_token); ++ formatter.field("name", &self.name); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for AngleBracketedGenericArguments { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("AngleBracketedGenericArguments"); ++ formatter.field("colon2_token", &self.colon2_token); ++ formatter.field("lt_token", &self.lt_token); ++ formatter.field("args", &self.args); ++ formatter.field("gt_token", &self.gt_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Arm { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Arm"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("pat", &self.pat); ++ formatter.field("guard", &self.guard); ++ formatter.field("fat_arrow_token", &self.fat_arrow_token); ++ formatter.field("body", &self.body); ++ formatter.field("comma", &self.comma); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for AttrStyle { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ AttrStyle::Outer => formatter.write_str("Outer"), ++ AttrStyle::Inner(v0) => { ++ let mut formatter = formatter.debug_tuple("Inner"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Attribute { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Attribute"); ++ formatter.field("pound_token", &self.pound_token); ++ formatter.field("style", &self.style); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("path", &self.path); ++ formatter.field("tokens", &self.tokens); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for BareFnArg { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("BareFnArg"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("name", &self.name); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for BinOp { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ BinOp::Add(v0) => { ++ let mut formatter = formatter.debug_tuple("Add"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Sub(v0) => { ++ let mut formatter = formatter.debug_tuple("Sub"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Mul(v0) => { ++ let mut formatter = formatter.debug_tuple("Mul"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Div(v0) => { ++ let mut formatter = formatter.debug_tuple("Div"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Rem(v0) => { ++ let mut formatter = formatter.debug_tuple("Rem"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::And(v0) => { ++ let mut formatter = formatter.debug_tuple("And"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Or(v0) => { ++ let mut formatter = formatter.debug_tuple("Or"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitXor(v0) => { ++ let mut formatter = formatter.debug_tuple("BitXor"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitAnd(v0) => { ++ let mut formatter = formatter.debug_tuple("BitAnd"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitOr(v0) => { ++ let mut formatter = formatter.debug_tuple("BitOr"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Shl(v0) => { ++ let mut formatter = formatter.debug_tuple("Shl"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Shr(v0) => { ++ let mut formatter = formatter.debug_tuple("Shr"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Eq(v0) => { ++ let mut formatter = formatter.debug_tuple("Eq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Lt(v0) => { ++ let mut formatter = formatter.debug_tuple("Lt"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Le(v0) => { ++ let mut formatter = formatter.debug_tuple("Le"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Ne(v0) => { ++ let mut formatter = formatter.debug_tuple("Ne"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Ge(v0) => { ++ let mut formatter = formatter.debug_tuple("Ge"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::Gt(v0) => { ++ let mut formatter = formatter.debug_tuple("Gt"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::AddEq(v0) => { ++ let mut formatter = formatter.debug_tuple("AddEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::SubEq(v0) => { ++ let mut formatter = formatter.debug_tuple("SubEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::MulEq(v0) => { ++ let mut formatter = formatter.debug_tuple("MulEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::DivEq(v0) => { ++ let mut formatter = formatter.debug_tuple("DivEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::RemEq(v0) => { ++ let mut formatter = formatter.debug_tuple("RemEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitXorEq(v0) => { ++ let mut formatter = formatter.debug_tuple("BitXorEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitAndEq(v0) => { ++ let mut formatter = formatter.debug_tuple("BitAndEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::BitOrEq(v0) => { ++ let mut formatter = formatter.debug_tuple("BitOrEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::ShlEq(v0) => { ++ let mut formatter = formatter.debug_tuple("ShlEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ BinOp::ShrEq(v0) => { ++ let mut formatter = formatter.debug_tuple("ShrEq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Binding { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Binding"); ++ formatter.field("ident", &self.ident); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Block { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Block"); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("stmts", &self.stmts); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for BoundLifetimes { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("BoundLifetimes"); ++ formatter.field("for_token", &self.for_token); ++ formatter.field("lt_token", &self.lt_token); ++ formatter.field("lifetimes", &self.lifetimes); ++ formatter.field("gt_token", &self.gt_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ConstParam { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ConstParam"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("const_token", &self.const_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("default", &self.default); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Constraint { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Constraint"); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "derive")] ++impl Debug for Data { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Data::Struct(v0) => { ++ let mut formatter = formatter.debug_tuple("Struct"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Data::Enum(v0) => { ++ let mut formatter = formatter.debug_tuple("Enum"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Data::Union(v0) => { ++ let mut formatter = formatter.debug_tuple("Union"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Debug for DataEnum { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("DataEnum"); ++ formatter.field("enum_token", &self.enum_token); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("variants", &self.variants); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "derive")] ++impl Debug for DataStruct { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("DataStruct"); ++ formatter.field("struct_token", &self.struct_token); ++ formatter.field("fields", &self.fields); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "derive")] ++impl Debug for DataUnion { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("DataUnion"); ++ formatter.field("union_token", &self.union_token); ++ formatter.field("fields", &self.fields); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "derive")] ++impl Debug for DeriveInput { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("DeriveInput"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("data", &self.data); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Expr { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ #[cfg(feature = "full")] ++ Expr::Array(v0) => { ++ let mut formatter = formatter.debug_tuple("Array"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Assign(v0) => { ++ let mut formatter = formatter.debug_tuple("Assign"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::AssignOp(v0) => { ++ let mut formatter = formatter.debug_tuple("AssignOp"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Async(v0) => { ++ let mut formatter = formatter.debug_tuple("Async"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Await(v0) => { ++ let mut formatter = formatter.debug_tuple("Await"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Binary(v0) => { ++ let mut formatter = formatter.debug_tuple("Binary"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Block(v0) => { ++ let mut formatter = formatter.debug_tuple("Block"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Box(v0) => { ++ let mut formatter = formatter.debug_tuple("Box"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Break(v0) => { ++ let mut formatter = formatter.debug_tuple("Break"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Call(v0) => { ++ let mut formatter = formatter.debug_tuple("Call"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Cast(v0) => { ++ let mut formatter = formatter.debug_tuple("Cast"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Closure(v0) => { ++ let mut formatter = formatter.debug_tuple("Closure"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Continue(v0) => { ++ let mut formatter = formatter.debug_tuple("Continue"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Field(v0) => { ++ let mut formatter = formatter.debug_tuple("Field"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::ForLoop(v0) => { ++ let mut formatter = formatter.debug_tuple("ForLoop"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Group(v0) => { ++ let mut formatter = formatter.debug_tuple("Group"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::If(v0) => { ++ let mut formatter = formatter.debug_tuple("If"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Index(v0) => { ++ let mut formatter = formatter.debug_tuple("Index"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Let(v0) => { ++ let mut formatter = formatter.debug_tuple("Let"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Lit(v0) => { ++ let mut formatter = formatter.debug_tuple("Lit"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Loop(v0) => { ++ let mut formatter = formatter.debug_tuple("Loop"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Match(v0) => { ++ let mut formatter = formatter.debug_tuple("Match"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::MethodCall(v0) => { ++ let mut formatter = formatter.debug_tuple("MethodCall"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Paren(v0) => { ++ let mut formatter = formatter.debug_tuple("Paren"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Path(v0) => { ++ let mut formatter = formatter.debug_tuple("Path"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Range(v0) => { ++ let mut formatter = formatter.debug_tuple("Range"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Reference(v0) => { ++ let mut formatter = formatter.debug_tuple("Reference"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Repeat(v0) => { ++ let mut formatter = formatter.debug_tuple("Repeat"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Return(v0) => { ++ let mut formatter = formatter.debug_tuple("Return"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Struct(v0) => { ++ let mut formatter = formatter.debug_tuple("Struct"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Try(v0) => { ++ let mut formatter = formatter.debug_tuple("Try"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::TryBlock(v0) => { ++ let mut formatter = formatter.debug_tuple("TryBlock"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Tuple(v0) => { ++ let mut formatter = formatter.debug_tuple("Tuple"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Unary(v0) => { ++ let mut formatter = formatter.debug_tuple("Unary"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Unsafe(v0) => { ++ let mut formatter = formatter.debug_tuple("Unsafe"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Expr::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::While(v0) => { ++ let mut formatter = formatter.debug_tuple("While"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ #[cfg(feature = "full")] ++ Expr::Yield(v0) => { ++ let mut formatter = formatter.debug_tuple("Yield"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprArray { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprArray"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("elems", &self.elems); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprAssign { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprAssign"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("left", &self.left); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("right", &self.right); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprAssignOp { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprAssignOp"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("left", &self.left); ++ formatter.field("op", &self.op); ++ formatter.field("right", &self.right); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprAsync { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprAsync"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("async_token", &self.async_token); ++ formatter.field("capture", &self.capture); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprAwait { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprAwait"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("base", &self.base); ++ formatter.field("dot_token", &self.dot_token); ++ formatter.field("await_token", &self.await_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprBinary { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprBinary"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("left", &self.left); ++ formatter.field("op", &self.op); ++ formatter.field("right", &self.right); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprBlock { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprBlock"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("label", &self.label); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprBox { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprBox"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("box_token", &self.box_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprBreak { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprBreak"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("break_token", &self.break_token); ++ formatter.field("label", &self.label); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprCall { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprCall"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("func", &self.func); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("args", &self.args); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprCast { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprCast"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("expr", &self.expr); ++ formatter.field("as_token", &self.as_token); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprClosure { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprClosure"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("asyncness", &self.asyncness); ++ formatter.field("movability", &self.movability); ++ formatter.field("capture", &self.capture); ++ formatter.field("or1_token", &self.or1_token); ++ formatter.field("inputs", &self.inputs); ++ formatter.field("or2_token", &self.or2_token); ++ formatter.field("output", &self.output); ++ formatter.field("body", &self.body); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprContinue { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprContinue"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("continue_token", &self.continue_token); ++ formatter.field("label", &self.label); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprField { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprField"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("base", &self.base); ++ formatter.field("dot_token", &self.dot_token); ++ formatter.field("member", &self.member); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprForLoop { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprForLoop"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("label", &self.label); ++ formatter.field("for_token", &self.for_token); ++ formatter.field("pat", &self.pat); ++ formatter.field("in_token", &self.in_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("body", &self.body); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprGroup { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprGroup"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("group_token", &self.group_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprIf { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprIf"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("if_token", &self.if_token); ++ formatter.field("cond", &self.cond); ++ formatter.field("then_branch", &self.then_branch); ++ formatter.field("else_branch", &self.else_branch); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprIndex { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprIndex"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("expr", &self.expr); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("index", &self.index); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprLet { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprLet"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("let_token", &self.let_token); ++ formatter.field("pat", &self.pat); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprLit { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprLit"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("lit", &self.lit); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprLoop { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprLoop"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("label", &self.label); ++ formatter.field("loop_token", &self.loop_token); ++ formatter.field("body", &self.body); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("mac", &self.mac); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprMatch { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprMatch"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("match_token", &self.match_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("arms", &self.arms); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprMethodCall { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprMethodCall"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("receiver", &self.receiver); ++ formatter.field("dot_token", &self.dot_token); ++ formatter.field("method", &self.method); ++ formatter.field("turbofish", &self.turbofish); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("args", &self.args); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprParen { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprParen"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprPath { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprPath"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("qself", &self.qself); ++ formatter.field("path", &self.path); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprRange { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprRange"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("from", &self.from); ++ formatter.field("limits", &self.limits); ++ formatter.field("to", &self.to); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprReference { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprReference"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("and_token", &self.and_token); ++ formatter.field("raw", &self.raw); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprRepeat { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprRepeat"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.field("len", &self.len); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprReturn { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprReturn"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("return_token", &self.return_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprStruct { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprStruct"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("path", &self.path); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("fields", &self.fields); ++ formatter.field("dot2_token", &self.dot2_token); ++ formatter.field("rest", &self.rest); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprTry { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprTry"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("expr", &self.expr); ++ formatter.field("question_token", &self.question_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprTryBlock { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprTryBlock"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("try_token", &self.try_token); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprTuple { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprTuple"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("elems", &self.elems); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("expr", &self.expr); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ExprUnary { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprUnary"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("op", &self.op); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprUnsafe { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprUnsafe"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("unsafe_token", &self.unsafe_token); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprWhile { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprWhile"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("label", &self.label); ++ formatter.field("while_token", &self.while_token); ++ formatter.field("cond", &self.cond); ++ formatter.field("body", &self.body); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ExprYield { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ExprYield"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("yield_token", &self.yield_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Field { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Field"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for FieldPat { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("FieldPat"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("member", &self.member); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("pat", &self.pat); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for FieldValue { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("FieldValue"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("member", &self.member); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Fields { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Fields::Named(v0) => { ++ let mut formatter = formatter.debug_tuple("Named"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Fields::Unnamed(v0) => { ++ let mut formatter = formatter.debug_tuple("Unnamed"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Fields::Unit => formatter.write_str("Unit"), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for FieldsNamed { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("FieldsNamed"); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("named", &self.named); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for FieldsUnnamed { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("FieldsUnnamed"); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("unnamed", &self.unnamed); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for File { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("File"); ++ formatter.field("shebang", &self.shebang); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("items", &self.items); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for FnArg { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ FnArg::Receiver(v0) => { ++ let mut formatter = formatter.debug_tuple("Receiver"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ FnArg::Typed(v0) => { ++ let mut formatter = formatter.debug_tuple("Typed"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ForeignItem { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ ForeignItem::Fn(v0) => { ++ let mut formatter = formatter.debug_tuple("Fn"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ForeignItem::Static(v0) => { ++ let mut formatter = formatter.debug_tuple("Static"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ForeignItem::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ForeignItem::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ForeignItem::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ForeignItemFn { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ForeignItemFn"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("sig", &self.sig); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ForeignItemMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ForeignItemMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("mac", &self.mac); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ForeignItemStatic { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ForeignItemStatic"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("static_token", &self.static_token); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ForeignItemType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ForeignItemType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("type_token", &self.type_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for GenericArgument { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ GenericArgument::Lifetime(v0) => { ++ let mut formatter = formatter.debug_tuple("Lifetime"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericArgument::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericArgument::Binding(v0) => { ++ let mut formatter = formatter.debug_tuple("Binding"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericArgument::Constraint(v0) => { ++ let mut formatter = formatter.debug_tuple("Constraint"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericArgument::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for GenericMethodArgument { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ GenericMethodArgument::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericMethodArgument::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for GenericParam { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ GenericParam::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericParam::Lifetime(v0) => { ++ let mut formatter = formatter.debug_tuple("Lifetime"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ GenericParam::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Generics { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Generics"); ++ formatter.field("lt_token", &self.lt_token); ++ formatter.field("params", &self.params); ++ formatter.field("gt_token", &self.gt_token); ++ formatter.field("where_clause", &self.where_clause); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ImplItem { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ ImplItem::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ImplItem::Method(v0) => { ++ let mut formatter = formatter.debug_tuple("Method"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ImplItem::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ImplItem::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ ImplItem::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ImplItemConst { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ImplItemConst"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("defaultness", &self.defaultness); ++ formatter.field("const_token", &self.const_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ImplItemMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ImplItemMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("mac", &self.mac); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ImplItemMethod { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ImplItemMethod"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("defaultness", &self.defaultness); ++ formatter.field("sig", &self.sig); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ImplItemType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ImplItemType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("defaultness", &self.defaultness); ++ formatter.field("type_token", &self.type_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Index { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Index"); ++ formatter.field("index", &self.index); ++ formatter.field("span", &self.span); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Item { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Item::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Enum(v0) => { ++ let mut formatter = formatter.debug_tuple("Enum"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::ExternCrate(v0) => { ++ let mut formatter = formatter.debug_tuple("ExternCrate"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Fn(v0) => { ++ let mut formatter = formatter.debug_tuple("Fn"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::ForeignMod(v0) => { ++ let mut formatter = formatter.debug_tuple("ForeignMod"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Impl(v0) => { ++ let mut formatter = formatter.debug_tuple("Impl"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Macro2(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro2"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Mod(v0) => { ++ let mut formatter = formatter.debug_tuple("Mod"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Static(v0) => { ++ let mut formatter = formatter.debug_tuple("Static"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Struct(v0) => { ++ let mut formatter = formatter.debug_tuple("Struct"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Trait(v0) => { ++ let mut formatter = formatter.debug_tuple("Trait"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::TraitAlias(v0) => { ++ let mut formatter = formatter.debug_tuple("TraitAlias"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Union(v0) => { ++ let mut formatter = formatter.debug_tuple("Union"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Use(v0) => { ++ let mut formatter = formatter.debug_tuple("Use"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Item::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemConst { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemConst"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("const_token", &self.const_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemEnum { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemEnum"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("enum_token", &self.enum_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("variants", &self.variants); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemExternCrate { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemExternCrate"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("extern_token", &self.extern_token); ++ formatter.field("crate_token", &self.crate_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("rename", &self.rename); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemFn { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemFn"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("sig", &self.sig); ++ formatter.field("block", &self.block); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemForeignMod { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemForeignMod"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("abi", &self.abi); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("items", &self.items); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemImpl { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemImpl"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("defaultness", &self.defaultness); ++ formatter.field("unsafety", &self.unsafety); ++ formatter.field("impl_token", &self.impl_token); ++ formatter.field("generics", &self.generics); ++ formatter.field("trait_", &self.trait_); ++ formatter.field("self_ty", &self.self_ty); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("items", &self.items); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("ident", &self.ident); ++ formatter.field("mac", &self.mac); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemMacro2 { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemMacro2"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("macro_token", &self.macro_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("rules", &self.rules); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemMod { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemMod"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("mod_token", &self.mod_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("content", &self.content); ++ formatter.field("semi", &self.semi); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemStatic { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemStatic"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("static_token", &self.static_token); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("expr", &self.expr); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemStruct { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemStruct"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("struct_token", &self.struct_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("fields", &self.fields); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemTrait { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemTrait"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("unsafety", &self.unsafety); ++ formatter.field("auto_token", &self.auto_token); ++ formatter.field("trait_token", &self.trait_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("supertraits", &self.supertraits); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("items", &self.items); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemTraitAlias { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemTraitAlias"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("trait_token", &self.trait_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("type_token", &self.type_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemUnion { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemUnion"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("union_token", &self.union_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("fields", &self.fields); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for ItemUse { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ItemUse"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("vis", &self.vis); ++ formatter.field("use_token", &self.use_token); ++ formatter.field("leading_colon", &self.leading_colon); ++ formatter.field("tree", &self.tree); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Label { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Label"); ++ formatter.field("name", &self.name); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.finish() ++ } ++} ++impl Debug for Lifetime { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Lifetime"); ++ formatter.field("apostrophe", &self.apostrophe); ++ formatter.field("ident", &self.ident); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for LifetimeDef { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("LifetimeDef"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("lifetime", &self.lifetime); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++impl Debug for Lit { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Lit::Str(v0) => { ++ let mut formatter = formatter.debug_tuple("Str"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::ByteStr(v0) => { ++ let mut formatter = formatter.debug_tuple("ByteStr"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Byte(v0) => { ++ let mut formatter = formatter.debug_tuple("Byte"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Char(v0) => { ++ let mut formatter = formatter.debug_tuple("Char"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Int(v0) => { ++ let mut formatter = formatter.debug_tuple("Int"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Float(v0) => { ++ let mut formatter = formatter.debug_tuple("Float"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Bool(v0) => { ++ let mut formatter = formatter.debug_tuple("Bool"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Lit::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Local { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Local"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("let_token", &self.let_token); ++ formatter.field("pat", &self.pat); ++ formatter.field("init", &self.init); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Macro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Macro"); ++ formatter.field("path", &self.path); ++ formatter.field("bang_token", &self.bang_token); ++ formatter.field("delimiter", &self.delimiter); ++ formatter.field("tokens", &self.tokens); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for MacroDelimiter { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ MacroDelimiter::Paren(v0) => { ++ let mut formatter = formatter.debug_tuple("Paren"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ MacroDelimiter::Brace(v0) => { ++ let mut formatter = formatter.debug_tuple("Brace"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ MacroDelimiter::Bracket(v0) => { ++ let mut formatter = formatter.debug_tuple("Bracket"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Member { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Member::Named(v0) => { ++ let mut formatter = formatter.debug_tuple("Named"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Member::Unnamed(v0) => { ++ let mut formatter = formatter.debug_tuple("Unnamed"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Meta { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Meta::Path(v0) => { ++ let mut formatter = formatter.debug_tuple("Path"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Meta::List(v0) => { ++ let mut formatter = formatter.debug_tuple("List"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Meta::NameValue(v0) => { ++ let mut formatter = formatter.debug_tuple("NameValue"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for MetaList { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("MetaList"); ++ formatter.field("path", &self.path); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("nested", &self.nested); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for MetaNameValue { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("MetaNameValue"); ++ formatter.field("path", &self.path); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("lit", &self.lit); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for MethodTurbofish { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("MethodTurbofish"); ++ formatter.field("colon2_token", &self.colon2_token); ++ formatter.field("lt_token", &self.lt_token); ++ formatter.field("args", &self.args); ++ formatter.field("gt_token", &self.gt_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for NestedMeta { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ NestedMeta::Meta(v0) => { ++ let mut formatter = formatter.debug_tuple("Meta"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ NestedMeta::Lit(v0) => { ++ let mut formatter = formatter.debug_tuple("Lit"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ParenthesizedGenericArguments { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("ParenthesizedGenericArguments"); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("inputs", &self.inputs); ++ formatter.field("output", &self.output); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Pat { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Pat::Box(v0) => { ++ let mut formatter = formatter.debug_tuple("Box"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Ident(v0) => { ++ let mut formatter = formatter.debug_tuple("Ident"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Lit(v0) => { ++ let mut formatter = formatter.debug_tuple("Lit"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Or(v0) => { ++ let mut formatter = formatter.debug_tuple("Or"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Path(v0) => { ++ let mut formatter = formatter.debug_tuple("Path"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Range(v0) => { ++ let mut formatter = formatter.debug_tuple("Range"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Reference(v0) => { ++ let mut formatter = formatter.debug_tuple("Reference"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Rest(v0) => { ++ let mut formatter = formatter.debug_tuple("Rest"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Slice(v0) => { ++ let mut formatter = formatter.debug_tuple("Slice"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Struct(v0) => { ++ let mut formatter = formatter.debug_tuple("Struct"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Tuple(v0) => { ++ let mut formatter = formatter.debug_tuple("Tuple"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::TupleStruct(v0) => { ++ let mut formatter = formatter.debug_tuple("TupleStruct"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Pat::Wild(v0) => { ++ let mut formatter = formatter.debug_tuple("Wild"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatBox { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatBox"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("box_token", &self.box_token); ++ formatter.field("pat", &self.pat); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatIdent { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatIdent"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("by_ref", &self.by_ref); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("ident", &self.ident); ++ formatter.field("subpat", &self.subpat); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatLit { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatLit"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("expr", &self.expr); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("mac", &self.mac); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatOr { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatOr"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("leading_vert", &self.leading_vert); ++ formatter.field("cases", &self.cases); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatPath { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatPath"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("qself", &self.qself); ++ formatter.field("path", &self.path); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatRange { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatRange"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("lo", &self.lo); ++ formatter.field("limits", &self.limits); ++ formatter.field("hi", &self.hi); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatReference { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatReference"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("and_token", &self.and_token); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("pat", &self.pat); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatRest { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatRest"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("dot2_token", &self.dot2_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatSlice { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatSlice"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("elems", &self.elems); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatStruct { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatStruct"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("path", &self.path); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("fields", &self.fields); ++ formatter.field("dot2_token", &self.dot2_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatTuple { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatTuple"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("elems", &self.elems); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatTupleStruct { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatTupleStruct"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("path", &self.path); ++ formatter.field("pat", &self.pat); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("pat", &self.pat); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for PatWild { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PatWild"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("underscore_token", &self.underscore_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Path { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Path"); ++ formatter.field("leading_colon", &self.leading_colon); ++ formatter.field("segments", &self.segments); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for PathArguments { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ PathArguments::None => formatter.write_str("None"), ++ PathArguments::AngleBracketed(v0) => { ++ let mut formatter = formatter.debug_tuple("AngleBracketed"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ PathArguments::Parenthesized(v0) => { ++ let mut formatter = formatter.debug_tuple("Parenthesized"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for PathSegment { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PathSegment"); ++ formatter.field("ident", &self.ident); ++ formatter.field("arguments", &self.arguments); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for PredicateEq { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PredicateEq"); ++ formatter.field("lhs_ty", &self.lhs_ty); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("rhs_ty", &self.rhs_ty); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for PredicateLifetime { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PredicateLifetime"); ++ formatter.field("lifetime", &self.lifetime); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for PredicateType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("PredicateType"); ++ formatter.field("lifetimes", &self.lifetimes); ++ formatter.field("bounded_ty", &self.bounded_ty); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for QSelf { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("QSelf"); ++ formatter.field("lt_token", &self.lt_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("position", &self.position); ++ formatter.field("as_token", &self.as_token); ++ formatter.field("gt_token", &self.gt_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for RangeLimits { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ RangeLimits::HalfOpen(v0) => { ++ let mut formatter = formatter.debug_tuple("HalfOpen"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ RangeLimits::Closed(v0) => { ++ let mut formatter = formatter.debug_tuple("Closed"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Receiver { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Receiver"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("reference", &self.reference); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("self_token", &self.self_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for ReturnType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ ReturnType::Default => formatter.write_str("Default"), ++ ReturnType::Type(v0, v1) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.field(v1); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Signature { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Signature"); ++ formatter.field("constness", &self.constness); ++ formatter.field("asyncness", &self.asyncness); ++ formatter.field("unsafety", &self.unsafety); ++ formatter.field("abi", &self.abi); ++ formatter.field("fn_token", &self.fn_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("inputs", &self.inputs); ++ formatter.field("variadic", &self.variadic); ++ formatter.field("output", &self.output); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for Stmt { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Stmt::Local(v0) => { ++ let mut formatter = formatter.debug_tuple("Local"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Stmt::Item(v0) => { ++ let mut formatter = formatter.debug_tuple("Item"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Stmt::Expr(v0) => { ++ let mut formatter = formatter.debug_tuple("Expr"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Stmt::Semi(v0, v1) => { ++ let mut formatter = formatter.debug_tuple("Semi"); ++ formatter.field(v0); ++ formatter.field(v1); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TraitBound { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TraitBound"); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("modifier", &self.modifier); ++ formatter.field("lifetimes", &self.lifetimes); ++ formatter.field("path", &self.path); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TraitBoundModifier { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ TraitBoundModifier::None => formatter.write_str("None"), ++ TraitBoundModifier::Maybe(v0) => { ++ let mut formatter = formatter.debug_tuple("Maybe"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for TraitItem { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ TraitItem::Const(v0) => { ++ let mut formatter = formatter.debug_tuple("Const"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ TraitItem::Method(v0) => { ++ let mut formatter = formatter.debug_tuple("Method"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ TraitItem::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ TraitItem::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ TraitItem::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for TraitItemConst { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TraitItemConst"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("const_token", &self.const_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("ty", &self.ty); ++ formatter.field("default", &self.default); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for TraitItemMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TraitItemMacro"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("mac", &self.mac); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for TraitItemMethod { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TraitItemMethod"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("sig", &self.sig); ++ formatter.field("default", &self.default); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for TraitItemType { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TraitItemType"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("type_token", &self.type_token); ++ formatter.field("ident", &self.ident); ++ formatter.field("generics", &self.generics); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.field("default", &self.default); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Type { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Type::Array(v0) => { ++ let mut formatter = formatter.debug_tuple("Array"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::BareFn(v0) => { ++ let mut formatter = formatter.debug_tuple("BareFn"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Group(v0) => { ++ let mut formatter = formatter.debug_tuple("Group"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::ImplTrait(v0) => { ++ let mut formatter = formatter.debug_tuple("ImplTrait"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Infer(v0) => { ++ let mut formatter = formatter.debug_tuple("Infer"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Macro(v0) => { ++ let mut formatter = formatter.debug_tuple("Macro"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Never(v0) => { ++ let mut formatter = formatter.debug_tuple("Never"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Paren(v0) => { ++ let mut formatter = formatter.debug_tuple("Paren"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Path(v0) => { ++ let mut formatter = formatter.debug_tuple("Path"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Ptr(v0) => { ++ let mut formatter = formatter.debug_tuple("Ptr"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Reference(v0) => { ++ let mut formatter = formatter.debug_tuple("Reference"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Slice(v0) => { ++ let mut formatter = formatter.debug_tuple("Slice"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::TraitObject(v0) => { ++ let mut formatter = formatter.debug_tuple("TraitObject"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Tuple(v0) => { ++ let mut formatter = formatter.debug_tuple("Tuple"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Type::Verbatim(v0) => { ++ let mut formatter = formatter.debug_tuple("Verbatim"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeArray { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeArray"); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("elem", &self.elem); ++ formatter.field("semi_token", &self.semi_token); ++ formatter.field("len", &self.len); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeBareFn { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeBareFn"); ++ formatter.field("lifetimes", &self.lifetimes); ++ formatter.field("unsafety", &self.unsafety); ++ formatter.field("abi", &self.abi); ++ formatter.field("fn_token", &self.fn_token); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("inputs", &self.inputs); ++ formatter.field("variadic", &self.variadic); ++ formatter.field("output", &self.output); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeGroup { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeGroup"); ++ formatter.field("group_token", &self.group_token); ++ formatter.field("elem", &self.elem); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeImplTrait { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeImplTrait"); ++ formatter.field("impl_token", &self.impl_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeInfer { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeInfer"); ++ formatter.field("underscore_token", &self.underscore_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeMacro { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeMacro"); ++ formatter.field("mac", &self.mac); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeNever { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeNever"); ++ formatter.field("bang_token", &self.bang_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeParam { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeParam"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon_token", &self.colon_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.field("eq_token", &self.eq_token); ++ formatter.field("default", &self.default); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeParamBound { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ TypeParamBound::Trait(v0) => { ++ let mut formatter = formatter.debug_tuple("Trait"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ TypeParamBound::Lifetime(v0) => { ++ let mut formatter = formatter.debug_tuple("Lifetime"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeParen { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeParen"); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("elem", &self.elem); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypePath { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypePath"); ++ formatter.field("qself", &self.qself); ++ formatter.field("path", &self.path); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypePtr { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypePtr"); ++ formatter.field("star_token", &self.star_token); ++ formatter.field("const_token", &self.const_token); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("elem", &self.elem); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeReference { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeReference"); ++ formatter.field("and_token", &self.and_token); ++ formatter.field("lifetime", &self.lifetime); ++ formatter.field("mutability", &self.mutability); ++ formatter.field("elem", &self.elem); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeSlice { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeSlice"); ++ formatter.field("bracket_token", &self.bracket_token); ++ formatter.field("elem", &self.elem); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeTraitObject { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeTraitObject"); ++ formatter.field("dyn_token", &self.dyn_token); ++ formatter.field("bounds", &self.bounds); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for TypeTuple { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("TypeTuple"); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("elems", &self.elems); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for UnOp { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ UnOp::Deref(v0) => { ++ let mut formatter = formatter.debug_tuple("Deref"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UnOp::Not(v0) => { ++ let mut formatter = formatter.debug_tuple("Not"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UnOp::Neg(v0) => { ++ let mut formatter = formatter.debug_tuple("Neg"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UseGlob { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("UseGlob"); ++ formatter.field("star_token", &self.star_token); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UseGroup { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("UseGroup"); ++ formatter.field("brace_token", &self.brace_token); ++ formatter.field("items", &self.items); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UseName { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("UseName"); ++ formatter.field("ident", &self.ident); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UsePath { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("UsePath"); ++ formatter.field("ident", &self.ident); ++ formatter.field("colon2_token", &self.colon2_token); ++ formatter.field("tree", &self.tree); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UseRename { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("UseRename"); ++ formatter.field("ident", &self.ident); ++ formatter.field("as_token", &self.as_token); ++ formatter.field("rename", &self.rename); ++ formatter.finish() ++ } ++} ++#[cfg(feature = "full")] ++impl Debug for UseTree { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ UseTree::Path(v0) => { ++ let mut formatter = formatter.debug_tuple("Path"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UseTree::Name(v0) => { ++ let mut formatter = formatter.debug_tuple("Name"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UseTree::Rename(v0) => { ++ let mut formatter = formatter.debug_tuple("Rename"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UseTree::Glob(v0) => { ++ let mut formatter = formatter.debug_tuple("Glob"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ UseTree::Group(v0) => { ++ let mut formatter = formatter.debug_tuple("Group"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Variadic { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Variadic"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("dots", &self.dots); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Variant { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("Variant"); ++ formatter.field("attrs", &self.attrs); ++ formatter.field("ident", &self.ident); ++ formatter.field("fields", &self.fields); ++ formatter.field("discriminant", &self.discriminant); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for VisCrate { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("VisCrate"); ++ formatter.field("crate_token", &self.crate_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for VisPublic { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("VisPublic"); ++ formatter.field("pub_token", &self.pub_token); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for VisRestricted { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("VisRestricted"); ++ formatter.field("pub_token", &self.pub_token); ++ formatter.field("paren_token", &self.paren_token); ++ formatter.field("in_token", &self.in_token); ++ formatter.field("path", &self.path); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for Visibility { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ Visibility::Public(v0) => { ++ let mut formatter = formatter.debug_tuple("Public"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Visibility::Crate(v0) => { ++ let mut formatter = formatter.debug_tuple("Crate"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Visibility::Restricted(v0) => { ++ let mut formatter = formatter.debug_tuple("Restricted"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ Visibility::Inherited => formatter.write_str("Inherited"), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for WhereClause { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ let mut formatter = formatter.debug_struct("WhereClause"); ++ formatter.field("where_token", &self.where_token); ++ formatter.field("predicates", &self.predicates); ++ formatter.finish() ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Debug for WherePredicate { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ match self { ++ WherePredicate::Type(v0) => { ++ let mut formatter = formatter.debug_tuple("Type"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ WherePredicate::Lifetime(v0) => { ++ let mut formatter = formatter.debug_tuple("Lifetime"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ WherePredicate::Eq(v0) => { ++ let mut formatter = formatter.debug_tuple("Eq"); ++ formatter.field(v0); ++ formatter.finish() ++ } ++ } ++ } ++} +diff --git a/third_party/rust/syn/src/gen/eq.rs b/third_party/rust/syn/src/gen/eq.rs +new file mode 100644 +index 0000000000..15b2bcbbde +--- /dev/null ++++ b/third_party/rust/syn/src/gen/eq.rs +@@ -0,0 +1,1930 @@ ++// This file is @generated by syn-internal-codegen. ++// It is not intended for manual editing. ++ ++#[cfg(any(feature = "derive", feature = "full"))] ++use crate::tt::TokenStreamHelper; ++use crate::*; ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Abi {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Abi { ++ fn eq(&self, other: &Self) -> bool { ++ self.name == other.name ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for AngleBracketedGenericArguments {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for AngleBracketedGenericArguments { ++ fn eq(&self, other: &Self) -> bool { ++ self.colon2_token == other.colon2_token && self.args == other.args ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Arm {} ++#[cfg(feature = "full")] ++impl PartialEq for Arm { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.pat == other.pat ++ && self.guard == other.guard ++ && self.body == other.body ++ && self.comma == other.comma ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for AttrStyle {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for AttrStyle { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (AttrStyle::Outer, AttrStyle::Outer) => true, ++ (AttrStyle::Inner(_), AttrStyle::Inner(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Attribute {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Attribute { ++ fn eq(&self, other: &Self) -> bool { ++ self.style == other.style ++ && self.path == other.path ++ && TokenStreamHelper(&self.tokens) == TokenStreamHelper(&other.tokens) ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for BareFnArg {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for BareFnArg { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.name == other.name && self.ty == other.ty ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for BinOp {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for BinOp { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (BinOp::Add(_), BinOp::Add(_)) => true, ++ (BinOp::Sub(_), BinOp::Sub(_)) => true, ++ (BinOp::Mul(_), BinOp::Mul(_)) => true, ++ (BinOp::Div(_), BinOp::Div(_)) => true, ++ (BinOp::Rem(_), BinOp::Rem(_)) => true, ++ (BinOp::And(_), BinOp::And(_)) => true, ++ (BinOp::Or(_), BinOp::Or(_)) => true, ++ (BinOp::BitXor(_), BinOp::BitXor(_)) => true, ++ (BinOp::BitAnd(_), BinOp::BitAnd(_)) => true, ++ (BinOp::BitOr(_), BinOp::BitOr(_)) => true, ++ (BinOp::Shl(_), BinOp::Shl(_)) => true, ++ (BinOp::Shr(_), BinOp::Shr(_)) => true, ++ (BinOp::Eq(_), BinOp::Eq(_)) => true, ++ (BinOp::Lt(_), BinOp::Lt(_)) => true, ++ (BinOp::Le(_), BinOp::Le(_)) => true, ++ (BinOp::Ne(_), BinOp::Ne(_)) => true, ++ (BinOp::Ge(_), BinOp::Ge(_)) => true, ++ (BinOp::Gt(_), BinOp::Gt(_)) => true, ++ (BinOp::AddEq(_), BinOp::AddEq(_)) => true, ++ (BinOp::SubEq(_), BinOp::SubEq(_)) => true, ++ (BinOp::MulEq(_), BinOp::MulEq(_)) => true, ++ (BinOp::DivEq(_), BinOp::DivEq(_)) => true, ++ (BinOp::RemEq(_), BinOp::RemEq(_)) => true, ++ (BinOp::BitXorEq(_), BinOp::BitXorEq(_)) => true, ++ (BinOp::BitAndEq(_), BinOp::BitAndEq(_)) => true, ++ (BinOp::BitOrEq(_), BinOp::BitOrEq(_)) => true, ++ (BinOp::ShlEq(_), BinOp::ShlEq(_)) => true, ++ (BinOp::ShrEq(_), BinOp::ShrEq(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Binding {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Binding { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Block {} ++#[cfg(feature = "full")] ++impl PartialEq for Block { ++ fn eq(&self, other: &Self) -> bool { ++ self.stmts == other.stmts ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for BoundLifetimes {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for BoundLifetimes { ++ fn eq(&self, other: &Self) -> bool { ++ self.lifetimes == other.lifetimes ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ConstParam {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ConstParam { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.ty == other.ty ++ && self.eq_token == other.eq_token ++ && self.default == other.default ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Constraint {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Constraint { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident && self.bounds == other.bounds ++ } ++} ++#[cfg(feature = "derive")] ++impl Eq for Data {} ++#[cfg(feature = "derive")] ++impl PartialEq for Data { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Data::Struct(self0), Data::Struct(other0)) => self0 == other0, ++ (Data::Enum(self0), Data::Enum(other0)) => self0 == other0, ++ (Data::Union(self0), Data::Union(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Eq for DataEnum {} ++#[cfg(feature = "derive")] ++impl PartialEq for DataEnum { ++ fn eq(&self, other: &Self) -> bool { ++ self.variants == other.variants ++ } ++} ++#[cfg(feature = "derive")] ++impl Eq for DataStruct {} ++#[cfg(feature = "derive")] ++impl PartialEq for DataStruct { ++ fn eq(&self, other: &Self) -> bool { ++ self.fields == other.fields && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "derive")] ++impl Eq for DataUnion {} ++#[cfg(feature = "derive")] ++impl PartialEq for DataUnion { ++ fn eq(&self, other: &Self) -> bool { ++ self.fields == other.fields ++ } ++} ++#[cfg(feature = "derive")] ++impl Eq for DeriveInput {} ++#[cfg(feature = "derive")] ++impl PartialEq for DeriveInput { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.data == other.data ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Expr {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Expr { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ #[cfg(feature = "full")] ++ (Expr::Array(self0), Expr::Array(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Assign(self0), Expr::Assign(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::AssignOp(self0), Expr::AssignOp(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Async(self0), Expr::Async(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Await(self0), Expr::Await(other0)) => self0 == other0, ++ (Expr::Binary(self0), Expr::Binary(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Block(self0), Expr::Block(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Box(self0), Expr::Box(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Break(self0), Expr::Break(other0)) => self0 == other0, ++ (Expr::Call(self0), Expr::Call(other0)) => self0 == other0, ++ (Expr::Cast(self0), Expr::Cast(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Closure(self0), Expr::Closure(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Continue(self0), Expr::Continue(other0)) => self0 == other0, ++ (Expr::Field(self0), Expr::Field(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::ForLoop(self0), Expr::ForLoop(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Group(self0), Expr::Group(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::If(self0), Expr::If(other0)) => self0 == other0, ++ (Expr::Index(self0), Expr::Index(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Let(self0), Expr::Let(other0)) => self0 == other0, ++ (Expr::Lit(self0), Expr::Lit(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Loop(self0), Expr::Loop(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Macro(self0), Expr::Macro(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Match(self0), Expr::Match(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::MethodCall(self0), Expr::MethodCall(other0)) => self0 == other0, ++ (Expr::Paren(self0), Expr::Paren(other0)) => self0 == other0, ++ (Expr::Path(self0), Expr::Path(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Range(self0), Expr::Range(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Reference(self0), Expr::Reference(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Repeat(self0), Expr::Repeat(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Return(self0), Expr::Return(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Struct(self0), Expr::Struct(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Try(self0), Expr::Try(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::TryBlock(self0), Expr::TryBlock(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Tuple(self0), Expr::Tuple(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Type(self0), Expr::Type(other0)) => self0 == other0, ++ (Expr::Unary(self0), Expr::Unary(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Unsafe(self0), Expr::Unsafe(other0)) => self0 == other0, ++ (Expr::Verbatim(self0), Expr::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ #[cfg(feature = "full")] ++ (Expr::While(self0), Expr::While(other0)) => self0 == other0, ++ #[cfg(feature = "full")] ++ (Expr::Yield(self0), Expr::Yield(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprArray {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprArray { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.elems == other.elems ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprAssign {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprAssign { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.left == other.left && self.right == other.right ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprAssignOp {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprAssignOp { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.left == other.left ++ && self.op == other.op ++ && self.right == other.right ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprAsync {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprAsync { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.capture == other.capture && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprAwait {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprAwait { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.base == other.base ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprBinary {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprBinary { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.left == other.left ++ && self.op == other.op ++ && self.right == other.right ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprBlock {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprBlock { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.label == other.label && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprBox {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprBox { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprBreak {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprBreak { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.label == other.label && self.expr == other.expr ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprCall {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprCall { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.func == other.func && self.args == other.args ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprCast {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprCast { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprClosure {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprClosure { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.asyncness == other.asyncness ++ && self.movability == other.movability ++ && self.capture == other.capture ++ && self.inputs == other.inputs ++ && self.output == other.output ++ && self.body == other.body ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprContinue {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprContinue { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.label == other.label ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprField {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprField { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.base == other.base && self.member == other.member ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprForLoop {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprForLoop { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.label == other.label ++ && self.pat == other.pat ++ && self.expr == other.expr ++ && self.body == other.body ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprGroup {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprGroup { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprIf {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprIf { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.cond == other.cond ++ && self.then_branch == other.then_branch ++ && self.else_branch == other.else_branch ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprIndex {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprIndex { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr && self.index == other.index ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprLet {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprLet { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.pat == other.pat && self.expr == other.expr ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprLit {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprLit { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.lit == other.lit ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprLoop {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprLoop { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.label == other.label && self.body == other.body ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mac == other.mac ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprMatch {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprMatch { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr && self.arms == other.arms ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprMethodCall {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprMethodCall { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.receiver == other.receiver ++ && self.method == other.method ++ && self.turbofish == other.turbofish ++ && self.args == other.args ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprParen {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprParen { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprPath {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprPath { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.qself == other.qself && self.path == other.path ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprRange {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprRange { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.from == other.from ++ && self.limits == other.limits ++ && self.to == other.to ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprReference {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprReference { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mutability == other.mutability && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprRepeat {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprRepeat { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr && self.len == other.len ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprReturn {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprReturn { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprStruct {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprStruct { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.path == other.path ++ && self.fields == other.fields ++ && self.dot2_token == other.dot2_token ++ && self.rest == other.rest ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprTry {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprTry { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprTryBlock {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprTryBlock { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprTuple {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprTuple { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.elems == other.elems ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprType {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr && self.ty == other.ty ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ExprUnary {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ExprUnary { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.op == other.op && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprUnsafe {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprUnsafe { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprWhile {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprWhile { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.label == other.label ++ && self.cond == other.cond ++ && self.body == other.body ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ExprYield {} ++#[cfg(feature = "full")] ++impl PartialEq for ExprYield { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Field {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Field { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.colon_token == other.colon_token ++ && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for FieldPat {} ++#[cfg(feature = "full")] ++impl PartialEq for FieldPat { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.member == other.member ++ && self.colon_token == other.colon_token ++ && self.pat == other.pat ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for FieldValue {} ++#[cfg(feature = "full")] ++impl PartialEq for FieldValue { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.member == other.member ++ && self.colon_token == other.colon_token ++ && self.expr == other.expr ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Fields {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Fields { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Fields::Named(self0), Fields::Named(other0)) => self0 == other0, ++ (Fields::Unnamed(self0), Fields::Unnamed(other0)) => self0 == other0, ++ (Fields::Unit, Fields::Unit) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for FieldsNamed {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for FieldsNamed { ++ fn eq(&self, other: &Self) -> bool { ++ self.named == other.named ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for FieldsUnnamed {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for FieldsUnnamed { ++ fn eq(&self, other: &Self) -> bool { ++ self.unnamed == other.unnamed ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for File {} ++#[cfg(feature = "full")] ++impl PartialEq for File { ++ fn eq(&self, other: &Self) -> bool { ++ self.shebang == other.shebang && self.attrs == other.attrs && self.items == other.items ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for FnArg {} ++#[cfg(feature = "full")] ++impl PartialEq for FnArg { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (FnArg::Receiver(self0), FnArg::Receiver(other0)) => self0 == other0, ++ (FnArg::Typed(self0), FnArg::Typed(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ForeignItem {} ++#[cfg(feature = "full")] ++impl PartialEq for ForeignItem { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (ForeignItem::Fn(self0), ForeignItem::Fn(other0)) => self0 == other0, ++ (ForeignItem::Static(self0), ForeignItem::Static(other0)) => self0 == other0, ++ (ForeignItem::Type(self0), ForeignItem::Type(other0)) => self0 == other0, ++ (ForeignItem::Macro(self0), ForeignItem::Macro(other0)) => self0 == other0, ++ (ForeignItem::Verbatim(self0), ForeignItem::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ForeignItemFn {} ++#[cfg(feature = "full")] ++impl PartialEq for ForeignItemFn { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.vis == other.vis && self.sig == other.sig ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ForeignItemMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for ForeignItemMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mac == other.mac && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ForeignItemStatic {} ++#[cfg(feature = "full")] ++impl PartialEq for ForeignItemStatic { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.mutability == other.mutability ++ && self.ident == other.ident ++ && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ForeignItemType {} ++#[cfg(feature = "full")] ++impl PartialEq for ForeignItemType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.vis == other.vis && self.ident == other.ident ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for GenericArgument {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for GenericArgument { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (GenericArgument::Lifetime(self0), GenericArgument::Lifetime(other0)) => { ++ self0 == other0 ++ } ++ (GenericArgument::Type(self0), GenericArgument::Type(other0)) => self0 == other0, ++ (GenericArgument::Binding(self0), GenericArgument::Binding(other0)) => self0 == other0, ++ (GenericArgument::Constraint(self0), GenericArgument::Constraint(other0)) => { ++ self0 == other0 ++ } ++ (GenericArgument::Const(self0), GenericArgument::Const(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for GenericMethodArgument {} ++#[cfg(feature = "full")] ++impl PartialEq for GenericMethodArgument { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (GenericMethodArgument::Type(self0), GenericMethodArgument::Type(other0)) => { ++ self0 == other0 ++ } ++ (GenericMethodArgument::Const(self0), GenericMethodArgument::Const(other0)) => { ++ self0 == other0 ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for GenericParam {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for GenericParam { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (GenericParam::Type(self0), GenericParam::Type(other0)) => self0 == other0, ++ (GenericParam::Lifetime(self0), GenericParam::Lifetime(other0)) => self0 == other0, ++ (GenericParam::Const(self0), GenericParam::Const(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Generics {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Generics { ++ fn eq(&self, other: &Self) -> bool { ++ self.lt_token == other.lt_token ++ && self.params == other.params ++ && self.gt_token == other.gt_token ++ && self.where_clause == other.where_clause ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ImplItem {} ++#[cfg(feature = "full")] ++impl PartialEq for ImplItem { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (ImplItem::Const(self0), ImplItem::Const(other0)) => self0 == other0, ++ (ImplItem::Method(self0), ImplItem::Method(other0)) => self0 == other0, ++ (ImplItem::Type(self0), ImplItem::Type(other0)) => self0 == other0, ++ (ImplItem::Macro(self0), ImplItem::Macro(other0)) => self0 == other0, ++ (ImplItem::Verbatim(self0), ImplItem::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ImplItemConst {} ++#[cfg(feature = "full")] ++impl PartialEq for ImplItemConst { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.defaultness == other.defaultness ++ && self.ident == other.ident ++ && self.ty == other.ty ++ && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ImplItemMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for ImplItemMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mac == other.mac && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ImplItemMethod {} ++#[cfg(feature = "full")] ++impl PartialEq for ImplItemMethod { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.defaultness == other.defaultness ++ && self.sig == other.sig ++ && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ImplItemType {} ++#[cfg(feature = "full")] ++impl PartialEq for ImplItemType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.defaultness == other.defaultness ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Item {} ++#[cfg(feature = "full")] ++impl PartialEq for Item { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Item::Const(self0), Item::Const(other0)) => self0 == other0, ++ (Item::Enum(self0), Item::Enum(other0)) => self0 == other0, ++ (Item::ExternCrate(self0), Item::ExternCrate(other0)) => self0 == other0, ++ (Item::Fn(self0), Item::Fn(other0)) => self0 == other0, ++ (Item::ForeignMod(self0), Item::ForeignMod(other0)) => self0 == other0, ++ (Item::Impl(self0), Item::Impl(other0)) => self0 == other0, ++ (Item::Macro(self0), Item::Macro(other0)) => self0 == other0, ++ (Item::Macro2(self0), Item::Macro2(other0)) => self0 == other0, ++ (Item::Mod(self0), Item::Mod(other0)) => self0 == other0, ++ (Item::Static(self0), Item::Static(other0)) => self0 == other0, ++ (Item::Struct(self0), Item::Struct(other0)) => self0 == other0, ++ (Item::Trait(self0), Item::Trait(other0)) => self0 == other0, ++ (Item::TraitAlias(self0), Item::TraitAlias(other0)) => self0 == other0, ++ (Item::Type(self0), Item::Type(other0)) => self0 == other0, ++ (Item::Union(self0), Item::Union(other0)) => self0 == other0, ++ (Item::Use(self0), Item::Use(other0)) => self0 == other0, ++ (Item::Verbatim(self0), Item::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemConst {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemConst { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.ty == other.ty ++ && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemEnum {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemEnum { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.variants == other.variants ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemExternCrate {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemExternCrate { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.rename == other.rename ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemFn {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemFn { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.sig == other.sig ++ && self.block == other.block ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemForeignMod {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemForeignMod { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.abi == other.abi && self.items == other.items ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemImpl {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemImpl { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.defaultness == other.defaultness ++ && self.unsafety == other.unsafety ++ && self.generics == other.generics ++ && self.trait_ == other.trait_ ++ && self.self_ty == other.self_ty ++ && self.items == other.items ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.mac == other.mac ++ && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemMacro2 {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemMacro2 { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && TokenStreamHelper(&self.rules) == TokenStreamHelper(&other.rules) ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemMod {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemMod { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.content == other.content ++ && self.semi == other.semi ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemStatic {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemStatic { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.mutability == other.mutability ++ && self.ident == other.ident ++ && self.ty == other.ty ++ && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemStruct {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemStruct { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.fields == other.fields ++ && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemTrait {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemTrait { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.unsafety == other.unsafety ++ && self.auto_token == other.auto_token ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.colon_token == other.colon_token ++ && self.supertraits == other.supertraits ++ && self.items == other.items ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemTraitAlias {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemTraitAlias { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.bounds == other.bounds ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemType {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemUnion {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemUnion { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.fields == other.fields ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for ItemUse {} ++#[cfg(feature = "full")] ++impl PartialEq for ItemUse { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.vis == other.vis ++ && self.leading_colon == other.leading_colon ++ && self.tree == other.tree ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Label {} ++#[cfg(feature = "full")] ++impl PartialEq for Label { ++ fn eq(&self, other: &Self) -> bool { ++ self.name == other.name ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for LifetimeDef {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for LifetimeDef { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.lifetime == other.lifetime ++ && self.colon_token == other.colon_token ++ && self.bounds == other.bounds ++ } ++} ++impl Eq for Lit {} ++impl PartialEq for Lit { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Lit::Str(self0), Lit::Str(other0)) => self0 == other0, ++ (Lit::ByteStr(self0), Lit::ByteStr(other0)) => self0 == other0, ++ (Lit::Byte(self0), Lit::Byte(other0)) => self0 == other0, ++ (Lit::Char(self0), Lit::Char(other0)) => self0 == other0, ++ (Lit::Int(self0), Lit::Int(other0)) => self0 == other0, ++ (Lit::Float(self0), Lit::Float(other0)) => self0 == other0, ++ (Lit::Bool(self0), Lit::Bool(other0)) => self0 == other0, ++ (Lit::Verbatim(self0), Lit::Verbatim(other0)) => { ++ self0.to_string() == other0.to_string() ++ } ++ _ => false, ++ } ++ } ++} ++impl Eq for LitBool {} ++impl PartialEq for LitBool { ++ fn eq(&self, other: &Self) -> bool { ++ self.value == other.value ++ } ++} ++impl Eq for LitByte {} ++impl Eq for LitByteStr {} ++impl Eq for LitChar {} ++impl Eq for LitFloat {} ++impl Eq for LitInt {} ++impl Eq for LitStr {} ++#[cfg(feature = "full")] ++impl Eq for Local {} ++#[cfg(feature = "full")] ++impl PartialEq for Local { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.pat == other.pat && self.init == other.init ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Macro {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Macro { ++ fn eq(&self, other: &Self) -> bool { ++ self.path == other.path ++ && self.delimiter == other.delimiter ++ && TokenStreamHelper(&self.tokens) == TokenStreamHelper(&other.tokens) ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for MacroDelimiter {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for MacroDelimiter { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (MacroDelimiter::Paren(_), MacroDelimiter::Paren(_)) => true, ++ (MacroDelimiter::Brace(_), MacroDelimiter::Brace(_)) => true, ++ (MacroDelimiter::Bracket(_), MacroDelimiter::Bracket(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Meta {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Meta { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Meta::Path(self0), Meta::Path(other0)) => self0 == other0, ++ (Meta::List(self0), Meta::List(other0)) => self0 == other0, ++ (Meta::NameValue(self0), Meta::NameValue(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for MetaList {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for MetaList { ++ fn eq(&self, other: &Self) -> bool { ++ self.path == other.path && self.nested == other.nested ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for MetaNameValue {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for MetaNameValue { ++ fn eq(&self, other: &Self) -> bool { ++ self.path == other.path && self.lit == other.lit ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for MethodTurbofish {} ++#[cfg(feature = "full")] ++impl PartialEq for MethodTurbofish { ++ fn eq(&self, other: &Self) -> bool { ++ self.args == other.args ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for NestedMeta {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for NestedMeta { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (NestedMeta::Meta(self0), NestedMeta::Meta(other0)) => self0 == other0, ++ (NestedMeta::Lit(self0), NestedMeta::Lit(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ParenthesizedGenericArguments {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ParenthesizedGenericArguments { ++ fn eq(&self, other: &Self) -> bool { ++ self.inputs == other.inputs && self.output == other.output ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Pat {} ++#[cfg(feature = "full")] ++impl PartialEq for Pat { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Pat::Box(self0), Pat::Box(other0)) => self0 == other0, ++ (Pat::Ident(self0), Pat::Ident(other0)) => self0 == other0, ++ (Pat::Lit(self0), Pat::Lit(other0)) => self0 == other0, ++ (Pat::Macro(self0), Pat::Macro(other0)) => self0 == other0, ++ (Pat::Or(self0), Pat::Or(other0)) => self0 == other0, ++ (Pat::Path(self0), Pat::Path(other0)) => self0 == other0, ++ (Pat::Range(self0), Pat::Range(other0)) => self0 == other0, ++ (Pat::Reference(self0), Pat::Reference(other0)) => self0 == other0, ++ (Pat::Rest(self0), Pat::Rest(other0)) => self0 == other0, ++ (Pat::Slice(self0), Pat::Slice(other0)) => self0 == other0, ++ (Pat::Struct(self0), Pat::Struct(other0)) => self0 == other0, ++ (Pat::Tuple(self0), Pat::Tuple(other0)) => self0 == other0, ++ (Pat::TupleStruct(self0), Pat::TupleStruct(other0)) => self0 == other0, ++ (Pat::Type(self0), Pat::Type(other0)) => self0 == other0, ++ (Pat::Verbatim(self0), Pat::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ (Pat::Wild(self0), Pat::Wild(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatBox {} ++#[cfg(feature = "full")] ++impl PartialEq for PatBox { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.pat == other.pat ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatIdent {} ++#[cfg(feature = "full")] ++impl PartialEq for PatIdent { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.by_ref == other.by_ref ++ && self.mutability == other.mutability ++ && self.ident == other.ident ++ && self.subpat == other.subpat ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatLit {} ++#[cfg(feature = "full")] ++impl PartialEq for PatLit { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.expr == other.expr ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for PatMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mac == other.mac ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatOr {} ++#[cfg(feature = "full")] ++impl PartialEq for PatOr { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.leading_vert == other.leading_vert ++ && self.cases == other.cases ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatPath {} ++#[cfg(feature = "full")] ++impl PartialEq for PatPath { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.qself == other.qself && self.path == other.path ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatRange {} ++#[cfg(feature = "full")] ++impl PartialEq for PatRange { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.lo == other.lo ++ && self.limits == other.limits ++ && self.hi == other.hi ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatReference {} ++#[cfg(feature = "full")] ++impl PartialEq for PatReference { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mutability == other.mutability && self.pat == other.pat ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatRest {} ++#[cfg(feature = "full")] ++impl PartialEq for PatRest { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatSlice {} ++#[cfg(feature = "full")] ++impl PartialEq for PatSlice { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.elems == other.elems ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatStruct {} ++#[cfg(feature = "full")] ++impl PartialEq for PatStruct { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.path == other.path ++ && self.fields == other.fields ++ && self.dot2_token == other.dot2_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatTuple {} ++#[cfg(feature = "full")] ++impl PartialEq for PatTuple { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.elems == other.elems ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatTupleStruct {} ++#[cfg(feature = "full")] ++impl PartialEq for PatTupleStruct { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.path == other.path && self.pat == other.pat ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatType {} ++#[cfg(feature = "full")] ++impl PartialEq for PatType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.pat == other.pat && self.ty == other.ty ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for PatWild {} ++#[cfg(feature = "full")] ++impl PartialEq for PatWild { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Path {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Path { ++ fn eq(&self, other: &Self) -> bool { ++ self.leading_colon == other.leading_colon && self.segments == other.segments ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for PathArguments {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for PathArguments { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (PathArguments::None, PathArguments::None) => true, ++ (PathArguments::AngleBracketed(self0), PathArguments::AngleBracketed(other0)) => { ++ self0 == other0 ++ } ++ (PathArguments::Parenthesized(self0), PathArguments::Parenthesized(other0)) => { ++ self0 == other0 ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for PathSegment {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for PathSegment { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident && self.arguments == other.arguments ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for PredicateEq {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for PredicateEq { ++ fn eq(&self, other: &Self) -> bool { ++ self.lhs_ty == other.lhs_ty && self.rhs_ty == other.rhs_ty ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for PredicateLifetime {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for PredicateLifetime { ++ fn eq(&self, other: &Self) -> bool { ++ self.lifetime == other.lifetime && self.bounds == other.bounds ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for PredicateType {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for PredicateType { ++ fn eq(&self, other: &Self) -> bool { ++ self.lifetimes == other.lifetimes ++ && self.bounded_ty == other.bounded_ty ++ && self.bounds == other.bounds ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for QSelf {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for QSelf { ++ fn eq(&self, other: &Self) -> bool { ++ self.ty == other.ty && self.position == other.position && self.as_token == other.as_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for RangeLimits {} ++#[cfg(feature = "full")] ++impl PartialEq for RangeLimits { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (RangeLimits::HalfOpen(_), RangeLimits::HalfOpen(_)) => true, ++ (RangeLimits::Closed(_), RangeLimits::Closed(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Receiver {} ++#[cfg(feature = "full")] ++impl PartialEq for Receiver { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.reference == other.reference ++ && self.mutability == other.mutability ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for ReturnType {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for ReturnType { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (ReturnType::Default, ReturnType::Default) => true, ++ (ReturnType::Type(_, self1), ReturnType::Type(_, other1)) => self1 == other1, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Signature {} ++#[cfg(feature = "full")] ++impl PartialEq for Signature { ++ fn eq(&self, other: &Self) -> bool { ++ self.constness == other.constness ++ && self.asyncness == other.asyncness ++ && self.unsafety == other.unsafety ++ && self.abi == other.abi ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.inputs == other.inputs ++ && self.variadic == other.variadic ++ && self.output == other.output ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for Stmt {} ++#[cfg(feature = "full")] ++impl PartialEq for Stmt { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Stmt::Local(self0), Stmt::Local(other0)) => self0 == other0, ++ (Stmt::Item(self0), Stmt::Item(other0)) => self0 == other0, ++ (Stmt::Expr(self0), Stmt::Expr(other0)) => self0 == other0, ++ (Stmt::Semi(self0, _), Stmt::Semi(other0, _)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TraitBound {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TraitBound { ++ fn eq(&self, other: &Self) -> bool { ++ self.paren_token == other.paren_token ++ && self.modifier == other.modifier ++ && self.lifetimes == other.lifetimes ++ && self.path == other.path ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TraitBoundModifier {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TraitBoundModifier { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (TraitBoundModifier::None, TraitBoundModifier::None) => true, ++ (TraitBoundModifier::Maybe(_), TraitBoundModifier::Maybe(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for TraitItem {} ++#[cfg(feature = "full")] ++impl PartialEq for TraitItem { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (TraitItem::Const(self0), TraitItem::Const(other0)) => self0 == other0, ++ (TraitItem::Method(self0), TraitItem::Method(other0)) => self0 == other0, ++ (TraitItem::Type(self0), TraitItem::Type(other0)) => self0 == other0, ++ (TraitItem::Macro(self0), TraitItem::Macro(other0)) => self0 == other0, ++ (TraitItem::Verbatim(self0), TraitItem::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for TraitItemConst {} ++#[cfg(feature = "full")] ++impl PartialEq for TraitItemConst { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.ty == other.ty ++ && self.default == other.default ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for TraitItemMacro {} ++#[cfg(feature = "full")] ++impl PartialEq for TraitItemMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs && self.mac == other.mac && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for TraitItemMethod {} ++#[cfg(feature = "full")] ++impl PartialEq for TraitItemMethod { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.sig == other.sig ++ && self.default == other.default ++ && self.semi_token == other.semi_token ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for TraitItemType {} ++#[cfg(feature = "full")] ++impl PartialEq for TraitItemType { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.generics == other.generics ++ && self.colon_token == other.colon_token ++ && self.bounds == other.bounds ++ && self.default == other.default ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Type {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Type { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Type::Array(self0), Type::Array(other0)) => self0 == other0, ++ (Type::BareFn(self0), Type::BareFn(other0)) => self0 == other0, ++ (Type::Group(self0), Type::Group(other0)) => self0 == other0, ++ (Type::ImplTrait(self0), Type::ImplTrait(other0)) => self0 == other0, ++ (Type::Infer(self0), Type::Infer(other0)) => self0 == other0, ++ (Type::Macro(self0), Type::Macro(other0)) => self0 == other0, ++ (Type::Never(self0), Type::Never(other0)) => self0 == other0, ++ (Type::Paren(self0), Type::Paren(other0)) => self0 == other0, ++ (Type::Path(self0), Type::Path(other0)) => self0 == other0, ++ (Type::Ptr(self0), Type::Ptr(other0)) => self0 == other0, ++ (Type::Reference(self0), Type::Reference(other0)) => self0 == other0, ++ (Type::Slice(self0), Type::Slice(other0)) => self0 == other0, ++ (Type::TraitObject(self0), Type::TraitObject(other0)) => self0 == other0, ++ (Type::Tuple(self0), Type::Tuple(other0)) => self0 == other0, ++ (Type::Verbatim(self0), Type::Verbatim(other0)) => { ++ TokenStreamHelper(self0) == TokenStreamHelper(other0) ++ } ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeArray {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeArray { ++ fn eq(&self, other: &Self) -> bool { ++ self.elem == other.elem && self.len == other.len ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeBareFn {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeBareFn { ++ fn eq(&self, other: &Self) -> bool { ++ self.lifetimes == other.lifetimes ++ && self.unsafety == other.unsafety ++ && self.abi == other.abi ++ && self.inputs == other.inputs ++ && self.variadic == other.variadic ++ && self.output == other.output ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeGroup {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeGroup { ++ fn eq(&self, other: &Self) -> bool { ++ self.elem == other.elem ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeImplTrait {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeImplTrait { ++ fn eq(&self, other: &Self) -> bool { ++ self.bounds == other.bounds ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeInfer {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeInfer { ++ fn eq(&self, _other: &Self) -> bool { ++ true ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeMacro {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeMacro { ++ fn eq(&self, other: &Self) -> bool { ++ self.mac == other.mac ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeNever {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeNever { ++ fn eq(&self, _other: &Self) -> bool { ++ true ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeParam {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeParam { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.colon_token == other.colon_token ++ && self.bounds == other.bounds ++ && self.eq_token == other.eq_token ++ && self.default == other.default ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeParamBound {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeParamBound { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (TypeParamBound::Trait(self0), TypeParamBound::Trait(other0)) => self0 == other0, ++ (TypeParamBound::Lifetime(self0), TypeParamBound::Lifetime(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeParen {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeParen { ++ fn eq(&self, other: &Self) -> bool { ++ self.elem == other.elem ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypePath {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypePath { ++ fn eq(&self, other: &Self) -> bool { ++ self.qself == other.qself && self.path == other.path ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypePtr {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypePtr { ++ fn eq(&self, other: &Self) -> bool { ++ self.const_token == other.const_token ++ && self.mutability == other.mutability ++ && self.elem == other.elem ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeReference {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeReference { ++ fn eq(&self, other: &Self) -> bool { ++ self.lifetime == other.lifetime ++ && self.mutability == other.mutability ++ && self.elem == other.elem ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeSlice {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeSlice { ++ fn eq(&self, other: &Self) -> bool { ++ self.elem == other.elem ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeTraitObject {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeTraitObject { ++ fn eq(&self, other: &Self) -> bool { ++ self.dyn_token == other.dyn_token && self.bounds == other.bounds ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for TypeTuple {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for TypeTuple { ++ fn eq(&self, other: &Self) -> bool { ++ self.elems == other.elems ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for UnOp {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for UnOp { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (UnOp::Deref(_), UnOp::Deref(_)) => true, ++ (UnOp::Not(_), UnOp::Not(_)) => true, ++ (UnOp::Neg(_), UnOp::Neg(_)) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UseGlob {} ++#[cfg(feature = "full")] ++impl PartialEq for UseGlob { ++ fn eq(&self, _other: &Self) -> bool { ++ true ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UseGroup {} ++#[cfg(feature = "full")] ++impl PartialEq for UseGroup { ++ fn eq(&self, other: &Self) -> bool { ++ self.items == other.items ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UseName {} ++#[cfg(feature = "full")] ++impl PartialEq for UseName { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UsePath {} ++#[cfg(feature = "full")] ++impl PartialEq for UsePath { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident && self.tree == other.tree ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UseRename {} ++#[cfg(feature = "full")] ++impl PartialEq for UseRename { ++ fn eq(&self, other: &Self) -> bool { ++ self.ident == other.ident && self.rename == other.rename ++ } ++} ++#[cfg(feature = "full")] ++impl Eq for UseTree {} ++#[cfg(feature = "full")] ++impl PartialEq for UseTree { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (UseTree::Path(self0), UseTree::Path(other0)) => self0 == other0, ++ (UseTree::Name(self0), UseTree::Name(other0)) => self0 == other0, ++ (UseTree::Rename(self0), UseTree::Rename(other0)) => self0 == other0, ++ (UseTree::Glob(self0), UseTree::Glob(other0)) => self0 == other0, ++ (UseTree::Group(self0), UseTree::Group(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Variadic {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Variadic { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Variant {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Variant { ++ fn eq(&self, other: &Self) -> bool { ++ self.attrs == other.attrs ++ && self.ident == other.ident ++ && self.fields == other.fields ++ && self.discriminant == other.discriminant ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for VisCrate {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for VisCrate { ++ fn eq(&self, _other: &Self) -> bool { ++ true ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for VisPublic {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for VisPublic { ++ fn eq(&self, _other: &Self) -> bool { ++ true ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for VisRestricted {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for VisRestricted { ++ fn eq(&self, other: &Self) -> bool { ++ self.in_token == other.in_token && self.path == other.path ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for Visibility {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for Visibility { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (Visibility::Public(self0), Visibility::Public(other0)) => self0 == other0, ++ (Visibility::Crate(self0), Visibility::Crate(other0)) => self0 == other0, ++ (Visibility::Restricted(self0), Visibility::Restricted(other0)) => self0 == other0, ++ (Visibility::Inherited, Visibility::Inherited) => true, ++ _ => false, ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for WhereClause {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for WhereClause { ++ fn eq(&self, other: &Self) -> bool { ++ self.predicates == other.predicates ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Eq for WherePredicate {} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl PartialEq for WherePredicate { ++ fn eq(&self, other: &Self) -> bool { ++ match (self, other) { ++ (WherePredicate::Type(self0), WherePredicate::Type(other0)) => self0 == other0, ++ (WherePredicate::Lifetime(self0), WherePredicate::Lifetime(other0)) => self0 == other0, ++ (WherePredicate::Eq(self0), WherePredicate::Eq(other0)) => self0 == other0, ++ _ => false, ++ } ++ } ++} +diff --git a/third_party/rust/syn/src/gen/fold.rs b/third_party/rust/syn/src/gen/fold.rs +index f51218b78c..d9dd32a420 100644 +--- a/third_party/rust/syn/src/gen/fold.rs ++++ b/third_party/rust/syn/src/gen/fold.rs +@@ -2,6 +2,7 @@ + // It is not intended for manual editing. + + #![allow(unreachable_code, unused_variables)] ++#![allow(clippy::match_wildcard_for_single_variants)] + #[cfg(any(feature = "full", feature = "derive"))] + use crate::gen::helper::fold::*; + #[cfg(any(feature = "full", feature = "derive"))] +@@ -26,7 +27,7 @@ macro_rules! full { + /// + /// [module documentation]: self + /// +-/// *This trait is available if Syn is built with the `"fold"` feature.* ++/// *This trait is available only if Syn is built with the `"fold"` feature.* + pub trait Fold { + #[cfg(any(feature = "derive", feature = "full"))] + fn fold_abi(&mut self, i: Abi) -> Abi { +@@ -433,35 +434,27 @@ pub trait Fold { + fn fold_lifetime_def(&mut self, i: LifetimeDef) -> LifetimeDef { + fold_lifetime_def(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit(&mut self, i: Lit) -> Lit { + fold_lit(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_bool(&mut self, i: LitBool) -> LitBool { + fold_lit_bool(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_byte(&mut self, i: LitByte) -> LitByte { + fold_lit_byte(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_byte_str(&mut self, i: LitByteStr) -> LitByteStr { + fold_lit_byte_str(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_char(&mut self, i: LitChar) -> LitChar { + fold_lit_char(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_float(&mut self, i: LitFloat) -> LitFloat { + fold_lit_float(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_int(&mut self, i: LitInt) -> LitInt { + fold_lit_int(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn fold_lit_str(&mut self, i: LitStr) -> LitStr { + fold_lit_str(self, i) + } +@@ -799,10 +792,10 @@ where + F: Fold + ?Sized, + { + AngleBracketedGenericArguments { +- colon2_token: (node.colon2_token).map(|it| Token ! [ :: ](tokens_helper(f, &it.spans))), +- lt_token: Token ! [ < ](tokens_helper(f, &node.lt_token.spans)), ++ colon2_token: (node.colon2_token).map(|it| Token ! [::](tokens_helper(f, &it.spans))), ++ lt_token: Token ! [<](tokens_helper(f, &node.lt_token.spans)), + args: FoldHelper::lift(node.args, |it| f.fold_generic_argument(it)), +- gt_token: Token ! [ > ](tokens_helper(f, &node.gt_token.spans)), ++ gt_token: Token ! [>](tokens_helper(f, &node.gt_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -819,9 +812,9 @@ where + Box::new(f.fold_expr(*(it).1)), + ) + }), +- fat_arrow_token: Token ! [ => ](tokens_helper(f, &node.fat_arrow_token.spans)), ++ fat_arrow_token: Token ! [=>](tokens_helper(f, &node.fat_arrow_token.spans)), + body: Box::new(f.fold_expr(*node.body)), +- comma: (node.comma).map(|it| Token ! [ , ](tokens_helper(f, &it.spans))), ++ comma: (node.comma).map(|it| Token ! [,](tokens_helper(f, &it.spans))), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -842,7 +835,7 @@ where + F: Fold + ?Sized, + { + Attribute { +- pound_token: Token ! [ # ](tokens_helper(f, &node.pound_token.spans)), ++ pound_token: Token ! [#](tokens_helper(f, &node.pound_token.spans)), + style: f.fold_attr_style(node.style), + bracket_token: Bracket(tokens_helper(f, &node.bracket_token.span)), + path: f.fold_path(node.path), +@@ -859,7 +852,7 @@ where + name: (node.name).map(|it| { + ( + f.fold_ident((it).0), +- Token ! [ : ](tokens_helper(f, &(it).1.spans)), ++ Token ! [:](tokens_helper(f, &(it).1.spans)), + ) + }), + ty: f.fold_type(node.ty), +@@ -871,59 +864,47 @@ where + F: Fold + ?Sized, + { + match node { +- BinOp::Add(_binding_0) => BinOp::Add(Token ! [ + ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Sub(_binding_0) => BinOp::Sub(Token ! [ - ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Mul(_binding_0) => BinOp::Mul(Token ! [ * ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Div(_binding_0) => BinOp::Div(Token ! [ / ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Rem(_binding_0) => BinOp::Rem(Token ! [ % ](tokens_helper(f, &_binding_0.spans))), +- BinOp::And(_binding_0) => BinOp::And(Token ! [ && ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Or(_binding_0) => BinOp::Or(Token ! [ || ](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Add(_binding_0) => BinOp::Add(Token ! [+](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Sub(_binding_0) => BinOp::Sub(Token ! [-](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Mul(_binding_0) => BinOp::Mul(Token ! [*](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Div(_binding_0) => BinOp::Div(Token ! [/](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Rem(_binding_0) => BinOp::Rem(Token ! [%](tokens_helper(f, &_binding_0.spans))), ++ BinOp::And(_binding_0) => BinOp::And(Token ! [&&](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Or(_binding_0) => BinOp::Or(Token ! [||](tokens_helper(f, &_binding_0.spans))), + BinOp::BitXor(_binding_0) => { +- BinOp::BitXor(Token ! [ ^ ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::BitXor(Token ! [^](tokens_helper(f, &_binding_0.spans))) + } + BinOp::BitAnd(_binding_0) => { +- BinOp::BitAnd(Token ! [ & ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::BitOr(_binding_0) => { +- BinOp::BitOr(Token ! [ | ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::Shl(_binding_0) => BinOp::Shl(Token ! [ << ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Shr(_binding_0) => BinOp::Shr(Token ! [ >> ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Eq(_binding_0) => BinOp::Eq(Token ! [ == ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Lt(_binding_0) => BinOp::Lt(Token ! [ < ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Le(_binding_0) => BinOp::Le(Token ! [ <= ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Ne(_binding_0) => BinOp::Ne(Token ! [ != ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Ge(_binding_0) => BinOp::Ge(Token ! [ >= ](tokens_helper(f, &_binding_0.spans))), +- BinOp::Gt(_binding_0) => BinOp::Gt(Token ! [ > ](tokens_helper(f, &_binding_0.spans))), +- BinOp::AddEq(_binding_0) => { +- BinOp::AddEq(Token ! [ += ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::SubEq(_binding_0) => { +- BinOp::SubEq(Token ! [ -= ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::MulEq(_binding_0) => { +- BinOp::MulEq(Token ! [ *= ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::DivEq(_binding_0) => { +- BinOp::DivEq(Token ! [ /= ](tokens_helper(f, &_binding_0.spans))) +- } +- BinOp::RemEq(_binding_0) => { +- BinOp::RemEq(Token ! [ %= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::BitAnd(Token ! [&](tokens_helper(f, &_binding_0.spans))) + } ++ BinOp::BitOr(_binding_0) => BinOp::BitOr(Token ! [|](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Shl(_binding_0) => BinOp::Shl(Token ! [<<](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Shr(_binding_0) => BinOp::Shr(Token ! [>>](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Eq(_binding_0) => BinOp::Eq(Token ! [==](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Lt(_binding_0) => BinOp::Lt(Token ! [<](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Le(_binding_0) => BinOp::Le(Token ! [<=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Ne(_binding_0) => BinOp::Ne(Token ! [!=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Ge(_binding_0) => BinOp::Ge(Token ! [>=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::Gt(_binding_0) => BinOp::Gt(Token ! [>](tokens_helper(f, &_binding_0.spans))), ++ BinOp::AddEq(_binding_0) => BinOp::AddEq(Token ! [+=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::SubEq(_binding_0) => BinOp::SubEq(Token ! [-=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::MulEq(_binding_0) => BinOp::MulEq(Token ! [*=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::DivEq(_binding_0) => BinOp::DivEq(Token ! [/=](tokens_helper(f, &_binding_0.spans))), ++ BinOp::RemEq(_binding_0) => BinOp::RemEq(Token ! [%=](tokens_helper(f, &_binding_0.spans))), + BinOp::BitXorEq(_binding_0) => { +- BinOp::BitXorEq(Token ! [ ^= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::BitXorEq(Token ! [^=](tokens_helper(f, &_binding_0.spans))) + } + BinOp::BitAndEq(_binding_0) => { +- BinOp::BitAndEq(Token ! [ &= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::BitAndEq(Token ! [&=](tokens_helper(f, &_binding_0.spans))) + } + BinOp::BitOrEq(_binding_0) => { +- BinOp::BitOrEq(Token ! [ |= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::BitOrEq(Token ! [|=](tokens_helper(f, &_binding_0.spans))) + } + BinOp::ShlEq(_binding_0) => { +- BinOp::ShlEq(Token ! [ <<= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::ShlEq(Token ! [<<=](tokens_helper(f, &_binding_0.spans))) + } + BinOp::ShrEq(_binding_0) => { +- BinOp::ShrEq(Token ! [ >>= ](tokens_helper(f, &_binding_0.spans))) ++ BinOp::ShrEq(Token ! [>>=](tokens_helper(f, &_binding_0.spans))) + } + } + } +@@ -934,7 +915,7 @@ where + { + Binding { + ident: f.fold_ident(node.ident), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + ty: f.fold_type(node.ty), + } + } +@@ -955,9 +936,9 @@ where + { + BoundLifetimes { + for_token: Token![for](tokens_helper(f, &node.for_token.span)), +- lt_token: Token ! [ < ](tokens_helper(f, &node.lt_token.spans)), ++ lt_token: Token ! [<](tokens_helper(f, &node.lt_token.spans)), + lifetimes: FoldHelper::lift(node.lifetimes, |it| f.fold_lifetime_def(it)), +- gt_token: Token ! [ > ](tokens_helper(f, &node.gt_token.spans)), ++ gt_token: Token ! [>](tokens_helper(f, &node.gt_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -969,9 +950,9 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + const_token: Token![const](tokens_helper(f, &node.const_token.span)), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: f.fold_type(node.ty), +- eq_token: (node.eq_token).map(|it| Token ! [ = ](tokens_helper(f, &it.spans))), ++ eq_token: (node.eq_token).map(|it| Token ! [=](tokens_helper(f, &it.spans))), + default: (node.default).map(|it| f.fold_expr(it)), + } + } +@@ -982,7 +963,7 @@ where + { + Constraint { + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_type_param_bound(it)), + } + } +@@ -1016,7 +997,7 @@ where + DataStruct { + struct_token: Token![struct](tokens_helper(f, &node.struct_token.span)), + fields: f.fold_fields(node.fields), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "derive")] +@@ -1112,7 +1093,7 @@ where + ExprAssign { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + left: Box::new(f.fold_expr(*node.left)), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + right: Box::new(f.fold_expr(*node.right)), + } + } +@@ -1148,7 +1129,7 @@ where + ExprAwait { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + base: Box::new(f.fold_expr(*node.base)), +- dot_token: Token ! [ . ](tokens_helper(f, &node.dot_token.spans)), ++ dot_token: Token ! [.](tokens_helper(f, &node.dot_token.spans)), + await_token: crate::token::Await(tokens_helper(f, &node.await_token.span)), + } + } +@@ -1232,9 +1213,9 @@ where + asyncness: (node.asyncness).map(|it| Token![async](tokens_helper(f, &it.span))), + movability: (node.movability).map(|it| Token![static](tokens_helper(f, &it.span))), + capture: (node.capture).map(|it| Token![move](tokens_helper(f, &it.span))), +- or1_token: Token ! [ | ](tokens_helper(f, &node.or1_token.spans)), ++ or1_token: Token ! [|](tokens_helper(f, &node.or1_token.spans)), + inputs: FoldHelper::lift(node.inputs, |it| f.fold_pat(it)), +- or2_token: Token ! [ | ](tokens_helper(f, &node.or2_token.spans)), ++ or2_token: Token ! [|](tokens_helper(f, &node.or2_token.spans)), + output: f.fold_return_type(node.output), + body: Box::new(f.fold_expr(*node.body)), + } +@@ -1258,7 +1239,7 @@ where + ExprField { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + base: Box::new(f.fold_expr(*node.base)), +- dot_token: Token ! [ . ](tokens_helper(f, &node.dot_token.spans)), ++ dot_token: Token ! [.](tokens_helper(f, &node.dot_token.spans)), + member: f.fold_member(node.member), + } + } +@@ -1327,7 +1308,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + let_token: Token![let](tokens_helper(f, &node.let_token.span)), + pat: f.fold_pat(node.pat), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + expr: Box::new(f.fold_expr(*node.expr)), + } + } +@@ -1384,7 +1365,7 @@ where + ExprMethodCall { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + receiver: Box::new(f.fold_expr(*node.receiver)), +- dot_token: Token ! [ . ](tokens_helper(f, &node.dot_token.spans)), ++ dot_token: Token ! [.](tokens_helper(f, &node.dot_token.spans)), + method: f.fold_ident(node.method), + turbofish: (node.turbofish).map(|it| f.fold_method_turbofish(it)), + paren_token: Paren(tokens_helper(f, &node.paren_token.span)), +@@ -1432,7 +1413,7 @@ where + { + ExprReference { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), +- and_token: Token ! [ & ](tokens_helper(f, &node.and_token.spans)), ++ and_token: Token ! [&](tokens_helper(f, &node.and_token.spans)), + raw: node.raw, + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + expr: Box::new(f.fold_expr(*node.expr)), +@@ -1447,7 +1428,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + bracket_token: Bracket(tokens_helper(f, &node.bracket_token.span)), + expr: Box::new(f.fold_expr(*node.expr)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + len: Box::new(f.fold_expr(*node.len)), + } + } +@@ -1484,7 +1465,7 @@ where + ExprTry { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + expr: Box::new(f.fold_expr(*node.expr)), +- question_token: Token ! [ ? ](tokens_helper(f, &node.question_token.spans)), ++ question_token: Token ! [?](tokens_helper(f, &node.question_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -1517,7 +1498,7 @@ where + ExprType { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + expr: Box::new(f.fold_expr(*node.expr)), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), + } + } +@@ -1576,7 +1557,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + vis: f.fold_visibility(node.vis), + ident: (node.ident).map(|it| f.fold_ident(it)), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + ty: f.fold_type(node.ty), + } + } +@@ -1588,7 +1569,7 @@ where + FieldPat { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + member: f.fold_member(node.member), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + pat: Box::new(f.fold_pat(*node.pat)), + } + } +@@ -1600,7 +1581,7 @@ where + FieldValue { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + member: f.fold_member(node.member), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + expr: f.fold_expr(node.expr), + } + } +@@ -1681,7 +1662,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + vis: f.fold_visibility(node.vis), + sig: f.fold_signature(node.sig), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -1692,7 +1673,7 @@ where + ForeignItemMacro { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + mac: f.fold_macro(node.mac), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -1706,9 +1687,9 @@ where + static_token: Token![static](tokens_helper(f, &node.static_token.span)), + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -1721,7 +1702,7 @@ where + vis: f.fold_visibility(node.vis), + type_token: Token![type](tokens_helper(f, &node.type_token.span)), + ident: f.fold_ident(node.ident), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -1779,9 +1760,9 @@ where + F: Fold + ?Sized, + { + Generics { +- lt_token: (node.lt_token).map(|it| Token ! [ < ](tokens_helper(f, &it.spans))), ++ lt_token: (node.lt_token).map(|it| Token ! [<](tokens_helper(f, &it.spans))), + params: FoldHelper::lift(node.params, |it| f.fold_generic_param(it)), +- gt_token: (node.gt_token).map(|it| Token ! [ > ](tokens_helper(f, &it.spans))), ++ gt_token: (node.gt_token).map(|it| Token ! [>](tokens_helper(f, &it.spans))), + where_clause: (node.where_clause).map(|it| f.fold_where_clause(it)), + } + } +@@ -1819,11 +1800,11 @@ where + defaultness: (node.defaultness).map(|it| Token![default](tokens_helper(f, &it.span))), + const_token: Token![const](tokens_helper(f, &node.const_token.span)), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: f.fold_type(node.ty), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + expr: f.fold_expr(node.expr), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -1834,7 +1815,7 @@ where + ImplItemMacro { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + mac: f.fold_macro(node.mac), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -1862,9 +1843,9 @@ where + type_token: Token![type](tokens_helper(f, &node.type_token.span)), + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + ty: f.fold_type(node.ty), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -1913,11 +1894,11 @@ where + vis: f.fold_visibility(node.vis), + const_token: Token![const](tokens_helper(f, &node.const_token.span)), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + expr: Box::new(f.fold_expr(*node.expr)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -1952,7 +1933,7 @@ where + f.fold_ident((it).1), + ) + }), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2011,7 +1992,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + ident: (node.ident).map(|it| f.fold_ident(it)), + mac: f.fold_macro(node.mac), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -2043,7 +2024,7 @@ where + FoldHelper::lift((it).1, |it| f.fold_item(it)), + ) + }), +- semi: (node.semi).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi: (node.semi).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -2057,11 +2038,11 @@ where + static_token: Token![static](tokens_helper(f, &node.static_token.span)), + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + expr: Box::new(f.fold_expr(*node.expr)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2076,7 +2057,7 @@ where + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), + fields: f.fold_fields(node.fields), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -2092,7 +2073,7 @@ where + trait_token: Token![trait](tokens_helper(f, &node.trait_token.span)), + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + supertraits: FoldHelper::lift(node.supertraits, |it| f.fold_type_param_bound(it)), + brace_token: Brace(tokens_helper(f, &node.brace_token.span)), + items: FoldHelper::lift(node.items, |it| f.fold_trait_item(it)), +@@ -2109,9 +2090,9 @@ where + trait_token: Token![trait](tokens_helper(f, &node.trait_token.span)), + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_type_param_bound(it)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2125,9 +2106,9 @@ where + type_token: Token![type](tokens_helper(f, &node.type_token.span)), + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2153,9 +2134,9 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + vis: f.fold_visibility(node.vis), + use_token: Token![use](tokens_helper(f, &node.use_token.span)), +- leading_colon: (node.leading_colon).map(|it| Token ! [ :: ](tokens_helper(f, &it.spans))), ++ leading_colon: (node.leading_colon).map(|it| Token ! [::](tokens_helper(f, &it.spans))), + tree: f.fold_use_tree(node.tree), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2165,7 +2146,7 @@ where + { + Label { + name: f.fold_lifetime(node.name), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + } + } + pub fn fold_lifetime<F>(f: &mut F, node: Lifetime) -> Lifetime +@@ -2185,11 +2166,10 @@ where + LifetimeDef { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + lifetime: f.fold_lifetime(node.lifetime), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_lifetime(it)), + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit<F>(f: &mut F, node: Lit) -> Lit + where + F: Fold + ?Sized, +@@ -2205,7 +2185,6 @@ where + Lit::Verbatim(_binding_0) => Lit::Verbatim(_binding_0), + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_bool<F>(f: &mut F, node: LitBool) -> LitBool + where + F: Fold + ?Sized, +@@ -2215,7 +2194,6 @@ where + span: f.fold_span(node.span), + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_byte<F>(f: &mut F, node: LitByte) -> LitByte + where + F: Fold + ?Sized, +@@ -2225,7 +2203,6 @@ where + node.set_span(span); + node + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_byte_str<F>(f: &mut F, node: LitByteStr) -> LitByteStr + where + F: Fold + ?Sized, +@@ -2235,7 +2212,6 @@ where + node.set_span(span); + node + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_char<F>(f: &mut F, node: LitChar) -> LitChar + where + F: Fold + ?Sized, +@@ -2245,7 +2221,6 @@ where + node.set_span(span); + node + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_float<F>(f: &mut F, node: LitFloat) -> LitFloat + where + F: Fold + ?Sized, +@@ -2255,7 +2230,6 @@ where + node.set_span(span); + node + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_int<F>(f: &mut F, node: LitInt) -> LitInt + where + F: Fold + ?Sized, +@@ -2265,7 +2239,6 @@ where + node.set_span(span); + node + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn fold_lit_str<F>(f: &mut F, node: LitStr) -> LitStr + where + F: Fold + ?Sized, +@@ -2286,11 +2259,11 @@ where + pat: f.fold_pat(node.pat), + init: (node.init).map(|it| { + ( +- Token ! [ = ](tokens_helper(f, &(it).0.spans)), ++ Token ! [=](tokens_helper(f, &(it).0.spans)), + Box::new(f.fold_expr(*(it).1)), + ) + }), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -2361,7 +2334,7 @@ where + { + MetaNameValue { + path: f.fold_path(node.path), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + lit: f.fold_lit(node.lit), + } + } +@@ -2371,10 +2344,10 @@ where + F: Fold + ?Sized, + { + MethodTurbofish { +- colon2_token: Token ! [ :: ](tokens_helper(f, &node.colon2_token.spans)), +- lt_token: Token ! [ < ](tokens_helper(f, &node.lt_token.spans)), ++ colon2_token: Token ! [::](tokens_helper(f, &node.colon2_token.spans)), ++ lt_token: Token ! [<](tokens_helper(f, &node.lt_token.spans)), + args: FoldHelper::lift(node.args, |it| f.fold_generic_method_argument(it)), +- gt_token: Token ! [ > ](tokens_helper(f, &node.gt_token.spans)), ++ gt_token: Token ! [>](tokens_helper(f, &node.gt_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -2449,7 +2422,7 @@ where + ident: f.fold_ident(node.ident), + subpat: (node.subpat).map(|it| { + ( +- Token ! [ @ ](tokens_helper(f, &(it).0.spans)), ++ Token ! [@](tokens_helper(f, &(it).0.spans)), + Box::new(f.fold_pat(*(it).1)), + ) + }), +@@ -2482,7 +2455,7 @@ where + { + PatOr { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), +- leading_vert: (node.leading_vert).map(|it| Token ! [ | ](tokens_helper(f, &it.spans))), ++ leading_vert: (node.leading_vert).map(|it| Token ! [|](tokens_helper(f, &it.spans))), + cases: FoldHelper::lift(node.cases, |it| f.fold_pat(it)), + } + } +@@ -2516,7 +2489,7 @@ where + { + PatReference { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), +- and_token: Token ! [ & ](tokens_helper(f, &node.and_token.spans)), ++ and_token: Token ! [&](tokens_helper(f, &node.and_token.spans)), + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + pat: Box::new(f.fold_pat(*node.pat)), + } +@@ -2585,7 +2558,7 @@ where + PatType { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + pat: Box::new(f.fold_pat(*node.pat)), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), + } + } +@@ -2605,7 +2578,7 @@ where + F: Fold + ?Sized, + { + Path { +- leading_colon: (node.leading_colon).map(|it| Token ! [ :: ](tokens_helper(f, &it.spans))), ++ leading_colon: (node.leading_colon).map(|it| Token ! [::](tokens_helper(f, &it.spans))), + segments: FoldHelper::lift(node.segments, |it| f.fold_path_segment(it)), + } + } +@@ -2641,7 +2614,7 @@ where + { + PredicateEq { + lhs_ty: f.fold_type(node.lhs_ty), +- eq_token: Token ! [ = ](tokens_helper(f, &node.eq_token.spans)), ++ eq_token: Token ! [=](tokens_helper(f, &node.eq_token.spans)), + rhs_ty: f.fold_type(node.rhs_ty), + } + } +@@ -2652,7 +2625,7 @@ where + { + PredicateLifetime { + lifetime: f.fold_lifetime(node.lifetime), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_lifetime(it)), + } + } +@@ -2664,7 +2637,7 @@ where + PredicateType { + lifetimes: (node.lifetimes).map(|it| f.fold_bound_lifetimes(it)), + bounded_ty: f.fold_type(node.bounded_ty), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_type_param_bound(it)), + } + } +@@ -2674,11 +2647,11 @@ where + F: Fold + ?Sized, + { + QSelf { +- lt_token: Token ! [ < ](tokens_helper(f, &node.lt_token.spans)), ++ lt_token: Token ! [<](tokens_helper(f, &node.lt_token.spans)), + ty: Box::new(f.fold_type(*node.ty)), + position: node.position, + as_token: (node.as_token).map(|it| Token![as](tokens_helper(f, &it.span))), +- gt_token: Token ! [ > ](tokens_helper(f, &node.gt_token.spans)), ++ gt_token: Token ! [>](tokens_helper(f, &node.gt_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2691,7 +2664,7 @@ where + RangeLimits::HalfOpen(Token![..](tokens_helper(f, &_binding_0.spans))) + } + RangeLimits::Closed(_binding_0) => { +- RangeLimits::Closed(Token ! [ ..= ](tokens_helper(f, &_binding_0.spans))) ++ RangeLimits::Closed(Token ! [..=](tokens_helper(f, &_binding_0.spans))) + } + } + } +@@ -2704,7 +2677,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + reference: (node.reference).map(|it| { + ( +- Token ! [ & ](tokens_helper(f, &(it).0.spans)), ++ Token ! [&](tokens_helper(f, &(it).0.spans)), + ((it).1).map(|it| f.fold_lifetime(it)), + ) + }), +@@ -2720,7 +2693,7 @@ where + match node { + ReturnType::Default => ReturnType::Default, + ReturnType::Type(_binding_0, _binding_1) => ReturnType::Type( +- Token ! [ -> ](tokens_helper(f, &_binding_0.spans)), ++ Token ! [->](tokens_helper(f, &_binding_0.spans)), + Box::new(f.fold_type(*_binding_1)), + ), + } +@@ -2761,7 +2734,7 @@ where + Stmt::Expr(_binding_0) => Stmt::Expr(f.fold_expr(_binding_0)), + Stmt::Semi(_binding_0, _binding_1) => Stmt::Semi( + f.fold_expr(_binding_0), +- Token ! [ ; ](tokens_helper(f, &_binding_1.spans)), ++ Token ! [;](tokens_helper(f, &_binding_1.spans)), + ), + } + } +@@ -2785,7 +2758,7 @@ where + match node { + TraitBoundModifier::None => TraitBoundModifier::None, + TraitBoundModifier::Maybe(_binding_0) => { +- TraitBoundModifier::Maybe(Token ! [ ? ](tokens_helper(f, &_binding_0.spans))) ++ TraitBoundModifier::Maybe(Token ! [?](tokens_helper(f, &_binding_0.spans))) + } + } + } +@@ -2812,15 +2785,15 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + const_token: Token![const](tokens_helper(f, &node.const_token.span)), + ident: f.fold_ident(node.ident), +- colon_token: Token ! [ : ](tokens_helper(f, &node.colon_token.spans)), ++ colon_token: Token ! [:](tokens_helper(f, &node.colon_token.spans)), + ty: f.fold_type(node.ty), + default: (node.default).map(|it| { + ( +- Token ! [ = ](tokens_helper(f, &(it).0.spans)), ++ Token ! [=](tokens_helper(f, &(it).0.spans)), + f.fold_expr((it).1), + ) + }), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -2831,7 +2804,7 @@ where + TraitItemMacro { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + mac: f.fold_macro(node.mac), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -2843,7 +2816,7 @@ where + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + sig: f.fold_signature(node.sig), + default: (node.default).map(|it| f.fold_block(it)), +- semi_token: (node.semi_token).map(|it| Token ! [ ; ](tokens_helper(f, &it.spans))), ++ semi_token: (node.semi_token).map(|it| Token ! [;](tokens_helper(f, &it.spans))), + } + } + #[cfg(feature = "full")] +@@ -2856,15 +2829,15 @@ where + type_token: Token![type](tokens_helper(f, &node.type_token.span)), + ident: f.fold_ident(node.ident), + generics: f.fold_generics(node.generics), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_type_param_bound(it)), + default: (node.default).map(|it| { + ( +- Token ! [ = ](tokens_helper(f, &(it).0.spans)), ++ Token ! [=](tokens_helper(f, &(it).0.spans)), + f.fold_type((it).1), + ) + }), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -2899,7 +2872,7 @@ where + TypeArray { + bracket_token: Bracket(tokens_helper(f, &node.bracket_token.span)), + elem: Box::new(f.fold_type(*node.elem)), +- semi_token: Token ! [ ; ](tokens_helper(f, &node.semi_token.spans)), ++ semi_token: Token ! [;](tokens_helper(f, &node.semi_token.spans)), + len: f.fold_expr(node.len), + } + } +@@ -2974,9 +2947,9 @@ where + TypeParam { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), + ident: f.fold_ident(node.ident), +- colon_token: (node.colon_token).map(|it| Token ! [ : ](tokens_helper(f, &it.spans))), ++ colon_token: (node.colon_token).map(|it| Token ! [:](tokens_helper(f, &it.spans))), + bounds: FoldHelper::lift(node.bounds, |it| f.fold_type_param_bound(it)), +- eq_token: (node.eq_token).map(|it| Token ! [ = ](tokens_helper(f, &it.spans))), ++ eq_token: (node.eq_token).map(|it| Token ! [=](tokens_helper(f, &it.spans))), + default: (node.default).map(|it| f.fold_type(it)), + } + } +@@ -3018,7 +2991,7 @@ where + F: Fold + ?Sized, + { + TypePtr { +- star_token: Token ! [ * ](tokens_helper(f, &node.star_token.spans)), ++ star_token: Token ! [*](tokens_helper(f, &node.star_token.spans)), + const_token: (node.const_token).map(|it| Token![const](tokens_helper(f, &it.span))), + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + elem: Box::new(f.fold_type(*node.elem)), +@@ -3030,7 +3003,7 @@ where + F: Fold + ?Sized, + { + TypeReference { +- and_token: Token ! [ & ](tokens_helper(f, &node.and_token.spans)), ++ and_token: Token ! [&](tokens_helper(f, &node.and_token.spans)), + lifetime: (node.lifetime).map(|it| f.fold_lifetime(it)), + mutability: (node.mutability).map(|it| Token![mut](tokens_helper(f, &it.span))), + elem: Box::new(f.fold_type(*node.elem)), +@@ -3072,9 +3045,9 @@ where + F: Fold + ?Sized, + { + match node { +- UnOp::Deref(_binding_0) => UnOp::Deref(Token ! [ * ](tokens_helper(f, &_binding_0.spans))), ++ UnOp::Deref(_binding_0) => UnOp::Deref(Token ! [*](tokens_helper(f, &_binding_0.spans))), + UnOp::Not(_binding_0) => UnOp::Not(Token![!](tokens_helper(f, &_binding_0.spans))), +- UnOp::Neg(_binding_0) => UnOp::Neg(Token ! [ - ](tokens_helper(f, &_binding_0.spans))), ++ UnOp::Neg(_binding_0) => UnOp::Neg(Token ! [-](tokens_helper(f, &_binding_0.spans))), + } + } + #[cfg(feature = "full")] +@@ -3083,7 +3056,7 @@ where + F: Fold + ?Sized, + { + UseGlob { +- star_token: Token ! [ * ](tokens_helper(f, &node.star_token.spans)), ++ star_token: Token ! [*](tokens_helper(f, &node.star_token.spans)), + } + } + #[cfg(feature = "full")] +@@ -3112,7 +3085,7 @@ where + { + UsePath { + ident: f.fold_ident(node.ident), +- colon2_token: Token ! [ :: ](tokens_helper(f, &node.colon2_token.spans)), ++ colon2_token: Token ! [::](tokens_helper(f, &node.colon2_token.spans)), + tree: Box::new(f.fold_use_tree(*node.tree)), + } + } +@@ -3147,7 +3120,7 @@ where + { + Variadic { + attrs: FoldHelper::lift(node.attrs, |it| f.fold_attribute(it)), +- dots: Token ! [ ... ](tokens_helper(f, &node.dots.spans)), ++ dots: Token ! [...](tokens_helper(f, &node.dots.spans)), + } + } + #[cfg(any(feature = "derive", feature = "full"))] +@@ -3161,7 +3134,7 @@ where + fields: f.fold_fields(node.fields), + discriminant: (node.discriminant).map(|it| { + ( +- Token ! [ = ](tokens_helper(f, &(it).0.spans)), ++ Token ! [=](tokens_helper(f, &(it).0.spans)), + f.fold_expr((it).1), + ) + }), +diff --git a/third_party/rust/syn/src/gen/hash.rs b/third_party/rust/syn/src/gen/hash.rs +new file mode 100644 +index 0000000000..9e9e84a7af +--- /dev/null ++++ b/third_party/rust/syn/src/gen/hash.rs +@@ -0,0 +1,2691 @@ ++// This file is @generated by syn-internal-codegen. ++// It is not intended for manual editing. ++ ++#[cfg(any(feature = "derive", feature = "full"))] ++use crate::tt::TokenStreamHelper; ++use crate::*; ++use std::hash::{Hash, Hasher}; ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Abi { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.name.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for AngleBracketedGenericArguments { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.colon2_token.hash(state); ++ self.args.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Arm { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.pat.hash(state); ++ self.guard.hash(state); ++ self.body.hash(state); ++ self.comma.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for AttrStyle { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ AttrStyle::Outer => { ++ state.write_u8(0u8); ++ } ++ AttrStyle::Inner(_) => { ++ state.write_u8(1u8); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Attribute { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.style.hash(state); ++ self.path.hash(state); ++ TokenStreamHelper(&self.tokens).hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for BareFnArg { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.name.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for BinOp { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ BinOp::Add(_) => { ++ state.write_u8(0u8); ++ } ++ BinOp::Sub(_) => { ++ state.write_u8(1u8); ++ } ++ BinOp::Mul(_) => { ++ state.write_u8(2u8); ++ } ++ BinOp::Div(_) => { ++ state.write_u8(3u8); ++ } ++ BinOp::Rem(_) => { ++ state.write_u8(4u8); ++ } ++ BinOp::And(_) => { ++ state.write_u8(5u8); ++ } ++ BinOp::Or(_) => { ++ state.write_u8(6u8); ++ } ++ BinOp::BitXor(_) => { ++ state.write_u8(7u8); ++ } ++ BinOp::BitAnd(_) => { ++ state.write_u8(8u8); ++ } ++ BinOp::BitOr(_) => { ++ state.write_u8(9u8); ++ } ++ BinOp::Shl(_) => { ++ state.write_u8(10u8); ++ } ++ BinOp::Shr(_) => { ++ state.write_u8(11u8); ++ } ++ BinOp::Eq(_) => { ++ state.write_u8(12u8); ++ } ++ BinOp::Lt(_) => { ++ state.write_u8(13u8); ++ } ++ BinOp::Le(_) => { ++ state.write_u8(14u8); ++ } ++ BinOp::Ne(_) => { ++ state.write_u8(15u8); ++ } ++ BinOp::Ge(_) => { ++ state.write_u8(16u8); ++ } ++ BinOp::Gt(_) => { ++ state.write_u8(17u8); ++ } ++ BinOp::AddEq(_) => { ++ state.write_u8(18u8); ++ } ++ BinOp::SubEq(_) => { ++ state.write_u8(19u8); ++ } ++ BinOp::MulEq(_) => { ++ state.write_u8(20u8); ++ } ++ BinOp::DivEq(_) => { ++ state.write_u8(21u8); ++ } ++ BinOp::RemEq(_) => { ++ state.write_u8(22u8); ++ } ++ BinOp::BitXorEq(_) => { ++ state.write_u8(23u8); ++ } ++ BinOp::BitAndEq(_) => { ++ state.write_u8(24u8); ++ } ++ BinOp::BitOrEq(_) => { ++ state.write_u8(25u8); ++ } ++ BinOp::ShlEq(_) => { ++ state.write_u8(26u8); ++ } ++ BinOp::ShrEq(_) => { ++ state.write_u8(27u8); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Binding { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Block { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.stmts.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for BoundLifetimes { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lifetimes.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ConstParam { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ self.eq_token.hash(state); ++ self.default.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Constraint { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ self.bounds.hash(state); ++ } ++} ++#[cfg(feature = "derive")] ++impl Hash for Data { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Data::Struct(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Data::Enum(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Data::Union(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(feature = "derive")] ++impl Hash for DataEnum { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.variants.hash(state); ++ } ++} ++#[cfg(feature = "derive")] ++impl Hash for DataStruct { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.fields.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "derive")] ++impl Hash for DataUnion { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.fields.hash(state); ++ } ++} ++#[cfg(feature = "derive")] ++impl Hash for DeriveInput { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.data.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Expr { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ #[cfg(feature = "full")] ++ Expr::Array(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Assign(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::AssignOp(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Async(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Await(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ Expr::Binary(v0) => { ++ state.write_u8(5u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Block(v0) => { ++ state.write_u8(6u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Box(v0) => { ++ state.write_u8(7u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Break(v0) => { ++ state.write_u8(8u8); ++ v0.hash(state); ++ } ++ Expr::Call(v0) => { ++ state.write_u8(9u8); ++ v0.hash(state); ++ } ++ Expr::Cast(v0) => { ++ state.write_u8(10u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Closure(v0) => { ++ state.write_u8(11u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Continue(v0) => { ++ state.write_u8(12u8); ++ v0.hash(state); ++ } ++ Expr::Field(v0) => { ++ state.write_u8(13u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::ForLoop(v0) => { ++ state.write_u8(14u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Group(v0) => { ++ state.write_u8(15u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::If(v0) => { ++ state.write_u8(16u8); ++ v0.hash(state); ++ } ++ Expr::Index(v0) => { ++ state.write_u8(17u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Let(v0) => { ++ state.write_u8(18u8); ++ v0.hash(state); ++ } ++ Expr::Lit(v0) => { ++ state.write_u8(19u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Loop(v0) => { ++ state.write_u8(20u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Macro(v0) => { ++ state.write_u8(21u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Match(v0) => { ++ state.write_u8(22u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::MethodCall(v0) => { ++ state.write_u8(23u8); ++ v0.hash(state); ++ } ++ Expr::Paren(v0) => { ++ state.write_u8(24u8); ++ v0.hash(state); ++ } ++ Expr::Path(v0) => { ++ state.write_u8(25u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Range(v0) => { ++ state.write_u8(26u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Reference(v0) => { ++ state.write_u8(27u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Repeat(v0) => { ++ state.write_u8(28u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Return(v0) => { ++ state.write_u8(29u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Struct(v0) => { ++ state.write_u8(30u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Try(v0) => { ++ state.write_u8(31u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::TryBlock(v0) => { ++ state.write_u8(32u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Tuple(v0) => { ++ state.write_u8(33u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Type(v0) => { ++ state.write_u8(34u8); ++ v0.hash(state); ++ } ++ Expr::Unary(v0) => { ++ state.write_u8(35u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Unsafe(v0) => { ++ state.write_u8(36u8); ++ v0.hash(state); ++ } ++ Expr::Verbatim(v0) => { ++ state.write_u8(37u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::While(v0) => { ++ state.write_u8(38u8); ++ v0.hash(state); ++ } ++ #[cfg(feature = "full")] ++ Expr::Yield(v0) => { ++ state.write_u8(39u8); ++ v0.hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprArray { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.elems.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprAssign { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.left.hash(state); ++ self.right.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprAssignOp { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.left.hash(state); ++ self.op.hash(state); ++ self.right.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprAsync { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.capture.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprAwait { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.base.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprBinary { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.left.hash(state); ++ self.op.hash(state); ++ self.right.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprBlock { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprBox { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprBreak { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprCall { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.func.hash(state); ++ self.args.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprCast { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprClosure { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.asyncness.hash(state); ++ self.movability.hash(state); ++ self.capture.hash(state); ++ self.inputs.hash(state); ++ self.output.hash(state); ++ self.body.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprContinue { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprField { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.base.hash(state); ++ self.member.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprForLoop { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ self.pat.hash(state); ++ self.expr.hash(state); ++ self.body.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprGroup { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprIf { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.cond.hash(state); ++ self.then_branch.hash(state); ++ self.else_branch.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprIndex { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ self.index.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprLet { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.pat.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprLit { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.lit.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprLoop { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ self.body.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mac.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprMatch { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ self.arms.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprMethodCall { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.receiver.hash(state); ++ self.method.hash(state); ++ self.turbofish.hash(state); ++ self.args.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprParen { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprPath { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.qself.hash(state); ++ self.path.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprRange { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.from.hash(state); ++ self.limits.hash(state); ++ self.to.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprReference { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mutability.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprRepeat { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ self.len.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprReturn { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprStruct { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.path.hash(state); ++ self.fields.hash(state); ++ self.dot2_token.hash(state); ++ self.rest.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprTry { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprTryBlock { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprTuple { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.elems.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ExprUnary { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.op.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprUnsafe { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprWhile { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.label.hash(state); ++ self.cond.hash(state); ++ self.body.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ExprYield { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Field { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.colon_token.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for FieldPat { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.member.hash(state); ++ self.colon_token.hash(state); ++ self.pat.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for FieldValue { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.member.hash(state); ++ self.colon_token.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Fields { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Fields::Named(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Fields::Unnamed(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Fields::Unit => { ++ state.write_u8(2u8); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for FieldsNamed { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.named.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for FieldsUnnamed { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.unnamed.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for File { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.shebang.hash(state); ++ self.attrs.hash(state); ++ self.items.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for FnArg { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ FnArg::Receiver(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ FnArg::Typed(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ForeignItem { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ ForeignItem::Fn(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ ForeignItem::Static(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ ForeignItem::Type(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ ForeignItem::Macro(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ ForeignItem::Verbatim(v0) => { ++ state.write_u8(4u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ForeignItemFn { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.sig.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ForeignItemMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mac.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ForeignItemStatic { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.mutability.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ForeignItemType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for GenericArgument { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ GenericArgument::Lifetime(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ GenericArgument::Type(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ GenericArgument::Binding(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ GenericArgument::Constraint(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ GenericArgument::Const(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for GenericMethodArgument { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ GenericMethodArgument::Type(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ GenericMethodArgument::Const(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for GenericParam { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ GenericParam::Type(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ GenericParam::Lifetime(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ GenericParam::Const(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Generics { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lt_token.hash(state); ++ self.params.hash(state); ++ self.gt_token.hash(state); ++ self.where_clause.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ImplItem { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ ImplItem::Const(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ ImplItem::Method(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ ImplItem::Type(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ ImplItem::Macro(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ ImplItem::Verbatim(v0) => { ++ state.write_u8(4u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ImplItemConst { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.defaultness.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ImplItemMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mac.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ImplItemMethod { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.defaultness.hash(state); ++ self.sig.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ImplItemType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.defaultness.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Item { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Item::Const(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Item::Enum(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Item::ExternCrate(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Item::Fn(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ Item::ForeignMod(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ Item::Impl(v0) => { ++ state.write_u8(5u8); ++ v0.hash(state); ++ } ++ Item::Macro(v0) => { ++ state.write_u8(6u8); ++ v0.hash(state); ++ } ++ Item::Macro2(v0) => { ++ state.write_u8(7u8); ++ v0.hash(state); ++ } ++ Item::Mod(v0) => { ++ state.write_u8(8u8); ++ v0.hash(state); ++ } ++ Item::Static(v0) => { ++ state.write_u8(9u8); ++ v0.hash(state); ++ } ++ Item::Struct(v0) => { ++ state.write_u8(10u8); ++ v0.hash(state); ++ } ++ Item::Trait(v0) => { ++ state.write_u8(11u8); ++ v0.hash(state); ++ } ++ Item::TraitAlias(v0) => { ++ state.write_u8(12u8); ++ v0.hash(state); ++ } ++ Item::Type(v0) => { ++ state.write_u8(13u8); ++ v0.hash(state); ++ } ++ Item::Union(v0) => { ++ state.write_u8(14u8); ++ v0.hash(state); ++ } ++ Item::Use(v0) => { ++ state.write_u8(15u8); ++ v0.hash(state); ++ } ++ Item::Verbatim(v0) => { ++ state.write_u8(16u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemConst { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemEnum { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.variants.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemExternCrate { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.rename.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemFn { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.sig.hash(state); ++ self.block.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemForeignMod { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.abi.hash(state); ++ self.items.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemImpl { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.defaultness.hash(state); ++ self.unsafety.hash(state); ++ self.generics.hash(state); ++ self.trait_.hash(state); ++ self.self_ty.hash(state); ++ self.items.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.mac.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemMacro2 { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ TokenStreamHelper(&self.rules).hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemMod { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.content.hash(state); ++ self.semi.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemStatic { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.mutability.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemStruct { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.fields.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemTrait { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.unsafety.hash(state); ++ self.auto_token.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.colon_token.hash(state); ++ self.supertraits.hash(state); ++ self.items.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemTraitAlias { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.bounds.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemUnion { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.fields.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for ItemUse { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.vis.hash(state); ++ self.leading_colon.hash(state); ++ self.tree.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Label { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.name.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for LifetimeDef { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.lifetime.hash(state); ++ self.colon_token.hash(state); ++ self.bounds.hash(state); ++ } ++} ++impl Hash for Lit { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Lit::Str(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Lit::ByteStr(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Lit::Byte(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Lit::Char(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ Lit::Int(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ Lit::Float(v0) => { ++ state.write_u8(5u8); ++ v0.hash(state); ++ } ++ Lit::Bool(v0) => { ++ state.write_u8(6u8); ++ v0.hash(state); ++ } ++ Lit::Verbatim(v0) => { ++ state.write_u8(7u8); ++ v0.to_string().hash(state); ++ } ++ } ++ } ++} ++impl Hash for LitBool { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.value.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Local { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.pat.hash(state); ++ self.init.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Macro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.path.hash(state); ++ self.delimiter.hash(state); ++ TokenStreamHelper(&self.tokens).hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for MacroDelimiter { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ MacroDelimiter::Paren(_) => { ++ state.write_u8(0u8); ++ } ++ MacroDelimiter::Brace(_) => { ++ state.write_u8(1u8); ++ } ++ MacroDelimiter::Bracket(_) => { ++ state.write_u8(2u8); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Meta { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Meta::Path(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Meta::List(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Meta::NameValue(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for MetaList { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.path.hash(state); ++ self.nested.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for MetaNameValue { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.path.hash(state); ++ self.lit.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for MethodTurbofish { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.args.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for NestedMeta { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ NestedMeta::Meta(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ NestedMeta::Lit(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ParenthesizedGenericArguments { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.inputs.hash(state); ++ self.output.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Pat { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Pat::Box(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Pat::Ident(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Pat::Lit(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Pat::Macro(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ Pat::Or(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ Pat::Path(v0) => { ++ state.write_u8(5u8); ++ v0.hash(state); ++ } ++ Pat::Range(v0) => { ++ state.write_u8(6u8); ++ v0.hash(state); ++ } ++ Pat::Reference(v0) => { ++ state.write_u8(7u8); ++ v0.hash(state); ++ } ++ Pat::Rest(v0) => { ++ state.write_u8(8u8); ++ v0.hash(state); ++ } ++ Pat::Slice(v0) => { ++ state.write_u8(9u8); ++ v0.hash(state); ++ } ++ Pat::Struct(v0) => { ++ state.write_u8(10u8); ++ v0.hash(state); ++ } ++ Pat::Tuple(v0) => { ++ state.write_u8(11u8); ++ v0.hash(state); ++ } ++ Pat::TupleStruct(v0) => { ++ state.write_u8(12u8); ++ v0.hash(state); ++ } ++ Pat::Type(v0) => { ++ state.write_u8(13u8); ++ v0.hash(state); ++ } ++ Pat::Verbatim(v0) => { ++ state.write_u8(14u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ Pat::Wild(v0) => { ++ state.write_u8(15u8); ++ v0.hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatBox { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.pat.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatIdent { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.by_ref.hash(state); ++ self.mutability.hash(state); ++ self.ident.hash(state); ++ self.subpat.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatLit { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.expr.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mac.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatOr { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.leading_vert.hash(state); ++ self.cases.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatPath { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.qself.hash(state); ++ self.path.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatRange { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.lo.hash(state); ++ self.limits.hash(state); ++ self.hi.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatReference { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mutability.hash(state); ++ self.pat.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatRest { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatSlice { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.elems.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatStruct { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.path.hash(state); ++ self.fields.hash(state); ++ self.dot2_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatTuple { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.elems.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatTupleStruct { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.path.hash(state); ++ self.pat.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.pat.hash(state); ++ self.ty.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for PatWild { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Path { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.leading_colon.hash(state); ++ self.segments.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for PathArguments { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ PathArguments::None => { ++ state.write_u8(0u8); ++ } ++ PathArguments::AngleBracketed(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ PathArguments::Parenthesized(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for PathSegment { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ self.arguments.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for PredicateEq { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lhs_ty.hash(state); ++ self.rhs_ty.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for PredicateLifetime { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lifetime.hash(state); ++ self.bounds.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for PredicateType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lifetimes.hash(state); ++ self.bounded_ty.hash(state); ++ self.bounds.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for QSelf { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ty.hash(state); ++ self.position.hash(state); ++ self.as_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for RangeLimits { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ RangeLimits::HalfOpen(_) => { ++ state.write_u8(0u8); ++ } ++ RangeLimits::Closed(_) => { ++ state.write_u8(1u8); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Receiver { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.reference.hash(state); ++ self.mutability.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for ReturnType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ ReturnType::Default => { ++ state.write_u8(0u8); ++ } ++ ReturnType::Type(_, v1) => { ++ state.write_u8(1u8); ++ v1.hash(state); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Signature { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.constness.hash(state); ++ self.asyncness.hash(state); ++ self.unsafety.hash(state); ++ self.abi.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.inputs.hash(state); ++ self.variadic.hash(state); ++ self.output.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for Stmt { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Stmt::Local(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Stmt::Item(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Stmt::Expr(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Stmt::Semi(v0, _) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TraitBound { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.paren_token.hash(state); ++ self.modifier.hash(state); ++ self.lifetimes.hash(state); ++ self.path.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TraitBoundModifier { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ TraitBoundModifier::None => { ++ state.write_u8(0u8); ++ } ++ TraitBoundModifier::Maybe(_) => { ++ state.write_u8(1u8); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for TraitItem { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ TraitItem::Const(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ TraitItem::Method(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ TraitItem::Type(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ TraitItem::Macro(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ TraitItem::Verbatim(v0) => { ++ state.write_u8(4u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for TraitItemConst { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.ty.hash(state); ++ self.default.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for TraitItemMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.mac.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for TraitItemMethod { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.sig.hash(state); ++ self.default.hash(state); ++ self.semi_token.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for TraitItemType { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.generics.hash(state); ++ self.colon_token.hash(state); ++ self.bounds.hash(state); ++ self.default.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Type { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Type::Array(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Type::BareFn(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Type::Group(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Type::ImplTrait(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ Type::Infer(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ Type::Macro(v0) => { ++ state.write_u8(5u8); ++ v0.hash(state); ++ } ++ Type::Never(v0) => { ++ state.write_u8(6u8); ++ v0.hash(state); ++ } ++ Type::Paren(v0) => { ++ state.write_u8(7u8); ++ v0.hash(state); ++ } ++ Type::Path(v0) => { ++ state.write_u8(8u8); ++ v0.hash(state); ++ } ++ Type::Ptr(v0) => { ++ state.write_u8(9u8); ++ v0.hash(state); ++ } ++ Type::Reference(v0) => { ++ state.write_u8(10u8); ++ v0.hash(state); ++ } ++ Type::Slice(v0) => { ++ state.write_u8(11u8); ++ v0.hash(state); ++ } ++ Type::TraitObject(v0) => { ++ state.write_u8(12u8); ++ v0.hash(state); ++ } ++ Type::Tuple(v0) => { ++ state.write_u8(13u8); ++ v0.hash(state); ++ } ++ Type::Verbatim(v0) => { ++ state.write_u8(14u8); ++ TokenStreamHelper(v0).hash(state); ++ } ++ _ => unreachable!(), ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeArray { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.elem.hash(state); ++ self.len.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeBareFn { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lifetimes.hash(state); ++ self.unsafety.hash(state); ++ self.abi.hash(state); ++ self.inputs.hash(state); ++ self.variadic.hash(state); ++ self.output.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeGroup { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.elem.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeImplTrait { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.bounds.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeInfer { ++ fn hash<H>(&self, _state: &mut H) ++ where ++ H: Hasher, ++ { ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeMacro { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.mac.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeNever { ++ fn hash<H>(&self, _state: &mut H) ++ where ++ H: Hasher, ++ { ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeParam { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.colon_token.hash(state); ++ self.bounds.hash(state); ++ self.eq_token.hash(state); ++ self.default.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeParamBound { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ TypeParamBound::Trait(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ TypeParamBound::Lifetime(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeParen { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.elem.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypePath { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.qself.hash(state); ++ self.path.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypePtr { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.const_token.hash(state); ++ self.mutability.hash(state); ++ self.elem.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeReference { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.lifetime.hash(state); ++ self.mutability.hash(state); ++ self.elem.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeSlice { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.elem.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeTraitObject { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.dyn_token.hash(state); ++ self.bounds.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for TypeTuple { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.elems.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for UnOp { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ UnOp::Deref(_) => { ++ state.write_u8(0u8); ++ } ++ UnOp::Not(_) => { ++ state.write_u8(1u8); ++ } ++ UnOp::Neg(_) => { ++ state.write_u8(2u8); ++ } ++ } ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UseGlob { ++ fn hash<H>(&self, _state: &mut H) ++ where ++ H: Hasher, ++ { ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UseGroup { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.items.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UseName { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UsePath { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ self.tree.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UseRename { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.ident.hash(state); ++ self.rename.hash(state); ++ } ++} ++#[cfg(feature = "full")] ++impl Hash for UseTree { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ UseTree::Path(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ UseTree::Name(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ UseTree::Rename(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ UseTree::Glob(v0) => { ++ state.write_u8(3u8); ++ v0.hash(state); ++ } ++ UseTree::Group(v0) => { ++ state.write_u8(4u8); ++ v0.hash(state); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Variadic { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Variant { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.attrs.hash(state); ++ self.ident.hash(state); ++ self.fields.hash(state); ++ self.discriminant.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for VisCrate { ++ fn hash<H>(&self, _state: &mut H) ++ where ++ H: Hasher, ++ { ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for VisPublic { ++ fn hash<H>(&self, _state: &mut H) ++ where ++ H: Hasher, ++ { ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for VisRestricted { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.in_token.hash(state); ++ self.path.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for Visibility { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ Visibility::Public(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ Visibility::Crate(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ Visibility::Restricted(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ Visibility::Inherited => { ++ state.write_u8(3u8); ++ } ++ } ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for WhereClause { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ self.predicates.hash(state); ++ } ++} ++#[cfg(any(feature = "derive", feature = "full"))] ++impl Hash for WherePredicate { ++ fn hash<H>(&self, state: &mut H) ++ where ++ H: Hasher, ++ { ++ match self { ++ WherePredicate::Type(v0) => { ++ state.write_u8(0u8); ++ v0.hash(state); ++ } ++ WherePredicate::Lifetime(v0) => { ++ state.write_u8(1u8); ++ v0.hash(state); ++ } ++ WherePredicate::Eq(v0) => { ++ state.write_u8(2u8); ++ v0.hash(state); ++ } ++ } ++ } ++} +diff --git a/third_party/rust/syn/src/gen/visit.rs b/third_party/rust/syn/src/gen/visit.rs +index b667f530c3..24d34b7480 100644 +--- a/third_party/rust/syn/src/gen/visit.rs ++++ b/third_party/rust/syn/src/gen/visit.rs +@@ -20,7 +20,6 @@ macro_rules! full { + unreachable!() + }; + } +-#[cfg(any(feature = "full", feature = "derive"))] + macro_rules! skip { + ($($tt:tt)*) => {}; + } +@@ -30,7 +29,7 @@ macro_rules! skip { + /// + /// [module documentation]: self + /// +-/// *This trait is available if Syn is built with the `"visit"` feature.* ++/// *This trait is available only if Syn is built with the `"visit"` feature.* + pub trait Visit<'ast> { + #[cfg(any(feature = "derive", feature = "full"))] + fn visit_abi(&mut self, i: &'ast Abi) { +@@ -434,35 +433,27 @@ pub trait Visit<'ast> { + fn visit_lifetime_def(&mut self, i: &'ast LifetimeDef) { + visit_lifetime_def(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit(&mut self, i: &'ast Lit) { + visit_lit(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_bool(&mut self, i: &'ast LitBool) { + visit_lit_bool(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_byte(&mut self, i: &'ast LitByte) { + visit_lit_byte(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_byte_str(&mut self, i: &'ast LitByteStr) { + visit_lit_byte_str(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_char(&mut self, i: &'ast LitChar) { + visit_lit_char(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_float(&mut self, i: &'ast LitFloat) { + visit_lit_float(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_int(&mut self, i: &'ast LitInt) { + visit_lit_int(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_str(&mut self, i: &'ast LitStr) { + visit_lit_str(self, i) + } +@@ -2537,7 +2528,6 @@ where + } + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit<'ast, V>(v: &mut V, node: &'ast Lit) + where + V: Visit<'ast> + ?Sized, +@@ -2569,7 +2559,6 @@ where + } + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_bool<'ast, V>(v: &mut V, node: &'ast LitBool) + where + V: Visit<'ast> + ?Sized, +@@ -2577,37 +2566,31 @@ where + skip!(node.value); + v.visit_span(&node.span); + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_byte<'ast, V>(v: &mut V, node: &'ast LitByte) + where + V: Visit<'ast> + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_byte_str<'ast, V>(v: &mut V, node: &'ast LitByteStr) + where + V: Visit<'ast> + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_char<'ast, V>(v: &mut V, node: &'ast LitChar) + where + V: Visit<'ast> + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_float<'ast, V>(v: &mut V, node: &'ast LitFloat) + where + V: Visit<'ast> + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_int<'ast, V>(v: &mut V, node: &'ast LitInt) + where + V: Visit<'ast> + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_str<'ast, V>(v: &mut V, node: &'ast LitStr) + where + V: Visit<'ast> + ?Sized, +diff --git a/third_party/rust/syn/src/gen/visit_mut.rs b/third_party/rust/syn/src/gen/visit_mut.rs +index 5cddb827c6..5ce11f0b2e 100644 +--- a/third_party/rust/syn/src/gen/visit_mut.rs ++++ b/third_party/rust/syn/src/gen/visit_mut.rs +@@ -20,7 +20,6 @@ macro_rules! full { + unreachable!() + }; + } +-#[cfg(any(feature = "full", feature = "derive"))] + macro_rules! skip { + ($($tt:tt)*) => {}; + } +@@ -31,7 +30,7 @@ macro_rules! skip { + /// + /// [module documentation]: self + /// +-/// *This trait is available if Syn is built with the `"visit-mut"` feature.* ++/// *This trait is available only if Syn is built with the `"visit-mut"` feature.* + pub trait VisitMut { + #[cfg(any(feature = "derive", feature = "full"))] + fn visit_abi_mut(&mut self, i: &mut Abi) { +@@ -438,35 +437,27 @@ pub trait VisitMut { + fn visit_lifetime_def_mut(&mut self, i: &mut LifetimeDef) { + visit_lifetime_def_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_mut(&mut self, i: &mut Lit) { + visit_lit_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_bool_mut(&mut self, i: &mut LitBool) { + visit_lit_bool_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_byte_mut(&mut self, i: &mut LitByte) { + visit_lit_byte_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_byte_str_mut(&mut self, i: &mut LitByteStr) { + visit_lit_byte_str_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_char_mut(&mut self, i: &mut LitChar) { + visit_lit_char_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_float_mut(&mut self, i: &mut LitFloat) { + visit_lit_float_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_int_mut(&mut self, i: &mut LitInt) { + visit_lit_int_mut(self, i) + } +- #[cfg(any(feature = "derive", feature = "full"))] + fn visit_lit_str_mut(&mut self, i: &mut LitStr) { + visit_lit_str_mut(self, i) + } +@@ -2543,7 +2534,6 @@ where + } + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_mut<V>(v: &mut V, node: &mut Lit) + where + V: VisitMut + ?Sized, +@@ -2575,7 +2565,6 @@ where + } + } + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_bool_mut<V>(v: &mut V, node: &mut LitBool) + where + V: VisitMut + ?Sized, +@@ -2583,37 +2572,31 @@ where + skip!(node.value); + v.visit_span_mut(&mut node.span); + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_byte_mut<V>(v: &mut V, node: &mut LitByte) + where + V: VisitMut + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_byte_str_mut<V>(v: &mut V, node: &mut LitByteStr) + where + V: VisitMut + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_char_mut<V>(v: &mut V, node: &mut LitChar) + where + V: VisitMut + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_float_mut<V>(v: &mut V, node: &mut LitFloat) + where + V: VisitMut + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_int_mut<V>(v: &mut V, node: &mut LitInt) + where + V: VisitMut + ?Sized, + { + } +-#[cfg(any(feature = "derive", feature = "full"))] + pub fn visit_lit_str_mut<V>(v: &mut V, node: &mut LitStr) + where + V: VisitMut + ?Sized, +diff --git a/third_party/rust/syn/src/generics.rs b/third_party/rust/syn/src/generics.rs +index 95ab2e404a..05e8ef5cdf 100644 +--- a/third_party/rust/syn/src/generics.rs ++++ b/third_party/rust/syn/src/generics.rs +@@ -1,13 +1,16 @@ + use super::*; + use crate::punctuated::{Iter, IterMut, Punctuated}; ++#[cfg(all(feature = "printing", feature = "extra-traits"))] ++use std::fmt::{self, Debug}; ++#[cfg(all(feature = "printing", feature = "extra-traits"))] ++use std::hash::{Hash, Hasher}; + + ast_struct! { + /// Lifetimes and type parameters attached to a declaration of a function, + /// enum, trait, etc. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- #[derive(Default)] + pub struct Generics { + pub lt_token: Option<Token![<]>, + pub params: Punctuated<GenericParam, Token![,]>, +@@ -20,7 +23,7 @@ ast_enum_of_structs! { + /// A generic type parameter, lifetime, or const generic: `T: Into<String>`, + /// `'a: 'b`, `const LEN: usize`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Syntax tree enum +@@ -28,9 +31,6 @@ ast_enum_of_structs! { + /// This type is a [syntax tree enum]. + /// + /// [syntax tree enum]: enum.Expr.html#syntax-tree-enums +- // +- // TODO: change syntax-tree-enum link to an intra rustdoc link, currently +- // blocked on https://github.com/rust-lang/rust/issues/62833 + pub enum GenericParam { + /// A generic type parameter: `T: Into<String>`. + Type(TypeParam), +@@ -46,7 +46,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A generic type parameter: `T: Into<String>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeParam { + pub attrs: Vec<Attribute>, +@@ -61,7 +61,7 @@ ast_struct! { + ast_struct! { + /// A lifetime definition: `'a: 'b + 'c + 'd`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct LifetimeDef { + pub attrs: Vec<Attribute>, +@@ -74,7 +74,7 @@ ast_struct! { + ast_struct! { + /// A const generic parameter: `const LENGTH: usize`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct ConstParam { + pub attrs: Vec<Attribute>, +@@ -87,6 +87,17 @@ ast_struct! { + } + } + ++impl Default for Generics { ++ fn default() -> Self { ++ Generics { ++ lt_token: None, ++ params: Punctuated::new(), ++ gt_token: None, ++ where_clause: None, ++ } ++ } ++} ++ + impl Generics { + /// Returns an + /// <code +@@ -280,29 +291,23 @@ impl<'a> Iterator for ConstParamsMut<'a> { + + /// Returned by `Generics::split_for_impl`. + /// +-/// *This type is available if Syn is built with the `"derive"` or `"full"` ++/// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature and the `"printing"` feature.* + #[cfg(feature = "printing")] +-#[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + pub struct ImplGenerics<'a>(&'a Generics); + + /// Returned by `Generics::split_for_impl`. + /// +-/// *This type is available if Syn is built with the `"derive"` or `"full"` ++/// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature and the `"printing"` feature.* + #[cfg(feature = "printing")] +-#[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + pub struct TypeGenerics<'a>(&'a Generics); + + /// Returned by `TypeGenerics::as_turbofish`. + /// +-/// *This type is available if Syn is built with the `"derive"` or `"full"` ++/// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature and the `"printing"` feature.* + #[cfg(feature = "printing")] +-#[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + pub struct Turbofish<'a>(&'a Generics); + + #[cfg(feature = "printing")] +@@ -314,9 +319,8 @@ impl Generics { + /// # use proc_macro2::{Span, Ident}; + /// # use quote::quote; + /// # +- /// # fn main() { +- /// # let generics: syn::Generics = Default::default(); +- /// # let name = Ident::new("MyType", Span::call_site()); ++ /// # let generics: syn::Generics = Default::default(); ++ /// # let name = Ident::new("MyType", Span::call_site()); + /// # + /// let (impl_generics, ty_generics, where_clause) = generics.split_for_impl(); + /// quote! { +@@ -324,11 +328,10 @@ impl Generics { + /// // ... + /// } + /// } +- /// # ; +- /// # } ++ /// # ; + /// ``` + /// +- /// *This method is available if Syn is built with the `"derive"` or ++ /// *This method is available only if Syn is built with the `"derive"` or + /// `"full"` feature and the `"printing"` feature.* + pub fn split_for_impl(&self) -> (ImplGenerics, TypeGenerics, Option<&WhereClause>) { + ( +@@ -339,11 +342,57 @@ impl Generics { + } + } + ++#[cfg(feature = "printing")] ++macro_rules! generics_wrapper_impls { ++ ($ty:ident) => { ++ #[cfg(feature = "clone-impls")] ++ impl<'a> Clone for $ty<'a> { ++ fn clone(&self) -> Self { ++ $ty(self.0) ++ } ++ } ++ ++ #[cfg(feature = "extra-traits")] ++ impl<'a> Debug for $ty<'a> { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ formatter ++ .debug_tuple(stringify!($ty)) ++ .field(self.0) ++ .finish() ++ } ++ } ++ ++ #[cfg(feature = "extra-traits")] ++ impl<'a> Eq for $ty<'a> {} ++ ++ #[cfg(feature = "extra-traits")] ++ impl<'a> PartialEq for $ty<'a> { ++ fn eq(&self, other: &Self) -> bool { ++ self.0 == other.0 ++ } ++ } ++ ++ #[cfg(feature = "extra-traits")] ++ impl<'a> Hash for $ty<'a> { ++ fn hash<H: Hasher>(&self, state: &mut H) { ++ self.0.hash(state); ++ } ++ } ++ }; ++} ++ ++#[cfg(feature = "printing")] ++generics_wrapper_impls!(ImplGenerics); ++#[cfg(feature = "printing")] ++generics_wrapper_impls!(TypeGenerics); ++#[cfg(feature = "printing")] ++generics_wrapper_impls!(Turbofish); ++ + #[cfg(feature = "printing")] + impl<'a> TypeGenerics<'a> { + /// Turn a type's generics like `<X, Y>` into a turbofish like `::<X, Y>`. + /// +- /// *This method is available if Syn is built with the `"derive"` or ++ /// *This method is available only if Syn is built with the `"derive"` or + /// `"full"` feature and the `"printing"` feature.* + pub fn as_turbofish(&self) -> Turbofish { + Turbofish(self.0) +@@ -353,9 +402,8 @@ impl<'a> TypeGenerics<'a> { + ast_struct! { + /// A set of bound lifetimes: `for<'a, 'b, 'c>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- #[derive(Default)] + pub struct BoundLifetimes { + pub for_token: Token![for], + pub lt_token: Token![<], +@@ -364,6 +412,17 @@ ast_struct! { + } + } + ++impl Default for BoundLifetimes { ++ fn default() -> Self { ++ BoundLifetimes { ++ for_token: Default::default(), ++ lt_token: Default::default(), ++ lifetimes: Punctuated::new(), ++ gt_token: Default::default(), ++ } ++ } ++} ++ + impl LifetimeDef { + pub fn new(lifetime: Lifetime) -> Self { + LifetimeDef { +@@ -391,7 +450,7 @@ impl From<Ident> for TypeParam { + ast_enum_of_structs! { + /// A trait or lifetime used as a bound on a type parameter. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum TypeParamBound { + Trait(TraitBound), +@@ -402,7 +461,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A trait used as a bound on a type parameter. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct TraitBound { + pub paren_token: Option<token::Paren>, +@@ -418,9 +477,8 @@ ast_enum! { + /// A modifier on a trait bound, currently only used for the `?` in + /// `?Sized`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- #[cfg_attr(feature = "clone-impls", derive(Copy))] + pub enum TraitBoundModifier { + None, + Maybe(Token![?]), +@@ -431,7 +489,7 @@ ast_struct! { + /// A `where` clause in a definition: `where T: Deserialize<'de>, D: + /// 'static`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct WhereClause { + pub where_token: Token![where], +@@ -442,7 +500,7 @@ ast_struct! { + ast_enum_of_structs! { + /// A single predicate in a `where` clause: `T: Deserialize<'de>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Syntax tree enum +@@ -450,9 +508,6 @@ ast_enum_of_structs! { + /// This type is a [syntax tree enum]. + /// + /// [syntax tree enum]: enum.Expr.html#syntax-tree-enums +- // +- // TODO: change syntax-tree-enum link to an intra rustdoc link, currently +- // blocked on https://github.com/rust-lang/rust/issues/62833 + pub enum WherePredicate { + /// A type predicate in a `where` clause: `for<'c> Foo<'c>: Trait<'c>`. + Type(PredicateType), +@@ -468,7 +523,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A type predicate in a `where` clause: `for<'c> Foo<'c>: Trait<'c>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct PredicateType { + /// Any lifetimes from a `for` binding +@@ -484,7 +539,7 @@ ast_struct! { + ast_struct! { + /// A lifetime predicate in a `where` clause: `'a: 'b + 'c`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct PredicateLifetime { + pub lifetime: Lifetime, +@@ -496,7 +551,7 @@ ast_struct! { + ast_struct! { + /// An equality predicate in a `where` clause (unsupported). + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct PredicateEq { + pub lhs_ty: Type, +@@ -521,7 +576,6 @@ pub mod parsing { + + let mut params = Punctuated::new(); + let mut allow_lifetime_param = true; +- let mut allow_type_param = true; + loop { + if input.peek(Token![>]) { + break; +@@ -534,7 +588,7 @@ pub mod parsing { + attrs, + ..input.parse()? + })); +- } else if allow_type_param && lookahead.peek(Ident) { ++ } else if lookahead.peek(Ident) { + allow_lifetime_param = false; + params.push_value(GenericParam::Type(TypeParam { + attrs, +@@ -542,7 +596,6 @@ pub mod parsing { + })); + } else if lookahead.peek(Token![const]) { + allow_lifetime_param = false; +- allow_type_param = false; + params.push_value(GenericParam::Const(ConstParam { + attrs, + ..input.parse()? +@@ -665,57 +718,53 @@ pub mod parsing { + + impl Parse for TypeParam { + fn parse(input: ParseStream) -> Result<Self> { +- let has_colon; +- let has_default; +- Ok(TypeParam { +- attrs: input.call(Attribute::parse_outer)?, +- ident: input.parse()?, +- colon_token: { +- if input.peek(Token![:]) { +- has_colon = true; +- Some(input.parse()?) +- } else { +- has_colon = false; +- None +- } +- }, +- bounds: { +- let mut bounds = Punctuated::new(); +- if has_colon { +- loop { +- if input.peek(Token![,]) +- || input.peek(Token![>]) +- || input.peek(Token![=]) +- { +- break; +- } +- let value = input.parse()?; +- bounds.push_value(value); +- if !input.peek(Token![+]) { +- break; +- } +- let punct = input.parse()?; +- bounds.push_punct(punct); +- } ++ let attrs = input.call(Attribute::parse_outer)?; ++ let ident: Ident = input.parse()?; ++ let colon_token: Option<Token![:]> = input.parse()?; ++ ++ let begin_bound = input.fork(); ++ let mut is_maybe_const = false; ++ let mut bounds = Punctuated::new(); ++ if colon_token.is_some() { ++ loop { ++ if input.peek(Token![,]) || input.peek(Token![>]) || input.peek(Token![=]) { ++ break; + } +- bounds +- }, +- eq_token: { +- if input.peek(Token![=]) { +- has_default = true; +- Some(input.parse()?) +- } else { +- has_default = false; +- None ++ if input.peek(Token![?]) && input.peek2(Token![const]) { ++ input.parse::<Token![?]>()?; ++ input.parse::<Token![const]>()?; ++ is_maybe_const = true; + } +- }, +- default: { +- if has_default { +- Some(input.parse()?) +- } else { +- None ++ let value: TypeParamBound = input.parse()?; ++ bounds.push_value(value); ++ if !input.peek(Token![+]) { ++ break; + } +- }, ++ let punct: Token![+] = input.parse()?; ++ bounds.push_punct(punct); ++ } ++ } ++ ++ let mut eq_token: Option<Token![=]> = input.parse()?; ++ let mut default = if eq_token.is_some() { ++ Some(input.parse::<Type>()?) ++ } else { ++ None ++ }; ++ ++ if is_maybe_const { ++ bounds.clear(); ++ eq_token = None; ++ default = Some(Type::Verbatim(verbatim::between(begin_bound, input))); ++ } ++ ++ Ok(TypeParam { ++ attrs, ++ ident, ++ colon_token, ++ bounds, ++ eq_token, ++ default, + }) + } + } +@@ -898,6 +947,8 @@ mod printing { + use super::*; + + use proc_macro2::TokenStream; ++ #[cfg(feature = "full")] ++ use proc_macro2::TokenTree; + use quote::{ToTokens, TokenStreamExt}; + + use crate::attr::FilterAttrs; +@@ -1080,9 +1131,25 @@ mod printing { + TokensOrDefault(&self.colon_token).to_tokens(tokens); + self.bounds.to_tokens(tokens); + } +- if self.default.is_some() { ++ if let Some(default) = &self.default { ++ #[cfg(feature = "full")] ++ { ++ if self.eq_token.is_none() { ++ if let Type::Verbatim(default) = default { ++ let mut iter = default.clone().into_iter(); ++ match (iter.next(), iter.next()) { ++ (Some(TokenTree::Punct(ref q)), Some(TokenTree::Ident(ref c))) ++ if q.as_char() == '?' && c == "const" => ++ { ++ return default.to_tokens(tokens); ++ } ++ _ => {} ++ } ++ } ++ } ++ } + TokensOrDefault(&self.eq_token).to_tokens(tokens); +- self.default.to_tokens(tokens); ++ default.to_tokens(tokens); + } + } + } +@@ -1117,9 +1184,9 @@ mod printing { + self.ident.to_tokens(tokens); + self.colon_token.to_tokens(tokens); + self.ty.to_tokens(tokens); +- if self.default.is_some() { ++ if let Some(default) = &self.default { + TokensOrDefault(&self.eq_token).to_tokens(tokens); +- self.default.to_tokens(tokens); ++ default.to_tokens(tokens); + } + } + } +diff --git a/third_party/rust/syn/src/item.rs b/third_party/rust/syn/src/item.rs +index ff4485ace9..0d8f7d3ddc 100644 +--- a/third_party/rust/syn/src/item.rs ++++ b/third_party/rust/syn/src/item.rs +@@ -1,17 +1,15 @@ + use super::*; +-use crate::derive::{Data, DeriveInput}; ++use crate::derive::{Data, DataEnum, DataStruct, DataUnion, DeriveInput}; + use crate::punctuated::Punctuated; + use proc_macro2::TokenStream; + +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; +-#[cfg(feature = "extra-traits")] +-use std::hash::{Hash, Hasher}; ++#[cfg(feature = "parsing")] ++use std::mem; + + ast_enum_of_structs! { + /// Things that can appear directly inside of a module or scope. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -21,7 +19,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum Item #manual_extra_traits { ++ pub enum Item { + /// A constant item: `const MAX: u16 = 65535`. + Const(ItemConst), + +@@ -83,7 +81,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A constant item: `const MAX: u16 = 65535`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemConst { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -100,7 +98,7 @@ ast_struct! { + ast_struct! { + /// An enum definition: `enum Foo<A, B> { A(A), B(B) }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemEnum { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -115,7 +113,7 @@ ast_struct! { + ast_struct! { + /// An `extern crate` item: `extern crate serde`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemExternCrate { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -131,7 +129,7 @@ ast_struct! { + /// A free-standing function: `fn process(n: usize) -> Result<()> { ... + /// }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemFn { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -143,7 +141,7 @@ ast_struct! { + ast_struct! { + /// A block of foreign items: `extern "C" { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemForeignMod { + pub attrs: Vec<Attribute>, + pub abi: Abi, +@@ -156,7 +154,7 @@ ast_struct! { + /// An impl block providing trait or associated items: `impl<A> Trait + /// for Data<A> { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemImpl { + pub attrs: Vec<Attribute>, + pub defaultness: Option<Token![default]>, +@@ -175,7 +173,7 @@ ast_struct! { + ast_struct! { + /// A macro invocation, which includes `macro_rules!` definitions. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemMacro { + pub attrs: Vec<Attribute>, + /// The `example` in `macro_rules! example { ... }`. +@@ -188,8 +186,8 @@ ast_struct! { + ast_struct! { + /// A 2.0-style declarative macro introduced by the `macro` keyword. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* +- pub struct ItemMacro2 #manual_extra_traits { ++ /// *This type is available only if Syn is built with the `"full"` feature.* ++ pub struct ItemMacro2 { + pub attrs: Vec<Attribute>, + pub vis: Visibility, + pub macro_token: Token![macro], +@@ -201,7 +199,7 @@ ast_struct! { + ast_struct! { + /// A module or module declaration: `mod m` or `mod m { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemMod { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -215,7 +213,7 @@ ast_struct! { + ast_struct! { + /// A static item: `static BIKE: Shed = Shed(42)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemStatic { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -233,7 +231,7 @@ ast_struct! { + ast_struct! { + /// A struct definition: `struct Foo<A> { x: A }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemStruct { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -248,7 +246,7 @@ ast_struct! { + ast_struct! { + /// A trait definition: `pub trait Iterator { ... }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemTrait { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -267,7 +265,7 @@ ast_struct! { + ast_struct! { + /// A trait alias: `pub trait SharableIterator = Iterator + Sync`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemTraitAlias { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -283,7 +281,7 @@ ast_struct! { + ast_struct! { + /// A type alias: `type Result<T> = std::result::Result<T, MyError>`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemType { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -299,7 +297,7 @@ ast_struct! { + ast_struct! { + /// A union definition: `union Foo<A, B> { x: A, y: B }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemUnion { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -313,7 +311,7 @@ ast_struct! { + ast_struct! { + /// A use declaration: `use std::collections::HashMap`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ItemUse { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -324,145 +322,32 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Item {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Item { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (Item::Const(this), Item::Const(other)) => this == other, +- (Item::Enum(this), Item::Enum(other)) => this == other, +- (Item::ExternCrate(this), Item::ExternCrate(other)) => this == other, +- (Item::Fn(this), Item::Fn(other)) => this == other, +- (Item::ForeignMod(this), Item::ForeignMod(other)) => this == other, +- (Item::Impl(this), Item::Impl(other)) => this == other, +- (Item::Macro(this), Item::Macro(other)) => this == other, +- (Item::Macro2(this), Item::Macro2(other)) => this == other, +- (Item::Mod(this), Item::Mod(other)) => this == other, +- (Item::Static(this), Item::Static(other)) => this == other, +- (Item::Struct(this), Item::Struct(other)) => this == other, +- (Item::Trait(this), Item::Trait(other)) => this == other, +- (Item::TraitAlias(this), Item::TraitAlias(other)) => this == other, +- (Item::Type(this), Item::Type(other)) => this == other, +- (Item::Union(this), Item::Union(other)) => this == other, +- (Item::Use(this), Item::Use(other)) => this == other, +- (Item::Verbatim(this), Item::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Item { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { ++impl Item { ++ #[cfg(feature = "parsing")] ++ pub(crate) fn replace_attrs(&mut self, new: Vec<Attribute>) -> Vec<Attribute> { + match self { +- Item::Const(item) => { +- state.write_u8(0); +- item.hash(state); +- } +- Item::Enum(item) => { +- state.write_u8(1); +- item.hash(state); +- } +- Item::ExternCrate(item) => { +- state.write_u8(2); +- item.hash(state); +- } +- Item::Fn(item) => { +- state.write_u8(3); +- item.hash(state); +- } +- Item::ForeignMod(item) => { +- state.write_u8(4); +- item.hash(state); +- } +- Item::Impl(item) => { +- state.write_u8(5); +- item.hash(state); +- } +- Item::Macro(item) => { +- state.write_u8(6); +- item.hash(state); +- } +- Item::Macro2(item) => { +- state.write_u8(7); +- item.hash(state); +- } +- Item::Mod(item) => { +- state.write_u8(8); +- item.hash(state); +- } +- Item::Static(item) => { +- state.write_u8(9); +- item.hash(state); +- } +- Item::Struct(item) => { +- state.write_u8(10); +- item.hash(state); +- } +- Item::Trait(item) => { +- state.write_u8(11); +- item.hash(state); +- } +- Item::TraitAlias(item) => { +- state.write_u8(12); +- item.hash(state); +- } +- Item::Type(item) => { +- state.write_u8(13); +- item.hash(state); +- } +- Item::Union(item) => { +- state.write_u8(14); +- item.hash(state); +- } +- Item::Use(item) => { +- state.write_u8(15); +- item.hash(state); +- } +- Item::Verbatim(item) => { +- state.write_u8(16); +- TokenStreamHelper(item).hash(state); +- } ++ Item::ExternCrate(ItemExternCrate { attrs, .. }) ++ | Item::Use(ItemUse { attrs, .. }) ++ | Item::Static(ItemStatic { attrs, .. }) ++ | Item::Const(ItemConst { attrs, .. }) ++ | Item::Fn(ItemFn { attrs, .. }) ++ | Item::Mod(ItemMod { attrs, .. }) ++ | Item::ForeignMod(ItemForeignMod { attrs, .. }) ++ | Item::Type(ItemType { attrs, .. }) ++ | Item::Struct(ItemStruct { attrs, .. }) ++ | Item::Enum(ItemEnum { attrs, .. }) ++ | Item::Union(ItemUnion { attrs, .. }) ++ | Item::Trait(ItemTrait { attrs, .. }) ++ | Item::TraitAlias(ItemTraitAlias { attrs, .. }) ++ | Item::Impl(ItemImpl { attrs, .. }) ++ | Item::Macro(ItemMacro { attrs, .. }) ++ | Item::Macro2(ItemMacro2 { attrs, .. }) => mem::replace(attrs, new), ++ Item::Verbatim(_) => Vec::new(), + Item::__Nonexhaustive => unreachable!(), + } + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for ItemMacro2 {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for ItemMacro2 { +- fn eq(&self, other: &Self) -> bool { +- self.attrs == other.attrs +- && self.vis == other.vis +- && self.macro_token == other.macro_token +- && self.ident == other.ident +- && TokenStreamHelper(&self.rules) == TokenStreamHelper(&other.rules) +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for ItemMacro2 { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- self.attrs.hash(state); +- self.vis.hash(state); +- self.macro_token.hash(state); +- self.ident.hash(state); +- TokenStreamHelper(&self.rules).hash(state); +- } +-} +- + impl From<DeriveInput> for Item { + fn from(input: DeriveInput) -> Item { + match input.data { +@@ -496,10 +381,57 @@ impl From<DeriveInput> for Item { + } + } + ++impl From<ItemStruct> for DeriveInput { ++ fn from(input: ItemStruct) -> DeriveInput { ++ DeriveInput { ++ attrs: input.attrs, ++ vis: input.vis, ++ ident: input.ident, ++ generics: input.generics, ++ data: Data::Struct(DataStruct { ++ struct_token: input.struct_token, ++ fields: input.fields, ++ semi_token: input.semi_token, ++ }), ++ } ++ } ++} ++ ++impl From<ItemEnum> for DeriveInput { ++ fn from(input: ItemEnum) -> DeriveInput { ++ DeriveInput { ++ attrs: input.attrs, ++ vis: input.vis, ++ ident: input.ident, ++ generics: input.generics, ++ data: Data::Enum(DataEnum { ++ enum_token: input.enum_token, ++ brace_token: input.brace_token, ++ variants: input.variants, ++ }), ++ } ++ } ++} ++ ++impl From<ItemUnion> for DeriveInput { ++ fn from(input: ItemUnion) -> DeriveInput { ++ DeriveInput { ++ attrs: input.attrs, ++ vis: input.vis, ++ ident: input.ident, ++ generics: input.generics, ++ data: Data::Union(DataUnion { ++ union_token: input.union_token, ++ fields: input.fields, ++ }), ++ } ++ } ++} ++ + ast_enum_of_structs! { + /// A suffix of an import tree in a `use` item: `Type as Renamed` or `*`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -530,7 +462,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A path prefix of imports in a `use` item: `std::...`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct UsePath { + pub ident: Ident, + pub colon2_token: Token![::], +@@ -541,7 +473,7 @@ ast_struct! { + ast_struct! { + /// An identifier imported by a `use` item: `HashMap`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct UseName { + pub ident: Ident, + } +@@ -550,7 +482,7 @@ ast_struct! { + ast_struct! { + /// An renamed identifier imported by a `use` item: `HashMap as Map`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct UseRename { + pub ident: Ident, + pub as_token: Token![as], +@@ -561,7 +493,7 @@ ast_struct! { + ast_struct! { + /// A glob import in a `use` item: `*`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct UseGlob { + pub star_token: Token![*], + } +@@ -570,7 +502,7 @@ ast_struct! { + ast_struct! { + /// A braced group of imports in a `use` item: `{A, B, C}`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct UseGroup { + pub brace_token: token::Brace, + pub items: Punctuated<UseTree, Token![,]>, +@@ -580,7 +512,7 @@ ast_struct! { + ast_enum_of_structs! { + /// An item within an `extern` block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -590,7 +522,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum ForeignItem #manual_extra_traits { ++ pub enum ForeignItem { + /// A foreign function in an `extern` block. + Fn(ForeignItemFn), + +@@ -614,7 +546,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A foreign function in an `extern` block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ForeignItemFn { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -626,7 +558,7 @@ ast_struct! { + ast_struct! { + /// A foreign static item in an `extern` block: `static ext: u8`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ForeignItemStatic { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -642,7 +574,7 @@ ast_struct! { + ast_struct! { + /// A foreign type in an `extern` block: `type void`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ForeignItemType { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -655,7 +587,7 @@ ast_struct! { + ast_struct! { + /// A macro invocation within an extern block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ForeignItemMacro { + pub attrs: Vec<Attribute>, + pub mac: Macro, +@@ -663,61 +595,10 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for ForeignItem {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for ForeignItem { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (ForeignItem::Fn(this), ForeignItem::Fn(other)) => this == other, +- (ForeignItem::Static(this), ForeignItem::Static(other)) => this == other, +- (ForeignItem::Type(this), ForeignItem::Type(other)) => this == other, +- (ForeignItem::Macro(this), ForeignItem::Macro(other)) => this == other, +- (ForeignItem::Verbatim(this), ForeignItem::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for ForeignItem { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- match self { +- ForeignItem::Fn(item) => { +- state.write_u8(0); +- item.hash(state); +- } +- ForeignItem::Static(item) => { +- state.write_u8(1); +- item.hash(state); +- } +- ForeignItem::Type(item) => { +- state.write_u8(2); +- item.hash(state); +- } +- ForeignItem::Macro(item) => { +- state.write_u8(3); +- item.hash(state); +- } +- ForeignItem::Verbatim(item) => { +- state.write_u8(4); +- TokenStreamHelper(item).hash(state); +- } +- ForeignItem::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + ast_enum_of_structs! { + /// An item declaration within the definition of a trait. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -727,7 +608,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum TraitItem #manual_extra_traits { ++ pub enum TraitItem { + /// An associated constant within the definition of a trait. + Const(TraitItemConst), + +@@ -751,7 +632,7 @@ ast_enum_of_structs! { + ast_struct! { + /// An associated constant within the definition of a trait. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct TraitItemConst { + pub attrs: Vec<Attribute>, + pub const_token: Token![const], +@@ -766,7 +647,7 @@ ast_struct! { + ast_struct! { + /// A trait method within the definition of a trait. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct TraitItemMethod { + pub attrs: Vec<Attribute>, + pub sig: Signature, +@@ -778,7 +659,7 @@ ast_struct! { + ast_struct! { + /// An associated type within the definition of a trait. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct TraitItemType { + pub attrs: Vec<Attribute>, + pub type_token: Token![type], +@@ -794,7 +675,7 @@ ast_struct! { + ast_struct! { + /// A macro invocation within the definition of a trait. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct TraitItemMacro { + pub attrs: Vec<Attribute>, + pub mac: Macro, +@@ -802,61 +683,10 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for TraitItem {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for TraitItem { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (TraitItem::Const(this), TraitItem::Const(other)) => this == other, +- (TraitItem::Method(this), TraitItem::Method(other)) => this == other, +- (TraitItem::Type(this), TraitItem::Type(other)) => this == other, +- (TraitItem::Macro(this), TraitItem::Macro(other)) => this == other, +- (TraitItem::Verbatim(this), TraitItem::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for TraitItem { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- match self { +- TraitItem::Const(item) => { +- state.write_u8(0); +- item.hash(state); +- } +- TraitItem::Method(item) => { +- state.write_u8(1); +- item.hash(state); +- } +- TraitItem::Type(item) => { +- state.write_u8(2); +- item.hash(state); +- } +- TraitItem::Macro(item) => { +- state.write_u8(3); +- item.hash(state); +- } +- TraitItem::Verbatim(item) => { +- state.write_u8(4); +- TokenStreamHelper(item).hash(state); +- } +- TraitItem::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + ast_enum_of_structs! { + /// An item within an impl block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -866,7 +696,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum ImplItem #manual_extra_traits { ++ pub enum ImplItem { + /// An associated constant within an impl block. + Const(ImplItemConst), + +@@ -890,7 +720,7 @@ ast_enum_of_structs! { + ast_struct! { + /// An associated constant within an impl block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ImplItemConst { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -908,7 +738,7 @@ ast_struct! { + ast_struct! { + /// A method within an impl block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ImplItemMethod { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -921,7 +751,7 @@ ast_struct! { + ast_struct! { + /// An associated type within an impl block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ImplItemType { + pub attrs: Vec<Attribute>, + pub vis: Visibility, +@@ -938,7 +768,7 @@ ast_struct! { + ast_struct! { + /// A macro invocation within an impl block. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct ImplItemMacro { + pub attrs: Vec<Attribute>, + pub mac: Macro, +@@ -946,62 +776,11 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for ImplItem {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for ImplItem { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (ImplItem::Const(this), ImplItem::Const(other)) => this == other, +- (ImplItem::Method(this), ImplItem::Method(other)) => this == other, +- (ImplItem::Type(this), ImplItem::Type(other)) => this == other, +- (ImplItem::Macro(this), ImplItem::Macro(other)) => this == other, +- (ImplItem::Verbatim(this), ImplItem::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for ImplItem { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- match self { +- ImplItem::Const(item) => { +- state.write_u8(0); +- item.hash(state); +- } +- ImplItem::Method(item) => { +- state.write_u8(1); +- item.hash(state); +- } +- ImplItem::Type(item) => { +- state.write_u8(2); +- item.hash(state); +- } +- ImplItem::Macro(item) => { +- state.write_u8(3); +- item.hash(state); +- } +- ImplItem::Verbatim(item) => { +- state.write_u8(4); +- TokenStreamHelper(item).hash(state); +- } +- ImplItem::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + ast_struct! { + /// A function signature in a trait or implementation: `unsafe fn + /// initialize(&self)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Signature { + pub constness: Option<Token![const]>, + pub asyncness: Option<Token![async]>, +@@ -1017,13 +796,34 @@ ast_struct! { + } + } + ++impl Signature { ++ /// A method's `self` receiver, such as `&self` or `self: Box<Self>`. ++ pub fn receiver(&self) -> Option<&FnArg> { ++ let arg = self.inputs.first()?; ++ match arg { ++ FnArg::Receiver(_) => Some(arg), ++ FnArg::Typed(PatType { pat, .. }) => { ++ if let Pat::Ident(PatIdent { ident, .. }) = &**pat { ++ if ident == "self" { ++ return Some(arg); ++ } ++ } ++ None ++ } ++ } ++ } ++} ++ + ast_enum_of_structs! { + /// An argument in a function signature: the `n: usize` in `fn f(n: usize)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub enum FnArg { + /// The `self` argument of an associated method, whether taken by value + /// or by reference. ++ /// ++ /// Note that `self` receivers with a specified type, such as `self: ++ /// Box<Self>`, are parsed as a `FnArg::Typed`. + Receiver(Receiver), + + /// A function argument accepted by pattern and type. +@@ -1035,7 +835,10 @@ ast_struct! { + /// The `self` argument of an associated method, whether taken by value + /// or by reference. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// Note that `self` receivers with a specified type, such as `self: ++ /// Box<Self>`, are parsed as a `FnArg::Typed`. ++ /// ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Receiver { + pub attrs: Vec<Attribute>, + pub reference: Option<(Token![&], Option<Lifetime>)>, +@@ -1056,7 +859,8 @@ pub mod parsing { + + use crate::ext::IdentExt; + use crate::parse::discouraged::Speculative; +- use crate::parse::{Parse, ParseStream, Result}; ++ use crate::parse::{Parse, ParseBuffer, ParseStream, Result}; ++ use crate::token::Brace; + use proc_macro2::{Delimiter, Group, Punct, Spacing, TokenTree}; + use std::iter::{self, FromIterator}; + +@@ -1064,18 +868,26 @@ pub mod parsing { + + impl Parse for Item { + fn parse(input: ParseStream) -> Result<Self> { ++ let begin = input.fork(); + let mut attrs = input.call(Attribute::parse_outer)?; + let ahead = input.fork(); + let vis: Visibility = ahead.parse()?; + + let lookahead = ahead.lookahead1(); +- let mut item = if lookahead.peek(Token![extern]) { ++ let mut item = if lookahead.peek(Token![fn]) || peek_signature(&ahead) { ++ let vis: Visibility = input.parse()?; ++ let sig = parse_signature(input)?; ++ if input.peek(Token![;]) { ++ input.parse::<Token![;]>()?; ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } else { ++ parse_rest_of_fn(input, Vec::new(), vis, sig).map(Item::Fn) ++ } ++ } else if lookahead.peek(Token![extern]) { + ahead.parse::<Token![extern]>()?; + let lookahead = ahead.lookahead1(); + if lookahead.peek(Token![crate]) { + input.parse().map(Item::ExternCrate) +- } else if lookahead.peek(Token![fn]) { +- input.parse().map(Item::Fn) + } else if lookahead.peek(token::Brace) { + input.parse().map(Item::ForeignMod) + } else if lookahead.peek(LitStr) { +@@ -1083,8 +895,6 @@ pub mod parsing { + let lookahead = ahead.lookahead1(); + if lookahead.peek(token::Brace) { + input.parse().map(Item::ForeignMod) +- } else if lookahead.peek(Token![fn]) { +- input.parse().map(Item::Fn) + } else { + Err(lookahead.error()) + } +@@ -1094,18 +904,61 @@ pub mod parsing { + } else if lookahead.peek(Token![use]) { + input.parse().map(Item::Use) + } else if lookahead.peek(Token![static]) { +- input.parse().map(Item::Static) ++ let vis = input.parse()?; ++ let static_token = input.parse()?; ++ let mutability = input.parse()?; ++ let ident = input.parse()?; ++ let colon_token = input.parse()?; ++ let ty = input.parse()?; ++ if input.peek(Token![;]) { ++ input.parse::<Token![;]>()?; ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(Item::Static(ItemStatic { ++ attrs: Vec::new(), ++ vis, ++ static_token, ++ mutability, ++ ident, ++ colon_token, ++ ty, ++ eq_token: input.parse()?, ++ expr: input.parse()?, ++ semi_token: input.parse()?, ++ })) ++ } + } else if lookahead.peek(Token![const]) { + ahead.parse::<Token![const]>()?; + let lookahead = ahead.lookahead1(); + if lookahead.peek(Ident) || lookahead.peek(Token![_]) { +- input.parse().map(Item::Const) +- } else if lookahead.peek(Token![unsafe]) +- || lookahead.peek(Token![async]) +- || lookahead.peek(Token![extern]) +- || lookahead.peek(Token![fn]) +- { +- input.parse().map(Item::Fn) ++ let vis = input.parse()?; ++ let const_token = input.parse()?; ++ let ident = { ++ let lookahead = input.lookahead1(); ++ if lookahead.peek(Ident) || lookahead.peek(Token![_]) { ++ input.call(Ident::parse_any)? ++ } else { ++ return Err(lookahead.error()); ++ } ++ }; ++ let colon_token = input.parse()?; ++ let ty = input.parse()?; ++ if input.peek(Token![;]) { ++ input.parse::<Token![;]>()?; ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(Item::Const(ItemConst { ++ attrs: Vec::new(), ++ vis, ++ const_token, ++ ident, ++ colon_token, ++ ty, ++ eq_token: input.parse()?, ++ expr: input.parse()?, ++ semi_token: input.parse()?, ++ })) ++ } + } else { + Err(lookahead.error()) + } +@@ -1117,21 +970,19 @@ pub mod parsing { + { + input.parse().map(Item::Trait) + } else if lookahead.peek(Token![impl]) { +- input.parse().map(Item::Impl) +- } else if lookahead.peek(Token![async]) +- || lookahead.peek(Token![extern]) +- || lookahead.peek(Token![fn]) +- { +- input.parse().map(Item::Fn) ++ let allow_const_impl = true; ++ if let Some(item) = parse_impl(input, allow_const_impl)? { ++ Ok(Item::Impl(item)) ++ } else { ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } + } else { + Err(lookahead.error()) + } +- } else if lookahead.peek(Token![async]) || lookahead.peek(Token![fn]) { +- input.parse().map(Item::Fn) + } else if lookahead.peek(Token![mod]) { + input.parse().map(Item::Mod) + } else if lookahead.peek(Token![type]) { +- input.parse().map(Item::Type) ++ parse_item_type(begin, input) + } else if lookahead.peek(existential) { + input.call(item_existential).map(Item::Verbatim) + } else if lookahead.peek(Token![struct]) { +@@ -1147,14 +998,18 @@ pub mod parsing { + } else if lookahead.peek(Token![impl]) + || lookahead.peek(Token![default]) && !ahead.peek2(Token![!]) + { +- input.parse().map(Item::Impl) ++ let allow_const_impl = true; ++ if let Some(item) = parse_impl(input, allow_const_impl)? { ++ Ok(Item::Impl(item)) ++ } else { ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } + } else if lookahead.peek(Token![macro]) { + input.parse().map(Item::Macro2) + } else if vis.is_inherited() + && (lookahead.peek(Ident) + || lookahead.peek(Token![self]) + || lookahead.peek(Token![super]) +- || lookahead.peek(Token![extern]) + || lookahead.peek(Token![crate]) + || lookahead.peek(Token![::])) + { +@@ -1163,32 +1018,64 @@ pub mod parsing { + Err(lookahead.error()) + }?; + +- { +- let item_attrs = match &mut item { +- Item::ExternCrate(item) => &mut item.attrs, +- Item::Use(item) => &mut item.attrs, +- Item::Static(item) => &mut item.attrs, +- Item::Const(item) => &mut item.attrs, +- Item::Fn(item) => &mut item.attrs, +- Item::Mod(item) => &mut item.attrs, +- Item::ForeignMod(item) => &mut item.attrs, +- Item::Type(item) => &mut item.attrs, +- Item::Struct(item) => &mut item.attrs, +- Item::Enum(item) => &mut item.attrs, +- Item::Union(item) => &mut item.attrs, +- Item::Trait(item) => &mut item.attrs, +- Item::TraitAlias(item) => &mut item.attrs, +- Item::Impl(item) => &mut item.attrs, +- Item::Macro(item) => &mut item.attrs, +- Item::Macro2(item) => &mut item.attrs, +- Item::Verbatim(_) => return Ok(item), +- Item::__Nonexhaustive => unreachable!(), +- }; +- attrs.extend(item_attrs.drain(..)); +- *item_attrs = attrs; ++ attrs.extend(item.replace_attrs(Vec::new())); ++ item.replace_attrs(attrs); ++ Ok(item) ++ } ++ } ++ ++ struct FlexibleItemType { ++ vis: Visibility, ++ defaultness: Option<Token![default]>, ++ type_token: Token![type], ++ ident: Ident, ++ generics: Generics, ++ colon_token: Option<Token![:]>, ++ bounds: Punctuated<TypeParamBound, Token![+]>, ++ ty: Option<(Token![=], Type)>, ++ semi_token: Token![;], ++ } ++ ++ impl Parse for FlexibleItemType { ++ fn parse(input: ParseStream) -> Result<Self> { ++ let vis: Visibility = input.parse()?; ++ let defaultness: Option<Token![default]> = input.parse()?; ++ let type_token: Token![type] = input.parse()?; ++ let ident: Ident = input.parse()?; ++ let mut generics: Generics = input.parse()?; ++ let colon_token: Option<Token![:]> = input.parse()?; ++ let mut bounds = Punctuated::new(); ++ if colon_token.is_some() { ++ loop { ++ bounds.push_value(input.parse::<TypeParamBound>()?); ++ if input.peek(Token![where]) || input.peek(Token![=]) || input.peek(Token![;]) { ++ break; ++ } ++ bounds.push_punct(input.parse::<Token![+]>()?); ++ if input.peek(Token![where]) || input.peek(Token![=]) || input.peek(Token![;]) { ++ break; ++ } ++ } + } ++ generics.where_clause = input.parse()?; ++ let ty = if let Some(eq_token) = input.parse()? { ++ Some((eq_token, input.parse::<Type>()?)) ++ } else { ++ None ++ }; ++ let semi_token: Token![;] = input.parse()?; + +- Ok(item) ++ Ok(FlexibleItemType { ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds, ++ ty, ++ semi_token, ++ }) + } + } + +@@ -1310,7 +1197,6 @@ pub mod parsing { + || lookahead.peek(Token![self]) + || lookahead.peek(Token![super]) + || lookahead.peek(Token![crate]) +- || lookahead.peek(Token![extern]) + { + let ident = input.call(Ident::parse_any)?; + if input.peek(Token![::]) { +@@ -1392,69 +1278,126 @@ pub mod parsing { + } + } + +- impl Parse for ItemFn { +- fn parse(input: ParseStream) -> Result<Self> { +- let outer_attrs = input.call(Attribute::parse_outer)?; +- let vis: Visibility = input.parse()?; +- let constness: Option<Token![const]> = input.parse()?; +- let asyncness: Option<Token![async]> = input.parse()?; +- let unsafety: Option<Token![unsafe]> = input.parse()?; +- let abi: Option<Abi> = input.parse()?; +- let fn_token: Token![fn] = input.parse()?; +- let ident: Ident = input.parse()?; +- let generics: Generics = input.parse()?; ++ fn pop_variadic(args: &mut Punctuated<FnArg, Token![,]>) -> Option<Variadic> { ++ let trailing_punct = args.trailing_punct(); + +- let content; +- let paren_token = parenthesized!(content in input); +- let inputs = content.parse_terminated(FnArg::parse)?; +- let variadic = inputs.last().as_ref().and_then(get_variadic); +- +- fn get_variadic(input: &&FnArg) -> Option<Variadic> { +- if let FnArg::Typed(PatType { ty, .. }) = input { +- if let Type::Verbatim(tokens) = &**ty { +- if let Ok(dots) = parse2(tokens.clone()) { +- return Some(Variadic { +- attrs: Vec::new(), +- dots, +- }); +- } +- } +- } +- None ++ let last = match args.last_mut()? { ++ FnArg::Typed(last) => last, ++ _ => return None, ++ }; ++ ++ let ty = match last.ty.as_ref() { ++ Type::Verbatim(ty) => ty, ++ _ => return None, ++ }; ++ ++ let mut variadic = Variadic { ++ attrs: Vec::new(), ++ dots: parse2(ty.clone()).ok()?, ++ }; ++ ++ if let Pat::Verbatim(pat) = last.pat.as_ref() { ++ if pat.to_string() == "..." && !trailing_punct { ++ variadic.attrs = mem::replace(&mut last.attrs, Vec::new()); ++ args.pop(); + } ++ } + +- let output: ReturnType = input.parse()?; +- let where_clause: Option<WhereClause> = input.parse()?; ++ Some(variadic) ++ } + +- let content; +- let brace_token = braced!(content in input); +- let inner_attrs = content.call(Attribute::parse_inner)?; +- let stmts = content.call(Block::parse_within)?; ++ fn variadic_to_tokens(dots: &Token![...]) -> TokenStream { ++ TokenStream::from_iter(vec![ ++ TokenTree::Punct({ ++ let mut dot = Punct::new('.', Spacing::Joint); ++ dot.set_span(dots.spans[0]); ++ dot ++ }), ++ TokenTree::Punct({ ++ let mut dot = Punct::new('.', Spacing::Joint); ++ dot.set_span(dots.spans[1]); ++ dot ++ }), ++ TokenTree::Punct({ ++ let mut dot = Punct::new('.', Spacing::Alone); ++ dot.set_span(dots.spans[2]); ++ dot ++ }), ++ ]) ++ } + +- Ok(ItemFn { +- attrs: private::attrs(outer_attrs, inner_attrs), +- vis, +- sig: Signature { +- constness, +- asyncness, +- unsafety, +- abi, +- fn_token, +- ident, +- paren_token, +- inputs, +- output, +- variadic, +- generics: Generics { +- where_clause, +- ..generics +- }, +- }, +- block: Box::new(Block { brace_token, stmts }), +- }) ++ fn peek_signature(input: ParseStream) -> bool { ++ let fork = input.fork(); ++ fork.parse::<Option<Token![const]>>().is_ok() ++ && fork.parse::<Option<Token![async]>>().is_ok() ++ && fork.parse::<Option<Token![unsafe]>>().is_ok() ++ && fork.parse::<Option<Abi>>().is_ok() ++ && fork.peek(Token![fn]) ++ } ++ ++ fn parse_signature(input: ParseStream) -> Result<Signature> { ++ let constness: Option<Token![const]> = input.parse()?; ++ let asyncness: Option<Token![async]> = input.parse()?; ++ let unsafety: Option<Token![unsafe]> = input.parse()?; ++ let abi: Option<Abi> = input.parse()?; ++ let fn_token: Token![fn] = input.parse()?; ++ let ident: Ident = input.parse()?; ++ let generics: Generics = input.parse()?; ++ ++ let content; ++ let paren_token = parenthesized!(content in input); ++ let mut inputs = parse_fn_args(&content)?; ++ let variadic = pop_variadic(&mut inputs); ++ ++ let output: ReturnType = input.parse()?; ++ let where_clause: Option<WhereClause> = input.parse()?; ++ ++ Ok(Signature { ++ constness, ++ asyncness, ++ unsafety, ++ abi, ++ fn_token, ++ ident, ++ paren_token, ++ inputs, ++ output, ++ variadic, ++ generics: Generics { ++ where_clause, ++ ..generics ++ }, ++ }) ++ } ++ ++ impl Parse for ItemFn { ++ fn parse(input: ParseStream) -> Result<Self> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; ++ let vis: Visibility = input.parse()?; ++ let sig = parse_signature(input)?; ++ parse_rest_of_fn(input, outer_attrs, vis, sig) + } + } + ++ fn parse_rest_of_fn( ++ input: ParseStream, ++ outer_attrs: Vec<Attribute>, ++ vis: Visibility, ++ sig: Signature, ++ ) -> Result<ItemFn> { ++ let content; ++ let brace_token = braced!(content in input); ++ let inner_attrs = content.call(Attribute::parse_inner)?; ++ let stmts = content.call(Block::parse_within)?; ++ ++ Ok(ItemFn { ++ attrs: private::attrs(outer_attrs, inner_attrs), ++ vis, ++ sig, ++ block: Box::new(Block { brace_token, stmts }), ++ }) ++ } ++ + impl Parse for FnArg { + fn parse(input: ParseStream) -> Result<Self> { + let attrs = input.call(Attribute::parse_outer)?; +@@ -1491,26 +1434,79 @@ pub mod parsing { + } + } + ++ fn parse_fn_args(input: ParseStream) -> Result<Punctuated<FnArg, Token![,]>> { ++ let mut args = Punctuated::new(); ++ let mut has_receiver = false; ++ ++ while !input.is_empty() { ++ let attrs = input.call(Attribute::parse_outer)?; ++ ++ let arg = if let Some(dots) = input.parse::<Option<Token![...]>>()? { ++ FnArg::Typed(PatType { ++ attrs, ++ pat: Box::new(Pat::Verbatim(variadic_to_tokens(&dots))), ++ colon_token: Token![:](dots.spans[0]), ++ ty: Box::new(Type::Verbatim(variadic_to_tokens(&dots))), ++ }) ++ } else { ++ let mut arg: FnArg = input.parse()?; ++ match &mut arg { ++ FnArg::Receiver(receiver) if has_receiver => { ++ return Err(Error::new( ++ receiver.self_token.span, ++ "unexpected second method receiver", ++ )); ++ } ++ FnArg::Receiver(receiver) if !args.is_empty() => { ++ return Err(Error::new( ++ receiver.self_token.span, ++ "unexpected method receiver", ++ )); ++ } ++ FnArg::Receiver(receiver) => { ++ has_receiver = true; ++ receiver.attrs = attrs; ++ } ++ FnArg::Typed(arg) => arg.attrs = attrs, ++ } ++ arg ++ }; ++ args.push_value(arg); ++ ++ if input.is_empty() { ++ break; ++ } ++ ++ let comma: Token![,] = input.parse()?; ++ args.push_punct(comma); ++ } ++ ++ Ok(args) ++ } ++ + fn fn_arg_typed(input: ParseStream) -> Result<PatType> { ++ // Hack to parse pre-2018 syntax in ++ // test/ui/rfc-2565-param-attrs/param-attrs-pretty.rs ++ // because the rest of the test case is valuable. ++ if input.peek(Ident) && input.peek2(Token![<]) { ++ let span = input.fork().parse::<Ident>()?.span(); ++ return Ok(PatType { ++ attrs: Vec::new(), ++ pat: Box::new(Pat::Wild(PatWild { ++ attrs: Vec::new(), ++ underscore_token: Token![_](span), ++ })), ++ colon_token: Token![:](span), ++ ty: input.parse()?, ++ }); ++ } ++ + Ok(PatType { + attrs: Vec::new(), +- pat: input.parse()?, ++ pat: Box::new(pat::parsing::multi_pat(input)?), + colon_token: input.parse()?, + ty: Box::new(match input.parse::<Option<Token![...]>>()? { +- Some(dot3) => { +- let args = vec![ +- TokenTree::Punct(Punct::new('.', Spacing::Joint)), +- TokenTree::Punct(Punct::new('.', Spacing::Joint)), +- TokenTree::Punct(Punct::new('.', Spacing::Alone)), +- ]; +- let tokens = TokenStream::from_iter(args.into_iter().zip(&dot3.spans).map( +- |(mut arg, span)| { +- arg.set_span(*span); +- arg +- }, +- )); +- Type::Verbatim(tokens) +- } ++ Some(dot3) => Type::Verbatim(variadic_to_tokens(&dot3)), + None => input.parse()?, + }), + }) +@@ -1581,22 +1577,60 @@ pub mod parsing { + + impl Parse for ForeignItem { + fn parse(input: ParseStream) -> Result<Self> { ++ let begin = input.fork(); + let mut attrs = input.call(Attribute::parse_outer)?; + let ahead = input.fork(); + let vis: Visibility = ahead.parse()?; + + let lookahead = ahead.lookahead1(); +- let mut item = if lookahead.peek(Token![fn]) { +- input.parse().map(ForeignItem::Fn) ++ let mut item = if lookahead.peek(Token![fn]) || peek_signature(&ahead) { ++ let vis: Visibility = input.parse()?; ++ let sig = parse_signature(input)?; ++ if input.peek(token::Brace) { ++ let content; ++ braced!(content in input); ++ content.call(Attribute::parse_inner)?; ++ content.call(Block::parse_within)?; ++ ++ Ok(ForeignItem::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(ForeignItem::Fn(ForeignItemFn { ++ attrs: Vec::new(), ++ vis, ++ sig, ++ semi_token: input.parse()?, ++ })) ++ } + } else if lookahead.peek(Token![static]) { +- input.parse().map(ForeignItem::Static) ++ let vis = input.parse()?; ++ let static_token = input.parse()?; ++ let mutability = input.parse()?; ++ let ident = input.parse()?; ++ let colon_token = input.parse()?; ++ let ty = input.parse()?; ++ if input.peek(Token![=]) { ++ input.parse::<Token![=]>()?; ++ input.parse::<Expr>()?; ++ input.parse::<Token![;]>()?; ++ Ok(ForeignItem::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(ForeignItem::Static(ForeignItemStatic { ++ attrs: Vec::new(), ++ vis, ++ static_token, ++ mutability, ++ ident, ++ colon_token, ++ ty, ++ semi_token: input.parse()?, ++ })) ++ } + } else if lookahead.peek(Token![type]) { +- input.parse().map(ForeignItem::Type) ++ parse_foreign_item_type(begin, input) + } else if vis.is_inherited() + && (lookahead.peek(Ident) + || lookahead.peek(Token![self]) + || lookahead.peek(Token![super]) +- || lookahead.peek(Token![extern]) + || lookahead.peek(Token![crate]) + || lookahead.peek(Token![::])) + { +@@ -1605,17 +1639,16 @@ pub mod parsing { + Err(lookahead.error()) + }?; + +- { +- let item_attrs = match &mut item { +- ForeignItem::Fn(item) => &mut item.attrs, +- ForeignItem::Static(item) => &mut item.attrs, +- ForeignItem::Type(item) => &mut item.attrs, +- ForeignItem::Macro(item) => &mut item.attrs, +- ForeignItem::Verbatim(_) | ForeignItem::__Nonexhaustive => unreachable!(), +- }; +- attrs.extend(item_attrs.drain(..)); +- *item_attrs = attrs; +- } ++ let item_attrs = match &mut item { ++ ForeignItem::Fn(item) => &mut item.attrs, ++ ForeignItem::Static(item) => &mut item.attrs, ++ ForeignItem::Type(item) => &mut item.attrs, ++ ForeignItem::Macro(item) => &mut item.attrs, ++ ForeignItem::Verbatim(_) => return Ok(item), ++ ForeignItem::__Nonexhaustive => unreachable!(), ++ }; ++ attrs.extend(item_attrs.drain(..)); ++ *item_attrs = attrs; + + Ok(item) + } +@@ -1625,55 +1658,12 @@ pub mod parsing { + fn parse(input: ParseStream) -> Result<Self> { + let attrs = input.call(Attribute::parse_outer)?; + let vis: Visibility = input.parse()?; +- let fn_token: Token![fn] = input.parse()?; +- let ident: Ident = input.parse()?; +- let generics: Generics = input.parse()?; +- +- let content; +- let paren_token = parenthesized!(content in input); +- let mut inputs = Punctuated::new(); +- let mut variadic = None; +- while !content.is_empty() { +- let attrs = content.call(Attribute::parse_outer)?; +- +- if let Some(dots) = content.parse()? { +- variadic = Some(Variadic { attrs, dots }); +- break; +- } +- +- let mut arg = content.call(fn_arg_typed)?; +- arg.attrs = attrs; +- inputs.push_value(FnArg::Typed(arg)); +- if content.is_empty() { +- break; +- } +- +- inputs.push_punct(content.parse()?); +- } +- +- let output: ReturnType = input.parse()?; +- let where_clause: Option<WhereClause> = input.parse()?; ++ let sig = parse_signature(input)?; + let semi_token: Token![;] = input.parse()?; +- + Ok(ForeignItemFn { + attrs, + vis, +- sig: Signature { +- constness: None, +- asyncness: None, +- unsafety: None, +- abi: None, +- fn_token, +- ident, +- paren_token, +- inputs, +- output, +- variadic, +- generics: Generics { +- where_clause, +- ..generics +- }, +- }, ++ sig, + semi_token, + }) + } +@@ -1706,6 +1696,37 @@ pub mod parsing { + } + } + ++ fn parse_foreign_item_type(begin: ParseBuffer, input: ParseStream) -> Result<ForeignItem> { ++ let FlexibleItemType { ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds: _, ++ ty, ++ semi_token, ++ } = input.parse()?; ++ ++ if defaultness.is_some() ++ || generics.lt_token.is_some() ++ || generics.where_clause.is_some() ++ || colon_token.is_some() ++ || ty.is_some() ++ { ++ Ok(ForeignItem::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(ForeignItem::Type(ForeignItemType { ++ attrs: Vec::new(), ++ vis, ++ type_token, ++ ident, ++ semi_token, ++ })) ++ } ++ } ++ + impl Parse for ForeignItemMacro { + fn parse(input: ParseStream) -> Result<Self> { + let attrs = input.call(Attribute::parse_outer)?; +@@ -1742,6 +1763,36 @@ pub mod parsing { + } + } + ++ fn parse_item_type(begin: ParseBuffer, input: ParseStream) -> Result<Item> { ++ let FlexibleItemType { ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds: _, ++ ty, ++ semi_token, ++ } = input.parse()?; ++ ++ if defaultness.is_some() || colon_token.is_some() || ty.is_none() { ++ Ok(Item::Verbatim(verbatim::between(begin, input))) ++ } else { ++ let (eq_token, ty) = ty.unwrap(); ++ Ok(Item::Type(ItemType { ++ attrs: Vec::new(), ++ vis, ++ type_token, ++ ident, ++ generics, ++ eq_token, ++ ty: Box::new(ty), ++ semi_token, ++ })) ++ } ++ } ++ + #[cfg(not(feature = "printing"))] + fn item_existential(input: ParseStream) -> Result<TokenStream> { + Err(input.error("existential type is not supported")) +@@ -1887,7 +1938,7 @@ pub mod parsing { + + impl Parse for ItemTrait { + fn parse(input: ParseStream) -> Result<Self> { +- let attrs = input.call(Attribute::parse_outer)?; ++ let outer_attrs = input.call(Attribute::parse_outer)?; + let vis: Visibility = input.parse()?; + let unsafety: Option<Token![unsafe]> = input.parse()?; + let auto_token: Option<Token![auto]> = input.parse()?; +@@ -1896,7 +1947,7 @@ pub mod parsing { + let generics: Generics = input.parse()?; + parse_rest_of_trait( + input, +- attrs, ++ outer_attrs, + vis, + unsafety, + auto_token, +@@ -1909,7 +1960,7 @@ pub mod parsing { + + fn parse_rest_of_trait( + input: ParseStream, +- attrs: Vec<Attribute>, ++ outer_attrs: Vec<Attribute>, + vis: Visibility, + unsafety: Option<Token![unsafe]>, + auto_token: Option<Token![auto]>, +@@ -1937,13 +1988,14 @@ pub mod parsing { + + let content; + let brace_token = braced!(content in input); ++ let inner_attrs = content.call(Attribute::parse_inner)?; + let mut items = Vec::new(); + while !content.is_empty() { + items.push(content.parse()?); + } + + Ok(ItemTrait { +- attrs, ++ attrs: private::attrs(outer_attrs, inner_attrs), + vis, + unsafety, + auto_token, +@@ -2014,14 +2066,19 @@ pub mod parsing { + + impl Parse for TraitItem { + fn parse(input: ParseStream) -> Result<Self> { ++ let begin = input.fork(); + let mut attrs = input.call(Attribute::parse_outer)?; ++ let vis: Visibility = input.parse()?; ++ let defaultness: Option<Token![default]> = input.parse()?; + let ahead = input.fork(); + + let lookahead = ahead.lookahead1(); +- let mut item = if lookahead.peek(Token![const]) { ++ let mut item = if lookahead.peek(Token![fn]) || peek_signature(&ahead) { ++ input.parse().map(TraitItem::Method) ++ } else if lookahead.peek(Token![const]) { + ahead.parse::<Token![const]>()?; + let lookahead = ahead.lookahead1(); +- if lookahead.peek(Ident) { ++ if lookahead.peek(Ident) || lookahead.peek(Token![_]) { + input.parse().map(TraitItem::Const) + } else if lookahead.peek(Token![async]) + || lookahead.peek(Token![unsafe]) +@@ -2032,18 +2089,11 @@ pub mod parsing { + } else { + Err(lookahead.error()) + } +- } else if lookahead.peek(Token![async]) +- || lookahead.peek(Token![unsafe]) +- || lookahead.peek(Token![extern]) +- || lookahead.peek(Token![fn]) +- { +- input.parse().map(TraitItem::Method) + } else if lookahead.peek(Token![type]) { +- input.parse().map(TraitItem::Type) ++ parse_trait_item_type(begin.fork(), input) + } else if lookahead.peek(Ident) + || lookahead.peek(Token![self]) + || lookahead.peek(Token![super]) +- || lookahead.peek(Token![extern]) + || lookahead.peek(Token![crate]) + || lookahead.peek(Token![::]) + { +@@ -2052,18 +2102,20 @@ pub mod parsing { + Err(lookahead.error()) + }?; + +- { +- let item_attrs = match &mut item { +- TraitItem::Const(item) => &mut item.attrs, +- TraitItem::Method(item) => &mut item.attrs, +- TraitItem::Type(item) => &mut item.attrs, +- TraitItem::Macro(item) => &mut item.attrs, +- TraitItem::Verbatim(_) | TraitItem::__Nonexhaustive => unreachable!(), +- }; +- attrs.extend(item_attrs.drain(..)); +- *item_attrs = attrs; ++ match (vis, defaultness) { ++ (Visibility::Inherited, None) => {} ++ _ => return Ok(TraitItem::Verbatim(verbatim::between(begin, input))), + } + ++ let item_attrs = match &mut item { ++ TraitItem::Const(item) => &mut item.attrs, ++ TraitItem::Method(item) => &mut item.attrs, ++ TraitItem::Type(item) => &mut item.attrs, ++ TraitItem::Macro(item) => &mut item.attrs, ++ TraitItem::Verbatim(_) | TraitItem::__Nonexhaustive => unreachable!(), ++ }; ++ attrs.extend(item_attrs.drain(..)); ++ *item_attrs = attrs; + Ok(item) + } + } +@@ -2073,7 +2125,14 @@ pub mod parsing { + Ok(TraitItemConst { + attrs: input.call(Attribute::parse_outer)?, + const_token: input.parse()?, +- ident: input.parse()?, ++ ident: { ++ let lookahead = input.lookahead1(); ++ if lookahead.peek(Ident) || lookahead.peek(Token![_]) { ++ input.call(Ident::parse_any)? ++ } else { ++ return Err(lookahead.error()); ++ } ++ }, + colon_token: input.parse()?, + ty: input.parse()?, + default: { +@@ -2093,20 +2152,7 @@ pub mod parsing { + impl Parse for TraitItemMethod { + fn parse(input: ParseStream) -> Result<Self> { + let outer_attrs = input.call(Attribute::parse_outer)?; +- let constness: Option<Token![const]> = input.parse()?; +- let asyncness: Option<Token![async]> = input.parse()?; +- let unsafety: Option<Token![unsafe]> = input.parse()?; +- let abi: Option<Abi> = input.parse()?; +- let fn_token: Token![fn] = input.parse()?; +- let ident: Ident = input.parse()?; +- let generics: Generics = input.parse()?; +- +- let content; +- let paren_token = parenthesized!(content in input); +- let inputs = content.parse_terminated(FnArg::parse)?; +- +- let output: ReturnType = input.parse()?; +- let where_clause: Option<WhereClause> = input.parse()?; ++ let sig = parse_signature(input)?; + + let lookahead = input.lookahead1(); + let (brace_token, inner_attrs, stmts, semi_token) = if lookahead.peek(token::Brace) { +@@ -2124,22 +2170,7 @@ pub mod parsing { + + Ok(TraitItemMethod { + attrs: private::attrs(outer_attrs, inner_attrs), +- sig: Signature { +- constness, +- asyncness, +- unsafety, +- abi, +- fn_token, +- ident, +- paren_token, +- inputs, +- output, +- variadic: None, +- generics: Generics { +- where_clause, +- ..generics +- }, +- }, ++ sig, + default: brace_token.map(|brace_token| Block { brace_token, stmts }), + semi_token, + }) +@@ -2188,6 +2219,35 @@ pub mod parsing { + } + } + ++ fn parse_trait_item_type(begin: ParseBuffer, input: ParseStream) -> Result<TraitItem> { ++ let FlexibleItemType { ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds, ++ ty, ++ semi_token, ++ } = input.parse()?; ++ ++ if defaultness.is_some() || vis.is_some() { ++ Ok(TraitItem::Verbatim(verbatim::between(begin, input))) ++ } else { ++ Ok(TraitItem::Type(TraitItemType { ++ attrs: Vec::new(), ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds, ++ default: ty, ++ semi_token, ++ })) ++ } ++ } ++ + impl Parse for TraitItemMacro { + fn parse(input: ParseStream) -> Result<Self> { + let attrs = input.call(Attribute::parse_outer)?; +@@ -2207,52 +2267,67 @@ pub mod parsing { + + impl Parse for ItemImpl { + fn parse(input: ParseStream) -> Result<Self> { +- let outer_attrs = input.call(Attribute::parse_outer)?; +- let defaultness: Option<Token![default]> = input.parse()?; +- let unsafety: Option<Token![unsafe]> = input.parse()?; +- let impl_token: Token![impl] = input.parse()?; +- +- let has_generics = input.peek(Token![<]) +- && (input.peek2(Token![>]) +- || input.peek2(Token![#]) +- || (input.peek2(Ident) || input.peek2(Lifetime)) +- && (input.peek3(Token![:]) +- || input.peek3(Token![,]) +- || input.peek3(Token![>]))); +- let generics: Generics = if has_generics { +- input.parse()? +- } else { +- Generics::default() +- }; +- +- let trait_ = { +- // TODO: optimize using advance_to +- let ahead = input.fork(); +- if ahead.parse::<Option<Token![!]>>().is_ok() +- && ahead.parse::<Path>().is_ok() +- && ahead.parse::<Token![for]>().is_ok() +- { +- let polarity: Option<Token![!]> = input.parse()?; +- let path: Path = input.parse()?; +- let for_token: Token![for] = input.parse()?; +- Some((polarity, path, for_token)) +- } else { +- None +- } +- }; +- let self_ty: Type = input.parse()?; +- let where_clause: Option<WhereClause> = input.parse()?; ++ let allow_const_impl = false; ++ parse_impl(input, allow_const_impl).map(Option::unwrap) ++ } ++ } ++ ++ fn parse_impl(input: ParseStream, allow_const_impl: bool) -> Result<Option<ItemImpl>> { ++ let outer_attrs = input.call(Attribute::parse_outer)?; ++ let defaultness: Option<Token![default]> = input.parse()?; ++ let unsafety: Option<Token![unsafe]> = input.parse()?; ++ let impl_token: Token![impl] = input.parse()?; ++ ++ let has_generics = input.peek(Token![<]) ++ && (input.peek2(Token![>]) ++ || input.peek2(Token![#]) ++ || (input.peek2(Ident) || input.peek2(Lifetime)) ++ && (input.peek3(Token![:]) ++ || input.peek3(Token![,]) ++ || input.peek3(Token![>])) ++ || input.peek2(Token![const])); ++ let generics: Generics = if has_generics { ++ input.parse()? ++ } else { ++ Generics::default() ++ }; + +- let content; +- let brace_token = braced!(content in input); +- let inner_attrs = content.call(Attribute::parse_inner)?; ++ let is_const_impl = allow_const_impl ++ && (input.peek(Token![const]) || input.peek(Token![?]) && input.peek2(Token![const])); ++ if is_const_impl { ++ input.parse::<Option<Token![?]>>()?; ++ input.parse::<Token![const]>()?; ++ } + +- let mut items = Vec::new(); +- while !content.is_empty() { +- items.push(content.parse()?); ++ let trait_ = (|| -> Option<_> { ++ let ahead = input.fork(); ++ let polarity: Option<Token![!]> = ahead.parse().ok()?; ++ let mut path: Path = ahead.parse().ok()?; ++ if path.segments.last().unwrap().arguments.is_empty() && ahead.peek(token::Paren) { ++ let parenthesized = PathArguments::Parenthesized(ahead.parse().ok()?); ++ path.segments.last_mut().unwrap().arguments = parenthesized; + } ++ let for_token: Token![for] = ahead.parse().ok()?; ++ input.advance_to(&ahead); ++ Some((polarity, path, for_token)) ++ })(); ++ ++ let self_ty: Type = input.parse()?; ++ let where_clause: Option<WhereClause> = input.parse()?; ++ ++ let content; ++ let brace_token = braced!(content in input); ++ let inner_attrs = content.call(Attribute::parse_inner)?; ++ ++ let mut items = Vec::new(); ++ while !content.is_empty() { ++ items.push(content.parse()?); ++ } + +- Ok(ItemImpl { ++ if is_const_impl { ++ Ok(None) ++ } else { ++ Ok(Some(ItemImpl { + attrs: private::attrs(outer_attrs, inner_attrs), + defaultness, + unsafety, +@@ -2265,12 +2340,13 @@ pub mod parsing { + self_ty: Box::new(self_ty), + brace_token, + items, +- }) ++ })) + } + } + + impl Parse for ImplItem { + fn parse(input: ParseStream) -> Result<Self> { ++ let begin = input.fork(); + let mut attrs = input.call(Attribute::parse_outer)?; + let ahead = input.fork(); + let vis: Visibility = ahead.parse()?; +@@ -2284,28 +2360,38 @@ pub mod parsing { + None + }; + +- let mut item = if lookahead.peek(Token![const]) { +- ahead.parse::<Token![const]>()?; ++ let mut item = if lookahead.peek(Token![fn]) || peek_signature(&ahead) { ++ input.parse().map(ImplItem::Method) ++ } else if lookahead.peek(Token![const]) { ++ let const_token: Token![const] = ahead.parse()?; + let lookahead = ahead.lookahead1(); +- if lookahead.peek(Ident) { +- input.parse().map(ImplItem::Const) +- } else if lookahead.peek(Token![unsafe]) +- || lookahead.peek(Token![async]) +- || lookahead.peek(Token![extern]) +- || lookahead.peek(Token![fn]) +- { +- input.parse().map(ImplItem::Method) ++ if lookahead.peek(Ident) || lookahead.peek(Token![_]) { ++ input.advance_to(&ahead); ++ let ident: Ident = input.call(Ident::parse_any)?; ++ let colon_token: Token![:] = input.parse()?; ++ let ty: Type = input.parse()?; ++ if let Some(eq_token) = input.parse()? { ++ return Ok(ImplItem::Const(ImplItemConst { ++ attrs, ++ vis, ++ defaultness, ++ const_token, ++ ident, ++ colon_token, ++ ty, ++ eq_token, ++ expr: input.parse()?, ++ semi_token: input.parse()?, ++ })); ++ } else { ++ input.parse::<Token![;]>()?; ++ return Ok(ImplItem::Verbatim(verbatim::between(begin, input))); ++ } + } else { + Err(lookahead.error()) + } +- } else if lookahead.peek(Token![unsafe]) +- || lookahead.peek(Token![async]) +- || lookahead.peek(Token![extern]) +- || lookahead.peek(Token![fn]) +- { +- input.parse().map(ImplItem::Method) + } else if lookahead.peek(Token![type]) { +- input.parse().map(ImplItem::Type) ++ parse_impl_item_type(begin, input) + } else if vis.is_inherited() && defaultness.is_none() && lookahead.peek(existential) { + input.call(item_existential).map(ImplItem::Verbatim) + } else if vis.is_inherited() +@@ -2313,7 +2399,6 @@ pub mod parsing { + && (lookahead.peek(Ident) + || lookahead.peek(Token![self]) + || lookahead.peek(Token![super]) +- || lookahead.peek(Token![extern]) + || lookahead.peek(Token![crate]) + || lookahead.peek(Token![::])) + { +@@ -2346,7 +2431,14 @@ pub mod parsing { + vis: input.parse()?, + defaultness: input.parse()?, + const_token: input.parse()?, +- ident: input.parse()?, ++ ident: { ++ let lookahead = input.lookahead1(); ++ if lookahead.peek(Ident) || lookahead.peek(Token![_]) { ++ input.call(Ident::parse_any)? ++ } else { ++ return Err(lookahead.error()); ++ } ++ }, + colon_token: input.parse()?, + ty: input.parse()?, + eq_token: input.parse()?, +@@ -2358,50 +2450,39 @@ pub mod parsing { + + impl Parse for ImplItemMethod { + fn parse(input: ParseStream) -> Result<Self> { +- let outer_attrs = input.call(Attribute::parse_outer)?; ++ let mut attrs = input.call(Attribute::parse_outer)?; + let vis: Visibility = input.parse()?; + let defaultness: Option<Token![default]> = input.parse()?; +- let constness: Option<Token![const]> = input.parse()?; +- let asyncness: Option<Token![async]> = input.parse()?; +- let unsafety: Option<Token![unsafe]> = input.parse()?; +- let abi: Option<Abi> = input.parse()?; +- let fn_token: Token![fn] = input.parse()?; +- let ident: Ident = input.parse()?; +- let generics: Generics = input.parse()?; +- +- let content; +- let paren_token = parenthesized!(content in input); +- let inputs = content.parse_terminated(FnArg::parse)?; +- +- let output: ReturnType = input.parse()?; +- let where_clause: Option<WhereClause> = input.parse()?; +- +- let content; +- let brace_token = braced!(content in input); +- let inner_attrs = content.call(Attribute::parse_inner)?; +- let stmts = content.call(Block::parse_within)?; ++ let sig = parse_signature(input)?; ++ ++ let block = if let Some(semi) = input.parse::<Option<Token![;]>>()? { ++ // Accept methods without a body in an impl block because ++ // rustc's *parser* does not reject them (the compilation error ++ // is emitted later than parsing) and it can be useful for macro ++ // DSLs. ++ let mut punct = Punct::new(';', Spacing::Alone); ++ punct.set_span(semi.span); ++ let tokens = TokenStream::from_iter(vec![TokenTree::Punct(punct)]); ++ Block { ++ brace_token: Brace::default(), ++ stmts: vec![Stmt::Item(Item::Verbatim(tokens))], ++ } ++ } else { ++ let content; ++ let brace_token = braced!(content in input); ++ attrs.extend(content.call(Attribute::parse_inner)?); ++ Block { ++ brace_token, ++ stmts: content.call(Block::parse_within)?, ++ } ++ }; + + Ok(ImplItemMethod { +- attrs: private::attrs(outer_attrs, inner_attrs), ++ attrs, + vis, + defaultness, +- sig: Signature { +- constness, +- asyncness, +- unsafety, +- abi, +- fn_token, +- ident, +- paren_token, +- inputs, +- output, +- variadic: None, +- generics: Generics { +- where_clause, +- ..generics +- }, +- }, +- block: Block { brace_token, stmts }, ++ sig, ++ block, + }) + } + } +@@ -2426,6 +2507,37 @@ pub mod parsing { + } + } + ++ fn parse_impl_item_type(begin: ParseBuffer, input: ParseStream) -> Result<ImplItem> { ++ let FlexibleItemType { ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ colon_token, ++ bounds: _, ++ ty, ++ semi_token, ++ } = input.parse()?; ++ ++ if colon_token.is_some() || ty.is_none() { ++ Ok(ImplItem::Verbatim(verbatim::between(begin, input))) ++ } else { ++ let (eq_token, ty) = ty.unwrap(); ++ Ok(ImplItem::Type(ImplItemType { ++ attrs: Vec::new(), ++ vis, ++ defaultness, ++ type_token, ++ ident, ++ generics, ++ eq_token, ++ ty, ++ semi_token, ++ })) ++ } ++ } ++ + impl Parse for ImplItemMacro { + fn parse(input: ParseStream) -> Result<Self> { + let attrs = input.call(Attribute::parse_outer)?; +@@ -2471,6 +2583,7 @@ mod printing { + + use crate::attr::FilterAttrs; + use crate::print::TokensOrDefault; ++ use crate::punctuated::Pair; + + impl ToTokens for ItemExternCrate { + fn to_tokens(&self, tokens: &mut TokenStream) { +@@ -2835,6 +2948,14 @@ mod printing { + self.vis.to_tokens(tokens); + self.defaultness.to_tokens(tokens); + self.sig.to_tokens(tokens); ++ if self.block.stmts.len() == 1 { ++ if let Stmt::Item(Item::Verbatim(verbatim)) = &self.block.stmts[0] { ++ if verbatim.to_string() == ";" { ++ verbatim.to_tokens(tokens); ++ return; ++ } ++ } ++ } + self.block.brace_token.surround(tokens, |tokens| { + tokens.append_all(self.attrs.inner()); + tokens.append_all(&self.block.stmts); +@@ -2905,6 +3026,33 @@ mod printing { + } + } + ++ fn maybe_variadic_to_tokens(arg: &FnArg, tokens: &mut TokenStream) -> bool { ++ let arg = match arg { ++ FnArg::Typed(arg) => arg, ++ FnArg::Receiver(receiver) => { ++ receiver.to_tokens(tokens); ++ return false; ++ } ++ }; ++ ++ match arg.ty.as_ref() { ++ Type::Verbatim(ty) if ty.to_string() == "..." => { ++ match arg.pat.as_ref() { ++ Pat::Verbatim(pat) if pat.to_string() == "..." => { ++ tokens.append_all(arg.attrs.outer()); ++ pat.to_tokens(tokens); ++ } ++ _ => arg.to_tokens(tokens), ++ } ++ true ++ } ++ _ => { ++ arg.to_tokens(tokens); ++ false ++ } ++ } ++ } ++ + impl ToTokens for Signature { + fn to_tokens(&self, tokens: &mut TokenStream) { + self.constness.to_tokens(tokens); +@@ -2915,11 +3063,24 @@ mod printing { + self.ident.to_tokens(tokens); + self.generics.to_tokens(tokens); + self.paren_token.surround(tokens, |tokens| { +- self.inputs.to_tokens(tokens); +- if self.variadic.is_some() && !self.inputs.empty_or_trailing() { +- <Token![,]>::default().to_tokens(tokens); ++ let mut last_is_variadic = false; ++ for input in self.inputs.pairs() { ++ match input { ++ Pair::Punctuated(input, comma) => { ++ maybe_variadic_to_tokens(input, tokens); ++ comma.to_tokens(tokens); ++ } ++ Pair::End(input) => { ++ last_is_variadic = maybe_variadic_to_tokens(input, tokens); ++ } ++ } ++ } ++ if self.variadic.is_some() && !last_is_variadic { ++ if !self.inputs.empty_or_trailing() { ++ <Token![,]>::default().to_tokens(tokens); ++ } ++ self.variadic.to_tokens(tokens); + } +- self.variadic.to_tokens(tokens); + }); + self.output.to_tokens(tokens); + self.generics.where_clause.to_tokens(tokens); +diff --git a/third_party/rust/syn/src/keyword.rs b/third_party/rust/syn/src/keyword.rs +deleted file mode 100644 +index e69de29bb2..0000000000 +diff --git a/third_party/rust/syn/src/lib.rs b/third_party/rust/syn/src/lib.rs +index c8ada7e638..3da506731e 100644 +--- a/third_party/rust/syn/src/lib.rs ++++ b/third_party/rust/syn/src/lib.rs +@@ -1,3 +1,11 @@ ++//! [![github]](https://github.com/dtolnay/syn) [![crates-io]](https://crates.io/crates/syn) [![docs-rs]](https://docs.rs/syn) ++//! ++//! [github]: https://img.shields.io/badge/github-8da0cb?style=for-the-badge&labelColor=555555&logo=github ++//! [crates-io]: https://img.shields.io/badge/crates.io-fc8d62?style=for-the-badge&labelColor=555555&logo=rust ++//! [docs-rs]: https://img.shields.io/badge/docs.rs-66c2a5?style=for-the-badge&labelColor=555555&logoColor=white&logo=data:image/svg+xml;base64,PHN2ZyByb2xlPSJpbWciIHhtbG5zPSJodHRwOi8vd3d3LnczLm9yZy8yMDAwL3N2ZyIgdmlld0JveD0iMCAwIDUxMiA1MTIiPjxwYXRoIGZpbGw9IiNmNWY1ZjUiIGQ9Ik00ODguNiAyNTAuMkwzOTIgMjE0VjEwNS41YzAtMTUtOS4zLTI4LjQtMjMuNC0zMy43bC0xMDAtMzcuNWMtOC4xLTMuMS0xNy4xLTMuMS0yNS4zIDBsLTEwMCAzNy41Yy0xNC4xIDUuMy0yMy40IDE4LjctMjMuNCAzMy43VjIxNGwtOTYuNiAzNi4yQzkuMyAyNTUuNSAwIDI2OC45IDAgMjgzLjlWMzk0YzAgMTMuNiA3LjcgMjYuMSAxOS45IDMyLjJsMTAwIDUwYzEwLjEgNS4xIDIyLjEgNS4xIDMyLjIgMGwxMDMuOS01MiAxMDMuOSA1MmMxMC4xIDUuMSAyMi4xIDUuMSAzMi4yIDBsMTAwLTUwYzEyLjItNi4xIDE5LjktMTguNiAxOS45LTMyLjJWMjgzLjljMC0xNS05LjMtMjguNC0yMy40LTMzLjd6TTM1OCAyMTQuOGwtODUgMzEuOXYtNjguMmw4NS0zN3Y3My4zek0xNTQgMTA0LjFsMTAyLTM4LjIgMTAyIDM4LjJ2LjZsLTEwMiA0MS40LTEwMi00MS40di0uNnptODQgMjkxLjFsLTg1IDQyLjV2LTc5LjFsODUtMzguOHY3NS40em0wLTExMmwtMTAyIDQxLjQtMTAyLTQxLjR2LS42bDEwMi0zOC4yIDEwMiAzOC4ydi42em0yNDAgMTEybC04NSA0Mi41di03OS4xbDg1LTM4Ljh2NzUuNHptMC0xMTJsLTEwMiA0MS40LTEwMi00MS40di0uNmwxMDItMzguMiAxMDIgMzguMnYuNnoiPjwvcGF0aD48L3N2Zz4K ++//! ++//! <br> ++//! + //! Syn is a parsing library for parsing a stream of Rust tokens into a syntax + //! tree of Rust source code. + //! +@@ -62,8 +70,8 @@ + //! ``` + //! + //! ``` +-//! extern crate proc_macro; +-//! ++//! # extern crate proc_macro; ++//! # + //! use proc_macro::TokenStream; + //! use quote::quote; + //! use syn::{parse_macro_input, DeriveInput}; +@@ -242,35 +250,48 @@ + //! dynamic library libproc_macro from rustc toolchain. + + // Syn types in rustdoc of other crates get linked to here. +-#![doc(html_root_url = "https://docs.rs/syn/1.0.5")] ++#![doc(html_root_url = "https://docs.rs/syn/1.0.40")] + #![deny(clippy::all, clippy::pedantic)] + // Ignored clippy lints. + #![allow( +- clippy::block_in_if_condition_stmt, ++ clippy::blocks_in_if_conditions, + clippy::cognitive_complexity, + clippy::doc_markdown, + clippy::eval_order_dependence, + clippy::inherent_to_string, + clippy::large_enum_variant, ++ clippy::manual_non_exhaustive, ++ clippy::match_like_matches_macro, ++ clippy::match_on_vec_items, ++ clippy::needless_doctest_main, + clippy::needless_pass_by_value, + clippy::never_loop, + clippy::suspicious_op_assign_impl, + clippy::too_many_arguments, +- clippy::trivially_copy_pass_by_ref ++ clippy::trivially_copy_pass_by_ref, ++ clippy::unnecessary_unwrap + )] + // Ignored clippy_pedantic lints. + #![allow( + clippy::cast_possible_truncation, ++ clippy::default_trait_access, + clippy::empty_enum, ++ clippy::expl_impl_clone_on_copy, + clippy::if_not_else, + clippy::items_after_statements, ++ clippy::match_same_arms, ++ clippy::missing_errors_doc, + clippy::module_name_repetitions, ++ clippy::must_use_candidate, ++ clippy::option_if_let_else, + clippy::shadow_unrelated, + clippy::similar_names, + clippy::single_match_else, ++ clippy::too_many_lines, + clippy::unseparated_literal_suffix, + clippy::use_self, +- clippy::used_underscore_binding ++ clippy::used_underscore_binding, ++ clippy::wildcard_imports + )] + + #[cfg(all( +@@ -284,7 +305,6 @@ extern crate unicode_xid; + #[cfg(feature = "printing")] + extern crate quote; + +-#[cfg(any(feature = "full", feature = "derive"))] + #[macro_use] + mod macros; + +@@ -307,7 +327,6 @@ pub use crate::attr::{ + AttrStyle, Attribute, AttributeArgs, Meta, MetaList, MetaNameValue, NestedMeta, + }; + +-#[cfg(any(feature = "full", feature = "derive"))] + mod bigint; + + #[cfg(any(feature = "full", feature = "derive"))] +@@ -364,9 +383,7 @@ pub use crate::file::File; + mod lifetime; + pub use crate::lifetime::Lifetime; + +-#[cfg(any(feature = "full", feature = "derive"))] + mod lit; +-#[cfg(any(feature = "full", feature = "derive"))] + pub use crate::lit::{ + Lit, LitBool, LitByte, LitByteStr, LitChar, LitFloat, LitInt, LitStr, StrStyle, + }; +@@ -441,6 +458,9 @@ pub mod parse_macro_input; + #[cfg(all(feature = "parsing", feature = "printing"))] + pub mod spanned; + ++#[cfg(all(feature = "parsing", feature = "full"))] ++mod whitespace; ++ + mod gen { + /// Syntax tree traversal to walk a shared borrow of a syntax tree. + /// +@@ -482,7 +502,7 @@ mod gen { + /// /* ... */ + /// ``` + /// +- /// *This module is available if Syn is built with the `"visit"` feature.* ++ /// *This module is available only if Syn is built with the `"visit"` feature.* + /// + /// <br> + /// +@@ -603,7 +623,7 @@ mod gen { + /// /* ... */ + /// ``` + /// +- /// *This module is available if Syn is built with the `"visit-mut"` ++ /// *This module is available only if Syn is built with the `"visit-mut"` + /// feature.* + /// + /// <br> +@@ -702,7 +722,7 @@ mod gen { + /// /* ... */ + /// ``` + /// +- /// *This module is available if Syn is built with the `"fold"` feature.* ++ /// *This module is available only if Syn is built with the `"fold"` feature.* + /// + /// <br> + /// +@@ -744,6 +764,22 @@ mod gen { + #[rustfmt::skip] + pub mod fold; + ++ #[cfg(feature = "clone-impls")] ++ #[rustfmt::skip] ++ mod clone; ++ ++ #[cfg(feature = "extra-traits")] ++ #[rustfmt::skip] ++ mod eq; ++ ++ #[cfg(feature = "extra-traits")] ++ #[rustfmt::skip] ++ mod hash; ++ ++ #[cfg(feature = "extra-traits")] ++ #[rustfmt::skip] ++ mod debug; ++ + #[cfg(any(feature = "full", feature = "derive"))] + #[path = "../gen_helper.rs"] + mod helper; +@@ -757,6 +793,8 @@ pub mod export; + mod custom_keyword; + mod custom_punctuation; + mod sealed; ++mod span; ++mod thread; + + #[cfg(feature = "parsing")] + mod lookahead; +@@ -764,13 +802,15 @@ mod lookahead; + #[cfg(feature = "parsing")] + pub mod parse; + +-mod span; ++#[cfg(feature = "full")] ++mod reserved; ++ ++#[cfg(all(any(feature = "full", feature = "derive"), feature = "parsing"))] ++mod verbatim; + + #[cfg(all(any(feature = "full", feature = "derive"), feature = "printing"))] + mod print; + +-mod thread; +- + //////////////////////////////////////////////////////////////////////////////// + + #[allow(dead_code, non_camel_case_types)] +@@ -800,14 +840,14 @@ pub use crate::error::{Error, Result}; + /// + /// [`syn::parse2`]: parse2 + /// +-/// *This function is available if Syn is built with both the `"parsing"` and ++/// *This function is available only if Syn is built with both the `"parsing"` and + /// `"proc-macro"` features.* + /// + /// # Examples + /// + /// ``` +-/// extern crate proc_macro; +-/// ++/// # extern crate proc_macro; ++/// # + /// use proc_macro::TokenStream; + /// use quote::quote; + /// use syn::DeriveInput; +@@ -847,7 +887,7 @@ pub fn parse<T: parse::Parse>(tokens: proc_macro::TokenStream) -> Result<T> { + /// + /// [`syn::parse`]: parse() + /// +-/// *This function is available if Syn is built with the `"parsing"` feature.* ++/// *This function is available only if Syn is built with the `"parsing"` feature.* + #[cfg(feature = "parsing")] + pub fn parse2<T: parse::Parse>(tokens: proc_macro2::TokenStream) -> Result<T> { + parse::Parser::parse2(T::parse, tokens) +@@ -855,7 +895,7 @@ pub fn parse2<T: parse::Parse>(tokens: proc_macro2::TokenStream) -> Result<T> { + + /// Parse a string of Rust code into the chosen syntax tree node. + /// +-/// *This function is available if Syn is built with the `"parsing"` feature.* ++/// *This function is available only if Syn is built with the `"parsing"` feature.* + /// + /// # Hygiene + /// +@@ -874,9 +914,7 @@ pub fn parse2<T: parse::Parse>(tokens: proc_macro2::TokenStream) -> Result<T> { + /// Ok(()) + /// } + /// # +-/// # fn main() { +-/// # run().unwrap(); +-/// # } ++/// # run().unwrap(); + /// ``` + #[cfg(feature = "parsing")] + pub fn parse_str<T: parse::Parse>(s: &str) -> Result<T> { +@@ -894,7 +932,7 @@ pub fn parse_str<T: parse::Parse>(s: &str) -> Result<T> { + /// + /// If present, either of these would be an error using `from_str`. + /// +-/// *This function is available if Syn is built with the `"parsing"` and ++/// *This function is available only if Syn is built with the `"parsing"` and + /// `"full"` features.* + /// + /// # Examples +@@ -918,9 +956,7 @@ pub fn parse_str<T: parse::Parse>(s: &str) -> Result<T> { + /// Ok(()) + /// } + /// # +-/// # fn main() { +-/// # run().unwrap(); +-/// # } ++/// # run().unwrap(); + /// ``` + #[cfg(all(feature = "parsing", feature = "full"))] + pub fn parse_file(mut content: &str) -> Result<File> { +@@ -931,13 +967,16 @@ pub fn parse_file(mut content: &str) -> Result<File> { + } + + let mut shebang = None; +- if content.starts_with("#!") && !content.starts_with("#![") { +- if let Some(idx) = content.find('\n') { +- shebang = Some(content[..idx].to_string()); +- content = &content[idx..]; +- } else { +- shebang = Some(content.to_string()); +- content = ""; ++ if content.starts_with("#!") { ++ let rest = whitespace::skip(&content[2..]); ++ if !rest.starts_with('[') { ++ if let Some(idx) = content.find('\n') { ++ shebang = Some(content[..idx].to_string()); ++ content = &content[idx..]; ++ } else { ++ shebang = Some(content.to_string()); ++ content = ""; ++ } + } + } + +diff --git a/third_party/rust/syn/src/lifetime.rs b/third_party/rust/syn/src/lifetime.rs +index d51c48e827..959cc5f9c6 100644 +--- a/third_party/rust/syn/src/lifetime.rs ++++ b/third_party/rust/syn/src/lifetime.rs +@@ -18,10 +18,8 @@ use crate::lookahead; + /// - All following characters must be Unicode code points with the XID_Continue + /// property. + /// +-/// *This type is available if Syn is built with the `"derive"` or `"full"` ++/// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +-#[cfg_attr(feature = "extra-traits", derive(Debug))] +-#[derive(Clone)] + pub struct Lifetime { + pub apostrophe: Span, + pub ident: Ident, +@@ -72,6 +70,15 @@ impl Display for Lifetime { + } + } + ++impl Clone for Lifetime { ++ fn clone(&self) -> Self { ++ Lifetime { ++ apostrophe: self.apostrophe, ++ ident: self.ident.clone(), ++ } ++ } ++} ++ + impl PartialEq for Lifetime { + fn eq(&self, other: &Lifetime) -> bool { + self.ident.eq(&other.ident) +diff --git a/third_party/rust/syn/src/lit.rs b/third_party/rust/syn/src/lit.rs +index f2209a2980..ee77e75bec 100644 +--- a/third_party/rust/syn/src/lit.rs ++++ b/third_party/rust/syn/src/lit.rs +@@ -22,9 +22,6 @@ use crate::{Error, Result}; + ast_enum_of_structs! { + /// A Rust literal such as a string or integer or boolean. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` +- /// feature.* +- /// + /// # Syntax tree enum + /// + /// This type is a [syntax tree enum]. +@@ -33,7 +30,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum Lit #manual_extra_traits { ++ pub enum Lit { + /// A UTF-8 string literal: `"foo"`. + Str(LitStr), + +@@ -64,61 +61,44 @@ ast_enum_of_structs! { + + ast_struct! { + /// A UTF-8 string literal: `"foo"`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitStr #manual_extra_traits_debug { +- repr: Box<LitStrRepr>, ++ pub struct LitStr { ++ repr: Box<LitRepr>, + } + } + +-#[cfg_attr(feature = "clone-impls", derive(Clone))] +-struct LitStrRepr { +- token: Literal, +- suffix: Box<str>, +-} +- + ast_struct! { + /// A byte string literal: `b"foo"`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitByteStr #manual_extra_traits_debug { +- token: Literal, ++ pub struct LitByteStr { ++ repr: Box<LitRepr>, + } + } + + ast_struct! { + /// A byte literal: `b'f'`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitByte #manual_extra_traits_debug { +- token: Literal, ++ pub struct LitByte { ++ repr: Box<LitRepr>, + } + } + + ast_struct! { + /// A character literal: `'a'`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitChar #manual_extra_traits_debug { +- token: Literal, ++ pub struct LitChar { ++ repr: Box<LitRepr>, + } + } + ++struct LitRepr { ++ token: Literal, ++ suffix: Box<str>, ++} ++ + ast_struct! { + /// An integer literal: `1` or `1u16`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitInt #manual_extra_traits_debug { ++ pub struct LitInt { + repr: Box<LitIntRepr>, + } + } + +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + struct LitIntRepr { + token: Literal, + digits: Box<str>, +@@ -129,15 +109,11 @@ ast_struct! { + /// A floating point literal: `1f64` or `1.0e10f64`. + /// + /// Must be finite. May not be infinte or NaN. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitFloat #manual_extra_traits_debug { ++ pub struct LitFloat { + repr: Box<LitFloatRepr>, + } + } + +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + struct LitFloatRepr { + token: Literal, + digits: Box<str>, +@@ -146,92 +122,27 @@ struct LitFloatRepr { + + ast_struct! { + /// A boolean literal: `true` or `false`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or +- /// `"full"` feature.* +- pub struct LitBool #manual_extra_traits_debug { ++ pub struct LitBool { + pub value: bool, + pub span: Span, + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Lit {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Lit { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (Lit::Str(this), Lit::Str(other)) => this == other, +- (Lit::ByteStr(this), Lit::ByteStr(other)) => this == other, +- (Lit::Byte(this), Lit::Byte(other)) => this == other, +- (Lit::Char(this), Lit::Char(other)) => this == other, +- (Lit::Int(this), Lit::Int(other)) => this == other, +- (Lit::Float(this), Lit::Float(other)) => this == other, +- (Lit::Bool(this), Lit::Bool(other)) => this == other, +- (Lit::Verbatim(this), Lit::Verbatim(other)) => this.to_string() == other.to_string(), +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Lit { +- fn hash<H>(&self, hash: &mut H) +- where +- H: Hasher, +- { +- match self { +- Lit::Str(lit) => { +- hash.write_u8(0); +- lit.hash(hash); +- } +- Lit::ByteStr(lit) => { +- hash.write_u8(1); +- lit.hash(hash); +- } +- Lit::Byte(lit) => { +- hash.write_u8(2); +- lit.hash(hash); +- } +- Lit::Char(lit) => { +- hash.write_u8(3); +- lit.hash(hash); +- } +- Lit::Int(lit) => { +- hash.write_u8(4); +- lit.hash(hash); +- } +- Lit::Float(lit) => { +- hash.write_u8(5); +- lit.hash(hash); +- } +- Lit::Bool(lit) => { +- hash.write_u8(6); +- lit.hash(hash); +- } +- Lit::Verbatim(lit) => { +- hash.write_u8(7); +- lit.to_string().hash(hash); +- } +- } +- } +-} +- + impl LitStr { + pub fn new(value: &str, span: Span) -> Self { +- let mut lit = Literal::string(value); +- lit.set_span(span); ++ let mut token = Literal::string(value); ++ token.set_span(span); + LitStr { +- repr: Box::new(LitStrRepr { +- token: lit, ++ repr: Box::new(LitRepr { ++ token, + suffix: Box::<str>::default(), + }), + } + } + + pub fn value(&self) -> String { +- let (value, _) = value::parse_lit_str(&self.repr.token.to_string()); ++ let repr = self.repr.token.to_string(); ++ let (value, _suffix) = value::parse_lit_str(&repr); + String::from(value) + } + +@@ -311,7 +222,7 @@ impl LitStr { + fn respan_token_tree(mut token: TokenTree, span: Span) -> TokenTree { + match &mut token { + TokenTree::Group(g) => { +- let stream = respan_token_stream(g.stream().clone(), span); ++ let stream = respan_token_stream(g.stream(), span); + *g = Group::new(g.delimiter(), stream); + g.set_span(span); + } +@@ -345,19 +256,30 @@ impl LitByteStr { + pub fn new(value: &[u8], span: Span) -> Self { + let mut token = Literal::byte_string(value); + token.set_span(span); +- LitByteStr { token } ++ LitByteStr { ++ repr: Box::new(LitRepr { ++ token, ++ suffix: Box::<str>::default(), ++ }), ++ } + } + + pub fn value(&self) -> Vec<u8> { +- value::parse_lit_byte_str(&self.token.to_string()) ++ let repr = self.repr.token.to_string(); ++ let (value, _suffix) = value::parse_lit_byte_str(&repr); ++ value + } + + pub fn span(&self) -> Span { +- self.token.span() ++ self.repr.token.span() + } + + pub fn set_span(&mut self, span: Span) { +- self.token.set_span(span) ++ self.repr.token.set_span(span) ++ } ++ ++ pub fn suffix(&self) -> &str { ++ &self.repr.suffix + } + } + +@@ -365,19 +287,30 @@ impl LitByte { + pub fn new(value: u8, span: Span) -> Self { + let mut token = Literal::u8_suffixed(value); + token.set_span(span); +- LitByte { token } ++ LitByte { ++ repr: Box::new(LitRepr { ++ token, ++ suffix: Box::<str>::default(), ++ }), ++ } + } + + pub fn value(&self) -> u8 { +- value::parse_lit_byte(&self.token.to_string()) ++ let repr = self.repr.token.to_string(); ++ let (value, _suffix) = value::parse_lit_byte(&repr); ++ value + } + + pub fn span(&self) -> Span { +- self.token.span() ++ self.repr.token.span() + } + + pub fn set_span(&mut self, span: Span) { +- self.token.set_span(span) ++ self.repr.token.set_span(span) ++ } ++ ++ pub fn suffix(&self) -> &str { ++ &self.repr.suffix + } + } + +@@ -385,36 +318,52 @@ impl LitChar { + pub fn new(value: char, span: Span) -> Self { + let mut token = Literal::character(value); + token.set_span(span); +- LitChar { token } ++ LitChar { ++ repr: Box::new(LitRepr { ++ token, ++ suffix: Box::<str>::default(), ++ }), ++ } + } + + pub fn value(&self) -> char { +- value::parse_lit_char(&self.token.to_string()) ++ let repr = self.repr.token.to_string(); ++ let (value, _suffix) = value::parse_lit_char(&repr); ++ value + } + + pub fn span(&self) -> Span { +- self.token.span() ++ self.repr.token.span() + } + + pub fn set_span(&mut self, span: Span) { +- self.token.set_span(span) ++ self.repr.token.set_span(span) ++ } ++ ++ pub fn suffix(&self) -> &str { ++ &self.repr.suffix + } + } + + impl LitInt { + pub fn new(repr: &str, span: Span) -> Self { +- if let Some((digits, suffix)) = value::parse_lit_int(repr) { +- let mut token = value::to_literal(repr); +- token.set_span(span); +- LitInt { +- repr: Box::new(LitIntRepr { +- token, +- digits, +- suffix, +- }), +- } +- } else { +- panic!("Not an integer literal: `{}`", repr); ++ let (digits, suffix) = match value::parse_lit_int(repr) { ++ Some(parse) => parse, ++ None => panic!("Not an integer literal: `{}`", repr), ++ }; ++ ++ let mut token = match value::to_literal(repr, &digits, &suffix) { ++ Some(token) => token, ++ None => panic!("Unsupported integer literal: `{}`", repr), ++ }; ++ ++ token.set_span(span); ++ LitInt { ++ repr: Box::new(LitIntRepr { ++ token, ++ digits, ++ suffix, ++ }), + } + } + +@@ -492,18 +441,23 @@ impl Display for LitInt { + + impl LitFloat { + pub fn new(repr: &str, span: Span) -> Self { +- if let Some((digits, suffix)) = value::parse_lit_float(repr) { +- let mut token = value::to_literal(repr); +- token.set_span(span); +- LitFloat { +- repr: Box::new(LitFloatRepr { +- token, +- digits, +- suffix, +- }), +- } +- } else { +- panic!("Not a float literal: `{}`", repr); ++ let (digits, suffix) = match value::parse_lit_float(repr) { ++ Some(parse) => parse, ++ None => panic!("Not a float literal: `{}`", repr), ++ }; ++ ++ let mut token = match value::to_literal(repr, &digits, &suffix) { ++ Some(token) => token, ++ None => panic!("Unsupported float literal: `{}`", repr), ++ }; ++ ++ token.set_span(span); ++ LitFloat { ++ repr: Box::new(LitFloatRepr { ++ token, ++ digits, ++ suffix, ++ }), + } + } + +@@ -575,7 +529,7 @@ mod debug_impls { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + formatter + .debug_struct("LitByteStr") +- .field("token", &format_args!("{}", self.token)) ++ .field("token", &format_args!("{}", self.repr.token)) + .finish() + } + } +@@ -584,7 +538,7 @@ mod debug_impls { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + formatter + .debug_struct("LitByte") +- .field("token", &format_args!("{}", self.token)) ++ .field("token", &format_args!("{}", self.repr.token)) + .finish() + } + } +@@ -593,7 +547,7 @@ mod debug_impls { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + formatter + .debug_struct("LitChar") +- .field("token", &format_args!("{}", self.token)) ++ .field("token", &format_args!("{}", self.repr.token)) + .finish() + } + } +@@ -626,15 +580,53 @@ mod debug_impls { + } + } + ++#[cfg(feature = "clone-impls")] ++impl Clone for LitRepr { ++ fn clone(&self) -> Self { ++ LitRepr { ++ token: self.token.clone(), ++ suffix: self.suffix.clone(), ++ } ++ } ++} ++ ++#[cfg(feature = "clone-impls")] ++impl Clone for LitIntRepr { ++ fn clone(&self) -> Self { ++ LitIntRepr { ++ token: self.token.clone(), ++ digits: self.digits.clone(), ++ suffix: self.suffix.clone(), ++ } ++ } ++} ++ ++#[cfg(feature = "clone-impls")] ++impl Clone for LitFloatRepr { ++ fn clone(&self) -> Self { ++ LitFloatRepr { ++ token: self.token.clone(), ++ digits: self.digits.clone(), ++ suffix: self.suffix.clone(), ++ } ++ } ++} ++ + macro_rules! lit_extra_traits { +- ($ty:ident, $($field:ident).+) => { +- #[cfg(feature = "extra-traits")] +- impl Eq for $ty {} ++ ($ty:ident) => { ++ #[cfg(feature = "clone-impls")] ++ impl Clone for $ty { ++ fn clone(&self) -> Self { ++ $ty { ++ repr: self.repr.clone(), ++ } ++ } ++ } + + #[cfg(feature = "extra-traits")] + impl PartialEq for $ty { + fn eq(&self, other: &Self) -> bool { +- self.$($field).+.to_string() == other.$($field).+.to_string() ++ self.repr.token.to_string() == other.repr.token.to_string() + } + } + +@@ -644,7 +636,7 @@ macro_rules! lit_extra_traits { + where + H: Hasher, + { +- self.$($field).+.to_string().hash(state); ++ self.repr.token.to_string().hash(state); + } + } + +@@ -657,20 +649,23 @@ macro_rules! lit_extra_traits { + }; + } + +-lit_extra_traits!(LitStr, repr.token); +-lit_extra_traits!(LitByteStr, token); +-lit_extra_traits!(LitByte, token); +-lit_extra_traits!(LitChar, token); +-lit_extra_traits!(LitInt, repr.token); +-lit_extra_traits!(LitFloat, repr.token); +-lit_extra_traits!(LitBool, value); ++lit_extra_traits!(LitStr); ++lit_extra_traits!(LitByteStr); ++lit_extra_traits!(LitByte); ++lit_extra_traits!(LitChar); ++lit_extra_traits!(LitInt); ++lit_extra_traits!(LitFloat); ++ ++#[cfg(feature = "parsing")] ++#[doc(hidden)] ++#[allow(non_snake_case)] ++pub fn LitBool(marker: lookahead::TokenMarker) -> LitBool { ++ match marker {} ++} + + ast_enum! { + /// The style of a string literal, either plain quoted or a raw string like + /// `r##"data"##`. +- /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` +- /// feature.* + pub enum StrStyle #no_visit { + /// An ordinary string like `"data"`. + Cooked, +@@ -691,7 +686,9 @@ pub fn Lit(marker: lookahead::TokenMarker) -> Lit { + #[cfg(feature = "parsing")] + pub mod parsing { + use super::*; ++ use crate::buffer::Cursor; + use crate::parse::{Parse, ParseStream, Result}; ++ use proc_macro2::Punct; + + impl Parse for Lit { + fn parse(input: ParseStream) -> Result<Self> { +@@ -699,25 +696,73 @@ pub mod parsing { + if let Some((lit, rest)) = cursor.literal() { + return Ok((Lit::new(lit), rest)); + } +- while let Some((ident, rest)) = cursor.ident() { +- let value = if ident == "true" { +- true +- } else if ident == "false" { +- false +- } else { +- break; +- }; +- let lit_bool = LitBool { +- value, +- span: ident.span(), +- }; +- return Ok((Lit::Bool(lit_bool), rest)); ++ ++ if let Some((ident, rest)) = cursor.ident() { ++ let value = ident == "true"; ++ if value || ident == "false" { ++ let lit_bool = LitBool { ++ value, ++ span: ident.span(), ++ }; ++ return Ok((Lit::Bool(lit_bool), rest)); ++ } + } ++ ++ if let Some((punct, rest)) = cursor.punct() { ++ if punct.as_char() == '-' { ++ if let Some((lit, rest)) = parse_negative_lit(punct, rest) { ++ return Ok((lit, rest)); ++ } ++ } ++ } ++ + Err(cursor.error("expected literal")) + }) + } + } + ++ fn parse_negative_lit(neg: Punct, cursor: Cursor) -> Option<(Lit, Cursor)> { ++ let (lit, rest) = cursor.literal()?; ++ ++ let mut span = neg.span(); ++ span = span.join(lit.span()).unwrap_or(span); ++ ++ let mut repr = lit.to_string(); ++ repr.insert(0, '-'); ++ ++ if !(repr.ends_with("f32") || repr.ends_with("f64")) { ++ if let Some((digits, suffix)) = value::parse_lit_int(&repr) { ++ if let Some(mut token) = value::to_literal(&repr, &digits, &suffix) { ++ token.set_span(span); ++ return Some(( ++ Lit::Int(LitInt { ++ repr: Box::new(LitIntRepr { ++ token, ++ digits, ++ suffix, ++ }), ++ }), ++ rest, ++ )); ++ } ++ } ++ } ++ ++ let (digits, suffix) = value::parse_lit_float(&repr)?; ++ let mut token = value::to_literal(&repr, &digits, &suffix)?; ++ token.set_span(span); ++ Some(( ++ Lit::Float(LitFloat { ++ repr: Box::new(LitFloatRepr { ++ token, ++ digits, ++ suffix, ++ }), ++ }), ++ rest, ++ )) ++ } ++ + impl Parse for LitStr { + fn parse(input: ParseStream) -> Result<Self> { + let head = input.fork(); +@@ -803,19 +848,19 @@ mod printing { + + impl ToTokens for LitByteStr { + fn to_tokens(&self, tokens: &mut TokenStream) { +- self.token.to_tokens(tokens); ++ self.repr.token.to_tokens(tokens); + } + } + + impl ToTokens for LitByte { + fn to_tokens(&self, tokens: &mut TokenStream) { +- self.token.to_tokens(tokens); ++ self.repr.token.to_tokens(tokens); + } + } + + impl ToTokens for LitChar { + fn to_tokens(&self, tokens: &mut TokenStream) { +- self.token.to_tokens(tokens); ++ self.repr.token.to_tokens(tokens); + } + } + +@@ -855,20 +900,29 @@ mod value { + b'"' | b'r' => { + let (_, suffix) = parse_lit_str(&repr); + return Lit::Str(LitStr { +- repr: Box::new(LitStrRepr { token, suffix }), ++ repr: Box::new(LitRepr { token, suffix }), + }); + } + b'b' => match byte(&repr, 1) { + b'"' | b'r' => { +- return Lit::ByteStr(LitByteStr { token }); ++ let (_, suffix) = parse_lit_byte_str(&repr); ++ return Lit::ByteStr(LitByteStr { ++ repr: Box::new(LitRepr { token, suffix }), ++ }); + } + b'\'' => { +- return Lit::Byte(LitByte { token }); ++ let (_, suffix) = parse_lit_byte(&repr); ++ return Lit::Byte(LitByte { ++ repr: Box::new(LitRepr { token, suffix }), ++ }); + } + _ => {} + }, + b'\'' => { +- return Lit::Char(LitChar { token }); ++ let (_, suffix) = parse_lit_char(&repr); ++ return Lit::Char(LitChar { ++ repr: Box::new(LitRepr { token, suffix }), ++ }); + } + b'0'..=b'9' | b'-' => { + if !(repr.ends_with("f32") || repr.ends_with("f64")) { +@@ -905,6 +959,44 @@ mod value { + + panic!("Unrecognized literal: `{}`", repr); + } ++ ++ pub fn suffix(&self) -> &str { ++ match self { ++ Lit::Str(lit) => lit.suffix(), ++ Lit::ByteStr(lit) => lit.suffix(), ++ Lit::Byte(lit) => lit.suffix(), ++ Lit::Char(lit) => lit.suffix(), ++ Lit::Int(lit) => lit.suffix(), ++ Lit::Float(lit) => lit.suffix(), ++ Lit::Bool(_) | Lit::Verbatim(_) => "", ++ } ++ } ++ ++ pub fn span(&self) -> Span { ++ match self { ++ Lit::Str(lit) => lit.span(), ++ Lit::ByteStr(lit) => lit.span(), ++ Lit::Byte(lit) => lit.span(), ++ Lit::Char(lit) => lit.span(), ++ Lit::Int(lit) => lit.span(), ++ Lit::Float(lit) => lit.span(), ++ Lit::Bool(lit) => lit.span, ++ Lit::Verbatim(lit) => lit.span(), ++ } ++ } ++ ++ pub fn set_span(&mut self, span: Span) { ++ match self { ++ Lit::Str(lit) => lit.set_span(span), ++ Lit::ByteStr(lit) => lit.set_span(span), ++ Lit::Byte(lit) => lit.set_span(span), ++ Lit::Char(lit) => lit.set_span(span), ++ Lit::Int(lit) => lit.set_span(span), ++ Lit::Float(lit) => lit.set_span(span), ++ Lit::Bool(lit) => lit.span = span, ++ Lit::Verbatim(lit) => lit.set_span(span), ++ } ++ } + } + + /// Get the byte at offset idx, or a default of `b'\0'` if we're looking +@@ -1004,19 +1096,18 @@ mod value { + pounds += 1; + } + assert_eq!(byte(s, pounds), b'"'); +- assert_eq!(byte(s, s.len() - pounds - 1), b'"'); +- for end in s[s.len() - pounds..].bytes() { ++ let close = s.rfind('"').unwrap(); ++ for end in s[close + 1..close + 1 + pounds].bytes() { + assert_eq!(end, b'#'); + } + +- let content = s[pounds + 1..s.len() - pounds - 1] +- .to_owned() +- .into_boxed_str(); +- let suffix = Box::<str>::default(); // todo ++ let content = s[pounds + 1..close].to_owned().into_boxed_str(); ++ let suffix = s[close + 1 + pounds..].to_owned().into_boxed_str(); + (content, suffix) + } + +- pub fn parse_lit_byte_str(s: &str) -> Vec<u8> { ++ // Returns (content, suffix). ++ pub fn parse_lit_byte_str(s: &str) -> (Vec<u8>, Box<str>) { + assert_eq!(byte(s, 0), b'b'); + match byte(s, 1) { + b'"' => parse_lit_byte_str_cooked(s), +@@ -1028,25 +1119,25 @@ mod value { + // Clippy false positive + // https://github.com/rust-lang-nursery/rust-clippy/issues/2329 + #[allow(clippy::needless_continue)] +- fn parse_lit_byte_str_cooked(mut s: &str) -> Vec<u8> { ++ fn parse_lit_byte_str_cooked(mut s: &str) -> (Vec<u8>, Box<str>) { + assert_eq!(byte(s, 0), b'b'); + assert_eq!(byte(s, 1), b'"'); + s = &s[2..]; + + // We're going to want to have slices which don't respect codepoint boundaries. +- let mut s = s.as_bytes(); ++ let mut v = s.as_bytes(); + + let mut out = Vec::new(); + 'outer: loop { +- let byte = match byte(s, 0) { ++ let byte = match byte(v, 0) { + b'"' => break, + b'\\' => { +- let b = byte(s, 1); +- s = &s[2..]; ++ let b = byte(v, 1); ++ v = &v[2..]; + match b { + b'x' => { +- let (b, rest) = backslash_x(s); +- s = rest; ++ let (b, rest) = backslash_x(v); ++ v = rest; + b + } + b'n' => b'\n', +@@ -1057,10 +1148,10 @@ mod value { + b'\'' => b'\'', + b'"' => b'"', + b'\r' | b'\n' => loop { +- let byte = byte(s, 0); ++ let byte = byte(v, 0); + let ch = char::from_u32(u32::from(byte)).unwrap(); + if ch.is_whitespace() { +- s = &s[1..]; ++ v = &v[1..]; + } else { + continue 'outer; + } +@@ -1069,42 +1160,45 @@ mod value { + } + } + b'\r' => { +- assert_eq!(byte(s, 1), b'\n', "Bare CR not allowed in string"); +- s = &s[2..]; ++ assert_eq!(byte(v, 1), b'\n', "Bare CR not allowed in string"); ++ v = &v[2..]; + b'\n' + } + b => { +- s = &s[1..]; ++ v = &v[1..]; + b + } + }; + out.push(byte); + } + +- assert_eq!(s, b"\""); +- out ++ assert_eq!(byte(v, 0), b'"'); ++ let suffix = s[s.len() - v.len() + 1..].to_owned().into_boxed_str(); ++ (out, suffix) + } + +- fn parse_lit_byte_str_raw(s: &str) -> Vec<u8> { ++ fn parse_lit_byte_str_raw(s: &str) -> (Vec<u8>, Box<str>) { + assert_eq!(byte(s, 0), b'b'); +- String::from(parse_lit_str_raw(&s[1..]).0).into_bytes() ++ let (value, suffix) = parse_lit_str_raw(&s[1..]); ++ (String::from(value).into_bytes(), suffix) + } + +- pub fn parse_lit_byte(s: &str) -> u8 { ++ // Returns (value, suffix). ++ pub fn parse_lit_byte(s: &str) -> (u8, Box<str>) { + assert_eq!(byte(s, 0), b'b'); + assert_eq!(byte(s, 1), b'\''); + + // We're going to want to have slices which don't respect codepoint boundaries. +- let mut s = s[2..].as_bytes(); ++ let mut v = s[2..].as_bytes(); + +- let b = match byte(s, 0) { ++ let b = match byte(v, 0) { + b'\\' => { +- let b = byte(s, 1); +- s = &s[2..]; ++ let b = byte(v, 1); ++ v = &v[2..]; + match b { + b'x' => { +- let (b, rest) = backslash_x(s); +- s = rest; ++ let (b, rest) = backslash_x(v); ++ v = rest; + b + } + b'n' => b'\n', +@@ -1118,16 +1212,18 @@ mod value { + } + } + b => { +- s = &s[1..]; ++ v = &v[1..]; + b + } + }; + +- assert_eq!(byte(s, 0), b'\''); +- b ++ assert_eq!(byte(v, 0), b'\''); ++ let suffix = s[s.len() - v.len() + 1..].to_owned().into_boxed_str(); ++ (b, suffix) + } + +- pub fn parse_lit_char(mut s: &str) -> char { ++ // Returns (value, suffix). ++ pub fn parse_lit_char(mut s: &str) -> (char, Box<str>) { + assert_eq!(byte(s, 0), b'\''); + s = &s[1..]; + +@@ -1163,8 +1259,9 @@ mod value { + ch + } + }; +- assert_eq!(s, "\'", "Expected end of char literal"); +- ch ++ assert_eq!(byte(s, 0), b'\''); ++ let suffix = s[1..].to_owned().into_boxed_str(); ++ (ch, suffix) + } + + fn backslash_x<S>(s: &S) -> (u8, &S) +@@ -1334,7 +1431,11 @@ mod value { + } + b'e' | b'E' => { + if has_e { +- return None; ++ if has_exponent { ++ break; ++ } else { ++ return None; ++ } + } + has_e = true; + bytes[write] = b'e'; +@@ -1372,11 +1473,33 @@ mod value { + } + } + +- pub fn to_literal(s: &str) -> Literal { +- let stream = s.parse::<TokenStream>().unwrap(); +- match stream.into_iter().next().unwrap() { +- TokenTree::Literal(l) => l, +- _ => unreachable!(), ++ pub fn to_literal(repr: &str, digits: &str, suffix: &str) -> Option<Literal> { ++ if repr.starts_with('-') { ++ if suffix == "f64" { ++ digits.parse().ok().map(Literal::f64_suffixed) ++ } else if suffix == "f32" { ++ digits.parse().ok().map(Literal::f32_suffixed) ++ } else if suffix == "i64" { ++ digits.parse().ok().map(Literal::i64_suffixed) ++ } else if suffix == "i32" { ++ digits.parse().ok().map(Literal::i32_suffixed) ++ } else if suffix == "i16" { ++ digits.parse().ok().map(Literal::i16_suffixed) ++ } else if suffix == "i8" { ++ digits.parse().ok().map(Literal::i8_suffixed) ++ } else if !suffix.is_empty() { ++ None ++ } else if digits.contains('.') { ++ digits.parse().ok().map(Literal::f64_unsuffixed) ++ } else { ++ digits.parse().ok().map(Literal::i64_unsuffixed) ++ } ++ } else { ++ let stream = repr.parse::<TokenStream>().unwrap(); ++ match stream.into_iter().next().unwrap() { ++ TokenTree::Literal(l) => Some(l), ++ _ => unreachable!(), ++ } + } + } + } +diff --git a/third_party/rust/syn/src/mac.rs b/third_party/rust/syn/src/mac.rs +index 6c3dcae92a..de288a34e1 100644 +--- a/third_party/rust/syn/src/mac.rs ++++ b/third_party/rust/syn/src/mac.rs +@@ -2,21 +2,17 @@ use super::*; + use crate::token::{Brace, Bracket, Paren}; + use proc_macro2::TokenStream; + #[cfg(feature = "parsing")] +-use proc_macro2::{Delimiter, Span, TokenTree}; ++use proc_macro2::{Delimiter, Group, Span, TokenTree}; + + #[cfg(feature = "parsing")] + use crate::parse::{Parse, ParseStream, Parser, Result}; +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; +-#[cfg(feature = "extra-traits")] +-use std::hash::{Hash, Hasher}; + + ast_struct! { + /// A macro invocation: `println!("{}", mac)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- pub struct Macro #manual_extra_traits { ++ pub struct Macro { + pub path: Path, + pub bang_token: Token![!], + pub delimiter: MacroDelimiter, +@@ -27,7 +23,7 @@ ast_struct! { + ast_enum! { + /// A grouping token that surrounds a macro body: `m!(...)` or `m!{...}` or `m![...]`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum MacroDelimiter { + Paren(Paren), +@@ -36,39 +32,20 @@ ast_enum! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Macro {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Macro { +- fn eq(&self, other: &Self) -> bool { +- self.path == other.path +- && self.bang_token == other.bang_token +- && self.delimiter == other.delimiter +- && TokenStreamHelper(&self.tokens) == TokenStreamHelper(&other.tokens) +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Macro { +- fn hash<H>(&self, state: &mut H) +- where +- H: Hasher, +- { +- self.path.hash(state); +- self.bang_token.hash(state); +- self.delimiter.hash(state); +- TokenStreamHelper(&self.tokens).hash(state); +- } +-} +- + #[cfg(feature = "parsing")] +-fn delimiter_span(delimiter: &MacroDelimiter) -> Span { +- match delimiter { ++fn delimiter_span_close(macro_delimiter: &MacroDelimiter) -> Span { ++ let delimiter = match macro_delimiter { ++ MacroDelimiter::Paren(_) => Delimiter::Parenthesis, ++ MacroDelimiter::Brace(_) => Delimiter::Brace, ++ MacroDelimiter::Bracket(_) => Delimiter::Bracket, ++ }; ++ let mut group = Group::new(delimiter, TokenStream::new()); ++ group.set_span(match macro_delimiter { + MacroDelimiter::Paren(token) => token.span, + MacroDelimiter::Brace(token) => token.span, + MacroDelimiter::Bracket(token) => token.span, +- } ++ }); ++ group.span_close() + } + + impl Macro { +@@ -163,9 +140,7 @@ impl Macro { + /// given parser. + #[cfg(feature = "parsing")] + pub fn parse_body_with<F: Parser>(&self, parser: F) -> Result<F::Output> { +- // TODO: see if we can get a group.span_close() span in here as the +- // scope, rather than the span of the whole group. +- let scope = delimiter_span(&self.delimiter); ++ let scope = delimiter_span_close(&self.delimiter); + crate::parse::parse_scoped(parser, scope, self.tokens.clone()) + } + } +diff --git a/third_party/rust/syn/src/macros.rs b/third_party/rust/syn/src/macros.rs +index 9cac5c15df..8060224381 100644 +--- a/third_party/rust/syn/src/macros.rs ++++ b/third_party/rust/syn/src/macros.rs +@@ -4,15 +4,11 @@ macro_rules! ast_struct { + struct $name:ident #full $($rest:tt)* + ) => { + #[cfg(feature = "full")] +- #[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] + $($attrs_pub)* struct $name $($rest)* + + #[cfg(not(feature = "full"))] +- #[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] + $($attrs_pub)* struct $name { +- _noconstruct: (), ++ _noconstruct: ::std::marker::PhantomData<::proc_macro2::Span>, + } + + #[cfg(all(not(feature = "full"), feature = "printing"))] +@@ -23,29 +19,10 @@ macro_rules! ast_struct { + } + }; + +- ( +- [$($attrs_pub:tt)*] +- struct $name:ident #manual_extra_traits $($rest:tt)* +- ) => { +- #[cfg_attr(feature = "extra-traits", derive(Debug))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] +- $($attrs_pub)* struct $name $($rest)* +- }; +- +- ( +- [$($attrs_pub:tt)*] +- struct $name:ident #manual_extra_traits_debug $($rest:tt)* +- ) => { +- #[cfg_attr(feature = "clone-impls", derive(Clone))] +- $($attrs_pub)* struct $name $($rest)* +- }; +- + ( + [$($attrs_pub:tt)*] + struct $name:ident $($rest:tt)* + ) => { +- #[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] + $($attrs_pub)* struct $name $($rest)* + }; + +@@ -63,21 +40,10 @@ macro_rules! ast_enum { + ast_enum!([$($attrs_pub)*] enum $name $($rest)*); + ); + +- ( +- [$($attrs_pub:tt)*] +- enum $name:ident #manual_extra_traits $($rest:tt)* +- ) => ( +- #[cfg_attr(feature = "extra-traits", derive(Debug))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] +- $($attrs_pub)* enum $name $($rest)* +- ); +- + ( + [$($attrs_pub:tt)*] + enum $name:ident $($rest:tt)* + ) => ( +- #[cfg_attr(feature = "extra-traits", derive(Debug, Eq, PartialEq, Hash))] +- #[cfg_attr(feature = "clone-impls", derive(Clone))] + $($attrs_pub)* enum $name $($rest)* + ); + +@@ -120,15 +86,9 @@ macro_rules! ast_enum_of_structs_impl { + check_keyword_matches!(pub $pub); + check_keyword_matches!(enum $enum); + +- $( +- $( +- impl From<$member> for $name { +- fn from(e: $member) -> $name { +- $name::$variant(e) +- } +- } +- )* +- )* ++ $($( ++ ast_enum_from_struct!($name::$variant, $member); ++ )*)* + + #[cfg(feature = "printing")] + generate_to_tokens! { +@@ -140,6 +100,19 @@ macro_rules! ast_enum_of_structs_impl { + }; + } + ++macro_rules! ast_enum_from_struct { ++ // No From<TokenStream> for verbatim variants. ++ ($name:ident::Verbatim, $member:ident) => {}; ++ ++ ($name:ident::$variant:ident, $member:ident) => { ++ impl From<$member> for $name { ++ fn from(e: $member) -> $name { ++ $name::$variant(e) ++ } ++ } ++ }; ++} ++ + #[cfg(feature = "printing")] + macro_rules! generate_to_tokens { + (do_not_generate_to_tokens $($foo:tt)*) => (); +diff --git a/third_party/rust/syn/src/op.rs b/third_party/rust/syn/src/op.rs +index 49fb853c79..d254673b40 100644 +--- a/third_party/rust/syn/src/op.rs ++++ b/third_party/rust/syn/src/op.rs +@@ -1,9 +1,8 @@ + ast_enum! { + /// A binary operator: `+`, `+=`, `&`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- #[cfg_attr(feature = "clone-impls", derive(Copy))] + pub enum BinOp { + /// The `+` operator (addition) + Add(Token![+]), +@@ -67,9 +66,8 @@ ast_enum! { + ast_enum! { + /// A unary operator: `*`, `!`, `-`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* +- #[cfg_attr(feature = "clone-impls", derive(Copy))] + pub enum UnOp { + /// The `*` operator for dereferencing + Deref(Token![*]), +diff --git a/third_party/rust/syn/src/parse.rs b/third_party/rust/syn/src/parse.rs +index 7c7b194308..abb4c4c14f 100644 +--- a/third_party/rust/syn/src/parse.rs ++++ b/third_party/rust/syn/src/parse.rs +@@ -26,8 +26,8 @@ + //! [`parse_macro_input!`]: ../macro.parse_macro_input.html + //! + //! ``` +-//! extern crate proc_macro; +-//! ++//! # extern crate proc_macro; ++//! # + //! use proc_macro::TokenStream; + //! use syn::{braced, parse_macro_input, token, Field, Ident, Result, Token}; + //! use syn::parse::{Parse, ParseStream}; +@@ -109,9 +109,7 @@ + //! # Ok(()) + //! # } + //! # +-//! # fn main() { +-//! # run_parser().unwrap(); +-//! # } ++//! # run_parser().unwrap(); + //! ``` + //! + //! The [`parse_quote!`] macro also uses this approach. +@@ -155,8 +153,8 @@ + //! [`Parser`]: trait.Parser.html + //! + //! ``` +-//! extern crate proc_macro; +-//! ++//! # extern crate proc_macro; ++//! # + //! use proc_macro::TokenStream; + //! use syn::parse::Parser; + //! use syn::punctuated::Punctuated; +@@ -186,7 +184,7 @@ + //! + //! --- + //! +-//! *This module is available if Syn is built with the `"parsing"` feature.* ++//! *This module is available only if Syn is built with the `"parsing"` feature.* + + #[path = "discouraged.rs"] + pub mod discouraged; +@@ -217,6 +215,11 @@ pub use crate::lookahead::{Lookahead1, Peek}; + + /// Parsing interface implemented by all types that can be parsed in a default + /// way from a token stream. ++/// ++/// Refer to the [module documentation] for details about implementing and using ++/// the `Parse` trait. ++/// ++/// [module documentation]: self + pub trait Parse: Sized { + fn parse(input: ParseStream) -> Result<Self>; + } +@@ -263,13 +266,16 @@ pub struct ParseBuffer<'a> { + // the cell. + cell: Cell<Cursor<'static>>, + marker: PhantomData<Cursor<'a>>, +- unexpected: Rc<Cell<Option<Span>>>, ++ unexpected: Cell<Option<Rc<Cell<Unexpected>>>>, + } + + impl<'a> Drop for ParseBuffer<'a> { + fn drop(&mut self) { +- if !self.is_empty() && self.unexpected.get().is_none() { +- self.unexpected.set(Some(self.cursor().span())); ++ if let Some(unexpected_span) = span_of_unexpected_ignoring_nones(self.cursor()) { ++ let (inner, old_span) = inner_unexpected(self); ++ if old_span.is_none() { ++ inner.set(Unexpected::Some(unexpected_span)); ++ } + } + } + } +@@ -324,15 +330,12 @@ impl<'a> Debug for ParseBuffer<'a> { + /// # input.parse() + /// # } + /// # +-/// # fn main() { +-/// # use syn::parse::Parser; +-/// # let remainder = remainder_after_skipping_past_next_at +-/// # .parse_str("a @ b c") +-/// # .unwrap(); +-/// # assert_eq!(remainder.to_string(), "b c"); +-/// # } ++/// # use syn::parse::Parser; ++/// # let remainder = remainder_after_skipping_past_next_at ++/// # .parse_str("a @ b c") ++/// # .unwrap(); ++/// # assert_eq!(remainder.to_string(), "b c"); + /// ``` +-#[derive(Copy, Clone)] + pub struct StepCursor<'c, 'a> { + scope: Span, + // This field is covariant in 'c. +@@ -356,6 +359,14 @@ impl<'c, 'a> Deref for StepCursor<'c, 'a> { + } + } + ++impl<'c, 'a> Copy for StepCursor<'c, 'a> {} ++ ++impl<'c, 'a> Clone for StepCursor<'c, 'a> { ++ fn clone(&self) -> Self { ++ *self ++ } ++} ++ + impl<'c, 'a> StepCursor<'c, 'a> { + /// Triggers an error at the current position of the parse stream. + /// +@@ -375,36 +386,81 @@ pub(crate) fn advance_step_cursor<'c, 'a>(proof: StepCursor<'c, 'a>, to: Cursor< + unsafe { mem::transmute::<Cursor<'c>, Cursor<'a>>(to) } + } + +-fn skip(input: ParseStream) -> bool { +- input +- .step(|cursor| { +- if let Some((_lifetime, rest)) = cursor.lifetime() { +- Ok((true, rest)) +- } else if let Some((_token, rest)) = cursor.token_tree() { +- Ok((true, rest)) +- } else { +- Ok((false, *cursor)) +- } +- }) +- .unwrap() +-} +- + pub(crate) fn new_parse_buffer( + scope: Span, + cursor: Cursor, +- unexpected: Rc<Cell<Option<Span>>>, ++ unexpected: Rc<Cell<Unexpected>>, + ) -> ParseBuffer { + ParseBuffer { + scope, + // See comment on `cell` in the struct definition. + cell: Cell::new(unsafe { mem::transmute::<Cursor, Cursor<'static>>(cursor) }), + marker: PhantomData, +- unexpected, ++ unexpected: Cell::new(Some(unexpected)), ++ } ++} ++ ++pub(crate) enum Unexpected { ++ None, ++ Some(Span), ++ Chain(Rc<Cell<Unexpected>>), ++} ++ ++impl Default for Unexpected { ++ fn default() -> Self { ++ Unexpected::None + } + } + +-pub(crate) fn get_unexpected(buffer: &ParseBuffer) -> Rc<Cell<Option<Span>>> { +- buffer.unexpected.clone() ++impl Clone for Unexpected { ++ fn clone(&self) -> Self { ++ match self { ++ Unexpected::None => Unexpected::None, ++ Unexpected::Some(span) => Unexpected::Some(*span), ++ Unexpected::Chain(next) => Unexpected::Chain(next.clone()), ++ } ++ } ++} ++ ++// We call this on Cell<Unexpected> and Cell<Option<T>> where temporarily ++// swapping in a None is cheap. ++fn cell_clone<T: Default + Clone>(cell: &Cell<T>) -> T { ++ let prev = cell.take(); ++ let ret = prev.clone(); ++ cell.set(prev); ++ ret ++} ++ ++fn inner_unexpected(buffer: &ParseBuffer) -> (Rc<Cell<Unexpected>>, Option<Span>) { ++ let mut unexpected = get_unexpected(buffer); ++ loop { ++ match cell_clone(&unexpected) { ++ Unexpected::None => return (unexpected, None), ++ Unexpected::Some(span) => return (unexpected, Some(span)), ++ Unexpected::Chain(next) => unexpected = next, ++ } ++ } ++} ++ ++pub(crate) fn get_unexpected(buffer: &ParseBuffer) -> Rc<Cell<Unexpected>> { ++ cell_clone(&buffer.unexpected).unwrap() ++} ++ ++fn span_of_unexpected_ignoring_nones(mut cursor: Cursor) -> Option<Span> { ++ if cursor.eof() { ++ return None; ++ } ++ while let Some((inner, _span, rest)) = cursor.group(Delimiter::None) { ++ if let Some(unexpected) = span_of_unexpected_ignoring_nones(inner) { ++ return Some(unexpected); ++ } ++ cursor = rest; ++ } ++ if cursor.eof() { ++ None ++ } else { ++ Some(cursor.span()) ++ } + } + + impl<'a> ParseBuffer<'a> { +@@ -566,14 +622,17 @@ impl<'a> ParseBuffer<'a> { + /// } + /// ``` + pub fn peek2<T: Peek>(&self, token: T) -> bool { +- let ahead = self.fork(); +- skip(&ahead) && ahead.peek(token) ++ let _ = token; ++ self.cursor().skip().map_or(false, T::Token::peek) + } + + /// Looks at the third-next token in the parse stream. + pub fn peek3<T: Peek>(&self, token: T) -> bool { +- let ahead = self.fork(); +- skip(&ahead) && skip(&ahead) && ahead.peek(token) ++ let _ = token; ++ self.cursor() ++ .skip() ++ .and_then(Cursor::skip) ++ .map_or(false, T::Token::peek) + } + + /// Parses zero or more occurrences of `T` separated by punctuation of type +@@ -615,12 +674,10 @@ impl<'a> ParseBuffer<'a> { + /// } + /// } + /// # +- /// # fn main() { +- /// # let input = quote! { +- /// # struct S(A, B); +- /// # }; +- /// # syn::parse2::<TupleStruct>(input).unwrap(); +- /// # } ++ /// # let input = quote! { ++ /// # struct S(A, B); ++ /// # }; ++ /// # syn::parse2::<TupleStruct>(input).unwrap(); + /// ``` + pub fn parse_terminated<T, P: Parse>( + &self, +@@ -847,8 +904,8 @@ impl<'a> ParseBuffer<'a> { + cell: self.cell.clone(), + marker: PhantomData, + // Not the parent's unexpected. Nothing cares whether the clone +- // parses all the way. +- unexpected: Rc::new(Cell::new(None)), ++ // parses all the way unless we `advance_to`. ++ unexpected: Cell::new(Some(Rc::new(Cell::new(Unexpected::None)))), + } + } + +@@ -923,13 +980,11 @@ impl<'a> ParseBuffer<'a> { + /// # input.parse() + /// # } + /// # +- /// # fn main() { +- /// # use syn::parse::Parser; +- /// # let remainder = remainder_after_skipping_past_next_at +- /// # .parse_str("a @ b c") +- /// # .unwrap(); +- /// # assert_eq!(remainder.to_string(), "b c"); +- /// # } ++ /// # use syn::parse::Parser; ++ /// # let remainder = remainder_after_skipping_past_next_at ++ /// # .parse_str("a @ b c") ++ /// # .unwrap(); ++ /// # assert_eq!(remainder.to_string(), "b c"); + /// ``` + pub fn step<F, R>(&self, function: F) -> Result<R> + where +@@ -961,6 +1016,18 @@ impl<'a> ParseBuffer<'a> { + Ok(node) + } + ++ /// Returns the `Span` of the next token in the parse stream, or ++ /// `Span::call_site()` if this parse stream has completely exhausted its ++ /// input `TokenStream`. ++ pub fn span(&self) -> Span { ++ let cursor = self.cursor(); ++ if cursor.eof() { ++ self.scope ++ } else { ++ crate::buffer::open_span_of_group(cursor) ++ } ++ } ++ + /// Provides low-level access to the token representation underlying this + /// parse stream. + /// +@@ -971,7 +1038,7 @@ impl<'a> ParseBuffer<'a> { + } + + fn check_unexpected(&self) -> Result<()> { +- match self.unexpected.get() { ++ match inner_unexpected(self).1 { + Some(span) => Err(Error::new(span, "unexpected token")), + None => Ok(()), + } +@@ -1048,7 +1115,7 @@ impl Parse for Literal { + /// + /// [module documentation]: self + /// +-/// *This trait is available if Syn is built with the `"parsing"` feature.* ++/// *This trait is available only if Syn is built with the `"parsing"` feature.* + pub trait Parser: Sized { + type Output; + +@@ -1063,7 +1130,7 @@ pub trait Parser: Sized { + /// This function will check that the input is fully parsed. If there are + /// any unparsed tokens at the end of the stream, an error is returned. + /// +- /// *This method is available if Syn is built with both the `"parsing"` and ++ /// *This method is available only if Syn is built with both the `"parsing"` and + /// `"proc-macro"` features.* + #[cfg(all( + not(all(target_arch = "wasm32", any(target_os = "unknown", target_os = "wasi"))), +@@ -1088,6 +1155,7 @@ pub trait Parser: Sized { + + // Not public API. + #[doc(hidden)] ++ #[cfg(any(feature = "full", feature = "derive"))] + fn __parse_scoped(self, scope: Span, tokens: TokenStream) -> Result<Self::Output> { + let _ = scope; + self.parse2(tokens) +@@ -1095,6 +1163,7 @@ pub trait Parser: Sized { + + // Not public API. + #[doc(hidden)] ++ #[cfg(any(feature = "full", feature = "derive"))] + fn __parse_stream(self, input: ParseStream) -> Result<Self::Output> { + input.parse().and_then(|tokens| self.parse2(tokens)) + } +@@ -1103,7 +1172,7 @@ pub trait Parser: Sized { + fn tokens_to_parse_buffer(tokens: &TokenBuffer) -> ParseBuffer { + let scope = Span::call_site(); + let cursor = tokens.begin(); +- let unexpected = Rc::new(Cell::new(None)); ++ let unexpected = Rc::new(Cell::new(Unexpected::None)); + new_parse_buffer(scope, cursor, unexpected) + } + +@@ -1118,38 +1187,42 @@ where + let state = tokens_to_parse_buffer(&buf); + let node = self(&state)?; + state.check_unexpected()?; +- if state.is_empty() { +- Ok(node) ++ if let Some(unexpected_span) = span_of_unexpected_ignoring_nones(state.cursor()) { ++ Err(Error::new(unexpected_span, "unexpected token")) + } else { +- Err(state.error("unexpected token")) ++ Ok(node) + } + } + + #[doc(hidden)] ++ #[cfg(any(feature = "full", feature = "derive"))] + fn __parse_scoped(self, scope: Span, tokens: TokenStream) -> Result<Self::Output> { + let buf = TokenBuffer::new2(tokens); + let cursor = buf.begin(); +- let unexpected = Rc::new(Cell::new(None)); ++ let unexpected = Rc::new(Cell::new(Unexpected::None)); + let state = new_parse_buffer(scope, cursor, unexpected); + let node = self(&state)?; + state.check_unexpected()?; +- if state.is_empty() { +- Ok(node) ++ if let Some(unexpected_span) = span_of_unexpected_ignoring_nones(state.cursor()) { ++ Err(Error::new(unexpected_span, "unexpected token")) + } else { +- Err(state.error("unexpected token")) ++ Ok(node) + } + } + + #[doc(hidden)] ++ #[cfg(any(feature = "full", feature = "derive"))] + fn __parse_stream(self, input: ParseStream) -> Result<Self::Output> { + self(input) + } + } + ++#[cfg(any(feature = "full", feature = "derive"))] + pub(crate) fn parse_scoped<F: Parser>(f: F, scope: Span, tokens: TokenStream) -> Result<F::Output> { + f.__parse_scoped(scope, tokens) + } + ++#[cfg(any(feature = "full", feature = "derive"))] + pub(crate) fn parse_stream<F: Parser>(f: F, input: ParseStream) -> Result<F::Output> { + f.__parse_stream(input) + } +@@ -1160,8 +1233,8 @@ pub(crate) fn parse_stream<F: Parser>(f: F, input: ParseStream) -> Result<F::Out + /// provided any attribute args. + /// + /// ``` +-/// extern crate proc_macro; +-/// ++/// # extern crate proc_macro; ++/// # + /// use proc_macro::TokenStream; + /// use syn::parse_macro_input; + /// use syn::parse::Nothing; +diff --git a/third_party/rust/syn/src/parse_macro_input.rs b/third_party/rust/syn/src/parse_macro_input.rs +index d6e0725c17..c8fc1cea37 100644 +--- a/third_party/rust/syn/src/parse_macro_input.rs ++++ b/third_party/rust/syn/src/parse_macro_input.rs +@@ -16,8 +16,8 @@ + /// #\[proc_macro_attribute\] attribute. + /// + /// ``` +-/// extern crate proc_macro; +-/// ++/// # extern crate proc_macro; ++/// # + /// use proc_macro::TokenStream; + /// use syn::{parse_macro_input, Result}; + /// use syn::parse::{Parse, ParseStream}; +@@ -43,7 +43,31 @@ + /// # "".parse().unwrap() + /// } + /// ``` +-#[macro_export(local_inner_macros)] ++/// ++/// <br> ++/// ++/// # Expansion ++/// ++/// `parse_macro_input!($variable as $Type)` expands to something like: ++/// ++/// ```no_run ++/// # extern crate proc_macro; ++/// # ++/// # macro_rules! doc_test { ++/// # ($variable:ident as $Type:ty) => { ++/// match syn::parse::<$Type>($variable) { ++/// Ok(syntax_tree) => syntax_tree, ++/// Err(err) => return proc_macro::TokenStream::from(err.to_compile_error()), ++/// } ++/// # }; ++/// # } ++/// # ++/// # fn test(input: proc_macro::TokenStream) -> proc_macro::TokenStream { ++/// # let _ = doc_test!(input as syn::Ident); ++/// # proc_macro::TokenStream::new() ++/// # } ++/// ``` ++#[macro_export] + macro_rules! parse_macro_input { + ($tokenstream:ident as $ty:ty) => { + match $crate::parse_macro_input::parse::<$ty>($tokenstream) { +@@ -54,7 +78,7 @@ macro_rules! parse_macro_input { + } + }; + ($tokenstream:ident) => { +- parse_macro_input!($tokenstream as _) ++ $crate::parse_macro_input!($tokenstream as _) + }; + } + +diff --git a/third_party/rust/syn/src/parse_quote.rs b/third_party/rust/syn/src/parse_quote.rs +index 18a47b95c7..66aa818cd0 100644 +--- a/third_party/rust/syn/src/parse_quote.rs ++++ b/third_party/rust/syn/src/parse_quote.rs +@@ -24,7 +24,7 @@ + /// } + /// ``` + /// +-/// *This macro is available if Syn is built with the `"parsing"` feature, ++/// *This macro is available only if Syn is built with the `"parsing"` feature, + /// although interpolation of syntax tree nodes into the quoted tokens is only + /// supported if Syn is built with the `"printing"` feature as well.* + /// +@@ -56,8 +56,10 @@ + /// or inner like `#![...]` + /// - [`Punctuated<T, P>`] — parses zero or more `T` separated by punctuation + /// `P` with optional trailing punctuation ++/// - [`Vec<Stmt>`] — parses the same as `Block::parse_within` + /// + /// [`Punctuated<T, P>`]: punctuated::Punctuated ++/// [`Vec<Stmt>`]: Block::parse_within + /// + /// # Panics + /// +@@ -67,7 +69,7 @@ + // + // TODO: allow Punctuated to be inferred as intra doc link, currently blocked on + // https://github.com/rust-lang/rust/issues/62834 +-#[macro_export(local_inner_macros)] ++#[macro_export] + macro_rules! parse_quote { + ($($tt:tt)*) => { + $crate::parse_quote::parse( +@@ -112,6 +114,8 @@ impl<T: Parse> ParseQuote for T { + use crate::punctuated::Punctuated; + #[cfg(any(feature = "full", feature = "derive"))] + use crate::{attr, Attribute}; ++#[cfg(feature = "full")] ++use crate::{Block, Stmt}; + + #[cfg(any(feature = "full", feature = "derive"))] + impl ParseQuote for Attribute { +@@ -129,3 +133,10 @@ impl<T: Parse, P: Parse> ParseQuote for Punctuated<T, P> { + Self::parse_terminated(input) + } + } ++ ++#[cfg(feature = "full")] ++impl ParseQuote for Vec<Stmt> { ++ fn parse(input: ParseStream) -> Result<Self> { ++ Block::parse_within(input) ++ } ++} +diff --git a/third_party/rust/syn/src/pat.rs b/third_party/rust/syn/src/pat.rs +index 9371e05493..e9576a2361 100644 +--- a/third_party/rust/syn/src/pat.rs ++++ b/third_party/rust/syn/src/pat.rs +@@ -1,16 +1,12 @@ + use super::*; + use crate::punctuated::Punctuated; +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; + use proc_macro2::TokenStream; +-#[cfg(feature = "extra-traits")] +-use std::hash::{Hash, Hasher}; + + ast_enum_of_structs! { + /// A pattern in a local binding, function signature, match expression, or + /// various other places. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + /// + /// # Syntax tree enum + /// +@@ -20,7 +16,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum Pat #manual_extra_traits { ++ pub enum Pat { + /// A box pattern: `box v`. + Box(PatBox), + +@@ -86,7 +82,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A box pattern: `box v`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatBox { + pub attrs: Vec<Attribute>, + pub box_token: Token![box], +@@ -97,7 +93,10 @@ ast_struct! { + ast_struct! { + /// A pattern that binds a new variable: `ref mut binding @ SUBPATTERN`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// It may also be a unit struct or struct variant (e.g. `None`), or a ++ /// constant; these cannot be distinguished syntactically. ++ /// ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatIdent { + pub attrs: Vec<Attribute>, + pub by_ref: Option<Token![ref]>, +@@ -113,7 +112,7 @@ ast_struct! { + /// This holds an `Expr` rather than a `Lit` because negative numbers + /// are represented as an `Expr::Unary`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatLit { + pub attrs: Vec<Attribute>, + pub expr: Box<Expr>, +@@ -123,7 +122,7 @@ ast_struct! { + ast_struct! { + /// A macro in pattern position. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatMacro { + pub attrs: Vec<Attribute>, + pub mac: Macro, +@@ -133,7 +132,7 @@ ast_struct! { + ast_struct! { + /// A pattern that matches any one of a set of cases. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatOr { + pub attrs: Vec<Attribute>, + pub leading_vert: Option<Token![|]>, +@@ -150,7 +149,7 @@ ast_struct! { + /// `<A>::B::C` and `<A as Trait>::B::C` can only legally refer to + /// associated constants. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatPath { + pub attrs: Vec<Attribute>, + pub qself: Option<QSelf>, +@@ -161,7 +160,7 @@ ast_struct! { + ast_struct! { + /// A range pattern: `1..=2`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatRange { + pub attrs: Vec<Attribute>, + pub lo: Box<Expr>, +@@ -173,7 +172,7 @@ ast_struct! { + ast_struct! { + /// A reference pattern: `&mut var`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatReference { + pub attrs: Vec<Attribute>, + pub and_token: Token![&], +@@ -185,7 +184,7 @@ ast_struct! { + ast_struct! { + /// The dots in a tuple or slice pattern: `[0, 1, ..]` + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatRest { + pub attrs: Vec<Attribute>, + pub dot2_token: Token![..], +@@ -195,7 +194,7 @@ ast_struct! { + ast_struct! { + /// A dynamically sized slice pattern: `[a, b, ref i @ .., y, z]`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatSlice { + pub attrs: Vec<Attribute>, + pub bracket_token: token::Bracket, +@@ -206,7 +205,7 @@ ast_struct! { + ast_struct! { + /// A struct or struct variant pattern: `Variant { x, y, .. }`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatStruct { + pub attrs: Vec<Attribute>, + pub path: Path, +@@ -219,7 +218,7 @@ ast_struct! { + ast_struct! { + /// A tuple pattern: `(a, b)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatTuple { + pub attrs: Vec<Attribute>, + pub paren_token: token::Paren, +@@ -230,7 +229,7 @@ ast_struct! { + ast_struct! { + /// A tuple struct or tuple variant pattern: `Variant(x, y, .., z)`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatTupleStruct { + pub attrs: Vec<Attribute>, + pub path: Path, +@@ -241,7 +240,7 @@ ast_struct! { + ast_struct! { + /// A type ascription pattern: `foo: f64`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatType { + pub attrs: Vec<Attribute>, + pub pat: Box<Pat>, +@@ -253,7 +252,7 @@ ast_struct! { + ast_struct! { + /// A pattern that matches any value: `_`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct PatWild { + pub attrs: Vec<Attribute>, + pub underscore_token: Token![_], +@@ -266,7 +265,7 @@ ast_struct! { + /// Patterns like the fields of Foo `{ x, ref y, ref mut z }` are treated + /// the same as `x: x, y: ref y, z: ref mut z` but there is no colon token. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct FieldPat { + pub attrs: Vec<Attribute>, + pub member: Member, +@@ -275,122 +274,17 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Pat {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Pat { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (Pat::Box(this), Pat::Box(other)) => this == other, +- (Pat::Ident(this), Pat::Ident(other)) => this == other, +- (Pat::Lit(this), Pat::Lit(other)) => this == other, +- (Pat::Macro(this), Pat::Macro(other)) => this == other, +- (Pat::Or(this), Pat::Or(other)) => this == other, +- (Pat::Path(this), Pat::Path(other)) => this == other, +- (Pat::Range(this), Pat::Range(other)) => this == other, +- (Pat::Reference(this), Pat::Reference(other)) => this == other, +- (Pat::Rest(this), Pat::Rest(other)) => this == other, +- (Pat::Slice(this), Pat::Slice(other)) => this == other, +- (Pat::Struct(this), Pat::Struct(other)) => this == other, +- (Pat::Tuple(this), Pat::Tuple(other)) => this == other, +- (Pat::TupleStruct(this), Pat::TupleStruct(other)) => this == other, +- (Pat::Type(this), Pat::Type(other)) => this == other, +- (Pat::Verbatim(this), Pat::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- (Pat::Wild(this), Pat::Wild(other)) => this == other, +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Pat { +- fn hash<H>(&self, hash: &mut H) +- where +- H: Hasher, +- { +- match self { +- Pat::Box(pat) => { +- hash.write_u8(0); +- pat.hash(hash); +- } +- Pat::Ident(pat) => { +- hash.write_u8(1); +- pat.hash(hash); +- } +- Pat::Lit(pat) => { +- hash.write_u8(2); +- pat.hash(hash); +- } +- Pat::Macro(pat) => { +- hash.write_u8(3); +- pat.hash(hash); +- } +- Pat::Or(pat) => { +- hash.write_u8(4); +- pat.hash(hash); +- } +- Pat::Path(pat) => { +- hash.write_u8(5); +- pat.hash(hash); +- } +- Pat::Range(pat) => { +- hash.write_u8(6); +- pat.hash(hash); +- } +- Pat::Reference(pat) => { +- hash.write_u8(7); +- pat.hash(hash); +- } +- Pat::Rest(pat) => { +- hash.write_u8(8); +- pat.hash(hash); +- } +- Pat::Slice(pat) => { +- hash.write_u8(9); +- pat.hash(hash); +- } +- Pat::Struct(pat) => { +- hash.write_u8(10); +- pat.hash(hash); +- } +- Pat::Tuple(pat) => { +- hash.write_u8(11); +- pat.hash(hash); +- } +- Pat::TupleStruct(pat) => { +- hash.write_u8(12); +- pat.hash(hash); +- } +- Pat::Type(pat) => { +- hash.write_u8(13); +- pat.hash(hash); +- } +- Pat::Verbatim(pat) => { +- hash.write_u8(14); +- TokenStreamHelper(pat).hash(hash); +- } +- Pat::Wild(pat) => { +- hash.write_u8(15); +- pat.hash(hash); +- } +- Pat::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + #[cfg(feature = "parsing")] +-mod parsing { ++pub mod parsing { + use super::*; + + use crate::ext::IdentExt; +- use crate::parse::{Parse, ParseStream, Result}; ++ use crate::parse::{Parse, ParseBuffer, ParseStream, Result}; + use crate::path; + + impl Parse for Pat { + fn parse(input: ParseStream) -> Result<Self> { ++ let begin = input.fork(); + let lookahead = input.lookahead1(); + if lookahead.peek(Ident) + && ({ +@@ -411,7 +305,6 @@ mod parsing { + || lookahead.peek(Token![<]) + || input.peek(Token![Self]) + || input.peek(Token![super]) +- || input.peek(Token![extern]) + || input.peek(Token![crate]) + { + pat_path_or_macro_or_struct_or_range(input) +@@ -434,7 +327,7 @@ mod parsing { + } else if lookahead.peek(token::Bracket) { + input.call(pat_slice).map(Pat::Slice) + } else if lookahead.peek(Token![..]) && !input.peek(Token![...]) { +- input.call(pat_rest).map(Pat::Rest) ++ pat_range_half_open(input, begin) + } else { + Err(lookahead.error()) + } +@@ -442,10 +335,11 @@ mod parsing { + } + + fn pat_path_or_macro_or_struct_or_range(input: ParseStream) -> Result<Pat> { ++ let begin = input.fork(); + let (qself, path) = path::parsing::qpath(input, true)?; + + if input.peek(Token![..]) { +- return pat_range(input, qself, path).map(Pat::Range); ++ return pat_range(input, begin, qself, path); + } + + if qself.is_some() { +@@ -487,7 +381,7 @@ mod parsing { + } else if input.peek(token::Paren) { + pat_tuple_struct(input, path).map(Pat::TupleStruct) + } else if input.peek(Token![..]) { +- pat_range(input, qself, path).map(Pat::Range) ++ pat_range(input, begin, qself, path) + } else { + Ok(Pat::Path(PatPath { + attrs: Vec::new(), +@@ -546,7 +440,7 @@ mod parsing { + while !content.is_empty() && !content.peek(Token![..]) { + let value = content.call(field_pat)?; + fields.push_value(value); +- if !content.peek(Token![,]) { ++ if content.is_empty() { + break; + } + let punct: Token![,] = content.parse()?; +@@ -578,6 +472,7 @@ mod parsing { + } + + fn field_pat(input: ParseStream) -> Result<FieldPat> { ++ let attrs = input.call(Attribute::parse_outer)?; + let boxed: Option<Token![box]> = input.parse()?; + let by_ref: Option<Token![ref]> = input.parse()?; + let mutability: Option<Token![mut]> = input.parse()?; +@@ -587,10 +482,10 @@ mod parsing { + || member.is_unnamed() + { + return Ok(FieldPat { +- attrs: Vec::new(), ++ attrs, + member, + colon_token: input.parse()?, +- pat: input.parse()?, ++ pat: Box::new(multi_pat(input)?), + }); + } + +@@ -610,30 +505,57 @@ mod parsing { + if let Some(boxed) = boxed { + pat = Pat::Box(PatBox { + attrs: Vec::new(), +- pat: Box::new(pat), + box_token: boxed, ++ pat: Box::new(pat), + }); + } + + Ok(FieldPat { ++ attrs, + member: Member::Named(ident), +- pat: Box::new(pat), +- attrs: Vec::new(), + colon_token: None, ++ pat: Box::new(pat), + }) + } + +- fn pat_range(input: ParseStream, qself: Option<QSelf>, path: Path) -> Result<PatRange> { +- Ok(PatRange { +- attrs: Vec::new(), +- lo: Box::new(Expr::Path(ExprPath { ++ fn pat_range( ++ input: ParseStream, ++ begin: ParseBuffer, ++ qself: Option<QSelf>, ++ path: Path, ++ ) -> Result<Pat> { ++ let limits: RangeLimits = input.parse()?; ++ let hi = input.call(pat_lit_expr)?; ++ if let Some(hi) = hi { ++ Ok(Pat::Range(PatRange { + attrs: Vec::new(), +- qself, +- path, +- })), +- limits: input.parse()?, +- hi: input.call(pat_lit_expr)?, +- }) ++ lo: Box::new(Expr::Path(ExprPath { ++ attrs: Vec::new(), ++ qself, ++ path, ++ })), ++ limits, ++ hi, ++ })) ++ } else { ++ Ok(Pat::Verbatim(verbatim::between(begin, input))) ++ } ++ } ++ ++ fn pat_range_half_open(input: ParseStream, begin: ParseBuffer) -> Result<Pat> { ++ let limits: RangeLimits = input.parse()?; ++ let hi = input.call(pat_lit_expr)?; ++ if hi.is_some() { ++ Ok(Pat::Verbatim(verbatim::between(begin, input))) ++ } else { ++ match limits { ++ RangeLimits::HalfOpen(dot2_token) => Ok(Pat::Rest(PatRest { ++ attrs: Vec::new(), ++ dot2_token, ++ })), ++ RangeLimits::Closed(_) => Err(input.error("expected range upper bound")), ++ } ++ } + } + + fn pat_tuple(input: ParseStream) -> Result<PatTuple> { +@@ -642,7 +564,7 @@ mod parsing { + + let mut elems = Punctuated::new(); + while !content.is_empty() { +- let value: Pat = content.parse()?; ++ let value = multi_pat(&content)?; + elems.push_value(value); + if content.is_empty() { + break; +@@ -668,14 +590,21 @@ mod parsing { + } + + fn pat_lit_or_range(input: ParseStream) -> Result<Pat> { +- let lo = input.call(pat_lit_expr)?; ++ let begin = input.fork(); ++ let lo = input.call(pat_lit_expr)?.unwrap(); + if input.peek(Token![..]) { +- Ok(Pat::Range(PatRange { +- attrs: Vec::new(), +- lo, +- limits: input.parse()?, +- hi: input.call(pat_lit_expr)?, +- })) ++ let limits: RangeLimits = input.parse()?; ++ let hi = input.call(pat_lit_expr)?; ++ if let Some(hi) = hi { ++ Ok(Pat::Range(PatRange { ++ attrs: Vec::new(), ++ lo, ++ limits, ++ hi, ++ })) ++ } else { ++ Ok(Pat::Verbatim(verbatim::between(begin, input))) ++ } + } else { + Ok(Pat::Lit(PatLit { + attrs: Vec::new(), +@@ -684,7 +613,17 @@ mod parsing { + } + } + +- fn pat_lit_expr(input: ParseStream) -> Result<Box<Expr>> { ++ fn pat_lit_expr(input: ParseStream) -> Result<Option<Box<Expr>>> { ++ if input.is_empty() ++ || input.peek(Token![|]) ++ || input.peek(Token![=>]) ++ || input.peek(Token![:]) && !input.peek(Token![::]) ++ || input.peek(Token![,]) ++ || input.peek(Token![;]) ++ { ++ return Ok(None); ++ } ++ + let neg: Option<Token![-]> = input.parse()?; + + let lookahead = input.lookahead1(); +@@ -696,7 +635,6 @@ mod parsing { + || lookahead.peek(Token![self]) + || lookahead.peek(Token![Self]) + || lookahead.peek(Token![super]) +- || lookahead.peek(Token![extern]) + || lookahead.peek(Token![crate]) + { + Expr::Path(input.parse()?) +@@ -704,7 +642,7 @@ mod parsing { + return Err(lookahead.error()); + }; + +- Ok(Box::new(if let Some(neg) = neg { ++ Ok(Some(Box::new(if let Some(neg) = neg { + Expr::Unary(ExprUnary { + attrs: Vec::new(), + op: UnOp::Neg(neg), +@@ -712,7 +650,7 @@ mod parsing { + }) + } else { + expr +- })) ++ }))) + } + + fn pat_slice(input: ParseStream) -> Result<PatSlice> { +@@ -721,7 +659,7 @@ mod parsing { + + let mut elems = Punctuated::new(); + while !content.is_empty() { +- let value: Pat = content.parse()?; ++ let value = multi_pat(&content)?; + elems.push_value(value); + if content.is_empty() { + break; +@@ -737,11 +675,35 @@ mod parsing { + }) + } + +- fn pat_rest(input: ParseStream) -> Result<PatRest> { +- Ok(PatRest { +- attrs: Vec::new(), +- dot2_token: input.parse()?, +- }) ++ pub fn multi_pat(input: ParseStream) -> Result<Pat> { ++ multi_pat_impl(input, None) ++ } ++ ++ pub fn multi_pat_with_leading_vert(input: ParseStream) -> Result<Pat> { ++ let leading_vert: Option<Token![|]> = input.parse()?; ++ multi_pat_impl(input, leading_vert) ++ } ++ ++ fn multi_pat_impl(input: ParseStream, leading_vert: Option<Token![|]>) -> Result<Pat> { ++ let mut pat: Pat = input.parse()?; ++ if leading_vert.is_some() ++ || input.peek(Token![|]) && !input.peek(Token![||]) && !input.peek(Token![|=]) ++ { ++ let mut cases = Punctuated::new(); ++ cases.push_value(pat); ++ while input.peek(Token![|]) && !input.peek(Token![||]) && !input.peek(Token![|=]) { ++ let punct = input.parse()?; ++ cases.push_punct(punct); ++ let pat: Pat = input.parse()?; ++ cases.push_value(pat); ++ } ++ pat = Pat::Or(PatOr { ++ attrs: Vec::new(), ++ leading_vert, ++ cases, ++ }); ++ } ++ Ok(pat) + } + } + +@@ -756,12 +718,14 @@ mod printing { + + impl ToTokens for PatWild { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.underscore_token.to_tokens(tokens); + } + } + + impl ToTokens for PatIdent { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.by_ref.to_tokens(tokens); + self.mutability.to_tokens(tokens); + self.ident.to_tokens(tokens); +@@ -774,6 +738,7 @@ mod printing { + + impl ToTokens for PatStruct { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.path.to_tokens(tokens); + self.brace_token.surround(tokens, |tokens| { + self.fields.to_tokens(tokens); +@@ -788,6 +753,7 @@ mod printing { + + impl ToTokens for PatTupleStruct { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.path.to_tokens(tokens); + self.pat.to_tokens(tokens); + } +@@ -804,12 +770,14 @@ mod printing { + + impl ToTokens for PatPath { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + private::print_path(tokens, &self.qself, &self.path); + } + } + + impl ToTokens for PatTuple { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.paren_token.surround(tokens, |tokens| { + self.elems.to_tokens(tokens); + }); +@@ -818,6 +786,7 @@ mod printing { + + impl ToTokens for PatBox { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.box_token.to_tokens(tokens); + self.pat.to_tokens(tokens); + } +@@ -825,6 +794,7 @@ mod printing { + + impl ToTokens for PatReference { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.and_token.to_tokens(tokens); + self.mutability.to_tokens(tokens); + self.pat.to_tokens(tokens); +@@ -833,18 +803,21 @@ mod printing { + + impl ToTokens for PatRest { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.dot2_token.to_tokens(tokens); + } + } + + impl ToTokens for PatLit { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.expr.to_tokens(tokens); + } + } + + impl ToTokens for PatRange { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.lo.to_tokens(tokens); + match &self.limits { + RangeLimits::HalfOpen(t) => t.to_tokens(tokens), +@@ -856,6 +829,7 @@ mod printing { + + impl ToTokens for PatSlice { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.bracket_token.surround(tokens, |tokens| { + self.elems.to_tokens(tokens); + }); +@@ -864,12 +838,14 @@ mod printing { + + impl ToTokens for PatMacro { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.mac.to_tokens(tokens); + } + } + + impl ToTokens for PatOr { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + self.leading_vert.to_tokens(tokens); + self.cases.to_tokens(tokens); + } +@@ -877,6 +853,7 @@ mod printing { + + impl ToTokens for FieldPat { + fn to_tokens(&self, tokens: &mut TokenStream) { ++ tokens.append_all(self.attrs.outer()); + if let Some(colon_token) = &self.colon_token { + self.member.to_tokens(tokens); + colon_token.to_tokens(tokens); +diff --git a/third_party/rust/syn/src/path.rs b/third_party/rust/syn/src/path.rs +index 8dda43ee67..15c0fcc664 100644 +--- a/third_party/rust/syn/src/path.rs ++++ b/third_party/rust/syn/src/path.rs +@@ -2,9 +2,9 @@ use super::*; + use crate::punctuated::Punctuated; + + ast_struct! { +- /// A path at which a named item is exported: `std::collections::HashMap`. ++ /// A path at which a named item is exported (e.g. `std::collections::HashMap`). + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Path { + pub leading_colon: Option<Token![::]>, +@@ -29,7 +29,7 @@ where + ast_struct! { + /// A segment of a path together with any path arguments on that segment. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct PathSegment { + pub ident: Ident, +@@ -52,7 +52,7 @@ where + ast_enum! { + /// Angle bracketed or parenthesized arguments of a path segment. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// ## Angle bracketed +@@ -98,7 +98,7 @@ impl PathArguments { + ast_enum! { + /// An individual generic argument, like `'a`, `T`, or `Item = T`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum GenericArgument { + /// A lifetime argument. +@@ -122,7 +122,7 @@ ast_struct! { + /// Angle bracketed arguments of a path segment: the `<K, V>` in `HashMap<K, + /// V>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct AngleBracketedGenericArguments { + pub colon2_token: Option<Token![::]>, +@@ -135,7 +135,7 @@ ast_struct! { + ast_struct! { + /// A binding (equality constraint) on an associated type: `Item = u8`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Binding { + pub ident: Ident, +@@ -147,7 +147,7 @@ ast_struct! { + ast_struct! { + /// An associated type bound: `Iterator<Item: Display>`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Constraint { + pub ident: Ident, +@@ -160,7 +160,7 @@ ast_struct! { + /// Arguments of a function path segment: the `(A, B) -> C` in `Fn(A,B) -> + /// C`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct ParenthesizedGenericArguments { + pub paren_token: token::Paren, +@@ -189,7 +189,7 @@ ast_struct! { + /// ty position = 0 + /// ``` + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct QSelf { + pub lt_token: Token![<], +@@ -291,11 +291,7 @@ pub mod parsing { + + impl PathSegment { + fn parse_helper(input: ParseStream, expr_style: bool) -> Result<Self> { +- if input.peek(Token![super]) +- || input.peek(Token![self]) +- || input.peek(Token![crate]) +- || input.peek(Token![extern]) +- { ++ if input.peek(Token![super]) || input.peek(Token![self]) || input.peek(Token![crate]) { + let ident = input.call(Ident::parse_any)?; + return Ok(PathSegment::from(ident)); + } +@@ -358,7 +354,7 @@ pub mod parsing { + impl Path { + /// Parse a `Path` containing no path arguments on any of its segments. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + /// + /// # Example +@@ -400,7 +396,6 @@ pub mod parsing { + && !input.peek(Token![self]) + && !input.peek(Token![Self]) + && !input.peek(Token![crate]) +- && !input.peek(Token![extern]) + { + break; + } +@@ -433,7 +428,7 @@ pub mod parsing { + /// path arguments, and + /// - the ident of the first path segment is equal to the given one. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + /// + /// # Example +@@ -472,7 +467,7 @@ pub mod parsing { + /// - the first path segment has no angle bracketed or parenthesized + /// path arguments. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + pub fn get_ident(&self) -> Option<&Ident> { + if self.leading_colon.is_none() +diff --git a/third_party/rust/syn/src/punctuated.rs b/third_party/rust/syn/src/punctuated.rs +index 38c7bf4e82..46c82a65b1 100644 +--- a/third_party/rust/syn/src/punctuated.rs ++++ b/third_party/rust/syn/src/punctuated.rs +@@ -22,6 +22,8 @@ + + #[cfg(feature = "extra-traits")] + use std::fmt::{self, Debug}; ++#[cfg(feature = "extra-traits")] ++use std::hash::{Hash, Hasher}; + #[cfg(any(feature = "full", feature = "derive"))] + use std::iter; + use std::iter::FromIterator; +@@ -41,8 +43,6 @@ use crate::token::Token; + /// Refer to the [module documentation] for details about punctuated sequences. + /// + /// [module documentation]: self +-#[cfg_attr(feature = "extra-traits", derive(Eq, PartialEq, Hash))] +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + pub struct Punctuated<T, P> { + inner: Vec<(T, P)>, + last: Option<Box<T>>, +@@ -76,22 +76,19 @@ impl<T, P> Punctuated<T, P> { + self.iter().next() + } + ++ /// Mutably borrows the first element in this sequence. ++ pub fn first_mut(&mut self) -> Option<&mut T> { ++ self.iter_mut().next() ++ } ++ + /// Borrows the last element in this sequence. + pub fn last(&self) -> Option<&T> { +- if self.last.is_some() { +- self.last.as_ref().map(Box::as_ref) +- } else { +- self.inner.last().map(|pair| &pair.0) +- } ++ self.iter().next_back() + } + + /// Mutably borrows the last element in this sequence. + pub fn last_mut(&mut self) -> Option<&mut T> { +- if self.last.is_some() { +- self.last.as_mut().map(Box::as_mut) +- } else { +- self.inner.last_mut().map(|pair| &mut pair.0) +- } ++ self.iter_mut().next_back() + } + + /// Returns an iterator over borrowed syntax tree nodes of type `&T`. +@@ -230,13 +227,19 @@ impl<T, P> Punctuated<T, P> { + } + } + ++ /// Clears the sequence of all values and punctuation, making it empty. ++ pub fn clear(&mut self) { ++ self.inner.clear(); ++ self.last = None; ++ } ++ + /// Parses zero or more occurrences of `T` separated by punctuation of type + /// `P`, with optional trailing punctuation. + /// + /// Parsing continues until the end of this parse stream. The entire content + /// of this parse stream must consist of `T` and `P`. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_terminated(input: ParseStream) -> Result<Self> +@@ -256,7 +259,7 @@ impl<T, P> Punctuated<T, P> { + /// + /// [`parse_terminated`]: Punctuated::parse_terminated + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_terminated_with( +@@ -292,7 +295,7 @@ impl<T, P> Punctuated<T, P> { + /// is not followed by a `P`, even if there are remaining tokens in the + /// stream. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_separated_nonempty(input: ParseStream) -> Result<Self> +@@ -312,7 +315,7 @@ impl<T, P> Punctuated<T, P> { + /// + /// [`parse_separated_nonempty`]: Punctuated::parse_separated_nonempty + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + #[cfg(feature = "parsing")] + pub fn parse_separated_nonempty_with( +@@ -338,6 +341,53 @@ impl<T, P> Punctuated<T, P> { + } + } + ++#[cfg(feature = "clone-impls")] ++impl<T, P> Clone for Punctuated<T, P> ++where ++ T: Clone, ++ P: Clone, ++{ ++ fn clone(&self) -> Self { ++ Punctuated { ++ inner: self.inner.clone(), ++ last: self.last.clone(), ++ } ++ } ++} ++ ++#[cfg(feature = "extra-traits")] ++impl<T, P> Eq for Punctuated<T, P> ++where ++ T: Eq, ++ P: Eq, ++{ ++} ++ ++#[cfg(feature = "extra-traits")] ++impl<T, P> PartialEq for Punctuated<T, P> ++where ++ T: PartialEq, ++ P: PartialEq, ++{ ++ fn eq(&self, other: &Self) -> bool { ++ let Punctuated { inner, last } = self; ++ *inner == other.inner && *last == other.last ++ } ++} ++ ++#[cfg(feature = "extra-traits")] ++impl<T, P> Hash for Punctuated<T, P> ++where ++ T: Hash, ++ P: Hash, ++{ ++ fn hash<H: Hasher>(&self, state: &mut H) { ++ let Punctuated { inner, last } = self; ++ inner.hash(state); ++ last.hash(state); ++ } ++} ++ + #[cfg(feature = "extra-traits")] + impl<T: Debug, P: Debug> Debug for Punctuated<T, P> { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +@@ -536,7 +586,6 @@ impl<'a, T, P> ExactSizeIterator for PairsMut<'a, T, P> { + /// Refer to the [module documentation] for details about punctuated sequences. + /// + /// [module documentation]: self +-#[derive(Clone)] + pub struct IntoPairs<T, P> { + inner: vec::IntoIter<(T, P)>, + last: option::IntoIter<T>, +@@ -572,12 +621,24 @@ impl<T, P> ExactSizeIterator for IntoPairs<T, P> { + } + } + ++impl<T, P> Clone for IntoPairs<T, P> ++where ++ T: Clone, ++ P: Clone, ++{ ++ fn clone(&self) -> Self { ++ IntoPairs { ++ inner: self.inner.clone(), ++ last: self.last.clone(), ++ } ++ } ++} ++ + /// An iterator over owned values of type `T`. + /// + /// Refer to the [module documentation] for details about punctuated sequences. + /// + /// [module documentation]: self +-#[derive(Clone)] + pub struct IntoIter<T> { + inner: vec::IntoIter<T>, + } +@@ -606,6 +667,17 @@ impl<T> ExactSizeIterator for IntoIter<T> { + } + } + ++impl<T> Clone for IntoIter<T> ++where ++ T: Clone, ++{ ++ fn clone(&self) -> Self { ++ IntoIter { ++ inner: self.inner.clone(), ++ } ++ } ++} ++ + /// An iterator over borrowed values of type `&T`. + /// + /// Refer to the [module documentation] for details about punctuated sequences. +@@ -799,7 +871,6 @@ impl<'a, T: 'a, I: 'a> IterMutTrait<'a, T> for I where + /// Refer to the [module documentation] for details about punctuated sequences. + /// + /// [module documentation]: self +-#[cfg_attr(feature = "clone-impls", derive(Clone))] + pub enum Pair<T, P> { + Punctuated(T, P), + End(T), +@@ -856,6 +927,20 @@ impl<T, P> Pair<T, P> { + } + } + ++#[cfg(feature = "clone-impls")] ++impl<T, P> Clone for Pair<T, P> ++where ++ T: Clone, ++ P: Clone, ++{ ++ fn clone(&self) -> Self { ++ match self { ++ Pair::Punctuated(t, p) => Pair::Punctuated(t.clone(), p.clone()), ++ Pair::End(t) => Pair::End(t.clone()), ++ } ++ } ++} ++ + impl<T, P> Index<usize> for Punctuated<T, P> { + type Output = T; + +diff --git a/third_party/rust/syn/src/reserved.rs b/third_party/rust/syn/src/reserved.rs +new file mode 100644 +index 0000000000..ccfb8b5ad0 +--- /dev/null ++++ b/third_party/rust/syn/src/reserved.rs +@@ -0,0 +1,42 @@ ++// Type for a syntax tree node that is reserved for future use. ++// ++// For example ExprReference contains a field `raw` of type Reserved. If `&raw ++// place` syntax becomes a thing as per https://github.com/rust-lang/rfcs/pull/2582, ++// we can backward compatibly change `raw`'s type to Option<Token![raw]> without ++// the possibility of breaking any code. ++ ++use proc_macro2::Span; ++use std::marker::PhantomData; ++ ++#[cfg(feature = "extra-traits")] ++use std::fmt::{self, Debug}; ++ ++ast_struct! { ++ pub struct Reserved { ++ _private: PhantomData<Span>, ++ } ++} ++ ++impl Default for Reserved { ++ fn default() -> Self { ++ Reserved { ++ _private: PhantomData, ++ } ++ } ++} ++ ++#[cfg(feature = "clone-impls")] ++impl Clone for Reserved { ++ fn clone(&self) -> Self { ++ Reserved { ++ _private: self._private, ++ } ++ } ++} ++ ++#[cfg(feature = "extra-traits")] ++impl Debug for Reserved { ++ fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { ++ formatter.debug_struct("Reserved").finish() ++ } ++} +diff --git a/third_party/rust/syn/src/spanned.rs b/third_party/rust/syn/src/spanned.rs +index 71ffe26b81..01591cedcb 100644 +--- a/third_party/rust/syn/src/spanned.rs ++++ b/third_party/rust/syn/src/spanned.rs +@@ -1,7 +1,7 @@ + //! A trait that can provide the `Span` of the complete contents of a syntax + //! tree node. + //! +-//! *This module is available if Syn is built with both the `"parsing"` and ++//! *This module is available only if Syn is built with both the `"parsing"` and + //! `"printing"` features.* + //! + //! <br> +@@ -97,7 +97,7 @@ use quote::spanned::Spanned as ToTokens; + /// + /// [module documentation]: self + /// +-/// *This trait is available if Syn is built with both the `"parsing"` and ++/// *This trait is available only if Syn is built with both the `"parsing"` and + /// `"printing"` features.* + pub trait Spanned { + /// Returns a `Span` covering the complete contents of this syntax tree +diff --git a/third_party/rust/syn/src/stmt.rs b/third_party/rust/syn/src/stmt.rs +index e4277fdbaa..b06e843d75 100644 +--- a/third_party/rust/syn/src/stmt.rs ++++ b/third_party/rust/syn/src/stmt.rs +@@ -3,7 +3,7 @@ use super::*; + ast_struct! { + /// A braced block containing Rust statements. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Block { + pub brace_token: token::Brace, + /// Statements in a block +@@ -14,7 +14,7 @@ ast_struct! { + ast_enum! { + /// A statement, usually ending in a semicolon. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub enum Stmt { + /// A local (let) binding. + Local(Local), +@@ -33,7 +33,7 @@ ast_enum! { + ast_struct! { + /// A local `let` binding: `let x: u64 = s.parse()?`. + /// +- /// *This type is available if Syn is built with the `"full"` feature.* ++ /// *This type is available only if Syn is built with the `"full"` feature.* + pub struct Local { + pub attrs: Vec<Attribute>, + pub let_token: Token![let], +@@ -47,14 +47,15 @@ ast_struct! { + pub mod parsing { + use super::*; + ++ use crate::parse::discouraged::Speculative; + use crate::parse::{Parse, ParseStream, Result}; +- use crate::punctuated::Punctuated; ++ use proc_macro2::TokenStream; + + impl Block { + /// Parse the body of a block as zero or more statements, possibly + /// including one trailing expression. + /// +- /// *This function is available if Syn is built with the `"parsing"` ++ /// *This function is available only if Syn is built with the `"parsing"` + /// feature.* + /// + /// # Example +@@ -106,8 +107,8 @@ pub mod parsing { + pub fn parse_within(input: ParseStream) -> Result<Vec<Stmt>> { + let mut stmts = Vec::new(); + loop { +- while input.peek(Token![;]) { +- input.parse::<Token![;]>()?; ++ while let Some(semi) = input.parse::<Option<Token![;]>>()? { ++ stmts.push(Stmt::Semi(Expr::Verbatim(TokenStream::new()), semi)); + } + if input.is_empty() { + break; +@@ -146,55 +147,55 @@ pub mod parsing { + } + + fn parse_stmt(input: ParseStream, allow_nosemi: bool) -> Result<Stmt> { +- // TODO: optimize using advance_to ++ let mut attrs = input.call(Attribute::parse_outer)?; ++ ++ // brace-style macros; paren and bracket macros get parsed as ++ // expression statements. + let ahead = input.fork(); +- ahead.call(Attribute::parse_outer)?; ++ if let Ok(path) = ahead.call(Path::parse_mod_style) { ++ if ahead.peek(Token![!]) && (ahead.peek2(token::Brace) || ahead.peek2(Ident)) { ++ input.advance_to(&ahead); ++ return stmt_mac(input, attrs, path); ++ } ++ } + +- if { +- let ahead = ahead.fork(); +- // Only parse braces here; paren and bracket will get parsed as +- // expression statements +- ahead.call(Path::parse_mod_style).is_ok() +- && ahead.parse::<Token![!]>().is_ok() +- && (ahead.peek(token::Brace) || ahead.peek(Ident)) +- } { +- stmt_mac(input) +- } else if ahead.peek(Token![let]) { +- stmt_local(input).map(Stmt::Local) +- } else if ahead.peek(Token![pub]) +- || ahead.peek(Token![crate]) && !ahead.peek2(Token![::]) +- || ahead.peek(Token![extern]) && !ahead.peek2(Token![::]) +- || ahead.peek(Token![use]) +- || ahead.peek(Token![static]) && (ahead.peek2(Token![mut]) || ahead.peek2(Ident)) +- || ahead.peek(Token![const]) +- || ahead.peek(Token![unsafe]) && !ahead.peek2(token::Brace) +- || ahead.peek(Token![async]) +- && (ahead.peek2(Token![unsafe]) +- || ahead.peek2(Token![extern]) +- || ahead.peek2(Token![fn])) +- || ahead.peek(Token![fn]) +- || ahead.peek(Token![mod]) +- || ahead.peek(Token![type]) +- || ahead.peek(item::parsing::existential) && ahead.peek2(Token![type]) +- || ahead.peek(Token![struct]) +- || ahead.peek(Token![enum]) +- || ahead.peek(Token![union]) && ahead.peek2(Ident) +- || ahead.peek(Token![auto]) && ahead.peek2(Token![trait]) +- || ahead.peek(Token![trait]) +- || ahead.peek(Token![default]) +- && (ahead.peek2(Token![unsafe]) || ahead.peek2(Token![impl])) +- || ahead.peek(Token![impl]) +- || ahead.peek(Token![macro]) ++ if input.peek(Token![let]) { ++ stmt_local(input, attrs).map(Stmt::Local) ++ } else if input.peek(Token![pub]) ++ || input.peek(Token![crate]) && !input.peek2(Token![::]) ++ || input.peek(Token![extern]) ++ || input.peek(Token![use]) ++ || input.peek(Token![static]) && (input.peek2(Token![mut]) || input.peek2(Ident)) ++ || input.peek(Token![const]) ++ || input.peek(Token![unsafe]) && !input.peek2(token::Brace) ++ || input.peek(Token![async]) ++ && (input.peek2(Token![unsafe]) ++ || input.peek2(Token![extern]) ++ || input.peek2(Token![fn])) ++ || input.peek(Token![fn]) ++ || input.peek(Token![mod]) ++ || input.peek(Token![type]) ++ || input.peek(item::parsing::existential) && input.peek2(Token![type]) ++ || input.peek(Token![struct]) ++ || input.peek(Token![enum]) ++ || input.peek(Token![union]) && input.peek2(Ident) ++ || input.peek(Token![auto]) && input.peek2(Token![trait]) ++ || input.peek(Token![trait]) ++ || input.peek(Token![default]) ++ && (input.peek2(Token![unsafe]) || input.peek2(Token![impl])) ++ || input.peek(Token![impl]) ++ || input.peek(Token![macro]) + { +- input.parse().map(Stmt::Item) ++ let mut item: Item = input.parse()?; ++ attrs.extend(item.replace_attrs(Vec::new())); ++ item.replace_attrs(attrs); ++ Ok(Stmt::Item(item)) + } else { +- stmt_expr(input, allow_nosemi) ++ stmt_expr(input, allow_nosemi, attrs) + } + } + +- fn stmt_mac(input: ParseStream) -> Result<Stmt> { +- let attrs = input.call(Attribute::parse_outer)?; +- let path = input.call(Path::parse_mod_style)?; ++ fn stmt_mac(input: ParseStream, attrs: Vec<Attribute>, path: Path) -> Result<Stmt> { + let bang_token: Token![!] = input.parse()?; + let ident: Option<Ident> = input.parse()?; + let (delimiter, tokens) = mac::parse_delimiter(input)?; +@@ -213,33 +214,12 @@ pub mod parsing { + }))) + } + +- fn stmt_local(input: ParseStream) -> Result<Local> { ++ fn stmt_local(input: ParseStream, attrs: Vec<Attribute>) -> Result<Local> { + Ok(Local { +- attrs: input.call(Attribute::parse_outer)?, ++ attrs, + let_token: input.parse()?, + pat: { +- let leading_vert: Option<Token![|]> = input.parse()?; +- let mut pat: Pat = input.parse()?; +- if leading_vert.is_some() +- || input.peek(Token![|]) && !input.peek(Token![||]) && !input.peek(Token![|=]) +- { +- let mut cases = Punctuated::new(); +- cases.push_value(pat); +- while input.peek(Token![|]) +- && !input.peek(Token![||]) +- && !input.peek(Token![|=]) +- { +- let punct = input.parse()?; +- cases.push_punct(punct); +- let pat: Pat = input.parse()?; +- cases.push_value(pat); +- } +- pat = Pat::Or(PatOr { +- attrs: Vec::new(), +- leading_vert, +- cases, +- }); +- } ++ let mut pat: Pat = pat::parsing::multi_pat_with_leading_vert(input)?; + if input.peek(Token![:]) { + let colon_token: Token![:] = input.parse()?; + let ty: Type = input.parse()?; +@@ -265,12 +245,19 @@ pub mod parsing { + }) + } + +- fn stmt_expr(input: ParseStream, allow_nosemi: bool) -> Result<Stmt> { +- let mut attrs = input.call(Attribute::parse_outer)?; ++ fn stmt_expr( ++ input: ParseStream, ++ allow_nosemi: bool, ++ mut attrs: Vec<Attribute>, ++ ) -> Result<Stmt> { + let mut e = expr::parsing::expr_early(input)?; + +- attrs.extend(e.replace_attrs(Vec::new())); +- e.replace_attrs(attrs); ++ let mut attr_target = &mut e; ++ while let Expr::Binary(e) = attr_target { ++ attr_target = &mut e.left; ++ } ++ attrs.extend(attr_target.replace_attrs(Vec::new())); ++ attr_target.replace_attrs(attrs); + + if input.peek(Token![;]) { + return Ok(Stmt::Semi(e, input.parse()?)); +diff --git a/third_party/rust/syn/src/token.rs b/third_party/rust/syn/src/token.rs +index 0b8c18192f..8539378c5e 100644 +--- a/third_party/rust/syn/src/token.rs ++++ b/third_party/rust/syn/src/token.rs +@@ -88,7 +88,6 @@ + //! [Printing]: https://docs.rs/quote/1.0/quote/trait.ToTokens.html + //! [`Span`]: https://docs.rs/proc-macro2/1.0/proc_macro2/struct.Span.html + +-use std; + #[cfg(feature = "extra-traits")] + use std::cmp; + #[cfg(feature = "extra-traits")] +@@ -97,13 +96,13 @@ use std::fmt::{self, Debug}; + use std::hash::{Hash, Hasher}; + use std::ops::{Deref, DerefMut}; + +-#[cfg(feature = "parsing")] +-use proc_macro2::Delimiter; + #[cfg(any(feature = "parsing", feature = "printing"))] + use proc_macro2::Ident; + use proc_macro2::Span; + #[cfg(feature = "printing")] + use proc_macro2::TokenStream; ++#[cfg(feature = "parsing")] ++use proc_macro2::{Delimiter, Literal, Punct, TokenTree}; + #[cfg(feature = "printing")] + use quote::{ToTokens, TokenStreamExt}; + +@@ -112,10 +111,8 @@ use self::private::WithSpan; + use crate::buffer::Cursor; + #[cfg(feature = "parsing")] + use crate::error::Result; +-#[cfg(any(feature = "full", feature = "derive"))] + #[cfg(feature = "parsing")] + use crate::lifetime::Lifetime; +-#[cfg(any(feature = "full", feature = "derive"))] + #[cfg(feature = "parsing")] + use crate::lit::{Lit, LitBool, LitByte, LitByteStr, LitChar, LitFloat, LitInt, LitStr}; + #[cfg(feature = "parsing")] +@@ -155,21 +152,20 @@ mod private { + #[cfg(feature = "parsing")] + impl private::Sealed for Ident {} + +-#[cfg(any(feature = "full", feature = "derive"))] + #[cfg(feature = "parsing")] + fn peek_impl(cursor: Cursor, peek: fn(ParseStream) -> bool) -> bool { ++ use crate::parse::Unexpected; + use std::cell::Cell; + use std::rc::Rc; + + let scope = Span::call_site(); +- let unexpected = Rc::new(Cell::new(None)); ++ let unexpected = Rc::new(Cell::new(Unexpected::None)); + let buffer = crate::parse::new_parse_buffer(scope, cursor, unexpected); + peek(&buffer) + } + +-#[cfg(any(feature = "full", feature = "derive"))] + macro_rules! impl_token { +- ($name:ident $display:expr) => { ++ ($display:tt $name:ty) => { + #[cfg(feature = "parsing")] + impl Token for $name { + fn peek(cursor: Cursor) -> bool { +@@ -189,24 +185,38 @@ macro_rules! impl_token { + }; + } + +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(Lifetime "lifetime"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(Lit "literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitStr "string literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitByteStr "byte string literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitByte "byte literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitChar "character literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitInt "integer literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitFloat "floating point literal"); +-#[cfg(any(feature = "full", feature = "derive"))] +-impl_token!(LitBool "boolean literal"); ++impl_token!("lifetime" Lifetime); ++impl_token!("literal" Lit); ++impl_token!("string literal" LitStr); ++impl_token!("byte string literal" LitByteStr); ++impl_token!("byte literal" LitByte); ++impl_token!("character literal" LitChar); ++impl_token!("integer literal" LitInt); ++impl_token!("floating point literal" LitFloat); ++impl_token!("boolean literal" LitBool); ++impl_token!("group token" proc_macro2::Group); ++ ++macro_rules! impl_low_level_token { ++ ($display:tt $ty:ident $get:ident) => { ++ #[cfg(feature = "parsing")] ++ impl Token for $ty { ++ fn peek(cursor: Cursor) -> bool { ++ cursor.$get().is_some() ++ } ++ ++ fn display() -> &'static str { ++ $display ++ } ++ } ++ ++ #[cfg(feature = "parsing")] ++ impl private::Sealed for $ty {} ++ }; ++} ++ ++impl_low_level_token!("punctuation token" Punct punct); ++impl_low_level_token!("literal" Literal literal); ++impl_low_level_token!("token" TokenTree token_tree); + + // Not public API. + #[doc(hidden)] +@@ -233,7 +243,6 @@ impl<T: CustomToken> Token for T { + macro_rules! define_keywords { + ($($token:tt pub struct $name:ident #[$doc:meta])*) => { + $( +- #[cfg_attr(feature = "clone-impls", derive(Copy, Clone))] + #[$doc] + /// + /// Don't try to remember the name of this type — use the +@@ -260,6 +269,16 @@ macro_rules! define_keywords { + } + } + ++ #[cfg(feature = "clone-impls")] ++ impl Copy for $name {} ++ ++ #[cfg(feature = "clone-impls")] ++ impl Clone for $name { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } ++ + #[cfg(feature = "extra-traits")] + impl Debug for $name { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +@@ -338,7 +357,6 @@ macro_rules! impl_deref_if_len_is_1 { + macro_rules! define_punctuation_structs { + ($($token:tt pub struct $name:ident/$len:tt #[$doc:meta])*) => { + $( +- #[cfg_attr(feature = "clone-impls", derive(Copy, Clone))] + #[repr(C)] + #[$doc] + /// +@@ -366,6 +384,16 @@ macro_rules! define_punctuation_structs { + } + } + ++ #[cfg(feature = "clone-impls")] ++ impl Copy for $name {} ++ ++ #[cfg(feature = "clone-impls")] ++ impl Clone for $name { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } ++ + #[cfg(feature = "extra-traits")] + impl Debug for $name { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +@@ -436,7 +464,6 @@ macro_rules! define_punctuation { + macro_rules! define_delimiters { + ($($token:tt pub struct $name:ident #[$doc:meta])*) => { + $( +- #[cfg_attr(feature = "clone-impls", derive(Copy, Clone))] + #[$doc] + pub struct $name { + pub span: Span, +@@ -458,6 +485,16 @@ macro_rules! define_delimiters { + } + } + ++ #[cfg(feature = "clone-impls")] ++ impl Copy for $name {} ++ ++ #[cfg(feature = "clone-impls")] ++ impl Clone for $name { ++ fn clone(&self) -> Self { ++ *self ++ } ++ } ++ + #[cfg(feature = "extra-traits")] + impl Debug for $name { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { +@@ -855,7 +892,7 @@ pub mod parsing { + } + + pub fn punct<S: FromSpans>(input: ParseStream, token: &str) -> Result<S> { +- let mut spans = [input.cursor().span(); 3]; ++ let mut spans = [input.span(); 3]; + punct_helper(input, token, &mut spans)?; + Ok(S::from_spans(&spans)) + } +diff --git a/third_party/rust/syn/src/tt.rs b/third_party/rust/syn/src/tt.rs +index f860eebb4f..8dba0627cd 100644 +--- a/third_party/rust/syn/src/tt.rs ++++ b/third_party/rust/syn/src/tt.rs +@@ -18,8 +18,8 @@ impl<'a> PartialEq for TokenTreeHelper<'a> { + _ => return false, + } + +- let s1 = g1.stream().clone().into_iter(); +- let mut s2 = g2.stream().clone().into_iter(); ++ let s1 = g1.stream().into_iter(); ++ let mut s2 = g2.stream().into_iter(); + + for item1 in s1 { + let item2 = match s2.next() { +@@ -60,7 +60,7 @@ impl<'a> Hash for TokenTreeHelper<'a> { + Delimiter::None => 3u8.hash(h), + } + +- for item in g.stream().clone() { ++ for item in g.stream() { + TokenTreeHelper(&item).hash(h); + } + 0xffu8.hash(h); // terminator w/ a variant we don't normally hash +diff --git a/third_party/rust/syn/src/ty.rs b/third_party/rust/syn/src/ty.rs +index 4ee59bda2a..fd7c97eab7 100644 +--- a/third_party/rust/syn/src/ty.rs ++++ b/third_party/rust/syn/src/ty.rs +@@ -1,15 +1,11 @@ + use super::*; + use crate::punctuated::Punctuated; +-#[cfg(feature = "extra-traits")] +-use crate::tt::TokenStreamHelper; + use proc_macro2::TokenStream; +-#[cfg(feature = "extra-traits")] +-use std::hash::{Hash, Hasher}; + + ast_enum_of_structs! { + /// The possible types that a Rust value could have. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + /// + /// # Syntax tree enum +@@ -20,7 +16,7 @@ ast_enum_of_structs! { + // + // TODO: change syntax-tree-enum link to an intra rustdoc link, currently + // blocked on https://github.com/rust-lang/rust/issues/62833 +- pub enum Type #manual_extra_traits { ++ pub enum Type { + /// A fixed size array type: `[T; n]`. + Array(TypeArray), + +@@ -77,7 +73,7 @@ ast_enum_of_structs! { + ast_struct! { + /// A fixed size array type: `[T; n]`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeArray { + pub bracket_token: token::Bracket, +@@ -90,7 +86,7 @@ ast_struct! { + ast_struct! { + /// A bare function type: `fn(usize) -> bool`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeBareFn { + pub lifetimes: Option<BoundLifetimes>, +@@ -107,7 +103,7 @@ ast_struct! { + ast_struct! { + /// A type contained within invisible delimiters. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeGroup { + pub group_token: token::Group, +@@ -119,7 +115,7 @@ ast_struct! { + /// An `impl Bound1 + Bound2 + Bound3` type where `Bound` is a trait or + /// a lifetime. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeImplTrait { + pub impl_token: Token![impl], +@@ -130,7 +126,7 @@ ast_struct! { + ast_struct! { + /// Indication that a type should be inferred by the compiler: `_`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeInfer { + pub underscore_token: Token![_], +@@ -140,7 +136,7 @@ ast_struct! { + ast_struct! { + /// A macro in the type position. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeMacro { + pub mac: Macro, +@@ -150,7 +146,7 @@ ast_struct! { + ast_struct! { + /// The never type: `!`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeNever { + pub bang_token: Token![!], +@@ -160,7 +156,7 @@ ast_struct! { + ast_struct! { + /// A parenthesized type equivalent to the inner type. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeParen { + pub paren_token: token::Paren, +@@ -172,7 +168,7 @@ ast_struct! { + /// A path like `std::slice::Iter`, optionally qualified with a + /// self-type as in `<Vec<T> as SomeTrait>::Associated`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypePath { + pub qself: Option<QSelf>, +@@ -183,7 +179,7 @@ ast_struct! { + ast_struct! { + /// A raw pointer type: `*const T` or `*mut T`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypePtr { + pub star_token: Token![*], +@@ -196,7 +192,7 @@ ast_struct! { + ast_struct! { + /// A reference type: `&'a T` or `&'a mut T`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeReference { + pub and_token: Token![&], +@@ -209,7 +205,7 @@ ast_struct! { + ast_struct! { + /// A dynamically sized slice type: `[T]`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeSlice { + pub bracket_token: token::Bracket, +@@ -221,7 +217,7 @@ ast_struct! { + /// A trait object type `Bound1 + Bound2 + Bound3` where `Bound` is a + /// trait or a lifetime. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeTraitObject { + pub dyn_token: Option<Token![dyn]>, +@@ -232,7 +228,7 @@ ast_struct! { + ast_struct! { + /// A tuple type: `(A, B, C, String)`. + /// +- /// *This type is available if Syn is built with the `"derive"` or ++ /// *This type is available only if Syn is built with the `"derive"` or + /// `"full"` feature.* + pub struct TypeTuple { + pub paren_token: token::Paren, +@@ -240,111 +236,10 @@ ast_struct! { + } + } + +-#[cfg(feature = "extra-traits")] +-impl Eq for Type {} +- +-#[cfg(feature = "extra-traits")] +-impl PartialEq for Type { +- fn eq(&self, other: &Self) -> bool { +- match (self, other) { +- (Type::Array(this), Type::Array(other)) => this == other, +- (Type::BareFn(this), Type::BareFn(other)) => this == other, +- (Type::Group(this), Type::Group(other)) => this == other, +- (Type::ImplTrait(this), Type::ImplTrait(other)) => this == other, +- (Type::Infer(this), Type::Infer(other)) => this == other, +- (Type::Macro(this), Type::Macro(other)) => this == other, +- (Type::Never(this), Type::Never(other)) => this == other, +- (Type::Paren(this), Type::Paren(other)) => this == other, +- (Type::Path(this), Type::Path(other)) => this == other, +- (Type::Ptr(this), Type::Ptr(other)) => this == other, +- (Type::Reference(this), Type::Reference(other)) => this == other, +- (Type::Slice(this), Type::Slice(other)) => this == other, +- (Type::TraitObject(this), Type::TraitObject(other)) => this == other, +- (Type::Tuple(this), Type::Tuple(other)) => this == other, +- (Type::Verbatim(this), Type::Verbatim(other)) => { +- TokenStreamHelper(this) == TokenStreamHelper(other) +- } +- _ => false, +- } +- } +-} +- +-#[cfg(feature = "extra-traits")] +-impl Hash for Type { +- fn hash<H>(&self, hash: &mut H) +- where +- H: Hasher, +- { +- match self { +- Type::Array(ty) => { +- hash.write_u8(0); +- ty.hash(hash); +- } +- Type::BareFn(ty) => { +- hash.write_u8(1); +- ty.hash(hash); +- } +- Type::Group(ty) => { +- hash.write_u8(2); +- ty.hash(hash); +- } +- Type::ImplTrait(ty) => { +- hash.write_u8(3); +- ty.hash(hash); +- } +- Type::Infer(ty) => { +- hash.write_u8(4); +- ty.hash(hash); +- } +- Type::Macro(ty) => { +- hash.write_u8(5); +- ty.hash(hash); +- } +- Type::Never(ty) => { +- hash.write_u8(6); +- ty.hash(hash); +- } +- Type::Paren(ty) => { +- hash.write_u8(7); +- ty.hash(hash); +- } +- Type::Path(ty) => { +- hash.write_u8(8); +- ty.hash(hash); +- } +- Type::Ptr(ty) => { +- hash.write_u8(9); +- ty.hash(hash); +- } +- Type::Reference(ty) => { +- hash.write_u8(10); +- ty.hash(hash); +- } +- Type::Slice(ty) => { +- hash.write_u8(11); +- ty.hash(hash); +- } +- Type::TraitObject(ty) => { +- hash.write_u8(12); +- ty.hash(hash); +- } +- Type::Tuple(ty) => { +- hash.write_u8(13); +- ty.hash(hash); +- } +- Type::Verbatim(ty) => { +- hash.write_u8(14); +- TokenStreamHelper(ty).hash(hash); +- } +- Type::__Nonexhaustive => unreachable!(), +- } +- } +-} +- + ast_struct! { + /// The binary interface of a function: `extern "C"`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Abi { + pub extern_token: Token![extern], +@@ -355,7 +250,7 @@ ast_struct! { + ast_struct! { + /// An argument in a function type: the `usize` in `fn(usize) -> bool`. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct BareFnArg { + pub attrs: Vec<Attribute>, +@@ -377,7 +272,7 @@ ast_struct! { + /// } + /// ``` + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub struct Variadic { + pub attrs: Vec<Attribute>, +@@ -388,7 +283,7 @@ ast_struct! { + ast_enum! { + /// Return type of a function signature. + /// +- /// *This type is available if Syn is built with the `"derive"` or `"full"` ++ /// *This type is available only if Syn is built with the `"derive"` or `"full"` + /// feature.* + pub enum ReturnType { + /// Return type is not specified. +@@ -407,10 +302,13 @@ pub mod parsing { + use crate::ext::IdentExt; + use crate::parse::{Parse, ParseStream, Result}; + use crate::path; ++ use proc_macro2::{Punct, Spacing, TokenTree}; ++ use std::iter::FromIterator; + + impl Parse for Type { + fn parse(input: ParseStream) -> Result<Self> { +- ambig_ty(input, true) ++ let allow_plus = true; ++ ambig_ty(input, allow_plus) + } + } + +@@ -421,15 +319,17 @@ pub mod parsing { + /// + /// This parser does not allow a `+`, while the default parser does. + pub fn without_plus(input: ParseStream) -> Result<Self> { +- ambig_ty(input, false) ++ let allow_plus = false; ++ ambig_ty(input, allow_plus) + } + } + + fn ambig_ty(input: ParseStream, allow_plus: bool) -> Result<Type> { +- if input.peek(token::Group) { ++ if input.peek(token::Group) && !input.peek2(Token![::]) && !input.peek2(Token![<]) { + return input.parse().map(Type::Group); + } + ++ let begin = input.fork(); + let mut lifetimes = None::<BoundLifetimes>; + let mut lookahead = input.lookahead1(); + if lookahead.peek(Token![for]) { +@@ -524,7 +424,7 @@ pub mod parsing { + ..trait_bound + }) + } +- other => other, ++ other @ TypeParamBound::Lifetime(_) => other, + } + } + _ => break, +@@ -549,17 +449,20 @@ pub mod parsing { + })) + } else if lookahead.peek(Token![fn]) + || lookahead.peek(Token![unsafe]) +- || lookahead.peek(Token![extern]) && !input.peek2(Token![::]) ++ || lookahead.peek(Token![extern]) + { +- let mut bare_fn: TypeBareFn = input.parse()?; +- bare_fn.lifetimes = lifetimes; +- Ok(Type::BareFn(bare_fn)) ++ let allow_mut_self = true; ++ if let Some(mut bare_fn) = parse_bare_fn(input, allow_mut_self)? { ++ bare_fn.lifetimes = lifetimes; ++ Ok(Type::BareFn(bare_fn)) ++ } else { ++ Ok(Type::Verbatim(verbatim::between(begin, input))) ++ } + } else if lookahead.peek(Ident) + || input.peek(Token![super]) + || input.peek(Token![self]) + || input.peek(Token![Self]) + || input.peek(Token![crate]) +- || input.peek(Token![extern]) + || lookahead.peek(Token![::]) + || lookahead.peek(Token![<]) + { +@@ -722,38 +625,58 @@ pub mod parsing { + + impl Parse for TypeBareFn { + fn parse(input: ParseStream) -> Result<Self> { +- let args; +- let allow_variadic; +- Ok(TypeBareFn { +- lifetimes: input.parse()?, +- unsafety: input.parse()?, +- abi: input.parse()?, +- fn_token: input.parse()?, +- paren_token: parenthesized!(args in input), +- inputs: { +- let mut inputs = Punctuated::new(); +- while !args.is_empty() && !args.peek(Token![...]) { +- inputs.push_value(args.parse()?); +- if args.is_empty() { +- break; +- } +- inputs.push_punct(args.parse()?); +- } +- allow_variadic = inputs.empty_or_trailing(); +- inputs +- }, +- variadic: { +- if allow_variadic && args.peek(Token![...]) { +- Some(Variadic { +- attrs: Vec::new(), ++ let allow_mut_self = false; ++ parse_bare_fn(input, allow_mut_self).map(Option::unwrap) ++ } ++ } ++ ++ fn parse_bare_fn(input: ParseStream, allow_mut_self: bool) -> Result<Option<TypeBareFn>> { ++ let args; ++ let mut variadic = None; ++ let mut has_mut_self = false; ++ ++ let bare_fn = TypeBareFn { ++ lifetimes: input.parse()?, ++ unsafety: input.parse()?, ++ abi: input.parse()?, ++ fn_token: input.parse()?, ++ paren_token: parenthesized!(args in input), ++ inputs: { ++ let mut inputs = Punctuated::new(); ++ ++ while !args.is_empty() { ++ let attrs = args.call(Attribute::parse_outer)?; ++ ++ if inputs.empty_or_trailing() && args.peek(Token![...]) { ++ variadic = Some(Variadic { ++ attrs, + dots: args.parse()?, +- }) ++ }); ++ break; ++ } ++ ++ if let Some(arg) = parse_bare_fn_arg(&args, allow_mut_self)? { ++ inputs.push_value(BareFnArg { attrs, ..arg }); + } else { +- None ++ has_mut_self = true; + } +- }, +- output: input.call(ReturnType::without_plus)?, +- }) ++ if args.is_empty() { ++ break; ++ } ++ ++ inputs.push_punct(args.parse()?); ++ } ++ ++ inputs ++ }, ++ variadic, ++ output: input.call(ReturnType::without_plus)?, ++ }; ++ ++ if has_mut_self { ++ Ok(None) ++ } else { ++ Ok(Some(bare_fn)) + } + } + +@@ -776,9 +699,27 @@ pub mod parsing { + impl Parse for TypeTuple { + fn parse(input: ParseStream) -> Result<Self> { + let content; ++ let paren_token = parenthesized!(content in input); ++ ++ if content.is_empty() { ++ return Ok(TypeTuple { ++ paren_token, ++ elems: Punctuated::new(), ++ }); ++ } ++ ++ let first: Type = content.parse()?; + Ok(TypeTuple { +- paren_token: parenthesized!(content in input), +- elems: content.parse_terminated(Type::parse)?, ++ paren_token, ++ elems: { ++ let mut elems = Punctuated::new(); ++ elems.push_value(first); ++ elems.push_punct(content.parse()?); ++ let rest: Punctuated<Type, Token![,]> = ++ content.parse_terminated(Parse::parse)?; ++ elems.extend(rest); ++ elems ++ }, + }) + } + } +@@ -807,9 +748,11 @@ pub mod parsing { + + impl ReturnType { + pub fn without_plus(input: ParseStream) -> Result<Self> { +- Self::parse(input, false) ++ let allow_plus = false; ++ Self::parse(input, allow_plus) + } + ++ #[doc(hidden)] + pub fn parse(input: ParseStream, allow_plus: bool) -> Result<Self> { + if input.peek(Token![->]) { + let arrow = input.parse()?; +@@ -844,10 +787,12 @@ pub mod parsing { + + impl TypeTraitObject { + pub fn without_plus(input: ParseStream) -> Result<Self> { +- Self::parse(input, false) ++ let allow_plus = false; ++ Self::parse(input, allow_plus) + } + + // Only allow multiple trait references if allow_plus is true. ++ #[doc(hidden)] + pub fn parse(input: ParseStream, allow_plus: bool) -> Result<Self> { + Ok(TypeTraitObject { + dyn_token: input.parse()?, +@@ -910,7 +855,8 @@ pub mod parsing { + + impl Parse for TypeParen { + fn parse(input: ParseStream) -> Result<Self> { +- Self::parse(input, false) ++ let allow_plus = false; ++ Self::parse(input, allow_plus) + } + } + +@@ -926,22 +872,72 @@ pub mod parsing { + + impl Parse for BareFnArg { + fn parse(input: ParseStream) -> Result<Self> { +- Ok(BareFnArg { +- attrs: input.call(Attribute::parse_outer)?, +- name: { +- if (input.peek(Ident) || input.peek(Token![_])) +- && input.peek2(Token![:]) +- && !input.peek2(Token![::]) +- { +- let name = input.call(Ident::parse_any)?; +- let colon: Token![:] = input.parse()?; +- Some((name, colon)) +- } else { +- None +- } +- }, +- ty: input.parse()?, +- }) ++ let allow_mut_self = false; ++ parse_bare_fn_arg(input, allow_mut_self).map(Option::unwrap) ++ } ++ } ++ ++ fn parse_bare_fn_arg( ++ input: ParseStream, ++ mut allow_mut_self: bool, ++ ) -> Result<Option<BareFnArg>> { ++ let mut has_mut_self = false; ++ let arg = BareFnArg { ++ attrs: input.call(Attribute::parse_outer)?, ++ name: { ++ if (input.peek(Ident) || input.peek(Token![_]) || input.peek(Token![self])) ++ && input.peek2(Token![:]) ++ && !input.peek2(Token![::]) ++ { ++ let name = input.call(Ident::parse_any)?; ++ let colon: Token![:] = input.parse()?; ++ Some((name, colon)) ++ } else if allow_mut_self ++ && input.peek(Token![mut]) ++ && input.peek2(Token![self]) ++ && input.peek3(Token![:]) ++ && !input.peek3(Token![::]) ++ { ++ has_mut_self = true; ++ allow_mut_self = false; ++ input.parse::<Token![mut]>()?; ++ input.parse::<Token![self]>()?; ++ input.parse::<Token![:]>()?; ++ None ++ } else { ++ None ++ } ++ }, ++ ty: if !has_mut_self && input.peek(Token![...]) { ++ let dot3 = input.parse::<Token![...]>()?; ++ let args = vec![ ++ TokenTree::Punct(Punct::new('.', Spacing::Joint)), ++ TokenTree::Punct(Punct::new('.', Spacing::Joint)), ++ TokenTree::Punct(Punct::new('.', Spacing::Alone)), ++ ]; ++ let tokens = TokenStream::from_iter(args.into_iter().zip(&dot3.spans).map( ++ |(mut arg, span)| { ++ arg.set_span(*span); ++ arg ++ }, ++ )); ++ Type::Verbatim(tokens) ++ } else if allow_mut_self && input.peek(Token![mut]) && input.peek2(Token![self]) { ++ has_mut_self = true; ++ input.parse::<Token![mut]>()?; ++ Type::Path(TypePath { ++ qself: None, ++ path: input.parse::<Token![self]>()?.into(), ++ }) ++ } else { ++ input.parse()? ++ }, ++ }; ++ ++ if has_mut_self { ++ Ok(None) ++ } else { ++ Ok(Some(arg)) + } + } + +diff --git a/third_party/rust/syn/src/verbatim.rs b/third_party/rust/syn/src/verbatim.rs +new file mode 100644 +index 0000000000..0686352f7a +--- /dev/null ++++ b/third_party/rust/syn/src/verbatim.rs +@@ -0,0 +1,15 @@ ++use crate::parse::{ParseBuffer, ParseStream}; ++use proc_macro2::TokenStream; ++use std::iter; ++ ++pub fn between<'a>(begin: ParseBuffer<'a>, end: ParseStream<'a>) -> TokenStream { ++ let end = end.cursor(); ++ let mut cursor = begin.cursor(); ++ let mut tokens = TokenStream::new(); ++ while cursor != end { ++ let (tt, next) = cursor.token_tree().unwrap(); ++ tokens.extend(iter::once(tt)); ++ cursor = next; ++ } ++ tokens ++} +diff --git a/third_party/rust/syn/src/whitespace.rs b/third_party/rust/syn/src/whitespace.rs +new file mode 100644 +index 0000000000..7be082e1a2 +--- /dev/null ++++ b/third_party/rust/syn/src/whitespace.rs +@@ -0,0 +1,65 @@ ++pub fn skip(mut s: &str) -> &str { ++ 'skip: while !s.is_empty() { ++ let byte = s.as_bytes()[0]; ++ if byte == b'/' { ++ if s.starts_with("//") ++ && (!s.starts_with("///") || s.starts_with("////")) ++ && !s.starts_with("//!") ++ { ++ if let Some(i) = s.find('\n') { ++ s = &s[i + 1..]; ++ continue; ++ } else { ++ return ""; ++ } ++ } else if s.starts_with("/**/") { ++ s = &s[4..]; ++ continue; ++ } else if s.starts_with("/*") ++ && (!s.starts_with("/**") || s.starts_with("/***")) ++ && !s.starts_with("/*!") ++ { ++ let mut depth = 0; ++ let bytes = s.as_bytes(); ++ let mut i = 0; ++ let upper = bytes.len() - 1; ++ while i < upper { ++ if bytes[i] == b'/' && bytes[i + 1] == b'*' { ++ depth += 1; ++ i += 1; // eat '*' ++ } else if bytes[i] == b'*' && bytes[i + 1] == b'/' { ++ depth -= 1; ++ if depth == 0 { ++ s = &s[i + 2..]; ++ continue 'skip; ++ } ++ i += 1; // eat '/' ++ } ++ i += 1; ++ } ++ return s; ++ } ++ } ++ match byte { ++ b' ' | 0x09..=0x0d => { ++ s = &s[1..]; ++ continue; ++ } ++ b if b <= 0x7f => {} ++ _ => { ++ let ch = s.chars().next().unwrap(); ++ if is_whitespace(ch) { ++ s = &s[ch.len_utf8()..]; ++ continue; ++ } ++ } ++ } ++ return s; ++ } ++ s ++} ++ ++fn is_whitespace(ch: char) -> bool { ++ // Rust treats left-to-right mark and right-to-left mark as whitespace ++ ch.is_whitespace() || ch == '\u{200e}' || ch == '\u{200f}' ++} +diff --git a/third_party/rust/syn/tests/clone.sh b/third_party/rust/syn/tests/clone.sh +deleted file mode 100755 +index 8e0863cba6..0000000000 +--- a/third_party/rust/syn/tests/clone.sh ++++ /dev/null +@@ -1,16 +0,0 @@ +-#!/bin/bash +- +-REV=4560cb830fce63fcffdc4558f4281aaac6a3a1ba +- +-set -euo pipefail +-cd "$(dirname "${BASH_SOURCE[0]}")" +-mkdir -p rust +-touch rust/COMMIT +- +-if [ "$(cat rust/COMMIT)" != "$REV" ]; then +- rm -rf rust +- mkdir rust +- curl -L "https://github.com/rust-lang/rust/archive/${REV}.tar.gz" \ +- | tar xz --directory rust --strip-components 1 +- echo "$REV" > rust/COMMIT +-fi +diff --git a/third_party/rust/syn/tests/common/eq.rs b/third_party/rust/syn/tests/common/eq.rs +index 13a6c36ae5..7589a07573 100644 +--- a/third_party/rust/syn/tests/common/eq.rs ++++ b/third_party/rust/syn/tests/common/eq.rs +@@ -1,36 +1,35 @@ ++extern crate rustc_ast; + extern crate rustc_data_structures; ++extern crate rustc_span; + extern crate rustc_target; +-extern crate syntax; +-extern crate syntax_pos; + + use std::mem; + +-use self::rustc_data_structures::sync::Lrc; +-use self::rustc_data_structures::thin_vec::ThinVec; +-use self::rustc_target::abi::FloatTy; +-use self::rustc_target::spec::abi::Abi; +-use self::syntax::ast::{ +- AngleBracketedArgs, AnonConst, Arg, Arm, AsmDialect, AssocTyConstraint, AssocTyConstraintKind, +- AttrId, AttrStyle, Attribute, BareFnTy, BinOpKind, BindingMode, Block, BlockCheckMode, +- CaptureBy, Constness, Crate, CrateSugar, Defaultness, EnumDef, Expr, ExprKind, Field, FieldPat, +- FnDecl, FnHeader, ForeignItem, ForeignItemKind, ForeignMod, FunctionRetTy, GenericArg, +- GenericArgs, GenericBound, GenericParam, GenericParamKind, Generics, GlobalAsm, Ident, +- ImplItem, ImplItemKind, ImplPolarity, InlineAsm, InlineAsmOutput, IntTy, IsAsync, IsAuto, Item, +- ItemKind, Label, Lifetime, Lit, LitIntType, LitKind, Local, Mac, MacDelimiter, MacStmtStyle, +- MacroDef, MethodSig, Mod, Movability, MutTy, Mutability, NodeId, ParenthesizedArgs, Pat, +- PatKind, Path, PathSegment, PolyTraitRef, QSelf, RangeEnd, RangeLimits, RangeSyntax, Stmt, +- StmtKind, StrStyle, StructField, TraitBoundModifier, TraitItem, TraitItemKind, +- TraitObjectSyntax, TraitRef, Ty, TyKind, UintTy, UnOp, UnsafeSource, Unsafety, UseTree, +- UseTreeKind, Variant, VariantData, VisibilityKind, WhereBoundPredicate, WhereClause, +- WhereEqPredicate, WherePredicate, WhereRegionPredicate, ++use rustc_ast::ast::{ ++ AngleBracketedArg, AngleBracketedArgs, AnonConst, Arm, AssocItemKind, AssocTyConstraint, ++ AssocTyConstraintKind, Async, AttrId, AttrItem, AttrKind, AttrStyle, Attribute, BareFnTy, ++ BinOpKind, BindingMode, Block, BlockCheckMode, BorrowKind, CaptureBy, Const, Crate, CrateSugar, ++ Defaultness, EnumDef, Expr, ExprKind, Extern, Field, FieldPat, FloatTy, FnDecl, FnHeader, ++ FnRetTy, FnSig, ForeignItemKind, ForeignMod, GenericArg, GenericArgs, GenericBound, ++ GenericParam, GenericParamKind, Generics, GlobalAsm, ImplPolarity, InlineAsm, InlineAsmOperand, ++ InlineAsmOptions, InlineAsmRegOrRegClass, InlineAsmTemplatePiece, IntTy, IsAuto, Item, ++ ItemKind, Label, Lifetime, Lit, LitFloatType, LitIntType, LitKind, LlvmAsmDialect, ++ LlvmInlineAsm, LlvmInlineAsmOutput, Local, MacArgs, MacCall, MacCallStmt, MacDelimiter, ++ MacStmtStyle, MacroDef, Mod, Movability, MutTy, Mutability, NodeId, Param, ParenthesizedArgs, ++ Pat, PatKind, Path, PathSegment, PolyTraitRef, QSelf, RangeEnd, RangeLimits, RangeSyntax, Stmt, ++ StmtKind, StrLit, StrStyle, StructField, TraitBoundModifier, TraitObjectSyntax, TraitRef, Ty, ++ TyKind, UintTy, UnOp, Unsafe, UnsafeSource, UseTree, UseTreeKind, Variant, VariantData, ++ VisibilityKind, WhereBoundPredicate, WhereClause, WhereEqPredicate, WherePredicate, ++ WhereRegionPredicate, + }; +-use self::syntax::parse::lexer::comments; +-use self::syntax::parse::token::{self, DelimToken, Token, TokenKind}; +-use self::syntax::ptr::P; +-use self::syntax::source_map::Spanned; +-use self::syntax::symbol::{sym, Symbol}; +-use self::syntax::tokenstream::{DelimSpan, TokenStream, TokenTree}; +-use self::syntax_pos::{Span, SyntaxContext, DUMMY_SP}; ++use rustc_ast::ptr::P; ++use rustc_ast::token::{self, CommentKind, DelimToken, Token, TokenKind}; ++use rustc_ast::tokenstream::{DelimSpan, TokenStream, TokenTree}; ++use rustc_data_structures::sync::Lrc; ++use rustc_data_structures::thin_vec::ThinVec; ++use rustc_span::source_map::Spanned; ++use rustc_span::symbol::Ident; ++use rustc_span::{Span, Symbol, SyntaxContext}; + + pub trait SpanlessEq { + fn eq(&self, other: &Self) -> bool; +@@ -86,14 +85,6 @@ impl<A: SpanlessEq, B: SpanlessEq> SpanlessEq for (A, B) { + } + } + +-impl<A: SpanlessEq, B: SpanlessEq, C: SpanlessEq> SpanlessEq for (A, B, C) { +- fn eq(&self, other: &Self) -> bool { +- SpanlessEq::eq(&self.0, &other.0) +- && SpanlessEq::eq(&self.1, &other.1) +- && SpanlessEq::eq(&self.2, &other.2) +- } +-} +- + macro_rules! spanless_eq_true { + ($name:ident) => { + impl SpanlessEq for $name { +@@ -126,17 +117,19 @@ spanless_eq_partial_eq!(u16); + spanless_eq_partial_eq!(u128); + spanless_eq_partial_eq!(usize); + spanless_eq_partial_eq!(char); ++spanless_eq_partial_eq!(String); + spanless_eq_partial_eq!(Symbol); +-spanless_eq_partial_eq!(Abi); ++spanless_eq_partial_eq!(CommentKind); + spanless_eq_partial_eq!(DelimToken); ++spanless_eq_partial_eq!(InlineAsmOptions); + + macro_rules! spanless_eq_struct { + { +- $name:ident; ++ $name:ident $(<$param:ident>)?; + $([$field:ident $other:ident])* + $(![$ignore:ident])* + } => { +- impl SpanlessEq for $name { ++ impl $(<$param: SpanlessEq>)* SpanlessEq for $name $(<$param>)* { + fn eq(&self, other: &Self) -> bool { + let $name { $($field,)* $($ignore: _,)* } = self; + let $name { $($field: $other,)* $($ignore: _,)* } = other; +@@ -146,14 +139,14 @@ macro_rules! spanless_eq_struct { + }; + + { +- $name:ident; ++ $name:ident $(<$param:ident>)?; + $([$field:ident $other:ident])* + $next:ident + $($rest:ident)* + $(!$ignore:ident)* + } => { + spanless_eq_struct! { +- $name; ++ $name $(<$param>)*; + $([$field $other])* + [$next other] + $($rest)* +@@ -162,14 +155,14 @@ macro_rules! spanless_eq_struct { + }; + + { +- $name:ident; ++ $name:ident $(<$param:ident>)?; + $([$field:ident $other:ident])* + $(![$ignore:ident])* + !$next:ident + $(!$rest:ident)* + } => { + spanless_eq_struct! { +- $name; ++ $name $(<$param>)*; + $([$field $other])* + $(![$ignore])* + ![$next] +@@ -263,119 +256,131 @@ macro_rules! spanless_eq_enum { + }; + } + +-spanless_eq_struct!(AngleBracketedArgs; span args constraints); ++spanless_eq_struct!(AngleBracketedArgs; span args); + spanless_eq_struct!(AnonConst; id value); +-spanless_eq_struct!(Arg; attrs ty pat id span); +-spanless_eq_struct!(Arm; attrs pats guard body span id); ++spanless_eq_struct!(Arm; attrs pat guard body span id is_placeholder); + spanless_eq_struct!(AssocTyConstraint; id ident kind span); +-spanless_eq_struct!(Attribute; id style path tokens span !is_sugared_doc); +-spanless_eq_struct!(BareFnTy; unsafety abi generic_params decl); ++spanless_eq_struct!(AttrItem; path args); ++spanless_eq_struct!(Attribute; kind id style span); ++spanless_eq_struct!(BareFnTy; unsafety ext generic_params decl); + spanless_eq_struct!(Block; stmts id rules span); +-spanless_eq_struct!(Crate; module attrs span); ++spanless_eq_struct!(Crate; module attrs span proc_macros); + spanless_eq_struct!(EnumDef; variants); +-spanless_eq_struct!(Expr; id node span attrs); +-spanless_eq_struct!(Field; ident expr span is_shorthand attrs id); +-spanless_eq_struct!(FieldPat; ident pat is_shorthand attrs id span); +-spanless_eq_struct!(FnDecl; inputs output c_variadic); +-spanless_eq_struct!(FnHeader; constness asyncness unsafety abi); +-spanless_eq_struct!(ForeignItem; ident attrs node id span vis); ++spanless_eq_struct!(Expr; id kind span attrs !tokens); ++spanless_eq_struct!(Field; attrs id span ident expr is_shorthand is_placeholder); ++spanless_eq_struct!(FieldPat; ident pat is_shorthand attrs id span is_placeholder); ++spanless_eq_struct!(FnDecl; inputs output); ++spanless_eq_struct!(FnHeader; constness asyncness unsafety ext); ++spanless_eq_struct!(FnSig; header decl span); + spanless_eq_struct!(ForeignMod; abi items); +-spanless_eq_struct!(GenericParam; id ident attrs bounds kind); ++spanless_eq_struct!(GenericParam; id ident attrs bounds is_placeholder kind); + spanless_eq_struct!(Generics; params where_clause span); + spanless_eq_struct!(GlobalAsm; asm); +-spanless_eq_struct!(ImplItem; id ident vis defaultness attrs generics node span !tokens); +-spanless_eq_struct!(InlineAsm; asm asm_str_style outputs inputs clobbers volatile alignstack dialect); +-spanless_eq_struct!(InlineAsmOutput; constraint expr is_rw is_indirect); +-spanless_eq_struct!(Item; ident attrs id node vis span !tokens); ++spanless_eq_struct!(InlineAsm; template operands options line_spans); ++spanless_eq_struct!(Item<K>; attrs id span vis ident kind !tokens); + spanless_eq_struct!(Label; ident); + spanless_eq_struct!(Lifetime; id ident); +-spanless_eq_struct!(Lit; token node span); ++spanless_eq_struct!(Lit; token kind span); ++spanless_eq_struct!(LlvmInlineAsm; asm asm_str_style outputs inputs clobbers volatile alignstack dialect); ++spanless_eq_struct!(LlvmInlineAsmOutput; constraint expr is_rw is_indirect); + spanless_eq_struct!(Local; pat ty init id span attrs); +-spanless_eq_struct!(Mac; path delim tts span prior_type_ascription); +-spanless_eq_struct!(MacroDef; tokens legacy); +-spanless_eq_struct!(MethodSig; header decl); ++spanless_eq_struct!(MacCall; path args prior_type_ascription); ++spanless_eq_struct!(MacCallStmt; mac style attrs); ++spanless_eq_struct!(MacroDef; body macro_rules); + spanless_eq_struct!(Mod; inner items inline); + spanless_eq_struct!(MutTy; ty mutbl); ++spanless_eq_struct!(Param; attrs ty pat id span is_placeholder); + spanless_eq_struct!(ParenthesizedArgs; span inputs output); +-spanless_eq_struct!(Pat; id node span); ++spanless_eq_struct!(Pat; id kind span tokens); + spanless_eq_struct!(Path; span segments); + spanless_eq_struct!(PathSegment; ident id args); + spanless_eq_struct!(PolyTraitRef; bound_generic_params trait_ref span); + spanless_eq_struct!(QSelf; ty path_span position); +-spanless_eq_struct!(Stmt; id node span); +-spanless_eq_struct!(StructField; span ident vis id ty attrs); ++spanless_eq_struct!(Stmt; id kind span); ++spanless_eq_struct!(StrLit; style symbol suffix span symbol_unescaped); ++spanless_eq_struct!(StructField; attrs id span vis ident ty is_placeholder); + spanless_eq_struct!(Token; kind span); +-spanless_eq_struct!(TraitItem; id ident attrs generics node span !tokens); + spanless_eq_struct!(TraitRef; path ref_id); +-spanless_eq_struct!(Ty; id node span); ++spanless_eq_struct!(Ty; id kind span); + spanless_eq_struct!(UseTree; prefix kind span); +-spanless_eq_struct!(Variant; ident attrs id data disr_expr span); ++spanless_eq_struct!(Variant; attrs id span vis ident data disr_expr is_placeholder); + spanless_eq_struct!(WhereBoundPredicate; span bound_generic_params bounded_ty bounds); +-spanless_eq_struct!(WhereClause; predicates span); ++spanless_eq_struct!(WhereClause; has_where_token predicates span); + spanless_eq_struct!(WhereEqPredicate; id span lhs_ty rhs_ty); + spanless_eq_struct!(WhereRegionPredicate; span lifetime bounds); +-spanless_eq_enum!(AsmDialect; Att Intel); ++spanless_eq_enum!(AngleBracketedArg; Arg(0) Constraint(0)); ++spanless_eq_enum!(AssocItemKind; Const(0 1 2) Fn(0 1 2 3) TyAlias(0 1 2 3) MacCall(0)); + spanless_eq_enum!(AssocTyConstraintKind; Equality(ty) Bound(bounds)); ++spanless_eq_enum!(Async; Yes(span closure_id return_impl_trait_id) No); ++spanless_eq_enum!(AttrKind; Normal(0) DocComment(0 1)); + spanless_eq_enum!(AttrStyle; Outer Inner); + spanless_eq_enum!(BinOpKind; Add Sub Mul Div Rem And Or BitXor BitAnd BitOr Shl Shr Eq Lt Le Ne Ge Gt); + spanless_eq_enum!(BindingMode; ByRef(0) ByValue(0)); + spanless_eq_enum!(BlockCheckMode; Default Unsafe(0)); ++spanless_eq_enum!(BorrowKind; Ref Raw); + spanless_eq_enum!(CaptureBy; Value Ref); +-spanless_eq_enum!(Constness; Const NotConst); ++spanless_eq_enum!(Const; Yes(0) No); + spanless_eq_enum!(CrateSugar; PubCrate JustCrate); +-spanless_eq_enum!(Defaultness; Default Final); ++spanless_eq_enum!(Defaultness; Default(0) Final); ++spanless_eq_enum!(Extern; None Implicit Explicit(0)); + spanless_eq_enum!(FloatTy; F32 F64); +-spanless_eq_enum!(ForeignItemKind; Fn(0 1) Static(0 1) Ty Macro(0)); +-spanless_eq_enum!(FunctionRetTy; Default(0) Ty(0)); ++spanless_eq_enum!(FnRetTy; Default(0) Ty(0)); ++spanless_eq_enum!(ForeignItemKind; Static(0 1 2) Fn(0 1 2 3) TyAlias(0 1 2 3) MacCall(0)); + spanless_eq_enum!(GenericArg; Lifetime(0) Type(0) Const(0)); + spanless_eq_enum!(GenericArgs; AngleBracketed(0) Parenthesized(0)); + spanless_eq_enum!(GenericBound; Trait(0 1) Outlives(0)); +-spanless_eq_enum!(GenericParamKind; Lifetime Type(default) Const(ty)); +-spanless_eq_enum!(ImplItemKind; Const(0 1) Method(0 1) TyAlias(0) OpaqueTy(0) Macro(0)); +-spanless_eq_enum!(ImplPolarity; Positive Negative); ++spanless_eq_enum!(GenericParamKind; Lifetime Type(default) Const(ty kw_span)); ++spanless_eq_enum!(ImplPolarity; Positive Negative(0)); ++spanless_eq_enum!(InlineAsmRegOrRegClass; Reg(0) RegClass(0)); ++spanless_eq_enum!(InlineAsmTemplatePiece; String(0) Placeholder(operand_idx modifier span)); + spanless_eq_enum!(IntTy; Isize I8 I16 I32 I64 I128); +-spanless_eq_enum!(IsAsync; Async(closure_id return_impl_trait_id) NotAsync); + spanless_eq_enum!(IsAuto; Yes No); ++spanless_eq_enum!(LitFloatType; Suffixed(0) Unsuffixed); + spanless_eq_enum!(LitIntType; Signed(0) Unsigned(0) Unsuffixed); ++spanless_eq_enum!(LlvmAsmDialect; Att Intel); ++spanless_eq_enum!(MacArgs; Empty Delimited(0 1 2) Eq(0 1)); + spanless_eq_enum!(MacDelimiter; Parenthesis Bracket Brace); + spanless_eq_enum!(MacStmtStyle; Semicolon Braces NoBraces); + spanless_eq_enum!(Movability; Static Movable); +-spanless_eq_enum!(Mutability; Mutable Immutable); ++spanless_eq_enum!(Mutability; Mut Not); + spanless_eq_enum!(RangeEnd; Included(0) Excluded); + spanless_eq_enum!(RangeLimits; HalfOpen Closed); +-spanless_eq_enum!(StmtKind; Local(0) Item(0) Expr(0) Semi(0) Mac(0)); ++spanless_eq_enum!(StmtKind; Local(0) Item(0) Expr(0) Semi(0) Empty MacCall(0)); + spanless_eq_enum!(StrStyle; Cooked Raw(0)); + spanless_eq_enum!(TokenTree; Token(0) Delimited(0 1 2)); +-spanless_eq_enum!(TraitBoundModifier; None Maybe); +-spanless_eq_enum!(TraitItemKind; Const(0 1) Method(0 1) Type(0 1) Macro(0)); ++spanless_eq_enum!(TraitBoundModifier; None Maybe MaybeConst MaybeConstMaybe); + spanless_eq_enum!(TraitObjectSyntax; Dyn None); + spanless_eq_enum!(UintTy; Usize U8 U16 U32 U64 U128); + spanless_eq_enum!(UnOp; Deref Not Neg); ++spanless_eq_enum!(Unsafe; Yes(0) No); + spanless_eq_enum!(UnsafeSource; CompilerGenerated UserProvided); +-spanless_eq_enum!(Unsafety; Unsafe Normal); + spanless_eq_enum!(UseTreeKind; Simple(0 1 2) Nested(0) Glob); + spanless_eq_enum!(VariantData; Struct(0 1) Tuple(0 1) Unit(0)); + spanless_eq_enum!(VisibilityKind; Public Crate(0) Restricted(path id) Inherited); + spanless_eq_enum!(WherePredicate; BoundPredicate(0) RegionPredicate(0) EqPredicate(0)); +-spanless_eq_enum!(ExprKind; Box(0) Array(0) Call(0 1) MethodCall(0 1) Tup(0) ++spanless_eq_enum!(ExprKind; Box(0) Array(0) Call(0 1) MethodCall(0 1 2) Tup(0) + Binary(0 1 2) Unary(0 1) Lit(0) Cast(0 1) Type(0 1) Let(0 1) If(0 1 2) + While(0 1 2) ForLoop(0 1 2 3) Loop(0 1) Match(0 1) Closure(0 1 2 3 4 5) +- Block(0 1) Async(0 1 2) Await(0) TryBlock(0) Assign(0 1) AssignOp(0 1 2) +- Field(0 1) Index(0 1) Range(0 1 2) Path(0 1) AddrOf(0 1) Break(0 1) +- Continue(0) Ret(0) InlineAsm(0) Mac(0) Struct(0 1 2) Repeat(0 1) Paren(0) +- Try(0) Yield(0) Err); +-spanless_eq_enum!(ItemKind; ExternCrate(0) Use(0) Static(0 1 2) Const(0 1) +- Fn(0 1 2 3) Mod(0) ForeignMod(0) GlobalAsm(0) TyAlias(0 1) OpaqueTy(0 1) +- Enum(0 1) Struct(0 1) Union(0 1) Trait(0 1 2 3 4) TraitAlias(0 1) +- Impl(0 1 2 3 4 5 6) Mac(0) MacroDef(0)); ++ Block(0 1) Async(0 1 2) Await(0) TryBlock(0) Assign(0 1 2) AssignOp(0 1 2) ++ Field(0 1) Index(0 1) Range(0 1 2) Path(0 1) AddrOf(0 1 2) Break(0 1) ++ Continue(0) Ret(0) InlineAsm(0) LlvmInlineAsm(0) MacCall(0) Struct(0 1 2) ++ Repeat(0 1) Paren(0) Try(0) Yield(0) Err); ++spanless_eq_enum!(InlineAsmOperand; In(reg expr) Out(reg late expr) ++ InOut(reg late expr) SplitInOut(reg late in_expr out_expr) Const(expr) ++ Sym(expr)); ++spanless_eq_enum!(ItemKind; ExternCrate(0) Use(0) Static(0 1 2) Const(0 1 2) ++ Fn(0 1 2 3) Mod(0) ForeignMod(0) GlobalAsm(0) TyAlias(0 1 2 3) Enum(0 1) ++ Struct(0 1) Union(0 1) Trait(0 1 2 3 4) TraitAlias(0 1) ++ Impl(unsafety polarity defaultness constness generics of_trait self_ty items) ++ MacCall(0) MacroDef(0)); + spanless_eq_enum!(LitKind; Str(0 1) ByteStr(0) Byte(0) Char(0) Int(0 1) +- Float(0 1) FloatUnsuffixed(0) Bool(0) Err(0)); ++ Float(0 1) Bool(0) Err(0)); + spanless_eq_enum!(PatKind; Wild Ident(0 1 2) Struct(0 1 2) TupleStruct(0 1) + Or(0) Path(0 1) Tuple(0) Box(0) Ref(0 1) Lit(0) Range(0 1 2) Slice(0) Rest +- Paren(0) Mac(0)); ++ Paren(0) MacCall(0)); + spanless_eq_enum!(TyKind; Slice(0) Array(0 1) Ptr(0) Rptr(0 1) BareFn(0) Never + Tup(0) Path(0 1) TraitObject(0 1) ImplTrait(0 1) Paren(0) Typeof(0) Infer +- ImplicitSelf Mac(0) Err CVarArgs); ++ ImplicitSelf MacCall(0) Err CVarArgs); + + impl SpanlessEq for Ident { + fn eq(&self, other: &Self) -> bool { +@@ -414,44 +419,20 @@ impl SpanlessEq for TokenKind { + + impl SpanlessEq for TokenStream { + fn eq(&self, other: &Self) -> bool { +- SpanlessEq::eq(&expand_tts(self), &expand_tts(other)) +- } +-} +- +-fn expand_tts(tts: &TokenStream) -> Vec<TokenTree> { +- let mut tokens = Vec::new(); +- for tt in tts.clone().into_trees() { +- let c = match tt { +- TokenTree::Token(Token { +- kind: TokenKind::DocComment(c), +- .. +- }) => c, +- _ => { +- tokens.push(tt); +- continue; ++ let mut this = self.clone().into_trees(); ++ let mut other = other.clone().into_trees(); ++ loop { ++ let this = match this.next() { ++ None => return other.next().is_none(), ++ Some(val) => val, ++ }; ++ let other = match other.next() { ++ None => return false, ++ Some(val) => val, ++ }; ++ if !SpanlessEq::eq(&this, &other) { ++ return false; + } +- }; +- let contents = comments::strip_doc_comment_decoration(&c.as_str()); +- let style = comments::doc_comment_style(&c.as_str()); +- tokens.push(TokenTree::token(TokenKind::Pound, DUMMY_SP)); +- if style == AttrStyle::Inner { +- tokens.push(TokenTree::token(TokenKind::Not, DUMMY_SP)); + } +- let lit = token::Lit { +- kind: token::LitKind::Str, +- symbol: Symbol::intern(&contents), +- suffix: None, +- }; +- let tts = vec![ +- TokenTree::token(TokenKind::Ident(sym::doc, false), DUMMY_SP), +- TokenTree::token(TokenKind::Eq, DUMMY_SP), +- TokenTree::token(TokenKind::Literal(lit), DUMMY_SP), +- ]; +- tokens.push(TokenTree::Delimited( +- DelimSpan::dummy(), +- DelimToken::Bracket, +- tts.into_iter().collect::<TokenStream>().into(), +- )); + } +- tokens + } +diff --git a/third_party/rust/syn/tests/common/mod.rs b/third_party/rust/syn/tests/common/mod.rs +index 8b784beed7..a1cc80a16f 100644 +--- a/third_party/rust/syn/tests/common/mod.rs ++++ b/third_party/rust/syn/tests/common/mod.rs +@@ -1,5 +1,6 @@ + #![allow(dead_code)] + ++use rayon::ThreadPoolBuilder; + use std::env; + + pub mod eq; +@@ -12,3 +13,15 @@ pub fn abort_after() -> usize { + Err(_) => usize::max_value(), + } + } ++ ++/// Configure Rayon threadpool. ++pub fn rayon_init() { ++ let stack_size = match env::var("RUST_MIN_STACK") { ++ Ok(s) => s.parse().expect("failed to parse RUST_MIN_STACK"), ++ Err(_) => 20 * 1024 * 1024, ++ }; ++ ThreadPoolBuilder::new() ++ .stack_size(stack_size) ++ .build_global() ++ .unwrap(); ++} +diff --git a/third_party/rust/syn/tests/common/parse.rs b/third_party/rust/syn/tests/common/parse.rs +index 41d192f6fb..192828fedd 100644 +--- a/third_party/rust/syn/tests/common/parse.rs ++++ b/third_party/rust/syn/tests/common/parse.rs +@@ -1,20 +1,20 @@ +-extern crate proc_macro2; +-extern crate syn; +-extern crate syntax; +-extern crate syntax_pos; ++extern crate rustc_ast; ++extern crate rustc_expand; ++extern crate rustc_parse as parse; ++extern crate rustc_session; ++extern crate rustc_span; + +-use self::syntax::ast; +-use self::syntax::parse::{self, ParseSess}; +-use self::syntax::ptr::P; +-use self::syntax::source_map::FilePathMapping; +-use self::syntax_pos::FileName; ++use rustc_ast::ast; ++use rustc_ast::ptr::P; ++use rustc_session::parse::ParseSess; ++use rustc_span::source_map::FilePathMapping; ++use rustc_span::FileName; + + use std::panic; + +-pub fn libsyntax_expr(input: &str) -> Option<P<ast::Expr>> { ++pub fn librustc_expr(input: &str) -> Option<P<ast::Expr>> { + match panic::catch_unwind(|| { + let sess = ParseSess::new(FilePathMapping::empty()); +- sess.span_diagnostic.set_continue_after_error(false); + let e = parse::new_parser_from_source_str( + &sess, + FileName::Custom("test_precedence".to_string()), +@@ -32,7 +32,7 @@ pub fn libsyntax_expr(input: &str) -> Option<P<ast::Expr>> { + Ok(Some(e)) => Some(e), + Ok(None) => None, + Err(_) => { +- errorf!("libsyntax panicked\n"); ++ errorf!("librustc panicked\n"); + None + } + } +diff --git a/third_party/rust/syn/tests/debug/gen.rs b/third_party/rust/syn/tests/debug/gen.rs +index 8450c09ecf..85a1a39079 100644 +--- a/third_party/rust/syn/tests/debug/gen.rs ++++ b/third_party/rust/syn/tests/debug/gen.rs +@@ -2,7 +2,7 @@ + // It is not intended for manual editing. + + use super::{Lite, RefCast}; +-use std::fmt::{self, Debug}; ++use std::fmt::{self, Debug, Display}; + impl Debug for Lite<syn::Abi> { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { + let _val = &self.value; +@@ -1039,9 +1039,9 @@ impl Debug for Lite<syn::Expr> { + } + syn::Expr::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + syn::Expr::While(_val) => { +@@ -2116,9 +2116,9 @@ impl Debug for Lite<syn::ForeignItem> { + } + syn::ForeignItem::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + _ => unreachable!(), +@@ -2432,9 +2432,9 @@ impl Debug for Lite<syn::ImplItem> { + } + syn::ImplItem::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + _ => unreachable!(), +@@ -2940,9 +2940,9 @@ impl Debug for Lite<syn::Item> { + } + syn::Item::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + _ => unreachable!(), +@@ -3437,9 +3437,9 @@ impl Debug for Lite<syn::Lit> { + } + syn::Lit::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + } +@@ -3878,9 +3878,9 @@ impl Debug for Lite<syn::Pat> { + } + syn::Pat::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + syn::Pat::Wild(_val) => { +@@ -4674,9 +4674,9 @@ impl Debug for Lite<syn::TraitItem> { + } + syn::TraitItem::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + _ => unreachable!(), +@@ -5040,9 +5040,9 @@ impl Debug for Lite<syn::Type> { + } + syn::Type::Verbatim(_val) => { + formatter.write_str("Verbatim")?; +- formatter.write_str("(")?; +- Debug::fmt(Lite(_val), formatter)?; +- formatter.write_str(")")?; ++ formatter.write_str("(`")?; ++ Display::fmt(_val, formatter)?; ++ formatter.write_str("`)")?; + Ok(()) + } + _ => unreachable!(), +diff --git a/third_party/rust/syn/tests/debug/mod.rs b/third_party/rust/syn/tests/debug/mod.rs +index c1180532ec..cefebacef7 100644 +--- a/third_party/rust/syn/tests/debug/mod.rs ++++ b/third_party/rust/syn/tests/debug/mod.rs +@@ -1,10 +1,7 @@ +-extern crate proc_macro2; +-extern crate ref_cast; +- + mod gen; + +-use self::proc_macro2::{Ident, Literal, TokenStream}; +-use self::ref_cast::RefCast; ++use proc_macro2::{Ident, Literal, TokenStream}; ++use ref_cast::RefCast; + use std::fmt::{self, Debug}; + use std::ops::Deref; + use syn::punctuated::Punctuated; +@@ -66,7 +63,15 @@ impl Debug for Lite<Literal> { + + impl Debug for Lite<TokenStream> { + fn fmt(&self, formatter: &mut fmt::Formatter) -> fmt::Result { +- write!(formatter, "`{}`", self.value) ++ let string = self.value.to_string(); ++ if string.len() <= 80 { ++ write!(formatter, "TokenStream(`{}`)", self.value) ++ } else { ++ formatter ++ .debug_tuple("TokenStream") ++ .field(&format_args!("`{}`", string)) ++ .finish() ++ } + } + } + +diff --git a/third_party/rust/syn/tests/features/error.rs b/third_party/rust/syn/tests/features/error.rs +deleted file mode 100644 +index 10ac88965d..0000000000 +--- a/third_party/rust/syn/tests/features/error.rs ++++ /dev/null +@@ -1 +0,0 @@ +-"Hello! You want: cargo test --release --all-features" +diff --git a/third_party/rust/syn/tests/features/mod.rs b/third_party/rust/syn/tests/features/mod.rs +deleted file mode 100644 +index 83fbe13e7e..0000000000 +--- a/third_party/rust/syn/tests/features/mod.rs ++++ /dev/null +@@ -1,22 +0,0 @@ +-#[allow(unused_macros)] +-macro_rules! hide_from_rustfmt { +- ($mod:item) => { +- $mod +- }; +-} +- +-#[cfg(not(all( +- feature = "derive", +- feature = "full", +- feature = "parsing", +- feature = "printing", +- feature = "visit", +- feature = "visit-mut", +- feature = "fold", +- feature = "clone-impls", +- feature = "extra-traits", +- feature = "proc-macro", +-)))] +-hide_from_rustfmt! { +- mod error; +-} +diff --git a/third_party/rust/syn/tests/macros/mod.rs b/third_party/rust/syn/tests/macros/mod.rs +index c72fd01058..3994615fc4 100644 +--- a/third_party/rust/syn/tests/macros/mod.rs ++++ b/third_party/rust/syn/tests/macros/mod.rs +@@ -1,5 +1,3 @@ +-extern crate proc_macro2; +- + #[path = "../debug/mod.rs"] + pub mod debug; + +@@ -42,18 +40,18 @@ macro_rules! snapshot_impl { + (($expr:ident) as $t:ty, @$snapshot:literal) => { + let $expr = crate::macros::Tokens::parse::<$t>($expr).unwrap(); + let debug = crate::macros::debug::Lite(&$expr); +- insta::assert_debug_snapshot_matches!(debug, @$snapshot); ++ insta::assert_debug_snapshot!(debug, @$snapshot); + }; + (($($expr:tt)*) as $t:ty, @$snapshot:literal) => {{ + let syntax_tree = crate::macros::Tokens::parse::<$t>($($expr)*).unwrap(); + let debug = crate::macros::debug::Lite(&syntax_tree); +- insta::assert_debug_snapshot_matches!(debug, @$snapshot); ++ insta::assert_debug_snapshot!(debug, @$snapshot); + syntax_tree + }}; + (($($expr:tt)*) , @$snapshot:literal) => {{ + let syntax_tree = $($expr)*; + let debug = crate::macros::debug::Lite(&syntax_tree); +- insta::assert_debug_snapshot_matches!(debug, @$snapshot); ++ insta::assert_debug_snapshot!(debug, @$snapshot); + syntax_tree + }}; + (($($expr:tt)*) $next:tt $($rest:tt)*) => { +diff --git a/third_party/rust/syn/tests/repo/mod.rs b/third_party/rust/syn/tests/repo/mod.rs +index c22cb03758..1d3e1f0e74 100644 +--- a/third_party/rust/syn/tests/repo/mod.rs ++++ b/third_party/rust/syn/tests/repo/mod.rs +@@ -1,8 +1,37 @@ +-extern crate walkdir; ++mod progress; + +-use std::process::Command; ++use self::progress::Progress; ++use anyhow::Result; ++use flate2::read::GzDecoder; ++use std::fs; ++use std::path::Path; ++use tar::Archive; ++use walkdir::DirEntry; + +-use self::walkdir::DirEntry; ++const REVISION: &str = "792c645ca7d11a8d254df307d019c5bf01445c37"; ++ ++#[rustfmt::skip] ++static EXCLUDE: &[&str] = &[ ++ // Compile-fail expr parameter in const generic position: f::<1 + 2>() ++ "test/ui/const-generics/const-expression-parameter.rs", ++ ++ // Deprecated anonymous parameter syntax in traits ++ "test/ui/issues/issue-13105.rs", ++ "test/ui/issues/issue-13775.rs", ++ "test/ui/issues/issue-34074.rs", ++ "test/ui/proc-macro/trait-fn-args-2015.rs", ++ ++ // Not actually test cases ++ "test/rustdoc-ui/test-compile-fail2.rs", ++ "test/rustdoc-ui/test-compile-fail3.rs", ++ "test/ui/include-single-expr-helper.rs", ++ "test/ui/include-single-expr-helper-1.rs", ++ "test/ui/issues/auxiliary/issue-21146-inc.rs", ++ "test/ui/json-bom-plus-crlf-multifile-aux.rs", ++ "test/ui/lint/expansion-time-include.rs", ++ "test/ui/macros/auxiliary/macro-comma-support.rs", ++ "test/ui/macros/auxiliary/macro-include-items-expr.rs", ++]; + + pub fn base_dir_filter(entry: &DirEntry) -> bool { + let path = entry.path(); +@@ -12,49 +41,95 @@ pub fn base_dir_filter(entry: &DirEntry) -> bool { + if path.extension().map(|e| e != "rs").unwrap_or(true) { + return false; + } +- let path_string = path.to_string_lossy(); +- let path_string = if cfg!(windows) { +- path_string.replace('\\', "/").into() ++ ++ let mut path_string = path.to_string_lossy(); ++ if cfg!(windows) { ++ path_string = path_string.replace('\\', "/").into(); ++ } ++ let path = if let Some(path) = path_string.strip_prefix("tests/rust/src/") { ++ path ++ } else if let Some(path) = path_string.strip_prefix("tests/rust/library/") { ++ path + } else { +- path_string ++ panic!("unexpected path in Rust dist: {}", path_string); + }; ++ + // TODO assert that parsing fails on the parse-fail cases +- if path_string.starts_with("tests/rust/src/test/parse-fail") +- || path_string.starts_with("tests/rust/src/test/compile-fail") +- || path_string.starts_with("tests/rust/src/test/rustfix") ++ if path.starts_with("test/parse-fail") ++ || path.starts_with("test/compile-fail") ++ || path.starts_with("test/rustfix") + { + return false; + } + +- if path_string.starts_with("tests/rust/src/test/ui") { +- let stderr_path = path.with_extension("stderr"); ++ if path.starts_with("test/ui") { ++ let stderr_path = entry.path().with_extension("stderr"); + if stderr_path.exists() { + // Expected to fail in some way + return false; + } + } + +- match path_string.as_ref() { +- // Deprecated placement syntax +- "tests/rust/src/test/ui/obsolete-in-place/bad.rs" | +- // Deprecated anonymous parameter syntax in traits +- "tests/rust/src/test/ui/error-codes/e0119/auxiliary/issue-23563-a.rs" | +- "tests/rust/src/test/ui/issues/issue-13105.rs" | +- "tests/rust/src/test/ui/issues/issue-13775.rs" | +- "tests/rust/src/test/ui/issues/issue-34074.rs" | +- // Deprecated await macro syntax +- "tests/rust/src/test/ui/async-await/await-macro.rs" | +- // 2015-style dyn that libsyntax rejects +- "tests/rust/src/test/ui/dyn-keyword/dyn-2015-no-warnings-without-lints.rs" | +- // not actually test cases +- "tests/rust/src/test/ui/macros/auxiliary/macro-comma-support.rs" | +- "tests/rust/src/test/ui/macros/auxiliary/macro-include-items-expr.rs" | +- "tests/rust/src/test/ui/issues/auxiliary/issue-21146-inc.rs" => false, +- _ => true, ++ !EXCLUDE.contains(&path) ++} ++ ++#[allow(dead_code)] ++pub fn edition(path: &Path) -> &'static str { ++ if path.ends_with("dyn-2015-no-warnings-without-lints.rs") { ++ "2015" ++ } else { ++ "2018" + } + } + + pub fn clone_rust() { +- let result = Command::new("tests/clone.sh").status().unwrap(); +- assert!(result.success()); ++ let needs_clone = match fs::read_to_string("tests/rust/COMMIT") { ++ Err(_) => true, ++ Ok(contents) => contents.trim() != REVISION, ++ }; ++ if needs_clone { ++ download_and_unpack().unwrap(); ++ } ++ let mut missing = String::new(); ++ let test_src = Path::new("tests/rust/src"); ++ for exclude in EXCLUDE { ++ if !test_src.join(exclude).exists() { ++ missing += "\ntests/rust/src/"; ++ missing += exclude; ++ } ++ } ++ if !missing.is_empty() { ++ panic!("excluded test file does not exist:{}\n", missing); ++ } ++} ++ ++fn download_and_unpack() -> Result<()> { ++ let url = format!( ++ "https://github.com/rust-lang/rust/archive/{}.tar.gz", ++ REVISION ++ ); ++ let response = reqwest::blocking::get(&url)?.error_for_status()?; ++ let progress = Progress::new(response); ++ let decoder = GzDecoder::new(progress); ++ let mut archive = Archive::new(decoder); ++ let prefix = format!("rust-{}", REVISION); ++ ++ let tests_rust = Path::new("tests/rust"); ++ if tests_rust.exists() { ++ fs::remove_dir_all(tests_rust)?; ++ } ++ ++ for entry in archive.entries()? { ++ let mut entry = entry?; ++ let path = entry.path()?; ++ if path == Path::new("pax_global_header") { ++ continue; ++ } ++ let relative = path.strip_prefix(&prefix)?; ++ let out = tests_rust.join(relative); ++ entry.unpack(&out)?; ++ } ++ ++ fs::write("tests/rust/COMMIT", REVISION)?; ++ Ok(()) + } +diff --git a/third_party/rust/syn/tests/repo/progress.rs b/third_party/rust/syn/tests/repo/progress.rs +new file mode 100644 +index 0000000000..28c8a44b12 +--- /dev/null ++++ b/third_party/rust/syn/tests/repo/progress.rs +@@ -0,0 +1,37 @@ ++use std::io::{Read, Result}; ++use std::time::{Duration, Instant}; ++ ++pub struct Progress<R> { ++ bytes: usize, ++ tick: Instant, ++ stream: R, ++} ++ ++impl<R> Progress<R> { ++ pub fn new(stream: R) -> Self { ++ Progress { ++ bytes: 0, ++ tick: Instant::now() + Duration::from_millis(2000), ++ stream, ++ } ++ } ++} ++ ++impl<R: Read> Read for Progress<R> { ++ fn read(&mut self, buf: &mut [u8]) -> Result<usize> { ++ let num = self.stream.read(buf)?; ++ self.bytes += num; ++ let now = Instant::now(); ++ if now > self.tick { ++ self.tick = now + Duration::from_millis(500); ++ errorf!("downloading... {} bytes\n", self.bytes); ++ } ++ Ok(num) ++ } ++} ++ ++impl<R> Drop for Progress<R> { ++ fn drop(&mut self) { ++ errorf!("done ({} bytes)\n", self.bytes); ++ } ++} +diff --git a/third_party/rust/syn/tests/test_asyncness.rs b/third_party/rust/syn/tests/test_asyncness.rs +index f868fbcc20..0efef5976f 100644 +--- a/third_party/rust/syn/tests/test_asyncness.rs ++++ b/third_party/rust/syn/tests/test_asyncness.rs +@@ -1,7 +1,3 @@ +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -12,16 +8,16 @@ fn test_async_fn() { + let input = "async fn process() {}"; + + snapshot!(input as Item, @r###" +- ⋮Item::Fn { +- ⋮ vis: Inherited, +- ⋮ sig: Signature { +- ⋮ asyncness: Some, +- ⋮ ident: "process", +- ⋮ generics: Generics, +- ⋮ output: Default, +- ⋮ }, +- ⋮ block: Block, +- ⋮} ++ Item::Fn { ++ vis: Inherited, ++ sig: Signature { ++ asyncness: Some, ++ ident: "process", ++ generics: Generics, ++ output: Default, ++ }, ++ block: Block, ++ } + "###); + } + +@@ -30,12 +26,12 @@ fn test_async_closure() { + let input = "async || {}"; + + snapshot!(input as Expr, @r###" +- ⋮Expr::Closure { +- ⋮ asyncness: Some, +- ⋮ output: Default, +- ⋮ body: Expr::Block { +- ⋮ block: Block, +- ⋮ }, +- ⋮} ++ Expr::Closure { ++ asyncness: Some, ++ output: Default, ++ body: Expr::Block { ++ block: Block, ++ }, ++ } + "###); + } +diff --git a/third_party/rust/syn/tests/test_attribute.rs b/third_party/rust/syn/tests/test_attribute.rs +index aff6294fc3..c26bd090ec 100644 +--- a/third_party/rust/syn/tests/test_attribute.rs ++++ b/third_party/rust/syn/tests/test_attribute.rs +@@ -1,7 +1,3 @@ +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -13,14 +9,14 @@ fn test_meta_item_word() { + let meta = test("#[foo]"); + + snapshot!(meta, @r###" +- ⋮Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮}) ++ Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }) + "###); + } + +@@ -29,17 +25,17 @@ fn test_meta_item_name_value() { + let meta = test("#[foo = 5]"); + + snapshot!(meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + } + +@@ -48,37 +44,37 @@ fn test_meta_item_bool_value() { + let meta = test("#[foo = true]"); + + snapshot!(meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: Lit::Bool { +- ⋮ value: true, +- ⋮ }, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: Lit::Bool { ++ value: true, ++ }, ++ } + "###); + + let meta = test("#[foo = false]"); + + snapshot!(meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: Lit::Bool { +- ⋮ value: false, +- ⋮ }, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: Lit::Bool { ++ value: false, ++ }, ++ } + "###); + } + +@@ -87,19 +83,19 @@ fn test_meta_item_list_lit() { + let meta = test("#[foo(5)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Lit(5), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Lit(5), ++ ], ++ } + "###); + } + +@@ -108,26 +104,26 @@ fn test_meta_item_list_word() { + let meta = test("#[foo(bar)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "bar", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "bar", ++ arguments: None, ++ }, ++ ], ++ })), ++ ], ++ } + "###); + } + +@@ -136,29 +132,29 @@ fn test_meta_item_list_name_value() { + let meta = test("#[foo(bar = 5)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "bar", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮ }), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "bar", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ }), ++ ], ++ } + "###); + } + +@@ -167,31 +163,31 @@ fn test_meta_item_list_bool_value() { + let meta = test("#[foo(bar = true)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "bar", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: Lit::Bool { +- ⋮ value: true, +- ⋮ }, +- ⋮ }), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "bar", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: Lit::Bool { ++ value: true, ++ }, ++ }), ++ ], ++ } + "###); + } + +@@ -200,68 +196,68 @@ fn test_meta_item_multiple() { + let meta = test("#[foo(word, name = 5, list(name2 = 6), word2)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮ }), +- ⋮ Meta(Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "list", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 6, +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word", ++ arguments: None, ++ }, ++ ], ++ })), ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ }), ++ Meta(Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "list", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name2", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 6, ++ }), ++ ], ++ }), ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word2", ++ arguments: None, ++ }, ++ ], ++ })), ++ ], ++ } + "###); + } + +@@ -270,21 +266,63 @@ fn test_bool_lit() { + let meta = test("#[foo(true)]"); + + snapshot!(meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Lit(Lit::Bool { +- ⋮ value: true, +- ⋮ }), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Lit(Lit::Bool { ++ value: true, ++ }), ++ ], ++ } ++ "###); ++} ++ ++#[test] ++fn test_negative_lit() { ++ let meta = test("#[form(min = -1, max = 200)]"); ++ ++ snapshot!(meta, @r###" ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "form", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "min", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: -1, ++ }), ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "max", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 200, ++ }), ++ ], ++ } + "###); + } + +diff --git a/third_party/rust/syn/tests/test_derive_input.rs b/third_party/rust/syn/tests/test_derive_input.rs +index de68240166..bf1ebdb67d 100644 +--- a/third_party/rust/syn/tests/test_derive_input.rs ++++ b/third_party/rust/syn/tests/test_derive_input.rs +@@ -1,8 +1,3 @@ +-extern crate quote; +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -16,15 +11,15 @@ fn test_unit() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "Unit", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "Unit", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -39,105 +34,105 @@ fn test_struct() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "derive", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: `( Debug , Clone )`, +- ⋮ }, +- ⋮ ], +- ⋮ vis: Visibility::Public, +- ⋮ ident: "Item", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Fields::Named { +- ⋮ named: [ +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ident: Some("ident"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Ident", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ident: Some("attrs"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Vec", +- ⋮ arguments: PathArguments::AngleBracketed { +- ⋮ args: [ +- ⋮ Type(Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Attribute", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }), +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "derive", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(`(Debug , Clone)`), ++ }, ++ ], ++ vis: Visibility::Public, ++ ident: "Item", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Named { ++ named: [ ++ Field { ++ vis: Visibility::Public, ++ ident: Some("ident"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Ident", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ Field { ++ vis: Visibility::Public, ++ ident: Some("attrs"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Vec", ++ arguments: PathArguments::AngleBracketed { ++ args: [ ++ Type(Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Attribute", ++ arguments: None, ++ }, ++ ], ++ }, ++ }), ++ ], ++ }, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ }, ++ } + "###); + + snapshot!(input.attrs[0].parse_meta().unwrap(), @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "derive", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Debug", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Clone", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "derive", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "Debug", ++ arguments: None, ++ }, ++ ], ++ })), ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "Clone", ++ arguments: None, ++ }, ++ ], ++ })), ++ ], ++ } + "###); + } + +@@ -151,46 +146,46 @@ fn test_union() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "MaybeUninit", +- ⋮ generics: Generics { +- ⋮ lt_token: Some, +- ⋮ params: [ +- ⋮ Type(TypeParam { +- ⋮ ident: "T", +- ⋮ }), +- ⋮ ], +- ⋮ gt_token: Some, +- ⋮ }, +- ⋮ data: Data::Union { +- ⋮ fields: FieldsNamed { +- ⋮ named: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ident: Some("uninit"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Tuple, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ident: Some("value"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "T", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "MaybeUninit", ++ generics: Generics { ++ lt_token: Some, ++ params: [ ++ Type(TypeParam { ++ ident: "T", ++ }), ++ ], ++ gt_token: Some, ++ }, ++ data: Data::Union { ++ fields: FieldsNamed { ++ named: [ ++ Field { ++ vis: Inherited, ++ ident: Some("uninit"), ++ colon_token: Some, ++ ty: Type::Tuple, ++ }, ++ Field { ++ vis: Inherited, ++ ident: Some("value"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "T", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ }, ++ } + "###); + } + +@@ -212,118 +207,118 @@ fn test_enum() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "doc", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: `= r" See the std::result module documentation for details."`, +- ⋮ }, +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "must_use", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: ``, +- ⋮ }, +- ⋮ ], +- ⋮ vis: Visibility::Public, +- ⋮ ident: "Result", +- ⋮ generics: Generics { +- ⋮ lt_token: Some, +- ⋮ params: [ +- ⋮ Type(TypeParam { +- ⋮ ident: "T", +- ⋮ }), +- ⋮ Type(TypeParam { +- ⋮ ident: "E", +- ⋮ }), +- ⋮ ], +- ⋮ gt_token: Some, +- ⋮ }, +- ⋮ data: Data::Enum { +- ⋮ variants: [ +- ⋮ Variant { +- ⋮ ident: "Ok", +- ⋮ fields: Fields::Unnamed { +- ⋮ unnamed: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "T", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ Variant { +- ⋮ ident: "Err", +- ⋮ fields: Fields::Unnamed { +- ⋮ unnamed: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "E", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ Variant { +- ⋮ ident: "Surprise", +- ⋮ fields: Unit, +- ⋮ discriminant: Some(Expr::Lit { +- ⋮ lit: 0isize, +- ⋮ }), +- ⋮ }, +- ⋮ Variant { +- ⋮ ident: "ProcMacroHack", +- ⋮ fields: Unit, +- ⋮ discriminant: Some(Expr::Field { +- ⋮ base: Expr::Tuple { +- ⋮ elems: [ +- ⋮ Expr::Lit { +- ⋮ lit: 0, +- ⋮ }, +- ⋮ Expr::Lit { +- ⋮ lit: "data", +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ member: Unnamed(Index { +- ⋮ index: 0, +- ⋮ }), +- ⋮ }), +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮} ++ DeriveInput { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "doc", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(`= r" See the std::result module documentation for details."`), ++ }, ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "must_use", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ vis: Visibility::Public, ++ ident: "Result", ++ generics: Generics { ++ lt_token: Some, ++ params: [ ++ Type(TypeParam { ++ ident: "T", ++ }), ++ Type(TypeParam { ++ ident: "E", ++ }), ++ ], ++ gt_token: Some, ++ }, ++ data: Data::Enum { ++ variants: [ ++ Variant { ++ ident: "Ok", ++ fields: Fields::Unnamed { ++ unnamed: [ ++ Field { ++ vis: Inherited, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "T", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ }, ++ Variant { ++ ident: "Err", ++ fields: Fields::Unnamed { ++ unnamed: [ ++ Field { ++ vis: Inherited, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "E", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ }, ++ Variant { ++ ident: "Surprise", ++ fields: Unit, ++ discriminant: Some(Expr::Lit { ++ lit: 0isize, ++ }), ++ }, ++ Variant { ++ ident: "ProcMacroHack", ++ fields: Unit, ++ discriminant: Some(Expr::Field { ++ base: Expr::Tuple { ++ elems: [ ++ Expr::Lit { ++ lit: 0, ++ }, ++ Expr::Lit { ++ lit: "data", ++ }, ++ ], ++ }, ++ member: Unnamed(Index { ++ index: 0, ++ }), ++ }), ++ }, ++ ], ++ }, ++ } + "###); + + let meta_items: Vec<_> = input +@@ -333,27 +328,27 @@ fn test_enum() { + .collect(); + + snapshot!(meta_items, @r###" +- ⋮[ +- ⋮ Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "doc", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: " See the std::result module documentation for details.", +- ⋮ }, +- ⋮ Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "must_use", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }), +- ⋮] ++ [ ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "doc", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: " See the std::result module documentation for details.", ++ }, ++ Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "must_use", ++ arguments: None, ++ }, ++ ], ++ }), ++ ] + "###); + } + +@@ -366,34 +361,34 @@ fn test_attr_with_path() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ leading_colon: Some, +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "attr_args", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ PathSegment { +- ⋮ ident: "identity", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: `fn main ( ) { assert_eq ! ( foo ( ) , "Hello, world!" ) ; }`, +- ⋮ }, +- ⋮ ], +- ⋮ vis: Inherited, +- ⋮ ident: "Dummy", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ leading_colon: Some, ++ segments: [ ++ PathSegment { ++ ident: "attr_args", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "identity", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(`fn main () { assert_eq ! (foo () , "Hello, world!") ; }`), ++ }, ++ ], ++ vis: Inherited, ++ ident: "Dummy", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + + assert!(input.attrs[0].parse_meta().is_err()); +@@ -407,29 +402,29 @@ fn test_attr_with_non_mod_style_path() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "inert", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: `< T >`, +- ⋮ }, +- ⋮ ], +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "inert", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(`< T >`), ++ }, ++ ], ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + + assert!(input.attrs[0].parse_meta().is_err()); +@@ -443,48 +438,48 @@ fn test_attr_with_mod_style_path_with_self() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ PathSegment { +- ⋮ ident: "self", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: ``, +- ⋮ }, +- ⋮ ], +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "self", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + + snapshot!(input.attrs[0].parse_meta().unwrap(), @r###" +- ⋮Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ PathSegment { +- ⋮ ident: "self", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮}) ++ Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "self", ++ arguments: None, ++ }, ++ ], ++ }) + "###); + } + +@@ -496,55 +491,55 @@ fn test_pub_restricted() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Visibility::Restricted { +- ⋮ in_token: Some, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "m", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ident: "Z", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Fields::Unnamed { +- ⋮ unnamed: [ +- ⋮ Field { +- ⋮ vis: Visibility::Restricted { +- ⋮ in_token: Some, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "m", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ PathSegment { +- ⋮ ident: "n", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "u8", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Visibility::Restricted { ++ in_token: Some, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "m", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ident: "Z", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Unnamed { ++ unnamed: [ ++ Field { ++ vis: Visibility::Restricted { ++ in_token: Some, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "m", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "n", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "u8", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -555,15 +550,15 @@ fn test_vis_crate() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Visibility::Crate, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Visibility::Crate, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -574,24 +569,24 @@ fn test_pub_restricted_crate() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Visibility::Restricted { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "crate", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Visibility::Restricted { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "crate", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -602,24 +597,24 @@ fn test_pub_restricted_super() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Visibility::Restricted { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "super", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Visibility::Restricted { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "super", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -630,25 +625,25 @@ fn test_pub_restricted_in_super() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Visibility::Restricted { +- ⋮ in_token: Some, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "super", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Visibility::Restricted { ++ in_token: Some, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "super", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + } + +@@ -659,15 +654,15 @@ fn test_fields_on_unit_struct() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + + let data = match input.data { +@@ -688,47 +683,47 @@ fn test_fields_on_named_struct() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Fields::Named { +- ⋮ named: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ident: Some("foo"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "i32", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ident: Some("bar"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "String", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Named { ++ named: [ ++ Field { ++ vis: Inherited, ++ ident: Some("foo"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "i32", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ Field { ++ vis: Visibility::Public, ++ ident: Some("bar"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "String", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ }, ++ } + "###); + + let data = match input.data { +@@ -737,38 +732,38 @@ fn test_fields_on_named_struct() { + }; + + snapshot!(data.fields.into_iter().collect::<Vec<_>>(), @r###" +- ⋮[ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ident: Some("foo"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "i32", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ident: Some("bar"), +- ⋮ colon_token: Some, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "String", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮] ++ [ ++ Field { ++ vis: Inherited, ++ ident: Some("foo"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "i32", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ Field { ++ vis: Visibility::Public, ++ ident: Some("bar"), ++ colon_token: Some, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "String", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ] + "###); + } + +@@ -779,44 +774,44 @@ fn test_fields_on_tuple_struct() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Fields::Unnamed { +- ⋮ unnamed: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "i32", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "String", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Unnamed { ++ unnamed: [ ++ Field { ++ vis: Inherited, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "i32", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ Field { ++ vis: Visibility::Public, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "String", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ semi_token: Some, ++ }, ++ } + "###); + + let data = match input.data { +@@ -825,34 +820,34 @@ fn test_fields_on_tuple_struct() { + }; + + snapshot!(data.fields.iter().collect::<Vec<_>>(), @r###" +- ⋮[ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "i32", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ Field { +- ⋮ vis: Visibility::Public, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "String", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮] ++ [ ++ Field { ++ vis: Inherited, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "i32", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ Field { ++ vis: Visibility::Public, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "String", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ] + "###); + } + +@@ -864,34 +859,34 @@ fn test_ambiguous_crate() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics, +- ⋮ data: Data::Struct { +- ⋮ fields: Fields::Unnamed { +- ⋮ unnamed: [ +- ⋮ Field { +- ⋮ vis: Inherited, +- ⋮ ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "crate", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ PathSegment { +- ⋮ ident: "X", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Unnamed { ++ unnamed: [ ++ Field { ++ vis: Inherited, ++ ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "crate", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "X", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ }, ++ semi_token: Some, ++ }, ++ } + "###); + } +diff --git a/third_party/rust/syn/tests/test_expr.rs b/third_party/rust/syn/tests/test_expr.rs +index c8a11cec2c..b2b65a254f 100644 +--- a/third_party/rust/syn/tests/test_expr.rs ++++ b/third_party/rust/syn/tests/test_expr.rs +@@ -1,40 +1,302 @@ +-extern crate proc_macro2; +-extern crate syn; +- + #[macro_use] + mod macros; + +-use std::str::FromStr; +- +-use proc_macro2::TokenStream; ++use proc_macro2::{Delimiter, Group, Ident, Punct, Spacing, Span, TokenStream, TokenTree}; ++use quote::quote; ++use std::iter::FromIterator; + use syn::{Expr, ExprRange}; + + #[test] + fn test_expr_parse() { +- let code = "..100u32"; +- let tt = TokenStream::from_str(code).unwrap(); +- let expr: Expr = syn::parse2(tt.clone()).unwrap(); +- let expr_range: ExprRange = syn::parse2(tt).unwrap(); +- assert_eq!(expr, Expr::Range(expr_range)); ++ let tokens = quote!(..100u32); ++ snapshot!(tokens as Expr, @r###" ++ Expr::Range { ++ limits: HalfOpen, ++ to: Some(Expr::Lit { ++ lit: 100u32, ++ }), ++ } ++ "###); ++ ++ let tokens = quote!(..100u32); ++ snapshot!(tokens as ExprRange, @r###" ++ ExprRange { ++ limits: HalfOpen, ++ to: Some(Expr::Lit { ++ lit: 100u32, ++ }), ++ } ++ "###); + } + + #[test] + fn test_await() { + // Must not parse as Expr::Field. +- let expr = syn::parse_str::<Expr>("fut.await").unwrap(); +- +- snapshot!(expr, @r###" +- ⋮Expr::Await { +- ⋮ base: Expr::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "fut", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮} ++ let tokens = quote!(fut.await); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Await { ++ base: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "fut", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ } ++ "###); ++} ++ ++#[rustfmt::skip] ++#[test] ++fn test_tuple_multi_index() { ++ for &input in &[ ++ "tuple.0.0", ++ "tuple .0.0", ++ "tuple. 0.0", ++ "tuple.0 .0", ++ "tuple.0. 0", ++ "tuple . 0 . 0", ++ ] { ++ snapshot!(input as Expr, @r###" ++ Expr::Field { ++ base: Expr::Field { ++ base: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "tuple", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ member: Unnamed(Index { ++ index: 0, ++ }), ++ }, ++ member: Unnamed(Index { ++ index: 0, ++ }), ++ } ++ "###); ++ } ++ ++ for tokens in vec![ ++ quote!(tuple.0.0), ++ quote!(tuple .0.0), ++ quote!(tuple. 0.0), ++ quote!(tuple.0 .0), ++ quote!(tuple.0. 0), ++ quote!(tuple . 0 . 0), ++ ] { ++ snapshot!(tokens as Expr, @r###" ++ Expr::Field { ++ base: Expr::Field { ++ base: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "tuple", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ member: Unnamed(Index { ++ index: 0, ++ }), ++ }, ++ member: Unnamed(Index { ++ index: 0, ++ }), ++ } ++ "###); ++ } ++} ++ ++#[test] ++fn test_macro_variable_func() { ++ // mimics the token stream corresponding to `$fn()` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { f })), ++ TokenTree::Group(Group::new(Delimiter::Parenthesis, TokenStream::new())), ++ ]); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Call { ++ func: Expr::Group { ++ expr: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "f", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ } ++ "###); ++ ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Punct(Punct::new('#', Spacing::Alone)), ++ TokenTree::Group(Group::new(Delimiter::Bracket, quote! { outside })), ++ TokenTree::Group(Group::new(Delimiter::None, quote! { #[inside] f })), ++ TokenTree::Group(Group::new(Delimiter::Parenthesis, TokenStream::new())), ++ ]); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Call { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "outside", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ func: Expr::Group { ++ expr: Expr::Path { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "inside", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "f", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }, ++ } ++ "###); ++} ++ ++#[test] ++fn test_macro_variable_macro() { ++ // mimics the token stream corresponding to `$macro!()` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { m })), ++ TokenTree::Punct(Punct::new('!', Spacing::Alone)), ++ TokenTree::Group(Group::new(Delimiter::Parenthesis, TokenStream::new())), ++ ]); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Macro { ++ mac: Macro { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "m", ++ arguments: None, ++ }, ++ ], ++ }, ++ delimiter: Paren, ++ tokens: TokenStream(``), ++ }, ++ } ++ "###); ++} ++ ++#[test] ++fn test_macro_variable_struct() { ++ // mimics the token stream corresponding to `$struct {}` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { S })), ++ TokenTree::Group(Group::new(Delimiter::Brace, TokenStream::new())), ++ ]); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Struct { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "S", ++ arguments: None, ++ }, ++ ], ++ }, ++ } ++ "###); ++} ++ ++#[test] ++fn test_macro_variable_match_arm() { ++ // mimics the token stream corresponding to `match v { _ => $expr }` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Ident(Ident::new("match", Span::call_site())), ++ TokenTree::Ident(Ident::new("v", Span::call_site())), ++ TokenTree::Group(Group::new( ++ Delimiter::Brace, ++ TokenStream::from_iter(vec![ ++ TokenTree::Punct(Punct::new('_', Spacing::Alone)), ++ TokenTree::Punct(Punct::new('=', Spacing::Joint)), ++ TokenTree::Punct(Punct::new('>', Spacing::Alone)), ++ TokenTree::Group(Group::new(Delimiter::None, quote! { #[a] () })), ++ ]), ++ )), ++ ]); ++ ++ snapshot!(tokens as Expr, @r###" ++ Expr::Match { ++ expr: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "v", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ arms: [ ++ Arm { ++ pat: Pat::Wild, ++ body: Expr::Group { ++ expr: Expr::Tuple { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "a", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ }, ++ }, ++ }, ++ ], ++ } + "###); + } +diff --git a/third_party/rust/syn/tests/test_generics.rs b/third_party/rust/syn/tests/test_generics.rs +index 55c79e066b..b29434a147 100644 +--- a/third_party/rust/syn/tests/test_generics.rs ++++ b/third_party/rust/syn/tests/test_generics.rs +@@ -1,8 +1,3 @@ +-extern crate quote; +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -16,90 +11,90 @@ fn test_split_for_impl() { + }; + + snapshot!(input as DeriveInput, @r###" +- ⋮DeriveInput { +- ⋮ vis: Inherited, +- ⋮ ident: "S", +- ⋮ generics: Generics { +- ⋮ lt_token: Some, +- ⋮ params: [ +- ⋮ Lifetime(LifetimeDef { +- ⋮ lifetime: Lifetime { +- ⋮ ident: "a", +- ⋮ }, +- ⋮ }), +- ⋮ Lifetime(LifetimeDef { +- ⋮ lifetime: Lifetime { +- ⋮ ident: "b", +- ⋮ }, +- ⋮ colon_token: Some, +- ⋮ bounds: [ +- ⋮ Lifetime { +- ⋮ ident: "a", +- ⋮ }, +- ⋮ ], +- ⋮ }), +- ⋮ Type(TypeParam { +- ⋮ attrs: [ +- ⋮ Attribute { +- ⋮ style: Outer, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "may_dangle", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ tokens: ``, +- ⋮ }, +- ⋮ ], +- ⋮ ident: "T", +- ⋮ colon_token: Some, +- ⋮ bounds: [ +- ⋮ Lifetime(Lifetime { +- ⋮ ident: "a", +- ⋮ }), +- ⋮ ], +- ⋮ eq_token: Some, +- ⋮ default: Some(Type::Tuple), +- ⋮ }), +- ⋮ ], +- ⋮ gt_token: Some, +- ⋮ where_clause: Some(WhereClause { +- ⋮ predicates: [ +- ⋮ Type(PredicateType { +- ⋮ bounded_ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "T", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ bounds: [ +- ⋮ Trait(TraitBound { +- ⋮ modifier: None, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Debug", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ }, +- ⋮ data: Data::Struct { +- ⋮ fields: Unit, +- ⋮ semi_token: Some, +- ⋮ }, +- ⋮} ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics { ++ lt_token: Some, ++ params: [ ++ Lifetime(LifetimeDef { ++ lifetime: Lifetime { ++ ident: "a", ++ }, ++ }), ++ Lifetime(LifetimeDef { ++ lifetime: Lifetime { ++ ident: "b", ++ }, ++ colon_token: Some, ++ bounds: [ ++ Lifetime { ++ ident: "a", ++ }, ++ ], ++ }), ++ Type(TypeParam { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "may_dangle", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ ident: "T", ++ colon_token: Some, ++ bounds: [ ++ Lifetime(Lifetime { ++ ident: "a", ++ }), ++ ], ++ eq_token: Some, ++ default: Some(Type::Tuple), ++ }), ++ ], ++ gt_token: Some, ++ where_clause: Some(WhereClause { ++ predicates: [ ++ Type(PredicateType { ++ bounded_ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "T", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ bounds: [ ++ Trait(TraitBound { ++ modifier: None, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Debug", ++ arguments: None, ++ }, ++ ], ++ }, ++ }), ++ ], ++ }), ++ ], ++ }), ++ }, ++ data: Data::Struct { ++ fields: Unit, ++ semi_token: Some, ++ }, ++ } + "###); + + let generics = input.generics; +@@ -131,46 +126,46 @@ fn test_split_for_impl() { + fn test_ty_param_bound() { + let tokens = quote!('a); + snapshot!(tokens as TypeParamBound, @r###" +- ⋮Lifetime(Lifetime { +- ⋮ ident: "a", +- ⋮}) ++ Lifetime(Lifetime { ++ ident: "a", ++ }) + "###); + + let tokens = quote!('_); + snapshot!(tokens as TypeParamBound, @r###" +- ⋮Lifetime(Lifetime { +- ⋮ ident: "_", +- ⋮}) ++ Lifetime(Lifetime { ++ ident: "_", ++ }) + "###); + + let tokens = quote!(Debug); + snapshot!(tokens as TypeParamBound, @r###" +- ⋮Trait(TraitBound { +- ⋮ modifier: None, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Debug", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮}) ++ Trait(TraitBound { ++ modifier: None, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Debug", ++ arguments: None, ++ }, ++ ], ++ }, ++ }) + "###); + + let tokens = quote!(?Sized); + snapshot!(tokens as TypeParamBound, @r###" +- ⋮Trait(TraitBound { +- ⋮ modifier: Maybe, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Sized", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮}) ++ Trait(TraitBound { ++ modifier: Maybe, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Sized", ++ arguments: None, ++ }, ++ ], ++ }, ++ }) + "###); + } + +@@ -187,76 +182,76 @@ fn test_fn_precedence_in_where_clause() { + }; + + snapshot!(input as ItemFn, @r###" +- ⋮ItemFn { +- ⋮ vis: Inherited, +- ⋮ sig: Signature { +- ⋮ ident: "f", +- ⋮ generics: Generics { +- ⋮ lt_token: Some, +- ⋮ params: [ +- ⋮ Type(TypeParam { +- ⋮ ident: "G", +- ⋮ }), +- ⋮ ], +- ⋮ gt_token: Some, +- ⋮ where_clause: Some(WhereClause { +- ⋮ predicates: [ +- ⋮ Type(PredicateType { +- ⋮ bounded_ty: Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "G", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ bounds: [ +- ⋮ Trait(TraitBound { +- ⋮ modifier: None, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "FnOnce", +- ⋮ arguments: PathArguments::Parenthesized { +- ⋮ output: Type( +- ⋮ Type::Path { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "i32", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }, +- ⋮ ), +- ⋮ }, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }), +- ⋮ Trait(TraitBound { +- ⋮ modifier: None, +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "Send", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ }, +- ⋮ output: Default, +- ⋮ }, +- ⋮ block: Block, +- ⋮} ++ ItemFn { ++ vis: Inherited, ++ sig: Signature { ++ ident: "f", ++ generics: Generics { ++ lt_token: Some, ++ params: [ ++ Type(TypeParam { ++ ident: "G", ++ }), ++ ], ++ gt_token: Some, ++ where_clause: Some(WhereClause { ++ predicates: [ ++ Type(PredicateType { ++ bounded_ty: Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "G", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ bounds: [ ++ Trait(TraitBound { ++ modifier: None, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "FnOnce", ++ arguments: PathArguments::Parenthesized { ++ output: Type( ++ Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "i32", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ ), ++ }, ++ }, ++ ], ++ }, ++ }), ++ Trait(TraitBound { ++ modifier: None, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "Send", ++ arguments: None, ++ }, ++ ], ++ }, ++ }), ++ ], ++ }), ++ ], ++ }), ++ }, ++ output: Default, ++ }, ++ block: Block, ++ } + "###); + + let where_clause = input.sig.generics.where_clause.as_ref().unwrap(); +@@ -270,7 +265,7 @@ fn test_fn_precedence_in_where_clause() { + assert_eq!(predicate.bounds.len(), 2, "{:#?}", predicate.bounds); + + let first_bound = &predicate.bounds[0]; +- assert_eq!(quote!(#first_bound).to_string(), "FnOnce ( ) -> i32"); ++ assert_eq!(quote!(#first_bound).to_string(), "FnOnce () -> i32"); + + let second_bound = &predicate.bounds[1]; + assert_eq!(quote!(#second_bound).to_string(), "Send"); +diff --git a/third_party/rust/syn/tests/test_grouping.rs b/third_party/rust/syn/tests/test_grouping.rs +index 1558a47b4b..a0fe716390 100644 +--- a/third_party/rust/syn/tests/test_grouping.rs ++++ b/third_party/rust/syn/tests/test_grouping.rs +@@ -1,8 +1,3 @@ +-extern crate proc_macro2; +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -28,31 +23,31 @@ fn test_grouping() { + TokenTree::Literal(Literal::i32_suffixed(4)), + ]); + +- assert_eq!(tokens.to_string(), "1i32 + 2i32 + 3i32 * 4i32"); ++ assert_eq!(tokens.to_string(), "1i32 + 2i32 + 3i32 * 4i32"); + + snapshot!(tokens as Expr, @r###" +- ⋮Expr::Binary { +- ⋮ left: Expr::Lit { +- ⋮ lit: 1i32, +- ⋮ }, +- ⋮ op: Add, +- ⋮ right: Expr::Binary { +- ⋮ left: Expr::Group { +- ⋮ expr: Expr::Binary { +- ⋮ left: Expr::Lit { +- ⋮ lit: 2i32, +- ⋮ }, +- ⋮ op: Add, +- ⋮ right: Expr::Lit { +- ⋮ lit: 3i32, +- ⋮ }, +- ⋮ }, +- ⋮ }, +- ⋮ op: Mul, +- ⋮ right: Expr::Lit { +- ⋮ lit: 4i32, +- ⋮ }, +- ⋮ }, +- ⋮} ++ Expr::Binary { ++ left: Expr::Lit { ++ lit: 1i32, ++ }, ++ op: Add, ++ right: Expr::Binary { ++ left: Expr::Group { ++ expr: Expr::Binary { ++ left: Expr::Lit { ++ lit: 2i32, ++ }, ++ op: Add, ++ right: Expr::Lit { ++ lit: 3i32, ++ }, ++ }, ++ }, ++ op: Mul, ++ right: Expr::Lit { ++ lit: 4i32, ++ }, ++ }, ++ } + "###); + } +diff --git a/third_party/rust/syn/tests/test_ident.rs b/third_party/rust/syn/tests/test_ident.rs +index bec00a70c9..ee01bfcc9f 100644 +--- a/third_party/rust/syn/tests/test_ident.rs ++++ b/third_party/rust/syn/tests/test_ident.rs +@@ -1,8 +1,3 @@ +-extern crate proc_macro2; +-extern crate syn; +- +-mod features; +- + use proc_macro2::{Ident, Span, TokenStream}; + use std::str::FromStr; + use syn::Result; +diff --git a/third_party/rust/syn/tests/test_item.rs b/third_party/rust/syn/tests/test_item.rs +new file mode 100644 +index 0000000000..74ac4baec6 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_item.rs +@@ -0,0 +1,45 @@ ++#[macro_use] ++mod macros; ++ ++use proc_macro2::{Delimiter, Group, Ident, Span, TokenStream, TokenTree}; ++use quote::quote; ++use std::iter::FromIterator; ++use syn::Item; ++ ++#[test] ++fn test_macro_variable_attr() { ++ // mimics the token stream corresponding to `$attr fn f() {}` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { #[test] })), ++ TokenTree::Ident(Ident::new("fn", Span::call_site())), ++ TokenTree::Ident(Ident::new("f", Span::call_site())), ++ TokenTree::Group(Group::new(Delimiter::Parenthesis, TokenStream::new())), ++ TokenTree::Group(Group::new(Delimiter::Brace, TokenStream::new())), ++ ]); ++ ++ snapshot!(tokens as Item, @r###" ++ Item::Fn { ++ attrs: [ ++ Attribute { ++ style: Outer, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "test", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(``), ++ }, ++ ], ++ vis: Inherited, ++ sig: Signature { ++ ident: "f", ++ generics: Generics, ++ output: Default, ++ }, ++ block: Block, ++ } ++ "###); ++} +diff --git a/third_party/rust/syn/tests/test_iterators.rs b/third_party/rust/syn/tests/test_iterators.rs +index 1cf7157e6f..2c8359c157 100644 +--- a/third_party/rust/syn/tests/test_iterators.rs ++++ b/third_party/rust/syn/tests/test_iterators.rs +@@ -1,10 +1,5 @@ + use syn::punctuated::{Pair, Punctuated}; +- +-extern crate quote; +-#[macro_use] +-extern crate syn; +- +-mod features; ++use syn::Token; + + #[macro_use] + mod macros; +diff --git a/third_party/rust/syn/tests/test_lit.rs b/third_party/rust/syn/tests/test_lit.rs +index 1e8f49d19b..e995f2287f 100644 +--- a/third_party/rust/syn/tests/test_lit.rs ++++ b/third_party/rust/syn/tests/test_lit.rs +@@ -1,13 +1,11 @@ +-extern crate proc_macro2; +-extern crate quote; +-extern crate syn; ++#[macro_use] ++mod macros; + +-mod features; +- +-use proc_macro2::{TokenStream, TokenTree}; ++use proc_macro2::{Delimiter, Group, Literal, Span, TokenStream, TokenTree}; + use quote::ToTokens; ++use std::iter::FromIterator; + use std::str::FromStr; +-use syn::Lit; ++use syn::{Lit, LitFloat, LitInt}; + + fn lit(s: &str) -> Lit { + match TokenStream::from_str(s) +@@ -50,6 +48,9 @@ fn strings() { + "contains\nnewlinesescaped newlines", + ); + test_string("r\"raw\nstring\\\nhere\"", "raw\nstring\\\nhere"); ++ test_string("\"...\"q", "..."); ++ test_string("r\"...\"q", "..."); ++ test_string("r##\"...\"##q", "..."); + } + + #[test] +@@ -79,6 +80,9 @@ fn byte_strings() { + b"contains\nnewlinesescaped newlines", + ); + test_byte_string("br\"raw\nstring\\\nhere\"", b"raw\nstring\\\nhere"); ++ test_byte_string("b\"...\"q", b"..."); ++ test_byte_string("br\"...\"q", b"..."); ++ test_byte_string("br##\"...\"##q", b"..."); + } + + #[test] +@@ -100,6 +104,7 @@ fn bytes() { + test_byte("b'\\t'", b'\t'); + test_byte("b'\\''", b'\''); + test_byte("b'\"'", b'"'); ++ test_byte("b'a'q", b'a'); + } + + #[test] +@@ -125,6 +130,7 @@ fn chars() { + test_char("'\\''", '\''); + test_char("'\"'", '"'); + test_char("'\\u{1F415}'", '\u{1F415}'); ++ test_char("'a'q", 'a'); + } + + #[test] +@@ -185,4 +191,59 @@ fn floats() { + test_float("5.5e12", 5.5e12, ""); + test_float("1.0__3e-12", 1.03e-12, ""); + test_float("1.03e+12", 1.03e12, ""); ++ test_float("9e99e99", 9e99, "e99"); ++} ++ ++#[test] ++fn negative() { ++ let span = Span::call_site(); ++ assert_eq!("-1", LitInt::new("-1", span).to_string()); ++ assert_eq!("-1i8", LitInt::new("-1i8", span).to_string()); ++ assert_eq!("-1i16", LitInt::new("-1i16", span).to_string()); ++ assert_eq!("-1i32", LitInt::new("-1i32", span).to_string()); ++ assert_eq!("-1i64", LitInt::new("-1i64", span).to_string()); ++ assert_eq!("-1.5", LitFloat::new("-1.5", span).to_string()); ++ assert_eq!("-1.5f32", LitFloat::new("-1.5f32", span).to_string()); ++ assert_eq!("-1.5f64", LitFloat::new("-1.5f64", span).to_string()); ++} ++ ++#[test] ++fn suffix() { ++ fn get_suffix(token: &str) -> String { ++ let lit = syn::parse_str::<Lit>(token).unwrap(); ++ match lit { ++ Lit::Str(lit) => lit.suffix().to_owned(), ++ Lit::ByteStr(lit) => lit.suffix().to_owned(), ++ Lit::Byte(lit) => lit.suffix().to_owned(), ++ Lit::Char(lit) => lit.suffix().to_owned(), ++ Lit::Int(lit) => lit.suffix().to_owned(), ++ Lit::Float(lit) => lit.suffix().to_owned(), ++ _ => unimplemented!(), ++ } ++ } ++ ++ assert_eq!(get_suffix("\"\"s"), "s"); ++ assert_eq!(get_suffix("r\"\"r"), "r"); ++ assert_eq!(get_suffix("b\"\"b"), "b"); ++ assert_eq!(get_suffix("br\"\"br"), "br"); ++ assert_eq!(get_suffix("r#\"\"#r"), "r"); ++ assert_eq!(get_suffix("'c'c"), "c"); ++ assert_eq!(get_suffix("b'b'b"), "b"); ++ assert_eq!(get_suffix("1i32"), "i32"); ++ assert_eq!(get_suffix("1_i32"), "i32"); ++ assert_eq!(get_suffix("1.0f32"), "f32"); ++ assert_eq!(get_suffix("1.0_f32"), "f32"); ++} ++ ++#[test] ++fn test_deep_group_empty() { ++ let tokens = TokenStream::from_iter(vec![TokenTree::Group(Group::new( ++ Delimiter::None, ++ TokenStream::from_iter(vec![TokenTree::Group(Group::new( ++ Delimiter::None, ++ TokenStream::from_iter(vec![TokenTree::Literal(Literal::string("hi"))]), ++ ))]), ++ ))]); ++ ++ snapshot!(tokens as Lit, @r#""hi""# ); + } +diff --git a/third_party/rust/syn/tests/test_meta.rs b/third_party/rust/syn/tests/test_meta.rs +index 547472d6f4..d37dda948a 100644 +--- a/third_party/rust/syn/tests/test_meta.rs ++++ b/third_party/rust/syn/tests/test_meta.rs +@@ -1,7 +1,3 @@ +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -12,14 +8,14 @@ fn test_parse_meta_item_word() { + let input = "hello"; + + snapshot!(input as Meta, @r###" +- ⋮Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "hello", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮}) ++ Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "hello", ++ arguments: None, ++ }, ++ ], ++ }) + "###); + } + +@@ -29,31 +25,31 @@ fn test_parse_meta_name_value() { + let (inner, meta) = (input, input); + + snapshot!(inner as MetaNameValue, @r###" +- ⋮MetaNameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ MetaNameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + snapshot!(meta as Meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + assert_eq!(meta, inner.into()); +@@ -65,31 +61,31 @@ fn test_parse_meta_name_value_with_keyword() { + let (inner, meta) = (input, input); + + snapshot!(inner as MetaNameValue, @r###" +- ⋮MetaNameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "static", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ MetaNameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "static", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + snapshot!(meta as Meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "static", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "static", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + assert_eq!(meta, inner.into()); +@@ -101,31 +97,31 @@ fn test_parse_meta_name_value_with_bool() { + let (inner, meta) = (input, input); + + snapshot!(inner as MetaNameValue, @r###" +- ⋮MetaNameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "true", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ MetaNameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "true", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + snapshot!(meta as Meta, @r###" +- ⋮Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "true", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮} ++ Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "true", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ } + "###); + + assert_eq!(meta, inner.into()); +@@ -137,35 +133,35 @@ fn test_parse_meta_item_list_lit() { + let (inner, meta) = (input, input); + + snapshot!(inner as MetaList, @r###" +- ⋮MetaList { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Lit(5), +- ⋮ ], +- ⋮} ++ MetaList { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Lit(5), ++ ], ++ } + "###); + + snapshot!(meta as Meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Lit(5), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Lit(5), ++ ], ++ } + "###); + + assert_eq!(meta, inner.into()); +@@ -177,133 +173,133 @@ fn test_parse_meta_item_multiple() { + let (inner, meta) = (input, input); + + snapshot!(inner as MetaList, @r###" +- ⋮MetaList { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮ }), +- ⋮ Meta(Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "list", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 6, +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ ], +- ⋮} ++ MetaList { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word", ++ arguments: None, ++ }, ++ ], ++ })), ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ }), ++ Meta(Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "list", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name2", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 6, ++ }), ++ ], ++ }), ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word2", ++ arguments: None, ++ }, ++ ], ++ })), ++ ], ++ } + "###); + + snapshot!(meta as Meta, @r###" +- ⋮Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "foo", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 5, +- ⋮ }), +- ⋮ Meta(Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "list", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 6, +- ⋮ }), +- ⋮ ], +- ⋮ }), +- ⋮ Meta(Path(Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "word2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ })), +- ⋮ ], +- ⋮} ++ Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "foo", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word", ++ arguments: None, ++ }, ++ ], ++ })), ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 5, ++ }), ++ Meta(Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "list", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name2", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 6, ++ }), ++ ], ++ }), ++ Meta(Path(Path { ++ segments: [ ++ PathSegment { ++ ident: "word2", ++ arguments: None, ++ }, ++ ], ++ })), ++ ], ++ } + "###); + + assert_eq!(meta, inner.into()); +@@ -316,28 +312,28 @@ fn test_parse_nested_meta() { + + let input = "list(name2 = 6)"; + snapshot!(input as NestedMeta, @r###" +- ⋮Meta(Meta::List { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "list", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ nested: [ +- ⋮ Meta(Meta::NameValue { +- ⋮ path: Path { +- ⋮ segments: [ +- ⋮ PathSegment { +- ⋮ ident: "name2", +- ⋮ arguments: None, +- ⋮ }, +- ⋮ ], +- ⋮ }, +- ⋮ lit: 6, +- ⋮ }), +- ⋮ ], +- ⋮}) ++ Meta(Meta::List { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "list", ++ arguments: None, ++ }, ++ ], ++ }, ++ nested: [ ++ Meta(Meta::NameValue { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "name2", ++ arguments: None, ++ }, ++ ], ++ }, ++ lit: 6, ++ }), ++ ], ++ }) + "###); + } +diff --git a/third_party/rust/syn/tests/test_parse_buffer.rs b/third_party/rust/syn/tests/test_parse_buffer.rs +index f09495187f..57a3c7c38c 100644 +--- a/third_party/rust/syn/tests/test_parse_buffer.rs ++++ b/third_party/rust/syn/tests/test_parse_buffer.rs +@@ -1,7 +1,7 @@ +-#[macro_use] +-extern crate syn; +- ++use proc_macro2::{Delimiter, Group, Punct, Spacing, TokenStream, TokenTree}; ++use std::iter::FromIterator; + use syn::parse::{discouraged::Speculative, Parse, ParseStream, Parser, Result}; ++use syn::{parenthesized, Token}; + + #[test] + #[should_panic(expected = "Fork was not derived from the advancing parse stream")] +@@ -53,3 +53,38 @@ fn smuggled_speculative_cursor_into_brackets() { + + syn::parse_str::<BreakRules>("()").unwrap(); + } ++ ++#[test] ++fn trailing_empty_none_group() { ++ fn parse(input: ParseStream) -> Result<()> { ++ input.parse::<Token![+]>()?; ++ ++ let content; ++ parenthesized!(content in input); ++ content.parse::<Token![+]>()?; ++ ++ Ok(()) ++ } ++ ++ // `+ ( + <Ø Ø> ) <Ø <Ø Ø> Ø>` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Punct(Punct::new('+', Spacing::Alone)), ++ TokenTree::Group(Group::new( ++ Delimiter::Parenthesis, ++ TokenStream::from_iter(vec![ ++ TokenTree::Punct(Punct::new('+', Spacing::Alone)), ++ TokenTree::Group(Group::new(Delimiter::None, TokenStream::new())), ++ ]), ++ )), ++ TokenTree::Group(Group::new(Delimiter::None, TokenStream::new())), ++ TokenTree::Group(Group::new( ++ Delimiter::None, ++ TokenStream::from_iter(vec![TokenTree::Group(Group::new( ++ Delimiter::None, ++ TokenStream::new(), ++ ))]), ++ )), ++ ]); ++ ++ parse.parse2(tokens).unwrap(); ++} +diff --git a/third_party/rust/syn/tests/test_parse_stream.rs b/third_party/rust/syn/tests/test_parse_stream.rs +new file mode 100644 +index 0000000000..76bd065777 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_parse_stream.rs +@@ -0,0 +1,12 @@ ++use syn::ext::IdentExt; ++use syn::parse::ParseStream; ++use syn::{Ident, Token}; ++ ++#[test] ++fn test_peek() { ++ let _ = |input: ParseStream| { ++ let _ = input.peek(Ident); ++ let _ = input.peek(Ident::peek_any); ++ let _ = input.peek(Token![::]); ++ }; ++} +diff --git a/third_party/rust/syn/tests/test_pat.rs b/third_party/rust/syn/tests/test_pat.rs +index 1343aa646f..73388dd79d 100644 +--- a/third_party/rust/syn/tests/test_pat.rs ++++ b/third_party/rust/syn/tests/test_pat.rs +@@ -1,10 +1,5 @@ +-extern crate quote; +-extern crate syn; +- +-mod features; +- + use quote::quote; +-use syn::Pat; ++use syn::{Item, Pat, Stmt}; + + #[test] + fn test_pat_ident() { +@@ -21,3 +16,23 @@ fn test_pat_path() { + value => panic!("expected PatPath, got {:?}", value), + } + } ++ ++#[test] ++fn test_leading_vert() { ++ // https://github.com/rust-lang/rust/blob/1.43.0/src/test/ui/or-patterns/remove-leading-vert.rs ++ ++ syn::parse_str::<Item>("fn f() {}").unwrap(); ++ syn::parse_str::<Item>("fn fun1(| A: E) {}").unwrap_err(); ++ syn::parse_str::<Item>("fn fun2(|| A: E) {}").unwrap_err(); ++ ++ syn::parse_str::<Stmt>("let | () = ();").unwrap(); ++ syn::parse_str::<Stmt>("let (| A): E;").unwrap_err(); ++ syn::parse_str::<Stmt>("let (|| A): (E);").unwrap_err(); ++ syn::parse_str::<Stmt>("let (| A,): (E,);").unwrap_err(); ++ syn::parse_str::<Stmt>("let [| A]: [E; 1];").unwrap_err(); ++ syn::parse_str::<Stmt>("let [|| A]: [E; 1];").unwrap_err(); ++ syn::parse_str::<Stmt>("let TS(| A): TS;").unwrap_err(); ++ syn::parse_str::<Stmt>("let TS(|| A): TS;").unwrap_err(); ++ syn::parse_str::<Stmt>("let NS { f: | A }: NS;").unwrap_err(); ++ syn::parse_str::<Stmt>("let NS { f: || A }: NS;").unwrap_err(); ++} +diff --git a/third_party/rust/syn/tests/test_path.rs b/third_party/rust/syn/tests/test_path.rs +new file mode 100644 +index 0000000000..2ce12066f5 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_path.rs +@@ -0,0 +1,52 @@ ++#[macro_use] ++mod macros; ++ ++use proc_macro2::{Delimiter, Group, Ident, Punct, Spacing, Span, TokenStream, TokenTree}; ++use quote::quote; ++use std::iter::FromIterator; ++use syn::{Expr, Type}; ++ ++#[test] ++fn parse_interpolated_leading_component() { ++ // mimics the token stream corresponding to `$mod::rest` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { first })), ++ TokenTree::Punct(Punct::new(':', Spacing::Joint)), ++ TokenTree::Punct(Punct::new(':', Spacing::Alone)), ++ TokenTree::Ident(Ident::new("rest", Span::call_site())), ++ ]); ++ ++ snapshot!(tokens.clone() as Expr, @r###" ++ Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "first", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "rest", ++ arguments: None, ++ }, ++ ], ++ }, ++ } ++ "###); ++ ++ snapshot!(tokens as Type, @r###" ++ Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "first", ++ arguments: None, ++ }, ++ PathSegment { ++ ident: "rest", ++ arguments: None, ++ }, ++ ], ++ }, ++ } ++ "###); ++} +diff --git a/third_party/rust/syn/tests/test_precedence.rs b/third_party/rust/syn/tests/test_precedence.rs +index 53ee66e372..a586b3fe48 100644 +--- a/third_party/rust/syn/tests/test_precedence.rs ++++ b/third_party/rust/syn/tests/test_precedence.rs +@@ -4,35 +4,26 @@ + + //! The tests in this module do the following: + //! +-//! 1. Parse a given expression in both `syn` and `libsyntax`. ++//! 1. Parse a given expression in both `syn` and `librustc`. + //! 2. Fold over the expression adding brackets around each subexpression (with +-//! some complications - see the `syn_brackets` and `libsyntax_brackets` ++//! some complications - see the `syn_brackets` and `librustc_brackets` + //! methods). + //! 3. Serialize the `syn` expression back into a string, and re-parse it with +-//! `libsyntax`. ++//! `librustc`. + //! 4. Respan all of the expressions, replacing the spans with the default + //! spans. + //! 5. Compare the expressions with one another, if they are not equal fail. + +-extern crate quote; +-extern crate rayon; +-extern crate regex; ++extern crate rustc_ast; + extern crate rustc_data_structures; +-extern crate smallvec; +-extern crate syn; +-extern crate syntax; +-extern crate syntax_pos; +-extern crate walkdir; +- +-mod features; ++extern crate rustc_span; + + use quote::quote; + use rayon::iter::{IntoParallelIterator, ParallelIterator}; + use regex::Regex; +-use smallvec::smallvec; +-use syntax::ast; +-use syntax::ptr::P; +-use syntax_pos::edition::Edition; ++use rustc_ast::ast; ++use rustc_ast::ptr::P; ++use rustc_span::edition::Edition; + use walkdir::{DirEntry, WalkDir}; + + use std::fs::File; +@@ -73,7 +64,7 @@ fn test_simple_precedence() { + continue; + }; + +- let pf = match test_expressions(vec![expr]) { ++ let pf = match test_expressions(Edition::Edition2018, vec![expr]) { + (1, 0) => "passed", + (0, 1) => { + failed += 1; +@@ -91,8 +82,8 @@ fn test_simple_precedence() { + + /// Test expressions from rustc, like in `test_round_trip`. + #[test] +-#[cfg_attr(target_os = "windows", ignore = "requires nix .sh")] + fn test_rustc_precedence() { ++ common::rayon_init(); + repo::clone_rust(); + let abort_after = common::abort_after(); + if abort_after == 0 { +@@ -118,15 +109,6 @@ fn test_rustc_precedence() { + return; + } + +- // Our version of `libsyntax` can't parse this tests +- if path +- .to_str() +- .unwrap() +- .ends_with("optional_comma_in_match_arm.rs") +- { +- return; +- } +- + let mut file = File::open(path).unwrap(); + let mut content = String::new(); + file.read_to_string(&mut content).unwrap(); +@@ -134,8 +116,9 @@ fn test_rustc_precedence() { + + let (l_passed, l_failed) = match syn::parse_file(&content) { + Ok(file) => { ++ let edition = repo::edition(path).parse().unwrap(); + let exprs = collect_exprs(file); +- test_expressions(exprs) ++ test_expressions(edition, exprs) + } + Err(msg) => { + errorf!("syn failed to parse\n{:?}\n", msg); +@@ -169,36 +152,36 @@ fn test_rustc_precedence() { + } + } + +-fn test_expressions(exprs: Vec<syn::Expr>) -> (usize, usize) { ++fn test_expressions(edition: Edition, exprs: Vec<syn::Expr>) -> (usize, usize) { + let mut passed = 0; + let mut failed = 0; + +- syntax::with_globals(Edition::Edition2018, || { ++ rustc_span::with_session_globals(edition, || { + for expr in exprs { + let raw = quote!(#expr).to_string(); + +- let libsyntax_ast = if let Some(e) = libsyntax_parse_and_rewrite(&raw) { ++ let librustc_ast = if let Some(e) = librustc_parse_and_rewrite(&raw) { + e + } else { + failed += 1; +- errorf!("\nFAIL - libsyntax failed to parse raw\n"); ++ errorf!("\nFAIL - librustc failed to parse raw\n"); + continue; + }; + + let syn_expr = syn_brackets(expr); +- let syn_ast = if let Some(e) = parse::libsyntax_expr("e!(#syn_expr).to_string()) { ++ let syn_ast = if let Some(e) = parse::librustc_expr("e!(#syn_expr).to_string()) { + e + } else { + failed += 1; +- errorf!("\nFAIL - libsyntax failed to parse bracketed\n"); ++ errorf!("\nFAIL - librustc failed to parse bracketed\n"); + continue; + }; + +- if SpanlessEq::eq(&syn_ast, &libsyntax_ast) { ++ if SpanlessEq::eq(&syn_ast, &librustc_ast) { + passed += 1; + } else { + failed += 1; +- errorf!("\nFAIL\n{:?}\n!=\n{:?}\n", syn_ast, libsyntax_ast); ++ errorf!("\nFAIL\n{:?}\n!=\n{:?}\n", syn_ast, librustc_ast); + } + } + }); +@@ -206,54 +189,106 @@ fn test_expressions(exprs: Vec<syn::Expr>) -> (usize, usize) { + (passed, failed) + } + +-fn libsyntax_parse_and_rewrite(input: &str) -> Option<P<ast::Expr>> { +- parse::libsyntax_expr(input).and_then(libsyntax_brackets) ++fn librustc_parse_and_rewrite(input: &str) -> Option<P<ast::Expr>> { ++ parse::librustc_expr(input).and_then(librustc_brackets) + } + + /// Wrap every expression which is not already wrapped in parens with parens, to + /// reveal the precidence of the parsed expressions, and produce a stringified + /// form of the resulting expression. + /// +-/// This method operates on libsyntax objects. +-fn libsyntax_brackets(mut libsyntax_expr: P<ast::Expr>) -> Option<P<ast::Expr>> { ++/// This method operates on librustc objects. ++fn librustc_brackets(mut librustc_expr: P<ast::Expr>) -> Option<P<ast::Expr>> { ++ use rustc_ast::ast::{ ++ Block, BorrowKind, Expr, ExprKind, Field, GenericArg, MacCall, Pat, Stmt, StmtKind, Ty, ++ }; ++ use rustc_ast::mut_visit::{noop_visit_generic_arg, MutVisitor}; ++ use rustc_data_structures::map_in_place::MapInPlace; + use rustc_data_structures::thin_vec::ThinVec; +- use smallvec::SmallVec; ++ use rustc_span::DUMMY_SP; + use std::mem; +- use syntax::ast::{Expr, ExprKind, Field, Mac, Pat, Stmt, StmtKind, Ty}; +- use syntax::mut_visit::{noop_visit_expr, MutVisitor}; +- use syntax_pos::DUMMY_SP; + + struct BracketsVisitor { + failed: bool, + }; + ++ fn flat_map_field<T: MutVisitor>(mut f: Field, vis: &mut T) -> Vec<Field> { ++ if f.is_shorthand { ++ noop_visit_expr(&mut f.expr, vis); ++ } else { ++ vis.visit_expr(&mut f.expr); ++ } ++ vec![f] ++ } ++ ++ fn flat_map_stmt<T: MutVisitor>(stmt: Stmt, vis: &mut T) -> Vec<Stmt> { ++ let kind = match stmt.kind { ++ // Don't wrap toplevel expressions in statements. ++ StmtKind::Expr(mut e) => { ++ noop_visit_expr(&mut e, vis); ++ StmtKind::Expr(e) ++ } ++ StmtKind::Semi(mut e) => { ++ noop_visit_expr(&mut e, vis); ++ StmtKind::Semi(e) ++ } ++ s => s, ++ }; ++ ++ vec![Stmt { kind, ..stmt }] ++ } ++ ++ fn noop_visit_expr<T: MutVisitor>(e: &mut Expr, vis: &mut T) { ++ use rustc_ast::mut_visit::{noop_visit_expr, visit_opt, visit_thin_attrs}; ++ match &mut e.kind { ++ ExprKind::AddrOf(BorrowKind::Raw, ..) => {} ++ ExprKind::Struct(path, fields, expr) => { ++ vis.visit_path(path); ++ fields.flat_map_in_place(|field| flat_map_field(field, vis)); ++ visit_opt(expr, |expr| vis.visit_expr(expr)); ++ vis.visit_id(&mut e.id); ++ vis.visit_span(&mut e.span); ++ visit_thin_attrs(&mut e.attrs, vis); ++ } ++ _ => noop_visit_expr(e, vis), ++ } ++ } ++ + impl MutVisitor for BracketsVisitor { + fn visit_expr(&mut self, e: &mut P<Expr>) { + noop_visit_expr(e, self); +- match e.node { ++ match e.kind { + ExprKind::If(..) | ExprKind::Block(..) | ExprKind::Let(..) => {} + _ => { + let inner = mem::replace( + e, + P(Expr { + id: ast::DUMMY_NODE_ID, +- node: ExprKind::Err, ++ kind: ExprKind::Err, + span: DUMMY_SP, + attrs: ThinVec::new(), ++ tokens: None, + }), + ); +- e.node = ExprKind::Paren(inner); ++ e.kind = ExprKind::Paren(inner); + } + } + } + +- fn flat_map_field(&mut self, mut f: Field) -> SmallVec<[Field; 1]> { +- if f.is_shorthand { +- noop_visit_expr(&mut f.expr, self); +- } else { +- self.visit_expr(&mut f.expr); ++ fn visit_generic_arg(&mut self, arg: &mut GenericArg) { ++ match arg { ++ // Don't wrap const generic arg as that's invalid syntax. ++ GenericArg::Const(arg) => noop_visit_expr(&mut arg.value, self), ++ _ => noop_visit_generic_arg(arg, self), + } +- SmallVec::from([f]) ++ } ++ ++ fn visit_block(&mut self, block: &mut P<Block>) { ++ self.visit_id(&mut block.id); ++ block ++ .stmts ++ .flat_map_in_place(|stmt| flat_map_stmt(stmt, self)); ++ self.visit_span(&mut block.span); + } + + // We don't want to look at expressions that might appear in patterns or +@@ -267,25 +302,8 @@ fn libsyntax_brackets(mut libsyntax_expr: P<ast::Expr>) -> Option<P<ast::Expr>> + let _ = ty; + } + +- fn flat_map_stmt(&mut self, stmt: Stmt) -> SmallVec<[Stmt; 1]> { +- let node = match stmt.node { +- // Don't wrap toplevel expressions in statements. +- StmtKind::Expr(mut e) => { +- noop_visit_expr(&mut e, self); +- StmtKind::Expr(e) +- } +- StmtKind::Semi(mut e) => { +- noop_visit_expr(&mut e, self); +- StmtKind::Semi(e) +- } +- s => s, +- }; +- +- smallvec![Stmt { node, ..stmt }] +- } +- +- fn visit_mac(&mut self, mac: &mut Mac) { +- // By default when folding over macros, libsyntax panics. This is ++ fn visit_mac(&mut self, mac: &mut MacCall) { ++ // By default when folding over macros, librustc panics. This is + // because it's usually not what you want, you want to run after + // macro expansion. We do want to do that (syn doesn't do macro + // expansion), so we implement visit_mac to just return the macro +@@ -295,11 +313,11 @@ fn libsyntax_brackets(mut libsyntax_expr: P<ast::Expr>) -> Option<P<ast::Expr>> + } + + let mut folder = BracketsVisitor { failed: false }; +- folder.visit_expr(&mut libsyntax_expr); ++ folder.visit_expr(&mut librustc_expr); + if folder.failed { + None + } else { +- Some(libsyntax_expr) ++ Some(librustc_expr) + } + } + +@@ -318,14 +336,33 @@ fn syn_brackets(syn_expr: syn::Expr) -> syn::Expr { + Expr::If(..) | Expr::Unsafe(..) | Expr::Block(..) | Expr::Let(..) => { + fold_expr(self, expr) + } +- node => Expr::Paren(ExprParen { ++ _ => Expr::Paren(ExprParen { + attrs: Vec::new(), +- expr: Box::new(fold_expr(self, node)), ++ expr: Box::new(fold_expr(self, expr)), + paren_token: token::Paren::default(), + }), + } + } + ++ fn fold_generic_argument(&mut self, arg: GenericArgument) -> GenericArgument { ++ match arg { ++ // Don't wrap const generic arg as that's invalid syntax. ++ GenericArgument::Const(a) => GenericArgument::Const(fold_expr(self, a)), ++ _ => fold_generic_argument(self, arg), ++ } ++ } ++ ++ fn fold_generic_method_argument( ++ &mut self, ++ arg: GenericMethodArgument, ++ ) -> GenericMethodArgument { ++ match arg { ++ // Don't wrap const generic arg as that's invalid syntax. ++ GenericMethodArgument::Const(a) => GenericMethodArgument::Const(fold_expr(self, a)), ++ _ => fold_generic_method_argument(self, arg), ++ } ++ } ++ + fn fold_stmt(&mut self, stmt: Stmt) -> Stmt { + match stmt { + // Don't wrap toplevel expressions in statements. +@@ -360,7 +397,10 @@ fn collect_exprs(file: syn::File) -> Vec<syn::Expr> { + struct CollectExprs(Vec<Expr>); + impl Fold for CollectExprs { + fn fold_expr(&mut self, expr: Expr) -> Expr { +- self.0.push(expr); ++ match expr { ++ Expr::Verbatim(tokens) if tokens.is_empty() => {} ++ _ => self.0.push(expr), ++ } + + Expr::Tuple(ExprTuple { + attrs: vec![], +diff --git a/third_party/rust/syn/tests/test_receiver.rs b/third_party/rust/syn/tests/test_receiver.rs +new file mode 100644 +index 0000000000..923df96ba9 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_receiver.rs +@@ -0,0 +1,127 @@ ++use syn::{parse_quote, FnArg, Receiver, TraitItemMethod}; ++ ++#[test] ++fn test_by_value() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn by_value(self: Self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_by_mut_value() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn by_mut(mut self: Self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_by_ref() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn by_ref(self: &Self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_by_box() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn by_box(self: Box<Self>); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_by_pin() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn by_pin(self: Pin<Self>); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_explicit_type() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn explicit_type(self: Pin<MyType>); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Typed(_)) => (), ++ value => panic!("expected FnArg::Typed, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_value_shorthand() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn value_shorthand(self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Receiver(Receiver { ++ reference: None, ++ mutability: None, ++ .. ++ })) => (), ++ value => panic!("expected FnArg::Receiver without ref/mut, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_mut_value_shorthand() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn mut_value_shorthand(mut self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Receiver(Receiver { ++ reference: None, ++ mutability: Some(_), ++ .. ++ })) => (), ++ value => panic!("expected FnArg::Receiver with mut, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_ref_shorthand() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn ref_shorthand(&self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Receiver(Receiver { ++ reference: Some(_), ++ mutability: None, ++ .. ++ })) => (), ++ value => panic!("expected FnArg::Receiver with ref, got {:?}", value), ++ } ++} ++ ++#[test] ++fn test_ref_mut_shorthand() { ++ let TraitItemMethod { sig, .. } = parse_quote! { ++ fn ref_mut_shorthand(&mut self); ++ }; ++ match sig.receiver() { ++ Some(FnArg::Receiver(Receiver { ++ reference: Some(_), ++ mutability: Some(_), ++ .. ++ })) => (), ++ value => panic!("expected FnArg::Receiver with ref+mut, got {:?}", value), ++ } ++} +diff --git a/third_party/rust/syn/tests/test_round_trip.rs b/third_party/rust/syn/tests/test_round_trip.rs +index 2fc9cecd86..260dd0c3d9 100644 +--- a/third_party/rust/syn/tests/test_round_trip.rs ++++ b/third_party/rust/syn/tests/test_round_trip.rs +@@ -2,22 +2,20 @@ + #![recursion_limit = "1024"] + #![feature(rustc_private)] + +-extern crate quote; +-extern crate rayon; +-extern crate syn; +-extern crate syntax; +-extern crate syntax_pos; +-extern crate walkdir; +- +-mod features; ++extern crate rustc_ast; ++extern crate rustc_errors; ++extern crate rustc_expand; ++extern crate rustc_parse as parse; ++extern crate rustc_session; ++extern crate rustc_span; + + use quote::quote; + use rayon::iter::{IntoParallelIterator, ParallelIterator}; +-use syntax::ast; +-use syntax::parse::{self, PResult, ParseSess}; +-use syntax::source_map::FilePathMapping; +-use syntax_pos::edition::Edition; +-use syntax_pos::FileName; ++use rustc_ast::ast; ++use rustc_errors::PResult; ++use rustc_session::parse::ParseSess; ++use rustc_span::source_map::FilePathMapping; ++use rustc_span::FileName; + use walkdir::{DirEntry, WalkDir}; + + use std::fs::File; +@@ -38,8 +36,8 @@ mod repo; + use common::eq::SpanlessEq; + + #[test] +-#[cfg_attr(target_os = "windows", ignore = "requires nix .sh")] + fn test_round_trip() { ++ common::rayon_init(); + repo::clone_rust(); + let abort_after = common::abort_after(); + if abort_after == 0 { +@@ -78,11 +76,12 @@ fn test_round_trip() { + } + }; + let back = quote!(#krate).to_string(); ++ let edition = repo::edition(path).parse().unwrap(); + + let equal = panic::catch_unwind(|| { +- syntax::with_globals(Edition::Edition2018, || { ++ rustc_span::with_session_globals(edition, || { + let sess = ParseSess::new(FilePathMapping::empty()); +- let before = match libsyntax_parse(content, &sess) { ++ let before = match librustc_parse(content, &sess) { + Ok(before) => before, + Err(mut diagnostic) => { + diagnostic.cancel(); +@@ -93,7 +92,7 @@ fn test_round_trip() { + errorf!("=== {}: ignore\n", path.display()); + } else { + errorf!( +- "=== {}: ignore - libsyntax failed to parse original content: {}\n", ++ "=== {}: ignore - librustc failed to parse original content: {}\n", + path.display(), + diagnostic.message() + ); +@@ -101,10 +100,10 @@ fn test_round_trip() { + return true; + } + }; +- let after = match libsyntax_parse(back, &sess) { ++ let after = match librustc_parse(back, &sess) { + Ok(after) => after, + Err(mut diagnostic) => { +- errorf!("=== {}: libsyntax failed to parse", path.display()); ++ errorf!("=== {}: librustc failed to parse", path.display()); + diagnostic.emit(); + return false; + } +@@ -130,7 +129,7 @@ fn test_round_trip() { + }) + }); + match equal { +- Err(_) => errorf!("=== {}: ignoring libsyntax panic\n", path.display()), ++ Err(_) => errorf!("=== {}: ignoring librustc panic\n", path.display()), + Ok(true) => {} + Ok(false) => { + let prev_failed = failed.fetch_add(1, Ordering::SeqCst); +@@ -147,7 +146,7 @@ fn test_round_trip() { + } + } + +-fn libsyntax_parse(content: String, sess: &ParseSess) -> PResult<ast::Crate> { ++fn librustc_parse(content: String, sess: &ParseSess) -> PResult<ast::Crate> { + let name = FileName::Custom("test_round_trip".to_string()); + parse::parse_crate_from_source_str(name, content, sess) + } +diff --git a/third_party/rust/syn/tests/test_shebang.rs b/third_party/rust/syn/tests/test_shebang.rs +new file mode 100644 +index 0000000000..dc26b9aab3 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_shebang.rs +@@ -0,0 +1,59 @@ ++#[macro_use] ++mod macros; ++ ++#[test] ++fn test_basic() { ++ let content = "#!/usr/bin/env rustx\nfn main() {}"; ++ let file = syn::parse_file(content).unwrap(); ++ snapshot!(file, @r###" ++ File { ++ shebang: Some("#!/usr/bin/env rustx"), ++ items: [ ++ Item::Fn { ++ vis: Inherited, ++ sig: Signature { ++ ident: "main", ++ generics: Generics, ++ output: Default, ++ }, ++ block: Block, ++ }, ++ ], ++ } ++ "###); ++} ++ ++#[test] ++fn test_comment() { ++ let content = "#!//am/i/a/comment\n[allow(dead_code)] fn main() {}"; ++ let file = syn::parse_file(content).unwrap(); ++ snapshot!(file, @r###" ++ File { ++ attrs: [ ++ Attribute { ++ style: Inner, ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "allow", ++ arguments: None, ++ }, ++ ], ++ }, ++ tokens: TokenStream(`(dead_code)`), ++ }, ++ ], ++ items: [ ++ Item::Fn { ++ vis: Inherited, ++ sig: Signature { ++ ident: "main", ++ generics: Generics, ++ output: Default, ++ }, ++ block: Block, ++ }, ++ ], ++ } ++ "###); ++} +diff --git a/third_party/rust/syn/tests/test_should_parse.rs b/third_party/rust/syn/tests/test_should_parse.rs +index aadf42e3af..180d859916 100644 +--- a/third_party/rust/syn/tests/test_should_parse.rs ++++ b/third_party/rust/syn/tests/test_should_parse.rs +@@ -1,7 +1,3 @@ +-extern crate syn; +- +-mod features; +- + macro_rules! should_parse { + ($name:ident, { $($in:tt)* }) => { + #[test] +diff --git a/third_party/rust/syn/tests/test_size.rs b/third_party/rust/syn/tests/test_size.rs +index 386d4df889..01e8401158 100644 +--- a/third_party/rust/syn/tests/test_size.rs ++++ b/third_party/rust/syn/tests/test_size.rs +@@ -1,7 +1,5 @@ + #![cfg(target_pointer_width = "64")] + +-mod features; +- + use std::mem; + use syn::*; + +diff --git a/third_party/rust/syn/tests/test_stmt.rs b/third_party/rust/syn/tests/test_stmt.rs +new file mode 100644 +index 0000000000..d68b47fd2f +--- /dev/null ++++ b/third_party/rust/syn/tests/test_stmt.rs +@@ -0,0 +1,44 @@ ++#[macro_use] ++mod macros; ++ ++use syn::Stmt; ++ ++#[test] ++fn test_raw_operator() { ++ let stmt = syn::parse_str::<Stmt>("let _ = &raw const x;").unwrap(); ++ ++ snapshot!(stmt, @r###" ++ Local(Local { ++ pat: Pat::Wild, ++ init: Some(Verbatim(`& raw const x`)), ++ }) ++ "###); ++} ++ ++#[test] ++fn test_raw_variable() { ++ let stmt = syn::parse_str::<Stmt>("let _ = &raw;").unwrap(); ++ ++ snapshot!(stmt, @r###" ++ Local(Local { ++ pat: Pat::Wild, ++ init: Some(Expr::Reference { ++ expr: Expr::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "raw", ++ arguments: None, ++ }, ++ ], ++ }, ++ }, ++ }), ++ }) ++ "###); ++} ++ ++#[test] ++fn test_raw_invalid() { ++ assert!(syn::parse_str::<Stmt>("let _ = &raw x;").is_err()); ++} +diff --git a/third_party/rust/syn/tests/test_token_trees.rs b/third_party/rust/syn/tests/test_token_trees.rs +index 70a9a72aab..5b00448af8 100644 +--- a/third_party/rust/syn/tests/test_token_trees.rs ++++ b/third_party/rust/syn/tests/test_token_trees.rs +@@ -1,9 +1,3 @@ +-extern crate proc_macro2; +-extern crate quote; +-extern crate syn; +- +-mod features; +- + #[macro_use] + mod macros; + +@@ -21,7 +15,11 @@ fn test_struct() { + } + "; + +- snapshot!(input as TokenStream, @"`# [ derive ( Debug , Clone ) ] pub struct Item { pub ident : Ident , pub attrs : Vec < Attribute >, }`"); ++ snapshot!(input as TokenStream, @r###" ++ TokenStream( ++ `# [derive (Debug , Clone)] pub struct Item { pub ident : Ident , pub attrs : Vec < Attribute >, }`, ++ ) ++ "###); + } + + #[test] +diff --git a/third_party/rust/syn/tests/test_ty.rs b/third_party/rust/syn/tests/test_ty.rs +new file mode 100644 +index 0000000000..9cbdcd6b99 +--- /dev/null ++++ b/third_party/rust/syn/tests/test_ty.rs +@@ -0,0 +1,53 @@ ++#[macro_use] ++mod macros; ++ ++use proc_macro2::{Delimiter, Group, Ident, Punct, Spacing, Span, TokenStream, TokenTree}; ++use quote::quote; ++use std::iter::FromIterator; ++use syn::Type; ++ ++#[test] ++fn test_mut_self() { ++ syn::parse_str::<Type>("fn(mut self)").unwrap(); ++ syn::parse_str::<Type>("fn(mut self: ())").unwrap(); ++ syn::parse_str::<Type>("fn(mut self: ...)").unwrap_err(); ++ syn::parse_str::<Type>("fn(mut self: mut self)").unwrap_err(); ++ syn::parse_str::<Type>("fn(mut self::T)").unwrap_err(); ++} ++ ++#[test] ++fn test_macro_variable_type() { ++ // mimics the token stream corresponding to `$ty<T>` ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, quote! { ty })), ++ TokenTree::Punct(Punct::new('<', Spacing::Alone)), ++ TokenTree::Ident(Ident::new("T", Span::call_site())), ++ TokenTree::Punct(Punct::new('>', Spacing::Alone)), ++ ]); ++ ++ snapshot!(tokens as Type, @r###" ++ Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "ty", ++ arguments: PathArguments::AngleBracketed { ++ args: [ ++ Type(Type::Path { ++ path: Path { ++ segments: [ ++ PathSegment { ++ ident: "T", ++ arguments: None, ++ }, ++ ], ++ }, ++ }), ++ ], ++ }, ++ }, ++ ], ++ }, ++ } ++ "###); ++} +diff --git a/third_party/rust/syn/tests/test_visibility.rs b/third_party/rust/syn/tests/test_visibility.rs +new file mode 100644 +index 0000000000..c3d0ac7a5b +--- /dev/null ++++ b/third_party/rust/syn/tests/test_visibility.rs +@@ -0,0 +1,145 @@ ++#[macro_use] ++mod macros; ++ ++use proc_macro2::{Delimiter, Group, Ident, Punct, Spacing, Span, TokenStream, TokenTree}; ++use std::iter::FromIterator; ++use syn::parse::{Parse, ParseStream}; ++use syn::{DeriveInput, Result, Visibility}; ++ ++#[derive(Debug)] ++struct VisRest { ++ vis: Visibility, ++ rest: TokenStream, ++} ++ ++impl Parse for VisRest { ++ fn parse(input: ParseStream) -> Result<Self> { ++ Ok(VisRest { ++ vis: input.parse()?, ++ rest: input.parse()?, ++ }) ++ } ++} ++ ++macro_rules! assert_vis_parse { ++ ($input:expr, Ok($p:pat)) => { ++ assert_vis_parse!($input, Ok($p) + ""); ++ }; ++ ++ ($input:expr, Ok($p:pat) + $rest:expr) => { ++ let expected = $rest.parse::<TokenStream>().unwrap(); ++ let parse: VisRest = syn::parse_str($input).unwrap(); ++ ++ match parse.vis { ++ $p => {} ++ _ => panic!("Expected {}, got {:?}", stringify!($p), parse.vis), ++ } ++ ++ // NOTE: Round-trips through `to_string` to avoid potential whitespace ++ // diffs. ++ assert_eq!(parse.rest.to_string(), expected.to_string()); ++ }; ++ ++ ($input:expr, Err) => { ++ syn::parse2::<VisRest>($input.parse().unwrap()).unwrap_err(); ++ }; ++} ++ ++#[test] ++fn test_pub() { ++ assert_vis_parse!("pub", Ok(Visibility::Public(_))); ++} ++ ++#[test] ++fn test_crate() { ++ assert_vis_parse!("crate", Ok(Visibility::Crate(_))); ++} ++ ++#[test] ++fn test_inherited() { ++ assert_vis_parse!("", Ok(Visibility::Inherited)); ++} ++ ++#[test] ++fn test_in() { ++ assert_vis_parse!("pub(in foo::bar)", Ok(Visibility::Restricted(_))); ++} ++ ++#[test] ++fn test_pub_crate() { ++ assert_vis_parse!("pub(crate)", Ok(Visibility::Restricted(_))); ++} ++ ++#[test] ++fn test_pub_self() { ++ assert_vis_parse!("pub(self)", Ok(Visibility::Restricted(_))); ++} ++ ++#[test] ++fn test_pub_super() { ++ assert_vis_parse!("pub(super)", Ok(Visibility::Restricted(_))); ++} ++ ++#[test] ++fn test_missing_in() { ++ assert_vis_parse!("pub(foo::bar)", Ok(Visibility::Public(_)) + "(foo::bar)"); ++} ++ ++#[test] ++fn test_missing_in_path() { ++ assert_vis_parse!("pub(in)", Err); ++} ++ ++#[test] ++fn test_crate_path() { ++ assert_vis_parse!("pub(crate::A, crate::B)", Ok(Visibility::Public(_)) + "(crate::A, crate::B)"); ++} ++ ++#[test] ++fn test_junk_after_in() { ++ assert_vis_parse!("pub(in some::path @@garbage)", Err); ++} ++ ++#[test] ++fn test_empty_group_vis() { ++ // mimics `struct S { $vis $field: () }` where $vis is empty ++ let tokens = TokenStream::from_iter(vec![ ++ TokenTree::Ident(Ident::new("struct", Span::call_site())), ++ TokenTree::Ident(Ident::new("S", Span::call_site())), ++ TokenTree::Group(Group::new( ++ Delimiter::Brace, ++ TokenStream::from_iter(vec![ ++ TokenTree::Group(Group::new(Delimiter::None, TokenStream::new())), ++ TokenTree::Group(Group::new( ++ Delimiter::None, ++ TokenStream::from_iter(vec![TokenTree::Ident(Ident::new( ++ "f", ++ Span::call_site(), ++ ))]), ++ )), ++ TokenTree::Punct(Punct::new(':', Spacing::Alone)), ++ TokenTree::Group(Group::new(Delimiter::Parenthesis, TokenStream::new())), ++ ]), ++ )), ++ ]); ++ ++ snapshot!(tokens as DeriveInput, @r###" ++ DeriveInput { ++ vis: Inherited, ++ ident: "S", ++ generics: Generics, ++ data: Data::Struct { ++ fields: Fields::Named { ++ named: [ ++ Field { ++ vis: Inherited, ++ ident: Some("f"), ++ colon_token: Some, ++ ty: Type::Tuple, ++ }, ++ ], ++ }, ++ }, ++ } ++ "###); ++} +diff --git a/third_party/rust/syn/tests/zzz_stable.rs b/third_party/rust/syn/tests/zzz_stable.rs +index a81b3df4d0..a1a670d9ed 100644 +--- a/third_party/rust/syn/tests/zzz_stable.rs ++++ b/third_party/rust/syn/tests/zzz_stable.rs +@@ -1,7 +1,5 @@ + #![cfg(syn_disable_nightly_tests)] + +-extern crate termcolor; +- + use std::io::{self, Write}; + use termcolor::{Color, ColorChoice, ColorSpec, StandardStream, WriteColor}; + +@@ -10,7 +8,7 @@ const MSG: &str = "\ + ‖ WARNING: + ‖ This is not a nightly compiler so not all tests were able to + ‖ run. Syn includes tests that compare Syn's parser against the +-‖ compiler's parser, which requires access to unstable libsyntax ++‖ compiler's parser, which requires access to unstable librustc + ‖ data structures and a nightly compiler. + ‖ + "; +-- +2.28.0 + diff --git a/source/xap/seamonkey/seamonkey.SlackBuild b/source/xap/seamonkey/seamonkey.SlackBuild index feacb5c1..cd4a58f2 100755 --- a/source/xap/seamonkey/seamonkey.SlackBuild +++ b/source/xap/seamonkey/seamonkey.SlackBuild @@ -68,8 +68,8 @@ else fi # Choose a compiler (gcc/g++ or clang/clang++): -export CC=${CC:-clang} -export CXX=${CXX:-clang++} +export CC=${CC:-gcc} +export CXX=${CXX:-g++} # -Wformat is needed for -Werror=format-security # -fno-delete-null-pointer-checks disables gcc6 optimization that leads to instability |