| use clippy_utils::attrs::is_doc_hidden; |
| use clippy_utils::diagnostics::{span_lint, span_lint_and_help, span_lint_and_note, span_lint_and_then}; |
| use clippy_utils::macros::{is_panic, root_macro_call_first_node}; |
| use clippy_utils::source::snippet_with_applicability; |
| use clippy_utils::ty::{implements_trait, is_type_diagnostic_item}; |
| use clippy_utils::{is_entrypoint_fn, method_chain_args, return_ty}; |
| use if_chain::if_chain; |
| use pulldown_cmark::Event::{ |
| Code, End, FootnoteReference, HardBreak, Html, Rule, SoftBreak, Start, TaskListMarker, Text, |
| }; |
| use pulldown_cmark::Tag::{CodeBlock, Heading, Item, Link, Paragraph}; |
| use pulldown_cmark::{BrokenLink, CodeBlockKind, CowStr, Options}; |
| use rustc_ast::ast::{Async, Attribute, Fn, FnRetTy, ItemKind}; |
| use rustc_data_structures::fx::FxHashSet; |
| use rustc_data_structures::sync::Lrc; |
| use rustc_errors::emitter::EmitterWriter; |
| use rustc_errors::{Applicability, Handler, SuggestionStyle}; |
| use rustc_hir as hir; |
| use rustc_hir::intravisit::{self, Visitor}; |
| use rustc_hir::{AnonConst, Expr}; |
| use rustc_lint::{LateContext, LateLintPass}; |
| use rustc_middle::hir::nested_filter; |
| use rustc_middle::lint::in_external_macro; |
| use rustc_middle::ty; |
| use rustc_parse::maybe_new_parser_from_source_str; |
| use rustc_parse::parser::ForceCollect; |
| use rustc_resolve::rustdoc::{ |
| add_doc_fragment, attrs_to_doc_fragments, main_body_opts, source_span_for_markdown_range, DocFragment, |
| }; |
| use rustc_session::parse::ParseSess; |
| use rustc_session::{declare_tool_lint, impl_lint_pass}; |
| use rustc_span::edition::Edition; |
| use rustc_span::{sym, BytePos, FileName, Pos, Span}; |
| use rustc_span::source_map::{FilePathMapping, SourceMap}; |
| use std::ops::Range; |
| use std::{io, thread}; |
| use url::Url; |
| |
| declare_clippy_lint! { |
| /// ### What it does |
| /// Checks for the presence of `_`, `::` or camel-case words |
| /// outside ticks in documentation. |
| /// |
| /// ### Why is this bad? |
| /// *Rustdoc* supports markdown formatting, `_`, `::` and |
| /// camel-case probably indicates some code which should be included between |
| /// ticks. `_` can also be used for emphasis in markdown, this lint tries to |
| /// consider that. |
| /// |
| /// ### Known problems |
| /// Lots of bad docs won’t be fixed, what the lint checks |
| /// for is limited, and there are still false positives. HTML elements and their |
| /// content are not linted. |
| /// |
| /// In addition, when writing documentation comments, including `[]` brackets |
| /// inside a link text would trip the parser. Therefore, documenting link with |
| /// `[`SmallVec<[T; INLINE_CAPACITY]>`]` and then [`SmallVec<[T; INLINE_CAPACITY]>`]: SmallVec |
| /// would fail. |
| /// |
| /// ### Examples |
| /// ```no_run |
| /// /// Do something with the foo_bar parameter. See also |
| /// /// that::other::module::foo. |
| /// // ^ `foo_bar` and `that::other::module::foo` should be ticked. |
| /// fn doit(foo_bar: usize) {} |
| /// ``` |
| /// |
| /// ```no_run |
| /// // Link text with `[]` brackets should be written as following: |
| /// /// Consume the array and return the inner |
| /// /// [`SmallVec<[T; INLINE_CAPACITY]>`][SmallVec]. |
| /// /// [SmallVec]: SmallVec |
| /// fn main() {} |
| /// ``` |
| #[clippy::version = "pre 1.29.0"] |
| pub DOC_MARKDOWN, |
| pedantic, |
| "presence of `_`, `::` or camel-case outside backticks in documentation" |
| } |
| |
| declare_clippy_lint! { |
| /// ### What it does |
| /// Checks for the doc comments of publicly visible |
| /// unsafe functions and warns if there is no `# Safety` section. |
| /// |
| /// ### Why is this bad? |
| /// Unsafe functions should document their safety |
| /// preconditions, so that users can be sure they are using them safely. |
| /// |
| /// ### Examples |
| /// ```no_run |
| ///# type Universe = (); |
| /// /// This function should really be documented |
| /// pub unsafe fn start_apocalypse(u: &mut Universe) { |
| /// unimplemented!(); |
| /// } |
| /// ``` |
| /// |
| /// At least write a line about safety: |
| /// |
| /// ```no_run |
| ///# type Universe = (); |
| /// /// # Safety |
| /// /// |
| /// /// This function should not be called before the horsemen are ready. |
| /// pub unsafe fn start_apocalypse(u: &mut Universe) { |
| /// unimplemented!(); |
| /// } |
| /// ``` |
| #[clippy::version = "1.39.0"] |
| pub MISSING_SAFETY_DOC, |
| style, |
| "`pub unsafe fn` without `# Safety` docs" |
| } |
| |
| declare_clippy_lint! { |
| /// ### What it does |
| /// Checks the doc comments of publicly visible functions that |
| /// return a `Result` type and warns if there is no `# Errors` section. |
| /// |
| /// ### Why is this bad? |
| /// Documenting the type of errors that can be returned from a |
| /// function can help callers write code to handle the errors appropriately. |
| /// |
| /// ### Examples |
| /// Since the following function returns a `Result` it has an `# Errors` section in |
| /// its doc comment: |
| /// |
| /// ```no_run |
| ///# use std::io; |
| /// /// # Errors |
| /// /// |
| /// /// Will return `Err` if `filename` does not exist or the user does not have |
| /// /// permission to read it. |
| /// pub fn read(filename: String) -> io::Result<String> { |
| /// unimplemented!(); |
| /// } |
| /// ``` |
| #[clippy::version = "1.41.0"] |
| pub MISSING_ERRORS_DOC, |
| pedantic, |
| "`pub fn` returns `Result` without `# Errors` in doc comment" |
| } |
| |
| declare_clippy_lint! { |
| /// ### What it does |
| /// Checks the doc comments of publicly visible functions that |
| /// may panic and warns if there is no `# Panics` section. |
| /// |
| /// ### Why is this bad? |
| /// Documenting the scenarios in which panicking occurs |
| /// can help callers who do not want to panic to avoid those situations. |
| /// |
| /// ### Examples |
| /// Since the following function may panic it has a `# Panics` section in |
| /// its doc comment: |
| /// |
| /// ```no_run |
| /// /// # Panics |
| /// /// |
| /// /// Will panic if y is 0 |
| /// pub fn divide_by(x: i32, y: i32) -> i32 { |
| /// if y == 0 { |
| /// panic!("Cannot divide by 0") |
| /// } else { |
| /// x / y |
| /// } |
| /// } |
| /// ``` |
| #[clippy::version = "1.51.0"] |
| pub MISSING_PANICS_DOC, |
| pedantic, |
| "`pub fn` may panic without `# Panics` in doc comment" |
| } |
| |
| declare_clippy_lint! { |
| /// ### What it does |
| /// Checks for `fn main() { .. }` in doctests |
| /// |
| /// ### Why is this bad? |
| /// The test can be shorter (and likely more readable) |
| /// if the `fn main()` is left implicit. |
| /// |
| /// ### Examples |
| /// ```no_run |
| /// /// An example of a doctest with a `main()` function |
| /// /// |
| /// /// # Examples |
| /// /// |
| /// /// ``` |
| /// /// fn main() { |
| /// /// // this needs not be in an `fn` |
| /// /// } |
| /// /// ``` |
| /// fn needless_main() { |
| /// unimplemented!(); |
| /// } |
| /// ``` |
| #[clippy::version = "1.40.0"] |
| pub NEEDLESS_DOCTEST_MAIN, |
| style, |
| "presence of `fn main() {` in code examples" |
| } |
| |
| declare_clippy_lint! { |
| /// ### What it does |
| /// Detects the syntax `['foo']` in documentation comments (notice quotes instead of backticks) |
| /// outside of code blocks |
| /// ### Why is this bad? |
| /// It is likely a typo when defining an intra-doc link |
| /// |
| /// ### Example |
| /// ```no_run |
| /// /// See also: ['foo'] |
| /// fn bar() {} |
| /// ``` |
| /// Use instead: |
| /// ```no_run |
| /// /// See also: [`foo`] |
| /// fn bar() {} |
| /// ``` |
| #[clippy::version = "1.63.0"] |
| pub DOC_LINK_WITH_QUOTES, |
| pedantic, |
| "possible typo for an intra-doc link" |
| } |
| |
| declare_clippy_lint! { |
| /// ### What it does |
| /// Checks for the doc comments of publicly visible |
| /// safe functions and traits and warns if there is a `# Safety` section. |
| /// |
| /// ### Why is this bad? |
| /// Safe functions and traits are safe to implement and therefore do not |
| /// need to describe safety preconditions that users are required to uphold. |
| /// |
| /// ### Examples |
| /// ```no_run |
| ///# type Universe = (); |
| /// /// # Safety |
| /// /// |
| /// /// This function should not be called before the horsemen are ready. |
| /// pub fn start_apocalypse_but_safely(u: &mut Universe) { |
| /// unimplemented!(); |
| /// } |
| /// ``` |
| /// |
| /// The function is safe, so there shouldn't be any preconditions |
| /// that have to be explained for safety reasons. |
| /// |
| /// ```no_run |
| ///# type Universe = (); |
| /// /// This function should really be documented |
| /// pub fn start_apocalypse(u: &mut Universe) { |
| /// unimplemented!(); |
| /// } |
| /// ``` |
| #[clippy::version = "1.67.0"] |
| pub UNNECESSARY_SAFETY_DOC, |
| restriction, |
| "`pub fn` or `pub trait` with `# Safety` docs" |
| } |
| |
| #[expect(clippy::module_name_repetitions)] |
| #[derive(Clone)] |
| pub struct DocMarkdown { |
| valid_idents: FxHashSet<String>, |
| in_trait_impl: bool, |
| } |
| |
| impl DocMarkdown { |
| pub fn new(valid_idents: FxHashSet<String>) -> Self { |
| Self { |
| valid_idents, |
| in_trait_impl: false, |
| } |
| } |
| } |
| |
| impl_lint_pass!(DocMarkdown => [ |
| DOC_LINK_WITH_QUOTES, |
| DOC_MARKDOWN, |
| MISSING_SAFETY_DOC, |
| MISSING_ERRORS_DOC, |
| MISSING_PANICS_DOC, |
| NEEDLESS_DOCTEST_MAIN, |
| UNNECESSARY_SAFETY_DOC, |
| ]); |
| |
| impl<'tcx> LateLintPass<'tcx> for DocMarkdown { |
| fn check_crate(&mut self, cx: &LateContext<'tcx>) { |
| let attrs = cx.tcx.hir().attrs(hir::CRATE_HIR_ID); |
| check_attrs(cx, &self.valid_idents, attrs); |
| } |
| |
| fn check_item(&mut self, cx: &LateContext<'tcx>, item: &'tcx hir::Item<'_>) { |
| let attrs = cx.tcx.hir().attrs(item.hir_id()); |
| let Some(headers) = check_attrs(cx, &self.valid_idents, attrs) else { |
| return; |
| }; |
| match item.kind { |
| hir::ItemKind::Fn(ref sig, _, body_id) => { |
| if !(is_entrypoint_fn(cx, item.owner_id.to_def_id()) || in_external_macro(cx.tcx.sess, item.span)) { |
| let body = cx.tcx.hir().body(body_id); |
| let mut fpu = FindPanicUnwrap { |
| cx, |
| typeck_results: cx.tcx.typeck(item.owner_id.def_id), |
| panic_span: None, |
| }; |
| fpu.visit_expr(body.value); |
| lint_for_missing_headers(cx, item.owner_id, sig, headers, Some(body_id), fpu.panic_span); |
| } |
| }, |
| hir::ItemKind::Impl(impl_) => { |
| self.in_trait_impl = impl_.of_trait.is_some(); |
| }, |
| hir::ItemKind::Trait(_, unsafety, ..) => match (headers.safety, unsafety) { |
| (false, hir::Unsafety::Unsafe) => span_lint( |
| cx, |
| MISSING_SAFETY_DOC, |
| cx.tcx.def_span(item.owner_id), |
| "docs for unsafe trait missing `# Safety` section", |
| ), |
| (true, hir::Unsafety::Normal) => span_lint( |
| cx, |
| UNNECESSARY_SAFETY_DOC, |
| cx.tcx.def_span(item.owner_id), |
| "docs for safe trait have unnecessary `# Safety` section", |
| ), |
| _ => (), |
| }, |
| _ => (), |
| } |
| } |
| |
| fn check_item_post(&mut self, _cx: &LateContext<'tcx>, item: &'tcx hir::Item<'_>) { |
| if let hir::ItemKind::Impl { .. } = item.kind { |
| self.in_trait_impl = false; |
| } |
| } |
| |
| fn check_trait_item(&mut self, cx: &LateContext<'tcx>, item: &'tcx hir::TraitItem<'_>) { |
| let attrs = cx.tcx.hir().attrs(item.hir_id()); |
| let Some(headers) = check_attrs(cx, &self.valid_idents, attrs) else { |
| return; |
| }; |
| if let hir::TraitItemKind::Fn(ref sig, ..) = item.kind { |
| if !in_external_macro(cx.tcx.sess, item.span) { |
| lint_for_missing_headers(cx, item.owner_id, sig, headers, None, None); |
| } |
| } |
| } |
| |
| fn check_impl_item(&mut self, cx: &LateContext<'tcx>, item: &'tcx hir::ImplItem<'_>) { |
| let attrs = cx.tcx.hir().attrs(item.hir_id()); |
| let Some(headers) = check_attrs(cx, &self.valid_idents, attrs) else { |
| return; |
| }; |
| if self.in_trait_impl || in_external_macro(cx.tcx.sess, item.span) { |
| return; |
| } |
| if let hir::ImplItemKind::Fn(ref sig, body_id) = item.kind { |
| let body = cx.tcx.hir().body(body_id); |
| let mut fpu = FindPanicUnwrap { |
| cx, |
| typeck_results: cx.tcx.typeck(item.owner_id.def_id), |
| panic_span: None, |
| }; |
| fpu.visit_expr(body.value); |
| lint_for_missing_headers(cx, item.owner_id, sig, headers, Some(body_id), fpu.panic_span); |
| } |
| } |
| } |
| |
| fn lint_for_missing_headers( |
| cx: &LateContext<'_>, |
| owner_id: hir::OwnerId, |
| sig: &hir::FnSig<'_>, |
| headers: DocHeaders, |
| body_id: Option<hir::BodyId>, |
| panic_span: Option<Span>, |
| ) { |
| if !cx.effective_visibilities.is_exported(owner_id.def_id) { |
| return; // Private functions do not require doc comments |
| } |
| |
| // do not lint if any parent has `#[doc(hidden)]` attribute (#7347) |
| if cx |
| .tcx |
| .hir() |
| .parent_iter(owner_id.into()) |
| .any(|(id, _node)| is_doc_hidden(cx.tcx.hir().attrs(id))) |
| { |
| return; |
| } |
| |
| let span = cx.tcx.def_span(owner_id); |
| match (headers.safety, sig.header.unsafety) { |
| (false, hir::Unsafety::Unsafe) => span_lint( |
| cx, |
| MISSING_SAFETY_DOC, |
| span, |
| "unsafe function's docs miss `# Safety` section", |
| ), |
| (true, hir::Unsafety::Normal) => span_lint( |
| cx, |
| UNNECESSARY_SAFETY_DOC, |
| span, |
| "safe function's docs have unnecessary `# Safety` section", |
| ), |
| _ => (), |
| } |
| if !headers.panics && panic_span.is_some() { |
| span_lint_and_note( |
| cx, |
| MISSING_PANICS_DOC, |
| span, |
| "docs for function which may panic missing `# Panics` section", |
| panic_span, |
| "first possible panic found here", |
| ); |
| } |
| if !headers.errors { |
| if is_type_diagnostic_item(cx, return_ty(cx, owner_id), sym::Result) { |
| span_lint( |
| cx, |
| MISSING_ERRORS_DOC, |
| span, |
| "docs for function returning `Result` missing `# Errors` section", |
| ); |
| } else { |
| if_chain! { |
| if let Some(body_id) = body_id; |
| if let Some(future) = cx.tcx.lang_items().future_trait(); |
| let typeck = cx.tcx.typeck_body(body_id); |
| let body = cx.tcx.hir().body(body_id); |
| let ret_ty = typeck.expr_ty(body.value); |
| if implements_trait(cx, ret_ty, future, &[]); |
| if let ty::Coroutine(_, subs, _) = ret_ty.kind(); |
| if is_type_diagnostic_item(cx, subs.as_coroutine().return_ty(), sym::Result); |
| then { |
| span_lint( |
| cx, |
| MISSING_ERRORS_DOC, |
| span, |
| "docs for function returning `Result` missing `# Errors` section", |
| ); |
| } |
| } |
| } |
| } |
| } |
| |
| #[derive(Copy, Clone)] |
| struct Fragments<'a> { |
| doc: &'a str, |
| fragments: &'a [DocFragment], |
| } |
| |
| impl Fragments<'_> { |
| fn span(self, cx: &LateContext<'_>, range: Range<usize>) -> Option<Span> { |
| source_span_for_markdown_range(cx.tcx, self.doc, &range, self.fragments) |
| } |
| } |
| |
| #[derive(Copy, Clone, Default)] |
| struct DocHeaders { |
| safety: bool, |
| errors: bool, |
| panics: bool, |
| } |
| |
| fn check_attrs(cx: &LateContext<'_>, valid_idents: &FxHashSet<String>, attrs: &[Attribute]) -> Option<DocHeaders> { |
| /// We don't want the parser to choke on intra doc links. Since we don't |
| /// actually care about rendering them, just pretend that all broken links |
| /// point to a fake address. |
| #[expect(clippy::unnecessary_wraps)] // we're following a type signature |
| fn fake_broken_link_callback<'a>(_: BrokenLink<'_>) -> Option<(CowStr<'a>, CowStr<'a>)> { |
| Some(("fake".into(), "fake".into())) |
| } |
| |
| if is_doc_hidden(attrs) { |
| return None; |
| } |
| |
| let (fragments, _) = attrs_to_doc_fragments(attrs.iter().map(|attr| (attr, None)), true); |
| let mut doc = String::new(); |
| for fragment in &fragments { |
| add_doc_fragment(&mut doc, fragment); |
| } |
| doc.pop(); |
| |
| if doc.is_empty() { |
| return Some(DocHeaders::default()); |
| } |
| |
| let mut cb = fake_broken_link_callback; |
| |
| // disable smart punctuation to pick up ['link'] more easily |
| let opts = main_body_opts() - Options::ENABLE_SMART_PUNCTUATION; |
| let parser = pulldown_cmark::Parser::new_with_broken_link_callback(&doc, opts, Some(&mut cb)); |
| |
| Some(check_doc( |
| cx, |
| valid_idents, |
| parser.into_offset_iter(), |
| Fragments { |
| fragments: &fragments, |
| doc: &doc, |
| }, |
| )) |
| } |
| |
| const RUST_CODE: &[&str] = &["rust", "no_run", "should_panic", "compile_fail"]; |
| |
| #[allow(clippy::too_many_lines)] // Only a big match statement |
| fn check_doc<'a, Events: Iterator<Item = (pulldown_cmark::Event<'a>, Range<usize>)>>( |
| cx: &LateContext<'_>, |
| valid_idents: &FxHashSet<String>, |
| events: Events, |
| fragments: Fragments<'_>, |
| ) -> DocHeaders { |
| // true if a safety header was found |
| let mut headers = DocHeaders::default(); |
| let mut in_code = false; |
| let mut in_link = None; |
| let mut in_heading = false; |
| let mut is_rust = false; |
| let mut no_test = false; |
| let mut edition = None; |
| let mut ticks_unbalanced = false; |
| let mut text_to_check: Vec<(CowStr<'_>, Range<usize>)> = Vec::new(); |
| let mut paragraph_range = 0..0; |
| for (event, range) in events { |
| match event { |
| Start(CodeBlock(ref kind)) => { |
| in_code = true; |
| if let CodeBlockKind::Fenced(lang) = kind { |
| for item in lang.split(',') { |
| if item == "ignore" { |
| is_rust = false; |
| break; |
| } else if item == "no_test" { |
| no_test = true; |
| } |
| if let Some(stripped) = item.strip_prefix("edition") { |
| is_rust = true; |
| edition = stripped.parse::<Edition>().ok(); |
| } else if item.is_empty() || RUST_CODE.contains(&item) { |
| is_rust = true; |
| } |
| } |
| } |
| }, |
| End(CodeBlock(_)) => { |
| in_code = false; |
| is_rust = false; |
| }, |
| Start(Link(_, url, _)) => in_link = Some(url), |
| End(Link(..)) => in_link = None, |
| Start(Heading(_, _, _) | Paragraph | Item) => { |
| if let Start(Heading(_, _, _)) = event { |
| in_heading = true; |
| } |
| ticks_unbalanced = false; |
| paragraph_range = range; |
| }, |
| End(Heading(_, _, _) | Paragraph | Item) => { |
| if let End(Heading(_, _, _)) = event { |
| in_heading = false; |
| } |
| if ticks_unbalanced && let Some(span) = fragments.span(cx, paragraph_range.clone()) { |
| span_lint_and_help( |
| cx, |
| DOC_MARKDOWN, |
| span, |
| "backticks are unbalanced", |
| None, |
| "a backtick may be missing a pair", |
| ); |
| } else { |
| for (text, range) in text_to_check { |
| if let Some(span) = fragments.span(cx, range) { |
| check_text(cx, valid_idents, &text, span); |
| } |
| } |
| } |
| text_to_check = Vec::new(); |
| }, |
| Start(_tag) | End(_tag) => (), // We don't care about other tags |
| Html(_html) => (), // HTML is weird, just ignore it |
| SoftBreak | HardBreak | TaskListMarker(_) | Code(_) | Rule => (), |
| FootnoteReference(text) | Text(text) => { |
| paragraph_range.end = range.end; |
| ticks_unbalanced |= text.contains('`') && !in_code; |
| if Some(&text) == in_link.as_ref() || ticks_unbalanced { |
| // Probably a link of the form `<http://example.com>` |
| // Which are represented as a link to "http://example.com" with |
| // text "http://example.com" by pulldown-cmark |
| continue; |
| } |
| let trimmed_text = text.trim(); |
| headers.safety |= in_heading && trimmed_text == "Safety"; |
| headers.safety |= in_heading && trimmed_text == "Implementation safety"; |
| headers.safety |= in_heading && trimmed_text == "Implementation Safety"; |
| headers.errors |= in_heading && trimmed_text == "Errors"; |
| headers.panics |= in_heading && trimmed_text == "Panics"; |
| if in_code { |
| if is_rust && !no_test { |
| let edition = edition.unwrap_or_else(|| cx.tcx.sess.edition()); |
| check_code(cx, &text, edition, range.clone(), fragments); |
| } |
| } else { |
| if in_link.is_some() { |
| check_link_quotes(cx, trimmed_text, range.clone(), fragments); |
| } |
| if let Some(link) = in_link.as_ref() |
| && let Ok(url) = Url::parse(link) |
| && (url.scheme() == "https" || url.scheme() == "http") |
| { |
| // Don't check the text associated with external URLs |
| continue; |
| } |
| text_to_check.push((text, range)); |
| } |
| }, |
| } |
| } |
| headers |
| } |
| |
| fn check_link_quotes(cx: &LateContext<'_>, trimmed_text: &str, range: Range<usize>, fragments: Fragments<'_>) { |
| if trimmed_text.starts_with('\'') |
| && trimmed_text.ends_with('\'') |
| && let Some(span) = fragments.span(cx, range) |
| { |
| span_lint( |
| cx, |
| DOC_LINK_WITH_QUOTES, |
| span, |
| "possible intra-doc link using quotes instead of backticks", |
| ); |
| } |
| } |
| |
| fn check_code(cx: &LateContext<'_>, text: &str, edition: Edition, range: Range<usize>, fragments: Fragments<'_>) { |
| fn has_needless_main(code: String, edition: Edition) -> bool { |
| rustc_driver::catch_fatal_errors(|| { |
| rustc_span::create_session_globals_then(edition, || { |
| let filename = FileName::anon_source_code(&code); |
| |
| let sm = Lrc::new(SourceMap::new(FilePathMapping::empty())); |
| let fallback_bundle = |
| rustc_errors::fallback_fluent_bundle(rustc_driver::DEFAULT_LOCALE_RESOURCES.to_vec(), false); |
| let emitter = EmitterWriter::new(Box::new(io::sink()), fallback_bundle); |
| let handler = Handler::with_emitter(Box::new(emitter)).disable_warnings(); |
| let sess = ParseSess::with_span_handler(handler, sm); |
| |
| let mut parser = match maybe_new_parser_from_source_str(&sess, filename, code) { |
| Ok(p) => p, |
| Err(errs) => { |
| drop(errs); |
| return false; |
| }, |
| }; |
| |
| let mut relevant_main_found = false; |
| loop { |
| match parser.parse_item(ForceCollect::No) { |
| Ok(Some(item)) => match &item.kind { |
| ItemKind::Fn(box Fn { |
| sig, body: Some(block), .. |
| }) if item.ident.name == sym::main => { |
| let is_async = matches!(sig.header.asyncness, Async::Yes { .. }); |
| let returns_nothing = match &sig.decl.output { |
| FnRetTy::Default(..) => true, |
| FnRetTy::Ty(ty) if ty.kind.is_unit() => true, |
| FnRetTy::Ty(_) => false, |
| }; |
| |
| if returns_nothing && !is_async && !block.stmts.is_empty() { |
| // This main function should be linted, but only if there are no other functions |
| relevant_main_found = true; |
| } else { |
| // This main function should not be linted, we're done |
| return false; |
| } |
| }, |
| // Tests with one of these items are ignored |
| ItemKind::Static(..) |
| | ItemKind::Const(..) |
| | ItemKind::ExternCrate(..) |
| | ItemKind::ForeignMod(..) |
| // Another function was found; this case is ignored |
| | ItemKind::Fn(..) => return false, |
| _ => {}, |
| }, |
| Ok(None) => break, |
| Err(e) => { |
| e.cancel(); |
| return false; |
| }, |
| } |
| } |
| |
| relevant_main_found |
| }) |
| }) |
| .ok() |
| .unwrap_or_default() |
| } |
| |
| let trailing_whitespace = text.len() - text.trim_end().len(); |
| |
| // Because of the global session, we need to create a new session in a different thread with |
| // the edition we need. |
| let text = text.to_owned(); |
| if thread::spawn(move || has_needless_main(text, edition)) |
| .join() |
| .expect("thread::spawn failed") |
| && let Some(span) = fragments.span(cx, range.start..range.end - trailing_whitespace) |
| { |
| span_lint(cx, NEEDLESS_DOCTEST_MAIN, span, "needless `fn main` in doctest"); |
| } |
| } |
| |
| fn check_text(cx: &LateContext<'_>, valid_idents: &FxHashSet<String>, text: &str, span: Span) { |
| for word in text.split(|c: char| c.is_whitespace() || c == '\'') { |
| // Trim punctuation as in `some comment (see foo::bar).` |
| // ^^ |
| // Or even as in `_foo bar_` which is emphasized. Also preserve `::` as a prefix/suffix. |
| let mut word = word.trim_matches(|c: char| !c.is_alphanumeric() && c != ':'); |
| |
| // Remove leading or trailing single `:` which may be part of a sentence. |
| if word.starts_with(':') && !word.starts_with("::") { |
| word = word.trim_start_matches(':'); |
| } |
| if word.ends_with(':') && !word.ends_with("::") { |
| word = word.trim_end_matches(':'); |
| } |
| |
| if valid_idents.contains(word) || word.chars().all(|c| c == ':') { |
| continue; |
| } |
| |
| // Adjust for the current word |
| let offset = word.as_ptr() as usize - text.as_ptr() as usize; |
| let span = Span::new( |
| span.lo() + BytePos::from_usize(offset), |
| span.lo() + BytePos::from_usize(offset + word.len()), |
| span.ctxt(), |
| span.parent(), |
| ); |
| |
| check_word(cx, word, span); |
| } |
| } |
| |
| fn check_word(cx: &LateContext<'_>, word: &str, span: Span) { |
| /// Checks if a string is upper-camel-case, i.e., starts with an uppercase and |
| /// contains at least two uppercase letters (`Clippy` is ok) and one lower-case |
| /// letter (`NASA` is ok). |
| /// Plurals are also excluded (`IDs` is ok). |
| fn is_camel_case(s: &str) -> bool { |
| if s.starts_with(|c: char| c.is_ascii_digit() | c.is_ascii_lowercase()) { |
| return false; |
| } |
| |
| let s = s.strip_suffix('s').unwrap_or(s); |
| |
| s.chars().all(char::is_alphanumeric) |
| && s.chars().filter(|&c| c.is_uppercase()).take(2).count() > 1 |
| && s.chars().filter(|&c| c.is_lowercase()).take(1).count() > 0 |
| } |
| |
| fn has_underscore(s: &str) -> bool { |
| s != "_" && !s.contains("\\_") && s.contains('_') |
| } |
| |
| fn has_hyphen(s: &str) -> bool { |
| s != "-" && s.contains('-') |
| } |
| |
| if let Ok(url) = Url::parse(word) { |
| // try to get around the fact that `foo::bar` parses as a valid URL |
| if !url.cannot_be_a_base() { |
| span_lint( |
| cx, |
| DOC_MARKDOWN, |
| span, |
| "you should put bare URLs between `<`/`>` or make a proper Markdown link", |
| ); |
| |
| return; |
| } |
| } |
| |
| // We assume that mixed-case words are not meant to be put inside backticks. (Issue #2343) |
| if has_underscore(word) && has_hyphen(word) { |
| return; |
| } |
| |
| if has_underscore(word) || word.contains("::") || is_camel_case(word) { |
| let mut applicability = Applicability::MachineApplicable; |
| |
| span_lint_and_then( |
| cx, |
| DOC_MARKDOWN, |
| span, |
| "item in documentation is missing backticks", |
| |diag| { |
| let snippet = snippet_with_applicability(cx, span, "..", &mut applicability); |
| diag.span_suggestion_with_style( |
| span, |
| "try", |
| format!("`{snippet}`"), |
| applicability, |
| // always show the suggestion in a separate line, since the |
| // inline presentation adds another pair of backticks |
| SuggestionStyle::ShowAlways, |
| ); |
| }, |
| ); |
| } |
| } |
| |
| struct FindPanicUnwrap<'a, 'tcx> { |
| cx: &'a LateContext<'tcx>, |
| panic_span: Option<Span>, |
| typeck_results: &'tcx ty::TypeckResults<'tcx>, |
| } |
| |
| impl<'a, 'tcx> Visitor<'tcx> for FindPanicUnwrap<'a, 'tcx> { |
| type NestedFilter = nested_filter::OnlyBodies; |
| |
| fn visit_expr(&mut self, expr: &'tcx Expr<'_>) { |
| if self.panic_span.is_some() { |
| return; |
| } |
| |
| if let Some(macro_call) = root_macro_call_first_node(self.cx, expr) { |
| if is_panic(self.cx, macro_call.def_id) |
| || matches!( |
| self.cx.tcx.item_name(macro_call.def_id).as_str(), |
| "assert" | "assert_eq" | "assert_ne" |
| ) |
| { |
| self.panic_span = Some(macro_call.span); |
| } |
| } |
| |
| // check for `unwrap` and `expect` for both `Option` and `Result` |
| if let Some(arglists) = method_chain_args(expr, &["unwrap"]).or(method_chain_args(expr, &["expect"])) { |
| let receiver_ty = self.typeck_results.expr_ty(arglists[0].0).peel_refs(); |
| if is_type_diagnostic_item(self.cx, receiver_ty, sym::Option) |
| || is_type_diagnostic_item(self.cx, receiver_ty, sym::Result) |
| { |
| self.panic_span = Some(expr.span); |
| } |
| } |
| |
| // and check sub-expressions |
| intravisit::walk_expr(self, expr); |
| } |
| |
| // Panics in const blocks will cause compilation to fail. |
| fn visit_anon_const(&mut self, _: &'tcx AnonConst) {} |
| |
| fn nested_visit_map(&mut self) -> Self::Map { |
| self.cx.tcx.hir() |
| } |
| } |