mirror of
https://github.com/rust-lang/rust.git
synced 2024-11-21 22:34:05 +00:00
Spellchecking compiler comments
This PR cleans up the rest of the spelling mistakes in the compiler comments. This PR does not change any literal or code spelling issues.
This commit is contained in:
parent
3e7514670d
commit
5160f8f843
@ -2646,7 +2646,7 @@ mod sig {
|
||||
|
||||
// Convert the result having "2 * precision" significant-bits back to the one
|
||||
// having "precision" significant-bits. First, move the radix point from
|
||||
// poision "2*precision - 1" to "precision - 1". The exponent need to be
|
||||
// position "2*precision - 1" to "precision - 1". The exponent need to be
|
||||
// adjusted by "2*precision - 1" - "precision - 1" = "precision".
|
||||
*exp -= precision as ExpInt + 1;
|
||||
|
||||
|
@ -295,7 +295,7 @@ impl AttrAnnotatedTokenStream {
|
||||
///
|
||||
/// For example, `#[cfg(FALSE)] struct Foo {}` would
|
||||
/// have an `attrs` field containing the `#[cfg(FALSE)]` attr,
|
||||
/// and a `tokens` field storing the (unparesd) tokens `struct Foo {}`
|
||||
/// and a `tokens` field storing the (unparsed) tokens `struct Foo {}`
|
||||
#[derive(Clone, Debug, Encodable, Decodable)]
|
||||
pub struct AttributesData {
|
||||
/// Attributes, both outer and inner.
|
||||
|
@ -158,7 +158,7 @@ impl<'tcx> TypeRelatingDelegate<'tcx> for NllTypeRelatingDelegate<'_, '_, 'tcx>
|
||||
.infcx
|
||||
.handle_opaque_type(a, b, a_is_expected, &cause, param_env)?
|
||||
.obligations,
|
||||
// These fields are filled in during exectuion of the operation
|
||||
// These fields are filled in during execution of the operation
|
||||
base_universe: None,
|
||||
region_constraints: None,
|
||||
},
|
||||
|
@ -23,7 +23,7 @@
|
||||
//! ss2 = explicit_slot 8 ; _4: (&&[u16],) size=8 align=8,8
|
||||
//! sig0 = (i64, i64, i64) system_v
|
||||
//! sig1 = (i64, i64, i64) system_v
|
||||
//! fn0 = colocated u0:6 sig1 ; Instance { def: Item(DefId(0/0:31 ~ example[8787]::{{impl}}[1]::call_mut[0])), substs: [ReErased, ReErased] }
|
||||
//! fn0 = collocated u0:6 sig1 ; Instance { def: Item(DefId(0/0:31 ~ example[8787]::{{impl}}[1]::call_mut[0])), substs: [ReErased, ReErased] }
|
||||
//!
|
||||
//! block0(v0: i64, v1: i64, v2: i64):
|
||||
//! v3 = stack_addr.i64 ss0
|
||||
|
@ -779,7 +779,7 @@ pub fn compute_per_cgu_lto_type(
|
||||
// we'll encounter later.
|
||||
let is_allocator = module_kind == ModuleKind::Allocator;
|
||||
|
||||
// We ignore a request for full crate grath LTO if the cate type
|
||||
// We ignore a request for full crate graph LTO if the crate type
|
||||
// is only an rlib, as there is no full crate graph to process,
|
||||
// that'll happen later.
|
||||
//
|
||||
|
@ -35,7 +35,7 @@ enum Frame {
|
||||
impl Frame {
|
||||
/// Construct a new frame around the delimited set of tokens.
|
||||
fn new(mut tts: Vec<mbe::TokenTree>) -> Frame {
|
||||
// Need to add empty delimeters.
|
||||
// Need to add empty delimiters.
|
||||
let open_tt = mbe::TokenTree::token(token::OpenDelim(token::NoDelim), DUMMY_SP);
|
||||
let close_tt = mbe::TokenTree::token(token::CloseDelim(token::NoDelim), DUMMY_SP);
|
||||
tts.insert(0, open_tt);
|
||||
@ -210,7 +210,7 @@ pub(super) fn transcribe<'a>(
|
||||
));
|
||||
}
|
||||
} else {
|
||||
// 0 is the initial counter (we have done 0 repretitions so far). `len`
|
||||
// 0 is the initial counter (we have done 0 repetitions so far). `len`
|
||||
// is the total number of repetitions we should generate.
|
||||
repeats.push((0, len));
|
||||
|
||||
|
@ -1717,7 +1717,7 @@ impl<R: Idx, C: Idx> SparseBitMatrix<R, C> {
|
||||
if let Some(Some(row)) = self.rows.get(row) { Some(row) } else { None }
|
||||
}
|
||||
|
||||
/// Interescts `row` with `set`. `set` can be either `BitSet` or
|
||||
/// Intersects `row` with `set`. `set` can be either `BitSet` or
|
||||
/// `HybridBitSet`. Has no effect if `row` does not exist.
|
||||
///
|
||||
/// Returns true if the row was changed.
|
||||
|
@ -2,7 +2,7 @@
|
||||
//! in particular to extract out the resulting region obligations and
|
||||
//! encode them therein.
|
||||
//!
|
||||
//! For an overview of what canonicaliation is and how it fits into
|
||||
//! For an overview of what canonicalization is and how it fits into
|
||||
//! rustc, check out the [chapter in the rustc dev guide][c].
|
||||
//!
|
||||
//! [c]: https://rust-lang.github.io/chalk/book/canonical_queries/canonicalization.html
|
||||
|
@ -1164,7 +1164,7 @@ impl<'a, 'tcx> InferCtxt<'a, 'tcx> {
|
||||
let len = sub1.len() - common_default_params;
|
||||
let consts_offset = len - sub1.consts().count();
|
||||
|
||||
// Only draw `<...>` if there're lifetime/type arguments.
|
||||
// Only draw `<...>` if there are lifetime/type arguments.
|
||||
if len > 0 {
|
||||
values.0.push_normal("<");
|
||||
values.1.push_normal("<");
|
||||
@ -1245,7 +1245,7 @@ impl<'a, 'tcx> InferCtxt<'a, 'tcx> {
|
||||
}
|
||||
|
||||
// Close the type argument bracket.
|
||||
// Only draw `<...>` if there're lifetime/type arguments.
|
||||
// Only draw `<...>` if there are lifetime/type arguments.
|
||||
if len > 0 {
|
||||
values.0.push_normal(">");
|
||||
values.1.push_normal(">");
|
||||
@ -1857,7 +1857,7 @@ impl<'a, 'tcx> InferCtxt<'a, 'tcx> {
|
||||
})
|
||||
.transpose();
|
||||
if output.is_some() {
|
||||
// We don't account for multiple `Future::Output = Ty` contraints.
|
||||
// We don't account for multiple `Future::Output = Ty` constraints.
|
||||
return output;
|
||||
}
|
||||
}
|
||||
|
@ -83,7 +83,7 @@ where
|
||||
// But if we did in reverse order, we would create a `v <:
|
||||
// LHS` (or vice versa) constraint and then instantiate
|
||||
// `v`. This would require further processing to achieve same
|
||||
// end-result; in partiular, this screws up some of the logic
|
||||
// end-result; in particular, this screws up some of the logic
|
||||
// in coercion, which expects LUB to figure out that the LHS
|
||||
// is (e.g.) `Box<i32>`. A more obvious solution might be to
|
||||
// iterate on the subtype obligations that are returned, but I
|
||||
|
@ -343,7 +343,7 @@ pub struct InferCtxt<'a, 'tcx> {
|
||||
|
||||
/// Track how many errors were reported when this infcx is created.
|
||||
/// If the number of errors increases, that's also a sign (line
|
||||
/// `tained_by_errors`) to avoid reporting certain kinds of errors.
|
||||
/// `tainted_by_errors`) to avoid reporting certain kinds of errors.
|
||||
// FIXME(matthewjasper) Merge into `tainted_by_errors_flag`
|
||||
err_count_on_creation: usize,
|
||||
|
||||
|
@ -550,7 +550,7 @@ where
|
||||
}
|
||||
|
||||
if a == b {
|
||||
// Subtle: if a or b has a bound variable that we are lazilly
|
||||
// Subtle: if a or b has a bound variable that we are lazily
|
||||
// substituting, then even if a == b, it could be that the values we
|
||||
// will substitute for those bound variables are *not* the same, and
|
||||
// hence returning `Ok(a)` is incorrect.
|
||||
|
@ -306,7 +306,7 @@ impl<'a, 'tcx> InferCtxt<'a, 'tcx> {
|
||||
/// # Constrain regions, not the hidden concrete type
|
||||
///
|
||||
/// Note that generating constraints on each region `Rc` is *not*
|
||||
/// the same as generating an outlives constraint on `Tc` iself.
|
||||
/// the same as generating an outlives constraint on `Tc` itself.
|
||||
/// For example, if we had a function like this:
|
||||
///
|
||||
/// ```rust
|
||||
|
@ -125,7 +125,7 @@ fn compute_components<'tcx>(
|
||||
// regionck more ways to prove that it holds. However,
|
||||
// regionck is not (at least currently) prepared to deal with
|
||||
// higher-ranked regions that may appear in the
|
||||
// trait-ref. Therefore, if we see any higher-ranke regions,
|
||||
// trait-ref. Therefore, if we see any higher-rank regions,
|
||||
// we simply fallback to the most restrictive rule, which
|
||||
// requires that `Pi: 'a` for all `i`.
|
||||
ty::Projection(ref data) => {
|
||||
|
@ -93,7 +93,7 @@ impl<'a, 'tcx> OutlivesEnvironment<'tcx> {
|
||||
&self.region_bound_pairs_map
|
||||
}
|
||||
|
||||
/// This is a hack to support the old-skool regionck, which
|
||||
/// This is a hack to support the old-school regionck, which
|
||||
/// processes region constraints from the main function and the
|
||||
/// closure together. In that context, when we enter a closure, we
|
||||
/// want to be able to "save" the state of the surrounding a
|
||||
|
@ -365,7 +365,7 @@ where
|
||||
debug!("projection_must_outlive: approx_env_bounds={:?}", approx_env_bounds);
|
||||
|
||||
// Remove outlives bounds that we get from the environment but
|
||||
// which are also deducable from the trait. This arises (cc
|
||||
// which are also deducible from the trait. This arises (cc
|
||||
// #55756) in cases where you have e.g., `<T as Foo<'a>>::Item:
|
||||
// 'a` in the environment but `trait Foo<'b> { type Item: 'b
|
||||
// }` in the trait definition.
|
||||
|
@ -33,7 +33,7 @@ impl<'tcx> RegionConstraintCollector<'_, 'tcx> {
|
||||
/// not entirely true. In particular, in the future, we may extend the
|
||||
/// environment with implied bounds or other info about how placeholders
|
||||
/// relate to regions in outer universes. In that case, `P1: R` for example
|
||||
/// might become solveable.
|
||||
/// might become solvable.
|
||||
///
|
||||
/// # Summary of the implementation
|
||||
///
|
||||
@ -210,7 +210,7 @@ impl<'me, 'tcx> LeakCheck<'me, 'tcx> {
|
||||
// * `scc_placeholder[scc1]` stores the placeholder that `scc1` must
|
||||
// be equal to (if any)
|
||||
//
|
||||
// For each succssor `scc2` where `scc1: scc2`:
|
||||
// For each successor `scc2` where `scc1: scc2`:
|
||||
//
|
||||
// * `scc_placeholder[scc2]` stores some placeholder `P` where
|
||||
// `scc2: P` (if any)
|
||||
@ -243,7 +243,7 @@ impl<'me, 'tcx> LeakCheck<'me, 'tcx> {
|
||||
// Update minimum universe of scc1.
|
||||
self.scc_universes[scc1] = scc1_universe;
|
||||
|
||||
// At this point, `scc_placholder[scc1]` stores the placeholder that
|
||||
// At this point, `scc_placeholder[scc1]` stores the placeholder that
|
||||
// `scc1` must be equal to, if any.
|
||||
if let Some(scc1_placeholder) = self.scc_placeholders[scc1] {
|
||||
debug!(
|
||||
|
@ -46,7 +46,7 @@ pub struct RegionConstraintStorage<'tcx> {
|
||||
/// exist). This prevents us from making many such regions.
|
||||
glbs: CombineMap<'tcx>,
|
||||
|
||||
/// When we add a R1 == R2 constriant, we currently add (a) edges
|
||||
/// When we add a R1 == R2 constraint, we currently add (a) edges
|
||||
/// R1 <= R2 and R2 <= R1 and (b) we unify the two regions in this
|
||||
/// table. You can then call `opportunistic_resolve_var` early
|
||||
/// which will map R1 and R2 to some common region (i.e., either
|
||||
|
@ -55,7 +55,7 @@ impl<'a, 'tcx> TypeFolder<'tcx> for OpportunisticVarResolver<'a, 'tcx> {
|
||||
|
||||
/// The opportunistic region resolver opportunistically resolves regions
|
||||
/// variables to the variable with the least variable id. It is used when
|
||||
/// normlizing projections to avoid hitting the recursion limit by creating
|
||||
/// normalizing projections to avoid hitting the recursion limit by creating
|
||||
/// many versions of a predicate for types that in the end have to unify.
|
||||
///
|
||||
/// If you want to resolve type and const variables as well, call
|
||||
|
@ -929,7 +929,7 @@ fn analysis(tcx: TyCtxt<'_>, (): ()) -> Result<()> {
|
||||
});
|
||||
},
|
||||
{
|
||||
// We force these querie to run,
|
||||
// We force these queries to run,
|
||||
// since they might not otherwise get called.
|
||||
// This marks the corresponding crate-level attributes
|
||||
// as used, and ensures that their values are valid.
|
||||
|
@ -64,7 +64,7 @@ impl<'tcx> LateLintPass<'tcx> for NoopMethodCall {
|
||||
let substs = cx.typeck_results().node_substs(expr.hir_id);
|
||||
if substs.needs_subst() {
|
||||
// We can't resolve on types that require monomorphization, so we don't handle them if
|
||||
// we need to perfom substitution.
|
||||
// we need to perform substitution.
|
||||
return;
|
||||
}
|
||||
let param_env = cx.tcx.param_env(trait_id);
|
||||
|
@ -2254,7 +2254,7 @@ declare_lint! {
|
||||
declare_lint! {
|
||||
/// The `nontrivial_structural_match` lint detects constants that are used in patterns,
|
||||
/// whose type is not structural-match and whose initializer body actually uses values
|
||||
/// that are not structural-match. So `Option<NotStruturalMatch>` is ok if the constant
|
||||
/// that are not structural-match. So `Option<NotStructuralMatch>` is ok if the constant
|
||||
/// is just `None`.
|
||||
///
|
||||
/// ### Example
|
||||
@ -2276,7 +2276,7 @@ declare_lint! {
|
||||
///
|
||||
/// ### Explanation
|
||||
///
|
||||
/// Previous versions of Rust accepted constants in patterns, even if those constants's types
|
||||
/// Previous versions of Rust accepted constants in patterns, even if those constants' types
|
||||
/// did not have `PartialEq` derived. Thus the compiler falls back to runtime execution of
|
||||
/// `PartialEq`, which can report that two constants are not equal even if they are
|
||||
/// bit-equivalent.
|
||||
@ -3626,7 +3626,7 @@ declare_lint! {
|
||||
/// The `deref_into_dyn_supertrait` lint is output whenever there is a use of the
|
||||
/// `Deref` implementation with a `dyn SuperTrait` type as `Output`.
|
||||
///
|
||||
/// These implementations will become shadowed when the `trait_upcasting` feature is stablized.
|
||||
/// These implementations will become shadowed when the `trait_upcasting` feature is stabilized.
|
||||
/// The `deref` functions will no longer be called implicitly, so there might be behavior change.
|
||||
///
|
||||
/// ### Example
|
||||
|
@ -580,7 +580,7 @@ impl<'a, 'tcx> Decodable<DecodeContext<'a, 'tcx>> for Span {
|
||||
//
|
||||
// Suppose that we're currently compiling crate A, and start deserializing
|
||||
// metadata from crate B. When we deserialize a Span from crate B's metadata,
|
||||
// there are two posibilites:
|
||||
// there are two possibilities:
|
||||
//
|
||||
// 1. The span references a file from crate B. This makes it a 'local' span,
|
||||
// which means that we can use crate B's serialized source map information.
|
||||
|
@ -271,7 +271,7 @@ impl<'a, 'tcx> Encodable<EncodeContext<'a, 'tcx>> for Span {
|
||||
// from. We use `TAG_VALID_SPAN_FOREIGN` to indicate that a `CrateNum` should
|
||||
// be deserialized after the rest of the span data, which tells the deserializer
|
||||
// which crate contains the source map information.
|
||||
// 2. This span comes from our own crate. No special hamdling is needed - we just
|
||||
// 2. This span comes from our own crate. No special handling is needed - we just
|
||||
// write `TAG_VALID_SPAN_LOCAL` to let the deserializer know that it should use
|
||||
// our own source map information.
|
||||
//
|
||||
|
@ -218,7 +218,7 @@ impl<'hir> Map<'hir> {
|
||||
}
|
||||
|
||||
pub fn iter_local_def_id(self) -> impl Iterator<Item = LocalDefId> + 'hir {
|
||||
// Create a dependency to the crate to be sure we reexcute this when the amount of
|
||||
// Create a dependency to the crate to be sure we re-execute this when the amount of
|
||||
// definitions change.
|
||||
self.tcx.ensure().hir_crate(());
|
||||
self.tcx.untracked_resolutions.definitions.iter_local_def_id()
|
||||
|
@ -220,7 +220,7 @@ pub enum Certainty {
|
||||
/// distinguish the two (e.g., due to our preference for where
|
||||
/// clauses over impls).
|
||||
///
|
||||
/// After some unifiations and things have been done, it makes
|
||||
/// After some unification and things have been done, it makes
|
||||
/// sense to try and prove again -- of course, at that point, the
|
||||
/// canonical form will be different, making this a distinct
|
||||
/// query.
|
||||
|
@ -340,7 +340,7 @@ pub fn struct_lint_level<'s, 'd>(
|
||||
(Level::Expect(expect_id), _) => {
|
||||
// This case is special as we actually allow the lint itself in this context, but
|
||||
// we can't return early like in the case for `Level::Allow` because we still
|
||||
// need the lint diagnostic to be emitted to `rustc_error::HanderInner`.
|
||||
// need the lint diagnostic to be emitted to `rustc_error::HandlerInner`.
|
||||
//
|
||||
// We can also not mark the lint expectation as fulfilled here right away, as it
|
||||
// can still be cancelled in the decorate function. All of this means that we simply
|
||||
|
@ -234,7 +234,7 @@ pub enum UndefinedBehaviorInfo<'tcx> {
|
||||
},
|
||||
/// Something was divided by 0 (x / 0).
|
||||
DivisionByZero,
|
||||
/// Something was "remainded" by 0 (x % 0).
|
||||
/// Something was "remaineded" by 0 (x % 0).
|
||||
RemainderByZero,
|
||||
/// Signed division overflowed (INT_MIN / -1).
|
||||
DivisionOverflow,
|
||||
|
@ -2840,7 +2840,7 @@ impl<'tcx> UserTypeProjections {
|
||||
|
||||
/// Encodes the effect of a user-supplied type annotation on the
|
||||
/// subcomponents of a pattern. The effect is determined by applying the
|
||||
/// given list of proejctions to some underlying base type. Often,
|
||||
/// given list of projections to some underlying base type. Often,
|
||||
/// the projection element list `projs` is empty, in which case this
|
||||
/// directly encodes a type in `base`. But in the case of complex patterns with
|
||||
/// subpatterns and bindings, we want to apply only a *part* of the type to a variable,
|
||||
|
@ -126,7 +126,7 @@ pub enum UnusedUnsafe {
|
||||
/// > ``… because it's nested under this `unsafe fn` ``
|
||||
///
|
||||
/// the second HirId here indicates the first usage of the `unsafe` block,
|
||||
/// which allows retrival of the LintLevelSource for why that operation would
|
||||
/// which allows retrieval of the LintLevelSource for why that operation would
|
||||
/// have been permitted without the block
|
||||
InUnsafeFn(hir::HirId, hir::HirId),
|
||||
}
|
||||
|
@ -5,7 +5,7 @@ use super::*;
|
||||
/// Preorder traversal of a graph.
|
||||
///
|
||||
/// Preorder traversal is when each node is visited after at least one of its predecessors. If you
|
||||
/// are familar with some basic graph theory, then this performs a depth first search and returns
|
||||
/// are familiar with some basic graph theory, then this performs a depth first search and returns
|
||||
/// nodes in order of discovery time.
|
||||
///
|
||||
/// ```text
|
||||
|
@ -656,7 +656,7 @@ pub enum PatKind<'tcx> {
|
||||
/// One of the following:
|
||||
/// * `&str`, which will be handled as a string pattern and thus exhaustiveness
|
||||
/// checking will detect if you use the same string twice in different patterns.
|
||||
/// * integer, bool, char or float, which will be handled by exhaustivenes to cover exactly
|
||||
/// * integer, bool, char or float, which will be handled by exhaustiveness to cover exactly
|
||||
/// its own value, similar to `&str`, but these values are much simpler.
|
||||
/// * Opaque constants, that must not be matched structurally. So anything that does not derive
|
||||
/// `PartialEq` and `Eq`.
|
||||
|
@ -524,7 +524,7 @@ pub type SelectionResult<'tcx, T> = Result<Option<T>, SelectionError<'tcx>>;
|
||||
/// // Case A: ImplSource points at a specific impl. Only possible when
|
||||
/// // type is concretely known. If the impl itself has bounded
|
||||
/// // type parameters, ImplSource will carry resolutions for those as well:
|
||||
/// concrete.clone(); // ImpleSource(Impl_1, [ImplSource(Impl_2, [ImplSource(Impl_3)])])
|
||||
/// concrete.clone(); // ImplSource(Impl_1, [ImplSource(Impl_2, [ImplSource(Impl_3)])])
|
||||
///
|
||||
/// // Case A: ImplSource points at a specific impl. Only possible when
|
||||
/// // type is concretely known. If the impl itself has bounded
|
||||
|
@ -4,7 +4,7 @@ use crate::ty::{PolyTraitRef, TyCtxt};
|
||||
|
||||
/// Given a PolyTraitRef, get the PolyTraitRefs of the trait's (transitive) supertraits.
|
||||
///
|
||||
/// A simplfied version of the same function at `rustc_infer::traits::util::supertraits`.
|
||||
/// A simplified version of the same function at `rustc_infer::traits::util::supertraits`.
|
||||
pub fn supertraits<'tcx>(
|
||||
tcx: TyCtxt<'tcx>,
|
||||
trait_ref: PolyTraitRef<'tcx>,
|
||||
|
@ -139,7 +139,7 @@ pub fn suggest_arbitrary_trait_bound(
|
||||
(Some(_), "Self") => return false,
|
||||
_ => {}
|
||||
}
|
||||
// Suggest a where clause bound for a non-type paremeter.
|
||||
// Suggest a where clause bound for a non-type parameter.
|
||||
let (action, prefix) = if generics.where_clause.predicates.is_empty() {
|
||||
("introducing a", " where ")
|
||||
} else {
|
||||
|
@ -2718,7 +2718,7 @@ impl<'tcx> ty::Instance<'tcx> {
|
||||
/// stack-based unwinding (the exact mechanism of which varies
|
||||
/// platform-by-platform).
|
||||
///
|
||||
/// Rust functions are classfied whether or not they can unwind based on the
|
||||
/// Rust functions are classified whether or not they can unwind based on the
|
||||
/// active "panic strategy". In other words Rust functions are considered to
|
||||
/// unwind in `-Cpanic=unwind` mode and cannot unwind in `-Cpanic=abort` mode.
|
||||
/// Note that Rust supports intermingling panic=abort and panic=unwind code, but
|
||||
@ -2773,7 +2773,7 @@ pub fn fn_can_unwind<'tcx>(
|
||||
// To fix this UB rustc would like to change in the future to catch unwinds
|
||||
// from function calls that may unwind within a Rust-defined `extern "C"`
|
||||
// function and forcibly abort the process, thereby respecting the
|
||||
// `nounwind` attribut emitted for `extern "C"`. This behavior change isn't
|
||||
// `nounwind` attribute emitted for `extern "C"`. This behavior change isn't
|
||||
// ready to roll out, so determining whether or not the `C` family of ABIs
|
||||
// unwinds is conditional not only on their definition but also whether the
|
||||
// `#![feature(c_unwind)]` feature gate is active.
|
||||
|
@ -5,7 +5,7 @@
|
||||
//! - [`rustc_middle::ty::Ty`], used to represent the semantics of a type.
|
||||
//! - [`rustc_middle::ty::TyCtxt`], the central data structure in the compiler.
|
||||
//!
|
||||
//! For more information, see ["The `ty` module: representing types"] in the ructc-dev-guide.
|
||||
//! For more information, see ["The `ty` module: representing types"] in the rustc-dev-guide.
|
||||
//!
|
||||
//! ["The `ty` module: representing types"]: https://rustc-dev-guide.rust-lang.org/ty.html
|
||||
|
||||
@ -1263,7 +1263,7 @@ pub type PlaceholderConst<'tcx> = Placeholder<BoundConst<'tcx>>;
|
||||
/// aren't allowed to call that query: it is equal to `type_of(const_param)` which is
|
||||
/// trivial to compute.
|
||||
///
|
||||
/// If we now want to use that constant in a place which potentionally needs its type
|
||||
/// If we now want to use that constant in a place which potentially needs its type
|
||||
/// we also pass the type of its `const_param`. This is the point of `WithOptConstParam`,
|
||||
/// except that instead of a `Ty` we bundle the `DefId` of the const parameter.
|
||||
/// Meaning that we need to use `type_of(const_param_did)` if `const_param_did` is `Some`
|
||||
|
@ -431,7 +431,7 @@ pub trait PrettyPrinter<'tcx>:
|
||||
// For example, take `std::os::unix::process::CommandExt`, this trait is actually
|
||||
// defined at `std::sys::unix::ext::process::CommandExt` (at time of writing).
|
||||
//
|
||||
// `std::os::unix` rexports the contents of `std::sys::unix::ext`. `std::sys` is
|
||||
// `std::os::unix` reexports the contents of `std::sys::unix::ext`. `std::sys` is
|
||||
// private so the "true" path to `CommandExt` isn't accessible.
|
||||
//
|
||||
// In this case, the `visible_parent_map` will look something like this:
|
||||
@ -1036,7 +1036,7 @@ pub trait PrettyPrinter<'tcx>:
|
||||
|
||||
let mut resugared = false;
|
||||
|
||||
// Special-case `Fn(...) -> ...` and resugar it.
|
||||
// Special-case `Fn(...) -> ...` and re-sugar it.
|
||||
let fn_trait_kind = cx.tcx().fn_trait_kind_from_lang_item(principal.def_id);
|
||||
if !cx.tcx().sess.verbose() && fn_trait_kind.is_some() {
|
||||
if let ty::Tuple(tys) = principal.substs.type_at(0).kind() {
|
||||
@ -2171,7 +2171,7 @@ impl<'tcx> FmtPrinter<'_, 'tcx> {
|
||||
define_scoped_cx!(self);
|
||||
|
||||
let mut region_index = self.region_index;
|
||||
// If we want to print verbosly, then print *all* binders, even if they
|
||||
// If we want to print verbosely, then print *all* binders, even if they
|
||||
// aren't named. Eventually, we might just want this as the default, but
|
||||
// this is not *quite* right and changes the ordering of some output
|
||||
// anyways.
|
||||
|
@ -284,7 +284,7 @@ static_assert_size!(TyKind<'_>, 32);
|
||||
/// - 'l0...'li and T0...Tj are the generic parameters
|
||||
/// in scope on the function that defined the closure,
|
||||
/// - CK represents the *closure kind* (Fn vs FnMut vs FnOnce). This
|
||||
/// is rather hackily encoded via a scalar type. See
|
||||
/// is rather hacky encoded via a scalar type. See
|
||||
/// `Ty::to_opt_closure_kind` for details.
|
||||
/// - CS represents the *closure signature*, representing as a `fn()`
|
||||
/// type. For example, `fn(u32, u32) -> u32` would mean that the closure
|
||||
|
@ -444,7 +444,7 @@ impl<'a, 'tcx> Builder<'a, 'tcx> {
|
||||
// we lower the guard.
|
||||
let target_block = self.cfg.start_new_block();
|
||||
let mut schedule_drops = true;
|
||||
// We keep a stack of all of the bindings and type asciptions
|
||||
// We keep a stack of all of the bindings and type descriptions
|
||||
// from the parent candidates that we visit, that also need to
|
||||
// be bound for each candidate.
|
||||
traverse_candidate(
|
||||
|
@ -234,7 +234,7 @@ impl<'a, 'tcx> Builder<'a, 'tcx> {
|
||||
};
|
||||
TerminatorKind::if_(self.tcx, Operand::Copy(place), true_bb, false_bb)
|
||||
} else {
|
||||
// The switch may be inexhaustive so we have a catch all block
|
||||
// The switch may be inexhaustible so we have a catch all block
|
||||
debug_assert_eq!(options.len() + 1, target_blocks.len());
|
||||
let otherwise_block = *target_blocks.last().unwrap();
|
||||
let switch_targets = SwitchTargets::new(
|
||||
|
@ -567,7 +567,7 @@ impl<'a, 'tcx> Builder<'a, 'tcx> {
|
||||
let tcx = self.tcx;
|
||||
if let LintLevel::Explicit(current_hir_id) = lint_level {
|
||||
// Use `maybe_lint_level_root_bounded` with `root_lint_level` as a bound
|
||||
// to avoid adding Hir dependences on our parents.
|
||||
// to avoid adding Hir dependencies on our parents.
|
||||
// We estimate the true lint roots here to avoid creating a lot of source scopes.
|
||||
|
||||
let parent_root = tcx.maybe_lint_level_root_bounded(
|
||||
@ -965,7 +965,7 @@ impl<'a, 'tcx> Builder<'a, 'tcx> {
|
||||
/// However, `_X` is still registered to be dropped, and so if we
|
||||
/// do nothing else, we would generate a `DROP(_X)` that occurs
|
||||
/// after the call. This will later be optimized out by the
|
||||
/// drop-elaboation code, but in the meantime it can lead to
|
||||
/// drop-elaboration code, but in the meantime it can lead to
|
||||
/// spurious borrow-check errors -- the problem, ironically, is
|
||||
/// not the `DROP(_X)` itself, but the (spurious) unwind pathways
|
||||
/// that it creates. See #64391 for an example.
|
||||
|
@ -1007,7 +1007,7 @@ impl<'tcx> SplitWildcard<'tcx> {
|
||||
{
|
||||
// `usize`/`isize` are not allowed to be matched exhaustively unless the
|
||||
// `precise_pointer_size_matching` feature is enabled. So we treat those types like
|
||||
// `#[non_exhaustive]` enums by returning a special unmatcheable constructor.
|
||||
// `#[non_exhaustive]` enums by returning a special unmatchable constructor.
|
||||
smallvec![NonExhaustive]
|
||||
}
|
||||
&ty::Int(ity) => {
|
||||
@ -1650,7 +1650,7 @@ impl<'p, 'tcx> fmt::Debug for DeconstructedPat<'p, 'tcx> {
|
||||
}
|
||||
|
||||
// Without `cx`, we can't know which field corresponds to which, so we can't
|
||||
// get the names of the fields. Instead we just display everything as a suple
|
||||
// get the names of the fields. Instead we just display everything as a simple
|
||||
// struct, which should be good enough.
|
||||
write!(f, "(")?;
|
||||
for p in self.iter_fields() {
|
||||
|
@ -62,7 +62,7 @@
|
||||
//!
|
||||
//! Note: we will often abbreviate "constructor" as "ctor".
|
||||
//!
|
||||
//! The idea that powers everything that is done in this file is the following: a (matcheable)
|
||||
//! The idea that powers everything that is done in this file is the following: a (matchable)
|
||||
//! value is made from a constructor applied to a number of subvalues. Examples of constructors are
|
||||
//! `Some`, `None`, `(,)` (the 2-tuple constructor), `Foo {..}` (the constructor for a struct
|
||||
//! `Foo`), and `2` (the constructor for the number `2`). This is natural when we think of
|
||||
@ -71,7 +71,7 @@
|
||||
//! Some of the ctors listed above might feel weird: `None` and `2` don't take any arguments.
|
||||
//! That's ok: those are ctors that take a list of 0 arguments; they are the simplest case of
|
||||
//! ctors. We treat `2` as a ctor because `u64` and other number types behave exactly like a huge
|
||||
//! `enum`, with one variant for each number. This allows us to see any matcheable value as made up
|
||||
//! `enum`, with one variant for each number. This allows us to see any matchable value as made up
|
||||
//! from a tree of ctors, each having a set number of children. For example: `Foo { bar: None,
|
||||
//! baz: Ok(0) }` is made from 4 different ctors, namely `Foo{..}`, `None`, `Ok` and `0`.
|
||||
//!
|
||||
@ -342,7 +342,7 @@ pub(super) struct PatCtxt<'a, 'p, 'tcx> {
|
||||
/// Whether the current pattern is the whole pattern as found in a match arm, or if it's a
|
||||
/// subpattern.
|
||||
pub(super) is_top_level: bool,
|
||||
/// Wether the current pattern is from a `non_exhaustive` enum.
|
||||
/// Whether the current pattern is from a `non_exhaustive` enum.
|
||||
pub(super) is_non_exhaustive: bool,
|
||||
}
|
||||
|
||||
|
@ -860,7 +860,7 @@ where
|
||||
// and this is ok because `open_drop` here can only be reached
|
||||
// within that own generator's resume function.
|
||||
// This should only happen for the self argument on the resume function.
|
||||
// It effetively only contains upvars until the generator transformation runs.
|
||||
// It effectively only contains upvars until the generator transformation runs.
|
||||
// See librustc_body/transform/generator.rs for more details.
|
||||
ty::Generator(_, substs, _) => {
|
||||
let tys: Vec<_> = substs.as_generator().upvar_tys().collect();
|
||||
|
@ -40,7 +40,7 @@ use crate::{AnalysisDomain, Backward, CallReturnPlaces, GenKill, GenKillAnalysis
|
||||
///
|
||||
/// This makes `MaybeLiveLocals` unsuitable for certain classes of optimization normally associated
|
||||
/// with a live variables analysis, notably dead-store elimination. It's a dirty hack, but it works
|
||||
/// okay for the generator state transform (currently the main consumuer of this analysis).
|
||||
/// okay for the generator state transform (currently the main consumer of this analysis).
|
||||
///
|
||||
/// [`MaybeBorrowedLocals`]: super::MaybeBorrowedLocals
|
||||
/// [flow-test]: https://github.com/rust-lang/rust/blob/a08c47310c7d49cbdc5d7afb38408ba519967ecd/src/test/ui/mir-dataflow/liveness-ptr.rs
|
||||
|
@ -428,7 +428,7 @@ impl<'tcx> AnalysisDomain<'tcx> for MaybeUninitializedPlaces<'_, 'tcx> {
|
||||
|
||||
// sets on_entry bits for Arg places
|
||||
fn initialize_start_block(&self, _: &mir::Body<'tcx>, state: &mut Self::Domain) {
|
||||
// set all bits to 1 (uninit) before gathering counterevidence
|
||||
// set all bits to 1 (uninit) before gathering counter-evidence
|
||||
state.insert_all();
|
||||
|
||||
drop_flag_effects_for_function_entry(self.tcx, self.body, self.mdpe, |path, s| {
|
||||
|
@ -149,7 +149,7 @@ impl<'tcx> Visitor<'tcx> for UnsafetyChecker<'_, 'tcx> {
|
||||
self.check_mut_borrowing_layout_constrained_field(*place, context.is_mutating_use());
|
||||
}
|
||||
|
||||
// Some checks below need the extra metainfo of the local declaration.
|
||||
// Some checks below need the extra meta info of the local declaration.
|
||||
let decl = &self.body.local_decls[place.local];
|
||||
|
||||
// Check the base local: it might be an unsafe-to-access static. We only check derefs of the
|
||||
|
@ -634,7 +634,7 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
|
||||
}
|
||||
}
|
||||
|
||||
// Attempt to use albegraic identities to eliminate constant expressions
|
||||
// Attempt to use algebraic identities to eliminate constant expressions
|
||||
fn eval_rvalue_with_identities(
|
||||
&mut self,
|
||||
rvalue: &Rvalue<'tcx>,
|
||||
|
@ -66,7 +66,7 @@ impl CoverageVisitor {
|
||||
// The operand ID is outside the known range of counter IDs and also outside the
|
||||
// known range of expression IDs. In either case, the result of a missing operand
|
||||
// (if and when used in an expression) will be zero, so from a computation
|
||||
// perspective, it doesn't matter whether it is interepretted as a counter or an
|
||||
// perspective, it doesn't matter whether it is interpreted as a counter or an
|
||||
// expression.
|
||||
//
|
||||
// However, the `num_counters` and `num_expressions` query results are used to
|
||||
|
@ -694,7 +694,7 @@ impl<'a, 'tcx> CoverageSpans<'a, 'tcx> {
|
||||
/// If prev.span() was split off to the right of a closure, prev.span().lo() will be
|
||||
/// greater than prev_original_span.lo(). The actual span of `prev_original_span` is
|
||||
/// not as important as knowing that `prev()` **used to have the same span** as `curr(),
|
||||
/// which means their sort order is still meaningful for determinating the dominator
|
||||
/// which means their sort order is still meaningful for determining the dominator
|
||||
/// relationship.
|
||||
///
|
||||
/// When two `CoverageSpan`s have the same `Span`, dominated spans can be discarded; but if
|
||||
@ -726,7 +726,7 @@ impl<'a, 'tcx> CoverageSpans<'a, 'tcx> {
|
||||
self.prev()
|
||||
);
|
||||
self.cutoff_prev_at_overlapping_curr();
|
||||
// If one span dominates the other, assocate the span with the code from the dominated
|
||||
// If one span dominates the other, associate the span with the code from the dominated
|
||||
// block only (`curr`), and discard the overlapping portion of the `prev` span. (Note
|
||||
// that if `prev.span` is wider than `prev_original_span`, a `CoverageSpan` will still
|
||||
// be created for `prev`s block, for the non-overlapping portion, left of `curr.span`.)
|
||||
|
@ -260,7 +260,7 @@ fn may_hoist<'tcx>(tcx: TyCtxt<'tcx>, body: &Body<'tcx>, place: Place<'tcx>) ->
|
||||
for (place, proj) in place.iter_projections() {
|
||||
match proj {
|
||||
// Dereferencing in the computation of `place` might cause issues from one of two
|
||||
// cateogires. First, the referrent might be invalid. We protect against this by
|
||||
// categories. First, the referent might be invalid. We protect against this by
|
||||
// dereferencing references only (not pointers). Second, the use of a reference may
|
||||
// invalidate other references that are used later (for aliasing reasons). Consider
|
||||
// where such an invalidated reference may appear:
|
||||
|
@ -500,7 +500,7 @@ fn locals_live_across_suspend_points<'tcx>(
|
||||
// The `liveness` variable contains the liveness of MIR locals ignoring borrows.
|
||||
// This is correct for movable generators since borrows cannot live across
|
||||
// suspension points. However for immovable generators we need to account for
|
||||
// borrows, so we conseratively assume that all borrowed locals are live until
|
||||
// borrows, so we conservatively assume that all borrowed locals are live until
|
||||
// we find a StorageDead statement referencing the locals.
|
||||
// To do this we just union our `liveness` result with `borrowed_locals`, which
|
||||
// contains all the locals which has been borrowed before this suspension point.
|
||||
|
@ -698,7 +698,7 @@ impl<'tcx> Inliner<'tcx> {
|
||||
// The `closure_ref` in our example above.
|
||||
let closure_ref_arg = iter::once(self_);
|
||||
|
||||
// The `tmp0`, `tmp1`, and `tmp2` in our example abonve.
|
||||
// The `tmp0`, `tmp1`, and `tmp2` in our example above.
|
||||
let tuple_tmp_args = tuple_tys.iter().enumerate().map(|(i, ty)| {
|
||||
// This is e.g., `tuple_tmp.0` in our example above.
|
||||
let tuple_field = Operand::Move(tcx.mk_place_field(tuple, Field::new(i), ty));
|
||||
|
@ -8,7 +8,7 @@ use rustc_middle::ty::{self, subst::SubstsRef, InstanceDef, TyCtxt};
|
||||
use rustc_session::Limit;
|
||||
|
||||
// FIXME: check whether it is cheaper to precompute the entire call graph instead of invoking
|
||||
// this query riddiculously often.
|
||||
// this query ridiculously often.
|
||||
#[instrument(level = "debug", skip(tcx, root, target))]
|
||||
crate fn mir_callgraph_reachable<'tcx>(
|
||||
tcx: TyCtxt<'tcx>,
|
||||
|
@ -64,7 +64,7 @@ impl<'tcx> MirPass<'tcx> for RenameReturnPlace {
|
||||
let (renamed_decl, ret_decl) =
|
||||
body.local_decls.pick2_mut(returned_local, mir::RETURN_PLACE);
|
||||
|
||||
// Sometimes, the return place is assigned a local of a different but coercable type, for
|
||||
// Sometimes, the return place is assigned a local of a different but coercible type, for
|
||||
// example `&mut T` instead of `&T`. Overwriting the `LocalInfo` for the return place means
|
||||
// its type may no longer match the return type of its function. This doesn't cause a
|
||||
// problem in codegen because these two types are layout-compatible, but may be unexpected.
|
||||
|
@ -36,7 +36,7 @@ impl RemoveNoopLandingPads {
|
||||
| StatementKind::AscribeUserType(..)
|
||||
| StatementKind::Coverage(..)
|
||||
| StatementKind::Nop => {
|
||||
// These are all nops in a landing pad
|
||||
// These are all noops in a landing pad
|
||||
}
|
||||
|
||||
StatementKind::Assign(box (place, Rvalue::Use(_) | Rvalue::Discriminant(_))) => {
|
||||
|
@ -12,7 +12,7 @@ use crate::MirPass;
|
||||
/// that point.
|
||||
///
|
||||
/// This is redundant with drop elaboration, but we need to do it prior to const-checking, and
|
||||
/// running const-checking after drop elaboration makes it opimization dependent, causing issues
|
||||
/// running const-checking after drop elaboration makes it optimization dependent, causing issues
|
||||
/// like [#90770].
|
||||
///
|
||||
/// [#90770]: https://github.com/rust-lang/rust/issues/90770
|
||||
|
@ -2,7 +2,7 @@
|
||||
//!
|
||||
//! When the MIR is built, we check `needs_drop` before emitting a `Drop` for a place. This pass is
|
||||
//! useful because (unlike MIR building) it runs after type checking, so it can make use of
|
||||
//! `Reveal::All` to provide more precies type information.
|
||||
//! `Reveal::All` to provide more precise type information.
|
||||
|
||||
use crate::MirPass;
|
||||
use rustc_middle::mir::*;
|
||||
|
@ -72,7 +72,7 @@ struct ArmIdentityInfo<'tcx> {
|
||||
/// (StorageLive index,, StorageDead index, Local)
|
||||
storage_stmts: Vec<(usize, usize, Local)>,
|
||||
|
||||
/// The statements that should be removed (turned into nops)
|
||||
/// The statements that should be removed (turned into noops)
|
||||
stmts_to_remove: Vec<usize>,
|
||||
|
||||
/// Indices of debug variables that need to be adjusted to point to
|
||||
|
@ -501,7 +501,7 @@ fn mono_item_visibility<'tcx>(
|
||||
// * First is weak lang items. These are basically mechanisms for
|
||||
// libcore to forward-reference symbols defined later in crates like
|
||||
// the standard library or `#[panic_handler]` definitions. The
|
||||
// definition of these weak lang items needs to be referenceable by
|
||||
// definition of these weak lang items needs to be referencable by
|
||||
// libcore, so we're no longer a candidate for internalization.
|
||||
// Removal of these functions can't be done by LLVM but rather must be
|
||||
// done by the linker as it's a non-local decision.
|
||||
|
@ -65,7 +65,7 @@ fn unused_generic_params<'tcx>(
|
||||
mark_used_by_default_parameters(tcx, def_id, generics, &mut unused_parameters);
|
||||
debug!(?unused_parameters, "(after default)");
|
||||
|
||||
// Visit MIR and accumululate used generic parameters.
|
||||
// Visit MIR and accumulate used generic parameters.
|
||||
let body = match tcx.hir().body_const_context(def_id.expect_local()) {
|
||||
// Const functions are actually called and should thus be considered for polymorphization
|
||||
// via their runtime MIR.
|
||||
|
@ -203,7 +203,7 @@ impl<'a> Parser<'a> {
|
||||
&self.sess.span_diagnostic
|
||||
}
|
||||
|
||||
/// Relace `self` with `snapshot.parser` and extend `unclosed_delims` with `snapshot.unclosed_delims`.
|
||||
/// Replace `self` with `snapshot.parser` and extend `unclosed_delims` with `snapshot.unclosed_delims`.
|
||||
/// This is to avoid losing unclosed delims errors `create_snapshot_for_diagnostic` clears.
|
||||
pub(super) fn restore_snapshot(&mut self, snapshot: SnapshotParser<'a>) {
|
||||
*self = snapshot.parser;
|
||||
|
@ -276,7 +276,7 @@ impl<'a> Parser<'a> {
|
||||
lhs = self.parse_assoc_op_ascribe(lhs, lhs_span)?;
|
||||
continue;
|
||||
} else if op == AssocOp::DotDot || op == AssocOp::DotDotEq {
|
||||
// If we didn’t have to handle `x..`/`x..=`, it would be pretty easy to
|
||||
// If we did not have to handle `x..`/`x..=`, it would be pretty easy to
|
||||
// generalise it to the Fixity::None code.
|
||||
lhs = self.parse_range_expr(prec, lhs, op, cur_op_span)?;
|
||||
break;
|
||||
|
@ -150,7 +150,7 @@ pub struct Parser<'a> {
|
||||
pub current_closure: Option<ClosureSpans>,
|
||||
}
|
||||
|
||||
/// Stores span informations about a closure.
|
||||
/// Stores span information about a closure.
|
||||
#[derive(Clone)]
|
||||
pub struct ClosureSpans {
|
||||
pub whole_closure: Span,
|
||||
|
@ -729,7 +729,7 @@ impl<'a> Parser<'a> {
|
||||
|
||||
/// Finds the indices of all characters that have been processed and differ between the actual
|
||||
/// written code (code snippet) and the `InternedString` that gets processed in the `Parser`
|
||||
/// in order to properly synthethise the intra-string `Span`s for error diagnostics.
|
||||
/// in order to properly synthesise the intra-string `Span`s for error diagnostics.
|
||||
fn find_skips_from_snippet(
|
||||
snippet: Option<string::String>,
|
||||
str_style: Option<usize>,
|
||||
|
@ -99,7 +99,7 @@ impl<'tcx> hir::itemlikevisit::ItemLikeVisitor<'tcx> for CheckConstTraitVisitor<
|
||||
} = *trait_item
|
||||
{
|
||||
// we can ignore functions that do not have default bodies:
|
||||
// if those are unimplemented it will be catched by typeck.
|
||||
// if those are unimplemented it will be caught by typeck.
|
||||
if !defaultness.has_value()
|
||||
|| self
|
||||
.tcx
|
||||
|
@ -291,7 +291,7 @@ fn resolve_expr<'tcx>(visitor: &mut RegionResolutionVisitor<'tcx>, expr: &'tcx h
|
||||
// like AddAssign is implemented).
|
||||
|
||||
// For primitive types (which, despite having a trait impl, don't actually
|
||||
// end up calling it), the evluation order is right-to-left. For example,
|
||||
// end up calling it), the evaluation order is right-to-left. For example,
|
||||
// the following code snippet:
|
||||
//
|
||||
// let y = &mut 0;
|
||||
|
@ -144,7 +144,7 @@ impl<'a, 'tcx> Annotator<'a, 'tcx> {
|
||||
}
|
||||
|
||||
if !self.tcx.features().staged_api {
|
||||
// Propagate unstability. This can happen even for non-staged-api crates in case
|
||||
// Propagate instability. This can happen even for non-staged-api crates in case
|
||||
// -Zforce-unstable-if-unmarked is set.
|
||||
if let Some(stab) = self.parent_stab {
|
||||
if inherit_deprecation.yes() && stab.level.is_unstable() {
|
||||
@ -344,7 +344,7 @@ impl<'a, 'tcx> Visitor<'tcx> for Annotator<'a, 'tcx> {
|
||||
match i.kind {
|
||||
// Inherent impls and foreign modules serve only as containers for other items,
|
||||
// they don't have their own stability. They still can be annotated as unstable
|
||||
// and propagate this unstability to children, but this annotation is completely
|
||||
// and propagate this instability to children, but this annotation is completely
|
||||
// optional. They inherit stability from their parents when unannotated.
|
||||
hir::ItemKind::Impl(hir::Impl { of_trait: None, .. })
|
||||
| hir::ItemKind::ForeignMod { .. } => {
|
||||
@ -557,7 +557,7 @@ impl<'tcx> Visitor<'tcx> for MissingStabilityAnnotations<'tcx> {
|
||||
fn visit_item(&mut self, i: &'tcx Item<'tcx>) {
|
||||
// Inherent impls and foreign modules serve only as containers for other items,
|
||||
// they don't have their own stability. They still can be annotated as unstable
|
||||
// and propagate this unstability to children, but this annotation is completely
|
||||
// and propagate this instability to children, but this annotation is completely
|
||||
// optional. They inherit stability from their parents when unannotated.
|
||||
if !matches!(
|
||||
i.kind,
|
||||
|
@ -559,9 +559,9 @@ impl<'tcx> EmbargoVisitor<'tcx> {
|
||||
}
|
||||
}
|
||||
|
||||
// Hygine isn't really implemented for `macro_rules!` macros at the
|
||||
// Hygiene isn't really implemented for `macro_rules!` macros at the
|
||||
// moment. Accordingly, marking them as reachable is unwise. `macro` macros
|
||||
// have normal hygine, so we can treat them like other items without type
|
||||
// have normal hygiene, so we can treat them like other items without type
|
||||
// privacy and mark them reachable.
|
||||
DefKind::Macro(_) => {
|
||||
let item = self.tcx.hir().expect_item(def_id);
|
||||
@ -1539,7 +1539,7 @@ impl<'a, 'tcx> Visitor<'tcx> for ObsoleteVisiblePrivateTypesVisitor<'a, 'tcx> {
|
||||
// 3. mentioned in the associated types of the impl
|
||||
//
|
||||
// Those in 1. can only occur if the trait is in
|
||||
// this crate and will've been warned about on the
|
||||
// this crate and will have been warned about on the
|
||||
// trait definition (there's no need to warn twice
|
||||
// so we don't check the methods).
|
||||
//
|
||||
@ -1999,7 +1999,7 @@ fn visibility(tcx: TyCtxt<'_>, def_id: DefId) -> ty::Visibility {
|
||||
}
|
||||
// - AST lowering may clone `use` items and the clones don't
|
||||
// get their entries in the resolver's visibility table.
|
||||
// - AST lowering also creates opaque type items with inherited visibilies.
|
||||
// - AST lowering also creates opaque type items with inherited visibilities.
|
||||
// Visibility on them should have no effect, but to avoid the visibility
|
||||
// query failing on some items, we provide it for opaque types as well.
|
||||
Node::Item(hir::Item {
|
||||
|
@ -7,7 +7,7 @@
|
||||
//! The serialisation is performed on-demand when each node is emitted. Using this
|
||||
//! scheme, we do not need to keep the current graph in memory.
|
||||
//!
|
||||
//! The deserisalisation is performed manually, in order to convert from the stored
|
||||
//! The deserialization is performed manually, in order to convert from the stored
|
||||
//! sequence of NodeInfos to the different arrays in SerializedDepGraph. Since the
|
||||
//! node and edge count are stored at the end of the file, all the arrays can be
|
||||
//! pre-allocated with the right length.
|
||||
|
@ -1603,7 +1603,7 @@ impl<'a, 'b> ImportResolver<'a, 'b> {
|
||||
// Remove the `removal_span`.
|
||||
corrections.push((removal_span, "".to_string()));
|
||||
|
||||
// Find the span after the crate name and if it has nested imports immediatately
|
||||
// Find the span after the crate name and if it has nested imports immediately
|
||||
// after the crate name already.
|
||||
// ie. `use a::b::{c, d};`
|
||||
// ^^^^^^^^^
|
||||
|
@ -74,7 +74,7 @@ crate struct Import<'a> {
|
||||
/// In the case where the `Import` was expanded from a "nested" use tree,
|
||||
/// this id is the ID of the leaf tree. For example:
|
||||
///
|
||||
/// ```ignore (pacify the mercilous tidy)
|
||||
/// ```ignore (pacify the merciless tidy)
|
||||
/// use foo::bar::{a, b}
|
||||
/// ```
|
||||
///
|
||||
|
@ -2151,7 +2151,7 @@ impl<'tcx> LifetimeContext<'_, 'tcx> {
|
||||
// but if we make a mistake elsewhere, mainly by keeping something in
|
||||
// `missing_named_lifetime_spots` that we shouldn't, like associated
|
||||
// `const`s or making a mistake in the AST lowering we would provide
|
||||
// non-sensical suggestions. Guard against that by skipping these.
|
||||
// nonsensical suggestions. Guard against that by skipping these.
|
||||
// (#74264)
|
||||
continue;
|
||||
}
|
||||
|
@ -261,7 +261,7 @@ enum Scope<'a> {
|
||||
s: ScopeRef<'a>,
|
||||
},
|
||||
|
||||
/// When we have nested trait refs, we concanetate late bound vars for inner
|
||||
/// When we have nested trait refs, we concatenate late bound vars for inner
|
||||
/// trait refs from outer ones. But we also need to include any HRTB
|
||||
/// lifetimes encountered when identifying the trait that an associated type
|
||||
/// is declared on.
|
||||
@ -282,8 +282,8 @@ enum BinderScopeType {
|
||||
/// Any non-concatenating binder scopes.
|
||||
Normal,
|
||||
/// Within a syntactic trait ref, there may be multiple poly trait refs that
|
||||
/// are nested (under the `associcated_type_bounds` feature). The binders of
|
||||
/// the innner poly trait refs are extended from the outer poly trait refs
|
||||
/// are nested (under the `associated_type_bounds` feature). The binders of
|
||||
/// the inner poly trait refs are extended from the outer poly trait refs
|
||||
/// and don't increase the late bound depth. If you had
|
||||
/// `T: for<'a> Foo<Bar: for<'b> Baz<'a, 'b>>`, then the `for<'b>` scope
|
||||
/// would be `Concatenating`. This also used in trait refs in where clauses
|
||||
@ -2930,7 +2930,7 @@ impl<'a, 'tcx> LifetimeContext<'a, 'tcx> {
|
||||
path_span: Span,
|
||||
generic_args: &'v hir::GenericArgs<'v>,
|
||||
) {
|
||||
// parenthesized args enter a new elison scope
|
||||
// parenthesized args enter a new elision scope
|
||||
if generic_args.parenthesized {
|
||||
return;
|
||||
}
|
||||
|
@ -506,7 +506,7 @@ impl ModuleKind {
|
||||
/// program) if all but one of them come from glob imports.
|
||||
#[derive(Copy, Clone, PartialEq, Eq, Hash, Debug)]
|
||||
struct BindingKey {
|
||||
/// The identifier for the binding, aways the `normalize_to_macros_2_0` version of the
|
||||
/// The identifier for the binding, always the `normalize_to_macros_2_0` version of the
|
||||
/// identifier.
|
||||
ident: Ident,
|
||||
ns: Namespace,
|
||||
|
@ -69,7 +69,7 @@ pub enum MacroRulesScope<'a> {
|
||||
/// The reason is that we update scopes with value `MacroRulesScope::Invocation(invoc_id)`
|
||||
/// in-place after `invoc_id` gets expanded.
|
||||
/// This helps to avoid uncontrollable growth of `macro_rules!` scope chains,
|
||||
/// which usually grow lineraly with the number of macro invocations
|
||||
/// which usually grow linearly with the number of macro invocations
|
||||
/// in a module (including derives) and hurt performance.
|
||||
pub(crate) type MacroRulesScopeRef<'a> = Interned<'a, Cell<MacroRulesScope<'a>>>;
|
||||
|
||||
|
@ -1437,7 +1437,7 @@ impl<'tcx> Visitor<'tcx> for DumpVisitor<'tcx> {
|
||||
self.process_macro_use(l.span);
|
||||
self.process_var_decl(&l.pat);
|
||||
|
||||
// Just walk the initialiser and type (don't want to walk the pattern again).
|
||||
// Just walk the initializer and type (don't want to walk the pattern again).
|
||||
walk_list!(self, visit_ty, &l.ty);
|
||||
walk_list!(self, visit_expr, &l.init);
|
||||
}
|
||||
|
@ -1177,7 +1177,7 @@ pub struct HygieneEncodeContext {
|
||||
/// that we don't accidentally try to encode any more `SyntaxContexts`
|
||||
serialized_ctxts: Lock<FxHashSet<SyntaxContext>>,
|
||||
/// The `SyntaxContexts` that we have serialized (e.g. as a result of encoding `Spans`)
|
||||
/// in the most recent 'round' of serializnig. Serializing `SyntaxContextData`
|
||||
/// in the most recent 'round' of serializing. Serializing `SyntaxContextData`
|
||||
/// may cause us to serialize more `SyntaxContext`s, so serialize in a loop
|
||||
/// until we reach a fixed point.
|
||||
latest_ctxts: Lock<FxHashSet<SyntaxContext>>,
|
||||
|
@ -1637,7 +1637,7 @@ impl fmt::Display for Ident {
|
||||
/// hygiene data, most importantly name of the crate it refers to.
|
||||
/// As a result we print `$crate` as `crate` if it refers to the local crate
|
||||
/// and as `::other_crate_name` if it refers to some other crate.
|
||||
/// Note, that this is only done if the ident token is printed from inside of AST pretty-pringing,
|
||||
/// Note, that this is only done if the ident token is printed from inside of AST pretty-printing,
|
||||
/// but not otherwise. Pretty-printing is the only way for proc macros to discover token contents,
|
||||
/// so we should not perform this lossy conversion if the top level call to the pretty-printer was
|
||||
/// done for a token stream or a single token.
|
||||
@ -1812,7 +1812,7 @@ pub(crate) struct Interner(Lock<InternerInner>);
|
||||
// revisited after further improvements to `indexmap`.
|
||||
//
|
||||
// This type is private to prevent accidentally constructing more than one
|
||||
// `Interner` on the same thread, which makes it easy to mixup `Symbol`s
|
||||
// `Interner` on the same thread, which makes it easy to mix up `Symbol`s
|
||||
// between `Interner`s.
|
||||
#[derive(Default)]
|
||||
struct InternerInner {
|
||||
|
@ -159,7 +159,7 @@ where
|
||||
// Unions and are always treated as a series of 64-bit integer chunks
|
||||
}
|
||||
abi::FieldsShape::Arbitrary { .. } => {
|
||||
// Stuctures with floating point numbers need special care.
|
||||
// Structures with floating point numbers need special care.
|
||||
|
||||
let mut data = parse_structure(
|
||||
cx,
|
||||
|
@ -49,7 +49,7 @@ pub fn target() -> Target {
|
||||
// setLibcallCallingConv(RTLIB::UREM_I64, CallingConv::X86_StdCall);
|
||||
// setLibcallCallingConv(RTLIB::MUL_I64, CallingConv::X86_StdCall);
|
||||
// }
|
||||
// The compiler intrisics should be implemented by compiler-builtins.
|
||||
// The compiler intrinsics should be implemented by compiler-builtins.
|
||||
// Unfortunately, compiler-builtins has not provided those intrinsics yet. Such as:
|
||||
// i386/divdi3.S
|
||||
// i386/lshrdi3.S
|
||||
@ -64,7 +64,7 @@ pub fn target() -> Target {
|
||||
// 2. Implement Intrinsics.
|
||||
// We evaluated all options.
|
||||
// #2 is hard because we need implement the intrinsics (_aulldiv) generated
|
||||
// from the other intrinscis (__udivdi3) implementation with the same
|
||||
// from the other intrinsics (__udivdi3) implementation with the same
|
||||
// functionality (udivmod_inner). If we let _aulldiv() call udivmod_inner!(),
|
||||
// then we are in loop. We may have to find another way to implement udivmod_inner!().
|
||||
// #1.2 may break the existing usage.
|
||||
@ -73,7 +73,7 @@ pub fn target() -> Target {
|
||||
// It uses cdecl, EAX/ECX/EDX as volatile register, and EAX/EDX as return value.
|
||||
// We also checked the LLVM X86TargetLowering, the differences between -gnu and -msvc
|
||||
// is fmodf(f32), longjmp() and TLS. None of them impacts the UEFI code.
|
||||
// As a result, we choose -gnu for i686 version before those intrisics are implemented in
|
||||
// As a result, we choose -gnu for i686 version before those intrinsics are implemented in
|
||||
// compiler-builtins. After compiler-builtins implements all required intrinsics, we may
|
||||
// remove -gnu and use the default one.
|
||||
Target {
|
||||
|
@ -1147,7 +1147,7 @@ pub struct TargetOptions {
|
||||
/// Linker arguments used in addition to `late_link_args` if at least one
|
||||
/// Rust dependency is dynamically linked.
|
||||
pub late_link_args_dynamic: LinkArgs,
|
||||
/// Linker arguments used in addition to `late_link_args` if aall Rust
|
||||
/// Linker arguments used in addition to `late_link_args` if all Rust
|
||||
/// dependencies are statically linked.
|
||||
pub late_link_args_static: LinkArgs,
|
||||
/// Linker arguments that are unconditionally passed *after* any
|
||||
@ -1255,7 +1255,7 @@ pub struct TargetOptions {
|
||||
/// handling COFF object files with more than 2<sup>15</sup> sections. Since each weak
|
||||
/// symbol needs its own COMDAT section, weak linkage implies a large
|
||||
/// number sections that easily exceeds the given limit for larger
|
||||
/// codebases. Consequently we want a way to disallow weak linkage on some
|
||||
/// codebase. Consequently we want a way to disallow weak linkage on some
|
||||
/// platforms.
|
||||
pub allows_weak_linkage: bool,
|
||||
/// Whether the linker support rpaths or not. Defaults to false.
|
||||
|
@ -109,7 +109,7 @@ pub fn options() -> TargetOptions {
|
||||
crt_objects_fallback: Some(CrtObjectsFallback::Wasm),
|
||||
|
||||
// This has no effect in LLVM 8 or prior, but in LLVM 9 and later when
|
||||
// PIC code is implemented this has quite a drastric effect if it stays
|
||||
// PIC code is implemented this has quite a drastic effect if it stays
|
||||
// at the default, `pic`. In an effort to keep wasm binaries as minimal
|
||||
// as possible we're defaulting to `static` for now, but the hope is
|
||||
// that eventually we can ship a `pic`-compatible standard library which
|
||||
|
@ -139,7 +139,7 @@ impl<'tcx> TypeFolder<'tcx> for ReverseMapper<'tcx> {
|
||||
match *r {
|
||||
// Ignore bound regions and `'static` regions that appear in the
|
||||
// type, we only need to remap regions that reference lifetimes
|
||||
// from the function declaraion.
|
||||
// from the function declaration.
|
||||
// This would ignore `'r` in a type like `for<'r> fn(&'r u32)`.
|
||||
ty::ReLateBound(..) | ty::ReStatic => return r,
|
||||
|
||||
|
@ -46,8 +46,8 @@ pub fn codegen_fulfill_obligation<'tcx>(
|
||||
Ok(Some(selection)) => selection,
|
||||
Ok(None) => {
|
||||
// Ambiguity can happen when monomorphizing during trans
|
||||
// expands to some humongo type that never occurred
|
||||
// statically -- this humongo type can then overflow,
|
||||
// expands to some humongous type that never occurred
|
||||
// statically -- this humongous type can then overflow,
|
||||
// leading to an ambiguous result. So report this as an
|
||||
// overflow bug, since I believe this is the only case
|
||||
// where ambiguity can result.
|
||||
|
@ -114,7 +114,7 @@ where
|
||||
}
|
||||
|
||||
// In the case where we detect an error, run the check again, but
|
||||
// this time tracking intercrate ambuiguity causes for better
|
||||
// this time tracking intercrate ambiguity causes for better
|
||||
// diagnostics. (These take time and can lead to false errors.)
|
||||
tcx.infer_ctxt().enter(|infcx| {
|
||||
let selcx = &mut SelectionContext::intercrate(&infcx);
|
||||
@ -762,7 +762,7 @@ fn ty_is_local_constructor(ty: Ty<'_>, in_crate: InCrate) -> bool {
|
||||
ty::Foreign(did) => def_id_is_local(did, in_crate),
|
||||
ty::Opaque(..) => {
|
||||
// This merits some explanation.
|
||||
// Normally, opaque types are not involed when performing
|
||||
// Normally, opaque types are not involved when performing
|
||||
// coherence checking, since it is illegal to directly
|
||||
// implement a trait on an opaque type. However, we might
|
||||
// end up looking at an opaque type during coherence checking
|
||||
|
@ -408,7 +408,7 @@ impl<'a, 'tcx> AbstractConstBuilder<'a, 'tcx> {
|
||||
}
|
||||
|
||||
/// Builds the abstract const by walking the thir and bailing out when
|
||||
/// encountering an unspported operation.
|
||||
/// encountering an unsupported operation.
|
||||
fn build(mut self) -> Result<&'tcx [Node<'tcx>], ErrorGuaranteed> {
|
||||
debug!("Abstractconstbuilder::build: body={:?}", &*self.body);
|
||||
self.recurse_build(self.body_id)?;
|
||||
@ -701,7 +701,7 @@ struct ConstUnifyCtxt<'tcx> {
|
||||
|
||||
impl<'tcx> ConstUnifyCtxt<'tcx> {
|
||||
// Substitutes generics repeatedly to allow AbstractConsts to unify where a
|
||||
// ConstKind::Unevalated could be turned into an AbstractConst that would unify e.g.
|
||||
// ConstKind::Unevaluated could be turned into an AbstractConst that would unify e.g.
|
||||
// Param(N) should unify with Param(T), substs: [Unevaluated("T2", [Unevaluated("T3", [Param(N)])])]
|
||||
#[inline]
|
||||
#[instrument(skip(self), level = "debug")]
|
||||
|
@ -230,7 +230,7 @@ fn suggest_restriction<'tcx>(
|
||||
{
|
||||
// We know we have an `impl Trait` that doesn't satisfy a required projection.
|
||||
|
||||
// Find all of the ocurrences of `impl Trait` for `Trait` in the function arguments'
|
||||
// Find all of the occurrences of `impl Trait` for `Trait` in the function arguments'
|
||||
// types. There should be at least one, but there might be *more* than one. In that
|
||||
// case we could just ignore it and try to identify which one needs the restriction,
|
||||
// but instead we choose to suggest replacing all instances of `impl Trait` with `T`
|
||||
|
@ -61,7 +61,7 @@ pub struct FulfillmentContext<'tcx> {
|
||||
// Should this fulfillment context register type-lives-for-region
|
||||
// obligations on its parent infcx? In some cases, region
|
||||
// obligations are either already known to hold (normalization) or
|
||||
// hopefully verifed elsewhere (type-impls-bound), and therefore
|
||||
// hopefully verified elsewhere (type-impls-bound), and therefore
|
||||
// should not be checked.
|
||||
//
|
||||
// Note that if we are normalizing a type that we already
|
||||
|
@ -539,7 +539,7 @@ fn prepare_vtable_segments<'tcx, T>(
|
||||
// the main traversal loop:
|
||||
// basically we want to cut the inheritance directed graph into a few non-overlapping slices of nodes
|
||||
// that each node is emitted after all its descendents have been emitted.
|
||||
// so we convert the directed graph into a tree by skipping all previously visted nodes using a visited set.
|
||||
// so we convert the directed graph into a tree by skipping all previously visited nodes using a visited set.
|
||||
// this is done on the fly.
|
||||
// Each loop run emits a slice - it starts by find a "childless" unvisited node, backtracking upwards, and it
|
||||
// stops after it finds a node that has a next-sibling node.
|
||||
|
@ -1566,7 +1566,7 @@ fn assemble_candidates_from_impls<'cx, 'tcx>(
|
||||
// fn bar<T:SomeTrait<Foo=usize>>(...) { ... }
|
||||
// ```
|
||||
//
|
||||
// Doesn't the `T : Sometrait<Foo=usize>` predicate help
|
||||
// Doesn't the `T : SomeTrait<Foo=usize>` predicate help
|
||||
// resolve `T::Foo`? And of course it does, but in fact
|
||||
// that single predicate is desugared into two predicates
|
||||
// in the compiler: a trait predicate (`T : SomeTrait`) and a
|
||||
@ -1989,7 +1989,7 @@ fn confirm_impl_candidate<'cx, 'tcx>(
|
||||
// Get obligations corresponding to the predicates from the where-clause of the
|
||||
// associated type itself.
|
||||
// Note: `feature(generic_associated_types)` is required to write such
|
||||
// predicates, even for non-generic associcated types.
|
||||
// predicates, even for non-generic associated types.
|
||||
fn assoc_ty_own_obligations<'cx, 'tcx>(
|
||||
selcx: &mut SelectionContext<'cx, 'tcx>,
|
||||
obligation: &ProjectionTyObligation<'tcx>,
|
||||
|
@ -58,7 +58,7 @@ impl<'cx, 'tcx> AtExt<'tcx> for At<'cx, 'tcx> {
|
||||
return InferOk { value: kinds, obligations };
|
||||
}
|
||||
|
||||
// Errors and ambiuity in dropck occur in two cases:
|
||||
// Errors and ambiguity in dropck occur in two cases:
|
||||
// - unresolved inference variables at the end of typeck
|
||||
// - non well-formed types where projections cannot be resolved
|
||||
// Either of these should have created an error before.
|
||||
|
@ -23,7 +23,7 @@ pub(crate) fn update<'tcx, T>(
|
||||
..tpred.trait_ref
|
||||
};
|
||||
|
||||
// Then contstruct a new obligation with Self = () added
|
||||
// Then construct a new obligation with Self = () added
|
||||
// to the ParamEnv, and see if it holds.
|
||||
let o = rustc_infer::traits::Obligation::new(
|
||||
ObligationCause::dummy(),
|
||||
|
@ -562,7 +562,7 @@ impl<'cx, 'tcx> SelectionContext<'cx, 'tcx> {
|
||||
{
|
||||
// If the result is something that we can cache, then mark this
|
||||
// entry as 'complete'. This will allow us to skip evaluating the
|
||||
// suboligations at all the next time we evaluate the projection
|
||||
// subobligations at all the next time we evaluate the projection
|
||||
// predicate.
|
||||
self.infcx
|
||||
.inner
|
||||
@ -751,7 +751,7 @@ impl<'cx, 'tcx> SelectionContext<'cx, 'tcx> {
|
||||
// provisional caches entries and inserting them into the evaluation cache
|
||||
//
|
||||
// This ensures that when a query reads this entry from the evaluation cache,
|
||||
// it will end up (transitively) dependening on all of the incr-comp dependencies
|
||||
// it will end up (transitively) depending on all of the incr-comp dependencies
|
||||
// created during the evaluation of this trait. For example, evaluating a trait
|
||||
// will usually require us to invoke `type_of(field_def_id)` to determine the
|
||||
// constituent types, and we want any queries reading from this evaluation
|
||||
@ -1475,7 +1475,7 @@ impl<'cx, 'tcx> SelectionContext<'cx, 'tcx> {
|
||||
/// `No` if it does not. Return `Ambiguous` in the case that the projection type is a GAT,
|
||||
/// and applying this env_predicate constrains any of the obligation's GAT substitutions.
|
||||
///
|
||||
/// This behavior is a somewhat of a hack to prevent overconstraining inference variables
|
||||
/// This behavior is a somewhat of a hack to prevent over-constraining inference variables
|
||||
/// in cases like #91762.
|
||||
pub(super) fn match_projection_projections(
|
||||
&mut self,
|
||||
@ -1729,7 +1729,7 @@ impl<'cx, 'tcx> SelectionContext<'cx, 'tcx> {
|
||||
// marker trait impls.
|
||||
//
|
||||
// Without this restriction, we could end up accidentally
|
||||
// constrainting inference variables based on an arbitrarily
|
||||
// constraining inference variables based on an arbitrarily
|
||||
// chosen trait impl.
|
||||
//
|
||||
// Imagine we have the following code:
|
||||
@ -1758,7 +1758,7 @@ impl<'cx, 'tcx> SelectionContext<'cx, 'tcx> {
|
||||
// some other means (e.g. type-checking of a function). We will
|
||||
// then be in a position to drop marker trait candidates
|
||||
// without constraining inference variables (since there are
|
||||
// none left to constrin)
|
||||
// none left to constrain)
|
||||
// 2) Be left with some unconstrained inference variables. We
|
||||
// will then correctly report an inference error, since the
|
||||
// existence of multiple marker trait impls tells us nothing
|
||||
@ -2518,7 +2518,7 @@ struct ProvisionalEvaluationCache<'tcx> {
|
||||
/// - `A B C` and we add a cache for the result of C (DFN 2)
|
||||
/// - Then we have a stack `A B D` where `D` has DFN 3
|
||||
/// - We try to solve D by evaluating E: `A B D E` (DFN 4)
|
||||
/// - `E` generates various cache entries which have cyclic dependices on `B`
|
||||
/// - `E` generates various cache entries which have cyclic dependencies on `B`
|
||||
/// - `A B D E F` and so forth
|
||||
/// - the DFN of `F` for example would be 5
|
||||
/// - then we determine that `E` is in error -- we will then clear
|
||||
|
@ -345,7 +345,7 @@ impl<'o, 'tcx> dyn AstConv<'tcx> + 'o {
|
||||
if potential_assoc_types.len() == assoc_items.len() {
|
||||
// Only suggest when the amount of missing associated types equals the number of
|
||||
// extra type arguments present, as that gives us a relatively high confidence
|
||||
// that the user forgot to give the associtated type's name. The canonical
|
||||
// that the user forgot to give the associated type's name. The canonical
|
||||
// example would be trying to use `Iterator<isize>` instead of
|
||||
// `Iterator<Item = isize>`.
|
||||
for (potential, item) in iter::zip(&potential_assoc_types, assoc_items) {
|
||||
|
@ -290,7 +290,7 @@ impl<'o, 'tcx> dyn AstConv<'tcx> + 'o {
|
||||
/// Given the type/lifetime/const arguments provided to some path (along with
|
||||
/// an implicit `Self`, if this is a trait reference), returns the complete
|
||||
/// set of substitutions. This may involve applying defaulted type parameters.
|
||||
/// Constraints on associated typess are created from `create_assoc_bindings_for_generic_args`.
|
||||
/// Constraints on associated types are created from `create_assoc_bindings_for_generic_args`.
|
||||
///
|
||||
/// Example:
|
||||
///
|
||||
@ -1435,7 +1435,7 @@ impl<'o, 'tcx> dyn AstConv<'tcx> + 'o {
|
||||
// If the projection output contains `Self`, force the user to
|
||||
// elaborate it explicitly to avoid a lot of complexity.
|
||||
//
|
||||
// The "classicaly useful" case is the following:
|
||||
// The "classically useful" case is the following:
|
||||
// ```
|
||||
// trait MyTrait: FnMut() -> <Self as MyTrait>::MyOutput {
|
||||
// type MyOutput;
|
||||
@ -1768,7 +1768,7 @@ impl<'o, 'tcx> dyn AstConv<'tcx> + 'o {
|
||||
// Will fail except for `T::A` and `Self::A`; i.e., if `qself_ty`/`qself_def` are not a type
|
||||
// parameter or `Self`.
|
||||
// NOTE: When this function starts resolving `Trait::AssocTy` successfully
|
||||
// it should also start reportint the `BARE_TRAIT_OBJECTS` lint.
|
||||
// it should also start reporting the `BARE_TRAIT_OBJECTS` lint.
|
||||
pub fn associated_path_to_ty(
|
||||
&self,
|
||||
hir_ref_id: hir::HirId,
|
||||
|
@ -721,7 +721,7 @@ impl<'a, 'tcx> FnCtxt<'a, 'tcx> {
|
||||
return None;
|
||||
}
|
||||
|
||||
// The `Future` trait has only one associted item, `Output`,
|
||||
// The `Future` trait has only one associated item, `Output`,
|
||||
// so check that this is what we see.
|
||||
let output_assoc_item = self.tcx.associated_item_def_ids(future_trait)[0];
|
||||
if output_assoc_item != predicate.projection_ty.item_def_id {
|
||||
|
@ -18,7 +18,7 @@
|
||||
//!
|
||||
//! ## Subtle note
|
||||
//!
|
||||
//! When infering the generic arguments of functions, the argument
|
||||
//! When inferring the generic arguments of functions, the argument
|
||||
//! order is relevant, which can lead to the following edge case:
|
||||
//!
|
||||
//! ```rust
|
||||
|
@ -590,7 +590,7 @@ impl<'a, 'tcx> FnCtxt<'a, 'tcx> {
|
||||
let opt_coerce_to = {
|
||||
// We should release `enclosing_breakables` before the `check_expr_with_hint`
|
||||
// below, so can't move this block of code to the enclosing scope and share
|
||||
// `ctxt` with the second `encloding_breakables` borrow below.
|
||||
// `ctxt` with the second `enclosing_breakables` borrow below.
|
||||
let mut enclosing_breakables = self.enclosing_breakables.borrow_mut();
|
||||
match enclosing_breakables.opt_find_breakable(target_id) {
|
||||
Some(ctxt) => ctxt.coerce.as_ref().map(|coerce| coerce.expected_ty()),
|
||||
@ -793,7 +793,7 @@ impl<'a, 'tcx> FnCtxt<'a, 'tcx> {
|
||||
self.tcx.types.never
|
||||
}
|
||||
|
||||
/// `explicit_return` is `true` if we're checkng an explicit `return expr`,
|
||||
/// `explicit_return` is `true` if we're checking an explicit `return expr`,
|
||||
/// and `false` if we're checking a trailing expression.
|
||||
pub(super) fn check_return_expr(
|
||||
&self,
|
||||
@ -1126,7 +1126,7 @@ impl<'a, 'tcx> FnCtxt<'a, 'tcx> {
|
||||
let method = match self.lookup_method(rcvr_t, segment, span, expr, rcvr, args) {
|
||||
Ok(method) => {
|
||||
// We could add a "consider `foo::<params>`" suggestion here, but I wasn't able to
|
||||
// trigger this codepath causing `structuraly_resolved_type` to emit an error.
|
||||
// trigger this codepath causing `structurally_resolved_type` to emit an error.
|
||||
|
||||
self.write_method_call(expr.hir_id, method);
|
||||
Ok(method)
|
||||
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue
Block a user