mirror of
https://github.com/rust-lang/rust.git
synced 2025-06-04 19:29:07 +00:00
Merge #1548
1548: use Parse in mbe r=matklad a=matklad Co-authored-by: Aleksey Kladov <aleksey.kladov@gmail.com>
This commit is contained in:
commit
58d4983ba5
@ -5,7 +5,7 @@ mod input;
|
|||||||
use std::{panic, sync::Arc};
|
use std::{panic, sync::Arc};
|
||||||
|
|
||||||
use ra_prof::profile;
|
use ra_prof::profile;
|
||||||
use ra_syntax::{Parse, SourceFile, TextRange, TextUnit};
|
use ra_syntax::{ast, Parse, SourceFile, TextRange, TextUnit};
|
||||||
use relative_path::RelativePathBuf;
|
use relative_path::RelativePathBuf;
|
||||||
|
|
||||||
pub use crate::{
|
pub use crate::{
|
||||||
@ -74,7 +74,7 @@ pub trait SourceDatabase: CheckCanceled + std::fmt::Debug {
|
|||||||
fn file_text(&self, file_id: FileId) -> Arc<String>;
|
fn file_text(&self, file_id: FileId) -> Arc<String>;
|
||||||
// Parses the file into the syntax tree.
|
// Parses the file into the syntax tree.
|
||||||
#[salsa::invoke(parse_query)]
|
#[salsa::invoke(parse_query)]
|
||||||
fn parse(&self, file_id: FileId) -> Parse;
|
fn parse(&self, file_id: FileId) -> Parse<ast::SourceFile>;
|
||||||
/// Path to a file, relative to the root of its source root.
|
/// Path to a file, relative to the root of its source root.
|
||||||
#[salsa::input]
|
#[salsa::input]
|
||||||
fn file_relative_path(&self, file_id: FileId) -> RelativePathBuf;
|
fn file_relative_path(&self, file_id: FileId) -> RelativePathBuf;
|
||||||
@ -98,7 +98,7 @@ fn source_root_crates(db: &impl SourceDatabase, id: SourceRootId) -> Arc<Vec<Cra
|
|||||||
Arc::new(res)
|
Arc::new(res)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn parse_query(db: &impl SourceDatabase, file_id: FileId) -> Parse {
|
fn parse_query(db: &impl SourceDatabase, file_id: FileId) -> Parse<ast::SourceFile> {
|
||||||
let _p = profile("parse_query");
|
let _p = profile("parse_query");
|
||||||
let text = db.file_text(file_id);
|
let text = db.file_text(file_id);
|
||||||
SourceFile::parse(&*text)
|
SourceFile::parse(&*text)
|
||||||
|
@ -2,7 +2,7 @@ use std::sync::Arc;
|
|||||||
|
|
||||||
use parking_lot::Mutex;
|
use parking_lot::Mutex;
|
||||||
use ra_db::{salsa, SourceDatabase};
|
use ra_db::{salsa, SourceDatabase};
|
||||||
use ra_syntax::{ast, SmolStr, SyntaxNode, TreeArc};
|
use ra_syntax::{ast, Parse, SmolStr, SyntaxNode, TreeArc};
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
adt::{EnumData, StructData},
|
adt::{EnumData, StructData},
|
||||||
@ -69,7 +69,7 @@ pub trait AstDatabase: InternDatabase {
|
|||||||
fn parse_or_expand(&self, file_id: HirFileId) -> Option<TreeArc<SyntaxNode>>;
|
fn parse_or_expand(&self, file_id: HirFileId) -> Option<TreeArc<SyntaxNode>>;
|
||||||
|
|
||||||
#[salsa::invoke(crate::ids::HirFileId::parse_macro_query)]
|
#[salsa::invoke(crate::ids::HirFileId::parse_macro_query)]
|
||||||
fn parse_macro(&self, macro_file: ids::MacroFile) -> Option<TreeArc<SyntaxNode>>;
|
fn parse_macro(&self, macro_file: ids::MacroFile) -> Option<Parse<SyntaxNode>>;
|
||||||
|
|
||||||
#[salsa::invoke(crate::ids::macro_def_query)]
|
#[salsa::invoke(crate::ids::macro_def_query)]
|
||||||
fn macro_def(&self, macro_id: MacroDefId) -> Option<Arc<mbe::MacroRules>>;
|
fn macro_def(&self, macro_id: MacroDefId) -> Option<Arc<mbe::MacroRules>>;
|
||||||
|
@ -6,7 +6,7 @@ use std::{
|
|||||||
use mbe::MacroRules;
|
use mbe::MacroRules;
|
||||||
use ra_db::{salsa, FileId};
|
use ra_db::{salsa, FileId};
|
||||||
use ra_prof::profile;
|
use ra_prof::profile;
|
||||||
use ra_syntax::{ast, AstNode, SyntaxNode, TreeArc};
|
use ra_syntax::{ast, AstNode, Parse, SyntaxNode, TreeArc};
|
||||||
|
|
||||||
use crate::{AstDatabase, AstId, DefDatabase, FileAstId, InternDatabase, Module, Source};
|
use crate::{AstDatabase, AstId, DefDatabase, FileAstId, InternDatabase, Module, Source};
|
||||||
|
|
||||||
@ -61,14 +61,16 @@ impl HirFileId {
|
|||||||
) -> Option<TreeArc<SyntaxNode>> {
|
) -> Option<TreeArc<SyntaxNode>> {
|
||||||
match file_id.0 {
|
match file_id.0 {
|
||||||
HirFileIdRepr::File(file_id) => Some(db.parse(file_id).tree().syntax().to_owned()),
|
HirFileIdRepr::File(file_id) => Some(db.parse(file_id).tree().syntax().to_owned()),
|
||||||
HirFileIdRepr::Macro(macro_file) => db.parse_macro(macro_file),
|
HirFileIdRepr::Macro(macro_file) => {
|
||||||
|
db.parse_macro(macro_file).map(|it| it.tree().to_owned())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn parse_macro_query(
|
pub(crate) fn parse_macro_query(
|
||||||
db: &impl AstDatabase,
|
db: &impl AstDatabase,
|
||||||
macro_file: MacroFile,
|
macro_file: MacroFile,
|
||||||
) -> Option<TreeArc<SyntaxNode>> {
|
) -> Option<Parse<SyntaxNode>> {
|
||||||
let _p = profile("parse_macro_query");
|
let _p = profile("parse_macro_query");
|
||||||
let macro_call_id = macro_file.macro_call_id;
|
let macro_call_id = macro_file.macro_call_id;
|
||||||
let tt = db
|
let tt = db
|
||||||
@ -85,10 +87,8 @@ impl HirFileId {
|
|||||||
})
|
})
|
||||||
.ok()?;
|
.ok()?;
|
||||||
match macro_file.macro_file_kind {
|
match macro_file.macro_file_kind {
|
||||||
MacroFileKind::Items => Some(mbe::token_tree_to_ast_item_list(&tt).syntax().to_owned()),
|
MacroFileKind::Items => Some(Parse::to_syntax(mbe::token_tree_to_ast_item_list(&tt))),
|
||||||
MacroFileKind::Expr => {
|
MacroFileKind::Expr => mbe::token_tree_to_expr(&tt).ok().map(Parse::to_syntax),
|
||||||
mbe::token_tree_to_expr(&tt).ok().map(|it| it.syntax().to_owned())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -43,7 +43,7 @@ pub(crate) struct CompletionContext<'a> {
|
|||||||
impl<'a> CompletionContext<'a> {
|
impl<'a> CompletionContext<'a> {
|
||||||
pub(super) fn new(
|
pub(super) fn new(
|
||||||
db: &'a db::RootDatabase,
|
db: &'a db::RootDatabase,
|
||||||
original_parse: &'a Parse,
|
original_parse: &'a Parse<ast::SourceFile>,
|
||||||
position: FilePosition,
|
position: FilePosition,
|
||||||
) -> Option<CompletionContext<'a>> {
|
) -> Option<CompletionContext<'a>> {
|
||||||
let module = source_binder::module_from_position(db, position);
|
let module = source_binder::module_from_position(db, position);
|
||||||
@ -83,7 +83,7 @@ impl<'a> CompletionContext<'a> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn fill(&mut self, original_parse: &'a Parse, offset: TextUnit) {
|
fn fill(&mut self, original_parse: &'a Parse<ast::SourceFile>, offset: TextUnit) {
|
||||||
// Insert a fake ident to get a valid parse tree. We will use this file
|
// Insert a fake ident to get a valid parse tree. We will use this file
|
||||||
// to determine context, though the original_file will be used for
|
// to determine context, though the original_file will be used for
|
||||||
// actual completion.
|
// actual completion.
|
||||||
|
@ -9,7 +9,7 @@ use ra_db::{
|
|||||||
FileTextQuery, SourceRootId,
|
FileTextQuery, SourceRootId,
|
||||||
};
|
};
|
||||||
use ra_prof::{memory_usage, Bytes};
|
use ra_prof::{memory_usage, Bytes};
|
||||||
use ra_syntax::{AstNode, Parse, SyntaxNode, TreeArc};
|
use ra_syntax::{ast, AstNode, Parse, SyntaxNode};
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
db::RootDatabase,
|
db::RootDatabase,
|
||||||
@ -79,10 +79,10 @@ impl fmt::Display for SyntaxTreeStats {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl FromIterator<TableEntry<FileId, Parse>> for SyntaxTreeStats {
|
impl FromIterator<TableEntry<FileId, Parse<ast::SourceFile>>> for SyntaxTreeStats {
|
||||||
fn from_iter<T>(iter: T) -> SyntaxTreeStats
|
fn from_iter<T>(iter: T) -> SyntaxTreeStats
|
||||||
where
|
where
|
||||||
T: IntoIterator<Item = TableEntry<FileId, Parse>>,
|
T: IntoIterator<Item = TableEntry<FileId, Parse<ast::SourceFile>>>,
|
||||||
{
|
{
|
||||||
let mut res = SyntaxTreeStats::default();
|
let mut res = SyntaxTreeStats::default();
|
||||||
for entry in iter {
|
for entry in iter {
|
||||||
@ -96,15 +96,15 @@ impl FromIterator<TableEntry<FileId, Parse>> for SyntaxTreeStats {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl FromIterator<TableEntry<MacroFile, Option<TreeArc<SyntaxNode>>>> for SyntaxTreeStats {
|
impl FromIterator<TableEntry<MacroFile, Option<Parse<SyntaxNode>>>> for SyntaxTreeStats {
|
||||||
fn from_iter<T>(iter: T) -> SyntaxTreeStats
|
fn from_iter<T>(iter: T) -> SyntaxTreeStats
|
||||||
where
|
where
|
||||||
T: IntoIterator<Item = TableEntry<MacroFile, Option<TreeArc<SyntaxNode>>>>,
|
T: IntoIterator<Item = TableEntry<MacroFile, Option<Parse<SyntaxNode>>>>,
|
||||||
{
|
{
|
||||||
let mut res = SyntaxTreeStats::default();
|
let mut res = SyntaxTreeStats::default();
|
||||||
for entry in iter {
|
for entry in iter {
|
||||||
res.total += 1;
|
res.total += 1;
|
||||||
if let Some(tree) = entry.value.and_then(|it| it) {
|
if let Some(tree) = entry.value.and_then(|it| it).map(|it| it.tree().to_owned()) {
|
||||||
res.retained += 1;
|
res.retained += 1;
|
||||||
res.retained_size += tree.memory_size_of_subtree();
|
res.retained_size += tree.memory_size_of_subtree();
|
||||||
}
|
}
|
||||||
|
@ -169,7 +169,9 @@ impl SymbolIndex {
|
|||||||
self.map.as_fst().size() + self.symbols.len() * mem::size_of::<FileSymbol>()
|
self.map.as_fst().size() + self.symbols.len() * mem::size_of::<FileSymbol>()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn for_files(files: impl ParallelIterator<Item = (FileId, Parse)>) -> SymbolIndex {
|
pub(crate) fn for_files(
|
||||||
|
files: impl ParallelIterator<Item = (FileId, Parse<ast::SourceFile>)>,
|
||||||
|
) -> SymbolIndex {
|
||||||
let symbols = files
|
let symbols = files
|
||||||
.flat_map(|(file_id, file)| source_file_to_file_symbols(file.tree(), file_id))
|
.flat_map(|(file_id, file)| source_file_to_file_symbols(file.tree(), file_id))
|
||||||
.collect::<Vec<_>>();
|
.collect::<Vec<_>>();
|
||||||
|
@ -2,8 +2,8 @@ use crate::subtree_source::SubtreeTokenSource;
|
|||||||
use crate::ExpandError;
|
use crate::ExpandError;
|
||||||
use ra_parser::{ParseError, TreeSink};
|
use ra_parser::{ParseError, TreeSink};
|
||||||
use ra_syntax::{
|
use ra_syntax::{
|
||||||
ast, AstNode, SmolStr, SyntaxElement, SyntaxKind, SyntaxKind::*, SyntaxNode, SyntaxTreeBuilder,
|
ast, AstNode, Parse, SmolStr, SyntaxElement, SyntaxKind, SyntaxKind::*, SyntaxNode,
|
||||||
TextRange, TextUnit, TreeArc, T,
|
SyntaxTreeBuilder, TextRange, TextUnit, T,
|
||||||
};
|
};
|
||||||
use tt::buffer::{Cursor, TokenBuffer};
|
use tt::buffer::{Cursor, TokenBuffer};
|
||||||
|
|
||||||
@ -45,7 +45,7 @@ pub fn syntax_node_to_token_tree(node: &SyntaxNode) -> Option<(tt::Subtree, Toke
|
|||||||
//
|
//
|
||||||
//
|
//
|
||||||
|
|
||||||
fn token_tree_to_syntax_node<F>(tt: &tt::Subtree, f: F) -> Result<TreeArc<SyntaxNode>, ExpandError>
|
fn token_tree_to_syntax_node<F>(tt: &tt::Subtree, f: F) -> Result<Parse<SyntaxNode>, ExpandError>
|
||||||
where
|
where
|
||||||
F: Fn(&mut dyn ra_parser::TokenSource, &mut dyn ra_parser::TreeSink),
|
F: Fn(&mut dyn ra_parser::TokenSource, &mut dyn ra_parser::TreeSink),
|
||||||
{
|
{
|
||||||
@ -58,50 +58,44 @@ where
|
|||||||
return Err(ExpandError::ConversionError);
|
return Err(ExpandError::ConversionError);
|
||||||
}
|
}
|
||||||
//FIXME: would be cool to report errors
|
//FIXME: would be cool to report errors
|
||||||
let (tree, _errors) = tree_sink.inner.finish();
|
let parse = tree_sink.inner.finish();
|
||||||
Ok(tree)
|
Ok(parse)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Parses the token tree (result of macro expansion) to an expression
|
/// Parses the token tree (result of macro expansion) to an expression
|
||||||
pub fn token_tree_to_expr(tt: &tt::Subtree) -> Result<TreeArc<ast::Expr>, ExpandError> {
|
pub fn token_tree_to_expr(tt: &tt::Subtree) -> Result<Parse<ast::Expr>, ExpandError> {
|
||||||
let syntax = token_tree_to_syntax_node(tt, ra_parser::parse_expr)?;
|
let parse = token_tree_to_syntax_node(tt, ra_parser::parse_expr)?;
|
||||||
ast::Expr::cast(&syntax)
|
parse.cast().ok_or_else(|| crate::ExpandError::ConversionError)
|
||||||
.map(|m| m.to_owned())
|
|
||||||
.ok_or_else(|| crate::ExpandError::ConversionError)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Parses the token tree (result of macro expansion) to a Pattern
|
/// Parses the token tree (result of macro expansion) to a Pattern
|
||||||
pub fn token_tree_to_pat(tt: &tt::Subtree) -> Result<TreeArc<ast::Pat>, ExpandError> {
|
pub fn token_tree_to_pat(tt: &tt::Subtree) -> Result<Parse<ast::Pat>, ExpandError> {
|
||||||
let syntax = token_tree_to_syntax_node(tt, ra_parser::parse_pat)?;
|
let parse = token_tree_to_syntax_node(tt, ra_parser::parse_pat)?;
|
||||||
ast::Pat::cast(&syntax).map(|m| m.to_owned()).ok_or_else(|| ExpandError::ConversionError)
|
parse.cast().ok_or_else(|| crate::ExpandError::ConversionError)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Parses the token tree (result of macro expansion) to a Type
|
/// Parses the token tree (result of macro expansion) to a Type
|
||||||
pub fn token_tree_to_ty(tt: &tt::Subtree) -> Result<TreeArc<ast::TypeRef>, ExpandError> {
|
pub fn token_tree_to_ty(tt: &tt::Subtree) -> Result<Parse<ast::TypeRef>, ExpandError> {
|
||||||
let syntax = token_tree_to_syntax_node(tt, ra_parser::parse_ty)?;
|
let parse = token_tree_to_syntax_node(tt, ra_parser::parse_ty)?;
|
||||||
ast::TypeRef::cast(&syntax).map(|m| m.to_owned()).ok_or_else(|| ExpandError::ConversionError)
|
parse.cast().ok_or_else(|| crate::ExpandError::ConversionError)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Parses the token tree (result of macro expansion) as a sequence of stmts
|
/// Parses the token tree (result of macro expansion) as a sequence of stmts
|
||||||
pub fn token_tree_to_macro_stmts(
|
pub fn token_tree_to_macro_stmts(tt: &tt::Subtree) -> Result<Parse<ast::MacroStmts>, ExpandError> {
|
||||||
tt: &tt::Subtree,
|
let parse = token_tree_to_syntax_node(tt, ra_parser::parse_macro_stmts)?;
|
||||||
) -> Result<TreeArc<ast::MacroStmts>, ExpandError> {
|
parse.cast().ok_or_else(|| crate::ExpandError::ConversionError)
|
||||||
let syntax = token_tree_to_syntax_node(tt, ra_parser::parse_macro_stmts)?;
|
|
||||||
ast::MacroStmts::cast(&syntax).map(|m| m.to_owned()).ok_or_else(|| ExpandError::ConversionError)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Parses the token tree (result of macro expansion) as a sequence of items
|
/// Parses the token tree (result of macro expansion) as a sequence of items
|
||||||
pub fn token_tree_to_macro_items(
|
pub fn token_tree_to_macro_items(tt: &tt::Subtree) -> Result<Parse<ast::MacroItems>, ExpandError> {
|
||||||
tt: &tt::Subtree,
|
let parse = token_tree_to_syntax_node(tt, ra_parser::parse_macro_items)?;
|
||||||
) -> Result<TreeArc<ast::MacroItems>, ExpandError> {
|
parse.cast().ok_or_else(|| crate::ExpandError::ConversionError)
|
||||||
let syntax = token_tree_to_syntax_node(tt, ra_parser::parse_macro_items)?;
|
|
||||||
ast::MacroItems::cast(&syntax).map(|m| m.to_owned()).ok_or_else(|| ExpandError::ConversionError)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Parses the token tree (result of macro expansion) as a sequence of items
|
/// Parses the token tree (result of macro expansion) as a sequence of items
|
||||||
pub fn token_tree_to_ast_item_list(tt: &tt::Subtree) -> TreeArc<ast::SourceFile> {
|
pub fn token_tree_to_ast_item_list(tt: &tt::Subtree) -> Parse<ast::SourceFile> {
|
||||||
let syntax = token_tree_to_syntax_node(tt, ra_parser::parse).unwrap();
|
let parse = token_tree_to_syntax_node(tt, ra_parser::parse).unwrap();
|
||||||
ast::SourceFile::cast(&syntax).unwrap().to_owned()
|
parse.cast().unwrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
impl TokenMap {
|
impl TokenMap {
|
||||||
|
@ -72,7 +72,7 @@ pub(crate) fn expand_to_items(
|
|||||||
invocation: &str,
|
invocation: &str,
|
||||||
) -> ra_syntax::TreeArc<ast::MacroItems> {
|
) -> ra_syntax::TreeArc<ast::MacroItems> {
|
||||||
let expanded = expand(rules, invocation);
|
let expanded = expand(rules, invocation);
|
||||||
token_tree_to_macro_items(&expanded).unwrap()
|
token_tree_to_macro_items(&expanded).unwrap().tree().to_owned()
|
||||||
}
|
}
|
||||||
|
|
||||||
#[allow(unused)]
|
#[allow(unused)]
|
||||||
@ -81,7 +81,7 @@ pub(crate) fn expand_to_stmts(
|
|||||||
invocation: &str,
|
invocation: &str,
|
||||||
) -> ra_syntax::TreeArc<ast::MacroStmts> {
|
) -> ra_syntax::TreeArc<ast::MacroStmts> {
|
||||||
let expanded = expand(rules, invocation);
|
let expanded = expand(rules, invocation);
|
||||||
token_tree_to_macro_stmts(&expanded).unwrap()
|
token_tree_to_macro_stmts(&expanded).unwrap().tree().to_owned()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn expand_to_expr(
|
pub(crate) fn expand_to_expr(
|
||||||
@ -89,7 +89,7 @@ pub(crate) fn expand_to_expr(
|
|||||||
invocation: &str,
|
invocation: &str,
|
||||||
) -> ra_syntax::TreeArc<ast::Expr> {
|
) -> ra_syntax::TreeArc<ast::Expr> {
|
||||||
let expanded = expand(rules, invocation);
|
let expanded = expand(rules, invocation);
|
||||||
token_tree_to_expr(&expanded).unwrap()
|
token_tree_to_expr(&expanded).unwrap().tree().to_owned()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn text_to_tokentree(text: &str) -> tt::Subtree {
|
pub(crate) fn text_to_tokentree(text: &str) -> tt::Subtree {
|
||||||
@ -164,22 +164,22 @@ pub(crate) fn assert_expansion(
|
|||||||
|
|
||||||
let (expanded_tree, expected_tree) = match kind {
|
let (expanded_tree, expected_tree) = match kind {
|
||||||
MacroKind::Items => {
|
MacroKind::Items => {
|
||||||
let expanded_tree = token_tree_to_macro_items(&expanded);
|
let expanded_tree = token_tree_to_macro_items(&expanded).unwrap().tree().to_owned();
|
||||||
let expected_tree = token_tree_to_macro_items(&expected);
|
let expected_tree = token_tree_to_macro_items(&expected).unwrap().tree().to_owned();
|
||||||
|
|
||||||
(
|
(
|
||||||
debug_dump_ignore_spaces(expanded_tree.unwrap().syntax()).trim().to_string(),
|
debug_dump_ignore_spaces(expanded_tree.syntax()).trim().to_string(),
|
||||||
debug_dump_ignore_spaces(expected_tree.unwrap().syntax()).trim().to_string(),
|
debug_dump_ignore_spaces(expected_tree.syntax()).trim().to_string(),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
MacroKind::Stmts => {
|
MacroKind::Stmts => {
|
||||||
let expanded_tree = token_tree_to_macro_stmts(&expanded);
|
let expanded_tree = token_tree_to_macro_stmts(&expanded).unwrap().tree().to_owned();
|
||||||
let expected_tree = token_tree_to_macro_stmts(&expected);
|
let expected_tree = token_tree_to_macro_stmts(&expected).unwrap().tree().to_owned();
|
||||||
|
|
||||||
(
|
(
|
||||||
debug_dump_ignore_spaces(expanded_tree.unwrap().syntax()).trim().to_string(),
|
debug_dump_ignore_spaces(expanded_tree.syntax()).trim().to_string(),
|
||||||
debug_dump_ignore_spaces(expected_tree.unwrap().syntax()).trim().to_string(),
|
debug_dump_ignore_spaces(expected_tree.syntax()).trim().to_string(),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
@ -419,9 +419,9 @@ fn test_expand_to_item_list() {
|
|||||||
",
|
",
|
||||||
);
|
);
|
||||||
let expansion = expand(&rules, "structs!(Foo, Bar);");
|
let expansion = expand(&rules, "structs!(Foo, Bar);");
|
||||||
let tree = token_tree_to_macro_items(&expansion);
|
let tree = token_tree_to_macro_items(&expansion).unwrap().tree().to_owned();
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
tree.unwrap().syntax().debug_dump().trim(),
|
tree.syntax().debug_dump().trim(),
|
||||||
r#"
|
r#"
|
||||||
MACRO_ITEMS@[0; 40)
|
MACRO_ITEMS@[0; 40)
|
||||||
STRUCT_DEF@[0; 20)
|
STRUCT_DEF@[0; 20)
|
||||||
@ -537,10 +537,10 @@ fn test_tt_to_stmts() {
|
|||||||
);
|
);
|
||||||
|
|
||||||
let expanded = expand(&rules, "foo!{}");
|
let expanded = expand(&rules, "foo!{}");
|
||||||
let stmts = token_tree_to_macro_stmts(&expanded);
|
let stmts = token_tree_to_macro_stmts(&expanded).unwrap().tree().to_owned();
|
||||||
|
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
stmts.unwrap().syntax().debug_dump().trim(),
|
stmts.syntax().debug_dump().trim(),
|
||||||
r#"MACRO_STMTS@[0; 15)
|
r#"MACRO_STMTS@[0; 15)
|
||||||
LET_STMT@[0; 7)
|
LET_STMT@[0; 7)
|
||||||
LET_KW@[0; 3) "let"
|
LET_KW@[0; 3) "let"
|
||||||
|
@ -43,8 +43,8 @@ pub use crate::{
|
|||||||
ptr::{AstPtr, SyntaxNodePtr},
|
ptr::{AstPtr, SyntaxNodePtr},
|
||||||
syntax_error::{Location, SyntaxError, SyntaxErrorKind},
|
syntax_error::{Location, SyntaxError, SyntaxErrorKind},
|
||||||
syntax_node::{
|
syntax_node::{
|
||||||
Direction, InsertPosition, SyntaxElement, SyntaxNode, SyntaxToken, SyntaxTreeBuilder,
|
Direction, InsertPosition, SyntaxElement, SyntaxNode, SyntaxNodeWrapper, SyntaxToken,
|
||||||
TreeArc, WalkEvent,
|
SyntaxTreeBuilder, TreeArc, WalkEvent,
|
||||||
},
|
},
|
||||||
syntax_text::SyntaxText,
|
syntax_text::SyntaxText,
|
||||||
};
|
};
|
||||||
@ -57,14 +57,24 @@ pub use rowan::{SmolStr, TextRange, TextUnit};
|
|||||||
///
|
///
|
||||||
/// Note that we always produce a syntax tree, even for completely invalid
|
/// Note that we always produce a syntax tree, even for completely invalid
|
||||||
/// files.
|
/// files.
|
||||||
#[derive(Debug, Clone, PartialEq, Eq)]
|
#[derive(Debug, PartialEq, Eq)]
|
||||||
pub struct Parse {
|
pub struct Parse<T: SyntaxNodeWrapper> {
|
||||||
tree: TreeArc<SourceFile>,
|
tree: TreeArc<T>,
|
||||||
errors: Arc<Vec<SyntaxError>>,
|
errors: Arc<Vec<SyntaxError>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Parse {
|
impl<T: SyntaxNodeWrapper> Clone for Parse<T> {
|
||||||
pub fn tree(&self) -> &SourceFile {
|
fn clone(&self) -> Parse<T> {
|
||||||
|
Parse { tree: self.tree.clone(), errors: self.errors.clone() }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: SyntaxNodeWrapper> Parse<T> {
|
||||||
|
fn new(tree: TreeArc<T>, errors: Vec<SyntaxError>) -> Parse<T> {
|
||||||
|
Parse { tree, errors: Arc::new(errors) }
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn tree(&self) -> &T {
|
||||||
&*self.tree
|
&*self.tree
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -72,18 +82,22 @@ impl Parse {
|
|||||||
&*self.errors
|
&*self.errors
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn ok(self) -> Result<TreeArc<SourceFile>, Arc<Vec<SyntaxError>>> {
|
pub fn ok(self) -> Result<TreeArc<T>, Arc<Vec<SyntaxError>>> {
|
||||||
if self.errors.is_empty() {
|
if self.errors.is_empty() {
|
||||||
Ok(self.tree)
|
Ok(self.tree)
|
||||||
} else {
|
} else {
|
||||||
Err(self.errors)
|
Err(self.errors)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
pub fn reparse(&self, edit: &AtomTextEdit) -> Parse {
|
impl<T: AstNode> Parse<T> {
|
||||||
self.incremental_reparse(edit).unwrap_or_else(|| self.full_reparse(edit))
|
pub fn to_syntax(this: Self) -> Parse<SyntaxNode> {
|
||||||
|
Parse { tree: this.tree().syntax().to_owned(), errors: this.errors }
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Parse<SourceFile> {
|
||||||
pub fn debug_dump(&self) -> String {
|
pub fn debug_dump(&self) -> String {
|
||||||
let mut buf = self.tree.syntax().debug_dump();
|
let mut buf = self.tree.syntax().debug_dump();
|
||||||
for err in self.errors.iter() {
|
for err in self.errors.iter() {
|
||||||
@ -92,7 +106,11 @@ impl Parse {
|
|||||||
buf
|
buf
|
||||||
}
|
}
|
||||||
|
|
||||||
fn incremental_reparse(&self, edit: &AtomTextEdit) -> Option<Parse> {
|
pub fn reparse(&self, edit: &AtomTextEdit) -> Parse<SourceFile> {
|
||||||
|
self.incremental_reparse(edit).unwrap_or_else(|| self.full_reparse(edit))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn incremental_reparse(&self, edit: &AtomTextEdit) -> Option<Parse<SourceFile>> {
|
||||||
// FIXME: validation errors are not handled here
|
// FIXME: validation errors are not handled here
|
||||||
parsing::incremental_reparse(self.tree.syntax(), edit, self.errors.to_vec()).map(
|
parsing::incremental_reparse(self.tree.syntax(), edit, self.errors.to_vec()).map(
|
||||||
|(green_node, errors, _reparsed_range)| Parse {
|
|(green_node, errors, _reparsed_range)| Parse {
|
||||||
@ -102,12 +120,19 @@ impl Parse {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn full_reparse(&self, edit: &AtomTextEdit) -> Parse {
|
fn full_reparse(&self, edit: &AtomTextEdit) -> Parse<SourceFile> {
|
||||||
let text = edit.apply(self.tree.syntax().text().to_string());
|
let text = edit.apply(self.tree.syntax().text().to_string());
|
||||||
SourceFile::parse(&text)
|
SourceFile::parse(&text)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl Parse<SyntaxNode> {
|
||||||
|
pub fn cast<T: AstNode>(self) -> Option<Parse<T>> {
|
||||||
|
let node = T::cast(&self.tree)?;
|
||||||
|
Some(Parse { tree: node.to_owned(), errors: self.errors })
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// `SourceFile` represents a parse tree for a single Rust file.
|
/// `SourceFile` represents a parse tree for a single Rust file.
|
||||||
pub use crate::ast::SourceFile;
|
pub use crate::ast::SourceFile;
|
||||||
|
|
||||||
@ -121,7 +146,7 @@ impl SourceFile {
|
|||||||
TreeArc::cast(root)
|
TreeArc::cast(root)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn parse(text: &str) -> Parse {
|
pub fn parse(text: &str) -> Parse<SourceFile> {
|
||||||
let (green, mut errors) = parsing::parse_text(text);
|
let (green, mut errors) = parsing::parse_text(text);
|
||||||
let tree = SourceFile::new(green);
|
let tree = SourceFile::new(green);
|
||||||
errors.extend(validation::validate(&tree));
|
errors.extend(validation::validate(&tree));
|
||||||
|
@ -18,7 +18,8 @@ use rowan::{GreenNodeBuilder, TransparentNewType};
|
|||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
syntax_error::{SyntaxError, SyntaxErrorKind},
|
syntax_error::{SyntaxError, SyntaxErrorKind},
|
||||||
AstNode, SmolStr, SourceFile, SyntaxKind, SyntaxNodePtr, SyntaxText, TextRange, TextUnit,
|
AstNode, Parse, SmolStr, SourceFile, SyntaxKind, SyntaxNodePtr, SyntaxText, TextRange,
|
||||||
|
TextUnit,
|
||||||
};
|
};
|
||||||
|
|
||||||
pub use rowan::WalkEvent;
|
pub use rowan::WalkEvent;
|
||||||
@ -594,13 +595,13 @@ impl SyntaxTreeBuilder {
|
|||||||
(green, self.errors)
|
(green, self.errors)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn finish(self) -> (TreeArc<SyntaxNode>, Vec<SyntaxError>) {
|
pub fn finish(self) -> Parse<SyntaxNode> {
|
||||||
let (green, errors) = self.finish_raw();
|
let (green, errors) = self.finish_raw();
|
||||||
let node = SyntaxNode::new(green);
|
let node = SyntaxNode::new(green);
|
||||||
if cfg!(debug_assertions) {
|
if cfg!(debug_assertions) {
|
||||||
crate::validation::validate_block_structure(&node);
|
crate::validation::validate_block_structure(&node);
|
||||||
}
|
}
|
||||||
(node, errors)
|
Parse::new(node, errors)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn token(&mut self, kind: SyntaxKind, text: SmolStr) {
|
pub fn token(&mut self, kind: SyntaxKind, text: SmolStr) {
|
||||||
|
Loading…
Reference in New Issue
Block a user