2018-09-16 09:54:24 +00:00
|
|
|
extern crate ra_syntax;
|
2018-08-12 15:50:16 +00:00
|
|
|
#[macro_use]
|
2018-08-25 11:26:34 +00:00
|
|
|
extern crate test_utils;
|
2018-08-13 14:42:43 +00:00
|
|
|
extern crate walkdir;
|
2018-01-07 11:56:08 +00:00
|
|
|
|
2018-07-30 12:25:52 +00:00
|
|
|
use std::{
|
2018-10-15 21:44:23 +00:00
|
|
|
fmt::Write,
|
2018-07-30 12:25:52 +00:00
|
|
|
fs,
|
2018-12-09 19:19:23 +00:00
|
|
|
path::{Path, PathBuf, Component},
|
2018-07-30 12:25:52 +00:00
|
|
|
};
|
2018-01-07 11:56:08 +00:00
|
|
|
|
2018-09-16 09:54:24 +00:00
|
|
|
use ra_syntax::{
|
2018-10-15 21:44:23 +00:00
|
|
|
utils::{check_fuzz_invariants, dump_tree},
|
2018-11-07 15:32:33 +00:00
|
|
|
SourceFileNode,
|
2018-08-25 11:45:17 +00:00
|
|
|
};
|
2018-08-25 09:10:35 +00:00
|
|
|
|
2018-08-11 07:03:03 +00:00
|
|
|
#[test]
|
|
|
|
fn lexer_tests() {
|
|
|
|
dir_tests(&["lexer"], |text| {
|
2018-09-16 09:54:24 +00:00
|
|
|
let tokens = ra_syntax::tokenize(text);
|
2018-08-11 07:03:03 +00:00
|
|
|
dump_tokens(&tokens, text)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2018-09-10 18:14:09 +00:00
|
|
|
#[test]
|
|
|
|
fn parser_tests() {
|
|
|
|
dir_tests(&["parser/inline", "parser/ok", "parser/err"], |text| {
|
2018-11-07 15:32:33 +00:00
|
|
|
let file = SourceFileNode::parse(text);
|
2018-09-10 18:14:09 +00:00
|
|
|
dump_tree(file.syntax())
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn parser_fuzz_tests() {
|
|
|
|
for (_, text) in collect_tests(&["parser/fuzz-failures"]) {
|
|
|
|
check_fuzz_invariants(&text)
|
|
|
|
}
|
2018-08-25 11:45:17 +00:00
|
|
|
}
|
|
|
|
|
2018-12-09 19:19:23 +00:00
|
|
|
/// Test that Rust-analyzer can parse and validate the rust-analyser
|
|
|
|
/// TODO: Use this as a benchmark
|
|
|
|
#[test]
|
|
|
|
fn self_hosting_parsing() {
|
|
|
|
let empty_vec = vec![];
|
|
|
|
let dir = project_dir();
|
|
|
|
for entry in walkdir::WalkDir::new(dir)
|
|
|
|
.into_iter()
|
|
|
|
.filter_entry(|entry| {
|
|
|
|
!entry
|
|
|
|
.path()
|
|
|
|
.components()
|
|
|
|
// TODO: this more neatly
|
|
|
|
.any(|component| {
|
|
|
|
// Get all files which are not in the crates/ra_syntax/tests/data folder
|
|
|
|
(component == Component::Normal(std::ffi::OsStr::new("data"))
|
|
|
|
// or the .git folder
|
|
|
|
|| component == Component::Normal(std::ffi::OsStr::new(".git")))
|
|
|
|
})
|
|
|
|
})
|
|
|
|
.map(|e| e.unwrap())
|
|
|
|
.filter(|entry| {
|
|
|
|
// Get all `.rs ` files
|
|
|
|
!entry.path().is_dir() && (entry.path().extension() == Some(std::ffi::OsStr::new("rs")))
|
|
|
|
})
|
|
|
|
{
|
|
|
|
let text = read_text(entry.path());
|
|
|
|
let node = SourceFileNode::parse(&text);
|
|
|
|
let errors = node.errors();
|
|
|
|
assert_eq!(
|
|
|
|
errors, empty_vec,
|
|
|
|
"There should be no errors in the file {:?}",
|
|
|
|
entry
|
|
|
|
);
|
|
|
|
}
|
|
|
|
}
|
2018-01-28 00:17:05 +00:00
|
|
|
/// Read file and normalize newlines.
|
|
|
|
///
|
|
|
|
/// `rustc` seems to always normalize `\r\n` newlines to `\n`:
|
|
|
|
///
|
|
|
|
/// ```
|
|
|
|
/// let s = "
|
|
|
|
/// ";
|
|
|
|
/// assert_eq!(s.as_bytes(), &[10]);
|
|
|
|
/// ```
|
|
|
|
///
|
|
|
|
/// so this should always be correct.
|
|
|
|
fn read_text(path: &Path) -> String {
|
2018-12-09 19:19:23 +00:00
|
|
|
fs::read_to_string(path)
|
|
|
|
.expect(&format!("File at {:?} should be valid", path))
|
|
|
|
.replace("\r\n", "\n")
|
2018-01-28 00:17:05 +00:00
|
|
|
}
|
|
|
|
|
2018-02-03 09:51:06 +00:00
|
|
|
pub fn dir_tests<F>(paths: &[&str], f: F)
|
2018-10-15 21:44:23 +00:00
|
|
|
where
|
|
|
|
F: Fn(&str) -> String,
|
2018-01-07 12:34:11 +00:00
|
|
|
{
|
2018-09-08 15:42:59 +00:00
|
|
|
for (path, input_code) in collect_tests(paths) {
|
2018-02-11 08:19:54 +00:00
|
|
|
let parse_tree = f(&input_code);
|
2018-01-07 12:34:11 +00:00
|
|
|
let path = path.with_extension("txt");
|
|
|
|
if !path.exists() {
|
|
|
|
println!("\nfile: {}", path.display());
|
2018-02-11 08:19:54 +00:00
|
|
|
println!("No .txt file with expected result, creating...\n");
|
|
|
|
println!("{}\n{}", input_code, parse_tree);
|
2018-07-30 12:25:52 +00:00
|
|
|
fs::write(&path, parse_tree).unwrap();
|
2018-01-07 12:34:11 +00:00
|
|
|
panic!("No expected result")
|
|
|
|
}
|
2018-01-28 00:17:05 +00:00
|
|
|
let expected = read_text(&path);
|
2018-01-07 12:34:11 +00:00
|
|
|
let expected = expected.as_str();
|
2018-02-11 08:19:54 +00:00
|
|
|
let parse_tree = parse_tree.as_str();
|
|
|
|
assert_equal_text(expected, parse_tree, &path);
|
2018-01-07 12:34:11 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-08-12 15:50:16 +00:00
|
|
|
const REWRITE: bool = false;
|
|
|
|
|
2018-02-03 09:51:06 +00:00
|
|
|
fn assert_equal_text(expected: &str, actual: &str, path: &Path) {
|
2018-08-12 15:50:16 +00:00
|
|
|
if expected == actual {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
let dir = project_dir();
|
2018-08-16 09:51:40 +00:00
|
|
|
let pretty_path = path.strip_prefix(&dir).unwrap_or_else(|_| path);
|
2018-08-12 15:50:16 +00:00
|
|
|
if expected.trim() == actual.trim() {
|
|
|
|
println!("whitespace difference, rewriting");
|
2018-08-16 09:51:40 +00:00
|
|
|
println!("file: {}\n", pretty_path.display());
|
2018-08-12 15:50:16 +00:00
|
|
|
fs::write(path, actual).unwrap();
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
if REWRITE {
|
2018-08-16 09:51:40 +00:00
|
|
|
println!("rewriting {}", pretty_path.display());
|
2018-08-12 15:50:16 +00:00
|
|
|
fs::write(path, actual).unwrap();
|
|
|
|
return;
|
2018-01-07 11:56:08 +00:00
|
|
|
}
|
2018-08-16 09:51:40 +00:00
|
|
|
assert_eq_text!(expected, actual, "file: {}", pretty_path.display());
|
2018-01-07 11:56:08 +00:00
|
|
|
}
|
|
|
|
|
2018-09-08 15:42:59 +00:00
|
|
|
fn collect_tests(paths: &[&str]) -> Vec<(PathBuf, String)> {
|
2018-02-03 09:51:06 +00:00
|
|
|
paths
|
|
|
|
.iter()
|
|
|
|
.flat_map(|path| {
|
|
|
|
let path = test_data_dir().join(path);
|
|
|
|
test_from_dir(&path).into_iter()
|
|
|
|
})
|
2018-09-08 15:42:59 +00:00
|
|
|
.map(|path| {
|
|
|
|
let text = read_text(&path);
|
|
|
|
(path, text)
|
|
|
|
})
|
2018-02-03 09:51:06 +00:00
|
|
|
.collect()
|
2018-01-07 11:56:08 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
fn test_from_dir(dir: &Path) -> Vec<PathBuf> {
|
|
|
|
let mut acc = Vec::new();
|
2018-07-30 12:25:52 +00:00
|
|
|
for file in fs::read_dir(&dir).unwrap() {
|
2018-01-07 11:56:08 +00:00
|
|
|
let file = file.unwrap();
|
|
|
|
let path = file.path();
|
|
|
|
if path.extension().unwrap_or_default() == "rs" {
|
|
|
|
acc.push(path);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
acc.sort();
|
|
|
|
acc
|
|
|
|
}
|
|
|
|
|
|
|
|
fn project_dir() -> PathBuf {
|
|
|
|
let dir = env!("CARGO_MANIFEST_DIR");
|
|
|
|
PathBuf::from(dir)
|
2018-02-03 09:51:06 +00:00
|
|
|
.parent()
|
|
|
|
.unwrap()
|
|
|
|
.parent()
|
|
|
|
.unwrap()
|
2018-01-07 11:56:08 +00:00
|
|
|
.to_owned()
|
|
|
|
}
|
|
|
|
|
|
|
|
fn test_data_dir() -> PathBuf {
|
2018-09-16 09:54:24 +00:00
|
|
|
project_dir().join("crates/ra_syntax/tests/data")
|
2018-08-11 07:03:03 +00:00
|
|
|
}
|
|
|
|
|
2018-09-16 09:54:24 +00:00
|
|
|
fn dump_tokens(tokens: &[ra_syntax::Token], text: &str) -> String {
|
2018-08-11 07:03:03 +00:00
|
|
|
let mut acc = String::new();
|
|
|
|
let mut offset = 0;
|
|
|
|
for token in tokens {
|
|
|
|
let len: u32 = token.len.into();
|
|
|
|
let len = len as usize;
|
|
|
|
let token_text = &text[offset..offset + len];
|
|
|
|
offset += len;
|
|
|
|
write!(acc, "{:?} {} {:?}\n", token.kind, token.len, token_text).unwrap()
|
|
|
|
}
|
|
|
|
acc
|
2018-02-03 09:51:06 +00:00
|
|
|
}
|