mirror of
https://github.com/rust-lang/rust.git
synced 2025-05-14 02:49:40 +00:00
auto merge of #9636 : alexcrichton/rust/rustdoc, r=huonw
Commits have all the juicy details. Import thing to note in this pull request is that `rustdoc html crate.rs` becomes `rustdoc crate.rs`
This commit is contained in:
commit
0dce112af7
@ -3,49 +3,84 @@
|
|||||||
rustdoc \- generate documentation from Rust source code
|
rustdoc \- generate documentation from Rust source code
|
||||||
.SH SYNOPSIS
|
.SH SYNOPSIS
|
||||||
.B rustdoc
|
.B rustdoc
|
||||||
[\fIOPTIONS\fR] \fICRATEFILE\fR
|
[\fIOPTIONS\fR] \fIINPUT\fR
|
||||||
|
|
||||||
.SH DESCRIPTION
|
.SH DESCRIPTION
|
||||||
This tool generates API reference documentation by extracting comments from
|
This tool generates API reference documentation by extracting comments from
|
||||||
source code written in the Rust language, available at <\fBhttps://www.rust-
|
source code written in the Rust language, available at
|
||||||
lang.org\fR>. It provides several output formats for the generated
|
<\fBhttps://www.rust-lang.org\fR>. It accepts several input formats and provides
|
||||||
documentation.
|
several output formats for the generated documentation.
|
||||||
|
|
||||||
.SH COMMANDS
|
.SH OPTIONS
|
||||||
|
|
||||||
.TP
|
.TP
|
||||||
--output-dir <val>
|
-r --input-format <val>
|
||||||
Put documents here (default: .)
|
html or json (default: inferred)
|
||||||
.TP
|
.TP
|
||||||
--output-format <val>
|
-w --output-format <val>
|
||||||
markdown or html (default: html)
|
html or json (default: html)
|
||||||
.TP
|
.TP
|
||||||
--output-style <val>
|
-o --output <val>
|
||||||
doc-per-crate or doc-per-mod (default: doc-per-mod)
|
where to place the output (default: doc/ for html, doc.json for json)
|
||||||
.TP
|
.TP
|
||||||
--pandoc-cmd <val>
|
--passes <val>
|
||||||
Command for running pandoc
|
space-separated list of passes to run (default: '')
|
||||||
|
.TP
|
||||||
|
--no-defaults
|
||||||
|
don't run the default passes
|
||||||
|
.TP
|
||||||
|
--plugins <val>
|
||||||
|
space-separated list of plugins to run (default: '')
|
||||||
|
.TP
|
||||||
|
--plugin-path <val>
|
||||||
|
directory to load plugins from (default: /tmp/rustdoc_ng/plugins)
|
||||||
|
.TP
|
||||||
|
-L --library-path <val>
|
||||||
|
directory to add to crate search path
|
||||||
.TP
|
.TP
|
||||||
-h, --help
|
-h, --help
|
||||||
Print help
|
Print help
|
||||||
|
|
||||||
.SH "OUTPUT FORMATS"
|
.SH "OUTPUT FORMATS"
|
||||||
|
|
||||||
The rustdoc tool can generate documentation in either the Markdown
|
The rustdoc tool can generate output in either an HTML or JSON format.
|
||||||
or HTML formats. It requires the pandoc tool
|
|
||||||
<\fBhttp://johnmacfarlane.net/pandoc/\fR> for conversion features.
|
If using an HTML format, then the specified output destination will be the root
|
||||||
|
directory of an HTML structure for all the documentation. Pages will be placed
|
||||||
|
into this directory, and source files will also possibly be rendered into it as
|
||||||
|
well.
|
||||||
|
|
||||||
|
If using a JSON format, then the specified output destination will have the
|
||||||
|
rustdoc output serialized as JSON into it. This output format exists to
|
||||||
|
pre-compile documentation for crates, and for usage in non-rustdoc tools. The
|
||||||
|
JSON output is the following hash:
|
||||||
|
|
||||||
|
{
|
||||||
|
"schema": VERSION,
|
||||||
|
"crate": ...,
|
||||||
|
"plugins": ...,
|
||||||
|
}
|
||||||
|
|
||||||
|
The schema version indicates what the structure of crate/plugins will look
|
||||||
|
like. Within a schema version the structure will remain the same. The `crate`
|
||||||
|
field will contain all relevant documentation for the source being documented,
|
||||||
|
and the `plugins` field will contain the output of the plugins run over the
|
||||||
|
crate.
|
||||||
|
|
||||||
.SH "EXAMPLES"
|
.SH "EXAMPLES"
|
||||||
|
|
||||||
To generate documentation for the source in the current directory:
|
To generate documentation for the source in the current directory:
|
||||||
$ rustdoc hello.rs
|
$ rustdoc hello.rs
|
||||||
|
|
||||||
To build documentation into a subdirectory named 'doc' in the Markdown
|
List all available passes that rustdoc has, along with default passes:
|
||||||
format:
|
$ rustdoc --passes list
|
||||||
$ rustdoc --output-dir doc --output-format markdown hello.rs
|
|
||||||
|
|
||||||
The generated HTML can be viewed with any standard web browser, while
|
To precompile the documentation for a crate, and then use it to render html at
|
||||||
the Markdown version is well-suited for conversion into other formats.
|
a later date:
|
||||||
|
$ rustdoc -w json hello.rs
|
||||||
|
$ rustdoc doc.json
|
||||||
|
|
||||||
|
The generated HTML can be viewed with any standard web browser.
|
||||||
|
|
||||||
.SH "SEE ALSO"
|
.SH "SEE ALSO"
|
||||||
|
|
||||||
|
@ -227,7 +227,7 @@ RUSTDOC = $(HBIN2_H_$(CFG_BUILD_TRIPLE))/rustdoc$(X_$(CFG_BUILD_TRIPLE))
|
|||||||
define libdoc
|
define libdoc
|
||||||
doc/$(1)/index.html: $$(RUSTDOC) $$(TLIB2_T_$(3)_H_$(3))/$(CFG_STDLIB_$(3))
|
doc/$(1)/index.html: $$(RUSTDOC) $$(TLIB2_T_$(3)_H_$(3))/$(CFG_STDLIB_$(3))
|
||||||
@$$(call E, rustdoc: $$@)
|
@$$(call E, rustdoc: $$@)
|
||||||
$(Q)$(RUSTDOC) html $(2)
|
$(Q)$(RUSTDOC) $(2)
|
||||||
|
|
||||||
DOCS += doc/$(1)/index.html
|
DOCS += doc/$(1)/index.html
|
||||||
endef
|
endef
|
||||||
|
@ -28,6 +28,7 @@ use std::vec;
|
|||||||
use extra::arc::RWArc;
|
use extra::arc::RWArc;
|
||||||
use extra::json::ToJson;
|
use extra::json::ToJson;
|
||||||
use extra::sort;
|
use extra::sort;
|
||||||
|
use extra::time;
|
||||||
|
|
||||||
use syntax::ast;
|
use syntax::ast;
|
||||||
use syntax::ast_util::is_local;
|
use syntax::ast_util::is_local;
|
||||||
@ -60,7 +61,7 @@ struct Cache {
|
|||||||
// typaram id => name of that typaram
|
// typaram id => name of that typaram
|
||||||
typarams: HashMap<ast::NodeId, ~str>,
|
typarams: HashMap<ast::NodeId, ~str>,
|
||||||
// type id => all implementations for that type
|
// type id => all implementations for that type
|
||||||
impls: HashMap<ast::NodeId, ~[clean::Impl]>,
|
impls: HashMap<ast::NodeId, ~[(clean::Impl, Option<~str>)]>,
|
||||||
// path id => (full qualified path, shortty) -- used to generate urls
|
// path id => (full qualified path, shortty) -- used to generate urls
|
||||||
paths: HashMap<ast::NodeId, (~[~str], &'static str)>,
|
paths: HashMap<ast::NodeId, (~[~str], &'static str)>,
|
||||||
// trait id => method name => dox
|
// trait id => method name => dox
|
||||||
@ -76,7 +77,7 @@ struct Cache {
|
|||||||
struct SourceCollector<'self> {
|
struct SourceCollector<'self> {
|
||||||
seen: HashSet<~str>,
|
seen: HashSet<~str>,
|
||||||
dst: Path,
|
dst: Path,
|
||||||
cx: &'self Context,
|
cx: &'self mut Context,
|
||||||
}
|
}
|
||||||
|
|
||||||
struct Item<'self> { cx: &'self Context, item: &'self clean::Item, }
|
struct Item<'self> { cx: &'self Context, item: &'self clean::Item, }
|
||||||
@ -179,7 +180,9 @@ pub fn run(mut crate: clean::Crate, dst: Path) {
|
|||||||
w.flush();
|
w.flush();
|
||||||
}
|
}
|
||||||
|
|
||||||
if cx.include_sources {
|
info2!("emitting source files");
|
||||||
|
let started = time::precise_time_ns();
|
||||||
|
{
|
||||||
let dst = cx.dst.push("src");
|
let dst = cx.dst.push("src");
|
||||||
mkdir(&dst);
|
mkdir(&dst);
|
||||||
let dst = dst.push(crate.name);
|
let dst = dst.push(crate.name);
|
||||||
@ -187,13 +190,18 @@ pub fn run(mut crate: clean::Crate, dst: Path) {
|
|||||||
let mut folder = SourceCollector {
|
let mut folder = SourceCollector {
|
||||||
dst: dst,
|
dst: dst,
|
||||||
seen: HashSet::new(),
|
seen: HashSet::new(),
|
||||||
cx: &cx,
|
cx: &mut cx,
|
||||||
};
|
};
|
||||||
crate = folder.fold_crate(crate);
|
crate = folder.fold_crate(crate);
|
||||||
}
|
}
|
||||||
|
let ended = time::precise_time_ns();
|
||||||
|
info2!("Took {:.03f}s", (ended as f64 - started as f64) / 1e9f64);
|
||||||
|
|
||||||
// Now render the whole crate.
|
info2!("rendering the whole crate");
|
||||||
|
let started = time::precise_time_ns();
|
||||||
cx.crate(crate, cache);
|
cx.crate(crate, cache);
|
||||||
|
let ended = time::precise_time_ns();
|
||||||
|
info2!("Took {:.03f}s", (ended as f64 - started as f64) / 1e9f64);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn write(dst: Path, contents: &str) {
|
fn write(dst: Path, contents: &str) {
|
||||||
@ -229,16 +237,28 @@ fn clean_srcpath(src: &str, f: &fn(&str)) {
|
|||||||
|
|
||||||
impl<'self> DocFolder for SourceCollector<'self> {
|
impl<'self> DocFolder for SourceCollector<'self> {
|
||||||
fn fold_item(&mut self, item: clean::Item) -> Option<clean::Item> {
|
fn fold_item(&mut self, item: clean::Item) -> Option<clean::Item> {
|
||||||
if !self.seen.contains(&item.source.filename) {
|
if self.cx.include_sources && !self.seen.contains(&item.source.filename) {
|
||||||
self.emit_source(item.source.filename);
|
// If it turns out that we couldn't read this file, then we probably
|
||||||
|
// can't read any of the files (generating html output from json or
|
||||||
|
// something like that), so just don't include sources for the
|
||||||
|
// entire crate. The other option is maintaining this mapping on a
|
||||||
|
// per-file basis, but that's probably not worth it...
|
||||||
|
self.cx.include_sources = self.emit_source(item.source.filename);
|
||||||
self.seen.insert(item.source.filename.clone());
|
self.seen.insert(item.source.filename.clone());
|
||||||
|
|
||||||
|
if !self.cx.include_sources {
|
||||||
|
println!("warning: source code was requested to be rendered, \
|
||||||
|
but `{}` is a missing source file.",
|
||||||
|
item.source.filename);
|
||||||
|
println!(" skipping rendering of source code");
|
||||||
|
}
|
||||||
}
|
}
|
||||||
self.fold_item_recur(item)
|
self.fold_item_recur(item)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<'self> SourceCollector<'self> {
|
impl<'self> SourceCollector<'self> {
|
||||||
fn emit_source(&self, filename: &str) {
|
fn emit_source(&mut self, filename: &str) -> bool {
|
||||||
let p = Path(filename);
|
let p = Path(filename);
|
||||||
|
|
||||||
// Read the contents of the file
|
// Read the contents of the file
|
||||||
@ -251,7 +271,11 @@ impl<'self> SourceCollector<'self> {
|
|||||||
// If we couldn't open this file, then just returns because it
|
// If we couldn't open this file, then just returns because it
|
||||||
// probably means that it's some standard library macro thing and we
|
// probably means that it's some standard library macro thing and we
|
||||||
// can't have the source to it anyway.
|
// can't have the source to it anyway.
|
||||||
let mut r = match r { Some(r) => r, None => return };
|
let mut r = match r {
|
||||||
|
Some(r) => r,
|
||||||
|
// eew macro hacks
|
||||||
|
None => return filename == "<std-macros>"
|
||||||
|
};
|
||||||
|
|
||||||
// read everything
|
// read everything
|
||||||
loop {
|
loop {
|
||||||
@ -273,7 +297,8 @@ impl<'self> SourceCollector<'self> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
let dst = cur.push(*p.components.last() + ".html");
|
let dst = cur.push(*p.components.last() + ".html");
|
||||||
let mut w = dst.open_writer(io::CreateOrTruncate);
|
let w = dst.open_writer(io::CreateOrTruncate);
|
||||||
|
let mut w = BufferedWriter::new(w);
|
||||||
|
|
||||||
let title = format!("{} -- source", *dst.components.last());
|
let title = format!("{} -- source", *dst.components.last());
|
||||||
let page = layout::Page {
|
let page = layout::Page {
|
||||||
@ -283,6 +308,8 @@ impl<'self> SourceCollector<'self> {
|
|||||||
};
|
};
|
||||||
layout::render(&mut w as &mut io::Writer, &self.cx.layout,
|
layout::render(&mut w as &mut io::Writer, &self.cx.layout,
|
||||||
&page, &(""), &Source(contents.as_slice()));
|
&page, &(""), &Source(contents.as_slice()));
|
||||||
|
w.flush();
|
||||||
|
return true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -427,21 +454,34 @@ impl DocFolder for Cache {
|
|||||||
// implementations elsewhere
|
// implementations elsewhere
|
||||||
let ret = match self.fold_item_recur(item) {
|
let ret = match self.fold_item_recur(item) {
|
||||||
Some(item) => {
|
Some(item) => {
|
||||||
match item.inner {
|
match item {
|
||||||
clean::ImplItem(i) => {
|
clean::Item{ attrs, inner: clean::ImplItem(i), _ } => {
|
||||||
match i.for_ {
|
match i.for_ {
|
||||||
clean::ResolvedPath { did, _ } if is_local(did) => {
|
clean::ResolvedPath { did, _ } if is_local(did) => {
|
||||||
let id = did.node;
|
let id = did.node;
|
||||||
let v = do self.impls.find_or_insert_with(id) |_| {
|
let v = do self.impls.find_or_insert_with(id) |_| {
|
||||||
~[]
|
~[]
|
||||||
};
|
};
|
||||||
v.push(i);
|
// extract relevant documentation for this impl
|
||||||
|
match attrs.move_iter().find(|a| {
|
||||||
|
match *a {
|
||||||
|
clean::NameValue(~"doc", _) => true,
|
||||||
|
_ => false
|
||||||
|
}
|
||||||
|
}) {
|
||||||
|
Some(clean::NameValue(_, dox)) => {
|
||||||
|
v.push((i, Some(dox)));
|
||||||
|
}
|
||||||
|
Some(*) | None => {
|
||||||
|
v.push((i, None));
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
_ => {}
|
_ => {}
|
||||||
}
|
}
|
||||||
None
|
None
|
||||||
}
|
}
|
||||||
_ => Some(item),
|
i => Some(i),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
i => i,
|
i => i,
|
||||||
@ -1050,10 +1090,24 @@ fn render_method(w: &mut io::Writer, meth: &clean::Item, withlink: bool) {
|
|||||||
|
|
||||||
fn item_struct(w: &mut io::Writer, it: &clean::Item, s: &clean::Struct) {
|
fn item_struct(w: &mut io::Writer, it: &clean::Item, s: &clean::Struct) {
|
||||||
write!(w, "<pre class='struct'>");
|
write!(w, "<pre class='struct'>");
|
||||||
render_struct(w, it, Some(&s.generics), s.struct_type, s.fields, "");
|
render_struct(w, it, Some(&s.generics), s.struct_type, s.fields, "", true);
|
||||||
write!(w, "</pre>");
|
write!(w, "</pre>");
|
||||||
|
|
||||||
document(w, it);
|
document(w, it);
|
||||||
|
match s.struct_type {
|
||||||
|
doctree::Plain => {
|
||||||
|
write!(w, "<h2 class='fields'>Fields</h2>\n<table>");
|
||||||
|
for field in s.fields.iter() {
|
||||||
|
write!(w, "<tr><td id='structfield.{name}'>\
|
||||||
|
<code>{name}</code></td><td>",
|
||||||
|
name = field.name.get_ref().as_slice());
|
||||||
|
document(w, field);
|
||||||
|
write!(w, "</td></tr>");
|
||||||
|
}
|
||||||
|
write!(w, "</table>");
|
||||||
|
}
|
||||||
|
_ => {}
|
||||||
|
}
|
||||||
render_methods(w, it);
|
render_methods(w, it);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1067,34 +1121,46 @@ fn item_enum(w: &mut io::Writer, it: &clean::Item, e: &clean::Enum) {
|
|||||||
} else {
|
} else {
|
||||||
write!(w, " \\{\n");
|
write!(w, " \\{\n");
|
||||||
for v in e.variants.iter() {
|
for v in e.variants.iter() {
|
||||||
let name = format!("<a name='variant.{0}'>{0}</a>",
|
write!(w, " ");
|
||||||
v.name.get_ref().as_slice());
|
let name = v.name.get_ref().as_slice();
|
||||||
match v.inner {
|
match v.inner {
|
||||||
clean::VariantItem(ref var) => {
|
clean::VariantItem(ref var) => {
|
||||||
match var.kind {
|
match var.kind {
|
||||||
clean::CLikeVariant => write!(w, " {},\n", name),
|
clean::CLikeVariant => write!(w, "{}", name),
|
||||||
clean::TupleVariant(ref tys) => {
|
clean::TupleVariant(ref tys) => {
|
||||||
write!(w, " {}(", name);
|
write!(w, "{}(", name);
|
||||||
for (i, ty) in tys.iter().enumerate() {
|
for (i, ty) in tys.iter().enumerate() {
|
||||||
if i > 0 { write!(w, ", ") }
|
if i > 0 { write!(w, ", ") }
|
||||||
write!(w, "{}", *ty);
|
write!(w, "{}", *ty);
|
||||||
}
|
}
|
||||||
write!(w, "),\n");
|
write!(w, ")");
|
||||||
}
|
}
|
||||||
clean::StructVariant(ref s) => {
|
clean::StructVariant(ref s) => {
|
||||||
render_struct(w, v, None, s.struct_type, s.fields,
|
render_struct(w, v, None, s.struct_type, s.fields,
|
||||||
" ");
|
" ", false);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
_ => unreachable!()
|
_ => unreachable!()
|
||||||
}
|
}
|
||||||
|
write!(w, ",\n");
|
||||||
}
|
}
|
||||||
write!(w, "\\}");
|
write!(w, "\\}");
|
||||||
}
|
}
|
||||||
write!(w, "</pre>");
|
write!(w, "</pre>");
|
||||||
|
|
||||||
document(w, it);
|
document(w, it);
|
||||||
|
if e.variants.len() > 0 {
|
||||||
|
write!(w, "<h2 class='variants'>Variants</h2>\n<table>");
|
||||||
|
for variant in e.variants.iter() {
|
||||||
|
write!(w, "<tr><td id='variant.{name}'><code>{name}</code></td><td>",
|
||||||
|
name = variant.name.get_ref().as_slice());
|
||||||
|
document(w, variant);
|
||||||
|
write!(w, "</td></tr>");
|
||||||
|
}
|
||||||
|
write!(w, "</table>");
|
||||||
|
|
||||||
|
}
|
||||||
render_methods(w, it);
|
render_methods(w, it);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1102,9 +1168,11 @@ fn render_struct(w: &mut io::Writer, it: &clean::Item,
|
|||||||
g: Option<&clean::Generics>,
|
g: Option<&clean::Generics>,
|
||||||
ty: doctree::StructType,
|
ty: doctree::StructType,
|
||||||
fields: &[clean::Item],
|
fields: &[clean::Item],
|
||||||
tab: &str) {
|
tab: &str,
|
||||||
write!(w, "{}struct {}",
|
structhead: bool) {
|
||||||
|
write!(w, "{}{}{}",
|
||||||
VisSpace(it.visibility),
|
VisSpace(it.visibility),
|
||||||
|
if structhead {"struct "} else {""},
|
||||||
it.name.get_ref().as_slice());
|
it.name.get_ref().as_slice());
|
||||||
match g {
|
match g {
|
||||||
Some(g) => write!(w, "{}", *g),
|
Some(g) => write!(w, "{}", *g),
|
||||||
@ -1112,16 +1180,15 @@ fn render_struct(w: &mut io::Writer, it: &clean::Item,
|
|||||||
}
|
}
|
||||||
match ty {
|
match ty {
|
||||||
doctree::Plain => {
|
doctree::Plain => {
|
||||||
write!(w, " \\{\n");
|
write!(w, " \\{\n{}", tab);
|
||||||
for field in fields.iter() {
|
for field in fields.iter() {
|
||||||
match field.inner {
|
match field.inner {
|
||||||
clean::StructFieldItem(ref ty) => {
|
clean::StructFieldItem(ref ty) => {
|
||||||
write!(w, " {}<a name='structfield.{name}'>{name}</a>: \
|
write!(w, " {}{}: {},\n{}",
|
||||||
{},\n{}",
|
|
||||||
VisSpace(field.visibility),
|
VisSpace(field.visibility),
|
||||||
|
field.name.get_ref().as_slice(),
|
||||||
ty.type_,
|
ty.type_,
|
||||||
tab,
|
tab);
|
||||||
name = field.name.get_ref().as_slice());
|
|
||||||
}
|
}
|
||||||
_ => unreachable!()
|
_ => unreachable!()
|
||||||
}
|
}
|
||||||
@ -1151,22 +1218,26 @@ fn render_methods(w: &mut io::Writer, it: &clean::Item) {
|
|||||||
do cache.read |c| {
|
do cache.read |c| {
|
||||||
match c.impls.find(&it.id) {
|
match c.impls.find(&it.id) {
|
||||||
Some(v) => {
|
Some(v) => {
|
||||||
let mut non_trait = v.iter().filter(|i| i.trait_.is_none());
|
let mut non_trait = v.iter().filter(|p| {
|
||||||
|
p.n0_ref().trait_.is_none()
|
||||||
|
});
|
||||||
let non_trait = non_trait.to_owned_vec();
|
let non_trait = non_trait.to_owned_vec();
|
||||||
let mut traits = v.iter().filter(|i| i.trait_.is_some());
|
let mut traits = v.iter().filter(|p| {
|
||||||
|
p.n0_ref().trait_.is_some()
|
||||||
|
});
|
||||||
let traits = traits.to_owned_vec();
|
let traits = traits.to_owned_vec();
|
||||||
|
|
||||||
if non_trait.len() > 0 {
|
if non_trait.len() > 0 {
|
||||||
write!(w, "<h2 id='methods'>Methods</h2>");
|
write!(w, "<h2 id='methods'>Methods</h2>");
|
||||||
for &i in non_trait.iter() {
|
for &(ref i, ref dox) in non_trait.move_iter() {
|
||||||
render_impl(w, i);
|
render_impl(w, i, dox);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if traits.len() > 0 {
|
if traits.len() > 0 {
|
||||||
write!(w, "<h2 id='implementations'>Trait \
|
write!(w, "<h2 id='implementations'>Trait \
|
||||||
Implementations</h2>");
|
Implementations</h2>");
|
||||||
for &i in traits.iter() {
|
for &(ref i, ref dox) in traits.move_iter() {
|
||||||
render_impl(w, i);
|
render_impl(w, i, dox);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1176,7 +1247,7 @@ fn render_methods(w: &mut io::Writer, it: &clean::Item) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn render_impl(w: &mut io::Writer, i: &clean::Impl) {
|
fn render_impl(w: &mut io::Writer, i: &clean::Impl, dox: &Option<~str>) {
|
||||||
write!(w, "<h3 class='impl'><code>impl{} ", i.generics);
|
write!(w, "<h3 class='impl'><code>impl{} ", i.generics);
|
||||||
let trait_id = match i.trait_ {
|
let trait_id = match i.trait_ {
|
||||||
Some(ref ty) => {
|
Some(ref ty) => {
|
||||||
@ -1189,6 +1260,13 @@ fn render_impl(w: &mut io::Writer, i: &clean::Impl) {
|
|||||||
None => None
|
None => None
|
||||||
};
|
};
|
||||||
write!(w, "{}</code></h3>", i.for_);
|
write!(w, "{}</code></h3>", i.for_);
|
||||||
|
match *dox {
|
||||||
|
Some(ref dox) => {
|
||||||
|
write!(w, "<div class='docblock'>{}</div>",
|
||||||
|
Markdown(dox.as_slice()));
|
||||||
|
}
|
||||||
|
None => {}
|
||||||
|
}
|
||||||
write!(w, "<div class='methods'>");
|
write!(w, "<div class='methods'>");
|
||||||
for meth in i.methods.iter() {
|
for meth in i.methods.iter() {
|
||||||
write!(w, "<h4 id='method.{}' class='method'><code>",
|
write!(w, "<h4 id='method.{}' class='method'><code>",
|
||||||
|
@ -224,8 +224,8 @@
|
|||||||
}, 20);
|
}, 20);
|
||||||
});
|
});
|
||||||
|
|
||||||
$(document).off('keyup.searchnav');
|
$(document).off('keypress.searchnav');
|
||||||
$(document).on('keyup.searchnav', function (e) {
|
$(document).on('keypress.searchnav', function (e) {
|
||||||
var $active = $results.filter('.highlighted');
|
var $active = $results.filter('.highlighted');
|
||||||
|
|
||||||
if (e.keyCode === 38) { // up
|
if (e.keyCode === 38) { // up
|
||||||
@ -321,8 +321,8 @@
|
|||||||
output += "</p>";
|
output += "</p>";
|
||||||
$('#main.content').addClass('hidden');
|
$('#main.content').addClass('hidden');
|
||||||
$('#search.content').removeClass('hidden').html(output);
|
$('#search.content').removeClass('hidden').html(output);
|
||||||
$('.search-results .desc').width($('.content').width() - 40 -
|
$('#search .desc').width($('#search').width() - 40 -
|
||||||
$('.content td:first-child').first().width());
|
$('#search td:first-child').first().width());
|
||||||
initSearchNav();
|
initSearchNav();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -21,13 +21,15 @@ extern mod syntax;
|
|||||||
extern mod rustc;
|
extern mod rustc;
|
||||||
extern mod extra;
|
extern mod extra;
|
||||||
|
|
||||||
use extra::serialize::Encodable;
|
|
||||||
use extra::time;
|
|
||||||
use extra::getopts::groups;
|
|
||||||
use std::cell::Cell;
|
use std::cell::Cell;
|
||||||
use std::rt::io;
|
|
||||||
use std::rt::io::Writer;
|
use std::rt::io::Writer;
|
||||||
use std::rt::io::file::FileInfo;
|
use std::rt::io::file::FileInfo;
|
||||||
|
use std::rt::io;
|
||||||
|
use extra::getopts;
|
||||||
|
use extra::getopts::groups;
|
||||||
|
use extra::json;
|
||||||
|
use extra::serialize::{Decodable, Encodable};
|
||||||
|
use extra::time;
|
||||||
|
|
||||||
pub mod clean;
|
pub mod clean;
|
||||||
pub mod core;
|
pub mod core;
|
||||||
@ -70,9 +72,7 @@ static DEFAULT_PASSES: &'static [&'static str] = &[
|
|||||||
|
|
||||||
local_data_key!(pub ctxtkey: @core::DocContext)
|
local_data_key!(pub ctxtkey: @core::DocContext)
|
||||||
|
|
||||||
enum OutputFormat {
|
type Output = (clean::Crate, ~[plugins::PluginJson]);
|
||||||
HTML, JSON
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn main() {
|
pub fn main() {
|
||||||
std::os::set_exit_status(main_args(std::os::args()));
|
std::os::set_exit_status(main_args(std::os::args()));
|
||||||
@ -81,6 +81,12 @@ pub fn main() {
|
|||||||
pub fn opts() -> ~[groups::OptGroup] {
|
pub fn opts() -> ~[groups::OptGroup] {
|
||||||
use extra::getopts::groups::*;
|
use extra::getopts::groups::*;
|
||||||
~[
|
~[
|
||||||
|
optflag("h", "help", "show this help message"),
|
||||||
|
optopt("r", "input-format", "the input type of the specified file",
|
||||||
|
"[rust|json]"),
|
||||||
|
optopt("w", "output-format", "the output type to write",
|
||||||
|
"[html|json]"),
|
||||||
|
optopt("o", "output", "where to place the output", "PATH"),
|
||||||
optmulti("L", "library-path", "directory to add to crate search path",
|
optmulti("L", "library-path", "directory to add to crate search path",
|
||||||
"DIR"),
|
"DIR"),
|
||||||
optmulti("", "plugin-path", "directory to load plugins from", "DIR"),
|
optmulti("", "plugin-path", "directory to load plugins from", "DIR"),
|
||||||
@ -89,32 +95,22 @@ pub fn opts() -> ~[groups::OptGroup] {
|
|||||||
"PASSES"),
|
"PASSES"),
|
||||||
optmulti("", "plugins", "space separated list of plugins to also load",
|
optmulti("", "plugins", "space separated list of plugins to also load",
|
||||||
"PLUGINS"),
|
"PLUGINS"),
|
||||||
optflag("h", "help", "show this help message"),
|
optflag("", "no-defaults", "don't run the default passes"),
|
||||||
optflag("", "nodefaults", "don't run the default passes"),
|
|
||||||
optopt("o", "output", "where to place the output", "PATH"),
|
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn usage(argv0: &str) {
|
pub fn usage(argv0: &str) {
|
||||||
println(groups::usage(format!("{} [options] [html|json] <crate>",
|
println(groups::usage(format!("{} [options] <input>", argv0), opts()));
|
||||||
argv0), opts()));
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn main_args(args: &[~str]) -> int {
|
pub fn main_args(args: &[~str]) -> int {
|
||||||
//use extra::getopts::groups::*;
|
|
||||||
|
|
||||||
let matches = groups::getopts(args.tail(), opts()).unwrap();
|
let matches = groups::getopts(args.tail(), opts()).unwrap();
|
||||||
|
|
||||||
if matches.opt_present("h") || matches.opt_present("help") {
|
if matches.opt_present("h") || matches.opt_present("help") {
|
||||||
usage(args[0]);
|
usage(args[0]);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut default_passes = !matches.opt_present("nodefaults");
|
if matches.opt_strs("passes") == ~[~"list"] {
|
||||||
let mut passes = matches.opt_strs("passes");
|
|
||||||
let mut plugins = matches.opt_strs("plugins");
|
|
||||||
|
|
||||||
if passes == ~[~"list"] {
|
|
||||||
println("Available passes for running rustdoc:");
|
println("Available passes for running rustdoc:");
|
||||||
for &(name, _, description) in PASSES.iter() {
|
for &(name, _, description) in PASSES.iter() {
|
||||||
println!("{:>20s} - {}", name, description);
|
println!("{:>20s} - {}", name, description);
|
||||||
@ -126,26 +122,69 @@ pub fn main_args(args: &[~str]) -> int {
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
let (format, cratefile) = match matches.free.clone() {
|
let (crate, res) = match acquire_input(&matches) {
|
||||||
[~"json", crate] => (JSON, crate),
|
Ok(pair) => pair,
|
||||||
[~"html", crate] => (HTML, crate),
|
Err(s) => {
|
||||||
[s, _] => {
|
println!("input error: {}", s);
|
||||||
println!("Unknown output format: `{}`", s);
|
|
||||||
usage(args[0]);
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
[_, .._] => {
|
|
||||||
println!("Expected exactly one crate to process");
|
|
||||||
usage(args[0]);
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
println!("Expected an output format and then one crate");
|
|
||||||
usage(args[0]);
|
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
info2!("going to format");
|
||||||
|
let started = time::precise_time_ns();
|
||||||
|
let output = matches.opt_str("o").map(|s| Path(*s));
|
||||||
|
match matches.opt_str("w") {
|
||||||
|
Some(~"html") | None => {
|
||||||
|
html::render::run(crate, output.unwrap_or(Path("doc")))
|
||||||
|
}
|
||||||
|
Some(~"json") => {
|
||||||
|
json_output(crate, res, output.unwrap_or(Path("doc.json")))
|
||||||
|
}
|
||||||
|
Some(s) => {
|
||||||
|
println!("unknown output format: {}", s);
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
let ended = time::precise_time_ns();
|
||||||
|
info2!("Took {:.03f}s", (ended as f64 - started as f64) / 1e9f64);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Looks inside the command line arguments to extract the relevant input format
|
||||||
|
/// and files and then generates the necessary rustdoc output for formatting.
|
||||||
|
fn acquire_input(matches: &getopts::Matches) -> Result<Output, ~str> {
|
||||||
|
if matches.free.len() == 0 {
|
||||||
|
return Err(~"expected an input file to act on");
|
||||||
|
} if matches.free.len() > 1 {
|
||||||
|
return Err(~"only one input file may be specified");
|
||||||
|
}
|
||||||
|
|
||||||
|
let input = matches.free[0].as_slice();
|
||||||
|
match matches.opt_str("r") {
|
||||||
|
Some(~"rust") => Ok(rust_input(input, matches)),
|
||||||
|
Some(~"json") => json_input(input),
|
||||||
|
Some(s) => Err("unknown input format: " + s),
|
||||||
|
None => {
|
||||||
|
if input.ends_with(".json") {
|
||||||
|
json_input(input)
|
||||||
|
} else {
|
||||||
|
Ok(rust_input(input, matches))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Interprets the input file as a rust source file, passing it through the
|
||||||
|
/// compiler all the way through the analysis passes. The rustdoc output is then
|
||||||
|
/// generated from the cleaned AST of the crate.
|
||||||
|
///
|
||||||
|
/// This form of input will run all of the plug/cleaning passes
|
||||||
|
fn rust_input(cratefile: &str, matches: &getopts::Matches) -> Output {
|
||||||
|
let mut default_passes = !matches.opt_present("no-defaults");
|
||||||
|
let mut passes = matches.opt_strs("passes");
|
||||||
|
let mut plugins = matches.opt_strs("plugins");
|
||||||
|
|
||||||
// First, parse the crate and extract all relevant information.
|
// First, parse the crate and extract all relevant information.
|
||||||
let libs = Cell::new(matches.opt_strs("L").map(|s| Path(*s)));
|
let libs = Cell::new(matches.opt_strs("L").map(|s| Path(*s)));
|
||||||
let cr = Cell::new(Path(cratefile));
|
let cr = Cell::new(Path(cratefile));
|
||||||
@ -188,7 +227,8 @@ pub fn main_args(args: &[~str]) -> int {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Load all plugins/passes into a PluginManager
|
// Load all plugins/passes into a PluginManager
|
||||||
let mut pm = plugins::PluginManager::new(Path("/tmp/rustdoc_ng/plugins"));
|
let path = matches.opt_str("plugin-path").unwrap_or(~"/tmp/rustdoc_ng/plugins");
|
||||||
|
let mut pm = plugins::PluginManager::new(Path(path));
|
||||||
for pass in passes.iter() {
|
for pass in passes.iter() {
|
||||||
let plugin = match PASSES.iter().position(|&(p, _, _)| p == *pass) {
|
let plugin = match PASSES.iter().position(|&(p, _, _)| p == *pass) {
|
||||||
Some(i) => PASSES[i].n1(),
|
Some(i) => PASSES[i].n1(),
|
||||||
@ -206,45 +246,73 @@ pub fn main_args(args: &[~str]) -> int {
|
|||||||
|
|
||||||
// Run everything!
|
// Run everything!
|
||||||
info2!("Executing passes/plugins");
|
info2!("Executing passes/plugins");
|
||||||
let (crate, res) = pm.run_plugins(crate);
|
return pm.run_plugins(crate);
|
||||||
|
|
||||||
info2!("going to format");
|
|
||||||
let started = time::precise_time_ns();
|
|
||||||
let output = matches.opt_str("o").map(|s| Path(*s));
|
|
||||||
match format {
|
|
||||||
HTML => { html::render::run(crate, output.unwrap_or(Path("doc"))) }
|
|
||||||
JSON => { jsonify(crate, res, output.unwrap_or(Path("doc.json"))) }
|
|
||||||
}
|
|
||||||
let ended = time::precise_time_ns();
|
|
||||||
info2!("Took {:.03f}s", (ended as f64 - started as f64) / 1000000000f64);
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn jsonify(crate: clean::Crate, res: ~[plugins::PluginJson], dst: Path) {
|
/// This input format purely deserializes the json output file. No passes are
|
||||||
|
/// run over the deserialized output.
|
||||||
|
fn json_input(input: &str) -> Result<Output, ~str> {
|
||||||
|
let input = match ::std::io::file_reader(&Path(input)) {
|
||||||
|
Ok(i) => i,
|
||||||
|
Err(s) => return Err(s),
|
||||||
|
};
|
||||||
|
match json::from_reader(input) {
|
||||||
|
Err(s) => Err(s.to_str()),
|
||||||
|
Ok(json::Object(obj)) => {
|
||||||
|
let mut obj = obj;
|
||||||
|
// Make sure the schema is what we expect
|
||||||
|
match obj.pop(&~"schema") {
|
||||||
|
Some(json::String(version)) => {
|
||||||
|
if version.as_slice() != SCHEMA_VERSION {
|
||||||
|
return Err(format!("sorry, but I only understand \
|
||||||
|
version {}", SCHEMA_VERSION))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Some(*) => return Err(~"malformed json"),
|
||||||
|
None => return Err(~"expected a schema version"),
|
||||||
|
}
|
||||||
|
let crate = match obj.pop(&~"crate") {
|
||||||
|
Some(json) => {
|
||||||
|
let mut d = json::Decoder(json);
|
||||||
|
Decodable::decode(&mut d)
|
||||||
|
}
|
||||||
|
None => return Err(~"malformed json"),
|
||||||
|
};
|
||||||
|
// XXX: this should read from the "plugins" field, but currently
|
||||||
|
// Json doesn't implement decodable...
|
||||||
|
let plugin_output = ~[];
|
||||||
|
Ok((crate, plugin_output))
|
||||||
|
}
|
||||||
|
Ok(*) => Err(~"malformed json input: expected an object at the top"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Outputs the crate/plugin json as a giant json blob at the specified
|
||||||
|
/// destination.
|
||||||
|
fn json_output(crate: clean::Crate, res: ~[plugins::PluginJson], dst: Path) {
|
||||||
// {
|
// {
|
||||||
// "schema": version,
|
// "schema": version,
|
||||||
// "crate": { parsed crate ... },
|
// "crate": { parsed crate ... },
|
||||||
// "plugins": { output of plugins ... }
|
// "plugins": { output of plugins ... }
|
||||||
// }
|
// }
|
||||||
let mut json = ~extra::treemap::TreeMap::new();
|
let mut json = ~extra::treemap::TreeMap::new();
|
||||||
json.insert(~"schema", extra::json::String(SCHEMA_VERSION.to_owned()));
|
json.insert(~"schema", json::String(SCHEMA_VERSION.to_owned()));
|
||||||
let plugins_json = ~res.move_iter().filter_map(|opt| opt).collect();
|
let plugins_json = ~res.move_iter().filter_map(|opt| opt).collect();
|
||||||
|
|
||||||
// FIXME #8335: yuck, Rust -> str -> JSON round trip! No way to .encode
|
// FIXME #8335: yuck, Rust -> str -> JSON round trip! No way to .encode
|
||||||
// straight to the Rust JSON representation.
|
// straight to the Rust JSON representation.
|
||||||
let crate_json_str = do std::io::with_str_writer |w| {
|
let crate_json_str = do std::io::with_str_writer |w| {
|
||||||
crate.encode(&mut extra::json::Encoder(w));
|
crate.encode(&mut json::Encoder(w));
|
||||||
};
|
};
|
||||||
let crate_json = match extra::json::from_str(crate_json_str) {
|
let crate_json = match json::from_str(crate_json_str) {
|
||||||
Ok(j) => j,
|
Ok(j) => j,
|
||||||
Err(_) => fail2!("Rust generated JSON is invalid??")
|
Err(_) => fail2!("Rust generated JSON is invalid??")
|
||||||
};
|
};
|
||||||
|
|
||||||
json.insert(~"crate", crate_json);
|
json.insert(~"crate", crate_json);
|
||||||
json.insert(~"plugins", extra::json::Object(plugins_json));
|
json.insert(~"plugins", json::Object(plugins_json));
|
||||||
|
|
||||||
let mut file = dst.open_writer(io::Create).unwrap();
|
let mut file = dst.open_writer(io::Create).unwrap();
|
||||||
let output = extra::json::Object(json).to_str();
|
let output = json::Object(json).to_str();
|
||||||
file.write(output.as_bytes());
|
file.write(output.as_bytes());
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user