| //! This module provides `StaticIndex` which is used for powering |
| //! read-only code browsers and emitting LSIF |
| |
| use hir::{db::HirDatabase, Crate, HirFileIdExt, Module, Semantics}; |
| use ide_db::{ |
| base_db::{FileId, FileRange, SourceDatabaseExt}, |
| defs::Definition, |
| documentation::Documentation, |
| famous_defs::FamousDefs, |
| helpers::get_definition, |
| FxHashMap, FxHashSet, RootDatabase, |
| }; |
| use syntax::{AstNode, SyntaxKind::*, SyntaxNode, TextRange, T}; |
| |
| use crate::inlay_hints::InlayFieldsToResolve; |
| use crate::navigation_target::UpmappingResult; |
| use crate::{ |
| hover::hover_for_definition, |
| inlay_hints::AdjustmentHintsMode, |
| moniker::{def_to_kind, def_to_moniker, MonikerResult, SymbolInformationKind}, |
| parent_module::crates_for, |
| Analysis, Fold, HoverConfig, HoverResult, InlayHint, InlayHintsConfig, TryToNav, |
| }; |
| |
| /// A static representation of fully analyzed source code. |
| /// |
| /// The intended use-case is powering read-only code browsers and emitting LSIF/SCIP. |
| #[derive(Debug)] |
| pub struct StaticIndex<'a> { |
| pub files: Vec<StaticIndexedFile>, |
| pub tokens: TokenStore, |
| analysis: &'a Analysis, |
| db: &'a RootDatabase, |
| def_map: FxHashMap<Definition, TokenId>, |
| } |
| |
| #[derive(Debug)] |
| pub struct ReferenceData { |
| pub range: FileRange, |
| pub is_definition: bool, |
| } |
| |
| #[derive(Debug)] |
| pub struct TokenStaticData { |
| pub documentation: Option<Documentation>, |
| pub hover: Option<HoverResult>, |
| pub definition: Option<FileRange>, |
| pub references: Vec<ReferenceData>, |
| pub moniker: Option<MonikerResult>, |
| pub display_name: Option<String>, |
| pub enclosing_moniker: Option<MonikerResult>, |
| pub signature: Option<String>, |
| pub kind: SymbolInformationKind, |
| } |
| |
| #[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)] |
| pub struct TokenId(usize); |
| |
| impl TokenId { |
| pub fn raw(self) -> usize { |
| self.0 |
| } |
| } |
| |
| #[derive(Default, Debug)] |
| pub struct TokenStore(Vec<TokenStaticData>); |
| |
| impl TokenStore { |
| pub fn insert(&mut self, data: TokenStaticData) -> TokenId { |
| let id = TokenId(self.0.len()); |
| self.0.push(data); |
| id |
| } |
| |
| pub fn get_mut(&mut self, id: TokenId) -> Option<&mut TokenStaticData> { |
| self.0.get_mut(id.0) |
| } |
| |
| pub fn get(&self, id: TokenId) -> Option<&TokenStaticData> { |
| self.0.get(id.0) |
| } |
| |
| pub fn iter(self) -> impl Iterator<Item = (TokenId, TokenStaticData)> { |
| self.0.into_iter().enumerate().map(|(id, data)| (TokenId(id), data)) |
| } |
| } |
| |
| #[derive(Debug)] |
| pub struct StaticIndexedFile { |
| pub file_id: FileId, |
| pub folds: Vec<Fold>, |
| pub inlay_hints: Vec<InlayHint>, |
| pub tokens: Vec<(TextRange, TokenId)>, |
| } |
| |
| fn all_modules(db: &dyn HirDatabase) -> Vec<Module> { |
| let mut worklist: Vec<_> = |
| Crate::all(db).into_iter().map(|krate| krate.root_module()).collect(); |
| let mut modules = Vec::new(); |
| |
| while let Some(module) = worklist.pop() { |
| modules.push(module); |
| worklist.extend(module.children(db)); |
| } |
| |
| modules |
| } |
| |
| fn documentation_for_definition( |
| sema: &Semantics<'_, RootDatabase>, |
| def: Definition, |
| scope_node: &SyntaxNode, |
| ) -> Option<Documentation> { |
| let famous_defs = match &def { |
| Definition::BuiltinType(_) => Some(FamousDefs(sema, sema.scope(scope_node)?.krate())), |
| _ => None, |
| }; |
| |
| def.docs(sema.db, famous_defs.as_ref()) |
| } |
| |
| impl StaticIndex<'_> { |
| fn add_file(&mut self, file_id: FileId) { |
| let current_crate = crates_for(self.db, file_id).pop().map(Into::into); |
| let folds = self.analysis.folding_ranges(file_id).unwrap(); |
| let inlay_hints = self |
| .analysis |
| .inlay_hints( |
| &InlayHintsConfig { |
| render_colons: true, |
| discriminant_hints: crate::DiscriminantHints::Fieldless, |
| type_hints: true, |
| parameter_hints: true, |
| chaining_hints: true, |
| closure_return_type_hints: crate::ClosureReturnTypeHints::WithBlock, |
| lifetime_elision_hints: crate::LifetimeElisionHints::Never, |
| adjustment_hints: crate::AdjustmentHints::Never, |
| adjustment_hints_mode: AdjustmentHintsMode::Prefix, |
| adjustment_hints_hide_outside_unsafe: false, |
| implicit_drop_hints: false, |
| hide_named_constructor_hints: false, |
| hide_closure_initialization_hints: false, |
| closure_style: hir::ClosureStyle::ImplFn, |
| param_names_for_lifetime_elision_hints: false, |
| binding_mode_hints: false, |
| max_length: Some(25), |
| closure_capture_hints: false, |
| closing_brace_hints_min_lines: Some(25), |
| fields_to_resolve: InlayFieldsToResolve::empty(), |
| range_exclusive_hints: false, |
| }, |
| file_id, |
| None, |
| ) |
| .unwrap(); |
| // hovers |
| let sema = hir::Semantics::new(self.db); |
| let tokens_or_nodes = sema.parse(file_id).syntax().clone(); |
| let tokens = tokens_or_nodes.descendants_with_tokens().filter_map(|it| match it { |
| syntax::NodeOrToken::Node(_) => None, |
| syntax::NodeOrToken::Token(it) => Some(it), |
| }); |
| let hover_config = HoverConfig { |
| links_in_hover: true, |
| memory_layout: None, |
| documentation: true, |
| keywords: true, |
| format: crate::HoverDocFormat::Markdown, |
| }; |
| let tokens = tokens.filter(|token| { |
| matches!( |
| token.kind(), |
| IDENT | INT_NUMBER | LIFETIME_IDENT | T![self] | T![super] | T![crate] | T![Self] |
| ) |
| }); |
| let mut result = StaticIndexedFile { file_id, inlay_hints, folds, tokens: vec![] }; |
| for token in tokens { |
| let range = token.text_range(); |
| let node = token.parent().unwrap(); |
| let def = match get_definition(&sema, token.clone()) { |
| Some(it) => it, |
| None => continue, |
| }; |
| let id = if let Some(it) = self.def_map.get(&def) { |
| *it |
| } else { |
| let it = self.tokens.insert(TokenStaticData { |
| documentation: documentation_for_definition(&sema, def, &node), |
| hover: hover_for_definition(&sema, file_id, def, &node, &hover_config), |
| definition: def.try_to_nav(self.db).map(UpmappingResult::call_site).map(|it| { |
| FileRange { file_id: it.file_id, range: it.focus_or_full_range() } |
| }), |
| references: vec![], |
| moniker: current_crate.and_then(|cc| def_to_moniker(self.db, def, cc)), |
| display_name: def.name(self.db).map(|name| name.display(self.db).to_string()), |
| enclosing_moniker: current_crate |
| .zip(def.enclosing_definition(self.db)) |
| .and_then(|(cc, enclosing_def)| def_to_moniker(self.db, enclosing_def, cc)), |
| signature: def.label(self.db), |
| kind: def_to_kind(self.db, def), |
| }); |
| self.def_map.insert(def, it); |
| it |
| }; |
| let token = self.tokens.get_mut(id).unwrap(); |
| token.references.push(ReferenceData { |
| range: FileRange { range, file_id }, |
| is_definition: match def.try_to_nav(self.db).map(UpmappingResult::call_site) { |
| Some(it) => it.file_id == file_id && it.focus_or_full_range() == range, |
| None => false, |
| }, |
| }); |
| result.tokens.push((range, id)); |
| } |
| self.files.push(result); |
| } |
| |
| pub fn compute(analysis: &Analysis) -> StaticIndex<'_> { |
| let db = &*analysis.db; |
| let work = all_modules(db).into_iter().filter(|module| { |
| let file_id = module.definition_source_file_id(db).original_file(db); |
| let source_root = db.file_source_root(file_id); |
| let source_root = db.source_root(source_root); |
| !source_root.is_library |
| }); |
| let mut this = StaticIndex { |
| files: vec![], |
| tokens: Default::default(), |
| analysis, |
| db, |
| def_map: Default::default(), |
| }; |
| let mut visited_files = FxHashSet::default(); |
| for module in work { |
| let file_id = module.definition_source_file_id(db).original_file(db); |
| if visited_files.contains(&file_id) { |
| continue; |
| } |
| this.add_file(file_id); |
| // mark the file |
| visited_files.insert(file_id); |
| } |
| this |
| } |
| } |
| |
| #[cfg(test)] |
| mod tests { |
| use crate::{fixture, StaticIndex}; |
| use ide_db::{base_db::FileRange, FxHashSet}; |
| use syntax::TextSize; |
| |
| fn check_all_ranges(ra_fixture: &str) { |
| let (analysis, ranges) = fixture::annotations_without_marker(ra_fixture); |
| let s = StaticIndex::compute(&analysis); |
| let mut range_set: FxHashSet<_> = ranges.iter().map(|it| it.0).collect(); |
| for f in s.files { |
| for (range, _) in f.tokens { |
| let it = FileRange { file_id: f.file_id, range }; |
| if !range_set.contains(&it) { |
| panic!("additional range {it:?}"); |
| } |
| range_set.remove(&it); |
| } |
| } |
| if !range_set.is_empty() { |
| panic!("unfound ranges {range_set:?}"); |
| } |
| } |
| |
| #[track_caller] |
| fn check_definitions(ra_fixture: &str) { |
| let (analysis, ranges) = fixture::annotations_without_marker(ra_fixture); |
| let s = StaticIndex::compute(&analysis); |
| let mut range_set: FxHashSet<_> = ranges.iter().map(|it| it.0).collect(); |
| for (_, t) in s.tokens.iter() { |
| if let Some(t) = t.definition { |
| if t.range.start() == TextSize::from(0) { |
| // ignore definitions that are whole of file |
| continue; |
| } |
| if !range_set.contains(&t) { |
| panic!("additional definition {t:?}"); |
| } |
| range_set.remove(&t); |
| } |
| } |
| if !range_set.is_empty() { |
| panic!("unfound definitions {range_set:?}"); |
| } |
| } |
| |
| #[test] |
| fn struct_and_enum() { |
| check_all_ranges( |
| r#" |
| struct Foo; |
| //^^^ |
| enum E { X(Foo) } |
| //^ ^ ^^^ |
| "#, |
| ); |
| check_definitions( |
| r#" |
| struct Foo; |
| //^^^ |
| enum E { X(Foo) } |
| //^ ^ |
| "#, |
| ); |
| } |
| |
| #[test] |
| fn multi_crate() { |
| check_definitions( |
| r#" |
| //- /main.rs crate:main deps:foo |
| |
| |
| use foo::func; |
| |
| fn main() { |
| //^^^^ |
| func(); |
| } |
| //- /foo/lib.rs crate:foo |
| |
| pub func() { |
| |
| } |
| "#, |
| ); |
| } |
| |
| #[test] |
| fn derives() { |
| check_all_ranges( |
| r#" |
| //- minicore:derive |
| #[rustc_builtin_macro] |
| //^^^^^^^^^^^^^^^^^^^ |
| pub macro Copy {} |
| //^^^^ |
| #[derive(Copy)] |
| //^^^^^^ ^^^^ |
| struct Hello(i32); |
| //^^^^^ ^^^ |
| "#, |
| ); |
| } |
| } |