| //! LSIF (language server index format) generator |
| |
| use std::collections::HashMap; |
| use std::env; |
| use std::time::Instant; |
| |
| use ide::{ |
| Analysis, FileId, FileRange, MonikerKind, PackageInformation, RootDatabase, StaticIndex, |
| StaticIndexedFile, TokenId, TokenStaticData, |
| }; |
| use ide_db::{ |
| base_db::salsa::{self, ParallelDatabase}, |
| line_index::WideEncoding, |
| LineIndexDatabase, |
| }; |
| use load_cargo::{load_workspace, LoadCargoConfig, ProcMacroServerChoice}; |
| use lsp_types::{self, lsif}; |
| use project_model::{CargoConfig, ProjectManifest, ProjectWorkspace, RustLibSource}; |
| use vfs::{AbsPathBuf, Vfs}; |
| |
| use crate::{ |
| cli::flags, |
| line_index::{LineEndings, LineIndex, PositionEncoding}, |
| to_proto, |
| version::version, |
| }; |
| |
| /// Need to wrap Snapshot to provide `Clone` impl for `map_with` |
| struct Snap<DB>(DB); |
| impl<DB: ParallelDatabase> Clone for Snap<salsa::Snapshot<DB>> { |
| fn clone(&self) -> Snap<salsa::Snapshot<DB>> { |
| Snap(self.0.snapshot()) |
| } |
| } |
| |
| struct LsifManager<'a> { |
| count: i32, |
| token_map: HashMap<TokenId, Id>, |
| range_map: HashMap<FileRange, Id>, |
| file_map: HashMap<FileId, Id>, |
| package_map: HashMap<PackageInformation, Id>, |
| analysis: &'a Analysis, |
| db: &'a RootDatabase, |
| vfs: &'a Vfs, |
| } |
| |
| #[derive(Clone, Copy)] |
| struct Id(i32); |
| |
| impl From<Id> for lsp_types::NumberOrString { |
| fn from(Id(it): Id) -> Self { |
| lsp_types::NumberOrString::Number(it) |
| } |
| } |
| |
| impl LsifManager<'_> { |
| fn new<'a>(analysis: &'a Analysis, db: &'a RootDatabase, vfs: &'a Vfs) -> LsifManager<'a> { |
| LsifManager { |
| count: 0, |
| token_map: HashMap::default(), |
| range_map: HashMap::default(), |
| file_map: HashMap::default(), |
| package_map: HashMap::default(), |
| analysis, |
| db, |
| vfs, |
| } |
| } |
| |
| fn add(&mut self, data: lsif::Element) -> Id { |
| let id = Id(self.count); |
| self.emit(&serde_json::to_string(&lsif::Entry { id: id.into(), data }).unwrap()); |
| self.count += 1; |
| id |
| } |
| |
| fn add_vertex(&mut self, vertex: lsif::Vertex) -> Id { |
| self.add(lsif::Element::Vertex(vertex)) |
| } |
| |
| fn add_edge(&mut self, edge: lsif::Edge) -> Id { |
| self.add(lsif::Element::Edge(edge)) |
| } |
| |
| // FIXME: support file in addition to stdout here |
| fn emit(&self, data: &str) { |
| println!("{data}"); |
| } |
| |
| fn get_token_id(&mut self, id: TokenId) -> Id { |
| if let Some(it) = self.token_map.get(&id) { |
| return *it; |
| } |
| let result_set_id = self.add_vertex(lsif::Vertex::ResultSet(lsif::ResultSet { key: None })); |
| self.token_map.insert(id, result_set_id); |
| result_set_id |
| } |
| |
| fn get_package_id(&mut self, package_information: PackageInformation) -> Id { |
| if let Some(it) = self.package_map.get(&package_information) { |
| return *it; |
| } |
| let pi = package_information.clone(); |
| let result_set_id = |
| self.add_vertex(lsif::Vertex::PackageInformation(lsif::PackageInformation { |
| name: pi.name, |
| manager: "cargo".to_string(), |
| uri: None, |
| content: None, |
| repository: pi.repo.map(|url| lsif::Repository { |
| url, |
| r#type: "git".to_string(), |
| commit_id: None, |
| }), |
| version: pi.version, |
| })); |
| self.package_map.insert(package_information, result_set_id); |
| result_set_id |
| } |
| |
| fn get_range_id(&mut self, id: FileRange) -> Id { |
| if let Some(it) = self.range_map.get(&id) { |
| return *it; |
| } |
| let file_id = id.file_id; |
| let doc_id = self.get_file_id(file_id); |
| let line_index = self.db.line_index(file_id); |
| let line_index = LineIndex { |
| index: line_index, |
| encoding: PositionEncoding::Wide(WideEncoding::Utf16), |
| endings: LineEndings::Unix, |
| }; |
| let range_id = self.add_vertex(lsif::Vertex::Range { |
| range: to_proto::range(&line_index, id.range), |
| tag: None, |
| }); |
| self.add_edge(lsif::Edge::Contains(lsif::EdgeDataMultiIn { |
| in_vs: vec![range_id.into()], |
| out_v: doc_id.into(), |
| })); |
| range_id |
| } |
| |
| fn get_file_id(&mut self, id: FileId) -> Id { |
| if let Some(it) = self.file_map.get(&id) { |
| return *it; |
| } |
| let path = self.vfs.file_path(id); |
| let path = path.as_path().unwrap(); |
| let doc_id = self.add_vertex(lsif::Vertex::Document(lsif::Document { |
| language_id: "rust".to_string(), |
| uri: lsp_types::Url::from_file_path(path).unwrap(), |
| })); |
| self.file_map.insert(id, doc_id); |
| doc_id |
| } |
| |
| fn add_token(&mut self, id: TokenId, token: TokenStaticData) { |
| let result_set_id = self.get_token_id(id); |
| if let Some(hover) = token.hover { |
| let hover_id = self.add_vertex(lsif::Vertex::HoverResult { |
| result: lsp_types::Hover { |
| contents: lsp_types::HoverContents::Markup(to_proto::markup_content( |
| hover.markup, |
| ide::HoverDocFormat::Markdown, |
| )), |
| range: None, |
| }, |
| }); |
| self.add_edge(lsif::Edge::Hover(lsif::EdgeData { |
| in_v: hover_id.into(), |
| out_v: result_set_id.into(), |
| })); |
| } |
| if let Some(moniker) = token.moniker { |
| let package_id = self.get_package_id(moniker.package_information); |
| let moniker_id = self.add_vertex(lsif::Vertex::Moniker(lsp_types::Moniker { |
| scheme: "rust-analyzer".to_string(), |
| identifier: moniker.identifier.to_string(), |
| unique: lsp_types::UniquenessLevel::Scheme, |
| kind: Some(match moniker.kind { |
| MonikerKind::Import => lsp_types::MonikerKind::Import, |
| MonikerKind::Export => lsp_types::MonikerKind::Export, |
| }), |
| })); |
| self.add_edge(lsif::Edge::PackageInformation(lsif::EdgeData { |
| in_v: package_id.into(), |
| out_v: moniker_id.into(), |
| })); |
| self.add_edge(lsif::Edge::Moniker(lsif::EdgeData { |
| in_v: moniker_id.into(), |
| out_v: result_set_id.into(), |
| })); |
| } |
| if let Some(def) = token.definition { |
| let result_id = self.add_vertex(lsif::Vertex::DefinitionResult); |
| let def_vertex = self.get_range_id(def); |
| self.add_edge(lsif::Edge::Item(lsif::Item { |
| document: (*self.file_map.get(&def.file_id).unwrap()).into(), |
| property: None, |
| edge_data: lsif::EdgeDataMultiIn { |
| in_vs: vec![def_vertex.into()], |
| out_v: result_id.into(), |
| }, |
| })); |
| self.add_edge(lsif::Edge::Definition(lsif::EdgeData { |
| in_v: result_id.into(), |
| out_v: result_set_id.into(), |
| })); |
| } |
| if !token.references.is_empty() { |
| let result_id = self.add_vertex(lsif::Vertex::ReferenceResult); |
| self.add_edge(lsif::Edge::References(lsif::EdgeData { |
| in_v: result_id.into(), |
| out_v: result_set_id.into(), |
| })); |
| let mut edges = token.references.iter().fold( |
| HashMap::<_, Vec<lsp_types::NumberOrString>>::new(), |
| |mut edges, it| { |
| let entry = |
| edges.entry((it.range.file_id, it.is_definition)).or_insert_with(Vec::new); |
| entry.push((*self.range_map.get(&it.range).unwrap()).into()); |
| edges |
| }, |
| ); |
| for it in token.references { |
| if let Some(vertices) = edges.remove(&(it.range.file_id, it.is_definition)) { |
| self.add_edge(lsif::Edge::Item(lsif::Item { |
| document: (*self.file_map.get(&it.range.file_id).unwrap()).into(), |
| property: Some(if it.is_definition { |
| lsif::ItemKind::Definitions |
| } else { |
| lsif::ItemKind::References |
| }), |
| edge_data: lsif::EdgeDataMultiIn { |
| in_vs: vertices, |
| out_v: result_id.into(), |
| }, |
| })); |
| } |
| } |
| } |
| } |
| |
| fn add_file(&mut self, file: StaticIndexedFile) { |
| let StaticIndexedFile { file_id, tokens, folds, .. } = file; |
| let doc_id = self.get_file_id(file_id); |
| let text = self.analysis.file_text(file_id).unwrap(); |
| let line_index = self.db.line_index(file_id); |
| let line_index = LineIndex { |
| index: line_index, |
| encoding: PositionEncoding::Wide(WideEncoding::Utf16), |
| endings: LineEndings::Unix, |
| }; |
| let result = folds |
| .into_iter() |
| .map(|it| to_proto::folding_range(&text, &line_index, false, it)) |
| .collect(); |
| let folding_id = self.add_vertex(lsif::Vertex::FoldingRangeResult { result }); |
| self.add_edge(lsif::Edge::FoldingRange(lsif::EdgeData { |
| in_v: folding_id.into(), |
| out_v: doc_id.into(), |
| })); |
| let tokens_id = tokens |
| .into_iter() |
| .map(|(range, id)| { |
| let range_id = self.add_vertex(lsif::Vertex::Range { |
| range: to_proto::range(&line_index, range), |
| tag: None, |
| }); |
| self.range_map.insert(FileRange { file_id, range }, range_id); |
| let result_set_id = self.get_token_id(id); |
| self.add_edge(lsif::Edge::Next(lsif::EdgeData { |
| in_v: result_set_id.into(), |
| out_v: range_id.into(), |
| })); |
| range_id.into() |
| }) |
| .collect(); |
| self.add_edge(lsif::Edge::Contains(lsif::EdgeDataMultiIn { |
| in_vs: tokens_id, |
| out_v: doc_id.into(), |
| })); |
| } |
| } |
| |
| impl flags::Lsif { |
| pub fn run(self) -> anyhow::Result<()> { |
| eprintln!("Generating LSIF started..."); |
| let now = Instant::now(); |
| let mut cargo_config = CargoConfig::default(); |
| cargo_config.sysroot = Some(RustLibSource::Discover); |
| let no_progress = &|_| (); |
| let load_cargo_config = LoadCargoConfig { |
| load_out_dirs_from_check: true, |
| with_proc_macro_server: ProcMacroServerChoice::Sysroot, |
| prefill_caches: false, |
| }; |
| let path = AbsPathBuf::assert(env::current_dir()?.join(&self.path)); |
| let manifest = ProjectManifest::discover_single(&path)?; |
| |
| let workspace = ProjectWorkspace::load(manifest, &cargo_config, no_progress)?; |
| |
| let (host, vfs, _proc_macro) = |
| load_workspace(workspace, &cargo_config.extra_env, &load_cargo_config)?; |
| let db = host.raw_database(); |
| let analysis = host.analysis(); |
| |
| let si = StaticIndex::compute(&analysis); |
| |
| let mut lsif = LsifManager::new(&analysis, db, &vfs); |
| lsif.add_vertex(lsif::Vertex::MetaData(lsif::MetaData { |
| version: String::from("0.5.0"), |
| project_root: lsp_types::Url::from_file_path(path).unwrap(), |
| position_encoding: lsif::Encoding::Utf16, |
| tool_info: Some(lsp_types::lsif::ToolInfo { |
| name: "rust-analyzer".to_string(), |
| args: vec![], |
| version: Some(version().to_string()), |
| }), |
| })); |
| for file in si.files { |
| lsif.add_file(file); |
| } |
| for (id, token) in si.tokens.iter() { |
| lsif.add_token(id, token); |
| } |
| eprintln!("Generating LSIF finished in {:?}", now.elapsed()); |
| Ok(()) |
| } |
| } |