mirror of
https://github.com/helix-editor/helix.git
synced 2025-04-04 19:37:54 +03:00
Merge branch 'master' into tree-sitter-helix
This commit is contained in:
commit
588e8bee1a
34 changed files with 1251 additions and 712 deletions
29
Cargo.lock
generated
29
Cargo.lock
generated
|
@ -272,7 +272,7 @@ checksum = "5041cc499144891f3790297212f32a74fb938e5136a14943f338ef9e0ae276cf"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"cfg-if",
|
"cfg-if",
|
||||||
"crossbeam-utils",
|
"crossbeam-utils",
|
||||||
"hashbrown",
|
"hashbrown 0.14.5",
|
||||||
"lock_api",
|
"lock_api",
|
||||||
"once_cell",
|
"once_cell",
|
||||||
"parking_lot_core",
|
"parking_lot_core",
|
||||||
|
@ -794,7 +794,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "189130bc372accd02e0520dc5ab1cef318dcc2bc829b76ab8d84bbe90ac212d1"
|
checksum = "189130bc372accd02e0520dc5ab1cef318dcc2bc829b76ab8d84bbe90ac212d1"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"gix-hash",
|
"gix-hash",
|
||||||
"hashbrown",
|
"hashbrown 0.14.5",
|
||||||
"parking_lot",
|
"parking_lot",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
@ -830,7 +830,7 @@ dependencies = [
|
||||||
"gix-traverse",
|
"gix-traverse",
|
||||||
"gix-utils",
|
"gix-utils",
|
||||||
"gix-validate",
|
"gix-validate",
|
||||||
"hashbrown",
|
"hashbrown 0.14.5",
|
||||||
"itoa",
|
"itoa",
|
||||||
"libc",
|
"libc",
|
||||||
"memmap2",
|
"memmap2",
|
||||||
|
@ -1286,6 +1286,12 @@ dependencies = [
|
||||||
"allocator-api2",
|
"allocator-api2",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "hashbrown"
|
||||||
|
version = "0.15.2"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "bf151400ff0baff5465007dd2f3e717f3fe502074ca563069ce3a6629d07b289"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "helix-core"
|
name = "helix-core"
|
||||||
version = "25.1.1"
|
version = "25.1.1"
|
||||||
|
@ -1299,7 +1305,7 @@ dependencies = [
|
||||||
"encoding_rs",
|
"encoding_rs",
|
||||||
"etcetera",
|
"etcetera",
|
||||||
"globset",
|
"globset",
|
||||||
"hashbrown",
|
"hashbrown 0.14.5",
|
||||||
"helix-loader",
|
"helix-loader",
|
||||||
"helix-parsec",
|
"helix-parsec",
|
||||||
"helix-stdx",
|
"helix-stdx",
|
||||||
|
@ -1349,7 +1355,7 @@ dependencies = [
|
||||||
"ahash",
|
"ahash",
|
||||||
"anyhow",
|
"anyhow",
|
||||||
"futures-executor",
|
"futures-executor",
|
||||||
"hashbrown",
|
"hashbrown 0.14.5",
|
||||||
"log",
|
"log",
|
||||||
"once_cell",
|
"once_cell",
|
||||||
"parking_lot",
|
"parking_lot",
|
||||||
|
@ -1427,6 +1433,7 @@ dependencies = [
|
||||||
"ropey",
|
"ropey",
|
||||||
"rustix",
|
"rustix",
|
||||||
"tempfile",
|
"tempfile",
|
||||||
|
"unicode-segmentation",
|
||||||
"which",
|
"which",
|
||||||
"windows-sys 0.59.0",
|
"windows-sys 0.59.0",
|
||||||
]
|
]
|
||||||
|
@ -1744,17 +1751,17 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "fc9da1a252bd44cd341657203722352efc9bc0c847d06ea6d2dc1cd1135e0a01"
|
checksum = "fc9da1a252bd44cd341657203722352efc9bc0c847d06ea6d2dc1cd1135e0a01"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"ahash",
|
"ahash",
|
||||||
"hashbrown",
|
"hashbrown 0.14.5",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "indexmap"
|
name = "indexmap"
|
||||||
version = "2.5.0"
|
version = "2.7.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "68b900aa2f7301e21c36462b170ee99994de34dff39a4a6a528e80e7376d07e5"
|
checksum = "8c9c992b02b5b4c94ea26e32fe5bccb7aa7d9f390ab5c1221ff895bc7ea8b652"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"equivalent",
|
"equivalent",
|
||||||
"hashbrown",
|
"hashbrown 0.15.2",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
|
@ -2217,9 +2224,9 @@ checksum = "719b953e2095829ee67db738b3bfa9fa368c94900df327b3f07fe6e794d2fe1f"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "rustix"
|
name = "rustix"
|
||||||
version = "0.38.43"
|
version = "0.38.44"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "a78891ee6bf2340288408954ac787aa063d8e8817e9f53abb37c695c6d834ef6"
|
checksum = "fdb5bc1ae2baa591800df16c9ca78619bf65c0488b41b96ccec5d11220d8c154"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"bitflags",
|
"bitflags",
|
||||||
"errno",
|
"errno",
|
||||||
|
|
|
@ -43,6 +43,8 @@ slotmap = "1.0.7"
|
||||||
thiserror = "2.0"
|
thiserror = "2.0"
|
||||||
tempfile = "3.15.0"
|
tempfile = "3.15.0"
|
||||||
bitflags = "2.8"
|
bitflags = "2.8"
|
||||||
|
unicode-segmentation = "1.2"
|
||||||
|
ropey = { version = "1.6.1", default-features = false, features = ["simd"] }
|
||||||
|
|
||||||
[workspace.package]
|
[workspace.package]
|
||||||
version = "25.1.1"
|
version = "25.1.1"
|
||||||
|
|
|
@ -64,6 +64,7 @@
|
||||||
| gdscript | ✓ | ✓ | ✓ | |
|
| gdscript | ✓ | ✓ | ✓ | |
|
||||||
| gemini | ✓ | | | |
|
| gemini | ✓ | | | |
|
||||||
| gherkin | ✓ | | | |
|
| gherkin | ✓ | | | |
|
||||||
|
| ghostty | ✓ | | | |
|
||||||
| git-attributes | ✓ | | | |
|
| git-attributes | ✓ | | | |
|
||||||
| git-commit | ✓ | ✓ | | |
|
| git-commit | ✓ | ✓ | | |
|
||||||
| git-config | ✓ | ✓ | | |
|
| git-config | ✓ | ✓ | | |
|
||||||
|
|
|
@ -20,10 +20,10 @@ helix-stdx = { path = "../helix-stdx" }
|
||||||
helix-loader = { path = "../helix-loader" }
|
helix-loader = { path = "../helix-loader" }
|
||||||
helix-parsec = { path = "../helix-parsec" }
|
helix-parsec = { path = "../helix-parsec" }
|
||||||
|
|
||||||
ropey = { version = "1.6.1", default-features = false, features = ["simd"] }
|
ropey.workspace = true
|
||||||
smallvec = "1.13"
|
smallvec = "1.13"
|
||||||
smartstring = "1.0.1"
|
smartstring = "1.0.1"
|
||||||
unicode-segmentation = "1.12"
|
unicode-segmentation.workspace = true
|
||||||
# unicode-width is changing width definitions
|
# unicode-width is changing width definitions
|
||||||
# that both break our logic and disagree with common
|
# that both break our logic and disagree with common
|
||||||
# width definitions in terminals, we need to replace it.
|
# width definitions in terminals, we need to replace it.
|
||||||
|
|
|
@ -19,10 +19,12 @@ mod test;
|
||||||
|
|
||||||
use unicode_segmentation::{Graphemes, UnicodeSegmentation};
|
use unicode_segmentation::{Graphemes, UnicodeSegmentation};
|
||||||
|
|
||||||
|
use helix_stdx::rope::{RopeGraphemes, RopeSliceExt};
|
||||||
|
|
||||||
use crate::graphemes::{Grapheme, GraphemeStr};
|
use crate::graphemes::{Grapheme, GraphemeStr};
|
||||||
use crate::syntax::Highlight;
|
use crate::syntax::Highlight;
|
||||||
use crate::text_annotations::TextAnnotations;
|
use crate::text_annotations::TextAnnotations;
|
||||||
use crate::{Position, RopeGraphemes, RopeSlice};
|
use crate::{Position, RopeSlice};
|
||||||
|
|
||||||
/// TODO make Highlight a u32 to reduce the size of this enum to a single word.
|
/// TODO make Highlight a u32 to reduce the size of this enum to a single word.
|
||||||
#[derive(Debug, Clone, Copy)]
|
#[derive(Debug, Clone, Copy)]
|
||||||
|
@ -219,7 +221,7 @@ impl<'t> DocumentFormatter<'t> {
|
||||||
text_fmt,
|
text_fmt,
|
||||||
annotations,
|
annotations,
|
||||||
visual_pos: Position { row: 0, col: 0 },
|
visual_pos: Position { row: 0, col: 0 },
|
||||||
graphemes: RopeGraphemes::new(text.slice(block_char_idx..)),
|
graphemes: text.slice(block_char_idx..).graphemes(),
|
||||||
char_pos: block_char_idx,
|
char_pos: block_char_idx,
|
||||||
exhausted: false,
|
exhausted: false,
|
||||||
indent_level: None,
|
indent_level: None,
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
//! Utility functions to traverse the unicode graphemes of a `Rope`'s text contents.
|
//! Utility functions to traverse the unicode graphemes of a `Rope`'s text contents.
|
||||||
//!
|
//!
|
||||||
//! Based on <https://github.com/cessen/led/blob/c4fa72405f510b7fd16052f90a598c429b3104a6/src/graphemes.rs>
|
//! Based on <https://github.com/cessen/led/blob/c4fa72405f510b7fd16052f90a598c429b3104a6/src/graphemes.rs>
|
||||||
use ropey::{iter::Chunks, str_utils::byte_to_char_idx, RopeSlice};
|
use ropey::{str_utils::byte_to_char_idx, RopeSlice};
|
||||||
use unicode_segmentation::{GraphemeCursor, GraphemeIncomplete};
|
use unicode_segmentation::{GraphemeCursor, GraphemeIncomplete};
|
||||||
use unicode_width::UnicodeWidthStr;
|
use unicode_width::UnicodeWidthStr;
|
||||||
|
|
||||||
|
@ -119,6 +119,9 @@ pub fn grapheme_width(g: &str) -> usize {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// NOTE: for byte indexing versions of these functions see `RopeSliceExt`'s
|
||||||
|
// `floor_grapheme_boundary` and `ceil_grapheme_boundary` and the rope grapheme iterators.
|
||||||
|
|
||||||
#[must_use]
|
#[must_use]
|
||||||
pub fn nth_prev_grapheme_boundary(slice: RopeSlice, char_idx: usize, n: usize) -> usize {
|
pub fn nth_prev_grapheme_boundary(slice: RopeSlice, char_idx: usize, n: usize) -> usize {
|
||||||
// Bounds check
|
// Bounds check
|
||||||
|
@ -208,43 +211,6 @@ pub fn nth_next_grapheme_boundary(slice: RopeSlice, char_idx: usize, n: usize) -
|
||||||
chunk_char_idx + tmp
|
chunk_char_idx + tmp
|
||||||
}
|
}
|
||||||
|
|
||||||
#[must_use]
|
|
||||||
pub fn nth_next_grapheme_boundary_byte(slice: RopeSlice, mut byte_idx: usize, n: usize) -> usize {
|
|
||||||
// Bounds check
|
|
||||||
debug_assert!(byte_idx <= slice.len_bytes());
|
|
||||||
|
|
||||||
// Get the chunk with our byte index in it.
|
|
||||||
let (mut chunk, mut chunk_byte_idx, mut _chunk_char_idx, _) = slice.chunk_at_byte(byte_idx);
|
|
||||||
|
|
||||||
// Set up the grapheme cursor.
|
|
||||||
let mut gc = GraphemeCursor::new(byte_idx, slice.len_bytes(), true);
|
|
||||||
|
|
||||||
// Find the nth next grapheme cluster boundary.
|
|
||||||
for _ in 0..n {
|
|
||||||
loop {
|
|
||||||
match gc.next_boundary(chunk, chunk_byte_idx) {
|
|
||||||
Ok(None) => return slice.len_bytes(),
|
|
||||||
Ok(Some(n)) => {
|
|
||||||
byte_idx = n;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
Err(GraphemeIncomplete::NextChunk) => {
|
|
||||||
chunk_byte_idx += chunk.len();
|
|
||||||
let (a, _, _c, _) = slice.chunk_at_byte(chunk_byte_idx);
|
|
||||||
chunk = a;
|
|
||||||
// chunk_char_idx = c;
|
|
||||||
}
|
|
||||||
Err(GraphemeIncomplete::PreContext(n)) => {
|
|
||||||
let ctx_chunk = slice.chunk_at_byte(n - 1).0;
|
|
||||||
gc.provide_context(ctx_chunk, n - ctx_chunk.len());
|
|
||||||
}
|
|
||||||
_ => unreachable!(),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
byte_idx
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Finds the next grapheme boundary after the given char position.
|
/// Finds the next grapheme boundary after the given char position.
|
||||||
#[must_use]
|
#[must_use]
|
||||||
#[inline(always)]
|
#[inline(always)]
|
||||||
|
@ -252,13 +218,6 @@ pub fn next_grapheme_boundary(slice: RopeSlice, char_idx: usize) -> usize {
|
||||||
nth_next_grapheme_boundary(slice, char_idx, 1)
|
nth_next_grapheme_boundary(slice, char_idx, 1)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Finds the next grapheme boundary after the given byte position.
|
|
||||||
#[must_use]
|
|
||||||
#[inline(always)]
|
|
||||||
pub fn next_grapheme_boundary_byte(slice: RopeSlice, byte_idx: usize) -> usize {
|
|
||||||
nth_next_grapheme_boundary_byte(slice, byte_idx, 1)
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Returns the passed char index if it's already a grapheme boundary,
|
/// Returns the passed char index if it's already a grapheme boundary,
|
||||||
/// or the next grapheme boundary char index if not.
|
/// or the next grapheme boundary char index if not.
|
||||||
#[must_use]
|
#[must_use]
|
||||||
|
@ -311,187 +270,6 @@ pub fn is_grapheme_boundary(slice: RopeSlice, char_idx: usize) -> bool {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns whether the given byte position is a grapheme boundary.
|
|
||||||
#[must_use]
|
|
||||||
pub fn is_grapheme_boundary_byte(slice: RopeSlice, byte_idx: usize) -> bool {
|
|
||||||
// Bounds check
|
|
||||||
debug_assert!(byte_idx <= slice.len_bytes());
|
|
||||||
|
|
||||||
// Get the chunk with our byte index in it.
|
|
||||||
let (chunk, chunk_byte_idx, _, _) = slice.chunk_at_byte(byte_idx);
|
|
||||||
|
|
||||||
// Set up the grapheme cursor.
|
|
||||||
let mut gc = GraphemeCursor::new(byte_idx, slice.len_bytes(), true);
|
|
||||||
|
|
||||||
// Determine if the given position is a grapheme cluster boundary.
|
|
||||||
loop {
|
|
||||||
match gc.is_boundary(chunk, chunk_byte_idx) {
|
|
||||||
Ok(n) => return n,
|
|
||||||
Err(GraphemeIncomplete::PreContext(n)) => {
|
|
||||||
let (ctx_chunk, ctx_byte_start, _, _) = slice.chunk_at_byte(n - 1);
|
|
||||||
gc.provide_context(ctx_chunk, ctx_byte_start);
|
|
||||||
}
|
|
||||||
Err(_) => unreachable!(),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// An iterator over the graphemes of a `RopeSlice`.
|
|
||||||
#[derive(Clone)]
|
|
||||||
pub struct RopeGraphemes<'a> {
|
|
||||||
text: RopeSlice<'a>,
|
|
||||||
chunks: Chunks<'a>,
|
|
||||||
cur_chunk: &'a str,
|
|
||||||
cur_chunk_start: usize,
|
|
||||||
cursor: GraphemeCursor,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl fmt::Debug for RopeGraphemes<'_> {
|
|
||||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
|
||||||
f.debug_struct("RopeGraphemes")
|
|
||||||
.field("text", &self.text)
|
|
||||||
.field("chunks", &self.chunks)
|
|
||||||
.field("cur_chunk", &self.cur_chunk)
|
|
||||||
.field("cur_chunk_start", &self.cur_chunk_start)
|
|
||||||
// .field("cursor", &self.cursor)
|
|
||||||
.finish()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl RopeGraphemes<'_> {
|
|
||||||
#[must_use]
|
|
||||||
pub fn new(slice: RopeSlice) -> RopeGraphemes {
|
|
||||||
let mut chunks = slice.chunks();
|
|
||||||
let first_chunk = chunks.next().unwrap_or("");
|
|
||||||
RopeGraphemes {
|
|
||||||
text: slice,
|
|
||||||
chunks,
|
|
||||||
cur_chunk: first_chunk,
|
|
||||||
cur_chunk_start: 0,
|
|
||||||
cursor: GraphemeCursor::new(0, slice.len_bytes(), true),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<'a> Iterator for RopeGraphemes<'a> {
|
|
||||||
type Item = RopeSlice<'a>;
|
|
||||||
|
|
||||||
fn next(&mut self) -> Option<RopeSlice<'a>> {
|
|
||||||
let a = self.cursor.cur_cursor();
|
|
||||||
let b;
|
|
||||||
loop {
|
|
||||||
match self
|
|
||||||
.cursor
|
|
||||||
.next_boundary(self.cur_chunk, self.cur_chunk_start)
|
|
||||||
{
|
|
||||||
Ok(None) => {
|
|
||||||
return None;
|
|
||||||
}
|
|
||||||
Ok(Some(n)) => {
|
|
||||||
b = n;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
Err(GraphemeIncomplete::NextChunk) => {
|
|
||||||
self.cur_chunk_start += self.cur_chunk.len();
|
|
||||||
self.cur_chunk = self.chunks.next().unwrap_or("");
|
|
||||||
}
|
|
||||||
Err(GraphemeIncomplete::PreContext(idx)) => {
|
|
||||||
let (chunk, byte_idx, _, _) = self.text.chunk_at_byte(idx.saturating_sub(1));
|
|
||||||
self.cursor.provide_context(chunk, byte_idx);
|
|
||||||
}
|
|
||||||
_ => unreachable!(),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if a < self.cur_chunk_start {
|
|
||||||
Some(self.text.byte_slice(a..b))
|
|
||||||
} else {
|
|
||||||
let a2 = a - self.cur_chunk_start;
|
|
||||||
let b2 = b - self.cur_chunk_start;
|
|
||||||
Some((&self.cur_chunk[a2..b2]).into())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// An iterator over the graphemes of a `RopeSlice` in reverse.
|
|
||||||
#[derive(Clone)]
|
|
||||||
pub struct RevRopeGraphemes<'a> {
|
|
||||||
text: RopeSlice<'a>,
|
|
||||||
chunks: Chunks<'a>,
|
|
||||||
cur_chunk: &'a str,
|
|
||||||
cur_chunk_start: usize,
|
|
||||||
cursor: GraphemeCursor,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl fmt::Debug for RevRopeGraphemes<'_> {
|
|
||||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
|
||||||
f.debug_struct("RevRopeGraphemes")
|
|
||||||
.field("text", &self.text)
|
|
||||||
.field("chunks", &self.chunks)
|
|
||||||
.field("cur_chunk", &self.cur_chunk)
|
|
||||||
.field("cur_chunk_start", &self.cur_chunk_start)
|
|
||||||
// .field("cursor", &self.cursor)
|
|
||||||
.finish()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl RevRopeGraphemes<'_> {
|
|
||||||
#[must_use]
|
|
||||||
pub fn new(slice: RopeSlice) -> RevRopeGraphemes {
|
|
||||||
let (mut chunks, mut cur_chunk_start, _, _) = slice.chunks_at_byte(slice.len_bytes());
|
|
||||||
chunks.reverse();
|
|
||||||
let first_chunk = chunks.next().unwrap_or("");
|
|
||||||
cur_chunk_start -= first_chunk.len();
|
|
||||||
RevRopeGraphemes {
|
|
||||||
text: slice,
|
|
||||||
chunks,
|
|
||||||
cur_chunk: first_chunk,
|
|
||||||
cur_chunk_start,
|
|
||||||
cursor: GraphemeCursor::new(slice.len_bytes(), slice.len_bytes(), true),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<'a> Iterator for RevRopeGraphemes<'a> {
|
|
||||||
type Item = RopeSlice<'a>;
|
|
||||||
|
|
||||||
fn next(&mut self) -> Option<RopeSlice<'a>> {
|
|
||||||
let a = self.cursor.cur_cursor();
|
|
||||||
let b;
|
|
||||||
loop {
|
|
||||||
match self
|
|
||||||
.cursor
|
|
||||||
.prev_boundary(self.cur_chunk, self.cur_chunk_start)
|
|
||||||
{
|
|
||||||
Ok(None) => {
|
|
||||||
return None;
|
|
||||||
}
|
|
||||||
Ok(Some(n)) => {
|
|
||||||
b = n;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
Err(GraphemeIncomplete::PrevChunk) => {
|
|
||||||
self.cur_chunk = self.chunks.next().unwrap_or("");
|
|
||||||
self.cur_chunk_start -= self.cur_chunk.len();
|
|
||||||
}
|
|
||||||
Err(GraphemeIncomplete::PreContext(idx)) => {
|
|
||||||
let (chunk, byte_idx, _, _) = self.text.chunk_at_byte(idx.saturating_sub(1));
|
|
||||||
self.cursor.provide_context(chunk, byte_idx);
|
|
||||||
}
|
|
||||||
_ => unreachable!(),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if a >= self.cur_chunk_start + self.cur_chunk.len() {
|
|
||||||
Some(self.text.byte_slice(b..a))
|
|
||||||
} else {
|
|
||||||
let a2 = a - self.cur_chunk_start;
|
|
||||||
let b2 = b - self.cur_chunk_start;
|
|
||||||
Some((&self.cur_chunk[b2..a2]).into())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// A highly compressed Cow<'a, str> that holds
|
/// A highly compressed Cow<'a, str> that holds
|
||||||
/// atmost u31::MAX bytes and is readonly
|
/// atmost u31::MAX bytes and is readonly
|
||||||
pub struct GraphemeStr<'a> {
|
pub struct GraphemeStr<'a> {
|
||||||
|
|
|
@ -8,7 +8,7 @@ use crate::{
|
||||||
graphemes::{grapheme_width, tab_width_at},
|
graphemes::{grapheme_width, tab_width_at},
|
||||||
syntax::{IndentationHeuristic, LanguageConfiguration, RopeProvider, Syntax},
|
syntax::{IndentationHeuristic, LanguageConfiguration, RopeProvider, Syntax},
|
||||||
tree_sitter::Node,
|
tree_sitter::Node,
|
||||||
Position, Rope, RopeGraphemes, RopeSlice, Tendril,
|
Position, Rope, RopeSlice, Tendril,
|
||||||
};
|
};
|
||||||
|
|
||||||
/// Enum representing indentation style.
|
/// Enum representing indentation style.
|
||||||
|
@ -200,7 +200,7 @@ pub fn indent_level_for_line(line: RopeSlice, tab_width: usize, indent_width: us
|
||||||
/// Create a string of tabs & spaces that has the same visual width as the given RopeSlice (independent of the tab width).
|
/// Create a string of tabs & spaces that has the same visual width as the given RopeSlice (independent of the tab width).
|
||||||
fn whitespace_with_same_width(text: RopeSlice) -> String {
|
fn whitespace_with_same_width(text: RopeSlice) -> String {
|
||||||
let mut s = String::new();
|
let mut s = String::new();
|
||||||
for grapheme in RopeGraphemes::new(text) {
|
for grapheme in text.graphemes() {
|
||||||
if grapheme == "\t" {
|
if grapheme == "\t" {
|
||||||
s.push('\t');
|
s.push('\t');
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -54,7 +54,6 @@ pub type Tendril = SmartString<smartstring::LazyCompact>;
|
||||||
#[doc(inline)]
|
#[doc(inline)]
|
||||||
pub use {regex, tree_sitter};
|
pub use {regex, tree_sitter};
|
||||||
|
|
||||||
pub use graphemes::RopeGraphemes;
|
|
||||||
pub use position::{
|
pub use position::{
|
||||||
char_idx_at_visual_offset, coords_at_pos, pos_at_coords, softwrapped_dimensions,
|
char_idx_at_visual_offset, coords_at_pos, pos_at_coords, softwrapped_dimensions,
|
||||||
visual_offset_from_anchor, visual_offset_from_block, Position, VisualOffsetError,
|
visual_offset_from_anchor, visual_offset_from_block, Position, VisualOffsetError,
|
||||||
|
|
|
@ -4,10 +4,12 @@ use std::{
|
||||||
ops::{Add, AddAssign, Sub, SubAssign},
|
ops::{Add, AddAssign, Sub, SubAssign},
|
||||||
};
|
};
|
||||||
|
|
||||||
|
use helix_stdx::rope::RopeSliceExt;
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
chars::char_is_line_ending,
|
chars::char_is_line_ending,
|
||||||
doc_formatter::{DocumentFormatter, TextFormat},
|
doc_formatter::{DocumentFormatter, TextFormat},
|
||||||
graphemes::{ensure_grapheme_boundary_prev, grapheme_width, RopeGraphemes},
|
graphemes::{ensure_grapheme_boundary_prev, grapheme_width},
|
||||||
line_ending::line_end_char_index,
|
line_ending::line_end_char_index,
|
||||||
text_annotations::TextAnnotations,
|
text_annotations::TextAnnotations,
|
||||||
RopeSlice,
|
RopeSlice,
|
||||||
|
@ -101,7 +103,7 @@ pub fn coords_at_pos(text: RopeSlice, pos: usize) -> Position {
|
||||||
|
|
||||||
let line_start = text.line_to_char(line);
|
let line_start = text.line_to_char(line);
|
||||||
let pos = ensure_grapheme_boundary_prev(text, pos);
|
let pos = ensure_grapheme_boundary_prev(text, pos);
|
||||||
let col = RopeGraphemes::new(text.slice(line_start..pos)).count();
|
let col = text.slice(line_start..pos).graphemes().count();
|
||||||
|
|
||||||
Position::new(line, col)
|
Position::new(line, col)
|
||||||
}
|
}
|
||||||
|
@ -126,7 +128,7 @@ pub fn visual_coords_at_pos(text: RopeSlice, pos: usize, tab_width: usize) -> Po
|
||||||
|
|
||||||
let mut col = 0;
|
let mut col = 0;
|
||||||
|
|
||||||
for grapheme in RopeGraphemes::new(text.slice(line_start..pos)) {
|
for grapheme in text.slice(line_start..pos).graphemes() {
|
||||||
if grapheme == "\t" {
|
if grapheme == "\t" {
|
||||||
col += tab_width - (col % tab_width);
|
col += tab_width - (col % tab_width);
|
||||||
} else {
|
} else {
|
||||||
|
@ -275,7 +277,7 @@ pub fn pos_at_coords(text: RopeSlice, coords: Position, limit_before_line_ending
|
||||||
};
|
};
|
||||||
|
|
||||||
let mut col_char_offset = 0;
|
let mut col_char_offset = 0;
|
||||||
for (i, g) in RopeGraphemes::new(text.slice(line_start..line_end)).enumerate() {
|
for (i, g) in text.slice(line_start..line_end).graphemes().enumerate() {
|
||||||
if i == col {
|
if i == col {
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
@ -306,7 +308,7 @@ pub fn pos_at_visual_coords(text: RopeSlice, coords: Position, tab_width: usize)
|
||||||
|
|
||||||
let mut col_char_offset = 0;
|
let mut col_char_offset = 0;
|
||||||
let mut cols_remaining = col;
|
let mut cols_remaining = col;
|
||||||
for grapheme in RopeGraphemes::new(text.slice(line_start..line_end)) {
|
for grapheme in text.slice(line_start..line_end).graphemes() {
|
||||||
let grapheme_width = if grapheme == "\t" {
|
let grapheme_width = if grapheme == "\t" {
|
||||||
tab_width - ((col - cols_remaining) % tab_width)
|
tab_width - ((col - cols_remaining) % tab_width)
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -9,7 +9,7 @@ use crate::{
|
||||||
},
|
},
|
||||||
line_ending::get_line_ending,
|
line_ending::get_line_ending,
|
||||||
movement::Direction,
|
movement::Direction,
|
||||||
Assoc, ChangeSet, RopeGraphemes, RopeSlice,
|
Assoc, ChangeSet, RopeSlice,
|
||||||
};
|
};
|
||||||
use helix_stdx::range::is_subset;
|
use helix_stdx::range::is_subset;
|
||||||
use helix_stdx::rope::{self, RopeSliceExt};
|
use helix_stdx::rope::{self, RopeSliceExt};
|
||||||
|
@ -379,7 +379,7 @@ impl Range {
|
||||||
|
|
||||||
/// Returns true if this Range covers a single grapheme in the given text
|
/// Returns true if this Range covers a single grapheme in the given text
|
||||||
pub fn is_single_grapheme(&self, doc: RopeSlice) -> bool {
|
pub fn is_single_grapheme(&self, doc: RopeSlice) -> bool {
|
||||||
let mut graphemes = RopeGraphemes::new(doc.slice(self.from()..self.to()));
|
let mut graphemes = doc.slice(self.from()..self.to()).graphemes();
|
||||||
let first = graphemes.next();
|
let first = graphemes.next();
|
||||||
let second = graphemes.next();
|
let second = graphemes.next();
|
||||||
first.is_some() && second.is_none()
|
first.is_some() && second.is_none()
|
||||||
|
|
|
@ -3,10 +3,11 @@ mod transport;
|
||||||
mod types;
|
mod types;
|
||||||
|
|
||||||
pub use client::{Client, ConnectionType};
|
pub use client::{Client, ConnectionType};
|
||||||
pub use events::Event;
|
|
||||||
pub use transport::{Payload, Response, Transport};
|
pub use transport::{Payload, Response, Transport};
|
||||||
pub use types::*;
|
pub use types::*;
|
||||||
|
|
||||||
|
use serde::de::DeserializeOwned;
|
||||||
|
|
||||||
use thiserror::Error;
|
use thiserror::Error;
|
||||||
#[derive(Error, Debug)]
|
#[derive(Error, Debug)]
|
||||||
pub enum Error {
|
pub enum Error {
|
||||||
|
@ -18,9 +19,84 @@ pub enum Error {
|
||||||
Timeout(u64),
|
Timeout(u64),
|
||||||
#[error("server closed the stream")]
|
#[error("server closed the stream")]
|
||||||
StreamClosed,
|
StreamClosed,
|
||||||
|
#[error("Unhandled")]
|
||||||
|
Unhandled,
|
||||||
#[error(transparent)]
|
#[error(transparent)]
|
||||||
ExecutableNotFound(#[from] helix_stdx::env::ExecutableNotFoundError),
|
ExecutableNotFound(#[from] helix_stdx::env::ExecutableNotFoundError),
|
||||||
#[error(transparent)]
|
#[error(transparent)]
|
||||||
Other(#[from] anyhow::Error),
|
Other(#[from] anyhow::Error),
|
||||||
}
|
}
|
||||||
pub type Result<T> = core::result::Result<T, Error>;
|
pub type Result<T> = core::result::Result<T, Error>;
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Request {
|
||||||
|
RunInTerminal(<requests::RunInTerminal as types::Request>::Arguments),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Request {
|
||||||
|
pub fn parse(command: &str, arguments: Option<serde_json::Value>) -> Result<Self> {
|
||||||
|
use crate::types::Request as _;
|
||||||
|
|
||||||
|
let arguments = arguments.unwrap_or_default();
|
||||||
|
let request = match command {
|
||||||
|
requests::RunInTerminal::COMMAND => Self::RunInTerminal(parse_value(arguments)?),
|
||||||
|
_ => return Err(Error::Unhandled),
|
||||||
|
};
|
||||||
|
|
||||||
|
Ok(request)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Event {
|
||||||
|
Initialized(<events::Initialized as events::Event>::Body),
|
||||||
|
Stopped(<events::Stopped as events::Event>::Body),
|
||||||
|
Continued(<events::Continued as events::Event>::Body),
|
||||||
|
Exited(<events::Exited as events::Event>::Body),
|
||||||
|
Terminated(<events::Terminated as events::Event>::Body),
|
||||||
|
Thread(<events::Thread as events::Event>::Body),
|
||||||
|
Output(<events::Output as events::Event>::Body),
|
||||||
|
Breakpoint(<events::Breakpoint as events::Event>::Body),
|
||||||
|
Module(<events::Module as events::Event>::Body),
|
||||||
|
LoadedSource(<events::LoadedSource as events::Event>::Body),
|
||||||
|
Process(<events::Process as events::Event>::Body),
|
||||||
|
Capabilities(<events::Capabilities as events::Event>::Body),
|
||||||
|
// ProgressStart(),
|
||||||
|
// ProgressUpdate(),
|
||||||
|
// ProgressEnd(),
|
||||||
|
// Invalidated(),
|
||||||
|
Memory(<events::Memory as events::Event>::Body),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Event {
|
||||||
|
pub fn parse(event: &str, body: Option<serde_json::Value>) -> Result<Self> {
|
||||||
|
use crate::events::Event as _;
|
||||||
|
|
||||||
|
let body = body.unwrap_or_default();
|
||||||
|
let event = match event {
|
||||||
|
events::Initialized::EVENT => Self::Initialized(parse_value(body)?),
|
||||||
|
events::Stopped::EVENT => Self::Stopped(parse_value(body)?),
|
||||||
|
events::Continued::EVENT => Self::Continued(parse_value(body)?),
|
||||||
|
events::Exited::EVENT => Self::Exited(parse_value(body)?),
|
||||||
|
events::Terminated::EVENT => Self::Terminated(parse_value(body)?),
|
||||||
|
events::Thread::EVENT => Self::Thread(parse_value(body)?),
|
||||||
|
events::Output::EVENT => Self::Output(parse_value(body)?),
|
||||||
|
events::Breakpoint::EVENT => Self::Breakpoint(parse_value(body)?),
|
||||||
|
events::Module::EVENT => Self::Module(parse_value(body)?),
|
||||||
|
events::LoadedSource::EVENT => Self::LoadedSource(parse_value(body)?),
|
||||||
|
events::Process::EVENT => Self::Process(parse_value(body)?),
|
||||||
|
events::Capabilities::EVENT => Self::Capabilities(parse_value(body)?),
|
||||||
|
events::Memory::EVENT => Self::Memory(parse_value(body)?),
|
||||||
|
_ => return Err(Error::Unhandled),
|
||||||
|
};
|
||||||
|
|
||||||
|
Ok(event)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn parse_value<T>(value: serde_json::Value) -> Result<T>
|
||||||
|
where
|
||||||
|
T: DeserializeOwned,
|
||||||
|
{
|
||||||
|
serde_json::from_value(value).map_err(|err| err.into())
|
||||||
|
}
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
use crate::{Error, Event, Result};
|
use crate::{Error, Result};
|
||||||
use anyhow::Context;
|
use anyhow::Context;
|
||||||
use log::{error, info, warn};
|
use log::{error, info, warn};
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
|
@ -32,11 +32,17 @@ pub struct Response {
|
||||||
pub body: Option<Value>,
|
pub body: Option<Value>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
|
pub struct Event {
|
||||||
|
pub event: String,
|
||||||
|
pub body: Option<Value>,
|
||||||
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, Deserialize, Serialize)]
|
#[derive(Debug, Clone, Deserialize, Serialize)]
|
||||||
#[serde(tag = "type", rename_all = "camelCase")]
|
#[serde(tag = "type", rename_all = "camelCase")]
|
||||||
pub enum Payload {
|
pub enum Payload {
|
||||||
// type = "event"
|
// type = "event"
|
||||||
Event(Box<Event>),
|
Event(Event),
|
||||||
// type = "response"
|
// type = "response"
|
||||||
Response(Response),
|
Response(Response),
|
||||||
// type = "request"
|
// type = "request"
|
||||||
|
|
|
@ -759,33 +759,30 @@ pub mod requests {
|
||||||
pub mod events {
|
pub mod events {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|
||||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
pub trait Event {
|
||||||
#[serde(rename_all = "camelCase")]
|
type Body: serde::de::DeserializeOwned + serde::Serialize;
|
||||||
#[serde(tag = "event", content = "body")]
|
const EVENT: &'static str;
|
||||||
// seq is omitted as unused and is not sent by some implementations
|
}
|
||||||
pub enum Event {
|
|
||||||
Initialized(Option<DebuggerCapabilities>),
|
#[derive(Debug)]
|
||||||
Stopped(Stopped),
|
pub enum Initialized {}
|
||||||
Continued(Continued),
|
|
||||||
Exited(Exited),
|
impl Event for Initialized {
|
||||||
Terminated(Option<Terminated>),
|
type Body = Option<DebuggerCapabilities>;
|
||||||
Thread(Thread),
|
const EVENT: &'static str = "initialized";
|
||||||
Output(Output),
|
}
|
||||||
Breakpoint(Breakpoint),
|
|
||||||
Module(Module),
|
#[derive(Debug)]
|
||||||
LoadedSource(LoadedSource),
|
pub enum Stopped {}
|
||||||
Process(Process),
|
|
||||||
Capabilities(Capabilities),
|
impl Event for Stopped {
|
||||||
// ProgressStart(),
|
type Body = StoppedBody;
|
||||||
// ProgressUpdate(),
|
const EVENT: &'static str = "stopped";
|
||||||
// ProgressEnd(),
|
|
||||||
// Invalidated(),
|
|
||||||
Memory(Memory),
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct Stopped {
|
pub struct StoppedBody {
|
||||||
pub reason: String,
|
pub reason: String,
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub description: Option<String>,
|
pub description: Option<String>,
|
||||||
|
@ -801,37 +798,77 @@ pub mod events {
|
||||||
pub hit_breakpoint_ids: Option<Vec<usize>>,
|
pub hit_breakpoint_ids: Option<Vec<usize>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Continued {}
|
||||||
|
|
||||||
|
impl Event for Continued {
|
||||||
|
type Body = ContinuedBody;
|
||||||
|
const EVENT: &'static str = "continued";
|
||||||
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct Continued {
|
pub struct ContinuedBody {
|
||||||
pub thread_id: ThreadId,
|
pub thread_id: ThreadId,
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub all_threads_continued: Option<bool>,
|
pub all_threads_continued: Option<bool>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug)]
|
||||||
#[serde(rename_all = "camelCase")]
|
pub enum Exited {}
|
||||||
pub struct Exited {
|
|
||||||
pub exit_code: usize,
|
impl Event for Exited {
|
||||||
|
type Body = ExitedBody;
|
||||||
|
const EVENT: &'static str = "exited";
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct Terminated {
|
pub struct ExitedBody {
|
||||||
|
pub exit_code: usize,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Terminated {}
|
||||||
|
|
||||||
|
impl Event for Terminated {
|
||||||
|
type Body = Option<TerminatedBody>;
|
||||||
|
const EVENT: &'static str = "terminated";
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct TerminatedBody {
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub restart: Option<Value>,
|
pub restart: Option<Value>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug)]
|
||||||
#[serde(rename_all = "camelCase")]
|
pub enum Thread {}
|
||||||
pub struct Thread {
|
|
||||||
pub reason: String,
|
impl Event for Thread {
|
||||||
pub thread_id: ThreadId,
|
type Body = ThreadBody;
|
||||||
|
const EVENT: &'static str = "thread";
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct Output {
|
pub struct ThreadBody {
|
||||||
|
pub reason: String,
|
||||||
|
pub thread_id: ThreadId,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Output {}
|
||||||
|
|
||||||
|
impl Event for Output {
|
||||||
|
type Body = OutputBody;
|
||||||
|
const EVENT: &'static str = "output";
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct OutputBody {
|
||||||
pub output: String,
|
pub output: String,
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub category: Option<String>,
|
pub category: Option<String>,
|
||||||
|
@ -849,30 +886,62 @@ pub mod events {
|
||||||
pub data: Option<Value>,
|
pub data: Option<Value>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Breakpoint {}
|
||||||
|
|
||||||
|
impl Event for Breakpoint {
|
||||||
|
type Body = BreakpointBody;
|
||||||
|
const EVENT: &'static str = "breakpoint";
|
||||||
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct Breakpoint {
|
pub struct BreakpointBody {
|
||||||
pub reason: String,
|
pub reason: String,
|
||||||
pub breakpoint: super::Breakpoint,
|
pub breakpoint: super::Breakpoint,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Module {}
|
||||||
|
|
||||||
|
impl Event for Module {
|
||||||
|
type Body = ModuleBody;
|
||||||
|
const EVENT: &'static str = "module";
|
||||||
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct Module {
|
pub struct ModuleBody {
|
||||||
pub reason: String,
|
pub reason: String,
|
||||||
pub module: super::Module,
|
pub module: super::Module,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug)]
|
||||||
#[serde(rename_all = "camelCase")]
|
pub enum LoadedSource {}
|
||||||
pub struct LoadedSource {
|
|
||||||
pub reason: String,
|
impl Event for LoadedSource {
|
||||||
pub source: super::Source,
|
type Body = LoadedSourceBody;
|
||||||
|
const EVENT: &'static str = "loadedSource";
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct Process {
|
pub struct LoadedSourceBody {
|
||||||
|
pub reason: String,
|
||||||
|
pub source: super::Source,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Process {}
|
||||||
|
|
||||||
|
impl Event for Process {
|
||||||
|
type Body = ProcessBody;
|
||||||
|
const EVENT: &'static str = "process";
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessBody {
|
||||||
pub name: String,
|
pub name: String,
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub system_process_id: Option<usize>,
|
pub system_process_id: Option<usize>,
|
||||||
|
@ -884,39 +953,55 @@ pub mod events {
|
||||||
pub pointer_size: Option<usize>,
|
pub pointer_size: Option<usize>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Capabilities {}
|
||||||
|
|
||||||
|
impl Event for Capabilities {
|
||||||
|
type Body = CapabilitiesBody;
|
||||||
|
const EVENT: &'static str = "capabilities";
|
||||||
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct Capabilities {
|
pub struct CapabilitiesBody {
|
||||||
pub capabilities: super::DebuggerCapabilities,
|
pub capabilities: super::DebuggerCapabilities,
|
||||||
}
|
}
|
||||||
|
|
||||||
// #[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
// #[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
// #[serde(rename_all = "camelCase")]
|
// #[serde(rename_all = "camelCase")]
|
||||||
// pub struct Invalidated {
|
// pub struct InvalidatedBody {
|
||||||
// pub areas: Vec<InvalidatedArea>,
|
// pub areas: Vec<InvalidatedArea>,
|
||||||
// pub thread_id: Option<ThreadId>,
|
// pub thread_id: Option<ThreadId>,
|
||||||
// pub stack_frame_id: Option<usize>,
|
// pub stack_frame_id: Option<usize>,
|
||||||
// }
|
// }
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Memory {}
|
||||||
|
|
||||||
|
impl Event for Memory {
|
||||||
|
type Body = MemoryBody;
|
||||||
|
const EVENT: &'static str = "memory";
|
||||||
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
#[derive(Debug, PartialEq, Eq, Clone, Deserialize, Serialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct Memory {
|
pub struct MemoryBody {
|
||||||
pub memory_reference: String,
|
pub memory_reference: String,
|
||||||
pub offset: usize,
|
pub offset: usize,
|
||||||
pub count: usize,
|
pub count: usize,
|
||||||
}
|
}
|
||||||
|
}
|
||||||
#[test]
|
|
||||||
fn test_deserialize_module_id_from_number() {
|
#[test]
|
||||||
let raw = r#"{"id": 0, "name": "Name"}"#;
|
fn test_deserialize_module_id_from_number() {
|
||||||
let module: super::Module = serde_json::from_str(raw).expect("Error!");
|
let raw = r#"{"id": 0, "name": "Name"}"#;
|
||||||
assert_eq!(module.id, "0");
|
let module: Module = serde_json::from_str(raw).expect("Error!");
|
||||||
}
|
assert_eq!(module.id, "0");
|
||||||
|
}
|
||||||
#[test]
|
|
||||||
fn test_deserialize_module_id_from_string() {
|
#[test]
|
||||||
let raw = r#"{"id": "0", "name": "Name"}"#;
|
fn test_deserialize_module_id_from_string() {
|
||||||
let module: super::Module = serde_json::from_str(raw).expect("Error!");
|
let raw = r#"{"id": "0", "name": "Name"}"#;
|
||||||
assert_eq!(module.id, "0");
|
let module: Module = serde_json::from_str(raw).expect("Error!");
|
||||||
}
|
assert_eq!(module.id, "0");
|
||||||
}
|
}
|
||||||
|
|
|
@ -14,12 +14,13 @@ homepage.workspace = true
|
||||||
[dependencies]
|
[dependencies]
|
||||||
dunce = "1.0"
|
dunce = "1.0"
|
||||||
etcetera = "0.8"
|
etcetera = "0.8"
|
||||||
ropey = { version = "1.6.1", default-features = false }
|
ropey.workspace = true
|
||||||
which = "7.0"
|
which = "7.0"
|
||||||
regex-cursor = "0.1.4"
|
regex-cursor = "0.1.4"
|
||||||
bitflags.workspace = true
|
bitflags.workspace = true
|
||||||
once_cell = "1.19"
|
once_cell = "1.19"
|
||||||
regex-automata = "0.4.9"
|
regex-automata = "0.4.9"
|
||||||
|
unicode-segmentation.workspace = true
|
||||||
|
|
||||||
[target.'cfg(windows)'.dependencies]
|
[target.'cfg(windows)'.dependencies]
|
||||||
windows-sys = { version = "0.59", features = ["Win32_Foundation", "Win32_Security", "Win32_Security_Authorization", "Win32_Storage_FileSystem", "Win32_System_Threading"] }
|
windows-sys = { version = "0.59", features = ["Win32_Foundation", "Win32_Security", "Win32_Security_Authorization", "Win32_Storage_FileSystem", "Win32_System_Threading"] }
|
||||||
|
|
|
@ -1,9 +1,12 @@
|
||||||
|
use std::fmt;
|
||||||
use std::ops::{Bound, RangeBounds};
|
use std::ops::{Bound, RangeBounds};
|
||||||
|
|
||||||
pub use regex_cursor::engines::meta::{Builder as RegexBuilder, Regex};
|
pub use regex_cursor::engines::meta::{Builder as RegexBuilder, Regex};
|
||||||
pub use regex_cursor::regex_automata::util::syntax::Config;
|
pub use regex_cursor::regex_automata::util::syntax::Config;
|
||||||
use regex_cursor::{Input as RegexInput, RopeyCursor};
|
use regex_cursor::{Input as RegexInput, RopeyCursor};
|
||||||
|
use ropey::iter::Chunks;
|
||||||
use ropey::RopeSlice;
|
use ropey::RopeSlice;
|
||||||
|
use unicode_segmentation::{GraphemeCursor, GraphemeIncomplete};
|
||||||
|
|
||||||
pub trait RopeSliceExt<'a>: Sized {
|
pub trait RopeSliceExt<'a>: Sized {
|
||||||
fn ends_with(self, text: &str) -> bool;
|
fn ends_with(self, text: &str) -> bool;
|
||||||
|
@ -52,6 +55,102 @@ pub trait RopeSliceExt<'a>: Sized {
|
||||||
/// assert_eq!(text.ceil_char_boundary(3), 3);
|
/// assert_eq!(text.ceil_char_boundary(3), 3);
|
||||||
/// ```
|
/// ```
|
||||||
fn ceil_char_boundary(self, byte_idx: usize) -> usize;
|
fn ceil_char_boundary(self, byte_idx: usize) -> usize;
|
||||||
|
/// Checks whether the given `byte_idx` lies on a character boundary.
|
||||||
|
///
|
||||||
|
/// # Example
|
||||||
|
///
|
||||||
|
/// ```
|
||||||
|
/// # use ropey::RopeSlice;
|
||||||
|
/// # use helix_stdx::rope::RopeSliceExt;
|
||||||
|
/// let text = RopeSlice::from("⌚"); // three bytes: e2 8c 9a
|
||||||
|
/// assert!(text.is_char_boundary(0));
|
||||||
|
/// assert!(!text.is_char_boundary(1));
|
||||||
|
/// assert!(!text.is_char_boundary(2));
|
||||||
|
/// assert!(text.is_char_boundary(3));
|
||||||
|
/// ```
|
||||||
|
#[allow(clippy::wrong_self_convention)]
|
||||||
|
fn is_char_boundary(self, byte_idx: usize) -> bool;
|
||||||
|
/// Finds the closest byte index not exceeding `byte_idx` which lies on a grapheme cluster
|
||||||
|
/// boundary.
|
||||||
|
///
|
||||||
|
/// If `byte_idx` already lies on a grapheme cluster boundary then it is returned as-is. When
|
||||||
|
/// `byte_idx` lies between two grapheme cluster boundaries, this function returns the byte
|
||||||
|
/// index of the lesser / earlier / left-hand-side boundary.
|
||||||
|
///
|
||||||
|
/// `byte_idx` does not need to be aligned to a character boundary.
|
||||||
|
///
|
||||||
|
/// # Example
|
||||||
|
///
|
||||||
|
/// ```
|
||||||
|
/// # use ropey::RopeSlice;
|
||||||
|
/// # use helix_stdx::rope::RopeSliceExt;
|
||||||
|
/// let text = RopeSlice::from("\r\n"); // U+000D U+000A, hex: 0d 0a
|
||||||
|
/// assert_eq!(text.floor_grapheme_boundary(0), 0);
|
||||||
|
/// assert_eq!(text.floor_grapheme_boundary(1), 0);
|
||||||
|
/// assert_eq!(text.floor_grapheme_boundary(2), 2);
|
||||||
|
/// ```
|
||||||
|
fn floor_grapheme_boundary(self, byte_idx: usize) -> usize;
|
||||||
|
/// Finds the closest byte index not exceeding `byte_idx` which lies on a grapheme cluster
|
||||||
|
/// boundary.
|
||||||
|
///
|
||||||
|
/// If `byte_idx` already lies on a grapheme cluster boundary then it is returned as-is. When
|
||||||
|
/// `byte_idx` lies between two grapheme cluster boundaries, this function returns the byte
|
||||||
|
/// index of the greater / later / right-hand-side boundary.
|
||||||
|
///
|
||||||
|
/// `byte_idx` does not need to be aligned to a character boundary.
|
||||||
|
///
|
||||||
|
/// # Example
|
||||||
|
///
|
||||||
|
/// ```
|
||||||
|
/// # use ropey::RopeSlice;
|
||||||
|
/// # use helix_stdx::rope::RopeSliceExt;
|
||||||
|
/// let text = RopeSlice::from("\r\n"); // U+000D U+000A, hex: 0d 0a
|
||||||
|
/// assert_eq!(text.ceil_grapheme_boundary(0), 0);
|
||||||
|
/// assert_eq!(text.ceil_grapheme_boundary(1), 2);
|
||||||
|
/// assert_eq!(text.ceil_grapheme_boundary(2), 2);
|
||||||
|
/// ```
|
||||||
|
fn ceil_grapheme_boundary(self, byte_idx: usize) -> usize;
|
||||||
|
/// Checks whether the `byte_idx` lies on a grapheme cluster boundary.
|
||||||
|
///
|
||||||
|
/// # Example
|
||||||
|
///
|
||||||
|
/// ```
|
||||||
|
/// # use ropey::RopeSlice;
|
||||||
|
/// # use helix_stdx::rope::RopeSliceExt;
|
||||||
|
/// let text = RopeSlice::from("\r\n"); // U+000D U+000A, hex: 0d 0a
|
||||||
|
/// assert!(text.is_grapheme_boundary(0));
|
||||||
|
/// assert!(!text.is_grapheme_boundary(1));
|
||||||
|
/// assert!(text.is_grapheme_boundary(2));
|
||||||
|
/// ```
|
||||||
|
#[allow(clippy::wrong_self_convention)]
|
||||||
|
fn is_grapheme_boundary(self, byte_idx: usize) -> bool;
|
||||||
|
/// Returns an iterator over the grapheme clusters in the slice.
|
||||||
|
///
|
||||||
|
/// # Example
|
||||||
|
///
|
||||||
|
/// ```
|
||||||
|
/// # use ropey::RopeSlice;
|
||||||
|
/// # use helix_stdx::rope::RopeSliceExt;
|
||||||
|
/// let text = RopeSlice::from("😶🌫️🏴☠️🖼️");
|
||||||
|
/// let graphemes: Vec<_> = text.graphemes().collect();
|
||||||
|
/// assert_eq!(graphemes.as_slice(), &["😶🌫️", "🏴☠️", "🖼️"]);
|
||||||
|
/// ```
|
||||||
|
fn graphemes(self) -> RopeGraphemes<'a>;
|
||||||
|
/// Returns an iterator over the grapheme clusters in the slice, reversed.
|
||||||
|
///
|
||||||
|
/// The returned iterator starts at the end of the slice and ends at the beginning of the
|
||||||
|
/// slice.
|
||||||
|
///
|
||||||
|
/// # Example
|
||||||
|
///
|
||||||
|
/// ```
|
||||||
|
/// # use ropey::RopeSlice;
|
||||||
|
/// # use helix_stdx::rope::RopeSliceExt;
|
||||||
|
/// let text = RopeSlice::from("😶🌫️🏴☠️🖼️");
|
||||||
|
/// let graphemes: Vec<_> = text.graphemes_rev().collect();
|
||||||
|
/// assert_eq!(graphemes.as_slice(), &["🖼️", "🏴☠️", "😶🌫️"]);
|
||||||
|
/// ```
|
||||||
|
fn graphemes_rev(self) -> RevRopeGraphemes<'a>;
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<'a> RopeSliceExt<'a> for RopeSlice<'a> {
|
impl<'a> RopeSliceExt<'a> for RopeSlice<'a> {
|
||||||
|
@ -112,7 +211,7 @@ impl<'a> RopeSliceExt<'a> for RopeSlice<'a> {
|
||||||
.map(|pos| self.len_chars() - pos - 1)
|
.map(|pos| self.len_chars() - pos - 1)
|
||||||
}
|
}
|
||||||
|
|
||||||
// These two are adapted from std's `round_char_boundary` functions:
|
// These three are adapted from std:
|
||||||
|
|
||||||
fn floor_char_boundary(self, byte_idx: usize) -> usize {
|
fn floor_char_boundary(self, byte_idx: usize) -> usize {
|
||||||
if byte_idx >= self.len_bytes() {
|
if byte_idx >= self.len_bytes() {
|
||||||
|
@ -140,6 +239,127 @@ impl<'a> RopeSliceExt<'a> for RopeSlice<'a> {
|
||||||
.map_or(upper_bound, |pos| pos + byte_idx)
|
.map_or(upper_bound, |pos| pos + byte_idx)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn is_char_boundary(self, byte_idx: usize) -> bool {
|
||||||
|
if byte_idx == 0 {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
if byte_idx >= self.len_bytes() {
|
||||||
|
byte_idx == self.len_bytes()
|
||||||
|
} else {
|
||||||
|
is_utf8_char_boundary(self.bytes_at(byte_idx).next().unwrap())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn floor_grapheme_boundary(self, mut byte_idx: usize) -> usize {
|
||||||
|
if byte_idx >= self.len_bytes() {
|
||||||
|
return self.len_bytes();
|
||||||
|
}
|
||||||
|
|
||||||
|
byte_idx = self.ceil_char_boundary(byte_idx + 1);
|
||||||
|
|
||||||
|
let (mut chunk, mut chunk_byte_idx, _, _) = self.chunk_at_byte(byte_idx);
|
||||||
|
|
||||||
|
let mut cursor = GraphemeCursor::new(byte_idx, self.len_bytes(), true);
|
||||||
|
|
||||||
|
loop {
|
||||||
|
match cursor.prev_boundary(chunk, chunk_byte_idx) {
|
||||||
|
Ok(None) => return 0,
|
||||||
|
Ok(Some(boundary)) => return boundary,
|
||||||
|
Err(GraphemeIncomplete::PrevChunk) => {
|
||||||
|
let (ch, ch_byte_idx, _, _) = self.chunk_at_byte(chunk_byte_idx - 1);
|
||||||
|
chunk = ch;
|
||||||
|
chunk_byte_idx = ch_byte_idx;
|
||||||
|
}
|
||||||
|
Err(GraphemeIncomplete::PreContext(n)) => {
|
||||||
|
let ctx_chunk = self.chunk_at_byte(n - 1).0;
|
||||||
|
cursor.provide_context(ctx_chunk, n - ctx_chunk.len());
|
||||||
|
}
|
||||||
|
_ => unreachable!(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn ceil_grapheme_boundary(self, mut byte_idx: usize) -> usize {
|
||||||
|
if byte_idx >= self.len_bytes() {
|
||||||
|
return self.len_bytes();
|
||||||
|
}
|
||||||
|
|
||||||
|
if byte_idx == 0 {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
byte_idx = self.floor_char_boundary(byte_idx - 1);
|
||||||
|
|
||||||
|
let (mut chunk, mut chunk_byte_idx, _, _) = self.chunk_at_byte(byte_idx);
|
||||||
|
|
||||||
|
let mut cursor = GraphemeCursor::new(byte_idx, self.len_bytes(), true);
|
||||||
|
|
||||||
|
loop {
|
||||||
|
match cursor.next_boundary(chunk, chunk_byte_idx) {
|
||||||
|
Ok(None) => return self.len_bytes(),
|
||||||
|
Ok(Some(boundary)) => return boundary,
|
||||||
|
Err(GraphemeIncomplete::NextChunk) => {
|
||||||
|
chunk_byte_idx += chunk.len();
|
||||||
|
chunk = self.chunk_at_byte(chunk_byte_idx).0;
|
||||||
|
}
|
||||||
|
Err(GraphemeIncomplete::PreContext(n)) => {
|
||||||
|
let ctx_chunk = self.chunk_at_byte(n - 1).0;
|
||||||
|
cursor.provide_context(ctx_chunk, n - ctx_chunk.len());
|
||||||
|
}
|
||||||
|
_ => unreachable!(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn is_grapheme_boundary(self, byte_idx: usize) -> bool {
|
||||||
|
// The byte must lie on a character boundary to lie on a grapheme cluster boundary.
|
||||||
|
if !self.is_char_boundary(byte_idx) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
let (chunk, chunk_byte_idx, _, _) = self.chunk_at_byte(byte_idx);
|
||||||
|
|
||||||
|
let mut cursor = GraphemeCursor::new(byte_idx, self.len_bytes(), true);
|
||||||
|
|
||||||
|
loop {
|
||||||
|
match cursor.is_boundary(chunk, chunk_byte_idx) {
|
||||||
|
Ok(n) => return n,
|
||||||
|
Err(GraphemeIncomplete::PreContext(n)) => {
|
||||||
|
let (ctx_chunk, ctx_byte_start, _, _) = self.chunk_at_byte(n - 1);
|
||||||
|
cursor.provide_context(ctx_chunk, ctx_byte_start);
|
||||||
|
}
|
||||||
|
Err(_) => unreachable!(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn graphemes(self) -> RopeGraphemes<'a> {
|
||||||
|
let mut chunks = self.chunks();
|
||||||
|
let first_chunk = chunks.next().unwrap_or("");
|
||||||
|
RopeGraphemes {
|
||||||
|
text: self,
|
||||||
|
chunks,
|
||||||
|
cur_chunk: first_chunk,
|
||||||
|
cur_chunk_start: 0,
|
||||||
|
cursor: GraphemeCursor::new(0, self.len_bytes(), true),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn graphemes_rev(self) -> RevRopeGraphemes<'a> {
|
||||||
|
let (mut chunks, mut cur_chunk_start, _, _) = self.chunks_at_byte(self.len_bytes());
|
||||||
|
chunks.reverse();
|
||||||
|
let first_chunk = chunks.next().unwrap_or("");
|
||||||
|
cur_chunk_start -= first_chunk.len();
|
||||||
|
RevRopeGraphemes {
|
||||||
|
text: self,
|
||||||
|
chunks,
|
||||||
|
cur_chunk: first_chunk,
|
||||||
|
cur_chunk_start,
|
||||||
|
cursor: GraphemeCursor::new(self.len_bytes(), self.len_bytes(), true),
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// copied from std
|
// copied from std
|
||||||
|
@ -149,6 +369,130 @@ const fn is_utf8_char_boundary(b: u8) -> bool {
|
||||||
(b as i8) >= -0x40
|
(b as i8) >= -0x40
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// An iterator over the graphemes of a `RopeSlice`.
|
||||||
|
#[derive(Clone)]
|
||||||
|
pub struct RopeGraphemes<'a> {
|
||||||
|
text: RopeSlice<'a>,
|
||||||
|
chunks: Chunks<'a>,
|
||||||
|
cur_chunk: &'a str,
|
||||||
|
cur_chunk_start: usize,
|
||||||
|
cursor: GraphemeCursor,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl fmt::Debug for RopeGraphemes<'_> {
|
||||||
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
|
f.debug_struct("RopeGraphemes")
|
||||||
|
.field("text", &self.text)
|
||||||
|
.field("chunks", &self.chunks)
|
||||||
|
.field("cur_chunk", &self.cur_chunk)
|
||||||
|
.field("cur_chunk_start", &self.cur_chunk_start)
|
||||||
|
// .field("cursor", &self.cursor)
|
||||||
|
.finish()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> Iterator for RopeGraphemes<'a> {
|
||||||
|
type Item = RopeSlice<'a>;
|
||||||
|
|
||||||
|
fn next(&mut self) -> Option<Self::Item> {
|
||||||
|
let a = self.cursor.cur_cursor();
|
||||||
|
let b;
|
||||||
|
loop {
|
||||||
|
match self
|
||||||
|
.cursor
|
||||||
|
.next_boundary(self.cur_chunk, self.cur_chunk_start)
|
||||||
|
{
|
||||||
|
Ok(None) => {
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
Ok(Some(n)) => {
|
||||||
|
b = n;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
Err(GraphemeIncomplete::NextChunk) => {
|
||||||
|
self.cur_chunk_start += self.cur_chunk.len();
|
||||||
|
self.cur_chunk = self.chunks.next().unwrap_or("");
|
||||||
|
}
|
||||||
|
Err(GraphemeIncomplete::PreContext(idx)) => {
|
||||||
|
let (chunk, byte_idx, _, _) = self.text.chunk_at_byte(idx.saturating_sub(1));
|
||||||
|
self.cursor.provide_context(chunk, byte_idx);
|
||||||
|
}
|
||||||
|
_ => unreachable!(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if a < self.cur_chunk_start {
|
||||||
|
Some(self.text.byte_slice(a..b))
|
||||||
|
} else {
|
||||||
|
let a2 = a - self.cur_chunk_start;
|
||||||
|
let b2 = b - self.cur_chunk_start;
|
||||||
|
Some((&self.cur_chunk[a2..b2]).into())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// An iterator over the graphemes of a `RopeSlice` in reverse.
|
||||||
|
#[derive(Clone)]
|
||||||
|
pub struct RevRopeGraphemes<'a> {
|
||||||
|
text: RopeSlice<'a>,
|
||||||
|
chunks: Chunks<'a>,
|
||||||
|
cur_chunk: &'a str,
|
||||||
|
cur_chunk_start: usize,
|
||||||
|
cursor: GraphemeCursor,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl fmt::Debug for RevRopeGraphemes<'_> {
|
||||||
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
|
f.debug_struct("RevRopeGraphemes")
|
||||||
|
.field("text", &self.text)
|
||||||
|
.field("chunks", &self.chunks)
|
||||||
|
.field("cur_chunk", &self.cur_chunk)
|
||||||
|
.field("cur_chunk_start", &self.cur_chunk_start)
|
||||||
|
// .field("cursor", &self.cursor)
|
||||||
|
.finish()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> Iterator for RevRopeGraphemes<'a> {
|
||||||
|
type Item = RopeSlice<'a>;
|
||||||
|
|
||||||
|
fn next(&mut self) -> Option<Self::Item> {
|
||||||
|
let a = self.cursor.cur_cursor();
|
||||||
|
let b;
|
||||||
|
loop {
|
||||||
|
match self
|
||||||
|
.cursor
|
||||||
|
.prev_boundary(self.cur_chunk, self.cur_chunk_start)
|
||||||
|
{
|
||||||
|
Ok(None) => {
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
Ok(Some(n)) => {
|
||||||
|
b = n;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
Err(GraphemeIncomplete::PrevChunk) => {
|
||||||
|
self.cur_chunk = self.chunks.next().unwrap_or("");
|
||||||
|
self.cur_chunk_start -= self.cur_chunk.len();
|
||||||
|
}
|
||||||
|
Err(GraphemeIncomplete::PreContext(idx)) => {
|
||||||
|
let (chunk, byte_idx, _, _) = self.text.chunk_at_byte(idx.saturating_sub(1));
|
||||||
|
self.cursor.provide_context(chunk, byte_idx);
|
||||||
|
}
|
||||||
|
_ => unreachable!(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if a >= self.cur_chunk_start + self.cur_chunk.len() {
|
||||||
|
Some(self.text.byte_slice(b..a))
|
||||||
|
} else {
|
||||||
|
let a2 = a - self.cur_chunk_start;
|
||||||
|
let b2 = b - self.cur_chunk_start;
|
||||||
|
Some((&self.cur_chunk[b2..a2]).into())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use ropey::RopeSlice;
|
use ropey::RopeSlice;
|
||||||
|
@ -166,12 +510,13 @@ mod tests {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn floor_ceil_char_boundary() {
|
fn char_boundaries() {
|
||||||
let ascii = RopeSlice::from("ascii");
|
let ascii = RopeSlice::from("ascii");
|
||||||
// When the given index lies on a character boundary, the index should not change.
|
// When the given index lies on a character boundary, the index should not change.
|
||||||
for byte_idx in 0..=ascii.len_bytes() {
|
for byte_idx in 0..=ascii.len_bytes() {
|
||||||
assert_eq!(ascii.floor_char_boundary(byte_idx), byte_idx);
|
assert_eq!(ascii.floor_char_boundary(byte_idx), byte_idx);
|
||||||
assert_eq!(ascii.ceil_char_boundary(byte_idx), byte_idx);
|
assert_eq!(ascii.ceil_char_boundary(byte_idx), byte_idx);
|
||||||
|
assert!(ascii.is_char_boundary(byte_idx));
|
||||||
}
|
}
|
||||||
|
|
||||||
// This is a polyfill of a method of this trait which was replaced by ceil_char_boundary.
|
// This is a polyfill of a method of this trait which was replaced by ceil_char_boundary.
|
||||||
|
@ -198,4 +543,44 @@ mod tests {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn grapheme_boundaries() {
|
||||||
|
let ascii = RopeSlice::from("ascii");
|
||||||
|
// When the given index lies on a grapheme boundary, the index should not change.
|
||||||
|
for byte_idx in 0..=ascii.len_bytes() {
|
||||||
|
assert_eq!(ascii.floor_char_boundary(byte_idx), byte_idx);
|
||||||
|
assert_eq!(ascii.ceil_char_boundary(byte_idx), byte_idx);
|
||||||
|
assert!(ascii.is_grapheme_boundary(byte_idx));
|
||||||
|
}
|
||||||
|
|
||||||
|
// 🏴☠️: U+1F3F4 U+200D U+2620 U+FE0F
|
||||||
|
// 13 bytes, hex: f0 9f 8f b4 + e2 80 8d + e2 98 a0 + ef b8 8f
|
||||||
|
let g = RopeSlice::from("🏴☠️\r\n");
|
||||||
|
let emoji_len = "🏴☠️".len();
|
||||||
|
let end = g.len_bytes();
|
||||||
|
|
||||||
|
for byte_idx in 0..emoji_len {
|
||||||
|
assert_eq!(g.floor_grapheme_boundary(byte_idx), 0);
|
||||||
|
}
|
||||||
|
for byte_idx in emoji_len..end {
|
||||||
|
assert_eq!(g.floor_grapheme_boundary(byte_idx), emoji_len);
|
||||||
|
}
|
||||||
|
assert_eq!(g.floor_grapheme_boundary(end), end);
|
||||||
|
|
||||||
|
assert_eq!(g.ceil_grapheme_boundary(0), 0);
|
||||||
|
for byte_idx in 1..=emoji_len {
|
||||||
|
assert_eq!(g.ceil_grapheme_boundary(byte_idx), emoji_len);
|
||||||
|
}
|
||||||
|
for byte_idx in emoji_len + 1..=end {
|
||||||
|
assert_eq!(g.ceil_grapheme_boundary(byte_idx), end);
|
||||||
|
}
|
||||||
|
|
||||||
|
assert!(g.is_grapheme_boundary(0));
|
||||||
|
assert!(g.is_grapheme_boundary(emoji_len));
|
||||||
|
assert!(g.is_grapheme_boundary(end));
|
||||||
|
for byte_idx in (1..emoji_len).chain(emoji_len + 1..end) {
|
||||||
|
assert!(!g.is_grapheme_boundary(byte_idx));
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -61,7 +61,7 @@ tokio-stream = "0.1"
|
||||||
futures-util = { version = "0.3", features = ["std", "async-await"], default-features = false }
|
futures-util = { version = "0.3", features = ["std", "async-await"], default-features = false }
|
||||||
arc-swap = { version = "1.7.1" }
|
arc-swap = { version = "1.7.1" }
|
||||||
termini = "1"
|
termini = "1"
|
||||||
indexmap = "2.5"
|
indexmap = "2.7"
|
||||||
|
|
||||||
# Logging
|
# Logging
|
||||||
fern = "0.7"
|
fern = "0.7"
|
||||||
|
|
|
@ -20,7 +20,7 @@ use helix_core::{
|
||||||
comment,
|
comment,
|
||||||
doc_formatter::TextFormat,
|
doc_formatter::TextFormat,
|
||||||
encoding, find_workspace,
|
encoding, find_workspace,
|
||||||
graphemes::{self, next_grapheme_boundary, RevRopeGraphemes},
|
graphemes::{self, next_grapheme_boundary},
|
||||||
history::UndoKind,
|
history::UndoKind,
|
||||||
increment,
|
increment,
|
||||||
indent::{self, IndentStyle},
|
indent::{self, IndentStyle},
|
||||||
|
@ -35,8 +35,8 @@ use helix_core::{
|
||||||
text_annotations::{Overlay, TextAnnotations},
|
text_annotations::{Overlay, TextAnnotations},
|
||||||
textobject,
|
textobject,
|
||||||
unicode::width::UnicodeWidthChar,
|
unicode::width::UnicodeWidthChar,
|
||||||
visual_offset_from_block, Deletion, LineEnding, Position, Range, Rope, RopeGraphemes,
|
visual_offset_from_block, Deletion, LineEnding, Position, Range, Rope, RopeReader, RopeSlice,
|
||||||
RopeReader, RopeSlice, Selection, SmallVec, Syntax, Tendril, Transaction,
|
Selection, SmallVec, Syntax, Tendril, Transaction,
|
||||||
};
|
};
|
||||||
use helix_view::{
|
use helix_view::{
|
||||||
document::{FormatterError, Mode, SCRATCH_BUFFER_NAME},
|
document::{FormatterError, Mode, SCRATCH_BUFFER_NAME},
|
||||||
|
@ -633,10 +633,17 @@ impl std::str::FromStr for MappableCommand {
|
||||||
.collect::<Vec<String>>();
|
.collect::<Vec<String>>();
|
||||||
typed::TYPABLE_COMMAND_MAP
|
typed::TYPABLE_COMMAND_MAP
|
||||||
.get(name)
|
.get(name)
|
||||||
.map(|cmd| MappableCommand::Typable {
|
.map(|cmd| {
|
||||||
name: cmd.name.to_owned(),
|
let doc = if args.is_empty() {
|
||||||
doc: format!(":{} {:?}", cmd.name, args),
|
cmd.doc.to_string()
|
||||||
args,
|
} else {
|
||||||
|
format!(":{} {:?}", cmd.name, args)
|
||||||
|
};
|
||||||
|
MappableCommand::Typable {
|
||||||
|
name: cmd.name.to_owned(),
|
||||||
|
doc,
|
||||||
|
args,
|
||||||
|
}
|
||||||
})
|
})
|
||||||
.ok_or_else(|| anyhow!("No TypableCommand named '{}'", s))
|
.ok_or_else(|| anyhow!("No TypableCommand named '{}'", s))
|
||||||
} else if let Some(suffix) = s.strip_prefix('@') {
|
} else if let Some(suffix) = s.strip_prefix('@') {
|
||||||
|
@ -1681,10 +1688,12 @@ fn replace(cx: &mut Context) {
|
||||||
if let Some(ch) = ch {
|
if let Some(ch) = ch {
|
||||||
let transaction = Transaction::change_by_selection(doc.text(), selection, |range| {
|
let transaction = Transaction::change_by_selection(doc.text(), selection, |range| {
|
||||||
if !range.is_empty() {
|
if !range.is_empty() {
|
||||||
let text: Tendril =
|
let text: Tendril = doc
|
||||||
RopeGraphemes::new(doc.text().slice(range.from()..range.to()))
|
.text()
|
||||||
.map(|_g| ch)
|
.slice(range.from()..range.to())
|
||||||
.collect();
|
.graphemes()
|
||||||
|
.map(|_g| ch)
|
||||||
|
.collect();
|
||||||
(range.from(), range.to(), Some(text))
|
(range.from(), range.to(), Some(text))
|
||||||
} else {
|
} else {
|
||||||
// No change.
|
// No change.
|
||||||
|
@ -3512,6 +3521,10 @@ async fn make_format_callback(
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
|
if write.is_none() {
|
||||||
|
editor.set_error(err.to_string());
|
||||||
|
return;
|
||||||
|
}
|
||||||
log::info!("failed to format '{}': {err}", doc.display_name());
|
log::info!("failed to format '{}': {err}", doc.display_name());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -3822,8 +3835,7 @@ fn goto_next_diag(cx: &mut Context) {
|
||||||
let diag = doc
|
let diag = doc
|
||||||
.diagnostics()
|
.diagnostics()
|
||||||
.iter()
|
.iter()
|
||||||
.find(|diag| diag.range.start > cursor_pos)
|
.find(|diag| diag.range.start > cursor_pos);
|
||||||
.or_else(|| doc.diagnostics().first());
|
|
||||||
|
|
||||||
let selection = match diag {
|
let selection = match diag {
|
||||||
Some(diag) => Selection::single(diag.range.start, diag.range.end),
|
Some(diag) => Selection::single(diag.range.start, diag.range.end),
|
||||||
|
@ -3850,8 +3862,7 @@ fn goto_prev_diag(cx: &mut Context) {
|
||||||
.diagnostics()
|
.diagnostics()
|
||||||
.iter()
|
.iter()
|
||||||
.rev()
|
.rev()
|
||||||
.find(|diag| diag.range.start < cursor_pos)
|
.find(|diag| diag.range.start < cursor_pos);
|
||||||
.or_else(|| doc.diagnostics().last());
|
|
||||||
|
|
||||||
let selection = match diag {
|
let selection = match diag {
|
||||||
// NOTE: the selection is reversed because we're jumping to the
|
// NOTE: the selection is reversed because we're jumping to the
|
||||||
|
@ -6488,6 +6499,7 @@ fn jump_to_label(cx: &mut Context, labels: Vec<Range>, behaviour: Movement) {
|
||||||
let alphabet = &cx.editor.config().jump_label_alphabet;
|
let alphabet = &cx.editor.config().jump_label_alphabet;
|
||||||
let Some(i) = event
|
let Some(i) = event
|
||||||
.char()
|
.char()
|
||||||
|
.filter(|_| event.modifiers.is_empty())
|
||||||
.and_then(|ch| alphabet.iter().position(|&it| it == ch))
|
.and_then(|ch| alphabet.iter().position(|&it| it == ch))
|
||||||
else {
|
else {
|
||||||
doc_mut!(cx.editor, &doc).remove_jump_labels(view);
|
doc_mut!(cx.editor, &doc).remove_jump_labels(view);
|
||||||
|
@ -6504,6 +6516,7 @@ fn jump_to_label(cx: &mut Context, labels: Vec<Range>, behaviour: Movement) {
|
||||||
let alphabet = &cx.editor.config().jump_label_alphabet;
|
let alphabet = &cx.editor.config().jump_label_alphabet;
|
||||||
let Some(inner) = event
|
let Some(inner) = event
|
||||||
.char()
|
.char()
|
||||||
|
.filter(|_| event.modifiers.is_empty())
|
||||||
.and_then(|ch| alphabet.iter().position(|&it| it == ch))
|
.and_then(|ch| alphabet.iter().position(|&it| it == ch))
|
||||||
else {
|
else {
|
||||||
return;
|
return;
|
||||||
|
@ -6572,7 +6585,9 @@ fn jump_to_word(cx: &mut Context, behaviour: Movement) {
|
||||||
// madeup of word characters. The latter condition is needed because
|
// madeup of word characters. The latter condition is needed because
|
||||||
// move_next_word_end simply treats a sequence of characters from
|
// move_next_word_end simply treats a sequence of characters from
|
||||||
// the same char class as a word so `=<` would also count as a word.
|
// the same char class as a word so `=<` would also count as a word.
|
||||||
let add_label = RevRopeGraphemes::new(text.slice(..cursor_fwd.head))
|
let add_label = text
|
||||||
|
.slice(..cursor_fwd.head)
|
||||||
|
.graphemes_rev()
|
||||||
.take(2)
|
.take(2)
|
||||||
.take_while(|g| g.chars().all(char_is_word))
|
.take_while(|g| g.chars().all(char_is_word))
|
||||||
.count()
|
.count()
|
||||||
|
@ -6598,7 +6613,9 @@ fn jump_to_word(cx: &mut Context, behaviour: Movement) {
|
||||||
// madeup of word characters. The latter condition is needed because
|
// madeup of word characters. The latter condition is needed because
|
||||||
// move_prev_word_start simply treats a sequence of characters from
|
// move_prev_word_start simply treats a sequence of characters from
|
||||||
// the same char class as a word so `=<` would also count as a word.
|
// the same char class as a word so `=<` would also count as a word.
|
||||||
let add_label = RopeGraphemes::new(text.slice(cursor_rev.head..))
|
let add_label = text
|
||||||
|
.slice(cursor_rev.head..)
|
||||||
|
.graphemes()
|
||||||
.take(2)
|
.take(2)
|
||||||
.take_while(|g| g.chars().all(char_is_word))
|
.take_while(|g| g.chars().all(char_is_word))
|
||||||
.count()
|
.count()
|
||||||
|
|
|
@ -61,14 +61,19 @@ macro_rules! language_server_with_feature {
|
||||||
}};
|
}};
|
||||||
}
|
}
|
||||||
|
|
||||||
/// A wrapper around `lsp::Location` that swaps out the LSP URI for `helix_core::Uri`.
|
/// A wrapper around `lsp::Location` that swaps out the LSP URI for `helix_core::Uri` and adds
|
||||||
|
/// the server's offset encoding.
|
||||||
#[derive(Debug, Clone, PartialEq, Eq)]
|
#[derive(Debug, Clone, PartialEq, Eq)]
|
||||||
struct Location {
|
struct Location {
|
||||||
uri: Uri,
|
uri: Uri,
|
||||||
range: lsp::Range,
|
range: lsp::Range,
|
||||||
|
offset_encoding: OffsetEncoding,
|
||||||
}
|
}
|
||||||
|
|
||||||
fn lsp_location_to_location(location: lsp::Location) -> Option<Location> {
|
fn lsp_location_to_location(
|
||||||
|
location: lsp::Location,
|
||||||
|
offset_encoding: OffsetEncoding,
|
||||||
|
) -> Option<Location> {
|
||||||
let uri = match location.uri.try_into() {
|
let uri = match location.uri.try_into() {
|
||||||
Ok(uri) => uri,
|
Ok(uri) => uri,
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
|
@ -79,13 +84,13 @@ fn lsp_location_to_location(location: lsp::Location) -> Option<Location> {
|
||||||
Some(Location {
|
Some(Location {
|
||||||
uri,
|
uri,
|
||||||
range: location.range,
|
range: location.range,
|
||||||
|
offset_encoding,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
struct SymbolInformationItem {
|
struct SymbolInformationItem {
|
||||||
location: Location,
|
location: Location,
|
||||||
symbol: lsp::SymbolInformation,
|
symbol: lsp::SymbolInformation,
|
||||||
offset_encoding: OffsetEncoding,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
struct DiagnosticStyles {
|
struct DiagnosticStyles {
|
||||||
|
@ -98,7 +103,6 @@ struct DiagnosticStyles {
|
||||||
struct PickerDiagnostic {
|
struct PickerDiagnostic {
|
||||||
location: Location,
|
location: Location,
|
||||||
diag: lsp::Diagnostic,
|
diag: lsp::Diagnostic,
|
||||||
offset_encoding: OffsetEncoding,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn location_to_file_location(location: &Location) -> Option<FileLocation> {
|
fn location_to_file_location(location: &Location) -> Option<FileLocation> {
|
||||||
|
@ -110,12 +114,7 @@ fn location_to_file_location(location: &Location) -> Option<FileLocation> {
|
||||||
Some((path.into(), line))
|
Some((path.into(), line))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn jump_to_location(
|
fn jump_to_location(editor: &mut Editor, location: &Location, action: Action) {
|
||||||
editor: &mut Editor,
|
|
||||||
location: &Location,
|
|
||||||
offset_encoding: OffsetEncoding,
|
|
||||||
action: Action,
|
|
||||||
) {
|
|
||||||
let (view, doc) = current!(editor);
|
let (view, doc) = current!(editor);
|
||||||
push_jump(view, doc);
|
push_jump(view, doc);
|
||||||
|
|
||||||
|
@ -124,7 +123,13 @@ fn jump_to_location(
|
||||||
editor.set_error(err);
|
editor.set_error(err);
|
||||||
return;
|
return;
|
||||||
};
|
};
|
||||||
jump_to_position(editor, path, location.range, offset_encoding, action);
|
jump_to_position(
|
||||||
|
editor,
|
||||||
|
path,
|
||||||
|
location.range,
|
||||||
|
location.offset_encoding,
|
||||||
|
action,
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn jump_to_position(
|
fn jump_to_position(
|
||||||
|
@ -220,9 +225,9 @@ fn diag_picker(
|
||||||
location: Location {
|
location: Location {
|
||||||
uri: uri.clone(),
|
uri: uri.clone(),
|
||||||
range: diag.range,
|
range: diag.range,
|
||||||
|
offset_encoding: ls.offset_encoding(),
|
||||||
},
|
},
|
||||||
diag,
|
diag,
|
||||||
offset_encoding: ls.offset_encoding(),
|
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -286,7 +291,7 @@ fn diag_picker(
|
||||||
flat_diag,
|
flat_diag,
|
||||||
styles,
|
styles,
|
||||||
move |cx, diag, action| {
|
move |cx, diag, action| {
|
||||||
jump_to_location(cx.editor, &diag.location, diag.offset_encoding, action);
|
jump_to_location(cx.editor, &diag.location, action);
|
||||||
let (view, doc) = current!(cx.editor);
|
let (view, doc) = current!(cx.editor);
|
||||||
view.diagnostics_handler
|
view.diagnostics_handler
|
||||||
.immediately_show_diagnostic(doc, view.id);
|
.immediately_show_diagnostic(doc, view.id);
|
||||||
|
@ -314,10 +319,10 @@ pub fn symbol_picker(cx: &mut Context) {
|
||||||
location: lsp::Location::new(file.uri.clone(), symbol.selection_range),
|
location: lsp::Location::new(file.uri.clone(), symbol.selection_range),
|
||||||
container_name: None,
|
container_name: None,
|
||||||
},
|
},
|
||||||
offset_encoding,
|
|
||||||
location: Location {
|
location: Location {
|
||||||
uri: uri.clone(),
|
uri: uri.clone(),
|
||||||
range: symbol.selection_range,
|
range: symbol.selection_range,
|
||||||
|
offset_encoding,
|
||||||
},
|
},
|
||||||
});
|
});
|
||||||
for child in symbol.children.into_iter().flatten() {
|
for child in symbol.children.into_iter().flatten() {
|
||||||
|
@ -355,9 +360,9 @@ pub fn symbol_picker(cx: &mut Context) {
|
||||||
location: Location {
|
location: Location {
|
||||||
uri: doc_uri.clone(),
|
uri: doc_uri.clone(),
|
||||||
range: symbol.location.range,
|
range: symbol.location.range,
|
||||||
|
offset_encoding,
|
||||||
},
|
},
|
||||||
symbol,
|
symbol,
|
||||||
offset_encoding,
|
|
||||||
})
|
})
|
||||||
.collect(),
|
.collect(),
|
||||||
lsp::DocumentSymbolResponse::Nested(symbols) => {
|
lsp::DocumentSymbolResponse::Nested(symbols) => {
|
||||||
|
@ -410,7 +415,7 @@ pub fn symbol_picker(cx: &mut Context) {
|
||||||
symbols,
|
symbols,
|
||||||
(),
|
(),
|
||||||
move |cx, item, action| {
|
move |cx, item, action| {
|
||||||
jump_to_location(cx.editor, &item.location, item.offset_encoding, action);
|
jump_to_location(cx.editor, &item.location, action);
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
.with_preview(move |_editor, item| location_to_file_location(&item.location))
|
.with_preview(move |_editor, item| location_to_file_location(&item.location))
|
||||||
|
@ -467,9 +472,9 @@ pub fn workspace_symbol_picker(cx: &mut Context) {
|
||||||
location: Location {
|
location: Location {
|
||||||
uri,
|
uri,
|
||||||
range: symbol.location.range,
|
range: symbol.location.range,
|
||||||
|
offset_encoding,
|
||||||
},
|
},
|
||||||
symbol,
|
symbol,
|
||||||
offset_encoding,
|
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
.collect();
|
.collect();
|
||||||
|
@ -521,7 +526,7 @@ pub fn workspace_symbol_picker(cx: &mut Context) {
|
||||||
[],
|
[],
|
||||||
(),
|
(),
|
||||||
move |cx, item, action| {
|
move |cx, item, action| {
|
||||||
jump_to_location(cx.editor, &item.location, item.offset_encoding, action);
|
jump_to_location(cx.editor, &item.location, action);
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
.with_preview(|_editor, item| location_to_file_location(&item.location))
|
.with_preview(|_editor, item| location_to_file_location(&item.location))
|
||||||
|
@ -853,17 +858,12 @@ impl Display for ApplyEditErrorKind {
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Precondition: `locations` should be non-empty.
|
/// Precondition: `locations` should be non-empty.
|
||||||
fn goto_impl(
|
fn goto_impl(editor: &mut Editor, compositor: &mut Compositor, locations: Vec<Location>) {
|
||||||
editor: &mut Editor,
|
|
||||||
compositor: &mut Compositor,
|
|
||||||
locations: Vec<Location>,
|
|
||||||
offset_encoding: OffsetEncoding,
|
|
||||||
) {
|
|
||||||
let cwdir = helix_stdx::env::current_working_dir();
|
let cwdir = helix_stdx::env::current_working_dir();
|
||||||
|
|
||||||
match locations.as_slice() {
|
match locations.as_slice() {
|
||||||
[location] => {
|
[location] => {
|
||||||
jump_to_location(editor, location, offset_encoding, Action::Replace);
|
jump_to_location(editor, location, Action::Replace);
|
||||||
}
|
}
|
||||||
[] => unreachable!("`locations` should be non-empty for `goto_impl`"),
|
[] => unreachable!("`locations` should be non-empty for `goto_impl`"),
|
||||||
_locations => {
|
_locations => {
|
||||||
|
@ -880,58 +880,76 @@ fn goto_impl(
|
||||||
},
|
},
|
||||||
)];
|
)];
|
||||||
|
|
||||||
let picker = Picker::new(columns, 0, locations, cwdir, move |cx, location, action| {
|
let picker = Picker::new(columns, 0, locations, cwdir, |cx, location, action| {
|
||||||
jump_to_location(cx.editor, location, offset_encoding, action)
|
jump_to_location(cx.editor, location, action)
|
||||||
})
|
})
|
||||||
.with_preview(move |_editor, location| location_to_file_location(location));
|
.with_preview(|_editor, location| location_to_file_location(location));
|
||||||
compositor.push(Box::new(overlaid(picker)));
|
compositor.push(Box::new(overlaid(picker)));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn to_locations(definitions: Option<lsp::GotoDefinitionResponse>) -> Vec<Location> {
|
|
||||||
match definitions {
|
|
||||||
Some(lsp::GotoDefinitionResponse::Scalar(location)) => {
|
|
||||||
lsp_location_to_location(location).into_iter().collect()
|
|
||||||
}
|
|
||||||
Some(lsp::GotoDefinitionResponse::Array(locations)) => locations
|
|
||||||
.into_iter()
|
|
||||||
.flat_map(lsp_location_to_location)
|
|
||||||
.collect(),
|
|
||||||
Some(lsp::GotoDefinitionResponse::Link(locations)) => locations
|
|
||||||
.into_iter()
|
|
||||||
.map(|location_link| {
|
|
||||||
lsp::Location::new(location_link.target_uri, location_link.target_range)
|
|
||||||
})
|
|
||||||
.flat_map(lsp_location_to_location)
|
|
||||||
.collect(),
|
|
||||||
None => Vec::new(),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn goto_single_impl<P, F>(cx: &mut Context, feature: LanguageServerFeature, request_provider: P)
|
fn goto_single_impl<P, F>(cx: &mut Context, feature: LanguageServerFeature, request_provider: P)
|
||||||
where
|
where
|
||||||
P: Fn(&Client, lsp::Position, lsp::TextDocumentIdentifier) -> Option<F>,
|
P: Fn(&Client, lsp::Position, lsp::TextDocumentIdentifier) -> Option<F>,
|
||||||
F: Future<Output = helix_lsp::Result<serde_json::Value>> + 'static + Send,
|
F: Future<Output = helix_lsp::Result<serde_json::Value>> + 'static + Send,
|
||||||
{
|
{
|
||||||
let (view, doc) = current!(cx.editor);
|
let (view, doc) = current_ref!(cx.editor);
|
||||||
|
let mut futures: FuturesOrdered<_> = doc
|
||||||
|
.language_servers_with_feature(feature)
|
||||||
|
.map(|language_server| {
|
||||||
|
let offset_encoding = language_server.offset_encoding();
|
||||||
|
let pos = doc.position(view.id, offset_encoding);
|
||||||
|
let future = request_provider(language_server, pos, doc.identifier()).unwrap();
|
||||||
|
async move {
|
||||||
|
let json = future.await?;
|
||||||
|
let response: Option<lsp::GotoDefinitionResponse> = serde_json::from_value(json)?;
|
||||||
|
anyhow::Ok((response, offset_encoding))
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.collect();
|
||||||
|
|
||||||
let language_server = language_server_with_feature!(cx.editor, doc, feature);
|
cx.jobs.callback(async move {
|
||||||
let offset_encoding = language_server.offset_encoding();
|
let mut locations = Vec::new();
|
||||||
let pos = doc.position(view.id, offset_encoding);
|
while let Some((response, offset_encoding)) = futures.try_next().await? {
|
||||||
let future = request_provider(language_server, pos, doc.identifier()).unwrap();
|
match response {
|
||||||
|
Some(lsp::GotoDefinitionResponse::Scalar(lsp_location)) => {
|
||||||
cx.callback(
|
locations.extend(lsp_location_to_location(lsp_location, offset_encoding));
|
||||||
future,
|
}
|
||||||
move |editor, compositor, response: Option<lsp::GotoDefinitionResponse>| {
|
Some(lsp::GotoDefinitionResponse::Array(lsp_locations)) => {
|
||||||
let items = to_locations(response);
|
locations.extend(
|
||||||
if items.is_empty() {
|
lsp_locations.into_iter().flat_map(|location| {
|
||||||
|
lsp_location_to_location(location, offset_encoding)
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
Some(lsp::GotoDefinitionResponse::Link(lsp_locations)) => {
|
||||||
|
locations.extend(
|
||||||
|
lsp_locations
|
||||||
|
.into_iter()
|
||||||
|
.map(|location_link| {
|
||||||
|
lsp::Location::new(
|
||||||
|
location_link.target_uri,
|
||||||
|
location_link.target_range,
|
||||||
|
)
|
||||||
|
})
|
||||||
|
.flat_map(|location| {
|
||||||
|
lsp_location_to_location(location, offset_encoding)
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
None => (),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
let call = move |editor: &mut Editor, compositor: &mut Compositor| {
|
||||||
|
if locations.is_empty() {
|
||||||
editor.set_error("No definition found.");
|
editor.set_error("No definition found.");
|
||||||
} else {
|
} else {
|
||||||
goto_impl(editor, compositor, items, offset_encoding);
|
goto_impl(editor, compositor, locations);
|
||||||
}
|
}
|
||||||
},
|
};
|
||||||
);
|
Ok(Callback::EditorCompositor(Box::new(call)))
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn goto_declaration(cx: &mut Context) {
|
pub fn goto_declaration(cx: &mut Context) {
|
||||||
|
@ -968,38 +986,48 @@ pub fn goto_implementation(cx: &mut Context) {
|
||||||
|
|
||||||
pub fn goto_reference(cx: &mut Context) {
|
pub fn goto_reference(cx: &mut Context) {
|
||||||
let config = cx.editor.config();
|
let config = cx.editor.config();
|
||||||
let (view, doc) = current!(cx.editor);
|
let (view, doc) = current_ref!(cx.editor);
|
||||||
|
|
||||||
// TODO could probably support multiple language servers,
|
let mut futures: FuturesOrdered<_> = doc
|
||||||
// not sure if there's a real practical use case for this though
|
.language_servers_with_feature(LanguageServerFeature::GotoReference)
|
||||||
let language_server =
|
.map(|language_server| {
|
||||||
language_server_with_feature!(cx.editor, doc, LanguageServerFeature::GotoReference);
|
let offset_encoding = language_server.offset_encoding();
|
||||||
let offset_encoding = language_server.offset_encoding();
|
let pos = doc.position(view.id, offset_encoding);
|
||||||
let pos = doc.position(view.id, offset_encoding);
|
let future = language_server
|
||||||
let future = language_server
|
.goto_reference(
|
||||||
.goto_reference(
|
doc.identifier(),
|
||||||
doc.identifier(),
|
pos,
|
||||||
pos,
|
config.lsp.goto_reference_include_declaration,
|
||||||
config.lsp.goto_reference_include_declaration,
|
None,
|
||||||
None,
|
)
|
||||||
)
|
.unwrap();
|
||||||
.unwrap();
|
async move {
|
||||||
|
let json = future.await?;
|
||||||
|
let locations: Option<Vec<lsp::Location>> = serde_json::from_value(json)?;
|
||||||
|
anyhow::Ok((locations, offset_encoding))
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.collect();
|
||||||
|
|
||||||
cx.callback(
|
cx.jobs.callback(async move {
|
||||||
future,
|
let mut locations = Vec::new();
|
||||||
move |editor, compositor, response: Option<Vec<lsp::Location>>| {
|
while let Some((lsp_locations, offset_encoding)) = futures.try_next().await? {
|
||||||
let items: Vec<Location> = response
|
locations.extend(
|
||||||
.into_iter()
|
lsp_locations
|
||||||
.flatten()
|
.into_iter()
|
||||||
.flat_map(lsp_location_to_location)
|
.flatten()
|
||||||
.collect();
|
.flat_map(|location| lsp_location_to_location(location, offset_encoding)),
|
||||||
if items.is_empty() {
|
);
|
||||||
|
}
|
||||||
|
let call = move |editor: &mut Editor, compositor: &mut Compositor| {
|
||||||
|
if locations.is_empty() {
|
||||||
editor.set_error("No references found.");
|
editor.set_error("No references found.");
|
||||||
} else {
|
} else {
|
||||||
goto_impl(editor, compositor, items, offset_encoding);
|
goto_impl(editor, compositor, locations);
|
||||||
}
|
}
|
||||||
},
|
};
|
||||||
);
|
Ok(Callback::EditorCompositor(Box::new(call)))
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn signature_help(cx: &mut Context) {
|
pub fn signature_help(cx: &mut Context) {
|
||||||
|
|
|
@ -137,9 +137,12 @@ impl Compositor {
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn handle_event(&mut self, event: &Event, cx: &mut Context) -> bool {
|
pub fn handle_event(&mut self, event: &Event, cx: &mut Context) -> bool {
|
||||||
// If it is a key event and a macro is being recorded, push the key event to the recording.
|
// If it is a key event, a macro is being recorded, and a macro isn't being replayed,
|
||||||
|
// push the key event to the recording.
|
||||||
if let (Event::Key(key), Some((_, keys))) = (event, &mut cx.editor.macro_recording) {
|
if let (Event::Key(key), Some((_, keys))) = (event, &mut cx.editor.macro_recording) {
|
||||||
keys.push(*key);
|
if cx.editor.macro_replaying.is_empty() {
|
||||||
|
keys.push(*key);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut callbacks = Vec::new();
|
let mut callbacks = Vec::new();
|
||||||
|
|
|
@ -72,6 +72,11 @@ pub(crate) fn path_completion(
|
||||||
return Vec::new();
|
return Vec::new();
|
||||||
};
|
};
|
||||||
|
|
||||||
|
let edit_diff = typed_file_name
|
||||||
|
.as_ref()
|
||||||
|
.map(|s| s.chars().count())
|
||||||
|
.unwrap_or_default();
|
||||||
|
|
||||||
read_dir
|
read_dir
|
||||||
.filter_map(Result::ok)
|
.filter_map(Result::ok)
|
||||||
.filter_map(|dir_entry| {
|
.filter_map(|dir_entry| {
|
||||||
|
@ -88,11 +93,6 @@ pub(crate) fn path_completion(
|
||||||
let kind = path_kind(&md);
|
let kind = path_kind(&md);
|
||||||
let documentation = path_documentation(&md, &dir_path.join(&file_name), kind);
|
let documentation = path_documentation(&md, &dir_path.join(&file_name), kind);
|
||||||
|
|
||||||
let edit_diff = typed_file_name
|
|
||||||
.as_ref()
|
|
||||||
.map(|f| f.len())
|
|
||||||
.unwrap_or_default();
|
|
||||||
|
|
||||||
let transaction = Transaction::change_by_selection(&text, &selection, |range| {
|
let transaction = Transaction::change_by_selection(&text, &selection, |range| {
|
||||||
let cursor = range.cursor(text.slice(..));
|
let cursor = range.cursor(text.slice(..));
|
||||||
(cursor - edit_diff, cursor, Some((&file_name).into()))
|
(cursor - edit_diff, cursor, Some((&file_name).into()))
|
||||||
|
|
|
@ -5,7 +5,6 @@ use helix_core::syntax;
|
||||||
use helix_lsp::lsp;
|
use helix_lsp::lsp;
|
||||||
use helix_view::graphics::{Margin, Rect, Style};
|
use helix_view::graphics::{Margin, Rect, Style};
|
||||||
use helix_view::input::Event;
|
use helix_view::input::Event;
|
||||||
use once_cell::sync::OnceCell;
|
|
||||||
use tui::buffer::Buffer;
|
use tui::buffer::Buffer;
|
||||||
use tui::widgets::{BorderType, Paragraph, Widget, Wrap};
|
use tui::widgets::{BorderType, Paragraph, Widget, Wrap};
|
||||||
|
|
||||||
|
@ -15,11 +14,8 @@ use crate::alt;
|
||||||
use crate::ui::Markdown;
|
use crate::ui::Markdown;
|
||||||
|
|
||||||
pub struct Hover {
|
pub struct Hover {
|
||||||
hovers: Vec<(String, lsp::Hover)>,
|
|
||||||
active_index: usize,
|
active_index: usize,
|
||||||
config_loader: Arc<ArcSwap<syntax::Loader>>,
|
contents: Vec<(Option<Markdown>, Markdown)>,
|
||||||
|
|
||||||
content: OnceCell<(Option<Markdown>, Markdown)>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Hover {
|
impl Hover {
|
||||||
|
@ -29,42 +25,42 @@ impl Hover {
|
||||||
hovers: Vec<(String, lsp::Hover)>,
|
hovers: Vec<(String, lsp::Hover)>,
|
||||||
config_loader: Arc<ArcSwap<syntax::Loader>>,
|
config_loader: Arc<ArcSwap<syntax::Loader>>,
|
||||||
) -> Self {
|
) -> Self {
|
||||||
|
let n_hovers = hovers.len();
|
||||||
|
let contents = hovers
|
||||||
|
.into_iter()
|
||||||
|
.enumerate()
|
||||||
|
.map(|(idx, (server_name, hover))| {
|
||||||
|
let header = (n_hovers > 1).then(|| {
|
||||||
|
Markdown::new(
|
||||||
|
format!("**[{}/{}] {}**", idx + 1, n_hovers, server_name),
|
||||||
|
config_loader.clone(),
|
||||||
|
)
|
||||||
|
});
|
||||||
|
let body = Markdown::new(
|
||||||
|
hover_contents_to_string(hover.contents),
|
||||||
|
config_loader.clone(),
|
||||||
|
);
|
||||||
|
(header, body)
|
||||||
|
})
|
||||||
|
.collect();
|
||||||
|
|
||||||
Self {
|
Self {
|
||||||
hovers,
|
|
||||||
active_index: usize::default(),
|
active_index: usize::default(),
|
||||||
config_loader,
|
contents,
|
||||||
content: OnceCell::new(),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn has_header(&self) -> bool {
|
||||||
|
self.contents.len() > 1
|
||||||
|
}
|
||||||
|
|
||||||
fn content(&self) -> &(Option<Markdown>, Markdown) {
|
fn content(&self) -> &(Option<Markdown>, Markdown) {
|
||||||
self.content.get_or_init(|| {
|
&self.contents[self.active_index]
|
||||||
let (server_name, hover) = &self.hovers[self.active_index];
|
|
||||||
// Only render the header when there is more than one hover response.
|
|
||||||
let header = (self.hovers.len() > 1).then(|| {
|
|
||||||
Markdown::new(
|
|
||||||
format!(
|
|
||||||
"**[{}/{}] {}**",
|
|
||||||
self.active_index + 1,
|
|
||||||
self.hovers.len(),
|
|
||||||
server_name
|
|
||||||
),
|
|
||||||
self.config_loader.clone(),
|
|
||||||
)
|
|
||||||
});
|
|
||||||
let body = Markdown::new(
|
|
||||||
hover_contents_to_string(&hover.contents),
|
|
||||||
self.config_loader.clone(),
|
|
||||||
);
|
|
||||||
(header, body)
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_index(&mut self, index: usize) {
|
fn set_index(&mut self, index: usize) {
|
||||||
assert!((0..self.hovers.len()).contains(&index));
|
assert!((0..self.contents.len()).contains(&index));
|
||||||
self.active_index = index;
|
self.active_index = index;
|
||||||
// Reset the cached markdown:
|
|
||||||
self.content.take();
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -100,13 +96,11 @@ impl Component for Hover {
|
||||||
|
|
||||||
// hover content
|
// hover content
|
||||||
let contents = contents.parse(Some(&cx.editor.theme));
|
let contents = contents.parse(Some(&cx.editor.theme));
|
||||||
let contents_area = area
|
let contents_area = area.clip_top(if self.has_header() {
|
||||||
.clip_top(if self.hovers.len() > 1 {
|
HEADER_HEIGHT + SEPARATOR_HEIGHT
|
||||||
HEADER_HEIGHT + SEPARATOR_HEIGHT
|
} else {
|
||||||
} else {
|
0
|
||||||
0
|
});
|
||||||
})
|
|
||||||
.clip_bottom(u16::from(cx.editor.popup_border()));
|
|
||||||
let contents_para = Paragraph::new(&contents)
|
let contents_para = Paragraph::new(&contents)
|
||||||
.wrap(Wrap { trim: false })
|
.wrap(Wrap { trim: false })
|
||||||
.scroll((cx.scroll.unwrap_or_default() as u16, 0));
|
.scroll((cx.scroll.unwrap_or_default() as u16, 0));
|
||||||
|
@ -132,7 +126,7 @@ impl Component for Hover {
|
||||||
crate::ui::text::required_size(&contents, max_text_width);
|
crate::ui::text::required_size(&contents, max_text_width);
|
||||||
|
|
||||||
let width = PADDING_HORIZONTAL + header_width.max(content_width);
|
let width = PADDING_HORIZONTAL + header_width.max(content_width);
|
||||||
let height = if self.hovers.len() > 1 {
|
let height = if self.has_header() {
|
||||||
PADDING_TOP + HEADER_HEIGHT + SEPARATOR_HEIGHT + content_height + PADDING_BOTTOM
|
PADDING_TOP + HEADER_HEIGHT + SEPARATOR_HEIGHT + content_height + PADDING_BOTTOM
|
||||||
} else {
|
} else {
|
||||||
PADDING_TOP + content_height + PADDING_BOTTOM
|
PADDING_TOP + content_height + PADDING_BOTTOM
|
||||||
|
@ -151,12 +145,12 @@ impl Component for Hover {
|
||||||
let index = self
|
let index = self
|
||||||
.active_index
|
.active_index
|
||||||
.checked_sub(1)
|
.checked_sub(1)
|
||||||
.unwrap_or(self.hovers.len() - 1);
|
.unwrap_or(self.contents.len() - 1);
|
||||||
self.set_index(index);
|
self.set_index(index);
|
||||||
EventResult::Consumed(None)
|
EventResult::Consumed(None)
|
||||||
}
|
}
|
||||||
alt!('n') => {
|
alt!('n') => {
|
||||||
self.set_index((self.active_index + 1) % self.hovers.len());
|
self.set_index((self.active_index + 1) % self.contents.len());
|
||||||
EventResult::Consumed(None)
|
EventResult::Consumed(None)
|
||||||
}
|
}
|
||||||
_ => EventResult::Ignored(None),
|
_ => EventResult::Ignored(None),
|
||||||
|
@ -164,13 +158,13 @@ impl Component for Hover {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn hover_contents_to_string(contents: &lsp::HoverContents) -> String {
|
fn hover_contents_to_string(contents: lsp::HoverContents) -> String {
|
||||||
fn marked_string_to_markdown(contents: &lsp::MarkedString) -> String {
|
fn marked_string_to_markdown(contents: lsp::MarkedString) -> String {
|
||||||
match contents {
|
match contents {
|
||||||
lsp::MarkedString::String(contents) => contents.clone(),
|
lsp::MarkedString::String(contents) => contents,
|
||||||
lsp::MarkedString::LanguageString(string) => {
|
lsp::MarkedString::LanguageString(string) => {
|
||||||
if string.language == "markdown" {
|
if string.language == "markdown" {
|
||||||
string.value.clone()
|
string.value
|
||||||
} else {
|
} else {
|
||||||
format!("```{}\n{}\n```", string.language, string.value)
|
format!("```{}\n{}\n```", string.language, string.value)
|
||||||
}
|
}
|
||||||
|
@ -180,10 +174,10 @@ fn hover_contents_to_string(contents: &lsp::HoverContents) -> String {
|
||||||
match contents {
|
match contents {
|
||||||
lsp::HoverContents::Scalar(contents) => marked_string_to_markdown(contents),
|
lsp::HoverContents::Scalar(contents) => marked_string_to_markdown(contents),
|
||||||
lsp::HoverContents::Array(contents) => contents
|
lsp::HoverContents::Array(contents) => contents
|
||||||
.iter()
|
.into_iter()
|
||||||
.map(marked_string_to_markdown)
|
.map(marked_string_to_markdown)
|
||||||
.collect::<Vec<_>>()
|
.collect::<Vec<_>>()
|
||||||
.join("\n\n"),
|
.join("\n\n"),
|
||||||
lsp::HoverContents::Markup(contents) => contents.value.clone(),
|
lsp::HoverContents::Markup(contents) => contents.value,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -793,3 +793,30 @@ fn foo() {
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[tokio::test(flavor = "multi_thread")]
|
||||||
|
async fn macro_play_within_macro_record() -> anyhow::Result<()> {
|
||||||
|
// <https://github.com/helix-editor/helix/issues/12697>
|
||||||
|
//
|
||||||
|
// * `"aQihello<esc>Q` record a macro to register 'a' which inserts "hello"
|
||||||
|
// * `Q"aq<space>world<esc>Q` record a macro to the default macro register which plays the
|
||||||
|
// macro in register 'a' and then inserts " world"
|
||||||
|
// * `%d` clear the buffer
|
||||||
|
// * `q` replay the macro in the default macro register
|
||||||
|
// * `i<ret>` add a newline at the end
|
||||||
|
//
|
||||||
|
// The inner macro in register 'a' should replay within the outer macro exactly once to insert
|
||||||
|
// "hello world".
|
||||||
|
test((
|
||||||
|
indoc! {"\
|
||||||
|
#[|]#
|
||||||
|
"},
|
||||||
|
r#""aQihello<esc>QQ"aqi<space>world<esc>Q%dqi<ret>"#,
|
||||||
|
indoc! {"\
|
||||||
|
hello world
|
||||||
|
#[|]#"},
|
||||||
|
))
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
|
@ -20,7 +20,7 @@ helix-core = { path = "../helix-core" }
|
||||||
|
|
||||||
bitflags.workspace = true
|
bitflags.workspace = true
|
||||||
cassowary = "0.3"
|
cassowary = "0.3"
|
||||||
unicode-segmentation = "1.12"
|
unicode-segmentation.workspace = true
|
||||||
crossterm = { version = "0.28", optional = true }
|
crossterm = { version = "0.28", optional = true }
|
||||||
termini = "1.0"
|
termini = "1.0"
|
||||||
serde = { version = "1", "optional" = true, features = ["derive"]}
|
serde = { version = "1", "optional" = true, features = ["derive"]}
|
||||||
|
|
|
@ -122,10 +122,11 @@ mod external {
|
||||||
Self::Tmux
|
Self::Tmux
|
||||||
} else if binary_exists("pbcopy") && binary_exists("pbpaste") {
|
} else if binary_exists("pbcopy") && binary_exists("pbpaste") {
|
||||||
Self::Pasteboard
|
Self::Pasteboard
|
||||||
} else if cfg!(feature = "term") {
|
|
||||||
Self::Termcode
|
|
||||||
} else {
|
} else {
|
||||||
Self::None
|
#[cfg(feature = "term")]
|
||||||
|
return Self::Termcode;
|
||||||
|
#[cfg(not(feature = "term"))]
|
||||||
|
return Self::None;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -363,7 +363,7 @@ pub struct Config {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, PartialEq, Deserialize, Serialize, Eq, PartialOrd, Ord)]
|
#[derive(Debug, Clone, PartialEq, Deserialize, Serialize, Eq, PartialOrd, Ord)]
|
||||||
#[serde(rename_all = "kebab-case", default)]
|
#[serde(default, rename_all = "kebab-case", deny_unknown_fields)]
|
||||||
pub struct SmartTabConfig {
|
pub struct SmartTabConfig {
|
||||||
pub enable: bool,
|
pub enable: bool,
|
||||||
pub supersede_menu: bool,
|
pub supersede_menu: bool,
|
||||||
|
|
|
@ -342,6 +342,7 @@ impl FromStr for UnderlineStyle {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[cfg(feature = "term")]
|
||||||
impl From<UnderlineStyle> for crossterm::style::Attribute {
|
impl From<UnderlineStyle> for crossterm::style::Attribute {
|
||||||
fn from(style: UnderlineStyle) -> Self {
|
fn from(style: UnderlineStyle) -> Self {
|
||||||
match style {
|
match style {
|
||||||
|
|
|
@ -5,6 +5,7 @@ use helix_core::Selection;
|
||||||
use helix_dap::{self as dap, Client, ConnectionType, Payload, Request, ThreadId};
|
use helix_dap::{self as dap, Client, ConnectionType, Payload, Request, ThreadId};
|
||||||
use helix_lsp::block_on;
|
use helix_lsp::block_on;
|
||||||
use log::warn;
|
use log::warn;
|
||||||
|
use serde_json::json;
|
||||||
use std::fmt::Write;
|
use std::fmt::Write;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
|
@ -141,7 +142,6 @@ pub fn breakpoints_changed(
|
||||||
|
|
||||||
impl Editor {
|
impl Editor {
|
||||||
pub async fn handle_debugger_message(&mut self, payload: helix_dap::Payload) -> bool {
|
pub async fn handle_debugger_message(&mut self, payload: helix_dap::Payload) -> bool {
|
||||||
use dap::requests::RunInTerminal;
|
|
||||||
use helix_dap::{events, Event};
|
use helix_dap::{events, Event};
|
||||||
|
|
||||||
let debugger = match self.debugger.as_mut() {
|
let debugger = match self.debugger.as_mut() {
|
||||||
|
@ -149,250 +149,260 @@ impl Editor {
|
||||||
None => return false,
|
None => return false,
|
||||||
};
|
};
|
||||||
match payload {
|
match payload {
|
||||||
Payload::Event(ev) => match *ev {
|
Payload::Event(event) => {
|
||||||
Event::Stopped(events::Stopped {
|
let event = match Event::parse(&event.event, event.body) {
|
||||||
thread_id,
|
Ok(event) => event,
|
||||||
description,
|
Err(dap::Error::Unhandled) => {
|
||||||
text,
|
log::info!("Discarding unknown DAP event '{}'", event.event);
|
||||||
reason,
|
|
||||||
all_threads_stopped,
|
|
||||||
..
|
|
||||||
}) => {
|
|
||||||
let all_threads_stopped = all_threads_stopped.unwrap_or_default();
|
|
||||||
|
|
||||||
if all_threads_stopped {
|
|
||||||
if let Ok(response) = debugger.request::<dap::requests::Threads>(()).await {
|
|
||||||
for thread in response.threads {
|
|
||||||
fetch_stack_trace(debugger, thread.id).await;
|
|
||||||
}
|
|
||||||
select_thread_id(self, thread_id.unwrap_or_default(), false).await;
|
|
||||||
}
|
|
||||||
} else if let Some(thread_id) = thread_id {
|
|
||||||
debugger.thread_states.insert(thread_id, reason.clone()); // TODO: dap uses "type" || "reason" here
|
|
||||||
|
|
||||||
// whichever thread stops is made "current" (if no previously selected thread).
|
|
||||||
select_thread_id(self, thread_id, false).await;
|
|
||||||
}
|
|
||||||
|
|
||||||
let scope = match thread_id {
|
|
||||||
Some(id) => format!("Thread {}", id),
|
|
||||||
None => "Target".to_owned(),
|
|
||||||
};
|
|
||||||
|
|
||||||
let mut status = format!("{} stopped because of {}", scope, reason);
|
|
||||||
if let Some(desc) = description {
|
|
||||||
write!(status, " {}", desc).unwrap();
|
|
||||||
}
|
|
||||||
if let Some(text) = text {
|
|
||||||
write!(status, " {}", text).unwrap();
|
|
||||||
}
|
|
||||||
if all_threads_stopped {
|
|
||||||
status.push_str(" (all threads stopped)");
|
|
||||||
}
|
|
||||||
|
|
||||||
self.set_status(status);
|
|
||||||
}
|
|
||||||
Event::Continued(events::Continued { thread_id, .. }) => {
|
|
||||||
debugger
|
|
||||||
.thread_states
|
|
||||||
.insert(thread_id, "running".to_owned());
|
|
||||||
if debugger.thread_id == Some(thread_id) {
|
|
||||||
debugger.resume_application();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Event::Thread(_) => {
|
|
||||||
// TODO: update thread_states, make threads request
|
|
||||||
}
|
|
||||||
Event::Breakpoint(events::Breakpoint { reason, breakpoint }) => {
|
|
||||||
match &reason[..] {
|
|
||||||
"new" => {
|
|
||||||
if let Some(source) = breakpoint.source {
|
|
||||||
self.breakpoints
|
|
||||||
.entry(source.path.unwrap()) // TODO: no unwraps
|
|
||||||
.or_default()
|
|
||||||
.push(Breakpoint {
|
|
||||||
id: breakpoint.id,
|
|
||||||
verified: breakpoint.verified,
|
|
||||||
message: breakpoint.message,
|
|
||||||
line: breakpoint.line.unwrap().saturating_sub(1), // TODO: no unwrap
|
|
||||||
column: breakpoint.column,
|
|
||||||
..Default::default()
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
"changed" => {
|
|
||||||
for breakpoints in self.breakpoints.values_mut() {
|
|
||||||
if let Some(i) =
|
|
||||||
breakpoints.iter().position(|b| b.id == breakpoint.id)
|
|
||||||
{
|
|
||||||
breakpoints[i].verified = breakpoint.verified;
|
|
||||||
breakpoints[i].message = breakpoint
|
|
||||||
.message
|
|
||||||
.clone()
|
|
||||||
.or_else(|| breakpoints[i].message.take());
|
|
||||||
breakpoints[i].line = breakpoint
|
|
||||||
.line
|
|
||||||
.map_or(breakpoints[i].line, |line| line.saturating_sub(1));
|
|
||||||
breakpoints[i].column =
|
|
||||||
breakpoint.column.or(breakpoints[i].column);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
"removed" => {
|
|
||||||
for breakpoints in self.breakpoints.values_mut() {
|
|
||||||
if let Some(i) =
|
|
||||||
breakpoints.iter().position(|b| b.id == breakpoint.id)
|
|
||||||
{
|
|
||||||
breakpoints.remove(i);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
reason => {
|
|
||||||
warn!("Unknown breakpoint event: {}", reason);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Event::Output(events::Output {
|
|
||||||
category, output, ..
|
|
||||||
}) => {
|
|
||||||
let prefix = match category {
|
|
||||||
Some(category) => {
|
|
||||||
if &category == "telemetry" {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
format!("Debug ({}):", category)
|
|
||||||
}
|
|
||||||
None => "Debug:".to_owned(),
|
|
||||||
};
|
|
||||||
|
|
||||||
log::info!("{}", output);
|
|
||||||
self.set_status(format!("{} {}", prefix, output));
|
|
||||||
}
|
|
||||||
Event::Initialized(_) => {
|
|
||||||
// send existing breakpoints
|
|
||||||
for (path, breakpoints) in &mut self.breakpoints {
|
|
||||||
// TODO: call futures in parallel, await all
|
|
||||||
let _ = breakpoints_changed(debugger, path.clone(), breakpoints);
|
|
||||||
}
|
|
||||||
// TODO: fetch breakpoints (in case we're attaching)
|
|
||||||
|
|
||||||
if debugger.configuration_done().await.is_ok() {
|
|
||||||
self.set_status("Debugged application started");
|
|
||||||
}; // TODO: do we need to handle error?
|
|
||||||
}
|
|
||||||
Event::Terminated(terminated) => {
|
|
||||||
let restart_args = if let Some(terminated) = terminated {
|
|
||||||
terminated.restart
|
|
||||||
} else {
|
|
||||||
None
|
|
||||||
};
|
|
||||||
|
|
||||||
let disconnect_args = Some(DisconnectArguments {
|
|
||||||
restart: Some(restart_args.is_some()),
|
|
||||||
terminate_debuggee: None,
|
|
||||||
suspend_debuggee: None,
|
|
||||||
});
|
|
||||||
|
|
||||||
if let Err(err) = debugger.disconnect(disconnect_args).await {
|
|
||||||
self.set_error(format!(
|
|
||||||
"Cannot disconnect debugger upon terminated event receival {:?}",
|
|
||||||
err
|
|
||||||
));
|
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
Err(err) => {
|
||||||
|
log::warn!("Discarding invalid DAP event '{}': {err}", event.event);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
match event {
|
||||||
|
Event::Stopped(events::StoppedBody {
|
||||||
|
thread_id,
|
||||||
|
description,
|
||||||
|
text,
|
||||||
|
reason,
|
||||||
|
all_threads_stopped,
|
||||||
|
..
|
||||||
|
}) => {
|
||||||
|
let all_threads_stopped = all_threads_stopped.unwrap_or_default();
|
||||||
|
|
||||||
match restart_args {
|
if all_threads_stopped {
|
||||||
Some(restart_args) => {
|
if let Ok(response) =
|
||||||
log::info!("Attempting to restart debug session.");
|
debugger.request::<dap::requests::Threads>(()).await
|
||||||
let connection_type = match debugger.connection_type() {
|
{
|
||||||
Some(connection_type) => connection_type,
|
for thread in response.threads {
|
||||||
None => {
|
fetch_stack_trace(debugger, thread.id).await;
|
||||||
self.set_error("No starting request found, to be used in restarting the debugging session.");
|
|
||||||
return false;
|
|
||||||
}
|
}
|
||||||
};
|
select_thread_id(self, thread_id.unwrap_or_default(), false).await;
|
||||||
|
}
|
||||||
|
} else if let Some(thread_id) = thread_id {
|
||||||
|
debugger.thread_states.insert(thread_id, reason.clone()); // TODO: dap uses "type" || "reason" here
|
||||||
|
|
||||||
let relaunch_resp = if let ConnectionType::Launch = connection_type {
|
// whichever thread stops is made "current" (if no previously selected thread).
|
||||||
debugger.launch(restart_args).await
|
select_thread_id(self, thread_id, false).await;
|
||||||
} else {
|
}
|
||||||
debugger.attach(restart_args).await
|
|
||||||
};
|
|
||||||
|
|
||||||
if let Err(err) = relaunch_resp {
|
let scope = match thread_id {
|
||||||
self.set_error(format!(
|
Some(id) => format!("Thread {}", id),
|
||||||
"Failed to restart debugging session: {:?}",
|
None => "Target".to_owned(),
|
||||||
err
|
};
|
||||||
));
|
|
||||||
|
let mut status = format!("{} stopped because of {}", scope, reason);
|
||||||
|
if let Some(desc) = description {
|
||||||
|
write!(status, " {}", desc).unwrap();
|
||||||
|
}
|
||||||
|
if let Some(text) = text {
|
||||||
|
write!(status, " {}", text).unwrap();
|
||||||
|
}
|
||||||
|
if all_threads_stopped {
|
||||||
|
status.push_str(" (all threads stopped)");
|
||||||
|
}
|
||||||
|
|
||||||
|
self.set_status(status);
|
||||||
|
}
|
||||||
|
Event::Continued(events::ContinuedBody { thread_id, .. }) => {
|
||||||
|
debugger
|
||||||
|
.thread_states
|
||||||
|
.insert(thread_id, "running".to_owned());
|
||||||
|
if debugger.thread_id == Some(thread_id) {
|
||||||
|
debugger.resume_application();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Event::Thread(_) => {
|
||||||
|
// TODO: update thread_states, make threads request
|
||||||
|
}
|
||||||
|
Event::Breakpoint(events::BreakpointBody { reason, breakpoint }) => {
|
||||||
|
match &reason[..] {
|
||||||
|
"new" => {
|
||||||
|
if let Some(source) = breakpoint.source {
|
||||||
|
self.breakpoints
|
||||||
|
.entry(source.path.unwrap()) // TODO: no unwraps
|
||||||
|
.or_default()
|
||||||
|
.push(Breakpoint {
|
||||||
|
id: breakpoint.id,
|
||||||
|
verified: breakpoint.verified,
|
||||||
|
message: breakpoint.message,
|
||||||
|
line: breakpoint.line.unwrap().saturating_sub(1), // TODO: no unwrap
|
||||||
|
column: breakpoint.column,
|
||||||
|
..Default::default()
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"changed" => {
|
||||||
|
for breakpoints in self.breakpoints.values_mut() {
|
||||||
|
if let Some(i) =
|
||||||
|
breakpoints.iter().position(|b| b.id == breakpoint.id)
|
||||||
|
{
|
||||||
|
breakpoints[i].verified = breakpoint.verified;
|
||||||
|
breakpoints[i].message = breakpoint
|
||||||
|
.message
|
||||||
|
.clone()
|
||||||
|
.or_else(|| breakpoints[i].message.take());
|
||||||
|
breakpoints[i].line =
|
||||||
|
breakpoint.line.map_or(breakpoints[i].line, |line| {
|
||||||
|
line.saturating_sub(1)
|
||||||
|
});
|
||||||
|
breakpoints[i].column =
|
||||||
|
breakpoint.column.or(breakpoints[i].column);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"removed" => {
|
||||||
|
for breakpoints in self.breakpoints.values_mut() {
|
||||||
|
if let Some(i) =
|
||||||
|
breakpoints.iter().position(|b| b.id == breakpoint.id)
|
||||||
|
{
|
||||||
|
breakpoints.remove(i);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
reason => {
|
||||||
|
warn!("Unknown breakpoint event: {}", reason);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
None => {
|
}
|
||||||
self.debugger = None;
|
Event::Output(events::OutputBody {
|
||||||
self.set_status(
|
category, output, ..
|
||||||
"Terminated debugging session and disconnected debugger.",
|
}) => {
|
||||||
);
|
let prefix = match category {
|
||||||
|
Some(category) => {
|
||||||
|
if &category == "telemetry" {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
format!("Debug ({}):", category)
|
||||||
|
}
|
||||||
|
None => "Debug:".to_owned(),
|
||||||
|
};
|
||||||
|
|
||||||
|
log::info!("{}", output);
|
||||||
|
self.set_status(format!("{} {}", prefix, output));
|
||||||
|
}
|
||||||
|
Event::Initialized(_) => {
|
||||||
|
// send existing breakpoints
|
||||||
|
for (path, breakpoints) in &mut self.breakpoints {
|
||||||
|
// TODO: call futures in parallel, await all
|
||||||
|
let _ = breakpoints_changed(debugger, path.clone(), breakpoints);
|
||||||
|
}
|
||||||
|
// TODO: fetch breakpoints (in case we're attaching)
|
||||||
|
|
||||||
|
if debugger.configuration_done().await.is_ok() {
|
||||||
|
self.set_status("Debugged application started");
|
||||||
|
}; // TODO: do we need to handle error?
|
||||||
|
}
|
||||||
|
Event::Terminated(terminated) => {
|
||||||
|
let restart_args = if let Some(terminated) = terminated {
|
||||||
|
terminated.restart
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
};
|
||||||
|
|
||||||
|
let disconnect_args = Some(DisconnectArguments {
|
||||||
|
restart: Some(restart_args.is_some()),
|
||||||
|
terminate_debuggee: None,
|
||||||
|
suspend_debuggee: None,
|
||||||
|
});
|
||||||
|
|
||||||
|
if let Err(err) = debugger.disconnect(disconnect_args).await {
|
||||||
|
self.set_error(format!(
|
||||||
|
"Cannot disconnect debugger upon terminated event receival {:?}",
|
||||||
|
err
|
||||||
|
));
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
match restart_args {
|
||||||
|
Some(restart_args) => {
|
||||||
|
log::info!("Attempting to restart debug session.");
|
||||||
|
let connection_type = match debugger.connection_type() {
|
||||||
|
Some(connection_type) => connection_type,
|
||||||
|
None => {
|
||||||
|
self.set_error("No starting request found, to be used in restarting the debugging session.");
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
let relaunch_resp = if let ConnectionType::Launch = connection_type
|
||||||
|
{
|
||||||
|
debugger.launch(restart_args).await
|
||||||
|
} else {
|
||||||
|
debugger.attach(restart_args).await
|
||||||
|
};
|
||||||
|
|
||||||
|
if let Err(err) = relaunch_resp {
|
||||||
|
self.set_error(format!(
|
||||||
|
"Failed to restart debugging session: {:?}",
|
||||||
|
err
|
||||||
|
));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
self.debugger = None;
|
||||||
|
self.set_status(
|
||||||
|
"Terminated debugging session and disconnected debugger.",
|
||||||
|
);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
Event::Exited(resp) => {
|
||||||
Event::Exited(resp) => {
|
let exit_code = resp.exit_code;
|
||||||
let exit_code = resp.exit_code;
|
if exit_code != 0 {
|
||||||
if exit_code != 0 {
|
self.set_error(format!(
|
||||||
self.set_error(format!(
|
"Debuggee failed to exit successfully (exit code: {exit_code})."
|
||||||
"Debuggee failed to exit successfully (exit code: {exit_code})."
|
));
|
||||||
));
|
}
|
||||||
|
}
|
||||||
|
ev => {
|
||||||
|
log::warn!("Unhandled event {:?}", ev);
|
||||||
|
return false; // return early to skip render
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
ev => {
|
}
|
||||||
log::warn!("Unhandled event {:?}", ev);
|
|
||||||
return false; // return early to skip render
|
|
||||||
}
|
|
||||||
},
|
|
||||||
Payload::Response(_) => unreachable!(),
|
Payload::Response(_) => unreachable!(),
|
||||||
Payload::Request(request) => match request.command.as_str() {
|
Payload::Request(request) => {
|
||||||
RunInTerminal::COMMAND => {
|
let reply = match Request::parse(&request.command, request.arguments) {
|
||||||
let arguments: dap::requests::RunInTerminalArguments =
|
Ok(Request::RunInTerminal(arguments)) => {
|
||||||
serde_json::from_value(request.arguments.unwrap_or_default()).unwrap();
|
let config = self.config();
|
||||||
// TODO: no unwrap
|
let Some(config) = config.terminal.as_ref() else {
|
||||||
|
|
||||||
let config = match self.config().terminal.clone() {
|
|
||||||
Some(config) => config,
|
|
||||||
None => {
|
|
||||||
self.set_error("No external terminal defined");
|
self.set_error("No external terminal defined");
|
||||||
return true;
|
return true;
|
||||||
}
|
};
|
||||||
};
|
|
||||||
|
|
||||||
// Re-borrowing debugger to avoid issues when loading config
|
let process = match std::process::Command::new(&config.command)
|
||||||
let debugger = match self.debugger.as_mut() {
|
.args(&config.args)
|
||||||
Some(debugger) => debugger,
|
.arg(arguments.args.join(" "))
|
||||||
None => return false,
|
.spawn()
|
||||||
};
|
{
|
||||||
|
Ok(process) => process,
|
||||||
|
Err(err) => {
|
||||||
|
self.set_error(format!(
|
||||||
|
"Error starting external terminal: {}",
|
||||||
|
err
|
||||||
|
));
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
let process = match std::process::Command::new(config.command)
|
Ok(json!(dap::requests::RunInTerminalResponse {
|
||||||
.args(config.args)
|
process_id: Some(process.id()),
|
||||||
.arg(arguments.args.join(" "))
|
shell_process_id: None,
|
||||||
.spawn()
|
}))
|
||||||
{
|
}
|
||||||
Ok(process) => process,
|
Err(err) => Err(err),
|
||||||
Err(err) => {
|
};
|
||||||
self.set_error(format!("Error starting external terminal: {}", err));
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let _ = debugger
|
if let Some(debugger) = self.debugger.as_mut() {
|
||||||
.reply(
|
debugger
|
||||||
request.seq,
|
.reply(request.seq, &request.command, reply)
|
||||||
dap::requests::RunInTerminal::COMMAND,
|
.await
|
||||||
serde_json::to_value(dap::requests::RunInTerminalResponse {
|
.ok();
|
||||||
process_id: Some(process.id()),
|
|
||||||
shell_process_id: None,
|
|
||||||
})
|
|
||||||
.map_err(|e| e.into()),
|
|
||||||
)
|
|
||||||
.await;
|
|
||||||
}
|
}
|
||||||
_ => log::error!("DAP reverse request not implemented: {:?}", request),
|
}
|
||||||
},
|
|
||||||
}
|
}
|
||||||
true
|
true
|
||||||
}
|
}
|
||||||
|
|
|
@ -1334,6 +1334,8 @@ scope = "source.svelte"
|
||||||
injection-regex = "svelte"
|
injection-regex = "svelte"
|
||||||
file-types = ["svelte"]
|
file-types = ["svelte"]
|
||||||
indent = { tab-width = 2, unit = " " }
|
indent = { tab-width = 2, unit = " " }
|
||||||
|
comment-token = "//"
|
||||||
|
block-comment-tokens = { start = "/*", end = "*/" }
|
||||||
language-servers = [ "svelteserver" ]
|
language-servers = [ "svelteserver" ]
|
||||||
|
|
||||||
[[grammar]]
|
[[grammar]]
|
||||||
|
@ -2199,7 +2201,7 @@ language-servers = [ "cairo-language-server" ]
|
||||||
|
|
||||||
[[grammar]]
|
[[grammar]]
|
||||||
name = "cairo"
|
name = "cairo"
|
||||||
source = { git = "https://github.com/starkware-libs/tree-sitter-cairo", rev = "e3a0212261c125cb38248458cd856c0ffee2b398" }
|
source = { git = "https://github.com/starkware-libs/tree-sitter-cairo", rev = "4c6a25680546761b80a710ead1dd34e76c203125" }
|
||||||
|
|
||||||
[[language]]
|
[[language]]
|
||||||
name = "cpon"
|
name = "cpon"
|
||||||
|
@ -2829,7 +2831,6 @@ file-types = [
|
||||||
"network",
|
"network",
|
||||||
{ glob = ".editorconfig" },
|
{ glob = ".editorconfig" },
|
||||||
{ glob = ".npmrc" },
|
{ glob = ".npmrc" },
|
||||||
{ glob = "ghostty/config" },
|
|
||||||
{ glob = "hgrc" },
|
{ glob = "hgrc" },
|
||||||
{ glob = "npmrc" },
|
{ glob = "npmrc" },
|
||||||
{ glob = "rclone.conf" },
|
{ glob = "rclone.conf" },
|
||||||
|
@ -4087,3 +4088,14 @@ file-types = ["multicursor"]
|
||||||
[[grammar]]
|
[[grammar]]
|
||||||
name = "multicursor"
|
name = "multicursor"
|
||||||
source = { git = "https://github.com/nik-rev/tree-sitter-multicursor", rev = "f6f868ecfe3de2fb2cd815cb40be55493f699a49" }
|
source = { git = "https://github.com/nik-rev/tree-sitter-multicursor", rev = "f6f868ecfe3de2fb2cd815cb40be55493f699a49" }
|
||||||
|
|
||||||
|
[[language]]
|
||||||
|
name = "ghostty"
|
||||||
|
scope = "source.ghostty"
|
||||||
|
file-types = [{ glob = "ghostty/config" }]
|
||||||
|
comment-tokens = "#"
|
||||||
|
indent = { tab-width = 2, unit = " " }
|
||||||
|
|
||||||
|
[[grammar]]
|
||||||
|
name = "ghostty"
|
||||||
|
source = { git = "https://github.com/bezhermoso/tree-sitter-ghostty" , rev = "8438a93b44367e962b2ea3a3b6511885bebd196a" }
|
||||||
|
|
|
@ -71,6 +71,8 @@
|
||||||
"<"
|
"<"
|
||||||
">"
|
">"
|
||||||
] @punctuation.bracket)
|
] @punctuation.bracket)
|
||||||
|
(closure_parameters
|
||||||
|
"|" @punctuation.bracket)
|
||||||
|
|
||||||
; ---
|
; ---
|
||||||
; Variables
|
; Variables
|
||||||
|
@ -92,6 +94,8 @@
|
||||||
(parameter
|
(parameter
|
||||||
pattern: (identifier) @variable.parameter)
|
pattern: (identifier) @variable.parameter)
|
||||||
|
|
||||||
|
(closure_parameters
|
||||||
|
(identifier) @variable.parameter)
|
||||||
; -------
|
; -------
|
||||||
; Keywords
|
; Keywords
|
||||||
; -------
|
; -------
|
||||||
|
|
|
@ -115,6 +115,16 @@
|
||||||
(#not-same-line? @expr-start @pattern-guard)
|
(#not-same-line? @expr-start @pattern-guard)
|
||||||
) @indent
|
) @indent
|
||||||
|
|
||||||
|
; Align closure parameters if they span more than one line
|
||||||
|
(closure_parameters
|
||||||
|
"|"
|
||||||
|
.
|
||||||
|
(_) @anchor
|
||||||
|
(_) @expr-end
|
||||||
|
.
|
||||||
|
(#not-same-line? @anchor @expr-end)
|
||||||
|
) @align
|
||||||
|
|
||||||
(for_expression
|
(for_expression
|
||||||
"in" @in
|
"in" @in
|
||||||
.
|
.
|
||||||
|
|
|
@ -7,6 +7,7 @@
|
||||||
(type_item)
|
(type_item)
|
||||||
(trait_item)
|
(trait_item)
|
||||||
(impl_item)
|
(impl_item)
|
||||||
|
(closure_expression)
|
||||||
(block)
|
(block)
|
||||||
] @local.scope
|
] @local.scope
|
||||||
|
|
||||||
|
@ -20,6 +21,8 @@
|
||||||
(constrained_type_parameter
|
(constrained_type_parameter
|
||||||
left: (type_identifier) @local.definition)
|
left: (type_identifier) @local.definition)
|
||||||
|
|
||||||
|
(closure_parameters (identifier) @local.definition)
|
||||||
|
|
||||||
; References
|
; References
|
||||||
(identifier) @local.reference
|
(identifier) @local.reference
|
||||||
(type_identifier) @local.reference
|
(type_identifier) @local.reference
|
||||||
|
|
|
@ -1,6 +1,9 @@
|
||||||
(function_item
|
(function_item
|
||||||
body: (_) @function.inside) @function.around
|
body: (_) @function.inside) @function.around
|
||||||
|
|
||||||
|
(closure_expression
|
||||||
|
body: (_) @function.inside) @function.around
|
||||||
|
|
||||||
(struct_item
|
(struct_item
|
||||||
body: (_) @class.inside) @class.around
|
body: (_) @class.inside) @class.around
|
||||||
|
|
||||||
|
@ -16,6 +19,9 @@
|
||||||
(parameters
|
(parameters
|
||||||
((_) @parameter.inside . ","? @parameter.around) @parameter.around)
|
((_) @parameter.inside . ","? @parameter.around) @parameter.around)
|
||||||
|
|
||||||
|
(closure_parameters
|
||||||
|
((_) @parameter.inside . ","? @parameter.around) @parameter.around)
|
||||||
|
|
||||||
(type_parameters
|
(type_parameters
|
||||||
((_) @parameter.inside . ","? @parameter.around) @parameter.around)
|
((_) @parameter.inside . ","? @parameter.around) @parameter.around)
|
||||||
|
|
||||||
|
|
68
runtime/queries/ghostty/highlights.scm
Normal file
68
runtime/queries/ghostty/highlights.scm
Normal file
|
@ -0,0 +1,68 @@
|
||||||
|
; extends
|
||||||
|
; Comments
|
||||||
|
(comment) @comment
|
||||||
|
|
||||||
|
; Keys
|
||||||
|
(property) @variable
|
||||||
|
|
||||||
|
; Values
|
||||||
|
(boolean) @constant.builtin.boolean
|
||||||
|
|
||||||
|
|
||||||
|
[
|
||||||
|
(number)
|
||||||
|
(adjustment)
|
||||||
|
] @constant.numeric
|
||||||
|
|
||||||
|
[
|
||||||
|
"+"
|
||||||
|
"="
|
||||||
|
(keybind_trigger ">")
|
||||||
|
] @operator
|
||||||
|
|
||||||
|
(":") @punctuation.delimiter
|
||||||
|
|
||||||
|
; (color) are hex values
|
||||||
|
(color "#" @punctuation.special
|
||||||
|
(#eq? @punctuation.special "#"))
|
||||||
|
|
||||||
|
(path_value "?" @keyword.control.conditional
|
||||||
|
(#eq? @keyword.control.conditional "?"))
|
||||||
|
|
||||||
|
; `palette`
|
||||||
|
(palette_index) @variable.other.member
|
||||||
|
|
||||||
|
; `path_directive`
|
||||||
|
(path_directive (property) @keyword.import)
|
||||||
|
(path_directive (path_value (string) @string.special.path ))
|
||||||
|
|
||||||
|
|
||||||
|
(action_name) @function.builtin
|
||||||
|
(action_argument (string) @variable.parameter )
|
||||||
|
|
||||||
|
; (tuple)
|
||||||
|
(tuple "," @punctuation.delimiter.special
|
||||||
|
(#eq? @punctuation.delimiter.special ","))
|
||||||
|
|
||||||
|
[
|
||||||
|
(string)
|
||||||
|
(color)
|
||||||
|
] @string
|
||||||
|
|
||||||
|
; clear is a special keyword that clear all existing keybind up to that point
|
||||||
|
((keybind_value) @keyword
|
||||||
|
(#eq? @keyword "clear"))
|
||||||
|
|
||||||
|
; `keybind`
|
||||||
|
(keybind_value) @string.special
|
||||||
|
|
||||||
|
; NOTE: The order here matters!
|
||||||
|
[
|
||||||
|
(key_qualifier)
|
||||||
|
(keybind_modifier)
|
||||||
|
] @attribute
|
||||||
|
|
||||||
|
[
|
||||||
|
(modifier_key)
|
||||||
|
(key)
|
||||||
|
] @constant.builtin
|
|
@ -203,6 +203,13 @@
|
||||||
(builtin_identifier) @keyword.control.import
|
(builtin_identifier) @keyword.control.import
|
||||||
(#any-of? @keyword.control.import "@import" "@cImport")))
|
(#any-of? @keyword.control.import "@import" "@cImport")))
|
||||||
|
|
||||||
|
(variable_declaration
|
||||||
|
(identifier) @variable ; TODO: module
|
||||||
|
(field_expression
|
||||||
|
object: (builtin_function
|
||||||
|
(builtin_identifier) @keyword.control.import
|
||||||
|
(#any-of? @keyword.control.import "@import" "@cImport"))))
|
||||||
|
|
||||||
; Functions
|
; Functions
|
||||||
|
|
||||||
(call_expression
|
(call_expression
|
||||||
|
@ -223,6 +230,10 @@
|
||||||
.
|
.
|
||||||
(identifier) @variable.other.member)
|
(identifier) @variable.other.member)
|
||||||
|
|
||||||
|
(field_expression
|
||||||
|
(_)
|
||||||
|
member: (identifier) @type (#match? @type "^[A-Z_][a-zA-Z0-9_]*"))
|
||||||
|
|
||||||
(field_expression
|
(field_expression
|
||||||
(_)
|
(_)
|
||||||
member: (identifier) @variable.other.member)
|
member: (identifier) @variable.other.member)
|
||||||
|
@ -266,7 +277,7 @@
|
||||||
type: (identifier) @type)
|
type: (identifier) @type)
|
||||||
|
|
||||||
((identifier) @type
|
((identifier) @type
|
||||||
(#lua-match? @type "^[A-Z_][a-zA-Z0-9_]*"))
|
(#match? @type "^[A-Z_][a-zA-Z0-9_]*"))
|
||||||
|
|
||||||
(variable_declaration
|
(variable_declaration
|
||||||
(identifier) @type
|
(identifier) @type
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue