mirror of
https://github.com/helix-editor/helix.git
synced 2025-04-03 19:07:44 +03:00
Create helix-stdx crate for stdlib extensions
helix-stdx is meant to carry extensions to the stdlib or low-level dependencies that are useful in all other crates. This commit starts with all of the path functions from helix-core and the CWD tracking that lived in helix-loader. The CWD tracking in helix-loader was previously unable to call the canonicalization functions in helix-core. Switching to our custom canonicalization code should make no noticeable difference though since `std::env::current_dir` returns a canonicalized path with symlinks resolved (at least on unix).
This commit is contained in:
parent
af8e524a7d
commit
1f916e65cf
27 changed files with 163 additions and 111 deletions
|
@ -3,7 +3,8 @@ use std::{
|
|||
ops::RangeInclusive,
|
||||
};
|
||||
|
||||
use helix_core::{diagnostic::Severity, path::get_normalized_path};
|
||||
use helix_core::diagnostic::Severity;
|
||||
use helix_stdx::path;
|
||||
use helix_view::doc;
|
||||
|
||||
use super::*;
|
||||
|
@ -23,7 +24,7 @@ async fn test_write_quit_fail() -> anyhow::Result<()> {
|
|||
assert_eq!(1, docs.len());
|
||||
|
||||
let doc = docs.pop().unwrap();
|
||||
assert_eq!(Some(&get_normalized_path(file.path())), doc.path());
|
||||
assert_eq!(Some(&path::normalize(file.path())), doc.path());
|
||||
assert_eq!(&Severity::Error, app.editor.get_status().unwrap().1);
|
||||
}),
|
||||
false,
|
||||
|
@ -269,7 +270,7 @@ async fn test_write_scratch_to_new_path() -> anyhow::Result<()> {
|
|||
assert_eq!(1, docs.len());
|
||||
|
||||
let doc = docs.pop().unwrap();
|
||||
assert_eq!(Some(&get_normalized_path(file.path())), doc.path());
|
||||
assert_eq!(Some(&path::normalize(file.path())), doc.path());
|
||||
}),
|
||||
false,
|
||||
)
|
||||
|
@ -341,7 +342,7 @@ async fn test_write_new_path() -> anyhow::Result<()> {
|
|||
Some(&|app| {
|
||||
let doc = doc!(app.editor);
|
||||
assert!(!app.editor.is_err());
|
||||
assert_eq!(&get_normalized_path(file1.path()), doc.path().unwrap());
|
||||
assert_eq!(&path::normalize(file1.path()), doc.path().unwrap());
|
||||
}),
|
||||
),
|
||||
(
|
||||
|
@ -349,7 +350,7 @@ async fn test_write_new_path() -> anyhow::Result<()> {
|
|||
Some(&|app| {
|
||||
let doc = doc!(app.editor);
|
||||
assert!(!app.editor.is_err());
|
||||
assert_eq!(&get_normalized_path(file2.path()), doc.path().unwrap());
|
||||
assert_eq!(&path::normalize(file2.path()), doc.path().unwrap());
|
||||
assert!(app.editor.document_by_path(file1.path()).is_none());
|
||||
}),
|
||||
),
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
use super::*;
|
||||
|
||||
use helix_core::path::get_normalized_path;
|
||||
use helix_stdx::path;
|
||||
|
||||
#[tokio::test(flavor = "multi_thread")]
|
||||
async fn test_split_write_quit_all() -> anyhow::Result<()> {
|
||||
|
@ -27,21 +27,21 @@ async fn test_split_write_quit_all() -> anyhow::Result<()> {
|
|||
|
||||
let doc1 = docs
|
||||
.iter()
|
||||
.find(|doc| doc.path().unwrap() == &get_normalized_path(file1.path()))
|
||||
.find(|doc| doc.path().unwrap() == &path::normalize(file1.path()))
|
||||
.unwrap();
|
||||
|
||||
assert_eq!("hello1", doc1.text().to_string());
|
||||
|
||||
let doc2 = docs
|
||||
.iter()
|
||||
.find(|doc| doc.path().unwrap() == &get_normalized_path(file2.path()))
|
||||
.find(|doc| doc.path().unwrap() == &path::normalize(file2.path()))
|
||||
.unwrap();
|
||||
|
||||
assert_eq!("hello2", doc2.text().to_string());
|
||||
|
||||
let doc3 = docs
|
||||
.iter()
|
||||
.find(|doc| doc.path().unwrap() == &get_normalized_path(file3.path()))
|
||||
.find(|doc| doc.path().unwrap() == &path::normalize(file3.path()))
|
||||
.unwrap();
|
||||
|
||||
assert_eq!("hello3", doc3.text().to_string());
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue