mirror of
https://github.com/Kozea/Radicale.git
synced 2025-04-04 21:57:43 +03:00
make flake8 happy
This commit is contained in:
parent
edd6d0a513
commit
24f5f9b98e
5 changed files with 9 additions and 15 deletions
|
@ -91,5 +91,5 @@ class Storage(
|
||||||
def __init__(self, configuration: config.Configuration) -> None:
|
def __init__(self, configuration: config.Configuration) -> None:
|
||||||
super().__init__(configuration)
|
super().__init__(configuration)
|
||||||
self._makedirs_synced(self._filesystem_folder)
|
self._makedirs_synced(self._filesystem_folder)
|
||||||
logger.info("storage location: %r", self._filesystem_folder);
|
logger.info("storage location: %r", self._filesystem_folder)
|
||||||
logger.info("storage cache subfolder usage for item: %s", self._use_cache_subfolder_for_item);
|
logger.info("storage cache subfolder usage for item: %s", self._use_cache_subfolder_for_item)
|
||||||
|
|
|
@ -85,7 +85,7 @@ class StorageBase(storage.BaseStorage):
|
||||||
return os.path.join(self._filesystem_folder, "collection-root")
|
return os.path.join(self._filesystem_folder, "collection-root")
|
||||||
|
|
||||||
def _get_collection_cache_folder(self, path, folder, subfolder) -> str:
|
def _get_collection_cache_folder(self, path, folder, subfolder) -> str:
|
||||||
if self._use_cache_subfolder_for_item == True and subfolder == "item":
|
if (self._use_cache_subfolder_for_item is True) and (subfolder == "item"):
|
||||||
path = path.replace(os.path.join(self._filesystem_folder, "collection-root"), os.path.join(self._filesystem_folder, "collection-cache"))
|
path = path.replace(os.path.join(self._filesystem_folder, "collection-root"), os.path.join(self._filesystem_folder, "collection-cache"))
|
||||||
return os.path.join(path, folder, subfolder)
|
return os.path.join(path, folder, subfolder)
|
||||||
|
|
||||||
|
|
|
@ -82,8 +82,7 @@ class CollectionPartCache(CollectionBase):
|
||||||
if not cache_hash:
|
if not cache_hash:
|
||||||
cache_hash = self._item_cache_hash(
|
cache_hash = self._item_cache_hash(
|
||||||
item.serialize().encode(self._encoding))
|
item.serialize().encode(self._encoding))
|
||||||
cache_folder = self._storage._get_collection_cache_folder(self._filesystem_path, ".Radicale.cache",
|
cache_folder = self._storage._get_collection_cache_folder(self._filesystem_path, ".Radicale.cache", "item")
|
||||||
"item")
|
|
||||||
content = self._item_cache_content(item)
|
content = self._item_cache_content(item)
|
||||||
self._storage._makedirs_synced(cache_folder)
|
self._storage._makedirs_synced(cache_folder)
|
||||||
# Race: Other processes might have created and locked the file.
|
# Race: Other processes might have created and locked the file.
|
||||||
|
@ -96,8 +95,7 @@ class CollectionPartCache(CollectionBase):
|
||||||
|
|
||||||
def _load_item_cache(self, href: str, cache_hash: str
|
def _load_item_cache(self, href: str, cache_hash: str
|
||||||
) -> Optional[CacheContent]:
|
) -> Optional[CacheContent]:
|
||||||
cache_folder = self._storage._get_collection_cache_folder(self._filesystem_path, ".Radicale.cache",
|
cache_folder = self._storage._get_collection_cache_folder(self._filesystem_path, ".Radicale.cache", "item")
|
||||||
"item")
|
|
||||||
try:
|
try:
|
||||||
with open(os.path.join(cache_folder, href), "rb") as f:
|
with open(os.path.join(cache_folder, href), "rb") as f:
|
||||||
hash_, *remainder = pickle.load(f)
|
hash_, *remainder = pickle.load(f)
|
||||||
|
@ -111,8 +109,7 @@ class CollectionPartCache(CollectionBase):
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def _clean_item_cache(self) -> None:
|
def _clean_item_cache(self) -> None:
|
||||||
cache_folder = self._storage._get_collection_cache_folder(self._filesystem_path, ".Radicale.cache",
|
cache_folder = self._storage._get_collection_cache_folder(self._filesystem_path, ".Radicale.cache", "item")
|
||||||
"item")
|
|
||||||
self._clean_cache(cache_folder, (
|
self._clean_cache(cache_folder, (
|
||||||
e.name for e in os.scandir(cache_folder) if not
|
e.name for e in os.scandir(cache_folder) if not
|
||||||
os.path.isfile(os.path.join(self._filesystem_path, e.name))))
|
os.path.isfile(os.path.join(self._filesystem_path, e.name))))
|
||||||
|
|
|
@ -42,10 +42,8 @@ class StoragePartMove(StorageBase):
|
||||||
if item.collection._filesystem_path != to_collection._filesystem_path:
|
if item.collection._filesystem_path != to_collection._filesystem_path:
|
||||||
self._sync_directory(item.collection._filesystem_path)
|
self._sync_directory(item.collection._filesystem_path)
|
||||||
# Move the item cache entry
|
# Move the item cache entry
|
||||||
cache_folder = self._get_collection_cache_folder(item.collection._filesystem_path,
|
cache_folder = self._get_collection_cache_folder(item.collection._filesystem_path, ".Radicale.cache", "item")
|
||||||
".Radicale.cache", "item")
|
to_cache_folder = self._get_collection_cache_folder(to_collection._filesystem_path, ".Radicale.cache", "item")
|
||||||
to_cache_folder = self._get_collection_cache_folder(to_collection._filesystem_path,
|
|
||||||
".Radicale.cache", "item")
|
|
||||||
self._makedirs_synced(to_cache_folder)
|
self._makedirs_synced(to_cache_folder)
|
||||||
try:
|
try:
|
||||||
os.replace(os.path.join(cache_folder, item.href),
|
os.replace(os.path.join(cache_folder, item.href),
|
||||||
|
|
|
@ -76,8 +76,7 @@ class CollectionPartUpload(CollectionPartGet, CollectionPartCache,
|
||||||
yield radicale_item.find_available_uid(
|
yield radicale_item.find_available_uid(
|
||||||
lambda href: not is_safe_free_href(href), suffix)
|
lambda href: not is_safe_free_href(href), suffix)
|
||||||
|
|
||||||
cache_folder = self._storage._get_collection_cache_folder(self._filesystem_path,
|
cache_folder = self._storage._get_collection_cache_folder(self._filesystem_path, ".Radicale.cache", "item")
|
||||||
".Radicale.cache", "item")
|
|
||||||
self._storage._makedirs_synced(cache_folder)
|
self._storage._makedirs_synced(cache_folder)
|
||||||
for item in items:
|
for item in items:
|
||||||
uid = item.uid
|
uid = item.uid
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue