implement PathRouterSyncer for file-based diff management, make RouterBuilder type ChainRouter
This commit is contained in:
parent
099ebad566
commit
16e2e94e2a
@ -1,5 +1,6 @@
|
|||||||
from execlog import util
|
from execlog import util
|
||||||
from execlog import routers
|
from execlog import routers
|
||||||
|
from execlog import syncers
|
||||||
from execlog import listeners
|
from execlog import listeners
|
||||||
|
|
||||||
from execlog.server import Server
|
from execlog.server import Server
|
||||||
|
@ -159,9 +159,13 @@ class PathListener(Listener[FileEvent]):
|
|||||||
logger.info(f'> Listening on path {path} for flags {iflags.from_mask(flags)}')
|
logger.info(f'> Listening on path {path} for flags {iflags.from_mask(flags)}')
|
||||||
|
|
||||||
for (callback, pattern, debounce, delay, *_) in self.router.routemap[path]:
|
for (callback, pattern, debounce, delay, *_) in self.router.routemap[path]:
|
||||||
|
callback_name = str(callback)
|
||||||
|
if hasattr(callback, '__name__'):
|
||||||
|
callback_name = callback.__name__
|
||||||
|
|
||||||
logger.info(
|
logger.info(
|
||||||
color_text(
|
color_text(
|
||||||
f'| > {pattern} -> {callback.__name__} (debounce {debounce}ms, delay {delay}ms)',
|
f'| > {pattern} -> {callback_name} (debounce {debounce}ms, delay {delay}ms)',
|
||||||
Style.DIM,
|
Style.DIM,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
@ -315,19 +319,17 @@ class PathListener(Listener[FileEvent]):
|
|||||||
If ``handle_events`` is called externally, note that this loop will block in the
|
If ``handle_events`` is called externally, note that this loop will block in the
|
||||||
calling thread until the jobs have been submitted. It will *not* block until
|
calling thread until the jobs have been submitted. It will *not* block until
|
||||||
jobs have completed, however, as a list of futures is returned. The calling
|
jobs have completed, however, as a list of futures is returned. The calling
|
||||||
Watcher instance may have already been started, in which case ``run()`` will
|
Listener instance may have already been started, in which case ``run()`` will
|
||||||
already be executing in a separate thread. Calling this method externally will
|
already be executing in a separate thread. Calling this method externally will
|
||||||
not interfere with this loop insofar as it adds jobs to the same thread pool.
|
not interfere with this loop insofar as it adds jobs to the same thread pool.
|
||||||
|
|
||||||
Because this method only submits jobs associated with the provided ``events``,
|
Because this method only submits jobs associated with the provided ``events``,
|
||||||
the calling thread can await the returned list of futures and be confident
|
the calling thread can await the returned list of futures and be confident
|
||||||
that top-level callbacks associated with these file events have completed. Do
|
that top-level callbacks associated with these file events have completed. Do
|
||||||
note that, if the Watcher has already been started, any propagating file
|
note that, if the Listener has already been started, any propagating file
|
||||||
events will be picked up and possibly processed simultaneously (although their
|
events will be picked up and possibly processed simultaneously (although their
|
||||||
associated callbacks will have nothing to do with the returned list of futures).
|
associated callbacks will have nothing to do with the returned list of futures).
|
||||||
'''
|
'''
|
||||||
from execlog.router import Event
|
|
||||||
|
|
||||||
for event in events:
|
for event in events:
|
||||||
# hard coded ignores
|
# hard coded ignores
|
||||||
if util.path.glob_match(event.name, util.path.IGNORE_PATTERNS): continue
|
if util.path.glob_match(event.name, util.path.IGNORE_PATTERNS): continue
|
||||||
@ -359,7 +361,7 @@ class PathListener(Listener[FileEvent]):
|
|||||||
|
|
||||||
logger.debug(f'Watcher fired for [{relpath}]: {mask_flags}')
|
logger.debug(f'Watcher fired for [{relpath}]: {mask_flags}')
|
||||||
|
|
||||||
route_event = Event(endpoint=str(path), name=str(relpath), action=mask_flags)
|
route_event = FileEvent(endpoint=str(path), name=str(relpath), action=mask_flags)
|
||||||
self.router.submit(route_event)
|
self.router.submit(route_event)
|
||||||
|
|
||||||
# handle renamed directories; old dir was being watched if these flags
|
# handle renamed directories; old dir was being watched if these flags
|
||||||
|
@ -592,7 +592,7 @@ class RouteRegistryMeta(type):
|
|||||||
|
|
||||||
return super().__new__(cls, name, bases, attrs)
|
return super().__new__(cls, name, bases, attrs)
|
||||||
|
|
||||||
class RouterBuilder(metaclass=RouteRegistryMeta):
|
class RouterBuilder(ChainRouter, metaclass=RouteRegistryMeta):
|
||||||
'''
|
'''
|
||||||
Builds a (Chain)Router using attached methods and passed options.
|
Builds a (Chain)Router using attached methods and passed options.
|
||||||
|
|
||||||
@ -656,9 +656,11 @@ class RouterBuilder(metaclass=RouteRegistryMeta):
|
|||||||
register_map: dict[str, tuple[Router, dict[str, tuple[tuple[str, str], dict[str, Any]]]]],
|
register_map: dict[str, tuple[Router, dict[str, tuple[tuple[str, str], dict[str, Any]]]]],
|
||||||
):
|
):
|
||||||
self.register_map = register_map
|
self.register_map = register_map
|
||||||
|
routers = []
|
||||||
|
|
||||||
# register
|
# register
|
||||||
for router_name, (router, router_options) in self.register_map.items():
|
for router_name, (router, router_options) in self.register_map.items():
|
||||||
|
routers.append(router)
|
||||||
for route_group, method_arg_list in self.route_registry[router_name].items():
|
for route_group, method_arg_list in self.route_registry[router_name].items():
|
||||||
# get post-callbacks for reserved key "post"
|
# get post-callbacks for reserved key "post"
|
||||||
# assumed no kwargs for passthrough
|
# assumed no kwargs for passthrough
|
||||||
@ -686,6 +688,10 @@ class RouterBuilder(metaclass=RouteRegistryMeta):
|
|||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
def get_router(self, router_key_list: list[str]):
|
super().__init__(routers)
|
||||||
return ChainRouter([self.register_map[k][0] for k in router_key_list])
|
|
||||||
|
# -- disabling for now to inherit from ChainRouter directly. Require the order to
|
||||||
|
# -- simply be specified by the order of the router keys in the register_map
|
||||||
|
# def get_router(self, router_key_list: list[str]):
|
||||||
|
# return ChainRouter([self.register_map[k][0] for k in router_key_list])
|
||||||
|
|
||||||
|
@ -46,20 +46,36 @@ class PathRouter(Router[FileEvent]):
|
|||||||
|
|
||||||
def filter(self, event, glob, **listen_kwargs) -> bool:
|
def filter(self, event, glob, **listen_kwargs) -> bool:
|
||||||
'''
|
'''
|
||||||
Note:
|
Filter path events based on the provided glob pattern and listen arguments.
|
||||||
If ``handle_events`` is called externally, note that this loop will block in the
|
|
||||||
calling thread until the jobs have been submitted. It will _not_ block until
|
|
||||||
jobs have completed, however, as a list of futures is returned. The calling
|
|
||||||
Watcher instance may have already been started, in which case ``run()`` will
|
|
||||||
already be executing in a separate thread. Calling this method externally will
|
|
||||||
not interfere with this loop insofar as it adds jobs to the same thread pool.
|
|
||||||
|
|
||||||
Because this method only submits jobs associated with the provided ``events``,
|
This method is needed due to the lack of granularity when you have separate router
|
||||||
the calling thread can await the returned list of futures and be confident
|
callbacks that listen to the same directory (or overlap on some nested directory
|
||||||
that top-level callbacks associated with these file events have completed. Do
|
therein) with *different listen flags*. The overlapping path in question will only
|
||||||
note that, if the Watcher has already been started, any propagating file
|
ever be assigned a single watch descriptor by iNotify, but will (or at least appears
|
||||||
events will be picked up and possibly process simultaneously (although their
|
to) add (via bitwise OR) new flags if the same path is registered. Thus, an event
|
||||||
associated callbacks will have nothing to do with the return list of futures).
|
fired by iNotify cannot be automatically propagated to the registered callbacks,
|
||||||
|
as the file event "action" may apply only to a subset of those functions. This is
|
||||||
|
the place for that final delineation, ensuring the exact action is matched before
|
||||||
|
callback execution. This has the benefit of being a suitable proxy for the actual
|
||||||
|
iNotify filtering that takes place when submitting synthetic events to the router
|
||||||
|
by hand.
|
||||||
|
|
||||||
|
**Bigger picture, and why we have to reproduce the work already done by an
|
||||||
|
event-based mechanism like iNotify**: Realistically, such a method is needed
|
||||||
|
regardless if we hope to connect to the threaded router model as we do not
|
||||||
|
canonically store callback associations at the listener level. If our responses
|
||||||
|
could be tied one-to-one to the sensitivities of iNotify events, then they could
|
||||||
|
be called directly in response to them. But they are not: we want to support
|
||||||
|
glob-based filtering, need to delineate by flags as explained above, and can have
|
||||||
|
separate endpoints for the same path. These are conditions *not* collapsed at the
|
||||||
|
iNotify level, and thus need to be fully re-implemented for callback matching.
|
||||||
|
(For example, imagine we had callback uniqueness on just endpoint and glob, i.e.,
|
||||||
|
no sensitivity to flags, then the flag-matching conditions implemented here would
|
||||||
|
not be needed to rightly pass iNotify events to their callbacks. In such a case,
|
||||||
|
we could rely fully on iNotify's flag response model to implicitly handle this
|
||||||
|
aspect of the filtering process. If the same could be said the remaining
|
||||||
|
constraints, then as mentioned, we could simply associate callbacks one-to-one and
|
||||||
|
avoid the auxiliary filtering altogether.)
|
||||||
|
|
||||||
Parameters:
|
Parameters:
|
||||||
event: Event instance
|
event: Event instance
|
||||||
|
1
execlog/syncers/__init__.py
Normal file
1
execlog/syncers/__init__.py
Normal file
@ -0,0 +1 @@
|
|||||||
|
from execlog.syncers.router import PathDiffer, PathRouterSyncer
|
148
execlog/syncers/router.py
Normal file
148
execlog/syncers/router.py
Normal file
@ -0,0 +1,148 @@
|
|||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
from co3.resources import DiskResource
|
||||||
|
from co3 import Differ, Syncer, Database
|
||||||
|
|
||||||
|
from execlog.event import Event
|
||||||
|
from execlog.routers import PathRouter
|
||||||
|
|
||||||
|
|
||||||
|
class PathDiffer(Differ[Path]):
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
database: Database,
|
||||||
|
):
|
||||||
|
super().__init__(DiskResource(), database)
|
||||||
|
|
||||||
|
def l_transform(self, item):
|
||||||
|
'''
|
||||||
|
Transform ``(path, head)`` tuple from ``DiskResource``.
|
||||||
|
'''
|
||||||
|
return Path(*item)
|
||||||
|
|
||||||
|
class PathRouterSyncer(Syncer[Path]):
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
differ: PathDiffer,
|
||||||
|
router: PathRouter,
|
||||||
|
):
|
||||||
|
super().__init__(differ)
|
||||||
|
self.router = router
|
||||||
|
|
||||||
|
def _construct_event(
|
||||||
|
self,
|
||||||
|
fpath: str | Path,
|
||||||
|
endpoint: str | Path,
|
||||||
|
action: bytes
|
||||||
|
):
|
||||||
|
return Event(
|
||||||
|
endpoint=str(endpoint),
|
||||||
|
name=str(Path(fpath).relative_to(endpoint)),
|
||||||
|
action=[action], # synthetic action to match any flag filters
|
||||||
|
)
|
||||||
|
|
||||||
|
def handle_l_excl(self, path: Path, disk_pairs: list):
|
||||||
|
'''
|
||||||
|
Handle disk exclusive paths (i.e., those added to disk since last sync).
|
||||||
|
'''
|
||||||
|
return [
|
||||||
|
self._construct_event(str(path), endpoint, iflags.CREATE)
|
||||||
|
for endpoint, _ in disk_pairs
|
||||||
|
]
|
||||||
|
|
||||||
|
def handle_r_excl(self, path: Path, db_vals: list):
|
||||||
|
'''
|
||||||
|
Handle database exclusive paths (i.e., those deleted from disk since last sync).
|
||||||
|
Searches for matching endpoints under the attached router and creates
|
||||||
|
corresponding events.
|
||||||
|
|
||||||
|
.. admonition:: On lack of endpoints
|
||||||
|
|
||||||
|
This method handles database exclusive items, i.e., paths no longer on disk
|
||||||
|
but still in the database. For typical Router designs, it is not important to
|
||||||
|
preserve possible endpoints of origin for this kind of event; what matters is
|
||||||
|
the absolute path of the file to be removed. In general, file events are
|
||||||
|
associated solely with a path, but we in some cases may be sensitive to the
|
||||||
|
base path seen to be "triggering" that file event, as router methods can hook
|
||||||
|
in to specific endpoints. This has somewhat dubious effects, as multiple
|
||||||
|
events (with the same action) are dispatched for the same file, purely to
|
||||||
|
observe the Router convention of endpoints and allowing independent
|
||||||
|
trajectories through the execution sequence.
|
||||||
|
|
||||||
|
One concern here is that you might, in theory, want to respond to the same
|
||||||
|
file deletion event in different ways under different endpoints. This will be
|
||||||
|
accessible when picking up such an event live, as endpoints are grouped by
|
||||||
|
watch descriptor and can be all be triggered from the single file event. This
|
||||||
|
is the one case where we can't *really* simulate the event taking place with
|
||||||
|
the available data, and instead have to peer into the router to see what root
|
||||||
|
paths the file could theoretically trigger. Most of the time, this won't be
|
||||||
|
too problematic, since we'll be watching the same paths and can tell where a
|
||||||
|
deleted file would've been. But there are cases where a watch path endpoint
|
||||||
|
may be abandoned, and thus no callback will be there to receive the DELETE
|
||||||
|
event. *Routers should heavily consider implementing a global DELETE handler
|
||||||
|
to prevent these cases if it's critical to respond to deletions.* Otherwise,
|
||||||
|
we still make an attempt to propagate under appropriate endpoints, allowing
|
||||||
|
for possible "deconstructor-like" behavior of specific filetypes (e.g.,
|
||||||
|
cleaning up auxiliary elements, writing to a log, creating a backup, etc).
|
||||||
|
'''
|
||||||
|
return [
|
||||||
|
self._construct_event(str(path), str(endpoint), iflags.DELETE)
|
||||||
|
for endpoint in self.router.routemap
|
||||||
|
if Path(path).is_relative_to(Path(endpoint))
|
||||||
|
]
|
||||||
|
|
||||||
|
def handle_lr_int(self, path: Path, path_tuples: tuple[list, list]):
|
||||||
|
'''
|
||||||
|
Handle paths reflected both in the database and on disk.
|
||||||
|
|
||||||
|
Paths only reach this method if still present after being passed through
|
||||||
|
``filter_diff_sets``, which will filter out those files that are up-to-date in the
|
||||||
|
database.
|
||||||
|
'''
|
||||||
|
return [
|
||||||
|
self._construct_event(str(path), endpoint, iflags.MODIFY)
|
||||||
|
for endpoint, _ in path_tuples[1]
|
||||||
|
]
|
||||||
|
|
||||||
|
def filter_diff_sets(self, l_excl, r_excl, lr_int):
|
||||||
|
total_disk_files = len(l_excl) + len(lr_int)
|
||||||
|
|
||||||
|
def file_out_of_sync(p):
|
||||||
|
db_el, disk_el = lr_int[p]
|
||||||
|
db_mtime = float(db_el[0].get('mtime','0'))
|
||||||
|
disk_mtime = File(p, disk_el[0]).mtime
|
||||||
|
return disk_mtime > db_mtime
|
||||||
|
|
||||||
|
lr_int = {p:v for p,v in lr_int.items() if file_out_of_sync(p)}
|
||||||
|
|
||||||
|
# compute out-of-sync details
|
||||||
|
oos_count = len(l_excl) + len(lr_int)
|
||||||
|
oos_prcnt = oos_count / max(total_disk_files, 1) * 100
|
||||||
|
|
||||||
|
logger.info(color_text(Fore.GREEN, f'{len(l_excl)} new files to add'))
|
||||||
|
logger.info(color_text(Fore.YELLOW, f'{len(lr_int)} modified files'))
|
||||||
|
logger.info(color_text(Fore.RED, f'{len(r_excl)} files to remove'))
|
||||||
|
logger.info(color_text(Style.DIM, f'({oos_prcnt:.2f}%) of disk files out-of-sync'))
|
||||||
|
|
||||||
|
return l_excl, r_excl, lr_int
|
||||||
|
|
||||||
|
def process_chunk(self, event_sets):
|
||||||
|
chunk_events = [e for event_set in event_sets for e in event_set]
|
||||||
|
|
||||||
|
# 1) flush synthetic events for the batch through the chained router
|
||||||
|
# 2) block until completed and sweep up the collected inserts
|
||||||
|
event_futures = self.router.submit(chunk_events)
|
||||||
|
|
||||||
|
# note: we structure this future waiting like this for the TQDM view
|
||||||
|
results = []
|
||||||
|
for future in tqdm(
|
||||||
|
as_completed(event_futures),
|
||||||
|
total=chunk_size,
|
||||||
|
desc=f'Awaiting chunk futures [submitted {len(event_futures)}/{chunk_size}]'
|
||||||
|
):
|
||||||
|
try:
|
||||||
|
results.append(future.result())
|
||||||
|
except Exception as e:
|
||||||
|
logger.warning(f"Sync job failed with exception {e}")
|
||||||
|
|
||||||
|
return results
|
Loading…
Reference in New Issue
Block a user