nwo
stringlengths
10
28
sha
stringlengths
40
40
path
stringlengths
11
97
identifier
stringlengths
1
64
parameters
stringlengths
2
2.24k
return_statement
stringlengths
0
2.17k
docstring
stringlengths
0
5.45k
docstring_summary
stringlengths
0
3.83k
func_begin
int64
1
13.4k
func_end
int64
2
13.4k
function
stringlengths
28
56.4k
url
stringlengths
106
209
project
int64
1
48
executed_lines
sequence
executed_lines_pc
float64
0
153
missing_lines
sequence
missing_lines_pc
float64
0
100
covered
bool
2 classes
filecoverage
float64
2.53
100
function_lines
int64
2
1.46k
mccabe
int64
1
253
coverage
float64
0
100
docstring_lines
int64
0
112
function_nodoc
stringlengths
9
56.4k
id
int64
0
29.8k
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/task.py
Task.__lt__
(self, other: Task)
return self.priority < other.priority
179
182
def __lt__(self, other: Task) -> bool: if self.priority is None or other.priority is None: return False return self.priority < other.priority
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/task.py#L179-L182
1
[ 0, 1, 3 ]
75
[ 2 ]
25
false
88.505747
4
3
75
0
def __lt__(self, other: Task) -> bool: if self.priority is None or other.priority is None: return False return self.priority < other.priority
100
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/task.py
Task.__eq__
(self, other: object)
return self.priority == other.priority
184
190
def __eq__(self, other: object) -> bool: if not isinstance(other, Task): return NotImplemented if not self.priority or not other.priority: # WTF??? return True return self.priority == other.priority
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/task.py#L184-L190
1
[ 0, 1, 3, 4, 5, 6 ]
85.714286
[ 2 ]
14.285714
false
88.505747
7
4
85.714286
0
def __eq__(self, other: object) -> bool: if not isinstance(other, Task): return NotImplemented if not self.priority or not other.priority: # WTF??? return True return self.priority == other.priority
101
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.__init__
( self, task_queue: None | BaseTaskQueue = None, thread_number: None | int = None, network_try_limit: None | int = None, task_try_limit: None | int = None, priority_mode: str = "random", meta: None | dict[str, Any] = None, config: None | dict[str, Any] = None, parser_requests_per_process: int = 10000, parser_pool_size: int = 1, network_service: None | BaseNetworkService = None, grab_transport: None | BaseTransport[HttpRequest, Document] | type[BaseTransport[HttpRequest, Document]] = None, )
Create Spider instance, duh. Arguments: * thread-number - Number of concurrent network streams * network_try_limit - How many times try to send request again if network error was occurred, use 0 to disable * task_try_limit - Limit of tries to execute some task this is not the same as network_try_limit network try limit limits the number of tries which are performed automatically in case of network timeout of some other physical error but task_try_limit limits the number of attempts which are scheduled manually in the spider business logic * priority_mode - could be "random" or "const" * meta - arbitrary user data
Create Spider instance, duh.
72
171
def __init__( self, task_queue: None | BaseTaskQueue = None, thread_number: None | int = None, network_try_limit: None | int = None, task_try_limit: None | int = None, priority_mode: str = "random", meta: None | dict[str, Any] = None, config: None | dict[str, Any] = None, parser_requests_per_process: int = 10000, parser_pool_size: int = 1, network_service: None | BaseNetworkService = None, grab_transport: None | BaseTransport[HttpRequest, Document] | type[BaseTransport[HttpRequest, Document]] = None, ) -> None: """Create Spider instance, duh. Arguments: * thread-number - Number of concurrent network streams * network_try_limit - How many times try to send request again if network error was occurred, use 0 to disable * task_try_limit - Limit of tries to execute some task this is not the same as network_try_limit network try limit limits the number of tries which are performed automatically in case of network timeout of some other physical error but task_try_limit limits the number of attempts which are scheduled manually in the spider business logic * priority_mode - could be "random" or "const" * meta - arbitrary user data """ self.fatal_error_queue: Queue[FatalErrorQueueItem] = Queue() self._started: None | float = None self.grab_transport = grab_transport self.parser_requests_per_process = parser_requests_per_process self.stat = Stat() self.runtime_events: dict[str, list[None | str]] = {} self.task_queue: BaseTaskQueue = task_queue if task_queue else MemoryTaskQueue() if config is not None: self.config = config else: self.config = {} if meta: self.meta = meta else: self.meta = {} self.thread_number = thread_number or int( self.config.get("thread_number", DEFAULT_NETWORK_STREAM_NUMBER) ) self.task_try_limit = task_try_limit or int( self.config.get("task_try_limit", DEFAULT_TASK_TRY_LIMIT) ) self.network_try_limit = network_try_limit or int( self.config.get("network_try_limit", DEFAULT_NETWORK_TRY_LIMIT) ) if priority_mode not in ["random", "const"]: raise SpiderMisuseError( 'Value of priority_mode option should be "random" or "const"' ) self.priority_mode = priority_mode self.work_allowed = True self.proxylist_enabled: None | bool = None self.proxylist: None | ProxyList = None self.proxy: None | ProxyServer = None self.proxy_auto_change = False self.parser_pool_size = parser_pool_size assert network_service is None or isinstance( network_service, BaseNetworkService ) self.network_service = ( network_service if network_service is not None else NetworkServiceThreaded( self.fatal_error_queue, self.thread_number, process_task=self.srv_process_task, get_task_from_queue=self.get_task_from_queue, ) ) self.task_dispatcher = TaskDispatcherService( self.fatal_error_queue, process_service_result=self.srv_process_service_result, ) self.parser_service = ParserService( fatal_error_queue=self.fatal_error_queue, pool_size=self.parser_pool_size, task_dispatcher=self.task_dispatcher, stat=self.stat, parser_requests_per_process=self.parser_requests_per_process, find_task_handler=self.find_task_handler, ) self.task_generator_service = TaskGeneratorService( self.fatal_error_queue, self.task_generator(), thread_number=self.thread_number, get_task_queue=self.get_task_queue, parser_service=self.parser_service, task_dispatcher=self.task_dispatcher, )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L72-L171
1
[ 0, 31, 32, 33, 34, 35, 36, 37, 38, 39, 41, 42, 43, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99 ]
68
[ 40, 44 ]
2
false
92.476489
100
9
98
15
def __init__( self, task_queue: None | BaseTaskQueue = None, thread_number: None | int = None, network_try_limit: None | int = None, task_try_limit: None | int = None, priority_mode: str = "random", meta: None | dict[str, Any] = None, config: None | dict[str, Any] = None, parser_requests_per_process: int = 10000, parser_pool_size: int = 1, network_service: None | BaseNetworkService = None, grab_transport: None | BaseTransport[HttpRequest, Document] | type[BaseTransport[HttpRequest, Document]] = None, ) -> None: self.fatal_error_queue: Queue[FatalErrorQueueItem] = Queue() self._started: None | float = None self.grab_transport = grab_transport self.parser_requests_per_process = parser_requests_per_process self.stat = Stat() self.runtime_events: dict[str, list[None | str]] = {} self.task_queue: BaseTaskQueue = task_queue if task_queue else MemoryTaskQueue() if config is not None: self.config = config else: self.config = {} if meta: self.meta = meta else: self.meta = {} self.thread_number = thread_number or int( self.config.get("thread_number", DEFAULT_NETWORK_STREAM_NUMBER) ) self.task_try_limit = task_try_limit or int( self.config.get("task_try_limit", DEFAULT_TASK_TRY_LIMIT) ) self.network_try_limit = network_try_limit or int( self.config.get("network_try_limit", DEFAULT_NETWORK_TRY_LIMIT) ) if priority_mode not in ["random", "const"]: raise SpiderMisuseError( 'Value of priority_mode option should be "random" or "const"' ) self.priority_mode = priority_mode self.work_allowed = True self.proxylist_enabled: None | bool = None self.proxylist: None | ProxyList = None self.proxy: None | ProxyServer = None self.proxy_auto_change = False self.parser_pool_size = parser_pool_size assert network_service is None or isinstance( network_service, BaseNetworkService ) self.network_service = ( network_service if network_service is not None else NetworkServiceThreaded( self.fatal_error_queue, self.thread_number, process_task=self.srv_process_task, get_task_from_queue=self.get_task_from_queue, ) ) self.task_dispatcher = TaskDispatcherService( self.fatal_error_queue, process_service_result=self.srv_process_service_result, ) self.parser_service = ParserService( fatal_error_queue=self.fatal_error_queue, pool_size=self.parser_pool_size, task_dispatcher=self.task_dispatcher, stat=self.stat, parser_requests_per_process=self.parser_requests_per_process, find_task_handler=self.find_task_handler, ) self.task_generator_service = TaskGeneratorService( self.fatal_error_queue, self.task_generator(), thread_number=self.thread_number, get_task_queue=self.get_task_queue, parser_service=self.parser_service, task_dispatcher=self.task_dispatcher, )
102
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.collect_runtime_event
(self, name: str, value: None | str)
173
174
def collect_runtime_event(self, name: str, value: None | str) -> None: self.runtime_events.setdefault(name, []).append(value)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L173-L174
1
[ 0, 1 ]
100
[]
0
true
92.476489
2
1
100
0
def collect_runtime_event(self, name: str, value: None | str) -> None: self.runtime_events.setdefault(name, []).append(value)
103
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.setup_queue
(self, *_args: Any, **_kwargs: Any)
Set up queue.
Set up queue.
178
184
def setup_queue(self, *_args: Any, **_kwargs: Any) -> None: """Set up queue.""" raise GrabFeatureIsDeprecated( """Method Spider.setup_queue is deprecated. Now MemoryTaskQueue is used by default. If you need custom task queue pass instance of queue class in task_queue parameter in constructor of Spider class.""" )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L178-L184
1
[ 0, 1 ]
28.571429
[ 2 ]
14.285714
false
92.476489
7
1
85.714286
1
def setup_queue(self, *_args: Any, **_kwargs: Any) -> None: raise GrabFeatureIsDeprecated( )
104
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.add_task
( self, task: Task, queue: None | BaseTaskQueue = None, raise_error: bool = False, )
return True
Add task to the task queue.
Add task to the task queue.
186
217
def add_task( self, task: Task, queue: None | BaseTaskQueue = None, raise_error: bool = False, ) -> bool: """Add task to the task queue.""" if queue is None: queue = self.task_queue if task.priority is None or not task.priority_set_explicitly: task.priority = self.generate_task_priority() task.priority_set_explicitly = False else: task.priority_set_explicitly = True if not task.request.url or not task.request.url.startswith( ("http://", "https://", "ftp://", "file://", "feed://") ): self.collect_runtime_event("task-with-invalid-url", task.request.url) msg = "Invalid task URL: %s" % task.request.url if raise_error: raise SpiderError(msg) logger.error( "%s\nTraceback:\n%s", msg, "".join(format_stack()), ) return False # TODO: keep original task priority if it was set explicitly # WTF the previous comment means? queue.put(task, priority=task.priority, schedule_time=task.schedule_time) return True
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L186-L217
1
[ 0, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31 ]
84.375
[]
0
false
92.476489
32
7
100
1
def add_task( self, task: Task, queue: None | BaseTaskQueue = None, raise_error: bool = False, ) -> bool: if queue is None: queue = self.task_queue if task.priority is None or not task.priority_set_explicitly: task.priority = self.generate_task_priority() task.priority_set_explicitly = False else: task.priority_set_explicitly = True if not task.request.url or not task.request.url.startswith( ("http://", "https://", "ftp://", "file://", "feed://") ): self.collect_runtime_event("task-with-invalid-url", task.request.url) msg = "Invalid task URL: %s" % task.request.url if raise_error: raise SpiderError(msg) logger.error( "%s\nTraceback:\n%s", msg, "".join(format_stack()), ) return False # TODO: keep original task priority if it was set explicitly # WTF the previous comment means? queue.put(task, priority=task.priority, schedule_time=task.schedule_time) return True
105
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.stop
(self)
Instruct spider to stop processing new tasks and start shutting down.
Instruct spider to stop processing new tasks and start shutting down.
219
221
def stop(self) -> None: """Instruct spider to stop processing new tasks and start shutting down.""" self.work_allowed = False
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L219-L221
1
[ 0, 1, 2 ]
100
[]
0
true
92.476489
3
1
100
1
def stop(self) -> None: self.work_allowed = False
106
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.load_proxylist
( self, source: str | BaseProxySource, source_type: None | str = None, proxy_type: str = "http", auto_init: bool = True, auto_change: bool = True, )
Load proxy list. :param source: Proxy source. Accepts string (file path, url) or ``BaseProxySource`` instance. :param source_type: The type of the specified source. Should be one of the following: 'text_file' or 'url'. :param proxy_type: Should be one of the following: 'socks4', 'socks5' or'http'. :param auto_change: If set to `True` then automatically random proxy rotation will be used. Proxy source format should be one of the following (for each line): - ip:port - ip:port:login:password
Load proxy list.
223
275
def load_proxylist( self, source: str | BaseProxySource, source_type: None | str = None, proxy_type: str = "http", auto_init: bool = True, auto_change: bool = True, ) -> None: """Load proxy list. :param source: Proxy source. Accepts string (file path, url) or ``BaseProxySource`` instance. :param source_type: The type of the specified source. Should be one of the following: 'text_file' or 'url'. :param proxy_type: Should be one of the following: 'socks4', 'socks5' or'http'. :param auto_change: If set to `True` then automatically random proxy rotation will be used. Proxy source format should be one of the following (for each line): - ip:port - ip:port:login:password """ if isinstance(source, BaseProxySource): self.proxylist = ProxyList(source) elif isinstance(source, str): if source_type == "text_file": self.proxylist = ProxyList.from_local_file( source, proxy_type=proxy_type ) elif source_type == "url": self.proxylist = ProxyList.from_network_file( source, proxy_type=proxy_type ) else: raise SpiderMisuseError( "Method `load_proxylist` received " "invalid `source_type` argument: %s" % source_type ) else: raise SpiderMisuseError( "Method `load_proxylist` received " "invalid `source` argument: %s" % source ) self.proxylist_enabled = True self.proxy = None if not auto_change and auto_init: self.proxy = self.proxylist.get_random_server() if not self.proxy.proxy_type: raise GrabMisuseError("Could not use proxy without defined proxy type") self.proxy_auto_change = auto_change
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L223-L275
1
[ 0, 23, 24, 26, 27, 28, 29, 30, 45, 46, 47, 48, 52 ]
24.528302
[ 25, 31, 32, 36, 41, 49, 50, 51 ]
15.09434
false
92.476489
53
8
84.90566
15
def load_proxylist( self, source: str | BaseProxySource, source_type: None | str = None, proxy_type: str = "http", auto_init: bool = True, auto_change: bool = True, ) -> None: if isinstance(source, BaseProxySource): self.proxylist = ProxyList(source) elif isinstance(source, str): if source_type == "text_file": self.proxylist = ProxyList.from_local_file( source, proxy_type=proxy_type ) elif source_type == "url": self.proxylist = ProxyList.from_network_file( source, proxy_type=proxy_type ) else: raise SpiderMisuseError( "Method `load_proxylist` received " "invalid `source_type` argument: %s" % source_type ) else: raise SpiderMisuseError( "Method `load_proxylist` received " "invalid `source` argument: %s" % source ) self.proxylist_enabled = True self.proxy = None if not auto_change and auto_init: self.proxy = self.proxylist.get_random_server() if not self.proxy.proxy_type: raise GrabMisuseError("Could not use proxy without defined proxy type") self.proxy_auto_change = auto_change
107
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.render_stats
(self)
return "\n".join(out) + "\n"
277
314
def render_stats(self) -> str: out = [ "------------ Stats: ------------", "Counters:", ] # Process counters items = sorted(self.stat.counters.items(), key=lambda x: x[0], reverse=True) for item in items: out.append(" %s: %s" % item) out.append("") out.append("Lists:") # Process event lists sorted by size in descendant order col_sizes = [(x, len(y)) for x, y in self.runtime_events.items()] col_sizes = sorted(col_sizes, key=lambda x: x[1], reverse=True) for col_size in col_sizes: out.append(" %s: %d" % col_size) out.append("") # Process extra metrics if "download-size" in self.stat.counters: out.append( "Network download: %s" % format_traffic_value(self.stat.counters["download-size"]) ) out.append( "Queue size: %d" % self.task_queue.size() if self.task_queue else "NA" ) out.append("Network streams: %d" % self.thread_number) elapsed = (time.time() - self._started) if self._started else 0 hours, seconds = divmod(elapsed, 3600) minutes, seconds = divmod(seconds, 60) out.append("Time elapsed: %d:%d:%d (H:M:S)" % (hours, minutes, seconds)) out.append( "End time: %s" % datetime.utcnow().strftime("%d %b %Y, %H:%M:%S UTC") ) return "\n".join(out) + "\n"
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L277-L314
1
[ 0, 1, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 18, 19, 20, 21, 26, 29, 30, 31, 32, 33, 34, 37 ]
65.789474
[ 17, 22 ]
5.263158
false
92.476489
38
5
94.736842
0
def render_stats(self) -> str: out = [ "------------ Stats: ------------", "Counters:", ] # Process counters items = sorted(self.stat.counters.items(), key=lambda x: x[0], reverse=True) for item in items: out.append(" %s: %s" % item) out.append("") out.append("Lists:") # Process event lists sorted by size in descendant order col_sizes = [(x, len(y)) for x, y in self.runtime_events.items()] col_sizes = sorted(col_sizes, key=lambda x: x[1], reverse=True) for col_size in col_sizes: out.append(" %s: %d" % col_size) out.append("") # Process extra metrics if "download-size" in self.stat.counters: out.append( "Network download: %s" % format_traffic_value(self.stat.counters["download-size"]) ) out.append( "Queue size: %d" % self.task_queue.size() if self.task_queue else "NA" ) out.append("Network streams: %d" % self.thread_number) elapsed = (time.time() - self._started) if self._started else 0 hours, seconds = divmod(elapsed, 3600) minutes, seconds = divmod(seconds, 60) out.append("Time elapsed: %d:%d:%d (H:M:S)" % (hours, minutes, seconds)) out.append( "End time: %s" % datetime.utcnow().strftime("%d %b %Y, %H:%M:%S UTC") ) return "\n".join(out) + "\n"
108
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.prepare
(self)
Do additional spider customization here. This method runs before spider has started working.
Do additional spider customization here.
320
324
def prepare(self) -> None: """Do additional spider customization here. This method runs before spider has started working. """
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L320-L324
1
[ 0, 1, 2, 3, 4 ]
100
[]
0
true
92.476489
5
1
100
3
def prepare(self) -> None:
109
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.shutdown
(self)
Override this method to do some final actions after parsing has been done.
Override this method to do some final actions after parsing has been done.
326
327
def shutdown(self) -> None: """Override this method to do some final actions after parsing has been done."""
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L326-L327
1
[ 0, 1 ]
100
[]
0
true
92.476489
2
1
100
1
def shutdown(self) -> None:
110
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.create_grab_instance
(self, **kwargs: Any)
return Grab(transport=self.grab_transport, **kwargs)
329
330
def create_grab_instance(self, **kwargs: Any) -> Grab: return Grab(transport=self.grab_transport, **kwargs)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L329-L330
1
[ 0, 1 ]
100
[]
0
true
92.476489
2
1
100
0
def create_grab_instance(self, **kwargs: Any) -> Grab: return Grab(transport=self.grab_transport, **kwargs)
111
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.task_generator
(self)
You can override this method to load new tasks. It will be used each time as number of tasks in task queue is less then number of threads multiplied on 2 This allows you to not overload all free memory if total number of tasks is big.
You can override this method to load new tasks.
332
340
def task_generator(self) -> Iterator[Task]: """You can override this method to load new tasks. It will be used each time as number of tasks in task queue is less then number of threads multiplied on 2 This allows you to not overload all free memory if total number of tasks is big. """ yield from ()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L332-L340
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 8 ]
100
[]
0
true
92.476489
9
1
100
6
def task_generator(self) -> Iterator[Task]: yield from ()
112
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.check_task_limits
(self, task: Task)
return True, "ok"
Check that task's network & try counters do not exceed limits. Returns: * if success: (True, None) * if error: (False, reason)
Check that task's network & try counters do not exceed limits.
346
360
def check_task_limits(self, task: Task) -> tuple[bool, str]: """Check that task's network & try counters do not exceed limits. Returns: * if success: (True, None) * if error: (False, reason) """ if task.task_try_count > self.task_try_limit: return False, "task-try-count" if task.network_try_count > self.network_try_limit: return False, "network-try-count" return True, "ok"
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L346-L360
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14 ]
100
[]
0
true
92.476489
15
3
100
5
def check_task_limits(self, task: Task) -> tuple[bool, str]: if task.task_try_count > self.task_try_limit: return False, "task-try-count" if task.network_try_count > self.network_try_limit: return False, "network-try-count" return True, "ok"
113
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.generate_task_priority
(self)
return system_random.randint(*RANDOM_TASK_PRIORITY_RANGE)
362
365
def generate_task_priority(self) -> int: if self.priority_mode == "const": return DEFAULT_TASK_PRIORITY return system_random.randint(*RANDOM_TASK_PRIORITY_RANGE)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L362-L365
1
[ 0, 1, 2, 3 ]
100
[]
0
true
92.476489
4
2
100
0
def generate_task_priority(self) -> int: if self.priority_mode == "const": return DEFAULT_TASK_PRIORITY return system_random.randint(*RANDOM_TASK_PRIORITY_RANGE)
114
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.process_initial_urls
(self)
367
370
def process_initial_urls(self) -> None: if self.initial_urls: for url in self.initial_urls: self.add_task(Task(name="initial", request=HttpRequest(url)))
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L367-L370
1
[ 0, 1, 2, 3 ]
100
[]
0
true
92.476489
4
3
100
0
def process_initial_urls(self) -> None: if self.initial_urls: for url in self.initial_urls: self.add_task(Task(name="initial", request=HttpRequest(url)))
115
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.get_task_from_queue
(self)
372
379
def get_task_from_queue(self) -> None | Literal[True] | Task: try: return self.task_queue.get() except Empty: size = self.task_queue.size() if size: return True return None
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L372-L379
1
[ 0, 1, 2, 3, 4, 5, 7 ]
87.5
[ 6 ]
12.5
false
92.476489
8
3
87.5
0
def get_task_from_queue(self) -> None | Literal[True] | Task: try: return self.task_queue.get() except Empty: size = self.task_queue.size() if size: return True return None
116
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.is_valid_network_response_code
(self, code: int, task: Task)
return code < 400 or code == 404 or code in task.valid_status
Test if response is valid. Valid response is handled with associated task handler. Failed respoosne is processed with error handler.
Test if response is valid.
381
387
def is_valid_network_response_code(self, code: int, task: Task) -> bool: """Test if response is valid. Valid response is handled with associated task handler. Failed respoosne is processed with error handler. """ return code < 400 or code == 404 or code in task.valid_status
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L381-L387
1
[ 0, 1, 2, 3, 4, 5, 6 ]
100
[]
0
true
92.476489
7
3
100
4
def is_valid_network_response_code(self, code: int, task: Task) -> bool: return code < 400 or code == 404 or code in task.valid_status
117
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.process_parser_error
( self, func_name: str, task: Task, exc_info: tuple[type[Exception], Exception, TracebackType], )
389
408
def process_parser_error( self, func_name: str, task: Task, exc_info: tuple[type[Exception], Exception, TracebackType], ) -> None: _, ex, _ = exc_info self.stat.inc("spider:error-%s" % ex.__class__.__name__.lower()) logger.error( "Task handler [%s] error\n%s", func_name, "".join(format_exception(*exc_info)), ) task_url = task.request.url if task else None self.collect_runtime_event( "fatal", "%s|%s|%s|%s" % (func_name, ex.__class__.__name__, str(ex), task_url), )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L389-L408
1
[ 0, 6, 7, 8, 9, 14, 15, 16 ]
40
[]
0
false
92.476489
20
1
100
0
def process_parser_error( self, func_name: str, task: Task, exc_info: tuple[type[Exception], Exception, TracebackType], ) -> None: _, ex, _ = exc_info self.stat.inc("spider:error-%s" % ex.__class__.__name__.lower()) logger.error( "Task handler [%s] error\n%s", func_name, "".join(format_exception(*exc_info)), ) task_url = task.request.url if task else None self.collect_runtime_event( "fatal", "%s|%s|%s|%s" % (func_name, ex.__class__.__name__, str(ex), task_url), )
118
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.find_task_handler
(self, task: Task)
410
423
def find_task_handler(self, task: Task) -> Callable[..., Any]: callback = task.get("callback") if callback: # pylint: disable=deprecated-typing-alias return cast(typing.Callable[..., Any], callback) # pylint: enable=deprecated-typing-alias try: # pylint: disable=deprecated-typing-alias return cast(typing.Callable[..., Any], getattr(self, "task_%s" % task.name)) # pylint: enable=deprecated-typing-alias except AttributeError as ex: raise NoTaskHandler( "No handler or callback defined for " "task %s" % task.name ) from ex
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L410-L423
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11 ]
85.714286
[]
0
false
92.476489
14
3
100
0
def find_task_handler(self, task: Task) -> Callable[..., Any]: callback = task.get("callback") if callback: # pylint: disable=deprecated-typing-alias return cast(typing.Callable[..., Any], callback) # pylint: enable=deprecated-typing-alias try: # pylint: disable=deprecated-typing-alias return cast(typing.Callable[..., Any], getattr(self, "task_%s" % task.name)) # pylint: enable=deprecated-typing-alias except AttributeError as ex: raise NoTaskHandler( "No handler or callback defined for " "task %s" % task.name ) from ex
119
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.log_network_result_stats
(self, res: NetworkResult, task: Task)
425
437
def log_network_result_stats(self, res: NetworkResult, task: Task) -> None: # Increase stat counters self.stat.inc("spider:request-processed") self.stat.inc("spider:task") self.stat.inc("spider:task-%s" % task.name) if task.network_try_count == 1 and task.task_try_count == 1: self.stat.inc("spider:task-%s-initial" % task.name) # Update traffic statistics if res["grab"] and res["doc"]: doc = res["doc"] self.stat.inc("spider:download-size", doc.download_size) self.stat.inc("spider:upload-size", doc.upload_size)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L425-L437
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12 ]
100
[]
0
true
92.476489
13
5
100
0
def log_network_result_stats(self, res: NetworkResult, task: Task) -> None: # Increase stat counters self.stat.inc("spider:request-processed") self.stat.inc("spider:task") self.stat.inc("spider:task-%s" % task.name) if task.network_try_count == 1 and task.task_try_count == 1: self.stat.inc("spider:task-%s-initial" % task.name) # Update traffic statistics if res["grab"] and res["doc"]: doc = res["doc"] self.stat.inc("spider:download-size", doc.download_size) self.stat.inc("spider:upload-size", doc.upload_size)
120
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.process_grab_proxy
(self, task: Task, grab: Grab)
Assign new proxy from proxylist to the task.
Assign new proxy from proxylist to the task.
439
445
def process_grab_proxy(self, task: Task, grab: Grab) -> None: """Assign new proxy from proxylist to the task.""" if task.use_proxylist and self.proxylist_enabled: if self.proxy_auto_change: self.change_active_proxy(task, grab) if self.proxy: raise Exception("Look like it is not called from tests")
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L439-L445
1
[ 0, 1, 2, 3, 5 ]
71.428571
[ 4, 6 ]
28.571429
false
92.476489
7
5
71.428571
1
def process_grab_proxy(self, task: Task, grab: Grab) -> None: if task.use_proxylist and self.proxylist_enabled: if self.proxy_auto_change: self.change_active_proxy(task, grab) if self.proxy: raise Exception("Look like it is not called from tests")
121
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.change_active_proxy
(self, task: Task, grab: Grab)
452
458
def change_active_proxy(self, task: Task, grab: Grab) -> None: # pylint: disable=unused-argument self.proxy = cast(ProxyList, self.proxylist).get_random_server() if not self.proxy.proxy_type: raise SpiderMisuseError( 'Value of priority_mode option should be "random" or "const"' )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L452-L458
1
[ 0, 1 ]
28.571429
[ 2, 3, 4 ]
42.857143
false
92.476489
7
2
57.142857
0
def change_active_proxy(self, task: Task, grab: Grab) -> None: # pylint: disable=unused-argument self.proxy = cast(ProxyList, self.proxylist).get_random_server() if not self.proxy.proxy_type: raise SpiderMisuseError( 'Value of priority_mode option should be "random" or "const"' )
122
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.get_task_queue
(self)
return self.task_queue
460
464
def get_task_queue(self) -> BaseTaskQueue: # this method is expected to be called # after "spider.run()" is called # i.e. the "self.task_queue" is set return self.task_queue
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L460-L464
1
[ 0, 1, 2, 3, 4 ]
100
[]
0
true
92.476489
5
1
100
0
def get_task_queue(self) -> BaseTaskQueue: # this method is expected to be called # after "spider.run()" is called # i.e. the "self.task_queue" is set return self.task_queue
123
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.is_idle_estimated
(self)
return ( not self.task_generator_service.is_alive() and not self.task_queue.size() and not self.task_dispatcher.input_queue.qsize() and not self.parser_service.input_queue.qsize() and not self.parser_service.is_busy() and not self.network_service.get_active_threads_number() and not self.network_service.is_busy() )
466
475
def is_idle_estimated(self) -> bool: return ( not self.task_generator_service.is_alive() and not self.task_queue.size() and not self.task_dispatcher.input_queue.qsize() and not self.parser_service.input_queue.qsize() and not self.parser_service.is_busy() and not self.network_service.get_active_threads_number() and not self.network_service.is_busy() )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L466-L475
1
[ 0, 1 ]
20
[]
0
false
92.476489
10
7
100
0
def is_idle_estimated(self) -> bool: return ( not self.task_generator_service.is_alive() and not self.task_queue.size() and not self.task_dispatcher.input_queue.qsize() and not self.parser_service.input_queue.qsize() and not self.parser_service.is_busy() and not self.network_service.get_active_threads_number() and not self.network_service.is_busy() )
124
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.is_idle_confirmed
(self, services: list[BaseService])
return False
Test if spider is fully idle. WARNING: As side effect it stops all services to get state of queues anaffected by sercies. Spider is full idle when all conditions are met: * all services are paused i.e. the do not change queues * all queues are empty * task generator is completed
Test if spider is fully idle.
477
495
def is_idle_confirmed(self, services: list[BaseService]) -> bool: """Test if spider is fully idle. WARNING: As side effect it stops all services to get state of queues anaffected by sercies. Spider is full idle when all conditions are met: * all services are paused i.e. the do not change queues * all queues are empty * task generator is completed """ if self.is_idle_estimated(): for srv in services: srv.pause() if self.is_idle_estimated(): return True for srv in services: srv.resume() return False
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L477-L495
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 18 ]
89.473684
[ 16, 17 ]
10.526316
false
92.476489
19
5
89.473684
9
def is_idle_confirmed(self, services: list[BaseService]) -> bool: if self.is_idle_estimated(): for srv in services: srv.pause() if self.is_idle_estimated(): return True for srv in services: srv.resume() return False
125
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.run
(self)
497
524
def run(self) -> None: self._started = time.time() services = [] try: self.prepare() self.process_initial_urls() services = [ self.task_dispatcher, self.task_generator_service, self.parser_service, self.network_service, ] for srv in services: srv.start() while self.work_allowed: try: exc_info = self.fatal_error_queue.get(True, 0.5) except Empty: pass else: # WTF: why? (see below) # The trackeback of fatal error MUST BE rendered by the sender raise exc_info[1] if self.is_idle_confirmed(services): break finally: self.shutdown_services(services) self.stat.shutdown(join_threads=True)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L497-L524
1
[ 0, 1, 2, 3, 4, 5, 6, 12, 13, 14, 15, 16, 17, 18, 21, 22, 23, 24, 26, 27 ]
71.428571
[]
0
false
92.476489
28
5
100
0
def run(self) -> None: self._started = time.time() services = [] try: self.prepare() self.process_initial_urls() services = [ self.task_dispatcher, self.task_generator_service, self.parser_service, self.network_service, ] for srv in services: srv.start() while self.work_allowed: try: exc_info = self.fatal_error_queue.get(True, 0.5) except Empty: pass else: # WTF: why? (see below) # The trackeback of fatal error MUST BE rendered by the sender raise exc_info[1] if self.is_idle_confirmed(services): break finally: self.shutdown_services(services) self.stat.shutdown(join_threads=True)
126
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.shutdown_services
(self, services: list[BaseService])
526
545
def shutdown_services(self, services: list[BaseService]) -> None: # TODO: for srv in services: # Resume service if it has been paused # to allow service to process stop signal srv.resume() srv.stop() start = time.time() while any(x.is_alive() for x in services): time.sleep(0.1) if time.time() - start > 10: break for srv in services: if srv.is_alive(): logger.error("The %s has not stopped :(", srv) self.stat.render_moment() self.shutdown() self.task_queue.clear() self.task_queue.close() logger.debug("Work done")
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L526-L545
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 12, 13, 15, 16, 17, 18, 19 ]
90
[ 11, 14 ]
10
false
92.476489
20
6
90
0
def shutdown_services(self, services: list[BaseService]) -> None: # TODO: for srv in services: # Resume service if it has been paused # to allow service to process stop signal srv.resume() srv.stop() start = time.time() while any(x.is_alive() for x in services): time.sleep(0.1) if time.time() - start > 10: break for srv in services: if srv.is_alive(): logger.error("The %s has not stopped :(", srv) self.stat.render_moment() self.shutdown() self.task_queue.clear() self.task_queue.close() logger.debug("Work done")
127
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.log_failed_network_result
(self, res: NetworkResult)
547
557
def log_failed_network_result(self, res: NetworkResult) -> None: orig_exc = ( res["exc"].original_exc if isinstance(res["exc"], OriginalExceptionGrabError) else res["exc"] ) msg = ( ("http-%s" % res["doc"].code) if res["ok"] else orig_exc.__class__.__name__ ) self.stat.inc("error:%s" % msg)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L547-L557
1
[ 0, 1, 6, 9, 10 ]
45.454545
[]
0
false
92.476489
11
1
100
0
def log_failed_network_result(self, res: NetworkResult) -> None: orig_exc = ( res["exc"].original_exc if isinstance(res["exc"], OriginalExceptionGrabError) else res["exc"] ) msg = ( ("http-%s" % res["doc"].code) if res["ok"] else orig_exc.__class__.__name__ ) self.stat.inc("error:%s" % msg)
128
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.log_rejected_task
(self, task: Task, reason: str)
559
565
def log_rejected_task(self, task: Task, reason: str) -> None: if reason == "task-try-count": self.collect_runtime_event("task-count-rejected", task.request.url) elif reason == "network-try-count": self.collect_runtime_event("network-count-rejected", task.request.url) else: raise SpiderError("Unknown response from check_task_limits: %s" % reason)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L559-L565
1
[ 0, 1, 2, 3, 4, 6 ]
85.714286
[]
0
false
92.476489
7
3
100
0
def log_rejected_task(self, task: Task, reason: str) -> None: if reason == "task-try-count": self.collect_runtime_event("task-count-rejected", task.request.url) elif reason == "network-try-count": self.collect_runtime_event("network-count-rejected", task.request.url) else: raise SpiderError("Unknown response from check_task_limits: %s" % reason)
129
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.get_fallback_handler
(self, task: Task)
return None
567
578
def get_fallback_handler(self, task: Task) -> None | Callable[..., Any]: if task.fallback_name: # pylint: disable=deprecated-typing-alias return cast(typing.Callable[..., Any], getattr(self, task.fallback_name)) # pylint: enable=deprecated-typing-alias if task.name: fb_name = "task_%s_fallback" % task.name if hasattr(self, fb_name): # pylint: disable=deprecated-typing-alias return cast(typing.Callable[..., Any], getattr(self, fb_name)) # pylint: enable=deprecated-typing-alias return None
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L567-L578
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11 ]
100
[]
0
true
92.476489
12
4
100
0
def get_fallback_handler(self, task: Task) -> None | Callable[..., Any]: if task.fallback_name: # pylint: disable=deprecated-typing-alias return cast(typing.Callable[..., Any], getattr(self, task.fallback_name)) # pylint: enable=deprecated-typing-alias if task.name: fb_name = "task_%s_fallback" % task.name if hasattr(self, fb_name): # pylint: disable=deprecated-typing-alias return cast(typing.Callable[..., Any], getattr(self, fb_name)) # pylint: enable=deprecated-typing-alias return None
130
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.srv_process_service_result
( self, result: Task | None | Exception | dict[str, Any], task: Task, meta: None | dict[str, Any] = None, )
Process result submitted from any service to task dispatcher service. Result could be: * Task * None * Task instance * ResponseNotValid-based exception * Arbitrary exception * Network response: {ok, ecode, emsg, exc, grab, grab_config_backup} Exception can come only from parser_service and it always has meta {"from": "parser", "exc_info": <...>}
Process result submitted from any service to task dispatcher service.
583
629
def srv_process_service_result( self, result: Task | None | Exception | dict[str, Any], task: Task, meta: None | dict[str, Any] = None, ) -> None: """Process result submitted from any service to task dispatcher service. Result could be: * Task * None * Task instance * ResponseNotValid-based exception * Arbitrary exception * Network response: {ok, ecode, emsg, exc, grab, grab_config_backup} Exception can come only from parser_service and it always has meta {"from": "parser", "exc_info": <...>} """ if meta is None: meta = {} if isinstance(result, Task): self.add_task(result) elif result is None: pass elif isinstance(result, ResponseNotValid): self.add_task(task.clone()) error_code = result.__class__.__name__.replace("_", "-") self.stat.inc("integrity:%s" % error_code) elif isinstance(result, Exception): if task: handler = self.find_task_handler(task) handler_name = getattr(handler, "__name__", "NONE") else: handler_name = "NA" self.process_parser_error( handler_name, task, meta["exc_info"], ) if isinstance(result, FatalError): self.fatal_error_queue.put(meta["exc_info"]) elif isinstance(result, dict) and "grab" in result: self.srv_process_network_result(result, task) else: raise SpiderError("Unknown result received from a service: %s" % result)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L583-L629
1
[ 0, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46 ]
59.574468
[ 35 ]
2.12766
false
92.476489
47
10
97.87234
13
def srv_process_service_result( self, result: Task | None | Exception | dict[str, Any], task: Task, meta: None | dict[str, Any] = None, ) -> None: if meta is None: meta = {} if isinstance(result, Task): self.add_task(result) elif result is None: pass elif isinstance(result, ResponseNotValid): self.add_task(task.clone()) error_code = result.__class__.__name__.replace("_", "-") self.stat.inc("integrity:%s" % error_code) elif isinstance(result, Exception): if task: handler = self.find_task_handler(task) handler_name = getattr(handler, "__name__", "NONE") else: handler_name = "NA" self.process_parser_error( handler_name, task, meta["exc_info"], ) if isinstance(result, FatalError): self.fatal_error_queue.put(meta["exc_info"]) elif isinstance(result, dict) and "grab" in result: self.srv_process_network_result(result, task) else: raise SpiderError("Unknown result received from a service: %s" % result)
131
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.srv_process_network_result
(self, result: NetworkResult, task: Task)
631
649
def srv_process_network_result(self, result: NetworkResult, task: Task) -> None: # TODO: Move to network service # starts self.log_network_result_stats(result, task) # ends is_valid = False if task.get("raw"): is_valid = True elif result["ok"]: res_code = result["doc"].code is_valid = self.is_valid_network_response_code(res_code, task) if is_valid: self.parser_service.input_queue.put((result, task)) else: self.log_failed_network_result(result) # Try to do network request one more time if self.network_try_limit > 0: self.add_task(task) self.stat.inc("spider:request")
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L631-L649
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 14, 15, 16, 17, 18 ]
94.736842
[]
0
false
92.476489
19
5
100
0
def srv_process_network_result(self, result: NetworkResult, task: Task) -> None: # TODO: Move to network service # starts self.log_network_result_stats(result, task) # ends is_valid = False if task.get("raw"): is_valid = True elif result["ok"]: res_code = result["doc"].code is_valid = self.is_valid_network_response_code(res_code, task) if is_valid: self.parser_service.input_queue.put((result, task)) else: self.log_failed_network_result(result) # Try to do network request one more time if self.network_try_limit > 0: self.add_task(task) self.stat.inc("spider:request")
132
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/base.py
Spider.srv_process_task
(self, task: Task)
651
686
def srv_process_task(self, task: Task) -> None: task.network_try_count += 1 is_valid, reason = self.check_task_limits(task) if is_valid: grab = self.create_grab_instance() self.process_grab_proxy(task, grab) self.stat.inc("spider:request-network") self.stat.inc("spider:task-%s-network" % task.name) try: result: dict[str, Any] = { "ok": True, "ecode": None, "emsg": None, "grab": grab, "task": task, "exc": None, "doc": None, } try: result["doc"] = grab.request(task.request) except ( GrabNetworkError, GrabInvalidUrl, GrabInvalidResponse, GrabTooManyRedirectsError, ) as ex: result.update({"ok": False, "exc": ex}) self.task_dispatcher.input_queue.put((result, task, None)) finally: pass else: self.log_rejected_task(task, reason) handler = self.get_fallback_handler(task) if handler: handler(task)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/base.py#L651-L686
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 19, 20, 21, 27, 28, 30, 32, 33, 34, 35 ]
58.333333
[]
0
false
92.476489
36
4
100
0
def srv_process_task(self, task: Task) -> None: task.network_try_count += 1 is_valid, reason = self.check_task_limits(task) if is_valid: grab = self.create_grab_instance() self.process_grab_proxy(task, grab) self.stat.inc("spider:request-network") self.stat.inc("spider:task-%s-network" % task.name) try: result: dict[str, Any] = { "ok": True, "ecode": None, "emsg": None, "grab": grab, "task": task, "exc": None, "doc": None, } try: result["doc"] = grab.request(task.request) except ( GrabNetworkError, GrabInvalidUrl, GrabInvalidResponse, GrabTooManyRedirectsError, ) as ex: result.update({"ok": False, "exc": ex}) self.task_dispatcher.input_queue.put((result, task, None)) finally: pass else: self.log_rejected_task(task, reason) handler = self.get_fallback_handler(task) if handler: handler(task)
133
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/memory.py
MemoryTaskQueue.__init__
(self)
12
15
def __init__(self) -> None: super().__init__() self.queue_object: PriorityQueue[tuple[int, Task]] = PriorityQueue() self.schedule_list: list[tuple[datetime, Task]] = []
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/memory.py#L12-L15
1
[ 0, 1, 2, 3 ]
100
[]
0
true
85.714286
4
1
100
0
def __init__(self) -> None: super().__init__() self.queue_object: PriorityQueue[tuple[int, Task]] = PriorityQueue() self.schedule_list: list[tuple[datetime, Task]] = []
134
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/memory.py
MemoryTaskQueue.put
( self, task: Task, priority: int, schedule_time: None | datetime = None )
17
23
def put( self, task: Task, priority: int, schedule_time: None | datetime = None ) -> None: if schedule_time is None: self.queue_object.put((priority, task)) else: self.schedule_list.append((schedule_time, task))
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/memory.py#L17-L23
1
[ 0, 3, 4 ]
42.857143
[ 6 ]
14.285714
false
85.714286
7
2
85.714286
0
def put( self, task: Task, priority: int, schedule_time: None | datetime = None ) -> None: if schedule_time is None: self.queue_object.put((priority, task)) else: self.schedule_list.append((schedule_time, task))
135
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/memory.py
MemoryTaskQueue.get
(self)
return task
25
40
def get(self) -> Task: now = datetime.utcnow() removed_indexes = [] for idx, item in enumerate(self.schedule_list): schedule_time, task = item if schedule_time <= now: self.put(task, 1) removed_indexes.append(idx) self.schedule_list = [ x for idx, x in enumerate(self.schedule_list) if idx not in removed_indexes ] _, task = self.queue_object.get(block=False) return task
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/memory.py#L25-L40
1
[ 0, 1, 2, 3, 4, 9, 10, 13, 14, 15 ]
62.5
[ 5, 6, 7, 8 ]
25
false
85.714286
16
4
75
0
def get(self) -> Task: now = datetime.utcnow() removed_indexes = [] for idx, item in enumerate(self.schedule_list): schedule_time, task = item if schedule_time <= now: self.put(task, 1) removed_indexes.append(idx) self.schedule_list = [ x for idx, x in enumerate(self.schedule_list) if idx not in removed_indexes ] _, task = self.queue_object.get(block=False) return task
136
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/memory.py
MemoryTaskQueue.size
(self)
return self.queue_object.qsize() + len(self.schedule_list)
42
43
def size(self) -> int: return self.queue_object.qsize() + len(self.schedule_list)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/memory.py#L42-L43
1
[ 0, 1 ]
100
[]
0
true
85.714286
2
1
100
0
def size(self) -> int: return self.queue_object.qsize() + len(self.schedule_list)
137
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/memory.py
MemoryTaskQueue.clear
(self)
45
49
def clear(self) -> None: with suppress(Empty): while True: self.queue_object.get(False) self.schedule_list = []
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/memory.py#L45-L49
1
[ 0, 1, 2, 3, 4 ]
100
[]
0
true
85.714286
5
3
100
0
def clear(self) -> None: with suppress(Empty): while True: self.queue_object.get(False) self.schedule_list = []
138
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/base.py
BaseTaskQueue.random_queue_name
(self)
return "task_queue_{}".format(str(uuid4()).replace("-", "_"))
15
16
def random_queue_name(self) -> str: return "task_queue_{}".format(str(uuid4()).replace("-", "_"))
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/base.py#L15-L16
1
[ 0 ]
50
[ 1 ]
50
false
91.666667
2
1
50
0
def random_queue_name(self) -> str: return "task_queue_{}".format(str(uuid4()).replace("-", "_"))
139
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/base.py
BaseTaskQueue.put
( self, task: Task, priority: int, schedule_time: None | datetime = None, )
18
24
def put( self, task: Task, priority: int, schedule_time: None | datetime = None, ) -> None: # pragma: no cover raise NotImplementedError
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/base.py#L18-L24
1
[]
0
[]
0
false
91.666667
7
1
100
0
def put( self, task: Task, priority: int, schedule_time: None | datetime = None, ) -> None: # pragma: no cover raise NotImplementedError
140
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/base.py
BaseTaskQueue.get
(self)
Return `Task` object or raise `Queue.Empty` exception. @returns: `grab.spider.task.Task` object @raises: `Queue.Empty` exception
Return `Task` object or raise `Queue.Empty` exception.
26
32
def get(self) -> Task: # pragma: no cover """Return `Task` object or raise `Queue.Empty` exception. @returns: `grab.spider.task.Task` object @raises: `Queue.Empty` exception """ raise NotImplementedError
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/base.py#L26-L32
1
[]
0
[]
0
false
91.666667
7
1
100
4
def get(self) -> Task: # pragma: no cover raise NotImplementedError
141
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/base.py
BaseTaskQueue.size
(self)
35
36
def size(self) -> int: # pragma: no cover raise NotImplementedError
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/base.py#L35-L36
1
[]
0
[]
0
false
91.666667
2
1
100
0
def size(self) -> int: # pragma: no cover raise NotImplementedError
142
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/base.py
BaseTaskQueue.clear
(self)
Remove all tasks from the queue.
Remove all tasks from the queue.
38
40
def clear(self) -> None: # pragma: no cover """Remove all tasks from the queue.""" raise NotImplementedError
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/base.py#L38-L40
1
[]
0
[]
0
false
91.666667
3
1
100
1
def clear(self) -> None: # pragma: no cover raise NotImplementedError
143
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/queue_backend/base.py
BaseTaskQueue.close
(self)
42
43
def close(self) -> None: # pragma: no cover raise NotImplementedError
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/queue_backend/base.py#L42-L43
1
[]
0
[]
0
false
91.666667
2
1
100
0
def close(self) -> None: # pragma: no cover raise NotImplementedError
144
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/network.py
BaseNetworkService.get_active_threads_number
(self)
18
19
def get_active_threads_number(self) -> int: # pragma: no cover raise NotImplementedError
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/network.py#L18-L19
1
[]
0
[]
0
false
94.594595
2
1
100
0
def get_active_threads_number(self) -> int: # pragma: no cover raise NotImplementedError
160
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/network.py
NetworkServiceThreaded.__init__
( self, fatal_error_queue: Queue[FatalErrorQueueItem], thread_number: int, process_task: Callable[[Task], None], get_task_from_queue: Callable[[], None | Literal[True] | Task], )
23
37
def __init__( self, fatal_error_queue: Queue[FatalErrorQueueItem], thread_number: int, process_task: Callable[[Task], None], get_task_from_queue: Callable[[], None | Literal[True] | Task], ) -> None: super().__init__(fatal_error_queue) self.thread_number = thread_number self.process_task = process_task self.get_task_from_queue = get_task_from_queue self.worker_pool = [] for _ in range(self.thread_number): self.worker_pool.append(self.create_worker(self.worker_callback)) self.register_workers(self.worker_pool)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/network.py#L23-L37
1
[ 0, 7, 8, 9, 10, 11, 12, 13, 14 ]
60
[]
0
false
94.594595
15
2
100
0
def __init__( self, fatal_error_queue: Queue[FatalErrorQueueItem], thread_number: int, process_task: Callable[[Task], None], get_task_from_queue: Callable[[], None | Literal[True] | Task], ) -> None: super().__init__(fatal_error_queue) self.thread_number = thread_number self.process_task = process_task self.get_task_from_queue = get_task_from_queue self.worker_pool = [] for _ in range(self.thread_number): self.worker_pool.append(self.create_worker(self.worker_callback)) self.register_workers(self.worker_pool)
161
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/network.py
NetworkServiceThreaded.get_active_threads_number
(self)
return sum( 1 for x in self.iterate_workers(self.worker_registry) if x.is_busy_event.is_set() )
39
44
def get_active_threads_number(self) -> int: return sum( 1 for x in self.iterate_workers(self.worker_registry) if x.is_busy_event.is_set() )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/network.py#L39-L44
1
[ 0, 1 ]
33.333333
[]
0
false
94.594595
6
1
100
0
def get_active_threads_number(self) -> int: return sum( 1 for x in self.iterate_workers(self.worker_registry) if x.is_busy_event.is_set() )
162
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/network.py
NetworkServiceThreaded.worker_callback
(self, worker: ServiceWorker)
47
62
def worker_callback(self, worker: ServiceWorker) -> None: while not worker.stop_event.is_set(): worker.process_pause_signal() try: task = self.get_task_from_queue() except Empty: time.sleep(0.1) else: if task is None or task is True: time.sleep(0.1) else: worker.is_busy_event.set() try: self.process_task(task) finally: worker.is_busy_event.clear()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/network.py#L47-L62
1
[ 0, 1, 2, 3, 4, 8, 9, 11, 12, 13, 15 ]
68.75
[ 5, 6 ]
12.5
false
94.594595
16
5
87.5
0
def worker_callback(self, worker: ServiceWorker) -> None: while not worker.stop_event.is_set(): worker.process_pause_signal() try: task = self.get_task_from_queue() except Empty: time.sleep(0.1) else: if task is None or task is True: time.sleep(0.1) else: worker.is_busy_event.set() try: self.process_task(task) finally: worker.is_busy_event.clear()
163
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/parser.py
ParserService.__init__
( self, fatal_error_queue: Queue[FatalErrorQueueItem], pool_size: int, task_dispatcher: TaskDispatcherService, stat: Stat, parser_requests_per_process: int, find_task_handler: Callable[[Task], Callable[..., None]], )
22
42
def __init__( self, fatal_error_queue: Queue[FatalErrorQueueItem], pool_size: int, task_dispatcher: TaskDispatcherService, stat: Stat, parser_requests_per_process: int, find_task_handler: Callable[[Task], Callable[..., None]], ) -> None: super().__init__(fatal_error_queue) self.task_dispatcher = task_dispatcher self.stat = stat self.parser_requests_per_process = parser_requests_per_process self.find_task_handler = find_task_handler self.input_queue: Queue[Any] = Queue() self.pool_size = pool_size self.workers_pool = [] for _ in range(self.pool_size): self.workers_pool.append(self.create_worker(self.worker_callback)) self.supervisor = self.create_worker(self.supervisor_callback) self.register_workers(self.workers_pool, self.supervisor)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/parser.py#L22-L42
1
[ 0, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20 ]
61.904762
[]
0
false
100
21
2
100
0
def __init__( self, fatal_error_queue: Queue[FatalErrorQueueItem], pool_size: int, task_dispatcher: TaskDispatcherService, stat: Stat, parser_requests_per_process: int, find_task_handler: Callable[[Task], Callable[..., None]], ) -> None: super().__init__(fatal_error_queue) self.task_dispatcher = task_dispatcher self.stat = stat self.parser_requests_per_process = parser_requests_per_process self.find_task_handler = find_task_handler self.input_queue: Queue[Any] = Queue() self.pool_size = pool_size self.workers_pool = [] for _ in range(self.pool_size): self.workers_pool.append(self.create_worker(self.worker_callback)) self.supervisor = self.create_worker(self.supervisor_callback) self.register_workers(self.workers_pool, self.supervisor)
164
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/parser.py
ParserService.check_pool_health
(self)
44
54
def check_pool_health(self) -> None: to_remove = [] for worker in self.workers_pool: if not worker.is_alive(): self.stat.inc("parser:worker-restarted") new_worker = self.create_worker(self.worker_callback) self.workers_pool.append(new_worker) new_worker.start() to_remove.append(worker) for worker in to_remove: self.workers_pool.remove(worker)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/parser.py#L44-L54
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 ]
100
[]
0
true
100
11
4
100
0
def check_pool_health(self) -> None: to_remove = [] for worker in self.workers_pool: if not worker.is_alive(): self.stat.inc("parser:worker-restarted") new_worker = self.create_worker(self.worker_callback) self.workers_pool.append(new_worker) new_worker.start() to_remove.append(worker) for worker in to_remove: self.workers_pool.remove(worker)
165
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/parser.py
ParserService.supervisor_callback
(self, worker: ServiceWorker)
56
60
def supervisor_callback(self, worker: ServiceWorker) -> None: while not worker.stop_event.is_set(): worker.process_pause_signal() self.check_pool_health() time.sleep(1)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/parser.py#L56-L60
1
[ 0, 1, 2, 3, 4 ]
100
[]
0
true
100
5
2
100
0
def supervisor_callback(self, worker: ServiceWorker) -> None: while not worker.stop_event.is_set(): worker.process_pause_signal() self.check_pool_health() time.sleep(1)
166
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/parser.py
ParserService.worker_callback
(self, worker: ServiceWorker)
62
92
def worker_callback(self, worker: ServiceWorker) -> None: process_request_count = 0 while not worker.stop_event.is_set(): worker.process_pause_signal() try: result, task = self.input_queue.get(True, 0.1) except Empty: pass else: worker.is_busy_event.set() try: process_request_count += 1 try: handler = self.find_task_handler(task) except NoTaskHandler as ex: self.task_dispatcher.input_queue.put( (ex, task, {"exc_info": sys.exc_info()}) ) self.stat.inc("parser:handler-not-found") else: self.execute_task_handler(handler, result, task) self.stat.inc("parser:handler-processed") if self.parser_requests_per_process and ( process_request_count >= self.parser_requests_per_process ): self.stat.inc( "parser:handler-req-limit", ) return finally: worker.is_busy_event.clear()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/parser.py#L62-L92
1
[ 0, 1, 2, 3, 4, 5, 6, 7, 9, 10, 11, 12, 13, 14, 15, 18, 20, 21, 22, 25, 28, 30 ]
70.967742
[]
0
false
100
31
6
100
0
def worker_callback(self, worker: ServiceWorker) -> None: process_request_count = 0 while not worker.stop_event.is_set(): worker.process_pause_signal() try: result, task = self.input_queue.get(True, 0.1) except Empty: pass else: worker.is_busy_event.set() try: process_request_count += 1 try: handler = self.find_task_handler(task) except NoTaskHandler as ex: self.task_dispatcher.input_queue.put( (ex, task, {"exc_info": sys.exc_info()}) ) self.stat.inc("parser:handler-not-found") else: self.execute_task_handler(handler, result, task) self.stat.inc("parser:handler-processed") if self.parser_requests_per_process and ( process_request_count >= self.parser_requests_per_process ): self.stat.inc( "parser:handler-req-limit", ) return finally: worker.is_busy_event.clear()
167
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/parser.py
ParserService.execute_task_handler
( self, handler: Callable[[Grab, Task], None], result: NetworkResult, task: Task )
94
116
def execute_task_handler( self, handler: Callable[[Grab, Task], None], result: NetworkResult, task: Task ) -> None: try: handler_result = handler(result["doc"], task) if handler_result is None: pass else: for item in handler_result: self.task_dispatcher.input_queue.put( (item, task, None), ) except Exception as ex: self.task_dispatcher.input_queue.put( ( ex, task, { "exc_info": sys.exc_info(), "from": "parser", }, ) )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/parser.py#L94-L116
1
[ 0, 3, 4, 5, 6, 8, 9, 12, 13 ]
39.130435
[]
0
false
100
23
4
100
0
def execute_task_handler( self, handler: Callable[[Grab, Task], None], result: NetworkResult, task: Task ) -> None: try: handler_result = handler(result["doc"], task) if handler_result is None: pass else: for item in handler_result: self.task_dispatcher.input_queue.put( (item, task, None), ) except Exception as ex: self.task_dispatcher.input_queue.put( ( ex, task, { "exc_info": sys.exc_info(), "from": "parser", }, ) )
168
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/task_dispatcher.py
TaskDispatcherService.__init__
( self, fatal_error_queue: Queue[FatalErrorQueueItem], process_service_result: Callable[[Any, Task, None | dict[str, Any]], Any], )
13
22
def __init__( self, fatal_error_queue: Queue[FatalErrorQueueItem], process_service_result: Callable[[Any, Task, None | dict[str, Any]], Any], ): super().__init__(fatal_error_queue) self.process_service_result = process_service_result self.input_queue: Queue[Any] = Queue() self.worker = self.create_worker(self.worker_callback) self.register_workers(self.worker)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/task_dispatcher.py#L13-L22
1
[ 0, 5, 6, 7, 8, 9 ]
60
[]
0
false
100
10
1
100
0
def __init__( self, fatal_error_queue: Queue[FatalErrorQueueItem], process_service_result: Callable[[Any, Task, None | dict[str, Any]], Any], ): super().__init__(fatal_error_queue) self.process_service_result = process_service_result self.input_queue: Queue[Any] = Queue() self.worker = self.create_worker(self.worker_callback) self.register_workers(self.worker)
169
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/task_dispatcher.py
TaskDispatcherService.start
(self)
24
25
def start(self) -> None: self.worker.start()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/task_dispatcher.py#L24-L25
1
[ 0, 1 ]
100
[]
0
true
100
2
1
100
0
def start(self) -> None: self.worker.start()
170
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/task_dispatcher.py
TaskDispatcherService.worker_callback
(self, worker: ServiceWorker)
27
35
def worker_callback(self, worker: ServiceWorker) -> None: while not worker.stop_event.is_set(): worker.process_pause_signal() try: result, task, meta = self.input_queue.get(True, 0.1) except Empty: pass else: self.process_service_result(result, task, meta)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/task_dispatcher.py#L27-L35
1
[ 0, 1, 2, 3, 4, 5, 6, 8 ]
88.888889
[]
0
false
100
9
3
100
0
def worker_callback(self, worker: ServiceWorker) -> None: while not worker.stop_event.is_set(): worker.process_pause_signal() try: result, task, meta = self.input_queue.get(True, 0.1) except Empty: pass else: self.process_service_result(result, task, meta)
171
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
ServiceWorker.__init__
( self, fatal_error_queue: Queue[FatalErrorQueueItem], worker_callback: Callable[..., Any], )
17
32
def __init__( self, fatal_error_queue: Queue[FatalErrorQueueItem], worker_callback: Callable[..., Any], ) -> None: self.fatal_error_queue = fatal_error_queue self.thread = Thread( target=self.worker_callback_wrapper(worker_callback), args=[self] ) self.thread.daemon = True self.thread.name = self.build_thread_name(worker_callback) self.pause_event = Event() self.stop_event = Event() self.resume_event = Event() self.activity_paused = Event() self.is_busy_event = Event()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L17-L32
1
[ 0, 5, 6, 9, 10, 11, 12, 13, 14, 15 ]
62.5
[]
0
false
100
16
1
100
0
def __init__( self, fatal_error_queue: Queue[FatalErrorQueueItem], worker_callback: Callable[..., Any], ) -> None: self.fatal_error_queue = fatal_error_queue self.thread = Thread( target=self.worker_callback_wrapper(worker_callback), args=[self] ) self.thread.daemon = True self.thread.name = self.build_thread_name(worker_callback) self.pause_event = Event() self.stop_event = Event() self.resume_event = Event() self.activity_paused = Event() self.is_busy_event = Event()
172
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
ServiceWorker.build_thread_name
(self, worker_callback: Callable[..., Any])
return "worker:%s:%s" % (cls_name, worker_callback.__name__)
34
40
def build_thread_name(self, worker_callback: Callable[..., Any]) -> str: cls_name = ( worker_callback.__self__.__class__.__name__ if hasattr(worker_callback, "__self__") else "NA" ) return "worker:%s:%s" % (cls_name, worker_callback.__name__)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L34-L40
1
[ 0, 1, 6 ]
42.857143
[]
0
false
100
7
1
100
0
def build_thread_name(self, worker_callback: Callable[..., Any]) -> str: cls_name = ( worker_callback.__self__.__class__.__name__ if hasattr(worker_callback, "__self__") else "NA" ) return "worker:%s:%s" % (cls_name, worker_callback.__name__)
173
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
ServiceWorker.worker_callback_wrapper
( self, callback: Callable[..., Any] )
return wrapper
42
57
def worker_callback_wrapper( self, callback: Callable[..., Any] ) -> Callable[..., None]: def wrapper(*args: Any, **kwargs: Any) -> None: try: callback(*args, **kwargs) except Exception as ex: logger.error("Spider Service Fatal Error", exc_info=ex) # pylint: disable=deprecated-typing-alias self.fatal_error_queue.put( cast( Tuple[Type[Exception], Exception, TracebackType], sys.exc_info() ) ) return wrapper
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L42-L57
1
[ 0, 3, 4, 5, 6, 7, 8, 9, 14, 15 ]
62.5
[]
0
false
100
16
3
100
0
def worker_callback_wrapper( self, callback: Callable[..., Any] ) -> Callable[..., None]: def wrapper(*args: Any, **kwargs: Any) -> None: try: callback(*args, **kwargs) except Exception as ex: logger.error("Spider Service Fatal Error", exc_info=ex) # pylint: disable=deprecated-typing-alias self.fatal_error_queue.put( cast( Tuple[Type[Exception], Exception, TracebackType], sys.exc_info() ) ) return wrapper
174
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
ServiceWorker.start
(self)
59
60
def start(self) -> None: self.thread.start()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L59-L60
1
[ 0, 1 ]
100
[]
0
true
100
2
1
100
0
def start(self) -> None: self.thread.start()
175
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
ServiceWorker.stop
(self)
62
63
def stop(self) -> None: self.stop_event.set()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L62-L63
1
[ 0, 1 ]
100
[]
0
true
100
2
1
100
0
def stop(self) -> None: self.stop_event.set()
176
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
ServiceWorker.process_pause_signal
(self)
65
68
def process_pause_signal(self) -> None: if self.pause_event.is_set(): self.activity_paused.set() self.resume_event.wait()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L65-L68
1
[ 0, 1, 2, 3 ]
100
[]
0
true
100
4
2
100
0
def process_pause_signal(self) -> None: if self.pause_event.is_set(): self.activity_paused.set() self.resume_event.wait()
177
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
ServiceWorker.pause
(self)
70
77
def pause(self) -> None: self.resume_event.clear() self.pause_event.set() while True: if self.activity_paused.wait(0.1): break if not self.is_alive(): break
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L70-L77
1
[ 0, 1, 2, 3, 4, 5, 6, 7 ]
100
[]
0
true
100
8
4
100
0
def pause(self) -> None: self.resume_event.clear() self.pause_event.set() while True: if self.activity_paused.wait(0.1): break if not self.is_alive(): break
178
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
ServiceWorker.resume
(self)
79
82
def resume(self) -> None: self.pause_event.clear() self.activity_paused.clear() self.resume_event.set()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L79-L82
1
[ 0, 1, 2, 3 ]
100
[]
0
true
100
4
1
100
0
def resume(self) -> None: self.pause_event.clear() self.activity_paused.clear() self.resume_event.set()
179
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
ServiceWorker.is_alive
(self)
return self.thread.is_alive()
84
85
def is_alive(self) -> bool: return self.thread.is_alive()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L84-L85
1
[ 0, 1 ]
100
[]
0
true
100
2
1
100
0
def is_alive(self) -> bool: return self.thread.is_alive()
180
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
BaseService.__init__
(self, fatal_error_queue: Queue[FatalErrorQueueItem])
89
91
def __init__(self, fatal_error_queue: Queue[FatalErrorQueueItem]) -> None: self.fatal_error_queue = fatal_error_queue self.worker_registry: list[ServiceWorker] = []
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L89-L91
1
[ 0, 1, 2 ]
100
[]
0
true
100
3
1
100
0
def __init__(self, fatal_error_queue: Queue[FatalErrorQueueItem]) -> None: self.fatal_error_queue = fatal_error_queue self.worker_registry: list[ServiceWorker] = []
181
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
BaseService.create_worker
(self, worker_action: Callable[..., None])
return ServiceWorker(self.fatal_error_queue, worker_action)
93
94
def create_worker(self, worker_action: Callable[..., None]) -> ServiceWorker: return ServiceWorker(self.fatal_error_queue, worker_action)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L93-L94
1
[ 0, 1 ]
100
[]
0
true
100
2
1
100
0
def create_worker(self, worker_action: Callable[..., None]) -> ServiceWorker: return ServiceWorker(self.fatal_error_queue, worker_action)
182
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
BaseService.iterate_workers
(self, objects: list[ServiceWorker])
96
102
def iterate_workers(self, objects: list[ServiceWorker]) -> Iterable[ServiceWorker]: for obj in objects: assert isinstance(obj, (ServiceWorker, list)) if isinstance(obj, ServiceWorker): yield obj elif isinstance(obj, list): yield from obj
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L96-L102
1
[ 0, 1, 2, 3, 4, 5, 6 ]
100
[]
0
true
100
7
5
100
0
def iterate_workers(self, objects: list[ServiceWorker]) -> Iterable[ServiceWorker]: for obj in objects: assert isinstance(obj, (ServiceWorker, list)) if isinstance(obj, ServiceWorker): yield obj elif isinstance(obj, list): yield from obj
183
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
BaseService.start
(self)
104
106
def start(self) -> None: for worker in self.iterate_workers(self.worker_registry): worker.start()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L104-L106
1
[ 0, 1, 2 ]
100
[]
0
true
100
3
2
100
0
def start(self) -> None: for worker in self.iterate_workers(self.worker_registry): worker.start()
184
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
BaseService.stop
(self)
108
110
def stop(self) -> None: for worker in self.iterate_workers(self.worker_registry): worker.stop()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L108-L110
1
[ 0, 1, 2 ]
100
[]
0
true
100
3
2
100
0
def stop(self) -> None: for worker in self.iterate_workers(self.worker_registry): worker.stop()
185
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
BaseService.pause
(self)
112
114
def pause(self) -> None: for worker in self.iterate_workers(self.worker_registry): worker.pause()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L112-L114
1
[ 0, 1, 2 ]
100
[]
0
true
100
3
2
100
0
def pause(self) -> None: for worker in self.iterate_workers(self.worker_registry): worker.pause()
186
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
BaseService.resume
(self)
116
118
def resume(self) -> None: for worker in self.iterate_workers(self.worker_registry): worker.resume()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L116-L118
1
[ 0, 1, 2 ]
100
[]
0
true
100
3
2
100
0
def resume(self) -> None: for worker in self.iterate_workers(self.worker_registry): worker.resume()
187
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
BaseService.register_workers
(self, *args: Any)
120
121
def register_workers(self, *args: Any) -> None: self.worker_registry = list(args)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L120-L121
1
[ 0, 1 ]
100
[]
0
true
100
2
1
100
0
def register_workers(self, *args: Any) -> None: self.worker_registry = list(args)
188
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
BaseService.is_busy
(self)
return any( x.is_busy_event.is_set() for x in self.iterate_workers(self.worker_registry) )
123
126
def is_busy(self) -> bool: return any( x.is_busy_event.is_set() for x in self.iterate_workers(self.worker_registry) )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L123-L126
1
[ 0, 1 ]
50
[]
0
false
100
4
1
100
0
def is_busy(self) -> bool: return any( x.is_busy_event.is_set() for x in self.iterate_workers(self.worker_registry) )
189
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/base.py
BaseService.is_alive
(self)
return any(x.is_alive() for x in self.iterate_workers(self.worker_registry))
128
129
def is_alive(self) -> bool: return any(x.is_alive() for x in self.iterate_workers(self.worker_registry))
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/base.py#L128-L129
1
[ 0, 1 ]
100
[]
0
true
100
2
1
100
0
def is_alive(self) -> bool: return any(x.is_alive() for x in self.iterate_workers(self.worker_registry))
190
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/task_generator.py
TaskGeneratorService.__init__
( self, fatal_error_queue: Queue[FatalErrorQueueItem], real_generator: Iterator[Task], thread_number: int, get_task_queue: Callable[[], BaseTaskQueue], parser_service: ParserService, task_dispatcher: TaskDispatcherService, )
16
32
def __init__( self, fatal_error_queue: Queue[FatalErrorQueueItem], real_generator: Iterator[Task], thread_number: int, get_task_queue: Callable[[], BaseTaskQueue], parser_service: ParserService, task_dispatcher: TaskDispatcherService, ) -> None: super().__init__(fatal_error_queue) self.real_generator = real_generator self.task_queue_threshold = max(200, thread_number * 2) self.get_task_queue = get_task_queue self.parser_service = parser_service self.task_dispatcher = task_dispatcher self.worker = self.create_worker(self.worker_callback) self.register_workers(self.worker)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/task_generator.py#L16-L32
1
[ 0, 9, 10, 11, 12, 13, 14, 15, 16 ]
52.941176
[]
0
false
94.285714
17
1
100
0
def __init__( self, fatal_error_queue: Queue[FatalErrorQueueItem], real_generator: Iterator[Task], thread_number: int, get_task_queue: Callable[[], BaseTaskQueue], parser_service: ParserService, task_dispatcher: TaskDispatcherService, ) -> None: super().__init__(fatal_error_queue) self.real_generator = real_generator self.task_queue_threshold = max(200, thread_number * 2) self.get_task_queue = get_task_queue self.parser_service = parser_service self.task_dispatcher = task_dispatcher self.worker = self.create_worker(self.worker_callback) self.register_workers(self.worker)
191
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/spider/service/task_generator.py
TaskGeneratorService.worker_callback
(self, worker: ServiceWorker)
34
56
def worker_callback(self, worker: ServiceWorker) -> None: # at this point I guess the task queue is set # i.e. "spider.run()" is called task_queue = self.get_task_queue() while not worker.stop_event.is_set(): worker.process_pause_signal() queue_size = max( task_queue.size(), self.parser_service.input_queue.qsize(), ) if queue_size < self.task_queue_threshold: try: for _ in range(self.task_queue_threshold - queue_size): if worker.pause_event.is_set(): return task = next(self.real_generator) self.task_dispatcher.input_queue.put( (task, None, {"source": "task_generator"}) ) except StopIteration: return else: time.sleep(0.1)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/spider/service/task_generator.py#L34-L56
1
[ 0, 1, 2, 3, 4, 5, 6, 10, 11, 12, 13, 15, 16, 19, 20 ]
65.217391
[ 14, 22 ]
8.695652
false
94.285714
23
6
91.304348
0
def worker_callback(self, worker: ServiceWorker) -> None: # at this point I guess the task queue is set # i.e. "spider.run()" is called task_queue = self.get_task_queue() while not worker.stop_event.is_set(): worker.process_pause_signal() queue_size = max( task_queue.size(), self.parser_service.input_queue.qsize(), ) if queue_size < self.task_queue_threshold: try: for _ in range(self.task_queue_threshold - queue_size): if worker.pause_event.is_set(): return task = next(self.real_generator) self.task_dispatcher.input_queue.put( (task, None, {"source": "task_generator"}) ) except StopIteration: return else: time.sleep(0.1)
192
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/metrics.py
in_unit
(num: int, unit: str)
return num
8
17
def in_unit(num: int, unit: str) -> int | float: if unit == "b": return num if unit == "kb": return round(num / float(KB), 2) if unit == "mb": return round(num / float(MB), 2) if unit == "gb": return round(num / float(GB), 2) return num
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/metrics.py#L8-L17
1
[ 0 ]
10
[ 1, 2, 3, 4, 5, 6, 7, 8, 9 ]
90
false
27.272727
10
5
10
0
def in_unit(num: int, unit: str) -> int | float: if unit == "b": return num if unit == "kb": return round(num / float(KB), 2) if unit == "mb": return round(num / float(MB), 2) if unit == "gb": return round(num / float(GB), 2) return num
193
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/metrics.py
format_traffic_value
(num: int)
return "%s GB" % in_unit(num, "gb")
20
27
def format_traffic_value(num: int) -> str: if num < KB: return "%s B" % in_unit(num, "b") if num < MB: return "%s KB" % in_unit(num, "kb") if num < GB: return "%s MB" % in_unit(num, "mb") return "%s GB" % in_unit(num, "gb")
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/metrics.py#L20-L27
1
[ 0 ]
12.5
[ 1, 2, 3, 4, 5, 6, 7 ]
87.5
false
27.272727
8
4
12.5
0
def format_traffic_value(num: int) -> str: if num < KB: return "%s B" % in_unit(num, "b") if num < MB: return "%s KB" % in_unit(num, "kb") if num < GB: return "%s MB" % in_unit(num, "mb") return "%s GB" % in_unit(num, "gb")
194
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/structures.py
merge_with_dict
( hdr1: MutableMapping[str, Any], hdr2: Mapping[str, Any], replace: bool, )
return hdr1
7
15
def merge_with_dict( hdr1: MutableMapping[str, Any], hdr2: Mapping[str, Any], replace: bool, ) -> MutableMapping[str, Any]: for key, val in hdr2.items(): if replace or key not in hdr1: hdr1[key] = val return hdr1
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/structures.py#L7-L15
1
[ 0, 5, 6, 7, 8 ]
55.555556
[]
0
false
100
9
4
100
0
def merge_with_dict( hdr1: MutableMapping[str, Any], hdr2: Mapping[str, Any], replace: bool, ) -> MutableMapping[str, Any]: for key, val in hdr2.items(): if replace or key not in hdr1: hdr1[key] = val return hdr1
196
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/types.py
resolve_entity
( base_type: type[T], entity: None | T | type[T], default: type[T], )
return cast(typing.Type[T], entity)()
16
32
def resolve_entity( base_type: type[T], entity: None | T | type[T], default: type[T], ) -> T: if entity and ( not isinstance(entity, base_type) and (not inspect.isclass(entity) or not issubclass(entity, base_type)) ): raise TypeError("Invalid {} entity: {}".format(base_type, entity)) if entity is None: assert issubclass(default, base_type) return default() if isinstance(entity, base_type): return entity # pylint: disable=deprecated-typing-alias return cast(typing.Type[T], entity)()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/types.py#L16-L32
1
[ 0, 5, 10, 11, 12, 13, 14, 15, 16 ]
52.941176
[ 9 ]
5.882353
false
92.857143
17
8
94.117647
0
def resolve_entity( base_type: type[T], entity: None | T | type[T], default: type[T], ) -> T: if entity and ( not isinstance(entity, base_type) and (not inspect.isclass(entity) or not issubclass(entity, base_type)) ): raise TypeError("Invalid {} entity: {}".format(base_type, entity)) if entity is None: assert issubclass(default, base_type) return default() if isinstance(entity, base_type): return entity # pylint: disable=deprecated-typing-alias return cast(typing.Type[T], entity)()
197
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/timeout.py
Timeout.__init__
( self, total: None | float | UndefinedParam = UNDEFINED_PARAM, connect: None | float | UndefinedParam = UNDEFINED_PARAM, read: None | float | UndefinedParam = UNDEFINED_PARAM, )
Timeout constructor. Unspecified total timeout is set to None. Unspecified connect timeout is set to total timeout. Unspecified read timeout is set to total timeout.
Timeout constructor.
17
31
def __init__( self, total: None | float | UndefinedParam = UNDEFINED_PARAM, connect: None | float | UndefinedParam = UNDEFINED_PARAM, read: None | float | UndefinedParam = UNDEFINED_PARAM, ): """Timeout constructor. Unspecified total timeout is set to None. Unspecified connect timeout is set to total timeout. Unspecified read timeout is set to total timeout. """ self.total = total if total is not UNDEFINED_PARAM else DEFAULT_TOTAL_TIMEOUT self.connect = connect if connect is not UNDEFINED_PARAM else self.total self.read = read if read is not UNDEFINED_PARAM else self.total
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/timeout.py#L17-L31
1
[ 0, 11, 12, 13, 14 ]
33.333333
[]
0
false
92.857143
15
1
100
5
def __init__( self, total: None | float | UndefinedParam = UNDEFINED_PARAM, connect: None | float | UndefinedParam = UNDEFINED_PARAM, read: None | float | UndefinedParam = UNDEFINED_PARAM, ): self.total = total if total is not UNDEFINED_PARAM else DEFAULT_TOTAL_TIMEOUT self.connect = connect if connect is not UNDEFINED_PARAM else self.total self.read = read if read is not UNDEFINED_PARAM else self.total
198
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/timeout.py
Timeout.__repr__
(self)
return "{}(connect={!r}, read={!r}, total={!r})".format( type(self).__name__, self.connect, self.read, self.total )
33
36
def __repr__(self) -> str: return "{}(connect={!r}, read={!r}, total={!r})".format( type(self).__name__, self.connect, self.read, self.total )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/timeout.py#L33-L36
1
[ 0 ]
25
[ 1 ]
25
false
92.857143
4
1
75
0
def __repr__(self) -> str: return "{}(connect={!r}, read={!r}, total={!r})".format( type(self).__name__, self.connect, self.read, self.total )
199
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
create_cookie
( # pylint: disable=too-many-arguments, too-many-locals *, name: str, value: str, domain: str, comment: None | str = None, comment_url: None | str = None, discard: bool = True, domain_initial_dot: None | bool = None, domain_specified: None | bool = None, expires: None | int = None, path: str = "/", path_specified: None | bool = None, port: None | int = None, port_specified: None | bool = None, rest: None | dict[str, Any] = None, rfc2109: bool = False, secure: bool = False, version: int = 0, httponly: None | bool = None, )
return Cookie( # from required scope name=name, value=value, domain=domain, # from non required scope comment=comment, comment_url=comment_url, discard=discard, domain_initial_dot=domain_initial_dot, domain_specified=domain_specified, expires=expires, path=path, path_specified=path_specified, port=str(port) if port else None, # typeshed bundled with mypy wants str type port_specified=port_specified, rest=new_rest, rfc2109=rfc2109, secure=secure, version=version, )
Create cookielib.Cookie instance.
Create cookielib.Cookie instance.
129
189
def create_cookie( # pylint: disable=too-many-arguments, too-many-locals *, name: str, value: str, domain: str, comment: None | str = None, comment_url: None | str = None, discard: bool = True, domain_initial_dot: None | bool = None, domain_specified: None | bool = None, expires: None | int = None, path: str = "/", path_specified: None | bool = None, port: None | int = None, port_specified: None | bool = None, rest: None | dict[str, Any] = None, rfc2109: bool = False, secure: bool = False, version: int = 0, httponly: None | bool = None, ) -> Cookie: """Create cookielib.Cookie instance.""" # See also type hints for Cookie at # https://github.com/python/typeshed/blob/main/stdlib/http/cookiejar.pyi if domain == "localhost": domain = "" if rest is None: new_rest = {} else: new_rest = copy(rest) if "HttpOnly" not in new_rest: new_rest["HttpOnly"] = httponly if port_specified is None: port_specified = port is not None if domain_specified is None: domain_specified = domain is not None if domain_initial_dot is None: domain_initial_dot = domain.startswith(".") if path_specified is None: path_specified = path is not None return Cookie( # from required scope name=name, value=value, domain=domain, # from non required scope comment=comment, comment_url=comment_url, discard=discard, domain_initial_dot=domain_initial_dot, domain_specified=domain_specified, expires=expires, path=path, path_specified=path_specified, port=str(port) if port else None, # typeshed bundled with mypy wants str type port_specified=port_specified, rest=new_rest, rfc2109=rfc2109, secure=secure, version=version, )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L129-L189
1
[ 0, 23, 24, 26, 27, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41 ]
24.590164
[ 25, 29, 30, 31 ]
6.557377
false
81.927711
61
8
93.442623
1
def create_cookie( # pylint: disable=too-many-arguments, too-many-locals *, name: str, value: str, domain: str, comment: None | str = None, comment_url: None | str = None, discard: bool = True, domain_initial_dot: None | bool = None, domain_specified: None | bool = None, expires: None | int = None, path: str = "/", path_specified: None | bool = None, port: None | int = None, port_specified: None | bool = None, rest: None | dict[str, Any] = None, rfc2109: bool = False, secure: bool = False, version: int = 0, httponly: None | bool = None, ) -> Cookie: # See also type hints for Cookie at # https://github.com/python/typeshed/blob/main/stdlib/http/cookiejar.pyi if domain == "localhost": domain = "" if rest is None: new_rest = {} else: new_rest = copy(rest) if "HttpOnly" not in new_rest: new_rest["HttpOnly"] = httponly if port_specified is None: port_specified = port is not None if domain_specified is None: domain_specified = domain is not None if domain_initial_dot is None: domain_initial_dot = domain.startswith(".") if path_specified is None: path_specified = path is not None return Cookie( # from required scope name=name, value=value, domain=domain, # from non required scope comment=comment, comment_url=comment_url, discard=discard, domain_initial_dot=domain_initial_dot, domain_specified=domain_specified, expires=expires, path=path, path_specified=path_specified, port=str(port) if port else None, # typeshed bundled with mypy wants str type port_specified=port_specified, rest=new_rest, rfc2109=rfc2109, secure=secure, version=version, )
200
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
build_cookie_header
( cookiejar: CookieJar, url: str, headers: Mapping[str, str] )
return mocked_req.get_new_headers().get("Cookie")
Build HTTP Cookie header value for given cookies.
Build HTTP Cookie header value for given cookies.
192
198
def build_cookie_header( cookiejar: CookieJar, url: str, headers: Mapping[str, str] ) -> None | str: """Build HTTP Cookie header value for given cookies.""" mocked_req = MockRequest(url, dict(headers)) cookiejar.add_cookie_header(cast(urllib.request.Request, mocked_req)) return mocked_req.get_new_headers().get("Cookie")
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L192-L198
1
[ 0, 3, 4, 5, 6 ]
71.428571
[]
0
false
81.927711
7
1
100
1
def build_cookie_header( cookiejar: CookieJar, url: str, headers: Mapping[str, str] ) -> None | str: mocked_req = MockRequest(url, dict(headers)) cookiejar.add_cookie_header(cast(urllib.request.Request, mocked_req)) return mocked_req.get_new_headers().get("Cookie")
201
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
build_jar
(cookies: Sequence[Cookie])
return jar
201
205
def build_jar(cookies: Sequence[Cookie]) -> CookieJar: jar = CookieJar() for item in cookies: jar.set_cookie(item) return jar
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L201-L205
1
[ 0 ]
20
[ 1, 2, 3, 4 ]
80
false
81.927711
5
2
20
0
def build_jar(cookies: Sequence[Cookie]) -> CookieJar: jar = CookieJar() for item in cookies: jar.set_cookie(item) return jar
202
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
extract_response_cookies
( req_url: str, req_headers: Mapping[str, Any] | HTTPMessage | HTTPHeaderDict, response_headers: HTTPMessage | HTTPHeaderDict, )
return list(jar)
208
218
def extract_response_cookies( req_url: str, req_headers: Mapping[str, Any] | HTTPMessage | HTTPHeaderDict, response_headers: HTTPMessage | HTTPHeaderDict, ) -> Sequence[Cookie]: jar = CookieJar() jar.extract_cookies( cast(HTTPResponse, MockResponse(response_headers)), cast(urllib.request.Request, MockRequest(req_url, dict(req_headers))), ) return list(jar)
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L208-L218
1
[ 0, 5, 6, 10 ]
36.363636
[]
0
false
81.927711
11
1
100
0
def extract_response_cookies( req_url: str, req_headers: Mapping[str, Any] | HTTPMessage | HTTPHeaderDict, response_headers: HTTPMessage | HTTPHeaderDict, ) -> Sequence[Cookie]: jar = CookieJar() jar.extract_cookies( cast(HTTPResponse, MockResponse(response_headers)), cast(urllib.request.Request, MockRequest(req_url, dict(req_headers))), ) return list(jar)
203
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.__init__
(self, url: str, headers: dict[str, str])
38
42
def __init__(self, url: str, headers: dict[str, str]): self._url = url self._headers = headers self._new_headers: dict[str, Any] = {} self.type = urlparse(self._url).scheme
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L38-L42
1
[ 0, 1, 2, 3, 4 ]
100
[]
0
true
81.927711
5
1
100
0
def __init__(self, url: str, headers: dict[str, str]): self._url = url self._headers = headers self._new_headers: dict[str, Any] = {} self.type = urlparse(self._url).scheme
204
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.get_type
(self)
return self.type
44
45
def get_type(self) -> str: return self.type
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L44-L45
1
[ 0 ]
50
[ 1 ]
50
false
81.927711
2
1
50
0
def get_type(self) -> str: return self.type
205
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.get_host
(self)
return urlparse(self._url).netloc
47
48
def get_host(self) -> str: return urlparse(self._url).netloc
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L47-L48
1
[ 0, 1 ]
100
[]
0
true
81.927711
2
1
100
0
def get_host(self) -> str: return urlparse(self._url).netloc
206
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.get_origin_req_host
(self)
return self.get_host()
50
51
def get_origin_req_host(self) -> str: return self.get_host()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L50-L51
1
[ 0, 1 ]
100
[]
0
true
81.927711
2
1
100
0
def get_origin_req_host(self) -> str: return self.get_host()
207
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.get_full_url
(self)
return urlunparse( [ parsed.scheme, host, parsed.path, parsed.params, parsed.query, parsed.fragment, ] )
53
71
def get_full_url(self) -> str: # Only return the response's URL if the user hadn't set the Host # header if not self._headers.get("Host"): return self._url # If they did set it, retrieve it and reconstruct the expected domain host = self._headers["Host"] parsed = urlparse(self._url) # Reconstruct the URL as we expect it return urlunparse( [ parsed.scheme, host, parsed.path, parsed.params, parsed.query, parsed.fragment, ] )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L53-L71
1
[ 0, 1, 2, 3, 4, 5 ]
31.578947
[ 6, 7, 9 ]
15.789474
false
81.927711
19
2
84.210526
0
def get_full_url(self) -> str: # Only return the response's URL if the user hadn't set the Host # header if not self._headers.get("Host"): return self._url # If they did set it, retrieve it and reconstruct the expected domain host = self._headers["Host"] parsed = urlparse(self._url) # Reconstruct the URL as we expect it return urlunparse( [ parsed.scheme, host, parsed.path, parsed.params, parsed.query, parsed.fragment, ] )
208
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.is_unverifiable
(self)
return True
73
74
def is_unverifiable(self) -> bool: return True
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L73-L74
1
[ 0, 1 ]
100
[]
0
true
81.927711
2
1
100
0
def is_unverifiable(self) -> bool: return True
209
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.has_header
(self, name: str)
return name in self._headers or name in self._new_headers
76
77
def has_header(self, name: str) -> bool: return name in self._headers or name in self._new_headers
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L76-L77
1
[ 0, 1 ]
100
[]
0
true
81.927711
2
2
100
0
def has_header(self, name: str) -> bool: return name in self._headers or name in self._new_headers
210
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.get_header
(self, name: str, default: Any = None)
return self._headers.get(name, self._new_headers.get(name, default))
79
80
def get_header(self, name: str, default: Any = None) -> str: return self._headers.get(name, self._new_headers.get(name, default))
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L79-L80
1
[ 0 ]
50
[ 1 ]
50
false
81.927711
2
1
50
0
def get_header(self, name: str, default: Any = None) -> str: return self._headers.get(name, self._new_headers.get(name, default))
211
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.add_header
(self, key: str, val: str)
Cookielib has no legitimate use for this method. Add it back if you find one.
Cookielib has no legitimate use for this method.
82
89
def add_header(self, key: str, val: str) -> None: """Cookielib has no legitimate use for this method. Add it back if you find one. """ raise NotImplementedError( "Cookie headers should be added with add_unredirected_header()" )
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L82-L89
1
[ 0, 1, 2, 3, 4 ]
62.5
[ 5 ]
12.5
false
81.927711
8
1
87.5
3
def add_header(self, key: str, val: str) -> None: raise NotImplementedError( "Cookie headers should be added with add_unredirected_header()" )
212
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.add_unredirected_header
(self, name: str, value: str)
91
92
def add_unredirected_header(self, name: str, value: str) -> None: self._new_headers[name] = value
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L91-L92
1
[ 0, 1 ]
100
[]
0
true
81.927711
2
1
100
0
def add_unredirected_header(self, name: str, value: str) -> None: self._new_headers[name] = value
213
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.get_new_headers
(self)
return self._new_headers
94
95
def get_new_headers(self) -> dict[str, str]: return self._new_headers
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L94-L95
1
[ 0, 1 ]
100
[]
0
true
81.927711
2
1
100
0
def get_new_headers(self) -> dict[str, str]: return self._new_headers
214
lorien/grab
2d170c31a3335c2e29578b42a5d62ef3efc5d7ee
grab/util/cookies.py
MockRequest.unverifiable
(self)
return self.is_unverifiable()
98
99
def unverifiable(self) -> bool: return self.is_unverifiable()
https://github.com/lorien/grab/blob/2d170c31a3335c2e29578b42a5d62ef3efc5d7ee/project1/grab/util/cookies.py#L98-L99
1
[ 0, 1 ]
100
[]
0
true
81.927711
2
1
100
0
def unverifiable(self) -> bool: return self.is_unverifiable()
215