Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
12 changes: 10 additions & 2 deletions crawlers/lib/crawl.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,11 +6,14 @@
import uuid
from typing import List, Generator
from flask import current_app
from requests import ConnectionError, Timeout, TooManyRedirects

from crawlers.constants import BLOCK_KEY_CALLBACK_URL

from crawlers.lib.platforms.i_crawler import ICrawler
from crawlers.lib.platforms import platforms
from urllib3.exceptions import MaxRetryError
from requests.exceptions import RequestException

logger = logging.getLogger(__name__)

Expand Down Expand Up @@ -54,7 +57,11 @@ def process_block_url(session, block_url) -> None:
f"skip crawl - no callback_url found! - key: {BLOCK_KEY_CALLBACK_URL}, block_data: {block_data}"
)
else:
repos = run_block(block_data)
try:
repos = run_block(block_data)
except (MaxRetryError, ConnectionError, Timeout, TooManyRedirects):
logger.exception("hosting service not reachable - no indexer callback issued")
return
_hoster_session_request(
"PUT", session, url=block_data[BLOCK_KEY_CALLBACK_URL], json=repos
)
Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this is the main part that fixes the potential issue - we simply dont issue the callback request if we encounter these specific exceptions.

All other exceptions are handled like previously, meaning they will only cause the chunk where they occur in to be empty - this one instead ignores any crawled content (if any) and just drops the block without a callback, and asks for the next block (if we're in the automated workflow like in docker-compose).

Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

im not sure anymore if we can deal with it like that.
since we only clean up dead blocks on add_blocks (when the crawler sends the "answer"), we would only ever hand out more blocks, but never clean up these dead blocks.
also, right now, in the indexer we are scheduling blocks like "hoster with the oldest run timestamp first", meaning that crawling would be stuck on this hoster the indexer never get answers for. (and even if we change the way we schedule, this hoster would only pile up more and more dead blocks, until redis uses up all the ram and crashes)

maybe we should think about a proper communication protocol first, something that wraps the repo list we return in some kind of state, so that the indexer has a way to, for example, pause a run on a hoster for a day on connection errors, or end the run completely without making an export - something like that?

Expand All @@ -76,8 +83,9 @@ def crawl(platform: ICrawler) -> Generator[List[dict], None, None]:
else:
# right now we dont want to emit failures (via yield) because that will send empty results back
# to the indexer, which can trigger a state reset (i.e. reached end, start over).
# TODO deal with failures - what are they?
# - complete connection failures and such should be handled via raised exceptions within crawlers!
pass

logger.debug(f"END block: {platform.type} - final state: {platform.state}")


Expand Down
5 changes: 2 additions & 3 deletions crawlers/lib/platforms/__init__.py
Original file line number Diff line number Diff line change
@@ -1,15 +1,14 @@
from typing import Dict, Any, Type, Union
from typing import Dict
from crawlers.lib.platforms.i_crawler import ICrawler
from crawlers.lib.platforms.gitea import GiteaCrawler
from crawlers.lib.platforms.gitlab import GitLabCrawler
from crawlers.lib.platforms.bitbucket import BitBucketCrawler
from crawlers.lib.platforms.github import GitHubV4Crawler, GitHubRESTCrawler
from crawlers.lib.platforms.github import GitHubV4Crawler

platforms: Dict[str, ICrawler] = {
GiteaCrawler.type: GiteaCrawler,
GitLabCrawler.type: GitLabCrawler,
GitHubV4Crawler.type: GitHubV4Crawler,
GitHubRESTCrawler.type: GitHubRESTCrawler,
BitBucketCrawler.type: BitBucketCrawler,
}

4 changes: 2 additions & 2 deletions crawlers/lib/platforms/bitbucket.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import logging
import time
import requests
from typing import List, Tuple
from typing import List, Tuple, Union
from urllib.parse import urljoin

from crawlers.lib.platforms.i_crawler import ICrawler
Expand Down Expand Up @@ -65,7 +65,7 @@ def crawl(self, state: dict = None) -> Tuple[bool, List[dict], dict]:
logger.error(e)
logger.error(e.response.reason)
logger.error(e.response.text)
return False, [], {}
return False, [], {}, e

response_json = response.json()
repos = response_json['values']
Expand Down
9 changes: 8 additions & 1 deletion crawlers/lib/platforms/gitea.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,8 @@
import logging
from typing import List, Tuple
from typing import List, Tuple, Union

from requests import ConnectionError, Timeout, TooManyRedirects
from urllib3.exceptions import MaxRetryError

from crawlers.constants import GITEA_PER_PAGE_MAX, DEFAULT_REQUEST_TIMEOUT
from crawlers.lib.platforms.i_crawler import ICrawler
Expand Down Expand Up @@ -40,6 +43,10 @@ def crawl(self, state: dict = None) -> Tuple[bool, List[dict], dict]:
f"- response not ok, status: {response.status_code}")
return False, [], state # nr.1 - we skip rest of this block, hope we get it next time
result = response.json()
except (MaxRetryError, ConnectionError, Timeout, TooManyRedirects) as e:
logger.exception(f"{self} - crawler cannot reach hoster")
# we re-raise these, as we want to avoid returning empty results to the indexer
raise e
except Exception as e:
logger.exception(f"(skipping block chunk) gitea crawler crashed")
return False, [], state # nr.2 - we skip rest of this block, hope we get it next time
Expand Down
1 change: 0 additions & 1 deletion crawlers/lib/platforms/github/__init__.py
Original file line number Diff line number Diff line change
@@ -1,2 +1 @@
from .github_v4 import GitHubV4Crawler
from .github_rest import GitHubRESTCrawler
249 changes: 0 additions & 249 deletions crawlers/lib/platforms/github/github_rest.py

This file was deleted.

Loading