mirror of
https://github.com/derrod/legendary.git
synced 2024-06-02 10:44:54 +12:00
Add a proxy option
This commit is contained in:
parent
03b21f49de
commit
dd07f87412
|
@ -204,6 +204,8 @@ optional arguments:
|
||||||
-J, --pretty-json Pretty-print JSON
|
-J, --pretty-json Pretty-print JSON
|
||||||
-A <seconds>, --api-timeout <seconds>
|
-A <seconds>, --api-timeout <seconds>
|
||||||
API HTTP request timeout (default: 10 seconds)
|
API HTTP request timeout (default: 10 seconds)
|
||||||
|
-p <proxy>, --proxy <proxy>
|
||||||
|
SOCKS5 proxy setting (default: 127.0.0.1:1080)
|
||||||
|
|
||||||
Commands:
|
Commands:
|
||||||
<command>
|
<command>
|
||||||
|
|
|
@ -34,16 +34,22 @@ class EPCAPI:
|
||||||
_store_gql_host = 'graphql.epicgames.com'
|
_store_gql_host = 'graphql.epicgames.com'
|
||||||
_artifact_service_host = 'artifact-public-service-prod.beee.live.use1a.on.epicgames.com'
|
_artifact_service_host = 'artifact-public-service-prod.beee.live.use1a.on.epicgames.com'
|
||||||
|
|
||||||
def __init__(self, lc='en', cc='US', timeout=10.0):
|
|
||||||
|
def __init__(self, lc='en', cc='US', timeout=10.0, proxy=""):
|
||||||
self.log = logging.getLogger('EPCAPI')
|
self.log = logging.getLogger('EPCAPI')
|
||||||
|
|
||||||
|
self.proxy = proxy
|
||||||
|
|
||||||
self.session = requests.session()
|
self.session = requests.session()
|
||||||
|
if proxy != "":
|
||||||
|
self.session.proxies.update({'http': f'socks5://{self.proxy}', 'https': f'socks5://{self.proxy}', 'socks5': f'socks5://{self.proxy}'})
|
||||||
self.session.headers['User-Agent'] = self._user_agent
|
self.session.headers['User-Agent'] = self._user_agent
|
||||||
# increase maximum pool size for multithreaded metadata requests
|
# increase maximum pool size for multithreaded metadata requests
|
||||||
self.session.mount('https://', requests.adapters.HTTPAdapter(pool_maxsize=16))
|
self.session.mount('https://', requests.adapters.HTTPAdapter(pool_maxsize=16))
|
||||||
|
|
||||||
self.unauth_session = requests.session()
|
self.unauth_session = requests.session()
|
||||||
self.unauth_session.headers['User-Agent'] = self._user_agent
|
if proxy != "":
|
||||||
|
self.unauth_session.proxies.update({'http': f'socks5://{self.proxy}', 'https': f'socks5://{self.proxy}', 'socks5': f'socks5://{self.proxy}'})
|
||||||
|
|
||||||
self._oauth_basic = HTTPBasicAuth(self._user_basic, self._pw_basic)
|
self._oauth_basic = HTTPBasicAuth(self._user_basic, self._pw_basic)
|
||||||
|
|
||||||
|
|
|
@ -39,8 +39,8 @@ logger = logging.getLogger('cli')
|
||||||
|
|
||||||
|
|
||||||
class LegendaryCLI:
|
class LegendaryCLI:
|
||||||
def __init__(self, override_config=None, api_timeout=None):
|
def __init__(self, override_config=None, api_timeout=None, proxy=""):
|
||||||
self.core = LegendaryCore(override_config, timeout=api_timeout)
|
self.core = LegendaryCore(override_config, timeout=api_timeout, proxy=proxy)
|
||||||
self.logger = logging.getLogger('cli')
|
self.logger = logging.getLogger('cli')
|
||||||
self.logging_queue = None
|
self.logging_queue = None
|
||||||
|
|
||||||
|
@ -2631,6 +2631,9 @@ def main():
|
||||||
parser.add_argument('-A', '--api-timeout', dest='api_timeout', action='store',
|
parser.add_argument('-A', '--api-timeout', dest='api_timeout', action='store',
|
||||||
type=float, default=10, metavar='<seconds>',
|
type=float, default=10, metavar='<seconds>',
|
||||||
help='API HTTP request timeout (default: 10 seconds)')
|
help='API HTTP request timeout (default: 10 seconds)')
|
||||||
|
parser.add_argument('-p', '--proxy', dest='proxy', action='store',
|
||||||
|
type=str, default="", metavar='<proxy>',
|
||||||
|
help='SOCKS5 proxy setting (ex. 127.0.0.1:1080)')
|
||||||
|
|
||||||
# all the commands
|
# all the commands
|
||||||
subparsers = parser.add_subparsers(title='Commands', dest='subparser_name', metavar='<command>')
|
subparsers = parser.add_subparsers(title='Commands', dest='subparser_name', metavar='<command>')
|
||||||
|
@ -3023,7 +3026,7 @@ def main():
|
||||||
subprocess.Popen(['cmd', '/K', 'echo>nul'])
|
subprocess.Popen(['cmd', '/K', 'echo>nul'])
|
||||||
return
|
return
|
||||||
|
|
||||||
cli = LegendaryCLI(override_config=args.config_file, api_timeout=args.api_timeout)
|
cli = LegendaryCLI(override_config=args.config_file, api_timeout=args.api_timeout, proxy=args.proxy)
|
||||||
ql = cli.setup_threaded_logging()
|
ql = cli.setup_threaded_logging()
|
||||||
|
|
||||||
config_ll = cli.core.lgd.config.get('Legendary', 'log_level', fallback='info')
|
config_ll = cli.core.lgd.config.get('Legendary', 'log_level', fallback='info')
|
||||||
|
|
|
@ -54,12 +54,13 @@ class LegendaryCore:
|
||||||
"""
|
"""
|
||||||
_egl_version = '11.0.1-14907503+++Portal+Release-Live'
|
_egl_version = '11.0.1-14907503+++Portal+Release-Live'
|
||||||
|
|
||||||
def __init__(self, override_config=None, timeout=10.0):
|
def __init__(self, override_config=None, timeout=10.0, proxy=""):
|
||||||
self.log = logging.getLogger('Core')
|
self.log = logging.getLogger('Core')
|
||||||
self.egs = EPCAPI(timeout=timeout)
|
self.egs = EPCAPI(timeout=timeout, proxy=proxy)
|
||||||
self.lgd = LGDLFS(config_file=override_config)
|
self.lgd = LGDLFS(config_file=override_config)
|
||||||
self.egl = EPCLFS()
|
self.egl = EPCLFS()
|
||||||
self.lgdapi = LGDAPI()
|
self.lgdapi = LGDAPI()
|
||||||
|
self.proxy = proxy
|
||||||
|
|
||||||
# on non-Windows load the programdata path from config
|
# on non-Windows load the programdata path from config
|
||||||
if os.name != 'nt':
|
if os.name != 'nt':
|
||||||
|
@ -99,6 +100,8 @@ class LegendaryCore:
|
||||||
:return: exchange code
|
:return: exchange code
|
||||||
"""
|
"""
|
||||||
s = session()
|
s = session()
|
||||||
|
if self.proxy != "":
|
||||||
|
s.proxies.update({'http': f'socks5://{self.proxy}', 'https': f'socks5://{self.proxy}', 'socks5': f'socks5://{self.proxy}'})
|
||||||
s.headers.update({
|
s.headers.update({
|
||||||
'X-Epic-Event-Action': 'login',
|
'X-Epic-Event-Action': 'login',
|
||||||
'X-Epic-Event-Category': 'login',
|
'X-Epic-Event-Category': 'login',
|
||||||
|
@ -1459,7 +1462,7 @@ class LegendaryCore:
|
||||||
|
|
||||||
dlm = DLManager(install_path, base_url, resume_file=resume_file, status_q=status_q,
|
dlm = DLManager(install_path, base_url, resume_file=resume_file, status_q=status_q,
|
||||||
max_shared_memory=max_shm * 1024 * 1024, max_workers=max_workers,
|
max_shared_memory=max_shm * 1024 * 1024, max_workers=max_workers,
|
||||||
dl_timeout=dl_timeout)
|
dl_timeout=dl_timeout, proxy=self.proxy)
|
||||||
anlres = dlm.run_analysis(manifest=new_manifest, old_manifest=old_manifest,
|
anlres = dlm.run_analysis(manifest=new_manifest, old_manifest=old_manifest,
|
||||||
patch=not disable_patching, resume=not force,
|
patch=not disable_patching, resume=not force,
|
||||||
file_prefix_filter=file_prefix_filter,
|
file_prefix_filter=file_prefix_filter,
|
||||||
|
|
|
@ -22,7 +22,7 @@ from legendary.models.manifest import ManifestComparison, Manifest
|
||||||
class DLManager(Process):
|
class DLManager(Process):
|
||||||
def __init__(self, download_dir, base_url, cache_dir=None, status_q=None,
|
def __init__(self, download_dir, base_url, cache_dir=None, status_q=None,
|
||||||
max_workers=0, update_interval=1.0, dl_timeout=10, resume_file=None,
|
max_workers=0, update_interval=1.0, dl_timeout=10, resume_file=None,
|
||||||
max_shared_memory=1024 * 1024 * 1024):
|
max_shared_memory=1024 * 1024 * 1024, proxy=""):
|
||||||
super().__init__(name='DLManager')
|
super().__init__(name='DLManager')
|
||||||
self.log = logging.getLogger('DLM')
|
self.log = logging.getLogger('DLM')
|
||||||
self.proc_debug = False
|
self.proc_debug = False
|
||||||
|
@ -30,6 +30,7 @@ class DLManager(Process):
|
||||||
self.base_url = base_url
|
self.base_url = base_url
|
||||||
self.dl_dir = download_dir
|
self.dl_dir = download_dir
|
||||||
self.cache_dir = cache_dir or os.path.join(download_dir, '.cache')
|
self.cache_dir = cache_dir or os.path.join(download_dir, '.cache')
|
||||||
|
self.proxy = proxy
|
||||||
|
|
||||||
# All the queues!
|
# All the queues!
|
||||||
self.logging_queue = None
|
self.logging_queue = None
|
||||||
|
@ -658,7 +659,7 @@ class DLManager(Process):
|
||||||
for i in range(self.max_workers):
|
for i in range(self.max_workers):
|
||||||
w = DLWorker(f'DLWorker {i + 1}', self.dl_worker_queue, self.dl_result_q,
|
w = DLWorker(f'DLWorker {i + 1}', self.dl_worker_queue, self.dl_result_q,
|
||||||
self.shared_memory.name, logging_queue=self.logging_queue,
|
self.shared_memory.name, logging_queue=self.logging_queue,
|
||||||
dl_timeout=self.dl_timeout)
|
dl_timeout=self.dl_timeout, proxy=self.proxy)
|
||||||
self.children.append(w)
|
self.children.append(w)
|
||||||
w.start()
|
w.start()
|
||||||
|
|
||||||
|
|
|
@ -20,11 +20,14 @@ from legendary.models.downloading import (
|
||||||
|
|
||||||
class DLWorker(Process):
|
class DLWorker(Process):
|
||||||
def __init__(self, name, queue, out_queue, shm, max_retries=7,
|
def __init__(self, name, queue, out_queue, shm, max_retries=7,
|
||||||
logging_queue=None, dl_timeout=10):
|
logging_queue=None, dl_timeout=10, proxy=""):
|
||||||
super().__init__(name=name)
|
super().__init__(name=name)
|
||||||
self.q = queue
|
self.q = queue
|
||||||
self.o_q = out_queue
|
self.o_q = out_queue
|
||||||
|
self.proxy = proxy
|
||||||
self.session = requests.session()
|
self.session = requests.session()
|
||||||
|
if proxy != "":
|
||||||
|
self.session.proxies.update({'http': f'socks5://{self.proxy}', 'https': f'socks5://{self.proxy}', 'socks5': f'socks5://{self.proxy}'})
|
||||||
self.session.headers.update({
|
self.session.headers.update({
|
||||||
'User-Agent': 'EpicGamesLauncher/11.0.1-14907503+++Portal+Release-Live Windows/10.0.19041.1.256.64bit'
|
'User-Agent': 'EpicGamesLauncher/11.0.1-14907503+++Portal+Release-Live Windows/10.0.19041.1.256.64bit'
|
||||||
})
|
})
|
||||||
|
|
Loading…
Reference in a new issue