mirror of
https://github.com/jellyfin/jellyfin-kodi.git
synced 2024-11-10 04:06:11 +00:00
free memory after a thread completes
-> prior all threads that fetched items from the server and their results stayed in memory until the sync was finished
This commit is contained in:
parent
09b0bdbc48
commit
37281f6ca7
1 changed files with 39 additions and 25 deletions
|
@ -7,7 +7,7 @@ import threading
|
|||
import concurrent.futures
|
||||
from datetime import date
|
||||
|
||||
from six.moves import range, queue as Queue, zip
|
||||
from six.moves import range, queue as Queue
|
||||
|
||||
from kodi_six import xbmc
|
||||
import requests
|
||||
|
@ -183,7 +183,6 @@ def get_item_count(parent_id, item_type=None, params=None):
|
|||
|
||||
return result.get('TotalRecordCount', 1)
|
||||
|
||||
|
||||
def get_items(parent_id, item_type=None, basic=False, params=None):
|
||||
|
||||
query = {
|
||||
|
@ -279,36 +278,51 @@ def _get_items(query, server_id=None):
|
|||
# complete all tasks before allowing any results to be processed. ThreadPoolExecutor
|
||||
# allows for completed tasks to be processed while other tasks are completed on other
|
||||
# threads. Dont be a dummy.Pool, be a ThreadPoolExecutor
|
||||
p = concurrent.futures.ThreadPoolExecutor(DTHREADS)
|
||||
with concurrent.futures.ThreadPoolExecutor(DTHREADS) as p:
|
||||
# dictionary for storing the jobs and their results
|
||||
jobs = {}
|
||||
|
||||
thread_buffer = threading.Semaphore(2 * LIMIT * DTHREADS)
|
||||
# semaphore to avoid fetching complete library to memory
|
||||
thread_buffer = threading.Semaphore(LIMIT * DTHREADS)
|
||||
|
||||
def get_wrapper(params):
|
||||
thread_buffer.acquire()
|
||||
return _get(url, params, server_id=server_id)
|
||||
# wrapper function for _get that uses a semaphore
|
||||
def get_wrapper(params):
|
||||
thread_buffer.acquire()
|
||||
return _get(url, params, server_id=server_id)
|
||||
|
||||
results = p.map(get_wrapper, query_params)
|
||||
# create jobs
|
||||
for param in query_params:
|
||||
job = p.submit(get_wrapper, param)
|
||||
# the query params are later needed again
|
||||
jobs[job] = param
|
||||
|
||||
for params, result in zip(query_params, results):
|
||||
query['params'] = params
|
||||
# process complete jobs
|
||||
for job in concurrent.futures.as_completed(jobs):
|
||||
# get the result
|
||||
result = job.result() or {'Items': []}
|
||||
query['params'] = jobs[job]
|
||||
|
||||
result = result or {'Items': []}
|
||||
# free job memory
|
||||
del jobs[job]
|
||||
del job
|
||||
|
||||
# Mitigates #216 till the server validates the date provided is valid
|
||||
if result['Items'][0].get('ProductionYear'):
|
||||
try:
|
||||
date(result['Items'][0]['ProductionYear'], 1, 1)
|
||||
except ValueError:
|
||||
LOG.info('#216 mitigation triggered. Setting ProductionYear to None')
|
||||
result['Items'][0]['ProductionYear'] = None
|
||||
# Mitigates #216 till the server validates the date provided is valid
|
||||
if result['Items'][0].get('ProductionYear'):
|
||||
try:
|
||||
date(result['Items'][0]['ProductionYear'], 1, 1)
|
||||
except ValueError:
|
||||
LOG.info('#216 mitigation triggered. Setting ProductionYear to None')
|
||||
result['Items'][0]['ProductionYear'] = None
|
||||
|
||||
items['Items'].extend(result['Items'])
|
||||
# Using items to return data and communicate a restore point back to the callee is
|
||||
# a violation of the SRP. TODO: Seperate responsibilities.
|
||||
items['RestorePoint'] = query
|
||||
yield items
|
||||
del items['Items'][:]
|
||||
thread_buffer.release()
|
||||
items['Items'].extend(result['Items'])
|
||||
# Using items to return data and communicate a restore point back to the callee is
|
||||
# a violation of the SRP. TODO: Seperate responsibilities.
|
||||
items['RestorePoint'] = query
|
||||
yield items
|
||||
del items['Items'][:]
|
||||
|
||||
# release the semaphore again
|
||||
thread_buffer.release()
|
||||
|
||||
|
||||
class GetItemWorker(threading.Thread):
|
||||
|
|
Loading…
Reference in a new issue