Handle missing files without errors
This commit is contained in:
parent
a726aee131
commit
4cbc13dd01
1 changed files with 61 additions and 44 deletions
|
@ -19,6 +19,28 @@ PLANNING_URL = 'https://amicale-insat.fr/event/json/list'
|
||||||
DASHBOARD_FILE = 'dashboard_data.json'
|
DASHBOARD_FILE = 'dashboard_data.json'
|
||||||
|
|
||||||
|
|
||||||
|
def get_news_feed():
|
||||||
|
"""
|
||||||
|
Get facebook news and truncate the data to 15 entries for faster loading
|
||||||
|
|
||||||
|
:return: an object containing the facebook feed data
|
||||||
|
"""
|
||||||
|
# Prevent concurrent access to file
|
||||||
|
while os.path.isfile(FACEBOOK_LOCK):
|
||||||
|
print("Waiting for Facebook lock")
|
||||||
|
try:
|
||||||
|
with open(FACEBOOK_FILE) as f:
|
||||||
|
data = json.load(f)
|
||||||
|
if 'data' in data and len(data['data']) > 15:
|
||||||
|
del data['data'][14:]
|
||||||
|
else:
|
||||||
|
data = {'data': []}
|
||||||
|
return data
|
||||||
|
except FileNotFoundError:
|
||||||
|
print("Could not find " + FACEBOOK_FILE)
|
||||||
|
return {'data': []}
|
||||||
|
|
||||||
|
|
||||||
def get_available_machines():
|
def get_available_machines():
|
||||||
"""
|
"""
|
||||||
Get the number of available washing/drying machines
|
Get the number of available washing/drying machines
|
||||||
|
@ -28,16 +50,20 @@ def get_available_machines():
|
||||||
# Prevent concurrent access to file
|
# Prevent concurrent access to file
|
||||||
while os.path.isfile(WASHINSA_LOCK):
|
while os.path.isfile(WASHINSA_LOCK):
|
||||||
print("Waiting for washinsa lock")
|
print("Waiting for washinsa lock")
|
||||||
with open(WASHINSA_FILE) as f:
|
try:
|
||||||
data = json.load(f)
|
with open(WASHINSA_FILE) as f:
|
||||||
available_dryers = 0
|
data = json.load(f)
|
||||||
available_washers = 0
|
available_dryers = 0
|
||||||
for machine in data['dryers']:
|
available_washers = 0
|
||||||
if machine['state'] == 'DISPONIBLE':
|
for machine in data['dryers']:
|
||||||
available_dryers += 1
|
if machine['state'] == 'DISPONIBLE':
|
||||||
for machine in data['washers']:
|
available_dryers += 1
|
||||||
if machine['state'] == 'DISPONIBLE':
|
for machine in data['washers']:
|
||||||
available_washers += 1
|
if machine['state'] == 'DISPONIBLE':
|
||||||
|
available_washers += 1
|
||||||
|
except FileNotFoundError:
|
||||||
|
print("Could not find " + WASHINSA_FILE)
|
||||||
|
return 0, 0
|
||||||
return available_dryers, available_washers
|
return available_dryers, available_washers
|
||||||
|
|
||||||
|
|
||||||
|
@ -50,14 +76,18 @@ def get_today_menu():
|
||||||
# Prevent concurrent access to file
|
# Prevent concurrent access to file
|
||||||
while os.path.isfile(MENU_LOCK):
|
while os.path.isfile(MENU_LOCK):
|
||||||
print("Waiting for menu lock")
|
print("Waiting for menu lock")
|
||||||
with open(MENU_FILE) as f:
|
try:
|
||||||
data = json.load(f)
|
with open(MENU_FILE) as f:
|
||||||
menu = []
|
data = json.load(f)
|
||||||
for i in range(0, len(data)):
|
menu = []
|
||||||
if data[i]['date'] == date.today().strftime('%Y-%m-%d'):
|
for i in range(0, len(data)):
|
||||||
menu = data[i]['meal'][0]['foodcategory']
|
if data[i]['date'] == date.today().strftime('%Y-%m-%d'):
|
||||||
break
|
menu = data[i]['meal'][0]['foodcategory']
|
||||||
return menu
|
break
|
||||||
|
return menu
|
||||||
|
except FileNotFoundError:
|
||||||
|
print("Could not find " + MENU_FILE)
|
||||||
|
return []
|
||||||
|
|
||||||
|
|
||||||
def get_proximo_article_number():
|
def get_proximo_article_number():
|
||||||
|
@ -66,13 +96,17 @@ def get_proximo_article_number():
|
||||||
|
|
||||||
:return: an integer representing the number of articles
|
:return: an integer representing the number of articles
|
||||||
"""
|
"""
|
||||||
with urllib.request.urlopen(PROXIMO_URL) as response:
|
try:
|
||||||
data = json.loads(response.read().decode())
|
with urllib.request.urlopen(PROXIMO_URL) as response:
|
||||||
count = 0
|
data = json.loads(response.read().decode())
|
||||||
for article in data['articles']:
|
count = 0
|
||||||
if int(article['quantity']) > 0:
|
for article in data['articles']:
|
||||||
count += 1
|
if int(article['quantity']) > 0:
|
||||||
return count
|
count += 1
|
||||||
|
return count
|
||||||
|
except:
|
||||||
|
print("Error processing following url: " + PROXIMO_URL)
|
||||||
|
return 0
|
||||||
|
|
||||||
|
|
||||||
def get_tutorinsa_tutoring_number():
|
def get_tutorinsa_tutoring_number():
|
||||||
|
@ -86,6 +120,7 @@ def get_tutorinsa_tutoring_number():
|
||||||
data = json.loads(response.read().decode())
|
data = json.loads(response.read().decode())
|
||||||
return int(data['tutorials_number'])
|
return int(data['tutorials_number'])
|
||||||
except:
|
except:
|
||||||
|
print("Error processing following url: " + TUTORINSA_URL)
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
|
@ -104,27 +139,9 @@ def get_today_events():
|
||||||
today_events.append(event)
|
today_events.append(event)
|
||||||
return today_events
|
return today_events
|
||||||
except:
|
except:
|
||||||
|
print("Error processing following url: " + PLANNING_URL)
|
||||||
return []
|
return []
|
||||||
|
|
||||||
|
|
||||||
def get_news_feed():
|
|
||||||
"""
|
|
||||||
Get facebook news and truncate the data to 15 entries for faster loading
|
|
||||||
|
|
||||||
:return: an object containing the facebook feed data
|
|
||||||
"""
|
|
||||||
# Prevent concurrent access to file
|
|
||||||
while os.path.isfile(FACEBOOK_LOCK):
|
|
||||||
print("Waiting for Facebook lock")
|
|
||||||
with open(FACEBOOK_FILE) as f:
|
|
||||||
data = json.load(f)
|
|
||||||
if 'data' in data and len(data['data']) > 15:
|
|
||||||
del data['data'][14:]
|
|
||||||
else:
|
|
||||||
data = {'data': []}
|
|
||||||
return data
|
|
||||||
|
|
||||||
|
|
||||||
def generate_dashboard_json():
|
def generate_dashboard_json():
|
||||||
"""
|
"""
|
||||||
Generate the actual dashboard
|
Generate the actual dashboard
|
||||||
|
|
Loading…
Reference in a new issue