forked from sch/Slixfeed
Fix command read; Improve code of module sqlite.
This commit is contained in:
parent
1b8254832d
commit
fd07ae865a
6 changed files with 96 additions and 379 deletions
|
@ -13,6 +13,7 @@ logger.debug('This is a debug message')
|
||||||
|
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
from datetime import datetime
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
|
|
||||||
|
@ -55,4 +56,6 @@ class Message:
|
||||||
|
|
||||||
|
|
||||||
def printer(text):
|
def printer(text):
|
||||||
print(text, end='\r')
|
now = datetime.now()
|
||||||
|
current_time = now.strftime("%H:%M:%S")
|
||||||
|
print('{} {}'.format(current_time, text), end='\r')
|
||||||
|
|
|
@ -498,23 +498,7 @@ async def add_metadata(db_file):
|
||||||
ixs = cur.execute(sql).fetchall()
|
ixs = cur.execute(sql).fetchall()
|
||||||
for ix in ixs:
|
for ix in ixs:
|
||||||
feed_id = ix[0]
|
feed_id = ix[0]
|
||||||
# insert_feed_properties(cur, feed_id)
|
# Set feed status
|
||||||
insert_feed_status(cur, feed_id)
|
|
||||||
insert_feed_preferences(cur, feed_id)
|
|
||||||
|
|
||||||
|
|
||||||
def insert_feed_status(cur, feed_id):
|
|
||||||
"""
|
|
||||||
Set feed status.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
cur : object
|
|
||||||
Cursor object.
|
|
||||||
"""
|
|
||||||
function_name = sys._getframe().f_code.co_name
|
|
||||||
logger.debug('{}: feed_id: {}'
|
|
||||||
.format(function_name, feed_id))
|
|
||||||
sql = (
|
sql = (
|
||||||
"""
|
"""
|
||||||
INSERT
|
INSERT
|
||||||
|
@ -531,20 +515,7 @@ def insert_feed_status(cur, feed_id):
|
||||||
logger.warning(
|
logger.warning(
|
||||||
"Skipping feed_id {} for table feeds_state".format(feed_id))
|
"Skipping feed_id {} for table feeds_state".format(feed_id))
|
||||||
logger.error(e)
|
logger.error(e)
|
||||||
|
# Set feed preferences.
|
||||||
|
|
||||||
def insert_feed_preferences(cur, feed_id):
|
|
||||||
"""
|
|
||||||
Set feed preferences.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
cur : object
|
|
||||||
Cursor object.
|
|
||||||
"""
|
|
||||||
function_name = sys._getframe().f_code.co_name
|
|
||||||
logger.debug('{}: feed_id: {}'
|
|
||||||
.format(function_name, feed_id))
|
|
||||||
sql = (
|
sql = (
|
||||||
"""
|
"""
|
||||||
INSERT
|
INSERT
|
||||||
|
@ -563,37 +534,6 @@ def insert_feed_preferences(cur, feed_id):
|
||||||
logger.error(e)
|
logger.error(e)
|
||||||
|
|
||||||
|
|
||||||
# TODO Test
|
|
||||||
def insert_feed_properties(cur, feed_id):
|
|
||||||
"""
|
|
||||||
Set feed properties.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
cur : object
|
|
||||||
Cursor object.
|
|
||||||
"""
|
|
||||||
function_name = sys._getframe().f_code.co_name
|
|
||||||
logger.debug('{}: feed_id: {}'
|
|
||||||
.format(function_name, feed_id))
|
|
||||||
sql = (
|
|
||||||
"""
|
|
||||||
INSERT
|
|
||||||
INTO feeds_properties(
|
|
||||||
id)
|
|
||||||
VALUES(
|
|
||||||
?)
|
|
||||||
"""
|
|
||||||
)
|
|
||||||
par = (feed_id,)
|
|
||||||
try:
|
|
||||||
cur.execute(sql, par)
|
|
||||||
except IntegrityError as e:
|
|
||||||
logger.warning(
|
|
||||||
"Skipping feed_id {} for table feeds_properties".format(feed_id))
|
|
||||||
logger.error(e)
|
|
||||||
|
|
||||||
|
|
||||||
async def insert_feed(db_file, url, title, identifier, entries=None, version=None,
|
async def insert_feed(db_file, url, title, identifier, entries=None, version=None,
|
||||||
encoding=None, language=None, status_code=None,
|
encoding=None, language=None, status_code=None,
|
||||||
updated=None):
|
updated=None):
|
||||||
|
@ -673,52 +613,6 @@ async def insert_feed(db_file, url, title, identifier, entries=None, version=Non
|
||||||
cur.execute(sql, par)
|
cur.execute(sql, par)
|
||||||
|
|
||||||
|
|
||||||
async def insert_feed_(db_file, url, title):
|
|
||||||
"""
|
|
||||||
Insert a new feed into the feeds table.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
db_file : str
|
|
||||||
Path to database file.
|
|
||||||
url : str
|
|
||||||
URL.
|
|
||||||
title : str, optional
|
|
||||||
Feed title. The default is None.
|
|
||||||
"""
|
|
||||||
function_name = sys._getframe().f_code.co_name
|
|
||||||
logger.debug('{}: db_file: {} url: {}'
|
|
||||||
.format(function_name, db_file, url))
|
|
||||||
async with DBLOCK:
|
|
||||||
with create_connection(db_file) as conn:
|
|
||||||
cur = conn.cursor()
|
|
||||||
sql = (
|
|
||||||
"""
|
|
||||||
INSERT
|
|
||||||
INTO feeds_properties(
|
|
||||||
title, url)
|
|
||||||
VALUES(
|
|
||||||
?, ?)
|
|
||||||
"""
|
|
||||||
)
|
|
||||||
par = (
|
|
||||||
title, url
|
|
||||||
)
|
|
||||||
cur.execute(sql, par)
|
|
||||||
sql = (
|
|
||||||
"""
|
|
||||||
SELECT id
|
|
||||||
FROM feeds_properties
|
|
||||||
WHERE url = :url
|
|
||||||
"""
|
|
||||||
)
|
|
||||||
par = (url,)
|
|
||||||
feed_id = cur.execute(sql, par).fetchone()[0]
|
|
||||||
# insert_feed_properties(cur, feed_id)
|
|
||||||
insert_feed_status(cur, feed_id)
|
|
||||||
insert_feed_preferences(cur, feed_id)
|
|
||||||
|
|
||||||
|
|
||||||
async def remove_feed_by_url(db_file, url):
|
async def remove_feed_by_url(db_file, url):
|
||||||
"""
|
"""
|
||||||
Delete a feed by feed URL.
|
Delete a feed by feed URL.
|
||||||
|
@ -1531,37 +1425,6 @@ def get_feed_id(db_file, url):
|
||||||
return feed_id
|
return feed_id
|
||||||
|
|
||||||
|
|
||||||
def is_entry_archived(cur, ix):
|
|
||||||
"""
|
|
||||||
Check whether a given entry is archived.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
cur : object
|
|
||||||
Cursor object.
|
|
||||||
ix : str
|
|
||||||
Index of entry.
|
|
||||||
|
|
||||||
Returns
|
|
||||||
-------
|
|
||||||
result : tuple
|
|
||||||
Entry ID.
|
|
||||||
"""
|
|
||||||
function_name = sys._getframe().f_code.co_name
|
|
||||||
logger.debug('{}: ix: {}'
|
|
||||||
.format(function_name, ix))
|
|
||||||
sql = (
|
|
||||||
"""
|
|
||||||
SELECT id
|
|
||||||
FROM entries_state
|
|
||||||
WHERE archived = 1 AND entry_id = ?
|
|
||||||
"""
|
|
||||||
)
|
|
||||||
par = (ix,)
|
|
||||||
result = cur.execute(sql, par).fetchone()
|
|
||||||
return result
|
|
||||||
|
|
||||||
|
|
||||||
def is_entry_read(db_file, ix):
|
def is_entry_read(db_file, ix):
|
||||||
"""
|
"""
|
||||||
Check whether a given entry is marked as read.
|
Check whether a given entry is marked as read.
|
||||||
|
@ -1977,53 +1840,9 @@ async def mark_feed_as_read(db_file, feed_id):
|
||||||
# cur.execute(sql, par)
|
# cur.execute(sql, par)
|
||||||
|
|
||||||
|
|
||||||
async def mark_entry_as_read(cur, ix):
|
|
||||||
"""
|
|
||||||
Set read status of entry as read.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
cur : object
|
|
||||||
Cursor object.
|
|
||||||
ix : str
|
|
||||||
Index of entry.
|
|
||||||
"""
|
|
||||||
function_name = sys._getframe().f_code.co_name
|
|
||||||
logger.debug('{}: ix: {}'
|
|
||||||
.format(function_name, ix))
|
|
||||||
sql = (
|
|
||||||
"""
|
|
||||||
UPDATE entries_state
|
|
||||||
SET read = 1
|
|
||||||
WHERE entry_id = ?
|
|
||||||
"""
|
|
||||||
)
|
|
||||||
par = (ix,)
|
|
||||||
cur.execute(sql, par)
|
|
||||||
|
|
||||||
|
|
||||||
async def mark_as_read(db_file, ix):
|
async def mark_as_read(db_file, ix):
|
||||||
function_name = sys._getframe().f_code.co_name
|
|
||||||
logger.debug('{}: db_file: {} ix: {}'
|
|
||||||
.format(function_name, db_file, ix))
|
|
||||||
async with DBLOCK:
|
|
||||||
with create_connection(db_file) as conn:
|
|
||||||
cur = conn.cursor()
|
|
||||||
# TODO While `async with DBLOCK` does work well from
|
|
||||||
# outside of functions, it would be better practice
|
|
||||||
# to place it within the functions.
|
|
||||||
# NOTE: We can use DBLOCK once for both
|
|
||||||
# functions, because, due to exclusive
|
|
||||||
# ID, only one can ever occur.
|
|
||||||
if is_entry_archived(cur, ix):
|
|
||||||
await delete_entry(cur, ix)
|
|
||||||
else:
|
|
||||||
await mark_entry_as_read(cur, ix)
|
|
||||||
|
|
||||||
|
|
||||||
async def delete_entry(cur, ix):
|
|
||||||
"""
|
"""
|
||||||
Delete entry.
|
Set read status of entry as read or delete entry.
|
||||||
|
|
||||||
Parameters
|
Parameters
|
||||||
----------
|
----------
|
||||||
|
@ -2033,8 +1852,23 @@ async def delete_entry(cur, ix):
|
||||||
Index of entry.
|
Index of entry.
|
||||||
"""
|
"""
|
||||||
function_name = sys._getframe().f_code.co_name
|
function_name = sys._getframe().f_code.co_name
|
||||||
logger.debug('{}: ix: {}'
|
logger.debug('{}: db_file: {} ix: {}'
|
||||||
.format(function_name, ix))
|
.format(function_name, db_file, ix))
|
||||||
|
async with DBLOCK:
|
||||||
|
with create_connection(db_file) as conn:
|
||||||
|
cur = conn.cursor()
|
||||||
|
# Check whether a given entry is archived.
|
||||||
|
sql = (
|
||||||
|
"""
|
||||||
|
SELECT id
|
||||||
|
FROM entries_state
|
||||||
|
WHERE archived = 1 AND entry_id = ?
|
||||||
|
"""
|
||||||
|
)
|
||||||
|
par = (ix,)
|
||||||
|
result = cur.execute(sql, par).fetchone()
|
||||||
|
# is_entry_archived
|
||||||
|
if result:
|
||||||
sql = (
|
sql = (
|
||||||
"""
|
"""
|
||||||
DELETE
|
DELETE
|
||||||
|
@ -2042,56 +1876,15 @@ async def delete_entry(cur, ix):
|
||||||
WHERE id = ?
|
WHERE id = ?
|
||||||
"""
|
"""
|
||||||
)
|
)
|
||||||
par = (ix,)
|
|
||||||
cur.execute(sql, par)
|
|
||||||
|
|
||||||
|
|
||||||
async def update_statistics(cur):
|
|
||||||
"""
|
|
||||||
Update table statistics.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
cur : object
|
|
||||||
Cursor object.
|
|
||||||
"""
|
|
||||||
function_name = sys._getframe().f_code.co_name
|
|
||||||
logger.debug('{}'.format(function_name))
|
|
||||||
stat_dict = {}
|
|
||||||
stat_dict["feeds"] = get_number_of_items(cur, 'feeds_properties')
|
|
||||||
stat_dict["entries"] = get_number_of_items(cur, 'entries_properties')
|
|
||||||
stat_dict["unread"] = get_number_of_entries_unread(cur=cur)
|
|
||||||
for i in stat_dict:
|
|
||||||
sql = (
|
|
||||||
"SELECT id "
|
|
||||||
"FROM statistics "
|
|
||||||
"WHERE title = ?"
|
|
||||||
)
|
|
||||||
par = (i,)
|
|
||||||
cur.execute(sql, par)
|
|
||||||
if cur.fetchone():
|
|
||||||
sql = (
|
|
||||||
"UPDATE statistics "
|
|
||||||
"SET number = :num "
|
|
||||||
"WHERE title = :title"
|
|
||||||
)
|
|
||||||
par = {
|
|
||||||
"title": i,
|
|
||||||
"num": stat_dict[i]
|
|
||||||
}
|
|
||||||
cur.execute(sql, par)
|
|
||||||
else:
|
else:
|
||||||
sql = (
|
sql = (
|
||||||
"SELECT count(id) "
|
"""
|
||||||
"FROM statistics"
|
UPDATE entries_state
|
||||||
|
SET read = 1
|
||||||
|
WHERE entry_id = ?
|
||||||
|
"""
|
||||||
)
|
)
|
||||||
count = cur.execute(sql).fetchone()[0]
|
par = (ix,)
|
||||||
ix = count + 1
|
|
||||||
sql = (
|
|
||||||
"INSERT INTO statistics "
|
|
||||||
"VALUES(?,?,?)"
|
|
||||||
)
|
|
||||||
par = (ix, i, stat_dict[i])
|
|
||||||
cur.execute(sql, par)
|
cur.execute(sql, par)
|
||||||
|
|
||||||
|
|
||||||
|
@ -2709,89 +2502,6 @@ def get_contents_by_entry_id(db_file, entry_id):
|
||||||
return result
|
return result
|
||||||
|
|
||||||
|
|
||||||
def get_invalid_entries(db_file, url, feed):
|
|
||||||
"""
|
|
||||||
List entries that do not exist in a given feed.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
db_file : str
|
|
||||||
Path to database file.
|
|
||||||
url : str
|
|
||||||
Feed URL.
|
|
||||||
feed : list
|
|
||||||
Parsed feed document.
|
|
||||||
|
|
||||||
Returns
|
|
||||||
-------
|
|
||||||
ixs : dict
|
|
||||||
List of indexes of invalid items.
|
|
||||||
"""
|
|
||||||
function_name = sys._getframe().f_code.co_name
|
|
||||||
logger.debug('{}: db_file: {} url: {}'.format(function_name, db_file, url))
|
|
||||||
feed_id = get_feed_id(db_file, url)
|
|
||||||
feed_id = feed_id[0]
|
|
||||||
items = get_entries_of_feed(db_file, feed_id)
|
|
||||||
entries = feed.entries
|
|
||||||
ixs = {}
|
|
||||||
for item in items:
|
|
||||||
ix, entry_title, entry_link, entry_id, timestamp = item
|
|
||||||
read_status = is_entry_read(db_file, ix)
|
|
||||||
read_status = read_status[0]
|
|
||||||
for entry in entries:
|
|
||||||
title = None
|
|
||||||
link = None
|
|
||||||
time = None
|
|
||||||
# TODO better check and don't repeat code
|
|
||||||
if entry.has_key("id") and entry_id:
|
|
||||||
if entry.id == entry_id:
|
|
||||||
# print(url)
|
|
||||||
# print("compare entry.id == entry_id:", entry.id)
|
|
||||||
# print("compare entry.id == entry_id:", entry_id)
|
|
||||||
# print("============")
|
|
||||||
# items_valid.append(ix)
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
# Prepare a title to compare
|
|
||||||
if entry.has_key("title"):
|
|
||||||
title = entry.title
|
|
||||||
else:
|
|
||||||
title = feed["feed"]["title"]
|
|
||||||
# Prepare a link to compare
|
|
||||||
if entry.has_key("link"):
|
|
||||||
link = Url.join_url(url, entry.link)
|
|
||||||
else:
|
|
||||||
link = url
|
|
||||||
# Compare date, link and title
|
|
||||||
if entry.has_key("published") and timestamp:
|
|
||||||
# print(url)
|
|
||||||
# print("compare published:", title, link, time)
|
|
||||||
# print("compare published:", entry_title, entry_link, timestamp)
|
|
||||||
# print("============")
|
|
||||||
time = DateAndTime.rfc2822_to_iso8601(entry.published)
|
|
||||||
if (entry_title == title and
|
|
||||||
entry_link == link and
|
|
||||||
timestamp == time):
|
|
||||||
# items_valid.append(ix)
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
# Compare link and title
|
|
||||||
if (entry_title == title and
|
|
||||||
entry_link == link):
|
|
||||||
# print(url)
|
|
||||||
# print("compare entry_link == link:", title, link)
|
|
||||||
# print("compare entry_title == title:", entry_title, entry_link)
|
|
||||||
# print("============")
|
|
||||||
# items_valid.append(ix)
|
|
||||||
break
|
|
||||||
# print('invalid entry:')
|
|
||||||
# print(entry)
|
|
||||||
# TODO better check and don't repeat code
|
|
||||||
ixs[ix] = read_status
|
|
||||||
# print(ixs)
|
|
||||||
return ixs
|
|
||||||
|
|
||||||
|
|
||||||
async def process_invalid_entries(db_file, ixs):
|
async def process_invalid_entries(db_file, ixs):
|
||||||
"""
|
"""
|
||||||
Batch process of invalid items.
|
Batch process of invalid items.
|
||||||
|
@ -2974,10 +2684,7 @@ def get_feeds_by_enabled_state(db_file, enabled_state):
|
||||||
function_name = sys._getframe().f_code.co_name
|
function_name = sys._getframe().f_code.co_name
|
||||||
logger.debug('{}: db_file: {} enabled_state: {}'
|
logger.debug('{}: db_file: {} enabled_state: {}'
|
||||||
.format(function_name, db_file, enabled_state))
|
.format(function_name, db_file, enabled_state))
|
||||||
if enabled_state:
|
enabled_state = 1 if enabled_state else 0
|
||||||
enabled_state = 1
|
|
||||||
else:
|
|
||||||
enabled_state = 0
|
|
||||||
with create_connection(db_file) as conn:
|
with create_connection(db_file) as conn:
|
||||||
cur = conn.cursor()
|
cur = conn.cursor()
|
||||||
sql = (
|
sql = (
|
||||||
|
|
|
@ -1,2 +1,2 @@
|
||||||
__version__ = '0.1.84'
|
__version__ = '0.1.85'
|
||||||
__version_info__ = (0, 1, 84)
|
__version_info__ = (0, 1, 85)
|
||||||
|
|
|
@ -444,7 +444,7 @@ class XmppCommands:
|
||||||
result['identifier'],
|
result['identifier'],
|
||||||
result['index']))
|
result['index']))
|
||||||
elif result['error']:
|
elif result['error']:
|
||||||
message = ('> {}\nFailed to find subscriptions. '
|
message = ('> {}\nNo subscriptions were found. '
|
||||||
'Reason: {} (status code: {})'
|
'Reason: {} (status code: {})'
|
||||||
.format(url, result['message'],
|
.format(url, result['message'],
|
||||||
result['code']))
|
result['code']))
|
||||||
|
@ -508,7 +508,7 @@ class XmppCommands:
|
||||||
result['name'],
|
result['name'],
|
||||||
result['index']))
|
result['index']))
|
||||||
elif result['error']:
|
elif result['error']:
|
||||||
message = ('> {}\nFailed to find subscriptions. '
|
message = ('> {}\nNo subscriptions were found. '
|
||||||
'Reason: {} (status code: {})'
|
'Reason: {} (status code: {})'
|
||||||
.format(url, result['message'],
|
.format(url, result['message'],
|
||||||
result['code']))
|
result['code']))
|
||||||
|
@ -740,7 +740,7 @@ class XmppCommands:
|
||||||
while True:
|
while True:
|
||||||
result = await fetch.http(url)
|
result = await fetch.http(url)
|
||||||
status = result['status_code']
|
status = result['status_code']
|
||||||
if not result['error']:
|
if result and not result['error']:
|
||||||
document = result['content']
|
document = result['content']
|
||||||
feed = parse(document)
|
feed = parse(document)
|
||||||
if Feed.is_feed(url, feed):
|
if Feed.is_feed(url, feed):
|
||||||
|
@ -760,6 +760,10 @@ class XmppCommands:
|
||||||
message += ('```\nTotal of {} feeds.'
|
message += ('```\nTotal of {} feeds.'
|
||||||
.format(len(results)))
|
.format(len(results)))
|
||||||
break
|
break
|
||||||
|
elif not result:
|
||||||
|
message = ('> {}\nNo subscriptions were found.'
|
||||||
|
.format(url))
|
||||||
|
break
|
||||||
else:
|
else:
|
||||||
url = result['link']
|
url = result['link']
|
||||||
else:
|
else:
|
||||||
|
@ -767,15 +771,13 @@ class XmppCommands:
|
||||||
.format(url, status))
|
.format(url, status))
|
||||||
break
|
break
|
||||||
else:
|
else:
|
||||||
message = ('No action has been taken.'
|
message = ('No action has been taken. Missing URL.')
|
||||||
'\n'
|
|
||||||
'Missing URL.')
|
|
||||||
case 2:
|
case 2:
|
||||||
num = data[1]
|
num = data[1]
|
||||||
if url.startswith('http'):
|
if url.startswith('http'):
|
||||||
while True:
|
while True:
|
||||||
result = await fetch.http(url)
|
result = await fetch.http(url)
|
||||||
if not result['error']:
|
if result and not result['error']:
|
||||||
document = result['content']
|
document = result['content']
|
||||||
status = result['status_code']
|
status = result['status_code']
|
||||||
feed = parse(document)
|
feed = parse(document)
|
||||||
|
@ -796,6 +798,10 @@ class XmppCommands:
|
||||||
message += ('```\nTotal of {} feeds.'
|
message += ('```\nTotal of {} feeds.'
|
||||||
.format(len(results)))
|
.format(len(results)))
|
||||||
break
|
break
|
||||||
|
elif not result:
|
||||||
|
message = ('> {}\nNo subscriptions were found.'
|
||||||
|
.format(url))
|
||||||
|
break
|
||||||
else:
|
else:
|
||||||
url = result['link']
|
url = result['link']
|
||||||
else:
|
else:
|
||||||
|
|
|
@ -14,8 +14,7 @@ TODO
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import asyncio
|
import asyncio
|
||||||
from slixfeed.utilities import DateAndTime
|
from slixfeed.log import Logger, Message
|
||||||
from slixfeed.log import Logger
|
|
||||||
from slixmpp.exceptions import IqTimeout, IqError
|
from slixmpp.exceptions import IqTimeout, IqError
|
||||||
from time import sleep
|
from time import sleep
|
||||||
|
|
||||||
|
@ -62,17 +61,17 @@ class XmppConnect:
|
||||||
|
|
||||||
def recover(self, message):
|
def recover(self, message):
|
||||||
logger.warning(message)
|
logger.warning(message)
|
||||||
print(DateAndTime.current_time(), message, 'Attempting to reconnect.')
|
Message.printer('Slixfeed ittempting to reconnect...')
|
||||||
self.connection_attempts += 1
|
self.connection_attempts += 1
|
||||||
# if self.connection_attempts <= self.max_connection_attempts:
|
# if self.connection_attempts <= self.max_connection_attempts:
|
||||||
# self.reconnect(wait=5.0) # wait a bit before attempting to reconnect
|
# self.reconnect(wait=5.0) # wait a bit before attempting to reconnect
|
||||||
# else:
|
# else:
|
||||||
# print(current_time(),"Maximum connection attempts exceeded.")
|
# print(current_time(),"Maximum connection attempts exceeded.")
|
||||||
# logging.error("Maximum connection attempts exceeded.")
|
# logging.error("Maximum connection attempts exceeded.")
|
||||||
print(DateAndTime.current_time(), 'Attempt number', self.connection_attempts)
|
Message.printer('Attempt number {}'.format(self.connection_attempts))
|
||||||
seconds = self.reconnect_timeout or 30
|
seconds = self.reconnect_timeout or 30
|
||||||
seconds = int(seconds)
|
seconds = int(seconds)
|
||||||
print(DateAndTime.current_time(), 'Next attempt within', seconds, 'seconds')
|
Message.printer('Next attempt be made within {} seconds'.format(seconds))
|
||||||
# NOTE asyncio.sleep doesn't interval as expected
|
# NOTE asyncio.sleep doesn't interval as expected
|
||||||
# await asyncio.sleep(seconds)
|
# await asyncio.sleep(seconds)
|
||||||
sleep(seconds)
|
sleep(seconds)
|
||||||
|
@ -80,12 +79,14 @@ class XmppConnect:
|
||||||
|
|
||||||
|
|
||||||
def inspect(self):
|
def inspect(self):
|
||||||
print('Disconnected\nReconnecting...')
|
Message.printer('Disconnected')
|
||||||
|
sleep(3)
|
||||||
|
Message.printer('Reconnecting...')
|
||||||
try:
|
try:
|
||||||
self.reconnect
|
self.reconnect
|
||||||
except:
|
except:
|
||||||
self.disconnect()
|
self.disconnect()
|
||||||
print('Problem reconnecting')
|
Message.printer('Problem reconnecting')
|
||||||
|
|
||||||
|
|
||||||
class XmppConnectTask:
|
class XmppConnectTask:
|
||||||
|
|
|
@ -34,8 +34,7 @@ class XmppGroupchat:
|
||||||
'bookmark {}'.format(bookmark['name']))
|
'bookmark {}'.format(bookmark['name']))
|
||||||
alias = bookmark["nick"]
|
alias = bookmark["nick"]
|
||||||
muc_jid = bookmark["jid"]
|
muc_jid = bookmark["jid"]
|
||||||
|
Message.printer('Joining to MUC {} ...'.format(muc_jid))
|
||||||
Message.printer('Joining MUC {} ...'.format(muc_jid))
|
|
||||||
result = await XmppMuc.join(self, muc_jid, alias)
|
result = await XmppMuc.join(self, muc_jid, alias)
|
||||||
if result == 'ban':
|
if result == 'ban':
|
||||||
await XmppBookmark.remove(self, muc_jid)
|
await XmppBookmark.remove(self, muc_jid)
|
||||||
|
@ -53,3 +52,4 @@ class XmppGroupchat:
|
||||||
elif not bookmark["jid"]:
|
elif not bookmark["jid"]:
|
||||||
logger.error('JID is missing for bookmark {}'
|
logger.error('JID is missing for bookmark {}'
|
||||||
.format(bookmark['name']))
|
.format(bookmark['name']))
|
||||||
|
print('Done')
|
||||||
|
|
Loading…
Reference in a new issue