Fix archiving. WARNING of complexity due to UNION of SQL

This commit is contained in:
Schimon Jehudah 2023-11-15 14:00:49 +00:00
parent 9a93d6de0e
commit cb2317b35a
7 changed files with 466 additions and 203 deletions

View file

@ -11,73 +11,7 @@ TODO
"""
import os
import sys
def get_default_dbdir():
"""
Determine the directory path where dbfile will be stored.
* If $XDG_DATA_HOME is defined, use it;
* else if $HOME exists, use it;
* else if the platform is Windows, use %APPDATA%;
* else use the current directory.
Returns
-------
str
Path to database file.
Note
----
This function was taken from project buku.
See https://github.com/jarun/buku
* Arun Prakash Jana (jarun)
* Dmitry Marakasov (AMDmi3)
"""
# data_home = xdg.BaseDirectory.xdg_data_home
data_home = os.environ.get('XDG_DATA_HOME')
if data_home is None:
if os.environ.get('HOME') is None:
if sys.platform == 'win32':
data_home = os.environ.get('APPDATA')
if data_home is None:
return os.path.abspath('.')
else:
return os.path.abspath('.')
else:
data_home = os.path.join(os.environ.get('HOME'), '.local', 'share')
return os.path.join(data_home, 'slixfeed')
def get_default_confdir():
"""
Determine the directory path where configuration will be stored.
* If $XDG_CONFIG_HOME is defined, use it;
* else if $HOME exists, use it;
* else if the platform is Windows, use %APPDATA%;
* else use the current directory.
Returns
-------
str
Path to configueation directory.
"""
# config_home = xdg.BaseDirectory.xdg_config_home
config_home = os.environ.get('XDG_CONFIG_HOME')
if config_home is None:
if os.environ.get('HOME') is None:
if sys.platform == 'win32':
config_home = os.environ.get('APPDATA')
if config_home is None:
return os.path.abspath('.')
else:
return os.path.abspath('.')
else:
config_home = os.path.join(os.environ.get('HOME'), '.config')
return os.path.join(config_home, 'slixfeed')
import filehandler
async def get_value_default(key):
@ -101,11 +35,11 @@ async def get_value_default(key):
case "filter-allow":
result = "hitler,sadam,saddam"
case "filter-deny":
result = "crim,dead,death,disaster,holocaust,murder,war"
result = "crim,dead,death,disaster,murder,war"
case "interval":
result = 30
result = 300
case "quantum":
result = 4
result = 3
case "random":
result = 0
return result
@ -121,7 +55,7 @@ def get_list():
Dictionary of pathnames.
"""
paths = []
cfg_dir = get_default_confdir()
cfg_dir = filehandler.get_default_confdir()
if not os.path.isdir(cfg_dir):
os.mkdir(cfg_dir)
cfg_file = os.path.join(cfg_dir, r"url_paths.txt")

View file

@ -4,7 +4,6 @@
import aiohttp
import asyncio
import feedparser
import os
import sqlitehandler
import confighandler
@ -22,39 +21,6 @@ from urllib.parse import urlunsplit
from lxml import html
# NOTE Perhaps this needs to be executed
# just once per program execution
async def initdb(jid, callback, message=None):
"""
Callback function to instantiate action on database.
Parameters
----------
jid : str
Jabber ID.
callback : ?
Function name.
message : str, optional
Optional kwarg when a message is a part or
required argument. The default is None.
Returns
-------
object
Coroutine object.
"""
db_dir = confighandler.get_default_dbdir()
if not os.path.isdir(db_dir):
os.mkdir(db_dir)
db_file = os.path.join(db_dir, r"{}.db".format(jid))
sqlitehandler.create_tables(db_file)
# await sqlitehandler.set_default_values(db_file)
if message:
return await callback(db_file, message)
else:
return await callback(db_file)
async def download_updates(db_file, url=None):
"""
Check feeds for new entries.
@ -207,7 +173,7 @@ async def download_updates(db_file, url=None):
string
)
if reject_list:
print(">>> REJECTED", title)
# print(">>> REJECTED", title)
summary = "REJECTED"
# summary = ""
read_status = 1
@ -630,21 +596,39 @@ async def feed_mode_request(db_file, url, tree):
res = await download_feed(address)
if res[1] == 200:
try:
title = feedparser.parse(res[0])["feed"]["title"]
feeds[address] = feedparser.parse(res[0])
# print(feeds)
except:
title = '*** No Title ***'
feeds[address] = title
continue
if len(feeds) > 1:
positive = 0
msg = (
"RSS URL discovery has found {} feeds:\n```\n"
).format(len(feeds))
for feed in feeds:
feed_name = feeds[feed]
feed_name = feeds[feed]["feed"]["title"]
feed_addr = feed
msg += "{}\n{}\n\n".format(feed_name, feed_addr)
feed_amnt = len(feeds[feed].entries)
if feed_amnt:
positive = 1
msg += (
"Title: {}\n"
" Link: {}\n"
"Count: {}\n"
"\n"
).format(
feed_name,
feed_addr,
feed_amnt
)
msg += (
"```\nThe above feeds were extracted from\n{}"
).format(url)
if not positive:
msg = (
"No feeds were found for {}."
).format(url)
return msg
elif feeds:
feed_addr = list(feeds)[0]
msg = await add_feed(db_file, feed_addr)
@ -709,11 +693,12 @@ async def feed_mode_scan(db_file, url, tree):
res = await download_feed(address)
if res[1] == 200:
try:
feeds[address] = feedparser.parse(res[0])["feed"]["title"]
print(feeds)
feeds[address] = feedparser.parse(res[0])
# print(feeds)
except:
continue
if len(feeds) > 1:
positive = 0
msg = (
"RSS URL scan has found {} feeds:\n```\n"
).format(len(feeds))
@ -722,12 +707,28 @@ async def feed_mode_scan(db_file, url, tree):
# res = await download_feed(feed)
# except:
# continue
feed_name = feeds[feed]
feed_name = feeds[feed]["feed"]["title"]
feed_addr = feed
msg += "{}\n{}\n\n".format(feed_name, feed_addr)
feed_amnt = len(feeds[feed].entries)
if feed_amnt:
positive = 1
msg += (
"Title: {}\n"
" Link: {}\n"
"Count: {}\n"
"\n"
).format(
feed_name,
feed_addr,
feed_amnt
)
msg += (
"```\nThe above feeds were extracted from\n{}"
).format(url)
if not positive:
msg = (
"No feeds were found for {}."
).format(url)
return msg
elif feeds:
feed_addr = list(feeds)[0]

105
slixfeed/filehandler.py Normal file
View file

@ -0,0 +1,105 @@
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
import os
import sys
import sqlitehandler
def get_default_dbdir():
"""
Determine the directory path where dbfile will be stored.
* If $XDG_DATA_HOME is defined, use it;
* else if $HOME exists, use it;
* else if the platform is Windows, use %APPDATA%;
* else use the current directory.
Returns
-------
str
Path to database file.
Note
----
This function was taken from project buku.
See https://github.com/jarun/buku
* Arun Prakash Jana (jarun)
* Dmitry Marakasov (AMDmi3)
"""
# data_home = xdg.BaseDirectory.xdg_data_home
data_home = os.environ.get('XDG_DATA_HOME')
if data_home is None:
if os.environ.get('HOME') is None:
if sys.platform == 'win32':
data_home = os.environ.get('APPDATA')
if data_home is None:
return os.path.abspath('.')
else:
return os.path.abspath('.')
else:
data_home = os.path.join(os.environ.get('HOME'), '.local', 'share')
return os.path.join(data_home, 'slixfeed')
def get_default_confdir():
"""
Determine the directory path where configuration will be stored.
* If $XDG_CONFIG_HOME is defined, use it;
* else if $HOME exists, use it;
* else if the platform is Windows, use %APPDATA%;
* else use the current directory.
Returns
-------
str
Path to configueation directory.
"""
# config_home = xdg.BaseDirectory.xdg_config_home
config_home = os.environ.get('XDG_CONFIG_HOME')
if config_home is None:
if os.environ.get('HOME') is None:
if sys.platform == 'win32':
config_home = os.environ.get('APPDATA')
if config_home is None:
return os.path.abspath('.')
else:
return os.path.abspath('.')
else:
config_home = os.path.join(os.environ.get('HOME'), '.config')
return os.path.join(config_home, 'slixfeed')
# NOTE Perhaps this needs to be executed
# just once per program execution
async def initdb(jid, callback, message=None):
"""
Callback function to instantiate action on database.
Parameters
----------
jid : str
Jabber ID.
callback : ?
Function name.
message : str, optional
Optional kwarg when a message is a part or
required argument. The default is None.
Returns
-------
object
Coroutine object.
"""
db_dir = get_default_dbdir()
if not os.path.isdir(db_dir):
os.mkdir(db_dir)
db_file = os.path.join(db_dir, r"{}.db".format(jid))
sqlitehandler.create_tables(db_file)
# await sqlitehandler.set_default_values(db_file)
if message:
return await callback(db_file, message)
else:
return await callback(db_file)

View file

@ -60,7 +60,7 @@ async def is_listed(db_file, type, string):
if not i or len(i) < 2:
continue
if i in string.lower():
print(">>> ACTIVATE", i)
# print(">>> ACTIVATE", i)
return 1
else:
return None

113
slixfeed/main.py Normal file
View file

@ -0,0 +1,113 @@
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
FIXME
1) Check feed duplication on runtime.
When feed is valid and is not yet in the database it is
posible to send a batch which would result in duplication.
Consequently, it might result in database lock error upon
feed removal attempt
TODO
1) SQL prepared statements
2) Machine Learning for scrapping Title, Link, Summary and Timstamp
3) Support MUC
4) Support categories
5) Default prepackaged list of feeds
6) XMPP commands
7) Bot as transport
8) OMEMO
9) Logging
10) Default feeds (e.g. Blacklisted News, TBOT etc.)
11) Download and upload/send article (xHTML, xHTMLZ, Markdown, MHTML, TXT)
Use Readability
12) Fetch summary from URL, instead of storing summary.
13) Support protocol Gopher
https://github.com/michael-lazar/pygopherd
https://github.com/gopherball/gb
"""
# vars and their meanings:
# jid = Jabber ID (XMPP)
# res = response (HTTP)
from argparse import ArgumentParser
from getpass import getpass
import logging
from datetime import date
import time
# from eliot import start_action, to_file
# # to_file(open("slixfeed.log", "w"))
# # with start_action(action_type="set_date()", jid=jid):
# # with start_action(action_type="message()", msg=msg):
#import irchandler
import xmpphandler
#import matrixhandler
if __name__ == '__main__':
# Setup the command line arguments.
parser = ArgumentParser(description=xmpphandler.Slixfeed.__doc__)
# Output verbosity options.
parser.add_argument(
"-q", "--quiet", help="set logging to ERROR",
action="store_const", dest="loglevel",
const=logging.ERROR, default=logging.INFO
)
parser.add_argument(
"-d", "--debug", help="set logging to DEBUG",
action="store_const", dest="loglevel",
const=logging.DEBUG, default=logging.INFO
)
# JID and password options.
parser.add_argument("-j", "--jid", dest="jid",
help="JID to use")
parser.add_argument("-p", "--password", dest="password",
help="password to use")
args = parser.parse_args()
# Setup logging.
logging.basicConfig(level=args.loglevel,
format='%(levelname)-8s %(message)s')
if args.jid is None:
args.jid = input("Username: ")
if args.password is None:
args.password = getpass("Password: ")
# Setup the Slixfeed and register plugins. Note that while plugins may
# have interdependencies, the order in which you register them does
# not matter.
xmpp = xmpphandler.Slixfeed(args.jid, args.password)
xmpp.register_plugin('xep_0004') # Data Forms
xmpp.register_plugin('xep_0030') # Service Discovery
xmpp.register_plugin('xep_0045') # Multi-User Chat
xmpp.register_plugin('xep_0060') # PubSub
xmpp.register_plugin('xep_0199') # XMPP Ping
# Connect to the XMPP server and start processing XMPP stanzas.
xmpp.connect()
xmpp.process()

View file

@ -249,6 +249,12 @@ async def remove_feed(db_file, ix):
"WHERE source = ?"
)
cur.execute(sql, (url,))
sql = (
"DELETE "
"FROM archive "
"WHERE source = ?"
)
cur.execute(sql, (url,))
sql = (
"DELETE FROM feeds "
"WHERE id = ?"
@ -360,9 +366,18 @@ async def get_number_of_entries_unread(db_file):
with create_connection(db_file) as conn:
cur = conn.cursor()
sql = (
"SELECT "
"("
"SELECT count(id) "
"FROM entries "
"WHERE read = 0"
") "
"+ "
"("
"SELECT count(id) "
"FROM archive"
") "
"AS total_count"
)
count = cur.execute(sql).fetchone()[0]
return count
@ -391,12 +406,32 @@ async def get_entry_unread(db_file, num=None):
num = int(num)
with create_connection(db_file) as conn:
cur = conn.cursor()
# sql = "SELECT id FROM entries WHERE read = 0 LIMIT 1"
# sql = "SELECT id FROM entries WHERE read = 0 ORDER BY timestamp DESC LIMIT 1"
# sql = (
# "SELECT id "
# "FROM entries "
# "WHERE read = 0 "
# "LIMIT 1"
# )
# sql = ("SELECT id "
# "FROM entries "
# "WHERE read = 0 "
# "ORDER BY timestamp DESC "
# "LIMIT 1"
# )
# sql = (
# "SELECT id, title, summary, link "
# "FROM entries "
# "WHERE read = 0 "
# "ORDER BY timestamp "
# "DESC LIMIT :num"
# )
sql = (
"SELECT id, title, summary, link "
"SELECT id, title, summary, link, timestamp "
"FROM entries "
"WHERE read = 0 "
"UNION ALL "
"SELECT id, title, summary, link, timestamp "
"FROM archive "
"ORDER BY timestamp "
"DESC LIMIT :num"
)
@ -444,7 +479,11 @@ async def get_entry_unread(db_file, num=None):
str(link)
)
async with DBLOCK:
# NOTE: We can use DBLOCK once for both
# functions, because, due to exclusive
# ID, only one can ever occur.
await mark_as_read(cur, ix)
await delete_entry(cur, ix)
return news_list
@ -467,6 +506,24 @@ async def mark_as_read(cur, ix):
cur.execute(sql, (ix,))
async def delete_entry(cur, ix):
"""
Delete entry from table archive.
Parameters
----------
db_file : str
Path to database file.
ix : str
Index of entry.
"""
sql = (
"DELETE FROM archive "
"WHERE id = ?"
)
cur.execute(sql, (ix,))
async def statistics(db_file):
"""
Return table statistics.
@ -803,8 +860,8 @@ async def remove_entry(db_file, source, length):
async def remove_nonexistent_entries(db_file, feed, source):
"""
Remove entries that don't exist in a given parsed feed.
Check the entries returned from feed and delete non
existing entries
Check the entries returned from feed and delete read non
existing entries, otherwise move to table archive, if unread.
Parameters
----------
@ -824,12 +881,20 @@ async def remove_nonexistent_entries(db_file, feed, source):
)
items = cur.execute(sql, (source,)).fetchall()
entries = feed.entries
for entry in entries:
valid = False
# breakpoint()
for item in items:
valid = False
for entry in entries:
title = None
link = None
time = None
# valid = False
# TODO better check and don't repeat code
if entry.has_key("id") and item[3]:
if entry.id == item[3]:
# print("compare1:", entry.id)
# print("compare2:", item[3])
# print("============")
valid = True
break
else:
@ -842,6 +907,9 @@ async def remove_nonexistent_entries(db_file, feed, source):
else:
link = source
if entry.has_key("published") and item[4]:
# print("compare11:", title, link, time)
# print("compare22:", item[1], item[2], item[4])
# print("============")
time = await datetimehandler.rfc2822_to_iso8601(entry.published)
if (item[1] == title and
item[2] == link and
@ -851,17 +919,41 @@ async def remove_nonexistent_entries(db_file, feed, source):
else:
if (item[1] == title and
item[2] == link):
# print("compare111:", title, link)
# print("compare222:", item[1], item[2])
# print("============")
valid = True
break
# TODO better check and don't repeat code
if not valid:
# print("id: ", item[0])
# if title:
# print("title: ", title)
# print("item[1]: ", item[1])
# if link:
# print("link: ", link)
# print("item[2]: ", item[2])
# if entry.id:
# print("last_entry:", entry.id)
# print("item[3]: ", item[3])
# if time:
# print("time: ", time)
# print("item[4]: ", item[4])
# print("read: ", item[5])
# breakpoint()
async with DBLOCK:
# TODO Send to table archive
# TODO Also make a regular/routine check for sources that
# have been changed (though that can only happen when
# manually editing)
ix = item[0]
print(">>> SOURCE: ", source)
print(">>> INVALID:", item[1])
# print("title:", item[1])
# print("link :", item[2])
# print("id :", item[3])
if item[5] == 1:
print(">>> DELETING:", item[1])
sql = (
"DELETE "
"FROM entries "
@ -869,16 +961,11 @@ async def remove_nonexistent_entries(db_file, feed, source):
)
cur.execute(sql, (ix,))
else:
print(">>> ARCHIVING:")
print("title:", item[1])
print("link :", item[2])
print("id :", item[3])
print(">>> ARCHIVING:", item[1])
sql = (
"INSERT "
"INTO archive "
"SELECT * "
# "SELECT title, summary, "
# "link, source, timestamp "
"FROM entries "
"WHERE entries.id = :ix"
)
@ -1015,13 +1102,18 @@ async def last_entries(db_file, num):
elif num < 1:
num = 1
cur = get_cursor(db_file)
# sql = "SELECT title, link FROM entries ORDER BY ROWID DESC LIMIT :num"
# sql = (
# "SELECT title, link "
# "FROM entries "
# "ORDER BY ROWID DESC "
# "LIMIT :num"
# )
sql = (
"SELECT title, link "
"FROM entries "
"WHERE read = 0 "
"ORDER BY timestamp "
"DESC LIMIT :num "
"ORDER BY timestamp DESC "
"LIMIT :num "
)
results = cur.execute(sql, (num,))
titles_list = "Recent {} titles:\n".format(num)
@ -1053,7 +1145,7 @@ async def search_feeds(db_file, query):
"""
cur = get_cursor(db_file)
sql = (
"SELECT name, id, address "
"SELECT name, address, id "
"FROM feeds "
"WHERE name LIKE ? "
"LIMIT 50"
@ -1066,7 +1158,10 @@ async def search_feeds(db_file, query):
for result in results:
counter += 1
results_list += (
"\n{} [{}]\n{}\n"
"\nName: {}"
"\n URL: {}"
"\n ID: {}"
"\n"
).format(
str(result[0]),
str(result[1]),
@ -1099,9 +1194,16 @@ async def search_entries(db_file, query):
"SELECT title, link "
"FROM entries "
"WHERE title LIKE ? "
"UNION ALL "
"SELECT title, link "
"FROM archive "
"WHERE title LIKE ? "
"LIMIT 50"
)
results = cur.execute(sql, [f'%{query}%'])
results = cur.execute(sql, (
f'%{query}%',
f'%{query}%'
))
results_list = (
"Search results for '{}':\n```"
).format(query)
@ -1168,11 +1270,15 @@ async def check_entry_exist(db_file, source, eid=None,
"link = :link and "
"timestamp = :date"
)
try:
result = cur.execute(sql, {
"title": title,
"link": link,
"timestamp": date
}).fetchone()
except:
print("this is source:", source)
print("this is date: ", date)
else:
sql = (
"SELECT id "
@ -1183,10 +1289,13 @@ async def check_entry_exist(db_file, source, eid=None,
"title": title,
"link": link
}).fetchone()
try:
if result:
return True
else:
None
except:
print("no result. this is source:", source)
async def set_settings_value(db_file, key_value):

View file

@ -19,9 +19,9 @@ import slixmpp
from slixmpp.plugins.xep_0363.http_upload import FileTooBig, HTTPError, UploadServiceNotFound
import confighandler
import datahandler
import datetimehandler
import filehandler
import filterhandler
import sqlitehandler
@ -115,7 +115,7 @@ class Slixfeed(slixmpp.ClientXMPP):
action = 0
jid = msg["from"].bare
db_dir = confighandler.get_default_dbdir()
db_dir = filehandler.get_default_dbdir()
os.chdir(db_dir)
if jid + ".db" not in os.listdir():
await self.task_jid(jid)
@ -140,7 +140,7 @@ class Slixfeed(slixmpp.ClientXMPP):
url = message.split(" ")[0]
title = " ".join(message.split(" ")[1:])
if url.startswith("http"):
action = await datahandler.initdb(
action = await filehandler.initdb(
jid,
datahandler.add_feed_no_check,
[url, title]
@ -152,7 +152,7 @@ class Slixfeed(slixmpp.ClientXMPP):
key = "filter-" + message[:5]
val = message[6:]
if val:
keywords = await datahandler.initdb(
keywords = await filehandler.initdb(
jid,
sqlitehandler.get_settings_value,
key
@ -161,7 +161,7 @@ class Slixfeed(slixmpp.ClientXMPP):
val,
keywords
)
await datahandler.initdb(
await filehandler.initdb(
jid,
sqlitehandler.set_settings_value,
[key, val]
@ -176,7 +176,7 @@ class Slixfeed(slixmpp.ClientXMPP):
key = "filter-" + message[:4]
val = message[5:]
if val:
keywords = await datahandler.initdb(
keywords = await filehandler.initdb(
jid,
sqlitehandler.get_settings_value,
key
@ -185,7 +185,7 @@ class Slixfeed(slixmpp.ClientXMPP):
val,
keywords
)
await datahandler.initdb(
await filehandler.initdb(
jid,
sqlitehandler.set_settings_value,
[key, val]
@ -198,7 +198,7 @@ class Slixfeed(slixmpp.ClientXMPP):
action = "Missing keywords."
case _ if message_lowercase.startswith("http"):
url = message
action = await datahandler.initdb(
action = await filehandler.initdb(
jid,
datahandler.add_feed,
url
@ -209,7 +209,7 @@ class Slixfeed(slixmpp.ClientXMPP):
query = message[6:]
if query:
if len(query) > 3:
action = await datahandler.initdb(
action = await filehandler.initdb(
jid,
sqlitehandler.search_feeds,
query
@ -219,7 +219,7 @@ class Slixfeed(slixmpp.ClientXMPP):
"Enter at least 4 characters to search"
)
else:
action = await datahandler.initdb(
action = await filehandler.initdb(
jid,
sqlitehandler.list_feeds
)
@ -235,7 +235,7 @@ class Slixfeed(slixmpp.ClientXMPP):
# action = (
# "Updates will be sent every {} minutes."
# ).format(action)
await datahandler.initdb(
await filehandler.initdb(
jid,
sqlitehandler.set_settings_value,
[key, val]
@ -263,7 +263,7 @@ class Slixfeed(slixmpp.ClientXMPP):
# action = (
# "Every update will contain {} news items."
# ).format(action)
await datahandler.initdb(
await filehandler.initdb(
jid,
sqlitehandler.set_settings_value,
[key, val]
@ -278,7 +278,7 @@ class Slixfeed(slixmpp.ClientXMPP):
case _ if message_lowercase.startswith("recent"):
num = message[7:]
if num:
action = await datahandler.initdb(
action = await filehandler.initdb(
jid,
sqlitehandler.last_entries,
num
@ -288,7 +288,7 @@ class Slixfeed(slixmpp.ClientXMPP):
case _ if message_lowercase.startswith("remove"):
ix = message[7:]
if ix:
action = await datahandler.initdb(
action = await filehandler.initdb(
jid,
sqlitehandler.remove_feed,
ix
@ -300,7 +300,7 @@ class Slixfeed(slixmpp.ClientXMPP):
query = message[7:]
if query:
if len(query) > 1:
action = await datahandler.initdb(
action = await filehandler.initdb(
jid,
sqlitehandler.search_entries,
query
@ -315,7 +315,7 @@ class Slixfeed(slixmpp.ClientXMPP):
# action = "Updates are enabled."
key = "enabled"
val = 1
await datahandler.initdb(
await filehandler.initdb(
jid,
sqlitehandler.set_settings_value,
[key, val]
@ -325,13 +325,13 @@ class Slixfeed(slixmpp.ClientXMPP):
# print(await datetimehandler.current_time(), "task_manager[jid]")
# print(task_manager[jid])
case "stats":
action = await datahandler.initdb(
action = await filehandler.initdb(
jid,
sqlitehandler.statistics
)
case _ if message_lowercase.startswith("status "):
ix = message[7:]
action = await datahandler.initdb(
action = await filehandler.initdb(
jid,
sqlitehandler.toggle_status,
ix
@ -349,7 +349,7 @@ class Slixfeed(slixmpp.ClientXMPP):
# task_manager[jid]["interval"].cancel()
# key = "enabled"
# val = 0
# action = await datahandler.initdb(
# action = await filehandler.initdb(
# jid,
# sqlitehandler.set_settings_value,
# [key, val]
@ -360,7 +360,7 @@ class Slixfeed(slixmpp.ClientXMPP):
# # await self.send_status(jid)
key = "enabled"
val = 0
await datahandler.initdb(
await filehandler.initdb(
jid,
sqlitehandler.set_settings_value,
[key, val]
@ -388,7 +388,7 @@ class Slixfeed(slixmpp.ClientXMPP):
Self.
"""
while True:
db_dir = confighandler.get_default_dbdir()
db_dir = filehandler.get_default_dbdir()
if not os.path.isdir(db_dir):
msg = (
"Slixfeed can not work without a database.\n"
@ -410,7 +410,8 @@ class Slixfeed(slixmpp.ClientXMPP):
# await jid_tasker[jid]
async with asyncio.TaskGroup() as tg:
for file in files:
if file.endswith(".db") and not file.endswith(".db-jour.db"):
if (file.endswith(".db") and
not file.endswith(".db-jour.db")):
jid = file[:-3]
main_task.extend([tg.create_task(self.task_jid(jid))])
# main_task = [tg.create_task(self.task_jid(jid))]
@ -428,7 +429,7 @@ class Slixfeed(slixmpp.ClientXMPP):
jid : str
Jabber ID.
"""
enabled = await datahandler.initdb(
enabled = await filehandler.initdb(
jid,
sqlitehandler.get_settings_value,
"enabled"
@ -476,7 +477,7 @@ class Slixfeed(slixmpp.ClientXMPP):
"""
# print("Starting send_update()")
# print(jid)
new = await datahandler.initdb(
new = await filehandler.initdb(
jid,
sqlitehandler.get_entry_unread,
num
@ -493,7 +494,7 @@ class Slixfeed(slixmpp.ClientXMPP):
self.send_update,
"interval"
)
# interval = await datahandler.initdb(
# interval = await filehandler.initdb(
# jid,
# sqlitehandler.get_settings_value,
# "interval"
@ -530,7 +531,7 @@ class Slixfeed(slixmpp.ClientXMPP):
Jabber ID.
"""
print(await datetimehandler.current_time(), "> SEND STATUS",jid)
enabled = await datahandler.initdb(
enabled = await filehandler.initdb(
jid,
sqlitehandler.get_settings_value,
"enabled"
@ -539,7 +540,7 @@ class Slixfeed(slixmpp.ClientXMPP):
status_mode = "xa"
status_text = "Send \"Start\" to receive news."
else:
feeds = await datahandler.initdb(
feeds = await filehandler.initdb(
jid,
sqlitehandler.get_number_of_items,
"feeds"
@ -550,7 +551,7 @@ class Slixfeed(slixmpp.ClientXMPP):
"📂️ Send a URL from a blog or a news website."
)
else:
unread = await datahandler.initdb(
unread = await filehandler.initdb(
jid,
sqlitehandler.get_number_of_entries_unread
)
@ -606,7 +607,7 @@ class Slixfeed(slixmpp.ClientXMPP):
Value. The default is None.
"""
if not val:
val = await datahandler.initdb(
val = await filehandler.initdb(
jid,
sqlitehandler.get_settings_value,
key
@ -644,7 +645,7 @@ async def check_updates(jid):
"""
while True:
print(await datetimehandler.current_time(), "> CHCK UPDATE",jid)
await datahandler.initdb(jid, datahandler.download_updates)
await filehandler.initdb(jid, datahandler.download_updates)
await asyncio.sleep(60 * 90)
# Schedule to call this function again in 90 minutes
# loop.call_at(