Fix archiving. WARNING of complexity due to UNION of SQL
This commit is contained in:
parent
9a93d6de0e
commit
cb2317b35a
7 changed files with 466 additions and 203 deletions
|
@ -11,73 +11,7 @@ TODO
|
|||
"""
|
||||
|
||||
import os
|
||||
import sys
|
||||
|
||||
def get_default_dbdir():
|
||||
"""
|
||||
Determine the directory path where dbfile will be stored.
|
||||
|
||||
* If $XDG_DATA_HOME is defined, use it;
|
||||
* else if $HOME exists, use it;
|
||||
* else if the platform is Windows, use %APPDATA%;
|
||||
* else use the current directory.
|
||||
|
||||
Returns
|
||||
-------
|
||||
str
|
||||
Path to database file.
|
||||
|
||||
Note
|
||||
----
|
||||
This function was taken from project buku.
|
||||
|
||||
See https://github.com/jarun/buku
|
||||
|
||||
* Arun Prakash Jana (jarun)
|
||||
* Dmitry Marakasov (AMDmi3)
|
||||
"""
|
||||
# data_home = xdg.BaseDirectory.xdg_data_home
|
||||
data_home = os.environ.get('XDG_DATA_HOME')
|
||||
if data_home is None:
|
||||
if os.environ.get('HOME') is None:
|
||||
if sys.platform == 'win32':
|
||||
data_home = os.environ.get('APPDATA')
|
||||
if data_home is None:
|
||||
return os.path.abspath('.')
|
||||
else:
|
||||
return os.path.abspath('.')
|
||||
else:
|
||||
data_home = os.path.join(os.environ.get('HOME'), '.local', 'share')
|
||||
return os.path.join(data_home, 'slixfeed')
|
||||
|
||||
|
||||
def get_default_confdir():
|
||||
"""
|
||||
Determine the directory path where configuration will be stored.
|
||||
|
||||
* If $XDG_CONFIG_HOME is defined, use it;
|
||||
* else if $HOME exists, use it;
|
||||
* else if the platform is Windows, use %APPDATA%;
|
||||
* else use the current directory.
|
||||
|
||||
Returns
|
||||
-------
|
||||
str
|
||||
Path to configueation directory.
|
||||
"""
|
||||
# config_home = xdg.BaseDirectory.xdg_config_home
|
||||
config_home = os.environ.get('XDG_CONFIG_HOME')
|
||||
if config_home is None:
|
||||
if os.environ.get('HOME') is None:
|
||||
if sys.platform == 'win32':
|
||||
config_home = os.environ.get('APPDATA')
|
||||
if config_home is None:
|
||||
return os.path.abspath('.')
|
||||
else:
|
||||
return os.path.abspath('.')
|
||||
else:
|
||||
config_home = os.path.join(os.environ.get('HOME'), '.config')
|
||||
return os.path.join(config_home, 'slixfeed')
|
||||
import filehandler
|
||||
|
||||
|
||||
async def get_value_default(key):
|
||||
|
@ -101,11 +35,11 @@ async def get_value_default(key):
|
|||
case "filter-allow":
|
||||
result = "hitler,sadam,saddam"
|
||||
case "filter-deny":
|
||||
result = "crim,dead,death,disaster,holocaust,murder,war"
|
||||
result = "crim,dead,death,disaster,murder,war"
|
||||
case "interval":
|
||||
result = 30
|
||||
result = 300
|
||||
case "quantum":
|
||||
result = 4
|
||||
result = 3
|
||||
case "random":
|
||||
result = 0
|
||||
return result
|
||||
|
@ -121,7 +55,7 @@ def get_list():
|
|||
Dictionary of pathnames.
|
||||
"""
|
||||
paths = []
|
||||
cfg_dir = get_default_confdir()
|
||||
cfg_dir = filehandler.get_default_confdir()
|
||||
if not os.path.isdir(cfg_dir):
|
||||
os.mkdir(cfg_dir)
|
||||
cfg_file = os.path.join(cfg_dir, r"url_paths.txt")
|
||||
|
|
|
@ -4,7 +4,6 @@
|
|||
import aiohttp
|
||||
import asyncio
|
||||
import feedparser
|
||||
import os
|
||||
|
||||
import sqlitehandler
|
||||
import confighandler
|
||||
|
@ -22,39 +21,6 @@ from urllib.parse import urlunsplit
|
|||
from lxml import html
|
||||
|
||||
|
||||
# NOTE Perhaps this needs to be executed
|
||||
# just once per program execution
|
||||
async def initdb(jid, callback, message=None):
|
||||
"""
|
||||
Callback function to instantiate action on database.
|
||||
|
||||
Parameters
|
||||
----------
|
||||
jid : str
|
||||
Jabber ID.
|
||||
callback : ?
|
||||
Function name.
|
||||
message : str, optional
|
||||
Optional kwarg when a message is a part or
|
||||
required argument. The default is None.
|
||||
|
||||
Returns
|
||||
-------
|
||||
object
|
||||
Coroutine object.
|
||||
"""
|
||||
db_dir = confighandler.get_default_dbdir()
|
||||
if not os.path.isdir(db_dir):
|
||||
os.mkdir(db_dir)
|
||||
db_file = os.path.join(db_dir, r"{}.db".format(jid))
|
||||
sqlitehandler.create_tables(db_file)
|
||||
# await sqlitehandler.set_default_values(db_file)
|
||||
if message:
|
||||
return await callback(db_file, message)
|
||||
else:
|
||||
return await callback(db_file)
|
||||
|
||||
|
||||
async def download_updates(db_file, url=None):
|
||||
"""
|
||||
Check feeds for new entries.
|
||||
|
@ -207,7 +173,7 @@ async def download_updates(db_file, url=None):
|
|||
string
|
||||
)
|
||||
if reject_list:
|
||||
print(">>> REJECTED", title)
|
||||
# print(">>> REJECTED", title)
|
||||
summary = "REJECTED"
|
||||
# summary = ""
|
||||
read_status = 1
|
||||
|
@ -630,21 +596,39 @@ async def feed_mode_request(db_file, url, tree):
|
|||
res = await download_feed(address)
|
||||
if res[1] == 200:
|
||||
try:
|
||||
title = feedparser.parse(res[0])["feed"]["title"]
|
||||
feeds[address] = feedparser.parse(res[0])
|
||||
# print(feeds)
|
||||
except:
|
||||
title = '*** No Title ***'
|
||||
feeds[address] = title
|
||||
continue
|
||||
if len(feeds) > 1:
|
||||
positive = 0
|
||||
msg = (
|
||||
"RSS URL discovery has found {} feeds:\n```\n"
|
||||
).format(len(feeds))
|
||||
for feed in feeds:
|
||||
feed_name = feeds[feed]
|
||||
feed_name = feeds[feed]["feed"]["title"]
|
||||
feed_addr = feed
|
||||
msg += "{}\n{}\n\n".format(feed_name, feed_addr)
|
||||
feed_amnt = len(feeds[feed].entries)
|
||||
if feed_amnt:
|
||||
positive = 1
|
||||
msg += (
|
||||
"Title: {}\n"
|
||||
" Link: {}\n"
|
||||
"Count: {}\n"
|
||||
"\n"
|
||||
).format(
|
||||
feed_name,
|
||||
feed_addr,
|
||||
feed_amnt
|
||||
)
|
||||
msg += (
|
||||
"```\nThe above feeds were extracted from\n{}"
|
||||
).format(url)
|
||||
if not positive:
|
||||
msg = (
|
||||
"No feeds were found for {}."
|
||||
).format(url)
|
||||
return msg
|
||||
elif feeds:
|
||||
feed_addr = list(feeds)[0]
|
||||
msg = await add_feed(db_file, feed_addr)
|
||||
|
@ -709,11 +693,12 @@ async def feed_mode_scan(db_file, url, tree):
|
|||
res = await download_feed(address)
|
||||
if res[1] == 200:
|
||||
try:
|
||||
feeds[address] = feedparser.parse(res[0])["feed"]["title"]
|
||||
print(feeds)
|
||||
feeds[address] = feedparser.parse(res[0])
|
||||
# print(feeds)
|
||||
except:
|
||||
continue
|
||||
if len(feeds) > 1:
|
||||
positive = 0
|
||||
msg = (
|
||||
"RSS URL scan has found {} feeds:\n```\n"
|
||||
).format(len(feeds))
|
||||
|
@ -722,12 +707,28 @@ async def feed_mode_scan(db_file, url, tree):
|
|||
# res = await download_feed(feed)
|
||||
# except:
|
||||
# continue
|
||||
feed_name = feeds[feed]
|
||||
feed_name = feeds[feed]["feed"]["title"]
|
||||
feed_addr = feed
|
||||
msg += "{}\n{}\n\n".format(feed_name, feed_addr)
|
||||
feed_amnt = len(feeds[feed].entries)
|
||||
if feed_amnt:
|
||||
positive = 1
|
||||
msg += (
|
||||
"Title: {}\n"
|
||||
" Link: {}\n"
|
||||
"Count: {}\n"
|
||||
"\n"
|
||||
).format(
|
||||
feed_name,
|
||||
feed_addr,
|
||||
feed_amnt
|
||||
)
|
||||
msg += (
|
||||
"```\nThe above feeds were extracted from\n{}"
|
||||
).format(url)
|
||||
if not positive:
|
||||
msg = (
|
||||
"No feeds were found for {}."
|
||||
).format(url)
|
||||
return msg
|
||||
elif feeds:
|
||||
feed_addr = list(feeds)[0]
|
||||
|
|
105
slixfeed/filehandler.py
Normal file
105
slixfeed/filehandler.py
Normal file
|
@ -0,0 +1,105 @@
|
|||
#!/usr/bin/env python3
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
import os
|
||||
import sys
|
||||
|
||||
import sqlitehandler
|
||||
|
||||
def get_default_dbdir():
|
||||
"""
|
||||
Determine the directory path where dbfile will be stored.
|
||||
|
||||
* If $XDG_DATA_HOME is defined, use it;
|
||||
* else if $HOME exists, use it;
|
||||
* else if the platform is Windows, use %APPDATA%;
|
||||
* else use the current directory.
|
||||
|
||||
Returns
|
||||
-------
|
||||
str
|
||||
Path to database file.
|
||||
|
||||
Note
|
||||
----
|
||||
This function was taken from project buku.
|
||||
|
||||
See https://github.com/jarun/buku
|
||||
|
||||
* Arun Prakash Jana (jarun)
|
||||
* Dmitry Marakasov (AMDmi3)
|
||||
"""
|
||||
# data_home = xdg.BaseDirectory.xdg_data_home
|
||||
data_home = os.environ.get('XDG_DATA_HOME')
|
||||
if data_home is None:
|
||||
if os.environ.get('HOME') is None:
|
||||
if sys.platform == 'win32':
|
||||
data_home = os.environ.get('APPDATA')
|
||||
if data_home is None:
|
||||
return os.path.abspath('.')
|
||||
else:
|
||||
return os.path.abspath('.')
|
||||
else:
|
||||
data_home = os.path.join(os.environ.get('HOME'), '.local', 'share')
|
||||
return os.path.join(data_home, 'slixfeed')
|
||||
|
||||
|
||||
def get_default_confdir():
|
||||
"""
|
||||
Determine the directory path where configuration will be stored.
|
||||
|
||||
* If $XDG_CONFIG_HOME is defined, use it;
|
||||
* else if $HOME exists, use it;
|
||||
* else if the platform is Windows, use %APPDATA%;
|
||||
* else use the current directory.
|
||||
|
||||
Returns
|
||||
-------
|
||||
str
|
||||
Path to configueation directory.
|
||||
"""
|
||||
# config_home = xdg.BaseDirectory.xdg_config_home
|
||||
config_home = os.environ.get('XDG_CONFIG_HOME')
|
||||
if config_home is None:
|
||||
if os.environ.get('HOME') is None:
|
||||
if sys.platform == 'win32':
|
||||
config_home = os.environ.get('APPDATA')
|
||||
if config_home is None:
|
||||
return os.path.abspath('.')
|
||||
else:
|
||||
return os.path.abspath('.')
|
||||
else:
|
||||
config_home = os.path.join(os.environ.get('HOME'), '.config')
|
||||
return os.path.join(config_home, 'slixfeed')
|
||||
|
||||
# NOTE Perhaps this needs to be executed
|
||||
# just once per program execution
|
||||
async def initdb(jid, callback, message=None):
|
||||
"""
|
||||
Callback function to instantiate action on database.
|
||||
|
||||
Parameters
|
||||
----------
|
||||
jid : str
|
||||
Jabber ID.
|
||||
callback : ?
|
||||
Function name.
|
||||
message : str, optional
|
||||
Optional kwarg when a message is a part or
|
||||
required argument. The default is None.
|
||||
|
||||
Returns
|
||||
-------
|
||||
object
|
||||
Coroutine object.
|
||||
"""
|
||||
db_dir = get_default_dbdir()
|
||||
if not os.path.isdir(db_dir):
|
||||
os.mkdir(db_dir)
|
||||
db_file = os.path.join(db_dir, r"{}.db".format(jid))
|
||||
sqlitehandler.create_tables(db_file)
|
||||
# await sqlitehandler.set_default_values(db_file)
|
||||
if message:
|
||||
return await callback(db_file, message)
|
||||
else:
|
||||
return await callback(db_file)
|
|
@ -60,7 +60,7 @@ async def is_listed(db_file, type, string):
|
|||
if not i or len(i) < 2:
|
||||
continue
|
||||
if i in string.lower():
|
||||
print(">>> ACTIVATE", i)
|
||||
# print(">>> ACTIVATE", i)
|
||||
return 1
|
||||
else:
|
||||
return None
|
||||
|
|
113
slixfeed/main.py
Normal file
113
slixfeed/main.py
Normal file
|
@ -0,0 +1,113 @@
|
|||
#!/usr/bin/env python3
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
"""
|
||||
|
||||
FIXME
|
||||
|
||||
1) Check feed duplication on runtime.
|
||||
When feed is valid and is not yet in the database it is
|
||||
posible to send a batch which would result in duplication.
|
||||
Consequently, it might result in database lock error upon
|
||||
feed removal attempt
|
||||
|
||||
TODO
|
||||
|
||||
1) SQL prepared statements
|
||||
|
||||
2) Machine Learning for scrapping Title, Link, Summary and Timstamp
|
||||
|
||||
3) Support MUC
|
||||
|
||||
4) Support categories
|
||||
|
||||
5) Default prepackaged list of feeds
|
||||
|
||||
6) XMPP commands
|
||||
|
||||
7) Bot as transport
|
||||
|
||||
8) OMEMO
|
||||
|
||||
9) Logging
|
||||
|
||||
10) Default feeds (e.g. Blacklisted News, TBOT etc.)
|
||||
|
||||
11) Download and upload/send article (xHTML, xHTMLZ, Markdown, MHTML, TXT)
|
||||
Use Readability
|
||||
|
||||
12) Fetch summary from URL, instead of storing summary.
|
||||
|
||||
13) Support protocol Gopher
|
||||
https://github.com/michael-lazar/pygopherd
|
||||
https://github.com/gopherball/gb
|
||||
|
||||
"""
|
||||
|
||||
# vars and their meanings:
|
||||
# jid = Jabber ID (XMPP)
|
||||
# res = response (HTTP)
|
||||
|
||||
from argparse import ArgumentParser
|
||||
from getpass import getpass
|
||||
import logging
|
||||
|
||||
from datetime import date
|
||||
import time
|
||||
|
||||
# from eliot import start_action, to_file
|
||||
# # to_file(open("slixfeed.log", "w"))
|
||||
# # with start_action(action_type="set_date()", jid=jid):
|
||||
# # with start_action(action_type="message()", msg=msg):
|
||||
|
||||
#import irchandler
|
||||
import xmpphandler
|
||||
#import matrixhandler
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
# Setup the command line arguments.
|
||||
parser = ArgumentParser(description=xmpphandler.Slixfeed.__doc__)
|
||||
|
||||
# Output verbosity options.
|
||||
parser.add_argument(
|
||||
"-q", "--quiet", help="set logging to ERROR",
|
||||
action="store_const", dest="loglevel",
|
||||
const=logging.ERROR, default=logging.INFO
|
||||
)
|
||||
parser.add_argument(
|
||||
"-d", "--debug", help="set logging to DEBUG",
|
||||
action="store_const", dest="loglevel",
|
||||
const=logging.DEBUG, default=logging.INFO
|
||||
)
|
||||
|
||||
# JID and password options.
|
||||
parser.add_argument("-j", "--jid", dest="jid",
|
||||
help="JID to use")
|
||||
parser.add_argument("-p", "--password", dest="password",
|
||||
help="password to use")
|
||||
|
||||
args = parser.parse_args()
|
||||
|
||||
# Setup logging.
|
||||
logging.basicConfig(level=args.loglevel,
|
||||
format='%(levelname)-8s %(message)s')
|
||||
|
||||
if args.jid is None:
|
||||
args.jid = input("Username: ")
|
||||
if args.password is None:
|
||||
args.password = getpass("Password: ")
|
||||
|
||||
# Setup the Slixfeed and register plugins. Note that while plugins may
|
||||
# have interdependencies, the order in which you register them does
|
||||
# not matter.
|
||||
xmpp = xmpphandler.Slixfeed(args.jid, args.password)
|
||||
xmpp.register_plugin('xep_0004') # Data Forms
|
||||
xmpp.register_plugin('xep_0030') # Service Discovery
|
||||
xmpp.register_plugin('xep_0045') # Multi-User Chat
|
||||
xmpp.register_plugin('xep_0060') # PubSub
|
||||
xmpp.register_plugin('xep_0199') # XMPP Ping
|
||||
|
||||
# Connect to the XMPP server and start processing XMPP stanzas.
|
||||
xmpp.connect()
|
||||
xmpp.process()
|
|
@ -249,6 +249,12 @@ async def remove_feed(db_file, ix):
|
|||
"WHERE source = ?"
|
||||
)
|
||||
cur.execute(sql, (url,))
|
||||
sql = (
|
||||
"DELETE "
|
||||
"FROM archive "
|
||||
"WHERE source = ?"
|
||||
)
|
||||
cur.execute(sql, (url,))
|
||||
sql = (
|
||||
"DELETE FROM feeds "
|
||||
"WHERE id = ?"
|
||||
|
@ -360,9 +366,18 @@ async def get_number_of_entries_unread(db_file):
|
|||
with create_connection(db_file) as conn:
|
||||
cur = conn.cursor()
|
||||
sql = (
|
||||
"SELECT "
|
||||
"("
|
||||
"SELECT count(id) "
|
||||
"FROM entries "
|
||||
"WHERE read = 0"
|
||||
") "
|
||||
"+ "
|
||||
"("
|
||||
"SELECT count(id) "
|
||||
"FROM archive"
|
||||
") "
|
||||
"AS total_count"
|
||||
)
|
||||
count = cur.execute(sql).fetchone()[0]
|
||||
return count
|
||||
|
@ -391,12 +406,32 @@ async def get_entry_unread(db_file, num=None):
|
|||
num = int(num)
|
||||
with create_connection(db_file) as conn:
|
||||
cur = conn.cursor()
|
||||
# sql = "SELECT id FROM entries WHERE read = 0 LIMIT 1"
|
||||
# sql = "SELECT id FROM entries WHERE read = 0 ORDER BY timestamp DESC LIMIT 1"
|
||||
# sql = (
|
||||
# "SELECT id "
|
||||
# "FROM entries "
|
||||
# "WHERE read = 0 "
|
||||
# "LIMIT 1"
|
||||
# )
|
||||
# sql = ("SELECT id "
|
||||
# "FROM entries "
|
||||
# "WHERE read = 0 "
|
||||
# "ORDER BY timestamp DESC "
|
||||
# "LIMIT 1"
|
||||
# )
|
||||
# sql = (
|
||||
# "SELECT id, title, summary, link "
|
||||
# "FROM entries "
|
||||
# "WHERE read = 0 "
|
||||
# "ORDER BY timestamp "
|
||||
# "DESC LIMIT :num"
|
||||
# )
|
||||
sql = (
|
||||
"SELECT id, title, summary, link "
|
||||
"SELECT id, title, summary, link, timestamp "
|
||||
"FROM entries "
|
||||
"WHERE read = 0 "
|
||||
"UNION ALL "
|
||||
"SELECT id, title, summary, link, timestamp "
|
||||
"FROM archive "
|
||||
"ORDER BY timestamp "
|
||||
"DESC LIMIT :num"
|
||||
)
|
||||
|
@ -444,7 +479,11 @@ async def get_entry_unread(db_file, num=None):
|
|||
str(link)
|
||||
)
|
||||
async with DBLOCK:
|
||||
# NOTE: We can use DBLOCK once for both
|
||||
# functions, because, due to exclusive
|
||||
# ID, only one can ever occur.
|
||||
await mark_as_read(cur, ix)
|
||||
await delete_entry(cur, ix)
|
||||
return news_list
|
||||
|
||||
|
||||
|
@ -467,6 +506,24 @@ async def mark_as_read(cur, ix):
|
|||
cur.execute(sql, (ix,))
|
||||
|
||||
|
||||
async def delete_entry(cur, ix):
|
||||
"""
|
||||
Delete entry from table archive.
|
||||
|
||||
Parameters
|
||||
----------
|
||||
db_file : str
|
||||
Path to database file.
|
||||
ix : str
|
||||
Index of entry.
|
||||
"""
|
||||
sql = (
|
||||
"DELETE FROM archive "
|
||||
"WHERE id = ?"
|
||||
)
|
||||
cur.execute(sql, (ix,))
|
||||
|
||||
|
||||
async def statistics(db_file):
|
||||
"""
|
||||
Return table statistics.
|
||||
|
@ -803,8 +860,8 @@ async def remove_entry(db_file, source, length):
|
|||
async def remove_nonexistent_entries(db_file, feed, source):
|
||||
"""
|
||||
Remove entries that don't exist in a given parsed feed.
|
||||
Check the entries returned from feed and delete non
|
||||
existing entries
|
||||
Check the entries returned from feed and delete read non
|
||||
existing entries, otherwise move to table archive, if unread.
|
||||
|
||||
Parameters
|
||||
----------
|
||||
|
@ -824,12 +881,20 @@ async def remove_nonexistent_entries(db_file, feed, source):
|
|||
)
|
||||
items = cur.execute(sql, (source,)).fetchall()
|
||||
entries = feed.entries
|
||||
for entry in entries:
|
||||
# breakpoint()
|
||||
for item in items:
|
||||
valid = False
|
||||
for item in items:
|
||||
for entry in entries:
|
||||
title = None
|
||||
link = None
|
||||
time = None
|
||||
# valid = False
|
||||
# TODO better check and don't repeat code
|
||||
if entry.has_key("id") and item[3]:
|
||||
if entry.id == item[3]:
|
||||
# print("compare1:", entry.id)
|
||||
# print("compare2:", item[3])
|
||||
# print("============")
|
||||
valid = True
|
||||
break
|
||||
else:
|
||||
|
@ -842,6 +907,9 @@ async def remove_nonexistent_entries(db_file, feed, source):
|
|||
else:
|
||||
link = source
|
||||
if entry.has_key("published") and item[4]:
|
||||
# print("compare11:", title, link, time)
|
||||
# print("compare22:", item[1], item[2], item[4])
|
||||
# print("============")
|
||||
time = await datetimehandler.rfc2822_to_iso8601(entry.published)
|
||||
if (item[1] == title and
|
||||
item[2] == link and
|
||||
|
@ -851,44 +919,63 @@ async def remove_nonexistent_entries(db_file, feed, source):
|
|||
else:
|
||||
if (item[1] == title and
|
||||
item[2] == link):
|
||||
# print("compare111:", title, link)
|
||||
# print("compare222:", item[1], item[2])
|
||||
# print("============")
|
||||
valid = True
|
||||
break
|
||||
# TODO better check and don't repeat code
|
||||
if not valid:
|
||||
async with DBLOCK:
|
||||
# TODO Send to table archive
|
||||
# TODO Also make a regular/routine check for sources that
|
||||
# have been changed (though that can only happen when
|
||||
# manually editing)
|
||||
ix = item[0]
|
||||
if item[5] == 1:
|
||||
sql = (
|
||||
"DELETE "
|
||||
"FROM entries "
|
||||
"WHERE id = :ix"
|
||||
)
|
||||
cur.execute(sql, (ix,))
|
||||
else:
|
||||
print(">>> ARCHIVING:")
|
||||
print("title:", item[1])
|
||||
print("link :", item[2])
|
||||
print("id :", item[3])
|
||||
sql = (
|
||||
"INSERT "
|
||||
"INTO archive "
|
||||
"SELECT * "
|
||||
# "SELECT title, summary, "
|
||||
# "link, source, timestamp "
|
||||
"FROM entries "
|
||||
"WHERE entries.id = :ix"
|
||||
)
|
||||
cur.execute(sql, (ix,))
|
||||
sql = (
|
||||
"DELETE "
|
||||
"FROM entries "
|
||||
"WHERE id = :ix"
|
||||
)
|
||||
cur.execute(sql, (ix,))
|
||||
if not valid:
|
||||
# print("id: ", item[0])
|
||||
# if title:
|
||||
# print("title: ", title)
|
||||
# print("item[1]: ", item[1])
|
||||
# if link:
|
||||
# print("link: ", link)
|
||||
# print("item[2]: ", item[2])
|
||||
# if entry.id:
|
||||
# print("last_entry:", entry.id)
|
||||
# print("item[3]: ", item[3])
|
||||
# if time:
|
||||
# print("time: ", time)
|
||||
# print("item[4]: ", item[4])
|
||||
# print("read: ", item[5])
|
||||
# breakpoint()
|
||||
async with DBLOCK:
|
||||
# TODO Send to table archive
|
||||
# TODO Also make a regular/routine check for sources that
|
||||
# have been changed (though that can only happen when
|
||||
# manually editing)
|
||||
ix = item[0]
|
||||
print(">>> SOURCE: ", source)
|
||||
print(">>> INVALID:", item[1])
|
||||
# print("title:", item[1])
|
||||
# print("link :", item[2])
|
||||
# print("id :", item[3])
|
||||
if item[5] == 1:
|
||||
print(">>> DELETING:", item[1])
|
||||
sql = (
|
||||
"DELETE "
|
||||
"FROM entries "
|
||||
"WHERE id = :ix"
|
||||
)
|
||||
cur.execute(sql, (ix,))
|
||||
else:
|
||||
print(">>> ARCHIVING:", item[1])
|
||||
sql = (
|
||||
"INSERT "
|
||||
"INTO archive "
|
||||
"SELECT * "
|
||||
"FROM entries "
|
||||
"WHERE entries.id = :ix"
|
||||
)
|
||||
cur.execute(sql, (ix,))
|
||||
sql = (
|
||||
"DELETE "
|
||||
"FROM entries "
|
||||
"WHERE id = :ix"
|
||||
)
|
||||
cur.execute(sql, (ix,))
|
||||
|
||||
|
||||
async def get_feeds(db_file):
|
||||
|
@ -1015,13 +1102,18 @@ async def last_entries(db_file, num):
|
|||
elif num < 1:
|
||||
num = 1
|
||||
cur = get_cursor(db_file)
|
||||
# sql = "SELECT title, link FROM entries ORDER BY ROWID DESC LIMIT :num"
|
||||
# sql = (
|
||||
# "SELECT title, link "
|
||||
# "FROM entries "
|
||||
# "ORDER BY ROWID DESC "
|
||||
# "LIMIT :num"
|
||||
# )
|
||||
sql = (
|
||||
"SELECT title, link "
|
||||
"FROM entries "
|
||||
"WHERE read = 0 "
|
||||
"ORDER BY timestamp "
|
||||
"DESC LIMIT :num "
|
||||
"ORDER BY timestamp DESC "
|
||||
"LIMIT :num "
|
||||
)
|
||||
results = cur.execute(sql, (num,))
|
||||
titles_list = "Recent {} titles:\n".format(num)
|
||||
|
@ -1053,7 +1145,7 @@ async def search_feeds(db_file, query):
|
|||
"""
|
||||
cur = get_cursor(db_file)
|
||||
sql = (
|
||||
"SELECT name, id, address "
|
||||
"SELECT name, address, id "
|
||||
"FROM feeds "
|
||||
"WHERE name LIKE ? "
|
||||
"LIMIT 50"
|
||||
|
@ -1066,7 +1158,10 @@ async def search_feeds(db_file, query):
|
|||
for result in results:
|
||||
counter += 1
|
||||
results_list += (
|
||||
"\n{} [{}]\n{}\n"
|
||||
"\nName: {}"
|
||||
"\n URL: {}"
|
||||
"\n ID: {}"
|
||||
"\n"
|
||||
).format(
|
||||
str(result[0]),
|
||||
str(result[1]),
|
||||
|
@ -1099,9 +1194,16 @@ async def search_entries(db_file, query):
|
|||
"SELECT title, link "
|
||||
"FROM entries "
|
||||
"WHERE title LIKE ? "
|
||||
"UNION ALL "
|
||||
"SELECT title, link "
|
||||
"FROM archive "
|
||||
"WHERE title LIKE ? "
|
||||
"LIMIT 50"
|
||||
)
|
||||
results = cur.execute(sql, [f'%{query}%'])
|
||||
results = cur.execute(sql, (
|
||||
f'%{query}%',
|
||||
f'%{query}%'
|
||||
))
|
||||
results_list = (
|
||||
"Search results for '{}':\n```"
|
||||
).format(query)
|
||||
|
@ -1168,11 +1270,15 @@ async def check_entry_exist(db_file, source, eid=None,
|
|||
"link = :link and "
|
||||
"timestamp = :date"
|
||||
)
|
||||
result = cur.execute(sql, {
|
||||
"title": title,
|
||||
"link": link,
|
||||
"timestamp": date
|
||||
}).fetchone()
|
||||
try:
|
||||
result = cur.execute(sql, {
|
||||
"title": title,
|
||||
"link": link,
|
||||
"timestamp": date
|
||||
}).fetchone()
|
||||
except:
|
||||
print("this is source:", source)
|
||||
print("this is date: ", date)
|
||||
else:
|
||||
sql = (
|
||||
"SELECT id "
|
||||
|
@ -1183,10 +1289,13 @@ async def check_entry_exist(db_file, source, eid=None,
|
|||
"title": title,
|
||||
"link": link
|
||||
}).fetchone()
|
||||
if result:
|
||||
return True
|
||||
else:
|
||||
None
|
||||
try:
|
||||
if result:
|
||||
return True
|
||||
else:
|
||||
None
|
||||
except:
|
||||
print("no result. this is source:", source)
|
||||
|
||||
|
||||
async def set_settings_value(db_file, key_value):
|
||||
|
|
|
@ -19,9 +19,9 @@ import slixmpp
|
|||
|
||||
from slixmpp.plugins.xep_0363.http_upload import FileTooBig, HTTPError, UploadServiceNotFound
|
||||
|
||||
import confighandler
|
||||
import datahandler
|
||||
import datetimehandler
|
||||
import filehandler
|
||||
import filterhandler
|
||||
import sqlitehandler
|
||||
|
||||
|
@ -115,7 +115,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
action = 0
|
||||
jid = msg["from"].bare
|
||||
|
||||
db_dir = confighandler.get_default_dbdir()
|
||||
db_dir = filehandler.get_default_dbdir()
|
||||
os.chdir(db_dir)
|
||||
if jid + ".db" not in os.listdir():
|
||||
await self.task_jid(jid)
|
||||
|
@ -140,7 +140,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
url = message.split(" ")[0]
|
||||
title = " ".join(message.split(" ")[1:])
|
||||
if url.startswith("http"):
|
||||
action = await datahandler.initdb(
|
||||
action = await filehandler.initdb(
|
||||
jid,
|
||||
datahandler.add_feed_no_check,
|
||||
[url, title]
|
||||
|
@ -152,7 +152,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
key = "filter-" + message[:5]
|
||||
val = message[6:]
|
||||
if val:
|
||||
keywords = await datahandler.initdb(
|
||||
keywords = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.get_settings_value,
|
||||
key
|
||||
|
@ -161,7 +161,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
val,
|
||||
keywords
|
||||
)
|
||||
await datahandler.initdb(
|
||||
await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.set_settings_value,
|
||||
[key, val]
|
||||
|
@ -176,7 +176,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
key = "filter-" + message[:4]
|
||||
val = message[5:]
|
||||
if val:
|
||||
keywords = await datahandler.initdb(
|
||||
keywords = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.get_settings_value,
|
||||
key
|
||||
|
@ -185,7 +185,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
val,
|
||||
keywords
|
||||
)
|
||||
await datahandler.initdb(
|
||||
await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.set_settings_value,
|
||||
[key, val]
|
||||
|
@ -198,7 +198,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
action = "Missing keywords."
|
||||
case _ if message_lowercase.startswith("http"):
|
||||
url = message
|
||||
action = await datahandler.initdb(
|
||||
action = await filehandler.initdb(
|
||||
jid,
|
||||
datahandler.add_feed,
|
||||
url
|
||||
|
@ -209,7 +209,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
query = message[6:]
|
||||
if query:
|
||||
if len(query) > 3:
|
||||
action = await datahandler.initdb(
|
||||
action = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.search_feeds,
|
||||
query
|
||||
|
@ -219,7 +219,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
"Enter at least 4 characters to search"
|
||||
)
|
||||
else:
|
||||
action = await datahandler.initdb(
|
||||
action = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.list_feeds
|
||||
)
|
||||
|
@ -235,7 +235,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
# action = (
|
||||
# "Updates will be sent every {} minutes."
|
||||
# ).format(action)
|
||||
await datahandler.initdb(
|
||||
await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.set_settings_value,
|
||||
[key, val]
|
||||
|
@ -263,7 +263,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
# action = (
|
||||
# "Every update will contain {} news items."
|
||||
# ).format(action)
|
||||
await datahandler.initdb(
|
||||
await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.set_settings_value,
|
||||
[key, val]
|
||||
|
@ -278,7 +278,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
case _ if message_lowercase.startswith("recent"):
|
||||
num = message[7:]
|
||||
if num:
|
||||
action = await datahandler.initdb(
|
||||
action = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.last_entries,
|
||||
num
|
||||
|
@ -288,7 +288,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
case _ if message_lowercase.startswith("remove"):
|
||||
ix = message[7:]
|
||||
if ix:
|
||||
action = await datahandler.initdb(
|
||||
action = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.remove_feed,
|
||||
ix
|
||||
|
@ -300,7 +300,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
query = message[7:]
|
||||
if query:
|
||||
if len(query) > 1:
|
||||
action = await datahandler.initdb(
|
||||
action = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.search_entries,
|
||||
query
|
||||
|
@ -315,7 +315,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
# action = "Updates are enabled."
|
||||
key = "enabled"
|
||||
val = 1
|
||||
await datahandler.initdb(
|
||||
await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.set_settings_value,
|
||||
[key, val]
|
||||
|
@ -325,13 +325,13 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
# print(await datetimehandler.current_time(), "task_manager[jid]")
|
||||
# print(task_manager[jid])
|
||||
case "stats":
|
||||
action = await datahandler.initdb(
|
||||
action = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.statistics
|
||||
)
|
||||
case _ if message_lowercase.startswith("status "):
|
||||
ix = message[7:]
|
||||
action = await datahandler.initdb(
|
||||
action = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.toggle_status,
|
||||
ix
|
||||
|
@ -349,7 +349,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
# task_manager[jid]["interval"].cancel()
|
||||
# key = "enabled"
|
||||
# val = 0
|
||||
# action = await datahandler.initdb(
|
||||
# action = await filehandler.initdb(
|
||||
# jid,
|
||||
# sqlitehandler.set_settings_value,
|
||||
# [key, val]
|
||||
|
@ -360,7 +360,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
# # await self.send_status(jid)
|
||||
key = "enabled"
|
||||
val = 0
|
||||
await datahandler.initdb(
|
||||
await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.set_settings_value,
|
||||
[key, val]
|
||||
|
@ -388,7 +388,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
Self.
|
||||
"""
|
||||
while True:
|
||||
db_dir = confighandler.get_default_dbdir()
|
||||
db_dir = filehandler.get_default_dbdir()
|
||||
if not os.path.isdir(db_dir):
|
||||
msg = (
|
||||
"Slixfeed can not work without a database.\n"
|
||||
|
@ -410,7 +410,8 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
# await jid_tasker[jid]
|
||||
async with asyncio.TaskGroup() as tg:
|
||||
for file in files:
|
||||
if file.endswith(".db") and not file.endswith(".db-jour.db"):
|
||||
if (file.endswith(".db") and
|
||||
not file.endswith(".db-jour.db")):
|
||||
jid = file[:-3]
|
||||
main_task.extend([tg.create_task(self.task_jid(jid))])
|
||||
# main_task = [tg.create_task(self.task_jid(jid))]
|
||||
|
@ -428,7 +429,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
jid : str
|
||||
Jabber ID.
|
||||
"""
|
||||
enabled = await datahandler.initdb(
|
||||
enabled = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.get_settings_value,
|
||||
"enabled"
|
||||
|
@ -476,7 +477,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
"""
|
||||
# print("Starting send_update()")
|
||||
# print(jid)
|
||||
new = await datahandler.initdb(
|
||||
new = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.get_entry_unread,
|
||||
num
|
||||
|
@ -493,7 +494,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
self.send_update,
|
||||
"interval"
|
||||
)
|
||||
# interval = await datahandler.initdb(
|
||||
# interval = await filehandler.initdb(
|
||||
# jid,
|
||||
# sqlitehandler.get_settings_value,
|
||||
# "interval"
|
||||
|
@ -530,7 +531,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
Jabber ID.
|
||||
"""
|
||||
print(await datetimehandler.current_time(), "> SEND STATUS",jid)
|
||||
enabled = await datahandler.initdb(
|
||||
enabled = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.get_settings_value,
|
||||
"enabled"
|
||||
|
@ -539,7 +540,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
status_mode = "xa"
|
||||
status_text = "Send \"Start\" to receive news."
|
||||
else:
|
||||
feeds = await datahandler.initdb(
|
||||
feeds = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.get_number_of_items,
|
||||
"feeds"
|
||||
|
@ -550,7 +551,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
"📂️ Send a URL from a blog or a news website."
|
||||
)
|
||||
else:
|
||||
unread = await datahandler.initdb(
|
||||
unread = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.get_number_of_entries_unread
|
||||
)
|
||||
|
@ -606,7 +607,7 @@ class Slixfeed(slixmpp.ClientXMPP):
|
|||
Value. The default is None.
|
||||
"""
|
||||
if not val:
|
||||
val = await datahandler.initdb(
|
||||
val = await filehandler.initdb(
|
||||
jid,
|
||||
sqlitehandler.get_settings_value,
|
||||
key
|
||||
|
@ -644,7 +645,7 @@ async def check_updates(jid):
|
|||
"""
|
||||
while True:
|
||||
print(await datetimehandler.current_time(), "> CHCK UPDATE",jid)
|
||||
await datahandler.initdb(jid, datahandler.download_updates)
|
||||
await filehandler.initdb(jid, datahandler.download_updates)
|
||||
await asyncio.sleep(60 * 90)
|
||||
# Schedule to call this function again in 90 minutes
|
||||
# loop.call_at(
|
||||
|
|
Loading…
Reference in a new issue