Add Hacker News to syndication
This commit is contained in:
@@ -3,11 +3,14 @@ from pathlib import Path
|
||||
import os
|
||||
import json
|
||||
import hashlib
|
||||
import requests
|
||||
|
||||
domain = "https://fundor333.com"
|
||||
rss_url_mastodon = "https://mastodon.social/@fundor333.rss"
|
||||
rss_url_bsky = "https://bsky.app/profile/did:plc:u7piwonv4s27ysugjaa6im2q/rss"
|
||||
|
||||
hacker_news_username = "fundor333"
|
||||
|
||||
|
||||
def clean_slug(slug: str):
|
||||
return hashlib.md5(
|
||||
@@ -15,6 +18,45 @@ def clean_slug(slug: str):
|
||||
).hexdigest()
|
||||
|
||||
|
||||
class HackerNewsFinder:
|
||||
def __init__(self, hacker_news_username):
|
||||
self.hacker_news_username = hacker_news_username
|
||||
self.path_folder = os.path.join("data", "syndication", "hacker_news")
|
||||
Path(self.path_folder).mkdir(parents=True, exist_ok=True)
|
||||
self.path_file = os.path.join(self.path_folder, self.hacker_news_username)
|
||||
if os.path.exists(self.path_file + ".json") is False:
|
||||
with open(self.path_file + ".json", "w") as fp:
|
||||
json.dump({"readed": []}, fp)
|
||||
|
||||
def get_articles_id(self):
|
||||
url = f"https://hacker-news.firebaseio.com/v0/user/{self.hacker_news_username}.json"
|
||||
data = requests.get(url).json()
|
||||
return data.get("submitted", [])
|
||||
|
||||
def get_article(self, article_id):
|
||||
url = f"https://hacker-news.firebaseio.com/v0/item/{article_id}.json"
|
||||
data = requests.get(url).json()
|
||||
if data.get("type", False) == "story":
|
||||
return (f"https://news.ycombinator.com/item?id={article_id}", data["url"])
|
||||
return False
|
||||
|
||||
def run(self, output: dict):
|
||||
with open(self.path_file + ".json") as fp:
|
||||
data = json.load(fp)
|
||||
for link in self.get_articles_id():
|
||||
if link not in data["readed"]:
|
||||
info = self.get_article(link)
|
||||
if info:
|
||||
hashed = clean_slug(info[1])
|
||||
if output.get(hashed, False):
|
||||
output[hashed].append(info[0])
|
||||
else:
|
||||
output[hashed] = [info[0]]
|
||||
data["readed"].append(link)
|
||||
with open(self.path_file + ".json", "w") as fp:
|
||||
json.dump(data, fp)
|
||||
|
||||
|
||||
class BskyFinder:
|
||||
def find_urls(self, string):
|
||||
x = string.split(" ")
|
||||
@@ -66,11 +108,18 @@ class MastodonFinder:
|
||||
|
||||
|
||||
class WriterSyndication:
|
||||
def __init__(self, rss_url_mastodon: str, rss_url_bsky: str, domain: str):
|
||||
def __init__(
|
||||
self,
|
||||
rss_url_mastodon: str,
|
||||
rss_url_bsky: str,
|
||||
domain: str,
|
||||
hacker_news_username: str,
|
||||
):
|
||||
self.output = {}
|
||||
self.rss_url_mastodon = rss_url_mastodon
|
||||
self.rss_url_bsky = rss_url_bsky
|
||||
self.domain = domain
|
||||
self.hacker_news_username = hacker_news_username
|
||||
|
||||
def data_gathering(self):
|
||||
m = MastodonFinder()
|
||||
@@ -79,6 +128,9 @@ class WriterSyndication:
|
||||
bs = BskyFinder()
|
||||
bs.run(self.rss_url_bsky, self.domain, self.output)
|
||||
|
||||
hn = HackerNewsFinder(self.hacker_news_username)
|
||||
hn.run(self.output)
|
||||
|
||||
def write(self):
|
||||
for key in self.output.keys():
|
||||
|
||||
@@ -109,5 +161,5 @@ class WriterSyndication:
|
||||
self.write()
|
||||
|
||||
|
||||
w = WriterSyndication(rss_url_mastodon, rss_url_bsky, domain)
|
||||
w = WriterSyndication(rss_url_mastodon, rss_url_bsky, domain, hacker_news_username)
|
||||
w.run()
|
||||
|
||||
Reference in New Issue
Block a user