Store domain in a JSON file (#46)

* Refactor domain function to store domain in a JSON file

* Revert "update version"

This reverts commit f866d9b42c38ff87eb9a0546c53f454ba42689c9.
This commit is contained in:
Fede14it 2024-02-28 07:27:29 +01:00 committed by GitHub
parent f866d9b42c
commit fec9df9695
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
4 changed files with 28 additions and 18 deletions

View File

@ -9,30 +9,34 @@ import requests, sys, json
from bs4 import BeautifulSoup
def domain_version():
console.print("[green]Get rules ...")
req_repo = requests.get("https://raw.githubusercontent.com/Ghost6446/Streaming_comunity_data/main/data.json", headers={'user-agent': get_headers()}, timeout=5)
domain = req_repo.json()['domain']
req_repo = None
try:
with open('data.json', 'r') as file:
req_repo = json.load(file)
except FileNotFoundError:
req_repo = {"domain": ""}
domain = req_repo['domain']
if req_repo.ok:
while True:
if not domain:
domain = input("Insert domain (streamingcommunity.?): ")
req_repo['domain'] = domain
with open('data.json', 'w') as file:
json.dump(req_repo, file)
console.print(f"[blue]Test domain [white]=> [red]{domain}")
site_url = f"https://streamingcommunity.{domain}"
site_request = requests.get(site_url, headers={'user-agent': get_headers()})
try:
site_request = requests.get(site_url, headers={'user-agent': get_headers()})
soup = BeautifulSoup(site_request.text, "lxml")
version = json.loads(soup.find("div", {"id": "app"}).get("data-page"))['version']
console.print(f"[blue]Rules [white]=> [red].{domain}")
return domain, version
except:
console.log("[red]Cant get version, problem with domain")
sys.exit(0)
else:
console.log(f"[red]Error: {req_repo.status_code}")
sys.exit(0)
except Exception as e:
console.log("[red]Cant get version, problem with domain. Try again.")
domain = None
continue
def search(title_search, domain):

View File

@ -193,7 +193,12 @@ class M3U8_Segments:
"""Single req to a ts file to get content"""
url_number = self.segments.index(ts_url)
is_valid = ( str(url_number) in failed_segments )
is_valid = True
for failde_seg in failed_segments:
if str(failde_seg) in ts_url:
is_valid = False
break
if is_valid:
@ -307,8 +312,8 @@ class M3U8_Segments:
with open(file_list_path, 'w') as f:
for ts_file in ts_files:
absolute_path = os.path.abspath(os.path.join(self.temp_folder, ts_file), current_dir)
f.write(f"file '{absolute_path}'\n")
relative_path = os.path.relpath(os.path.join(self.temp_folder, ts_file), current_dir)
f.write(f"file '{relative_path}'\n")
console.log("[cyan]Start join all file")
try:

View File

@ -1,5 +1,5 @@
__title__ = 'Streaming_community'
__version__ = 'v0.8.6'
__version__ = 'v0.8.5'
__author__ = 'Ghost6446'
__description__ = 'A command-line program to download film'
__license__ = 'MIT License'

1
data.json Normal file
View File

@ -0,0 +1 @@
{"domain": "express"}