add watch scripts for PyPI projects
This commit is contained in:
parent
5859eecca0
commit
c92af8eedf
20
python3-agate-dbf/.watch
Executable file
20
python3-agate-dbf/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/agate-dbf/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-agate-excel/.watch
Executable file
20
python3-agate-excel/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/agate-excel/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-agate-sql/.watch
Executable file
20
python3-agate-sql/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/agate-sql/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-agate/.watch
Executable file
20
python3-agate/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/agate/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-csvkit/.watch
Executable file
20
python3-csvkit/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/csvkit/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-dbfread/.watch
Executable file
20
python3-dbfread/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/dbfread/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-et_xmlfile/.watch
Executable file
20
python3-et_xmlfile/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/et_xmlfile/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-leather/.watch
Executable file
20
python3-leather/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/leather/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-olefile/.watch
Executable file
20
python3-olefile/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/olefile/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-openpyxl/.watch
Executable file
20
python3-openpyxl/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/openpyxl/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-pyicu/.watch
Executable file
20
python3-pyicu/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/pyicu/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-pytimeparse/.watch
Executable file
20
python3-pytimeparse/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/pytimeparse/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-slugify/.watch
Executable file
20
python3-slugify/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/slugify/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-sqlalchemy/.watch
Executable file
20
python3-sqlalchemy/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/sqlalchemy/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
20
python3-xlrd/.watch
Executable file
20
python3-xlrd/.watch
Executable file
@ -0,0 +1,20 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from subprocess import Popen,PIPE
|
||||
from bs4 import BeautifulSoup as bs
|
||||
|
||||
def get_html(url):
|
||||
try:
|
||||
curlpipe = Popen(["/usr/bin/curl","-s","-L",url],stdout=PIPE)
|
||||
except:
|
||||
print("- Request error: unable to retrieve {}".format(url))
|
||||
return None
|
||||
curltup = curlpipe.communicate()
|
||||
return curltup[0]
|
||||
|
||||
page = get_html("https://pypi.org/project/xlrd/")
|
||||
soup = bs(page, 'lxml')
|
||||
headers = soup.body.find_all('h1')
|
||||
for h in headers[:1]:
|
||||
name_version = h.get_text().split()
|
||||
print(name_version[-1])
|
Loading…
Reference in New Issue
Block a user