├── .gitignore ├── README.md ├── compressor.py ├── mover.py └── scrapper.py /.gitignore: -------------------------------------------------------------------------------- 1 | env 2 | .DS_Store 3 | -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- 1 | # Organizr 2 | A couple of organising scripts. Enjoy :) 3 | -------------------------------------------------------------------------------- /compressor.py: -------------------------------------------------------------------------------- 1 | from PIL import Image # python3 -m pip install Pillow 2 | 3 | import os 4 | 5 | downloadsFolder = "/Users/nicolasschurmann/Downloads/" 6 | 7 | if __name__ == "__main__": 8 | for filename in os.listdir(downloadsFolder): 9 | name, extension = os.path.splitext(downloadsFolder + filename) 10 | 11 | if extension in [".jpg", ".jpeg", ".png"]: 12 | picture = Image.open(downloadsFolder + filename) 13 | picture.save(downloadsFolder + "compressed_"+filename, optimize=True, quality=60) 14 | -------------------------------------------------------------------------------- /mover.py: -------------------------------------------------------------------------------- 1 | from PIL import Image 2 | import os 3 | 4 | downloadsFolder = "/Users/nicolasschurmann/Downloads/" 5 | picturesFolder = "/Users/nicolasschurmann/Pictures/" 6 | 7 | if __name__ == "__main__": 8 | for filename in os.listdir(downloadsFolder): 9 | name, extension = os.path.splitext(downloadsFolder + filename) 10 | 11 | if extension in [".jpg", ".jpeg", ".png"]: 12 | picture = Image.open(downloadsFolder + filename) 13 | picture.save(picturesFolder + "compressed_"+filename, optimize=True, quality=60) 14 | os.remove(downloadsFolder + filename) 15 | print(name + ": " + extension) 16 | 17 | if extension in [".mp3"]: 18 | musicFolder = "/Users/nicolasschurmann/Music/" 19 | os.rename(downloadsFolder + filename, musicFolder + filename) 20 | 21 | 22 | 23 | -------------------------------------------------------------------------------- /scrapper.py: -------------------------------------------------------------------------------- 1 | import requests # python3 -m pip install requests beautifulsoup4 2 | from bs4 import BeautifulSoup 3 | 4 | url = "https://www.seek.co.nz/python-jobs?salaryrange=100000-999999&salarytype=annual" 5 | 6 | if "__main__" == __name__: 7 | page = requests.get(url) 8 | soup = BeautifulSoup(page.content, "html.parser") 9 | 10 | def has_data_search(tag): 11 | return tag.has_attr("data-search-sol-meta") 12 | 13 | results = soup.find_all(has_data_search) 14 | 15 | for job in results: 16 | try: 17 | titleElement = job.find("a", attrs={"data-automation": "jobTitle"}) 18 | title = titleElement.get_text() 19 | company = job.find("a", attrs={"data-automation": "jobCompany"}).get_text() 20 | joblink = "https://www.seek.co.nz" + titleElement["href"] 21 | salary = job.find("span", attrs={"data-automation": "jobSalary"}) 22 | salary = salary.get_text() if salary else 'n/a' 23 | 24 | job = "Titulo: {}\nEmpresa: {}\nSalario: {}\nLink: {}a\n" 25 | 26 | job = job.format(title, company, salary, joblink) 27 | 28 | print(job) 29 | except Exception as e: 30 | print("Exception: {}".format(e)) 31 | pass 32 | --------------------------------------------------------------------------------