Scrapes the Sun Life of Canada Philippines web site for historical prices of their investment funds and then saves them as CSV files.

Related tags

Web Crawlingpython
Overview

slocpi-scraper

Sun Life of Canada Philippines Inc Investment Funds Scraper

Install dependencies

pip install -r requirements.txt

Usage

General format:

./api_client.py 
    
    
      (end_date)

    
   

The fund codes are listed in the next section.

The dates should use the YYYY-MM-DD format.

End dates are optional and default to today if not explicitly specified.

To get historical data for their Peso Balanced Fund from Jan 1, 2021 to Feb 14, 2021:

./api_client.py SLPBA 2021-01-01 2021-02-14

To get historical data for all of their VULs from Jan 1, 2021 to Feb 14, 2021:

./api_client.py VUL 2021-01-01 2021-02-14

To get historical data for all of their mutual funds from Jan 1, 2021 to Feb 14, 2021:

./api_client.py MF 2021-01-01 2021-02-14

To get historical data for all of their funds from Jan 1, 2021 to Feb 14, 2021:

./api_client.py ALL 2021-01-01 2021-02-14

List of VULs and their codes

  • SLPBA: Sun Life Phils - Peso Balanced Fund
  • SLPBF: Sun Life Phils - Peso Bond Fund
  • SLPCP: Sun Life Phils - Captains Fund
  • SLPDF: Dynamic Fund
  • SLPEF: Sun Life Phils - Peso Equity Fund
  • SLPGF: Sun Life Phils - Peso Growth Fund
  • SLPGP: Sun Life Phils - Growth Plus Fund
  • SLPIF: Sun Life Phils - Peso Income Fund
  • SLPIN: Sun Life Phils - Index Fund
  • SLPMM: Sun Life Phils - Money Market Fund
  • SLPOF: Sun Life Phils - Peso Opportunity Fund
  • SLPOT: Sun Life Phils - Opportunity Tracker Fund
  • SLPP1: Sun Peso Maximizer - Fund
  • SLPP2: Sun Peso Maximizer - Primo 2 Fund
  • TDF20: Sun Life Phils - Peso MyFuture 2020
  • TDF25: Sun Life Phils - Peso MyFuture 2025
  • TDF30: Sun Life Phils - Peso MyFuture 2030
  • TDF35: Sun Life Phils - Peso MyFuture 2035
  • TDF40: Sun Life Phils - Peso MyFuture 2040
  • SLUBF: Sun Life Phils - Dollar Bond Fund
  • SLUD7: Sun Life Phils - Sun Dollar Maximizer - WT
  • SLUD8: Sun Life Phils - Sun Dollar Maximizer - PriMO
  • SLUD9: Sun Life Phils - Sun Dollar Maximizer - PriMO 2
  • SLUGF: Sun Life Phils - Global Growth Fund
  • SLUIF: Sun Life Phils - Global Income Fund
  • SLUOF: Sun Life Phils - Global Opportunity Fund
  • SLUMM: Sun Life Phils - Dollar Money Market Fund

List of mutual funds and their codes

  • CF0001: Sun Life Prosperity Bond Fund
  • CF0002: Sun Life Prosperity Balanced Fund
  • CF0003: Sun Life Prosperity Philippine Equity Fund
  • CF0004: Sun Life Prosperity Dollar Advantage Fund
  • CF0005: Sun Life Prosperity Money Market Fund
  • CF0006: Sun Life Prosperity Dollar Abundance Fund
  • CF0007: Sun Life Prosperity Government Securities (GS) Fund
  • CF0008: Sun Life Prosperity Dynamic Fund
  • CF0009: Sun Life Prosperity Philippine Stock Index Fund
  • CF0010: Sun Life Prosperity Dollar Wellspring Fund
  • CF0011: Sun Life Prosperity World Voyager Fund
  • CF0012: Sun Life Prosperity Dollar Starter Fund
  • CF0013: Sun Life Prosperity Achiever Fund 2028
  • CF0014: Sun Life Prosperity Achiever Fund 2038
  • CF0015: Sun Life Prosperity Achiever Fund 2048
  • CF0016: Sun Life Prosperity World Equity Index Feeder Fund
Owner
Daryl Yu
Daryl Yu
An arxiv spider

An Arxiv Spider 做为一个cser,杰出男孩深知内核对连接到计算机上的硬件设备进行管理的高效方式是中断而不是轮询。每当小伙伴发来一篇刚挂在arxiv上的”热乎“好文章时,杰出男孩都会感叹道:”师兄这是每天都挂在arxiv上呀,跑的好快~“。于是杰出男孩找了找 github,借鉴了一下其

Jie Liu 11 Sep 09, 2022
simple http & https proxy scraper and checker

simple http & https proxy scraper and checker

Neospace 11 Nov 15, 2021
A distributed crawler for weibo, building with celery and requests.

A distributed crawler for weibo, building with celery and requests.

SpiderClub 4.8k Jan 03, 2023
河南工业大学 完美校园 自动校外打卡

HAUT-checkin 河南工业大学自动校外打卡 由于github actions存在明显延迟,建议直接使用腾讯云函数 特点 多人打卡 使用简单,仅需账号密码以及用于微信推送的uid 自动获取上一次打卡信息用于打卡 向所有成员微信单独推送打卡状态 完美校园服务器繁忙时造成打卡失败会自动重新打卡

36 Oct 27, 2022
A module for CME that spiders hashes across the domain with a given hash.

hash_spider A module for CME that spiders hashes across the domain with a given hash. Installation Simply copy hash_spider.py to your CME module folde

37 Sep 08, 2022
This is a web scraper, using Python framework Scrapy, built to extract data from the Deals of the Day section on Mercado Livre website.

Deals of the Day This is a web scraper, using the Python framework Scrapy, built to extract data such as price and product name from the Deals of the

David Souza 1 Jan 12, 2022
Binance Smart Chain Contract Scraper + Contract Evaluator

Pulls Binance Smart Chain feed of newly-verified contracts every 30 seconds, then checks their contract code for links to socials.Returns only those with socials information included, and then submit

14 Dec 09, 2022
Dude is a very simple framework for writing web scrapers using Python decorators

Dude is a very simple framework for writing web scrapers using Python decorators. The design, inspired by Flask, was to easily build a web scraper in just a few lines of code. Dude has an easy-to-lea

Ronie Martinez 326 Dec 15, 2022
Scrapy uses Request and Response objects for crawling web sites.

Requests and Responses¶ Scrapy uses Request and Response objects for crawling web sites. Typically, Request objects are generated in the spiders and p

Md Rashidul Islam 1 Nov 03, 2021
PaperRobot: a paper crawler that can quickly download numerous papers, facilitating paper studying and management

PaperRobot PaperRobot 是一个论文抓取工具,可以快速批量下载大量论文,方便后期进行持续的论文管理与学习。 PaperRobot通过多个接口抓取论文,目前抓取成功率维持在90%以上。通过配置Config文件,可以抓取任意计算机领域相关会议的论文。 Installation Down

moxiaoxi 47 Nov 23, 2022
A dead simple crawler to get books information from Douban.

Introduction A dead simple crawler to get books information from Douban. Pre-requesites Python 3 Install dependencies from requirements.txt (Optional)

Yun Wang 1 Jan 10, 2022
Visual scraping for Scrapy

Portia Portia is a tool that allows you to visually scrape websites without any programming knowledge required. With Portia you can annotate a web pag

Scrapinghub 8.7k Jan 05, 2023
Libextract: extract data from websites

Libextract is a statistics-enabled data extraction library that works on HTML and XML documents and written in Python

499 Dec 09, 2022
Introduction to WebScraping Workshop - Semcomp 24 Beta

Extrair informações da internet de forma automatizada. Existem diversas maneiras de fazer isso, nesse tutorial vamos ver algumas delas, por meio de bibliotecas de python.

Luísa Moura 19 Sep 11, 2022
FilmMikirAPI - A simple rest-api which is used for scrapping on the Kincir website using the Python and Flask package

FilmMikirAPI - A simple rest-api which is used for scrapping on the Kincir website using the Python and Flask package

UserGhost411 1 Nov 17, 2022
Pelican plugin that adds site search capability

Search: A Plugin for Pelican This plugin generates an index for searching content on a Pelican-powered site. Why would you want this? Static sites are

22 Nov 21, 2022
自动完成每日体温上报(Github Actions)

体温上报助手 简介 每天 10:30 GMT+8 自动完成体温上报,如想修改定时运行的时间,可修改 .github/workflows/SduHealthReport.yml 中 schedule 属性。 如果当日有异常,请手动在小程序端/PC 端填写!

Teng Zhang 23 Sep 15, 2022
Telegram group scraper tool

Telegram Group Scrapper

Wahyusaputra 2 Jan 11, 2022
a way to scrape a database of all of the isef projects

ISEF Database This is a simple web scraper which gets all of the projects and abstract information from here. My goal for this is for someone to get i

William Kaiser 1 Mar 18, 2022
Subscrape - A Python scraper for substrate chains

subscrape A Python scraper for substrate chains that uses Subscan. Usage copy co

ChaosDAO 14 Dec 15, 2022