Scrapes the Sun Life of Canada Philippines web site for historical prices of their investment funds and then saves them as CSV files.

Related tags

Web Crawlingpython
Overview

slocpi-scraper

Sun Life of Canada Philippines Inc Investment Funds Scraper

Install dependencies

pip install -r requirements.txt

Usage

General format:

./api_client.py 
    
    
      (end_date)

    
   

The fund codes are listed in the next section.

The dates should use the YYYY-MM-DD format.

End dates are optional and default to today if not explicitly specified.

To get historical data for their Peso Balanced Fund from Jan 1, 2021 to Feb 14, 2021:

./api_client.py SLPBA 2021-01-01 2021-02-14

To get historical data for all of their VULs from Jan 1, 2021 to Feb 14, 2021:

./api_client.py VUL 2021-01-01 2021-02-14

To get historical data for all of their mutual funds from Jan 1, 2021 to Feb 14, 2021:

./api_client.py MF 2021-01-01 2021-02-14

To get historical data for all of their funds from Jan 1, 2021 to Feb 14, 2021:

./api_client.py ALL 2021-01-01 2021-02-14

List of VULs and their codes

  • SLPBA: Sun Life Phils - Peso Balanced Fund
  • SLPBF: Sun Life Phils - Peso Bond Fund
  • SLPCP: Sun Life Phils - Captains Fund
  • SLPDF: Dynamic Fund
  • SLPEF: Sun Life Phils - Peso Equity Fund
  • SLPGF: Sun Life Phils - Peso Growth Fund
  • SLPGP: Sun Life Phils - Growth Plus Fund
  • SLPIF: Sun Life Phils - Peso Income Fund
  • SLPIN: Sun Life Phils - Index Fund
  • SLPMM: Sun Life Phils - Money Market Fund
  • SLPOF: Sun Life Phils - Peso Opportunity Fund
  • SLPOT: Sun Life Phils - Opportunity Tracker Fund
  • SLPP1: Sun Peso Maximizer - Fund
  • SLPP2: Sun Peso Maximizer - Primo 2 Fund
  • TDF20: Sun Life Phils - Peso MyFuture 2020
  • TDF25: Sun Life Phils - Peso MyFuture 2025
  • TDF30: Sun Life Phils - Peso MyFuture 2030
  • TDF35: Sun Life Phils - Peso MyFuture 2035
  • TDF40: Sun Life Phils - Peso MyFuture 2040
  • SLUBF: Sun Life Phils - Dollar Bond Fund
  • SLUD7: Sun Life Phils - Sun Dollar Maximizer - WT
  • SLUD8: Sun Life Phils - Sun Dollar Maximizer - PriMO
  • SLUD9: Sun Life Phils - Sun Dollar Maximizer - PriMO 2
  • SLUGF: Sun Life Phils - Global Growth Fund
  • SLUIF: Sun Life Phils - Global Income Fund
  • SLUOF: Sun Life Phils - Global Opportunity Fund
  • SLUMM: Sun Life Phils - Dollar Money Market Fund

List of mutual funds and their codes

  • CF0001: Sun Life Prosperity Bond Fund
  • CF0002: Sun Life Prosperity Balanced Fund
  • CF0003: Sun Life Prosperity Philippine Equity Fund
  • CF0004: Sun Life Prosperity Dollar Advantage Fund
  • CF0005: Sun Life Prosperity Money Market Fund
  • CF0006: Sun Life Prosperity Dollar Abundance Fund
  • CF0007: Sun Life Prosperity Government Securities (GS) Fund
  • CF0008: Sun Life Prosperity Dynamic Fund
  • CF0009: Sun Life Prosperity Philippine Stock Index Fund
  • CF0010: Sun Life Prosperity Dollar Wellspring Fund
  • CF0011: Sun Life Prosperity World Voyager Fund
  • CF0012: Sun Life Prosperity Dollar Starter Fund
  • CF0013: Sun Life Prosperity Achiever Fund 2028
  • CF0014: Sun Life Prosperity Achiever Fund 2038
  • CF0015: Sun Life Prosperity Achiever Fund 2048
  • CF0016: Sun Life Prosperity World Equity Index Feeder Fund
Owner
Daryl Yu
Daryl Yu
API to parse tibia.com content into python objects.

Tibia.py An API to parse Tibia.com content into object oriented data. No fetching is done by this module, you must provide the html content. Features:

Allan Galarza 25 Oct 31, 2022
A python module to parse the Open Graph Protocol

OpenGraph is a module of python for parsing the Open Graph Protocol, you can read more about the specification at http://ogp.me/ Installation $ pip in

Erik Rivera 213 Nov 12, 2022
A package designed to scrape data from Yahoo Finance.

yahoostock A package designed to scrape data from Yahoo Finance. Installation The most simple installation method is through PIP. pip install yahoosto

Rohan Singh 2 May 28, 2022
Deep Web Miner Python | Spyder Crawler

Webcrawler written in Python. This crawler does dig in till the 3 level of inside addressed and mine the respective data accordingly

Karan Arora 17 Jan 24, 2022
Visual scraping for Scrapy

Portia Portia is a tool that allows you to visually scrape websites without any programming knowledge required. With Portia you can annotate a web pag

Scrapinghub 8.7k Jan 05, 2023
A web crawler for recording posts in "sina weibo"

Web Crawler for "sina weibo" A web crawler for recording posts in "sina weibo" Introduction This script helps collect attributes of posts in "sina wei

4 Aug 20, 2022
Universal Reddit Scraper - A comprehensive Reddit scraping command-line tool written in Python.

Universal Reddit Scraper - A comprehensive Reddit scraping command-line tool written in Python.

Joseph Lai 543 Jan 03, 2023
python+selenium实现的web端自动打卡 + 每日邮件发送 + 金山词霸 每日一句 + 毒鸡汤(从2月份稳定运行至今)

python+selenium实现的web端自动打卡 说明 本打卡脚本适用于郑州大学健康打卡,其他web端打卡也可借鉴学习。(自己用的,从2月分稳定运行至今) 仅供学习交流使用,请勿依赖。开发者对使用本脚本造成的问题不负任何责任,不对脚本执行效果做出任何担保,原则上不提供任何形式的技术支持。 为防止

Sunday 1 Aug 27, 2022
This repo has the source code for the crawler and data crawled from auto-data.net

This repo contains the source code for crawler and crawled data of cars specifications from autodata. The data has roughly 45k cars

Tô Đức Anh 5 Nov 22, 2022
用python爬取江苏几大高校的就业网站,并提供3种方式通知给用户,分别是通过微信发送、命令行直接输出、windows气泡通知。

crawler_for_university 用python爬取江苏几大高校的就业网站,并提供3种方式通知给用户,分别是通过微信发送、命令行直接输出、windows气泡通知。 环境依赖 wxpy,requests,bs4等库 功能描述 该项目基于python,通过爬虫爬各高校的就业信息网,爬取招聘信

8 Aug 16, 2021
Github scraper app is used to scrape data for a specific user profile created using streamlit and BeautifulSoup python packages

Github Scraper Github scraper app is used to scrape data for a specific user profile. Github scraper app gets a github profile name and check whether

Siva Prakash 6 Apr 05, 2022
High available distributed ip proxy pool, powerd by Scrapy and Redis

高可用IP代理池 README | 中文文档 本项目所采集的IP资源都来自互联网,愿景是为大型爬虫项目提供一个高可用低延迟的高匿IP代理池。 项目亮点 代理来源丰富 代理抓取提取精准 代理校验严格合理 监控完备,鲁棒性强 架构灵活,便于扩展 各个组件分布式部署 快速开始 注意,代码请在release

SpiderClub 5.2k Jan 03, 2023
Scraping weather data using Python to receive umbrella reminders

A Python package which scrapes weather data from google and sends umbrella reminders to specified email at specified time daily.

Edula Vinay Kumar Reddy 1 Aug 23, 2022
Simple proxy scraper made by using ProxyScrape's api.

What is Moon? Moon is a lightweight and fast proxy scraper made by using ProxyScrape's api. What can i do with this? You can use proxies for varietys

1 Jul 04, 2022
A Web Scraper built with beautiful soup, that fetches udemy course information. Get udemy course information and convert it to json, csv or xml file

Udemy Scraper A Web Scraper built with beautiful soup, that fetches udemy course information. Installation Virtual Environment Firstly, it is recommen

Aditya Gupta 15 May 17, 2022
Scrapy uses Request and Response objects for crawling web sites.

Requests and Responses¶ Scrapy uses Request and Response objects for crawling web sites. Typically, Request objects are generated in the spiders and p

Md Rashidul Islam 1 Nov 03, 2021
This program will help you to properly scrape all data from a specific website

This program will help you to properly scrape all data from a specific website

MD. MINHAZ 0 May 15, 2022
Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Django and Vue.js

Gerapy Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Scrapyd-Client, Scrapyd-API, Django and Vue.js. Documentation Documentation

Gerapy 2.9k Jan 03, 2023
A crawler of doubamovie

豆瓣电影 A crawler of doubamovie 一个小小的入门级scrapy框架的应用,选取豆瓣电影对排行榜前1000的电影数据进行爬取。 spider.py start_requests方法为scrapy的方法,我们对它进行重写。 def start_requests(self):

Cats without dried fish 1 Oct 05, 2021
This Scrapy project uses Redis and Kafka to create a distributed on demand scraping cluster

This Scrapy project uses Redis and Kafka to create a distributed on demand scraping cluster.

IST Research 1.1k Jan 06, 2023