Pull data from Surfline, MagicSeaweed & Spitcast APIs to display an aggregated surf forecast.
- Install dependencies using Homebrew:
brew bundle
- If on Linux:
pg_ctl -D /home/linuxbrew/.linuxbrew/var/postgres start
mkcert -install
mkdir config/ssl && cd config/ssl && mkcert surf.localhost surf-packs.localhost localhost 127.0.0.1 ::1 && cd ../../
gem install bundler -v '>= 2.0.2'
bundle
yarn
cp config/database.yml.example config/database.yml
bin/rails db:setup
rvmsudo bin/invoker setup --tld localhost
(help available here)- Grab some Spitcast data:
bin/rails spitcast:update
- Grab some Surfline v2 data:
SURFLINE_EMAIL=xxx SURFLINE_PASSWORD=yyy bin/rails surfline_v2:update
- Grab some MagicSeaweed data (requires a valid API key):
MSW_API_KEY=xxx bin/rails msw:update
(replacexxx
with your key) - Refresh the materialized Postgres view that collates all forecast data into one table:
bin/rails database_views:refresh
bin/server
- Open https://surf.localhost
- Score!
Pull requests welcome, especially around new data sources/better data visualization (see TODO for suggestions)
Contributing new spots is easy! Make sure you're signed into your Github account and edit the seeds file:
- Create a new Region/Subregion if necessary. For example, Los Angeles is created like so:
You can get valid timezone names from this list.
CA = Region.find_or_create_by(name: 'California') LA = Subregion.find_or_create_by(name: 'Los Angeles', region: CA) LA.timezone = 'America/Los_Angeles' LA.save!
- Use this tool to draw a bounding box around the area you want to find spots for.
- Choose CSV from the dropdown at the bottom & copy the coordinates string.
- Run
rails console
, then runSpotFinder.new('{string}').formatted_spots
. - Copy the output and paste it into the spots array in
seeds.rb
, then make sure to assign each spot to the right subregion & delete extraneous fields (msw_name
,match_type
,distance
). - If the spot is in California, get the Spitcast spot id & slug (unique text id) their spot list API (you can change the county at the end of the URL). The slug is
spot_id_char
in their API. - It's strongly encouraged to add all spots for a particular county or region rather than just a single one. Be a pal!
- Submit a pull request and I'll get it on the site ASAP!
Use the following as a template. Delete the lines for surfline_v2_id
, msw_id
, etc, if that spot doesn't exist on that particular site.
{
name: 'County Line',
lat: 34.051,
lon: -118.964,
surfline_v2_id: '590927576a2e4300134fbed8',
msw_id: 277,
spitcast_id: 207,
spitcast_slug: 'county-line-malibu-ca',
subregion: LA,
},
Surfline's new API is undocumented but easy to reverse engineer using their new website's code. Thankfully its structure is much more sane than the old API.
https://services.surfline.com/kbyg/spots/forecasts/{type}?{params}
For reference, I believe kbyg
stands for "Know Before You Go," which is their tagline.
Type | Data |
---|---|
wave | array of min/max sizes & optimal scores |
wind | array of wind directions/speeds & optimal scores |
tides | array of types & heights |
weather | array of sunrise/set times, array of temperatures/weather conditions |
Param | Values | Effect |
---|---|---|
spotId | string | Surfline spot id that you want data for. A typical Surfline URL is https://www.surfline.com/surf-report/venice-breakwater/590927576a2e4300134fbed8 where 590927576a2e4300134fbed8 is the spotId |
days | integer | Number of forecast days to get (Max 6 w/o access token, Max 17 w/ premium token) |
intervalHours | integer | Minimum of 1 (hour) |
maxHeights | boolean | true seems to remove min & optimal values from the wave data output |
accesstoken | string | Auth token to get premium data access (optional) |
Anywhere there is an optimalScore
the value can be interpreted as follows:
Value | Meaning |
---|---|
0 | Suboptimal |
1 | Good |
2 | Optimal |
Surfline's old API is undocumented and unauthenticated, but was used via javascript on their website, so it was fairly easy to reverse-engineer. However, they have updated their site & apps to use the new API, and it appears that they've stopped including some critical data in the responses for the old API, so it's disabled in this app for now (and probably forever).
It returned JSON, but with a very odd structure, with each item that is time-sensitive containing an array of daily arrays of values that correspond to timestamps provided in a separate set of arrays. For example (lots of data left out for brevity):
"Surf": {
"dateStamp": [
[
"January 24, 2016 04:00:00",
"January 24, 2016 10:00:00",
"January 24, 2016 16:00:00",
"January 24, 2016 22:00:00"
],
[
"January 25, 2016 04:00:00",
"January 25, 2016 10:00:00",
"January 25, 2016 16:00:00",
"January 25, 2016 22:00:00"
]
],
"surf_min": [
[
2.15,
1.8,
1.4,
1
],
[
0.7,
0.4,
0.3,
0.3
]
],
}
Requests are structured as follows:
https://api.surfline.com/v1/forecasts/{spot_id}?{params}
This is a breakdown of the params available:
Param | Values | Effect |
---|---|---|
spot_id | integer | Surfline spot id that you want data for. A typical legacy Surfline URL is https://www.surfline.com/surf-report/venice-beach-southern-california_4211/ where 4211 is the spot_id . You can also get this from the response's id property. Unfortunately I'm no longer aware of a way to get legacy spot ids. |
resources | string | Any comma-separated list of "surf,analysis,wind,weather,tide,sort". There could be more available that I haven't discovered. "Sort" gives an array of swells, periods & heights that are used for the tables on spot forecast pages. To see the whole list, just set 'all'. |
days | integer | Number of days of forecast to get. This seems to cap out at 16 for Wind and 25 for Surf. |
getAllSpots | boolean | false returns an object containing the single spot you requested, true returns an array of data for all spots in the same region as your spot, in this case "South Los Angeles" |
units | string | e returns American units (ft/mi), m uses metric |
usenearshore | boolean | The best that I can gather, you want this set to true to use the more accurate nearshore models that take into account how each spot's unique bathymetry affects the incoming swells. |
interpolate | boolean | Provide "forecasts" every 3 hours instead of ever 6. These interpolations seem to be simple averages of the values of the 6-hour forecasts. |
showOptimal | boolean | Includes arrays of 0's & 1's indicating whether each wind & swell forecast is optimal for this spot or not. Unfortunately the optimal swell data is only provided if you include the "sort" resource - it is not included in the "surf" resource. |
callback | string | jsonp callback function name |
MagicSeaweed has a well-documented JSON API that requires requesting an API key via email. This was a straightforward process and they got back to me quickly with my key.
I've asked MagicSeaweed a few questions and added their responses below:
- "Our API provides 5 days of forecast data, with segments of data provided for each 3 hour interval during that 5 day time span."
- "Our data is updated every 3 hours."
Spitcast only provides a list of API endpoints, but the data is sanely-structured JSON so it's pretty easy to parse.
I've asked Jack from Spitcast a few questions and added his responses below:
- To get more than the default 24 hour forecast for a spot, add
dcat=week
to the querystring. - Why does the site show a size range, but the API only returns one
size
value? "I actually take the API number and create the max by adding 1/6 the height (in feet), and then create the min by subtracting 1/6 the height." - All possible values for shape:
- Poor
- Poor-Fair
- Fair
- Fair-Good
- Good
All of the forecasting services (including Surfline v1 vs v2) use different systems for rating waves. I've attempted to normalize them all to a 0-5 (6-point) scale as best as possible, which is perhaps easier to understand when mapped onto the commonly-used Poor-Good scale (some throw Epic in there at the top end, but I went with Very Good):
- 0 => Poor
- 1 => Poor - Fair
- 2 => Fair
- 3 => Fair - Good
- 4 => Good
- 5 => Very Good
Each forecasting service is massaged onto that scale as follows:
-
MagicSeaweed: integer
fadedRating
(0-5) &solidRating
(0-5). I simply subtract fadedRating (which is essentially the negative effect of wind) from solidRating. -
Spitcast: ratings in text form:
- Poor
- Poor-Fair
- Fair
- Fair-Good
- Good
I massage these by assigning them a number from 0.5-4.5
-
Surfline v1: decimal ratings (0-1) for up to 6 different swells at each spot, as well as an
optimalWind
boolean. I take the max swell rating at any given time for that spot, multiply it by 5, and then halve it if the wind is not optimal. -
Surfline v2: integer
optimalScore
s for both swell & wind (0-2). Adding these together gives a 0-4 scale, and adding 0.5 puts it on the same 0.5-4.5 scale as Spitcast.
It took me a long time to land on a solution here, but I've finally settled on storing all timestamps in the database in the spot's local time. This defies Rails convention, but makes intuitive sense. If you pull up the forecasts table and look at the timestamp, that's the actual local time at that spot that it's forecast for (even though Rails & Postgres both think it's being stored in UTC). This is typically the format that the forecasting service gives it to us in, and what users want to see it in, so there's no point in doing all sorts of fancy conversion when it should be the same all the way through the pipeline. Now, you may ask, why am I still using the Rails default of TIMESTAMP WITHOUT TIMEZONE
, and the answer is that shockingly enough, TIMESTAMP WITH TIMEZONE
doesn't actually store timezone data!
- Improve charts:
- Fix timestamp formatting.
- Account for min/max size forecast. Currently charts just reflect the max.
- Display forecast quality ratings. Perhaps color each bar different depending on how good the rating is. Surfline also has an
optimal_wind
boolean that is being crudely integrated into thedisplay_swell_rating
method - improvements welcome.
- Refresh data on a schedule based on when new data is available (refreshing all forecast sources hourly)
- Support multiple timezones as opposed to Pacific Time only
- New Surfline API
- Stop manually seeding the db and figure out a way to pull all spots from each data source and automatically associate them to a canonical spot record (probably using geocoding)
- Dark Theme
- Remove asset pipeline & process CSS w/ webpacker
- Explore lazy-loading components
- Explore SSR for possibly faster browser paint
- Explore Preact
- Fetch & display tide/wind/water temperature data from NOAA (they actually have a decent API!)
- Fetch & display recent buoy trends that are relevant to each spot to give an idea of when swell is actually arriving.