Upgrading
Instructions to migrate existing geoprocessing projects to newer versions.
Upgrade Geoprocessing
Your project will be pinned to a specific version of the geoprocessing library in package.json, under devDependencies
. To update to the latest version, first, make sure you don't have any unsaved work and your git repository is in a clean state. Then update the version, and run the upgrade script:
npm update @seasketch/geoprocessing@latest
npm install
npm run upgrade
The upgrade command may alter or overwrite a significant number of files including configuration, scripts, package.json dependencies and scripts, etc. The upgrade command may overwrite any customizations you have made such as to scripts, in which case you would need to manually recover them by studying the git diffs.
If you are upgrading to a new major version (e.g 6.1.x to 7.0.x), then there will be breaking changes that may affect your project, covered in this document. Check the release notes for any additional required steps.
It is suggested that you upgrade only one major version at a time if there are significant manual steps required to migrate your project. If you need to upgrade 2 or more major versions, consider simply creating a new project from scratch, them migrating everything over one at a time (datasources, metric groups, functions and report clients).
After upgrading you should always rerun your tests and storybook to verify everything is working properly and test output changes are as expected.
npm test
npm run storybook
Finally, rebuild and redeploy your project.
npm run build
npm run deploy
After deploy, breaking changes may require you to take additional steps. Follow the steps for your particular migration. If you are seeing errors or unexpected behavior, try any one of the following steps:
- Rebuild dependencies:
rm -rf node_modules && rm package-lock.json && npm install
- Reimport all datasources:
npm run reimport:data
- Republish all datasources:
npm run publish:data
- Clear AWS cache:
npm run clear-results
, thenall
Test Stack
If you'd like to deploy to a test stack first, alongside your existing production stack, to make sure everything is working properly:
- Change the
name
field of your project in package.json (e.g."name": "my-reports"
becomes `"name": "my-reports-7x". - Then
npm run build
andnpm run deploy
. This will deploy to a new AWS CloudFormation stack, separate from your production stack. - Once deployed, repoint your existing sketch classes to the new geoprocessing service, or create separate "admin only" sketch classes that point to your new test service. Make sure that all required sketch class attributes are in place.
- When you are ready to update production, change the
name
in package.json back, and rerunbuild
anddeploy
.
Migration
6.x to 7.x
Numerous manual migration steps are required, including a number of breaking changes:
- Testing framework switched from Jest to Vite
- Required use of new Node features (e.g. switch from __dirname to import.meta.dirname)
- Required use of ES Module structure and import style
- Lambda functions are now maintained in one or more LambdaStack(s) nested within the root CloudFormation stack. The number of nested stacks auto-scales as functions are added to the project (about 15-20 functions per LambdaStack). This change typically will cause an error when re-deploying an existing project, because the logical ID's for lambda functions are unique and fixed and migrating functions from the root stack to a nested LambdaStack causes a
duplicate identifier
error. If you see this error, the solution is to runnpm run destroy
to delete your stack first. Then rundeploy
again. You will need topublish
your datasources again after deploy. Follow the instructions up top for how to use atest stack
See fsm-reports for a migration from 6.x to 7.x and brazil-reports for migration from 3.x to 7.x which was done by starting with a freshly created project in a git branch and slowly migrating the datasources, metric groups, functions, and report clients over.
Upgrade Devcontainer
If you're using geoprocessing-devcontainer
to develop in a Docker environment, you will need to update this repo and the underlying geoprocessing-workspace
docker image to the latest. First, make sure Docker Desktop is running, then:
cd geoprocessing-devcontainer
git pull
docker pull seasketch/geoprocessing-workspace
You should now be able to start a Docker container shell using the latest image and test that everything is up to date
sudo docker run -it --entrypoint /bin/bash seasketch/geoprocessing-workspace
(base) vscode ➜ / $ node -v
v20.12.1
(base) vscode ➜ / $ npm -v
10.5.0
(base) vscode ➜ / $ gdalinfo --version
GDAL 3.8.5, released 2024/04/02
Exit back out of this shell when done
The latest version of the geoprocessing-workspace
will only work with geoprocessing 7.x projects. This is due to a change in how GDAL produces flatgeobuf files. If you suddenly see errors of "Not a FlatGeobuf file"
when trying to read your file, this is likely the reason. In order to continue to develop older 6.x and lower geoprocessing projects you will need to start your devcontainer using the local-dev-pre-7x
environment. This is pinned to an older version of the docker image - seasketch/geoprocessing-workspace:sha-69bb889
If you're maintaining your own development environment then you should look to have at least the following versions at minimum:
- Node 20.12.1
- NPM 10.5.0
- GDAL 3.5.0
Upgrade Windows Geoprocessing
Upgrade Geoprocessing Distribution
If you previously installed a version of the Geoprocessing distribution for running in WSL, and now want to replace it with a new one, first backup your existing Geoprocessing distribution (if needed):
mkdir C:\tmp\WslBackups\Geoprocessing
wsl --export Geoprocessing C:\tmp\WslBackups\20241104_Geoprocessing.tar
Then unregister it:
wsl --unregister Geoprocessing
Then follow the instructions above to install WSL Geoprocessing image again
Upgrade Script
As of v7.0-beta.5 there is a new upgrade
script that automates installing/updating assets in your project from the gp library. As of now it upgrades:
- package.json - updates scripts, dependencies, devDependencies
- .storybook - installs directory
- .vscode - overwrites
.vscode
directory with new files - i18n - creates a new
project/i18n.json
file. overwritessrc/i18n
directory with new files.
Some of these upgrades are destructive and will simply overwrite the director in your project (.storybook, .vscode, src/i18n). If you have customized any of these parts of your project, then be sure to look at the git changelog and bring back any of your work.
This includes: package.json, i18n, storybook, vscode.
Add it to your projects package.json scripts
section
"upgrade": "geoprocessing upgrade"
Then run it:
npm run upgrade
Convert project to ES Modules
This is the biggest breaking change in v7. You will need to change your project to an ES Module
structure or ESM
for short. With this change, Node will use an ESM
runtime, instead of the original CommonJS
, which will necessitate some additional code changes covered below.
Use the base project that gets installed when you init a new geoprocessing project as a guide.
First, in package.json:
- Add
"type": "module",
Then reload your VSCode window to make sure it picks up that your project is now an ESM project. You can do this with Cmd-Shift-P
on Mac or Ctrl-Shift-P
on Windows and then start typing reload
and select the Developer: Reload Window
option. Or just close and restart the VSCode app as you normally would.
Now you need to update all of your projects source files to be ESM and Node v20 compliant. VSCode should give you hints along the way, so basically just click through all the source files looking for red squiggle underlined text. You will focus in the src
directory.
- For each import of a local module (e.g.
import project from ../project
), use a full explicit path and include a.js
extension on the end, even if you are importing a.ts
file. The example would becomeimport project from ../project/projectClient.js
. - NodeJS when using the ES Module engine now requires explicit paths to code files. No longer can you import a module from a directory (e.g.
import foo from ./my/directory
) and expect it will look for an index.js file. You have to change this toimport foo form ./my/directory/index.js
.__dirname
built-in must be changed toimport.meta.dirname
Migrate asset imports
require
is no longer allowed for importing images and other static assets. Vite expects you to import the assets directly as urls. SizeCard.tsx is one component installed by default with projects that will need to be updated.
Change:
<img
src={require("../assets/img/territorial_waters.png")}
style={{ maxWidth: "100%" }}
/>
to:
import watersImgUrl from "../assets/img/territorial_waters.png";
...
{<img src={watersImgUrl} style={{ maxWidth: "100%" }} />}
At this point, VSCode will complain about your image import, it doesn't support importing anything other than code and JSON files by default. The code bundler now used by your project, Vite, knows how to do this however, you just need to load its capabilities by creating a file called vite-env.d.ts
at the top-level of your project with the following:
/// <reference types="vite/client" />
// Add Vite types to project
// https://vitejs.dev/guide/features.html#client-types
Other Changes
- Rename babel.config.js to babel.config.cjs. This babel config is used only by the translation library.
- update project/projectClient.ts with type assertion syntax
import datasources from "./datasources.json" with { type: "json" };
import metrics from "./metrics.json" with { type: "json" };
import precalc from "./precalc.json" with { type: "json" };
import objectives from "./objectives.json" with { type: "json" };
import geographies from "./geographies.json" with { type: "json" };
import basic from "./basic.json" with { type: "json" };
import projectPackage from "../package.json" with { type: "json" };
import gp from "../project/geoprocessing.json" with { type: "json" };
import { ProjectClientBase } from "@seasketch/geoprocessing/client-core";
const projectClient = new ProjectClientBase({
datasources,
metricGroups: metrics,
precalc,
objectives,
geographies,
basic,
package: projectPackage,
geoprocessing: gp,
});
export default projectClient;
Migrate styled-components
- If you have report components that use styled-components for its styling, you will need to change all code imports of
styled-components
from
import styled from "styled-components";
to use of the named export
import { styled } from "styled-components";
- Also when you run storybook or load your reports in production you may start to see React console warnings about extra attributes being present.
`React does not recognize the rowTotals prop on a DOM element. If you intentionally want it to appear in the DOM as a custom attribute, spell it as lowercase rowtotals instead. If you accidentally passed it from a parent component, remove it from the DOM element.
The solution is to switch to using transient
prop names, or component prop names that start with a dollar sign (e.g. $rowTotals
instead of rowTotals
). Styled-components will automatically filter these props out before passing to React to render them as element attributes in the browser.
Stop importing directly from @seasketch/geoprocessing in report clients
- Report client code must no longer import from geoprocessing libraries top level entry point
@seasketch/geoprocessing
or you may see a "require is not defined" error or other errors related to Node specific modules not found. The solution is to switch from for example:
import { ProjectClientBase } from "@seasketch/geoprocessing";
to:
import { ProjectClientBase } from "@seasketch/geoprocessing/client-core";
The use of the top-level entry point has persisted in some code because the previous Webpack code bundler did some extra magic to not let Node modules be bundled into report client browser code. The new Vite code bundler does not do this magic and leaves it to you to track your imports. The geoprocessing library offers both the client-core
and client-ui
entry points which should be used. These should offer everything you need.
6.0 to 6.1
- Run
reimport:data
to ensure that all raster datasources indata/dist
are in an equal area projection. - Run
precalc:data
for all raster datasources to precalculate additional metrics includingsum
,area
,valid
,count
. - Run
publish:data
for all raster datasources to ensure equal area version is published to S3 storage. - Migrate geoprocessing functions from
otverlapRaster()
(now deprecated) torasterMetrics()
as you have time, and need to calculate additional stats like area.rasterStats()
andgetArea()
are available as lower level alternatives for constructing your own functions. - any use of geoblaze directly, that passes a polygon feature for overlap, must reproject the feature to an equal area projection first, using
toRasterProjection
. See getSum for an example. - any use of the deprecated
loadCogWindow()
should be replaced with the newerloadCog()
. The former doesnt' appear to work correctly with functions likerasterStats()
andrasterMetrics()
.
5.x to 6.x
- Add
explodeMulti: true
to all vector datasources inproject/datasources.json
. You can set this to false if you know for sure you need to maintain multipolygons in the datasource. Otherwise breaking them up can speed up geoprocessing function by not fetching an operating on extra polygons outside the bounding box of a sketch.
4.x to 5.x
package.json
- Update package.json to latest 5.x version of geoprocessing library and run
npm install
- Add the
precalc:data
andprecalc:data:clean
cli commands topackage.json
:
{
"precalc:data": "start-server-and-test 'http-server data/dist -c-1 -p 8001' http://localhost:8001 precalc:data_",
"precalc:data_": "geoprocessing precalc:data",
"precalc:data:clean": "geoprocessing precalc:data:clean"
}
- Drop use of web server from
import:data
andreimport:data
{
"import:data": "geoprocessing import:data",
"reimport:data": "geoprocessing reimport:data"
}
- The ProjectClient now takes precalc metrics and geographies as input. Update
project/projectClient.ts
to the following:
import datasources from "./datasources.json" with { type: "json" };
import metrics from "./metrics.json" with { type: "json" };
import precalc from "./precalc.json" with { type: "json" };
import objectives from "./objectives.json" with { type: "json" };
import geographies from "./geographies.json" with { type: "json" };
import basic from "./basic.json" with { type: "json" };
import projectPackage from "../package.json" with { type: "json" };
import gp from "../project/geoprocessing.json" with { type: "json" };
import { ProjectClientBase } from "@seasketch/geoprocessing" with { type: "json" };
const projectClient = new ProjectClientBase({
datasources,
metricGroups: metrics,
precalc: precalc,
objectives,
geographies,
basic,
package: projectPackage,
geoprocessing: gp,
});
export default projectClient;
Geographies
Geographies are a new construct, most commonly used for planning boundaries. You are required to define at least one per project and you can have more than one. Projects have always had them, but they were implicitly defined based on how data was clipped, which was both unclear to the report developer and very limiting. Geographies are explicit. There is no longer confusion about whether and how to clip datasources to one or more planning boundaries. You just define what the geography boundaries are, by associating it with a datasource. Then the precalc command will clip the datasource (whether vector or raster) to each geographies features (intersection) and precompute metrics with what remains (total area, count, sum). This replaces what was keyStats
in datasources.json. Preclac metrics are typically used as the denominator when calculating % sketch overlap in reports. Geoprocessing functions also clip the current sketch to one or more geographies at runtime when calculating metrics. These are often used as the numerator when when calculating sketch % overlap in reports.
To setup your projects default geography, create a new file project/geographies.json
. Choose from one of the options below for your default geography. Just make sure that the geography is assigned to the default-boundary
group, and precalc
is set to true
- If you already have a local datasource with your planning boundary, then just define a geography that uses that datasource.
[
{
"geographyId": "nearshore",
"datasourceId": "6nm_boundary",
"display": "Azores",
"layerId": "x6kSfK6Lb",
"groups": ["default-boundary"],
"precalc": true
}
]
- If your planning boundary is a Marine Regions EEZ, you can define an
eez
geography that uses the newglobal-eez-mr-v12
datasource (see below on how to add this datasource to your project), which is the default for a new project when you choose the Ocean EEZ template. You just need to apply the correctbboxFilter
andpropertyFilter
for your EEZ or EEZ's of choice. [TODO: ADD WEB LINK]
[
{
"geographyId": "eez",
"datasourceId": "global-eez-mr-v12",
"display": "Samoan EEZ",
"propertyFilter": {
"property": "GEONAME",
"values": ["Samoan Exclusive Economic Zone"]
},
"bboxFilter": [
-174.51139447157757, -15.878383591829206, -170.54265693017294,
-10.960825304544073
],
"groups": ["default-boundary"],
"precalc": true
}
]
- If you don't have a planning boundary or want to use the entire world as your planning boundary you can use the world geography which uses the world datasource (see below for how to add this datasource). world is the new default geography for all new
blank
geoprocessing projects.
[
{
"geographyId": "world",
"datasourceId": "world",
"display": "World",
"groups": ["default-boundary"],
"precalc": true
}
]
Datasources
Based on your geography choice above, add the corresponding datasource for this geography to your datasources.json
file.
World datasource published by global-datasources:
[
{
"datasourceId": "world",
"geo_type": "vector",
"formats": ["json", "fgb"],
"layerName": "world",
"classKeys": [],
"url": "https://gp-global-datasources-datasets.s3.us-west-1.amazonaws.com/world.fgb",
"propertiesToKeep": [],
"metadata": {
"name": "World Outline Polygon",
"description": "World polygon for default project geography in seasketch geoprocessing proejcts",
"version": "1.0",
"publisher": "SeaSketch",
"publishDate": "20231018",
"publishLink": ""
},
"precalc": false
}
]
Global EEZ datasource published by global-datasources (with filters set to for Samoa EEZ)
[
{
"datasourceId": "global-eez-mr-v12",
"geo_type": "vector",
"formats": ["fgb", "json"],
"metadata": {
"name": "World EEZ v11",
"description": "World EEZ boundaries and disputed areas",
"version": "11.0",
"publisher": "Flanders Marine Institute (VLIZ)",
"publishDate": "2019118",
"publishLink": "https://marineregions.org/"
},
"idProperty": "GEONAME",
"nameProperty": "GEONAME",
"classKeys": [],
"url": "https://gp-global-datasources-datasets.s3.us-west-1.amazonaws.com/global-eez-mr-v12.fgb",
"propertyFilter": {
"property": "GEONAME",
"values": ["Samoan Exclusive Economic Zone"]
},
"bboxFilter": [
-174.51139447157757, -15.878383591829206, -170.54265693017294,
-10.960825304544073
],
"precalc": false
}
]
Finally, you need to add a precalc
setting to all other datasources in your datasources.json
file. This property is required, and you will see validation errors when running any data commands or smoke tests.
- Add
"precalc": true
to all datasources inproject/datasources.json
that metrics should be precalculated for. This is typically limited to datasources that you need to precalc the overall summary metric of the datasource (total area, total feature count) so that you can report the % of the total that a sketch overlaps with. Otherwise you don't need to precalc. - Set all other datasources to
"precalc": false
. This includes global datasources or datasources that are only a source forgeography
features and otherwise aren't used in reports. Setting these to true will at best just precalculate extra metrics that won't be used. At worst it will try to fetch entire global datasources and fail at the task, because the necessary filters aren't in place.
Precalc
Once you have your geographies
and datasources
properly configured, you're ready to run the precalc:data
command.
- Create a new file
project/precalc.json
populated with an empty array[]
npm run precalc:data
This will precompute metrics for all combinations of geographies and datasources. It will also strip any existing keyStats
properties from datasources.json
and populate precalc.json
.
Geoprocessing functions
- Update
clipToGeography
function, to allow geographies to use external datasources. To copy the file from the gp library to your project space, run the following from a terminal in the top-level of your project.
mkdir -p src/util && cp -r node_modules/@seasketch/geoprocessing/dist/base-project/src/util/clipToGeography.ts src/util/clipToGeography.ts
- update all geoprocessing functions to use the following boilerplate:
- look for
geographyIds
passed in extraParams - get the geography using project client and fallback to default geography
- clip the current sketch to the geography
- see azores-nearshore-reports for examples.
- look for
Report Clients
Reports clients should migrate to new boilerplate code. See azores-nearshore-reports for examples, specifically:
- receive the current geography as a
geographyId
passed in props. - get the current geography using
project.getGeographyById
. - pass geographyId to
project.getPrecalcMetrics()
- pass geographyId to ResultsCard as
extraParams
for gp function to use - use curGeography.display string as appropriate in client to tell user which geography is being reported on.
- Update any calls to
toPercentMetric()
, now overriding metricIds are passed in through an object, instead of directly. (i.e.toPercentMetric(singleMetrics, precalcMetrics, project.getMetricGroupPercId(metricGroup)
becomestoPercentMetric(singleMetrics, precalcMetrics, {metricIdOverride: project.getMetricGroupPercId(metricGroup)})
)
If you would like to allow the user to switch between planning geographies from the report UI, you can add a GeographySwitcher
at the top level of the report client (see azores-nearshore-reports example). The user chosen geography ID is then passed into each ResultsCard
and on to each geoprocessing function as extraParams
Language Translation
- Geography display names in geographies.json are now extracted with
npm run extract:translations
. Then translate these strings per your current workflow and GeographySwitcher will use them automatically. The same is true for objectives.json and metrics.json. - Update src/i18n/i18nAsync.ts to latest. Copy the following file in the gp library to your project space to overwrite.
cp -r node_modules/@seasketch/geoprocessing/dist/base-project/src/i18n/i18nAsync.ts src/i18n/i18nAsync.ts