-
Notifications
You must be signed in to change notification settings - Fork 56
Add Segment migration guide #1348
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
✅ Deploy Preview for snowplow-docs ready!
To edit notification comments on pull requests, go to your Netlify project configuration. |
7e419cd
to
6358029
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@mscwilson overall this is better, and I like having the text shorter. It does feel a bit too terse and low-level at the moment though - it only focuses on tracking, and assumes a lot of knowledge about Snowplow. I wouldn't mind adding a bit more context about our platform, bearing in mind the likely audience is someone who is looking to understand the overall differences and what they should be thinking about.
@@ -8,22 +8,7 @@ This guide helps technical implementers migrate from Segment to Snowplow. | |||
|
|||
## Platform differences | |||
|
|||
There are a number of differences between Segment and Snowplow as a data platform. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think this feature comparison table is actually quite important - we can probably make the text a bit less marketing-speaky and more neutral, but it's important to lay out the differences at a higher level.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
that depends what this document is for. I think overall comparisons of the platforms is a marketing thing, they can already read about that on our marketing website https://snowplow.io/comparisons/snowplow-vs-segment
we could link to that comparison?
this is about how to technically migrate - what concepts map to what, what do they need to action to make it work
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
yup, let's at least link to the comparison
2. Implement and validate | ||
3. Cutover and finalize | ||
|
||
### Assess and plan |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The section from here on out feels a bit too bullet-point-y - definitely AI-generated, but with not quite enough context or detail to actually be useful. We should assume that people reading this migration guide are not actually familiar with Snowplow, so taking a bit of extra time to explain concepts in more detail could be helpful.
I also find myself wanting numbers? e.g. "1. Assess and plan", "1.1 Audit existing implementation" etc
|
||
Segment loads each custom event type into separate tables, for example, `order_completed`, or `product_viewed` tables. Analysts must `UNION` tables together to reconstruct user journeys. | ||
|
||
Snowplow uses a single [`atomic.events`](/docs/fundamentals/canonical-event/index.md) table in warehouses like Snowflake and BigQuery. Events and entities are stored as structured columns within that table, simplifying analysis. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I would also mention that we provide OOTB data models to model common analytics use-cases, and link to our DBT packages. Otherwise we risk this migration guide focusing solely on the upstream tracking bit, vs comparing one data platform to another.
- Migrate BI dashboards to query Snowplow tables | ||
- Test all data-dependent workflows | ||
|
||
#### Configure integrations |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think we could say more here - both of these are full products in their own right, and ProfServ has mentioned that these have been quite tricky to do in the past.
- Translate Segment events into Snowplow self-describing events | ||
- The [Snowplow CLI](/docs/data-product-studio/snowplow-cli/index.md) MCP server can help with this | ||
- Identify reusable entities that can replace repeated properties | ||
- Create JSON schemas for all events and entities |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We should mention Data Products here - it is about more than just events and entities. And this isn't quite the right order of operations - the MCP server does actually create the JSON schemas.
No description provided.