Giant-scale Person Sequences at Pinterest | by Pinterest Engineering | Pinterest Engineering Weblog | Could, 2023

Person Understanding group: Zefan Fu, Minzhe Zhou, Neng Gu, Leo Zhang, Kimmie Hua, Sufyan Suliman | Software program Engineer, Yitong Zhou | Software program Engineering Supervisor

Index Core Entity group: Dumitru Daniliuc, Jisong Liu, Kangnan Li | Software program Engineer, Shunping Chiu | Software program Engineering Supervisor

User Signal Service Platform

Understanding and responding to consumer actions and preferences is crucial to delivering a personalised, top quality consumer expertise. On this weblog publish, we’ll talk about how a number of groups joined collectively to construct a brand new large-scale, highly-flexible, and cost-efficient consumer sign platform service, which indexes the related consumer occasions in close to real-time, constructs them into consumer sequences, and makes it tremendous straightforward to make use of each for on-line service requests and for ML coaching & inferences.

Person sequence is one kind of ML function composed as a time-ordered record of consumer engagement actions. The sequence captures one’s latest actions in real-time, reflecting their newest pursuits in addition to their shift of focus. This type of sign performs a crucial position in varied ML purposes, particularly for large-scale sequential modeling purposes (see instance).

To make the real-time consumer sequence extra accessible throughout the Pinterest ML ecosystem, and to empower our every day metrics enchancment, we record the next key options to ship for ML purposes:

  • Actual-time: on common < 2 seconds latency from a consumer’s newest motion to the service response
  • Flexibility: knowledge might be fetched and reused by a mix-and-use sample to allow sooner iterations for ML engineers specializing in fast growth time
  • Platform: serve all completely different wants and requests with a uniform knowledge API layer
  • Price Environment friendly: enhance infra shareability and reusability, and keep away from duplications in storage or computation wherever doable


  1. Sign: the info inputs for downstream purposes particularly in machine studying purposes
  2. Person Sequence: a selected sort of consumer alerts that arranges consumer’s previous actions in a strict temporal order and joins every exercise with enrichment knowledge
  3. Unified Function Illustration: or “UFR” is a function format for all Pinterest mannequin options
Realtime indexing pipeline, offline indexing pipeline, serving side

Our infrastructure adopts a lambda architecture: the real-time indexing pipeline, the offline indexing pipeline, and the serving facet elements.

Actual-Time Indexing Pipeline

The principle purpose of the real-time indexing pipeline is to counterpoint, retailer, and serve the previous couple of related consumer actions as they arrive in. At Pinterest, most of our streaming jobs are constructed on prime of Apache Flink, as a result of Flink is a mature streaming framework with a variety of adoption within the business. So our consumer sequence real-time indexing pipeline consists of a Flink job that reads the related occasions as they arrive into our Kafka streams, fetches the specified options for every occasion from our function companies, and shops the enriched occasions into our KV retailer system. We arrange a separate dataset for every occasion kind listed by our system, as a result of we need to have the pliability to scale these datasets independently. For instance, if a consumer is more likely to click on on pins than to repin them, it is likely to be sufficient to retailer the final 10 repins per consumer, and on the similar time we’d need to retailer the final 100 “close-ups.”

repins and closeups

It’s price noting that the selection of the KV retailer expertise is extraordinarily essential, as a result of it may possibly have a big effect on the general effectivity (and in the end, value) of the complete infrastructure, in addition to the complexity of the real-time indexing job. Specifically, we wished our KV retailer datasets to have the next properties:

  1. Permits inserts. We’d like every dataset to retailer the final N occasions for a consumer. Nevertheless, after we course of a brand new occasion for a consumer, we don’t need to learn the prevailing N occasions, replace them, after which write all of them again to the respective dataset. That is inefficient (processing every occasion takes O(N) time as an alternative of O(1)), and it may possibly result in concurrent modification points if two hosts course of two completely different occasions for a similar consumer on the similar time. Due to this fact, our most essential requirement for our storage layer was to have the ability to deal with inserts.
  2. Handles out-of-order inserts. We wish our datasets to retailer the occasions for every consumer ordered in reverse chronological order (latest occasions first), as a result of then we are able to fetch them in essentially the most environment friendly manner. Nevertheless, we can’t assure the order by which our real-time indexing job will course of the occasions, and we don’t need to introduce a man-made processing delay (to order the occasions), as a result of we wish an infrastructure that enables us to right away react to any consumer motion. Due to this fact, it was crucial that the storage layer is ready to deal with out-of-order inserts.
  3. Handles duplicate values. Delegating the deduplication duty to the storage layer has allowed us to run our real-time indexing job with “not less than as soon as” semantic, which has enormously lowered its complexity and the variety of failure situations we would have liked to deal with.

Happily, Pinterest’s inside large column storage system (constructed on prime of RocksDB) may fulfill all these necessities, which has allowed us to maintain our real-time indexing job pretty easy.

Price Environment friendly Storage

Within the ML world, there isn’t any achieve that may be sustained with out caring for the fee. Irrespective of how fancy an ML mannequin is, it should perform inside cheap infrastructure prices. As well as, a price saving infra normally comes with optimized computing and storage which in flip contribute to the stableness of the system.

After we designed and carried out this technique, we saved value effectivity in thoughts from day one. To construct up this technique, the fee comes from two components: computing and storage. We carried out varied methods to cut back the fee from these two components with out sacrificing system efficiency.

  • Computing value effectivity: Throughout indexing time, at a excessive degree, Flink jobs ought to eat from the newest new occasions and apply these updates to the prevailing storage, representing the historic consumer sequence. As a substitute of learn, modify and write again, our Flink job is designed to solely append new occasions to the tip of consumer sequence and depend on storage periodical clean-up thread to keep up consumer sequence size beneath limitation. In contrast with read-modify-write, which has to load all earlier consumer sequence into Flink job, this strategy makes use of far much less reminiscence and CPU. This optimization additionally permits this job to deal with extra quantity after we need to index extra consumer occasions.
  • Storage value effectivity: To chase down storage prices, we encourage knowledge sharing throughout completely different use sequence use circumstances and solely retailer the enrichment of a consumer occasion when a number of use circumstances want it. For instance, let’s say use case 1 must click_event and view_event with enrichment A and B, and use case 2 must click_event with enrichment A solely. Use case 1 and a pair of will fetch click_event from the identical dataset, and solely enrichment A is built-in. Use case 1 must fetch view_event from one other dataset and fetch enrichment B within the serving time. This precept helps us maximize the info sharing throughout completely different use circumstances.

Offline Indexing Pipeline

Having a real-time indexing pipeline is crucial, as a result of it permits us to react to consumer actions and alter our suggestions in real-time. Nevertheless, it has some limitations. For instance, we can’t use it so as to add new alerts to the occasions that had been already listed. That’s the reason we additionally constructed an offline pipeline of Spark jobs to assist us:

  1. Enrich and retailer occasions every day. If the real-time pipeline missed or incorrectly enriched some occasions (because of some sudden points), the offline pipeline will right them.
  2. Bootstrap a dataset for a brand new related occasion kind. Each time we have to bootstrap a dataset for a brand new occasion kind, we are able to run the offline pipeline for that occasion kind for the final N days, as an alternative of ready for N days for the real-time indexing pipeline to supply knowledge.
  3. Add new enrichments to listed occasions. Each time a brand new function turns into obtainable, we are able to simply replace our offline indexing pipeline to counterpoint all listed occasions with the brand new function.
  4. Check out varied occasion choice algorithms. For now, our consumer sequences are primarily based on the final N occasions of a consumer. Nevertheless, sooner or later, we’d wish to experiment with our occasion choice algorithm (for instance, as an alternative of choosing the final N occasions, we may choose the “most related” N occasions). Since our real-time indexing pipeline wants to counterpoint and index occasions as quick as doable, we’d not have the ability to add refined occasion choice algorithms to it. Nevertheless, it will be very straightforward to experiment with the occasion choice algorithm in our offline indexing pipeline.

Lastly, since we wish our infrastructure to offer as a lot flexibility as doable to our product groups, we’d like our offline indexing pipeline to counterpoint and retailer as many occasions as doable. On the similar time, we’ve got to be conscious of our storage and operational prices. For now, we’ve got determined to retailer the previous couple of thousand occasions for every consumer, which makes our offline indexing pipeline course of PBs of knowledge. Nevertheless, our offline pipeline is designed to have the ability to course of far more knowledge, and we are able to simply scale up the variety of occasions saved per consumer sooner or later, if wanted.

Serving Layer

Our API is constructed on prime of the Galaxy framework (i.e. Pinterest’s inside sign processing and serving stack) and presents two varieties of responses: Thrift and UFR . Thrift permits for better flexibility by permitting the return of uncooked or aggregated options. UFR is right for direct consumption by fashions.

Our serving layer has a number of options that make it helpful for experiments and testing new concepts. Tenant separation ensures that use circumstances are remoted from one another, stopping issues from propagating. Tenant separation is carried out in function registration, logging and sign degree logic isolation. We make sure the heavy processing of 1 use case doesn’t have an effect on others. Whereas options might be simply shared, the enter parameters are strictly tied to function definition so no different use case can mess up the info. Well being metrics and built-in validations guarantee stability and reliability. The serving layer can also be versatile, permitting for straightforward experimentation at low value. Purchasers can take a look at a number of approaches inside a single experiment and rapidly iterate to seek out the very best answer. We offer tuning configurations in some ways, completely different sequence mixtures, function size, filtering thresholds, and so on, all of which may change instantly on-the-fly.

Extra particularly, on the serving layer, decoupled modules deal with completely different duties in the course of the processing of a request. The primary module retrieves key-value knowledge from the storage system. This knowledge is then handed by means of a filter, which removes any pointless or duplicate data. Subsequent, the enricher module provides further embedding to the info by becoming a member of from varied sources. The sizer module trims the info to a constant dimension, and the featurizer module converts the info right into a format that may be simply consumed by fashions. By separating these duties into distinct modules, we are able to extra simply preserve and replace the serving layer as wanted.

The choice to counterpoint embedding knowledge at indexing time or serving time can have a major influence on each the dimensions we retailer in kv and the time it takes to retrieve knowledge throughout serving. This trade-off between indexing time and serving time is basically a balancing act between storage value and latency. Transferring heavy joins to indexing time might end in smaller serving latency, but it surely additionally will increase storage value.

Our decision-making guidelines have advanced to emphasise chopping storage dimension as follows:

  • If it’s an experimental consumer sequence, it’s added to the serving time enricher
  • If it’s not shared with a number of surfaces, additionally it is added to the serving time enricher
  • If a timeout is reached throughout serving time, it’s added to the indexing time enricher

Constructing and successfully utilizing a generic infrastructure of this scale requires dedication from a number of groups. Historically, product engineers have to be uncovered to the infra complexity, together with knowledge schema, useful resource provisions, and storage allocations, which includes a number of groups. For instance, when product engineers need to make use of a brand new enrichment of their fashions, they should work with the indexing group to ensure that the enrichment is added to the related knowledge, and in flip, the indexing group must work with the storage group to ensure that our knowledge shops have the required capability. Due to this fact, you will need to have a collaboration mannequin that hides the complexity by clearly defining the obligations of every group and the best way groups talk necessities to one another.

Lowering the variety of dependencies for every group is essential to creating that group as environment friendly as doable. For this reason we’ve got divided our consumer sequence infrastructure into a number of horizontal layers, and we devised a collaboration mannequin that requires every layer to speak solely to the layer straight above and the one straight under.

On this mannequin, the Person Understanding group takes possession of the serving-side elements and is the one group that interacts with the product groups. On one hand, we conceal the complexity of this infrastructure from the product groups and supply the product groups with a single level of contact for all their requests. Alternatively, it offers the Person Understanding group visibility into all product necessities, which permits them to design generic serving-side elements that may be reused by a number of product groups. Equally, if a brand new product requirement can’t be happy on the serving facet and desires some indexing-side modifications, the Person Understanding group is chargeable for speaking these necessities to the Indexing Core Entities group, which owns the indexing elements. The Indexing Core Entities group then communicates with the “core companies” groups as wanted, with the intention to create new datasets, provision extra processing assets, and so on., with out exposing all these particulars to the groups larger up within the stack.

Having this “collaboration chain” (quite than a tree or graph of dependencies at every degree) additionally makes it a lot simpler for us to maintain monitor of all work that must be performed to onboard new use circumstances onto this infrastructure: at any time limit, any new use case is blocked by one and just one group, and as soon as that blocker is resolved, we routinely know which group must work on the subsequent steps.

UFR logging is commonly used each for mannequin coaching and mannequin serving. Most fashions maintain the info at serving time and use it for coaching functions to ensure they’re the identical.

Inside Mannequin construction, consumer sequence options are fed into sequence transformer and merged at function cross layer

For extra element data, please try this engineering article on HomeFeed mannequin taking in Person Sequence and increase Engagement Quantity

On this weblog, we introduced a brand new consumer sequence infra that introduces important enhancements on real-time responsiveness, flexibility, and value effectivity. Totally different than our earlier real-time consumer sign infra, this platform has been far more scalable and maximizes storage reusability. We’ve had profitable adoptions resembling in homefeed advice driving important consumer engagement positive factors. This platform can also be a key part for PinnerFormer work offering real-time consumer sequence knowledge.

For future work, we’re trying into each extra environment friendly and scalable knowledge storage options, resembling occasion compression or online-offline lambda structure, in addition to extra scalable on-line mannequin inference functionality built-in into the streaming platform. In the long term, we envision the real-time consumer sign sequence platform serving as a necessary infrastructure basis for all advice programs at Pinterest.

Contributors to consumer sequence adoption:

  • HomeFeed Rating
  • HomeFeed Candidate Technology
  • Notifications Relevance
  • Activation Basis
  • Search Rating and Mixing
  • Closeup Rating & Mixing
  • Advertisements Entire Web page Optimization
  • ATG Utilized Science
  • Advertisements Engagement
  • Advertisements Ocpm
  • Advertisements Retrieval
  • Advertisements Relevance
  • Residence Product
  • Galaxy
  • KV Storage Crew
  • Realtime Knowledge Warehouse Crew

To study extra about engineering at Pinterest, try the remainder of our Engineering Weblog and go to our Pinterest Labs website. To discover life at Pinterest, go to our Careers web page.