Output preset - split when document reaches X number of pages, but without splitting a document in two

Given a datamapper that has several thousands records and a template that creates variable sections for each (the number of pages of each record is unknown before), is it possible to create and output preset that creates merged PDFs with no more than, let’s say, 1000 pages but without splitting the pages of a record between two files?

For example, given a limit of 15 pages per file:

  • Record A has 10 pages
  • Record B has 17 pages
  • Record C has 3 pages

Expected output:

  • File One with 17 pages (record A and B), even if the number of pages is higher than the threshold, a record cannot be split, so all the record B is included
  • File Two with 3 pages (record C)

I second this feature request. I want to split by SHEET or PAGE count, WITHOUT “breaking” a Document. And I’d rather go UNDER my count that OVER. In other words, if the NEXT document would exceed the split count, do not include that document in the current split.

I have done this in the past by using a Metadata Sequencer to split the metadata given to the Create Output step. That way you can set it 1,000 items at ‘document’ level for example.

This has at the added benefit of outputting said metadata to a folder and then using a threaded process to pick the back up and run multiple WeaverEngine processes simultaneously.

It’s already possible in an Output Preset to split at “x” number of Documents (or Document Sets, or Job Segments), OR by a specific sheet count. The request here is to split on a sheet count WITHOUT breaking a Document.

The use case is if the document needs to be cut or trimmed, and the cutter can physically only handle 2500 sheets of paper. You want to break the job based on a sheet count of 2500, but not if that leaves you with a partial document at the bottom of the stack. We want to break on document boundaries into sets of “up to” 2500 sheets. Note that in many jobs the number of sheets in a document is variable.

TDGreer, what about using the “size grouping after normal grouping” option? I’d expect the document boundary to be respected first, and then the size grouping to take place, which should result in document sets of complete documents up to the page or sheet size limit you set on each size grouping. One of the OL moderators will correct me if I’m mistaken…

Hi James,

I guess this was your resolution - Out of memory during Output Creation - Workflow - Upland OL User community

We’re having out of memory issue on Metadata sequencer because of the amount of data per record. Sorting out to Output preset may not feasible for us as we have to write companion text file per batch/iteration. OL support is suggesting us to stay out of sequencer. I’m still looking for any other alternative. Any chance you encountered the same issue?

Thanks,R

I also second this feature request, so pushing it to the top of the thread.