Preview backup for drive-backed services (#4699)
Add logic and tests for preview backups in drive-backed services. Does slightly change a few of the options for preview backup limits --- #### Does this PR need a docs update or release note? - [ ] ✅ Yes, it's included - [ ] 🕐 Yes, but in a later PR - [x] ⛔ No #### Type of change - [x] 🌻 Feature - [ ] 🐛 Bugfix - [ ] 🗺️ Documentation - [ ] 🤖 Supportability/Tests - [ ] 💻 CI/Deployment - [ ] 🧹 Tech Debt/Cleanup #### Test Plan - [ ] 💪 Manual - [x] ⚡ Unit test - [ ] 💚 E2E
This commit is contained in:
parent
d45842fe5c
commit
bcf290d628
@ -215,6 +215,22 @@ func (oc *Collection) IsEmpty() bool {
|
|||||||
return len(oc.driveItems) == 0
|
return len(oc.driveItems) == 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ContainsItem returns true if the collection has the given item as one of its
|
||||||
|
// children.
|
||||||
|
func (oc Collection) ContainsItem(item models.DriveItemable) bool {
|
||||||
|
_, ok := oc.driveItems[ptr.Val(item.GetId())]
|
||||||
|
return ok
|
||||||
|
}
|
||||||
|
|
||||||
|
// AddedItems returns the number of non-deleted items in the collection.
|
||||||
|
func (oc Collection) CountAddedItems() int {
|
||||||
|
// Subtract one since the folder is added to the collection so we get folder
|
||||||
|
// metadata. The collection of the root folder of the drive doesn't have its
|
||||||
|
// own folder reference since it doesn't have permissions the user can change,
|
||||||
|
// but it's close enough for our purposes.
|
||||||
|
return len(oc.driveItems) - 1
|
||||||
|
}
|
||||||
|
|
||||||
// Items() returns the channel containing M365 Exchange objects
|
// Items() returns the channel containing M365 Exchange objects
|
||||||
func (oc *Collection) Items(
|
func (oc *Collection) Items(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
|
|||||||
@ -29,7 +29,15 @@ import (
|
|||||||
"github.com/alcionai/corso/src/pkg/services/m365/api/pagers"
|
"github.com/alcionai/corso/src/pkg/services/m365/api/pagers"
|
||||||
)
|
)
|
||||||
|
|
||||||
const restrictedDirectory = "Site Pages"
|
const (
|
||||||
|
restrictedDirectory = "Site Pages"
|
||||||
|
|
||||||
|
defaultPreviewNumContainers = 5
|
||||||
|
defaultPreviewNumItemsPerContainer = 10
|
||||||
|
defaultPreviewNumItems = defaultPreviewNumContainers * defaultPreviewNumItemsPerContainer
|
||||||
|
defaultPreviewNumBytes int64 = 100 * 1024 * 1024
|
||||||
|
defaultPreviewNumPages = 50
|
||||||
|
)
|
||||||
|
|
||||||
// Collections is used to retrieve drive data for a
|
// Collections is used to retrieve drive data for a
|
||||||
// resource owner, which can be either a user or a sharepoint site.
|
// resource owner, which can be either a user or a sharepoint site.
|
||||||
@ -742,6 +750,83 @@ func (c *Collections) getCollectionPath(
|
|||||||
return collectionPath, nil
|
return collectionPath, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type driveEnumerationStats struct {
|
||||||
|
numPages int
|
||||||
|
numAddedFiles int
|
||||||
|
numContainers int
|
||||||
|
numBytes int64
|
||||||
|
}
|
||||||
|
|
||||||
|
func newPagerLimiter(opts control.Options) *pagerLimiter {
|
||||||
|
res := &pagerLimiter{limits: opts.PreviewLimits}
|
||||||
|
|
||||||
|
if res.limits.MaxContainers == 0 {
|
||||||
|
res.limits.MaxContainers = defaultPreviewNumContainers
|
||||||
|
}
|
||||||
|
|
||||||
|
if res.limits.MaxItemsPerContainer == 0 {
|
||||||
|
res.limits.MaxItemsPerContainer = defaultPreviewNumItemsPerContainer
|
||||||
|
}
|
||||||
|
|
||||||
|
if res.limits.MaxItems == 0 {
|
||||||
|
res.limits.MaxItems = defaultPreviewNumItems
|
||||||
|
}
|
||||||
|
|
||||||
|
if res.limits.MaxBytes == 0 {
|
||||||
|
res.limits.MaxBytes = defaultPreviewNumBytes
|
||||||
|
}
|
||||||
|
|
||||||
|
if res.limits.MaxPages == 0 {
|
||||||
|
res.limits.MaxPages = defaultPreviewNumPages
|
||||||
|
}
|
||||||
|
|
||||||
|
return res
|
||||||
|
}
|
||||||
|
|
||||||
|
type pagerLimiter struct {
|
||||||
|
limits control.PreviewItemLimits
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l pagerLimiter) enabled() bool {
|
||||||
|
return l.limits.Enabled
|
||||||
|
}
|
||||||
|
|
||||||
|
// sizeLimit returns the total number of bytes this backup should try to
|
||||||
|
// contain.
|
||||||
|
func (l pagerLimiter) sizeLimit() int64 {
|
||||||
|
return l.limits.MaxBytes
|
||||||
|
}
|
||||||
|
|
||||||
|
// atItemLimit returns true if the limiter is enabled and has reached the limit
|
||||||
|
// for individual items added to collections for this backup.
|
||||||
|
func (l pagerLimiter) atItemLimit(stats *driveEnumerationStats) bool {
|
||||||
|
return l.enabled() &&
|
||||||
|
(stats.numAddedFiles >= l.limits.MaxItems ||
|
||||||
|
stats.numBytes >= l.limits.MaxBytes)
|
||||||
|
}
|
||||||
|
|
||||||
|
// atContainerItemsLimit returns true if the limiter is enabled and the current
|
||||||
|
// number of items is above the limit for the number of items for a container
|
||||||
|
// for this backup.
|
||||||
|
func (l pagerLimiter) atContainerItemsLimit(numItems int) bool {
|
||||||
|
return l.enabled() && numItems >= l.limits.MaxItemsPerContainer
|
||||||
|
}
|
||||||
|
|
||||||
|
// atContainerPageLimit returns true if the limiter is enabled and the number of
|
||||||
|
// pages processed so far is beyond the limit for this backup.
|
||||||
|
func (l pagerLimiter) atPageLimit(stats *driveEnumerationStats) bool {
|
||||||
|
return l.enabled() && stats.numPages >= l.limits.MaxPages
|
||||||
|
}
|
||||||
|
|
||||||
|
// atLimit returns true if the limiter is enabled and meets any of the
|
||||||
|
// conditions for max items, containers, etc for this backup.
|
||||||
|
func (l pagerLimiter) atLimit(stats *driveEnumerationStats) bool {
|
||||||
|
return l.enabled() &&
|
||||||
|
(l.atItemLimit(stats) ||
|
||||||
|
stats.numContainers >= l.limits.MaxContainers ||
|
||||||
|
stats.numPages >= l.limits.MaxPages)
|
||||||
|
}
|
||||||
|
|
||||||
// PopulateDriveCollections initializes and adds the provided drive items to Collections
|
// PopulateDriveCollections initializes and adds the provided drive items to Collections
|
||||||
// A new collection is created for every drive folder.
|
// A new collection is created for every drive folder.
|
||||||
// Along with populating the collection items and updating the excluded item IDs, this func
|
// Along with populating the collection items and updating the excluded item IDs, this func
|
||||||
@ -772,9 +857,13 @@ func (c *Collections) PopulateDriveCollections(
|
|||||||
// already seen. This will help us track in case a folder was
|
// already seen. This will help us track in case a folder was
|
||||||
// recreated multiple times in between a run.
|
// recreated multiple times in between a run.
|
||||||
seenFolders = map[string]string{}
|
seenFolders = map[string]string{}
|
||||||
|
|
||||||
|
limiter = newPagerLimiter(c.ctrl)
|
||||||
|
stats = &driveEnumerationStats{}
|
||||||
)
|
)
|
||||||
|
|
||||||
ctx = clues.Add(ctx, "invalid_prev_delta", invalidPrevDelta)
|
ctx = clues.Add(ctx, "invalid_prev_delta", invalidPrevDelta)
|
||||||
|
logger.Ctx(ctx).Infow("running backup with limiter", "limiter", limiter)
|
||||||
|
|
||||||
if !invalidPrevDelta {
|
if !invalidPrevDelta {
|
||||||
maps.Copy(newPrevPaths, oldPrevPaths)
|
maps.Copy(newPrevPaths, oldPrevPaths)
|
||||||
@ -788,6 +877,12 @@ func (c *Collections) PopulateDriveCollections(
|
|||||||
Select: api.DefaultDriveItemProps(),
|
Select: api.DefaultDriveItemProps(),
|
||||||
})
|
})
|
||||||
|
|
||||||
|
// Needed since folders are mixed in with items. This allows us to handle
|
||||||
|
// hitting the maxContainer limit while (hopefully) still adding items to the
|
||||||
|
// container we reached the limit on. It may not behave as expected across
|
||||||
|
// page boundaries if items in other folders have also changed.
|
||||||
|
var lastContainerID string
|
||||||
|
|
||||||
for page, reset, done := pager.NextPage(); !done; page, reset, done = pager.NextPage() {
|
for page, reset, done := pager.NextPage(); !done; page, reset, done = pager.NextPage() {
|
||||||
if el.Failure() != nil {
|
if el.Failure() != nil {
|
||||||
break
|
break
|
||||||
@ -805,7 +900,10 @@ func (c *Collections) PopulateDriveCollections(
|
|||||||
c.CollectionMap[driveID] = map[string]*Collection{}
|
c.CollectionMap[driveID] = map[string]*Collection{}
|
||||||
invalidPrevDelta = true
|
invalidPrevDelta = true
|
||||||
|
|
||||||
|
// Reset collections and stats counts since we're starting over.
|
||||||
c.resetStats()
|
c.resetStats()
|
||||||
|
|
||||||
|
stats = &driveEnumerationStats{}
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, item := range page {
|
for _, item := range page {
|
||||||
@ -813,6 +911,24 @@ func (c *Collections) PopulateDriveCollections(
|
|||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Check if we got the max number of containers we're looking for and also
|
||||||
|
// processed items for the final container.
|
||||||
|
if item.GetFolder() != nil || item.GetPackageEscaped() != nil {
|
||||||
|
id := ptr.Val(item.GetId())
|
||||||
|
|
||||||
|
// Don't check for containers we've already seen.
|
||||||
|
if _, ok := c.CollectionMap[driveID][id]; !ok {
|
||||||
|
if id != lastContainerID {
|
||||||
|
if limiter.atLimit(stats) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
lastContainerID = id
|
||||||
|
stats.numContainers++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
err := c.processItem(
|
err := c.processItem(
|
||||||
ctx,
|
ctx,
|
||||||
item,
|
item,
|
||||||
@ -826,13 +942,45 @@ func (c *Collections) PopulateDriveCollections(
|
|||||||
topLevelPackages,
|
topLevelPackages,
|
||||||
invalidPrevDelta,
|
invalidPrevDelta,
|
||||||
counter,
|
counter,
|
||||||
|
stats,
|
||||||
|
limiter,
|
||||||
el)
|
el)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
el.AddRecoverable(ctx, clues.Stack(err))
|
el.AddRecoverable(ctx, clues.Stack(err))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Check if we reached the item or size limit while processing this page.
|
||||||
|
// The check after this loop will get us out of the pager.
|
||||||
|
// We don't want to check all limits because it's possible we've reached
|
||||||
|
// the container limit but haven't reached the item limit or really added
|
||||||
|
// items to the last container we found.
|
||||||
|
if limiter.atItemLimit(stats) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
stats.numPages++
|
||||||
|
|
||||||
|
// Stop enumeration early if we've reached the item or page limit. Do this
|
||||||
|
// at the end of the loop so we don't request another page in the
|
||||||
|
// background.
|
||||||
|
//
|
||||||
|
// We don't want to break on just the container limit here because it's
|
||||||
|
// possible that there's more items in the current (final) container that
|
||||||
|
// we're processing. We need to see the next page to determine if we've
|
||||||
|
// reached the end of the container. Note that this doesn't take into
|
||||||
|
// account the number of items in the current container, so it's possible it
|
||||||
|
// will fetch more data when it doesn't really need to.
|
||||||
|
if limiter.atPageLimit(stats) || limiter.atItemLimit(stats) {
|
||||||
|
break
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Always cancel the pager so that even if we exit early from the loop above
|
||||||
|
// we don't deadlock. Cancelling a pager that's already completed is
|
||||||
|
// essentially a noop.
|
||||||
|
pager.Cancel()
|
||||||
|
|
||||||
du, err := pager.Results()
|
du, err := pager.Results()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return du, nil, clues.Stack(err)
|
return du, nil, clues.Stack(err)
|
||||||
@ -853,6 +1001,8 @@ func (c *Collections) processItem(
|
|||||||
topLevelPackages map[string]struct{},
|
topLevelPackages map[string]struct{},
|
||||||
invalidPrevDelta bool,
|
invalidPrevDelta bool,
|
||||||
counter *count.Bus,
|
counter *count.Bus,
|
||||||
|
stats *driveEnumerationStats,
|
||||||
|
limiter *pagerLimiter,
|
||||||
skipper fault.AddSkipper,
|
skipper fault.AddSkipper,
|
||||||
) error {
|
) error {
|
||||||
var (
|
var (
|
||||||
@ -1037,6 +1187,23 @@ func (c *Collections) processItem(
|
|||||||
return clues.NewWC(ctx, "item seen before parent folder").Label(count.ItemBeforeParent)
|
return clues.NewWC(ctx, "item seen before parent folder").Label(count.ItemBeforeParent)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Don't move items if the new collection's already reached it's limit. This
|
||||||
|
// helps ensure we don't get some pathological case where we end up dropping
|
||||||
|
// a bunch of items that got moved.
|
||||||
|
//
|
||||||
|
// We need to check if the collection already contains the item though since
|
||||||
|
// it could be an item update instead of a move.
|
||||||
|
if !collection.ContainsItem(item) &&
|
||||||
|
limiter.atContainerItemsLimit(collection.CountAddedItems()) {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Skip large files that don't fit within the size limit.
|
||||||
|
if limiter.enabled() &&
|
||||||
|
limiter.sizeLimit() < ptr.Val(item.GetSize())+stats.numBytes {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
// This will only kick in if the file was moved multiple times
|
// This will only kick in if the file was moved multiple times
|
||||||
// within a single delta query. We delete the file from the previous
|
// within a single delta query. We delete the file from the previous
|
||||||
// collection so that it doesn't appear in two places.
|
// collection so that it doesn't appear in two places.
|
||||||
@ -1061,6 +1228,8 @@ func (c *Collections) processItem(
|
|||||||
if collection.Add(item) && !alreadyAdded {
|
if collection.Add(item) && !alreadyAdded {
|
||||||
c.NumItems++
|
c.NumItems++
|
||||||
c.NumFiles++
|
c.NumFiles++
|
||||||
|
stats.numAddedFiles++
|
||||||
|
stats.numBytes += ptr.Val(item.GetSize())
|
||||||
}
|
}
|
||||||
|
|
||||||
// Do this after adding the file to the collection so if we fail to add
|
// Do this after adding the file to the collection so if we fail to add
|
||||||
|
|||||||
@ -129,6 +129,17 @@ func driveItem(
|
|||||||
return coreItem(id, name, parentPath, parentID, it)
|
return coreItem(id, name, parentPath, parentID, it)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func driveItemWithSize(
|
||||||
|
id, name, parentPath, parentID string,
|
||||||
|
size int64,
|
||||||
|
it itemType,
|
||||||
|
) models.DriveItemable {
|
||||||
|
res := coreItem(id, name, parentPath, parentID, it)
|
||||||
|
res.SetSize(ptr.To(size))
|
||||||
|
|
||||||
|
return res
|
||||||
|
}
|
||||||
|
|
||||||
func fileItem(
|
func fileItem(
|
||||||
id, name, parentPath, parentID, url string,
|
id, name, parentPath, parentID, url string,
|
||||||
deleted bool,
|
deleted bool,
|
||||||
@ -3553,6 +3564,704 @@ func (suite *CollectionsUnitSuite) TestGet() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TestGet_PreviewLimits checks that the limits set for preview backups in
|
||||||
|
// control.Options.ItemLimits are respected. These tests run a reduced set of
|
||||||
|
// checks that don't examine metadata, collection states, etc. They really just
|
||||||
|
// check the expected items appear.
|
||||||
|
func (suite *CollectionsUnitSuite) TestGet_PreviewLimits() {
|
||||||
|
metadataPath, err := path.BuildMetadata(
|
||||||
|
tenant,
|
||||||
|
user,
|
||||||
|
path.OneDriveService,
|
||||||
|
path.FilesCategory,
|
||||||
|
false)
|
||||||
|
require.NoError(suite.T(), err, "making metadata path", clues.ToCore(err))
|
||||||
|
|
||||||
|
drive1 := models.NewDrive()
|
||||||
|
drive1.SetId(ptr.To(idx(drive, 1)))
|
||||||
|
drive1.SetName(ptr.To(namex(drive, 1)))
|
||||||
|
|
||||||
|
drive2 := models.NewDrive()
|
||||||
|
drive2.SetId(ptr.To(idx(drive, 2)))
|
||||||
|
drive2.SetName(ptr.To(namex(drive, 2)))
|
||||||
|
|
||||||
|
table := []struct {
|
||||||
|
name string
|
||||||
|
limits control.PreviewItemLimits
|
||||||
|
drives []models.Driveable
|
||||||
|
enumerator mock.EnumerateItemsDeltaByDrive
|
||||||
|
// Collection name -> set of item IDs. We can't check item data because
|
||||||
|
// that's not mocked out. Metadata is checked separately.
|
||||||
|
expectedCollections map[string][]string
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "OneDrive SinglePage ExcludeItemsOverMaxSize",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 999,
|
||||||
|
MaxItemsPerContainer: 999,
|
||||||
|
MaxContainers: 999,
|
||||||
|
MaxBytes: 5,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItemWithSize(idx(file, 1), namex(file, 1), parent(1), rootID, 7, isFile),
|
||||||
|
driveItemWithSize(idx(file, 2), namex(file, 2), parent(1), rootID, 1, isFile),
|
||||||
|
driveItemWithSize(idx(file, 3), namex(file, 3), parent(1), rootID, 1, isFile),
|
||||||
|
},
|
||||||
|
}},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 2), idx(file, 3)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive SinglePage SingleFolder ExcludeCombinedItemsOverMaxSize",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 999,
|
||||||
|
MaxItemsPerContainer: 999,
|
||||||
|
MaxContainers: 999,
|
||||||
|
MaxBytes: 3,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItemWithSize(idx(file, 1), namex(file, 1), parent(1), rootID, 1, isFile),
|
||||||
|
driveItemWithSize(idx(file, 2), namex(file, 2), parent(1), rootID, 2, isFile),
|
||||||
|
driveItemWithSize(idx(file, 3), namex(file, 3), parent(1), rootID, 1, isFile),
|
||||||
|
},
|
||||||
|
}},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive SinglePage MultipleFolders ExcludeCombinedItemsOverMaxSize",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 999,
|
||||||
|
MaxItemsPerContainer: 999,
|
||||||
|
MaxContainers: 999,
|
||||||
|
MaxBytes: 3,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItemWithSize(idx(file, 1), namex(file, 1), parent(1), rootID, 1, isFile),
|
||||||
|
driveItemWithSize(idx(folder, 1), namex(folder, 1), parent(1), rootID, 1, isFolder),
|
||||||
|
driveItemWithSize(idx(file, 2), namex(file, 2), parent(1, namex(folder, 1)), idx(folder, 1), 2, isFile),
|
||||||
|
driveItemWithSize(idx(file, 3), namex(file, 3), parent(1, namex(folder, 1)), idx(folder, 1), 1, isFile),
|
||||||
|
},
|
||||||
|
}},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1)},
|
||||||
|
fullPath(1, namex(folder, 1)): {idx(folder, 1), idx(file, 2)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive SinglePage SingleFolder ItemLimit",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 3,
|
||||||
|
MaxItemsPerContainer: 999,
|
||||||
|
MaxContainers: 999,
|
||||||
|
MaxBytes: 999999,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 5), namex(file, 5), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 6), namex(file, 6), parent(1), rootID, isFile),
|
||||||
|
},
|
||||||
|
}},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2), idx(file, 3)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive MultiplePages MultipleFolders ItemLimit WithRepeatedItem",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 3,
|
||||||
|
MaxItemsPerContainer: 999,
|
||||||
|
MaxContainers: 999,
|
||||||
|
MaxBytes: 999999,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1), rootID, isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
// Repeated items shouldn't count against the limit.
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(folder, 1), namex(folder, 1), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
driveItem(idx(file, 5), namex(file, 5), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
driveItem(idx(file, 6), namex(file, 6), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2)},
|
||||||
|
fullPath(1, namex(folder, 1)): {idx(folder, 1), idx(file, 3)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive MultiplePages PageLimit",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 999,
|
||||||
|
MaxItemsPerContainer: 999,
|
||||||
|
MaxContainers: 999,
|
||||||
|
MaxBytes: 999999,
|
||||||
|
MaxPages: 1,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1), rootID, isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 1), namex(folder, 1), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
driveItem(idx(file, 5), namex(file, 5), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
driveItem(idx(file, 6), namex(file, 6), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive MultiplePages PerContainerItemLimit",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 999,
|
||||||
|
MaxItemsPerContainer: 1,
|
||||||
|
MaxContainers: 999,
|
||||||
|
MaxBytes: 999999,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1), rootID, isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 1), namex(folder, 1), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
driveItem(idx(file, 5), namex(file, 5), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
// Root has an additional item. It's hard to fix that in the code
|
||||||
|
// though.
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2)},
|
||||||
|
fullPath(1, namex(folder, 1)): {idx(folder, 1), idx(file, 4)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive MultiplePages PerContainerItemLimit ItemUpdated",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 999,
|
||||||
|
MaxItemsPerContainer: 3,
|
||||||
|
MaxContainers: 999,
|
||||||
|
MaxBytes: 999999,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 0), namex(folder, 0), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1, namex(folder, 0)), idx(folder, 0), isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1, namex(folder, 0)), idx(folder, 0), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 0), namex(folder, 0), parent(1), rootID, isFolder),
|
||||||
|
// Updated item that shouldn't count against the limit a second time.
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1, namex(folder, 0)), idx(folder, 0), isFile),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1, namex(folder, 0)), idx(folder, 0), isFile),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1, namex(folder, 0)), idx(folder, 0), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {},
|
||||||
|
fullPath(1, namex(folder, 0)): {idx(folder, 0), idx(file, 1), idx(file, 2), idx(file, 3)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive MultiplePages PerContainerItemLimit MoveItemBetweenFolders",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 999,
|
||||||
|
MaxItemsPerContainer: 2,
|
||||||
|
MaxContainers: 999,
|
||||||
|
MaxBytes: 999999,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1), rootID, isFile),
|
||||||
|
// Put folder 0 at limit.
|
||||||
|
driveItem(idx(folder, 0), namex(folder, 0), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1, namex(folder, 0)), idx(folder, 0), isFile),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1, namex(folder, 0)), idx(folder, 0), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 0), namex(folder, 0), parent(1), rootID, isFolder),
|
||||||
|
// Try to move item from root to folder 0 which is already at the limit.
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1, namex(folder, 0)), idx(folder, 0), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2)},
|
||||||
|
fullPath(1, namex(folder, 0)): {idx(folder, 0), idx(file, 3), idx(file, 4)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive MultiplePages ContainerLimit LastContainerSplitAcrossPages",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 999,
|
||||||
|
MaxItemsPerContainer: 999,
|
||||||
|
MaxContainers: 2,
|
||||||
|
MaxBytes: 999999,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1), rootID, isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 1), namex(folder, 1), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 1), namex(folder, 1), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 5), namex(file, 5), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2), idx(file, 3)},
|
||||||
|
fullPath(1, namex(folder, 1)): {idx(folder, 1), idx(file, 4), idx(file, 5)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive MultiplePages ContainerLimit NextContainerOnSamePage",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 999,
|
||||||
|
MaxItemsPerContainer: 999,
|
||||||
|
MaxContainers: 2,
|
||||||
|
MaxBytes: 999999,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1), rootID, isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 1), namex(folder, 1), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
driveItem(idx(file, 5), namex(file, 5), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
// This container shouldn't be returned.
|
||||||
|
driveItem(idx(folder, 2), namex(folder, 2), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 7), namex(file, 7), parent(1, namex(folder, 2)), idx(folder, 2), isFile),
|
||||||
|
driveItem(idx(file, 8), namex(file, 8), parent(1, namex(folder, 2)), idx(folder, 2), isFile),
|
||||||
|
driveItem(idx(file, 9), namex(file, 9), parent(1, namex(folder, 2)), idx(folder, 2), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2), idx(file, 3)},
|
||||||
|
fullPath(1, namex(folder, 1)): {idx(folder, 1), idx(file, 4), idx(file, 5)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive MultiplePages ContainerLimit NextContainerOnNextPage",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 999,
|
||||||
|
MaxItemsPerContainer: 999,
|
||||||
|
MaxContainers: 2,
|
||||||
|
MaxBytes: 999999,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1), rootID, isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 1), namex(folder, 1), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
driveItem(idx(file, 5), namex(file, 5), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
// This container shouldn't be returned.
|
||||||
|
driveItem(idx(folder, 2), namex(folder, 2), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 7), namex(file, 7), parent(1, namex(folder, 2)), idx(folder, 2), isFile),
|
||||||
|
driveItem(idx(file, 8), namex(file, 8), parent(1, namex(folder, 2)), idx(folder, 2), isFile),
|
||||||
|
driveItem(idx(file, 9), namex(file, 9), parent(1, namex(folder, 2)), idx(folder, 2), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2), idx(file, 3)},
|
||||||
|
fullPath(1, namex(folder, 1)): {idx(folder, 1), idx(file, 4), idx(file, 5)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "TwoDrives SeparateLimitAccounting",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
Enabled: true,
|
||||||
|
MaxItems: 3,
|
||||||
|
MaxItemsPerContainer: 999,
|
||||||
|
MaxContainers: 999,
|
||||||
|
MaxBytes: 999999,
|
||||||
|
MaxPages: 999,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1, drive2},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 5), namex(file, 5), parent(1), rootID, isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
idx(drive, 2): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(2), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(2), rootID, isFile),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(2), rootID, isFile),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(2), rootID, isFile),
|
||||||
|
driveItem(idx(file, 5), namex(file, 5), parent(2), rootID, isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2), idx(file, 3)},
|
||||||
|
fullPath(2): {idx(file, 1), idx(file, 2), idx(file, 3)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "OneDrive PreviewDisabled MinimumLimitsIgnored",
|
||||||
|
limits: control.PreviewItemLimits{
|
||||||
|
MaxItems: 1,
|
||||||
|
MaxItemsPerContainer: 1,
|
||||||
|
MaxContainers: 1,
|
||||||
|
MaxBytes: 1,
|
||||||
|
MaxPages: 1,
|
||||||
|
},
|
||||||
|
drives: []models.Driveable{drive1},
|
||||||
|
enumerator: mock.EnumerateItemsDeltaByDrive{
|
||||||
|
DrivePagers: map[string]*mock.DriveItemsDeltaPager{
|
||||||
|
idx(drive, 1): {
|
||||||
|
Pages: []mock.NextPage{
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(file, 1), namex(file, 1), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 2), namex(file, 2), parent(1), rootID, isFile),
|
||||||
|
driveItem(idx(file, 3), namex(file, 3), parent(1), rootID, isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 1), namex(folder, 1), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 4), namex(file, 4), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Items: []models.DriveItemable{
|
||||||
|
driveRootItem(rootID), // will be present, not needed
|
||||||
|
driveItem(idx(folder, 1), namex(folder, 1), parent(1), rootID, isFolder),
|
||||||
|
driveItem(idx(file, 5), namex(file, 5), parent(1, namex(folder, 1)), idx(folder, 1), isFile),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
DeltaUpdate: pagers.DeltaUpdate{URL: id(delta)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedCollections: map[string][]string{
|
||||||
|
fullPath(1): {idx(file, 1), idx(file, 2), idx(file, 3)},
|
||||||
|
fullPath(1, namex(folder, 1)): {idx(folder, 1), idx(file, 4), idx(file, 5)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
for _, test := range table {
|
||||||
|
suite.Run(test.name, func() {
|
||||||
|
t := suite.T()
|
||||||
|
|
||||||
|
ctx, flush := tester.NewContext(t)
|
||||||
|
defer flush()
|
||||||
|
|
||||||
|
mockDrivePager := &apiMock.Pager[models.Driveable]{
|
||||||
|
ToReturn: []apiMock.PagerResult[models.Driveable]{
|
||||||
|
{Values: test.drives},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
mbh := mock.DefaultOneDriveBH(user)
|
||||||
|
mbh.DrivePagerV = mockDrivePager
|
||||||
|
mbh.DriveItemEnumeration = test.enumerator
|
||||||
|
|
||||||
|
opts := control.DefaultOptions()
|
||||||
|
opts.PreviewLimits = test.limits
|
||||||
|
|
||||||
|
c := NewCollections(
|
||||||
|
mbh,
|
||||||
|
tenant,
|
||||||
|
idname.NewProvider(user, user),
|
||||||
|
func(*support.ControllerOperationStatus) {},
|
||||||
|
opts,
|
||||||
|
count.New())
|
||||||
|
|
||||||
|
errs := fault.New(true)
|
||||||
|
|
||||||
|
delList := prefixmatcher.NewStringSetBuilder()
|
||||||
|
|
||||||
|
cols, canUsePreviousBackup, err := c.Get(ctx, nil, delList, errs)
|
||||||
|
require.NoError(t, err, clues.ToCore(err))
|
||||||
|
|
||||||
|
assert.True(t, canUsePreviousBackup, "can use previous backup")
|
||||||
|
assert.Empty(t, errs.Skipped())
|
||||||
|
|
||||||
|
collPaths := []string{}
|
||||||
|
|
||||||
|
for _, baseCol := range cols {
|
||||||
|
// There shouldn't be any deleted collections.
|
||||||
|
if !assert.NotEqual(
|
||||||
|
t,
|
||||||
|
data.DeletedState,
|
||||||
|
baseCol.State(),
|
||||||
|
"collection marked deleted") {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
folderPath := baseCol.FullPath().String()
|
||||||
|
|
||||||
|
if folderPath == metadataPath.String() {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
collPaths = append(collPaths, folderPath)
|
||||||
|
|
||||||
|
// TODO: We should really be getting items in the collection
|
||||||
|
// via the Items() channel. The lack of that makes this check a bit more
|
||||||
|
// bittle since internal details can change. The wiring to support
|
||||||
|
// mocked GetItems is available. We just haven't plugged it in yet.
|
||||||
|
col, ok := baseCol.(*Collection)
|
||||||
|
require.True(t, ok, "getting onedrive.Collection handle")
|
||||||
|
|
||||||
|
itemIDs := make([]string, 0, len(col.driveItems))
|
||||||
|
|
||||||
|
for id := range col.driveItems {
|
||||||
|
itemIDs = append(itemIDs, id)
|
||||||
|
}
|
||||||
|
|
||||||
|
assert.ElementsMatchf(
|
||||||
|
t,
|
||||||
|
test.expectedCollections[folderPath],
|
||||||
|
itemIDs,
|
||||||
|
"expected elements to match in collection with path %q",
|
||||||
|
folderPath)
|
||||||
|
}
|
||||||
|
|
||||||
|
assert.ElementsMatch(
|
||||||
|
t,
|
||||||
|
maps.Keys(test.expectedCollections),
|
||||||
|
collPaths,
|
||||||
|
"collection paths")
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func (suite *CollectionsUnitSuite) TestAddURLCacheToDriveCollections() {
|
func (suite *CollectionsUnitSuite) TestAddURLCacheToDriveCollections() {
|
||||||
drive1 := models.NewDrive()
|
drive1 := models.NewDrive()
|
||||||
drive1.SetId(ptr.To(idx(drive, 1)))
|
drive1.SetId(ptr.To(idx(drive, 1)))
|
||||||
|
|||||||
@ -331,6 +331,8 @@ func (edi *DriveItemsDeltaPager) NextPage() ([]models.DriveItemable, bool, bool)
|
|||||||
return np.Items, np.Reset, false
|
return np.Items, np.Reset, false
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (edi *DriveItemsDeltaPager) Cancel() {}
|
||||||
|
|
||||||
func (edi *DriveItemsDeltaPager) Results() (pagers.DeltaUpdate, error) {
|
func (edi *DriveItemsDeltaPager) Results() (pagers.DeltaUpdate, error) {
|
||||||
return edi.DeltaUpdate, edi.Err
|
return edi.DeltaUpdate, edi.Err
|
||||||
}
|
}
|
||||||
|
|||||||
@ -43,7 +43,8 @@ type PreviewItemLimits struct {
|
|||||||
MaxItems int
|
MaxItems int
|
||||||
MaxItemsPerContainer int
|
MaxItemsPerContainer int
|
||||||
MaxContainers int
|
MaxContainers int
|
||||||
MaxBytes int
|
MaxBytes int64
|
||||||
|
MaxPages int
|
||||||
Enabled bool
|
Enabled bool
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@ -51,6 +51,7 @@ type nextPage[T any] struct {
|
|||||||
type NextPageResulter[T any] interface {
|
type NextPageResulter[T any] interface {
|
||||||
NextPager[T]
|
NextPager[T]
|
||||||
|
|
||||||
|
Cancel()
|
||||||
Results() (DeltaUpdate, error)
|
Results() (DeltaUpdate, error)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user