Skip to main content
Help the Python Software Foundation raise $60,000 USD by December 31st!  Building the PSF Q4 Fundraiser

Ignore all indexing operations

Project description


Goal and usage

Add collective.noindexing to the eggs in your buildout (and to zcml on Plone 3.2 or earlier). This makes two browser views available on the Plone Site root: @@collective-noindexing-apply and @@collective-noindexing-unapply. The first applies some patches and the second undoes the patching. Both can be called multiple times safely.

This patches some catalog methods so no indexing, reindexing or unindexing is done at all. The idea is that you use this package so you can quickly move a big part of your Plone Site to a different folder without having to worry about indexing. It really makes moving a lot faster. You do the indexing later, probably by doing a catalog clear and rebuild; you have a bit more control there about subtransactions, to help avoid a MemoryError or [Errno 24] Too many open files. A script to run the catalog clear and rebuild with some intermediate commits can help here for large sites; see for example

Note: the patches do not apply to ATBTreeFolders like the standard Members, events and news Large Plone Folders. This is not deliberate, so it might change in the future, but for now I do not mind.


  • Go to the archetype_tool object in the ZMI, and then to the Catalogs tab. Switching off portal_catalog in all the types there should have basically the same effect.

  • Add Products.QueueCatalog and Products.PloneQueueCatalog to the eggs of your buildout. In the portal_quickinstaller install PloneQueueCatalog. This renames the portal_catalog to portal_catalog_real and creates a ZCatalog Queue with the id portal_catalog. The standard settings worked fine for me. You now do that large move. In the fresh portal_catalog you go to the Queue tab. It should say you have lots of items in the queue, in my case around 12,000. Clicking the ‘Process Queue’ button will by default process just twenty items of that queue. You can increase that number. This is an easy way of avoiding MemoryErrors during indexing, as the total number of objects reindexed in one go will be as low as you want.

    Note that I tried this but ran into problems as this website had the portal_catalog in a separate CatalogData.fs, which worked fine until I restarted the zeoclient. With some tinkering it should work, but I did not want to bother with that. We can revisit that when indexing becomes a problem all the time instead of just once for a clear and rebuild.


Maurits van Rees


1.1 (2011-01-04)

  • Moved most logging to debug level as it quickly becomes noise in situations where you most need this package. [maurits]

1.0 (2010-04-09)

  • Initial release

Project details

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Files for collective.noindexing, version 1.1
Filename, size File type Python version Upload date Hashes
Filename, size (18.7 kB) File type Source Python version None Upload date Hashes View

Supported by

Pingdom Pingdom Monitoring Google Google Object Storage and Download Analytics Sentry Sentry Error logging AWS AWS Cloud computing DataDog DataDog Monitoring Fastly Fastly CDN DigiCert DigiCert EV certificate StatusPage StatusPage Status page