Differences
This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision | ||
litespeed_wiki:cache:lscjoomla:crawler [2020/01/22 22:27] Joshua Reynolds [Run crawler once manually] Changed to Title case |
litespeed_wiki:cache:lscjoomla:crawler [2020/07/14 17:09] Lisa Clarke Redirect to new Documentation Site |
||
---|---|---|---|
Line 1: | Line 1: | ||
- | ====== How to Set Up Crawler Cron Job for LSCJoomla ====== | + | ~~REDIRECT>https://docs.litespeedtech.com/lscache/lscjoomla/settings/~~ |
- | + | ||
- | ===== Crawler Function===== | + | |
- | The latest LSCJoomla provides a crawler function at **Advanced -> Web interface to rebuild all LiteSpeed Cache**: | + | |
- | + | ||
- | {{ :litespeed_wiki:cache:lscjoomla:lscache-joomla-crawler-1.png?800 |}} | + | |
- | + | ||
- | Web interface to rebuild all LiteSpeed Cache: https://yourdomain.com/index.php?option=com_lscache&recache=5dcfad26ad657984064c731ee774a57f | + | |
- | + | ||
- | ===== Prerequisite for Crawler ===== | + | |
- | To run the above crawler, your ''curl_exec'' cannot be disabled in your ''php.ini'' settings. Also you will need to [[litespeed_wiki:cache:lscwp:configuration:enabling_the_crawler|enable crawler feature on LSWS]] which is most likely disabled on shared hosting server by default. | + | |
- | + | ||
- | If ''curl_exec'' was disabled from ''php.ini'', you may see similar error like: | + | |
- | Warning: curl_exec() has been disabled for security reasons in /home/joomla/public_html/plugins/system/lscache/lscache.php on line 1564 | + | |
- | + | ||
- | If crawler feature is not enabled on LSWS, you may see an error like the following: | + | |
- | {{ :litespeed_wiki:cache:lscjoomla:lscache-joomla-crawler-2.png?400 |}} | + | |
- | ===== Run Crawler Once Manually ===== | + | |
- | You can click ''Web interface to rebuild all LiteSpeed Cache link'' to run the crawler manually. It will only run once until completion. | + | |
- | + | ||
- | ===== Run Crawler as Cron Job ===== | + | |
- | + | ||
- | To setup crawler as a cron job, you can simply use wget to retrieve the cachelink URL (please replace the URL with your own value): | + | |
- | wget https://yourdomain.com/index.php?option=com_lscache&recache=5dcfad26ad657984064c731ee774a57f | + | |
- | + | ||
- | For example, setup the above cron job daily to run at 0:00, in crontab -e user: | + | |
- | 0 0 * * * wget wget https://yourdomain.com/index.php?option=com_lscache&recache=5dcfad26ad657984064c731ee774a57 | + | |
- | + | ||
- | ===== How to Keep the Page Cached More Often ===== | + | |
- | The ''Homepage Public Cache TTL (minutes)'' defines how long before the cache will be expired. The default settings for it is ''2000'' minutes, which is around 33 hours. | + | |
- | + | ||
- | To keep the page cached most of the time, you can either set a longer Time to Live (TTL) to make the cache stay longer, or run the crawler periodically before the TTL is up and the cache expires. For example, for 2000 minutes TTL, running the crawler cron job daily will keep the pages cached all the time. | + |