Algolia Crawler Management

To deliver great search experiences, aggregating multiple data sources is crucial. Our challenge was to create an Algolia crawler capable of swiftly combining public and private data, to help showcasing the platform’s capabilities.

Context

The Algolia Crawler serves as a versatile tool for efficiently extracting content from multiple websites, formatting it as needed, and seamlessly uploading it to Algolia. Its functionality includes the rapid aggregation of distributed content, ensuring that your data is consolidated effectively.

Challenges

The obstacle I encountered revolved around transforming this internal tool into a pivotal asset for driving our product-led growth strategy. This entailed smoothing out any rough edges, from streamlining the customer onboarding and configuration process to ensuring that subsequent users could easily navigate existing configurations even without prior setup experience.

A quick summary of the crawler instance’s health status.
A quick summary of the crawler instance’s health status.

Inline guidance

Offering comprehensive guidance was essential, avoiding assumptions about customer knowledge and instead providing context for every aspect of the product. This tool is typically a ‘set and forget’ experience, rather than a frequently used tool in a developer’s toolkit, so ensuring clarity and support throughout the user journey is paramount.

Path Explore enables users to understand why certain content was skipped and resolve their issues.

Onboarding

Establishing a straightforward method for new customers to authenticate the domains they want to crawl was crucial. Demonstrating value promptly to users is paramount, as it allows us to then offer additional recommendations on how to deploy this tool effectively to meet their specific needs.