Ben Gross

Simulating Searches

This project’s goal was to figure out if someone can perform a good web search. Collaborating with content specialists, psychometricians, and assessment developers, I lead the design and UI/UX to create a simulated search engine.

Focus

After initial pitches and consideration between research, content, and design teams, the focus settled on two specific skills used when performing a web search:

  1. The initial search engine query.
  2. Choosing the best site to explore from the standardized search engine results list (comprised of page titles, URLs, and summaries).

We would create a simulated search engine to learn how good people are at these two skills. Our client, the National Center for Educational Statistics (NCES), accepted this design and we moved forward!

Design

I lead the design by proposing a collection of systems to improve the simulative qualities of our search engine. These systems were also chosen based on an earlier market analysis, which found that comparable simulated search engines were particularly weak in fixing user input errors and providing dynamic search results (some engines would spit out the same canned results as long as the users query included one key term).

Weight System

The weighting system was designed to deliver relevant results based on the unique term(s) searched. Content specialists can assign a 0 to 1 weight value for each term and source, ensuring that search results are relevant to the terms searched.

For example, let’s say the weighting was set up as follows

Term 1*Term 2Term 3Term 4
Source 110.30.750.9
Source 20.510.330.75
Source 30.650.510.33
Source 40.30.750.851
* "Term" can be any text, and multiple forms of a Term can also be inserted. For example, if your Term was "meme", you could enter "meme, memes, dank memes", and the algorithm would know that those 3 comma-separated items would all count as getting that that Term.

I then perform a search for Term 2 and Term 3.

Image of a search engine with term 2 and term 3 typed in.

The total weights for those two Terms would be calculated

Term 2Term 3Total
Source 10.30 0.751.05
Source 21.00 0.331.33
Source 30.501.001.50
Source 40.75 0.851.60

and the user would be displayed the source results in order based on highest-to-lowest total weight values.

Image of a search engine with term 2 and term 3 typed in.

While actual search engines like Google are much smarter and multifactorial, it would have been impossible to recreate or use these engines due to their proprietary nature and the constraints of the project (such as needing a walled-garden system that limited the number of sources and Terms so it would be easier to quantify and automate measurement). By implementing this weighting system, we could simulate the core functionality of search engines—matching search queries to sources based on weighted relevance.

In addition to the weighting system, I designed error correction systems that included spell-checking and whitelisting, a suite of front-end parameters to allow for customization, data capture logic, and the overall UI/UX.

Results

The final deliverable, a new form of assessment focused on measuring the first steps of a good web search, was completed on-time and accepted by our client. I wrote an end-to-end design document that covered all aspects of this tool, including

Image of a search engine with term 2 and term 3 typed in.

The Simulated Search tool went through user testing, cog labs, a large scale pilot, and is now a part of the National Assessment of Educational Progress. I wrote follow-up documentation covering all aspects of setting up the weighting system and changing various unique parameters, ensuring a clean hand-off.