The technology allows researchers to ask a blend of survey-style questions followed by open-ended questions for clarity and depth at the scale of hundreds of participants. Qualitative research methods like in-depth interviews are typically conducted one-on-one or in focus groups of a handful of participants. Remesh's response scale paired with a depth of detail creates a unique blend of qualitative and quantitative analysis opportunities.
I was hired to understand why it was so difficult for customers to analyze a Remesh dataset after completing a study, and to develop a solution that supported them in their analysis and reporting.
Qualitative researchers have a lot of data to make sense of after running a study with Remesh. Their next phase of work is to sort through this data to uncover relevant insights to report to their stakeholders.
Instead of being able to spend their time understanding the results of their research, our customer analysts and internal researchers were spending countless hours reorganizing their Remesh data output and creating basic data visualizations before even starting their analysis.
Researchers were taking their data off the Remesh platform and forgoing the analysis tools provided since they were unhelpful.
Sales and Customer Support teams actively avoided pitching Remesh’s analysis features and training new customers on them further indicating that the analysis capabilities were missing the mark.
Qualitative research begins with a research objective. By stating an objective, researchers outline the parameters of what they want to understand from a target population.
Remesh excelled at helping researchers plan their research and gather data, but struggled when it came to analyzing the results. While we considered the entire research workflow, we focused our efforts here.
In addition to our user objective, we were also in the middle of a strategic shift from operating as a service-based business to a digital technology one. We framed this business objective to align our project: Develop a customer self-serve analysis solution that supports a service-to-technology transition.
From our discovery research we knew that our existing analysis pages omitted some of the data from a research project which led to an unusable experience.
Our first step was to provide a complete view of results from the research. We augmented that baseline with methods to understand those results: data visualizations and comparison methods. Finally we considered the analysis and reporting workflows after leaving the Remesh platform.
MVP Release
The new data-complete and visual analysis product showed healthy customer engagement from launch and steady usage in following months.
Notable was new engagement from executive stakeholders and project supporters who were informed but not in charge of executing the research. Our visual and intuitive solution not only supported research analysts, but also provided much desired visibility to other team members.
With new confidence, our customer-facing teams highlighted and provided training on the new product area to customers.
Analysis methods vary widely even among researchers from the same team. However there is a general framework shared by researchers from different organizations, teams, and use cases.
We reviewed our prior and current analysis offerings to understand what had been attempted and why they had failed. We noted ideas that still seemed relevant to pursue.
We prioritized our work based on the analysis workflow and also by understanding which features brought the most immediate value to researchers.
We pushed new features as they were ready to internal researchers for live testing. We then batched them into complete offerings for formal customer releases.
Guidelines
We shared ideas with our internal researchers before validating design concepts with external customers in multiple rounds of feedback.
Prioritizing Features: We asked customers to rank a list of possible features to better understand their potential value. This informed how features were presented and which ones made it into our inital release.
Confirming Value: On average customers interviewed rated our conceptual designs as medium-to-high value (3 on a 4-point scale). Having a way to visually consume results helped researchers clarify the story of the research and orient towards areas to spend their time doing deeper analysis.
Future Opportunity: We also gathered ideas for additional data views, filtering capabilities, and other settings that would be useful in future versions. We would go on to make 3 additional major releases after our initial MVP.
This alone made our most recent analysis offering miss "viable" in the definition of MVP.
Users care about what the data is telling thPm and how Uhey might craft that into a relatable story.CThe data is elemental to Vhe design.
The data must appear familiar and usable in the various forms and views it may take.
The data is often given shape and meaning by its metadata which should also be considered when presenting the data.
Having done the appropriate research and testing we along with our leaders were confident that we were building a truly useful product. We also set ourselves up to quickly make innovative iterations. View the next project
My Role
Tasks
Team
Duration