Christian Wiklund, unitQ | AWS Startup Showcase S2 E3
(upbeat music) >> Hello, everyone. Welcome to the theCUBE's presentation of the AWS Startup Showcase. The theme, this showcase is MarTech, the emerging cloud scale customer experiences. Season two of episode three, the ongoing series covering the startups, the hot startups, talking about analytics, data, all things MarTech. I'm your host, John Furrier, here joined by Christian Wiklund, founder and CEO of unitQ here, talk about harnessing the power of user feedback to empower marketing. Thanks for joining us today. >> Thank you so much, John. Happy to be here. >> In these new shifts in the market, when you got cloud scale, open source software is completely changing the software business. We know that. There's no longer a software category. It's cloud, integration, data. That's the new normal. That's the new category, right? So as companies are building their products, and want to do a good job, it used to be, you send out surveys, you try to get the product market fit. And if you were smart, you got it right the third, fourth, 10th time. If you were lucky, like some companies, you get it right the first time. But the holy grail is to get it right the first time. And now, this new data acquisition opportunities that you guys in the middle of that can tap customers or prospects or end users to get data before things are shipped, or built, or to iterate on products. This is the customer feedback loop or data, voice of the customer journey. It's a gold mine. And it's you guys, it's your secret weapon. Take us through what this is about now. I mean, it's not just surveys. What's different? >> So yeah, if we go back to why are we building unitQ? Which is we want to build a quality company. Which is basically, how do we enable other companies to build higher quality experiences by tapping into all of the existing data assets? And the one we are in particularly excited about is user feedback. So me and my co-founder, Nik, and we're doing now the second company together. We spent 14 years. So we're like an old married couple. We accept each other, and we don't fight anymore, which is great. We did a consumer company called Skout, which was sold five years ago. And Skout was kind of early in the whole mobile first. I guess, we were actually mobile first company. And when we launched this one, we immediately had the entire world as our marketplace, right? Like any modern company. We launch a product, we have support for many languages. It's multiple platforms. We have Android, iOS, web, big screens, small screens, and that brings some complexities as it relates to staying on top of the quality of the experience because how do I test everything? >> John: Yeah. >> Pre-production. How do I make sure that our Polish Android users are having a good day? And we found at Skout, personally, like I could discover million dollar bugs by just drinking coffee and reading feedback. And we're like, "Well, there's got to be a better way to actually harness the end user feedback. That they are leaving in so many different places." So, you know what, what unitQ does is that we basically aggregate all different sources of user feedback, which can be app store reviews, Reddit posts, Tweets, comments on your Facebook ads. It can be better Business Bureau Reports. We don't like to get to many of those, of course. But really, anything on the public domain that mentions or refers to your product, we want to ingest that data in this machine, and then all the private sources. So you probably have a support system deployed, a Zendesk, or an Intercom. You might have a chatbot like an Ada, or and so forth. And your end user is going to leave a lot of feedback there as well. So we take all of these channels, plug it into the machine, and then we're able to take this qualitative data. Which and I actually think like, when an end user leaves a piece of feedback, it's an act of love. They took time out of the day, and they're going to tell you, "Hey, this is not working for me," or, "Hey, this is working for me," and they're giving you feedback. But how do we package these very messy, multi-channel, multiple languages, all over the place data? How can we distill it into something that's quantifiable? Because I want to be able to monitor these different signals. So I want to turn user feedback into time series. 'Cause with time series, I can now treat this the same way as Datadog treats machine logs. I want to be able to see anomalies, and I want to know when something breaks. So what we do here is that we break down your data in something called quality monitors, which is basically machine learning models that can aggregate the same type of feedback data in this very fine grained and discrete buckets. And we deploy up to a thousand of these quality monitors per product. And so we can get down to the root cause. Let's say, passive reset link is not working. And it's in that root cause, the granularity that we see that companies take action on the data. And I think historically, there has been like the workflow between marketing and support, and engineering and product has been a bit broken. They've been siloed from a data perspective. They've been siloed from a workflow perspective, where support will get a bunch of tickets around some issue in production. And they're trained to copy and paste some examples, and throw it over the wall, file a Jira ticket, and then they don't know what happens. So what we see with the platform we built is that these teams are able to rally around the single source of troop or like, yes, passive recent link seems to have broken. This is not a user error. It's not a fix later, or I can't reproduce. We're looking at the data, and yes, something broke. We need to fix it. >> I mean, the data silos a huge issue. Different channels, omnichannel. Now, there's more and more channels that people are talking in. So that's huge. I want to get to that. But also, you said that it's a labor of love to leave a comment or a feedback. But also, I remember from my early days, breaking into the business at IBM and Hewlett-Packard, where I worked. People who complain are the most loyal customers, if you service them. So it's complaints. >> Christian: Yeah. >> It's leaving feedback. And then, there's also reading between the lines with app errors or potentially what's going on under the covers that people may not be complaining about, but they're leaving maybe gesture data or some sort of digital trail. >> Yeah. >> So this is the confluence of the multitude of data sources. And then you got the siloed locations. >> Siloed locations. >> It's complicated problem. >> It's very complicated. And when you think about, so I started, I came to Bay Area in 2005. My dream was to be a quant analyst on Wall Street, and I ended up in QA at VMware. So I started at VMware in Palo Alto, and didn't have a driver's license. I had to bike around, which was super exciting. And we were shipping box software, right? This was literally a box with a DVD that's been burned, and if that DVD had bugs in it, guess what it'll be very costly to then have to ship out, and everything. So I love the VMware example because the test cycles were long and brutal. It was like a six month deal to get through all these different cases, and they couldn't be any bugs. But then as the industry moved into the cloud, CI/CD, ship at will. And if you look at the modern company, you'll have at least 20 plus integrations into your product. Analytics, add that's the case, authentication, that's the case, and so forth. And these integrations, they morph, and they break. And you have connectivity issues. Is your product working as well on Caltrain, when you're driving up and down, versus wifi? You have language specific bugs that happen. Android is also quite a fragmented market. The binary may not perform as well on that device, or is that device. So how do we make sure that we test everything before we ship? The answer is, we can't. There's no company today that can test everything before the ship. In particular, in consumer. And the epiphany we had at our last company, Skout, was that, "Hey, wait a minute. The end user, they're testing every configuration." They're sitting on the latest device, the oldest device. They're sitting on Japanese language, on Swedish language. >> John: Yeah. >> They are in different code paths because our product executed differently, depending on if you were a paid user, or a freemium user, or if you were certain demographical data. There's so many ways that you would have to test. And PagerDuty actually had a study they came out with recently, where they said 51% of all end user impacting issues are discovered first by the end user, when they serve with a bunch of customers. And again, like the cool part is, they will tell you what's not working. So now, how do we tap into that? >> Yeah. >> So what I'd like to say is, "Hey, your end user is like your ultimate test group, and unitQ is the layer that converts them into your extended test team." Now, the signals they're producing, it's making it through to the different teams in the organization. >> I think that's the script that you guys are flipping. If I could just interject. Because to me, when I hear you talking, I hear, "Okay, you're letting the customers be an input into the product development process." And there's many different pipelines of that development. And that could be whether you're iterating, or geography, releases, all kinds of different pipelines to get to the market. But in the old days, it was like just customer satisfaction. Complain in a call center. >> Christian: Yeah. >> Or I'm complaining, how do I get support? Nothing made itself into the product improvement, except for slow moving, waterfall-based processes. And then, maybe six months later, a small tweak could be improved. >> Yes. >> Here, you're taking direct input from collective intelligence. Okay. >> Is that have input and on timing is very important here, right? So how do you know if the product is working as it should in all these different flavors and configurations right now? How do you know if it's working well? And how do you know if you're improving or not improving over time? And I think the industry, what can we look at, as far as when it relates to quality? So I can look at star ratings, right? So what's the star rating in the app store? Well, star ratings, that's an average over time. So that's something that you may have a lot of issues in production today, and you're going to get dinged on star ratings over the next few months. And then, it brings down the score. NPS is another one, where we're not going to run NPS surveys every day. We're going to run it once a quarter, maybe once a month, if we're really, really aggressive. That's also a snapshot in time. And we need to have the finger on the pulse of product quality today. I need to know if this release is good or not good. I need to know if anything broke. And I think that real time aspect, what we see as stuff sort of bubbles up the stack, and not into production, we see up to a 50% reduction in time to fix these end user impacting issues. And I think, we also need to appreciate when someone takes time out of the day to write an app review, or email support, or write that Reddit post, it's pretty serious. It's not going to be like, "Oh, I don't like the shade of blue on this button." It's going to be something like, "I got double billed," or "Hey, someone took over my account," or, "I can't reset my password anymore. The CAPTCHA, I'm solving it, but I can't get through to the next phase." And we see a lot of these trajectory impacting bugs and quality issues in these work, these flows in the product that you're not testing every day. So if you work at Snapchat, your employees probably going to use Snapchat every day. Are they going to sign up every day? No. Are they going to do passive reset every day? No. And these things are very hard to instrument, lower in the stack. >> Yeah, I think this is, and again, back to these big problems. It's smoke before fire, and you're essentially seeing it early with your process. Can you give an example of how this new focus or new mindset of user feedback data can help customers increase their experience? Can you give some examples, 'cause folks watching and be like, "Okay, I love this value. Sell me on this idea, I'm sold. Okay, I want to tap into my prospects, and my customers, my end users to help me improve my product." 'Cause again, we can measure everything now with data. >> Yeah. We can measure everything. we can even measure quality these days. So when we started this company, I went out to talk to a bunch of friends, who are entrepreneurs, and VCs, and board members, and I asked them this very simple question. So in your board meetings, or on all hands, how do you talk about quality of the product? Do you have a metric? And everyone said, no. Okay. So are you data driven company? Yes, we're very data driven. >> John: Yeah. Go data driven. >> But you're not really sure if quality, how do you compare against competition? Are you doing as good as them, worse, better? Are you improving over time, and how do you measure it? And they're like, "Well, it's kind of like a blind spot of the company." And then you ask, "Well, do you think quality of experience is important?" And they say, "Yeah." "Well, why?" "Well, top of fund and growth. Higher quality products going to spread faster organically, we're going to make better store ratings. We're going to have the storefronts going to look better." And of course, more importantly, they said the different conversion cycles in the product box itself. That if you have bugs and friction, or an interface that's hard to use, then the inputs, the signups, it's not going to convert as well. So you're going to get dinged on retention, engagement, conversion to paid, and so forth. And that's what we've seen with the companies we work with. It is that poor quality acts as a filter function for the entire business, if you're a product led company. So if you think about product led company, where the product is really the centerpiece. And if it performs really, really well, then it allows you to hire more engineers, you can spend more on marketing. Everything is fed by this product at them in the middle, and then quality can make that thing perform worse or better. And we developed a metric actually called the unitQ Score. So if you go to our website, unitq.com, we have indexed the 5,000 largest apps in the world. And we're able to then, on a daily basis, update the score. Because the score is not something you do once a month or once a quarter. It's something that changes continuously. So now, you can get a score between zero and 100. If you get the score 100, that means that our AI doesn't find any quality issues reported in that data set. And if your score is 90, that means that 10% will be a quality issue. So now you can do a lot of fun stuff. You can start benchmarking against competition. So you can see, "Well, I'm Spotify. How do I rank against Deezer, or SoundCloud, or others in my space?" And what we've seen is that as the score goes up, we see this real big impact on KPI, such as conversion, organic growth, retention, ultimately, revenue, right? And so that was very satisfying for us, when we launched it. quality actually still really, really matters. >> Yeah. >> And I think we all agree at test, but how do we make a science out of it? And that's so what we've done. And when we were very lucky early on to get some incredible brands that we work with. So Pinterest is a big customer of ours. We have Spotify. We just signed new bank, Chime. So like we even signed BetterHelp recently, and the world's largest Bible app. So when you look at the types of businesses that we work with, it's truly a universal, very broad field, where if you have a digital exhaust or feedback, I can guarantee you, there are insights in there that are being neglected. >> John: So Chris, I got to. >> So these manual workflows. Yeah, please go ahead. >> I got to ask you, because this is a really great example of this new shift, right? The new shift of leveraging data, flipping the script. Everything's flipping the script here, right? >> Yeah. >> So you're talking about, what the value proposition is? "Hey, board example's a good one. How do you measure quality? There's no KPI for that." So it's almost category creating in its own way. In that, this net new things, it's okay to be new, it's just new. So the question is, if I'm a customer, I buy it. I can see my product teams engaging with this. I can see how it can changes my marketing, and customer experience teams. How do I operationalize this? Okay. So what do I do? So do I reorganize my marketing team? So take me through the impact to the customer that you're seeing. What are they resonating towards? Obviously, getting that data is key, and that's holy gray, we all know that. But what do I got to do to change my environment? What's my operationalization piece of it? >> Yeah, and that's one of the coolest parts I think, and that is, let's start with your user base. We're not going to ask your users to ask your users to do something differently. They're already producing this data every day. They are tweeting about it. They're putting in app produce. They're emailing support. They're engaging with your support chatbot. They're already doing it. And every day that you're not leveraging that data, the data that was produced today is less valuable tomorrow. And in 30 days, I would argue, it's probably useless. >> John: Unless it's same guy commenting. >> Yeah. (Christian and John laughing) The first, we need to make everyone understand. Well, yeah, the data is there, and we don't need to do anything differently with the end user. And then, what we do is we ask the customer to tell us, "Where should we listen in the public domain? So do you want the Reddit post, the Trustpilot? What channels should we listen to?" And then, our machine basically starts ingesting that data. So we have integration with all these different sites. And then, to get access to private data, it'll be, if you're on Zendesk, you have to issue a Zendesk token, right? So you don't need any engineering hours, except your IT person will have to grant us access to the data source. And then, when we go live. We basically build up this taxonomy with the customers. So we don't we don't want to try and impose our view of the world, of how do you describe the product with these buckets, these quality monitors? So we work with the company to then build out this taxonomy. So it's almost like a bespoke solution that we can bootstrap with previous work we've done, where you don't have these very, very fine buckets of where stuff could go wrong. And then what we do is there are different ways to hook this into the workflow. So one is just to use our products. It's a SaaS product as anything else. So you log in, and you can then get this overview of how is quality trending in different markets, on different platforms, different languages, and what is impacting them? What is driving this unitQ Score that's not good enough? And all of these different signals, we can then hook into Jira for instance. We have a Jira integration. We have a PagerDuty integration. We can wake up engineers if certain things break. We also tag tickets in your support system, which is actually quite cool. Where, let's say, you have 200 people, who wrote into support, saying, "I got double billed on Android." It turns out, there are some bugs that double billed them. Well, now we can tag all of these users in Zendesk, and then the support team can then reach out to that segment of users and say, "Hey, we heard that you had this bug with double billing. We're so sorry. We're working on it." And then when we push fix, we can then email the same group again, and maybe give them a little gift card or something, for the thank you. So you can have, even big companies can have that small company experience. So, so it's groups that use us, like at Pinterest, we have 800 accounts. So it's really through marketing has vested interest because they want to know what is impacting the end user. Because brand and product, the lines are basically gone, right? >> John: Yeah. >> So if the product is not working, then my spend into this machine is going to be less efficient. The reputation of our company is going to be worse. And the challenge for marketers before unitQ was, how do I engage with engineering and product? I'm dealing with anecdotal data, and my own experience of like, "Hey, I've never seen these type of complaints before. I think something is going on." >> John: Yeah. >> And then engineering will be like, "Ah, you know, well, I have 5,000 bugs in Jira. Why does this one matter? When did it start? Is this a growing issue?" >> John: You have to replicate the problem, right? >> Replicate it then. >> And then it goes on and on and on. >> And a lot of times, reproducing bugs, it's really hard because it works on my device. Because you don't sit on that device that it happened on. >> Yup. >> So now, when marketing can come with indisputable data, and say, "Hey, something broke here." And we see the same with support. Product engineering, of course, for them, we talk about, "Hey, listen, you you've invested a lot in observability of your stack, haven't you?" "Yeah, yeah, yeah." "So you have a Datadog in the bottom?" "Absolutely." "And you have an APP D on the client?" "Absolutely." "Well, what about the last mile? How the product manifests itself? Shouldn't you monitor that as well using machines?" They're like, "Yeah, that'd be really cool." (John laughs) And we see this. There's no way to instrument everything, lowering the stack to capture these bugs that leak out. So it resonates really well there. And even for the engineers who's going to fix it. >> Yeah. >> I call it like empathy data. >> Yup. >> Where I get assigned a bug to fix. Well, now, I can read all the feedback. I can actually see, and I can see the feedback coming in. >> Yeah. >> Oh, there's users out there, suffering from this bug. And then when I fix it and I deploy the fix, and I see the trend go down to zero, and then I can celebrate it. So that whole feedback loop is (indistinct). >> And that's real time. It's usually missed too. This is the power of user feedback. You guys got a great product, unitQ. Great to have you on. Founder and CEO, Christian Wiklund. Thanks for coming on and sharing, and showcase. >> Thank you, John. For the last 30 seconds, the minute we have left, put a plug in for the company. What are you guys looking for? Give a quick pitch for the company, real quick, for the folks out there. Looking for more people, funding status, number of employees. Give a quick plug. >> Yes. So we raised our A Round from Google, and then we raised our B from Excel that we closed late last year. So we're not raising money. We are hiring across go-to-markets, engineering. And we love to work with people, who are passionate about quality and data. We're always, of course, looking for customers, who are interested in upping their game. And hey, listen, competing with features is really hard because you can copy features very quickly. Competing with content. Content is commodity. You're going to get the same movies more or less on all these different providers. And competing on price, we're not willing to do. You're going to pay 10 bucks a month for music. So how do you compete today? And if your competitor has a better fine tuned piano than your competitor will have better efficiencies, and they're going to retain customers and users better. And you don't want to lose on quality because it is actually a deterministic and fixable problem. So yeah, come talk to us if you want to up the game there. >> Great stuff. The iteration lean startup model, some say took craft out of building the product. But this is now bringing the craftsmanship into the product cycle, when you can get that data from customers and users. >> Yeah. >> Who are going to be happy that you fixed it, that you're listening. >> Yeah. >> And that the product got better. So it's a flywheel of loyalty, quality, brand, all off you can figure it out. It's the holy grail. >> I think it is. It's a gold mine. And every day you're not leveraging this assets, your use of feedback that's there, is a missed opportunity. >> Christian, thanks so much for coming on. Congratulations to you and your startup. You guys back together. The band is back together, up into the right, doing well. >> Yeah. We we'll check in with you later. Thanks for coming on this showcase. Appreciate it. >> Thank you, John. Appreciate it very much. >> Okay. AWS Startup Showcase. This is season two, episode three, the ongoing series. This one's about MarTech, cloud experiences are scaling. I'm John Furrier, your host. Thanks for watching. (upbeat music)
SUMMARY :
of the AWS Startup Showcase. Thank you so much, John. But the holy grail is to And the one we are in And so we can get down to the root cause. I mean, the data silos a huge issue. reading between the lines And then you got the siloed locations. And the epiphany we had at And again, like the cool part is, in the organization. But in the old days, it was the product improvement, Here, you're taking direct input And how do you know if you're improving Can you give an example So are you data driven company? And then you ask, And I think we all agree at test, So these manual workflows. I got to ask you, So the question is, if And every day that you're ask the customer to tell us, So if the product is not working, And then engineering will be like, And a lot of times, And even for the engineers Well, now, I can read all the feedback. and I see the trend go down to zero, Great to have you on. the minute we have left, So how do you compete today? of building the product. happy that you fixed it, And that the product got better. And every day you're not Congratulations to you and your startup. We we'll check in with you later. Appreciate it very much. I'm John Furrier, your host.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Chris | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Christian Wiklund | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
2005 | DATE | 0.99+ |
Hewlett-Packard | ORGANIZATION | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
10% | QUANTITY | 0.99+ |
six month | QUANTITY | 0.99+ |
third | QUANTITY | 0.99+ |
fourth | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
800 accounts | QUANTITY | 0.99+ |
5,000 bugs | QUANTITY | 0.99+ |
51% | QUANTITY | 0.99+ |
14 years | QUANTITY | 0.99+ |
Bay Area | LOCATION | 0.99+ |
90 | QUANTITY | 0.99+ |
Android | TITLE | 0.99+ |
200 people | QUANTITY | 0.99+ |
Nik | PERSON | 0.99+ |
Skout | ORGANIZATION | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
iOS | TITLE | 0.99+ |
Excel | TITLE | 0.99+ |
tomorrow | DATE | 0.99+ |
first time | QUANTITY | 0.99+ |
Christian | PERSON | 0.99+ |
today | DATE | 0.99+ |
unitQ | ORGANIZATION | 0.99+ |
5,000 largest apps | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
second company | QUANTITY | 0.99+ |
100 | QUANTITY | 0.99+ |
Jira | TITLE | 0.99+ |
Spotify | ORGANIZATION | 0.99+ |
Bible | TITLE | 0.99+ |
30 days | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
Zendesk | ORGANIZATION | 0.99+ |
Intercom | ORGANIZATION | 0.98+ |
Chime | ORGANIZATION | 0.98+ |
first | QUANTITY | 0.98+ |
Wall Street | LOCATION | 0.98+ |
once a month | QUANTITY | 0.98+ |
ORGANIZATION | 0.98+ | |
once a quarter | QUANTITY | 0.98+ |
five years ago | DATE | 0.98+ |
million dollar | QUANTITY | 0.97+ |
first company | QUANTITY | 0.97+ |
six months later | DATE | 0.97+ |
zero | QUANTITY | 0.97+ |
Swedish | OTHER | 0.97+ |
Japanese | OTHER | 0.97+ |
late last year | DATE | 0.96+ |
PagerDuty | ORGANIZATION | 0.95+ |
AWS | ORGANIZATION | 0.95+ |
10th time | QUANTITY | 0.95+ |
Christian Wiklund, unitQ | CUBE Conversation
>>Welcome everyone to this cube conversation featuring unit Q. I'm your host, Lisa Martin. And we are excited to be joined by Christian Vickle, the founder and CEO of unit Q Christian. Thank you so much for joining me today. >>Thank you so much, Lisa pleasure to be here. >>Let's talk a little bit about unit Q. You guys were founded in 2018, so pretty recent. What is it that unit Q does. And what were some of the gaps in the market that led you to founding the company? >>Yep. So me and my co-founder Nick, we're actually doing our second company now is the unit Q is number two, and our first company was called scout years ago. We were back ES wicks and it was very different from unit Q. It's a social network for meeting people. And it was really during that experience where we saw the impact that quality of the experience quality of the product can have on your growth trajectory and the challenges we faced. How do we test everything before we ship it? And in reality, a modern company will have, let's say, 20 languages supported you support Android, Iowas, web big screen, small screen, you have 20 plus integrations and you have lots of different devices out there that might run your binary a little differently. So who is the ultimate test group of all of these different permutation and that's the end user. >>And we, we saw the, the big gap in the market, sort of the dream platform for us was unit queue. So if, if this would've existed back in the day, we would've been a, a happy purchaser and customer, and it really comes down to how do we, how do we harness the power of user feedback? You know, the end user, that's testing your product every single day in all different configurations. And then they're telling you that, Hey, something didn't work for me. I got double build or the passive recent link didn't work, or I couldn't, you know, when music, when the ad is finished playing on, on my app, the music doesn't resume. So how do we capture those signals into something that the company and different teams can align on? So that's where, you know, unit Q the, the vision here is to build a quality company, to help other companies build higher quality products. >>So really empowering companies to take a data driven approach to product quality. I was looking on your website and noticed that Pandora is one of your customers, but talk to me a little bit about a customer example that you think really articulates the value of what Q unit he was delivering. >>Right? So maybe we should just go back one little step and talk about what is quality. And I think quality is something that is, is a bit subjective. It's something that we live and breathe every day. It's something that can be formed in an instant first impressions. Last it's something that can be built over time that, Hey, I'm using this product and it's just not working for me. Maybe it's missing features. Maybe there are performance related bots. Maybe there is there's even fulfillment related issues. Like we work with Uber and hello, fresh and, and other types of more hybrid type companies in addition to the Pandoras and, and Pinterest and, and Spotify, and these more digital, only products, but the, the end users I'm producing this data, the reporting, what is working and not working out there in many different channels. So they will leave app produce. >>They will write into support. They might engage with a chat support bot. They will post stuff on Reddit on Twitter. They will comment on Facebook ads. So like this data is dispersed everywhere. The end user is not gonna fill out a perfect bug report in a form somewhere that gets filed into gr like they're, they're producing this content everywhere in different languages. So the first value of what we do is to just ingest all of that data. So all the entire surface area of use of feedback, we ingest into a machine and then we clean the data. We normalize it, and then we translate everything into English. And it was actually a surprise to us when we started this company, that there are quite a few companies out there that they're only looking at feedback in English. So what about my Spanish speaking users? What about my French speaking users? >>And when, when, when that is done, like when all of that data is, is need to organized, we extract signals from that around what is impacting the user experience right now. So we break these, all of this data down into something called quality monitors. So quality monitor is basically a topic which can be again, passive reset, link noting, or really anything that that's impacting the end user. And the important part here is that we need to have specific actionable data. For instance, if I tell you, Hey, Lisa music stops playing is a growing trend that our users are reporting. You will tell me, well, what can I do with that? Like what specifically is breaking? So we deploy up to 1500 unique quality monitors per customer. So we can then alert different teams inside of the organization of like, Hey, something broke and you should take a look at it. >>So it's really breaking down data silos within the company. It aligns cross-functional teams to agree on what should be fixed next. Cause there's typically a lot of confusion, you know, marketing, they might say, Hey, we want this fixed engineering. They're like, well, I can't reproduce, or that's not a high priority for us. The support teams might also have stuff that they want to get fixed. And what we've seen is that these teams, they struggle to communicate. So how do we align them around the single source of truth? And I think that's for unit two is early identification of stuff. That's not working in production and it's also aligning the teams so they can quickly triage and say, yes, we gotta fix this right before it snowballs into something. We say, you know, we wanna, we wanna cap catch issues before you go into crisis PR mode, right? So we want to get this, we wanna address it early in the cycle. >>Talk to me about when you're in customer conversations, Christian, the MarTech landscape is competitive. There's nearly 10,000 different solutions out there, and it's growing really quickly quality monitors that you just described is that one of the key things that, that you talk to customers about, that's a differentiator for unit Q. >>Yeah. So I mean, it, it, it comes down to, as you're building your product, right, you, you have, you have a few different options. One is to build new features and we need to build new features and innovate and, and, and that's all great. We also need to make sure that the foundation of the product is working and that we keep improving quality and what, what we see with, with basically every customer that we work with, that, that when quality goes up, it's supercharges the growth machine. So quality goes up, you're gonna see less support tickets. You're gonna see less one star reviews, less one star reviews is of course good for making the store front convert better. You know, I, I want install a 4.5 star app, not a 3.9 star app. We also see that sentiment. So for those who are interested in getting that NPS score up for the next time we measure it, we see that quality is of course a very important piece of that. >>And maybe even more importantly, so sort of inside of the product machine, the different conversion steps, let's say sign up to activate it to coming back in second day, 30 day, 90 day, and so forth. We see a dramatic impact on how quality sort of moves that up and down the retention function, if you will. So it, it really, if you think about a modern company, like the product is sort of the center of the existence of the company, and if the product performs really well, then you can spend more money in marketing because it converts really good. You can hire more engineers, you can hire, you can hire more support people and so forth. So it's, it's really cool to see that when quality improves its supercharges, everything else I think for marketing it's how do you know if you're spending into a broken product or not? >>And I, and I, I feel like marketing has, they have their insights, but it's, it's not deep enough where they can go to engineering and say, Hey, these 10 issues are impacting my MPS score and they're impacting my conversion and I would love for you to fix it. And when you can bring tangible impact, when you can bring real data to, to engineering and product, they move on it cause they also wanna help build the company. And, and so I think that's, that's how we stand out from the more traditional MarTech, because we need to fix the core of, of sort of this growth engine, which is the quality of the product >>Quality of the product. And obviously that's directly related to the customer experience. And we know these days, one of the things I think that's been in short supply the last couple of years is patience. We know when customers are unhappy with the product or service, and you talked about it a minute ago, they're gonna go right to, to Reddit or other sources to complain about that. So being able to, for uniq, to help companies to improve the customer experience, isn't I think table stakes for businesses it's mission critical these days. Yeah, >>It is mission critical. So if you look at the, let's say that we were gonna start a, a music app. Okay. So how do we, how do we compete as a music app? Well, if you, if you were to analyze all different music apps out there, they have more or less the same features app. Like they, the feature differentiation is minimal. And, and if you launch a new cool feature than your competitor will probably copy that pretty quickly as well. So competing with features is really hard. What about content? Well, I'm gonna get the same content on Spotify as apple SD. So competing with content is also really hard. What about price? So it turns out you'll pay 9 99 a month for music, but there's no, there's no 1 99. It's gonna be 9 99. So quality of the experience is one of the like last vectors or areas where you can actually compete. >>And we see consistently that if you' beating your competition on quality, you will do better. Like the best companies out there also have the highest quality experience. So it's, it's been, you know, for us at our last company, measuring quality was something that was very hard. How do we talk about it? And when we started this company, I went out and talked to a bunch of CEOs and product leaders and board members. And I said, how do you talk about quality in a board meeting? And they were, they said, well, we don't, we don't have any metrics. So actually the first thing we did was to define a metrics. We have, we have this thing called this unit Q score, which is on our website as well, where we can base it's like the credit score. So you can see your score between zero and a hundred. >>And if your score is 100, it means that we're finding no quality issues in the public domain. If your score is 90, it means that 10% of the data we look at refers to a quality issue. And the definition of a quality issue is quite simple. It is when the user experience doesn't match the user expectation. There is a gap in between, and we've actually indexed the 5,000 largest apps out there. So we're then looking at all the public review. So on our website, you can go in and, and look up the unit Q score for the 5,000 largest products. And we republish these every night. So it's an operational metric that changes all the time. >>Hugely impactful. Christian, thank you so much for joining me today, talking to the audience about unit Q, how you're turning qualitative feedback into pretty significant product improvements for your customers. We appreciate your insights. >>Thank you, Lisa, have a great day. >>You as well, per Christian Lin, I'm Lisa Martin. You're watching a cube conversation.
SUMMARY :
And we are excited to be joined by Christian Vickle, the founder and CEO of And what were some of the gaps in the market that led you to founding the company? the challenges we faced. So that's where, you know, unit Q the, So really empowering companies to take a data driven approach to product quality. So maybe we should just go back one little step and talk about what is quality. So the first value of what we do And the important part here is that we need to have specific actionable data. So how do we align them around the single source of truth? that you just described is that one of the key things that, that you talk to customers about, that's a differentiator for unit the next time we measure it, we see that quality is of course a very important piece of that. and if the product performs really well, then you can spend more money in marketing because it converts And when you can bring tangible And we know these days, one of the things I think that's been in short supply the last couple of years is So quality of the experience is one of the like So actually the first thing we did was to So it's an operational metric that changes all the time. Christian, thank you so much for joining me today, talking to the audience about unit Q, You as well, per Christian Lin, I'm Lisa Martin.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
2018 | DATE | 0.99+ |
Christian Wiklund | PERSON | 0.99+ |
3.9 star | QUANTITY | 0.99+ |
10% | QUANTITY | 0.99+ |
Nick | PERSON | 0.99+ |
Christian Vickle | PERSON | 0.99+ |
4.5 star | QUANTITY | 0.99+ |
Christian | PERSON | 0.99+ |
one star | QUANTITY | 0.99+ |
Pandora | ORGANIZATION | 0.99+ |
10 issues | QUANTITY | 0.99+ |
90 day | QUANTITY | 0.99+ |
second company | QUANTITY | 0.99+ |
Pandoras | ORGANIZATION | 0.99+ |
Lisa | PERSON | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
90 | QUANTITY | 0.99+ |
30 day | QUANTITY | 0.99+ |
100 | QUANTITY | 0.99+ |
Christian Lin | PERSON | 0.99+ |
20 languages | QUANTITY | 0.99+ |
second day | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
first company | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
Spotify | ORGANIZATION | 0.99+ |
English | OTHER | 0.99+ |
ORGANIZATION | 0.99+ | |
today | DATE | 0.99+ |
MarTech | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
20 plus integrations | QUANTITY | 0.98+ |
Spanish | OTHER | 0.98+ |
Android | TITLE | 0.97+ |
ORGANIZATION | 0.97+ | |
ORGANIZATION | 0.97+ | |
9 99 | QUANTITY | 0.97+ |
5,000 largest apps | QUANTITY | 0.96+ |
apple | ORGANIZATION | 0.96+ |
unit Q | ORGANIZATION | 0.96+ |
first value | QUANTITY | 0.95+ |
first impressions | QUANTITY | 0.95+ |
9 99 a month | QUANTITY | 0.94+ |
One | QUANTITY | 0.94+ |
5,000 largest products | QUANTITY | 0.93+ |
scout | ORGANIZATION | 0.86+ |
nearly 10,000 different solutions | QUANTITY | 0.85+ |
single source | QUANTITY | 0.85+ |
up to 1500 unique quality monitors | QUANTITY | 0.85+ |
Iowas | LOCATION | 0.84+ |
a minute ago | DATE | 0.84+ |
unit two | QUANTITY | 0.82+ |
ES wicks | ORGANIZATION | 0.8+ |
French | OTHER | 0.78+ |
years ago | DATE | 0.75+ |
one little | QUANTITY | 0.75+ |
zero and | QUANTITY | 0.75+ |
a hundred | QUANTITY | 0.73+ |
1 99 | QUANTITY | 0.72+ |
single day | QUANTITY | 0.7+ |
last couple of years | DATE | 0.68+ |
number two | OTHER | 0.62+ |
unit | ORGANIZATION | 0.61+ |
uniq | ORGANIZATION | 0.55+ |