Edo Liberty: How Vector Data Is Changing The Way We Recommend Everything
Edo Liberty is the CEO of Pinecone, a company hiring exceptional scientists and engineers to solve some of the hardest and most impactful machine learning challenges of our times. Edo also worked at Amazon Web Services where he managed the algorithms group at Amazon AI.
As Senior Manager of Research, Amazon SageMaker, Edo and his team built scalable machine learning systems and algorithms used both internally and externally by customers of SageMaker, AWS’s flagship machine learning platform.
Edo served as Senior Research Director at Yahoo where he was the head of Yahoo’s Independent Research in New York with focus on scalable machine learning and data mining for Yahoo critical applications.
Edo is a Post Doctoral Research fellow in Applied Mathematics from Yale University. His research focused on randomized algorithms for data mining. In particular: dimensionality reduction, numerical linear algebra, and clustering. He is also interested in the concentration of measure phenomenon.
Please support this podcast by checking out our sponsors:
Edo Liberty LinkedIn: https://www.linkedin.com/in/edo-liberty-4380164/
Edo Liberty Twitter: https://twitter.com/pinecone
Edo Liberty Website: https://www.pinecone.io
Podcast website: https://www.humainpodcast.com
YouTube Full Episodes: https://www.youtube.com/channel/UCxvclFvpPvFM9_RxcNg1rag
Support and Social Media:
– Check out the sponsors above, it’s the best way to support this podcast
– Support on Patreon: https://www.patreon.com/humain/creators
– Twitter: https://twitter.com/dyakobovitch
– Instagram: https://www.instagram.com/humainpodcast/
– LinkedIn: https://www.linkedin.com/in/davidyakobovitch/
– Facebook: https://www.facebook.com/HumainPodcast/
– HumAIn Website Articles: https://www.humainpodcast.com/blog/
Here’s the timestamps for the episode:
(06:02)- It’s funny how being a scientist and building applications and building platforms are so different. It’s kind of like for me it’s just by analogy, I mean, kind of a scientist, if you’re looking at some achievement, like technical achievement as being a top of a mountain and a scientist is trying to like hike, they’re trying to be the first person to the summit.
(06:28)- When you build an application, you kind of have to build a road, you have to be able to drive them with a car. And when you’re building a platform on AWS or at Pinecone, you have to like build a city there. You have to really like, completely like to cover it. For me, the experience of building platforms and AWS was transformational because the way we think about problems is completely different. It’s not about proving that something is possible, it is building the mechanisms that make it possible always for, in any circumstance.
(13:43)- And so on and today with machine learning, you don’t really have to do any of that. You have pre-trained NLP models that convert a string, like a, take a sentence in English to an embedding, to a high dimensional vector, such that the similarity or either the distance or the angle between them is analogous to the similarity between them in terms of like conceptual smelts semantic similarity.
(18:17)- Almost always Pinecone ends up being a lot easier, a lot faster and a lot more production ready than what they would build in house. A lot more functional. We’ve spent two and a half years now baking a lot of really great features into Pinecone. And we’re, we’ve just launched a version 2.0 that contains all sorts of filtering capabilities and cost reduction measures and you name it.
(21:22)- And so I’m a great believer in knowing your own data and knowing your own customers and training your own models. It doesn’t mean that you have to train them from scratch. It doesn’t mean you don’t have to use the right tools. You don’t have to reinvent the wheel, but I’m not a big believer in completely pre-trained, plucked off of a random place in the internet models. I do want to say that there are great models for just feature engineering for objects that don’t change so much. So we have language models like BERT that transform text and create great embeddings and they’re a good starting point.
(31:01)- So I think you’ll see two things. First of all, with Pinecone specifically, we’re focused on really only two things; making it easy to use and get value out of Pinecone and making it cheaper. That’s it! I mean that, those are the only two things we care about. Like if you can get a ton of value out of it and it doesn’t cost you too much, that’s it, you’re a happy customer and we’re happy to get you there. So that pretty much sums up all of our focus.