Dec. 1, 2022

How far are we from general AI in healthcare? Amit Garg - Tau Ventures

How far are we from general AI in healthcare? Amit Garg - Tau Ventures
The player is loading ...
How far are we from general AI in healthcare? Amit Garg - Tau Ventures
Apple Podcasts podcast player iconSpotify podcast player iconRSS Feed podcast player iconYouTube podcast player icon
Apple Podcasts podcast player iconSpotify podcast player iconRSS Feed podcast player iconYouTube podcast player icon

Amit is a founding partner at Tau Venture, an $85 million fund which invests in early stage healthcare and enterprise software startups with an AI focus.

00:00 Introduction

02:50 Future of the fund

04:15 How are you thinking about allocation?

08:15 Time to exit

09:10 Secondary market

12:15 Healthcare and profits

15:25 Value based care

18:25 One piece of advice for your past self

23:00 What tailwinds are you investing on?

28:45 AI and biotech

31:15 Will AI replace us?

40:00 What would you do if money was not an issue

44:00 What makes you resilient?

1
00:00:00,000 --> 00:00:02,960
Amit, thanks so much for being here.

2
00:00:02,960 --> 00:00:08,240
You have been very generous with your time with me and I'm honored for that.

3
00:00:08,240 --> 00:00:11,560
Thanks for taking the time to come to this podcast today.

4
00:00:11,560 --> 00:00:15,640
I think to start, if you can give us a brief introduction and then we can get into it.

5
00:00:15,640 --> 00:00:18,760
Well, Rishad, thank you for the kind words.

6
00:00:18,760 --> 00:00:20,040
Have the same words back to you.

7
00:00:20,040 --> 00:00:25,520
It's been a pleasure getting to know you and collaborating and at some point we will do

8
00:00:25,520 --> 00:00:26,520
more deals together.

9
00:00:26,520 --> 00:00:30,920
But for today, I'm just excited to be here and have a conversation.

10
00:00:30,920 --> 00:00:35,480
My background for those listening in, for those watching, I am a venture capitalist

11
00:00:35,480 --> 00:00:36,920
in Silicon Valley.

12
00:00:36,920 --> 00:00:39,600
I run a fund called Tau Ventures.

13
00:00:39,600 --> 00:00:45,320
As of recording this, we have 85 million total that we're investing primarily at the seed

14
00:00:45,320 --> 00:00:48,680
stage writing typically $500,000 checks.

15
00:00:48,680 --> 00:00:50,800
I focus on the healthcare side.

16
00:00:50,800 --> 00:00:56,000
My co-founder and partner Sanjay focuses on enterprise and everything we invest in, we

17
00:00:56,000 --> 00:00:59,320
look for AI, artificial intelligence.

18
00:00:59,320 --> 00:01:03,480
Companies we have invested in include Iterative Health, which is computer vision for colon

19
00:01:03,480 --> 00:01:04,480
cancer.

20
00:01:04,480 --> 00:01:06,480
Now it does a lot more than that.

21
00:01:06,480 --> 00:01:10,400
We have companies that do machine learning for drug discovery, companies that help with

22
00:01:10,400 --> 00:01:11,400
pre-auth.

23
00:01:11,400 --> 00:01:18,960
In general, what we're looking for is how can AI really empower both folks in healthcare

24
00:01:18,960 --> 00:01:22,960
and technology and the intersection to make a really big difference?

25
00:01:22,960 --> 00:01:25,040
We have so many problems in healthcare.

26
00:01:25,040 --> 00:01:29,640
I know Drishad, you're based in Canada, but I'm going to pick on the US here.

27
00:01:29,640 --> 00:01:33,880
In the US, we spent just over 18% of our GDP in healthcare.

28
00:01:33,880 --> 00:01:38,280
We have worse outcomes than countries that are comparable to the US.

29
00:01:38,280 --> 00:01:45,160
We honestly have created a very tragic situation in the US where we have both the best treatments

30
00:01:45,160 --> 00:01:50,160
and the best doctors in the world, but the costs are just out of hand.

31
00:01:50,160 --> 00:01:52,400
The weights are out of hand.

32
00:01:52,400 --> 00:01:53,400
There's a lot to be done.

33
00:01:53,400 --> 00:01:57,320
I think technology is in many ways a very powerful tool.

34
00:01:57,320 --> 00:01:59,760
It's not the only tool to make a difference.

35
00:01:59,760 --> 00:02:03,800
Before all of this, I co-founded a startup called Health IQ.

36
00:02:03,800 --> 00:02:08,320
Publicly, you'll see that the company raised $140 million.

37
00:02:08,320 --> 00:02:12,200
Before that, I worked at a couple of other VC funds, started my career at Google.

38
00:02:12,200 --> 00:02:16,960
This was pre-IPO days at Google, and I guess I'm dating myself.

39
00:02:16,960 --> 00:02:21,360
I was a product manager there, learned a whole bunch and very grateful for my experience

40
00:02:21,360 --> 00:02:22,520
there.

41
00:02:22,520 --> 00:02:25,400
My training is in computer science and biology.

42
00:02:25,400 --> 00:02:28,080
My master's is at the intersection of those two.

43
00:02:28,080 --> 00:02:29,920
Then I also went to business school.

44
00:02:29,920 --> 00:02:36,200
I'm trying to bring all of this, my life experiences, having worked in a big company, having started

45
00:02:36,200 --> 00:02:40,040
companies, having worked at VC funds, and trying to bring it all together here at Tau

46
00:02:40,040 --> 00:02:46,520
Ventures so that we can truly, truly help our entrepreneurs succeed, make an impact,

47
00:02:46,520 --> 00:02:47,520
and make money.

48
00:02:47,520 --> 00:02:50,680
I believe in the intersection of all of those.

49
00:02:50,680 --> 00:02:51,680
Perfect.

50
00:02:51,680 --> 00:02:52,680
Thanks for that introduction.

51
00:02:52,680 --> 00:02:59,160
I told Sametra when I was talking to him that you're going to have more than a billion under

52
00:02:59,160 --> 00:03:00,760
management in five years.

53
00:03:00,760 --> 00:03:02,640
Wow, you are very kind.

54
00:03:02,640 --> 00:03:05,400
I don't think we will, to be honest.

55
00:03:05,400 --> 00:03:12,840
Not trying to be falsely modest here, but raising funds takes time and you can't run

56
00:03:12,840 --> 00:03:14,280
before you walk.

57
00:03:14,280 --> 00:03:18,320
I don't think it's even the right thing for us to get to a billion within five years.

58
00:03:18,320 --> 00:03:23,160
I think more likely we'll raise another fund and then another fund every two or three years.

59
00:03:23,160 --> 00:03:24,560
That's the norm.

60
00:03:24,560 --> 00:03:27,400
You roughly maybe double in size.

61
00:03:27,400 --> 00:03:32,400
There's exceptions and there's reasons to do something different.

62
00:03:32,400 --> 00:03:39,160
If you follow the trajectory here, that is the biggest expectation, we will grow in size

63
00:03:39,160 --> 00:03:40,160
for sure.

64
00:03:40,160 --> 00:03:43,680
I always want to be an early stage fund.

65
00:03:43,680 --> 00:03:48,560
At least that's what we're thinking right now that our differentiator is in how we help

66
00:03:48,560 --> 00:03:51,440
entrepreneurs build companies at the late stage.

67
00:03:51,440 --> 00:03:57,000
It's a lot more about financial modeling and there's a lot of value in that too for sure,

68
00:03:57,000 --> 00:03:59,680
but it's not what we are focused on.

69
00:03:59,680 --> 00:04:05,040
It's also much easier to do a 10x on a $500 million fund than on a $5 billion fund.

70
00:04:05,040 --> 00:04:11,400
There's practical reasons to also keep your size within a certain range.

71
00:04:11,400 --> 00:04:17,280
Let's talk about when you're thinking of allocating this $85 million, how do you manage risk and

72
00:04:17,280 --> 00:04:18,280
reward?

73
00:04:18,280 --> 00:04:23,320
Are you looking for say a 10x return on every single deal you put into?

74
00:04:23,320 --> 00:04:28,680
Are you okay with maybe a 1000x return, but a 5% chance of that on some deals?

75
00:04:28,680 --> 00:04:34,840
How are you thinking about that risk reward to eventually return back, as you said, 10x

76
00:04:34,840 --> 00:04:36,880
on the whole $85 million?

77
00:04:36,880 --> 00:04:40,200
10x by the way is very ambitious.

78
00:04:40,200 --> 00:04:47,080
It says 10x, but if you look at the data, overwhelmingly good funds do 3x and exceptionally

79
00:04:47,080 --> 00:04:49,320
good fund does 5x or higher.

80
00:04:49,320 --> 00:04:56,680
So there is a long tail of distributions for sure, but what we were hoping is 3x at least

81
00:04:56,680 --> 00:04:58,280
and 5x ambitiously.

82
00:04:58,280 --> 00:05:02,360
Anything over that I'll be extremely, extremely happy about.

83
00:05:02,360 --> 00:05:08,280
I can share today our fund is at 2.5x in just over two years.

84
00:05:08,280 --> 00:05:09,780
That's the first fund.

85
00:05:09,780 --> 00:05:11,600
So we seem to be on track.

86
00:05:11,600 --> 00:05:14,480
Now the law of small numbers helps me.

87
00:05:14,480 --> 00:05:20,560
As I was talking earlier, if you have a smaller fund, it's easier to actually get outsized

88
00:05:20,560 --> 00:05:21,560
returns.

89
00:05:21,560 --> 00:05:24,320
You have more flexibility when you buy and sell.

90
00:05:24,320 --> 00:05:30,560
You're also hungry and you also need to get a few exits in order for the needle to really

91
00:05:30,560 --> 00:05:31,560
move.

92
00:05:31,560 --> 00:05:38,500
So for all those three reasons, having a manageable fund size is a good thing.

93
00:05:38,500 --> 00:05:41,680
So when we look at deals, that's partly what we look at also.

94
00:05:41,680 --> 00:05:45,860
We're investing primarily at the seed stage and specifically late seed.

95
00:05:45,860 --> 00:05:49,880
So we're looking for companies that are a little bit more than two people in a garage.

96
00:05:49,880 --> 00:05:52,600
They typically have a pipeline of customers.

97
00:05:52,600 --> 00:05:54,320
That's the key thing to look for.

98
00:05:54,320 --> 00:05:55,800
If you have pilots, wonderful.

99
00:05:55,800 --> 00:05:58,080
If you have pay pilots, even better.

100
00:05:58,080 --> 00:06:02,360
So if you're making money, revenue is amazing, but that's not our expectation.

101
00:06:02,360 --> 00:06:07,340
Our expectation is that you have a roster of potential clients and that you're able

102
00:06:07,340 --> 00:06:12,040
to close on them and get to recurring contracts and eventually to a Series A within nine to

103
00:06:12,040 --> 00:06:13,040
18 months.

104
00:06:13,040 --> 00:06:16,920
A Series A is when you have product market fit here in the US at least.

105
00:06:16,920 --> 00:06:19,400
A million ARR is kind of what people look for.

106
00:06:19,400 --> 00:06:24,080
So we'll look for, can this company get there?

107
00:06:24,080 --> 00:06:27,120
And can we really help them get there?

108
00:06:27,120 --> 00:06:30,760
And can this company have an explosive growth?

109
00:06:30,760 --> 00:06:35,620
So what you want is not just a company that has good revenues and good profitability,

110
00:06:35,620 --> 00:06:38,120
but in what time horizon it does that.

111
00:06:38,120 --> 00:06:42,960
If a company is raising a seed and they have been around for 10 years, it's not a good

112
00:06:42,960 --> 00:06:43,960
fit for me.

113
00:06:43,960 --> 00:06:46,720
They may be a very good company, but it's not what I'm looking for.

114
00:06:46,720 --> 00:06:52,880
I'm looking for perhaps a Silicon Valley mold of companies where you're raising every couple

115
00:06:52,880 --> 00:06:54,160
of years.

116
00:06:54,160 --> 00:07:01,400
You are ideally doubling every year in terms of your revenues or very soon in your ARR,

117
00:07:01,400 --> 00:07:03,160
annual recurring revenue.

118
00:07:03,160 --> 00:07:06,760
But we don't expect every company to hit 10X, obviously.

119
00:07:06,760 --> 00:07:12,320
If every company hit 10X, then there would be no need for venture capitalists, I should

120
00:07:12,320 --> 00:07:13,320
say.

121
00:07:13,320 --> 00:07:15,560
But we look for the odds that a company could get there.

122
00:07:15,560 --> 00:07:22,240
So in our portfolio construction, we do hope and expect, and so far are seeing this, that

123
00:07:22,240 --> 00:07:25,680
about 10% of the companies will actually do 10X or better.

124
00:07:25,680 --> 00:07:33,080
And then maybe 50% or 60% will do somewhere between three or four X and five and six X.

125
00:07:33,080 --> 00:07:39,080
Some companies may do one or two X, and you may have some companies that lose money.

126
00:07:39,080 --> 00:07:40,080
It's possible.

127
00:07:40,080 --> 00:07:43,680
And in our portfolio construction, we said maybe 10% of the companies will actually end

128
00:07:43,680 --> 00:07:44,680
up making less.

129
00:07:44,680 --> 00:07:50,680
I'm happy to say that so far we have way over indexed on the successes.

130
00:07:50,680 --> 00:07:55,520
And once again, that's in some ways a function of having a small fund and having the flexibility,

131
00:07:55,520 --> 00:08:00,700
how we play Parada, when we play Parada, how we help other portfolio companies in terms

132
00:08:00,700 --> 00:08:05,480
of getting customer traction and in terms of getting investor traction and then helping

133
00:08:05,480 --> 00:08:06,480
them with exits.

134
00:08:06,480 --> 00:08:07,720
I'm very proud to say this.

135
00:08:07,720 --> 00:08:11,780
We have had four exits so far, and the first fund is just about two years old, just over

136
00:08:11,780 --> 00:08:13,240
two years old.

137
00:08:13,240 --> 00:08:14,240
Congratulations.

138
00:08:14,240 --> 00:08:15,240
That is impressive.

139
00:08:15,240 --> 00:08:20,440
I was reading some statistics on AngelList, and the average time to exit after VC money

140
00:08:20,440 --> 00:08:22,440
was 5.6 years.

141
00:08:22,440 --> 00:08:26,200
So you guys are doing better than half that.

142
00:08:26,200 --> 00:08:27,200
So it depends.

143
00:08:27,200 --> 00:08:31,920
First of all, the 5.6, I wasn't familiar with that figure, I've heard higher figures than

144
00:08:31,920 --> 00:08:32,920
that.

145
00:08:32,920 --> 00:08:35,060
Different industries have different time horizons.

146
00:08:35,060 --> 00:08:37,560
When you invest also has a different time horizon.

147
00:08:37,560 --> 00:08:42,120
Like if you're investing at a series C, well, you're probably looking more like a three

148
00:08:42,120 --> 00:08:47,200
or four X and probably within three or four years rather than a 10X in 10 years when you

149
00:08:47,200 --> 00:08:48,640
invest at the seed stage, right?

150
00:08:48,640 --> 00:08:51,240
It's risk reward based on time.

151
00:08:51,240 --> 00:08:57,160
But a couple of things that have been beneficial to us is we have had M&As and whatever comes

152
00:08:57,160 --> 00:09:01,360
out of companies that got acquired actually the acquirement IPO, and we have stock in

153
00:09:01,360 --> 00:09:03,000
the acquire also.

154
00:09:03,000 --> 00:09:08,840
But the other instrument that we have besides IPO and M&A is to do secondaries.

155
00:09:08,840 --> 00:09:14,880
So I'm open to selling my shares to somebody else, and I'm open to buying shares from somebody

156
00:09:14,880 --> 00:09:15,880
else.

157
00:09:15,880 --> 00:09:19,640
I've actually bought a lot more so far than I've sold.

158
00:09:19,640 --> 00:09:25,880
I've actually bought four or five times more so far from angel investors, from other VC

159
00:09:25,880 --> 00:09:29,360
investors, and I'm open to selling to somebody else.

160
00:09:29,360 --> 00:09:34,160
I have only done it once so far, but in the near future, I'll do more of it.

161
00:09:34,160 --> 00:09:36,600
And that's what's called a secondary.

162
00:09:36,600 --> 00:09:41,840
So the advantage of a secondary is that you can recognize exits a little bit earlier and

163
00:09:41,840 --> 00:09:46,080
return money to your LPs, your own investors a little bit earlier.

164
00:09:46,080 --> 00:09:50,080
Now how and when you do that, there's art and science to it.

165
00:09:50,080 --> 00:09:55,440
And how much you sell is also obviously there's a lot of art and science to it.

166
00:09:55,440 --> 00:10:00,160
So inherently, I would think if your company is doing well, you would want to buy as much

167
00:10:00,160 --> 00:10:01,160
as you can.

168
00:10:01,160 --> 00:10:02,160
When?

169
00:10:02,160 --> 00:10:03,160
Not necessarily.

170
00:10:03,160 --> 00:10:04,280
Not necessarily.

171
00:10:04,280 --> 00:10:07,200
We are big believers in supporting our companies.

172
00:10:07,200 --> 00:10:11,800
But mind you, we are at the moment 85 AUM.

173
00:10:11,800 --> 00:10:16,120
So if my company is already worth a billion, even if I put in 2 million, 3 million, which

174
00:10:16,120 --> 00:10:19,960
for me is a big check right now, it doesn't move the needle that much.

175
00:10:19,960 --> 00:10:23,120
It's not the kind of capital that that company is looking for.

176
00:10:23,120 --> 00:10:27,000
And it may also not be the amount of risk reward that I want.

177
00:10:27,000 --> 00:10:34,040
So when I decide to play my prorata, and a prorata is just a fancy word that means investing

178
00:10:34,040 --> 00:10:38,980
enough to maintain your ownership in the next round, we like doing that.

179
00:10:38,980 --> 00:10:43,040
But sometimes we don't, even if the company is doing well, because there's enough interest

180
00:10:43,040 --> 00:10:47,160
around the table and we want to make sure really good investors come in.

181
00:10:47,160 --> 00:10:53,000
Or because the company is already valued so highly that the opportunity cost for me is

182
00:10:53,000 --> 00:10:54,000
too high.

183
00:10:54,000 --> 00:10:56,960
I may say, look, I could put more money here, but I could also put it in a company that's

184
00:10:56,960 --> 00:11:02,220
worth one-tenth the valuation where I may have a higher risk reward.

185
00:11:02,220 --> 00:11:07,200
So there's many motivations and many things you have to consider, not just when you make

186
00:11:07,200 --> 00:11:09,680
the investment, but when you do follow-ups.

187
00:11:09,680 --> 00:11:10,680
That makes sense.

188
00:11:10,680 --> 00:11:16,840
Amit, you've had a window into healthcare in various different countries.

189
00:11:16,840 --> 00:11:22,760
You grew up in Brazil, you've invested in India, and obviously North America as well.

190
00:11:22,760 --> 00:11:29,080
The question I want to ask is, there's a theory that I've been playing around with that primary

191
00:11:29,080 --> 00:11:35,280
care and first access to healthcare is where profits should lie, and acute care, chronic

192
00:11:35,280 --> 00:11:41,240
care to an extent, and cancer care likely should not be where profits should lie to

193
00:11:41,240 --> 00:11:43,760
build the best model of healthcare.

194
00:11:43,760 --> 00:11:50,100
I think we still haven't figured out healthcare in terms of pricing and reimbursement anywhere.

195
00:11:50,100 --> 00:11:58,160
How do you look at if you were to design your health system, would it be profitable or not,

196
00:11:58,160 --> 00:12:01,320
and where would most of the profits be derived from if so?

197
00:12:01,320 --> 00:12:05,960
Yeah, no, that's a very simple and very hard question.

198
00:12:05,960 --> 00:12:09,960
Healthcare has obviously a moral component to it.

199
00:12:09,960 --> 00:12:16,160
I think many people, perhaps you included, Rishad, you're a doctor, would agree that

200
00:12:16,160 --> 00:12:21,040
you have to provide some kind of baseline of care for everyone.

201
00:12:21,040 --> 00:12:23,000
I certainly believe in that.

202
00:12:23,000 --> 00:12:29,360
But at the same time, I do see the benefits of having a profit motive because that ensures

203
00:12:29,360 --> 00:12:35,980
accountability, that ensures innovation, that ensures alignment of interests in many ways.

204
00:12:35,980 --> 00:12:38,960
Can also do misalignment of interests.

205
00:12:38,960 --> 00:12:39,960
I am a capitalist.

206
00:12:39,960 --> 00:12:44,800
I mean, venture capitalist has two words, and I did go to business school.

207
00:12:44,800 --> 00:12:51,240
I do believe actually that if you do capitalism in the right way, it is the single most powerful

208
00:12:51,240 --> 00:12:56,480
way of impacting our societies, creating jobs.

209
00:12:56,480 --> 00:13:00,120
I do think there's a way to do that in medicine.

210
00:13:00,120 --> 00:13:01,840
It's not easy.

211
00:13:01,840 --> 00:13:05,840
If I were to design a healthcare system from scratch, first of all, I would surround myself

212
00:13:05,840 --> 00:13:08,400
with a lot of good people because I don't know everything.

213
00:13:08,400 --> 00:13:09,400
I don't know enough.

214
00:13:09,400 --> 00:13:13,980
Healthcare is too big for any one person to try to design by themselves.

215
00:13:13,980 --> 00:13:17,800
What I would try to do is align the incentives as much as possible.

216
00:13:17,800 --> 00:13:22,520
There are companies out there, med device companies, pharma companies, where you need

217
00:13:22,520 --> 00:13:24,080
to provide a profit motive.

218
00:13:24,080 --> 00:13:25,080
Absolutely.

219
00:13:25,080 --> 00:13:26,520
Otherwise, there will not be innovation.

220
00:13:26,520 --> 00:13:29,280
There will not be discoveries.

221
00:13:29,280 --> 00:13:37,100
There is a part of healthcare where you're providing care to perhaps disadvantaged populations,

222
00:13:37,100 --> 00:13:43,400
people who are below the poverty line, where it doesn't make sense to necessarily charge

223
00:13:43,400 --> 00:13:45,680
them money.

224
00:13:45,680 --> 00:13:47,680
Perhaps what you do is you create tiers.

225
00:13:47,680 --> 00:13:52,000
Different countries have tried doing that, including here in the US, or Brazil, or India,

226
00:13:52,000 --> 00:13:56,240
or Canada, or UK.

227
00:13:56,240 --> 00:14:01,260
I think that the answer is you have to have a public system that is good, that attracts

228
00:14:01,260 --> 00:14:08,960
the best talent, that does pay well, that does serve everyone, and that perhaps is managed

229
00:14:08,960 --> 00:14:11,040
at a national level.

230
00:14:11,040 --> 00:14:17,120
I know there's pluses and minuses to this, but I do think there's no way around it.

231
00:14:17,120 --> 00:14:22,120
At the same time, you have to have a good private system because there's other things

232
00:14:22,120 --> 00:14:25,960
in healthcare that absolutely need a good private system.

233
00:14:25,960 --> 00:14:28,320
I think you have to attack from both fronts.

234
00:14:28,320 --> 00:14:34,520
If I look at the countries or the states that have done this the best, it's usually a combination.

235
00:14:34,520 --> 00:14:37,760
This requires far more than just policymakers.

236
00:14:37,760 --> 00:14:42,720
It actually requires what I call all the P's, the letter P, in healthcare.

237
00:14:42,720 --> 00:14:48,440
It requires providers, it requires payers, it requires patients, it requires policymakers

238
00:14:48,440 --> 00:14:51,340
or politicians, it requires pharma.

239
00:14:51,340 --> 00:14:55,880
There's other P's out there, but those are the five big ones.

240
00:14:55,880 --> 00:15:01,680
I think that you have to create a regulatory framework that allows people to actually make

241
00:15:01,680 --> 00:15:03,480
the right choices.

242
00:15:03,480 --> 00:15:08,920
If you create frameworks where the interests are misaligned, no matter how good people

243
00:15:08,920 --> 00:15:11,800
are, they will take decisions that are suboptimal.

244
00:15:11,800 --> 00:15:17,000
This is something I've been thinking about in value-based care, is the incentives are

245
00:15:17,000 --> 00:15:24,360
often aligned to outcomes, which actually creates perverse processes to obtain those

246
00:15:24,360 --> 00:15:25,360
outcomes.

247
00:15:25,360 --> 00:15:29,560
A way better example, if you incentivize a strict BMI, people will starve themselves.

248
00:15:29,560 --> 00:15:33,920
You should incentivize the process or the structures in place instead.

249
00:15:33,920 --> 00:15:39,840
What are your thoughts on value-based care and how to best incentivize the right processes

250
00:15:39,840 --> 00:15:42,560
so the outcomes are achieved that we desire?

251
00:15:42,560 --> 00:15:44,520
You said it better than me, Rishad.

252
00:15:44,520 --> 00:15:49,360
If you pick just one variable and you optimize around that variable, it doesn't necessarily

253
00:15:49,360 --> 00:15:52,160
capture everything.

254
00:15:52,160 --> 00:15:58,840
You could optimize on price, you could optimize on outcomes, you could optimize on the amount

255
00:15:58,840 --> 00:16:02,080
of time for you to get seen quickly.

256
00:16:02,080 --> 00:16:06,840
You could optimize on any one variable and you will not actually optimize for everyone

257
00:16:06,840 --> 00:16:08,280
in every single situation.

258
00:16:08,280 --> 00:16:11,000
That's the problem in healthcare, honestly.

259
00:16:11,000 --> 00:16:13,960
The answer is you can't optimize just on one variable.

260
00:16:13,960 --> 00:16:19,520
I think for better or for worse, outcomes is the least worst metric that I can think

261
00:16:19,520 --> 00:16:21,840
of, but it's not perfect.

262
00:16:21,840 --> 00:16:26,840
You have to temper it exactly as you said by looking at process.

263
00:16:26,840 --> 00:16:29,160
Let me create another situation here.

264
00:16:29,160 --> 00:16:33,520
Doctors that deal, and not just doctors, healthcare providers really, doctors, nurses, physician

265
00:16:33,520 --> 00:16:37,880
assistants, everyone who's involved in the care of a patient that deal with complicated

266
00:16:37,880 --> 00:16:44,560
cases are going to have worse outcomes, very presumably, than doctors who are focused on

267
00:16:44,560 --> 00:16:47,840
easier cases as a percentage.

268
00:16:47,840 --> 00:16:53,160
Doctors who are perceived, and this one I will pick on doctors, who are perceived as nice,

269
00:16:53,160 --> 00:16:59,640
bedside manners, get rated higher even if they're not necessarily the best doctors.

270
00:16:59,640 --> 00:17:04,680
The best doctors may actually be a little bit rough on the edges and they may have patients

271
00:17:04,680 --> 00:17:07,600
do things that the patients don't like.

272
00:17:07,600 --> 00:17:11,720
It creates for a little bit of friction.

273
00:17:11,720 --> 00:17:14,560
If you optimize just on patient satisfaction, it's not correct.

274
00:17:14,560 --> 00:17:16,520
If you just optimize on outcome, it's not correct.

275
00:17:16,520 --> 00:17:20,280
If you optimize on keeping costs low, it's not correct.

276
00:17:20,280 --> 00:17:26,320
Unfortunately, you have to pick one of them and focus on it, but not lose sight of everything

277
00:17:26,320 --> 00:17:27,320
else.

278
00:17:27,320 --> 00:17:31,360
I think that's why you need to look at patients holistically and also at healthcare systems

279
00:17:31,360 --> 00:17:32,520
holistically.

280
00:17:32,520 --> 00:17:39,360
I'm fully in favor of healthcare administrators being much more proficient about the challenge

281
00:17:39,360 --> 00:17:45,120
they're dealing with, for providers to be much more proficient about the administrative

282
00:17:45,120 --> 00:17:51,960
challenges, for patients to be very engaged in their own care, for pharma companies to

283
00:17:51,960 --> 00:17:56,320
change their business models, improve their business models, which are honestly a stretch

284
00:17:56,320 --> 00:18:00,840
very crazy right now, like $10 billion or $1 billion to develop a drug.

285
00:18:00,840 --> 00:18:01,840
That's crazy.

286
00:18:01,840 --> 00:18:05,160
What we need is to really listen into each other.

287
00:18:05,160 --> 00:18:09,060
I know this sounds a little bit cliche, but it is really true to listen into each other

288
00:18:09,060 --> 00:18:11,480
to see things from each other's perspective.

289
00:18:11,480 --> 00:18:15,920
I think if there was one variable, one other P I would focus on, you said process and I

290
00:18:15,920 --> 00:18:17,240
would say perspective.

291
00:18:17,240 --> 00:18:23,180
If you could go back in time to yourself 10 years ago, 20 years ago, what is one piece

292
00:18:23,180 --> 00:18:25,600
of advice you would give yourself?

293
00:18:25,600 --> 00:18:30,160
Oh boy, only one?

294
00:18:30,160 --> 00:18:32,000
As many as you want, Amit.

295
00:18:32,000 --> 00:18:37,800
I think this gets a little philosophical, but what you think now and what ends up happening

296
00:18:37,800 --> 00:18:42,080
10 years later are invariably very different things.

297
00:18:42,080 --> 00:18:46,320
Practically every plan I have made in my life didn't work out or didn't work out the way

298
00:18:46,320 --> 00:18:50,880
I thought it would, but having the plan was very important.

299
00:18:50,880 --> 00:18:56,340
If you'd asked me 10 years ago, I could have never predicted that I would be sitting today

300
00:18:56,340 --> 00:18:59,840
running Tao Ventures, doing early stage investments.

301
00:18:59,840 --> 00:19:06,920
However, if you had asked me what I would like to do in terms of principles, those principles

302
00:19:06,920 --> 00:19:09,480
to stay very same.

303
00:19:09,480 --> 00:19:14,200
I hear this from a lot of people that my principles have stayed constant.

304
00:19:14,200 --> 00:19:20,080
Obviously you evolve over time, but the principle is what you're grounded on.

305
00:19:20,080 --> 00:19:25,240
I've always wanted to make an impact, do good, do well.

306
00:19:25,240 --> 00:19:29,880
I've always believed more specifically that the intersection of life sciences and technology

307
00:19:29,880 --> 00:19:33,640
is something that I can make a disproportionate impact.

308
00:19:33,640 --> 00:19:40,680
In my engineering mindset and my life sciences mindset, I could bring both those two and

309
00:19:40,680 --> 00:19:42,680
provide the best treatments for everyone.

310
00:19:42,680 --> 00:19:44,000
Make a lot of money in the process.

311
00:19:44,000 --> 00:19:46,760
I don't think those are mutually exclusive, by the way.

312
00:19:46,760 --> 00:19:51,560
I don't think that you have to choose between making money and doing good.

313
00:19:51,560 --> 00:19:55,200
There are ways in which you can actually do both.

314
00:19:55,200 --> 00:19:57,600
That principle has stayed constant for me.

315
00:19:57,600 --> 00:20:02,540
If I were to remind myself 10 years ago, it would be everything you're thinking right

316
00:20:02,540 --> 00:20:06,480
now specifically will probably not work like you're thinking.

317
00:20:06,480 --> 00:20:11,440
But the underlying motivations, make sure you keep holding onto those.

318
00:20:11,440 --> 00:20:13,640
I agree with that.

319
00:20:13,640 --> 00:20:18,940
I would add money is status and status brings the power to affect change.

320
00:20:18,940 --> 00:20:24,440
Take me back to when you were a child, when you were five years old, 10 years old.

321
00:20:24,440 --> 00:20:27,720
What did you want to become at that point?

322
00:20:27,720 --> 00:20:34,000
To make an assumption there was, some people would argue is the mindset of a child.

323
00:20:34,000 --> 00:20:40,720
But jokes aside, as far as I can remember, I wanted to do something in healthcare.

324
00:20:40,720 --> 00:20:42,480
This is for many reasons.

325
00:20:42,480 --> 00:20:46,280
I did grow up in a smaller city in Brazil.

326
00:20:46,280 --> 00:20:48,440
I did my high school in a larger city.

327
00:20:48,440 --> 00:20:52,540
I've spent some time in India, which is where my roots are.

328
00:20:52,540 --> 00:20:58,280
When you grow up in emerging countries, you do see lots of social problems.

329
00:20:58,280 --> 00:21:01,600
You see both the best and the worst of humanity in some ways.

330
00:21:01,600 --> 00:21:05,080
I wanted to try fixing some problems.

331
00:21:05,080 --> 00:21:09,200
Growing up, I do remember very specifically people not having access to education and

332
00:21:09,200 --> 00:21:10,800
healthcare.

333
00:21:10,800 --> 00:21:14,080
My parents came from education and they're professors.

334
00:21:14,080 --> 00:21:17,960
I always thought maybe something along those lines or something along the lines of healthcare.

335
00:21:17,960 --> 00:21:23,560
I would gravitate more towards healthcare because I felt that if you don't have good health,

336
00:21:23,560 --> 00:21:25,800
then you can't do anything in life really.

337
00:21:25,800 --> 00:21:32,200
Healthcare is really a foundation of who you are.

338
00:21:32,200 --> 00:21:36,520
If you're not healthy yourself, you can't afford even to go to school in some ways.

339
00:21:36,520 --> 00:21:41,200
It's truly, truly foundational for you as an individual, for a family, and for society

340
00:21:41,200 --> 00:21:42,200
in general.

341
00:21:42,200 --> 00:21:46,320
I'm pretty sure my parents would have said that I wanted to be a doctor.

342
00:21:46,320 --> 00:21:54,320
I'm sure I entertained being an astronaut or being a jet pilot fighter, but those were

343
00:21:54,320 --> 00:21:55,320
whims.

344
00:21:55,320 --> 00:21:56,320
They were not real plans.

345
00:21:56,320 --> 00:22:00,160
I think I truly wanted to be a doctor.

346
00:22:00,160 --> 00:22:02,480
As I grew older, I actually applied to be a doctor.

347
00:22:02,480 --> 00:22:06,520
I got into med school, but I decided not to.

348
00:22:06,520 --> 00:22:15,320
It wasn't a change in goals in some ways because my ultimate goal wasn't to be a doctor.

349
00:22:15,320 --> 00:22:18,720
My ultimate goal was to make an impact.

350
00:22:18,720 --> 00:22:20,640
My ultimate goal was to fix problems.

351
00:22:20,640 --> 00:22:25,600
My ultimate goal was to create a better life for myself and for other people.

352
00:22:25,600 --> 00:22:28,280
I don't want to come across here as not being self-interested.

353
00:22:28,280 --> 00:22:29,980
I'm very self-interested.

354
00:22:29,980 --> 00:22:31,320
I do want to make money.

355
00:22:31,320 --> 00:22:32,600
I do want to live comfortably.

356
00:22:32,600 --> 00:22:37,880
I do want to have the words you mentioned, power and status, but I don't want those necessarily

357
00:22:37,880 --> 00:22:41,040
at the exclusion of being able to make an impact.

358
00:22:41,040 --> 00:22:45,280
In fact, if anything, I want to use both in service of each other.

359
00:22:45,280 --> 00:22:49,200
In my mind, at least those things are aligned.

360
00:22:49,200 --> 00:22:54,720
My ultimate motivation behind it, that has stayed constant as far as I can remember.

361
00:22:54,720 --> 00:22:55,720
Okay, perfect.

362
00:22:55,720 --> 00:22:58,240
Thanks for sharing that, Amit.

363
00:22:58,240 --> 00:23:04,880
The next question I want to ask is, building on Bill Gross's idea lab study of why now

364
00:23:04,880 --> 00:23:10,920
is the best predictor of success in a startup, one could argue tailwinds drive a considerable

365
00:23:10,920 --> 00:23:16,080
amount of success, COVID, regulatory changes, wars to an extent.

366
00:23:16,080 --> 00:23:22,680
A tailwind I'm banking on is hybrid at-home care, really growing over the next five years,

367
00:23:22,680 --> 00:23:25,600
including hospital care at home.

368
00:23:25,600 --> 00:23:28,840
What are some tailwinds you're banking on right now?

369
00:23:28,840 --> 00:23:31,760
I'm going to ask you to predict the future.

370
00:23:31,760 --> 00:23:37,440
What are some tailwinds you think will emerge in the next five to 10 years?

371
00:23:37,440 --> 00:23:38,440
Sure.

372
00:23:38,440 --> 00:23:41,520
I'm going to invest in a company called Irritative Health, which I know you're familiar with

373
00:23:41,520 --> 00:23:42,520
also, Rishad.

374
00:23:42,520 --> 00:23:48,320
It was very coincidentally the very first investment we did through the fund, officially.

375
00:23:48,320 --> 00:23:51,800
They've done phenomenally well.

376
00:23:51,800 --> 00:23:57,960
They've raised more than $200 million, and this is all public in just over two years.

377
00:23:57,960 --> 00:24:03,360
What the company had started off with was training their algorithm to detect cancer

378
00:24:03,360 --> 00:24:04,600
using computer vision.

379
00:24:04,600 --> 00:24:10,920
More specifically, taking video feeds of a colonoscopy and in real time, being able to

380
00:24:10,920 --> 00:24:16,760
detect if there was a polyp that was cancerous, and helping physicians, specialists, especially

381
00:24:16,760 --> 00:24:22,680
gastroenterologists, be able to detect colon cancer earlier, quicker, and better.

382
00:24:22,680 --> 00:24:28,320
If you don't detect it early enough, cancer is, unfortunately, all of us have been touched

383
00:24:28,320 --> 00:24:33,360
by it in some way or shape or form, and I think all of us will resonate with this, that

384
00:24:33,360 --> 00:24:36,400
it can mean the difference between making it or not making it.

385
00:24:36,400 --> 00:24:38,080
Detect early, you save a patient's life.

386
00:24:38,080 --> 00:24:42,040
You detect it too late, maybe you can't do anything about it.

387
00:24:42,040 --> 00:24:47,760
Where I'm going with this is we betted on a company that was able to take massive amounts

388
00:24:47,760 --> 00:24:52,520
of data and make sense of it very quickly, and that would not have been possible even

389
00:24:52,520 --> 00:24:53,520
a few years back.

390
00:24:53,520 --> 00:24:57,360
I dare say even three or four years back, because now we have a lot more computational

391
00:24:57,360 --> 00:24:58,360
power.

392
00:24:58,360 --> 00:24:59,360
We have a lot more data.

393
00:24:59,360 --> 00:25:03,440
We have a lot more understanding of how to analyze this data.

394
00:25:03,440 --> 00:25:05,280
We have algorithms that have been built by others.

395
00:25:05,280 --> 00:25:10,000
We have tech stacks that you can leverage from, and I'm giving the example of iterative

396
00:25:10,000 --> 00:25:13,240
health, but that's symptomatic of most of our portfolio.

397
00:25:13,240 --> 00:25:15,720
We have found really good folks.

398
00:25:15,720 --> 00:25:17,600
Sometimes they're doctors, sometimes they're technologists.

399
00:25:17,600 --> 00:25:23,120
More often than not, there are partnerships between these people who are able to take

400
00:25:23,120 --> 00:25:26,200
cutting edge tools and apply it to solve these problems.

401
00:25:26,200 --> 00:25:28,240
I'll give you another example.

402
00:25:28,240 --> 00:25:37,840
We have a company called Signos, which repurposed a hardware device, a continuous glucose monitor,

403
00:25:37,840 --> 00:25:47,240
that is typically used for you to track and hopefully avoid a glucose shortage.

404
00:25:47,240 --> 00:25:54,080
It's affiliated with an insulin pump and diabetics use it, but they have repurposed it for obesity.

405
00:25:54,080 --> 00:25:56,880
You can see what's going on inside your body in real time.

406
00:25:56,880 --> 00:26:00,680
You eat, drink, or exercise and help you take better decisions.

407
00:26:00,680 --> 00:26:05,080
If you can measure it, you can monitor it, and if you can monitor it, you can modify

408
00:26:05,080 --> 00:26:08,880
it, your behaviors.

409
00:26:08,880 --> 00:26:15,640
Once again, if you think about it, we bet it on a company that has done really well

410
00:26:15,640 --> 00:26:20,080
so far and we continue very much believing in it, that can take massive amounts of data

411
00:26:20,080 --> 00:26:22,800
and try to make sense of it.

412
00:26:22,800 --> 00:26:27,360
No single human being could have made sense of all of this until recently.

413
00:26:27,360 --> 00:26:28,920
You do need machines.

414
00:26:28,920 --> 00:26:34,080
I am a firm believer that machines are not here to replace us.

415
00:26:34,080 --> 00:26:43,760
They're here to help us, to augment what we can do as physicians, as nurses, as investors,

416
00:26:43,760 --> 00:26:45,120
as a society.

417
00:26:45,120 --> 00:26:46,320
It will create jobs.

418
00:26:46,320 --> 00:26:47,680
It will destroy jobs.

419
00:26:47,680 --> 00:26:54,800
Overall, what technology has done throughout human history is make us more efficient, is

420
00:26:54,800 --> 00:26:58,840
able to do things that we couldn't do as well at all.

421
00:26:58,840 --> 00:27:06,560
I do believe that AI done in a smart way can make our societies incredibly better.

422
00:27:06,560 --> 00:27:12,560
It can lead us, the whole gospel of prosperity, it can lead us to a brighter future.

423
00:27:12,560 --> 00:27:13,940
The key is to do it well.

424
00:27:13,940 --> 00:27:21,280
If you do it in the wrong ways, we can create a lot of imbalances in society.

425
00:27:21,280 --> 00:27:23,480
That is going on to your second question.

426
00:27:23,480 --> 00:27:27,320
What I do expect in the next 10 years, I expect more of this.

427
00:27:27,320 --> 00:27:29,080
I expect a lot more of this.

428
00:27:29,080 --> 00:27:37,960
I expect AI to really make a transformational change in how we develop drugs, how we discover

429
00:27:37,960 --> 00:27:39,480
drugs.

430
00:27:39,480 --> 00:27:45,000
The amount of molecules that we know today is a fraction of all the possible molecules

431
00:27:45,000 --> 00:27:49,000
that can exist in the universe.

432
00:27:49,000 --> 00:27:52,120
I've heard the figure 0.01%, by the way.

433
00:27:52,120 --> 00:27:59,360
That's the amount of molecules that we know today of all the possibilities out there.

434
00:27:59,360 --> 00:28:04,240
I'm hinting here at a revolution in biotech, which I would love to do more of.

435
00:28:04,240 --> 00:28:05,880
I'm not there yet.

436
00:28:05,880 --> 00:28:12,000
We invest in digital health, but in the near future, I do expect biotech to be as fast

437
00:28:12,000 --> 00:28:19,000
in terms of development with CRISPR and CAR-T and with AI and with all of these new technologies

438
00:28:19,000 --> 00:28:25,680
that are emerging to become more similar to how tech is done today.

439
00:28:25,680 --> 00:28:27,400
Digital health is already there, by the way.

440
00:28:27,400 --> 00:28:32,760
Digital health, two people in a garage can now build digital health companies the same

441
00:28:32,760 --> 00:28:35,680
way that Google and Apple were created.

442
00:28:35,680 --> 00:28:40,160
So digital health already operates very much like tech, and I expect another industry to

443
00:28:40,160 --> 00:28:43,960
start operating in the next 10 years.

444
00:28:43,960 --> 00:28:49,920
With biotech in particular, the time to exit on average from what I remember is over eight

445
00:28:49,920 --> 00:28:51,320
to 10 years.

446
00:28:51,320 --> 00:28:56,640
A lot of that is because of the trials that need to happen in the stepwise fashion waiting

447
00:28:56,640 --> 00:28:58,000
for results to come.

448
00:28:58,000 --> 00:29:01,800
Do you think that process will be expedited through AI as well?

449
00:29:01,800 --> 00:29:02,800
Absolutely.

450
00:29:02,800 --> 00:29:03,800
Absolutely.

451
00:29:03,800 --> 00:29:11,920
I have a PhD in biochemistry, so I'm remiss in saying that, but I am operating at the

452
00:29:11,920 --> 00:29:17,240
intersection of AI and life sciences, and that's what I've built my career for the last

453
00:29:17,240 --> 00:29:18,760
20 plus years.

454
00:29:18,760 --> 00:29:26,840
I firmly believe that what in silico allows us to speed things up, not by one or two X,

455
00:29:26,840 --> 00:29:28,880
by 10, 100 X.

456
00:29:28,880 --> 00:29:35,680
Think about being able to simulate molecules, to be able to simulate protein-protein bonds,

457
00:29:35,680 --> 00:29:43,600
to be able to simulate how drug-to-drug interactions would be, to be able to simulate how a particular

458
00:29:43,600 --> 00:29:46,080
drug would behave in your body.

459
00:29:46,080 --> 00:29:49,000
Ultimately, you will still need human clinical trials.

460
00:29:49,000 --> 00:29:50,000
Absolutely.

461
00:29:50,000 --> 00:29:51,000
Animal trials.

462
00:29:51,000 --> 00:29:52,000
Absolutely.

463
00:29:52,000 --> 00:30:01,040
But you'll be able to hone in instead of looking at a search space that is the entire universe.

464
00:30:01,040 --> 00:30:03,480
You'll be able to hone in, and these are the things we need to test.

465
00:30:03,480 --> 00:30:05,040
This is how we need to test.

466
00:30:05,040 --> 00:30:07,840
These are the things you have discovered that are not issues, and these are potential things

467
00:30:07,840 --> 00:30:08,840
we need to go deeper in.

468
00:30:08,840 --> 00:30:09,840
Okay.

469
00:30:09,840 --> 00:30:10,840
Yeah.

470
00:30:10,840 --> 00:30:14,720
I think what AI will allow us to do, and is allowing us to do, is to be able to focus

471
00:30:14,720 --> 00:30:16,040
a lot better.

472
00:30:16,040 --> 00:30:18,040
We do have some portfolio companies doing this already.

473
00:30:18,040 --> 00:30:24,240
We have one called ArpeggioBio that's focused on mRNA and helping develop more around mRNA

474
00:30:24,240 --> 00:30:25,240
as a platform.

475
00:30:25,240 --> 00:30:26,880
RNA in general, I should say.

476
00:30:26,880 --> 00:30:34,720
We have another one called Teco.Bio that is focused on how do drugs interact with your

477
00:30:34,720 --> 00:30:37,680
blood and with cancer in general.

478
00:30:37,680 --> 00:30:39,640
How do you make sense of all of that?

479
00:30:39,640 --> 00:30:46,520
Because the blood is actually the way your entire body really transports substances to

480
00:30:46,520 --> 00:30:47,520
each other.

481
00:30:47,520 --> 00:30:51,160
Ultimately, it's the blood for the vast majority of things.

482
00:30:51,160 --> 00:30:55,320
How you deliver drugs to the right place and how they interact with each other is absolutely

483
00:30:55,320 --> 00:30:57,560
crucial for how we make sense of cancer.

484
00:30:57,560 --> 00:30:58,560
Yeah.

485
00:30:58,560 --> 00:31:03,960
I think targeted chemo, especially immunotherapy, has just made such big advancements over the

486
00:31:03,960 --> 00:31:04,960
past two decades.

487
00:31:04,960 --> 00:31:10,280
I don't want to get too deep into this because a good chunk of our listeners are not physicians.

488
00:31:10,280 --> 00:31:18,280
I think there's a comfort with AI replacing analytical tasks, but there is a discomfort

489
00:31:18,280 --> 00:31:26,640
with AI replacing creative tasks, especially with Dali and what it's doing in the art world.

490
00:31:26,640 --> 00:31:33,760
We seem to protect the artistic side of humanity from AI, and we seem to want to say what makes

491
00:31:33,760 --> 00:31:37,520
us human is our creative endeavors.

492
00:31:37,520 --> 00:31:44,480
What are your thoughts if AI can be more creative than us, if AI can be more human than humans?

493
00:31:44,480 --> 00:31:48,520
Should AI replace creative tasks?

494
00:31:48,520 --> 00:31:54,820
Art music are probably easier to answer than being a counselor, being a friend, or even

495
00:31:54,820 --> 00:31:56,440
being a parent.

496
00:31:56,440 --> 00:32:00,520
If AI can do it better than us, should we let it?

497
00:32:00,520 --> 00:32:03,260
Once again, very simple and very hard question.

498
00:32:03,260 --> 00:32:06,680
You talk to different folks, they'll have different opinions, and these are folks who

499
00:32:06,680 --> 00:32:07,680
are experts.

500
00:32:07,680 --> 00:32:11,360
You talk to people who have spent their entire lives with AI, and they'll give you different

501
00:32:11,360 --> 00:32:13,240
answers around this.

502
00:32:13,240 --> 00:32:18,440
There's very famous Canadians, or people based in Canada, I should say, Jeffrey Hinton and

503
00:32:18,440 --> 00:32:23,480
Yasha Bengio, and some of the foremost names in AI are actually based in Canada or working

504
00:32:23,480 --> 00:32:25,200
out of Canada.

505
00:32:25,200 --> 00:32:31,680
But I fall under the school of thought that general AI is really, really far.

506
00:32:31,680 --> 00:32:34,280
It's not going to happen in my lifetime.

507
00:32:34,280 --> 00:32:38,400
An intelligence that is more human than human is very far.

508
00:32:38,400 --> 00:32:43,800
An intelligence that is better than humans in specific tasks, that's already here, and

509
00:32:43,800 --> 00:32:45,260
that's going to expand.

510
00:32:45,260 --> 00:32:48,340
But think about how AI in general learns today.

511
00:32:48,340 --> 00:32:53,920
By and large, it's because we throw a lot of data, and we do reinforcement learning,

512
00:32:53,920 --> 00:32:54,920
supervised learning.

513
00:32:54,920 --> 00:33:00,200
There is such a thing called unsupervised learning, but by and large, the inherent intelligence

514
00:33:00,200 --> 00:33:06,360
of the fastest, strongest, biggest computer in the world, which you could argue is the

515
00:33:06,360 --> 00:33:13,320
internet itself, is not bigger than an invertebrate, than a worm.

516
00:33:13,320 --> 00:33:18,520
There are tasks that human babies learn by themselves or by watching that computers can't

517
00:33:18,520 --> 00:33:19,520
do.

518
00:33:19,520 --> 00:33:24,760
Fundamentally, how human brains are structured and learn is very different from how computers

519
00:33:24,760 --> 00:33:25,760
are structured.

520
00:33:25,760 --> 00:33:28,940
It doesn't mean that we need to build computers like humans, by the way.

521
00:33:28,940 --> 00:33:32,560
It leads to a whole philosophical question of what is intelligence, how do you define

522
00:33:32,560 --> 00:33:35,680
it, and can intelligence be replicated in different ways?

523
00:33:35,680 --> 00:33:41,200
I fall under the school of thought that there's not a single path towards intelligence.

524
00:33:41,200 --> 00:33:47,080
Going to your question really more specifically, look, we've been going through revolutions

525
00:33:47,080 --> 00:33:49,160
for thousands of years at this point.

526
00:33:49,160 --> 00:33:54,480
We started off thinking that the Earth was the center of the universe, and then we realized,

527
00:33:54,480 --> 00:33:56,760
no, the Earth orbits a star.

528
00:33:56,760 --> 00:33:59,000
We thought that that star was the center of the universe.

529
00:33:59,000 --> 00:34:02,520
Then we said, no, no, that star is just one in a galaxy.

530
00:34:02,520 --> 00:34:04,960
Then we said that galaxy is the only thing in the universe.

531
00:34:04,960 --> 00:34:09,640
Now we know that galaxy is one of thousands, millions, trillions in the universe.

532
00:34:09,640 --> 00:34:14,560
We have gone from an anthropocentric view of the world to a heliocentric view of the

533
00:34:14,560 --> 00:34:22,040
world to a perhaps a pantheistic centric view of the world.

534
00:34:22,040 --> 00:34:23,680
Why not?

535
00:34:23,680 --> 00:34:30,760
Why can't we also accept that things that we create, AI is a creation of human beings,

536
00:34:30,760 --> 00:34:32,860
can actually create things that we can't?

537
00:34:32,860 --> 00:34:40,440
We already do automation in factories, and those factories will do productions at a higher

538
00:34:40,440 --> 00:34:45,520
quality and at higher speed than we do.

539
00:34:45,520 --> 00:34:47,000
Think about cars.

540
00:34:47,000 --> 00:34:50,600
We invented cars, and then we used machines to perfect those cars.

541
00:34:50,600 --> 00:34:54,560
We built machines that built machines that built the machines that built the cars.

542
00:34:54,560 --> 00:34:59,240
We have created layers upon layers upon layers upon layers.

543
00:34:59,240 --> 00:35:07,800
I'm not personally afraid of AI diminishing my humanity if AI can do something better.

544
00:35:07,800 --> 00:35:08,920
I don't celebrate that.

545
00:35:08,920 --> 00:35:13,420
We created AI, and AI can do things that I couldn't even dream of.

546
00:35:13,420 --> 00:35:16,160
What I think needs to happen is for us not to lose purpose.

547
00:35:16,160 --> 00:35:17,720
I think that's a different question.

548
00:35:17,720 --> 00:35:22,880
If humans lose their purpose as individuals and as a society and as a civilization, then

549
00:35:22,880 --> 00:35:24,400
yes, then they're screwed.

550
00:35:24,400 --> 00:35:30,280
There is a danger then in recognizing that things that we create are better than us in

551
00:35:30,280 --> 00:35:35,480
so many different ways that we start losing purpose, but it doesn't need to be that way.

552
00:35:35,480 --> 00:35:39,360
I'm totally okay with AI doing things better than me.

553
00:35:39,360 --> 00:35:44,800
Sure, I just don't need to lose sight of purpose in that conversation.

554
00:35:44,800 --> 00:35:46,160
I'm a big fan of you, man.

555
00:35:46,160 --> 00:35:47,920
I wish more people thought like that.

556
00:35:47,920 --> 00:35:50,280
There's a lot of fear in this space.

557
00:35:50,280 --> 00:35:54,580
You said something very interesting that I want to pick at more, that we don't need to

558
00:35:54,580 --> 00:35:57,440
build computers like humans.

559
00:35:57,440 --> 00:36:03,200
As per my understanding of the current regulatory space, there's still a need for us to understand

560
00:36:03,200 --> 00:36:05,480
what the AI is doing.

561
00:36:05,480 --> 00:36:10,280
And I'm not a computer engineer, but there needs to be some transparency over the neural

562
00:36:10,280 --> 00:36:12,640
networks or whatever the learning process is.

563
00:36:12,640 --> 00:36:18,920
At what point do we let go of that need for control and say, if there is correlation between

564
00:36:18,920 --> 00:36:25,960
outcomes or the outcomes the AI is producing are great and better than ours, then we don't

565
00:36:25,960 --> 00:36:30,680
need to understand the process because we might not be able to, and maybe we're limiting

566
00:36:30,680 --> 00:36:36,120
the scope of AI by saying, okay, we need to understand what's going on here on the backend.

567
00:36:36,120 --> 00:36:41,720
Yeah, so you're talking about explainability and that's tough.

568
00:36:41,720 --> 00:36:46,120
Every single one of your questions has been tough because they're not clear cut answers.

569
00:36:46,120 --> 00:36:47,480
Once again, things evolve.

570
00:36:47,480 --> 00:36:53,080
You have to also be humble enough, and I'm speaking at myself, that with new data, with

571
00:36:53,080 --> 00:36:57,800
new knowledge, with new expertise, you factor that into your thinking and your thinking

572
00:36:57,800 --> 00:36:58,800
may change.

573
00:36:58,800 --> 00:37:01,680
Once again, thinking will change, will evolve.

574
00:37:01,680 --> 00:37:06,800
I think the underlying motivations and principles, those can stay constant.

575
00:37:06,800 --> 00:37:15,200
So I think explainability is crucial for us to be able to justify, rationalize, make sense

576
00:37:15,200 --> 00:37:16,440
of things.

577
00:37:16,440 --> 00:37:23,520
We're not there in terms of being comfortable with the unexplainable when we have created

578
00:37:23,520 --> 00:37:26,000
the unexplainable.

579
00:37:26,000 --> 00:37:29,720
There's plenty of things that happen in the world that are unexplainable to us.

580
00:37:29,720 --> 00:37:36,480
Life is in many ways, arguably, random.

581
00:37:36,480 --> 00:37:44,440
There's so many variables to track for that we don't have the capacity to decide that

582
00:37:44,440 --> 00:37:45,920
this caused this, this caused this.

583
00:37:45,920 --> 00:37:48,240
There's so many different factors at play.

584
00:37:48,240 --> 00:37:49,880
We're comfortable with randomness.

585
00:37:49,880 --> 00:37:55,000
We're comfortable with non-explainability to a certain degree when we are not behind

586
00:37:55,000 --> 00:37:56,760
it.

587
00:37:56,760 --> 00:38:03,680
When we think about, oh, I was born in a rich family or I was born in a poor family.

588
00:38:03,680 --> 00:38:05,960
Well, I can't explain that.

589
00:38:05,960 --> 00:38:07,760
It happened.

590
00:38:07,760 --> 00:38:09,360
We're reasonably comfortable with that.

591
00:38:09,360 --> 00:38:10,760
We may argue about it.

592
00:38:10,760 --> 00:38:11,760
We may debate about it.

593
00:38:11,760 --> 00:38:17,000
We may at the end of the day say, I don't understand how it happened, but we know that

594
00:38:17,000 --> 00:38:18,000
it happens.

595
00:38:18,000 --> 00:38:23,080
And we have over the course of hundreds of generations gotten comfortable with that.

596
00:38:23,080 --> 00:38:27,360
But we're not comfortable when we create that unexplainability.

597
00:38:27,360 --> 00:38:32,720
We're not comfortable with the fact that I created a box and I can't explain what the

598
00:38:32,720 --> 00:38:33,900
box does.

599
00:38:33,900 --> 00:38:38,280
So that is ultimately on us eventually getting comfortable.

600
00:38:38,280 --> 00:38:43,920
If the AI is doing a good job and we have tested it hundreds and millions of times and

601
00:38:43,920 --> 00:38:49,140
it is leading to better patient outcomes, it's making our societies better.

602
00:38:49,140 --> 00:38:50,500
Maybe we should consider it.

603
00:38:50,500 --> 00:38:56,960
Maybe we should consider it that I can't explain how, but I know that this is better for us.

604
00:38:56,960 --> 00:39:00,760
We're not there and I'm not advocating we get there anytime soon.

605
00:39:00,760 --> 00:39:05,800
Right now I think it is very important for us to go from one step to another step to

606
00:39:05,800 --> 00:39:06,800
another step.

607
00:39:06,800 --> 00:39:09,760
I think it is, I'm not comfortable with that.

608
00:39:09,760 --> 00:39:14,240
If I were being treated for something and an AI did something and we couldn't explain

609
00:39:14,240 --> 00:39:15,960
how it did it, I would worry.

610
00:39:15,960 --> 00:39:17,960
I would be grateful, but I would worry.

611
00:39:17,960 --> 00:39:21,180
What if we do follow-ups here?

612
00:39:21,180 --> 00:39:26,160
If something happens down the road, if we can explain what the AI did, then how do I

613
00:39:26,160 --> 00:39:28,520
ensure my future treatment goes well?

614
00:39:28,520 --> 00:39:33,520
I think we do need the regulation and we need to create frameworks.

615
00:39:33,520 --> 00:39:35,040
This will be an evolving conversation.

616
00:39:35,040 --> 00:39:40,280
It will take perhaps far more than your or my lifetime.

617
00:39:40,280 --> 00:39:44,680
I'll ask what I think hopefully is an easier question.

618
00:39:44,680 --> 00:39:51,680
If your LPs came to you tomorrow and said, we don't want any more returns, keep all the

619
00:39:51,680 --> 00:39:54,960
money we've committed, what would you do the day after?

620
00:39:54,960 --> 00:39:58,880
What if that amount was a billion?

621
00:39:58,880 --> 00:40:01,000
Wow.

622
00:40:01,000 --> 00:40:05,400
That is not a situation that will happen.

623
00:40:05,400 --> 00:40:10,640
That I have a billion tomorrow and my LPs say keep it.

624
00:40:10,640 --> 00:40:12,240
I know where you're going with this.

625
00:40:12,240 --> 00:40:15,440
You're asking the hypothetical question is, if money wasn't an object, what would you

626
00:40:15,440 --> 00:40:16,440
do?

627
00:40:16,440 --> 00:40:17,440
That's the question you're asking.

628
00:40:17,440 --> 00:40:23,120
If money wasn't an object and you had a lot of money at your disposal, it's not just that

629
00:40:23,120 --> 00:40:26,840
you don't need any more money for whatever you want to do, but you have quite a bit of

630
00:40:26,840 --> 00:40:27,840
capital that-

631
00:40:27,840 --> 00:40:33,400
The second part of your question is an assumption that the world is the same as it is right

632
00:40:33,400 --> 00:40:37,140
now, meaning tomorrow hasn't changed anything.

633
00:40:37,140 --> 00:40:44,080
If tomorrow's world is significantly different than today's world, then I would perhaps take

634
00:40:44,080 --> 00:40:45,840
different actions.

635
00:40:45,840 --> 00:40:50,440
If there's an asteroid hurtling toward the earth tomorrow and I have a billion dollars

636
00:40:50,440 --> 00:40:56,280
and I can do something about it, well, yes, then that's what I should do.

637
00:40:56,280 --> 00:40:58,640
Because the principle is to value human life.

638
00:40:58,640 --> 00:41:00,160
The principle is to value intelligence.

639
00:41:00,160 --> 00:41:05,440
The principle is to make sure that all of us on this little spaceship traveling through

640
00:41:05,440 --> 00:41:10,400
the universe, that we have the ability to reach the stars one day.

641
00:41:10,400 --> 00:41:15,260
Getting a little poetic here, but the principle here is to preserve intelligence and human

642
00:41:15,260 --> 00:41:18,680
life in general, and not just human life, life in general.

643
00:41:18,680 --> 00:41:25,160
Being more specific here, I would continue running what I do today as a VC fund and I

644
00:41:25,160 --> 00:41:26,580
would do a lot more of it.

645
00:41:26,580 --> 00:41:31,760
If I have a billion dollars, well, then I can do almost 10 times more than I'm doing

646
00:41:31,760 --> 00:41:32,920
right now with 85.

647
00:41:32,920 --> 00:41:37,080
I would try to invest beyond what I'm investing in right now.

648
00:41:37,080 --> 00:41:40,600
I'm focused on AI in healthcare, my partner AI in enterprise.

649
00:41:40,600 --> 00:41:44,840
We might start doing AI in biotech, AI in climate tech.

650
00:41:44,840 --> 00:41:46,080
We're focused on the US right now.

651
00:41:46,080 --> 00:41:47,560
We're open to Canada.

652
00:41:47,560 --> 00:41:49,080
We might start doing other geographies.

653
00:41:49,080 --> 00:41:51,920
I would love to invest more in emerging countries.

654
00:41:51,920 --> 00:41:56,440
India and Brazil are national fits for me, given my heritage, given my connections, given

655
00:41:56,440 --> 00:41:58,000
my knowledge of those two markets.

656
00:41:58,000 --> 00:42:02,920
But I would love to be able to do more things in Singapore, in the UK, perhaps even in other

657
00:42:02,920 --> 00:42:05,840
countries where we are comfortable operating.

658
00:42:05,840 --> 00:42:08,200
I would love to build a bigger team.

659
00:42:08,200 --> 00:42:11,280
That's a necessity actually, not a desire.

660
00:42:11,280 --> 00:42:13,320
It would be absolutely a requirement.

661
00:42:13,320 --> 00:42:15,640
I would like to perhaps expand.

662
00:42:15,640 --> 00:42:19,640
We want to be early stage, but with a billion dollars, it becomes tricky to be an early

663
00:42:19,640 --> 00:42:20,760
stage fund.

664
00:42:20,760 --> 00:42:25,280
Maybe we would become a seed series A and series B and start leading those deals.

665
00:42:25,280 --> 00:42:29,240
With a billion, you probably will have to go beyond series B. That is not what I was

666
00:42:29,240 --> 00:42:31,960
thinking, but we would have to seriously consider it.

667
00:42:31,960 --> 00:42:35,560
I would like to build the type of fund that that's what I'm focused on right now that

668
00:42:35,560 --> 00:42:37,720
does really good and really well.

669
00:42:37,720 --> 00:42:44,640
So 10X returns, but also in some ways, the type of companies we're investing in create

670
00:42:44,640 --> 00:42:45,640
value.

671
00:42:45,640 --> 00:42:50,960
I believe very much that if you create value, you get valuation.

672
00:42:50,960 --> 00:42:56,120
There's unfortunate for capitalism, those two things are not one and the same, but we

673
00:42:56,120 --> 00:43:00,840
can make sure that we operate so that they both help each other.

674
00:43:00,840 --> 00:43:04,320
There are ways of making money in this world that do not create value.

675
00:43:04,320 --> 00:43:09,600
It sounds like you have found your eikigai, which is the Japanese concept of purpose.

676
00:43:09,600 --> 00:43:13,720
And it's the intersection of where you're good at, where you love to do, and what people

677
00:43:13,720 --> 00:43:14,960
will pay you for.

678
00:43:14,960 --> 00:43:17,160
I want to be mindful of the time, Amit.

679
00:43:17,160 --> 00:43:18,760
Do you have time for one more question?

680
00:43:18,760 --> 00:43:20,480
If not, we can end it here.

681
00:43:20,480 --> 00:43:21,480
Let's do it.

682
00:43:21,480 --> 00:43:22,480
Let's do it.

683
00:43:22,480 --> 00:43:23,480
It's been an honor, Shad.

684
00:43:23,480 --> 00:43:24,480
You're very kind.

685
00:43:24,480 --> 00:43:27,120
You're very tough on your questions and very kind on your comments.

686
00:43:27,120 --> 00:43:28,360
What makes you resilient?

687
00:43:28,360 --> 00:43:31,240
This is something I've been thinking about for my kids.

688
00:43:31,240 --> 00:43:37,440
My previous answer would be you have to go through adversity to be resilient.

689
00:43:37,440 --> 00:43:42,720
Looking at the studies and the data out there, from what I've found, resilience comes from

690
00:43:42,720 --> 00:43:49,880
a good internal and external support system and doesn't necessarily require experiencing

691
00:43:49,880 --> 00:43:55,240
obstacles and overcoming them in a healthy fashion without maladaptive behavior.

692
00:43:55,240 --> 00:43:57,960
What makes you resilient?

693
00:43:57,960 --> 00:44:00,240
I might be changing your answer by saying that.

694
00:44:00,240 --> 00:44:01,240
No, no, no.

695
00:44:01,240 --> 00:44:02,940
I think you're right.

696
00:44:02,940 --> 00:44:04,720
No man is an island.

697
00:44:04,720 --> 00:44:10,200
I think part of it is yes, you're internal, somewhat shaped by your experiences, not always.

698
00:44:10,200 --> 00:44:14,880
I do think you can learn from other people's experiences, by the way, just because I'll

699
00:44:14,880 --> 00:44:21,400
take a very stupid example, but you wouldn't go and jump into a well.

700
00:44:21,400 --> 00:44:23,440
That's by the way a proverb in Hindi.

701
00:44:23,440 --> 00:44:26,640
The reason you don't do that is because you know that jumping into a well for the last

702
00:44:26,640 --> 00:44:29,000
guy who did it didn't turn out as well.

703
00:44:29,000 --> 00:44:31,640
You learn from somebody else's experience.

704
00:44:31,640 --> 00:44:36,440
I think there's a component here of what you experience yourself, what you learn from others,

705
00:44:36,440 --> 00:44:39,760
what you learn from others by watching them, what you learn from others by reading in them,

706
00:44:39,760 --> 00:44:43,800
what you learn from others by just what other people share with you.

707
00:44:43,800 --> 00:44:49,240
There's a component of resilience that is absolutely the village around you.

708
00:44:49,240 --> 00:44:54,520
Now there's obviously situations, incredible stories of people who beat all kinds of odds

709
00:44:54,520 --> 00:44:56,960
with very little support systems.

710
00:44:56,960 --> 00:45:01,100
There are also stories of people with a lot of support systems who don't get the amount

711
00:45:01,100 --> 00:45:04,800
of resilience perhaps that they were hoping for.

712
00:45:04,800 --> 00:45:05,800
There's a spectrum.

713
00:45:05,800 --> 00:45:06,840
It's very contextual.

714
00:45:06,840 --> 00:45:10,560
If you have resilience in one area, you may not have as much resilience in another.

715
00:45:10,560 --> 00:45:16,000
I'm not aware of those studies you might be, but I am willing to bet you that physical

716
00:45:16,000 --> 00:45:21,720
and mental resilience, there's a correlation, but they're not necessarily completely connected.

717
00:45:21,720 --> 00:45:29,600
I may be great at handling stress at work, but be terrible at handling stress when I'm

718
00:45:29,600 --> 00:45:30,600
running.

719
00:45:30,600 --> 00:45:32,800
It's not a perfect correlation.

720
00:45:32,800 --> 00:45:36,200
For me personally, I think it's all of the above.

721
00:45:36,200 --> 00:45:38,400
I think resilience is a muscle in some ways.

722
00:45:38,400 --> 00:45:40,320
You have to keep exercising it.

723
00:45:40,320 --> 00:45:43,080
It's easy to fall too comfortable.

724
00:45:43,080 --> 00:45:45,760
I'm grateful for all the people around me.

725
00:45:45,760 --> 00:45:52,000
First and foremost, my wife, she's my rock and she didn't pay me to say all of this.

726
00:45:52,000 --> 00:45:56,200
When she hears this, I'll hopefully make some brownie points, but it's really true.

727
00:45:56,200 --> 00:46:01,820
She gives me a lot of wisdom and she gives me a lot of direction and she helps me how

728
00:46:01,820 --> 00:46:02,820
to be better.

729
00:46:02,820 --> 00:46:08,000
Obviously, my parents, they were who gave me the foundation.

730
00:46:08,000 --> 00:46:13,320
My teachers, my mentors, both in the past and in the present, my friends, both in the

731
00:46:13,320 --> 00:46:17,040
past and in the present.

732
00:46:17,040 --> 00:46:22,400
There's people that I've never met, some of them alive, some of them not alive, who are

733
00:46:22,400 --> 00:46:23,800
role models.

734
00:46:23,800 --> 00:46:27,960
Obviously, I'm building Tao Ventures with a team.

735
00:46:27,960 --> 00:46:29,640
My co-founder for sure.

736
00:46:29,640 --> 00:46:34,600
The reason we are building this fund together is because we know we are a good team.

737
00:46:34,600 --> 00:46:35,920
We are a good partnership.

738
00:46:35,920 --> 00:46:41,400
We can keep each other both accountable, but also bring the best in each other.

739
00:46:41,400 --> 00:46:43,120
Big shout out here to you, Sanjay.

740
00:46:43,120 --> 00:46:46,160
I don't think I'm perfect at this, Rishad.

741
00:46:46,160 --> 00:46:47,760
Nobody is, to be honest.

742
00:46:47,760 --> 00:46:53,760
The day I believe I am or that I've hit my limits, then that means that I'll start failing.

743
00:46:53,760 --> 00:46:56,120
It's a good reminder to myself, there's always more to learn.

744
00:46:56,120 --> 00:46:57,920
There's always more to unlearn.

745
00:46:57,920 --> 00:47:04,880
I discover every day as I go by that something I knew, there's far more to learn about it.

746
00:47:04,880 --> 00:47:05,880
Resilience included.

747
00:47:05,880 --> 00:47:07,760
It's been great talking to you, Amit.

748
00:47:07,760 --> 00:47:10,480
Thanks so much for coming on the show today.

749
00:47:10,480 --> 00:47:15,760
We didn't get to talk too much about investing or other topics I had in mind.

750
00:47:15,760 --> 00:47:18,760
Would love to do it again in the new year.

751
00:47:18,760 --> 00:47:19,760
Absolutely.

752
00:47:19,760 --> 00:47:20,760
Thank you for having me.

753
00:47:20,760 --> 00:47:22,520
Thank you to all of you watching us.

754
00:47:22,520 --> 00:47:25,080
We are TaoVentures.com.

755
00:47:25,080 --> 00:47:28,480
Feel free to check us out.

756
00:47:28,480 --> 00:47:31,480
We read everything that reaches our inbox.

757
00:47:31,480 --> 00:47:32,960
So welcome to reach out.

758
00:47:32,960 --> 00:47:36,640
I'm not able to respond to everyone, but I will certainly read it.

759
00:47:36,640 --> 00:47:40,840
Once again, we're a focused seed stage, primarily enterprise and healthcare, investing in the

760
00:47:40,840 --> 00:47:42,680
US, but very much open to Canada.

761
00:47:42,680 --> 00:47:43,680
Awesome.

762
00:47:43,680 --> 00:47:55,320
Thanks, Amit.