You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: content/posts/personal-assistants.md
+29-21Lines changed: 29 additions & 21 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -10,11 +10,11 @@ images:
10
10
tags:
11
11
---
12
12
13
-
okay I’ll admit that was clickbait. an LLM did not find me a gf or bf at all.
13
+
okay I’ll admit it: clickbait. no LLM found me gf or bf _(yet)_
14
14
15
-
so clawd is the shiny new thing this week on twitter. so what? why is that interesting?
15
+
so clawd is the shiny new thing on X this week. so what? why is that interesting?
16
16
17
-
I think this new wave of adoption of agents will drive some very interesting changes in the AI bubble.
17
+
I think this new wave of adoption is going to warp pricing + reliability in ways people aren't pricing in yet.
18
18
19
19
so let’s discuss that instead of why I don’t have a partner ◝(ᵔᗜᵔ)◜
20
20
@@ -30,7 +30,7 @@ tags:
30
30
31
31
the thing with clawd is that you _can_ own your data, even if you don't right now. or at least you have the possibility.
32
32
33
-
why do i say this? clawd let's you use any LLM provider, and holds your data in whatever device you're running it on. it also holds the credentials to all the services you connect it to.
33
+
why do i say this? clawd lets you use any LLM provider, and holds your data in whatever device you're running it on. it also holds the credentials to all the services you connect it to._(yes, scary)_
34
34
35
35
so even if openai, anthropic or whoever has the hottest SOTA this week goes down, you can just swap to another model. so if you’re filthy rich you could just host some open-source SOTA yourself and keep it running even if we nuke each other out for some time.
36
36
@@ -42,29 +42,27 @@ TLDR:
42
42
43
43
# providers will price this in
44
44
45
-
if you didn't know it already let me break you the news. the AI industry is heavily subsidized by almost anyone they can borrow money from by selling them the promise of a path to profitability in the future.
45
+
if you didn't know it already let me break the news to you. the AI industry is heavily subsidized by almost anyone who'll lend them money by selling them the promise of a path to profitability in the future.
46
46
47
47
that's why Microsoft burns cash like you could just print money.. oh wait. sorry, I meant that they burn money like Microsoft _owned_ the federal reserve.
48
48
49
-
it's a long-term bet. it's an arms-race that requires patience and a huge pile of cash, or rather a pile of heads willing to borrow you that cash.
49
+
it's a long-term bet. it's an arms-race that requires patience and a huge pile of cash, or rather a pile of heads willing to lend you that cash.
50
50
51
-
so how do they even convince those heads to borrow money from them for a completely uncertain endeavor?
51
+
so how do they even convince those heads to lend them money for a completely uncertain endeavor?
52
52
53
-
they promise green pastures far far away...
53
+
the pitch is basically: _"subsidize usage now, capture the workflow, monetize later."_
54
54
55
-
and what lies in these promised lands?
56
-
57
-
hoards of users that use the harness you tailor made for your model. locked into your ecosystem.
55
+
if you don't own the harness, you don't own the user. and if you don't own the user, those subsidies look less like strategy and more like lighting money on fire.
58
56
59
-
oh right, your model, that massive pile of floats (weights) that can be hold on an SD card and costs dozens of billions, or even hundreds of billions to train.
57
+
and the 'asset' they're trying to amortize is your model: you know, that massive pile of floats (weights) that can be held on an SD card and costs dozens of billions, or even hundreds of billions to train. and can be just copied infinitely.
60
58
61
59
so we don't really know how much cash they're burning vs revenue.
62
60
63
-
but using an h100 for a whole day for $200 a month? that's a steal right?
61
+
but using an h100 for a whole day for $200 a month? that's a steal, right?
64
62
65
63
bc it kinda is, there's no way they're making money right now. everyone knows this.
66
64
67
-
everyone on twitter and their mom are tweeting on how to use the Oauth credentials of low-price/high-quota plans to avoid paying hundreds or thousands in api credits.
65
+
everyone on X and their mom are posting on how to use the OAuth credentials of low-price/high-quota plans to avoid paying hundreds or thousands in api credits.
68
66
69
67
all of that cash they're burning?
70
68
all of that is so you use their harness, buy into their ecosystem and they collect data about how you use it for fine-tuning and improving their models. that’s their moat.
@@ -73,9 +71,9 @@ TLDR:
73
71
74
72
how do you think they're gonna justify to their investors subsidizing **_you_** if they don't own the harness you use and you give them scattered usage data?
75
73
76
-
I mean, you use their platform so they _do_ have your data when you generate tokens, but you can just switch up to another provider to another SOTA model without a massive impact in your day to day use.
74
+
I mean, you use their platform so they _do_ have your data when you generate tokens, but you can just switch to another provider that has a SOTA model without a massive impact in your day to day use.
77
75
78
-
this is somehting that has been said for a while, but now more than ever it's going to be felt in inference provider's pricing.
76
+
this is something that has been said for a while, but now more than ever it's going to be felt in inference providers' pricing.
79
77
80
78
that $200 subscription for unlimited model use? who knows how long it's going to last.
81
79
@@ -85,21 +83,31 @@ TLDR:
85
83
86
84
so we will probably see unlimited or high-usage flat tiers going up or disappearing, and API prices going up. my bet is before 2027.
87
85
88
-
# all models will be more reliable in high-value tasks for real users
86
+
# all models will be more reliable and secure in high-value tasks for real users
89
87
90
88
yes, we do have some standards at this point to establish how a model should talk to a tool, or API (ex: MCP, openai standard), but the agentic flow and the way they perform tasks are not the same along most providers. and the range of tasks each model can do differs.
91
89
92
-
some models perform agentic tasks better than others based solely on the harness. for example codex performs way better on the codex-cli harness than on say claude-code or open code.
90
+
some models perform agentic tasks better than others based solely on the harness. for example codex performs way better on the codex-cli harness than on say claude-code or opencode.
93
91
94
92
the cool thing with apps like clawd is that it will generate new benchmarks of real high-value tasks for the end users that all providers will try to optimize for.
95
93
96
94
if it ends up being something people keep using companies are gonna end up training on that data and optimizing for those use cases.
97
95
98
96
so we will end up with a very robust ecosystem where agents can reliably perform a lot of tasks users really want done by their agents daily.
99
97
100
-
it will enforce a soft-default standard on how agents should act, and what minimum range of tasks a model should be able to bring to the table to even be considered to be SOTA-ish.
98
+
it will enforce a soft-default standard on how agents should act, and what minimum range of tasks a model should be able to bring to the table to even be considered SOTA-ish.
99
+
100
+
# the security tradeoffs won't turn off mainstream adoption
101
+
102
+
since clawd got so popular a lot of inexperienced or non-technical users tried it. a lot of vulnerabilities were discovered and a lot of people got hacked.
103
+
104
+
lots of people are getting hacked, and that _will_ keep happening. but i don't think this will be a turnoff for adoption. there's so much to gain that people won't care the tradeoff.
105
+
106
+
ofc this also means that there are going to be a lot more guardrails to approximate 'foolproof' usage of these tools. and eventually security oriented clawdbot copies will be made into products for non-technical people.
107
+
108
+
guardrails will be tested in the real world and will get more and more robust.
101
109
102
-
# the best thing about owning the shoggoth harness :D
110
+
# the best thing about owning the shoggoth harness :D
103
111
104
112
so I wanted to have a personal assistant for gamification of goals for some time.
105
113
@@ -123,4 +131,4 @@ TLDR:
123
131
124
132
so even though _it is_ the flashy new thing, it's a good thing!
125
133
126
-
and this new massive wave of adoption it's a really good first step on real world agent-reliability and data ownership.
134
+
and this new massive wave of adoption is a really good first step on real world agent-reliability and data ownership.
0 commit comments