summaryrefslogtreecommitdiffstats
path: root/2025/captions/emacsconf-2025-private-ai--emacs-and-private-ai-a-great-match--aaron-grothe--...
diff options
context:
space:
mode:
authorSacha Chua <sacha@sachachua.com>2025-12-28 21:23:56 -0500
committerSacha Chua <sacha@sachachua.com>2025-12-28 21:23:56 -0500
commitff518d86cf4eabbcb1603596d836ce28cee7367b (patch)
tree425fba7a315d5fe3df712c6672cdbd744981af61 /2025/captions/emacsconf-2025-private-ai--emacs-and-private-ai-a-great-match--aaron-grothe--main.vtt
parent41d07038edc8d9b9f1af70ec75e5a1c62d17bb60 (diff)
downloademacsconf-wiki-ff518d86cf4eabbcb1603596d836ce28cee7367b.tar.xz
emacsconf-wiki-ff518d86cf4eabbcb1603596d836ce28cee7367b.zip
updatesHEADmaster
Diffstat (limited to '2025/captions/emacsconf-2025-private-ai--emacs-and-private-ai-a-great-match--aaron-grothe--main.vtt')
-rw-r--r--2025/captions/emacsconf-2025-private-ai--emacs-and-private-ai-a-great-match--aaron-grothe--main.vtt573
1 files changed, 323 insertions, 250 deletions
diff --git a/2025/captions/emacsconf-2025-private-ai--emacs-and-private-ai-a-great-match--aaron-grothe--main.vtt b/2025/captions/emacsconf-2025-private-ai--emacs-and-private-ai-a-great-match--aaron-grothe--main.vtt
index a8e76986..dc4ba1b1 100644
--- a/2025/captions/emacsconf-2025-private-ai--emacs-and-private-ai-a-great-match--aaron-grothe--main.vtt
+++ b/2025/captions/emacsconf-2025-private-ai--emacs-and-private-ai-a-great-match--aaron-grothe--main.vtt
@@ -1,5 +1,7 @@
WEBVTT
+NOTE Introduction
+
00:00:00.000 --> 00:00:04.859
Hey, everybody. Welcome from frigid Omaha, Nebraska.
@@ -10,7 +12,7 @@ I'm just going to kick off my talk here,
and we'll see how it all goes. Thanks for attending.
00:00:23.900 --> 00:00:26.939
-So the slides will be available on my site, growthy.us,
+So the slides will be available on my site, https://grothe.us,
00:00:26.940 --> 00:00:29.899
in the presentation section tonight or tomorrow.
@@ -39,8 +41,10 @@ Really hope some of you give it a shot
00:00:45.820 --> 00:00:48.179
and learn something along the way.
+NOTE Overview of talk
+
00:00:48.180 --> 00:00:50.379
-So the overview of the talk.
+So the overview of the talk
00:00:50.380 --> 00:00:54.939
broke down these basic bullet points of why private AI,
@@ -54,6 +58,8 @@ pieces for an AI Emacs solution,
00:01:02.740 --> 00:01:08.059
a demo of a minimal viable product, and the summary.
+NOTE Why private AI?
+
00:01:08.060 --> 00:01:10.779
Why private AI? This is pretty simple.
@@ -67,7 +73,7 @@ for any AI system you're currently using.
If you're using the free tiers, your queries,
00:01:17.020 --> 00:01:18.619
-code uploaded information
+code, uploaded information
00:01:18.620 --> 00:01:20.699
is being used to train the models.
@@ -139,7 +145,7 @@ The problem with that is now
there's less data going to Stack Overflow
00:02:10.380 --> 00:02:12.979
-for the AI to get. vicious cycle,
+for the AI to get. Vicious cycle,
00:02:12.980 --> 00:02:14.619
especially when you start looking at
@@ -160,7 +166,7 @@ Right now, these services are being heavily subsidized.
If you're paying Claude $20 a month,
00:02:29.420 --> 00:02:32.579
-it is not costing Claude, those guys $20 a month
+it is not costing Claude, those guys, $20 a month
00:02:32.580 --> 00:02:34.099
to host all the infrastructure
@@ -219,6 +225,8 @@ are going to have no option but to move to private AI
00:03:11.900 --> 00:03:16.019
or give up on AI overall.
+NOTE What do I need for private AI?
+
00:03:16.020 --> 00:03:18.659
What do you need to be able to do private AI?
@@ -316,7 +324,7 @@ And Raspberry Pi's new AI top is about 18 teraflops
and is $70 on top of the cost of Raspberry Pi 5.
00:04:51.220 --> 00:04:56.059
-Keep in mind Raspberry recently
+Keep in mind, Raspberry recently
00:04:56.060 --> 00:04:59.499
raised the cost of their Pi 5s because of RAM pricing,
@@ -342,8 +350,13 @@ A lot of people are going to have PCs on their desks.
00:05:11.180 --> 00:05:13.459
They're going to run a decent private AI
-00:05:13.460 --> 00:05:18.059
-without much issue. So for Emacs and private AI,
+00:05:13.460 --> 00:05:16.347
+without much issue.
+
+NOTE Emacs and private AI
+
+00:05:16.348 --> 00:05:18.059
+So for Emacs and private AI,
00:05:18.060 --> 00:05:20.139
there's a couple popular solutions.
@@ -364,7 +377,7 @@ It's just, quite honestly, chef's kiss,
just a beautifully well-done piece of software.
00:05:31.060 --> 00:05:33.859
-OllamaBuddy has more features,
+Ollama Buddy has more features,
00:05:33.860 --> 00:05:36.259
a menu interface, has quick access
@@ -382,7 +395,7 @@ This is the one that you spend a little more time with,
but you also get a little bit more back from it.
00:05:43.940 --> 00:05:49.419
-Elama is another one, has some really good features to it,
+Ellama is another one, has some really good features to it,
00:05:49.420 --> 00:05:51.059
more different capabilities,
@@ -391,13 +404,13 @@ more different capabilities,
but it's a different set of rules and capabilities to it.
00:05:54.980 --> 00:05:59.179
-Itermac, which is programming with your AI and Emacs.
+Aidermac, which is programming with your AI and Emacs.
00:05:59.180 --> 00:06:01.219
The closest thing I can come up
00:06:01.220 --> 00:06:04.139
-to comparing this to is Cursor, except it's an Emacs.
+to comparing this to is Cursor, except it's in Emacs.
00:06:04.140 --> 00:06:05.659
It's really quite well done.
@@ -414,6 +427,8 @@ If you go out to GitHub, type Emacs AI,
00:06:10.820 --> 00:06:13.219
you'll find a lot of different options.
+NOTE Pieces for an AI Emacs solution
+
00:06:13.220 --> 00:06:18.459
So what is a minimal viable product that can be done?
@@ -453,16 +468,16 @@ while it runs on a bunch of different systems.
00:06:48.700 --> 00:06:51.299
And Gptel, which is an easy plug-in for Emacs,
-00:06:51.300 --> 00:06:54.979
+00:06:51.300 --> 00:06:56.339
which we talked about in the last slide a bit.
-00:06:54.980 --> 00:07:00.179
+00:06:56.340 --> 00:07:00.179
So setting up the LLM, you have to just go out
-00:07:00.180 --> 00:07:01.699
-and just hit the a page for it
+00:07:00.180 --> 00:07:03.542
+and just hit a page for it
-00:07:01.700 --> 00:07:05.099
+00:07:03.543 --> 00:07:05.099
and go out and do a wget of it.
00:07:05.100 --> 00:07:07.099
@@ -487,7 +502,7 @@ And let's just take a look at it.
I've actually downloaded several of them,
00:07:22.900 --> 00:07:25.699
-but let's go ahead and just run lava 3.2b
+but let's go ahead and just run llama 3.2-1b
00:07:25.700 --> 00:07:31.179
with the 3 billion instructions. And that's it firing up.
@@ -505,10 +520,10 @@ So once you do that, you have to install gptel and emacs.
That's as simple as firing up emacs,
00:07:45.660 --> 00:07:48.339
-doing the meta x install package,
+doing the M-x install-package,
00:07:48.340 --> 00:07:49.779
-and then just typing gptel
+and then just typing gptel,
00:07:49.780 --> 00:07:51.499
if you have your repository set up right,
@@ -516,17 +531,19 @@ if you have your repository set up right,
00:07:51.500 --> 00:07:52.299
which hopefully you do.
-00:07:52.300 --> 00:07:54.499
+00:07:52.300 --> 00:07:56.339
And then you just go ahead and have it.
-00:07:54.500 --> 00:07:58.139
+NOTE Config file
+
+00:07:56.340 --> 00:07:58.139
You also have to set up a config file.
00:07:58.140 --> 00:08:01.739
Here's my example config file as it currently set up,
00:08:01.740 --> 00:08:04.019
-requiring ensuring Gptel is loaded,
+requiring, ensuring Gptel is loaded,
00:08:04.020 --> 00:08:05.899
defining the Llamafile backend.
@@ -585,6 +602,8 @@ which is really kind of cool.
00:08:48.700 --> 00:08:52.099
But, uh, that's all that takes.
+NOTE Demo: Who was David Bowie?
+
00:08:52.100 --> 00:09:03.779
So let's go ahead and go to a quick test of it.
@@ -619,10 +638,10 @@ And we can just do, we can either do
the alt X and send the gptel-send,
00:09:36.060 --> 00:09:37.979
-or we can just do control C and hit enter.
+or we can just do C-c and hit enter.
00:09:37.980 --> 00:09:39.139
-We'll just do control C and enter.
+We'll just do C-c and enter.
00:09:39.140 --> 00:09:43.659
And now it's going ahead and hitting our local AI system
@@ -648,6 +667,8 @@ and other information about him.
00:10:10.140 --> 00:10:21.699
So very, very happy with that.
+NOTE Hallucinations
+
00:10:21.700 --> 00:10:23.539
One thing to keep in mind is you look at things
@@ -672,6 +693,8 @@ Is it not actually on this one?
00:10:39.740 --> 00:10:42.179
Alright, so that's the first question I always ask one.
+NOTE Next question: What are sea monkeys?
+
00:10:42.180 --> 00:10:44.659
The next one is what are sea monkeys?
@@ -702,10 +725,10 @@ I did not see that coming.
00:11:21.460 --> 00:11:24.139
Here's one that some people say is a really good one
-00:11:24.140 --> 00:11:25.739
-to ask ours in strawberry.
+00:11:24.140 --> 00:11:42.779
+to ask. Rs in "strawberry."
-00:11:25.740 --> 00:11:46.179
+00:11:42.780 --> 00:11:46.179
All right, now she's going off the reservation.
00:11:46.180 --> 00:11:48.139
@@ -714,11 +737,13 @@ She's going in a different direction.
00:11:48.140 --> 00:11:49.979
Let me go ahead and reopen that again,
-00:11:49.980 --> 00:11:52.979
-because it's went down a bad hole there for a second.
+00:11:49.980 --> 00:11:57.179
+because it went down a bad hole there for a second.
+
+NOTE Writing Hello World in Emacs Lisp
-00:11:52.980 --> 00:11:58.419
-Let me ask it to do write hello world in Emacs list.
+00:11:57.180 --> 00:11:58.419
+Let me ask it to write hello world in Emacs Lisp.
00:11:58.420 --> 00:12:10.419
Yep, that works. So the point being here,
@@ -741,6 +766,8 @@ We also have the default system here as well.
00:12:25.260 --> 00:12:32.579
So not that bad.
+NOTE Pieces for a better solution
+
00:12:32.580 --> 00:12:35.379
That's a basic solution, that's a basic setup
@@ -778,7 +805,7 @@ But if you want a better solution,
I recommend Ollama or LM Studio.
00:12:59.940 --> 00:13:01.899
-They're both more capable than llama file.
+They're both more capable than Llamafile.
00:13:01.900 --> 00:13:03.859
They can accept a lot of different models.
@@ -805,7 +832,7 @@ so you can start uploading your code, your information,
and actually being able to do analysis of it.
00:13:20.140 --> 00:13:23.539
-OpenWebUI provides more capabilities.
+Open WebUI provides more capabilities.
00:13:23.540 --> 00:13:24.859
It provides an interface that's similar
@@ -814,7 +841,7 @@ It provides an interface that's similar
to what you're used to seeing
00:13:25.900 --> 00:13:28.179
-for chat, GPT, and the other systems.
+for ChatGPT and the other systems.
00:13:28.180 --> 00:13:29.419
It's really quite well done.
@@ -826,7 +853,9 @@ And once again, gptel, I have to mention that
because that's the one I really kind of like.
00:13:34.780 --> 00:13:36.899
-And OlamaBuddy is also another really nice one.
+And Ollama Buddy is also another really nice one.
+
+NOTE What about the license?
00:13:36.900 --> 00:13:41.019
So what about the licensing of these models?
@@ -841,7 +870,7 @@ a model and doing this stuff.
Let's take a look at a couple of highlights
00:13:46.580 --> 00:13:49.379
-from the MetaLlama 3 community license scale.
+from the Meta Llama 3 community license scale.
00:13:49.380 --> 00:13:52.579
If your service exceeds 700 million monthly users,
@@ -892,7 +921,7 @@ And there are some very cool ones out there.
There's people who've done things
00:14:25.260 --> 00:14:29.579
-to try and make the llama bee less, what's the phrase,
+to try and make the Llama be less, what's the phrase,
00:14:29.580 --> 00:14:31.939
ethical if you're doing penetration testing research
@@ -918,19 +947,21 @@ It's designed to keep it to research and development.
00:14:45.220 --> 00:14:46.739
You can't use it commercially.
-00:14:46.740 --> 00:14:50.419
+00:14:46.740 --> 00:14:51.792
So it's designed to clearly delineate
-00:14:50.420 --> 00:14:52.939
+00:14:51.793 --> 00:14:52.939
between research and development
00:14:52.940 --> 00:14:54.259
and somebody trying to actually build
-00:14:54.260 --> 00:14:55.379
+00:14:54.260 --> 00:14:56.579
something on top of it.
-00:14:55.380 --> 00:14:57.979
+NOTE Are there open source data model options?
+
+00:14:56.580 --> 00:14:57.979
And another question I get asked is,
00:14:57.980 --> 00:14:59.899
@@ -951,17 +982,19 @@ but it's very cool to see where it's going.
00:15:09.020 --> 00:15:11.339
And it's another thing that's just going forward.
-00:15:11.340 --> 00:15:13.379
+00:15:11.340 --> 00:15:14.519
It's under the MIT license.
-00:15:13.380 --> 00:15:15.819
+NOTE Things to know
+
+00:15:14.520 --> 00:15:15.819
Some things to know to help you
00:15:15.820 --> 00:15:17.499
have a better experience with this.
00:15:17.500 --> 00:15:21.059
-Get a Llama and OpenWebUI working by themselves,
+Get ollama and Open WebUI working by themselves,
00:15:21.060 --> 00:15:22.659
then set up your config file.
@@ -970,7 +1003,7 @@ then set up your config file.
I was fighting both at the same time,
00:15:24.820 --> 00:15:26.699
-and it turned out I had a problem with my LLAMA.
+and it turned out I had a problem with my ollama.
00:15:26.700 --> 00:15:28.899
I had a conflict, so that was what my problem is.
@@ -994,7 +1027,7 @@ You can create models, you can put things like
help desk developers and stuff like that, breaking it out.
00:15:46.420 --> 00:15:51.019
-The Hacker News has a how to build a $300 AI computer.
+The Hacker Noon has a how to build a $300 AI computer.
00:15:51.020 --> 00:15:52.859
This is for March 2024,
@@ -1015,7 +1048,7 @@ inside my Acer Aspire,
that's where I got the idea doing that.
00:16:04.420 --> 00:16:06.739
-Make sure you do the RockM stuff correctly
+Make sure you do the ROCm stuff correctly
00:16:06.740 --> 00:16:09.899
to get the GUI extensions. But it's just really good stuff.
@@ -1024,37 +1057,37 @@ to get the GUI extensions. But it's just really good stuff.
You don't need a great GPU or CPU to get started.
00:16:13.060 --> 00:16:14.819
-Smaller models like Tiny Llama
+Smaller models like tinyllama
-00:16:14.820 --> 00:16:16.179
+00:16:14.820 --> 00:16:16.819
can run on very small systems.
-00:16:16.180 --> 00:16:18.499
+00:16:16.820 --> 00:16:19.042
It gets you the ability to start playing with it
-00:16:18.500 --> 00:16:21.619
+00:16:19.043 --> 00:16:21.619
and start experimenting and figure out if that's for you
00:16:21.620 --> 00:16:23.379
and to move forward with it.
00:16:23.380 --> 00:16:29.219
-The AMD Ryzen AI Max 395 plus is a mini PC
+The AMD Ryzen AI Max+ 395 is a mini PC
00:16:29.220 --> 00:16:31.179
makes it really nice dedicated host.
-00:16:31.180 --> 00:16:34.619
-You used to be able to buy these for about $1200 now
+00:16:31.180 --> 00:16:34.078
+You used to be able to buy these for about $1200.
-00:16:34.620 --> 00:16:35.579
-with the RAM price increase,
+00:16:34.079 --> 00:16:35.579
+Now with the RAM price increase,
-00:16:35.580 --> 00:16:38.779
-you want to get 120 gig when you're pushing two brands so.
+00:16:35.580 --> 00:16:38.458
+you want to get 120 gig when you're pushing two brands,
-00:16:38.780 --> 00:16:40.739
-It gets a little tighter.
+00:16:38.459 --> 00:16:40.739
+so it gets a little tighter.
00:16:40.740 --> 00:16:44.099
Macs work remarkably well with AI.
@@ -1093,16 +1126,16 @@ So Mac Minis and the Mac Ultras and stuff like that
tend to work really well for that.
00:17:06.100 --> 00:17:09.779
-Alex Ziskin on YouTube has a channel.
+Alex Ziskind on YouTube has a channel.
00:17:09.780 --> 00:17:11.899
He does a lot of AI performance benchmarking,
00:17:11.900 --> 00:17:14.819
-like I load a 70 billion parameter model
+like "I load a 70 billion parameter model
00:17:14.820 --> 00:17:16.699
-on this mini PC and stuff like that.
+on this mini PC" and stuff like that.
00:17:16.700 --> 00:17:19.019
It's a lot of fun and interesting stuff there.
@@ -1129,7 +1162,7 @@ It gives you capabilities to start training stuff
that you couldn't do with like the big ones.
00:17:35.900 --> 00:17:38.059
-Even with in terms of fine tuning and stuff,
+Even with in terms of fine-tuning and stuff,
00:17:38.060 --> 00:17:40.539
it's remarkable to see where that space is coming along
@@ -1138,12 +1171,12 @@ it's remarkable to see where that space is coming along
in the next year or so.
00:17:41.740 --> 00:17:46.219
-Hugging Face Co has pointers to tons of AI models.
+HuggingFace.co has pointers to tons of AI models.
-00:17:46.220 --> 00:17:49.259
+00:17:46.220 --> 00:17:48.417
You'll find the one that works for you, hopefully there.
-00:17:49.260 --> 00:17:50.539
+00:17:48.418 --> 00:17:50.539
If you're doing cybersecurity,
00:17:50.540 --> 00:17:52.059
@@ -1174,7 +1207,7 @@ and run into those problems. So, That is my talk.
What I would like you to get out of that is,
00:18:17.220 --> 00:18:21.859
-if you haven't tried it, give GPTEL and LlamaFile a shot.
+if you haven't tried it, give Gptel and LlamaFile a shot.
00:18:21.860 --> 00:18:23.979
Fire up a little small AI instance,
@@ -1201,7 +1234,7 @@ Let me see if we got anything you want, Pat. You do.
You've got a few questions.
00:18:43.900 --> 00:18:48.059
-Hey, this is Corwin. Thank you so much. Thank you, Aaron.
+[Corwin]: Hey, this is Corwin. Thank you so much. Thank you, Aaron.
00:18:48.060 --> 00:18:50.339
What an awesome talk this was, actually.
@@ -1213,12 +1246,12 @@ If you don't have a camera,
I can get away with not having one too.
00:18:54.340 --> 00:18:56.299
-I've got, I'll turn the camera on.
+[Aaron]: I've got, I'll turn the camera on.
-00:18:56.300 --> 00:19:01.499
-Okay. All right. I'll turn mine back on. Here I come.
+00:18:56.300 --> 00:18:59.833
+[Corwin]: Okay. All right. I'll turn mine back on. Here I come.
-00:19:01.500 --> 00:19:03.139
+00:18:59.834 --> 00:19:03.139
Yeah, so there are a few questions,
00:19:03.140 --> 00:19:04.579
@@ -1252,7 +1285,7 @@ put infrastructure we have at home to use
and get more done with less.
00:19:34.020 --> 00:19:37.499
-Yeah, the data center impact's interesting
+[Aaron]: Yeah, the data center impact's interesting
00:19:37.500 --> 00:19:39.979
because there was a study a while ago.
@@ -1264,7 +1297,7 @@ Someone said every time you do a Gemini query,
it's like boiling a cup of water.
00:19:45.020 --> 00:19:48.619
-Yeah, I've heard that one too. So do you want to, you know,
+[Corwin]: Yeah, I've heard that one too. So do you want to, you know,
00:19:48.620 --> 00:19:51.699
I don't know how much direction you want.
@@ -1273,19 +1306,21 @@ I don't know how much direction you want.
I'd be very happy to read out the questions for you.
00:19:53.860 --> 00:19:55.219
-Yeah, that would be great.
+[Aaron]: Yeah, that would be great.
00:19:55.220 --> 00:19:57.619
I'm having trouble getting to that tab.
00:19:57.620 --> 00:20:02.779
-Okay, I'm there, so I'll put it into our chat too,
+[Corwin]: Okay, I'm there, so I'll put it into our chat too,
00:20:02.780 --> 00:20:07.419
so you can follow along if you'd like.
+NOTE Q: Why is the David Bowie question a good one for testing a model? e.g. does it fail in interesting ways?
+
00:20:07.420 --> 00:20:11.219
-The first question was, why is the David Bowie question
+[Corwin]: The first question was, why is the David Bowie question
00:20:11.220 --> 00:20:12.219
a good one to start with?
@@ -1293,23 +1328,23 @@ a good one to start with?
00:20:12.220 --> 00:20:14.419
Does it have interesting failure conditions
-00:20:14.420 --> 00:20:17.299
+00:20:14.420 --> 00:20:16.639
or what made you choose that?
-00:20:17.300 --> 00:20:21.979
-First off, huge fan of David Bowie.
+00:20:16.640 --> 00:20:21.979
+[Aaron]: First off, huge fan of David Bowie.
00:20:21.980 --> 00:20:24.499
But I came down to it really taught me a few things
00:20:24.500 --> 00:20:26.299
-about how old the models work
+about how the models work
00:20:26.300 --> 00:20:28.819
in terms of things like how many kids he had,
00:20:28.820 --> 00:20:31.779
-because deep seek, which is a very popular Chinese model
+because Deepseek, which is a very popular Chinese model
00:20:31.780 --> 00:20:33.179
that a lot of people are using now,
@@ -1324,37 +1359,37 @@ and he has like one son and one, one, I think,
two sons and a daughter or something like that.
00:20:40.900 --> 00:20:43.659
-so there's differences on that and it just goes over
+so there's differences on that, and it just goes over...
00:20:43.660 --> 00:20:45.299
there's a whole lot of stuff
00:20:45.300 --> 00:20:47.779
-because his story spans like 60 years
+because his story spans like 60 years,
00:20:47.780 --> 00:20:49.659
-so it gives a good good feedback
+so it gives good feedback.
00:20:49.660 --> 00:20:51.539
-that's the real main reason I asked that question
+That's the real main reason I asked that question
00:20:51.540 --> 00:20:53.699
-because I just needed one that sea monkeys I just picked
+because I just needed one... That sea monkeys, I just picked
00:20:53.700 --> 00:20:56.579
-because it was obscure and just always have right
+because it was obscure, and just always have, write,
00:20:56.580 --> 00:20:58.939
-I used to have it right hello world and forth
+I used to have it write hello world in forth
00:20:58.940 --> 00:21:01.019
-because I thought was an interesting one as well so
+because I thought was an interesting one as well.
00:21:01.020 --> 00:21:03.899
It's just picking random ones like that.
00:21:03.900 --> 00:21:06.499
-One question asked, sorry, a lot of models is,
+One question I ask a lot of models is,
00:21:06.500 --> 00:21:09.419
what is the closest star to the Earth?
@@ -1366,7 +1401,7 @@ Because most of them will say Alpha Centauri
or Proxima Centauri and not the sun.
00:21:13.740 --> 00:21:15.899
-And I have a whole nother talk
+And I have a whole 'nother talk
00:21:15.900 --> 00:21:17.899
where I just argue with the LLM
@@ -1377,19 +1412,21 @@ trying to say, hey, the sun is a star.
00:21:20.020 --> 00:21:26.579
And he just wouldn't accept it, so. What?
-00:21:26.580 --> 00:21:28.419
-Oh, I can hear that.
+00:21:26.580 --> 00:21:30.739
+Oh, I can... You're there.
+
+NOTE Q: What specific tasks do you use local AI for?
-00:21:28.420 --> 00:21:34.379
-So what specific tasks do you like to use your local AI?
+00:21:30.740 --> 00:21:34.379
+[Corwin]: So what specific tasks do you like to use your local AI?
00:21:34.380 --> 00:21:37.459
-I like to load a lot of my code into
+[Aaron]: I like to load a lot of my code into
-00:21:37.460 --> 00:21:39.739
+00:21:37.460 --> 00:21:39.099
and actually have it do analysis of it.
-00:21:39.740 --> 00:21:42.339
+00:21:39.100 --> 00:21:42.339
I was actually going through some code
00:21:42.340 --> 00:21:45.619
@@ -1428,17 +1465,19 @@ if you're doing cyber security researching.
00:22:04.980 --> 00:22:06.819
and you have your white papers
-00:22:06.820 --> 00:22:10.779
+00:22:06.820 --> 00:22:08.417
and stuff like that and stuff in there.
-00:22:10.780 --> 00:22:13.979
+00:22:08.418 --> 00:22:10.625
I've got a lot of that loaded into RAG
-00:22:13.980 --> 00:22:15.659
-in one model on my OpenWebUI system.
+00:22:10.626 --> 00:22:16.879
+in one model on my Open WebUI system.
-00:22:15.660 --> 00:22:21.059
-Neat. Have you used have you used
+NOTE Q: Have you used any small domain-specific LLMs? What are the kinds of tasks they specialize in, and how do I find and use them?
+
+00:22:16.880 --> 00:22:21.059
+[Corwin]: Neat. Have you used have you used
00:22:21.060 --> 00:22:25.739
any small domain specific LLMs? What kind of tasks?
@@ -1450,7 +1489,7 @@ If so, what kind of tasks that they specialize in?
And you know, how?
00:22:32.140 --> 00:22:34.979
-Not to be honest, but there are some out there like once again,
+[Aaron]: Not to be honest, but there are some out there like once again,
00:22:34.980 --> 00:22:36.779
for cybersecurity and stuff like that,
@@ -1461,17 +1500,19 @@ that I really need to dig into that's on my to do list.
00:22:39.740 --> 00:22:41.699
I've got a couple weeks off at the end of the year.
-00:22:41.700 --> 00:22:43.779
+00:22:41.700 --> 00:22:46.539
And that's a big part of my plan for that.
-00:22:43.780 --> 00:22:49.379
-Are the various models updated pretty regularly?
+NOTE Q: Are the various models updated regularly? Can you add your own data to pre-built models?
+
+00:22:46.540 --> 00:22:49.379
+[Corwin]: Are the various models updated pretty regularly?
00:22:49.380 --> 00:22:52.059
Can you add your own data to the pre-built models?
00:22:52.060 --> 00:22:56.699
-Yes. The models are updated pretty reasonably.
+[Aaron]: Yes. The models are updated pretty reasonably.
00:22:56.700 --> 00:22:59.699
You can add data to a model in a couple of different ways.
@@ -1492,7 +1533,7 @@ You can do retrieval augmentation generation,
which is you load your data on top of the system
00:23:09.500 --> 00:23:11.299
-and puts inside a database
+and put inside a database,
00:23:11.300 --> 00:23:12.859
and you can actually scan that and stuff.
@@ -1510,34 +1551,34 @@ I load the talk into the engine
and I ask questions against that.
00:23:20.100 --> 00:23:22.179
-I would have one more time would have done that
+If I would have had time, I would have done that,
-00:23:22.180 --> 00:23:26.499
-but it comes down to how many That's that's rag rag
+00:23:22.180 --> 00:23:25.796
+but it comes down to how many... That's RAG.
-00:23:26.500 --> 00:23:29.419
-is pretty easy to do through open web UI or LM studio
+00:23:25.797 --> 00:23:29.419
+RAG is pretty easy to do through Open WebUI or LM studio.
00:23:29.420 --> 00:23:31.419
-It's a great way you just like point a folder
+It's a great way, you just, like,
00:23:31.420 --> 00:23:34.099
-point it to a folder and it just sucks all that state into
+point it to a folder and it just sucks all that state into...
00:23:34.100 --> 00:23:35.499
-and it'll hit that data first
+and it'll hit that data first.
00:23:35.500 --> 00:23:36.859
-you have like helpdesk and stuff and
+You have like helpdesk and stuff and...
00:23:36.860 --> 00:23:39.619
-The other options there's vector databases,
+The other options: there's vector databases,
00:23:39.620 --> 00:23:41.819
-which is like if you use PostgreSQL.
+which is, like, if you use PostgreSQL,
00:23:41.820 --> 00:23:43.699
-It has a PG vector I can do a lot of that stuff.
+it has a pg vector that can do a lot of that stuff.
00:23:43.700 --> 00:23:44.739
I've not dug into that yet,
@@ -1545,11 +1586,13 @@ I've not dug into that yet,
00:23:44.740 --> 00:23:46.099
but that is also on that to-do list
-00:23:46.100 --> 00:23:48.459
-I've got a lot of stuff planned for Cool.
+00:23:46.100 --> 00:23:48.055
+I've got a lot of stuff planned for...
+
+NOTE Q: What is your experience with RAG? Are you using them and how have they helped?
-00:23:48.460 --> 00:23:51.819
-So what are your experience with rags?
+00:23:48.056 --> 00:23:51.819
+[Corwin]: Cool. So what are your experience with RAGs?
00:23:51.820 --> 00:23:54.339
I don't even know what that means.
@@ -1561,10 +1604,10 @@ Do you know what that means?
Do you remember this question again?
00:23:59.620 --> 00:24:03.979
-What is your experience with RAGs? RAGs is great.
+What is your experience with RAGs?
00:24:03.980 --> 00:24:07.459
-That's Retrieval Augmentation Generation.
+[Aaron]: RAGs is great. That's Retrieval Augmentation Generation.
00:24:07.460 --> 00:24:09.739
That loads your data first, and it hits yours,
@@ -1579,7 +1622,7 @@ There's a guy who wrote a RAG in 100 lines of Python,
and it's an impressive piece of software.
00:24:16.900 --> 00:24:18.779
-I think if you hit one of my site,
+I think if you hit one of my sites,
00:24:18.780 --> 00:24:22.099
I've got a private AI talk where I actually refer to that.
@@ -1599,20 +1642,22 @@ That's one of the great things about AI,
00:24:32.660 --> 00:24:33.619
especially private AI,
-00:24:33.620 --> 00:24:37.739
+00:24:33.620 --> 00:24:35.625
is you can do whatever you want to with it
-00:24:37.740 --> 00:24:43.179
+00:24:35.626 --> 00:24:38.833
and build up with it as you get more experience.
-00:24:43.180 --> 00:24:44.219
-Any thoughts on running things
+NOTE Q: Thoughts on running things on AWS/digital ocean instances, etc?
+
+00:24:38.834 --> 00:24:44.219
+[Corwin]: Any thoughts on running things
00:24:44.220 --> 00:24:49.179
on AWS, DigitalOcean, and so on?
00:24:49.180 --> 00:24:50.619
-AWS is not bad.
+[Aaron]: AWS is not bad.
00:24:50.620 --> 00:24:52.659
The DigitalOcean, they have some of their GPUs.
@@ -1662,11 +1707,13 @@ And Google's also has it,
00:25:23.180 --> 00:25:26.739
but I still tend to keep more stuff on local PCs,
-00:25:26.740 --> 00:25:33.299
-because I just paranoid that way. Gotcha.
+00:25:26.740 --> 00:25:31.077
+because I'm just paranoid that way.
+
+NOTE Q: What has your experience been using AI for cyber security applications? What do you usually use it for?
-00:25:33.300 --> 00:25:35.579
-What has your experience been using AI?
+00:25:31.078 --> 00:25:35.579
+[Corwin]: Gotcha. What has your experience been using AI?
00:25:35.580 --> 00:25:40.139
Do you want to get into that, using AI for cybersecurity?
@@ -1675,16 +1722,16 @@ Do you want to get into that, using AI for cybersecurity?
You might have already touched on this.
00:25:42.020 --> 00:25:44.379
-Yeah, really, for cybersecurity,
+[Aaron]: Yeah, really, for cybersecurity,
00:25:44.380 --> 00:25:46.259
what I've had to do is I've dumped logs
00:25:46.260 --> 00:25:47.299
-to have a due correlation.
+to have it do correlation.
00:25:47.300 --> 00:25:49.859
-Keep in mind, the size of that LLAMA file we were using
+Keep in mind, the size of that Llama file we were using
00:25:49.860 --> 00:25:52.059
for figuring out David Bowie, writing the hello world,
@@ -1722,11 +1769,11 @@ But I want to work on something to do that more locally
00:26:15.300 --> 00:26:19.019
and be able to actually drive this stuff over that.
-00:26:19.020 --> 00:26:21.979
+00:26:19.020 --> 00:26:24.659
That's one also on the long-term goals.
-00:26:21.980 --> 00:26:26.059
-So we got any other questions or?
+00:26:24.660 --> 00:26:26.059
+[Corwin]: So we got any other questions or?
00:26:26.060 --> 00:26:29.099
Those are the questions that I see.
@@ -1738,13 +1785,13 @@ I want to just read out a couple of comments
that I saw in IRC though.
00:26:33.420 --> 00:26:36.699
-Jay Rutabaga says, it went very well
+jrootabaga says, it went very well
00:26:36.700 --> 00:26:39.259
from an audience perspective.
00:26:39.260 --> 00:26:43.619
-And G Gundam says, respect your commitment to privacy.
+And GGundam says, respect your commitment to privacy.
00:26:43.620 --> 00:26:45.619
And then somebody is telling us
@@ -1764,6 +1811,8 @@ I just didn't type in the answer here's
00:26:57.660 --> 00:26:59.659
and there's a couple more questions coming in so
+NOTE Q: Is there a disparity where you go to paid models becouse they are better and what problems would those be?
+
00:26:59.660 --> 00:27:04.699
Is there a disparity where you go to paid models
@@ -1834,61 +1883,60 @@ going into these AIs and stuff,
but in terms of the ability to get a decent one,
00:27:56.220 --> 00:27:57.979
-like the llama, llama three, two,
+like the llama, llama 3.2,
-00:27:57.980 --> 00:28:01.699
+00:27:57.980 --> 00:28:01.239
and load your data into it, you can be pretty competitive.
-00:28:01.700 --> 00:28:04.779
+00:28:01.240 --> 00:28:02.792
You're not going to get all the benefits,
-00:28:04.780 --> 00:28:07.299
+00:28:02.793 --> 00:28:04.333
but you have more control over it.
-00:28:07.300 --> 00:28:11.819
-So it's, it's a, this and that it's a,
+00:28:04.334 --> 00:28:11.000
+So it's a balancing act.
-00:28:11.820 --> 00:28:13.139
-it's a balancing act.
+00:28:11.001 --> 00:28:14.125
+[Corwin]: Okay, and I think I see a couple more questions coming in.
-00:28:13.140 --> 00:28:15.539
-Okay, and I think I see a couple more questions coming in.
+NOTE Q: What's the largest (in parameter size) local model you've been able to successfully run locally, and do you run into issues with limited context window size?
-00:28:15.540 --> 00:28:19.619
+00:28:14.126 --> 00:28:19.619
What is the largest parameter size for local models
00:28:19.620 --> 00:28:22.459
that you've been able to successfully run locally
00:28:22.460 --> 00:28:26.059
-and do run into issues with limited context window size?
+and do you run into issues with limited context window size?
00:28:26.060 --> 00:28:29.659
-The top eight models will tend to have a larger ceiling.
+The top paid models will tend to have a larger ceiling.
00:28:29.660 --> 00:28:32.859
-Yes, yes, yes, yes, yes.
+[Aaron]: Yes, yes, yes, yes, yes.
00:28:32.860 --> 00:28:37.019
By default, the context size is I think 1024.
-00:28:37.020 --> 00:28:44.619
-But I've upped it to 8192 on the on this box, the Pangolin
+00:28:37.020 --> 00:28:41.160
+But I've upped it to 8192 on this box, the Pangolin,
-00:28:44.620 --> 00:28:46.939
-because it seems to be some reason
+00:28:41.161 --> 00:28:43.542
+because it seems to be, for some reason,
-00:28:46.940 --> 00:28:49.459
-it's just a very working quite well.
+00:28:43.543 --> 00:28:45.208
+it's just a very... working quite well.
-00:28:49.460 --> 00:28:52.219
-But the largest ones I've loaded have been in
+00:28:45.209 --> 00:28:49.750
+But the largest ones I've loaded have been in the...
-00:28:52.220 --> 00:28:54.059
-the have not been that huge.
+00:28:49.751 --> 00:28:51.333
+have not been that huge.
-00:28:54.060 --> 00:28:55.699
-I've loaded this the last biggest one I've done.
+00:28:51.334 --> 00:28:55.699
+I've loaded this... the last biggest one I've done...
00:28:55.700 --> 00:28:57.459
That's the reason why I'm planning
@@ -1927,7 +1975,7 @@ and the ones that are more quantitized
because it just tends to work better for me.
00:29:26.620 --> 00:29:29.179
-We've still got over 10 minutes before we're cutting away,
+[Corwin]: We've still got over 10 minutes before we're cutting away,
00:29:29.180 --> 00:29:30.179
but I'm just anticipating
@@ -1956,6 +2004,8 @@ Okay. And we've got 10 minutes
00:29:49.740 --> 00:29:52.379
where we're still going to stay live.
+NOTE Q: Are there "Free" as in FSF/open source issues with the data?
+
00:29:52.380 --> 00:30:00.139
So next question coming in, I see, are there free as in freedom,
@@ -1963,7 +2013,7 @@ So next question coming in, I see, are there free as in freedom,
free as in FSF issues with the data?
00:30:05.740 --> 00:30:11.699
-Yes, where's the data coming from is a huge question with AI.
+[Aaron]: Yes, where's the data coming from is a huge question with AI.
00:30:11.700 --> 00:30:13.739
It's astonishing you can ask questions
@@ -1990,7 +2040,7 @@ I can't remember who it was.
Somebody was actually out torrenting books
00:30:28.660 --> 00:30:30.939
-just to be able to build into their AI system.
+just to be able to build it into their AI system.
00:30:30.940 --> 00:30:32.339
I think it might've been Meta.
@@ -2011,13 +2061,13 @@ like the mobile guys have got their own license,
but where they're getting their data from,
00:30:42.740 --> 00:30:45.499
-I'm not sure on so that that's a huge question.
+I'm not sure, so that's a huge question.
00:30:45.500 --> 00:30:47.979
-That's a that's a talk in itself.
+That's a talk in itself.
00:30:47.980 --> 00:30:51.979
-But yeah, but you if you train on your RAG and your data,
+But yeah, if you train on your RAG and your data,
00:30:51.980 --> 00:30:53.499
you know what it's come, you know,
@@ -2032,16 +2082,21 @@ but the other stuff is just
more lines of supplement
00:30:56.740 --> 00:31:01.379
-if you're using a smaller model,
+if you're using a smaller model.
00:31:01.380 --> 00:31:05.419
-but the comment online, I see a couple of them.
+[Corwin]: The comments online, I see a couple of them.
00:31:05.420 --> 00:31:08.339
I'll read them out in order here. Really interesting stuff.
-00:31:08.340 --> 00:31:11.659
-Thank you for your talk. Given that large AI companies
+00:31:08.340 --> 00:31:09.556
+Thank you for your talk.
+
+NOTE Q: Given that large AI companies are openly stealing IP and copyright, thereby eroding the authority of such law (and eroding truth itself as well), can you see a future where IP & copyright flaw become untenable and what sort of onwards effect might that have?
+
+00:31:09.557 --> 00:31:11.659
+Given that large AI companies
00:31:11.660 --> 00:31:14.899
are openly stealing intellectual property and copyright
@@ -2098,7 +2153,7 @@ and I do not have money.
So don't sue me, is there's going to be
00:31:58.860 --> 00:32:02.899
-the current administration tends is very AI pro AI.
+the current administration tends is very AI, pro AI.
00:32:02.900 --> 00:32:05.499
And there's very a great deal of lobbying by those groups.
@@ -2115,11 +2170,13 @@ what happens to copyright the next 510 years.
00:32:11.700 --> 00:32:13.339
I just don't know how it keeps up
-00:32:13.340 --> 00:32:16.059
+00:32:13.340 --> 00:32:18.059
without there being some adjustments and stuff.
-00:32:16.060 --> 00:32:20.419
-Okay, and then another comment I saw,
+NOTE Comment: File size is not going to be the bottleneck, your RAM is.
+
+00:32:18.060 --> 00:32:20.419
+[Corwin]: Okay, and then another comment I saw,
00:32:20.420 --> 00:32:23.219
file size is not going to be a bottleneck.
@@ -2136,11 +2193,11 @@ and 512 gigabytes of RAM to run the larger ones.
00:32:31.980 --> 00:32:35.059
You'll need a GPU with that much memory
-00:32:35.060 --> 00:32:39.099
-if you want it to run quickly. Yeah. Oh no.
+00:32:35.060 --> 00:32:38.318
+if you want it to run quickly.
-00:32:39.100 --> 00:32:41.259
-It also depends upon how your memory is laid out.
+00:32:38.319 --> 00:32:41.259
+[Aaron]: Yeah. Oh no. It also depends upon how your memory is laid out.
00:32:41.260 --> 00:32:45.699
Like example being the Ultra i285H
@@ -2167,7 +2224,7 @@ So it's able to do some additional stuff with it
as opposed to come off disk.
00:33:00.820 --> 00:33:03.699
-It's all balancing act. If you hit Zyskin's website,
+It's all balancing act. If you hit Ziskind's website,
00:33:03.700 --> 00:33:05.819
that guy's done some great work on it.
@@ -2206,7 +2263,7 @@ But if you want to, Network Chuck had a great video
talking about building his own system
00:33:30.940 --> 00:33:34.379
-with a couple really powerful NVIDIA cards
+with a couple really powerful Nvidia cards
00:33:34.380 --> 00:33:35.379
and stuff like that in it.
@@ -2226,20 +2283,20 @@ which models you want to use,
00:33:44.900 --> 00:33:46.219
how much information you need,
-00:33:46.220 --> 00:33:48.019
+00:33:46.220 --> 00:33:49.579
but it's part of the learning.
-00:33:48.020 --> 00:33:52.899
-And you can run models, even as a Raspberry PI fives,
+00:33:49.580 --> 00:33:52.899
+And you can run models, even on Raspberry Pi 5s,
00:33:52.900 --> 00:33:54.499
if you want to, they'll run slow.
-00:33:54.500 --> 00:33:56.459
+00:33:54.500 --> 00:33:59.339
Don't get me wrong, but they're possible.
-00:33:56.460 --> 00:34:02.179
-Okay, and I think there's other questions coming in too,
+00:33:59.340 --> 00:34:02.179
+[Corwin]: Okay, and I think there's other questions coming in too,
00:34:02.180 --> 00:34:04.019
so I'll just bam for another second.
@@ -2263,7 +2320,7 @@ This is another one that I'm going to
have to study after the conference.
00:34:18.340 --> 00:34:21.099
-We greatly appreciate, all of us appreciate
+[Aaron]: We greatly appreciate, all of us appreciate
00:34:21.100 --> 00:34:22.459
you guys putting on the conference.
@@ -2272,25 +2329,27 @@ you guys putting on the conference.
It's a great conference. It's well done.
00:34:26.300 --> 00:34:28.019
-It's an honor to be on the stage
+[Corwin]: It's an honor to be on the stage
-00:34:28.020 --> 00:34:30.899
+00:34:28.020 --> 00:34:33.124
with the brains of the project, which is you.
-00:34:30.900 --> 00:34:34.699
-So what else we got? Question wise.
+00:34:33.125 --> 00:34:34.699
+[Aaron]: So what else we got? Question wise.
+
+00:34:34.700 --> 00:34:46.899
+[Corwin]: Okay, so just scanning here.
-00:34:34.700 --> 00:34:39.499
-Okay, so just scanning here.
+NOTE Q: Have you used local models capable of tool-calling?
-00:34:39.500 --> 00:34:50.699
+00:34:46.900 --> 00:34:50.699
Have you used local models capable of tool calling?
00:34:50.700 --> 00:34:54.779
-I'm, I'm scared of agentic.
+I'm scared of agentic.
00:34:54.780 --> 00:34:58.739
-I, I am, I'm going to be a slow adopter of that.
+I'm going to be a slow adopter of that.
00:34:58.740 --> 00:35:02.459
I want to do it, but I just don't have the, uh,
@@ -2299,7 +2358,7 @@ I want to do it, but I just don't have the, uh,
four decimal fortitude right now to do it.
00:35:04.340 --> 00:35:07.179
-I, I, I've had to give me the commands,
+I've had to give me the commands,
00:35:07.180 --> 00:35:08.739
but I still run the commands by hand.
@@ -2307,11 +2366,11 @@ but I still run the commands by hand.
00:35:08.740 --> 00:35:10.539
I'm looking into it and it's on once again,
-00:35:10.540 --> 00:35:14.139
+00:35:10.540 --> 00:35:20.899
it's on that list, but I just, that's a big step for me.
-00:35:14.140 --> 00:35:23.139
-So. Awesome. All right.
+00:35:20.900 --> 00:35:23.139
+[Corwin]: So. Awesome. All right.
00:35:23.140 --> 00:35:27.179
Well, maybe it's, let me just scroll through
@@ -2331,17 +2390,19 @@ So this was piggybacking on the question
00:35:41.140 --> 00:35:44.859
about model updates and adding data.
+NOTE Q: Will the models reach out to the web if they need to for more info?
+
00:35:44.860 --> 00:35:46.579
And will models reach out to the web
00:35:46.580 --> 00:35:47.819
if they need more info?
-00:35:47.820 --> 00:35:51.779
+00:35:47.820 --> 00:35:52.479
Or have you worked with any models that work that way?
-00:35:51.780 --> 00:35:55.259
-No, I've not seen any models to do that
+00:35:52.480 --> 00:35:55.259
+[Aaron]: No, I've not seen any models to do that
00:35:55.260 --> 00:35:57.739
There's there was like a group
@@ -2353,19 +2414,22 @@ working on something like a package updater
that would do different diffs on it,
00:36:02.500 --> 00:36:03.939
-but it's so Models change so much
+but it's so... Models change so much,
00:36:03.940 --> 00:36:05.739
-even who make minor changes and fine-tuning.
+even who make minor changes and fine-tuning,
00:36:05.740 --> 00:36:07.659
-It's hard just to update them in place
+It's hard just to update them in place.
00:36:07.660 --> 00:36:10.099
So I haven't seen one, but that doesn't mean
-00:36:10.100 --> 00:36:16.259
-they're not out there. I'm curious topic though Awesome
+00:36:10.100 --> 00:36:15.713
+they're not out there. Curious topic though.
+
+00:36:15.714 --> 00:36:16.259
+[Corwin]: Awesome.
00:36:16.260 --> 00:36:19.539
Well, it's probably pretty good timing.
@@ -2385,17 +2449,19 @@ I want to make sure while we're still live, though,
00:36:28.300 --> 00:36:31.299
I give you a chance to offer any closing thoughts.
+NOTE Q: What scares you most about agentic tools? How would you think about putting a sandbox around it if you adopt an agentic workflow?
+
00:36:31.300 --> 00:36:35.779
So what scares you most about the agentic tools?
00:36:35.780 --> 00:36:38.419
How would you think about putting a sandbox around that
-00:36:38.420 --> 00:36:42.139
+00:36:38.420 --> 00:36:41.619
if you did adopt an agentic workflow?
-00:36:42.140 --> 00:36:42.899
-That is a great question.
+00:36:41.620 --> 00:36:42.899
+[Aaron]: That is a great question.
00:36:42.900 --> 00:36:45.939
In terms of that, I would just control
@@ -2419,7 +2485,7 @@ because that's just kind of the way it works out for us.
So agentic, it's just going to take a while to get trust.
00:36:58.500 --> 00:37:01.059
-I want to want to see more stuff happening.
+I want to see more stuff happening.
00:37:01.060 --> 00:37:02.819
Humans screw up stuff enough.
@@ -2452,16 +2518,18 @@ but it's going to be, it's happening.
It's just, I'm going to be one of the laggards on that one.
00:37:25.820 --> 00:37:29.259
-So airgab, jail, extremely locked down environments,
+So air gap, jail, extremely locked down environments,
00:37:29.260 --> 00:37:34.899
like we're talking about separate physicals, not Docker.
-00:37:34.900 --> 00:37:37.499
-Yeah, hopefully. Right, fair.
+00:37:34.900 --> 00:37:36.577
+Yeah, hopefully.
+
+NOTE Q: Tool calling can be read-only, such as giving models the ability to search the web before answersing your question. (No write access or execute access) I'm interested to know if local models are any good at calling tools, though.
-00:37:37.500 --> 00:37:39.899
-So tool calling can be read-only,
+00:37:36.578 --> 00:37:39.899
+[Corwin]: Right, fair. So tool calling can be read-only,
00:37:39.900 --> 00:37:42.539
such as giving models the ability to search the web
@@ -2479,7 +2547,7 @@ I'm interested to know if local models
are any good at that.
00:37:51.420 --> 00:37:55.579
-Yes, local models can do a lot of that stuff.
+[Aaron]: Yes, local models can do a lot of that stuff.
00:37:55.580 --> 00:37:56.819
It's their capabilities.
@@ -2488,13 +2556,13 @@ It's their capabilities.
If you load LM studio, you can do a lot of wonderful stuff
00:37:59.020 --> 00:38:02.419
-with that or with open web UI with a llama.
+with that or with Open Web UI with ollama.
00:38:02.420 --> 00:38:05.739
It's a lot of capabilities. It's amazing.
00:38:05.740 --> 00:38:08.139
-Open web UI is actually what a lot of companies are using now
+Open Web UI is actually what a lot of companies are using now
00:38:08.140 --> 00:38:10.259
to put their data behind that.
@@ -2505,8 +2573,11 @@ They're curated data and stuff like that. So works well.
00:38:12.140 --> 00:38:15.819
I can confirm that from my own professional experience.
-00:38:15.820 --> 00:38:19.659
-Excellent. Okay, well, our timing should be just perfect
+00:38:15.820 --> 00:38:16.915
+Excellent.
+
+00:38:16.916 --> 00:38:19.659
+[Corwin]: Okay, well, our timing should be just perfect
00:38:19.660 --> 00:38:22.659
if you want to give us like a 30-second, 45-second wrap-up.
@@ -2521,7 +2592,7 @@ Thank you again so much for preparing this talk
and for entertaining all of our questions.
00:38:30.500 --> 00:38:33.299
-Yeah, let me just thank you guys for the conference again.
+[Aaron]: Yeah, let me just thank you guys for the conference again.
00:38:33.300 --> 00:38:35.179
This is a great one. I've enjoyed a lot of it.
@@ -2532,6 +2603,8 @@ I've only had a couple of talks so far,
00:38:37.340 --> 00:38:41.659
but I'm looking forward to hitting the ones after this and tomorrow.
+NOTE Wrapping up
+
00:38:41.660 --> 00:38:44.739
But the AI stuff is coming. Get on board.
@@ -2545,7 +2618,7 @@ and get a little taste of it,
what my minimal viable product
00:38:49.780 --> 00:38:51.619
-with just LlamaFile and GPTEL
+with just Llamafile and gptel
00:38:51.620 --> 00:38:53.139
will get you to the point where you start figuring out.
@@ -2554,34 +2627,34 @@ will get you to the point where you start figuring out.
Gptel is an amazing thing. It just gets out of your way,
00:38:55.580 --> 00:39:00.459
-but it works solo with Emacs. Design because it takes
+but it works so well with Emacs's design because
00:39:00.460 --> 00:39:01.699
-doesn't take your hands off the keyboard.
+it doesn't take your hands off the keyboard.
00:39:01.700 --> 00:39:02.499
-It's just another buffer
+It's just another buffer,
00:39:02.500 --> 00:39:04.059
and you just put information in there.
00:39:04.060 --> 00:39:06.979
-It's quite quite a wonderful It's a wonderful time.
+It's quite a wonderful time.
-00:39:06.980 --> 00:39:10.819
-Let's put that way That's all I got Thank you
+00:39:06.980 --> 00:39:10.501
+Let's put that way. That's all I got.
-00:39:10.820 --> 00:39:14.339
-so much for once again, and we're we're just cut away.
+00:39:10.502 --> 00:39:14.339
+[Corwin]: Thank you so much for once again, and we've just cut away.
00:39:14.340 --> 00:39:15.779
So I'll stop the recording
00:39:15.780 --> 00:39:18.259
-and you're on your own recognizance
+and you're on your own recognizance.
00:39:18.260 --> 00:39:19.699
-Well, I'm gonna punch out
+[Aaron]: Well, I'm gonna punch out
00:39:19.700 --> 00:39:21.059
if anybody has any questions or anything
@@ -2590,7 +2663,7 @@ if anybody has any questions or anything
my email address is ajgrothe@yahoo.com or at gmail and
00:39:24.700 --> 00:39:26.779
-Thank you all for attending
+thank you all for attending,
00:39:26.780 --> 00:39:29.939
and thanks again for the conference