summaryrefslogtreecommitdiffstats
path: root/2023/talks/llm.md
diff options
context:
space:
mode:
Diffstat (limited to '2023/talks/llm.md')
-rw-r--r--2023/talks/llm.md77
1 files changed, 77 insertions, 0 deletions
diff --git a/2023/talks/llm.md b/2023/talks/llm.md
index e1723796..cb76d783 100644
--- a/2023/talks/llm.md
+++ b/2023/talks/llm.md
@@ -33,6 +33,83 @@ Andrew Hyatt has contributed the Emacs websocket package, the triples
(making a triple-based DB library) and the ekg package (a tag-based
note-taking application). He has been using various other LLM
integrations, and ss part of extending ekg, he's been working on his own.
+# Discussion
+
+## Questions and answers
+
+- Q: What is your use case for Embedding? Mainly for searching? 
+ - A:
+ - I got you. It\'s kinda expand our memory capcity. 
+- Q: What do you think about \"Embed Emacs manual\" VS \"GPTs  Emacs
+ manual?
+ - A: 
+ - yes GPTS actually how it\'s kind of embedding your document
+ into its memory and then using the logic that provided by
+ GPT-4 or other versions. I never tried that one but I\'m
+ just wondering if you have ever tried the difference
+- Q: When deferring commit messages to an LLM, what (if anything) do
+ you find you have lost?
+ - A:
+- Q: Can you share your font settings in your emacs config? :) (Yeah,
+ those are some nice fonts for reading)
+ - A: I think it was Menlo, but I\'ve sinced changed it (I\'m
+ experimenting with Monaspace
+- Q: In terms of standardisation, do you see a need for a
+ medium-to-large scale effort needed?
+ - A:
+ - I mean, as a user case, the interface is quite simple
+ because we\'re just providing an API to a server. I\'m not
+ sure what standardization we are really looking at. I mean,
+ it\'s more like the how we use those callback from the llm.
+- Q: What are your thoughts on the carbon footprint of LLM useage?
+ - A:
+- Q: LLMs are slow in responding. Do you think Emacs should provide
+ more async primitives to keep it responsive? E.g. url-retrieve is
+ quite bad at building API clients with it.
+ - A:
+ - Gptel.el is async. And very good at tracking the point. 
+- Q: Speaking of which, anyone trained/fined-tuned/prompted a model
+ with their Org data yet and applied it to interesting use cases
+ (planning/scheduling, etc) and care to comment?
+ - A:
+ - I use GPTS doing weekly review. I\'m not purely rely on it.
+ It\'s help me to find something I never thought about and I
+ just using as alternateive way to do the reviewing.  I find
+ it\'s kind of interesting to do so.
+
+### Notes and discussion
+
+- gptel is another package doing a good job is flexible configuration and choice over LLM/API
+- I came across this adapter to run multiple LLM's, apache 2.0 license too! https://github.com/predibase/lorax
+- It will turn out the escape-hatch for AGI will be someone's integration of LLMs into their Emacs and enabling M-x control.
+- i don't know what question to ask but i found presentation extremely useful thank you
+- I think we are close to getting semantic search down for our own files
+ - yeah, khoj uses embeddings to search Org, I think
+ - I tried it a couple of times, latest about a month ago. The search was quite bad unfortunately
+ - did you try the GPT version or just the PyTorch version?
+ - just the local ones. For GPT I used a couple of other packages to embed in OpenAI APIs. But I am too shy to send all my notes :D
+ - Same for me. But I really suspect that GPT will be way better. They now also support LLama, which is hopeful
+ - I keep meaning to revisit the idea of the Remembrance Agent and see if it can be updated for these times (and maybe local HuggingFace embeddings)
+- I think Andrew is right that Emacs is uniquely positioned, being a unified integrated interface with good universal abstractions (buffers, text manipulation, etc), and across all uses cases and notably one's Org data. Should be interesting...!
+- Speaking of which, anyone trained/fined-tuned/prompted a model with their Org data yet and applied it to interesting use cases (planning/scheduling, etc) and care to comment?
+- The ubiquitous integration of LLMs (multi-modal) for anything and everything in/across Emacs and Org is both 1) exciting, 2) scary.
+- I could definitely use semantic search across all of my stored notes. Can't remember what words I used to capture things.
+- Indeed. A "working group" / "birds of a feather" type of thing around the potential usages and integration of LLMs and other models into Emacs and Org-mode would be interesting, especially as this is what pulls people into other platforms these days.
+- To that end, Andrew is right that we'll want to abstract it into the right abstractions and interfaces. And not just LLMs by vendor/models, but what comes after LLMs/GPTs in terms of approach.
+- I lean toward thinking that LLMs may have some value but to me a potentially wrong result is worse than no result
+ - I think it would depend on the use case. A quasi-instant first approximation that can readily be fixed/tweaked can be quite useful in some contexts.
+- not to mention the "summarization" use cases (for papers, and even across papers I've found, like a summarization across abstracts/contents of a multiplicity of papers and publications around a topic or in a field - weeks of grunt work saved, not to mention of procrastination avoided)
+ - IMHO summarization is exactly where LLMs can't be useful because they can't be trusted to be accurate
+- <https://dindi.garjola.net/ai-assistants.html>; A friend wrote this <https://www.jordiinglada.net/sblog/llm.html>; < https://blogs.microsoft.com/on-the-issues/2023/09/07/copilot-copyright-commitment-ai-legal-concerns/>
+- I have a feeling this is one of the 'em "if you can't beat them join them" scenario. I don't see that ending with a bit global rollback due to such issues anytime soon...
+- (discussion about LLMs, copyright, privacy)
+- I spent more time than I was hoping to setting up some custom Marginalia(s?) the other day, notably for cases where the "category" is dynamic, the annotation/affixation function varies, the candidates are an alist of key-value pairs and not just directly the value, and many little specificities like that. Idem for org-ql many moons back, org-agenda, etc. That sort of workflow always involves the same things: learning/reading, examples, trials, etc. I wonder if LLMs could be integrated at various points in that recurring exercise, to take just a sample case.
+- that's yet another great use case for LLMs : externalizing one's thinking for its own sake, if only to hear back the echo of one's "voice", and do so with an infinitely patient quasi-omniscient second party.
+ - oooh, might be a good one for blog post writing: generate some follow-up questions people might have
+ - Yeah, a "rubber duck" LLM could be very handy
+ - I'm sure there would be great demand for such a thing, to dry-run one's presentations (video or text) and generate anticipate questions and so on. Great take.
+ - I've seen some journaling prompts along those lines. I think it'll get even more interesting as the text-to-speech and speech-to-text parts get better. Considering how much people bonded with Eliza, might be interesting to see what people can do with a Socratic assistant...
+
[[!inline pages="internal(2023/info/llm-after)" raw="yes"]]