From 6ebf0ccbfab1b51b51d3417a482222386a5ceca6 Mon Sep 17 00:00:00 2001 From: Sacha Chua Date: Fri, 8 Dec 2023 10:21:03 -0500 Subject: add all the comments from IRC and the pad --- 2023/talks/matplotllm.md | 24 ------------------------ 1 file changed, 24 deletions(-) (limited to '2023/talks/matplotllm.md') diff --git a/2023/talks/matplotllm.md b/2023/talks/matplotllm.md index 784c9463..dfa32233 100644 --- a/2023/talks/matplotllm.md +++ b/2023/talks/matplotllm.md @@ -63,30 +63,6 @@ Emacs. - Repository link . A connected blog post here -- gptel is another package doing a good job is flexible configuration and choice over LLM/API -- I came across this adapter to run multiple LLM's, apache 2.0 license too! https://github.com/predibase/lorax -- It will turn out the escape-hatch for AGI will be someone's integration of LLMs into their Emacs and enabling M-x control. -- i don't know what question to ask but i found presentation extremely useful thank you -- I think we are close to getting semantic search down for our own files - - yeah, khoj uses embeddings to search Org, I think - - I tried it a couple of times, latest about a month ago. The search was quite bad unfortunately - - did you try the GPT version or just the PyTorch version? - - just the local ones. For GPT I used a couple of other packages to embed in OpenAI APIs. But I am too shy to send all my notes :D - - Same for me. But I really suspect that GPT will be way better. They now also support LLama, which is hopeful - - I keep meaning to revisit the idea of the Remembrance Agent and see if it can be updated for these times (and maybe local HuggingFace embeddings) -- I think Andrew is right that Emacs is uniquely positioned, being a unified integrated interface with good universal abstractions (buffers, text manipulation, etc), and across all uses cases and notably one's Org data. Should be interesting...! -- Speaking of which, anyone trained/fined-tuned/prompted a model with their Org data yet and applied it to interesting use cases (planning/scheduling, etc) and care to comment? -- The ubiquitous integration of LLMs (multi-modal) for anything and everything in/across Emacs and Org is both 1) exciting, 2) scary. -- I could definitely use semantic search across all of my stored notes. Can't remember what words I used to capture things. -- Indeed. A "working group" / "birds of a feather" type of thing around the potential usages and integration of LLMs and other models into Emacs and Org-mode would be interesting, especially as this is what pulls people into other platforms these days. -- To that end, Andrew is right that we'll want to abstract it into the right abstractions and interfaces. And not just LLMs by vendor/models, but what comes after LLMs/GPTs in terms of approach. -- I lean toward thinking that LLMs may have some value but to me a potentially wrong result is worse than no result - - I think it would depend on the use case. A quasi-instant first approximation that can readily be fixed/tweaked can be quite useful in some contexts. -- not to mention the "summarization" use cases (for papers, and even across papers I've found, like a summarization across abstracts/contents of a multiplicity of papers and publications around a topic or in a field - weeks of grunt work saved, not to mention of procrastination avoided) - - IMHO summarization is exactly where LLMs can't be useful because they can't be trusted to be accurate -- ; A friend wrote this ; < https://blogs.microsoft.com/on-the-issues/2023/09/07/copilot-copyright-commitment-ai-legal-concerns/> -- I have a feeling this is one of the 'em "if you can't beat them join them" scenario. I don't see that ending with a bit global rollback due to such issues anytime soon... -- (discussion about LLMs, copyright, privacy) -- cgit v1.2.3