Skip to content

Latest commit

 

History

History
131 lines (75 loc) · 4.08 KB

personalized_anything.md

File metadata and controls

131 lines (75 loc) · 4.08 KB

personalized anything

labels: productivity, personalization, llm, gdpr_data_export

here's how this would work e.g. for mastodon. generalized from there.

  1. consume feeds of mastodon people i follow
  2. use plugin adaptors to expose content to an LLM, add CLIP embeddings + images
  3. track my likes
  4. predict likes
  5. track what I have and have not seen
  6. potential emergent behavior: recommends content to me because it knows i haven't seen it and that the people around me are talking about it

  7. i maintain my own personal searchable database
  8. i use it for a feed
  9. it RLHFs. assign a rank to each candidate document (separate model for classifying stuff i don't even need to archive)

so i'm thinking about mastodon right now, but this could really be a feed of anything. subreddits. hackernews. github repos.

just let an LLM see your content interaction history and ask it directly if a piece of content is something you'd find interesting.

NTS: crack open the twitter data export


i've already got my twitter export, i just need to finetune on that. grab some pre-trained multimodal model, finetune it on whether a tweet was authored, liked, or retweeted by me. for quote tweets, fill the contents... naw i'm making this too complicated, just generic LLM autoregressive next word prediction objective should be fine.


relevant:

anyway, multimodal model. LLaVA or whatever.

"instruct" tune it, predicting my reply tweet in response to the context "prompt"

... this has to already be a thing by now, right?

close: https://github.com/recalign/RecAlign

another strong candidate: https://github.com/salesforce/LAVIS

...ok looks like LLAVA has already been improved upon too - https://llavar.github.io/


this looks promising: https://github.com/georgia-tech-db/evadb

also this: https://github.com/kvablack/LLaVA-server

while we're at it: https://github.com/pgvector/pgvector

if i go llama family: https://github.com/coreylowman/llama-dfdx

LoRA trainer: https://github.com/official-elinas/zeus-llm-trainer

another llm finetuner/server: https://github.com/OptimalScale/LMFlow

https://github.com/mlc-ai/web-llm

lol https://github.com/jdagdelen/hyperDB

https://github.com/lm-sys/FastChat

another LoRA trainer - https://docs.adapterhub.ml/model_overview.html


fallback plan: augment a conventional LLM with visual descriptions

maybe some shortlist of content like important AI papers:

Another approach: EVAPORATE

and another: https://github.com/VPGTrans/VPGTrans

another take: align a frozen vision encoder with a frozen llm - https://github.com/Vision-CAIR/MiniGPT-4

oh yeah, there's always full-on hugginggpt esque orchestrator approach: https://github.com/lupantech/chameleon-llm

https://github.com/amazon-science/mm-cot

MAGMA - https://arxiv.org/abs/2112.05253


ok now we're talking:

i wonderif maybe we could benefit from a "model merge" of those open flamingo checkpoints

this is what i was looking for i think: https://github.com/haotian-liu/LLaVA#LLaVA-MPT-7b

looks like it's a llama model. still need that checkpoint. way overdue.

maybe can convert llama base model's to openllama? https://github.com/openlm-research/open_llama

oh great, whole shitload of other ones here too: https://github.com/microsoft/unilm


weakly related? https://github.com/google-research/pix2struct


separate project from personalized policies: persist papers from arxiv (plus code) as knowledge plugins for personalized research assistant


uh... hard maybe... https://github.com/princeton-nlp/MeZO

another hard maybe: https://github.com/mlc-ai/mlc-llm another h