Waseem AlShikh
wassemgtk
AI & ML interests
Multi-modal, Palmyra LLMs, Knowledge Graph
Recent Activity
replied to
their
post
9 days ago
# GESAL: Real-Time Adaptation for LLMs
We’re excited to unveil **Graph-Enhanced Singular Adaptive Learning (GESAL)**, a framework that lets LLMs like `meta-llama/Llama-3.2-1B` adapt in real time using user feedback. Check out the code and white paper on GitHub!
🔗 **Code**: [https://github.com/writer/AI-Adaptive-Learning-GESAL](https://github.com/writer/AI-Adaptive-Learning-GESAL)
---
## Why GESAL?
Static LLMs struggle to adapt without heavy retraining. GESAL solves this with:
- **SVF**: Adapts weights via \( W' = U (\Sigma \cdot z) V^T \), using few parameters.
- **Graph Memory**: Stores adaptations in nodes for scalability.
- **RL**: Updates via \( J(z) = \mathbb{E}[\log \pi_z(y|x) r] \) based on feedback.
---
## How It Works
Ask "How many R’s in ‘strawberry’?" If it says "2" and you say "no," GESAL learns to say "3" next time, avoiding repeats.
---
## Try It
Built with Hugging Face’s `transformers`:
```bash
pip install transformers torch numpy
python Adaptive_Learning_(GESAL).py
```
Needs a Hugging Face token for Llama-3.2-1B.
---
## Results
GESAL hits 95% accuracy after 5 feedbacks vs. LoRA’s 70%. It’s efficient (~0.5M params) and scalable.
replied to
their
post
9 days ago
# GESAL: Real-Time Adaptation for LLMs
We’re excited to unveil **Graph-Enhanced Singular Adaptive Learning (GESAL)**, a framework that lets LLMs like `meta-llama/Llama-3.2-1B` adapt in real time using user feedback. Check out the code and white paper on GitHub!
🔗 **Code**: [https://github.com/writer/AI-Adaptive-Learning-GESAL](https://github.com/writer/AI-Adaptive-Learning-GESAL)
---
## Why GESAL?
Static LLMs struggle to adapt without heavy retraining. GESAL solves this with:
- **SVF**: Adapts weights via \( W' = U (\Sigma \cdot z) V^T \), using few parameters.
- **Graph Memory**: Stores adaptations in nodes for scalability.
- **RL**: Updates via \( J(z) = \mathbb{E}[\log \pi_z(y|x) r] \) based on feedback.
---
## How It Works
Ask "How many R’s in ‘strawberry’?" If it says "2" and you say "no," GESAL learns to say "3" next time, avoiding repeats.
---
## Try It
Built with Hugging Face’s `transformers`:
```bash
pip install transformers torch numpy
python Adaptive_Learning_(GESAL).py
```
Needs a Hugging Face token for Llama-3.2-1B.
---
## Results
GESAL hits 95% accuracy after 5 feedbacks vs. LoRA’s 70%. It’s efficient (~0.5M params) and scalable.
replied to
their
post
9 days ago
# GESAL: Real-Time Adaptation for LLMs
We’re excited to unveil **Graph-Enhanced Singular Adaptive Learning (GESAL)**, a framework that lets LLMs like `meta-llama/Llama-3.2-1B` adapt in real time using user feedback. Check out the code and white paper on GitHub!
🔗 **Code**: [https://github.com/writer/AI-Adaptive-Learning-GESAL](https://github.com/writer/AI-Adaptive-Learning-GESAL)
---
## Why GESAL?
Static LLMs struggle to adapt without heavy retraining. GESAL solves this with:
- **SVF**: Adapts weights via \( W' = U (\Sigma \cdot z) V^T \), using few parameters.
- **Graph Memory**: Stores adaptations in nodes for scalability.
- **RL**: Updates via \( J(z) = \mathbb{E}[\log \pi_z(y|x) r] \) based on feedback.
---
## How It Works
Ask "How many R’s in ‘strawberry’?" If it says "2" and you say "no," GESAL learns to say "3" next time, avoiding repeats.
---
## Try It
Built with Hugging Face’s `transformers`:
```bash
pip install transformers torch numpy
python Adaptive_Learning_(GESAL).py
```
Needs a Hugging Face token for Llama-3.2-1B.
---
## Results
GESAL hits 95% accuracy after 5 feedbacks vs. LoRA’s 70%. It’s efficient (~0.5M params) and scalable.
Organizations
wassemgtk's activity
Added explanation of how combined score is calculated
#2 opened 19 days ago
by
ecw429
Update app.py
1
#1 opened 19 days ago
by
karsitu
Add link to paper and update citation.
#2 opened 25 days ago
by
nielsr

Any details on this and the base model?
1
#1 opened 4 months ago
by
jukofyork

Base model?
5
#4 opened 7 months ago
by
KevalRx
Update src/lib/data.ts
#11 opened 6 months ago
by
wassemgtk

Access the model via Ollama
1
#3 opened 7 months ago
by
gileneo
Quantization allowed?
1
#1 opened 7 months ago
by
bartowski

Adding `safetensors` variant of this model
#1 opened about 1 year ago
by
SFconvertbot

Update src/lib/data.ts
1
#2 opened 8 months ago
by
wassemgtk

Adding Writer LLMs
1
#1 opened 8 months ago
by
wassemgtk

70B submissions are just disappearing after a while
3
#10 opened 10 months ago
by
abhinand
It's broken
12
#9 opened 10 months ago
by
tosaddler

Create README.md
2
#2 opened over 1 year ago
by
wassemgtk

Create handler.py
#4 opened over 1 year ago
by
wassemgtk

add example for text-generation-inference
#1 opened over 1 year ago
by
nbroad

add text-generation-inference example
#1 opened over 1 year ago
by
nbroad

add example for text-generation-inference
#3 opened over 1 year ago
by
nbroad

changed palmyra-med -> InstructPalmyra-20b
#2 opened over 1 year ago
by
nbroad
