Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
21
2
30
Friedrich Marty
Smorty100
Follow
Maisal02's profile picture
1 follower
·
1 following
https://gitlab.com/users/Marty_Friedrich/projects
AI & ML interests
I'm most interested in content rerouting between LLM and VLLM agens for automation possibilities. Using templates for each agent which is then filled in by another agents inputs seems really useful.
Recent Activity
liked
a model
2 days ago
PRIME-RL/Eurus-2-7B-PRIME
reacted
to
Severian
's
post
with 👍
3 days ago
Interesting Solution to the Problem of Misguided Attention So I've been fascinated by the problem of Misguided Attention for a few weeks. I am trying to build an inference algorithm to help LLMs address that issue; but in the process, I found a cool short-term fix I call "Mindful Attention" using just prompt-engineering. Have you ever thought about how our brains filter reality through layers of past experiences, concepts, and mental images? For example, when you look at an oak tree, are you truly seeing that oak tree in all its unique details, or are you overlaying it with a generalized idea of "oak tree"? This phenomenon inspired the new approach. LLMs often fall into a similar trap, hence the Misguided Attention problem. They process input not as it’s uniquely presented but through patterns and templates they’ve seen before. This leads to responses that can feel "off," like missing the point of a carefully crafted prompt or defaulting to familiar but irrelevant solutions. I wanted to address this head-on by encouraging LLMs to slow down, focus, and engage directly with the input—free of assumptions. This is the core of the Mindful Attention Directive, a prompt designed to steer models away from over-generalization and back into the moment. You can read more about the broader issue here: https://github.com/cpldcpu/MisguidedAttention And if you want to try this mindful approach in action, check out the LLM I’ve set up for testing: https://hf.co/chat/assistant/677e7ebcb0f26b87340f032e. It works about 80% of the time to counteract these issues, and the results are pretty cool. I'll add the Gist with the full prompt. I admit, it is quite verbose but it's the most effective one I have landed on yet. I am working on a smaller version that can be appended to any System Prompt to harness the Mindful Attention. Feel free to experiment to find a better version for the community! Here is the Gist: https://gist.github.com/severian42/6dd96a94e546a38642278aeb4537cfb3
reacted
to
Severian
's
post
with 👀
3 days ago
Interesting Solution to the Problem of Misguided Attention So I've been fascinated by the problem of Misguided Attention for a few weeks. I am trying to build an inference algorithm to help LLMs address that issue; but in the process, I found a cool short-term fix I call "Mindful Attention" using just prompt-engineering. Have you ever thought about how our brains filter reality through layers of past experiences, concepts, and mental images? For example, when you look at an oak tree, are you truly seeing that oak tree in all its unique details, or are you overlaying it with a generalized idea of "oak tree"? This phenomenon inspired the new approach. LLMs often fall into a similar trap, hence the Misguided Attention problem. They process input not as it’s uniquely presented but through patterns and templates they’ve seen before. This leads to responses that can feel "off," like missing the point of a carefully crafted prompt or defaulting to familiar but irrelevant solutions. I wanted to address this head-on by encouraging LLMs to slow down, focus, and engage directly with the input—free of assumptions. This is the core of the Mindful Attention Directive, a prompt designed to steer models away from over-generalization and back into the moment. You can read more about the broader issue here: https://github.com/cpldcpu/MisguidedAttention And if you want to try this mindful approach in action, check out the LLM I’ve set up for testing: https://hf.co/chat/assistant/677e7ebcb0f26b87340f032e. It works about 80% of the time to counteract these issues, and the results are pretty cool. I'll add the Gist with the full prompt. I admit, it is quite verbose but it's the most effective one I have landed on yet. I am working on a smaller version that can be appended to any System Prompt to harness the Mindful Attention. Feel free to experiment to find a better version for the community! Here is the Gist: https://gist.github.com/severian42/6dd96a94e546a38642278aeb4537cfb3
View all activity
Organizations
None yet
Smorty100
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
liked
a model
2 days ago
PRIME-RL/Eurus-2-7B-PRIME
Text Generation
•
Updated
6 days ago
•
669
•
51
liked
a model
5 days ago
nvidia/Cosmos-1.0-Diffusion-7B-Text2World
Updated
3 days ago
•
1.44k
•
124
liked
a dataset
11 days ago
cfahlgren1/react-code-instructions
Viewer
•
Updated
1 day ago
•
61.2k
•
606
•
106
liked
a model
12 days ago
answerdotai/ModernBERT-base
Fill-Mask
•
Updated
2 days ago
•
3.57M
•
653
liked
a model
15 days ago
hexgrad/Kokoro-82M
Text-to-Speech
•
Updated
6 days ago
•
9.72k
•
830
liked
a model
23 days ago
moxin-org/moxin-chat-7b
Updated
23 days ago
•
1.04k
•
29
liked
a model
25 days ago
rombodawg/Rombos-LLM-70b-Llama-3.3
Text Generation
•
Updated
25 days ago
•
152
•
5
liked
a model
about 1 month ago
PrimeIntellect/INTELLECT-1-Instruct
Text Generation
•
Updated
Nov 29, 2024
•
985
•
117
liked
a model
about 2 months ago
Qwen/QwQ-32B-Preview
Text Generation
•
Updated
1 day ago
•
134k
•
1.53k
liked
2 Spaces
about 2 months ago
Running
644
👁
PR Puppet Sora
Running
5
👄
Lip
liked
a model
2 months ago
rombodawg/Rombos-LLM-V2.5-Qwen-32b
Text Generation
•
Updated
Oct 6, 2024
•
5.16k
•
48
liked
2 models
3 months ago
microsoft/OmniParser
Image-Text-to-Text
•
Updated
Dec 2, 2024
•
1.03k
•
1.52k
ostris/OpenFLUX.1
Text-to-Image
•
Updated
Oct 3, 2024
•
4.67k
•
602
liked
5 models
4 months ago
meta-llama/Llama-3.2-11B-Vision-Instruct
Image-Text-to-Text
•
Updated
Dec 4, 2024
•
2.51M
•
•
1.21k
meta-llama/Llama-3.2-11B-Vision
Image-Text-to-Text
•
Updated
Sep 27, 2024
•
57.1k
•
427
peakji/peak-reasoning-7b-gguf
Updated
Oct 21, 2024
•
178
•
4
HF1BitLLM/Llama3-8B-1.58-100B-tokens
Text Generation
•
Updated
Sep 19, 2024
•
1.69k
•
166
G-reen/gpt5o-reflexion-q-agi-llama-3.1-8b
Text Generation
•
Updated
Sep 13, 2024
•
195
•
64
liked
a Space
4 months ago
Running
on
L4
417
🏆
Fish Speech 1
Load more