Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
21
2
30
Friedrich Marty
Smorty100
Follow
Maisal02's profile picture
NickyNicky's profile picture
2 followers
·
1 following
https://gitlab.com/users/Marty_Friedrich/projects
AI & ML interests
I'm most interested in content rerouting between LLM and VLLM agens for automation possibilities. Using templates for each agent which is then filled in by another agents inputs seems really useful.
Recent Activity
replied
to
MonsterMMORPG
's
post
2 days ago
It is now possible to generate 16 Megapixel (4096x4096) raw images with SANA 4K model using under 8GB VRAM, 4 Megapixel (2048x2048) images using under 6GB VRAM, and 1 Megapixel (1024x1024) images using under 4GB VRAM thanks to new optimizations 13 January 2024 Update Installers : https://www.patreon.com/posts/from-nvidia-labs-116474081 New 4K Tutorial Video : https://youtu.be/GjENQfHF4W8 Now the APP will use Diffusers Pipeline and it has huge VRAM optimizations You need to reinstall The models will be downloaded into your Hugging Face cache folder when you first time generate something How to Get Installation Logs and How to Change Hugging Face Cache Folder : https://www.patreon.com/posts/108419878 Please make a fresh install When you enable all 4 optimizations the VRAM usages are like below Make sure shared VRAM is enabled because initial loading of the model need more VRAM Enable VAE Tiling + Enable VAE Slicing + Enable Model CPU Offload + Enable Sequential CPU Offload 1K (1024x1024) : 4 GB GPUs 2K (2048x2048) : 6 GB GPUs 4K (4096x4096) : 8 GB GPUs Still in any case may work on your GPU test it Just Enable VAE Tiling + Enable Model CPU Offload works great in many cases All below attached images are generated via SANA 4K model, they are RAW and their resolution is 5376x3072 Official repo page : https://github.com/NVlabs/Sana
liked
a model
4 days ago
PRIME-RL/Eurus-2-7B-PRIME
reacted
to
Severian
's
post
with 👍
5 days ago
Interesting Solution to the Problem of Misguided Attention So I've been fascinated by the problem of Misguided Attention for a few weeks. I am trying to build an inference algorithm to help LLMs address that issue; but in the process, I found a cool short-term fix I call "Mindful Attention" using just prompt-engineering. Have you ever thought about how our brains filter reality through layers of past experiences, concepts, and mental images? For example, when you look at an oak tree, are you truly seeing that oak tree in all its unique details, or are you overlaying it with a generalized idea of "oak tree"? This phenomenon inspired the new approach. LLMs often fall into a similar trap, hence the Misguided Attention problem. They process input not as it’s uniquely presented but through patterns and templates they’ve seen before. This leads to responses that can feel "off," like missing the point of a carefully crafted prompt or defaulting to familiar but irrelevant solutions. I wanted to address this head-on by encouraging LLMs to slow down, focus, and engage directly with the input—free of assumptions. This is the core of the Mindful Attention Directive, a prompt designed to steer models away from over-generalization and back into the moment. You can read more about the broader issue here: https://github.com/cpldcpu/MisguidedAttention And if you want to try this mindful approach in action, check out the LLM I’ve set up for testing: https://hf.co/chat/assistant/677e7ebcb0f26b87340f032e. It works about 80% of the time to counteract these issues, and the results are pretty cool. I'll add the Gist with the full prompt. I admit, it is quite verbose but it's the most effective one I have landed on yet. I am working on a smaller version that can be appended to any System Prompt to harness the Mindful Attention. Feel free to experiment to find a better version for the community! Here is the Gist: https://gist.github.com/severian42/6dd96a94e546a38642278aeb4537cfb3
View all activity
Organizations
None yet
Smorty100
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
liked
a model
4 days ago
PRIME-RL/Eurus-2-7B-PRIME
Text Generation
•
Updated
1 day ago
•
995
•
53
liked
a model
7 days ago
nvidia/Cosmos-1.0-Diffusion-7B-Text2World
Updated
5 days ago
•
86.4k
•
157
liked
a dataset
14 days ago
cfahlgren1/react-code-instructions
Viewer
•
Updated
about 1 hour ago
•
63.2k
•
784
•
121
liked
a model
14 days ago
answerdotai/ModernBERT-base
Fill-Mask
•
Updated
4 days ago
•
4.54M
•
670
liked
a model
17 days ago
hexgrad/Kokoro-82M
Text-to-Speech
•
Updated
1 day ago
•
15.6k
•
1.37k
liked
a model
25 days ago
moxin-org/moxin-chat-7b
Updated
25 days ago
•
1.04k
•
29
liked
a model
27 days ago
rombodawg/Rombos-LLM-70b-Llama-3.3
Text Generation
•
Updated
27 days ago
•
152
•
5
liked
a model
about 1 month ago
PrimeIntellect/INTELLECT-1-Instruct
Text Generation
•
Updated
Nov 29, 2024
•
997
•
117
liked
a model
about 2 months ago
Qwen/QwQ-32B-Preview
Text Generation
•
Updated
3 days ago
•
145k
•
•
1.55k
liked
a Space
about 2 months ago
Running
644
👁
PR Puppet Sora
liked
a Space
2 months ago
Running
5
👄
Lip
liked
3 models
3 months ago
rombodawg/Rombos-LLM-V2.5-Qwen-32b
Text Generation
•
Updated
Oct 6, 2024
•
5.05k
•
50
microsoft/OmniParser
Image-Text-to-Text
•
Updated
Dec 2, 2024
•
1.11k
•
1.52k
ostris/OpenFLUX.1
Text-to-Image
•
Updated
Oct 3, 2024
•
4.53k
•
603
liked
5 models
4 months ago
meta-llama/Llama-3.2-11B-Vision-Instruct
Image-Text-to-Text
•
Updated
Dec 4, 2024
•
2.42M
•
•
1.22k
meta-llama/Llama-3.2-11B-Vision
Image-Text-to-Text
•
Updated
Sep 27, 2024
•
53k
•
430
peakji/peak-reasoning-7b-gguf
Updated
Oct 21, 2024
•
197
•
4
HF1BitLLM/Llama3-8B-1.58-100B-tokens
Text Generation
•
Updated
Sep 19, 2024
•
1.87k
•
167
G-reen/gpt5o-reflexion-q-agi-llama-3.1-8b
Text Generation
•
Updated
Sep 13, 2024
•
160
•
64
liked
a Space
4 months ago
Running
on
L4
419
🏆
Fish Speech 1
Load more