Comments

Log in with itch.io to leave a comment.

Viewing most recent comments 1 to 16 of 60 · Next page · Last page

I have to say, after I played with it for a while, it's rather disappointing.  The usage of Ollama makes the performance far worse that it could otherwise be and the default models never work correctly. Sometimes they work, sometimes they just decide to give you no letters or a letter or 2-3 words and then stop generating . Nothing but closing the app can save the performance at that point. The available models are also far from the best RP models out there, with many being neither RP nor Story models. The website is decent but the promise of a desktop application is not fulfilled by this app. It also doesn't help that the Ollama back end is slower both to first token and to generate than alternatives like KoboldCCP and LM Studio. Overall, a disappointing program that would probably have potential were it not so fundamentally based around Ollama.


But this is just after a week of random usage, I'm sure other people that have been using it for longer might be able to give both more constructive criticism and more pertinent reviews. Better than Sillytavern on account of being actually designed for desktop/laptop use as opposed to server usage, but that's about it.

Thanks for the feedback! So right now our Ollama version is actually really old. Does it work better if you use this beta version? It updates Ollama and should be MUCH faster: https://github.com/hammer-ai/hammerai/releases/tag/v0.0.206

Yes and no. On the one hand when the models when they run they do run better, on the other hand it's still impossible to run the model set as character model and I have to use the custom model option. The good news is that it does run faster and that at least on custom I haven't encountered the issues I had before where it just wound't run.


BTW, did you put the linux AMD ROCm support in just for me as your one known linux user or is HammerAI actually detecting that I have an AMD AI capable CPU with iGPU besides the Nvidia GPU it's using right now? Because if the latter that's actually impressive - ROCm support is so spotty on linux is might as well not be there. The 780m from AMD is a lot weaker than the 4060 so I don't think It will see much usage, but I might try bigger models just to see how it behaves if Hammer can actually use the AMD iGPU natively.

PS. Please add a few newer RP models. Some of your competitors have a few finetunes that are open source licenses. ArliAI, Latitudegames, Dreamgen. Please add a few newer Nemo finetunes. Also, and this is just up to you, consider IQ quants and do not offload KV cache to VRAM. IQ quants can make 8GB fully enough to 100% offload most non-gemma 12B models, as long as one doesn't also try to offload KV cache to the Vram. That's in case you're not doing this already.


Anyways, cheers and thanks for the new Ollama update, it did in fact help.

(1 edit)

Glad to hear it's better! I really need to get this update out to users. But there is one bug I know about that before I can launch.

Linux AMD ROCm support was just in case I had any users, I wanted to make sure it was awesome for them! Glad to hear that day happened faster than I expected.

Will definitely add some more models, I'm pretty behind. Any specific suggestions? Would love to hear what the best stuff is nowadays.

I will learn more about IQ quants and the KV cache offloading. Is that suggestion for the local LLMs, or the cloud-hosted ones?

Anyways, happy it's better. If you want to chat more, I'm hammer_ai on Discord - would be fun to chat more about finetunes to add / any other suggestions you have.

(+1)

For the desktop version. Basically, there are two way to use AI, one with slowly building KV cache over yeach new prompt and reply and one where the user sends the entire conversation back to the AI to be processed with each new prompt to get a new response. On desktop it's faster to use KV cache than it is to reprocess the entire conversation again and again. Thing is, the KV cache can be separate from from rest of the model. IF the offload to VRAM option is used and there is enough VRAM it's always faster BUT if there isn't enough VRAM for the desired KV cache size then part of the model and part of the KV cache are in VRAM and the rest are in RAM and this is always slower. 

If you can fit everything in VRAM you're at  say 21 tokens per second, with only the model in VRAM and KV cache in RAM you'd be at around 15, and with part of the model and part of the KV cache in VRAM and the rest in RAM you could go as low as 10 or even 5 tokens per second. So it's always preferable to only load the model in VRAM and let the KV cache in RAM if you can't fit everything in VRAM. For the website, since you're only using 4k context window as long as everything fits in VRAM, I wouldn't touch it - if it ain't broke don't fix it and whatnot. But on desktop, allowing us to keep KV cache only in RAM or offload it to VRAM, can significantly increase performance.


As for recommended models. I'd say move the Nous models to Hermes 3 (nonthinking) look into the ArliAI RPMax v.1.3 series of models (4 models at 4 sizes, based on 3 different bases, Llama 3.1, qwen 2.5 and Mistral Nemo), and the latest Latitudegames models. I'm using Wayfarer 12b for RP and Muse 12b for story writing (latitudegames models) but they have larger models and again, all open source and on Huggingface. Dreamgen is also doing interesting stuff, but their older stuff is, well, older, and the new model - Lucid - is still in beta and fairly bad at following instruction. 

But yeah, try Wayfarer, at least for me it's significantly superior to the Drummer Rocinante you have as a default option. I get actual RP responses from it wile Rocinante 12b wants to just continue my own posts 90% of the time. Also, I'd probably remove the thinking models from the default options. Honestly, most people are not going to have the kind of hardware to run them at high enough speeds to make the thinking steps worth it - at least not on desktop. Especially for smaller models that even unquantize can still catch themselves in an infinite thinking loop. 

Overall, I'd try to find finetunes and test them if I were you. What I recommended is what I tested and found to be an improvement over what came before. I'd stay away from mergers, ablated and uncensored models. Just try to find RP and story finetunes that are open source and on huggingface to test. Also, and you did not hear this from me, try IBM's Granite 3.3 8B model... for a model designed for office work and which was instruction trained to be harmless and safe boy does it follow nsfw instructions well. And I do mean NSFW. And it's Apache 2.0 :)


As for IQ quants, they can offer similar performance to KV cache at smaller sized - but are only similarly fast under ROCm and CUDA - significant slowdowns under Vulkan and CPU. I know Ollama supports them, though I don't think you can DL IQ quant from their site directly. An IQ4_XS should be very similar to a Q4_KS in output - within margin or error for RP and story purposes - but substantially smaller. 

I wrote earlier that I couldn't get the characters to respond offline. I now have it working. I noticed on reddit some had the same problem. What I did was delete the default model and down the lemon one. The characters all now respond in chat offline. Awesome program. If there is a better payment plan or perhaps donation set I'd be interested in donating a little. 

Nice! I do have this if you'd like, but no need, your nice words are enough! https://www.patreon.com/HammerAI

Stupid question: why provide a flatpack in place an appimage if you're not going use flathub?

Oh, so I just use the default Electron Forge makers:  https://www.electronforge.io/config/makers/flatpak

I can look into putting it on flathub, but I don't have a linux machine, so just haven't actually tested any of the Linux apps myself. Sorry about that. Anything I need to fix with them?

Not really, they work. I would probably prefer a more LM Studio like approach where we can select if we want to run local models via CUDA, Vulkan or CPU but no, the applications works. 

That being said, just out of curiosity, aren't you running the website on a linux servers? I mean it's probably a docker container but still, I'd be actually shocked if you told me you're running the website on a windows servers.

(1 edit)

Okay interesting. I think the new update (which is in beta right now at https://github.com/hammer-ai/hammerai/releases/tag/v0.0.206) might help. 

The website is linux servers + Runpod. But it's a different codepath than the desktop app. So doing the website doesn't really help with the Electron app.

"Installation has failed: There was an error while installing the application" Nothing in the log and subsequent attempts to install fail as well, as trying to install previous versions.

Can you try closing and trying again? Maybe turning on and off the computer? Sorry about that.

I've done that and more: terminated processes, uninstalled, reinstalled, nothing helps. Does your installer leave anything in the registry? It looks like a failed installation will leave something that doesn't get cleaned up. I can't seem to find it.

i got a small feedback. Could you also make it, that you can not only create a person as a character but also an RP? like with scenery and not only the one character based creation

Would love if you can join the Discord and post there! Then we can chat more about the feature.

so im using this offline but the ai text back really slow one word at a time it takes a few mins for it to say a line or to is there a way i can fix this and speed it up so im not w8 so long for a short reply ?

You probably don't have a powerful enough computer for the model you chose. Can you try a smaller one?

ant the 8b a small one ?  im not sure how that works

i some what fix it but after the up date im getting the ai randomly saying  <[im end]>  <[im start}>  <[im done]>  thin it just go crazy and go on about its own story and makes me and character go on a long thing making both sides talk when i dint type . what is going on and how can i fix ?

Could you join the Discord to chat more? We can help you in there!

HI, I'm a developer and I want to know, how I can help you with this project?

Sure, DM me on Discord, hammer_ai is my username. Share your resume + Github + some projects you've worked on, please!

whats the extent of the NSFW content?

Uncensored! You can see the content policy here: https://www.hammerai.com/terms

Hello, could you at least roughly say when local image generation will be available?

So I have it mostly working already, but not yet polished up. I do want it really badly, and will update you when it's ready. Sorry for the long delay.

uninstalled an older version and went to install a newer verion got this 

16900> 2024-12-08 05:31:10> Program: Starting Squirrel Updater: --install .

16900> 2024-12-08 05:31:11> Program: Starting install, writing to C:\Users\jrski\AppData\Local\SquirrelTemp

16900> 2024-12-08 05:31:11> Program: About to install to: C:\Users\jrski\AppData\Local\Discord

16900> 2024-12-08 05:31:11> SingleGlobalInstance: Grabbing lockfile with timeout of 00:00:10

16900> 2024-12-08 05:31:11> CheckForUpdateImpl: Reading RELEASES file from C:\Users\jrski\AppData\Local\SquirrelTemp

16900> 2024-12-08 05:31:11> CheckForUpdateImpl: Remote version 1.0.9059 differs from local 

16900> 2024-12-08 05:31:11> CheckForUpdateImpl: First run or local directory is corrupt, starting from scratch

16900> 2024-12-08 05:31:11> ApplyReleasesImpl: Writing files to app directory: C:\Users\jrski\AppData\Local\Discord\app-1.0.9059

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\chrome_200_percent.pak to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\chrome_200_percent.pak

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\chrome_100_percent.pak to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\chrome_100_percent.pak

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\app.ico to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\app.ico

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\d3dcompiler_47.dll to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\d3dcompiler_47.dll

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\Discord.exe.sig to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\Discord.exe.sig

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\Discord.exe to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\Discord.exe

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\ffmpeg.dll to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\ffmpeg.dll

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\icudtl.dat to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\icudtl.dat

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\installer.db to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\installer.db

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\libEGL.dll to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\libEGL.dll

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\libGLESv2.dll to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\libGLESv2.dll

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\resources.pak to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\resources.pak

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\snapshot_blob.bin to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\snapshot_blob.bin

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\v8_context_snapshot.bin to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\v8_context_snapshot.bin

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\vk_swiftshader_icd.json to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\vk_swiftshader_icd.json

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\vk_swiftshader.dll to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\vk_swiftshader.dll

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\updater.node to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\updater.node

16900> 2024-12-08 05:31:14> ApplyReleasesImpl: Moving file C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\lib\net45\vulkan-1.dll to C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\vulkan-1.dll

16900> 2024-12-08 05:31:15> ApplyReleasesImpl: Squirrel Enabled Apps: [C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\Discord.exe]

16900> 2024-12-08 05:31:19> Utility: Process Started: C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\Discord.exe --squirrel-install 1.0.9059, pid 9240

16900> 2024-12-08 05:31:30> Utility: Received exitcode 0 from process C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\Discord.exe

16900> 2024-12-08 05:31:30> ApplyReleasesImpl: ran C:\Users\jrski\AppData\Local\Discord\app-1.0.9059\Discord.exe, pid 18364

16900> 2024-12-08 05:31:30> ApplyReleasesImpl: Starting fixPinnedExecutables

16900> 2024-12-08 05:31:30> ApplyReleasesImpl: fixPinnedExecutables: newCurrentFolder: app-1.0.9059

16900> 2024-12-08 05:31:30> ApplyReleasesImpl: File 'C:\Users\jrski\AppData\Roaming\Microsoft\Windows\Start Menu\Programs\Accessibility\LiveCaptions.lnk' could not be converted into a valid ShellLink: System.IO.FileNotFoundException: The system cannot find the file specified. (Exception from HRESULT: 0x80070002)

   at Squirrel.Shell.ShellLink.IShellLinkW.Resolve(IntPtr hWnd, UInt32 fFlags)

   at Squirrel.Shell.ShellLink.Open(String linkFile, IntPtr hWnd, EShellLinkResolveFlags resolveFlags, UInt16 timeOut)

   at Squirrel.UpdateManager.ApplyReleasesImpl.<fixPinnedExecutables>b__11_0(FileInfo file)

16900> 2024-12-08 05:31:30> ApplyReleasesImpl: File 'C:\Users\jrski\AppData\Roaming\Microsoft\Windows\Start Menu\Programs\Accessibility\Narrator.lnk' could not be converted into a valid ShellLink: System.IO.FileNotFoundException: The system cannot find the file specified. (Exception from HRESULT: 0x80070002)

   at Squirrel.Shell.ShellLink.IShellLinkW.Resolve(IntPtr hWnd, UInt32 fFlags)

   at Squirrel.Shell.ShellLink.Open(String linkFile, IntPtr hWnd, EShellLinkResolveFlags resolveFlags, UInt16 timeOut)

   at Squirrel.UpdateManager.ApplyReleasesImpl.<fixPinnedExecutables>b__11_0(FileInfo file)

16900> 2024-12-08 05:31:30> ApplyReleasesImpl: File 'C:\Users\jrski\AppData\Roaming\Microsoft\Windows\Start Menu\Programs\Accessibility\On-Screen Keyboard.lnk' could not be converted into a valid ShellLink: System.IO.FileNotFoundException: The system cannot find the file specified. (Exception from HRESULT: 0x80070002)

   at Squirrel.Shell.ShellLink.IShellLinkW.Resolve(IntPtr hWnd, UInt32 fFlags)

   at Squirrel.Shell.ShellLink.Open(String linkFile, IntPtr hWnd, EShellLinkResolveFlags resolveFlags, UInt16 timeOut)

   at Squirrel.UpdateManager.ApplyReleasesImpl.<fixPinnedExecutables>b__11_0(FileInfo file)

16900> 2024-12-08 05:31:30> ApplyReleasesImpl: File 'C:\Users\jrski\AppData\Roaming\Microsoft\Windows\Start Menu\Programs\Accessibility\VoiceAccess.lnk' could not be converted into a valid ShellLink: System.IO.FileNotFoundException: The system cannot find the file specified. (Exception from HRESULT: 0x80070002)

   at Squirrel.Shell.ShellLink.IShellLinkW.Resolve(IntPtr hWnd, UInt32 fFlags)

   at Squirrel.Shell.ShellLink.Open(String linkFile, IntPtr hWnd, EShellLinkResolveFlags resolveFlags, UInt16 timeOut)

   at Squirrel.UpdateManager.ApplyReleasesImpl.<fixPinnedExecutables>b__11_0(FileInfo file)

16900> 2024-12-08 05:31:30> ApplyReleasesImpl: Updating shortcut C:\Users\jrski\AppData\Roaming\Microsoft\Windows\Start Menu\Programs\Discord Inc\Discord.lnk

16900> 2024-12-08 05:31:31> ApplyReleasesImpl: Updating shortcut C:\Users\jrski\OneDrive\Desktop\Discord.lnk

16900> 2024-12-08 05:31:31> ApplyReleasesImpl: Fixing up tray icons

16900> 2024-12-08 05:31:31> ApplyReleasesImpl: cleanDeadVersions: for version 1.0.9059

16900> 2024-12-08 05:31:31> ApplyReleasesImpl: cleanDeadVersions: exclude current version folder app-1.0.9059

any help would be appreciated so i can install a newer version

Sorry about that! It's a weird issue people get into related to Discord interfering with our update. The solution is to either restart your computer or kill all "HammerAI" and "Ollama" processes under Task Manager. Sorry again.

You need to start hammer Ai as admin. then the downloade works for me. otherwise same issue here

(1 edit)

When is it going to be available for phones? Just wanna know, not stress you out or something.

(+1)

No specific timeframe, sorry! As a one-person project, I can only do so much 😭 

PS. If I could find someone to work with me, I'd definitely go faster, so if anyone reading here is a dev, please reach out!

(-3)

can you make everything free pls

(1 edit)

everytime i say hi the ai breaks chraracter and be a another chatgpt clone can you fix that i dont like paying to just have a roleplay thank you besides it is a little paywall like

(1 edit) (+1)

Hi! Sorry about that, usually that's because the character you're chatting with wasn't written very well. If you try with one of these is it any better? https://www.hammerai.com/characters?tag=Featured

In terms of paywall like, it is 100% free to chat with the cloud-hosted LLM Smart Lemon Cookie 7B, or with any local LLM! But I am a solo dev building this, so I made saving chats is a paid feature, sorry about that. If it makes you feel better, I spend all the money I make to pay other contractors to help build it with me. And I have a 100% no questions asked refund policy, if you end up not being happy with it. Again, sorry for the issues.

Hello, what is the memory system like on this? Does it forget things or does it remember everything?

It depends! The lower parameter (i.e. 7B and 8B) models forget more, and the higher (20B and 70B) forget less.

Can I trust it?

(1 edit)

Yes, it is safe! But you don't need to take my word for it, you can also ask in the Discord, or maybe read this review from someone on Reddit? They said:

> All in all it is one of the best options for a locally installed AI chatbot to use privately. Using wireshark, iftop, and other tools I didn't notice any unnecessary calls or shady traffic. Which is awesome. However, please be aware you lose some of that privacy as you need to log to discord to access basic docs for the app.

https://www.reddit.com/r/HammerAI/comments/1i2a9tp/60ish_day_review/

PS. I'm working on adding docs to the site to help address their privacy concerns (they don't like Discord)

Can you have multiple ai bots chatting to you ie: 
you say something then both bots type in their separate responses

Not yet, sorry! But on the roadmap.

Viewing most recent comments 1 to 16 of 60 · Next page · Last page