Do you intend to reach feature parity with something like MacWhisper? I'd love to switch to something open source, but automated meeting detection, push to transcribe (with custom rewrite actions) are two features I've learned to love, beside basic transcript. I also enjoy the automatic transcription from an audio, video or a even a YouTube link.
But because MacWhisper does not store transcripts or do much with them (other than giving you export options), there are some missed opportunities: I'd love to be able to add project tags to transcripts, so that any new transcript is summarized with the context of all previous transcript summaries that share the same tag. Thinking about it maybe I should build a Logseq extension to do that myself as I store all my meeting summaries there anyway.
Speaker detection is not great in MacWhisper (at least in my context where I work mostly with non native English speakers), so that would be a good differentiation too.
I'm always amazed at these relatively tiny projects that "launch" with a "customers" list that reads like they've spent 10 years doing hard outbound enterprise sales: Google, Intel, Apple, Amazon, Deloitte, IBM, Ford, Meta, Uber, Tencent, etc.
Congrats on the launch. I never understood why an AI meeting notetaker needed sota LLMs and subscriptions (talking about literally all the other notetakers) - thanks for making it local first.
I use a locally patched up whisperx + qwen3:1.7 + nomic embed (ofcourse with a swift script that picks up the audio buffer from microphone) and it works just fine.
Rarely i create next steps / sop from the transcript - i use gemini 2.5 and export it as pdf. I’ll give Hyprnote a try soon.
I hope, since it’s opensource, you are thinking about exposing api / hooks for downstream tasks.
This is perfect timing! I just cancelled my fireflies.ai subscription yesterday because it just felt unnecessary. I prefer using less platforms and more tools, especially those that can work under the surface.
Would be great if you could include in your launch message how you plan to monetize this. Everybody likes open source software and local-first is excellent too, but if you mention YC too then everybody also knows that there is no free lunch, so what's coming down the line would be good to know before deciding whether to give it a shot or just move on.
How are you balancing accuracy vs. time-to-word-on-live-transcript? Is this something you're actively balancing, or can allow an end user to tune?
I find myself often using otter.ai - because while it's inferior to Whisper in many ways, and anything but on-device, it's able to show words on the live transcript with minimal delay, rather than waiting for a moment of silence or for a multi-second buffer to fill. That's vital if I'm using my live transcription both to drive async summarization/notes and for my operational use in the same call, to let me speed-read to catch up to a question that was just posed to me while I was multitasking (or doing research for a prior question!)
It sometimes boggles me that we consider the latency of keypress-to-character-on-screen to be sacrosanct, but are fine with waiting for a phrase or paragraph or even an entire conversation to be complete before visualizing its transcription. Being able to control this would be incredible.
I just downloaded on mac M4 pro mini. I installed the apple silicon version and try to launch it and it fails. No error message or anything. Just the icon keep bouncing on the dock. I assumed it needs some privacy and screen recording and audio permissions and explicitly gave them, however still just jumps on the dock and the app does not open. (OS, mac sequoia 15.5)
Looks really cool - I noticed Enterprise has smart consent management?
The thing I think some enterprise customers are worried about in this space is that in many jurisdictions you legally need to disclose recording - having a bot join the call can do that disclosure - but users hate the bot and it takes up too much visibility on many of these calls.
Would love to learn more about your approach there
Congrats on the launch! I'm very bullish on how powerful <10B-param models are becoming, so the on-device angle is cool (and great for your bottom line too, as it's cheaper for you to run).
Something that I think is interesting about AI note taking products is focus. How does it choose what's important vs what isn't? The better it is at distinguishing the signal from the noise, the more powerful it is. I wonder if there is an in-context learning angle here where you can update the model weights (either directly or via LoRA) as you get to know the user better. And, of course, everything stays private and on-device.
I was talking about this a week ago. One person wanted to make a pdf tutorial of how to use a software. I asked him to record himself in teams and share his screen and have AI take notes. It will create a fabulous summary with snapshots of everything he is going over.
Look forward to testing the Windows version. Hope it has ability to also upload recordings, etc.
Meetily is nice but setup feels too convoluted, with a backend and frontend being required to separately install...
This is really cool! I've been using Obsidian more and more as a second brain and getting data in has consistently been the point of failure, so I've been wanting something just like this. Specifically something that runs locally and offline.
Is the future goal of Hyprnote specifically meeting notes and leaning into features around meeting notes, or more general note taking and recall features?
Since this isn't available yet on Windows, what would be the glue & duct tape alternative? Record audio and dump it in chatGPT? Or do you need to create some kind of automation with n8n / Zapier? I don't have that many meetings but it could be nice to have
The only issue I have with those tools, and I have not seen a single one even acknowledge this, is that it becomes completely useless when holding meetings in a hybrid fashion where some people are remote and others are in the office with a shared mic.
Almost all of our meetings are hybrid in this way, and it's a real pain having almost half of the meeting be identified as a single individual talking because the mic is hooked up to their machine.
It's a total dealbreaker for us, and we won't use such tools until that problem is solved.
another free tier (but not opensource) recording tool Ive been using is MacWhisper. Does this and more all locally too. Will try hyprnote out because its neat to do the transcription in real time and its note taking purposes
Looks great. From my experience Tauri team has no clue what mobile is and they're not interested in fixing mobile issues. I can already tell the mobile version will be a disappointment.
Super cool, congrats on the launch - will be trying this soon! I noticed it’s using Tauri - what are your main takeaways from building a local inference desktop app with it?
Congrats! I'm currently a Granola user, and wanted to build this myself a while back. But I probably wouldn't have gone as far as fine-tuning a small model for meeting summarization. Can't wait to try it out!
Do you intend to reach feature parity with something like MacWhisper? I'd love to switch to something open source, but automated meeting detection, push to transcribe (with custom rewrite actions) are two features I've learned to love, beside basic transcript. I also enjoy the automatic transcription from an audio, video or a even a YouTube link.
But because MacWhisper does not store transcripts or do much with them (other than giving you export options), there are some missed opportunities: I'd love to be able to add project tags to transcripts, so that any new transcript is summarized with the context of all previous transcript summaries that share the same tag. Thinking about it maybe I should build a Logseq extension to do that myself as I store all my meeting summaries there anyway.
Speaker detection is not great in MacWhisper (at least in my context where I work mostly with non native English speakers), so that would be a good differentiation too.
I'm always amazed at these relatively tiny projects that "launch" with a "customers" list that reads like they've spent 10 years doing hard outbound enterprise sales: Google, Intel, Apple, Amazon, Deloitte, IBM, Ford, Meta, Uber, Tencent, etc.
Congrats on the launch. I never understood why an AI meeting notetaker needed sota LLMs and subscriptions (talking about literally all the other notetakers) - thanks for making it local first. I use a locally patched up whisperx + qwen3:1.7 + nomic embed (ofcourse with a swift script that picks up the audio buffer from microphone) and it works just fine. Rarely i create next steps / sop from the transcript - i use gemini 2.5 and export it as pdf. I’ll give Hyprnote a try soon.
I hope, since it’s opensource, you are thinking about exposing api / hooks for downstream tasks.
This is perfect timing! I just cancelled my fireflies.ai subscription yesterday because it just felt unnecessary. I prefer using less platforms and more tools, especially those that can work under the surface.
Nice!
Would be great if you could include in your launch message how you plan to monetize this. Everybody likes open source software and local-first is excellent too, but if you mention YC too then everybody also knows that there is no free lunch, so what's coming down the line would be good to know before deciding whether to give it a shot or just move on.
How are you balancing accuracy vs. time-to-word-on-live-transcript? Is this something you're actively balancing, or can allow an end user to tune?
I find myself often using otter.ai - because while it's inferior to Whisper in many ways, and anything but on-device, it's able to show words on the live transcript with minimal delay, rather than waiting for a moment of silence or for a multi-second buffer to fill. That's vital if I'm using my live transcription both to drive async summarization/notes and for my operational use in the same call, to let me speed-read to catch up to a question that was just posed to me while I was multitasking (or doing research for a prior question!)
It sometimes boggles me that we consider the latency of keypress-to-character-on-screen to be sacrosanct, but are fine with waiting for a phrase or paragraph or even an entire conversation to be complete before visualizing its transcription. Being able to control this would be incredible.
I would like to try this on Linux
I just downloaded on mac M4 pro mini. I installed the apple silicon version and try to launch it and it fails. No error message or anything. Just the icon keep bouncing on the dock. I assumed it needs some privacy and screen recording and audio permissions and explicitly gave them, however still just jumps on the dock and the app does not open. (OS, mac sequoia 15.5)
Looks really cool - I noticed Enterprise has smart consent management?
The thing I think some enterprise customers are worried about in this space is that in many jurisdictions you legally need to disclose recording - having a bot join the call can do that disclosure - but users hate the bot and it takes up too much visibility on many of these calls.
Would love to learn more about your approach there
Congrats on the launch! I'm very bullish on how powerful <10B-param models are becoming, so the on-device angle is cool (and great for your bottom line too, as it's cheaper for you to run).
Something that I think is interesting about AI note taking products is focus. How does it choose what's important vs what isn't? The better it is at distinguishing the signal from the noise, the more powerful it is. I wonder if there is an in-context learning angle here where you can update the model weights (either directly or via LoRA) as you get to know the user better. And, of course, everything stays private and on-device.
Congratulations! Is there a mobile version as well, especially for Android?
I was talking about this a week ago. One person wanted to make a pdf tutorial of how to use a software. I asked him to record himself in teams and share his screen and have AI take notes. It will create a fabulous summary with snapshots of everything he is going over.
Look forward to testing the Windows version. Hope it has ability to also upload recordings, etc. Meetily is nice but setup feels too convoluted, with a backend and frontend being required to separately install...
This is really cool! I've been using Obsidian more and more as a second brain and getting data in has consistently been the point of failure, so I've been wanting something just like this. Specifically something that runs locally and offline.
Is the future goal of Hyprnote specifically meeting notes and leaning into features around meeting notes, or more general note taking and recall features?
Congrats on the launch. Is there a reason why the app isn't sandboxed?
Since this isn't available yet on Windows, what would be the glue & duct tape alternative? Record audio and dump it in chatGPT? Or do you need to create some kind of automation with n8n / Zapier? I don't have that many meetings but it could be nice to have
The only issue I have with those tools, and I have not seen a single one even acknowledge this, is that it becomes completely useless when holding meetings in a hybrid fashion where some people are remote and others are in the office with a shared mic.
Almost all of our meetings are hybrid in this way, and it's a real pain having almost half of the meeting be identified as a single individual talking because the mic is hooked up to their machine.
It's a total dealbreaker for us, and we won't use such tools until that problem is solved.
another free tier (but not opensource) recording tool Ive been using is MacWhisper. Does this and more all locally too. Will try hyprnote out because its neat to do the transcription in real time and its note taking purposes
https://goodsnooze.gumroad.com/l/macwhisper
Looks great. From my experience Tauri team has no clue what mobile is and they're not interested in fixing mobile issues. I can already tell the mobile version will be a disappointment.
Really cool - how does it compare to Granola outside of the OSS part?
congrats, app looks gorgeous. def a good tauri codebase to study ( been using https://deepwiki.com/fastrepl/hyprnote)
any interest in the Cluely-style live conversation help/overlay?
Nicely done, I or someone can push the translation option too. Well done.
why use whisper over parakeet? how will you monetise?
Super cool, congrats on the launch - will be trying this soon! I noticed it’s using Tauri - what are your main takeaways from building a local inference desktop app with it?
Congrats! I'm currently a Granola user, and wanted to build this myself a while back. But I probably wouldn't have gone as far as fine-tuning a small model for meeting summarization. Can't wait to try it out!
Why in the world is there _background music_ when I start the app?!
Looks promising, but "Linux maybe"? Signing off.