Show stories

mbitsnbites 3 days ago

Show HN: A luma dependent chroma compression algorithm (image compression)

This article presents a novel image compression algorithm that adapts the block size and luma-dependent chroma compression based on the spatial domain, leading to improved compression efficiency while maintaining image quality.

bitsnbites.eu
4 0
Summary
Show HN: I saw this cool navigation reveal, so I made a simple HTML+CSS version
momciloo about 4 hours ago

Show HN: I saw this cool navigation reveal, so I made a simple HTML+CSS version

Two clip-paths, over the navigation:

- The first clip-path is a circle (top-left corner) - The second clip-path is a polygon, that acts like a ray (hardcoded, can be improved)

The original work by Iventions Events https://iventions.com/ uses JavaScript, but I found CSS-only approach more fun

Here's a demo and the codebase: https://github.com/Momciloo/fun-with-clip-path

github.com
30 5
Summary
latentio about 1 hour ago

Show HN: PalettePoint – AI color palette generator from text or images

Hey HN, I built PalettePoint (https://palettepoint.com) because picking colors for projects was always the part I dreaded most. I'm not a designer, and every time I'd start a new project I'd spend way too long on color palette websites refreshing until something looked okay. So I built a tool where you just describe what you want, "warm coffee shop vibes" or "clean SaaS dashboard", and it generates a proper palette with color theory behind it. You can also upload an image and it'll extract a palette from that or get inspired by it. What it does: Text-to-palette: describe a mood/theme and get 3-7 colors Image-to-palette: upload a photo and extract its colors or get inspired by it. 10 style modes (analogous, triadic, complementary, pastel, etc.) Pin base colors and generate around them Gallery with 100K+ palettes Free tools: contrast checker, color mixer, gradient generator, color converter Export to CSS, SCSS, Tailwind, JSON, (more to come) Palette visualiser on logos, layouts etc .. (more to come)

Would love feedback, especially on the palette quality and if the styles actually feel distinct from each other.

palettepoint.com
2 0
Summary
Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox
isitcontent 1 day ago

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

Example repo: https://github.com/valdanylchuk/breezydemo

The underlying ESP-IDF component: https://github.com/valdanylchuk/breezybox

It is something like Raspberry Pi, but without the overhead of a full server-grade OS.

It captures a lot of the old school DOS era coding experience. I created a custom fast text mode driver, plan to add VGA-like graphics next. ANSI text demos run smooth, as you can see in the demo video featured in the Readme.

App installs also work smoothly. The first time it installed 6 apps from my git repo with one command, felt like, "OMG, I got homebrew to run on a toaster!" And best of all, it can install from any repo, no approvals or waiting, you just publish a compatible ELF file in your release.

Coverage:

Hackaday: https://hackaday.com/2026/02/06/breezybox-a-busybox-like-she...

Hackster.io: https://www.hackster.io/news/valentyn-danylchuk-s-breezybox-...

Reddit: https://www.reddit.com/r/esp32/comments/1qq503c/i_made_an_in...

github.com
287 38
Summary
Show HN: Kappal – CLI to Run Docker Compose YML on Kubernetes for Local Dev
sandGorgon 3 days ago

Show HN: Kappal – CLI to Run Docker Compose YML on Kubernetes for Local Dev

Kappal is an open-source software project that provides a platform for building and deploying containerized applications. It offers a simplified workflow for managing infrastructure, orchestrating services, and automating deployments across multiple environments.

github.com
32 15
Summary
Show HN: If you lose your memory, how to regain access to your computer?
eljojo 1 day ago

Show HN: If you lose your memory, how to regain access to your computer?

Due to bike-induced concussions, I've been worried for a while about losing my memory and not being able to log back in.

I combined shamir secret sharing (hashicorp vault's implementation) with age-encryption, and packaged it using WASM for a neat in-browser offline UX.

The idea is that if something happens to me, my friends and family would help me get back access to the data that matters most to me. 5 out of 7 friends need to agree for the vault to unlock.

Try out the demo in the website, it runs entirely in your browser!

eljojo.github.io
357 216
Summary
vecti 1 day ago

Show HN: I spent 4 years building a UI design tool with only the features I use

Hello everyone!

I'm a solo developer who's been doing UI/UX work since 2007. Over the years, I watched design tools evolve from lightweight products into bloated feature-heavy platforms. I kept finding myself using a small amount of the features while the rest just mostly got in the way.

So a few years ago I set out to build a design tool just like I wanted. So I built Vecti with what I actually need: pixel-perfect grid snapping, a performant canvas renderer, shared asset libraries, and export/presentation features. No collaborative whiteboarding. No plugin ecosystem. No enterprise features. Just the design loop.

Four years later, I can proudly show it off. Built and hosted in the EU with European privacy regulations. Free tier available (no credit card, one editor forever).

On privacy: I use some basic analytics (page views, referrers) but zero tracking inside the app itself. No session recordings, no behavior analytics, no third-party scripts beyond the essentials.

If you're a solo designer or small team who wants a tool that stays out of your way, I'd genuinely appreciate your feedback: https://vecti.com

Happy to answer questions about the tech stack, architecture decisions, why certain features didn't make the cut, or what's next.

vecti.com
371 169
Summary
Show HN: I built a <400ms latency voice agent that runs on a 4gb vram GTX 1650"
shubham-coder about 3 hours ago

Show HN: I built a <400ms latency voice agent that runs on a 4gb vram GTX 1650"

I built a Voice agent platform my drobotics lab of my university..which is already being cloned by 330+ people within 12hrs .. I am a first year cse student and so I tried to figure out a way to actually run everything on my laptop and working on it currently to completely transform to edge ai voice assistants for the robotics and 100% private and local control of robotics related project of my lab..

The intersting features are : 1> I used json rag with real time embeddings so that for a few specs and info we don't need to set a whole pipeline..

I have already built " Hierarchical Agentic Rag with Hybrid Search ( knowledge graph + vector search) u can view that on my profile ...

I am actively trying to share as much as possible related to it but that project is actually linked with a huge set of files it's 693k points of data with pgvector+ postgress .. give a visit u will get more idea from that

2> I had tried every sort of whisper models.. faster whisper .. turbo or anything u can u think of ..even with a self c++ engine .. but that model itself was hallucintion prone architecture..

Then I moved to parakeet tdt with silero vad and not parakeet rnn for better speed and optimisations .. repo has further details ..

3> fine tuned a dataset from anthropic rlhf through space and glinner and convert that to a perfect training dataset of the Lama 3.2 3b ..

I will attach the dataset of u need or will upload that to hugging face if u want to use it for yourself..

4> attached phonetic correctors for both output from parakeet and llama for better tts working .

5> I used setfit to route the queries and confidence based semantic search for faster and accurate as much as possible

6> I am using sherpa onxx and qued the tts and stt and everything but as a experimentation I have also achieved llama generating respond and kokora processing as a batch with a full nyc working as well and everything on my laptop...

7> along with these my frontend also relies on heavy three.js and 3d view files but I had applied optimisations there which works perfectly with everything together on the laptop..

8> I also applied glued interaction to the llm model .. implemented FIFO with 5 interactions and storing them for future fine tuning and phonetic words additions.

Pls give a visit it and let me know if I should learn something new ..

One kind note : as a enthusiast spending so much energy on these things things .. I have taken help from ai for the md files and expansion or explanations in the codes for better help of every single person...

github.com
2 1
Summary
Show HN: Smooth CLI – Token-efficient browser for AI agents
antves 2 days ago

Show HN: Smooth CLI – Token-efficient browser for AI agents

Hi HN! Smooth CLI (https://www.smooth.sh) is a browser that agents like Claude Code can use to navigate the web reliably, quickly, and affordably. It lets agents specify tasks using natural language, hiding UI complexity, and allowing them to focus on higher-level intents to carry out complex web tasks. It can also use your IP address while running browsers in the cloud, which helps a lot with roadblocks like captchas (https://docs.smooth.sh/features/use-my-ip).

Here’s a demo: https://www.youtube.com/watch?v=62jthcU705k Docs start at https://docs.smooth.sh.

Agents like Claude Code, etc are amazing but mostly restrained to the CLI, while a ton of valuable work needs a browser. This is a fundamental limitation to what these agents can do.

So far, attempts to add browsers to these agents (Claude’s built-in --chrome, Playwright MCP, agent-browser, etc.) all have interfaces that are unnatural for browsing. They expose hundreds of tools - e.g. click, type, select, etc - and the action space is too complex. (For an example, see the low-level details listed at https://github.com/vercel-labs/agent-browser). Also, they don’t handle the billion edge cases of the internet like iframes nested in iframes nested in shadow-doms and so on. The internet is super messy! Tools that rely on the accessibility tree, in particular, unfortunately do not work for a lot of websites.

We believe that these tools are at the wrong level of abstraction: they make the agent focus on UI details instead of the task to be accomplished.

Using a giant general-purpose model like Opus to click on buttons and fill out forms ends up being slow and expensive. The context window gets bogged down with details like clicks and keystrokes, and the model has to figure out how to do browser navigation each time. A smaller model in a system specifically designed for browsing can actually do this much better and at a fraction of the cost and latency.

Security matters too - probably more than people realize. When you run an agent on the web, you should treat it like an untrusted actor. It should access the web using a sandboxed machine and have minimal permissions by default. Virtual browsers are the perfect environment for that. There’s a good write up by Paul Kinlan that explains this very well (see https://aifoc.us/the-browser-is-the-sandbox and https://news.ycombinator.com/item?id=46762150). Browsers were built to interact with untrusted software safely. They’re an isolation boundary that already works.

Smooth CLI is a browser designed for agents based on what they’re good at. We expose a higher-level interface to let the agent think in terms of goals and tasks, not low-level details.

For example, instead of this:

  click(x=342, y=128)
  type("search query")
  click(x=401, y=130)
  scroll(down=500)
  click(x=220, y=340)
  ...50 more steps
Your agent just says:

  Search for flights from NYC to LA and find the cheapest option
Agents like Claude Code can use the Smooth CLI to extract hard-to-reach data, fill-in forms, download files, interact with dynamic content, handle authentication, vibe-test apps, and a lot more.

Smooth enables agents to launch as many browsers and tasks as they want, autonomously, and on-demand. If the agent is carrying out work on someone’s behalf, the agent’s browser presents itself to the web as a device on the user’s network. The need for this feature may diminish over time, but for now it’s a necessary primitive. To support this, Smooth offers a “self” proxy that creates a secure tunnel and routes all browser traffic through your machine’s IP address (https://docs.smooth.sh/features/use-my-ip). This is one of our favorite features because it makes the agent look like it’s running on your machine, while keeping all the benefits of running in the cloud.

We also take away as much security responsibility from the agent as possible. The agent should not be aware of authentication details or be responsible for handling malicious behavior such as prompt injections. While some security responsibility will always remain with the agent, the browser should minimize this burden as much as possible.

We’re biased of course, but in our tests, running Claude with Smooth CLI has been 20x faster and 5x cheaper than Claude Code with the --chrome flag (https://www.smooth.sh/images/comparison.gif). Happy to explain further how we’ve tested this and to answer any questions about it!

Instructions to install: https://docs.smooth.sh/cli. Plans and pricing: https://docs.smooth.sh/pricing.

It’s free to try, and we'd love to get feedback/ideas if you give it a go :)

We’d love to hear what you think, especially if you’ve tried using browsers with AI agents. Happy to answer questions, dig into tradeoffs, or explain any part of the design and implementation!

docs.smooth.sh
96 70
Summary
Show HN: R3forth, a ColorForth-inspired language with a tiny VM
phreda4 about 24 hours ago

Show HN: R3forth, a ColorForth-inspired language with a tiny VM

r3 is a high-performance, open-source programming language and environment that focuses on simplicity, efficiency, and creativity. It provides a powerful set of tools for developing a wide range of applications, from games and graphics to data visualization and automation.

github.com
84 16
Summary
Keyframe about 4 hours ago

Show HN: Stacky – certain block game clone

As a long-time programmer this all just feels all sorts of wrong, but also invigorating. Vibe "coded" the whole thing from 0-100 over the course of few days, on and off. I have no intentions of developing it further since it's obvious what it is; I would absolutely love to work on a licensed game and do it proper with all the various ideas I have, since this is maybe 10% of what I want in such a game, but I heard somewhere licensing is cost-prohibitive.

Putting AI shame aside, it really allowed me to explore so many things in a short amount of time that it feels good, almost enough to compensate the feeling of shame using AI to begin with.

WebGPU isn't in there, although it's in another experimental version, part are indeed written in Rust (game logic).

It has:

- lock delay / grace period (allowing for 15 moves)

- DAS (Delayed Auto Shift) and ARR (Auto Repeat Rate for continuous movement) for horizontal and soft drop movements

- SRS wall kicks (Super Rotation System) to rotate pieces in-place

- Shift+Enter "hidden" level select on the main screen

- Shift+D for debug/performance indicator panel

- Several ranodmizers including 7-bag and NES ones

- combo system with difficulty (time) modes (easy by default) - x2: DOUBLE STRIKE, x5: CHAIN REACTION, x7: MEGA COMBO, x9: PHOSPHOR OVERLOAD, x10+: CRITICAL MASS

- backgrounds which change over time or you can change them with SHIFT+B (B turns it off/on) which react both to music (FFT!) and to your game play when you clear lines

- normal and two phosphor rendering modes of game field (R to toggle)

- CRT Filter (shift+c to toggle)

- F for full screen toggle

- A for previous song, S for pause song, D for next song (all songs made with Suno, of course)

and many more. It was a fun experience for sure, just not sure how to feel about it. On one hand I understand it wouldn't look like it does without my input, and it was a lot of what felt like work (intense sessions looking over the output, correcting etc), yet it doesn't feel like I really made anything by myself. I had fun though.

While at it, created a small demo as well which isn't a game yet: https://www.susmel.com/rolly/ and also something to play with parametric curves here: https://www.susmel.com/graphy/

all within a span of a couple of days while we were having our third baby. The future is weird, and I'm still not sure whether I like it or not. One thing is sure - it's here to stay. Peace out, my friends!

susmel.com
3 0
Summary
xeouz about 5 hours ago

Show HN: A toy compiler I built in high school (runs in browser)

Hey HN,

Indian high schooler here, currently prepping for JEE, thought itd be nice to share here.

Three years ago in 9th/10th grade I got a knack for coding, I taught myself and made a custom compiler with LLVM to try to learn C++. So I spent a lot of time learning LLVM from the docs and also C++. It's not some marvelous piece of engineering,

I designed the syntax to be a mix of C and what I wished C looked like back in 9th grade.

It has:

  - Basic types like bool, int, double, float, char etc. with type casting
  - Variables, Arrays, Assign operators & Shorthands
  - Conditionals (if/else-if/else), Operators (and/or), arithmetics (parenthesis etc)
  - Arrays and indexing stuff
  - C style Loops (for/while) and break/continue
  - Structs and dot accessing
  - extern C interop with the "extern" keyword
Some challenges I faced:

  - Emscripten and WASM, as I also had to make it run on my demo website
  - Learning typescript and all for the website (lol)
  - Custom parser with basic error reporting and Semantic analysis was a PITA for my undeveloped brain (I was 15)
  - Learning LLVM from the docs
Important Learnings:

  - Testing is a very important aspect of making software, I skipped it - big regret
  - Learning how computers interpret text
  - Programming in general was a new tour for me
  - I appreciate unique_ptrs and ownership
Github: https://github.com/xeouz/virec

Its on my github and there's a link to my web demo (https://vire-lang.web.app/), it might take some time to load the binary from firebase.

Very monolithic, ~7500 lines of code, I’d really appreciate any feedback, criticism, or pointers on how I could’ve done this better.

vire-lang.web.app
3 1
Summary
Show HN: Artifact Keeper – Open-Source Artifactory/Nexus Alternative in Rust
bsgeraci 1 day ago

Show HN: Artifact Keeper – Open-Source Artifactory/Nexus Alternative in Rust

I'm a software engineer who keeps getting pulled into DevOps no matter how hard I try to escape it. I recently moved into a Lead DevOps Engineer role writing tooling to automate a lot of the pain away. On my own time outside of work, I built Artifact Keeper — a self-hosted artifact registry that supports 45+ package formats. Security scanning, SSO, replication, WASM plugins — it's all in the MIT-licensed release. No enterprise tier. No feature gates. No surprise invoices.

Your package managers — pip, npm, docker, cargo, helm, go, all of them — talk directly to it using their native protocols. Security scanning with Trivy, Grype, and OpenSCAP is built in, with a policy engine that can quarantine bad artifacts before they hit your builds. And if you need a format it doesn't support yet, there's a WASM plugin system so you can add your own without forking the backend.

Why I built it:

Part of what pulled me into computers in the first place was open source. I grew up poor in New Orleans, and the only hardware I had access to in the early 2000s were some Compaq Pentium IIs my dad brought home after his work was tossing them out. I put Linux on them, and it ran circles around Windows 2000 and Millennium on that low-end hardware. That experience taught me that the best software is software that's open for everyone to see, use, and that actually runs well on whatever you've got.

Fast forward to today, and I see the same pattern everywhere: GitLab, JFrog, Harbor, and others ship a limited "community" edition and then hide the features teams actually need behind some paywall. I get it — paychecks have to come from somewhere. But I wanted to prove that a fully-featured artifact registry could exist as genuinely open-source software. Every feature. No exceptions.

The specific features came from real pain points. Artifactory's search is painfully slow — that's why I integrated Meilisearch. Security scanning that doesn't require a separate enterprise license was another big one. And I wanted replication that didn't need a central coordinator — so I built a peer mesh where any node can replicate to any other node. I haven't deployed this at work yet — right now I'm running it at home for my personal projects — but I'd love to see it tested at scale, and that's a big part of why I'm sharing it here.

The AI story (I'm going to be honest about this):

I built this in about three weeks using Claude Code. I know a lot of you will say this is probably vibe coding garbage — but if that's the case, it's an impressive pile of vibe coding garbage. Go look at the codebase. The backend is ~80% Rust with 429 unit tests, 33 PostgreSQL migrations, a layered architecture, and a full CI/CD pipeline with E2E tests, stress testing, and failure injection.

AI didn't make the design decisions for me. I still had to design the WASM plugin system, figure out how the scanning engines complement each other, and architect the mesh replication. Years of domain knowledge drove the design — AI just let me build it way faster. I'm floored at what these tools make possible for a tinkerer and security nerd like me.

Tech stack: Rust on Axum, PostgreSQL 16, Meilisearch, Trivy + Grype + OpenSCAP, Wasmtime WASM plugins (hot-reloadable), mesh replication with chunked transfers. Frontend is Next.js 15 plus native Swift (iOS/macOS) and Kotlin (Android) apps. OpenAPI 3.1 spec with auto-generated TypeScript and Rust SDKs.

Try it:

  git clone https://github.com/artifact-keeper/artifact-keeper.git
  cd artifact-keeper
  docker compose up -d
Then visit http://localhost:30080

Live demo: https://demo.artifactkeeper.com Docs: https://artifactkeeper.com/docs/

I'd love any feedback — what you think of the approach, what you'd want to see, what you hate about Artifactory or Nexus that you wish someone would just fix. It doesn't have to be a PR. Open an issue, start a discussion, or just tell me here.

https://github.com/artifact-keeper

github.com
155 64
Show HN: BioTradingArena – Benchmark for LLMs to predict biotech stock movements
dchu17 1 day ago

Show HN: BioTradingArena – Benchmark for LLMs to predict biotech stock movements

Hi HN,

My friend and I have been experimenting with using LLMs to reason about biotech stocks. Unlike many other sectors, Biotech trading is largely event-driven: FDA decisions, clinical trial readouts, safety updates, or changes in trial design can cause a stock to 3x in a single day (https://www.biotradingarena.com/cases/MDGL_2023-12-14_Resmet...).

Interpreting these ‘catalysts,’ which comes in the form of a press release, usually requires analysts with previous expertise in biology or medicine. A catalyst that sounds “positive” can still lead to a selloff if, for example: the effect size is weaker than expected

- results apply only to a narrow subgroup

- endpoints don’t meaningfully de-risk later phases,

- the readout doesn’t materially change approval odds.

To explore this, we built BioTradingArena, a benchmark for evaluating how well LLMs can interpret biotech catalysts and predict stock reactions. Given only the catalyst and the information available before the date of the press release (trial design, prior data, PubMed articles, and market expectations), the benchmark tests to see how accurate the model is at predicting the stock movement for when the catalyst is released.

The benchmark currently includes 317 historical catalysts. We also created subsets for specific indications (with the largest in Oncology) as different indications often have different patterns. We plan to add more catalysts to the public dataset over the next few weeks. The dataset spans companies of different sizes and creates an adjusted score, since large-cap biotech tends to exhibit much lower volatility than small and mid-cap names.

Each row of data includes:

- Real historical biotech catalysts (Phase 1–3 readouts, FDA actions, etc.) and pricing data from the day before, and the day of the catalyst

- Linked Clinical Trial data, and PubMed pdfs

Note, there are may exist some fairly obvious problems with our approach. First, many clinical trial press releases are likely already included in the LLMs’ pretraining data. While we try to reduce this by ‘de-identifying each press release’, and providing only the data available to the LLM up to the date of the catalyst, there are obviously some uncertainties about whether this is sufficient.

We’ve been using this benchmark to test prompting strategies and model families. Results so far are mixed but interesting as the most reliable approach we found was to use LLMs to quantify qualitative features and then a linear regression of these features, rather than direct price prediction.

Just wanted to share this with HN. I built a playground link for those of you who would like to play around with it in a sandbox. Would love to hear some ideas and hope people can play around with this!

biotradingarena.com
29 12
Show HN: Slack CLI for Agents
nwparker 2 days ago

Show HN: Slack CLI for Agents

Our team lives in Slack, but we don’t have access to the Slack MCP and couldn’t find anything out there that worked for us, so we coded our own agent-slack CLI

  * Can paste in Slack URLs
  * Token efficient
  * Zero-config (auto auth if you use Slack Desktop)
Auto downloads files/snippets. Also can read Slack canvases as markdown!

MIT License

github.com
53 12
Show HN: Nginx-defender – realtime abuse blocking for Nginx
anipaleja about 7 hours ago

Show HN: Nginx-defender – realtime abuse blocking for Nginx

I built nginx-defender after repeatedly seeing small and mid-sized NGINX servers get hammered by automated abuse (credential stuffing, path probing, aggressive scraping).

Existing tools like fail2ban or CrowdSec felt either too slow to react, too heavy for low resource servers, or painful to tune for modern traffic patterns.

nginx-defender runs inline with NGINX and blocks abusive IPs in real time based on request behavior rather than static rules. It’s designed to be lightweight, simple to deploy, and usable on small VPS setups.

I’ve been running it on my own servers and have seen thousands of abusive requests blocked within hours with minimal overhead.

Would love feedback from people running NGINX in production, especially on detection logic, false positives, or missing use cases.

github.com
3 0
Summary
Show HN: ARM64 Android Dev Kit
denuoweb 2 days ago

Show HN: ARM64 Android Dev Kit

GUI-first, multi-service gRPC scaffold for an Android Development Kit style workflow on an AArch64 system.

github.com
18 2
Summary
Show HN: Gigacode – Use OpenCode's UI with Claude Code/Codex/Amp
NathanFlurry 1 day ago

Show HN: Gigacode – Use OpenCode's UI with Claude Code/Codex/Amp

Gigacode is an experimental, just-for-fun project that makes OpenCode's TUI + web + SDK work with Claude Code, Codex, and Amp.

It's not a fork of OpenCode. Instead, it implements the OpenCode protocol and just runs `opencode attach` to the server that converts API calls to the underlying agents.

We build this to scratch our itch of being able to rapidly switch between coding agents based on the task at hand. For example, we find that:

- Claude Code is the best executor & fast iterator - Codex (high) is the best for complex or long-running tasks - OpenCode for fine-tuned, do-exactly-as-I-say edits

I personally believe that harnesses matter almost as much as the models in 2026. OpenCode lets you swap out models already, but the CC & Codex harnesses + system prompts make a big difference in practice.

Under the hood, this is all powered by our Sandbox Agent SDK:

- Sandbox Agent SDK provides a universal HTTP API for controlling Claude Code, Codex, and Amp - Sandbox Agent SDK exposes an OpenCode-compatible endpoint so OpenCode can talk to any agent - OpenCode connects to Sandbox Agent SDK via attach

I want to emphasize: the Anomaly folks are doing awesome work with OpenCode agent + Zen + Black. I use OC regularly alongside CC & Codex depending on the task. Gigacode is only possible because OpenCode is insanely flexible, hackable, and well documented.

Give it a try:

$ curl -fsSL https://releases.rivet.dev/sandbox-agent/latest/gigacode-ins... | sh

Check out the project, architecture, and other install options:

https://github.com/rivet-dev/sandbox-agent/tree/main/gigacod...

github.com
22 10
Show HN: MCP App to play backgammon with your LLM
sam256 about 9 hours ago

Show HN: MCP App to play backgammon with your LLM

I learned to play backgammon the same week the MCP App Extension standard came out. So...

You can now play backgammon with any AI client that supports MCP App (which includes Claude Desktop, VSCode, ChatGPT these days). Actually you can do more than that -- you can play against a friend and have they AI comment; you can watch the AI play itself (boring); you can play against the LLM and have it teach you the finer points of the game (my use case!).

The one problem--and this is a limitation of the current spec--is that the board gets redrawn each time the AI takes a turn. That's actually not that bad, but when the spec adds persistent/reusable views it will be even cooler.

github.com
3 1
Summary
Show HN: I'm 75, building an OSS Virtual Protest Protocol for digital activism
sakanakana00 about 10 hours ago

Show HN: I'm 75, building an OSS Virtual Protest Protocol for digital activism

Hi HN,

I’m a 75-year-old former fishmonger from Japan, currently working on compensation claims for victims of the Fukushima nuclear disaster. Witnessing social divisions and bureaucratic limitations firsthand, I realized we need a new way for people to express their will without being “disposable.”

To address this, I designed the Virtual Protest Protocol (VPP) ? an open-source framework for large-scale, 2D avatar-based digital demonstrations.

Key Features:

Beyond Yes/No: Adds an "Observe" option for the silent majority

Economic Sustainability: Funds global activism through U.S. commercial operations and avatar creator royalties

AI Moderation: LLMs maintain civil discourse in real-time

Privacy First: Minimal data retention ? only anonymous attributes, no personal IDs after the event

I shared this with the Open Technology Fund (OTF) and received positive feedback. Now, I’m looking for software engineers, designers, and OSS collaborators to help implement this as a robust project. I am not seeking personal gain; my goal is to leave this infrastructure for the next generation.

Links:

GitHub: https://github.com/voice-of-japan/Virtual-Protest-Protocol/b...

Project Site: https://voice-of-japan.net

Technical Notes:

Scalable 2D Rendering: 3?4 static frames per avatar, looped for movement

Cell-Based Grid System: Manages thousands of avatars efficiently, instantiates new cells as participation grows

Low Barrier to Entry: Accessible on low-spec smartphones and low-bandwidth environments

We are looking for collaborators with expertise in:

Backend/Real-Time Architecture: Node.js, Go, etc.

Frontend/Canvas Rendering: Handling thousands of avatars

AI Moderation / LLM Integration

OSS Governance & Project Management

If you’re interested, have technical advice, or want to join the build, please check the GitHub link and reach out. Your feedback and contribution can help make this infrastructure real and sustainable.

github.com
7 1
Summary
Show HN: I built Divvy to split restaurant bills from a photo
pieterdy about 10 hours ago

Show HN: I built Divvy to split restaurant bills from a photo

I built Divvy to make splitting restaurant bills less annoying. You take a photo of the bill, tap which items belong to each person, and it calculates totals including tax and tips.

divvyai.app
3 1
Summary
Show HN: Micropolis/SimCity Clone in Emacs Lisp
vkazanov 3 days ago

Show HN: Micropolis/SimCity Clone in Emacs Lisp

This is a little game implemented over a week of tinkering and targeting Emacs.

The point is both to have fun with this kind of simulations, and also explore the "functional core / imperative shell" approach to architecture. I also developed a tile and tile effect definition DSL, which makes this even easier to extend. From this point of view it's a success: easy testing, easy extension,

Gameplay-wise the simulation is too simplistic, and needs input from people interested in this kind of toys. The original Micropolis/SimSity is the last time I built a virtual city.

github.com
173 49
Summary
Show HN: Horizons – OSS agent execution engine
JoshPurtell 2 days ago

Show HN: Horizons – OSS agent execution engine

I'm Josh, founder of Synth. We've been working on coding agent optimization with method like GEPA and MIPRO (the latter of which, I helped to originally develop), agent evaluation via methods like RLMs, and large scale deployment for training and inference. We've also worked on patterns for memory, processing live context, and managing agent actions, combining it all in a single stack called Horizons. With the release of OpenAI's Frontier and the consumer excitement around OpenClaw, we think the timing is right to release a v0.

It integrates with our sdk for evaluation and optimization but also comes batteries-included with self-hosted implementations. We think Horizons will make building agent-based products a lot easier and help builders focus on their proprietary data, context, and algorithms

Some notes:

- you can configure claude code, codex, opencode to run in the engine. on-demand or on a cron

- we're striving to make it simple to integrate with existing backends via a 2-way event driven interface, but I'm 99.9% sure it'll change as there are a ton of unknown unknowns

- support for mcp, and we are building with authentication (rbac) in mind, although it's a long-journey

- all self-host able via docker

A very simplistic way to think about it - an OSS take on Frontier, or maybe OpenClaw for prod

github.com
27 5
Summary
p-s-v about 5 hours ago

Show HN: Which chef knife steels are good? Data from 540 Reddit tread

I wanted to see if "super-steel" marketing actually matched reality, so I scraped ~500 threads from r/chefknives and ran sentiment analysis on the specific steels mentioned.

The results:

MagnaCut: 28:1 positive ratio. The hype is real.

Ginsan: The sleeper favorite. High satisfaction because it almost never chips.

VG-10: Most controversial. High volume, but highest statistical ratio of "micro-chipping" complaints.

How I saved tokens (Inverse Masking): Feeding raw threads into an LLM just to find common terms like "Wüsthof" is a waste of money. I built a hybrid pipeline instead:

Fuzzy Match: Fuse.js catches 80% of common brands for $0.

Mask: Replace those entities with placeholders.

LLM: Feed the "masked" text to the model to catch obscure artisan makers and nuanced sentiment the fuzzy matcher missed.

Stack is Node.js + MongoDB. Full charts and the steel-by-steel breakdown are here: https://new.knife.day/blog/knife-steel-comparisons/all

Would love to hear thoughts on the methodology or if the data misses your favorite steel.

new.knife.day
2 0
Summary
Show HN: Daily-updated database of malicious browser extensions
toborrm9 1 day ago

Show HN: Daily-updated database of malicious browser extensions

Hey HN, I built an automated system that tracks malicious Chrome/Edge extensions daily.

The database updates automatically by monitoring chrome-stats for removed extensions and scanning security blogs. Currently tracking 1000+ known malicious extensions with extension IDs, names, and dates.

I'm working on detection tools (GUI + CLI) to scan locally installed extensions against this database, but wanted to share the raw data first since maintained threat intelligence lists like this are hard to find.

The automation runs 24/7 and pushes updates to GitHub. Free to use for research, integration into security tools, or whatever you need.

Happy to answer questions about the scraping approach or data collection methods.

github.com
14 8
melvinzammit about 12 hours ago

Show HN: I Hacked My Family's Meal Planning with an App

Me and my wife have been meal planning for the last 5 years. We used google keep. It was working for us, but during the years we needed to streamline the process. We tried other methods but nothing worked, so I spent the last 1 month hacking together this custom app. It includes all we needed to make our meal planning at least 5x more efficient. That is: syncing, one tap import of recipes, groceries, shopping mode, weekly meal plan, custom meals (like leftover, veg, eating out..)

We managed to do last Sunday's meal plan in under a minute, since all our favorite (100+) recipes are in one place. We also tagged them by daily food themes ( Monday-pasta, Tuesday- meat..). So we can quickly & mindlessly select a meal for each day.

For the app I used AI to classify groceries by aisle, but not generative, since I found that simple ML Models do a better job.

I would love any feedback from other hackers.

Feel free to use it. It's free, apart from syncing, which I had to add a subscription for due to server costs. I tried to make it generous: one subscription per 10 people.

mealjar.app
2 0
keepamovin about 15 hours ago

Show HN: Slop News – HN front page now, but it's all slop

The article discusses the emergence of 'Slop News,' a new type of news content that prioritizes speed and engagement over accuracy and depth. It explores the societal and ethical implications of this shift in the news industry.

dosaygo-studio.github.io
21 6
Summary
rahuljaguste about 24 hours ago

Show HN: Falcon's Eye (isometric NetHack) running in the browser via WebAssembly

The article discusses the Nethack Falcon's Eye, a graphical user interface (GUI) for the classic roguelike game Nethack. It highlights the Falcon's Eye's features, including improved graphics, streamlined gameplay, and enhanced functionality compared to the original terminal-based version of the game.

rahuljaguste.github.io
7 1
Summary
Show HN: I built a free UCP checker – see if AI agents can find your store
vladeta about 12 hours ago

Show HN: I built a free UCP checker – see if AI agents can find your store

The article discusses the launch of UCP Store Check, an AI-powered tool that helps retailers optimize their store operations by providing insights into customer behavior, inventory management, and staff performance.

ucphub.ai
2 2
Summary
Show HN: Local task classifier and dispatcher on RTX 3080
Shubham_Amb 2 days ago

Show HN: Local task classifier and dispatcher on RTX 3080

Hi HN, I am shubham a 3d artist who learned coding in college as an I.T. graduate know logics but not an expert as i just wanna try my hands on to ai

So i built Resilient Workflow Sentinel this is offline ai agent which classify urgency (Low,Medium and HIgh) and dispatches to the candidates based on availability Well i want an offline system like a person can trust with its sensitive data to stay completely locally

Did use ai to code for speeding and cutting labor.

Its works on RTX 3080 system (this is an basic affordable setup not heavy ai machinery) which i want it to make it reliable without heavy upgrade This is full system doesn't require ollama(I am not against it)

I see in companies tickets are raised on jira and slack. Currently people or manager (self) have to sort those things either manually read one by one or send them to the cloud. But the issue is you can't send everything like there is a lot of sensitive data out there which they do not trust and makes it harder and manual sorting through thousands is likely a nightmare.

But then just imagine u get all the task classified like its urgency and distribution u can selectively see which task is urgent and needs immediate attention and last of all information doesn't leave your building totally secure Also Api sending is not the only issue u are paying per token cost for task for each may be monthly 100$ to 1000$ which can like save hassle for startup a lot or companies as well

There was several biases like positional bias also json out put bias also have issues in attention At start i tried just prompting things like Chain of thoughts,RISE(evaluate negative first), given negative examples,Positive examples, somewhere it was struggling with commonsense issue so examples for that (Later changed the approach)

Well prompting did give the output and worked well but took too much time to process for single task like 70 to 90secs for a task

Then i tried batching and the biases got worst like it got stronger it always use to like favour alice also more prompts are like ignored and more

For json output i used constrain so model can only generate json and if fails there is a as well parser i used when i implemented prompting only

This reduce time from 90sec to nearly 15 to 30secs per task I used steering vector to correct the attention i seen issues happening

Stack: Language: Python 3.10 Model: qwen2.5-7b-instruct Libraries: Pytorch, Hugging Face Transformers (No Langchain, No Ollama) API: Fast API UI: NiceGUI Hardware: Ryzen 5, 16Gb ram RTX 3080

Implementation:

Quantization: Load model in nf4 quantization so models like 7b can fit on vram of 10gb which is on rtx 3080 also my hardware

Steering Vectors: Standard prompting wasn't enough. I need to block or direct certain things on a certain layer of llm to make it reliable.

Json Constraints: Used constraint to make model strictly give json and also stop from over explanation this happens at logits level where token are blocked which are not required etc

github : https://github.com/resilientworkflowsentinel/resilient-workf...

Youtube: https://youtu.be/tky3eURLzWo

github.com
25 2
Summary