Preset
Background
Text
Font
Size
Width
Account Monday, April 20, 2026

The Git Times

“Tools for conviviality are those which give each person who uses them the greatest opportunity to enrich the environment with the fruits of their vision.” — Ivan Illich

AI Models
Claude Sonnet 4.6 $15/M GPT-5.4 $15/M Gemini 3.1 Pro $12/M Grok 4.20 $6/M DeepSeek V3.2 $0.89/M Llama 4 Maverick $0.60/M
Full Markets →

OpenMythos Reconstructs Claude Mythos Architecture 🔗

Python library implements recurrent-depth transformers with looped blocks and sparse MoE

kyegomez/OpenMythos · Python · 2.4k stars 1d old

An independent developer has released OpenMythos, a Python package that reconstructs the Claude Mythos architecture from publicly available research literature. The project is unaffiliated with Anthropic and focuses on theoretical exploration rather than production deployment.

The core implementation is a Recurrent-Depth Transformer (RDT) organized in three stages.

An independent developer has released OpenMythos, a Python package that reconstructs the Claude Mythos architecture from publicly available research literature. The project is unaffiliated with Anthropic and focuses on theoretical exploration rather than production deployment.

The core implementation is a Recurrent-Depth Transformer (RDT) organized in three stages. A Prelude of standard transformer blocks processes initial input, followed by a looped Recurrent Block that iterates up to a configurable max_loop_iters value. A final Coda applies additional layers to produce output. Attention switches between MLA and GQA modes, while the feed-forward network uses a sparse Mixture of Experts with both routed and shared experts.

Configuration occurs through the MythosConfig class, which accepts parameters for vocabulary size, model dimension, head counts, expert counts, and LoRA ranks. Pre-built variants span 1B to 1T parameters. The library supports forward passes, token generation, and analysis of the recurrent injection matrix A, confirming its spectral radius stays below 1 for stability.

The project supplies concrete tools for examining compute-adaptive, depth-variable reasoning. Installation uses pip install open-mythos. Example code initializes the model with 256-dimensional embeddings and eight heads, runs inference with four or eight loops, and returns logits or generated sequences.

This work makes advanced architectural concepts accessible to the research community for local experimentation.

Use Cases
  • AI researchers testing recurrent-depth transformers with MoE layers
  • Engineers analyzing spectral radius stability in looped blocks
  • Developers experimenting with compute-adaptive reasoning models
Similar Projects
  • RWKV - uses linear recurrence instead of looped transformer blocks
  • Mixtral - applies MoE routing without recurrent depth stages
  • Megatron-LM - scales standard transformers but omits prelude-coda design

More Stories

CC Design Skill Equips AI With HTML Prototyping Tools 🔗

Context-first workflow and brand libraries raise quality of AI-generated interfaces

ZeroZ-lab/cc-design · JavaScript · 368 stars 1d old

ZeroZ-lab/cc-design supplies a JavaScript skill that turns Claude Code into an expert product designer. The system embeds a structured workflow covering requirement clarification, design-system research, component assembly and delivery of polished HTML artifacts.

Two operating principles shape its behavior.

ZeroZ-lab/cc-design supplies a JavaScript skill that turns Claude Code into an expert product designer. The system embeds a structured workflow covering requirement clarification, design-system research, component assembly and delivery of polished HTML artifacts.

Two operating principles shape its behavior. Context-first design instructs the agent to locate and reuse existing brand assets, component libraries or live product code before inventing new visuals. Progressive disclosure keeps the core skill compact while loading any of 12 supporting references on demand, controlling token usage.

Capabilities span multiple deliverables: interactive prototypes, slide decks, landing pages, UI mockups, wireframes, animated motion studies and full design systems. It progressively loads 68 brand libraries including Stripe, Vercel, Notion, Linear and Apple, supplies a catalog of proven layout patterns, and applies a quality framework that enforces hierarchy, emotional tone and anti-slop rules. The skill generates three or more variations across layout, interaction, visual intensity and motion dimensions.

Implementation uses inline React with pinned Babel, scoped component management and starter scaffolds. Playwright MCP delivers screenshot verification; local scripts handle clean export. Adapted from the Claude Artifacts environment, the package runs natively inside Claude Code.

For builders, the project matters because it replaces generic AI output with brand-consistent, verifiable interfaces that respect established design languages.

Use Cases
  • Engineers generating brand-aligned interactive UI components using Claude Code
  • Designers cloning corporate systems into new landing page mockups
  • Teams iterating on verified animated prototypes and motion studies
Similar Projects
  • vercel/v0 - generates React UIs from prompts but lacks mandatory brand-context steps
  • anthropic/artifacts - original isolated HTML canvas this project adapts for Claude Code
  • tldraw/make-real - converts sketches to code without the quality framework or 68-brand library

Go Proxy Links Freebuff Models to OpenAI Clients 🔗

Server translates requests with token rotation, dynamic fingerprints and Docker deployment for reliable access

Quorinex/Freebuff2API · Go · 350 stars 2d old

A Go application called Freebuff2API serves as an intermediary that converts standard OpenAI API calls into the format required by Freebuff's backend. The proxy enables any OpenAI-compatible client, SDK or CLI tool to consume Freebuff's free models without code changes.

The server implements three core technical capabilities.

A Go application called Freebuff2API serves as an intermediary that converts standard OpenAI API calls into the format required by Freebuff's backend. The proxy enables any OpenAI-compatible client, SDK or CLI tool to consume Freebuff's free models without code changes.

The server implements three core technical capabilities. It generates randomized client fingerprints that replicate official Freebuff SDK behavior to evade detection. Multiple authentication tokens can be supplied and are automatically cycled on a configurable schedule to improve throughput. All outbound requests can be routed through an upstream HTTP proxy when needed.

Users obtain required auth tokens by visiting the Freebuff web interface after login or by extracting the authToken value from the JSON credentials file created by the official CLI. Configuration is handled through a config.json file or equivalent environment variables, keeping setup straightforward.

Ready-made Docker images allow deployment in a single command on local machines or cloud servers. By bridging the two ecosystems, the proxy lowers the cost of building and testing AI applications while preserving compatibility with the broader OpenAI tooling landscape.

Freebuff2API was created on April 18, 2026 and remains under active development.

Use Cases
  • Engineers integrating free models into existing OpenAI SDK applications
  • Teams cycling multiple accounts to scale request volume safely
  • DevOps staff deploying containerized proxies behind corporate firewalls
Similar Projects
  • litellm - broader multi-provider support but lacks Freebuff-specific fingerprinting
  • openai-forward - simpler Python proxy without automatic token rotation
  • vllm - serves local models with OpenAI API but requires self-hosting

New Collection Curates Prompts for GPT-Image-2 🔗

Repository gathers community examples across portraits, posters, UI mockups and character sheets

EvoLinkAI/awesome-gpt-image-2-prompts · Unknown · 694 stars 1d old

The EvoLinkAI/awesome-gpt-image-2-prompts repository collects prompts for the GPT-Image-2 image generator available on Evolink. It pairs each prompt with its output image. Sources include X/Twitter, creator communities and public demos.

The EvoLinkAI/awesome-gpt-image-2-prompts repository collects prompts for the GPT-Image-2 image generator available on Evolink. It pairs each prompt with its output image. Sources include X/Twitter, creator communities and public demos.

The collection organizes material into distinct sections. Portrait and photography cases feature convenience store neon portraits, cinematic minimal compositions, Japanese onsen ryokan scenes, 35mm flash editorials and mirror selfie bedroom shots.

Nine poster and illustration examples range from Boston Spring city posters to vintage Amalfi travel designs, Chengdu food maps, minimalist S-shaped posters and futuristic mandala illustrations.

Character design cases include anime snapshot conversions, Persona 5 character reference cards, gal game introduction pages and chibi character reference sheets.

First released on April 18, 2026, the repository received 10 new prompts the following day. These additions covered additional poster, UI and comparison cases.

The project focuses on reusable prompt patterns rather than single uses. Users can reference these concrete cases when generating portraits, posters, UI mockups or character sheets with the tool.

Such structured examples help identify which descriptive elements influence specific aspects of the output image.

Use Cases
  • Digital artists generate consistent anime character sheets
  • Designers produce UI mockups using optimized prompts
  • Marketers create city posters from reference examples
Similar Projects
  • lexica-art - searchable database rather than curated categories
  • awesome-midjourney - targets competing commercial image model
  • openai-cookbook - includes DALL-E prompts without Evolink focus

Create Mod Gains Physics Contraptions for Vehicles 🔗

Simulated Project offers assembly tools, redstone components and specialized flying modules

Creators-of-Aeronautics/Simulated-Project · Java · 350 stars 5d old

The Simulated Project delivers a suite of NeoForge mods that extend the Create mechanical framework with physics simulation. Written in Java, the tools allow players to construct moving contraptions that obey consistent rules for thrust, lift, collision and gravity.

Simulated forms the foundation.

The Simulated Project delivers a suite of NeoForge mods that extend the Create mechanical framework with physics simulation. Written in Java, the tools allow players to construct moving contraptions that obey consistent rules for thrust, lift, collision and gravity.

Simulated forms the foundation. It supplies assembly interfaces, redstone components and interaction utilities required to manipulate physics objects. Aeronautics adds flight systems built around hot-air lift, propeller thrust and buoyant magic floating rocks. Offroad converts almost any wheel-shaped block into functional drive components for land vehicles.

These modules operate as a single cohesive system rather than isolated features. Builders can therefore combine stationary Create machinery with dynamic vehicles without relying on command blocks or crude pistons. The result is reliable behavior for everything from lightweight scout planes to heavy cargo airships and multi-legged experimental walkers.

The project was created on 15 April 2026 and received further updates by 20 April. Its focused development improves simulation stability and redstone responsiveness, giving technical players a precise toolkit for engineering moving contraptions inside Minecraft.

Use Cases
  • Redstone engineers integrating controls into physics-based aircraft
  • Vehicle builders assembling cars from arbitrary wheel-shaped blocks
  • Technical players constructing hot-air airships with propeller arrays
Similar Projects
  • Valkyrien Skies - offers independent physics simulation without Create's mechanical integration
  • Immersive Aircraft - supplies pre-built planes with far less modular assembly
  • Eureka Balloons - focuses on simple buoyant airships lacking redstone and wheel systems

Repository Delivers Design Blueprints for Claude Workspace 🔗

Collection of 68 DESIGN.md files generates complete UI systems including tokens, components and kits in one step

VoltAgent/awesome-claude-design · Unknown · 536 stars 2d old

VoltAgent/awesome-claude-design supplies 68 ready-to-use DESIGN.md files for Anthropic's Claude Design workspace. Drop any file into the tool and it immediately scaffolds color tokens, typography scales, buttons, cards, navigation and a working UI kit.

VoltAgent/awesome-claude-design supplies 68 ready-to-use DESIGN.md files for Anthropic's Claude Design workspace. Drop any file into the tool and it immediately scaffolds color tokens, typography scales, buttons, cards, navigation and a working UI kit. All assets land in a persistent Design System review tab, keeping every subsequent screen on-brand.

Claude Design maintains a single source of truth for a project's visual rules rather than generating disconnected mockups in chat. The DESIGN.md format captures both concrete specifications and the reasoning behind them, allowing the AI to make consistent decisions on novel elements. This sits between rigid Figma libraries that omit rationale and loose brand PDFs that offer no machine-readable guidance.

The collection covers multiple aesthetics so teams can match the exact feel required, then iterate inside the organized workspace. It pairs naturally with AGENTS.md files that instruct coding agents, creating aligned briefs for full-stack AI development.

By removing the blank-page phase of design-system creation, the project lets engineers and designers move directly to implementation while preserving coherence. The markdown files remain editable, version-controlled documents that travel with the codebase.

**

Use Cases
  • Engineers generating persistent UI kits from markdown files
  • Teams establishing brand-consistent tokens in Claude Design
  • Developers scaffolding interface systems without manual setup
Similar Projects
  • getdesign.md - originated the comparable collection of DESIGN.md files
  • Google Stitch - first introduced the DESIGN.md format for agents
  • Figma AI - produces isolated screens instead of persistent systems

Open Source Crafts Modular Skills Ecosystem for AI Agents 🔗

Community-created design systems, domain expertise packs, and memory tools are transforming LLMs into autonomous, context-aware engineering collaborators.

An unmistakable pattern is crystallizing across open source: the systematic construction of modular, composable skills and supporting infrastructure that turns generic large language models into specialized, autonomous AI agents. Rather than treating agents as simple prompt responders, developers are packaging reusable capabilities—design systems, domain knowledge, memory layers, and orchestration primitives—that allow agents to operate as genuine teammates in software creation, research, and operations.

Evidence appears in the surge of skill repositories.

An unmistakable pattern is crystallizing across open source: the systematic construction of modular, composable skills and supporting infrastructure that turns generic large language models into specialized, autonomous AI agents. Rather than treating agents as simple prompt responders, developers are packaging reusable capabilities—design systems, domain knowledge, memory layers, and orchestration primitives—that allow agents to operate as genuine teammates in software creation, research, and operations.

Evidence appears in the surge of skill repositories. alirezarezvani/claude-skills ships more than 232 plugins spanning engineering, marketing, compliance, and executive advisory, while addyosmani/agent-skills and coreyhaines31/marketingskills deliver production-grade tools for coding, CRO, SEO, and growth engineering. Cybersecurity receives similar treatment in mukul975/Anthropic-Cybersecurity-Skills, which maps 754 structured capabilities to MITRE ATT&CK, NIST CSF, and other standards. These collections share a common technical philosophy: skills are versioned, discoverable, and portable across Claude Code, Cursor, Gemini CLI, and OpenAI-compatible endpoints.

Design and context layers are evolving in parallel. VoltAgent/awesome-claude-design and VoltAgent/awesome-design-md curate ready-to-use DESIGN.md files that let agents scaffold pixel-perfect UIs from established systems in one shot. ZeroZ-lab/cc-design supplies high-fidelity HTML prototypes, and mksglu/context-mode demonstrates aggressive context optimization—sandboxing tool output for a reported 98% reduction in token usage across 12 platforms.

Memory, autonomy, and collaboration complete the picture. thedotmack/claude-mem automatically captures, compresses via Claude’s own SDK, and reinjects session history. multica-ai/multica lets teams assign GitHub issues to agents that self-report blockers and update statuses like human colleagues. openai/openai-agents-python and obra/superpowers supply lightweight multi-agent orchestration, while HKUDS/CLI-Anything and Tracer-Cloud/opensre push the “agent-native” boundary by making existing CLIs and SRE workflows consumable by autonomous loops.

Collectively these projects reveal where open source is heading: from model hosting toward a rich agent operating system composed of standardized skill interfaces, persistent memory architectures, cross-platform hooks, and shared design corpora. The ecosystem mirrors npm’s package model but for agent behaviors—developers will increasingly compose rather than hand-craft intelligence, accelerating a shift to agentic engineering where humans orchestrate fleets of specialized, self-improving collaborators. This modular approach also lowers the barrier for domain experts to encode their knowledge (kepano/obsidian-skills, mvanhorn/last30days-skill), suggesting AI agents will soon possess deep, updatable expertise across every vertical.

The pattern is still early, yet its technical direction is clear: open source is building the middleware layer that will make truly autonomous software agents not science fiction, but daily infrastructure.

Use Cases
  • Engineers injecting domain skills into coding agents
  • Teams assigning GitHub issues to autonomous AI teammates
  • Developers scaffolding UIs via standardized design systems
Similar Projects
  • LangChain - Delivers composable agent tools and memory that align with the modular skills pattern but focuses more on Python orchestration than domain-specific packs
  • CrewAI - Specializes in role-based multi-agent teams similar to multica-ai/multica yet emphasizes conversational coordination over CLI and design.md integration
  • AutoGen - Enables dynamic multi-agent conversations like openai-agents-python while offering less emphasis on portable skill libraries and context optimization techniques

Agent-Native Web Frameworks Emerge to Elevate AI Coding Quality 🔗

Open source projects are building specialized starters, design systems, and LLM proxies that help AI agents produce sophisticated, non-generic web applications.

Open source is shifting from traditional web frameworks toward agent-native ecosystems that treat AI coding agents as first-class users. Rather than simply offering libraries for human developers, this new generation of projects focuses on reducing ambiguity, encoding design taste, and providing universal LLM backends so autonomous agents can generate production-grade frontend and full-stack code with minimal human correction.

Evidence appears across the cluster.

Open source is shifting from traditional web frameworks toward agent-native ecosystems that treat AI coding agents as first-class users. Rather than simply offering libraries for human developers, this new generation of projects focuses on reducing ambiguity, encoding design taste, and providing universal LLM backends so autonomous agents can generate production-grade frontend and full-stack code with minimal human correction.

Evidence appears across the cluster. Leonxlnx/taste-skill explicitly targets the "high-agency frontend" problem, training agents to avoid the generic "slop" that large language models typically produce when left unsupervised. Complementing this, VoltAgent/awesome-design-md supplies ready-to-use DESIGN.md files distilled from popular websites, giving agents concrete, parseable specifications they can follow to replicate sophisticated UI systems.

Starter templates have evolved accordingly. jpedroschmitz/typescript-nextjs-starter delivers a deliberately non-opinionated yet fully equipped foundation for Next.js 16, supplying every modern tool an agent might need without forcing architectural decisions. On the backend, nhost/nhost offers a GraphQL-first Firebase alternative that agents can scaffold quickly, while maplibre/maplibre-gl-js provides battle-tested vector mapping components ready for browser-based interactive applications.

Powering these agents is an explosion of compatibility layers. Projects like Quorinex/Freebuff2API, router-for-me/CLIProxyAPI, Wei-Shaw/sub2api, QuantumNous/new-api, and mnfst/awesome-free-llm-apis create unified gateways that normalize access to Gemini, Claude, OpenAI, and 200+ models. Gitlawb/openclaude and badlogic/pi-mono further integrate these backends into purpose-built coding-agent CLIs, TUI interfaces, and web UI libraries. Even web-platform-tests/wpt and WebKit/WebKit contribute by strengthening the underlying platform guarantees that agents can rely upon.

This cluster signals a deeper technical transition: web frameworks are gaining formalisms (design manifests, token-rotating proxies, standardized skill interfaces) that make generative models dramatically more effective. The pattern suggests open source is moving toward symbiotic human-AI development environments where the framework itself becomes an active participant in guiding agent behavior, accelerating the creation of complex web experiences while raising their baseline quality.

The result is an infrastructure layer purpose-built for the coming wave of autonomous coding agents that will increasingly own large portions of web application development.

Use Cases
  • AI agents generating tasteful frontend code from design specs
  • Developers routing multiple LLMs through unified web API gateways
  • Teams scaffolding full-stack GraphQL apps with Next.js starters
Similar Projects
  • Vercel AI SDK - Provides React-focused AI streaming components but lacks the broad LLM proxy and design-system focus of this cluster.
  • LangChain.js - Offers agent orchestration tools yet does not emphasize frontend taste enforcement or Next.js-specific starter kits.
  • Supabase - Delivers an open Firebase alternative with strong realtime features but without the agent-native DESIGN.md and CLI proxy integrations.

Open Source Builds Modular LLM Tooling Layer for Agents 🔗

Proxies, skill libraries, and interoperability frameworks are turning proprietary models into customizable, cost-efficient development infrastructure.

An unmistakable pattern is emerging in open source: the rapid construction of a middleware layer for large language models that emphasizes compatibility, efficiency, and extensibility. Rather than training new foundation models, developers are focusing on tools that wrap, optimize, augment, and orchestrate existing ones—particularly coding agents and CLI interfaces.

This cluster reveals three technical pillars.

An unmistakable pattern is emerging in open source: the rapid construction of a middleware layer for large language models that emphasizes compatibility, efficiency, and extensibility. Rather than training new foundation models, developers are focusing on tools that wrap, optimize, augment, and orchestrate existing ones—particularly coding agents and CLI interfaces.

This cluster reveals three technical pillars. First, API compatibility proxies are proliferating. router-for-me/CLIProxyAPI, Quorinex/Freebuff2API, Wei-Shaw/sub2api, and QuantumNous/new-api convert free-tier access to Gemini, Claude, and other models into standard OpenAI-compatible endpoints. These projects use token rotation, dynamic routing, and subscription sharing to slash costs while maintaining seamless integration with existing toolchains.

Second, specialized skill and plugin ecosystems are maturing. alirezarezvani/claude-skills and mukul975/Anthropic-Cybersecurity-Skills package hundreds of structured capabilities—ranging from engineering workflows to 754 MITRE ATT&CK-mapped cybersecurity techniques—designed to work across Claude Code, Cursor, Gemini CLI, and Codex. hesreallyhim/awesome-claude-code and EvoLinkAI/awesome-gpt-image-2-prompts further demonstrate how curated prompts, slash commands, and agent hooks are becoming first-class artifacts.

Third, agent frameworks and efficiency layers are closing the stack. openai/openai-agents-python delivers lightweight multi-agent orchestration, badlogic/pi-mono combines CLI, TUI, web UI, and Slack bots into one toolkit, while rtk-ai/rtk achieves 60-90% token reduction on common dev commands through a single Rust binary. On the reasoning side, verl-project/verl, infiniflow/ragflow, and educational notebooks from unslothai and LLMForEverybody show reinforcement learning, RAG-agent hybrids, and fine-tuning techniques being productized for practical deployment.

Collectively, these repositories signal where open source is heading: toward a composable AI operating system. By decoupling model access from model ownership, the community is creating plug-and-play components that let developers swap providers, inject domain expertise via plugins, and orchestrate agents without reinventing infrastructure. The focus has shifted from raw capability to leverage—making frontier models faster, cheaper, and dramatically more useful inside real software workflows.

This pattern suggests the next wave of innovation will live in the glue between models, not the models themselves.

Use Cases
  • Developers routing CLI commands through cost-optimized LLM proxies
  • Security teams equipping agents with structured cybersecurity skills
  • Engineers orchestrating multi-agent workflows across model providers
Similar Projects
  • LangChain - Delivers broader orchestration abstractions while this cluster emphasizes lightweight CLI proxies and domain skill packs
  • LlamaIndex - Focuses primarily on data indexing for RAG unlike the agent plugin and API compatibility emphasis here
  • AutoGen - Specializes in conversational multi-agent patterns but lacks the token-optimization proxies and cybersecurity skill libraries seen in this trend

Deep Cuts

Unlocking WeChat Favorites with Interactive HTML Reports 🔗

End-to-end Python pipeline decrypts databases and builds engaging data visualizations

zhuyansen/wx-favorites-report · Python · 497 stars

While hunting for overlooked tools that solve real personal-data problems, I discovered wx-favorites-report, a remarkably complete Python project that turns WeChat’s encrypted favorites database into a polished, interactive HTML dashboard.

The tool handles the entire journey: it decrypts WeChat’s proprietary storage format, parses years of saved articles, links, images, and notes, then generates a single-file HTML report packed with timelines, tag clouds, full-text search, reading-pattern analytics, and clickable cards. What feels like magic is the seamless engineering—careful reverse-engineering wrapped in clean, maintainable Python that anyone can extend.

While hunting for overlooked tools that solve real personal-data problems, I discovered wx-favorites-report, a remarkably complete Python project that turns WeChat’s encrypted favorites database into a polished, interactive HTML dashboard.

The tool handles the entire journey: it decrypts WeChat’s proprietary storage format, parses years of saved articles, links, images, and notes, then generates a single-file HTML report packed with timelines, tag clouds, full-text search, reading-pattern analytics, and clickable cards. What feels like magic is the seamless engineering—careful reverse-engineering wrapped in clean, maintainable Python that anyone can extend.

Builders should pay attention because this project demonstrates production-grade techniques for closed-platform data liberation. It shows how to move beyond simple export scripts into rich, browser-based knowledge interfaces. The architecture serves as a template for anyone building personal data vaults, research archives, or habit-visualization tools. Once you see your scattered WeChat saves transformed into an explorable knowledge map, it becomes obvious how many similar “walled garden” datasets are waiting for the same thoughtful treatment.

In short, wx-favorites-report is less a utility and more a blueprint for turning forgotten app data into living insight.

Use Cases
  • Researchers mapping long-term WeChat reading patterns and interests
  • Developers prototyping encrypted database decryption and visualization pipelines
  • Power users creating searchable interactive reports of saved WeChat content
Similar Projects
  • wx-dump - raw JSON exporter without interactive HTML layer
  • telegram-favorites-viz - similar concept but lacks WeChat decryption skills
  • personal-knowledge-dashboard - generic visualizer missing the end-to-end encrypted pipeline

Quick Hits

verl Verl delivers a scalable RL framework to efficiently train and fine-tune LLMs with Volcano Engine optimizations. 20.8k
permiso Permiso adds Codex-style permission dialogs for macOS accessibility settings directly into your Swift apps. 346
spring-ai Spring AI gives Java developers a full framework to build, integrate, and deploy production AI applications. 8.5k
tinygrad Tinygrad provides a minimalist, hackable tensor library for building neural nets with zero bloat. 32.5k
chains Chains supplies standardized metadata for EVM networks to simplify blockchain integration and configuration. 9.8k
core :house_with_garden: Open source home automation that puts local control and privacy first. 86.1k

FaceSwap v3.0 Simplifies Neural Network Face Swaps 🔗

Automated installer configures PyTorch, CUDA and ROCm environments for immediate GUI access.

deepfakes/faceswap · Python · 55.2k stars Est. 2017 · Latest: v3.0.0

FaceSwap has released version 3.0.0, centered on an installer that removes the previous manual configuration burden for its deep-learning face swapping pipeline.

FaceSwap has released version 3.0.0, centered on an installer that removes the previous manual configuration burden for its deep-learning face swapping pipeline. The faceswap_setup_x64.exe automatically installs Git, MiniConda and PyTorch, creates a Conda environment, and places a desktop shortcut that launches straight into the GUI.

Nvidia users receive a local CUDA 11.8+ and cuDNN stack provided the graphics card meets the requirement. AMD users must pre-install ROCm 6.0-6.4; on Windows, AMD operation routes through WSL2. The project explicitly notes that only Nvidia and CPU backends are natively supported under Windows without the Linux subsystem.

The core workflow remains extract, train and convert. Faces are detected and aligned from source images or video, a neural network is trained on paired data using models such as Phaze-A or Villain, then the trained model generates the swapped output. Examples circulating in the community include high-fidelity swaps of well-known actors that demonstrate the models' ability to preserve lighting, skin tone and expression.

Now more than eight years old, the project continues to function as an on-ramp for builders who want to experiment with generative neural networks without advanced mathematics degrees. Its maintainers maintain a manifesto that distinguishes legitimate applications in visual effects, historical reconstruction and academic research from misuse. The active forum and Discord remain primary support channels, where users share training tips and artifact-reduction techniques.

The v3.0 installer reflects sustained maintenance rather than reinvention, lowering the hardware and software friction that once limited participation while preserving the same Python-based, locally runnable architecture that distinguished the project at its 2017 debut.

Use Cases
  • VFX artists swap faces in independent film post-production
  • Researchers train custom models for facial recognition testing
  • Educators demonstrate generative AI ethics through controlled swaps
Similar Projects
  • DeepFaceLab - offers more video-centric training controls
  • Roop - provides one-click inference without model training
  • SimSwap - focuses on faster real-time swapping algorithms

More Stories

TensorFlow 2.21 Sharpens Edge AI Quantization 🔗

Release drops Python 3.9 and TensorBoard while adding int2 and int4 support

tensorflow/tensorflow · C++ · 194.8k stars Est. 2015

TensorFlow 2.21.0 introduces targeted changes that prioritize efficient on-device inference over backward compatibility.

TensorFlow 2.21.0 introduces targeted changes that prioritize efficient on-device inference over backward compatibility. The release removes Python 3.9 support and eliminates the TensorBoard dependency from the core package, forcing teams to modernize their environments and decouple visualization tools.

Improvements concentrate on tf.lite. The update adds int8 and int16x8 implementations for the SQRT operator, extends int16x8 coverage to EQUAL and NOT_EQUAL, and introduces native int2 and uint4 types. New capabilities include int2/int4 casting, SRQ int2 in fully connected layers, and int4 slicing. These additions enable smaller model footprints and faster execution on phones, microcontrollers, and embedded accelerators where every byte matters.

tf.image gains JPEG XL decoding, allowing direct ingestion of the high-compression format now used by major browsers and content delivery networks. In tf.data, NoneTensorSpec becomes part of the public API so developers can reliably test for optional tensors in pipelines.

More than fifty contributors, including Google engineers and independent developers, delivered these updates. Installation remains straightforward—pip install tensorflow for GPU or pip install tensorflow-cpu for lightweight builds—yet existing codebases using deprecated Python or implicit TensorBoard hooks will require immediate adjustments.

The changes reflect TensorFlow’s long-term shift toward production efficiency as quantization moves from research curiosity to standard practice.

Use Cases
  • Mobile engineers quantizing neural networks for embedded systems
  • Vision teams integrating JPEG XL decoding into production pipelines
  • ML practitioners migrating code after Python 3.9 deprecation
Similar Projects
  • PyTorch - offers dynamic computation graphs instead of static optimization
  • JAX - emphasizes functional transformations for research-scale autodiff
  • ONNX Runtime - focuses on cross-framework inference without full training

Netdata v2.10.2 Hardens Real-Time Observability 🔗

Patch release eliminates ZFS crashes, SNMP instability and false configuration warnings

netdata/netdata · C · 78.5k stars Est. 2013

Netdata has shipped v2.10.2, a targeted maintenance release that fixes three classes of operational friction for users running its real-time monitoring agent.

Netdata has shipped v2.10.2, a targeted maintenance release that fixes three classes of operational friction for users running its real-time monitoring agent.

The update guards the diskspace.plugin against NULL filesystem pointers on ZFS, swapping a blocking pool-capacity collector for a lightweight cache fed by existing statvfs calls. The change prevents coredumps on degraded or exporting pools. SNMP polling becomes more reliable by lowering the default MaxOIDs from 60 to 20 and removing 32-bit counter fallbacks that caused type switching between collection cycles. Finally, the dynamic configuration subsystem no longer emits false-positive “timed out waiting for enable/disable decision” warnings; back-pressure now propagates upstream instead of surfacing 503 errors.

These fixes arrive as infrastructure operators demand per-second visibility without adding overhead. Written in C, Netdata collects thousands of metrics per second from Linux kernels, cgroups, Docker, Kubernetes, MySQL, PostgreSQL, MongoDB and Prometheus endpoints. Its built-in machine-learning models flag anomalies and forecast capacity while the agent itself rarely exceeds 3 % CPU and a few hundred megabytes of RAM. Data stays local by default, eliminating the need for central collectors.

For lean teams already running the agent, v2.10.2 simply removes reasons to restart it. The project, now over a decade old, continues to prove that high-resolution observability and low resource usage are not mutually exclusive.

Use Cases
  • SREs monitoring Kubernetes clusters at one-second resolution
  • DevOps teams detecting Linux anomalies with zero configuration
  • DBAs troubleshooting MySQL performance with per-second metrics
Similar Projects
  • Prometheus - focuses on long-term storage but lacks Netdata's real-time visuals and ML
  • Grafana - provides dashboards while Netdata supplies the high-frequency data and alerting
  • Zabbix - delivers enterprise monitoring yet consumes more resources than Netdata's lightweight agent

Quick Hits

transformers Transformers equips builders to train and deploy SOTA models across text, vision, audio, and multimodal tasks in one framework. 159.6k
scikit-learn Scikit-learn delivers battle-tested Python tools for classification, regression, clustering, and the full classical ML workflow. 65.9k
tesseract Tesseract extracts accurate text from images and scans across languages, powering production OCR applications. 73.6k
dify Dify provides a production-ready platform to build, orchestrate, and ship complex agentic AI workflows. 138.4k
supervision Supervision gives you reusable computer vision tools that instantly accelerate detection, tracking, and annotation pipelines. 38.2k
ragflow RAGFlow is a leading open-source Retrieval-Augmented Generation (RAG) engine that fuses cutting-edge RAG with Agent capabilities to create a superior context layer for LLMs 78.5k

Stable Baselines3 v2.8.0 Updates RL Toolbox for Modern Python 🔗

Latest release drops Python 3.9 support, adds 3.13 compatibility and fixes Torch compilation and PPO edge cases for reliable reinforcement learning research

DLR-RM/stable-baselines3 · Python · 13.1k stars Est. 2020 · Latest: v2.8.0

Stable Baselines3 remains the standard PyTorch implementation of reliable reinforcement learning algorithms six years after its creation. Version 2.8.

Stable Baselines3 remains the standard PyTorch implementation of reliable reinforcement learning algorithms six years after its creation. Version 2.8.0, released this week, modernizes the library for current Python runtimes while correcting several issues that affected production and research workflows.

The library delivers tested implementations of core RL methods including PPO, SAC, TD3, and A2C. It builds directly on the original Stable Baselines project, shifting the codebase to PyTorch and establishing a consistent interface that supports custom environments, custom policies, and Dict observation spaces. Every algorithm ships with documented performance benchmarks, Tensorboard logging, and IPython-friendly APIs. The maintainers emphasize that users need working knowledge of reinforcement learning; the library is not a teaching tool but a production-grade foundation for extending ideas or comparing new approaches.

The v2.8.0 release contains three categories of change. Breaking updates remove Python 3.9 support—developers must now target Python 3.10 or higher—and enforce strict=True on every zip() call. The extras installation now uses pygame-ce instead of the deprecated pygame package. On the feature side, the library now officially supports Python 3.13, ensuring compatibility with the latest language improvements and standard library changes.

Bug fixes address real-world pain points. Saving and loading of Torch compiled models (th.compile()) now works correctly after updates to get_parameters(). The environment checker issues a warning for multidiscrete spaces containing multi-dimensional arrays. Empty dataframes are filtered before pandas.concat operations, eliminating future-warning spam that cluttered logs in automated training pipelines.

SB3-Contrib received parallel treatment. MaskablePPO and RecurrentPPO now count n_updates accurately when target_kl triggers early exit. Action reshaping bugs in forward() and predict() methods were corrected, and a numerical stability issue in MaskableCategorical.apply_masking() was fixed for large action spaces under Torch 2.9+. Both the main library and the RL Zoo have migrated their documentation to Markdown using the MyST parser, improving maintainability.

These changes matter now because reinforcement learning workloads increasingly run on updated Python stacks and compiled models. Robotics teams, autonomous systems developers, and research labs depend on SB3 as a stable baseline; small numerical or compatibility bugs can invalidate weeks of training runs. By pruning legacy support and fixing edge cases, the maintainers keep the toolbox lean and trustworthy.

The project’s emphasis on high code coverage, type hints, PEP8 compliance, and reproducible benchmarks has made it the default starting point for anyone building on Gymnasium or custom robotic simulators. Version 2.8.0 ensures that foundation remains solid as the broader Python and PyTorch ecosystems continue to evolve.

Use Cases
  • Robotics engineers training manipulation policies
  • Researchers replicating RL benchmark experiments
  • Developers prototyping agents in custom Gym environments
Similar Projects
  • CleanRL - Single-file implementations that prioritize research reproducibility over SB3's modular toolbox approach
  • Ray RLlib - Distributed training framework offering greater scale at the expense of SB3's simplicity and tight benchmarking
  • Tianshou - Modular PyTorch RL library focused on vectorized environments rather than SB3's emphasis on algorithm reliability

More Stories

OpenArm 1.1 Refines Teleoperation for Contact Tasks 🔗

Automated calibration and modular mounts address community feedback in 7DOF humanoid arm

enactic/openarm · MDX · 2.2k stars Est. 2024

OpenArm has shipped version 1.1 of its open-source 7DOF humanoid arm, the second release of the OpenArm 01 platform. The update focuses on hardware reliability, teleoperation performance, assembly clarity, and vendor transparency after 18 months of community use in physical AI labs.

OpenArm has shipped version 1.1 of its open-source 7DOF humanoid arm, the second release of the OpenArm 01 platform. The update focuses on hardware reliability, teleoperation performance, assembly clarity, and vendor transparency after 18 months of community use in physical AI labs.

Key changes tackle previously reported pain points. An automated zero-position calibration routine replaces error-prone manual alignment, ensuring repeatable joint referencing across sessions. A new modular camera mount provides verified CAD for Realsense D435 chest cameras and D405 wrist cameras, standardizing data collection pipelines that previously required custom fabrication.

On the leader arm, engineers redesigned the J5 casing and added a rubber-band coupling interface. This enables natural elbow tracking during bilateral teleoperation, eliminating earlier ambiguity caused by redundant joint configurations in 7-DoF motion.

The arm retains its defining characteristics: human-scale proportions, high backdrivability, and sufficient payload for contact-rich manipulation while remaining compliant for safe human interaction. A complete bimanual system costs $6,500. Multiple repositories deliver hardware files under CERN-OHL-S-2.0, URDF descriptions, CAN motor control, ROS2 nodes, and simulation support for MuJoCo and Genesis.

The release demonstrates steady iteration on an accessible platform for imitation learning, reinforcement learning, and force-feedback research rather than radical redesign.

Use Cases
  • Researchers collecting reproducible teleoperation datasets for imitation learning
  • Engineers deploying reinforcement learning policies in compliant contact tasks
  • Developers integrating ROS2 force-feedback with MuJoCo simulation transfer
Similar Projects
  • LeRobot - software-focused imitation learning stack without matching physical hardware
  • OpenManipulator-X - lower-cost 6DOF arm lacking human-scale proportions and compliance
  • Shadow Dexterous Hand - higher-dexterity gripper at greater cost with narrower research scope

Kornia Shifts Toward End-to-End Vision Models 🔗

Differentiable PyTorch library integrates VLMs and VLAs for spatial AI pipelines

kornia/kornia · Python · 11.2k stars Est. 2018

Kornia has announced a strategic shift toward end-to-end vision solutions, prioritizing integration of state-of-the-art Vision Language Models and Vision Language Agents. The move extends the project's original mandate as a differentiable computer vision library built on PyTorch.

Since its creation in 2018, Kornia has supplied over 500 differentiable operators that slot directly into neural network training.

Kornia has announced a strategic shift toward end-to-end vision solutions, prioritizing integration of state-of-the-art Vision Language Models and Vision Language Agents. The move extends the project's original mandate as a differentiable computer vision library built on PyTorch.

Since its creation in 2018, Kornia has supplied over 500 differentiable operators that slot directly into neural network training. These include Gaussian, Sobel and Median filters, affine and homography transformations, histogram equalization, CLAHE, and edge detectors such as Canny and Laplacian. All operations support batch processing, automatic differentiation and GPU acceleration.

The library's augmentation tools—AugmentationSequential, PatchSequential, RandAugment and TrivialAugment—enable complex, differentiable data pipelines. Pre-trained models already cover face detection with YuNet, feature matching via LoFTR and LightGlue, descriptor extraction with DISK, and segmentation through SAM.

Version 0.8.2 delivers mostly maintenance: expanded documentation with SEO meta descriptions, dependency bumps, link fixes and pre-commit standardization. The project has also migrated its community chat to Discord.

The timing matters. As robotics and spatial AI systems demand unified perception pipelines, Kornia's differentiable geometry combined with emerging VLMs offers a single framework for both low-level pixel operations and high-level reasoning. Researchers can now backpropagate through the entire stack without switching libraries.

**

Use Cases
  • Robotics engineers running differentiable geometric transforms in PyTorch
  • Researchers training models with fully differentiable augmentation pipelines
  • Developers integrating SAM and LoFTR into end-to-end vision agents
Similar Projects
  • torchvision - PyTorch native utilities lacking Kornia's geometric depth
  • OpenCV - mature computer vision library without differentiability
  • albumentations - fast augmentation toolkit outside PyTorch autograd

Gazebo Jetty Release Refines Robot Simulation Tools 🔗

Version 10.0.0 tightens physics integration, rendering and remote capabilities for ROS2 workflows

gazebosim/gz-sim · C++ · 1.3k stars Est. 2020

The Gazebo project has released Jetty, corresponding to gz-sim 10.0.0.

The Gazebo project has released Jetty, corresponding to gz-sim 10.0.0. The update refines the modular libraries that separate physics, rendering, sensing and transport functions, allowing each to evolve independently while maintaining tight integration.

Dynamics simulation now routes through the Gazebo Physics library with clearer API boundaries for selecting and swapping engines. Rendering employs OGRE v2 via Gazebo Rendering to produce accurate lighting, shadows and texture detail on complex meshes. Sensor output, handled by Gazebo Sensors, includes calibrated models for laser range finders, 2D/3D cameras, IMUs, GPS and force-torque units, with configurable noise profiles.

Plugin interfaces received incremental upgrades for robot, sensor and world control. The Gazebo GUI supports additional plugin entry points for real-time introspection. Command-line tools gained options for headless operation and automated scenario scripting. TCP/IP transport via Gazebo Transport improves asynchronous message passing, making it simpler to run heavy simulation loads on remote servers while keeping lightweight client interfaces local.

Compatibility with ROS 2 remains unchanged, preserving existing control stacks. Pre-built models of the PR2, TurtleBot, Pioneer 2DX and iRobot Create continue to be available through Gazebo Fuel, with SDF used for custom construction. Builds target Ubuntu Noble, with maintained Homebrew and Windows support.

The release reflects steady iteration on 16 years of robotics simulation experience rather than wholesale redesign.

Use Cases
  • Autonomous vehicle developers simulating complex traffic scenarios before real-world testing
  • Academic research labs designing control systems for advanced legged robots
  • Industrial engineers testing multi-robot warehouse automation solutions in simulation
Similar Projects
  • Webots - similar multi-robot simulation with stronger hardware-in-the-loop focus
  • MuJoCo - emphasizes contact-rich physics for reinforcement learning research
  • CoppeliaSim - offers rapid prototyping through embedded scripting languages

Quick Hits

XRobot XRobot's Python toolkit auto-generates production-ready robot code from high-level specs, slashing dev time for complex automation. 221
cddp-cpp High-performance C++ solver delivers constrained differential dynamic programming for precise trajectory optimization and real-time MPC. 89
robotmk Robotmk fuses Robot Framework test automation directly into Checkmk, enabling powerful robotic process monitoring and validation. 58
cloisim CLOiSim instantly builds multi-robot Unity environments from SDF files and bridges them to ROS2 for rapid realistic testing. 172
BotBrain BotBrain gives legged robots a modular open-source brain with web UI for teleop, autonomy, mapping, monitoring and 3D-printable ROS2 hardware. 169

AI Agents Acquire Senior Analyst Skills Through Framework-Mapped Library 🔗

754 structured cybersecurity capabilities linked to MITRE ATT&CK, NIST CSF, ATLAS, D3FEND and AI RMF accelerate security automation

mukul975/Anthropic-Cybersecurity-Skills · Python · 5k stars 1mo old · Latest: v1.2.0

Builders integrating large language models into security operations repeatedly encounter the same limitation: agents lack the procedural knowledge that junior analysts take for granted. An AI assistant can generate code but cannot independently select the correct Volatility3 plugin for memory analysis, identify the Sigma rules that detect Kerberoasting, or properly scope a multi-cloud breach. The mukul975/Anthropic-Cybersecurity-Skills repository directly addresses this gap.

Builders integrating large language models into security operations repeatedly encounter the same limitation: agents lack the procedural knowledge that junior analysts take for granted. An AI assistant can generate code but cannot independently select the correct Volatility3 plugin for memory analysis, identify the Sigma rules that detect Kerberoasting, or properly scope a multi-cloud breach. The mukul975/Anthropic-Cybersecurity-Skills repository directly addresses this gap.

The project contains 754 production-grade skills organized across 26 security domains. Each skill follows the open agentskills.io standard and ships as a self-contained SKILL.md file. The repository is licensed under Apache 2.0 and remains an independent community effort unaffiliated with Anthropic.

Its defining technical contribution is complete mapping of every skill to five established frameworks. Version 1.2.0, released in April 2026, added the final three mappings—MITRE ATLAS v5.5, MITRE D3FEND v1.3 and NIST AI RMF 1.0—joining existing coverage of MITRE ATT&CK Enterprise and NIST CSF 2.0. No other open-source library currently provides unified cross-framework coverage at this granularity.

The new mappings bring specific capabilities. ATLAS v5.5 contributes 81 AI-specific adversarial techniques, including model poisoning, prompt injection defense, AI supply chain attacks and agentic escape-to-host scenarios. D3FEND v1.3 supplies 139 defensive techniques across seven categories: Model, Harden, Detect, Isolate, Deceive, Evict and Restore. NIST AI RMF 1.0 maps 85 skills to the Govern, Map, Measure and Manage functions that govern the AI system lifecycle.

Frontmatter in each skill file now includes explicit YAML fields:

atlas_techniques: [AML.T0051, AML.T0054]
d3fend_techniques: [D3-NTA, D3-PA]
nist_ai_rmf: [MEASURE-2.7, GOVERN-6.1]
nist_csf: [DE.CM-01, RS.AN-03]

A concrete example is the skill analyzing-network-traffic-of-malware, which simultaneously references ATT&CK technique T1071, multiple NIST CSF detection and response categories, ATLAS adversarial methods, D3FEND network traffic analysis countermeasures and AI risk measurement subcategories.

Developers simply clone the repository and instruct their agent—whether Claude Code, GitHub Copilot, Cursor, Gemini CLI or any of 20 compatible platforms—to consult the skill library during investigations. The structured format allows agents to retrieve precise, framework-aligned guidance within seconds.

For security engineering teams building autonomous agents, the library removes weeks of manual knowledge encoding while ensuring every action remains traceable to accepted industry standards. It represents a practical step toward agentic security operations that combine the speed of AI with the rigor of established frameworks.

(Word count: 378)

Use Cases
  • DevSecOps engineers equip AI agents for cloud breach scoping
  • Red team operators train agents on penetration testing tactics
  • Threat hunters integrate skills into autonomous incident response
Similar Projects
  • mitre/cti - Supplies raw ATT&CK STIX data but lacks agent-formatted skills and multi-framework YAML mappings
  • SigmaHQ/sigma - Delivers detection rules for specific threats without structured agent skills or cross-framework coverage
  • AtomicRedTeam - Provides execution tests for ATT&CK techniques but does not map senior-analyst procedures for LLM consumption

More Stories

Juice Shop v19.2.1 Refines Security Training Tools 🔗

Build automation improvements keep OWASP Top 10 challenges current for trainers and pentesters

juice-shop/juice-shop · TypeScript · 13k stars Est. 2014

OWASP Juice Shop has released version 19.2.1 with changes focused on its build pipeline.

OWASP Juice Shop has released version 19.2.1 with changes focused on its build pipeline. The update automatically syncs coding challenge snippets with the project's documentation repository and fixes generation of frontend bundle analysis diagrams.

Written in TypeScript, the application presents a realistic online shop riddled with vulnerabilities drawn from the full OWASP Top 10 and additional real-world flaws. These include SQL injection in product search, broken access control in administrative functions, insecure deserialization, and GraphQL endpoint weaknesses.

The project supports multiple deployment methods. Users can clone the repository, run npm install and npm start, or use packaged 64-bit distributions for Windows, macOS and Linux. Docker images and Vagrant configurations further simplify setup for training environments.

First published in 2014, Juice Shop remains actively maintained precisely because web application attacks continue to evolve. The latest release reduces manual work for maintainers while preserving the application's value as a safe testbed. Security teams deploy it to let participants exploit flaws firsthand before learning corresponding defenses.

Its modular challenge architecture allows both free-form hacking and structured learning paths. This flexibility explains its enduring role in corporate workshops, university courses and capture-the-flag events.

Concrete impact: trainers gain fresh challenge data without extra steps, while tool vendors can reliably benchmark scanners against consistent vulnerable endpoints.

Use Cases
  • Security trainers demonstrating OWASP Top 10 exploits in workshops
  • Penetration testers validating scanning tools against realistic flaws
  • CTF organizers hosting web application hacking competitions
Similar Projects
  • WebGoat - focuses on guided Java lessons rather than open exploitation
  • DVWA - supplies simpler PHP-based challenges with fewer vulnerability types
  • bWAPP - offers basic vulnerable PHP app but lacks modern tech coverage

CISO Assistant Release Refines Risk and Compliance Workflows 🔗

Version 3.15.9 improves EBIOS navigation, data import precision and asset visualization for GRC teams

intuitem/ciso-assistant-community · Python · 4k stars Est. 2023

The latest release of intuitem/ciso-assistant-community tightens core functions in its open-source GRC platform. Version 3.15.

The latest release of intuitem/ciso-assistant-community tightens core functions in its open-source GRC platform. Version 3.15.9 focuses on usability and data integrity rather than broad new capabilities, addressing friction points reported by practitioners managing overlapping regulatory demands.

Key changes target the EBIOS RM module with a direct navigation button from strategic scenario details back to parent studies. Vulnerabilities can now link to follow-up tasks inside the data wizard. Risk assessment import and export routines were revised for reliable round-tripping, while finding exports received sanitization to prevent formatting errors on re-import. The CLI gained a name flag for assessments, easing automation scripts.

Interface updates sort domains within the asset graph and correct multi-level filtering. Task email templates now support links and full variable lists. Hardening fixes align the community edition with enterprise requirements.

These incremental improvements reinforce the project's deliberate architecture. It maintains strict separation between compliance requirements and reusable security controls, enabling the same evidence to satisfy multiple frameworks without duplication. The platform ships with 130 frameworks—including ISO 27001, NIST CSF, SOC 2, PCI DSS, NIS2, DORA, GDPR and HIPAA—plus automatic control mapping, built-in threat libraries, risk quantification, and remediation tracking.

An API-first design supports both interactive use and external automation via CLI, Kafka or direct calls. Custom frameworks remain editable through a lightweight open format. For teams buried in spreadsheets and point tools, the release reduces manual reconciliation work that previously consumed audit cycles.

**

Use Cases
  • GRC teams automating control mapping across ISO 27001 and NIS2
  • Risk officers performing EBIOS RM assessments with remediation tracking
  • Compliance leads importing and exporting risk data via API and CLI
Similar Projects
  • OpenSCAP - focuses on automated scanning but lacks unified risk workflows
  • OSCAL - standardizes compliance formats without built-in assessment engine
  • Wazuh - delivers SIEM and endpoint detection separate from GRC integration

Trivy 0.70 Refines Multi-Target Security Scanning 🔗

Latest release sharpens SBOM output, IaC checks and secret detection accuracy

aquasecurity/trivy · Go · 34.6k stars Est. 2019

Trivy v0.70.0, released this month, delivers targeted improvements to its unified scanning engine rather than broad new feature additions.

Trivy v0.70.0, released this month, delivers targeted improvements to its unified scanning engine rather than broad new feature additions. The Go application consolidates vulnerability, misconfiguration, secret and SBOM detection across container images, Kubernetes clusters, Git repositories, VM images and filesystems.

Notable changes include refined misconfiguration rules for current IaC tools, expanded secret detection patterns for cloud credentials, and more consistent SBOM generation in SPDX and CycloneDX formats. Scan performance on large Kubernetes deployments has been optimised, reducing memory usage during cluster-wide audits. The vulnerability database refresh now pulls from additional upstream sources, tightening coverage for recent CVEs in popular language ecosystems.

Usage remains unchanged at the command level. A typical invocation looks like trivy image python:3.4-alpine or trivy k8s --report summary. The binary, available via Homebrew, direct download or the aquasec/trivy Docker image, requires no daemon and runs equally well in CI pipelines or developer workstations.

For teams already using Trivy, version 0.70.0 removes several deprecated scanner flags and standardises output schemas, forcing minor workflow updates but delivering cleaner integration with downstream reporting tools. The project continues to ship canary builds from main for early testing, though the maintainers explicitly advise against production use of those images.

The net result is a more precise tool for organisations that have moved beyond basic CVE scanning and now require combined SBOM, license and secret visibility across their entire supply chain.

Use Cases
  • DevOps engineers scan Docker images for CVEs and SBOM data
  • Platform teams audit Kubernetes clusters for IaC misconfigurations
  • Compliance officers generate license and secret reports from repos
Similar Projects
  • Grype - narrower focus on container vulnerabilities without native IaC scanning
  • Clair - container image scanner that lacks unified secret and SBOM support
  • Kubescape - Kubernetes-centric with stronger runtime posture but weaker language coverage

Quick Hits

mitmproxy Intercept, inspect, and modify TLS HTTP traffic in real time with mitmproxy's interactive toolkit for power debugging and pentesting. 43.2k
PROXY-List Grab fresh proxy lists that update daily to instantly power your scrapers, testers, and anonymity pipelines. 5.5k
BrowserBox Spin up secure remote browsers anywhere you need them for safe, flexible, and fully controlled web sessions. 3.8k
Azure-Sentinel Detect, investigate, and respond to threats across your entire enterprise with Azure Sentinel's cloud-native intelligent SIEM. 5.6k
nDPI Identify protocols and analyze packets at wire speed with nDPI's open-source deep packet inspection engine. 4.4k

Lazygit Refines GitHub Pull Request Integration in v0.61.1 🔗

Latest release polishes recently added PR features with targeted fixes and maintenance while preserving the tool's core terminal workflow

jesseduffield/lazygit · Go · 76.7k stars Est. 2018 · Latest: v0.61.1

Lazygit has never pretended git is pleasant. Its value has always been in making the painful parts less excruciating. Version 0.

Lazygit has never pretended git is pleasant. Its value has always been in making the painful parts less excruciating. Version 0.61.1, released this week, continues that mission by addressing friction in the GitHub pull requests feature added in recent months.

The update hides closed pull requests when viewing main branches, normalizes repository owner casing to prevent integration failures, and stops defaulting the base repository to "origin". These changes, contributed by first-time contributors bradly0cjw and stefanhaller, remove small but persistent annoyances for developers who move fluidly between GitHub's web interface and local work.

A security fix avoids ${{ }} variable interpolation in workflow steps, and the project has added a justfile to simplify common development tasks. While modest, these maintenance items reflect the project's seven-year pattern of steady, practical iteration rather than headline-grabbing rewrites.

The tool's enduring appeal lies in its terminal UI for everyday git operations that traditionally demand arcane commands or manual file editing. Instead of crafting patch files by hand to stage selected changes, users navigate hunks visually. Interactive rebasing becomes a matter of cursor movement and keypresses rather than editing a TODO file in an external editor. The frustration of being forced to stash changes only to discover no conflicts existed is largely eliminated.

Built in Go, lazygit delivers responsive performance even on large repositories. Its feature set targets precisely where git's power becomes a liability: partial staging, commit amending across history, bisect operations, worktree management, and what the project calls "rebase magic" — custom patches and rebasing from a marked base commit. The commit graph view and two-commit comparison tools provide clarity that raw git log output rarely achieves.

For builders who live primarily in the terminal, these capabilities reduce context switching. When a rebase goes sideways, the undo functionality offers a safety net. When reviewing contributions, the PR integration now connects GitHub metadata more cleanly to local branches.

The 0.61.1 changes matter because they demonstrate the project's focus remains on eliminating real workflow friction rather than adding complexity. In an ecosystem filled with git wrappers and graphical clients, lazygit's insistence on staying lightweight while steadily improving integration with modern platforms like GitHub keeps it relevant for developers who value speed and keyboard-driven efficiency.

As repositories grow more complex and teams ship more frequently, the gap between what git can do and what humans can reliably do without error remains wide. Tools that narrow that gap without introducing new abstractions deserve attention.

Use Cases
  • Terminal developers staging partial file hunks visually
  • Engineers performing interactive rebases without TODO files
  • Contributors syncing GitHub PRs with local branch workflows
Similar Projects
  • gitui - Rust-based terminal git UI focused on performance but with less emphasis on advanced rebase workflows
  • tig - Text-mode git interface that browses history effectively yet lacks lazygit's interactive editing capabilities
  • magit - Emacs-based git porcelain offering deep integration for editor-centric users rather than standalone terminal operation

More Stories

Gitea 1.26 Adds Terraform Registry and Actions Controls 🔗

Latest release brings concurrency support, maintenance mode, and workflow visualization to self-hosted Git environments

go-gitea/gitea · Go · 55k stars Est. 2016

Gitea v1.26.0 delivers concrete upgrades for teams running their own development infrastructure.

Gitea v1.26.0 delivers concrete upgrades for teams running their own development infrastructure. The Go-based platform, which combines Git hosting, code review, package registries, and CI/CD in a single binary, now supports the full Actions concurrency syntax, giving developers finer control over parallel workflow execution and resource usage.

Notable additions include a built-in Terraform state registry, instance-wide informational banners, and a maintenance mode that administrators can activate without disrupting running services. Workflow dependencies can now be visualized directly in the UI, and failed jobs gain a one-click re-run button. The release also introduces automatic changelog generation, keyboard shortcuts for repository search, OpenAPI spec rendering, and non-zipped artifact support for newer runners.

Security fixes bound page sizes in repository listings, while several performance improvements target the Actions backend. Breaking changes modernize API annotations and default PUBLIC_URL_DETECTION to auto.

For organizations wary of vendor lock-in, these updates narrow the gap with commercial platforms while preserving Gitea's lightweight footprint and straightforward deployment. The binary runs on everything from Raspberry Pi to enterprise servers with minimal configuration.

**

Use Cases
  • DevOps teams self-hosting Git with integrated CI/CD pipelines
  • Infrastructure groups managing Terraform states in private registries
  • Engineering orgs running Actions workflows from internal repositories
Similar Projects
  • GitLab CE - broader feature set but significantly higher resource demands
  • Gogs - lighter predecessor offering fewer CI/CD and registry capabilities
  • Forgejo - community fork emphasizing independence with similar Go architecture

RustDesk 1.4.6 Adds ARM64 and Flatpak Support 🔗

Latest release expands architecture coverage and deployment options for self-hosted remote access

rustdesk/rustdesk · Rust · 112.4k stars Est. 2020

RustDesk version 1.4.6 delivers native binaries for AArch64 alongside existing x86_64 support, reflecting the shift toward ARM hardware in servers, laptops and mobile devices.

RustDesk version 1.4.6 delivers native binaries for AArch64 alongside existing x86_64 support, reflecting the shift toward ARM hardware in servers, laptops and mobile devices. The release provides .deb packages for Ubuntu on both architectures, Apple Silicon DMGs for macOS, signed APKs for Android, and a Flatpak bundle for Linux desktop environments. An official iOS build is now listed on the App Store, while a web client enables browser-based sessions without installation.

The Rust codebase continues to emphasize direct P2P connectivity when possible, with optional self-hosted rendezvous and relay servers for environments that prohibit external traffic. Administrators can deploy these components via official Docker images, keeping all session data inside their own infrastructure. Video encoding relies on libvpx, aom and opus, managed through vcpkg during builds.

The desktop GUI has fully migrated to Flutter, retiring the earlier Sciter implementation for better cross-platform consistency and faster iteration. Build instructions now target stable Rust toolchains and document the exact dependency steps required for clean compilation on Ubuntu 18 and newer.

These updates arrive as organizations audit remote-access tools for supply-chain risk and regulatory compliance. By removing mandatory cloud intermediaries, 1.4.6 lowers both cost and exposure for teams that already run their own identity and networking infrastructure. (178 words)

Use Cases
  • Enterprise admins self-hosting remote support on private networks
  • Developers accessing ARM-based test machines without vendor accounts
  • Support engineers connecting to iOS and Android user devices
Similar Projects
  • TeamViewer - proprietary cloud service versus RustDesk's self-hosted control
  • AnyDesk - closed-source commercial tool lacking open auditability
  • FreeRDP - protocol implementation but without complete cross-platform client

Nanobrew Delivers Millisecond Installs for macOS 🔗

Zig-based package manager achieves 3ms warm installs with full Homebrew compatibility on macOS and Linux

justrach/nanobrew · Zig · 979 stars 2mo old

nanobrew is a package manager for macOS and Linux written in Zig. It reports warm installs of cached packages in roughly 3.5 milliseconds and cold installs approximately nine times faster than Homebrew.

nanobrew is a package manager for macOS and Linux written in Zig. It reports warm installs of cached packages in roughly 3.5 milliseconds and cold installs approximately nine times faster than Homebrew. The project reuses Homebrew formulas, bottles and casks while shipping as a single 1.2 MB static binary with no Ruby runtime or bootstrapping step.

Design choices emphasize speed and predictability. nb install performs no automatic update, parallelizing all dependency downloads and extractions by default. Cask installs omit the com.apple.quarantine attribute, eliminating Gatekeeper prompts. Third-party taps work without extra configuration. On Linux and in Docker containers, native .deb support delivers up to 13× faster warm installs than apt-get.

Version 0.1.191, released 20 April 2026, fixed hardlink extraction in bottles such as unzip, perl and postgresql@17 by replacing the previous fallback with a native USTAR/GNU tar parser. Apple Silicon binaries are now Developer ID signed and notarized. Command benchmarks improved markedly: nb leaves on a cold cache of roughly 100 packages dropped from 10.0 s to 0.83 s, now 1.37× faster than Homebrew on the same hardware. Search and resolver times also fell by nearly half.

The tool covers the common fast path but explicitly omits post_install hooks, source builds and Mac App Store integration. nb bundle install returns instantly when a Brewfile is satisfied.

Use Cases
  • macOS developers installing cached dependencies in under 4 ms
  • DevOps engineers building Linux Docker images with native debs
  • Teams running Brewfiles without automatic update delays
Similar Projects
  • Homebrew - shares formulas but adds Ruby overhead and auto-updates
  • apt-get - Linux default that nanobrew beats 13× on warm installs
  • MacPorts - alternative macOS manager lacking Homebrew bottle reuse

Quick Hits

syncthing Syncthing syncs files continuously across devices via peer-to-peer encryption, giving builders private, cloud-free control over their data. 81.9k
zed Zed delivers lightning-fast code editing with real-time multiplayer collaboration, built for developers who want to think and ship at speed. 79.4k
codex-auth Codex-auth instantly switches and manages multiple Codex accounts from the terminal, streamlining auth workflows for power users. 1.1k
react-native React Native builds truly native iOS and Android apps from a single React codebase, letting builders ship fast with native performance. 125.7k
yabai yabai automatically tiles macOS windows using binary space partitioning, delivering keyboard-driven window management that boosts developer focus. 28.7k

ESPectre 2.7 Adds BLE Control to WiFi CSI Motion Sensing 🔗

Latest release enables standalone deployments, runtime configuration, and robust CSI normalization across ESP32 variants and dual software stacks.

francescopace/espectre · Python · 7.1k stars 5mo old · Latest: 2.7.0

Six months after its debut, ESPectre continues to mature. Version 2.7.

Six months after its debut, ESPectre continues to mature. Version 2.7.0 delivers practical upgrades that broaden its appeal beyond the Home Assistant environment while sharpening its core technical performance.

The project detects motion by analyzing Wi-Fi Channel State Information (CSI) rather than relying on cameras or microphones. A standard 2.4 GHz router and an ESP32 board costing roughly €10 become the complete sensor array. Movement disturbs the wireless multipath environment, producing measurable shifts in amplitude and phase that the firmware interprets as human presence. ESP32-S3 and ESP32-C6 variants are recommended, though the original ESP32, C3 and others remain supported after reviewing the platform comparison table.

Previous releases established native integration through ESPHome, allowing 10–15 minute setups using only YAML configuration. Version 2.5 introduced an on-device neural network detector that eliminates manual calibration. Release 2.7.0 now unlocks that capability for users who prefer to operate without Home Assistant entirely.

The headline addition is BLE control. A command channel permits live adjustment of detection thresholds and opens the door for custom clients. The demonstration web game, previously limited to Web Serial, has migrated to Web Bluetooth, providing a ready-made example for developers building their own integrations. These changes reflect a deliberate two-platform strategy: a full-featured ESPHome/C++ component alongside Micro-ESPectre, a Python implementation for lighter deployments.

Substantial engineering effort addressed CSI payload handling. The new code consistently normalizes 256→128 (double HT-LTF), 228→114 and 114→128 variants before HT20 processing. Packet drop rates have fallen on boards that emit shorter or non-standard CSI lengths. Both stacks now share identical normalization logic, verified by new unit tests covering all documented payload scenarios.

Documentation remains a strength. SETUP.md, TUNING.md, a sensor placement guide and a technical deep dive give builders the signal required to deploy reliable sensors. A dedicated security and privacy section confirms that no visual or audio data is ever captured or transmitted.

For developers and privacy-conscious integrators, these updates matter. ESPectre demonstrates that meaningful sensing can be extracted from existing wireless infrastructure at minimal cost and zero visual intrusion. The addition of BLE runtime control and hardened cross-stack compatibility suggests the project is moving from interesting experiment toward dependable infrastructure component.

**

Use Cases
  • Homeowners triggering automations via invisible WiFi motion sensing
  • Developers building custom BLE clients for standalone presence detection
  • Privacy advocates deploying calibration-free ML sensors in existing homes
Similar Projects
  • nexmon-csi - Extracts raw CSI data from Broadcom chips but demands router firmware patches and lacks native ML or HA integration
  • ESP32-PIR - Depends on traditional infrared hardware instead of leveraging existing WiFi signals for through-wall detection
  • mmwave-radars - Offers precise ranging with dedicated radar modules yet requires more expensive components than ESPectre's €10 approach

More Stories

NWinfo 1.6.2 Sharpens GPU and PCIe Diagnostics 🔗

Latest release adds memory frequency reporting, link speeds and audio metrics without WMI

a1ive/nwinfo · C · 523 stars Est. 2021

NWinfo has shipped version 1.6.2, its most substantial update in months.

NWinfo has shipped version 1.6.2, its most substantial update in months. The command-line and GUI utility, written in C, continues to read hardware directly rather than routing queries through WMI, delivering lower latency and fewer permission hurdles on locked-down Windows installs.

The new release focuses on modern bus and graphics details. It now reports GPU memory frequency, PCIe link speeds and current link width for every PCI device. Storage diagnostics gained AAM/APM value display in S.M.A.R.T. output, while the audio subsystem can measure device loudness. GUI users will see HVCI status, expanded sound-card information and an option to export a compact summary report.

Other fixes address NULL dereferences in DMI strings, ARM64 build stability, mainboard miscellaneous data and garbled characters in localized output. PCI vendor and device IDs are now stored as uint16_t, and the bundled pci_ids.h database has been refreshed.

These changes matter as PCIe 5.0 hardware and memory-heavy GPUs become standard in both workstations and gaming rigs. Engineers and system integrators who rely on nwinfo for JSON or YAML inventory scripts gain more precise data without installing heavier commercial tools.

The project remains under the Unlicense and incorporates libcpuid, CrystalDiskInfo routines and Nuklear for its lightweight interface.

Use Cases
  • Engineers auditing PCIe link widths on new motherboards
  • Support technicians exporting JSON hardware inventories
  • Overclockers monitoring GPU memory frequency and SMART data
Similar Projects
  • HWiNFO - commercial Windows tool with broader real-time sensors
  • CPU-Z - GUI-focused CPU and memory viewer lacking YAML export
  • Open Hardware Monitor - sensor-oriented but without deep SMBIOS parsing

WLED Wemos Shield v3.0 Refines Hardware Access 🔗

Rear solder jumpers, updated pinout and added PWM fan circuit improve custom LED controller builds

srg74/WLED-wemos-shield · Unknown · 556 stars Est. 2020

The WLED-wemos-shield project has shipped version 3.0, relocating all solder jumpers to the back of the PCB for easier configuration after assembly. The pinout has been revised and a PWM fan circuit added, addressing thermal demands in enclosed or high-output installations.

The WLED-wemos-shield project has shipped version 3.0, relocating all solder jumpers to the back of the PCB for easier configuration after assembly. The pinout has been revised and a PWM fan circuit added, addressing thermal demands in enclosed or high-output installations.

Designed for the Wemos D1 Mini (ESP8266) or ESP32 D1 Mini, the shield converts these compact boards into full-featured WLED controllers. It supplies a 3.3 V to 5 V level shifter for stable signalling to long runs of WS2812, WS2815, SK6812 or APA102 LEDs, supports both single-wire and clocked data protocols, and includes a power selector for 5 V, 12 V or 24 V strips.

Other onboard elements remain: analog and digital audio inputs for sound-reactive firmware, a relay to eliminate phantom power draw when LEDs are off, I2C header for OLED displays or environmental sensors, I2S on ESP32 variants, optional IR receiver, Dallas temperature probe footprint, and auxiliary 5 V output. The project supplies ready binaries from the main WLED repository plus Atuline and MoonModules sound-reactive forks.

For builders integrating lighting into permanent fixtures or audio installations, these incremental hardware changes reduce assembly friction and broaden environmental tolerance without increasing board size.

PCBs cost $5 for ten boards; fully assembled units are available on Tindie.

Use Cases
  • Makers wiring sound-reactive LED arrays in home theaters
  • DIYers adding fan-cooled controllers to outdoor light sculptures
  • Builders integrating temperature-monitored holiday lighting displays
Similar Projects
  • QuinLED-Dig2 - supplies more robust power handling at larger size
  • Athom-WLED - delivers pre-flashed controllers that skip soldering
  • ESPHome-LED-boards - offers greater firmware flexibility but needs more setup

OSM Chef Repo Adapts to Evolving Infrastructure Needs 🔗

Recent updates to roles and cookbooks support expanded global mapping operations

openstreetmap/chef · Ruby · 112 stars Est. 2014

Following its most recent commits in April 2026, the openstreetmap/chef repository continues to orchestrate configuration for every machine run by the OpenStreetMap Foundation's Operations Working Group. The project employs CINC, the open-source Chef platform, to enforce consistent state across the entire fleet.

All servers carry dragon names and are assembled through layered roles.

Following its most recent commits in April 2026, the openstreetmap/chef repository continues to orchestrate configuration for every machine run by the OpenStreetMap Foundation's Operations Working Group. The project employs CINC, the open-source Chef platform, to enforce consistent state across the entire fleet.

All servers carry dragon names and are assembled through layered roles. Server-specific files such as faffy.rb declare IP addresses, included services and local quirks. Hardware roles like hp-g9.rb capture motherboard and firmware settings reusable across identical machines. Location roles form a hierarchy—equinix-dub.rb sits inside datacentre, organisation and country definitions—while service roles such as web-frontend pull in the exact recipes, configuration values and dependent roles required.

The team follows the organization-repository model: every cookbook lives inside this single repository with no external dependencies. This simplifies testing, accelerates rollouts and guarantees that knife commands operate against a known, self-contained codebase.

The setup matters now because OpenStreetMap's traffic—tile serving, API calls and planet-file distribution—keeps rising. Small changes to a hardware or service role can be tested locally, peer-reviewed and deployed with minimal risk, keeping the global infrastructure reliable without manual intervention on distant machines.

Contributing remains straightforward; the operations team stays reachable on #osmf-operations and via Matrix bridge.

Use Cases
  • OSM operations team applies hierarchical roles to new servers
  • Engineers maintain hardware roles for consistent HP deployments
  • Developers update service cookbooks for API and tile servers
Similar Projects
  • ansible/ansible - uses agentless YAML playbooks instead of Ruby cookbooks
  • puppetlabs/puppet - declarative DSL focused on convergence rather than roles
  • saltstack/salt - event-driven model for large-scale remote execution

Quick Hits

Classic-Repair-Toolbox Cross-platform C# toolbox diagnoses, troubleshoots and repairs vintage Commodore and Amstrad hardware on Windows, Linux and macOS. 172
IceNav-v3 ESP32 GPS navigator delivers offline OSM maps and multi-GNSS reception for building reliable connectivity-free navigation devices. 347
Circuitry-Based-Sound Teaching repository explores DIY circuitry, experimental sound generation and live electronic performance techniques from HfG Karlsruhe workshops. 52
awesome-fabrication Curated collection of fabrication software, tools and resources to accelerate your digital manufacturing and maker projects. 35
node-feature-discovery Automatically detects and labels node hardware features in Kubernetes so workloads can schedule onto exactly the right machines. 1k

Magpie 0.12.1 Refines Window Upscaling With Auto-Hide Cursor 🔗

Latest release stabilizes auto-scaling, enforces topmost layering and fixes cropping and cursor bugs for sharper Windows 10 and 11 workflows.

Blinue/Magpie · HLSL · 13.6k stars Est. 2021 · Latest: v0.12.1

Magpie has never been a simple magnifier. The open-source utility captures any window and re-renders it at higher resolution using a palette of high-quality algorithms. Version 0.

Magpie has never been a simple magnifier. The open-source utility captures any window and re-renders it at higher resolution using a palette of high-quality algorithms. Version 0.12.1, released this week, tightens that capability with pragmatic improvements that matter to anyone shipping or maintaining Windows software.

The headline addition is an idle cursor auto-hide function. Developers can set a custom delay before the cursor disappears, eliminating visual distraction during video, gameplay or presentation sessions. The change directly addresses years of user requests for cleaner scaled output.

Equally significant is the revised auto-scaling logic. Pop-up dialogs and notification windows no longer interrupt the scaling pipeline. Combined with fixes for unexpected scaling termination, the update produces more predictable behavior across applications that spawn transient UI elements.

Several layering and rendering bugs have been closed. The scaled window is now unconditionally topmost; the former “Keep scaled window on top” toggle has been removed because it proved unreliable. Title-bar cropping failures on certain legacy applications have been corrected, monochrome cursor handling no longer freezes the output, and a toolbar menu ID conflict that affected screenshots has been resolved. These changes, while incremental, remove friction that previously forced users to choose between native resolution and visual quality.

Under the hood Magpie remains a lean DirectX 11 application written in HLSL. It offers both fullscreen and bordered window modes, multi-monitor awareness, and a WinUI interface that respects system light and dark themes. The built-in effect library includes Anime4K for line preservation in 2D animation, AMD FSR for temporal upscaling, assorted CRT shaders, and additional filters contributed by the community. Because the entire stack is GPU-driven, overhead stays low even at 4K output.

For builders the project is instructive. It demonstrates practical use of cppwinrt, XAML Islands and Fluent Design while exposing a shader pipeline that can be extended without touching core capture logic. The GPLv3 license and active contribution history lower the barrier for adding new algorithms or platform integrations.

As desktop resolutions climb and older codebases persist, Magpie’s approach—capture, upscale, composite—offers a system-level solution that neither requires per-application rewrites nor forces users into fullscreen exclusive mode. The 0.12.1 release makes that solution noticeably more robust.

System requirements remain modest: Windows 10 v1903 or Windows 11 with DirectX feature level 11. The binary footprint stays under 10 MB, reflecting the project’s continued focus on efficiency over feature bloat.

Use Cases
  • PC gamers upscaling legacy titles with FSR and CRT shaders
  • Developers debugging high-DPI rendering in older Windows apps
  • 2D artists applying Anime4K filters to any desktop window
Similar Projects
  • Lossless Scaling - Commercial Steam tool with similar real-time upscaling but paid algorithms and Steam Deck focus
  • ReShade - Injects post-process shaders per application instead of system-wide window capture
  • Anime4KCPP - Standalone Anime4K implementation lacking Magpie’s full window management and multi-monitor support

More Stories

Super Mario Remastered Adds EU ROM and Checkpoint Tools 🔗

Version 1.0.2 expands compatibility, custom level limits and visual fidelity for Godot builders

JHDev2006/Super-Mario-Bros.-Remastered-Public · GDScript · 2.6k stars 7mo old

The Super Mario Bros. Remastered project has shipped version 1.0.

The Super Mario Bros. Remastered project has shipped version 1.0.2, bringing targeted improvements for both gameplay and content creation. The update adds official support for the European SMB1 ROM, alongside the ability to regenerate assets and re-verify ROM integrity when graphics become corrupted.

Custom level authors benefit most. The restriction on checkpoints has been removed, allowing unlimited placement across all subareas. Boo colour unlocks now scale with completion time rather than repeated runs, culminating in Golden Boo. New optional character animations are documented in the project wiki, and creators can set a precise frame-rate limit in the options menu.

Portable operation is now activated by simply dropping a portable.txt file next to the executable. Resource packs gain .ogg music support, Firebars can toggle their original “snappy” movement, and mushroom ejection direction now respects which half of the block is struck. Level Share Square browsing displays difficulty with skulls and ratings with stars; completed levels restore the previous view state.

Built in GDScript for Godot 4.6, the project continues to recreate the original NES titles and Lost Levels with refined physics while requiring a legitimate ROM. The changes reflect steady iteration by its maintainer and contributors rather than wholesale redesign.

Use Cases
  • Godot developers prototyping precise 2D platformer physics
  • Retro fans building and sharing custom Mario-style levels
  • Modders creating and distributing visual resource packs
Similar Projects
  • SMBX2 - Lua-based Mario fan game with deeper scripting but closed editor
  • sm64-port - Focuses on Mario 64 decompilation and modding ecosystem
  • SuperTux - Open-source platformer with built-in level editor but original IP

Bliss Shader Refines Variable Minecraft Lighting 🔗

Release11 integrates voxel floodfill coloring and enhanced depth of field from community contributors

X0nk/Bliss-Shader · GLSL · 944 stars Est. 2023

Bliss Shader's release11 refines its core promise of inconsistent, context-aware illumination in Minecraft. The GLSL edit of Chocapic v9 avoids static lighting models, generating scenes whose mood and color temperature shift depending on biome, time, and player position.

The most significant recent change is Null's voxel floodfill colored lighting system.

Bliss Shader's release11 refines its core promise of inconsistent, context-aware illumination in Minecraft. The GLSL edit of Chocapic v9 avoids static lighting models, generating scenes whose mood and color temperature shift depending on biome, time, and player position.

The most significant recent change is Null's voxel floodfill colored lighting system. It propagates light in three dimensions with proper color bleed, sharply reducing leaks. Emin and Gri573 contributed practical fixes that further tighten light containment. WoMspace's depth-of-field overhaul delivers smoother focal transitions useful for both gameplay and recording.

Customization remains extensive. Dozens of sliders control shadow softness, water caustics, atmospheric scattering, and cloud behavior. Three update channels exist: release versions appear on Modrinth and Curseforge once stable; the stable branch receives regular tested builds; the unstable branch tracks daily commits for users willing to file issues.

Installation is direct. Download the repository zip from the main branch and drop the archive into the shader packs folder—no extraction required.

These incremental improvements matter as Minecraft 1.21 players seek visual variety without external renderers. The project shows how sustained community editing can evolve an established shader into a distinctly moody visual language.

Use Cases
  • Experienced players adjusting dynamic lighting across biomes and structures
  • Content creators capturing cinematic scenes with refined depth of field
  • Builders testing voxel-based colored lighting in large survival bases
Similar Projects
  • Chocapic13 - Original base shader that Bliss extensively modifies and extends
  • BSL Shaders - Performance-focused alternative emphasizing smoothness over variability
  • Complementary Reimagined - Vibrant color palette contrasting Bliss's moody scene shifts

LibGDX 1.14.0 Refines Cross-Platform Java Tools 🔗

Version 1.14.0 delivers Tiled class support, Java 21 compatibility and Android stability fixes

libgdx/libgdx · Java · 25k stars Est. 2012

LibGDX 1.14.0 focuses on incremental modernization of the veteran Java game framework.

LibGDX 1.14.0 focuses on incremental modernization of the veteran Java game framework. The release improves compatibility with current toolchains while addressing long-standing pain points for developers shipping to desktop, Android, iOS, HTML5, Windows, Linux and macOS.

Notable changes include native class support for Tiled map files, allowing cleaner integration of tile sets and object layers. FreeType has been updated to 2.13.3 for more consistent font rendering. Android users gain concrete fixes: crash prevention when calculating soft-button bar height, Pools API adjustments to eliminate desugar conflicts, and an extracted createGraphics method that simplifies custom AndroidGraphics subclasses.

Build compatibility now extends cleanly to Java 21 after the Spotless Gradle plugin update. Convenience additions comprise static Vector.One fields and a new JsonValue#toJson overload accepting a Writer. The release also replaces deprecated Android audio and cursor APIs, corrects misspellings in AsyncExecutor, and supplies a dark-mode logo variant.

These updates, drawn from more than a dozen community pull requests, demonstrate the project's continued maintenance without altering its core contract: full OpenGL ES access and no enforced architecture or coding style. The Apache 2.0 framework therefore remains a stable base for teams iterating on 2D and 3D titles while leveraging its mature third-party ecosystem.

Use Cases
  • Java engineers shipping 2D titles to Android and iOS
  • Studios integrating Tiled maps across desktop and web
  • Teams building 3D prototypes with custom OpenGL backends
Similar Projects
  • jMonkeyEngine - Java-first 3D engine with scene graph instead of raw GL access
  • Godot - offers visual editor and GDScript but requires different language binding
  • LWJGL - supplies lower-level Java OpenGL bindings that power libGDX desktop layer

Quick Hits

love LÖVE's Lua framework lets builders craft 2D games with effortless graphics, physics, and audio tools. 8.2k
tabletop-club Tabletop Club delivers physics-based 3D virtual tabletops where builders can create and play games across platforms with Godot. 1.4k
gozen Gozen gives builders a minimalistic Godot-powered video editor focused on simplicity and speed without unnecessary complexity. 398
Revelation Revelation transforms Minecraft Java worlds with explorative GLSL shaders that deliver stunning atmospheric and artistic visuals. 511
bgfx bgfx equips builders with a cross-platform rendering library that abstracts graphics APIs so you bring your own engine. 17k